Skip to content

Conversation

@MadelineAu
Copy link
Collaborator

No description provided.

@vercel
Copy link

vercel bot commented Dec 1, 2025

The latest updates on your projects. Learn more about Vercel for GitHub.

Project Deployment Preview Updated (UTC)
eigencloud-docs Ready Ready Preview Dec 4, 2025 11:59pm


:::note
- OpenAI-compatible refers to the [messages-based Chat Completions API](https://platform.openai.com/docs/api-reference/chat/create).
- Deterministic refers to providing one request (prompt, seed/parameters) to the EigenAI API multiple times will produce the same output bit-by-bit.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
- Deterministic refers to providing one request (prompt, seed/parameters) to the EigenAI API multiple times will produce the same output bit-by-bit.
- Deterministic behavior refers to providing one request (prompt, seed/parameters) to the EigenAI API multiple times and receiving the same output bit-by-bit every time. The EigenAI inference stack was designed with this determinism guarantee in mind. **This behavior is not based on caching the results.** Safety-critical systems cannot be vulnerable to potential cache misses.
- A directionally similar work was published by [Thinking Machines, titled Defeating Nondeterminism in LLMs](https://thinkingmachines.ai/blog/defeating-nondeterminism-in-llm-inference/). EigenAI achieves the same outcome as this research, making EigenAI the world's first production-level deterministic LLM API.

Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Updated

Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I broke the piece about not caching into a separate bullet point and reworded slightly - you're saying that 'determinism requires not caching the results' right? Or something different to that?

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Not that it requires not caching, it's that the deterministic behavior is not a result of caching, since typically caching gives you deterministic behavior.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants