EigenAI Overview
Go to deTERMinal and get a 1M inference token credit for free by connecting your X account.
To integrate deTERMinal token credit with your client, refer to the Grant API Integration README.
1M token creditFor more tokens or different models, you can contact us here.
What is EigenAI?
EigenAI is a verifiable LLM inference service that provides deterministic execution of open source LLMs. Unlike traditional AI services where you trust the provider's outputs, EigenAI enables cryptographic verification that inference results are executed using the specified model and input.
EigenAI enables reproducible, and auditable AI inference. Developers access these guarantees through an OpenAI-compatible API and support for frontier open-source models.
- OpenAI-compatible refers to the messages-based Chat Completions API.
- Deterministic refers to providing one request (prompt, parameters, etc) to the EigenAI API multiple times will produce the same output bit-by-bit.
Why build with EigenAI?
Build verifiable applications leveraging LLM inference without wondering if the same LLM request might produce different results on different runs, or whether your prompts, models, or responses are modified in any way. EigenAI offers:
Verifiable AI increases trust in the AI quality of service provided to applications, and increased user trust in agentic workflows. Verifiable AI is made possible by determinism, and EigenAI provides deterministic execution.
How EigenAI works?
EigenAI delivers verifiable LLM inference by making GPU execution a deterministic pipeline.
Deterministic GPU inference
EigenAI constrains GPU execution so that the same inputs always produce the same outputs. EigenAI removes typical nondeterministic behavior found in AI systems, such as batching, kernel race conditions, and opportunistic memory reuse.
Isolated per-request execution
Each query runs in its own clean environment. The KV cache is reset, the full context is loaded, and tokens are generated sequentially with no batching or shared GPU state. This ensures that no other workload can influence the execution path or final output.
Seed-controlled sampling
Randomness is governed through strict seed management. Users can provide a seed or rely on deterministic defaults. This makes every result reproducible and enables users, or third parties, to re-run the exact same request to confirm correctness.
If you want non-determinism for your application, introduce non-determinism by setting a different seed for requests but otherwise keep the request the same. The API will produce a different output.
Model and hardware integrity
EigenAI provides a consistent, verifiable execution stack. Model weights, quantization levels, and GPU types are fixed. Only H100 GPUs are used, with ECC memory enabled, providing stable, integrity-preserving computation.
Verifiability Roadmap
EigenAI’s deterministic execution makes verification possible. As we move through mainnet alpha into general availability, the verification pathways expand.
Self-verification (Mainnet Alpha)
EigenAI will open source its inference stack. Anyone with access to suitable GPUs can re-run a request locally using the same model, inputs, and seed, and confirm that the output matches bit-for-bit.
Third-party verification (GA Target)
A separate verification API will allow independent operators to re-execute requests and return attestations. Applications can use this to spot-check results or provide external proof that an inference was executed correctly.
As the EigenAI roadmap is delivered, the level of required trust decreases while the strength of guarantees increases.