SecretLLM
What is SecretLLM
SecretLLM is a set of OpenAI-compatible APIs that run AI models within a Trusted Execution Environment (TEE). You can use SecretLLM to build new private AI applications or migrate existing ones to run in a secure SecretLLM environment where your data remains private.
How SecretLLM Works
- Send your prompt to SecretLLM over HTTPS (just like any secure web request)
- Your prompt runs inside a Trusted Execution Environment where it is hidden from the cloud provider
- Get your result back over HTTPS, complete with a cryptographic signature
Available Models
Model | Parameters | Features | Best For | Access |
---|---|---|---|---|
meta-llama/Llama-3.2-3B-Instruct | 3B | Chat completion, tool support | Quick responses, simple Q&A, basic chat. Faster inference with lower resource usage. | Testnet |
meta-llama/Llama-3.1-8B-Instruct | 8B | Chat completion, tool support | More complex tasks, detailed responses, code generation. Higher quality but slower inference. | Testnet |
deepseek-ai/DeepSeek-R1-Distill-Qwen-14B | 14B | Chat completion | Reasoning and complex tasks that require a long output. Also, slowest of the three | Apply for access |
For most use cases, start with the 3B model and move up if you need more capability.
Key Features
- Drop-in Privacy: Keep your OpenAI-style code, just point it to SecretLLM
- Privacy Through TEEs: All processing happens within a TEE, built on NVIDIA Confidential Computing
- Standard Security: HTTPS encryption provides end-to-end protection for all data in transit between your application and the TEE environment
- Cryptographic Attestation: Verify the integrity of your TEE environment through our attestation API
- Signed Responses: Every model response includes a cryptographic signature for verification
Attestation & Verification
SecretLLM provides two layers of cryptographic proof:
-
Environment Attestation: Verify your TEE's integrity through the attestation endpoint, including:
- CPU environment verification
- GPU environment verification
- Service verification through a verifying key
-
Response Signatures: Every chat completion response includes a cryptographic signature, allowing you to verify that the response came from the attested environment.
Getting Started
- Get access to SecretLLM by obtaining your NilAI API key and node URL
- Use your API key to check the
/models
endpoint for available models - Start running private AI using your chosen model
- Optionally, verify your environment using the attestation API