Install
Drop in the OpenAI SDK. No proprietary client, no vendor lock. One dependency you already know.
Zero logs. Zero limits.
An encrypted AI gateway for every model.
Integration
Inference leading LLMs run inside verifiably secure runtimes, powered by Intel TDX and NVIDIA Confidential Computing architectures.
1import OpenAI from "openai"
2
3const zima = new OpenAI({
4 baseURL: "https://www.zima.chat/api/v1",
5 apiKey: process.env.ZIMA_KEY,
6})
7
8const response = await zima.chat.completions.create({
9 model: "qwen3-next-80B-a3B-instruct",
10 messages: [{ role: "user", content: "..." }],
11})
12
13// encrypted in hardware. zero data retained.Quickstart
Use your existing OpenAI client. Swap the base URL and API key to start encrypting every inference request today.
Drop in the OpenAI SDK. No proprietary client, no vendor lock. One dependency you already know.
Swap the base URL. Every request now routes through our secure enclave — encrypted in hardware, zero data retained.
Same models, same latency, hardware-level encryption. Your compliance team will thank you.
From SDK request to protected output, without storing prompts or responses.
Architecture
Infrastructure
From key exchange to model output, every step runs inside hardware-attested enclaves.
Intel TDX and NVIDIA confidential compute enclaves protect every request made through Zima.
No prompts, outputs, or metadata survive past completion. GPU memory gets wiped after every request.
Point your existing OpenAI client at Zima's base URL. Fully encrypted inference with a single line swap.
OpenAI, Anthropic, Mistral and more. Access the models you already use, through one secure endpoint.
| Model | Provider | Input / 1M | Output / 1M |
|---|---|---|---|
| GPT-4o | OpenAI | $2.50 | $10.00 |
| Claude Sonnet 4 | Anthropic | $3.00 | $15.00 |
| Gemini 2.5 Pro | $1.25 | $5.00 | |
| Llama 3.1 70B | Meta | $0.50 | $0.75 |
| Mistral Large | Mistral | $2.00 | $6.00 |
Models
Provider pricing with zero intermediary markup. Hardware-grade encryption at no extra cost.
Every request executes inside a hardware-isolated enclave. No prompts, outputs, or metadata survive past completion.
No prompts, outputs, or metadata survive past completion
Data sealed in silicon. Encrypted at rest, in transit, and in use.
Cryptographic proof before any key is released.
“We evaluated six AI gateways. Zima was the only one where our security team signed off without a single objection.”
“Switched from direct OpenAI calls in an afternoon. Same SDK, same format — just encrypted now.”
Security, compatibility, and pricing.
No. Requests exist only for the duration of execution inside the enclave. No prompts, outputs, metadata, or logs are persisted. This is enforced by hardware, not policy.
Point the OpenAI SDK at Zima, keep model choice and pricing visibility, and move sensitive traffic into attested hardware.