Compress every LLM call automatically
Drop-in proxy for OpenAI & Anthropic. Just change your base URL — save 30-60% on tokens with persistent context compression.
🔌 Drop-in Replacement
Change base_url to api.nexil.co/v1 — your existing OpenAI or Anthropic code works unchanged. Zero refactoring.
🗜️ Persistent Compression
Shared Context Table (SCT) persists across requests. The more you use it, the more you save — progressive compression that learns your patterns.
📊 Analytics Dashboard
Track tokens saved, compression ratios, and usage per API key. See exactly how much you're saving in real time.
🔑 Secure by Default
Your upstream API key is never stored — passed per-request via header. NEXIL keys are SHA-256 hashed. Enterprise-grade security.
⚡ Fast Proxy
Minimal latency overhead. Compression happens in microseconds. Responses stream back to your app in the standard OpenAI format.
💰 Simple Pricing
Three tiers. No hidden fees. Pay for what you need — from startups to enterprise.
Pricing
Save more than you spend. Every plan pays for itself. Start with a 7-day free trial.
Starter
7-day free trial
- 500K tokens/day
- 1 API key
- OpenAI + Anthropic
- Dashboard + analytics
Growth
7-day free trial
- 5M tokens/day
- 5 API keys
- OpenAI + Anthropic
- Priority support
Enterprise
7-day free trial
- Unlimited tokens
- Unlimited API keys
- OpenAI + Anthropic
- Dedicated support