The composable AI gateway. One env var. Infinite context.
Plug in intelligence, not just guardrails.
Get early accessPowered by ekkOS
How it works
Replace one env var. Zero code changes. Every app, every framework, every model — it just works.
Every conversation forges patterns. Outcomes are tracked. Failures retire. Good solutions reinforce.
Patterns inject before each request. Context never resets. Your AI bill goes down over time.
Composable modules
Smart tool selection. 67k of MCP tokens become 6k. Your $200 plan stops dying after 19 minutes.
Progressive compression by age. Sessions never reset. Your AI sees everything that happened.
Forges fixes from successful conversations. Injects relevant solutions before each request.
Cosine-similar prompts return cached responses. Free latency. Free spend.
Per-key spend ceilings. 429 before you blow past. Daily, weekly, monthly.
Byte-for-byte matches return instantly. TTL configurable per pipe.
vs the field
| prxy.monster | OpenRouter | Portkey | Helicone | LiteLLM | |
|---|---|---|---|---|---|
| Multi-provider routing | YES | YES | YES | YES | YES |
| MCP token optimization | YES | NO | NO | NO | NO |
| Infinite context (compressed) | YES | NO | NO | NO | NO |
| Pattern learning across sessions | YES | NO | NO | NO | NO |
| Semantic cache | YES | NO | YES | NO | NO |
| Self-host (MIT) | YES | NO | NO | YES | YES |
| Composable modules | YES | NO | NO | NO | NO |
| AWS Bedrock supported | YES | YES | YES | NO | YES |
Plays nice
Same wire format as Anthropic and OpenAI. Most integrations are a single env var. Zero code change for the SDKs you already have wired up.
AI Coding Tools
SDKs
Frameworks
Deploy
Hosted gateway. Zero ops. Pay-as-you-go credits.
Single binary. SQLite + filesystem. MIT licensed.
One-command deploy to your AWS account. Bedrock-native.
npm run deploy --context domain=prxy.example.com