Helicone’s AI Gateway is the open-source, lightweight Rust-based solution for intelligent routing to 100+ language models.
Still using our legacy cloud AI Gateway? Access the documentation here. This version is being phased out in favor of a new cloud offering based on this improved self-hosted solution (coming soon).
The AI development space is progressing at an exponential rate. We see it every day at Helicone—over 90 % of our users run 5 + LLMs in production, each with its own SDK, auth scheme, rate limits, and quirks.
Keeping up today means rewriting integrations for every new model, managing a maze of API keys, engineering custom fallbacks for provider outages, and constantly tuning traffic for cost or compliance.
Helicone AI Gateway is our answer. It is a lightweight Rust router inspired by NGINX that removes the integration tax so you can focus on shipping features.
A unified interface for every LLM provider using familiar OpenAI syntax
Load balance to always hit the fastest, cheapest, or most reliable option
Rate limit to prevent runaway costs and usage abuse
Cache responses to reduce costs and latency by up to 95%
Fallback to a different provider or model if the primary one is down
Store all API keys securely to end credential chaos and security risks
Monitor your AI workflows with built-in Helicone integration
Deploy to Fly.io, AWS, Porter, or any other platform in minutes
Built in Rust, the Gateway ships as one lightweight binary you can run anywhere:
Get started in 1 minute with our quickstart guide