Helicone OSS LLM Observability home page
Search...
⌘K
Ask AI
Discord
Github
Dashboard
Dashboard
Search...
Navigation
Cloud AI Gateway
Cloud AI Gateway
Documentation
API Reference
Guides
FAQ
Getting Started
Introduction
Self Host
Integrations
LLM & Inference Providers
Frameworks & Libraries
Tools Logging
Any Vector DB
Tracing
Caching
Custom Properties
Sessions
User Metrics
Omit Logs
Prompts & Evals
Prompt Management
Eval Scores
User Feedback
Webhooks
Cloud AI Gateway
Overview
Custom Rate Limits
Gateway Fallbacks
LLM Security
Moderations
Retries
Key Vault
References
Prompts (Legacy)
Async Logging
Cost Calculation
Data Security & Privacy
Generic Gateway
Prompts (Legacy)
Header Directory
Latency
Open Source
Proxy vs Async
Reliability
On this page
Cloud AI Gateway Features
Core Gateway Features
Cloud AI Gateway
Cloud AI Gateway
Documentation for Helicone’s cloud-based AI Gateway
Cloud AI Gateway Features
Here’s the complete documentation for all available features:
Core Gateway Features
Custom Rate Limits
Configure request rate limiting and spending controls for your applications
Gateway Fallbacks
Set up automatic fallback providers when your primary LLM provider fails
LLM Security
Implement security measures and content filtering for your LLM requests
Content Moderation
Automatically detect and filter inappropriate content in requests and responses
Automatic Retries
Configure retry logic for failed requests with exponential backoff
Was this page helpful?
Yes
No
Suggest edits
Raise issue
Webhooks Local Testing
Custom Rate Limits
Assistant
Responses are generated using AI and may contain mistakes.