Helicone OSS LLM Observability home pagelight logodark logo
  • Discord
  • Github
  • Dashboard
  • Dashboard
  • Documentation
  • API Reference
  • Guides
  • FAQ
  • Getting Started
    • Introduction
    • Self Host
    Integrations
    • LLM & Inference Providers
    • Frameworks & Libraries
    • Tools Logging
    • Any Vector DB
    Tracing
    • Caching
    • Custom Properties
    • Sessions
    • User Metrics
    • Omit Logs
    Prompts & Evals
    • Prompt Management
    • Eval Scores
    • User Feedback
    • Webhooks
    Cloud AI Gateway
    • Overview
    • Custom Rate Limits
    • Gateway Fallbacks
    • LLM Security
    • Moderations
    • Retries
    • Key Vault
    References
    • Prompts (Legacy)
    • Async Logging
    • Cost Calculation
    • Data Security & Privacy
    • Generic Gateway
    • Prompts (Legacy)
    • Header Directory
    • Latency
    • Open Source
    • Proxy vs Async
    • Reliability
    On this page
    • Cloud AI Gateway Features
    • Core Gateway Features
    Cloud AI Gateway

    Cloud AI Gateway

    Documentation for Helicone’s cloud-based AI Gateway

    ​
    Cloud AI Gateway Features

    Here’s the complete documentation for all available features:

    ​
    Core Gateway Features

    Custom Rate Limits

    Configure request rate limiting and spending controls for your applications

    Gateway Fallbacks

    Set up automatic fallback providers when your primary LLM provider fails

    LLM Security

    Implement security measures and content filtering for your LLM requests

    Content Moderation

    Automatically detect and filter inappropriate content in requests and responses

    Automatic Retries

    Configure retry logic for failed requests with exponential backoff

    Was this page helpful?

    Suggest editsRaise issue
    Webhooks Local TestingCustom Rate Limits
    twitterlinkedingithubdiscord
    Powered by Mintlify
    Assistant
    Responses are generated using AI and may contain mistakes.