Configure LLM provider endpoints and models for custom deployments
The AI Gateway ships with built-in configurations for all major LLM providers, including their endpoints, API versions, and supported models. Most users never need to configure providers manually.
Provider configuration becomes necessary when you need custom endpoints, enterprise deployments, or want to add models not yet included in the defaults.
Provider configuration defines how the AI Gateway connects to and communicates with LLM providers. Each provider has specific requirements for:
The gateway includes comprehensive defaults that work out-of-the-box for standard deployments.
You’ll need to configure providers when you have:
Provider configuration only defines endpoints and models. API key management is handled separately in Secret Management.
Providers are enabled by simply setting their API keys as environment variables. See Secret Management for details on API key configuration.
Use case: Limit available models for cost control or compliance.
Use case: Limit available models for cost control or compliance.
Use case: Self-hosted Ollama instance on custom port or remote server.
Use case: AWS Bedrock with specific regional endpoint.
For complete configuration options and syntax, see the Configuration Reference.
The AI Gateway supports the following LLM providers with built-in configurations:
Provider | Default Endpoint | Custom Endpoint Support |
---|---|---|
OpenAI | https://api.openai.com | ✅ (regional variants) |
Anthropic | https://api.anthropic.com | ✅ (regional variants) |
Gemini | https://generativelanguage.googleapis.com | ✅ (Gemini) |
AWS Bedrock | Regional AWS endpoints | ✅ (cross-region) |
VertexAI | Regional GCP endpoints | ✅ (cross-region) |
Ollama | http://localhost:11434 | ✅ (any host/port) |
For current model support by provider, see the embedded provider configuration which is automatically updated as new models are released.
Model categories supported: