Transactional

Supported Providers

Complete list of supported AI providers and their models.

Overview

AI Gateway supports multiple LLM providers through a unified OpenAI-compatible API. Add your provider API keys in the dashboard to enable access to their models.

Provider Matrix

ProviderStatusModelsFeatures
OpenAIFully SupportedGPT-4o, GPT-4-turbo, o1, o1-miniStreaming, Functions, Vision
AnthropicFully SupportedClaude Opus 4, Claude Sonnet 4, Claude 3.5, Claude 3Streaming, Tool Use
Google AIComing SoonGemini 2.0, Gemini 1.5 Pro, Gemini 1.5 Flash-
AWS BedrockComing SoonClaude, Llama 3, Titan-
Azure OpenAIComing SoonGPT-4, GPT-3.5-

Adding Provider Keys

  1. Navigate to AI Gateway Settings
  2. Under "Provider Keys", click Add Key
  3. Select your provider from the dropdown
  4. Paste your API key
  5. Click Save

Key Security

  • Provider keys are encrypted at rest
  • Keys are never exposed in logs or responses
  • You can rotate keys at any time
  • Deleting a key immediately revokes access

Model Naming

Use the provider's model names directly:

// OpenAI
model: 'gpt-4o'
model: 'gpt-4-turbo'
model: 'gpt-3.5-turbo'
model: 'o1'
model: 'o1-mini'
 
// Anthropic
model: 'claude-3-5-sonnet'
model: 'claude-3-opus'
model: 'claude-3-haiku'
model: 'claude-sonnet-4'
model: 'claude-opus-4'

Provider Priority

When you have multiple providers configured, you can set priority for fallback:

  1. Go to Settings > Provider Keys
  2. Drag providers to reorder priority
  3. The first provider is used by default
  4. If it fails, the next provider is tried

Provider-Specific Limits

ProviderDefault Rate LimitMax TokensContext Window
OpenAI3,500 RPM4,096-128K128K (GPT-4o)
Anthropic1,000 RPM4,096-200K200K (Claude 3.5)

These limits are from the provider. AI Gateway respects and passes through provider rate limits.

Using Multiple Providers

Fallback Configuration

Set up automatic failover between providers:

// Primary request goes to OpenAI
// If it fails, AI Gateway tries Anthropic
const response = await openai.chat.completions.create({
  model: 'gpt-4o', // Primary
  messages: [...],
});

Configure fallback mapping in Settings:

Primary ModelFallback Model
gpt-4oclaude-3-5-sonnet
gpt-3.5-turboclaude-3-haiku

Load Balancing

Coming soon: Distribute requests across providers based on cost, latency, or availability.

Checking Provider Status

View provider health in the dashboard:

  1. Go to AI Gateway > Analytics
  2. See real-time status for each provider
  3. View error rates and latency by provider

Next Steps