Skip to content

Endpoints Reference

LLMProxy supports 15 LLM providers with automatic request/response translation.

Supported Providers

ProviderBase URLAuthModels
OpenAIapi.openai.com/v1Bearergpt-4o, gpt-4o-mini, gpt-4.1, o3-mini, embeddings
Anthropicapi.anthropic.com/v1x-api-keyclaude-sonnet-4, claude-haiku-4.5, claude-opus-4
Googlegenerativelanguage.googleapis.comAPI keygemini-2.5-pro, gemini-2.5-flash, embeddings
Azure{resource}.openai.azure.comapi-keygpt-4o, gpt-4o-mini
Ollamalocalhost:11434Nonellama3.3, qwen3, phi-4, gemma3, embeddings
Groqapi.groq.com/openai/v1Bearerllama-3.3-70b, mixtral-8x7b
Togetherapi.together.xyz/v1BearerLlama-3.3-70B, Mixtral-8x7B
Mistralapi.mistral.ai/v1Bearermistral-large, mistral-small, codestral
DeepSeekapi.deepseek.com/v1Bearerdeepseek-chat, deepseek-reasoner
xAIapi.x.ai/v1Bearergrok-3, grok-3-mini
Perplexityapi.perplexity.aiBearersonar-pro, sonar
OpenRouteropenrouter.ai/api/v1BearerAll models via unified API
Fireworksapi.fireworks.ai/inference/v1Bearerllama-v3p3-70b-instruct
SambaNovaapi.sambanova.ai/v1BearerMeta-Llama-3.3-70B-Instruct
OpenAI-CompatibleCustomBearerAny OpenAI-compatible API

Configuration Examples

OpenAI

yaml
endpoints:
  openai:
    provider: "openai"
    base_url: "https://api.openai.com/v1"
    api_key_env: "OPENAI_API_KEY"
    models: ["gpt-4o", "gpt-4o-mini", "text-embedding-3-small"]
    rate_limit: { rpm: 3500, tpm: 60000 }

Anthropic

yaml
  anthropic:
    provider: "anthropic"
    base_url: "https://api.anthropic.com/v1"
    api_key_env: "ANTHROPIC_API_KEY"
    models: ["claude-sonnet-4-20250514", "claude-haiku-4-5-20251001"]
    rate_limit: { rpm: 1000 }

Google

yaml
  google:
    provider: "google"
    base_url: "https://generativelanguage.googleapis.com/v1beta"
    api_key_env: "GOOGLE_API_KEY"
    models: ["gemini-2.5-pro", "gemini-2.5-flash", "text-embedding-004"]

Ollama (Local)

yaml
  ollama:
    provider: "ollama"
    base_url: "http://localhost:11434"
    auth_type: "none"
    models: ["llama3.3", "qwen3", "phi-4", "nomic-embed-text"]

OpenAI-Compatible (Custom)

For any provider with an OpenAI-compatible API:

yaml
  infercom:
    provider: "openai-compatible"
    base_url: "https://api.infercom.ai/v1"
    api_key_env: "INFERCOM_API_KEY"
    models: ["MiniMax-M2.5", "DeepSeek-R1"]

Format Translation

LLMProxy automatically translates between provider formats:

  • OpenAIAnthropic: Messages format, system prompt handling, streaming events
  • OpenAIGoogle: Content parts, role mapping, safety settings
  • OpenAIAzure: Deployment URL construction, API version headers
  • OpenAIOllama: Direct pass-through (Ollama uses OpenAI format)

Multimodal content (images) is also translated:

  • Anthropic: base64 or urlsource format
  • Google: inlineData or fileData format
  • MIME type auto-detection

MIT License