LiteLLM

Universal LLM routing library and proxy server that provides a single OpenAI-compatible interface to 100+ LLM providers with cost tracking, fallbacks, and load balancing.

Evaluated Mar 06, 2026 (0d ago) vcurrent
Homepage ↗ Repo ↗ AI & Machine Learning llm python proxy openai-compatible routing cost-tracking
⚙ Agent Friendliness
64
/ 100
Can an agent use this?
🔒 Security
57
/ 100
Is it safe for agents?
⚡ Reliability
57
/ 100
Does it work consistently?

Score Breakdown

⚙ Agent Friendliness

MCP Quality
--
Documentation
82
Error Messages
78
Auth Simplicity
95
Rate Limits
90

🔒 Security

TLS Enforcement
0
Auth Strength
70
Scope Granularity
60
Dep. Hygiene
75
Secret Handling
80

API keys for upstream providers are passed as environment variables or YAML config; ensure config files are not committed to version control. Proxy master key provides coarse access control only. No built-in secret rotation.

⚡ Reliability

Uptime/SLA
0
Version Stability
75
Breaking Changes
70
Error Recovery
82
AF Security Reliability

Best When

You need a single call site that can target any of 100+ LLM providers and want automatic fallback, retries, and cost logging without rewriting agent code per provider.

Avoid When

You rely on provider-specific streaming events, function-call schemas, or response fields that LiteLLM's translation layer does not yet map correctly.

Use Cases

  • Route agent LLM calls across multiple providers (OpenAI, Anthropic, Bedrock, Vertex) with a single unified API surface
  • Implement automatic fallback chains so agents continue operating when a primary LLM provider is unavailable
  • Track per-agent token usage and cost across providers without instrumenting each provider SDK separately
  • Deploy a shared LiteLLM proxy server so multiple agents share rate limit budgets and a single set of API keys
  • Load-balance inference across multiple deployments of the same model to increase throughput for high-QPS agents

Not For

  • Fine-tuning or training models — LiteLLM is inference routing only
  • Storing or retrieving conversation history — no built-in memory or persistence layer
  • Applications requiring vendor-specific features not exposed through the OpenAI-compatible interface

Interface

REST API
Yes
GraphQL
No
gRPC
No
MCP Server
No
SDK
Yes
Webhooks
No

Authentication

Methods: api_key none
OAuth: No Scopes: No

No auth required when used as a local Python library. Proxy server mode supports a LITELLM_MASTER_KEY or per-virtual-key auth for team use. Upstream provider keys are passed via environment variables or config.

Pricing

Model: open_source
Free tier: Yes
Requires CC: No

The core library and proxy are free and open source. Enterprise tier is optional for large team deployments.

Agent Metadata

Pagination
none
Idempotent
Full
Retry Guidance
Documented

Known Gotchas

  • Provider-specific parameters (e.g. Anthropic top_k, Bedrock guardrail IDs) must be passed via extra_body or provider-prefixed kwargs; they are silently ignored otherwise
  • Streaming response chunks differ subtly between providers even through the adapter — agents parsing raw chunks may break on provider switch
  • Router fallback order is defined in config, not inferred; if no fallback list is set, a provider outage raises immediately rather than trying alternatives
  • Cost tracking requires setting up a cost database (Redis or Postgres); without it, usage() returns None silently
  • Virtual key rate limits in proxy mode are enforced in-memory per pod — horizontal scaling without shared Redis will allow each pod its own full quota

Alternatives

Full Evaluation Report

Detailed scoring breakdown, competitive positioning, security analysis, and improvement recommendations for LiteLLM.

$99

Scores are editorial opinions as of 2026-03-06.

5178
Packages Evaluated
26151
Need Evaluation
173
Need Re-evaluation
Community Powered