About LiteLLM
One gateway in front of every model you call.
LiteLLM is an open-source LLM gateway built by BerriAI, a Y Combinator company started in 2023 by Krrish Dholakia and Ishaan Jaff. The repo is on GitHub at BerriAI/litellm with 44k+ stars and a thousand-plus contributors. Stripe, Netflix, Google ADK, Greptile and OpenHands run it in production, alongside the long tail of teams that wanted one endpoint instead of ten provider SDKs.
The product has two shapes. The Python SDK lets code call litellm.completion() with the same call signature regardless of provider. The Proxy is a self-hosted server (Docker, Kubernetes or the LiteLLM CLI) that exposes one OpenAI-compatible REST endpoint and routes the call to OpenAI, Anthropic, Azure OpenAI, Amazon Bedrock, Google Vertex AI, Cohere, Mistral, Hugging Face, Groq or any of the 100+ providers it speaks. Around that proxy sit virtual API keys per team or user, per-key and per-team budgets, rate limits, automatic fallback and retry across deployments, response caching and structured spend logs. The Postgres tables behind the proxy (LiteLLM_VerificationToken, LiteLLM_TeamTable, LiteLLM_UserTable, LiteLLM_SpendLogs, LiteLLM_BudgetTable) hold every request with the api_key, user, team_id, end_user, model, model_group, prompt_tokens, completion_tokens, total_tokens, spend, request_tags and metadata attached, which is what turns a stack of provider invoices into something a finance and ML team can query.