Change8

LiteLLM

AI & LLMs

Python SDK, Proxy Server (AI Gateway) to call 100+ LLM APIs in OpenAI (or native) format, with cost tracking, guardrails, loadbalancing and logging. [Bedrock, Azure, OpenAI, VertexAI, Cohere, Anthropic, Sagemaker, HuggingFace, VLLM, NVIDIA NIM]

Latest: v1.81.6.rc.5100 releases1 breaking changes23 common errorsView on GitHub

Release History

v1.81.6.rc.52 fixes
Feb 24, 2026

This release focuses on bug fixes, specifically removing a duplicate schema definition and introducing spend logging within the Release Candidate build.

v1.81.6.rc.43 fixes
Feb 24, 2026

LiteLLM v1.81.6.rc.4 focuses on stability by fixing several bugs related to parameter passing (like max_tokens for Azure) and response streaming across various providers.

v1.81.12-stable.11 fix
Feb 24, 2026

This patch release backports a stateless fix for StreamableHTTPSessionManager to the v1.81.12-stable branch.

v1.81.12-stable_gpt-5.3
Feb 24, 2026

This release appears to be a maintenance or internal version bump, indicated by the comparison range v1.81.12-stable to v1.81.12-stable_gpt-5.3, with no detailed changes provided in the snippet.

v1.81.15-nightly
Feb 24, 2026

This release provides an update to the nightly build, linking to the full comparison between the previous development version and the new nightly version.

CVEv1.81.59 fixes
Feb 24, 2026

This release focuses primarily on bug fixes across various integrations, including recovery from zombie processes, improved pricing application, and fixes related to rate limiting and authentication flows. Several recent fixes were also reverted.

litellm-CVE-v1.81.152 fixes
Feb 24, 2026

This release addresses a critical CVE by fixing issues related to the identification and handling of the 'presidio' model integration.

RCE-v1.81.15
Feb 24, 2026

This release primarily contains internal updates and links to the full comparison view for detailed changes.

litellm-v1.81.14.pre-call-hook-fix.dev16 fixes12 features
Feb 24, 2026

This release introduces new features like custom favicon support, configurable semantic cache dimensions, and guardrail policy versioning. Numerous bug fixes address issues across Bedrock, budget configuration, testing stability, and model cost mapping.

litellmv1.81.15.presidio.dev1 fix3 features
Feb 24, 2026

This release introduces new model pricing, enhanced content filtering with topic blockers, and a new guardrail hook for voice transcription, alongside a fix for Replicate server communication.

litellm-precall-dev-v1.81.1417 fixes12 features
Feb 24, 2026

This release introduces several new features, including custom favicon support, configurable semantic cache dimensions, and new guardrail policies. Numerous bug fixes address issues across Bedrock integration, budget calculation, testing stability, and data normalization.

litellm-pres-dev-v1.81.1521 fixes12 features
Feb 24, 2026

This release introduces new features like custom proxy favicons, configurable semantic cache dimensions, and guardrail policy versioning. Numerous bug fixes address issues across Bedrock integration, budget tracking, testing stability, and data normalization.

litellm_presidio-dev-v1.81.1516 fixes8 features
Feb 24, 2026

This release introduces new features like custom favicon support for the proxy, configurable semantic cache dimensions, and guardrail monitoring. It also includes numerous bug fixes across providers like Bedrock and Anthropic, and utility improvements like parameter key normalization.

v1.81.14.rc.134 fixes15 features
Feb 23, 2026

This release introduces significant feature additions, including Gemini 3.1 Pro preview support, new AI policy templates, and method-based routing for passthrough endpoints. Numerous bug fixes address issues across proxying, Bedrock integration, Anthropic translation, and CI/testing stability.

v1.81.14-nightly27 fixes16 features
Feb 23, 2026

This release introduces support for Gemini 3.1 Pro Preview, adds various policy template features, and includes extensive bug fixes across proxying, CI/CD workflows, and model parameter handling. Several tests requiring external services were skipped to improve CI stability.

litellm_idx-db-dev-v1.81.1439 fixes13 features
Feb 23, 2026

This release focuses heavily on UI enhancements, introducing new guardrail features, performance optimizations across streaming and cost calculation, and significant stability fixes for proxy connections and testing infrastructure.

v1.81.12-stable3 fixes
Feb 21, 2026

LiteLLM v1.81.12-stable focuses on stability by fixing several bugs related to parameter passing, streaming responses, and context length validation across various providers.

litellm_pro-mgmnt-dev-v1.81.1329 fixes13 features
Feb 21, 2026

This release introduces a new Prompt Management API, adds support for several new models and compliance features like EU AI Act templates, and includes numerous performance improvements and bug fixes across routing, proxy, and testing infrastructure.

litellm_langfuse-dev-v1.81.1344 fixes7 features
Feb 20, 2026

This release introduces a new Prompt Management API and significant UI enhancements, including Access Group selection and Usage filtering. It also contains extensive bug fixes focused heavily on improving test stability and isolation across various integrations.

litellm_sso-dev-v1.81.1329 fixes3 features
Feb 20, 2026

This release focuses heavily on improving test stability across various integrations (Langfuse, Vertex AI, SSO) and introduces UI enhancements for Access Groups. It also includes fixes for parameter handling in chat models and updates to managed batch functionality.

v1.81.9-stable.gemini.3.1-pro.sonnet-4.61 fix
Feb 20, 2026

This patch release corrects the model name mapping for Gemini 1.5 Flash to use the correct preview identifier.

v1.81.9-stable.gemini.3.1-pro-patch1 fix
Feb 20, 2026

This patch release fixes a bug where Gemini 1.5 Pro incorrectly errored when the response_format parameter was used.

v1.81.9-stable.patch.1
Feb 20, 2026

This release appears to be a patch update (v1.81.9-stable.patch.1) following a nightly build, with details available via the provided comparison link.

v1.81.13.dev132 fixes15 features
Feb 19, 2026

This release introduces new features like IBM watsonx.ai rerank support, Bedrock structured output API, and performance improvements in routing logic. Numerous bug fixes address issues across proxy headers, metrics counting, model parameter handling, and CI/testing stability.

v1.81.9-stable.gemini.3.1-pro1 fix
Feb 19, 2026

This patch release addresses a critical bug where streaming responses for Gemini 1.5 Pro were failing, ensuring correct behavior when streaming is enabled.

v1.81.3.oauth.dev2 features
Feb 19, 2026

This release introduces OAuth flow support for Azure OpenAI, configurable via environment variables.

litellm_1.81.13-dev24 fixes8 features
Feb 19, 2026

This release introduces a new Prompt Management API and significant UI enhancements, including access group selectors and usage filtering. It also includes extensive bug fixes focused heavily on improving test stability and isolation across various providers and features.

1.81.3-stable.dev.oauth
Feb 19, 2026

This release appears to be a development build focused on OAuth integration, as indicated by the version tag 1.81.3-stable.dev.oauth.

v1.80.15-stable.2
Feb 19, 2026

This release primarily consists of internal updates and bug fixes, with the full details available in the comparison link.

v1.80.11.gemini-metadata.dev1 feature
Feb 18, 2026

This release introduces support for extracting metadata specifically for the Gemini 1.5 Pro model.

v1.81.3-stable.sonnet-4-62 fixes
Feb 17, 2026

This release primarily addresses a bug in the Anthropic provider related to incorrect model name setting for Sonnet, ensuring proper model identification.

v1.81.13-nightly13 fixes3 features
Feb 17, 2026

This release focuses heavily on infrastructure improvements, particularly stabilizing tests across multiple providers and fixing various bugs related to API parameters, SSO, and managed batches. New UI features include Access Group selection and Content Filter management.

v1.81.1337 fixes5 features
Feb 17, 2026

This release focuses heavily on improving test stability across various integrations (Langfuse, Vertex AI, SSO) and fixing numerous minor bugs. New features include UI enhancements for Access Groups and Content Filters, alongside updates to reasoning handling and model parameter exclusions.

v1.81.10.dev1
Feb 16, 2026

This release primarily contains internal testing and development builds, indicated by the version tag v1.81.10.dev1.

v1.81.12.rc.1
Feb 15, 2026

This release contains the changelog from v1.81.12-nightly to v1.81.12.rc.1, detailed in the provided comparison link.

v1.81.12-nightly30 fixes17 features
Feb 15, 2026

This release focuses heavily on bug fixes across various providers (Azure, OCI, Anthropic, Vertex AI) and internal systems like the scheduler and proxy. It also introduces several new features, including enhanced Guardrails, new model support (MiniMax, Vertex GLM-5), and UI improvements for MCP management and Zscaler AI Guard.

v1.81.9-stable29 fixes17 features
Feb 15, 2026

This release focuses heavily on bug fixes across various providers (including Anthropic, Vertex AI, and GigaChat) and significant enhancements to the UI, including new budget management features and improved admin panels. It also introduces support for new models like Qwen3-235B and Claude Opus 4.6.

litellm_dev-test-v0.1
Feb 14, 2026

This release primarily serves as a development test build, linking to the full comparison changelog for details.

litellm-readonly-v0.1
Feb 14, 2026

This release primarily links to the full changelog for version v0.1, indicating a minor version update or specific build tag promotion.

v1.81.11-nightly23 fixes8 features
Feb 13, 2026

This release introduces new configuration options, enhanced guardrail capabilities, and support for new models like MiniMax. Numerous bug fixes address issues related to streaming errors, queue management, Azure compatibility, and model metadata synchronization.

v1.81.0-patch43 fixes
Feb 13, 2026

This patch release focuses on stability, fixing streaming issues with Azure OpenAI and correcting token usage calculation for Gemini 1.5 Flash.

litellm_dev-test-v025 fixes16 features
Feb 12, 2026

This release focuses heavily on stability, security fixes (including CVEs), and extensive UI/UX improvements across logging, teams, and authentication flows. New features include support for Websearch interception hooks and expanded model support for Anthropic, Sarvam AI, and Perplexity.

v1.81.3-stable.opus-4-61 fix
Feb 11, 2026

This patch release fixes a bug where streaming requests failed for the Anthropic Claude 3 Opus model via the completion endpoint.

litellm_stable_build-test-v0.124 fixes17 features
Feb 11, 2026

This release focuses heavily on stability, bug fixes across various providers (including Bedrock, Cohere, Vertex AI), and significant improvements to the LiteLLM UI and OAuth2 support. New features include Websearch interception and support for Sarvam AI and Perplexity Research API.

v1.81.10-nightly19 fixes12 features
Feb 11, 2026

This release focuses heavily on stability, security fixes (including critical CVEs), and extensive UI improvements and refactoring. New features include support for Websearch interception hooks and expanded model/API support for Anthropic, Sarvam AI, and Perplexity.

v1.81.3-stable
Feb 8, 2026

This release primarily serves as a stable tag for version v1.81.3, referencing the changes made since the previous development build.

v1.81.9.rc.128 fixes17 features
Feb 8, 2026

This release focuses heavily on bug fixes across various providers (including Anthropic, Vertex AI, and GigaChat) and significant enhancements to the UI, including new budget management features and model cost map updates. Key improvements were made to tracing, caching performance, and general stability.

v1.81.9-nightly29 fixes15 features
Feb 7, 2026

This release focuses heavily on bug fixes across various providers (including Vertex AI, GigaChat, and Anthropic), significant UI enhancements for administration and usage tracking, and expanded model support, notably for Claude Opus 4.6 and new OpenRouter/ElevenLabs models.

v1.81.0-stable.11 fix
Feb 7, 2026

This hotfix release addresses a specific user_id issue found in the /key/list endpoint.

v1.81.0-patch32 fixes
Feb 6, 2026

This patch release addresses critical bugs related to the Gemini 1.5 Flash model, specifically fixing incorrect empty tool_calls responses and 400 errors when using tool_code.

v1.80.0-stable.opus-4-6
Feb 5, 2026

This release primarily contains a link to the full changelog comparing versions v1.81.0-stable and v1.80.0-stable.opus-4-6.

v1.81.8-nightly5 fixes3 features
Feb 5, 2026

This release focuses on UI enhancements, adding new models via OpenRouter, and fixing several internal regressions and inconsistencies, particularly around provider responses and typing.

v1.81.6.rc.12 fixes
Feb 5, 2026

This release focuses on bug fixes for the Ollama provider, specifically addressing issues with empty choices and incorrect finish reasons in the completion endpoint.

1.81.3.rc.62 fixes
Feb 4, 2026

Release v1.81.3.rc.6 primarily focuses on bug fixes, addressing incorrect parameter passing for Azure OpenAI and improving streaming response handling across providers.

v1.81.7.dev133 fixes22 features
Feb 4, 2026

This release focuses heavily on bug fixes across various providers like Gemini, Vertex AI, and AWS Bedrock, alongside significant feature additions to the UI, Search API, and LiteLLM Vector Stores, including new model support.

v1.81.7-nightly37 fixes16 features
Feb 3, 2026

This release focuses heavily on bug fixes across various providers (Gemini, AWS, Azure, Anthropic) and significant enhancements to the UI, including new usage reporting features and improved search capabilities. New models and integrations, such as the Claude Agent SDK, were also added.

v1.81.6-nightly33 fixes22 features
Feb 1, 2026

This release focuses heavily on bug fixes across various providers like Gemini, AWS Bedrock, and Vertex AI, alongside significant feature additions to the UI, Search API, and LiteLLM Vector Stores, including new model support.

v1.81.0-patch23 fixes
Jan 30, 2026

This patch release focuses on stability improvements, primarily fixing issues related to API key handling and token reporting for the Azure OpenAI provider, along with resolving streaming response truncation bugs.

v1.81.3.dev1
Jan 29, 2026

This release appears to be a development build (v1.81.3.dev1) following a release candidate (v1.81.3.rc.4), with the full changelog available via the provided comparison link.

v1.81.5-nightly7 fixes4 features
Jan 29, 2026

This release focuses on stability, introducing features like Bedrock converse caching and configurable CLI JWT expiration, alongside numerous bug fixes and infrastructure improvements to stabilize testing.

v1.81.3.rc.532 fixes17 features
Jan 29, 2026

This release introduces significant enhancements to observability (Datadog, Prometheus metrics), adds support for Vercel AI Gateway embeddings, and includes numerous bug fixes across model integrations and UI functionality.

v1.81.0-patch
Jan 29, 2026

This release contains patch updates for LiteLLM, detailed in the full changelog link.

v1.81.3.rc.4
Jan 28, 2026

This release primarily consists of internal updates and bug fixes, as indicated by the comparison between release candidates v1.81.3.rc.3 and v1.81.3.rc.4.

v1.81.3.rc.3
Jan 28, 2026

This release primarily contains internal updates and bug fixes, detailed in the linked comparison view.

v1.81.3.rc.2
Jan 26, 2026

This release (v1.81.3.rc.2) is a release candidate update, with the full changelog available via the provided comparison link.

v1.81.3.rc.1
Jan 25, 2026

This release primarily points to the full changelog for version v1.81.3.rc.1, indicating minor updates or release candidate finalization.

v1.81.3-nightly
Jan 25, 2026

This release primarily serves as a nightly update, linking to the full comparison between v1.81.3.rc and v1.81.3-nightly.

v1.81.3.rc40 fixes13 features
Jan 25, 2026

This release focuses heavily on bug fixes across various providers (Bedrock, Ollama, Google) and infrastructure components like logging and routing. Key features include support for Azure OpenAI v1 API, enhanced retry mechanisms, and updates to Gemini and Volcengine integrations.

v1.81.0-stable
Jan 24, 2026

This release marks the stable release of version 1.81.0, following the release candidate 5.

v1.81.0.rc.5
Jan 22, 2026

This release primarily consists of minor updates and bug fixes, detailed in the full changelog comparing versions v1.81.0.rc.4 and v1.81.0.rc.5.

v1.81.0.rc.4
Jan 21, 2026

This release contains the full changelog available via the provided comparison link between v1.81.0-nightly and v1.81.0.rc.4.

v1.81.1-nightly39 fixes12 features
Jan 21, 2026

This release focuses heavily on bug fixes across various providers (Bedrock, Ollama, Groq, Gemini) and internal systems like logging and routing. New features include enhanced retry logic for completions and support for Azure OpenAI v1 API and KEDA in Helm charts.

v1.81.0.rc.14 fixes6 features
Jan 20, 2026

This release focuses heavily on UI improvements, including new status queries and views for deleted resources, alongside adding Websearch support for Claude Code and fixing reliability issues like OOMs.

v1.81.0-nightly4 fixes6 features
Jan 18, 2026

This release focuses heavily on UI enhancements, including new status queries and views for deleted resources, alongside adding Websearch support for Claude Code and fixing reliability issues with image handling.

1.80.15-stable.1
Jan 17, 2026

This release incorporates a merge from a pull request initiated by dominicfeliton and merged by yuneng-jiang.

v1.80.15-stable2 fixes
Jan 17, 2026

This release primarily focuses on bug fixes, including ensuring correct parameter passing for Azure OpenAI and improving streaming response handling.

v1.80.17-nightly5 fixes3 features
Jan 17, 2026

This release focuses on UI enhancements for team management, infrastructure updates including schema migrations for keys and teams, and several bug fixes across Vertex AI integration and logging. Several recent changes were also reverted to maintain stability.

v1.80.16.dev639 fixes8 features
Jan 16, 2026

This release introduces new model pricing, enhances guardrail functionality with a default failopen option, and includes numerous fixes across model integrations, pricing accuracy, and UI components. Performance bottlenecks under heavy load were also addressed.

v1.80.11-stable.11 fix
Jan 15, 2026

This hotfix release addresses a critical issue related to Azure Storage circular references.

v1.80.16-nightly27 fixes7 features
Jan 13, 2026

This release focuses heavily on bug fixes across various providers (Gemini, Anthropic, OCI) and internal systems like guardrails and pricing. It also introduces several UI enhancements and infrastructure updates, including Helm chart versioning sync.

v1.80.15.dev123 fixes4 features
Jan 12, 2026

This release focuses heavily on bug fixes across various providers like Gemini, Anthropic, and OCI, alongside adding new features such as Bedrock service tier translation and support for replicate chat models. It also includes documentation updates regarding load balancing and Redis requirements.

v1.80.15.rc.2
Jan 12, 2026

This release contains the changelog link for version v1.80.15.rc.2 compared to the previous nightly build.

v1.80.15.rc.134 fixes18 features
Jan 11, 2026

This release introduces several new provider integrations, enhanced Prometheus metrics for monitoring, and numerous bug fixes across proxying, routing, and provider configurations. Performance improvements were also made to provider configuration lookups.

v1.80.15-nightly32 fixes18 features
Jan 11, 2026

This release introduces numerous new features, including provider additions (abliteration.ai, Bedrock for token counting), enhanced Prometheus metrics, and UI improvements for the Playground. Numerous bug fixes address issues across proxy streaming, provider configurations (Gemini, Azure), security, and internal workflows.

v1.80.11-stable20 fixes14 features
Jan 10, 2026

This release introduces significant new features like Guardrails Load Balancing, Unified Skills API, and a new Rag Search API. It also includes numerous bug fixes across pricing, integrations (like Langfuse and Datadog), and UI elements, alongside performance improvements via lazy loading.

v1.80.13.rc.1
Jan 9, 2026
v1.80.13-nightly
Jan 9, 2026
v1.80.12-nightly17 fixes31 features
Jan 7, 2026

This release introduces significant feature enhancements, including support for image tokens, AWS Polly TTS, Minimax integration, and major UI improvements like error code filtering and key management updates. It also includes substantial internal refactoring to lazy load configuration classes for better performance.

v1.80.8-stable.1-patch01
Jan 6, 2026

No changelog details were provided for this patch release.

v1.80.11.rc.125 fixes21 features
Dec 22, 2025

This release adds several new capabilities such as Guardrails load‑balancing, a unified Skills API, Rag Search and Query APIs with rerankers, and expanded model support, while also delivering numerous bug fixes and performance improvements across CI/CD, UI, and observability components.

v1.80.11-nightly28 fixes20 features
Dec 22, 2025

This release adds numerous new features such as Guardrails load balancing, a unified Skills API, and expanded model support, while delivering a large set of bug fixes and performance improvements across CI/CD, UI, and observability components.

v1.80.10.dev.1Breaking24 fixes22 features
Dec 19, 2025

This release adds extensive new features—including Gemini 3 flash preview support, Azure Sentinel logging, Guardrails enhancements, and a new LinkUp Search provider—while delivering numerous bug fixes and a breaking change that renames `extra_headers` to `additional_headers`.

v1.80.10.rc.5
Dec 19, 2025

No specific changes were listed in the release notes; refer to the full changelog for details.

v1.80.10.rc.4
Dec 17, 2025

No changelog details were provided beyond the link to the full comparison.

v1.80.8-stable.1
Dec 17, 2025

No changes were documented for this release.

v1.80.10.rc.317 fixes21 features
Dec 17, 2025

This release adds numerous new providers and features—including Stability AI, Azure Cohere reranking, and VertexAI Agent Engine—while fixing a wide range of bugs and refactoring lazy imports for better performance.

v1.80.10.rc.2
Dec 14, 2025

No detailed changelog items were provided; see the full changelog link for specifics.

v1.80.10.rc.125 fixes22 features
Dec 14, 2025

LiteLLM v1.80.10 introduces Azure GPT‑5.2 models, new security guardrail evidence, expanded UI features, and numerous bug fixes and documentation updates.

v1.80.10-nightly28 fixes26 features
Dec 14, 2025

Litellm v1.80.10 introduces Azure GPT‑5.2 model support, new security guardrail evidence, JWT team‑id selection, OTEL latency metrics, UI enhancements, and a large set of bug fixes and documentation updates.

v1.80.8-stable
Dec 14, 2025

Release v1.80.8 stable was published; specific changes are not listed in the provided notes. See the full changelog for details.

Common Errors

ServiceUnavailableError7 reports

ServiceUnavailableError usually indicates the LLM provider is overloaded or temporarily unavailable. Implement retry logic with exponential backoff using `retry` decorator in `litellm` around the failing function calls, and/or check the LLM provider's status page to confirm an outage before retrying. You might also need to increase your rate limits or switch to a different model if the issue persists.

NotFoundError5 reports

The "NotFoundError" in LiteLLM often arises from an incorrectly specified model name or an issue with the provider configuration. Double-check that the `model` parameter matches exactly what's expected by the provider (including any provider prefixes like "bedrock/"). Verify your provider's authentication and configuration are set correctly using `litellm.api_key` or appropriate environment variables as described in the LiteLLM documentation for the specific provider.

ModuleNotFoundError3 reports

"ModuleNotFoundError" usually arises when a required Python package isn't installed, or the Python environment is misconfigured. Resolve this by first ensuring the necessary package is installed using `pip install <package_name>` (e.g., `pip install fastapi`). If the package is installed yet still causing problems, verify your Python environment and that the package version is compatible.

ContentPolicyViolationError3 reports

ContentPolicyViolationError usually arises from AI models flagging input or output text as violating their content policies. To fix it, carefully review your prompt and the model's response for potentially harmful or sensitive content. Redraft prompts to be less suggestive or controversial and implement output filtering to sanitize model responses, ensuring compliance with content safety guidelines.

BadRequestError2 reports

BadRequestError in litellm usually stems from incompatibility between the input format (e.g., tool call IDs, API parameters) expected by a destination model (like OpenAI or Mistral) and that provided by the originating model (e.g., MiniMax, VertexAI Claude). To fix, sanitize or transform the model's output to strictly adhere to the API specifications of the destination model before sending the request; this often involves modifying tool call IDs or other parameter formats. Consider using litellm's `input_formatter` to adjust the request parameters.

MidStreamFallbackError2 reports

MidStreamFallbackError in litellm usually arises when a streaming response fails mid-flight during a fallback scenario, obscuring the original error code. To fix this, ensure that the `MidStreamFallbackError` preserves the original exception's status code and message, allowing litellm to correctly handle errors like rate limits or other service issues further upstream - log this internal exception and its trace! Also ensure that any stream that throws an exception is immediately closed to prevent resource leaks.

Related AI & LLMs Packages

Subscribe to Updates

Get notified when new versions are released

RSS Feed