litellm-v1.81.14.pre-call-hook-fix.dev
📦 litellmView on GitHub →
✨ 12 features🐛 16 fixes🔧 6 symbols
Summary
This release introduces new features like custom favicon support, configurable semantic cache dimensions, and guardrail policy versioning. Numerous bug fixes address issues across Bedrock, budget configuration, testing stability, and model cost mapping.
Migration Steps
- If you rely on camelCase keys for thinking parameters, they are now normalized to snake_case; update any code expecting camelCase.
- If you were using hardcoded timezone maps for budget configuration, note that this has been replaced with ZoneInfo usage.
✨ New Features
- Show proxy URL in ModelHub.
- Add custom favicon support for the proxy.
- Support configurable vector dimensions for semantic-cache.
- Add optional digest mode for Slack alert types.
- Enable context-1m-2025-08-07 beta header for vertex_ai provider (after a revert).
- Guardrail Policy Versioning implemented.
- Add Noma guardrails v2 based on custom guardrails.
- LiteLLM network mock capability added.
- Implement Guardrail Monitor to measure guardrail reliability in production.
- Add Singapore guardrail policies (PDPA + MAS AI Risk Management).
- Add variant parameter to video content download.
- UI: Add Blog Dropdown in Navbar.
🐛 Bug Fixes
- Correct modelInput format for Bedrock Converse API batch models.
- Only tag selected deployment in access group creation.
- Prevent double UUID in create_file S3 key for Bedrock.
- Normalize camelCase thinking param keys to snake_case in utils.
- Fix timezone config lookup and replace hardcoded timezone map with ZoneInfo in budget module.
- Add missing return type annotations to iterator protocol methods in streaming_handler.
- Avoid mutating caller-owned dicts in SpendUpdateQueue aggregation.
- Fix model cost map for anthropic fast and inference_geo.
- Add Priority PayGo cost tracking for gemini/vertex ai.
- Fix StopIteration in prisma self-heal cooldown test.
- Use absolute path for model_prices JSON validation test.
- Add INCOMPLETE to interactions status enum expected values in tests.
- Make RPM limit test sequential to fix race condition.
- Fix Anthropic model wildcard access issue.
- Pass api_key from litellm_params to video remix handlers.
- Fix Spend Update Queue Aggregation Never Triggers with Default Presets.