v1.74.3-stable
Deploy this version​
- Docker
- Pip
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:v1.74.3-stable
pip install litellm==1.74.3.post1
Key Highlights​
- MCP: Model Access Groups - Add mcp servers to access groups, for easily managing access to users and teams.
- MCP: Tool Cost Tracking - Set prices for each MCP tool.
- Model Hub v2 - New OSS Model Hub for telling developers what models are available on the proxy.
- Bytez - New LLM API Provider.
- Dashscope API - Call Alibaba's qwen models via new Dashscope API Provider.
MCP Gateway: Model Access Groups​
v1.74.3-stable adds support for adding MCP servers to access groups, this makes it easier for Proxy Admins to manage access to MCP servers across users and teams.
For developers, this means you can now connect to multiple MCP servers by passing the access group name in the x-mcp-servers
header.
Read more here
MCP Gateway: Tool Cost Tracking​
This release adds cost tracking for MCP tool calls. This is great for Proxy Admins giving MCP access to developers as you can now attribute MCP tool call costs to specific LiteLLM keys and teams.
You can set:
- Uniform server cost: Set a uniform cost for all tools from a server
- Individual tool cost: Define individual costs for specific tools (e.g., search_tool costs $10, get_weather costs $5).
- Dynamic costs: For use cases where you want to set costs based on the MCP's response, you can write a custom post mcp call hook to parse responses and set costs dynamically.
Model Hub v2​
v1.74.3-stable introduces a new OSS Model Hub for telling developers what models are available on the proxy.
This is great for Proxy Admins as you can now tell developers what models are available on the proxy.
This improves on the previous model hub by enabling:
- The ability to show Developers models, even if they don't have a LiteLLM key.
- The ability for Proxy Admins to select specific models to be public on the model hub.
- Improved search and filtering capabilities:
- search for models by partial name (e.g.
xai grok-4
) - filter by provider and feature (e.g. 'vision' models)
- sort by cost (e.g. cheapest vision model from OpenAI)
- search for models by partial name (e.g.
New Models / Updated Models​
Pricing / Context Window Updates​
Provider | Model | Context Window | Input ($/1M tokens) | Output ($/1M tokens) | Type |
---|---|---|---|---|---|
Xai | xai/grok-4 | 256k | $3.00 | $15.00 | New |
Xai | xai/grok-4-0709 | 256k | $3.00 | $15.00 | New |
Xai | xai/grok-4-latest | 256k | $3.00 | $15.00 | New |
Mistral | mistral/devstral-small-2507 | 128k | $0.1 | $0.3 | New |
Mistral | mistral/devstral-medium-2507 | 128k | $0.4 | $2 | New |
Azure OpenAI | azure/o3-deep-research | 200k | $10 | $40 | New |
Features​
- Xinference
- Image generation API support - PR
- Bedrock
- API Key Auth support for AWS Bedrock API - PR
- 🆕 Dashscope
- New integration from Alibaba (enables qwen usage) - PR
- 🆕 Bytez
- New /chat/completion integration - PR
Bugs​
- Github Copilot
- Fix API base url for Github Copilot - PR
- Bedrock
- XAI
- ensure finish_reason includes tool calls when xai responses with tool calls - PR
LLM API Endpoints​
Features​
- /completions
- Return ‘reasoning_content’ on streaming - PR
- /chat/completions
- Add 'thinking blocks' to stream chunk builder - PR
- /v1/messages
MCP Gateway​
Features​
- Cost Tracking
- Auth
- MCP Server
- Allow using stdio MCPs with LiteLLM (enables using Circle CI MCP w/ LiteLLM) - PR, Get Started
Bugs​
- General
- Fix task group is not initialized error - PR s/o @juancarlosm
- MCP Server
- Fix mcp tool separator to work with Claude code - PR, Get Started
- Add validation to mcp server name to not allow "-" (enables namespaces to work) - PR
Management Endpoints / UI​
Features​
- Model Hub
- MCP
- Test Key
- Usage
- Teams
- SSO
- Allow users to run a custom sso login handler - PR
- Navbar
- improve user dropdown UI with premium badge and cleaner layout - PR
- General
Bugs​
- Model Hub
- fix duplicates in /model_group/info - PR
- MCP
- Fix UI not syncing MCP access groups properly with object permissions - PR
Logging / Guardrail Integrations​
Features​
- Langfuse
- Bedrock Guardrails
- Raise Bedrock output text on 'BLOCKED' actions from guardrail - PR
- OTEL
OTEL_RESOURCE_ATTRIBUTES
support - PR
- Guardrails AI
- pre-call + logging only guardrail (pii detection/competitor names) support - PR
- Guardrails
- [Enterprise] Support tag based mode for guardrails - PR, Get Started
- OpenAI Moderations API
- New guardrail integration - PR
- Prometheus
- support tag based metrics (enables prometheus metrics for measuring roo-code/cline/claude code engagement) - PR, Get Started
- Datadog LLM Observability
- Added
total_cost
field to track costs in DataDog LLM observability metrics - PR
- Added
Bugs​
- Prometheus
- Remove experimental
_by_tag
metrics (fixes cardinality issue) - PR
- Remove experimental
- Slack Alerting
- Fix slack alerting for outage and region outage alerts - PR, Get Started
Performance / Loadbalancing / Reliability improvements​
Bugs​
- Responses API Bridge
- add image support for Responses API when falling back on Chat Completions - PR s/o @ryan-castner
- aiohttp
- Properly close aiohttp client sessions to prevent resource leaks - PR
- Router
- don't add invalid deployment to router pattern match - PR
General Proxy Improvements​
Bugs​
- S3
- s3 config.yaml file - ensure yaml safe load is used - PR
- Audit Logs
- Add audit logs for model updates - PR
- Startup
- Multiple API Keys Created on Startup when max_budget is enabled - PR
- Auth
- Resolve model group alias on Auth (if user has access to underlying model, allow alias request to work) - PR
- config.yaml
- fix parsing environment_variables from config.yaml - PR
- Security
- Log hashed jwt w/ prefix instead of actual value - PR
Features​
- MCP
- Bump mcp version on docker img - PR
- Request Headers
- Forward ‘anthropic-beta’ header when forward_client_headers_to_llm_api is true - PR
New Contributors​
- @kanaka made their first contribution in https://github.com/BerriAI/litellm/pull/12418
- @juancarlosm made their first contribution in https://github.com/BerriAI/litellm/pull/12411
- @DmitriyAlergant made their first contribution in https://github.com/BerriAI/litellm/pull/12356
- @Rayshard made their first contribution in https://github.com/BerriAI/litellm/pull/12487
- @minghao51 made their first contribution in https://github.com/BerriAI/litellm/pull/12361
- @jdietzsch91 made their first contribution in https://github.com/BerriAI/litellm/pull/12488
- @iwinux made their first contribution in https://github.com/BerriAI/litellm/pull/12473
- @andresC98 made their first contribution in https://github.com/BerriAI/litellm/pull/12413
- @EmaSuriano made their first contribution in https://github.com/BerriAI/litellm/pull/12509
- @strawgate made their first contribution in https://github.com/BerriAI/litellm/pull/12528
- @inf3rnus made their first contribution in https://github.com/BerriAI/litellm/pull/12121