v1.74.3-stable
Deploy this version​
- Docker
- Pip
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
docker.litellm.ai/berriai/litellm:v1.74.3-stable
pip install litellm==1.74.3.post1
Key Highlights​
- MCP: Model Access Groups - Add mcp servers to access groups, for easily managing access to users and teams.
- MCP: Tool Cost Tracking - Set prices for each MCP tool.
- Model Hub v2 - New OSS Model Hub for telling developers what models are available on the proxy.
- Bytez - New LLM API Provider.
- Dashscope API - Call Alibaba's qwen models via new Dashscope API Provider.
MCP Gateway: Model Access Groups​
v1.74.3-stable adds support for adding MCP servers to access groups, this makes it easier for Proxy Admins to manage access to MCP servers across users and teams.
For developers, this means you can now connect to multiple MCP servers by passing the access group name in the x-mcp-servers header.
Read more here
MCP Gateway: Tool Cost Tracking​
This release adds cost tracking for MCP tool calls. This is great for Proxy Admins giving MCP access to developers as you can now attribute MCP tool call costs to specific LiteLLM keys and teams.
You can set:
- Uniform server cost: Set a uniform cost for all tools from a server
- Individual tool cost: Define individual costs for specific tools (e.g., search_tool costs $10, get_weather costs $5).
- Dynamic costs: For use cases where you want to set costs based on the MCP's response, you can write a custom post mcp call hook to parse responses and set costs dynamically.
Model Hub v2​
v1.74.3-stable introduces a new OSS Model Hub for telling developers what models are available on the proxy.
This is great for Proxy Admins as you can now tell developers what models are available on the proxy.
This improves on the previous model hub by enabling:
- The ability to show Developers models, even if they don't have a LiteLLM key.
- The ability for Proxy Admins to select specific models to be public on the model hub.
- Improved search and filtering capabilities:
- search for models by partial name (e.g.
xai grok-4) - filter by provider and feature (e.g. 'vision' models)
- sort by cost (e.g. cheapest vision model from OpenAI)
- search for models by partial name (e.g.
New Models / Updated Models​
Pricing / Context Window Updates​
| Provider | Model | Context Window | Input ($/1M tokens) | Output ($/1M tokens) | Type |
|---|---|---|---|---|---|
| Xai | xai/grok-4 | 256k | $3.00 | $15.00 | New |
| Xai | xai/grok-4-0709 | 256k | $3.00 | $15.00 | New |
| Xai | xai/grok-4-latest | 256k | $3.00 | $15.00 | New |
| Mistral | mistral/devstral-small-2507 | 128k | $0.1 | $0.3 | New |
| Mistral | mistral/devstral-medium-2507 | 128k | $0.4 | $2 | New |
| Azure OpenAI | azure/o3-deep-research | 200k | $10 | $40 | New |
Features​
- Xinference
- Image generation API support - PR
- Bedrock
- API Key Auth support for AWS Bedrock API - PR
- 🆕 Dashscope
- New integration from Alibaba (enables qwen usage) - PR
- 🆕 Bytez
- New /chat/completion integration - PR
Bugs​
- Github Copilot
- Fix API base url for Github Copilot - PR
- Bedrock
- XAI
- ensure finish_reason includes tool calls when xai responses with tool calls - PR
LLM API Endpoints​
Features​
- /completions
- Return ‘reasoning_content’ on streaming - PR
- /chat/completions
- Add 'thinking blocks' to stream chunk builder - PR
- /v1/messages
MCP Gateway​
Features​
- Cost Tracking
- Auth
- MCP Server
- Allow using stdio MCPs with LiteLLM (enables using Circle CI MCP w/ LiteLLM) - PR, Get Started
Bugs​
- General
- Fix task group is not initialized error - PR s/o @juancarlosm
- MCP Server
- Fix mcp tool separator to work with Claude code - PR, Get Started
- Add validation to mcp server name to not allow "-" (enables namespaces to work) - PR