v1.80.15 - Manus API Support
Deploy this version
- Docker
- Pip
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
docker.litellm.ai/berriai/litellm:v1.80.15.rc.1
pip install litellm==1.80.15
Key Highlights
- Manus API Support - New provider support for Manus API on /responses and GET /responses endpoints
- MiniMax Provider - Full support for MiniMax chat completions, TTS, and Anthropic native endpoint
- AWS Polly TTS - New TTS provider using AWS Polly API
- SSO Role Mapping - Configure role mappings for SSO providers directly in the UI
- Cost Estimator - New UI tool for estimating costs across multiple models and requests
- MCP Global Mode - Configure MCP servers globally with visibility controls
- Interactions API Bridge - Use all LiteLLM providers with the Interactions API
- RAG Query Endpoint - New RAG Search/Query endpoint for retrieval-augmented generation
- UI Usage - Endpoint Activity - Users can now see Endpoint Activity Metrics in the UI
- 50% Overhead Reduction - LiteLLM now sends 2.5× more requests to LLM providers
Performance - 50% Overhead Reduction
LiteLLM now sends 2.5× more requests to LLM providers by replacing sequential if/elif chains with O(1) dictionary lookups for provider configuration resolution (92.7% faster). This optimization has a high impact because it runs inside the client decorator, which is invoked on every HTTP request made to the proxy server.
Before
Note: Worse-looking provider metrics are a good sign here—they indicate requests spend less time inside LiteLLM.
============================================================
Fake LLM Provider Stats (When called by LiteLLM)
============================================================
Total Time: 0.56s
Requests/Second: 10746.68
Latency Statistics (seconds):
Mean: 0.2039s
Median (p50): 0.2310s
Min: 0.0323s
Max: 0.3928s
Std Dev: 0.1166s
p95: 0.3574s
p99: 0.3748s
Status Codes:
200: 6000
After
============================================================
Fake LLM Provider Stats (When called by LiteLLM)
============================================================
Total Time: 1.42s
Requests/Second: 4224.49
Latency Statistics (seconds):
Mean: 0.5300s
Median (p50): 0.5871s
Min: 0.0885s
Max: 1.0482s
Std Dev: 0.3065s
p95: 0.9750s
p99: 1.0444s
Status Codes:
200: 6000
The benchmarks run LiteLLM locally with a lightweight LLM provider to eliminate network latency, isolating internal overhead and bottlenecks so we can focus on reducing pure LiteLLM overhead on a single instance.
UI Usage - Endpoint Activity
Users can now see Endpoint Activity Metrics in the UI.
New Providers and Endpoints
New Providers (11 new providers)
| Provider | Supported LiteLLM Endpoints | Description |
|---|---|---|
| Manus | /responses | Manus API for agentic workflows |
| Manus | GET /responses | Manus API for retrieving responses |
| Manus | /files | Manus API for file management |
| MiniMax | /chat/completions | MiniMax chat completions |
| MiniMax | /audio/speech | MiniMax text-to-speech |
| AWS Polly | /audio/speech | AWS Polly text-to-speech API |
| GigaChat | /chat/completions | GigaChat provider for Russian language AI |
| LlamaGate | /chat/completions | LlamaGate chat completions |
| LlamaGate | /embeddings | LlamaGate embeddings |
| Abliteration AI | /chat/completions | Abliteration.ai provider support |
| Bedrock | /v1/messages/count_tokens | Bedrock as new provider for token counting |
New LLM API Endpoints (3 new endpoints)
| Endpoint | Method | Description | Documentation |
|---|---|---|---|
/responses/compact | POST | Compact responses API endpoint | Docs |
/rag/query | POST | RAG Search/Query endpoint | Docs |
/containers/{id}/files | POST | Upload files to containers | Docs |
New Models / Updated Models
New Model Support (100+ new models)
| Provider | Model | Context Window | Input ($/1M tokens) | Output ($/1M tokens) | Features |
|---|---|---|---|---|---|
| Azure | azure/gpt-5.2 | 400K | $1.75 | $14.00 | Reasoning, vision, caching |
| Azure | azure/gpt-5.2-chat | 128K | $1.75 | $14.00 | Reasoning, vision |
| Azure | azure/gpt-5.2-pro | 400K | $21.00 | $168.00 | Reasoning, vision, web search |
| Azure | azure/gpt-image-1.5 | - | Token-based | Token-based | Image generation/editing |
| Azure AI | azure_ai/gpt-oss-120b | 131K | $0.15 | $0.60 | Function calling |
| Azure AI | azure_ai/flux.2-pro | - | - | $0.04/image | Image generation |
| Azure AI | azure_ai/deepseek-v3.2 | 164K | $0.58 | $1.68 | Reasoning, function calling |
| Bedrock | amazon.nova-2-multimodal-embeddings-v1:0 | 8K | $0.135 | - | Multimodal embeddings |
| Bedrock | writer.palmyra-x4-v1:0 | 128K | $2.50 | $10.00 | Function calling, PDF |
| Bedrock | writer.palmyra-x5-v1:0 | 1M | $0.60 | $6.00 | Function calling, PDF |
| Bedrock | moonshot.kimi-k2-v1:0 | - | - | - | Kimi K2 model |
| Cerebras | cerebras/zai-glm-4.6 | 128K | $2.25 | $2.75 | Reasoning, function calling |
| GigaChat | gigachat/GigaChat-2-Lite | - | - | - | Chat completions |
| GigaChat | gigachat/GigaChat-2-Max | - | - | - | Chat completions |
| GigaChat | gigachat/GigaChat-2-Pro | - | - | - | Chat completions |
| Gemini | gemini/veo-3.1-generate-001 | - | - | - | Video generation |
| Gemini | gemini/veo-3.1-fast-generate-001 | - | - | - | Video generation |
| GitHub Copilot | 25+ models | Various | - | - | Chat completions |
| LlamaGate | 15+ models | Various | - | - | Chat, vision, embeddings |
| MiniMax | minimax/abab7-chat-preview | - | - | - | Chat completions |
| Novita | 80+ models | Various | Various | Various | Chat, vision, embeddings |
| OpenRouter | openrouter/google/gemini-3-flash-preview | - | - | - | Chat completions |
| Together AI | Multiple models | Various | Various | Various | Response schema support |
| Vertex AI | vertex_ai/zai-glm-4.7 | - | - | - | GLM 4.7 support |
Features
- Gemini
- Add image tokens in chat completion - PR #18327
- Add usage object in image generation - PR #18328
- Add thought signature support via tool call id - PR #18374
- Add thought signature for non tool call requests - PR #18581
- Preserve system instructions - PR #18585
- Fix Gemini 3 images in tool response - PR #18190
- Support snake_case for google_search tool parameters - PR #18451
- Google GenAI adapter inline data support - PR #18477
- Add deprecation_date for discontinued Google models - PR #18550
- Vertex AI
- Add centralized get_vertex_base_url() helper for global location support - PR #18410
- Convert image URLs to base64 for Vertex AI Anthropic - PR #18497
- Separate Tool objects for each tool type per API spec - PR #18514
- Add thought_signatures to VertexGeminiConfig - PR #18853
- Add support for Vertex AI API keys - PR #18806
- Add zai glm-4.7 model support - PR #18782
- Azure
- Add Azure gpt-image-1.5 pricing to cost map - PR #18347
- Add azure/gpt-5.2-chat model - PR #18361
- Add support for image generation via Azure AD token - PR #18413
- Add logprobs support for Azure OpenAI GPT-5.2 model - PR #18856
- Add Azure BFL Flux 2 models for image generation and editing - PR #18764, PR #18766
- Bedrock
- Databricks
- Add enhanced authentication, security features, and custom user-agent support - PR #18349
- MiniMax
- Together AI
- Add supports_response_schema to all supported Together AI models - PR #18368
- OpenRouter
- Add OpenRouter embeddings API support - PR #18391
- Anthropic
- Ollama
- Watsonx
- SAP
- Fireworks AI
- Correct deepseek-v3p2 pricing - PR #18483
- ZAI
- Add GLM-4.7 model with reasoning support - PR #18476
- Codestral
- Correctly route codestral chat and FIM endpoints - PR #18467
- Azure AI
- Fix authentication errors at messages API via azure_ai - PR #18500
New Provider Support
- AWS Polly - Add AWS Polly API for TTS - PR #18326
- GigaChat - Add GigaChat provider support - PR #18564
- LlamaGate - Add LlamaGate as a new provider - PR #18673
- Abliteration AI - Add abliteration.ai provider - PR #18678
- Manus - Add Manus API support on /responses, GET /responses - PR #18804
- 5 AI Providers via openai_like - Add 5 AI providers using openai_like - PR #18362
Bug Fixes
- Gemini
- Groq
- Remove deprecated Groq models and update model registry - PR #18062
- Vertex AI
- Handle unsupported region for Vertex AI count tokens endpoint - PR #18665
- General
- Fix request body for image embedding request - PR #18336
- Fix lost tool_calls when streaming has both text and tool_calls - PR #18316
- Add all resolution for gpt-image-1.5 - PR #18586
- Fix gpt-image-1 cost calculation using token-based pricing - PR #17906
- Fix response_format leaking into extra_body - PR #18859
- Align max_tokens with max_output_tokens for consistency - PR #18820
LLM API Endpoints
Features
- Responses API
- Interactions API
- Allow using all LiteLLM providers (interactions -> responses API bridge) - PR #18373
- RAG Search API
- Add RAG Search/Query endpoint - PR #18376
- CountTokens API
- Add Bedrock as a new provider for
/v1/messages/count_tokens- PR #18858
- Add Bedrock as a new provider for
- Generate Content
- Add generate content in LLM route - PR #18405
- General
Bugs
- General
- Handle empty error objects in response conversion - PR #18493
- Preserve client error status codes in streaming mode - PR #18698
- Return json error response instead of SSE format for initial streaming errors - PR #18757
- Fix auth header for custom api base in generateContent request - PR #18637
- Tool content should be string for Deepinfra - PR #18739
- Fix incomplete usage in response object passed - PR #18799
- Unify model names to provider-defined names - PR #18573
Management Endpoints / UI
Features
- SSO Configuration
- Virtual Keys
- Allow deleting key expiry - PR #18278
- Add optional query param "expand" to /key/list - PR #18502
- Key Table Loading Skeleton - PR #18527
- Allow column resizing on Keys Table - PR #18424
- Virtual Keys Table Loading State Between Pages - PR #18619
- Key and Team Router Setting - PR #18790
- Allow router_settings on Keys and Teams - PR #18675
- Use timedelta to calculate key expiry on generate - PR #18666
- Models + Endpoints
- Organizations & Teams
- Allow Organization Admins to See Organization Tab - PR #18400
- Resolve Organization Alias on Team Table - PR #18401
- Resolve Team Alias in Organization Info View - PR #18404
- Allow Organization Admins to View Their Organization Info - PR #18417
- Allow editing team_member_budget_duration in /team/update - PR #18735
- Reusable Duration Select + Team Update Member Budget Duration - PR #18736
- Usage & Spend
- Cost Estimator
- CloudZero
- Playground
- General UI
- UI styling improvements and fixes - PR #18310
- Add reusable "New" badge component for feature highlights - PR #18537
- Hide New Badges - PR #18547
- Change Budget page to Have Tabs - PR #18576
- Clicking on Logo Directs to Correct URL - PR #18575
- Add UI support for configuring meta URLs - PR #18580
- Expire Previous UI Session Tokens on Login - PR #18557
- Add license endpoint - PR #18311
- Router Fields Endpoint + React Query for Router Fields - PR #18880
Bugs
- UI Fixes
- Fix Key Creation MCP Settings Submit Form Unintentionally - PR #18355
- Fix UI Disappears in Development Environments - PR #18399
- Fix Disable Admin UI Flag - PR #18397
- Remove Model Analytics From Model Page - PR #18552
- Useful Links Remove Modal on Adding Links - PR #18602
- SSO Edit Modal Clear Role Mapping Values on Provider Change - PR #18680
- UI Login Case Sensitivity fix - PR #18877
- API Fixes
AI Integrations
New Integrations (4 new integrations)
| Integration | Type | Description |
|---|---|---|
| Focus | Logging | Focus export support for observability - PR #18802 |
| SigNoz | Logging | SigNoz integration for observability - PR #18726 |
| Qualifire | Guardrails | Qualifire guardrails and eval webhook - PR #18594 |
| Levo AI | Guardrails | Levo AI integration for security - PR #18529 |
Logging
- DataDog
- Fix span kind fallback when parent_id missing - PR #18418
- Langfuse
- Map Gemini cached_tokens to Langfuse cache_read_input_tokens - PR #18614
- Prometheus
- Braintrust
- Pass span_attributes in async logging and skip tags on non-root spans - PR #18409
- CloudZero
- Add user email to CloudZero - PR #18584
- OpenTelemetry
- LangSmith
- Add support for LangSmith organization-scoped API keys with tenant ID - PR #18623
- Generic API Logger
- Add log_format option to GenericAPILogger - PR #18587
Guardrails
- Content Filter
- Qualifire
- Add Qualifire eval webhook - PR #18836
- Lasso Security
- Add Lasso guardrail API docs - PR #18652
- Noma Security
- Add MCP guardrail support for Noma - PR #18668
- Bedrock Guardrails
- Remove redundant Bedrock guardrail block handling - PR #18634
- General
- Generic guardrail API update - PR #18647
- Prevent proxy startup failures from case-sensitive tool permission guardrail validation - PR #18662
- Extend case normalization to ALL guardrail types - PR #18664
- Fix MCP handling in unified guardrail - PR #18630
- Fix embeddings calltype for guardrail precallhook - PR #18740
Spend Tracking, Budgets and Rate Limiting
- Platform Fee / Margins - Add support for Platform Fee / Margins - PR #18427
- Negative Budget Validation - Add validation for negative budget - PR #18583
- Cost Calculation Fixes
- Tag Routing - Support toggling tag matching between ANY and ALL - PR #18776
MCP Gateway
- MCP Global Mode - Add MCP global mode - PR #18639
- MCP Server Visibility - Add configurable MCP server visibility - PR #18681
- MCP Registry - Add MCP registry - PR #18850
- MCP Stdio Header - Support MCP stdio header env overrides - PR #18324
- Parallel Tool Fetching - Parallelize tool fetching from multiple MCP servers - PR #18627
- Optimize MCP Server Listing - Separate health checks for optimized listing - PR #18530
- Auth Improvements
- Bug Fixes
- Migrate MCP Fetching Logic to React Query - PR #18352
Performance / Loadbalancing / Reliability improvements
- 92.7% Faster Provider Config Lookup - LiteLLM now stresses LLM providers 2.5x more - PR #18867
- Lazy Loading Improvements
- Consolidate lazy import handlers with registry pattern - PR #18389
- Complete lazy loading migration for all 180+ LLM config classes - PR #18392
- Lazy load additional components (types, callbacks, utilities) - PR #18396
- Add lazy loading for get_llm_provider - PR #18591
- Lazy-load heavy audio library and loggers - PR #18592
- Lazy load 9 heavy imports in litellm/utils.py - PR #18595
- Lazy load heavy imports to improve import time and memory usage - PR #18610
- Implement lazy loading for provider configs, model info classes, streaming handlers - PR #18611
- Lazy load 15 additional imports - PR #18613
- Lazy load 15+ unused imports - PR #18616
- Lazy load DatadogLLMObsInitParams - PR #18658
- Migrate utils.py lazy imports to registry pattern - PR #18657
- Lazy load get_llm_provider and remove_index_from_tool_calls - PR #18608
- Router Improvements
- Memory Improvements
- Database
- Docker
- Helm
- Add lifecycle support to Helm charts - PR #18517
- Authentication
- Logging Worker
- Handle event loop changes in multiprocessing - PR #18423
- Security
- Misc
- Remove double imports in main.py - PR #18406
- Add LITELLM_DISABLE_LAZY_LOADING env var to fix VCR cassette creation issue - PR #18725
- Add xiaomi_mimo to LlmProviders enum to fix router support - PR #18819
- Allow installation with current grpcio on old Python - PR #18473
- Add Custom CA certificates to boto3 clients - PR #18852
- Fix bedrock_cache, metadata and max_model_budget - PR #18872
- Fix LiteLLM SDK embedding headers missing field - PR #18844
- Put automatic reasoning summary inclusion behind feat flag - PR #18688
- turn_off_message_logging Does Not Redact Request Messages in proxy_server_request Field - PR #18897
Documentation Updates
- Provider Documentation
- Update MiniMax docs to be in proper format - PR #18403
- Add docs for 5 AI providers - PR #18388
- Fix gpt-5-mini reasoning_effort supported values - PR #18346
- Fix PDF documentation inconsistency in Anthropic page - PR #18816
- Update OpenRouter docs to include embedding support - PR #18874
- Add LITELLM_REASONING_AUTO_SUMMARY in doc - PR #18705
- MCP Documentation
- Guardrails Documentation
- Qualifire docs hotfix - PR #18724
- Infrastructure Documentation
- General
New Contributors
- @prasadkona made their first contribution in PR #18349
- @lucasrothman made their first contribution in PR #18283
- @aggeentik made their first contribution in PR #18317
- @mihidumh made their first contribution in PR #18361
- @Prazeina made their first contribution in PR #18498
- @systec-dk made their first contribution in PR #18500
- @xuan07t2 made their first contribution in PR #18514
- @RensDimmendaal made their first contribution in PR #18190
- @yurekami made their first contribution in PR #18483
- @agertz7 made their first contribution in PR #18556
- @yudelevi made their first contribution in PR #18550
- @smallp made their first contribution in PR #18536
- @kevinpauer made their first contribution in PR #18569
- @cansakiroglu made their first contribution in PR #18517
- @dee-walia20 made their first contribution in PR #18432
- @luxinfeng made their first contribution in PR #18477
- @cantalupo555 made their first contribution in PR #18476
- @andersk made their first contribution in PR #18473
- @majiayu000 made their first contribution in PR #18467
- @amangupta-20 made their first contribution in PR #18529
- @hamzaq453 made their first contribution in PR #18480
- @ktsaou made their first contribution in PR #18627
- @FlibbertyGibbitz made their first contribution in PR #18624
- @drorIvry made their first contribution in PR #18594
- @urainshah made their first contribution in PR #18524
- @mangabits made their first contribution in PR #18279
- @0717376 made their first contribution in PR #18564
- @nmgarza5 made their first contribution in PR #17330
- @wileykestner made their first contribution in PR #18445
- @minijeong-log made their first contribution in PR #14440
- @Isaac4real made their first contribution in PR #18710
- @marukaz made their first contribution in PR #18711
- @rohitravirane made their first contribution in PR #18712
- @lizzzcai made their first contribution in PR #18714
- @hkd987 made their first contribution in PR #18673
- @Mr-Pepe made their first contribution in PR #18674
- @gkarthi-signoz made their first contribution in PR #18726
- @Tianduo16 made their first contribution in PR #18723
- @wilsonjr made their first contribution in PR #18721
- @abliteration-ai made their first contribution in PR #18678
- @danialkhan02 made their first contribution in PR #18770
- @ihower made their first contribution in PR #18409
- @elkkhan made their first contribution in PR #18391
- @runixer made their first contribution in PR #18435
- @choby-shun made their first contribution in PR #18776
- @jutaz made their first contribution in PR #18853
- @sjmatta made their first contribution in PR #18250
- @andres-ortizl made their first contribution in PR #18856
- @gauthiermartin made their first contribution in PR #18844
- @mel2oo made their first contribution in PR #18845
- @DominikHallab made their first contribution in PR #18846
- @ji-chuan-che made their first contribution in PR #18540
- @raghav-stripe made their first contribution in PR #18858
- @akraines made their first contribution in PR #18629
- @otaviofbrito made their first contribution in PR #18665
- @chetanchoudhary-sumo made their first contribution in PR #18587
- @pascalwhoop made their first contribution in PR #13328
- @orgersh92 made their first contribution in PR #18652
- @DevajMody made their first contribution in PR #18497
- @matt-greathouse made their first contribution in PR #18247
- @emerzon made their first contribution in PR #18290
- @Eric84626 made their first contribution in PR #18281
- @LukasdeBoer made their first contribution in PR #18055
- @LingXuanYin made their first contribution in PR #18513
- @krisxia0506 made their first contribution in PR #18698
- @LouisShark made their first contribution in PR #18414

