v1.61.20-stable
These are the changes since v1.61.13-stable.
This release is primarily focused on:
- LLM Translation improvements (claude-3-7-sonnet + 'thinking'/'reasoning_content' support)
 - UI improvements (add model flow, user management, etc)
 
Demo Instance​
Here's a Demo Instance to test changes:
- Instance: https://demo.litellm.ai/
 - Login Credentials:
- Username: admin
 - Password: sk-1234
 
 
New Models / Updated Models​
- Anthropic 3-7 sonnet support + cost tracking (Anthropic API + Bedrock + Vertex AI + OpenRouter)
- Anthropic API Start here
 - Bedrock API Start here
 - Vertex AI API See here
 - OpenRouter See here
 
 - Gpt-4.5-preview support + cost tracking See here
 - Azure AI - Phi-4 cost tracking See here
 - Claude-3.5-sonnet - vision support updated on Anthropic API See here
 - Bedrock llama vision support See here
 - Cerebras llama3.3-70b pricing See here
 
LLM Translation​
- Infinity Rerank - support returning documents when return_documents=True Start here
 - Amazon Deepseek - 
<think>param extraction into ‘reasoning_content’ Start here - Amazon Titan Embeddings - filter out ‘aws_’ params from request body Start here
 - Anthropic ‘thinking’ + ‘reasoning_content’ translation support (Anthropic API, Bedrock, Vertex AI) Start here
 - VLLM - support ‘video_url’ Start here
 - Call proxy via litellm SDK: Support 
litellm_proxy/for embedding, image_generation, transcription, speech, rerank Start here - OpenAI Pass-through - allow using Assistants GET, DELETE on /openai pass through routes Start here
 - Message Translation - fix openai message for assistant msg if role is missing - openai allows this
 - O1/O3 - support ‘drop_params’ for o3-mini and o1 parallel_tool_calls param (not supported currently) See here
 
Spend Tracking Improvements​
- Cost tracking for rerank via Bedrock See PR
 - Anthropic pass-through - fix race condition causing cost to not be tracked See PR
 - Anthropic pass-through: Ensure accurate token counting See PR
 
Management Endpoints / UI​
- Models Page - Allow sorting models by ‘created at’
 - Models Page - Edit Model Flow Improvements
 - Models Page - Fix Adding Azure, Azure AI Studio models on UI
 - Internal Users Page - Allow Bulk Adding Internal Users on UI
 - Internal Users Page - Allow sorting users by ‘created at’
 - Virtual Keys Page - Allow searching for UserIDs on the dropdown when assigning a user to a team See PR
 - Virtual Keys Page - allow creating a user when assigning keys to users See PR
 - Model Hub Page - fix text overflow issue See PR
 - Admin Settings Page - Allow adding MSFT SSO on UI
 - Backend - don't allow creating duplicate internal users in DB
 
Helm​
- support ttlSecondsAfterFinished on the migration job - See PR
 - enhance migrations job with additional configurable properties - See PR
 
Logging / Guardrail Integrations​
- Arize Phoenix support
 - ‘No-log’ - fix ‘no-log’ param support on embedding calls
 
Performance / Loadbalancing / Reliability improvements​
- Single Deployment Cooldown logic - Use allowed_fails or allowed_fail_policy if set Start here
 
General Proxy Improvements​
- Hypercorn - fix reading / parsing request body
 - Windows - fix running proxy in windows
 - DD-Trace - fix dd-trace enablement on proxy
 
Complete Git Diff​
View the complete git diff here.