Skip to main content

The LiteLLM Blog

Guides, announcements, and best practices from the LiteLLM team.

Latest

Realtime WebRTC HTTP Endpoints

Use the LiteLLM proxy to route OpenAI-style WebRTC realtime via HTTP: client_secrets and SDP exchange.

realtimewebrtcproxy

Day 0 Support: GPT-5.4

GPT-5.4 model support in LiteLLM

openaigpt-5.4completion

DAY 0 Support: Gemini 3.1 Flash Lite Preview on LiteLLM

Guide to using Gemini 3.1 Flash Lite Preview on LiteLLM Proxy and SDK with day 0 support.

geminiday 0 supportllms

Incident Report: Cache Eviction Closes In-Use httpx Clients

Date: February 27, 2026

incident-reportcachingstability

Day 0 Support: GPT-5.3-Codex

Day 0 support for GPT-5.3-Codex on LiteLLM, including phase parameter handling for Responses API.

openaigpt-5.3-codexcodex

Incident Report: Encrypted Content Failures in Multi-Region Responses API Load Balancing

Date: Feb 24, 2026

incident-reportproxyresponses-api

Incident Report: Wildcard Blocking New Models After Cost Map Reload

Date: Feb 23, 2026

incident-reportproxyauth

Incident Report: SERVER_ROOT_PATH regression broke UI routing

Date: January 22, 2026

incident-reportuistability

DAY 0 Support: Gemini 3.1 Pro on LiteLLM

Guide to using Gemini 3.1 Pro on LiteLLM Proxy and SDK with day 0 support.

geminiday 0 supportllms

Incident Report: vLLM Embeddings Broken by encoding_format Parameter

Date: Feb 16, 2026

incident-reportembeddingsvllm