Continue browsing this topic cluster with SEO-safe static pagination.
LiteLLM LiteLLM Updated May 15, 2026
Fix LiteLLM showing zero costs for Anthropic batch API requests Includes evidence for LiteLLM troubleshooting demand.
Anthropic batch costs always 0 — transform_file_content_request routes msgbatch_* IDs to wrong endpoint LiteLLM LiteLLM Updated May 15, 2026
Fix LiteLLM Azure OpenAI authentication failure after upgrading to 1.84.0 Includes evidence for LiteLLM troubleshooting demand.
Azure OAI Models Auth Broken in LiteLLM 1.84.0 LiteLLM LiteLLM Updated May 15, 2026
Fix LiteLLM streaming fallback sending unsupported assistant prefill to Claude models causing 400 error Includes evidence for LiteLLM troubleshooting demand.
HTTP 400: This model does not support assistant message prefill. The conversation must end with a user message. LiteLLM LiteLLM Updated May 15, 2026
Fix LiteLLM crashing with nacl.exceptions.ValueError nonce must be exactly 24 bytes when configuring SMTP Includes evidence for LiteLLM troubleshooting demand.
nacl.exceptions.ValueError: The nonce must be exactly 24 bytes long LiteLLM LiteLLM Updated May 14, 2026
Fix LiteLLM Azure OpenAI auth broken after upgrade to 1.84.0 Includes evidence for LiteLLM troubleshooting demand.
litellm.AuthenticationError: AzureException AuthenticationError - The api_key client option must be set LiteLLM LiteLLM Updated May 14, 2026
Fix LiteLLM max_parallel_requests counter not decrementing on cancelled streams Includes evidence for LiteLLM troubleshooting demand.
max_parallel_requests Current limit: N, Remaining: 0 LiteLLM LiteLLM Updated May 14, 2026
Fix LiteLLM not capturing reasoning/thinking content from Ollama models like Qwen3 and DeepSeek-R1 Includes evidence for LiteLLM troubleshooting demand.
LiteLLM Ollama reasoning_content always null — /api/generate doesn't return thinking field LiteLLM LiteLLM / Anthropic API Updated May 14, 2026
Fix LiteLLM reasoning_content missing after Anthropic to OpenAI format conversion Includes evidence for LiteLLM / Anthropic API troubleshooting demand.
LiteLLM Anthropic to OpenAI conversion drops reasoning_content, breaks multi-turn with reasoning models LiteLLM LiteLLM / Anthropic API Updated May 14, 2026
Fix LiteLLM cache_control not forwarding to Anthropic when using Responses API Includes evidence for LiteLLM / Anthropic API troubleshooting demand.
LiteLLM Responses API silently drops cache_control on input_text content blocks during transformation LiteLLM LiteLLM Updated May 14, 2026
Fix LiteLLM Azure OAI authentication error after upgrading to 1.84.0 when using enable_azure_ad_token_refresh Includes evidence for LiteLLM troubleshooting demand.
LiteLLM AuthenticationError: AzureException AuthenticationError - The api_key client option must be set either by passing api_key to the client or by setting the OPENAI_API_KEY environment variable LiteLLM LiteLLM Updated May 14, 2026
Fix LiteLLM global_max_parallel_requests not enforcing the configured limit, allowing unlimited concurrent requests through Includes evidence for LiteLLM troubleshooting demand.
set global_max_parallel_requests not work LiteLLM LiteLLM Updated May 14, 2026
Fix LiteLLM 401 Unauthorized when querying single model info via /v1/models/{model} Includes evidence for LiteLLM troubleshooting demand.
Only proxy admin can be used to generate, delete, update info for new keys/users/teams. Route=/v1/models/{model} LiteLLM LiteLLM Updated May 14, 2026
Fix LiteLLM A2A /a2a/{id}/message/send returning -32603 Stream consumed error Includes evidence for LiteLLM troubleshooting demand.
-32603 Internal error: Stream consumed LiteLLM LiteLLM Updated May 14, 2026
Fix LiteLLM budget enforcement blocking free model discovery API access Includes evidence for LiteLLM troubleshooting demand.
Budget enforcement blocks model discovery endpoints (allow accessing free models) LiteLLM LiteLLM Updated May 14, 2026
Fix LiteLLM ghost models not being cleared when using --num_workers > 1 Includes evidence for LiteLLM troubleshooting demand.
Ghost models with --num_workers > 1: Deleted models not cleared from other workers' local cache LiteLLM LiteLLM Updated May 14, 2026
Fix LiteLLM not returning Retry-After header when all deployments are in cooldown Includes evidence for LiteLLM troubleshooting demand.
No Retry-After header on RouterRateLimitError (all deployments in cooldown) LiteLLM LiteLLM Updated May 14, 2026
Fix LiteLLM proxy error 'context_management: Extra inputs are not permitted' when using Claude Code with Bedrock InvokeModel Includes evidence for LiteLLM troubleshooting demand.
context_management: Extra inputs are not permitted LiteLLM LiteLLM Updated May 14, 2026
Fix LiteLLM /metrics endpoint returning 401 unauthorized after upgrading to 1.84.0 Includes evidence for LiteLLM troubleshooting demand.
LiteLLM /metrics endpoint returning 401 after upgrade to 1.84.0 LiteLLM LiteLLM Updated May 14, 2026
Fix LiteLLM virtual keys not enforcing TPM rate limits correctly Includes evidence for LiteLLM troubleshooting demand.
Incorrect TPM limiting for virtual keys — rate limit not enforced correctly LiteLLM LiteLLM Updated May 14, 2026
Fix LiteLLM passthrough to AWS Bedrock failing when using Anthropic's /v1/messages endpoint with context_management feature Includes evidence for LiteLLM troubleshooting demand.
LiteLLM /v1/messages passthrough to Bedrock InvokeModel fails — missing compact-2026-01-12 beta header