28312 Commits

Author SHA1 Message Date
yuneng-jiang
5439f03bfc bump: version 0.4.9 → 0.4.10 2025-12-04 12:56:43 -08:00
yuneng-jiang
2e65c464ad Adding tests 2025-12-04 12:36:15 -08:00
yuneng-jiang
562afb208d v0 customer usage, pending tests + extras version bump 2025-12-04 12:30:08 -08:00
Krrish Dholakia
5aeba81538 docs(multi_tenant_architecture.md): add new architecture doc 2025-12-04 11:13:50 -08:00
Ishaan Jaff
8ee8cad4f1 feat - new model amazon.titan-image-generator-v2:0 (#17488) 2025-12-04 09:26:55 -08:00
Sameer Kankute
31f1bf0539 Merge pull request #17485 from BerriAI/litellm_ragflow_test_fix
fix: test_default_api_base for ragfow
2025-12-04 21:57:18 +05:30
Sameer Kankute
f2c0029939 Merge pull request #17470 from BerriAI/litellm_batches_bedrock_content
Add support for file content download for bedrock batches
2025-12-04 21:57:04 +05:30
Sameer Kankute
5b4542304d Merge pull request #17461 from BerriAI/litellm_qwen2_imported_model_support
Add support for bedrock qwen 2 imported model
2025-12-04 21:56:22 +05:30
Sameer Kankute
1bd4050921 Merge pull request #17458 from BerriAI/litellm_map_output_tokens_details
Map output_tokens_details of responses API to completion_tokens_details
2025-12-04 21:55:21 +05:30
Sameer Kankute
6751badf3a fix: test_default_api_base for ragfow 2025-12-04 21:49:33 +05:30
Sameer Kankute
46ebf425d5 Fix : test_vertexai_model_garden_model_completion 2025-12-04 21:41:02 +05:30
Sameer Kankute
edd392b50d Add support for file content download for bedrock batches 2025-12-04 13:27:53 +05:30
yuneng-jiang
94b944e8ec Merge pull request #17465 from BerriAI/litellm_ui_model_page_perf_2
[Infra] UI - Change Credentials to use React Query
2025-12-03 23:53:40 -08:00
yuneng-jiang
ad86dfbf7d Merge pull request #17451 from BerriAI/litellm_new_login_page_sso_changes
[Feature] Add Auto Redirect to SSO to New Login Page
2025-12-03 23:53:21 -08:00
yuneng-jiang
5496e622c1 Unit tests 2025-12-03 23:46:09 -08:00
yuneng-jiang
59b32c28a8 Change selector in test 2025-12-03 23:37:15 -08:00
yuneng-jiang
0483336c53 change test to follow other logins 2025-12-03 23:21:09 -08:00
yuneng-jiang
19dc5d750c Adding waits 2025-12-03 23:10:32 -08:00
yuneng-jiang
45a9723467 view_user_info debugging 2025-12-03 22:20:04 -08:00
yuneng-jiang
b8190c6b6d resolving build issues 2025-12-03 22:15:19 -08:00
yuneng-jiang
7519a3e30b Change credentials to use react-query 2025-12-03 22:08:41 -08:00
Krish Dholakia
dc7c2b9b05 Update docs to link agent hub (#17462)
* Docs: Add AI Hub agent registry documentation

Co-authored-by: krrishdholakia <krrishdholakia@gmail.com>

* Fix: Update AI Hub link in A2A documentation

Co-authored-by: krrishdholakia <krrishdholakia@gmail.com>

---------

Co-authored-by: Cursor Agent <cursoragent@cursor.com>
2025-12-03 21:59:45 -08:00
yuneng-jiang
c634c08818 Find inputs by placeholder 2025-12-03 21:51:45 -08:00
Sameer Kankute
4710e772be Add support for bedrock qwen 2 imported model 2025-12-04 11:08:57 +05:30
Daniel Klein
3134fc825b fix: make bedrock converse messages respect modify_params as expected (#17427) 2025-12-03 21:34:50 -08:00
Cesar Garcia
664b384dce fix(pricing): add missing Moonshot turbo models and fix incorrect pricing (#17432)
Fixes #17417

- Add moonshot/kimi-k2-turbo-preview ($1.15/$8.00 input/output)
- Add moonshot/kimi-k2-thinking-turbo ($1.15/$8.00 input/output)
- Add moonshot/kimi-k2-0905-preview ($0.60/$2.50 input/output, 262K context)
- Fix moonshot/kimi-thinking-preview pricing ($30/$30 -> $0.60/$2.50)
- Fix wandb/moonshotai/Kimi-K2-Instruct absurd pricing ($135K/$400K -> $0.60/$2.50)
2025-12-03 21:31:49 -08:00
Cesar Garcia
b05f6d3cfd fix(groq): check supports_response_schema before using json_tool_call workaround (#17438)
Fixes issue where LiteLLM used the json_tool_call workaround for all Groq
models, even those that support native json_schema (e.g., gpt-oss-120b,
llama-4, kimi-k2). This caused errors when users passed their own tools
alongside structured outputs.

Changes:
- Check `supports_response_schema()` before using the workaround
- Only use json_tool_call workaround for models without native support
- Add clear error message when using workaround with user-provided tools
- Update model config to set `supports_response_schema: false` for models
  that don't support native json_schema
- Add unit tests for structured output handling
2025-12-03 21:31:13 -08:00
Cesar Garcia
6d831ffff4 fix(ocr): add missing ocr and aocr to CallTypes enum (#17435)
Add `ocr` and `aocr` entries to the CallTypes enum to fix the
ValueError that occurs when using the /v1/ocr endpoint with
guardrails enabled.

The OCR endpoint uses route_type="aocr", but the CallTypes enum
was missing these values, causing guardrail hooks to fail when
trying to instantiate CallTypes("aocr").

Fixes #17381
2025-12-03 21:28:13 -08:00
Andrew Truong
1879623502 feat: Basic Weave OTEL integration (#17439)
* test

* testfix
2025-12-03 21:26:22 -08:00
Haiyi
06d42fbd30 Fix: Skip empty text blocks in Anthropic system messages (#17442)
When translating system messages for the Anthropic API, empty text
content blocks cause the error "messages: text content blocks must be
non-empty". This fix skips empty string content and empty text blocks
in list content to prevent this error.

Fixes issue with Vertex AI Anthropic API calls.

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-authored-by: Claude <noreply@anthropic.com>
2025-12-03 21:25:06 -08:00
Devaj Mody
a6ce118989 fix(openai): add clear error message for empty LLM endpoint responses (#17445)
Fixes #17209

  - Add validation after response.parse() to detect empty/invalid responses
  - Raise clear OpenAIError with actionable message
  - Add 3 unit tests
2025-12-03 21:23:09 -08:00
codgician
adfbb1c308 docs: document responses and embedding api for github copilot (#17456) 2025-12-03 21:22:08 -08:00
yuneng-jiang
636f26d8a1 e2e changes 2025-12-03 21:21:27 -08:00
Krish Dholakia
32013f63a0 Guardrail API - support tool call checks on OpenAI /chat/completions, OpenAI /responses, Anthropic /v1/messages (#17459)
* fix(unified_guardrail.py): correctly map a v1/messages call to the anthropic unified guardrail

* fix: add more rigorous call type checks

* fix(anthropic_endpoints/endpoints.py): initialize logging object at the beginning of endpoint

ensures call id + trace id are emitted to guardrail api

* feat(anthropic/chat/guardrail_translation): support streaming guardrails

sample on every 5 chunks

* fix(openai/chat/guardrail_translation): support openai streaming guardrails

* fix: initial commit fixing output guardrails for responses api

* feat(openai/responses/guardrail_translation): handler.py - fix output checks on responses api

* fix(openai/responses/guardrail_translation/handler.py): ensure responses api guardrails work on streaming

* test: update tests

* test: update tests

* fix: support multiple kinds of input to the guardrail api

* feat(guardrail_translation/handler.py): support extracting tool calls from openai chat completions for guardrail api's

* feat(generic_guardrail_api.py): support extracting + returning modified tool calls on generic_guardrails_api

allows guardrail api to analyze tool call being sent to provider - to run any analysis on it

* fix(guardrails.py): support anthropic /v1/messages tool calls

* feat(responses_api/): extract tool calls for guardrail processing

* docs(generic_guardrail_api.md): document tools param support

* docs: generic_guardrail_api.md

improve documentation
2025-12-03 21:20:39 -08:00
yuneng-jiang
a4341ccf83 ci/cd changes for debugging 2025-12-03 21:00:49 -08:00
Krish Dholakia
be0530a6b3 fix(unified_guardrail.py): correctly map a v1/messages call to the anthropic unified guardrail (#17424)
* fix(unified_guardrail.py): correctly map a v1/messages call to the anthropic unified guardrail

* fix: add more rigorous call type checks

* fix(anthropic_endpoints/endpoints.py): initialize logging object at the beginning of endpoint

ensures call id + trace id are emitted to guardrail api

* feat(anthropic/chat/guardrail_translation): support streaming guardrails

sample on every 5 chunks

* fix(openai/chat/guardrail_translation): support openai streaming guardrails

* fix: initial commit fixing output guardrails for responses api

* feat(openai/responses/guardrail_translation): handler.py - fix output checks on responses api

* fix(openai/responses/guardrail_translation/handler.py): ensure responses api guardrails work on streaming

* test: update tests

* test: update tests

* test: update tests

* fix(bedrock_guardrails.py): fix post call streaming iterator logic

* fix: fix return

* fix(bedrock_guardrails.py): fix
v1.80.7.dev.3
2025-12-03 20:54:56 -08:00
Sameer Kankute
2072dcefa8 Map output_tokens_details of responses API to completion_tokens_details 2025-12-04 10:20:45 +05:30
yuneng-jiang
3de84b3f8b e2e tests 2025-12-03 20:34:41 -08:00
Xander Song
a711b63b06 fix: clean arize-phoenix traces (#16611)
* fix

* clean

* clean

* clean

* clean

* clean

* lint
2025-12-03 19:50:20 -08:00
Ishaan Jaff
607d517c01 Revert "Bump mcp from 1.10.1 to 1.23.0 in /.circleci (#17363)" (#17457)
This reverts commit 462d423d86.
2025-12-03 19:44:24 -08:00
Ishaan Jaffer
8a824b7c17 fix mypy linting 2025-12-03 19:43:29 -08:00
Ishaan Jaffer
92cca5d515 test_cyberark_write_and_read_secret 2025-12-03 19:43:29 -08:00
Ishaan Jaffer
e7f066ea7d test_hosted_vllm_full_workflow 2025-12-03 19:43:29 -08:00
Sameer Kankute
50fc30644e Merge pull request #17311 from Chesars/fix/azure-gpt5.1-reasoning-effort-none
Fix: Allow reasoning_effort='none' for Azure gpt-5.1 models
2025-12-04 09:08:32 +05:30
Ishaan Jaff
b96657a083 [Feat] UI - Agent Gateway - allow testing agents on UI (#17455)
* init litellm A2a client

* simpler a2a client interface

* test a2a

* move a2a invoking tests

* test fix

* ensure a2a send message is tracked n logs

* rename tags

* add streaming handlng

* add a2a invocation

* add a2a invocation i cost calc

* test_a2a_logging_payload

* update invoke_agent_a2a

* test_invoke_agent_a2a_adds_litellm_data

* add A2a agent

* fix endpoints on A2a

* UI allow testing a2a endpoints
2025-12-03 18:57:57 -08:00
Ishaan Jaff
e4f954b354 [Docs] Agent Gateway (#17454)
* init litellm A2a client

* simpler a2a client interface

* test a2a

* move a2a invoking tests

* test fix

* ensure a2a send message is tracked n logs

* rename tags

* add streaming handlng

* add a2a invocation

* add a2a invocation i cost calc

* test_a2a_logging_payload

* update invoke_agent_a2a

* test_invoke_agent_a2a_adds_litellm_data

* add A2a agent

* fix endpoints on A2a

* UI allow testing a2a endpoints

* add agent imgs

* add a2a as an endpoint

* add a2a

* docs a2a invoke

* docs a2a

* docs A2a invoke
2025-12-03 18:57:41 -08:00
Ishaan Jaff
585aee2ae4 [Feat] Agent Gateway - Allow tracking request / response in "Logs" Page (#17449)
* init litellm A2a client

* simpler a2a client interface

* test a2a

* move a2a invoking tests

* test fix

* ensure a2a send message is tracked n logs

* rename tags

* add streaming handlng

* add a2a invocation

* add a2a invocation i cost calc

* test_a2a_logging_payload

* update invoke_agent_a2a

* test_invoke_agent_a2a_adds_litellm_data

* add A2a agent
2025-12-03 18:57:18 -08:00
Ishaan Jaff
4370f6fb74 [Feat] Agent Gateway - Allow invoking agents through AI Gateway (#17440)
* init litellm A2a client

* simpler a2a client interface

* test a2a

* move a2a invoking tests

* test fix

* ensure a2a send message is tracked n logs

* rename tags

* add streaming handlng

* add a2a invocation
2025-12-03 18:54:55 -08:00
Ishaan Jaff
f035984dd7 fix: cyberark allow setting ssl verfiy to false (#17433) 2025-12-03 18:54:31 -08:00
yuneng-jiang
e29acb2f64 Merge pull request #16850 from BerriAI/litellm_user_info_404
[Fix] Return 404 when a user is not found on /user/info
2025-12-03 17:15:55 -08:00