This commit is contained in:
Ishaan Jaffer
2025-12-05 15:39:22 -08:00
parent 2ffe8ee204
commit ae065525ea
2 changed files with 81 additions and 0 deletions

View File

@@ -520,6 +520,7 @@ perplexity_models: Set = set()
watsonx_models: Set = set()
gemini_models: Set = set()
xai_models: Set = set()
zai_models: Set = set()
deepseek_models: Set = set()
runwayml_models: Set = set()
azure_ai_models: Set = set()
@@ -711,6 +712,8 @@ def add_known_models():
text_completion_codestral_models.add(key)
elif value.get("litellm_provider") == "xai":
xai_models.add(key)
elif value.get("litellm_provider") == "zai":
zai_models.add(key)
elif value.get("litellm_provider") == "fal_ai":
fal_ai_models.add(key)
elif value.get("litellm_provider") == "deepseek":
@@ -872,6 +875,7 @@ model_list = list(
| gemini_models
| text_completion_codestral_models
| xai_models
| zai_models
| fal_ai_models
| deepseek_models
| azure_ai_models
@@ -960,6 +964,7 @@ models_by_provider: dict = {
"aleph_alpha": aleph_alpha_models,
"text-completion-codestral": text_completion_codestral_models,
"xai": xai_models,
"zai": zai_models,
"fal_ai": fal_ai_models,
"deepseek": deepseek_models,
"runwayml": runwayml_models,
@@ -1497,10 +1502,58 @@ def set_global_gitlab_config(config: Dict[str, Any]) -> None:
# Lazy loading system for heavy modules to reduce initial import time and memory usage
if TYPE_CHECKING:
from litellm.types.utils import ModelInfo
# Cost calculator functions
cost_per_token: Callable[..., Tuple[float, float]]
completion_cost: Callable[..., float]
response_cost_calculator: Any
modify_integration: Any
# Utils functions - type stubs for lazy loaded functions
exception_type: Callable[..., Any]
get_optional_params: Callable[..., dict]
get_response_string: Callable[..., str]
token_counter: Callable[..., int]
create_pretrained_tokenizer: Callable[..., Any]
create_tokenizer: Callable[..., Any]
supports_function_calling: Callable[..., bool]
supports_web_search: Callable[..., bool]
supports_url_context: Callable[..., bool]
supports_response_schema: Callable[..., bool]
supports_parallel_function_calling: Callable[..., bool]
supports_vision: Callable[..., bool]
supports_audio_input: Callable[..., bool]
supports_audio_output: Callable[..., bool]
supports_system_messages: Callable[..., bool]
supports_reasoning: Callable[..., bool]
get_litellm_params: Callable[..., dict]
acreate: Callable[..., Any]
get_max_tokens: Callable[..., int]
get_model_info: Callable[..., ModelInfo]
register_prompt_template: Callable[..., None]
validate_environment: Callable[..., dict]
check_valid_key: Callable[..., bool]
register_model: Callable[..., None]
encode: Callable[..., list]
decode: Callable[..., str]
_calculate_retry_after: Callable[..., float]
_should_retry: Callable[[int], bool]
get_supported_openai_params: Callable[..., Optional[list]]
get_api_base: Callable[..., Optional[str]]
get_first_chars_messages: Callable[..., str]
get_provider_fields: Callable[..., dict]
get_valid_models: Callable[..., list]
# Response types - lazy loaded
ModelResponse: Type[Any]
ModelResponseStream: Type[Any]
EmbeddingResponse: Type[Any]
ImageResponse: Type[Any]
TranscriptionResponse: Type[Any]
TextCompletionResponse: Type[Any]
ModelResponseListIterator: Type[Any]
Logging: Type[Any]
def __getattr__(name: str) -> Any:

View File

@@ -5164,6 +5164,19 @@
"supports_function_calling": true,
"supports_tool_choice": true
},
"azure_ai/mistral-large-3": {
"input_cost_per_token": 5e-07,
"litellm_provider": "azure_ai",
"max_input_tokens": 256000,
"max_output_tokens": 8191,
"max_tokens": 8191,
"mode": "chat",
"output_cost_per_token": 1.5e-06,
"source": "https://azure.microsoft.com/en-us/blog/introducing-mistral-large-3-in-microsoft-foundry-open-capable-and-ready-for-production-workloads/",
"supports_function_calling": true,
"supports_tool_choice": true,
"supports_vision": true
},
"azure_ai/mistral-medium-2505": {
"input_cost_per_token": 4e-07,
"litellm_provider": "azure_ai",
@@ -18745,6 +18758,21 @@
"supports_response_schema": true,
"supports_tool_choice": true
},
"mistral/mistral-large-3": {
"input_cost_per_token": 5e-07,
"litellm_provider": "mistral",
"max_input_tokens": 256000,
"max_output_tokens": 8191,
"max_tokens": 8191,
"mode": "chat",
"output_cost_per_token": 1.5e-06,
"source": "https://docs.mistral.ai/models/mistral-large-3-25-12",
"supports_assistant_prefill": true,
"supports_function_calling": true,
"supports_response_schema": true,
"supports_tool_choice": true,
"supports_vision": true
},
"mistral/mistral-medium": {
"input_cost_per_token": 2.7e-06,
"litellm_provider": "mistral",