From 375a80f1ed3f3002820d52ab969fa9a184b23709 Mon Sep 17 00:00:00 2001 From: Ammar Date: Mon, 24 Nov 2025 14:23:59 -0600 Subject: [PATCH 1/2] =?UTF-8?q?=F0=9F=A4=96=20chore:=20update=20models.jso?= =?UTF-8?q?n=20from=20LiteLLM?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit _Generated with `mux`_ --- src/common/utils/tokens/models.json | 9432 ++++++++++++++++++--------- 1 file changed, 6264 insertions(+), 3168 deletions(-) diff --git a/src/common/utils/tokens/models.json b/src/common/utils/tokens/models.json index a0142b393b..5400a23151 100644 --- a/src/common/utils/tokens/models.json +++ b/src/common/utils/tokens/models.json @@ -1,26 +1,32 @@ { "sample_spec": { - "code_interpreter_cost_per_session": 0.0, - "computer_use_input_cost_per_1k_tokens": 0.0, - "computer_use_output_cost_per_1k_tokens": 0.0, + "code_interpreter_cost_per_session": 0, + "computer_use_input_cost_per_1k_tokens": 0, + "computer_use_output_cost_per_1k_tokens": 0, "deprecation_date": "date when the model becomes deprecated in the format YYYY-MM-DD", - "file_search_cost_per_1k_calls": 0.0, - "file_search_cost_per_gb_per_day": 0.0, - "input_cost_per_audio_token": 0.0, - "input_cost_per_token": 0.0, + "file_search_cost_per_1k_calls": 0, + "file_search_cost_per_gb_per_day": 0, + "input_cost_per_audio_token": 0, + "input_cost_per_token": 0, "litellm_provider": "one of https://docs.litellm.ai/docs/providers", "max_input_tokens": "max input tokens, if the provider specifies it. if not default to max_tokens", "max_output_tokens": "max output tokens, if the provider specifies it. if not default to max_tokens", "max_tokens": "LEGACY parameter. set to max_output_tokens if provider specifies it. IF not set to max_input_tokens, if provider specifies it.", "mode": "one of: chat, embedding, completion, image_generation, audio_transcription, audio_speech, image_generation, moderation, rerank, search", - "output_cost_per_reasoning_token": 0.0, - "output_cost_per_token": 0.0, + "output_cost_per_reasoning_token": 0, + "output_cost_per_token": 0, "search_context_cost_per_query": { - "search_context_size_high": 0.0, - "search_context_size_low": 0.0, - "search_context_size_medium": 0.0 + "search_context_size_high": 0, + "search_context_size_low": 0, + "search_context_size_medium": 0 }, - "supported_regions": ["global", "us-west-2", "eu-west-1", "ap-southeast-1", "ap-northeast-1"], + "supported_regions": [ + "global", + "us-west-2", + "eu-west-1", + "ap-southeast-1", + "ap-northeast-1" + ], "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -31,7 +37,7 @@ "supports_system_messages": true, "supports_vision": true, "supports_web_search": true, - "vector_store_cost_per_gb_per_day": 0.0 + "vector_store_cost_per_gb_per_day": 0 }, "1024-x-1024/50-steps/bedrock/amazon.nova-canvas-v1:0": { "litellm_provider": "bedrock", @@ -50,7 +56,7 @@ "input_cost_per_pixel": 1.9e-8, "litellm_provider": "openai", "mode": "image_generation", - "output_cost_per_pixel": 0.0 + "output_cost_per_pixel": 0 }, "1024-x-1024/max-steps/stability.stable-diffusion-xl-v1": { "litellm_provider": "bedrock", @@ -63,7 +69,7 @@ "input_cost_per_pixel": 2.4414e-7, "litellm_provider": "openai", "mode": "image_generation", - "output_cost_per_pixel": 0.0 + "output_cost_per_pixel": 0 }, "512-x-512/50-steps/stability.stable-diffusion-xl-v0": { "litellm_provider": "bedrock", @@ -76,7 +82,7 @@ "input_cost_per_pixel": 6.86e-8, "litellm_provider": "openai", "mode": "image_generation", - "output_cost_per_pixel": 0.0 + "output_cost_per_pixel": 0 }, "512-x-512/max-steps/stability.stable-diffusion-xl-v0": { "litellm_provider": "bedrock", @@ -86,31 +92,31 @@ "output_cost_per_image": 0.036 }, "ai21.j2-mid-v1": { - "input_cost_per_token": 1.25e-5, + "input_cost_per_token": 0.0000125, "litellm_provider": "bedrock", "max_input_tokens": 8191, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 1.25e-5 + "output_cost_per_token": 0.0000125 }, "ai21.j2-ultra-v1": { - "input_cost_per_token": 1.88e-5, + "input_cost_per_token": 0.0000188, "litellm_provider": "bedrock", "max_input_tokens": 8191, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 1.88e-5 + "output_cost_per_token": 0.0000188 }, "ai21.jamba-1-5-large-v1:0": { - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "bedrock", "max_input_tokens": 256000, "max_output_tokens": 256000, "max_tokens": 256000, "mode": "chat", - "output_cost_per_token": 8e-6 + "output_cost_per_token": 0.000008 }, "ai21.jamba-1-5-mini-v1:0": { "input_cost_per_token": 2e-7, @@ -139,7 +145,9 @@ "mode": "image_generation", "output_cost_per_image": 0.021, "source": "https://docs.aimlapi.com/", - "supported_endpoints": ["/v1/images/generations"] + "supported_endpoints": [ + "/v1/images/generations" + ] }, "aiml/dall-e-3": { "litellm_provider": "aiml", @@ -149,7 +157,9 @@ "mode": "image_generation", "output_cost_per_image": 0.042, "source": "https://docs.aimlapi.com/", - "supported_endpoints": ["/v1/images/generations"] + "supported_endpoints": [ + "/v1/images/generations" + ] }, "aiml/flux-pro": { "litellm_provider": "aiml", @@ -159,19 +169,25 @@ "mode": "image_generation", "output_cost_per_image": 0.053, "source": "https://docs.aimlapi.com/", - "supported_endpoints": ["/v1/images/generations"] + "supported_endpoints": [ + "/v1/images/generations" + ] }, "aiml/flux-pro/v1.1": { "litellm_provider": "aiml", "mode": "image_generation", "output_cost_per_image": 0.042, - "supported_endpoints": ["/v1/images/generations"] + "supported_endpoints": [ + "/v1/images/generations" + ] }, "aiml/flux-pro/v1.1-ultra": { "litellm_provider": "aiml", "mode": "image_generation", "output_cost_per_image": 0.063, - "supported_endpoints": ["/v1/images/generations"] + "supported_endpoints": [ + "/v1/images/generations" + ] }, "aiml/flux-realism": { "litellm_provider": "aiml", @@ -181,7 +197,9 @@ "mode": "image_generation", "output_cost_per_image": 0.037, "source": "https://docs.aimlapi.com/", - "supported_endpoints": ["/v1/images/generations"] + "supported_endpoints": [ + "/v1/images/generations" + ] }, "aiml/flux/dev": { "litellm_provider": "aiml", @@ -191,7 +209,9 @@ "mode": "image_generation", "output_cost_per_image": 0.026, "source": "https://docs.aimlapi.com/", - "supported_endpoints": ["/v1/images/generations"] + "supported_endpoints": [ + "/v1/images/generations" + ] }, "aiml/flux/kontext-max/text-to-image": { "litellm_provider": "aiml", @@ -201,7 +221,9 @@ "mode": "image_generation", "output_cost_per_image": 0.084, "source": "https://docs.aimlapi.com/", - "supported_endpoints": ["/v1/images/generations"] + "supported_endpoints": [ + "/v1/images/generations" + ] }, "aiml/flux/kontext-pro/text-to-image": { "litellm_provider": "aiml", @@ -211,7 +233,9 @@ "mode": "image_generation", "output_cost_per_image": 0.042, "source": "https://docs.aimlapi.com/", - "supported_endpoints": ["/v1/images/generations"] + "supported_endpoints": [ + "/v1/images/generations" + ] }, "aiml/flux/schnell": { "litellm_provider": "aiml", @@ -221,7 +245,9 @@ "mode": "image_generation", "output_cost_per_image": 0.003, "source": "https://docs.aimlapi.com/", - "supported_endpoints": ["/v1/images/generations"] + "supported_endpoints": [ + "/v1/images/generations" + ] }, "amazon.nova-lite-v1:0": { "input_cost_per_token": 6e-8, @@ -256,7 +282,7 @@ "max_output_tokens": 10000, "max_tokens": 10000, "mode": "chat", - "output_cost_per_token": 3.2e-6, + "output_cost_per_token": 0.0000032, "supports_function_calling": true, "supports_pdf_input": true, "supports_prompt_caching": true, @@ -265,7 +291,7 @@ }, "amazon.rerank-v1:0": { "input_cost_per_query": 0.001, - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "bedrock", "max_document_chunks_per_query": 100, "max_input_tokens": 32000, @@ -274,10 +300,10 @@ "max_tokens": 32000, "max_tokens_per_document_chunk": 512, "mode": "rerank", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "amazon.titan-embed-image-v1": { - "input_cost_per_image": 6e-5, + "input_cost_per_image": 0.00006, "input_cost_per_token": 8e-7, "litellm_provider": "bedrock", "max_input_tokens": 128, @@ -286,7 +312,7 @@ "notes": "'supports_image_input' is a deprecated field. Use 'supports_embedding_image_input' instead." }, "mode": "embedding", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "output_vector_size": 1024, "source": "https://us-east-1.console.aws.amazon.com/bedrock/home?region=us-east-1#/providers?model=amazon.titan-image-generator-v1", "supports_embedding_image_input": true, @@ -298,7 +324,7 @@ "max_input_tokens": 8192, "max_tokens": 8192, "mode": "embedding", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "output_vector_size": 1536 }, "amazon.titan-embed-text-v2:0": { @@ -307,11 +333,11 @@ "max_input_tokens": 8192, "max_tokens": 8192, "mode": "embedding", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "output_vector_size": 1024 }, "amazon.titan-image-generator-v1": { - "input_cost_per_image": 0.0, + "input_cost_per_image": 0, "output_cost_per_image": 0.008, "output_cost_per_image_premium_image": 0.01, "output_cost_per_image_above_512_and_512_pixels": 0.01, @@ -320,7 +346,7 @@ "mode": "image_generation" }, "amazon.titan-image-generator-v2": { - "input_cost_per_image": 0.0, + "input_cost_per_image": 0, "output_cost_per_image": 0.008, "output_cost_per_image_premium_image": 0.01, "output_cost_per_image_above_1024_and_1024_pixels": 0.01, @@ -329,18 +355,18 @@ "mode": "image_generation" }, "twelvelabs.marengo-embed-2-7-v1:0": { - "input_cost_per_token": 7e-5, + "input_cost_per_token": 0.00007, "litellm_provider": "bedrock", "max_input_tokens": 77, "max_tokens": 77, "mode": "embedding", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "output_vector_size": 1024, "supports_embedding_image_input": true, "supports_image_input": true }, "us.twelvelabs.marengo-embed-2-7-v1:0": { - "input_cost_per_token": 7e-5, + "input_cost_per_token": 0.00007, "input_cost_per_video_per_second": 0.0007, "input_cost_per_audio_per_second": 0.00014, "input_cost_per_image": 0.0001, @@ -348,13 +374,13 @@ "max_input_tokens": 77, "max_tokens": 77, "mode": "embedding", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "output_vector_size": 1024, "supports_embedding_image_input": true, "supports_image_input": true }, "eu.twelvelabs.marengo-embed-2-7-v1:0": { - "input_cost_per_token": 7e-5, + "input_cost_per_token": 0.00007, "input_cost_per_video_per_second": 0.0007, "input_cost_per_audio_per_second": 0.00014, "input_cost_per_image": 0.0001, @@ -362,40 +388,40 @@ "max_input_tokens": 77, "max_tokens": 77, "mode": "embedding", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "output_vector_size": 1024, "supports_embedding_image_input": true, "supports_image_input": true }, "twelvelabs.pegasus-1-2-v1:0": { "input_cost_per_video_per_second": 0.00049, - "output_cost_per_token": 7.5e-6, + "output_cost_per_token": 0.0000075, "litellm_provider": "bedrock", "mode": "chat", "supports_video_input": true }, "us.twelvelabs.pegasus-1-2-v1:0": { "input_cost_per_video_per_second": 0.00049, - "output_cost_per_token": 7.5e-6, + "output_cost_per_token": 0.0000075, "litellm_provider": "bedrock", "mode": "chat", "supports_video_input": true }, "eu.twelvelabs.pegasus-1-2-v1:0": { "input_cost_per_video_per_second": 0.00049, - "output_cost_per_token": 7.5e-6, + "output_cost_per_token": 0.0000075, "litellm_provider": "bedrock", "mode": "chat", "supports_video_input": true }, "amazon.titan-text-express-v1": { - "input_cost_per_token": 1.3e-6, + "input_cost_per_token": 0.0000013, "litellm_provider": "bedrock", "max_input_tokens": 42000, "max_output_tokens": 8000, "max_tokens": 8000, "mode": "chat", - "output_cost_per_token": 1.7e-6 + "output_cost_per_token": 0.0000017 }, "amazon.titan-text-lite-v1": { "input_cost_per_token": 3e-7, @@ -413,10 +439,10 @@ "max_output_tokens": 32000, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 1.5e-6 + "output_cost_per_token": 0.0000015 }, "anthropic.claude-3-5-haiku-20241022-v1:0": { - "cache_creation_input_token_cost": 1e-6, + "cache_creation_input_token_cost": 0.000001, "cache_read_input_token_cost": 8e-8, "input_cost_per_token": 8e-7, "litellm_provider": "bedrock", @@ -424,7 +450,7 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 4e-6, + "output_cost_per_token": 0.000004, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_pdf_input": true, @@ -433,51 +459,57 @@ "supports_tool_choice": true }, "anthropic.claude-haiku-4-5-20251001-v1:0": { - "cache_creation_input_token_cost": 1.25e-6, + "cache_creation_input_token_cost": 0.00000125, "cache_read_input_token_cost": 1e-7, - "input_cost_per_token": 1e-6, - "litellm_provider": "bedrock", + "input_cost_per_token": 0.000001, + "litellm_provider": "bedrock_converse", "max_input_tokens": 200000, - "max_output_tokens": 8192, - "max_tokens": 8192, + "max_output_tokens": 64000, + "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 5e-6, + "output_cost_per_token": 0.000005, "source": "https://aws.amazon.com/about-aws/whats-new/2025/10/claude-4-5-haiku-anthropic-amazon-bedrock", "supports_assistant_prefill": true, + "supports_computer_use": true, "supports_function_calling": true, "supports_pdf_input": true, "supports_prompt_caching": true, "supports_reasoning": true, "supports_response_schema": true, - "supports_tool_choice": true + "supports_tool_choice": true, + "supports_vision": true, + "tool_use_system_prompt_tokens": 346 }, "anthropic.claude-haiku-4-5@20251001": { - "cache_creation_input_token_cost": 1.25e-6, + "cache_creation_input_token_cost": 0.00000125, "cache_read_input_token_cost": 1e-7, - "input_cost_per_token": 1e-6, - "litellm_provider": "bedrock", + "input_cost_per_token": 0.000001, + "litellm_provider": "bedrock_converse", "max_input_tokens": 200000, - "max_output_tokens": 8192, - "max_tokens": 8192, + "max_output_tokens": 64000, + "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 5e-6, + "output_cost_per_token": 0.000005, "source": "https://aws.amazon.com/about-aws/whats-new/2025/10/claude-4-5-haiku-anthropic-amazon-bedrock", "supports_assistant_prefill": true, + "supports_computer_use": true, "supports_function_calling": true, "supports_pdf_input": true, "supports_prompt_caching": true, "supports_reasoning": true, "supports_response_schema": true, - "supports_tool_choice": true + "supports_tool_choice": true, + "supports_vision": true, + "tool_use_system_prompt_tokens": 346 }, "anthropic.claude-3-5-sonnet-20240620-v1:0": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "bedrock", "max_input_tokens": 200000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "supports_function_calling": true, "supports_pdf_input": true, "supports_response_schema": true, @@ -485,15 +517,15 @@ "supports_vision": true }, "anthropic.claude-3-5-sonnet-20241022-v2:0": { - "cache_creation_input_token_cost": 3.75e-6, + "cache_creation_input_token_cost": 0.00000375, "cache_read_input_token_cost": 3e-7, - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "bedrock", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "supports_assistant_prefill": true, "supports_computer_use": true, "supports_function_calling": true, @@ -504,15 +536,15 @@ "supports_vision": true }, "anthropic.claude-3-7-sonnet-20240620-v1:0": { - "cache_creation_input_token_cost": 4.5e-6, + "cache_creation_input_token_cost": 0.0000045, "cache_read_input_token_cost": 3.6e-7, - "input_cost_per_token": 3.6e-6, + "input_cost_per_token": 0.0000036, "litellm_provider": "bedrock", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 1.8e-5, + "output_cost_per_token": 0.000018, "supports_assistant_prefill": true, "supports_computer_use": true, "supports_function_calling": true, @@ -524,15 +556,15 @@ "supports_vision": true }, "anthropic.claude-3-7-sonnet-20250219-v1:0": { - "cache_creation_input_token_cost": 3.75e-6, + "cache_creation_input_token_cost": 0.00000375, "cache_read_input_token_cost": 3e-7, - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "bedrock_converse", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "supports_assistant_prefill": true, "supports_computer_use": true, "supports_function_calling": true, @@ -550,7 +582,7 @@ "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 1.25e-6, + "output_cost_per_token": 0.00000125, "supports_function_calling": true, "supports_pdf_input": true, "supports_response_schema": true, @@ -558,26 +590,26 @@ "supports_vision": true }, "anthropic.claude-3-opus-20240229-v1:0": { - "input_cost_per_token": 1.5e-5, + "input_cost_per_token": 0.000015, "litellm_provider": "bedrock", "max_input_tokens": 200000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 7.5e-5, + "output_cost_per_token": 0.000075, "supports_function_calling": true, "supports_response_schema": true, "supports_tool_choice": true, "supports_vision": true }, "anthropic.claude-3-sonnet-20240229-v1:0": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "bedrock", "max_input_tokens": 200000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "supports_function_calling": true, "supports_pdf_input": true, "supports_response_schema": true, @@ -591,19 +623,19 @@ "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 2.4e-6, + "output_cost_per_token": 0.0000024, "supports_tool_choice": true }, "anthropic.claude-opus-4-1-20250805-v1:0": { - "cache_creation_input_token_cost": 1.875e-5, - "cache_read_input_token_cost": 1.5e-6, - "input_cost_per_token": 1.5e-5, + "cache_creation_input_token_cost": 0.00001875, + "cache_read_input_token_cost": 0.0000015, + "input_cost_per_token": 0.000015, "litellm_provider": "bedrock_converse", "max_input_tokens": 200000, "max_output_tokens": 32000, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 7.5e-5, + "output_cost_per_token": 0.000075, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -621,15 +653,15 @@ "tool_use_system_prompt_tokens": 159 }, "anthropic.claude-opus-4-20250514-v1:0": { - "cache_creation_input_token_cost": 1.875e-5, - "cache_read_input_token_cost": 1.5e-6, - "input_cost_per_token": 1.5e-5, + "cache_creation_input_token_cost": 0.00001875, + "cache_read_input_token_cost": 0.0000015, + "input_cost_per_token": 0.000015, "litellm_provider": "bedrock_converse", "max_input_tokens": 200000, "max_output_tokens": 32000, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 7.5e-5, + "output_cost_per_token": 0.000075, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -647,19 +679,49 @@ "tool_use_system_prompt_tokens": 159 }, "anthropic.claude-sonnet-4-20250514-v1:0": { - "cache_creation_input_token_cost": 3.75e-6, + "cache_creation_input_token_cost": 0.00000375, "cache_read_input_token_cost": 3e-7, - "input_cost_per_token": 3e-6, - "input_cost_per_token_above_200k_tokens": 6e-6, - "output_cost_per_token_above_200k_tokens": 2.25e-5, - "cache_creation_input_token_cost_above_200k_tokens": 7.5e-6, + "input_cost_per_token": 0.000003, + "input_cost_per_token_above_200k_tokens": 0.000006, + "output_cost_per_token_above_200k_tokens": 0.0000225, + "cache_creation_input_token_cost_above_200k_tokens": 0.0000075, "cache_read_input_token_cost_above_200k_tokens": 6e-7, "litellm_provider": "bedrock_converse", "max_input_tokens": 1000000, "max_output_tokens": 64000, "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, + "search_context_cost_per_query": { + "search_context_size_high": 0.01, + "search_context_size_low": 0.01, + "search_context_size_medium": 0.01 + }, + "supports_assistant_prefill": true, + "supports_computer_use": true, + "supports_function_calling": true, + "supports_pdf_input": true, + "supports_prompt_caching": true, + "supports_reasoning": true, + "supports_response_schema": true, + "supports_tool_choice": true, + "supports_vision": true, + "tool_use_system_prompt_tokens": 159 + }, + "anthropic.claude-sonnet-4-5-20250929-v1:0": { + "cache_creation_input_token_cost": 0.00000375, + "cache_read_input_token_cost": 3e-7, + "input_cost_per_token": 0.000003, + "input_cost_per_token_above_200k_tokens": 0.000006, + "output_cost_per_token_above_200k_tokens": 0.0000225, + "cache_creation_input_token_cost_above_200k_tokens": 0.0000075, + "cache_read_input_token_cost_above_200k_tokens": 6e-7, + "litellm_provider": "bedrock_converse", + "max_input_tokens": 200000, + "max_output_tokens": 64000, + "max_tokens": 64000, + "mode": "chat", + "output_cost_per_token": 0.000015, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -677,22 +739,22 @@ "tool_use_system_prompt_tokens": 159 }, "anthropic.claude-v1": { - "input_cost_per_token": 8e-6, + "input_cost_per_token": 0.000008, "litellm_provider": "bedrock", "max_input_tokens": 100000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 2.4e-5 + "output_cost_per_token": 0.000024 }, "anthropic.claude-v2:1": { - "input_cost_per_token": 8e-6, + "input_cost_per_token": 0.000008, "litellm_provider": "bedrock", "max_input_tokens": 100000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 2.4e-5, + "output_cost_per_token": 0.000024, "supports_tool_choice": true }, "anyscale/HuggingFaceH4/zephyr-7b-beta": { @@ -705,22 +767,22 @@ "output_cost_per_token": 1.5e-7 }, "anyscale/codellama/CodeLlama-34b-Instruct-hf": { - "input_cost_per_token": 1e-6, + "input_cost_per_token": 0.000001, "litellm_provider": "anyscale", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 1e-6 + "output_cost_per_token": 0.000001 }, "anyscale/codellama/CodeLlama-70b-Instruct-hf": { - "input_cost_per_token": 1e-6, + "input_cost_per_token": 0.000001, "litellm_provider": "anyscale", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 1e-6, + "output_cost_per_token": 0.000001, "source": "https://docs.anyscale.com/preview/endpoints/text-generation/supported-models/codellama-CodeLlama-70b-Instruct-hf" }, "anyscale/google/gemma-7b-it": { @@ -743,13 +805,13 @@ "output_cost_per_token": 2.5e-7 }, "anyscale/meta-llama/Llama-2-70b-chat-hf": { - "input_cost_per_token": 1e-6, + "input_cost_per_token": 0.000001, "litellm_provider": "anyscale", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 1e-6 + "output_cost_per_token": 0.000001 }, "anyscale/meta-llama/Llama-2-7b-chat-hf": { "input_cost_per_token": 1.5e-7, @@ -761,13 +823,13 @@ "output_cost_per_token": 1.5e-7 }, "anyscale/meta-llama/Meta-Llama-3-70B-Instruct": { - "input_cost_per_token": 1e-6, + "input_cost_per_token": 0.000001, "litellm_provider": "anyscale", "max_input_tokens": 8192, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 1e-6, + "output_cost_per_token": 0.000001, "source": "https://docs.anyscale.com/preview/endpoints/text-generation/supported-models/meta-llama-Meta-Llama-3-70B-Instruct" }, "anyscale/meta-llama/Meta-Llama-3-8B-Instruct": { @@ -846,7 +908,7 @@ "max_output_tokens": 10000, "max_tokens": 10000, "mode": "chat", - "output_cost_per_token": 3.36e-6, + "output_cost_per_token": 0.00000336, "supports_function_calling": true, "supports_pdf_input": true, "supports_prompt_caching": true, @@ -854,13 +916,13 @@ "supports_vision": true }, "apac.anthropic.claude-3-5-sonnet-20240620-v1:0": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "bedrock", "max_input_tokens": 200000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "supports_function_calling": true, "supports_pdf_input": true, "supports_response_schema": true, @@ -868,15 +930,15 @@ "supports_vision": true }, "apac.anthropic.claude-3-5-sonnet-20241022-v2:0": { - "cache_creation_input_token_cost": 3.75e-6, + "cache_creation_input_token_cost": 0.00000375, "cache_read_input_token_cost": 3e-7, - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "bedrock", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "supports_assistant_prefill": true, "supports_computer_use": true, "supports_function_calling": true, @@ -893,7 +955,7 @@ "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 1.25e-6, + "output_cost_per_token": 0.00000125, "supports_function_calling": true, "supports_pdf_input": true, "supports_response_schema": true, @@ -901,32 +963,35 @@ "supports_vision": true }, "apac.anthropic.claude-haiku-4-5-20251001-v1:0": { - "cache_creation_input_token_cost": 1.375e-6, + "cache_creation_input_token_cost": 0.000001375, "cache_read_input_token_cost": 1.1e-7, - "input_cost_per_token": 1.1e-6, - "litellm_provider": "bedrock", + "input_cost_per_token": 0.0000011, + "litellm_provider": "bedrock_converse", "max_input_tokens": 200000, - "max_output_tokens": 8192, - "max_tokens": 8192, + "max_output_tokens": 64000, + "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 5.5e-6, + "output_cost_per_token": 0.0000055, "source": "https://aws.amazon.com/about-aws/whats-new/2025/10/claude-4-5-haiku-anthropic-amazon-bedrock", "supports_assistant_prefill": true, + "supports_computer_use": true, "supports_function_calling": true, "supports_pdf_input": true, "supports_prompt_caching": true, "supports_reasoning": true, "supports_response_schema": true, - "supports_tool_choice": true + "supports_tool_choice": true, + "supports_vision": true, + "tool_use_system_prompt_tokens": 346 }, "apac.anthropic.claude-3-sonnet-20240229-v1:0": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "bedrock", "max_input_tokens": 200000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "supports_function_calling": true, "supports_pdf_input": true, "supports_response_schema": true, @@ -934,19 +999,19 @@ "supports_vision": true }, "apac.anthropic.claude-sonnet-4-20250514-v1:0": { - "cache_creation_input_token_cost": 3.75e-6, + "cache_creation_input_token_cost": 0.00000375, "cache_read_input_token_cost": 3e-7, - "input_cost_per_token": 3e-6, - "input_cost_per_token_above_200k_tokens": 6e-6, - "output_cost_per_token_above_200k_tokens": 2.25e-5, - "cache_creation_input_token_cost_above_200k_tokens": 7.5e-6, + "input_cost_per_token": 0.000003, + "input_cost_per_token_above_200k_tokens": 0.000006, + "output_cost_per_token_above_200k_tokens": 0.0000225, + "cache_creation_input_token_cost_above_200k_tokens": 0.0000075, "cache_read_input_token_cost_above_200k_tokens": 6e-7, "litellm_provider": "bedrock_converse", "max_input_tokens": 1000000, "max_output_tokens": 64000, "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -964,31 +1029,31 @@ "tool_use_system_prompt_tokens": 159 }, "assemblyai/best": { - "input_cost_per_second": 3.333e-5, + "input_cost_per_second": 0.00003333, "litellm_provider": "assemblyai", "mode": "audio_transcription", - "output_cost_per_second": 0.0 + "output_cost_per_second": 0 }, "assemblyai/nano": { "input_cost_per_second": 0.00010278, "litellm_provider": "assemblyai", "mode": "audio_transcription", - "output_cost_per_second": 0.0 + "output_cost_per_second": 0 }, "au.anthropic.claude-sonnet-4-5-20250929-v1:0": { - "cache_creation_input_token_cost": 4.125e-6, + "cache_creation_input_token_cost": 0.000004125, "cache_read_input_token_cost": 3.3e-7, - "input_cost_per_token": 3.3e-6, - "input_cost_per_token_above_200k_tokens": 6.6e-6, - "output_cost_per_token_above_200k_tokens": 2.475e-5, - "cache_creation_input_token_cost_above_200k_tokens": 8.25e-6, + "input_cost_per_token": 0.0000033, + "input_cost_per_token_above_200k_tokens": 0.0000066, + "output_cost_per_token_above_200k_tokens": 0.00002475, + "cache_creation_input_token_cost_above_200k_tokens": 0.00000825, "cache_read_input_token_cost_above_200k_tokens": 6.6e-7, "litellm_provider": "bedrock_converse", "max_input_tokens": 200000, "max_output_tokens": 64000, "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 1.65e-5, + "output_cost_per_token": 0.0000165, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -1011,20 +1076,27 @@ "max_input_tokens": 8191, "max_tokens": 8191, "mode": "embedding", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "azure/codex-mini": { "cache_read_input_token_cost": 3.75e-7, - "input_cost_per_token": 1.5e-6, + "input_cost_per_token": 0.0000015, "litellm_provider": "azure", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "responses", - "output_cost_per_token": 6e-6, - "supported_endpoints": ["/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "output_cost_per_token": 0.000006, + "supported_endpoints": [ + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_pdf_input": true, @@ -1036,26 +1108,33 @@ "supports_vision": true }, "azure/command-r-plus": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "supports_function_calling": true }, "azure/computer-use-preview": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "azure", "max_input_tokens": 8192, "max_output_tokens": 1024, "max_tokens": 1024, "mode": "chat", - "output_cost_per_token": 1.2e-5, - "supported_endpoints": ["/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "output_cost_per_token": 0.000012, + "supported_endpoints": [ + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": false, @@ -1072,14 +1151,14 @@ }, "azure/eu/gpt-4o-2024-08-06": { "deprecation_date": "2026-02-27", - "cache_read_input_token_cost": 1.375e-6, - "input_cost_per_token": 2.75e-6, + "cache_read_input_token_cost": 0.000001375, + "input_cost_per_token": 0.00000275, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 1.1e-5, + "output_cost_per_token": 0.000011, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -1089,14 +1168,14 @@ }, "azure/eu/gpt-4o-2024-11-20": { "deprecation_date": "2026-03-01", - "cache_creation_input_token_cost": 1.38e-6, - "input_cost_per_token": 2.75e-6, + "cache_creation_input_token_cost": 0.00000138, + "input_cost_per_token": 0.00000275, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 1.1e-5, + "output_cost_per_token": 0.000011, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_response_schema": true, @@ -1122,15 +1201,15 @@ "azure/eu/gpt-4o-mini-realtime-preview-2024-12-17": { "cache_creation_input_audio_token_cost": 3.3e-7, "cache_read_input_token_cost": 3.3e-7, - "input_cost_per_audio_token": 1.1e-5, + "input_cost_per_audio_token": 0.000011, "input_cost_per_token": 6.6e-7, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_audio_token": 2.2e-5, - "output_cost_per_token": 2.64e-6, + "output_cost_per_audio_token": 0.000022, + "output_cost_per_token": 0.00000264, "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -1139,17 +1218,17 @@ "supports_tool_choice": true }, "azure/eu/gpt-4o-realtime-preview-2024-10-01": { - "cache_creation_input_audio_token_cost": 2.2e-5, - "cache_read_input_token_cost": 2.75e-6, + "cache_creation_input_audio_token_cost": 0.000022, + "cache_read_input_token_cost": 0.00000275, "input_cost_per_audio_token": 0.00011, - "input_cost_per_token": 5.5e-6, + "input_cost_per_token": 0.0000055, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", "output_cost_per_audio_token": 0.00022, - "output_cost_per_token": 2.2e-5, + "output_cost_per_token": 0.000022, "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -1158,19 +1237,25 @@ "supports_tool_choice": true }, "azure/eu/gpt-4o-realtime-preview-2024-12-17": { - "cache_read_input_audio_token_cost": 2.5e-6, - "cache_read_input_token_cost": 2.75e-6, - "input_cost_per_audio_token": 4.4e-5, - "input_cost_per_token": 5.5e-6, + "cache_read_input_audio_token_cost": 0.0000025, + "cache_read_input_token_cost": 0.00000275, + "input_cost_per_audio_token": 0.000044, + "input_cost_per_token": 0.0000055, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_audio_token": 8e-5, - "output_cost_per_token": 2.2e-5, - "supported_modalities": ["text", "audio"], - "supported_output_modalities": ["text", "audio"], + "output_cost_per_audio_token": 0.00008, + "output_cost_per_token": 0.000022, + "supported_modalities": [ + "text", + "audio" + ], + "supported_output_modalities": [ + "text", + "audio" + ], "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -1180,16 +1265,25 @@ }, "azure/eu/gpt-5-2025-08-07": { "cache_read_input_token_cost": 1.375e-7, - "input_cost_per_token": 1.375e-6, + "input_cost_per_token": 0.000001375, "litellm_provider": "azure", "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 1.1e-5, - "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "output_cost_per_token": 0.000011, + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -1209,10 +1303,85 @@ "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 2.2e-6, - "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "output_cost_per_token": 0.0000022, + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], + "supports_function_calling": true, + "supports_native_streaming": true, + "supports_parallel_function_calling": true, + "supports_pdf_input": true, + "supports_prompt_caching": true, + "supports_reasoning": true, + "supports_response_schema": true, + "supports_system_messages": true, + "supports_tool_choice": true, + "supports_vision": true + }, + "azure/eu/gpt-5.1": { + "cache_read_input_token_cost": 1.4e-7, + "input_cost_per_token": 0.00000138, + "litellm_provider": "azure", + "max_input_tokens": 272000, + "max_output_tokens": 128000, + "max_tokens": 128000, + "mode": "chat", + "output_cost_per_token": 0.000011, + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text", + "image" + ], + "supports_function_calling": true, + "supports_native_streaming": true, + "supports_parallel_function_calling": true, + "supports_pdf_input": true, + "supports_prompt_caching": true, + "supports_reasoning": true, + "supports_response_schema": true, + "supports_system_messages": true, + "supports_tool_choice": true, + "supports_vision": true + }, + "azure/eu/gpt-5.1-chat": { + "cache_read_input_token_cost": 1.4e-7, + "input_cost_per_token": 0.00000138, + "litellm_provider": "azure", + "max_input_tokens": 272000, + "max_output_tokens": 128000, + "max_tokens": 128000, + "mode": "chat", + "output_cost_per_token": 0.000011, + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text", + "image" + ], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -1224,6 +1393,66 @@ "supports_tool_choice": true, "supports_vision": true }, + "azure/eu/gpt-5.1-codex": { + "cache_read_input_token_cost": 1.4e-7, + "input_cost_per_token": 0.00000138, + "litellm_provider": "azure", + "max_input_tokens": 272000, + "max_output_tokens": 128000, + "max_tokens": 128000, + "mode": "responses", + "output_cost_per_token": 0.000011, + "supported_endpoints": [ + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], + "supports_function_calling": true, + "supports_native_streaming": true, + "supports_parallel_function_calling": true, + "supports_pdf_input": true, + "supports_prompt_caching": true, + "supports_reasoning": true, + "supports_response_schema": true, + "supports_system_messages": false, + "supports_tool_choice": true, + "supports_vision": true + }, + "azure/eu/gpt-5.1-codex-mini": { + "cache_read_input_token_cost": 2.8e-8, + "input_cost_per_token": 2.75e-7, + "litellm_provider": "azure", + "max_input_tokens": 272000, + "max_output_tokens": 128000, + "max_tokens": 128000, + "mode": "responses", + "output_cost_per_token": 0.0000022, + "supported_endpoints": [ + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], + "supports_function_calling": true, + "supports_native_streaming": true, + "supports_parallel_function_calling": true, + "supports_pdf_input": true, + "supports_prompt_caching": true, + "supports_reasoning": true, + "supports_response_schema": true, + "supports_system_messages": false, + "supports_tool_choice": true, + "supports_vision": true + }, "azure/eu/gpt-5-nano-2025-08-07": { "cache_read_input_token_cost": 5.5e-9, "input_cost_per_token": 5.5e-8, @@ -1233,9 +1462,18 @@ "max_tokens": 128000, "mode": "chat", "output_cost_per_token": 4.4e-7, - "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -1248,14 +1486,14 @@ "supports_vision": true }, "azure/eu/o1-2024-12-17": { - "cache_read_input_token_cost": 8.25e-6, - "input_cost_per_token": 1.65e-5, + "cache_read_input_token_cost": 0.00000825, + "input_cost_per_token": 0.0000165, "litellm_provider": "azure", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 6.6e-5, + "output_cost_per_token": 0.000066, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -1264,29 +1502,29 @@ }, "azure/eu/o1-mini-2024-09-12": { "cache_read_input_token_cost": 6.05e-7, - "input_cost_per_token": 1.21e-6, + "input_cost_per_token": 0.00000121, "input_cost_per_token_batches": 6.05e-7, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 65536, "max_tokens": 65536, "mode": "chat", - "output_cost_per_token": 4.84e-6, - "output_cost_per_token_batches": 2.42e-6, + "output_cost_per_token": 0.00000484, + "output_cost_per_token_batches": 0.00000242, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, "supports_vision": false }, "azure/eu/o1-preview-2024-09-12": { - "cache_read_input_token_cost": 8.25e-6, - "input_cost_per_token": 1.65e-5, + "cache_read_input_token_cost": 0.00000825, + "input_cost_per_token": 0.0000165, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 6.6e-5, + "output_cost_per_token": 0.000066, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -1294,30 +1532,30 @@ }, "azure/eu/o3-mini-2025-01-31": { "cache_read_input_token_cost": 6.05e-7, - "input_cost_per_token": 1.21e-6, + "input_cost_per_token": 0.00000121, "input_cost_per_token_batches": 6.05e-7, "litellm_provider": "azure", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 4.84e-6, - "output_cost_per_token_batches": 2.42e-6, + "output_cost_per_token": 0.00000484, + "output_cost_per_token_batches": 0.00000242, "supports_prompt_caching": true, "supports_reasoning": true, "supports_tool_choice": true, "supports_vision": false }, "azure/global-standard/gpt-4o-2024-08-06": { - "cache_read_input_token_cost": 1.25e-6, + "cache_read_input_token_cost": 0.00000125, "deprecation_date": "2026-02-27", - "input_cost_per_token": 2.5e-6, + "input_cost_per_token": 0.0000025, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 1e-5, + "output_cost_per_token": 0.00001, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -1326,15 +1564,15 @@ "supports_vision": true }, "azure/global-standard/gpt-4o-2024-11-20": { - "cache_read_input_token_cost": 1.25e-6, + "cache_read_input_token_cost": 0.00000125, "deprecation_date": "2026-03-01", - "input_cost_per_token": 2.5e-6, + "input_cost_per_token": 0.0000025, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 1e-5, + "output_cost_per_token": 0.00001, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_response_schema": true, @@ -1357,14 +1595,14 @@ }, "azure/global/gpt-4o-2024-08-06": { "deprecation_date": "2026-02-27", - "cache_read_input_token_cost": 1.25e-6, - "input_cost_per_token": 2.5e-6, + "cache_read_input_token_cost": 0.00000125, + "input_cost_per_token": 0.0000025, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 1e-5, + "output_cost_per_token": 0.00001, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -1374,14 +1612,14 @@ }, "azure/global/gpt-4o-2024-11-20": { "deprecation_date": "2026-03-01", - "cache_read_input_token_cost": 1.25e-6, - "input_cost_per_token": 2.5e-6, + "cache_read_input_token_cost": 0.00000125, + "input_cost_per_token": 0.0000025, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 1e-5, + "output_cost_per_token": 0.00001, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -1389,46 +1627,172 @@ "supports_tool_choice": true, "supports_vision": true }, - "azure/gpt-3.5-turbo": { - "input_cost_per_token": 5e-7, + "azure/global/gpt-5.1": { + "cache_read_input_token_cost": 1.25e-7, + "input_cost_per_token": 0.00000125, "litellm_provider": "azure", - "max_input_tokens": 4097, - "max_output_tokens": 4096, - "max_tokens": 4096, + "max_input_tokens": 272000, + "max_output_tokens": 128000, + "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 1.5e-6, + "output_cost_per_token": 0.00001, + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text", + "image" + ], "supports_function_calling": true, - "supports_tool_choice": true + "supports_native_streaming": true, + "supports_parallel_function_calling": true, + "supports_pdf_input": true, + "supports_prompt_caching": true, + "supports_reasoning": true, + "supports_response_schema": true, + "supports_system_messages": true, + "supports_tool_choice": true, + "supports_vision": true }, - "azure/gpt-3.5-turbo-0125": { - "deprecation_date": "2025-03-31", - "input_cost_per_token": 5e-7, + "azure/global/gpt-5.1-chat": { + "cache_read_input_token_cost": 1.25e-7, + "input_cost_per_token": 0.00000125, "litellm_provider": "azure", - "max_input_tokens": 16384, - "max_output_tokens": 4096, - "max_tokens": 4096, + "max_input_tokens": 272000, + "max_output_tokens": 128000, + "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 1.5e-6, + "output_cost_per_token": 0.00001, + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text", + "image" + ], "supports_function_calling": true, + "supports_native_streaming": true, "supports_parallel_function_calling": true, - "supports_tool_choice": true - }, - "azure/gpt-3.5-turbo-instruct-0914": { - "input_cost_per_token": 1.5e-6, - "litellm_provider": "azure_text", - "max_input_tokens": 4097, - "max_tokens": 4097, - "mode": "completion", - "output_cost_per_token": 2e-6 + "supports_pdf_input": true, + "supports_prompt_caching": true, + "supports_reasoning": true, + "supports_response_schema": true, + "supports_system_messages": true, + "supports_tool_choice": true, + "supports_vision": true }, - "azure/gpt-35-turbo": { - "input_cost_per_token": 5e-7, + "azure/global/gpt-5.1-codex": { + "cache_read_input_token_cost": 1.25e-7, + "input_cost_per_token": 0.00000125, "litellm_provider": "azure", - "max_input_tokens": 4097, - "max_output_tokens": 4096, - "max_tokens": 4096, + "max_input_tokens": 272000, + "max_output_tokens": 128000, + "max_tokens": 128000, + "mode": "responses", + "output_cost_per_token": 0.00001, + "supported_endpoints": [ + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], + "supports_function_calling": true, + "supports_native_streaming": true, + "supports_parallel_function_calling": true, + "supports_pdf_input": true, + "supports_prompt_caching": true, + "supports_reasoning": true, + "supports_response_schema": true, + "supports_system_messages": false, + "supports_tool_choice": true, + "supports_vision": true + }, + "azure/global/gpt-5.1-codex-mini": { + "cache_read_input_token_cost": 2.5e-8, + "input_cost_per_token": 2.5e-7, + "litellm_provider": "azure", + "max_input_tokens": 272000, + "max_output_tokens": 128000, + "max_tokens": 128000, + "mode": "responses", + "output_cost_per_token": 0.000002, + "supported_endpoints": [ + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], + "supports_function_calling": true, + "supports_native_streaming": true, + "supports_parallel_function_calling": true, + "supports_pdf_input": true, + "supports_prompt_caching": true, + "supports_reasoning": true, + "supports_response_schema": true, + "supports_system_messages": false, + "supports_tool_choice": true, + "supports_vision": true + }, + "azure/gpt-3.5-turbo": { + "input_cost_per_token": 5e-7, + "litellm_provider": "azure", + "max_input_tokens": 4097, + "max_output_tokens": 4096, + "max_tokens": 4096, + "mode": "chat", + "output_cost_per_token": 0.0000015, + "supports_function_calling": true, + "supports_tool_choice": true + }, + "azure/gpt-3.5-turbo-0125": { + "deprecation_date": "2025-03-31", + "input_cost_per_token": 5e-7, + "litellm_provider": "azure", + "max_input_tokens": 16384, + "max_output_tokens": 4096, + "max_tokens": 4096, + "mode": "chat", + "output_cost_per_token": 0.0000015, + "supports_function_calling": true, + "supports_parallel_function_calling": true, + "supports_tool_choice": true + }, + "azure/gpt-3.5-turbo-instruct-0914": { + "input_cost_per_token": 0.0000015, + "litellm_provider": "azure_text", + "max_input_tokens": 4097, + "max_tokens": 4097, + "mode": "completion", + "output_cost_per_token": 0.000002 + }, + "azure/gpt-35-turbo": { + "input_cost_per_token": 5e-7, + "litellm_provider": "azure", + "max_input_tokens": 4097, + "max_output_tokens": 4096, + "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 1.5e-6, + "output_cost_per_token": 0.0000015, "supports_function_calling": true, "supports_tool_choice": true }, @@ -1440,7 +1804,7 @@ "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 1.5e-6, + "output_cost_per_token": 0.0000015, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_tool_choice": true @@ -1453,122 +1817,122 @@ "max_output_tokens": 4096, "max_tokens": 4097, "mode": "chat", - "output_cost_per_token": 2e-6, + "output_cost_per_token": 0.000002, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_tool_choice": true }, "azure/gpt-35-turbo-0613": { "deprecation_date": "2025-02-13", - "input_cost_per_token": 1.5e-6, + "input_cost_per_token": 0.0000015, "litellm_provider": "azure", "max_input_tokens": 4097, "max_output_tokens": 4096, "max_tokens": 4097, "mode": "chat", - "output_cost_per_token": 2e-6, + "output_cost_per_token": 0.000002, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_tool_choice": true }, "azure/gpt-35-turbo-1106": { "deprecation_date": "2025-03-31", - "input_cost_per_token": 1e-6, + "input_cost_per_token": 0.000001, "litellm_provider": "azure", "max_input_tokens": 16384, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 2e-6, + "output_cost_per_token": 0.000002, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_tool_choice": true }, "azure/gpt-35-turbo-16k": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "azure", "max_input_tokens": 16385, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 4e-6, + "output_cost_per_token": 0.000004, "supports_tool_choice": true }, "azure/gpt-35-turbo-16k-0613": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "azure", "max_input_tokens": 16385, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 4e-6, + "output_cost_per_token": 0.000004, "supports_function_calling": true, "supports_tool_choice": true }, "azure/gpt-35-turbo-instruct": { - "input_cost_per_token": 1.5e-6, + "input_cost_per_token": 0.0000015, "litellm_provider": "azure_text", "max_input_tokens": 4097, "max_tokens": 4097, "mode": "completion", - "output_cost_per_token": 2e-6 + "output_cost_per_token": 0.000002 }, "azure/gpt-35-turbo-instruct-0914": { - "input_cost_per_token": 1.5e-6, + "input_cost_per_token": 0.0000015, "litellm_provider": "azure_text", "max_input_tokens": 4097, "max_tokens": 4097, "mode": "completion", - "output_cost_per_token": 2e-6 + "output_cost_per_token": 0.000002 }, "azure/gpt-4": { - "input_cost_per_token": 3e-5, + "input_cost_per_token": 0.00003, "litellm_provider": "azure", "max_input_tokens": 8192, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 6e-5, + "output_cost_per_token": 0.00006, "supports_function_calling": true, "supports_tool_choice": true }, "azure/gpt-4-0125-preview": { - "input_cost_per_token": 1e-5, + "input_cost_per_token": 0.00001, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 3e-5, + "output_cost_per_token": 0.00003, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_tool_choice": true }, "azure/gpt-4-0613": { - "input_cost_per_token": 3e-5, + "input_cost_per_token": 0.00003, "litellm_provider": "azure", "max_input_tokens": 8192, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 6e-5, + "output_cost_per_token": 0.00006, "supports_function_calling": true, "supports_tool_choice": true }, "azure/gpt-4-1106-preview": { - "input_cost_per_token": 1e-5, + "input_cost_per_token": 0.00001, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 3e-5, + "output_cost_per_token": 0.00003, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_tool_choice": true }, "azure/gpt-4-32k": { - "input_cost_per_token": 6e-5, + "input_cost_per_token": 0.00006, "litellm_provider": "azure", "max_input_tokens": 32768, "max_output_tokens": 4096, @@ -1578,7 +1942,7 @@ "supports_tool_choice": true }, "azure/gpt-4-32k-0613": { - "input_cost_per_token": 6e-5, + "input_cost_per_token": 0.00006, "litellm_provider": "azure", "max_input_tokens": 32768, "max_output_tokens": 4096, @@ -1588,55 +1952,64 @@ "supports_tool_choice": true }, "azure/gpt-4-turbo": { - "input_cost_per_token": 1e-5, + "input_cost_per_token": 0.00001, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 3e-5, + "output_cost_per_token": 0.00003, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_tool_choice": true }, "azure/gpt-4-turbo-2024-04-09": { - "input_cost_per_token": 1e-5, + "input_cost_per_token": 0.00001, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 3e-5, + "output_cost_per_token": 0.00003, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_tool_choice": true, "supports_vision": true }, "azure/gpt-4-turbo-vision-preview": { - "input_cost_per_token": 1e-5, + "input_cost_per_token": 0.00001, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 3e-5, + "output_cost_per_token": 0.00003, "supports_tool_choice": true, "supports_vision": true }, "azure/gpt-4.1": { "cache_read_input_token_cost": 5e-7, - "input_cost_per_token": 2e-6, - "input_cost_per_token_batches": 1e-6, + "input_cost_per_token": 0.000002, + "input_cost_per_token_batches": 0.000001, "litellm_provider": "azure", "max_input_tokens": 1047576, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 8e-6, - "output_cost_per_token_batches": 4e-6, - "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "output_cost_per_token": 0.000008, + "output_cost_per_token_batches": 0.000004, + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -1650,18 +2023,27 @@ "azure/gpt-4.1-2025-04-14": { "deprecation_date": "2026-11-04", "cache_read_input_token_cost": 5e-7, - "input_cost_per_token": 2e-6, - "input_cost_per_token_batches": 1e-6, + "input_cost_per_token": 0.000002, + "input_cost_per_token_batches": 0.000001, "litellm_provider": "azure", "max_input_tokens": 1047576, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 8e-6, - "output_cost_per_token_batches": 4e-6, - "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "output_cost_per_token": 0.000008, + "output_cost_per_token_batches": 0.000004, + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -1681,11 +2063,20 @@ "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 1.6e-6, + "output_cost_per_token": 0.0000016, "output_cost_per_token_batches": 8e-7, - "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -1706,11 +2097,20 @@ "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 1.6e-6, + "output_cost_per_token": 0.0000016, "output_cost_per_token_batches": 8e-7, - "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -1732,9 +2132,18 @@ "mode": "chat", "output_cost_per_token": 4e-7, "output_cost_per_token_batches": 2e-7, - "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -1756,9 +2165,18 @@ "mode": "chat", "output_cost_per_token": 4e-7, "output_cost_per_token_batches": 2e-7, - "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -1769,16 +2187,16 @@ "supports_vision": true }, "azure/gpt-4.5-preview": { - "cache_read_input_token_cost": 3.75e-5, - "input_cost_per_token": 7.5e-5, - "input_cost_per_token_batches": 3.75e-5, + "cache_read_input_token_cost": 0.0000375, + "input_cost_per_token": 0.000075, + "input_cost_per_token_batches": 0.0000375, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", "output_cost_per_token": 0.00015, - "output_cost_per_token_batches": 7.5e-5, + "output_cost_per_token_batches": 0.000075, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -1788,14 +2206,14 @@ "supports_vision": true }, "azure/gpt-4o": { - "cache_read_input_token_cost": 1.25e-6, - "input_cost_per_token": 2.5e-6, + "cache_read_input_token_cost": 0.00000125, + "input_cost_per_token": 0.0000025, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 1e-5, + "output_cost_per_token": 0.00001, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -1804,13 +2222,13 @@ "supports_vision": true }, "azure/gpt-4o-2024-05-13": { - "input_cost_per_token": 5e-6, + "input_cost_per_token": 0.000005, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -1819,14 +2237,14 @@ }, "azure/gpt-4o-2024-08-06": { "deprecation_date": "2026-02-27", - "cache_read_input_token_cost": 1.25e-6, - "input_cost_per_token": 2.5e-6, + "cache_read_input_token_cost": 0.00000125, + "input_cost_per_token": 0.0000025, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 1e-5, + "output_cost_per_token": 0.00001, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -1836,14 +2254,14 @@ }, "azure/gpt-4o-2024-11-20": { "deprecation_date": "2026-03-01", - "cache_read_input_token_cost": 1.25e-6, - "input_cost_per_token": 2.75e-6, + "cache_read_input_token_cost": 0.00000125, + "input_cost_per_token": 0.00000275, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 1.1e-5, + "output_cost_per_token": 0.000011, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -1851,19 +2269,89 @@ "supports_tool_choice": true, "supports_vision": true }, + "azure/gpt-audio-2025-08-28": { + "input_cost_per_audio_token": 0.00004, + "input_cost_per_token": 0.0000025, + "litellm_provider": "azure", + "max_input_tokens": 128000, + "max_output_tokens": 16384, + "max_tokens": 16384, + "mode": "chat", + "output_cost_per_audio_token": 0.00008, + "output_cost_per_token": 0.00001, + "supported_endpoints": [ + "/v1/chat/completions" + ], + "supported_modalities": [ + "text", + "audio" + ], + "supported_output_modalities": [ + "text", + "audio" + ], + "supports_function_calling": true, + "supports_native_streaming": true, + "supports_parallel_function_calling": true, + "supports_prompt_caching": false, + "supports_reasoning": false, + "supports_response_schema": false, + "supports_system_messages": true, + "supports_tool_choice": true, + "supports_vision": false + }, + "azure/gpt-audio-mini-2025-10-06": { + "input_cost_per_audio_token": 0.00001, + "input_cost_per_token": 6e-7, + "litellm_provider": "azure", + "max_input_tokens": 128000, + "max_output_tokens": 16384, + "max_tokens": 16384, + "mode": "chat", + "output_cost_per_audio_token": 0.00002, + "output_cost_per_token": 0.0000024, + "supported_endpoints": [ + "/v1/chat/completions" + ], + "supported_modalities": [ + "text", + "audio" + ], + "supported_output_modalities": [ + "text", + "audio" + ], + "supports_function_calling": true, + "supports_native_streaming": true, + "supports_parallel_function_calling": true, + "supports_prompt_caching": false, + "supports_reasoning": false, + "supports_response_schema": false, + "supports_system_messages": true, + "supports_tool_choice": true, + "supports_vision": false + }, "azure/gpt-4o-audio-preview-2024-12-17": { - "input_cost_per_audio_token": 4e-5, - "input_cost_per_token": 2.5e-6, + "input_cost_per_audio_token": 0.00004, + "input_cost_per_token": 0.0000025, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_audio_token": 8e-5, - "output_cost_per_token": 1e-5, - "supported_endpoints": ["/v1/chat/completions"], - "supported_modalities": ["text", "audio"], - "supported_output_modalities": ["text", "audio"], + "output_cost_per_audio_token": 0.00008, + "output_cost_per_token": 0.00001, + "supported_endpoints": [ + "/v1/chat/completions" + ], + "supported_modalities": [ + "text", + "audio" + ], + "supported_output_modalities": [ + "text", + "audio" + ], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -1907,18 +2395,26 @@ "supports_vision": true }, "azure/gpt-4o-mini-audio-preview-2024-12-17": { - "input_cost_per_audio_token": 4e-5, - "input_cost_per_token": 2.5e-6, + "input_cost_per_audio_token": 0.00004, + "input_cost_per_token": 0.0000025, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_audio_token": 8e-5, - "output_cost_per_token": 1e-5, - "supported_endpoints": ["/v1/chat/completions"], - "supported_modalities": ["text", "audio"], - "supported_output_modalities": ["text", "audio"], + "output_cost_per_audio_token": 0.00008, + "output_cost_per_token": 0.00001, + "supported_endpoints": [ + "/v1/chat/completions" + ], + "supported_modalities": [ + "text", + "audio" + ], + "supported_output_modalities": [ + "text", + "audio" + ], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -1932,15 +2428,79 @@ "azure/gpt-4o-mini-realtime-preview-2024-12-17": { "cache_creation_input_audio_token_cost": 3e-7, "cache_read_input_token_cost": 3e-7, - "input_cost_per_audio_token": 1e-5, + "input_cost_per_audio_token": 0.00001, "input_cost_per_token": 6e-7, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_audio_token": 2e-5, - "output_cost_per_token": 2.4e-6, + "output_cost_per_audio_token": 0.00002, + "output_cost_per_token": 0.0000024, + "supports_audio_input": true, + "supports_audio_output": true, + "supports_function_calling": true, + "supports_parallel_function_calling": true, + "supports_system_messages": true, + "supports_tool_choice": true + }, + "azure/gpt-realtime-2025-08-28": { + "cache_creation_input_audio_token_cost": 0.000004, + "cache_read_input_token_cost": 0.000004, + "input_cost_per_audio_token": 0.000032, + "input_cost_per_image": 0.000005, + "input_cost_per_token": 0.000004, + "litellm_provider": "azure", + "max_input_tokens": 32000, + "max_output_tokens": 4096, + "max_tokens": 4096, + "mode": "chat", + "output_cost_per_audio_token": 0.000064, + "output_cost_per_token": 0.000016, + "supported_endpoints": [ + "/v1/realtime" + ], + "supported_modalities": [ + "text", + "image", + "audio" + ], + "supported_output_modalities": [ + "text", + "audio" + ], + "supports_audio_input": true, + "supports_audio_output": true, + "supports_function_calling": true, + "supports_parallel_function_calling": true, + "supports_system_messages": true, + "supports_tool_choice": true + }, + "azure/gpt-realtime-mini-2025-10-06": { + "cache_creation_input_audio_token_cost": 3e-7, + "cache_read_input_token_cost": 6e-8, + "input_cost_per_audio_token": 0.00001, + "input_cost_per_image": 8e-7, + "input_cost_per_token": 6e-7, + "litellm_provider": "azure", + "max_input_tokens": 32000, + "max_output_tokens": 4096, + "max_tokens": 4096, + "mode": "chat", + "output_cost_per_audio_token": 0.00002, + "output_cost_per_token": 0.0000024, + "supported_endpoints": [ + "/v1/realtime" + ], + "supported_modalities": [ + "text", + "image", + "audio" + ], + "supported_output_modalities": [ + "text", + "audio" + ], "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -1949,38 +2509,47 @@ "supports_tool_choice": true }, "azure/gpt-4o-mini-transcribe": { - "input_cost_per_audio_token": 3e-6, - "input_cost_per_token": 1.25e-6, + "input_cost_per_audio_token": 0.000003, + "input_cost_per_token": 0.00000125, "litellm_provider": "azure", "max_input_tokens": 16000, "max_output_tokens": 2000, "mode": "audio_transcription", - "output_cost_per_token": 5e-6, - "supported_endpoints": ["/v1/audio/transcriptions"] + "output_cost_per_token": 0.000005, + "supported_endpoints": [ + "/v1/audio/transcriptions" + ] }, "azure/gpt-4o-mini-tts": { - "input_cost_per_token": 2.5e-6, + "input_cost_per_token": 0.0000025, "litellm_provider": "azure", "mode": "audio_speech", - "output_cost_per_audio_token": 1.2e-5, + "output_cost_per_audio_token": 0.000012, "output_cost_per_second": 0.00025, - "output_cost_per_token": 1e-5, - "supported_endpoints": ["/v1/audio/speech"], - "supported_modalities": ["text", "audio"], - "supported_output_modalities": ["audio"] + "output_cost_per_token": 0.00001, + "supported_endpoints": [ + "/v1/audio/speech" + ], + "supported_modalities": [ + "text", + "audio" + ], + "supported_output_modalities": [ + "audio" + ] }, "azure/gpt-4o-realtime-preview-2024-10-01": { - "cache_creation_input_audio_token_cost": 2e-5, - "cache_read_input_token_cost": 2.5e-6, + "cache_creation_input_audio_token_cost": 0.00002, + "cache_read_input_token_cost": 0.0000025, "input_cost_per_audio_token": 0.0001, - "input_cost_per_token": 5e-6, + "input_cost_per_token": 0.000005, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", "output_cost_per_audio_token": 0.0002, - "output_cost_per_token": 2e-5, + "output_cost_per_token": 0.00002, "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -1989,18 +2558,24 @@ "supports_tool_choice": true }, "azure/gpt-4o-realtime-preview-2024-12-17": { - "cache_read_input_token_cost": 2.5e-6, - "input_cost_per_audio_token": 4e-5, - "input_cost_per_token": 5e-6, + "cache_read_input_token_cost": 0.0000025, + "input_cost_per_audio_token": 0.00004, + "input_cost_per_token": 0.000005, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_audio_token": 8e-5, - "output_cost_per_token": 2e-5, - "supported_modalities": ["text", "audio"], - "supported_output_modalities": ["text", "audio"], + "output_cost_per_audio_token": 0.00008, + "output_cost_per_token": 0.00002, + "supported_modalities": [ + "text", + "audio" + ], + "supported_output_modalities": [ + "text", + "audio" + ], "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -2009,27 +2584,187 @@ "supports_tool_choice": true }, "azure/gpt-4o-transcribe": { - "input_cost_per_audio_token": 6e-6, - "input_cost_per_token": 2.5e-6, + "input_cost_per_audio_token": 0.000006, + "input_cost_per_token": 0.0000025, + "litellm_provider": "azure", + "max_input_tokens": 16000, + "max_output_tokens": 2000, + "mode": "audio_transcription", + "output_cost_per_token": 0.00001, + "supported_endpoints": [ + "/v1/audio/transcriptions" + ] + }, + "azure/gpt-4o-transcribe-diarize": { + "input_cost_per_audio_token": 0.000006, + "input_cost_per_token": 0.0000025, "litellm_provider": "azure", "max_input_tokens": 16000, "max_output_tokens": 2000, "mode": "audio_transcription", - "output_cost_per_token": 1e-5, - "supported_endpoints": ["/v1/audio/transcriptions"] + "output_cost_per_token": 0.00001, + "supported_endpoints": [ + "/v1/audio/transcriptions" + ] + }, + "azure/gpt-5.1-2025-11-13": { + "cache_read_input_token_cost": 1.25e-7, + "cache_read_input_token_cost_priority": 2.5e-7, + "input_cost_per_token": 0.00000125, + "input_cost_per_token_priority": 0.0000025, + "litellm_provider": "azure", + "max_input_tokens": 272000, + "max_output_tokens": 128000, + "max_tokens": 128000, + "mode": "chat", + "output_cost_per_token": 0.00001, + "output_cost_per_token_priority": 0.00002, + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text", + "image" + ], + "supports_function_calling": true, + "supports_native_streaming": true, + "supports_parallel_function_calling": true, + "supports_pdf_input": true, + "supports_prompt_caching": true, + "supports_reasoning": true, + "supports_response_schema": true, + "supports_system_messages": true, + "supports_tool_choice": true, + "supports_service_tier": true, + "supports_vision": true + }, + "azure/gpt-5.1-chat-2025-11-13": { + "cache_read_input_token_cost": 1.25e-7, + "cache_read_input_token_cost_priority": 2.5e-7, + "input_cost_per_token": 0.00000125, + "input_cost_per_token_priority": 0.0000025, + "litellm_provider": "azure", + "max_input_tokens": 128000, + "max_output_tokens": 16384, + "max_tokens": 16384, + "mode": "chat", + "output_cost_per_token": 0.00001, + "output_cost_per_token_priority": 0.00002, + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text", + "image" + ], + "supports_function_calling": false, + "supports_native_streaming": true, + "supports_parallel_function_calling": false, + "supports_pdf_input": true, + "supports_prompt_caching": true, + "supports_reasoning": true, + "supports_response_schema": true, + "supports_system_messages": true, + "supports_tool_choice": false, + "supports_vision": true + }, + "azure/gpt-5.1-codex-2025-11-13": { + "cache_read_input_token_cost": 1.25e-7, + "cache_read_input_token_cost_priority": 2.5e-7, + "input_cost_per_token": 0.00000125, + "input_cost_per_token_priority": 0.0000025, + "litellm_provider": "azure", + "max_input_tokens": 272000, + "max_output_tokens": 128000, + "max_tokens": 128000, + "mode": "responses", + "output_cost_per_token": 0.00001, + "output_cost_per_token_priority": 0.00002, + "supported_endpoints": [ + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], + "supports_function_calling": true, + "supports_native_streaming": true, + "supports_parallel_function_calling": true, + "supports_pdf_input": true, + "supports_prompt_caching": true, + "supports_reasoning": true, + "supports_response_schema": true, + "supports_system_messages": false, + "supports_tool_choice": true, + "supports_vision": true + }, + "azure/gpt-5.1-codex-mini-2025-11-13": { + "cache_read_input_token_cost": 2.5e-8, + "cache_read_input_token_cost_priority": 4.5e-8, + "input_cost_per_token": 2.5e-7, + "input_cost_per_token_priority": 4.5e-7, + "litellm_provider": "azure", + "max_input_tokens": 272000, + "max_output_tokens": 128000, + "max_tokens": 128000, + "mode": "responses", + "output_cost_per_token": 0.000002, + "output_cost_per_token_priority": 0.0000036, + "supported_endpoints": [ + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], + "supports_function_calling": true, + "supports_native_streaming": true, + "supports_parallel_function_calling": true, + "supports_pdf_input": true, + "supports_prompt_caching": true, + "supports_reasoning": true, + "supports_response_schema": true, + "supports_system_messages": false, + "supports_tool_choice": true, + "supports_vision": true }, "azure/gpt-5": { "cache_read_input_token_cost": 1.25e-7, - "input_cost_per_token": 1.25e-6, + "input_cost_per_token": 0.00000125, "litellm_provider": "azure", "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 1e-5, - "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "output_cost_per_token": 0.00001, + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -2043,16 +2778,25 @@ }, "azure/gpt-5-2025-08-07": { "cache_read_input_token_cost": 1.25e-7, - "input_cost_per_token": 1.25e-6, + "input_cost_per_token": 0.00000125, "litellm_provider": "azure", "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 1e-5, - "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "output_cost_per_token": 0.00001, + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -2066,17 +2810,26 @@ }, "azure/gpt-5-chat": { "cache_read_input_token_cost": 1.25e-7, - "input_cost_per_token": 1.25e-6, + "input_cost_per_token": 0.00000125, "litellm_provider": "azure", "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 1e-5, + "output_cost_per_token": 0.00001, "source": "https://azure.microsoft.com/en-us/blog/gpt-5-in-azure-ai-foundry-the-future-of-ai-apps-and-agents-starts-here/", - "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -2090,16 +2843,25 @@ }, "azure/gpt-5-chat-latest": { "cache_read_input_token_cost": 1.25e-7, - "input_cost_per_token": 1.25e-6, + "input_cost_per_token": 0.00000125, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 1e-5, - "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "output_cost_per_token": 0.00001, + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -2113,16 +2875,23 @@ }, "azure/gpt-5-codex": { "cache_read_input_token_cost": 1.25e-7, - "input_cost_per_token": 1.25e-6, + "input_cost_per_token": 0.00000125, "litellm_provider": "azure", "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "responses", - "output_cost_per_token": 1e-5, - "supported_endpoints": ["/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "output_cost_per_token": 0.00001, + "supported_endpoints": [ + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -2142,10 +2911,19 @@ "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 2e-6, - "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "output_cost_per_token": 0.000002, + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -2165,10 +2943,19 @@ "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 2e-6, - "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "output_cost_per_token": 0.000002, + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -2189,9 +2976,18 @@ "max_tokens": 128000, "mode": "chat", "output_cost_per_token": 4e-7, - "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -2212,9 +3008,18 @@ "max_tokens": 128000, "mode": "chat", "output_cost_per_token": 4e-7, - "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -2227,7 +3032,7 @@ "supports_vision": true }, "azure/gpt-5-pro": { - "input_cost_per_token": 1.5e-5, + "input_cost_per_token": 0.000015, "litellm_provider": "azure", "max_input_tokens": 272000, "max_output_tokens": 128000, @@ -2235,10 +3040,83 @@ "mode": "responses", "output_cost_per_token": 0.00012, "source": "https://learn.microsoft.com/en-us/azure/ai-foundry/foundry-models/concepts/models-sold-directly-by-azure?pivots=azure-openai&tabs=global-standard-aoai%2Cstandard-chat-completions%2Cglobal-standard#gpt-5", - "supported_endpoints": ["/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "supported_endpoints": [ + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], + "supports_function_calling": true, + "supports_parallel_function_calling": true, + "supports_pdf_input": true, + "supports_prompt_caching": true, + "supports_reasoning": true, + "supports_response_schema": true, + "supports_system_messages": true, + "supports_tool_choice": true, + "supports_vision": true + }, + "azure/gpt-5.1": { + "cache_read_input_token_cost": 1.25e-7, + "input_cost_per_token": 0.00000125, + "litellm_provider": "azure", + "max_input_tokens": 272000, + "max_output_tokens": 128000, + "max_tokens": 128000, + "mode": "chat", + "output_cost_per_token": 0.00001, + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text", + "image" + ], + "supports_function_calling": true, + "supports_native_streaming": true, + "supports_parallel_function_calling": true, + "supports_pdf_input": true, + "supports_prompt_caching": true, + "supports_reasoning": true, + "supports_response_schema": true, + "supports_system_messages": true, + "supports_tool_choice": true, + "supports_vision": true + }, + "azure/gpt-5.1-chat": { + "cache_read_input_token_cost": 1.25e-7, + "input_cost_per_token": 0.00000125, + "litellm_provider": "azure", + "max_input_tokens": 272000, + "max_output_tokens": 128000, + "max_tokens": 128000, + "mode": "chat", + "output_cost_per_token": 0.00001, + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text", + "image" + ], "supports_function_calling": true, + "supports_native_streaming": true, "supports_parallel_function_calling": true, "supports_pdf_input": true, "supports_prompt_caching": true, @@ -2248,191 +3126,291 @@ "supports_tool_choice": true, "supports_vision": true }, + "azure/gpt-5.1-codex": { + "cache_read_input_token_cost": 1.25e-7, + "input_cost_per_token": 0.00000125, + "litellm_provider": "azure", + "max_input_tokens": 272000, + "max_output_tokens": 128000, + "max_tokens": 128000, + "mode": "responses", + "output_cost_per_token": 0.00001, + "supported_endpoints": [ + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], + "supports_function_calling": true, + "supports_native_streaming": true, + "supports_parallel_function_calling": true, + "supports_pdf_input": true, + "supports_prompt_caching": true, + "supports_reasoning": true, + "supports_response_schema": true, + "supports_system_messages": false, + "supports_tool_choice": true, + "supports_vision": true + }, + "azure/gpt-5.1-codex-mini": { + "cache_read_input_token_cost": 2.5e-8, + "input_cost_per_token": 2.5e-7, + "litellm_provider": "azure", + "max_input_tokens": 272000, + "max_output_tokens": 128000, + "max_tokens": 128000, + "mode": "responses", + "output_cost_per_token": 0.000002, + "supported_endpoints": [ + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], + "supports_function_calling": true, + "supports_native_streaming": true, + "supports_parallel_function_calling": true, + "supports_pdf_input": true, + "supports_prompt_caching": true, + "supports_reasoning": true, + "supports_response_schema": true, + "supports_system_messages": false, + "supports_tool_choice": true, + "supports_vision": true + }, "azure/gpt-image-1": { "input_cost_per_pixel": 4.0054321e-8, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_pixel": 0.0, - "supported_endpoints": ["/v1/images/generations"] + "output_cost_per_pixel": 0, + "supported_endpoints": [ + "/v1/images/generations" + ] }, "azure/hd/1024-x-1024/dall-e-3": { "input_cost_per_pixel": 7.629e-8, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "azure/hd/1024-x-1792/dall-e-3": { "input_cost_per_pixel": 6.539e-8, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "azure/hd/1792-x-1024/dall-e-3": { "input_cost_per_pixel": 6.539e-8, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "azure/high/1024-x-1024/gpt-image-1": { "input_cost_per_pixel": 1.59263611e-7, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_pixel": 0.0, - "supported_endpoints": ["/v1/images/generations"] + "output_cost_per_pixel": 0, + "supported_endpoints": [ + "/v1/images/generations" + ] }, "azure/high/1024-x-1536/gpt-image-1": { "input_cost_per_pixel": 1.58945719e-7, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_pixel": 0.0, - "supported_endpoints": ["/v1/images/generations"] + "output_cost_per_pixel": 0, + "supported_endpoints": [ + "/v1/images/generations" + ] }, "azure/high/1536-x-1024/gpt-image-1": { "input_cost_per_pixel": 1.58945719e-7, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_pixel": 0.0, - "supported_endpoints": ["/v1/images/generations"] + "output_cost_per_pixel": 0, + "supported_endpoints": [ + "/v1/images/generations" + ] }, "azure/low/1024-x-1024/gpt-image-1": { "input_cost_per_pixel": 1.0490417e-8, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_pixel": 0.0, - "supported_endpoints": ["/v1/images/generations"] + "output_cost_per_pixel": 0, + "supported_endpoints": [ + "/v1/images/generations" + ] }, "azure/low/1024-x-1536/gpt-image-1": { "input_cost_per_pixel": 1.0172526e-8, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_pixel": 0.0, - "supported_endpoints": ["/v1/images/generations"] + "output_cost_per_pixel": 0, + "supported_endpoints": [ + "/v1/images/generations" + ] }, "azure/low/1536-x-1024/gpt-image-1": { "input_cost_per_pixel": 1.0172526e-8, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_pixel": 0.0, - "supported_endpoints": ["/v1/images/generations"] + "output_cost_per_pixel": 0, + "supported_endpoints": [ + "/v1/images/generations" + ] }, "azure/medium/1024-x-1024/gpt-image-1": { "input_cost_per_pixel": 4.0054321e-8, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_pixel": 0.0, - "supported_endpoints": ["/v1/images/generations"] + "output_cost_per_pixel": 0, + "supported_endpoints": [ + "/v1/images/generations" + ] }, "azure/medium/1024-x-1536/gpt-image-1": { "input_cost_per_pixel": 4.0054321e-8, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_pixel": 0.0, - "supported_endpoints": ["/v1/images/generations"] + "output_cost_per_pixel": 0, + "supported_endpoints": [ + "/v1/images/generations" + ] }, "azure/medium/1536-x-1024/gpt-image-1": { "input_cost_per_pixel": 4.0054321e-8, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_pixel": 0.0, - "supported_endpoints": ["/v1/images/generations"] + "output_cost_per_pixel": 0, + "supported_endpoints": [ + "/v1/images/generations" + ] }, "azure/gpt-image-1-mini": { "input_cost_per_pixel": 8.0566406e-9, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_pixel": 0.0, - "supported_endpoints": ["/v1/images/generations"] + "output_cost_per_pixel": 0, + "supported_endpoints": [ + "/v1/images/generations" + ] }, "azure/low/1024-x-1024/gpt-image-1-mini": { "input_cost_per_pixel": 2.0751953125e-9, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_pixel": 0.0, - "supported_endpoints": ["/v1/images/generations"] + "output_cost_per_pixel": 0, + "supported_endpoints": [ + "/v1/images/generations" + ] }, "azure/low/1024-x-1536/gpt-image-1-mini": { "input_cost_per_pixel": 2.0751953125e-9, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_pixel": 0.0, - "supported_endpoints": ["/v1/images/generations"] + "output_cost_per_pixel": 0, + "supported_endpoints": [ + "/v1/images/generations" + ] }, "azure/low/1536-x-1024/gpt-image-1-mini": { "input_cost_per_pixel": 2.0345052083e-9, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_pixel": 0.0, - "supported_endpoints": ["/v1/images/generations"] + "output_cost_per_pixel": 0, + "supported_endpoints": [ + "/v1/images/generations" + ] }, "azure/medium/1024-x-1024/gpt-image-1-mini": { "input_cost_per_pixel": 8.056640625e-9, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_pixel": 0.0, - "supported_endpoints": ["/v1/images/generations"] + "output_cost_per_pixel": 0, + "supported_endpoints": [ + "/v1/images/generations" + ] }, "azure/medium/1024-x-1536/gpt-image-1-mini": { "input_cost_per_pixel": 8.056640625e-9, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_pixel": 0.0, - "supported_endpoints": ["/v1/images/generations"] + "output_cost_per_pixel": 0, + "supported_endpoints": [ + "/v1/images/generations" + ] }, "azure/medium/1536-x-1024/gpt-image-1-mini": { "input_cost_per_pixel": 7.9752604167e-9, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_pixel": 0.0, - "supported_endpoints": ["/v1/images/generations"] + "output_cost_per_pixel": 0, + "supported_endpoints": [ + "/v1/images/generations" + ] }, "azure/high/1024-x-1024/gpt-image-1-mini": { "input_cost_per_pixel": 3.173828125e-8, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_pixel": 0.0, - "supported_endpoints": ["/v1/images/generations"] + "output_cost_per_pixel": 0, + "supported_endpoints": [ + "/v1/images/generations" + ] }, "azure/high/1024-x-1536/gpt-image-1-mini": { "input_cost_per_pixel": 3.173828125e-8, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_pixel": 0.0, - "supported_endpoints": ["/v1/images/generations"] + "output_cost_per_pixel": 0, + "supported_endpoints": [ + "/v1/images/generations" + ] }, "azure/high/1536-x-1024/gpt-image-1-mini": { "input_cost_per_pixel": 3.1575520833e-8, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_pixel": 0.0, - "supported_endpoints": ["/v1/images/generations"] + "output_cost_per_pixel": 0, + "supported_endpoints": [ + "/v1/images/generations" + ] }, "azure/mistral-large-2402": { - "input_cost_per_token": 8e-6, + "input_cost_per_token": 0.000008, "litellm_provider": "azure", "max_input_tokens": 32000, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 2.4e-5, + "output_cost_per_token": 0.000024, "supports_function_calling": true }, "azure/mistral-large-latest": { - "input_cost_per_token": 8e-6, + "input_cost_per_token": 0.000008, "litellm_provider": "azure", "max_input_tokens": 32000, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 2.4e-5, + "output_cost_per_token": 0.000024, "supports_function_calling": true }, "azure/o1": { - "cache_read_input_token_cost": 7.5e-6, - "input_cost_per_token": 1.5e-5, + "cache_read_input_token_cost": 0.0000075, + "input_cost_per_token": 0.000015, "litellm_provider": "azure", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 6e-5, + "output_cost_per_token": 0.00006, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -2441,14 +3419,14 @@ "supports_vision": true }, "azure/o1-2024-12-17": { - "cache_read_input_token_cost": 7.5e-6, - "input_cost_per_token": 1.5e-5, + "cache_read_input_token_cost": 0.0000075, + "input_cost_per_token": 0.000015, "litellm_provider": "azure", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 6e-5, + "output_cost_per_token": 0.00006, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -2458,13 +3436,13 @@ }, "azure/o1-mini": { "cache_read_input_token_cost": 6.05e-7, - "input_cost_per_token": 1.21e-6, + "input_cost_per_token": 0.00000121, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 65536, "max_tokens": 65536, "mode": "chat", - "output_cost_per_token": 4.84e-6, + "output_cost_per_token": 0.00000484, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -2473,13 +3451,13 @@ }, "azure/o1-mini-2024-09-12": { "cache_read_input_token_cost": 5.5e-7, - "input_cost_per_token": 1.1e-6, + "input_cost_per_token": 0.0000011, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 65536, "max_tokens": 65536, "mode": "chat", - "output_cost_per_token": 4.4e-6, + "output_cost_per_token": 0.0000044, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -2487,14 +3465,14 @@ "supports_vision": false }, "azure/o1-preview": { - "cache_read_input_token_cost": 7.5e-6, - "input_cost_per_token": 1.5e-5, + "cache_read_input_token_cost": 0.0000075, + "input_cost_per_token": 0.000015, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 6e-5, + "output_cost_per_token": 0.00006, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -2502,14 +3480,14 @@ "supports_vision": false }, "azure/o1-preview-2024-09-12": { - "cache_read_input_token_cost": 7.5e-6, - "input_cost_per_token": 1.5e-5, + "cache_read_input_token_cost": 0.0000075, + "input_cost_per_token": 0.000015, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 6e-5, + "output_cost_per_token": 0.00006, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_pdf_input": true, @@ -2519,16 +3497,25 @@ }, "azure/o3": { "cache_read_input_token_cost": 5e-7, - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "azure", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 8e-6, - "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "output_cost_per_token": 0.000008, + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_parallel_function_calling": false, "supports_prompt_caching": true, @@ -2540,16 +3527,25 @@ "azure/o3-2025-04-16": { "deprecation_date": "2026-04-16", "cache_read_input_token_cost": 5e-7, - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "azure", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 8e-6, - "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "output_cost_per_token": 0.000008, + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_parallel_function_calling": false, "supports_prompt_caching": true, @@ -2559,17 +3555,26 @@ "supports_vision": true }, "azure/o3-deep-research": { - "cache_read_input_token_cost": 2.5e-6, - "input_cost_per_token": 1e-5, + "cache_read_input_token_cost": 0.0000025, + "input_cost_per_token": 0.00001, "litellm_provider": "azure", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "responses", - "output_cost_per_token": 4e-5, - "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "output_cost_per_token": 0.00004, + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_pdf_input": true, @@ -2583,13 +3588,13 @@ }, "azure/o3-mini": { "cache_read_input_token_cost": 5.5e-7, - "input_cost_per_token": 1.1e-6, + "input_cost_per_token": 0.0000011, "litellm_provider": "azure", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 4.4e-6, + "output_cost_per_token": 0.0000044, "supports_prompt_caching": true, "supports_reasoning": true, "supports_response_schema": true, @@ -2598,31 +3603,40 @@ }, "azure/o3-mini-2025-01-31": { "cache_read_input_token_cost": 5.5e-7, - "input_cost_per_token": 1.1e-6, + "input_cost_per_token": 0.0000011, "litellm_provider": "azure", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 4.4e-6, + "output_cost_per_token": 0.0000044, "supports_prompt_caching": true, "supports_reasoning": true, "supports_tool_choice": true, "supports_vision": false }, "azure/o3-pro": { - "input_cost_per_token": 2e-5, - "input_cost_per_token_batches": 1e-5, + "input_cost_per_token": 0.00002, + "input_cost_per_token_batches": 0.00001, "litellm_provider": "azure", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "responses", - "output_cost_per_token": 8e-5, - "output_cost_per_token_batches": 4e-5, - "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "output_cost_per_token": 0.00008, + "output_cost_per_token_batches": 0.00004, + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_parallel_function_calling": false, "supports_prompt_caching": false, @@ -2632,18 +3646,27 @@ "supports_vision": true }, "azure/o3-pro-2025-06-10": { - "input_cost_per_token": 2e-5, - "input_cost_per_token_batches": 1e-5, + "input_cost_per_token": 0.00002, + "input_cost_per_token_batches": 0.00001, "litellm_provider": "azure", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "responses", - "output_cost_per_token": 8e-5, - "output_cost_per_token_batches": 4e-5, - "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "output_cost_per_token": 0.00008, + "output_cost_per_token_batches": 0.00004, + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_parallel_function_calling": false, "supports_prompt_caching": false, @@ -2654,16 +3677,25 @@ }, "azure/o4-mini": { "cache_read_input_token_cost": 2.75e-7, - "input_cost_per_token": 1.1e-6, + "input_cost_per_token": 0.0000011, "litellm_provider": "azure", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 4.4e-6, - "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "output_cost_per_token": 0.0000044, + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_parallel_function_calling": false, "supports_prompt_caching": true, @@ -2674,13 +3706,13 @@ }, "azure/o4-mini-2025-04-16": { "cache_read_input_token_cost": 2.75e-7, - "input_cost_per_token": 1.1e-6, + "input_cost_per_token": 0.0000011, "litellm_provider": "azure", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 4.4e-6, + "output_cost_per_token": 0.0000044, "supports_function_calling": true, "supports_parallel_function_calling": false, "supports_prompt_caching": true, @@ -2690,28 +3722,28 @@ "supports_vision": true }, "azure/standard/1024-x-1024/dall-e-2": { - "input_cost_per_pixel": 0.0, + "input_cost_per_pixel": 0, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "azure/standard/1024-x-1024/dall-e-3": { "input_cost_per_pixel": 3.81469e-8, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "azure/standard/1024-x-1792/dall-e-3": { "input_cost_per_pixel": 4.359e-8, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "azure/standard/1792-x-1024/dall-e-3": { "input_cost_per_pixel": 4.359e-8, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "azure/text-embedding-3-large": { "input_cost_per_token": 1.3e-7, @@ -2719,7 +3751,7 @@ "max_input_tokens": 8191, "max_tokens": 8191, "mode": "embedding", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "azure/text-embedding-3-small": { "deprecation_date": "2026-04-30", @@ -2728,7 +3760,7 @@ "max_input_tokens": 8191, "max_tokens": 8191, "mode": "embedding", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "azure/text-embedding-ada-002": { "input_cost_per_token": 1e-7, @@ -2736,45 +3768,54 @@ "max_input_tokens": 8191, "max_tokens": 8191, "mode": "embedding", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "azure/speech/azure-tts": { - "input_cost_per_character": 15e-6, + "input_cost_per_character": 0.000015, "litellm_provider": "azure", "mode": "audio_speech", "source": "https://azure.microsoft.com/en-us/pricing/calculator/" }, "azure/speech/azure-tts-hd": { - "input_cost_per_character": 30e-6, + "input_cost_per_character": 0.00003, "litellm_provider": "azure", "mode": "audio_speech", "source": "https://azure.microsoft.com/en-us/pricing/calculator/" }, "azure/tts-1": { - "input_cost_per_character": 1.5e-5, + "input_cost_per_character": 0.000015, "litellm_provider": "azure", "mode": "audio_speech" }, "azure/tts-1-hd": { - "input_cost_per_character": 3e-5, + "input_cost_per_character": 0.00003, "litellm_provider": "azure", "mode": "audio_speech" }, "azure/us/gpt-4.1-2025-04-14": { "deprecation_date": "2026-11-04", "cache_read_input_token_cost": 5.5e-7, - "input_cost_per_token": 2.2e-6, - "input_cost_per_token_batches": 1.1e-6, + "input_cost_per_token": 0.0000022, + "input_cost_per_token_batches": 0.0000011, "litellm_provider": "azure", "max_input_tokens": 1047576, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 8.8e-6, - "output_cost_per_token_batches": 4.4e-6, - "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "output_cost_per_token": 0.0000088, + "output_cost_per_token_batches": 0.0000044, + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -2795,11 +3836,20 @@ "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 1.76e-6, + "output_cost_per_token": 0.00000176, "output_cost_per_token_batches": 8.8e-7, - "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -2822,9 +3872,18 @@ "mode": "chat", "output_cost_per_token": 4.4e-7, "output_cost_per_token_batches": 2.2e-7, - "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -2836,14 +3895,14 @@ }, "azure/us/gpt-4o-2024-08-06": { "deprecation_date": "2026-02-27", - "cache_read_input_token_cost": 1.375e-6, - "input_cost_per_token": 2.75e-6, + "cache_read_input_token_cost": 0.000001375, + "input_cost_per_token": 0.00000275, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 1.1e-5, + "output_cost_per_token": 0.000011, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -2853,14 +3912,14 @@ }, "azure/us/gpt-4o-2024-11-20": { "deprecation_date": "2026-03-01", - "cache_creation_input_token_cost": 1.38e-6, - "input_cost_per_token": 2.75e-6, + "cache_creation_input_token_cost": 0.00000138, + "input_cost_per_token": 0.00000275, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 1.1e-5, + "output_cost_per_token": 0.000011, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_response_schema": true, @@ -2886,15 +3945,15 @@ "azure/us/gpt-4o-mini-realtime-preview-2024-12-17": { "cache_creation_input_audio_token_cost": 3.3e-7, "cache_read_input_token_cost": 3.3e-7, - "input_cost_per_audio_token": 1.1e-5, + "input_cost_per_audio_token": 0.000011, "input_cost_per_token": 6.6e-7, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_audio_token": 2.2e-5, - "output_cost_per_token": 2.64e-6, + "output_cost_per_audio_token": 0.000022, + "output_cost_per_token": 0.00000264, "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -2903,17 +3962,17 @@ "supports_tool_choice": true }, "azure/us/gpt-4o-realtime-preview-2024-10-01": { - "cache_creation_input_audio_token_cost": 2.2e-5, - "cache_read_input_token_cost": 2.75e-6, + "cache_creation_input_audio_token_cost": 0.000022, + "cache_read_input_token_cost": 0.00000275, "input_cost_per_audio_token": 0.00011, - "input_cost_per_token": 5.5e-6, + "input_cost_per_token": 0.0000055, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", "output_cost_per_audio_token": 0.00022, - "output_cost_per_token": 2.2e-5, + "output_cost_per_token": 0.000022, "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -2922,19 +3981,25 @@ "supports_tool_choice": true }, "azure/us/gpt-4o-realtime-preview-2024-12-17": { - "cache_read_input_audio_token_cost": 2.5e-6, - "cache_read_input_token_cost": 2.75e-6, - "input_cost_per_audio_token": 4.4e-5, - "input_cost_per_token": 5.5e-6, + "cache_read_input_audio_token_cost": 0.0000025, + "cache_read_input_token_cost": 0.00000275, + "input_cost_per_audio_token": 0.000044, + "input_cost_per_token": 0.0000055, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_audio_token": 8e-5, - "output_cost_per_token": 2.2e-5, - "supported_modalities": ["text", "audio"], - "supported_output_modalities": ["text", "audio"], + "output_cost_per_audio_token": 0.00008, + "output_cost_per_token": 0.000022, + "supported_modalities": [ + "text", + "audio" + ], + "supported_output_modalities": [ + "text", + "audio" + ], "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -2944,16 +4009,25 @@ }, "azure/us/gpt-5-2025-08-07": { "cache_read_input_token_cost": 1.375e-7, - "input_cost_per_token": 1.375e-6, + "input_cost_per_token": 0.000001375, "litellm_provider": "azure", "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 1.1e-5, - "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "output_cost_per_token": 0.000011, + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -2973,10 +4047,19 @@ "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 2.2e-6, - "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "output_cost_per_token": 0.0000022, + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -2997,9 +4080,84 @@ "max_tokens": 128000, "mode": "chat", "output_cost_per_token": 4.4e-7, - "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], + "supports_function_calling": true, + "supports_native_streaming": true, + "supports_parallel_function_calling": true, + "supports_pdf_input": true, + "supports_prompt_caching": true, + "supports_reasoning": true, + "supports_response_schema": true, + "supports_system_messages": true, + "supports_tool_choice": true, + "supports_vision": true + }, + "azure/us/gpt-5.1": { + "cache_read_input_token_cost": 1.4e-7, + "input_cost_per_token": 0.00000138, + "litellm_provider": "azure", + "max_input_tokens": 272000, + "max_output_tokens": 128000, + "max_tokens": 128000, + "mode": "chat", + "output_cost_per_token": 0.000011, + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text", + "image" + ], + "supports_function_calling": true, + "supports_native_streaming": true, + "supports_parallel_function_calling": true, + "supports_pdf_input": true, + "supports_prompt_caching": true, + "supports_reasoning": true, + "supports_response_schema": true, + "supports_system_messages": true, + "supports_tool_choice": true, + "supports_vision": true + }, + "azure/us/gpt-5.1-chat": { + "cache_read_input_token_cost": 1.4e-7, + "input_cost_per_token": 0.00000138, + "litellm_provider": "azure", + "max_input_tokens": 272000, + "max_output_tokens": 128000, + "max_tokens": 128000, + "mode": "chat", + "output_cost_per_token": 0.000011, + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text", + "image" + ], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -3011,15 +4169,75 @@ "supports_tool_choice": true, "supports_vision": true }, + "azure/us/gpt-5.1-codex": { + "cache_read_input_token_cost": 1.4e-7, + "input_cost_per_token": 0.00000138, + "litellm_provider": "azure", + "max_input_tokens": 272000, + "max_output_tokens": 128000, + "max_tokens": 128000, + "mode": "responses", + "output_cost_per_token": 0.000011, + "supported_endpoints": [ + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], + "supports_function_calling": true, + "supports_native_streaming": true, + "supports_parallel_function_calling": true, + "supports_pdf_input": true, + "supports_prompt_caching": true, + "supports_reasoning": true, + "supports_response_schema": true, + "supports_system_messages": false, + "supports_tool_choice": true, + "supports_vision": true + }, + "azure/us/gpt-5.1-codex-mini": { + "cache_read_input_token_cost": 2.8e-8, + "input_cost_per_token": 2.75e-7, + "litellm_provider": "azure", + "max_input_tokens": 272000, + "max_output_tokens": 128000, + "max_tokens": 128000, + "mode": "responses", + "output_cost_per_token": 0.0000022, + "supported_endpoints": [ + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], + "supports_function_calling": true, + "supports_native_streaming": true, + "supports_parallel_function_calling": true, + "supports_pdf_input": true, + "supports_prompt_caching": true, + "supports_reasoning": true, + "supports_response_schema": true, + "supports_system_messages": false, + "supports_tool_choice": true, + "supports_vision": true + }, "azure/us/o1-2024-12-17": { - "cache_read_input_token_cost": 8.25e-6, - "input_cost_per_token": 1.65e-5, + "cache_read_input_token_cost": 0.00000825, + "input_cost_per_token": 0.0000165, "litellm_provider": "azure", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 6.6e-5, + "output_cost_per_token": 0.000066, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -3028,29 +4246,29 @@ }, "azure/us/o1-mini-2024-09-12": { "cache_read_input_token_cost": 6.05e-7, - "input_cost_per_token": 1.21e-6, + "input_cost_per_token": 0.00000121, "input_cost_per_token_batches": 6.05e-7, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 65536, "max_tokens": 65536, "mode": "chat", - "output_cost_per_token": 4.84e-6, - "output_cost_per_token_batches": 2.42e-6, + "output_cost_per_token": 0.00000484, + "output_cost_per_token_batches": 0.00000242, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, "supports_vision": false }, "azure/us/o1-preview-2024-09-12": { - "cache_read_input_token_cost": 8.25e-6, - "input_cost_per_token": 1.65e-5, + "cache_read_input_token_cost": 0.00000825, + "input_cost_per_token": 0.0000165, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 6.6e-5, + "output_cost_per_token": 0.000066, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -3059,16 +4277,25 @@ "azure/us/o3-2025-04-16": { "deprecation_date": "2026-04-16", "cache_read_input_token_cost": 5.5e-7, - "input_cost_per_token": 2.2e-6, + "input_cost_per_token": 0.0000022, "litellm_provider": "azure", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 8.8e-6, - "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "output_cost_per_token": 0.0000088, + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_parallel_function_calling": false, "supports_prompt_caching": true, @@ -3079,15 +4306,15 @@ }, "azure/us/o3-mini-2025-01-31": { "cache_read_input_token_cost": 6.05e-7, - "input_cost_per_token": 1.21e-6, + "input_cost_per_token": 0.00000121, "input_cost_per_token_batches": 6.05e-7, "litellm_provider": "azure", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 4.84e-6, - "output_cost_per_token_batches": 2.42e-6, + "output_cost_per_token": 0.00000484, + "output_cost_per_token_batches": 0.00000242, "supports_prompt_caching": true, "supports_reasoning": true, "supports_tool_choice": true, @@ -3095,13 +4322,13 @@ }, "azure/us/o4-mini-2025-04-16": { "cache_read_input_token_cost": 3.1e-7, - "input_cost_per_token": 1.21e-6, + "input_cost_per_token": 0.00000121, "litellm_provider": "azure", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 4.84e-6, + "output_cost_per_token": 0.00000484, "supports_function_calling": true, "supports_parallel_function_calling": false, "supports_prompt_caching": true, @@ -3122,7 +4349,7 @@ "max_input_tokens": 512, "max_tokens": 512, "mode": "embedding", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "output_vector_size": 1024, "source": "https://azuremarketplace.microsoft.com/en-us/marketplace/apps/cohere.cohere-embed-v3-english-offer?tab=PlansAndPrice", "supports_embedding_image_input": true @@ -3133,7 +4360,7 @@ "max_input_tokens": 512, "max_tokens": 512, "mode": "embedding", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "output_vector_size": 1024, "source": "https://azuremarketplace.microsoft.com/en-us/marketplace/apps/cohere.cohere-embed-v3-english-offer?tab=PlansAndPrice", "supports_embedding_image_input": true @@ -3143,14 +4370,18 @@ "mode": "image_generation", "output_cost_per_image": 0.04, "source": "https://techcommunity.microsoft.com/blog/azure-ai-foundry-blog/black-forest-labs-flux-1-kontext-pro-and-flux1-1-pro-now-available-in-azure-ai-f/4434659", - "supported_endpoints": ["/v1/images/generations"] + "supported_endpoints": [ + "/v1/images/generations" + ] }, "azure_ai/FLUX.1-Kontext-pro": { "litellm_provider": "azure_ai", "mode": "image_generation", "output_cost_per_image": 0.04, "source": "https://azuremarketplace.microsoft.com/pt-br/marketplace/apps/cohere.cohere-embed-4-offer?tab=PlansAndPrice", - "supported_endpoints": ["/v1/images/generations"] + "supported_endpoints": [ + "/v1/images/generations" + ] }, "azure_ai/Llama-3.2-11B-Vision-Instruct": { "input_cost_per_token": 3.7e-7, @@ -3166,13 +4397,13 @@ "supports_vision": true }, "azure_ai/Llama-3.2-90B-Vision-Instruct": { - "input_cost_per_token": 2.04e-6, + "input_cost_per_token": 0.00000204, "litellm_provider": "azure_ai", "max_input_tokens": 128000, "max_output_tokens": 2048, "max_tokens": 2048, "mode": "chat", - "output_cost_per_token": 2.04e-6, + "output_cost_per_token": 0.00000204, "source": "https://azuremarketplace.microsoft.com/en/marketplace/apps/metagenai.meta-llama-3-2-90b-vision-instruct-offer?tab=Overview", "supports_function_calling": true, "supports_tool_choice": true, @@ -3191,7 +4422,7 @@ "supports_tool_choice": true }, "azure_ai/Llama-4-Maverick-17B-128E-Instruct-FP8": { - "input_cost_per_token": 1.41e-6, + "input_cost_per_token": 0.00000141, "litellm_provider": "azure_ai", "max_input_tokens": 1000000, "max_output_tokens": 16384, @@ -3217,7 +4448,7 @@ "supports_vision": true }, "azure_ai/Meta-Llama-3-70B-Instruct": { - "input_cost_per_token": 1.1e-6, + "input_cost_per_token": 0.0000011, "litellm_provider": "azure_ai", "max_input_tokens": 8192, "max_output_tokens": 2048, @@ -3227,24 +4458,24 @@ "supports_tool_choice": true }, "azure_ai/Meta-Llama-3.1-405B-Instruct": { - "input_cost_per_token": 5.33e-6, + "input_cost_per_token": 0.00000533, "litellm_provider": "azure_ai", "max_input_tokens": 128000, "max_output_tokens": 2048, "max_tokens": 2048, "mode": "chat", - "output_cost_per_token": 1.6e-5, + "output_cost_per_token": 0.000016, "source": "https://azuremarketplace.microsoft.com/en-us/marketplace/apps/metagenai.meta-llama-3-1-405b-instruct-offer?tab=PlansAndPrice", "supports_tool_choice": true }, "azure_ai/Meta-Llama-3.1-70B-Instruct": { - "input_cost_per_token": 2.68e-6, + "input_cost_per_token": 0.00000268, "litellm_provider": "azure_ai", "max_input_tokens": 128000, "max_output_tokens": 2048, "max_tokens": 2048, "mode": "chat", - "output_cost_per_token": 3.54e-6, + "output_cost_per_token": 0.00000354, "source": "https://azuremarketplace.microsoft.com/en-us/marketplace/apps/metagenai.meta-llama-3-1-70b-instruct-offer?tab=PlansAndPrice", "supports_tool_choice": true }, @@ -3392,7 +4623,7 @@ "supports_function_calling": true }, "azure_ai/Phi-4-multimodal-instruct": { - "input_cost_per_audio_token": 4e-6, + "input_cost_per_audio_token": 0.000004, "input_cost_per_token": 8e-8, "litellm_provider": "azure_ai", "max_input_tokens": 131072, @@ -3431,108 +4662,116 @@ }, "azure_ai/mistral-document-ai-2505": { "litellm_provider": "azure_ai", - "ocr_cost_per_page": 3e-3, + "ocr_cost_per_page": 0.003, "mode": "ocr", - "supported_endpoints": ["/v1/ocr"], + "supported_endpoints": [ + "/v1/ocr" + ], "source": "https://devblogs.microsoft.com/foundry/whats-new-in-azure-ai-foundry-august-2025/#mistral-document-ai-(ocr)-%E2%80%94-serverless-in-foundry" }, "azure_ai/doc-intelligence/prebuilt-read": { "litellm_provider": "azure_ai", - "ocr_cost_per_page": 1.5e-3, + "ocr_cost_per_page": 0.0015, "mode": "ocr", - "supported_endpoints": ["/v1/ocr"], + "supported_endpoints": [ + "/v1/ocr" + ], "source": "https://azure.microsoft.com/en-us/pricing/details/ai-document-intelligence/" }, "azure_ai/doc-intelligence/prebuilt-layout": { "litellm_provider": "azure_ai", - "ocr_cost_per_page": 1e-2, + "ocr_cost_per_page": 0.01, "mode": "ocr", - "supported_endpoints": ["/v1/ocr"], + "supported_endpoints": [ + "/v1/ocr" + ], "source": "https://azure.microsoft.com/en-us/pricing/details/ai-document-intelligence/" }, "azure_ai/doc-intelligence/prebuilt-document": { "litellm_provider": "azure_ai", - "ocr_cost_per_page": 1e-2, + "ocr_cost_per_page": 0.01, "mode": "ocr", - "supported_endpoints": ["/v1/ocr"], + "supported_endpoints": [ + "/v1/ocr" + ], "source": "https://azure.microsoft.com/en-us/pricing/details/ai-document-intelligence/" }, "azure_ai/MAI-DS-R1": { - "input_cost_per_token": 1.35e-6, + "input_cost_per_token": 0.00000135, "litellm_provider": "azure_ai", "max_input_tokens": 128000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 5.4e-6, + "output_cost_per_token": 0.0000054, "source": "https://azure.microsoft.com/en-us/pricing/details/ai-foundry-models/microsoft/", "supports_reasoning": true, "supports_tool_choice": true }, "azure_ai/cohere-rerank-v3-english": { "input_cost_per_query": 0.002, - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "azure_ai", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_query_tokens": 2048, "max_tokens": 4096, "mode": "rerank", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "azure_ai/cohere-rerank-v3-multilingual": { "input_cost_per_query": 0.002, - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "azure_ai", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_query_tokens": 2048, "max_tokens": 4096, "mode": "rerank", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "azure_ai/cohere-rerank-v3.5": { "input_cost_per_query": 0.002, - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "azure_ai", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_query_tokens": 2048, "max_tokens": 4096, "mode": "rerank", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "azure_ai/deepseek-r1": { - "input_cost_per_token": 1.35e-6, + "input_cost_per_token": 0.00000135, "litellm_provider": "azure_ai", "max_input_tokens": 128000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 5.4e-6, + "output_cost_per_token": 0.0000054, "source": "https://techcommunity.microsoft.com/blog/machinelearningblog/deepseek-r1-improved-performance-higher-limits-and-transparent-pricing/4386367", "supports_reasoning": true, "supports_tool_choice": true }, "azure_ai/deepseek-v3": { - "input_cost_per_token": 1.14e-6, + "input_cost_per_token": 0.00000114, "litellm_provider": "azure_ai", "max_input_tokens": 128000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 4.56e-6, + "output_cost_per_token": 0.00000456, "source": "https://techcommunity.microsoft.com/blog/machinelearningblog/announcing-deepseek-v3-on-azure-ai-foundry-and-github/4390438", "supports_tool_choice": true }, "azure_ai/deepseek-v3-0324": { - "input_cost_per_token": 1.14e-6, + "input_cost_per_token": 0.00000114, "litellm_provider": "azure_ai", "max_input_tokens": 128000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 4.56e-6, + "output_cost_per_token": 0.00000456, "source": "https://techcommunity.microsoft.com/blog/machinelearningblog/announcing-deepseek-v3-on-azure-ai-foundry-and-github/4390438", "supports_function_calling": true, "supports_tool_choice": true @@ -3543,21 +4782,26 @@ "max_input_tokens": 128000, "max_tokens": 128000, "mode": "embedding", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "output_vector_size": 3072, "source": "https://azuremarketplace.microsoft.com/pt-br/marketplace/apps/cohere.cohere-embed-4-offer?tab=PlansAndPrice", - "supported_endpoints": ["/v1/embeddings"], - "supported_modalities": ["text", "image"], + "supported_endpoints": [ + "/v1/embeddings" + ], + "supported_modalities": [ + "text", + "image" + ], "supports_embedding_image_input": true }, "azure_ai/global/grok-3": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "azure_ai", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "source": "https://devblogs.microsoft.com/foundry/announcing-grok-3-and-grok-3-mini-on-azure-ai-foundry/", "supports_function_calling": true, "supports_response_schema": false, @@ -3571,7 +4815,7 @@ "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 1.27e-6, + "output_cost_per_token": 0.00000127, "source": "https://devblogs.microsoft.com/foundry/announcing-grok-3-and-grok-3-mini-on-azure-ai-foundry/", "supports_function_calling": true, "supports_reasoning": true, @@ -3580,13 +4824,13 @@ "supports_web_search": true }, "azure_ai/grok-3": { - "input_cost_per_token": 3.3e-6, + "input_cost_per_token": 0.0000033, "litellm_provider": "azure_ai", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 1.65e-5, + "output_cost_per_token": 0.0000165, "source": "https://devblogs.microsoft.com/foundry/announcing-grok-3-and-grok-3-mini-on-azure-ai-foundry/", "supports_function_calling": true, "supports_response_schema": false, @@ -3600,7 +4844,7 @@ "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 1.38e-6, + "output_cost_per_token": 0.00000138, "source": "https://devblogs.microsoft.com/foundry/announcing-grok-3-and-grok-3-mini-on-azure-ai-foundry/", "supports_function_calling": true, "supports_reasoning": true, @@ -3609,13 +4853,13 @@ "supports_web_search": true }, "azure_ai/grok-4": { - "input_cost_per_token": 5.5e-6, + "input_cost_per_token": 0.0000055, "litellm_provider": "azure_ai", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 2.75e-5, + "output_cost_per_token": 0.0000275, "source": "https://azure.microsoft.com/en-us/blog/grok-4-is-now-available-in-azure-ai-foundry-unlock-frontier-intelligence-and-business-ready-capabilities/", "supports_function_calling": true, "supports_response_schema": true, @@ -3623,8 +4867,8 @@ "supports_web_search": true }, "azure_ai/grok-4-fast-non-reasoning": { - "input_cost_per_token": 0.43e-6, - "output_cost_per_token": 1.73e-6, + "input_cost_per_token": 4.3e-7, + "output_cost_per_token": 0.00000173, "litellm_provider": "azure_ai", "max_input_tokens": 131072, "max_output_tokens": 131072, @@ -3636,8 +4880,8 @@ "supports_web_search": true }, "azure_ai/grok-4-fast-reasoning": { - "input_cost_per_token": 0.43e-6, - "output_cost_per_token": 1.73e-6, + "input_cost_per_token": 4.3e-7, + "output_cost_per_token": 0.00000173, "litellm_provider": "azure_ai", "max_input_tokens": 131072, "max_output_tokens": 131072, @@ -3650,13 +4894,13 @@ "supports_web_search": true }, "azure_ai/grok-code-fast-1": { - "input_cost_per_token": 3.5e-6, + "input_cost_per_token": 0.0000035, "litellm_provider": "azure_ai", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 1.75e-5, + "output_cost_per_token": 0.0000175, "source": "https://azure.microsoft.com/en-us/blog/grok-4-is-now-available-in-azure-ai-foundry-unlock-frontier-intelligence-and-business-ready-capabilities/", "supports_function_calling": true, "supports_response_schema": true, @@ -3696,36 +4940,36 @@ "supports_tool_choice": true }, "azure_ai/mistral-large": { - "input_cost_per_token": 4e-6, + "input_cost_per_token": 0.000004, "litellm_provider": "azure_ai", "max_input_tokens": 32000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 1.2e-5, + "output_cost_per_token": 0.000012, "supports_function_calling": true, "supports_tool_choice": true }, "azure_ai/mistral-large-2407": { - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "azure_ai", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 6e-6, + "output_cost_per_token": 0.000006, "source": "https://azuremarketplace.microsoft.com/en/marketplace/apps/000-000.mistral-ai-large-2407-offer?tab=Overview", "supports_function_calling": true, "supports_tool_choice": true }, "azure_ai/mistral-large-latest": { - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "azure_ai", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 6e-6, + "output_cost_per_token": 0.000006, "source": "https://azuremarketplace.microsoft.com/en/marketplace/apps/000-000.mistral-ai-large-2407-offer?tab=Overview", "supports_function_calling": true, "supports_tool_choice": true @@ -3737,7 +4981,7 @@ "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 2e-6, + "output_cost_per_token": 0.000002, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_tool_choice": true @@ -3754,24 +4998,24 @@ "supports_function_calling": true }, "azure_ai/mistral-small": { - "input_cost_per_token": 1e-6, + "input_cost_per_token": 0.000001, "litellm_provider": "azure_ai", "max_input_tokens": 32000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 3e-6, + "output_cost_per_token": 0.000003, "supports_function_calling": true, "supports_tool_choice": true }, "azure_ai/mistral-small-2503": { - "input_cost_per_token": 1e-6, + "input_cost_per_token": 0.000001, "litellm_provider": "azure_ai", "max_input_tokens": 128000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 3e-6, + "output_cost_per_token": 0.000003, "supports_function_calling": true, "supports_tool_choice": true, "supports_vision": true @@ -3884,43 +5128,43 @@ "supports_tool_choice": true }, "bedrock/ap-northeast-1/anthropic.claude-instant-v1": { - "input_cost_per_token": 2.23e-6, + "input_cost_per_token": 0.00000223, "litellm_provider": "bedrock", "max_input_tokens": 100000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 7.55e-6, + "output_cost_per_token": 0.00000755, "supports_tool_choice": true }, "bedrock/ap-northeast-1/anthropic.claude-v1": { - "input_cost_per_token": 8e-6, + "input_cost_per_token": 0.000008, "litellm_provider": "bedrock", "max_input_tokens": 100000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 2.4e-5, + "output_cost_per_token": 0.000024, "supports_tool_choice": true }, "bedrock/ap-northeast-1/anthropic.claude-v2:1": { - "input_cost_per_token": 8e-6, + "input_cost_per_token": 0.000008, "litellm_provider": "bedrock", "max_input_tokens": 100000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 2.4e-5, + "output_cost_per_token": 0.000024, "supports_tool_choice": true }, "bedrock/ap-south-1/meta.llama3-70b-instruct-v1:0": { - "input_cost_per_token": 3.18e-6, + "input_cost_per_token": 0.00000318, "litellm_provider": "bedrock", "max_input_tokens": 8192, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 4.2e-6 + "output_cost_per_token": 0.0000042 }, "bedrock/ap-south-1/meta.llama3-8b-instruct-v1:0": { "input_cost_per_token": 3.6e-7, @@ -3932,13 +5176,13 @@ "output_cost_per_token": 7.2e-7 }, "bedrock/ca-central-1/meta.llama3-70b-instruct-v1:0": { - "input_cost_per_token": 3.05e-6, + "input_cost_per_token": 0.00000305, "litellm_provider": "bedrock", "max_input_tokens": 8192, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 4.03e-6 + "output_cost_per_token": 0.00000403 }, "bedrock/ca-central-1/meta.llama3-8b-instruct-v1:0": { "input_cost_per_token": 3.5e-7, @@ -4008,42 +5252,42 @@ "supports_tool_choice": true }, "bedrock/eu-central-1/anthropic.claude-instant-v1": { - "input_cost_per_token": 2.48e-6, + "input_cost_per_token": 0.00000248, "litellm_provider": "bedrock", "max_input_tokens": 100000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 8.38e-6, + "output_cost_per_token": 0.00000838, "supports_tool_choice": true }, "bedrock/eu-central-1/anthropic.claude-v1": { - "input_cost_per_token": 8e-6, + "input_cost_per_token": 0.000008, "litellm_provider": "bedrock", "max_input_tokens": 100000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 2.4e-5 + "output_cost_per_token": 0.000024 }, "bedrock/eu-central-1/anthropic.claude-v2:1": { - "input_cost_per_token": 8e-6, + "input_cost_per_token": 0.000008, "litellm_provider": "bedrock", "max_input_tokens": 100000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 2.4e-5, + "output_cost_per_token": 0.000024, "supports_tool_choice": true }, "bedrock/eu-west-1/meta.llama3-70b-instruct-v1:0": { - "input_cost_per_token": 2.86e-6, + "input_cost_per_token": 0.00000286, "litellm_provider": "bedrock", "max_input_tokens": 8192, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 3.78e-6 + "output_cost_per_token": 0.00000378 }, "bedrock/eu-west-1/meta.llama3-8b-instruct-v1:0": { "input_cost_per_token": 3.2e-7, @@ -4055,13 +5299,13 @@ "output_cost_per_token": 6.5e-7 }, "bedrock/eu-west-2/meta.llama3-70b-instruct-v1:0": { - "input_cost_per_token": 3.45e-6, + "input_cost_per_token": 0.00000345, "litellm_provider": "bedrock", "max_input_tokens": 8192, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 4.55e-6 + "output_cost_per_token": 0.00000455 }, "bedrock/eu-west-2/meta.llama3-8b-instruct-v1:0": { "input_cost_per_token": 3.9e-7, @@ -4083,13 +5327,13 @@ "supports_tool_choice": true }, "bedrock/eu-west-3/mistral.mistral-large-2402-v1:0": { - "input_cost_per_token": 1.04e-5, + "input_cost_per_token": 0.0000104, "litellm_provider": "bedrock", "max_input_tokens": 32000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 3.12e-5, + "output_cost_per_token": 0.0000312, "supports_function_calling": true }, "bedrock/eu-west-3/mistral.mixtral-8x7b-instruct-v0:1": { @@ -4103,7 +5347,7 @@ "supports_tool_choice": true }, "bedrock/invoke/anthropic.claude-3-5-sonnet-20240620-v1:0": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "bedrock", "max_input_tokens": 200000, "max_output_tokens": 4096, @@ -4112,20 +5356,20 @@ "notes": "Anthropic via Invoke route does not currently support pdf input." }, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "supports_function_calling": true, "supports_response_schema": true, "supports_tool_choice": true, "supports_vision": true }, "bedrock/sa-east-1/meta.llama3-70b-instruct-v1:0": { - "input_cost_per_token": 4.45e-6, + "input_cost_per_token": 0.00000445, "litellm_provider": "bedrock", "max_input_tokens": 8192, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 5.88e-6 + "output_cost_per_token": 0.00000588 }, "bedrock/sa-east-1/meta.llama3-8b-instruct-v1:0": { "input_cost_per_token": 5e-7, @@ -4134,7 +5378,7 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 1.01e-6 + "output_cost_per_token": 0.00000101 }, "bedrock/us-east-1/1-month-commitment/anthropic.claude-instant-v1": { "input_cost_per_second": 0.011, @@ -4201,37 +5445,37 @@ "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 2.4e-6, + "output_cost_per_token": 0.0000024, "supports_tool_choice": true }, "bedrock/us-east-1/anthropic.claude-v1": { - "input_cost_per_token": 8e-6, + "input_cost_per_token": 0.000008, "litellm_provider": "bedrock", "max_input_tokens": 100000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 2.4e-5, + "output_cost_per_token": 0.000024, "supports_tool_choice": true }, "bedrock/us-east-1/anthropic.claude-v2:1": { - "input_cost_per_token": 8e-6, + "input_cost_per_token": 0.000008, "litellm_provider": "bedrock", "max_input_tokens": 100000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 2.4e-5, + "output_cost_per_token": 0.000024, "supports_tool_choice": true }, "bedrock/us-east-1/meta.llama3-70b-instruct-v1:0": { - "input_cost_per_token": 2.65e-6, + "input_cost_per_token": 0.00000265, "litellm_provider": "bedrock", "max_input_tokens": 8192, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 3.5e-6 + "output_cost_per_token": 0.0000035 }, "bedrock/us-east-1/meta.llama3-8b-instruct-v1:0": { "input_cost_per_token": 3e-7, @@ -4253,13 +5497,13 @@ "supports_tool_choice": true }, "bedrock/us-east-1/mistral.mistral-large-2402-v1:0": { - "input_cost_per_token": 8e-6, + "input_cost_per_token": 0.000008, "litellm_provider": "bedrock", "max_input_tokens": 32000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 2.4e-5, + "output_cost_per_token": 0.000024, "supports_function_calling": true }, "bedrock/us-east-1/mistral.mixtral-8x7b-instruct-v0:1": { @@ -4279,7 +5523,7 @@ "max_output_tokens": 10000, "max_tokens": 10000, "mode": "chat", - "output_cost_per_token": 3.84e-6, + "output_cost_per_token": 0.00000384, "supports_function_calling": true, "supports_pdf_input": true, "supports_prompt_caching": true, @@ -4292,7 +5536,7 @@ "max_input_tokens": 8192, "max_tokens": 8192, "mode": "embedding", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "output_vector_size": 1536 }, "bedrock/us-gov-east-1/amazon.titan-embed-text-v2:0": { @@ -4301,17 +5545,17 @@ "max_input_tokens": 8192, "max_tokens": 8192, "mode": "embedding", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "output_vector_size": 1024 }, "bedrock/us-gov-east-1/amazon.titan-text-express-v1": { - "input_cost_per_token": 1.3e-6, + "input_cost_per_token": 0.0000013, "litellm_provider": "bedrock", "max_input_tokens": 42000, "max_output_tokens": 8000, "max_tokens": 8000, "mode": "chat", - "output_cost_per_token": 1.7e-6 + "output_cost_per_token": 0.0000017 }, "bedrock/us-gov-east-1/amazon.titan-text-lite-v1": { "input_cost_per_token": 3e-7, @@ -4329,16 +5573,16 @@ "max_output_tokens": 32000, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 1.5e-6 + "output_cost_per_token": 0.0000015 }, "bedrock/us-gov-east-1/anthropic.claude-3-5-sonnet-20240620-v1:0": { - "input_cost_per_token": 3.6e-6, + "input_cost_per_token": 0.0000036, "litellm_provider": "bedrock", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 1.8e-5, + "output_cost_per_token": 0.000018, "supports_function_calling": true, "supports_pdf_input": true, "supports_response_schema": true, @@ -4352,21 +5596,39 @@ "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 1.5e-6, + "output_cost_per_token": 0.0000015, + "supports_function_calling": true, + "supports_pdf_input": true, + "supports_response_schema": true, + "supports_tool_choice": true, + "supports_vision": true + }, + "bedrock/us-gov-east-1/claude-sonnet-4-5-20250929-v1:0": { + "input_cost_per_token": 0.0000033, + "litellm_provider": "bedrock", + "max_input_tokens": 200000, + "max_output_tokens": 4096, + "max_tokens": 4096, + "mode": "chat", + "output_cost_per_token": 0.0000165, + "supports_assistant_prefill": true, + "supports_computer_use": true, "supports_function_calling": true, "supports_pdf_input": true, + "supports_prompt_caching": true, + "supports_reasoning": true, "supports_response_schema": true, "supports_tool_choice": true, "supports_vision": true }, "bedrock/us-gov-east-1/meta.llama3-70b-instruct-v1:0": { - "input_cost_per_token": 2.65e-6, + "input_cost_per_token": 0.00000265, "litellm_provider": "bedrock", "max_input_tokens": 8000, "max_output_tokens": 2048, "max_tokens": 2048, "mode": "chat", - "output_cost_per_token": 3.5e-6, + "output_cost_per_token": 0.0000035, "supports_pdf_input": true }, "bedrock/us-gov-east-1/meta.llama3-8b-instruct-v1:0": { @@ -4376,7 +5638,7 @@ "max_output_tokens": 2048, "max_tokens": 2048, "mode": "chat", - "output_cost_per_token": 2.65e-6, + "output_cost_per_token": 0.00000265, "supports_pdf_input": true }, "bedrock/us-gov-west-1/amazon.nova-pro-v1:0": { @@ -4386,7 +5648,7 @@ "max_output_tokens": 10000, "max_tokens": 10000, "mode": "chat", - "output_cost_per_token": 3.84e-6, + "output_cost_per_token": 0.00000384, "supports_function_calling": true, "supports_pdf_input": true, "supports_prompt_caching": true, @@ -4399,7 +5661,7 @@ "max_input_tokens": 8192, "max_tokens": 8192, "mode": "embedding", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "output_vector_size": 1536 }, "bedrock/us-gov-west-1/amazon.titan-embed-text-v2:0": { @@ -4408,17 +5670,17 @@ "max_input_tokens": 8192, "max_tokens": 8192, "mode": "embedding", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "output_vector_size": 1024 }, "bedrock/us-gov-west-1/amazon.titan-text-express-v1": { - "input_cost_per_token": 1.3e-6, + "input_cost_per_token": 0.0000013, "litellm_provider": "bedrock", "max_input_tokens": 42000, "max_output_tokens": 8000, "max_tokens": 8000, "mode": "chat", - "output_cost_per_token": 1.7e-6 + "output_cost_per_token": 0.0000017 }, "bedrock/us-gov-west-1/amazon.titan-text-lite-v1": { "input_cost_per_token": 3e-7, @@ -4436,18 +5698,18 @@ "max_output_tokens": 32000, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 1.5e-6 + "output_cost_per_token": 0.0000015 }, "bedrock/us-gov-west-1/anthropic.claude-3-7-sonnet-20250219-v1:0": { - "cache_creation_input_token_cost": 4.5e-6, + "cache_creation_input_token_cost": 0.0000045, "cache_read_input_token_cost": 3.6e-7, - "input_cost_per_token": 3.6e-6, + "input_cost_per_token": 0.0000036, "litellm_provider": "bedrock", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 1.8e-5, + "output_cost_per_token": 0.000018, "supports_assistant_prefill": true, "supports_computer_use": true, "supports_function_calling": true, @@ -4459,13 +5721,13 @@ "supports_vision": true }, "bedrock/us-gov-west-1/anthropic.claude-3-5-sonnet-20240620-v1:0": { - "input_cost_per_token": 3.6e-6, + "input_cost_per_token": 0.0000036, "litellm_provider": "bedrock", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 1.8e-5, + "output_cost_per_token": 0.000018, "supports_function_calling": true, "supports_pdf_input": true, "supports_response_schema": true, @@ -4479,21 +5741,39 @@ "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 1.5e-6, + "output_cost_per_token": 0.0000015, + "supports_function_calling": true, + "supports_pdf_input": true, + "supports_response_schema": true, + "supports_tool_choice": true, + "supports_vision": true + }, + "bedrock/us-gov-west-1/claude-sonnet-4-5-20250929-v1:0": { + "input_cost_per_token": 0.0000033, + "litellm_provider": "bedrock", + "max_input_tokens": 200000, + "max_output_tokens": 4096, + "max_tokens": 4096, + "mode": "chat", + "output_cost_per_token": 0.0000165, + "supports_assistant_prefill": true, + "supports_computer_use": true, "supports_function_calling": true, "supports_pdf_input": true, + "supports_prompt_caching": true, + "supports_reasoning": true, "supports_response_schema": true, "supports_tool_choice": true, "supports_vision": true }, "bedrock/us-gov-west-1/meta.llama3-70b-instruct-v1:0": { - "input_cost_per_token": 2.65e-6, + "input_cost_per_token": 0.00000265, "litellm_provider": "bedrock", "max_input_tokens": 8000, "max_output_tokens": 2048, "max_tokens": 2048, "mode": "chat", - "output_cost_per_token": 3.5e-6, + "output_cost_per_token": 0.0000035, "supports_pdf_input": true }, "bedrock/us-gov-west-1/meta.llama3-8b-instruct-v1:0": { @@ -4503,17 +5783,17 @@ "max_output_tokens": 2048, "max_tokens": 2048, "mode": "chat", - "output_cost_per_token": 2.65e-6, + "output_cost_per_token": 0.00000265, "supports_pdf_input": true }, "bedrock/us-west-1/meta.llama3-70b-instruct-v1:0": { - "input_cost_per_token": 2.65e-6, + "input_cost_per_token": 0.00000265, "litellm_provider": "bedrock", "max_input_tokens": 8192, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 3.5e-6 + "output_cost_per_token": 0.0000035 }, "bedrock/us-west-1/meta.llama3-8b-instruct-v1:0": { "input_cost_per_token": 3e-7, @@ -4589,27 +5869,27 @@ "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 2.4e-6, + "output_cost_per_token": 0.0000024, "supports_tool_choice": true }, "bedrock/us-west-2/anthropic.claude-v1": { - "input_cost_per_token": 8e-6, + "input_cost_per_token": 0.000008, "litellm_provider": "bedrock", "max_input_tokens": 100000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 2.4e-5, + "output_cost_per_token": 0.000024, "supports_tool_choice": true }, "bedrock/us-west-2/anthropic.claude-v2:1": { - "input_cost_per_token": 8e-6, + "input_cost_per_token": 0.000008, "litellm_provider": "bedrock", "max_input_tokens": 100000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 2.4e-5, + "output_cost_per_token": 0.000024, "supports_tool_choice": true }, "bedrock/us-west-2/mistral.mistral-7b-instruct-v0:2": { @@ -4623,13 +5903,13 @@ "supports_tool_choice": true }, "bedrock/us-west-2/mistral.mistral-large-2402-v1:0": { - "input_cost_per_token": 8e-6, + "input_cost_per_token": 0.000008, "litellm_provider": "bedrock", "max_input_tokens": 32000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 2.4e-5, + "output_cost_per_token": 0.000024, "supports_function_calling": true }, "bedrock/us-west-2/mistral.mixtral-8x7b-instruct-v0:1": { @@ -4643,7 +5923,7 @@ "supports_tool_choice": true }, "bedrock/us.anthropic.claude-3-5-haiku-20241022-v1:0": { - "cache_creation_input_token_cost": 1e-6, + "cache_creation_input_token_cost": 0.000001, "cache_read_input_token_cost": 8e-8, "input_cost_per_token": 8e-7, "litellm_provider": "bedrock", @@ -4651,7 +5931,7 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 4e-6, + "output_cost_per_token": 0.000004, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_pdf_input": true, @@ -4666,7 +5946,7 @@ "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 1.2e-6, + "output_cost_per_token": 0.0000012, "supports_function_calling": true, "supports_tool_choice": true }, @@ -4692,7 +5972,7 @@ "supports_function_calling": true, "supports_tool_choice": true }, - "cerebras/openai/gpt-oss-120b": { + "cerebras/gpt-oss-120b": { "input_cost_per_token": 2.5e-7, "litellm_provider": "cerebras", "max_input_tokens": 131072, @@ -4795,13 +6075,13 @@ "output_cost_per_token": 5e-7 }, "chatgpt-4o-latest": { - "input_cost_per_token": 5e-6, + "input_cost_per_token": 0.000005, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_pdf_input": true, @@ -4811,8 +6091,8 @@ "supports_vision": true }, "claude-3-5-haiku-20241022": { - "cache_creation_input_token_cost": 1e-6, - "cache_creation_input_token_cost_above_1hr": 6e-6, + "cache_creation_input_token_cost": 0.000001, + "cache_creation_input_token_cost_above_1hr": 0.000006, "cache_read_input_token_cost": 8e-8, "deprecation_date": "2025-10-01", "input_cost_per_token": 8e-7, @@ -4821,7 +6101,7 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 4e-6, + "output_cost_per_token": 0.000004, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -4838,17 +6118,17 @@ "tool_use_system_prompt_tokens": 264 }, "claude-3-5-haiku-latest": { - "cache_creation_input_token_cost": 1.25e-6, - "cache_creation_input_token_cost_above_1hr": 6e-6, + "cache_creation_input_token_cost": 0.00000125, + "cache_creation_input_token_cost_above_1hr": 0.000006, "cache_read_input_token_cost": 1e-7, "deprecation_date": "2025-10-01", - "input_cost_per_token": 1e-6, + "input_cost_per_token": 0.000001, "litellm_provider": "anthropic", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 5e-6, + "output_cost_per_token": 0.000005, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -4865,16 +6145,16 @@ "tool_use_system_prompt_tokens": 264 }, "claude-haiku-4-5-20251001": { - "cache_creation_input_token_cost": 1.25e-6, - "cache_creation_input_token_cost_above_1hr": 2e-6, + "cache_creation_input_token_cost": 0.00000125, + "cache_creation_input_token_cost_above_1hr": 0.000002, "cache_read_input_token_cost": 1e-7, - "input_cost_per_token": 1e-6, + "input_cost_per_token": 0.000001, "litellm_provider": "anthropic", "max_input_tokens": 200000, "max_output_tokens": 64000, "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 5e-6, + "output_cost_per_token": 0.000005, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_computer_use": true, @@ -4886,16 +6166,16 @@ "supports_vision": true }, "claude-haiku-4-5": { - "cache_creation_input_token_cost": 1.25e-6, - "cache_creation_input_token_cost_above_1hr": 2e-6, + "cache_creation_input_token_cost": 0.00000125, + "cache_creation_input_token_cost_above_1hr": 0.000002, "cache_read_input_token_cost": 1e-7, - "input_cost_per_token": 1e-6, + "input_cost_per_token": 0.000001, "litellm_provider": "anthropic", "max_input_tokens": 200000, "max_output_tokens": 64000, "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 5e-6, + "output_cost_per_token": 0.000005, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_computer_use": true, @@ -4907,17 +6187,17 @@ "supports_vision": true }, "claude-3-5-sonnet-20240620": { - "cache_creation_input_token_cost": 3.75e-6, - "cache_creation_input_token_cost_above_1hr": 6e-6, + "cache_creation_input_token_cost": 0.00000375, + "cache_creation_input_token_cost_above_1hr": 0.000006, "cache_read_input_token_cost": 3e-7, "deprecation_date": "2025-06-01", - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "anthropic", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_pdf_input": true, @@ -4928,17 +6208,17 @@ "tool_use_system_prompt_tokens": 159 }, "claude-3-5-sonnet-20241022": { - "cache_creation_input_token_cost": 3.75e-6, - "cache_creation_input_token_cost_above_1hr": 6e-6, + "cache_creation_input_token_cost": 0.00000375, + "cache_creation_input_token_cost_above_1hr": 0.000006, "cache_read_input_token_cost": 3e-7, "deprecation_date": "2025-10-01", - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "anthropic", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -4956,17 +6236,17 @@ "tool_use_system_prompt_tokens": 159 }, "claude-3-5-sonnet-latest": { - "cache_creation_input_token_cost": 3.75e-6, - "cache_creation_input_token_cost_above_1hr": 6e-6, + "cache_creation_input_token_cost": 0.00000375, + "cache_creation_input_token_cost_above_1hr": 0.000006, "cache_read_input_token_cost": 3e-7, "deprecation_date": "2025-06-01", - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "anthropic", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -4984,17 +6264,17 @@ "tool_use_system_prompt_tokens": 159 }, "claude-3-7-sonnet-20250219": { - "cache_creation_input_token_cost": 3.75e-6, - "cache_creation_input_token_cost_above_1hr": 6e-6, + "cache_creation_input_token_cost": 0.00000375, + "cache_creation_input_token_cost_above_1hr": 0.000006, "cache_read_input_token_cost": 3e-7, "deprecation_date": "2026-02-19", - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "anthropic", "max_input_tokens": 200000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -5013,17 +6293,17 @@ "tool_use_system_prompt_tokens": 159 }, "claude-3-7-sonnet-latest": { - "cache_creation_input_token_cost": 3.75e-6, - "cache_creation_input_token_cost_above_1hr": 6e-6, + "cache_creation_input_token_cost": 0.00000375, + "cache_creation_input_token_cost_above_1hr": 0.000006, "cache_read_input_token_cost": 3e-7, "deprecation_date": "2025-06-01", - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "anthropic", "max_input_tokens": 200000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -5042,7 +6322,7 @@ }, "claude-3-haiku-20240307": { "cache_creation_input_token_cost": 3e-7, - "cache_creation_input_token_cost_above_1hr": 6e-6, + "cache_creation_input_token_cost_above_1hr": 0.000006, "cache_read_input_token_cost": 3e-8, "input_cost_per_token": 2.5e-7, "litellm_provider": "anthropic", @@ -5050,7 +6330,7 @@ "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 1.25e-6, + "output_cost_per_token": 0.00000125, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_prompt_caching": true, @@ -5060,17 +6340,17 @@ "tool_use_system_prompt_tokens": 264 }, "claude-3-opus-20240229": { - "cache_creation_input_token_cost": 1.875e-5, - "cache_creation_input_token_cost_above_1hr": 6e-6, - "cache_read_input_token_cost": 1.5e-6, + "cache_creation_input_token_cost": 0.00001875, + "cache_creation_input_token_cost_above_1hr": 0.000006, + "cache_read_input_token_cost": 0.0000015, "deprecation_date": "2026-05-01", - "input_cost_per_token": 1.5e-5, + "input_cost_per_token": 0.000015, "litellm_provider": "anthropic", "max_input_tokens": 200000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 7.5e-5, + "output_cost_per_token": 0.000075, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_prompt_caching": true, @@ -5080,17 +6360,17 @@ "tool_use_system_prompt_tokens": 395 }, "claude-3-opus-latest": { - "cache_creation_input_token_cost": 1.875e-5, - "cache_creation_input_token_cost_above_1hr": 6e-6, - "cache_read_input_token_cost": 1.5e-6, + "cache_creation_input_token_cost": 0.00001875, + "cache_creation_input_token_cost_above_1hr": 0.000006, + "cache_read_input_token_cost": 0.0000015, "deprecation_date": "2025-03-01", - "input_cost_per_token": 1.5e-5, + "input_cost_per_token": 0.000015, "litellm_provider": "anthropic", "max_input_tokens": 200000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 7.5e-5, + "output_cost_per_token": 0.000075, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_prompt_caching": true, @@ -5100,15 +6380,15 @@ "tool_use_system_prompt_tokens": 395 }, "claude-4-opus-20250514": { - "cache_creation_input_token_cost": 1.875e-5, - "cache_read_input_token_cost": 1.5e-6, - "input_cost_per_token": 1.5e-5, + "cache_creation_input_token_cost": 0.00001875, + "cache_read_input_token_cost": 0.0000015, + "input_cost_per_token": 0.000015, "litellm_provider": "anthropic", "max_input_tokens": 200000, "max_output_tokens": 32000, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 7.5e-5, + "output_cost_per_token": 0.000075, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -5126,19 +6406,19 @@ "tool_use_system_prompt_tokens": 159 }, "claude-4-sonnet-20250514": { - "cache_creation_input_token_cost": 3.75e-6, - "cache_creation_input_token_cost_above_200k_tokens": 7.5e-6, + "cache_creation_input_token_cost": 0.00000375, + "cache_creation_input_token_cost_above_200k_tokens": 0.0000075, "cache_read_input_token_cost": 3e-7, "cache_read_input_token_cost_above_200k_tokens": 6e-7, - "input_cost_per_token": 3e-6, - "input_cost_per_token_above_200k_tokens": 6e-6, + "input_cost_per_token": 0.000003, + "input_cost_per_token_above_200k_tokens": 0.000006, "litellm_provider": "anthropic", "max_input_tokens": 1000000, "max_output_tokens": 64000, "max_tokens": 1000000, "mode": "chat", - "output_cost_per_token": 1.5e-5, - "output_cost_per_token_above_200k_tokens": 2.25e-5, + "output_cost_per_token": 0.000015, + "output_cost_per_token_above_200k_tokens": 0.0000225, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -5156,19 +6436,19 @@ "tool_use_system_prompt_tokens": 159 }, "claude-sonnet-4-5": { - "cache_creation_input_token_cost": 3.75e-6, + "cache_creation_input_token_cost": 0.00000375, "cache_read_input_token_cost": 3e-7, - "input_cost_per_token": 3e-6, - "input_cost_per_token_above_200k_tokens": 6e-6, - "output_cost_per_token_above_200k_tokens": 2.25e-5, - "cache_creation_input_token_cost_above_200k_tokens": 7.5e-6, + "input_cost_per_token": 0.000003, + "input_cost_per_token_above_200k_tokens": 0.000006, + "output_cost_per_token_above_200k_tokens": 0.0000225, + "cache_creation_input_token_cost_above_200k_tokens": 0.0000075, "cache_read_input_token_cost_above_200k_tokens": 6e-7, "litellm_provider": "anthropic", "max_input_tokens": 200000, "max_output_tokens": 64000, "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -5186,19 +6466,19 @@ "tool_use_system_prompt_tokens": 346 }, "claude-sonnet-4-5-20250929": { - "cache_creation_input_token_cost": 3.75e-6, + "cache_creation_input_token_cost": 0.00000375, "cache_read_input_token_cost": 3e-7, - "input_cost_per_token": 3e-6, - "input_cost_per_token_above_200k_tokens": 6e-6, - "output_cost_per_token_above_200k_tokens": 2.25e-5, - "cache_creation_input_token_cost_above_200k_tokens": 7.5e-6, + "input_cost_per_token": 0.000003, + "input_cost_per_token_above_200k_tokens": 0.000006, + "output_cost_per_token_above_200k_tokens": 0.0000225, + "cache_creation_input_token_cost_above_200k_tokens": 0.0000075, "cache_read_input_token_cost_above_200k_tokens": 6e-7, "litellm_provider": "anthropic", "max_input_tokens": 200000, "max_output_tokens": 64000, "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -5216,17 +6496,42 @@ "supports_web_search": true, "tool_use_system_prompt_tokens": 346 }, + "claude-sonnet-4-5-20250929-v1:0": { + "cache_creation_input_token_cost": 0.00000375, + "cache_read_input_token_cost": 3e-7, + "input_cost_per_token": 0.000003, + "input_cost_per_token_above_200k_tokens": 0.000006, + "output_cost_per_token_above_200k_tokens": 0.0000225, + "cache_creation_input_token_cost_above_200k_tokens": 0.0000075, + "cache_read_input_token_cost_above_200k_tokens": 6e-7, + "litellm_provider": "bedrock", + "max_input_tokens": 200000, + "max_output_tokens": 64000, + "max_tokens": 64000, + "mode": "chat", + "output_cost_per_token": 0.000015, + "supports_assistant_prefill": true, + "supports_computer_use": true, + "supports_function_calling": true, + "supports_pdf_input": true, + "supports_prompt_caching": true, + "supports_reasoning": true, + "supports_response_schema": true, + "supports_tool_choice": true, + "supports_vision": true, + "tool_use_system_prompt_tokens": 159 + }, "claude-opus-4-1": { - "cache_creation_input_token_cost": 1.875e-5, - "cache_creation_input_token_cost_above_1hr": 3e-5, - "cache_read_input_token_cost": 1.5e-6, - "input_cost_per_token": 1.5e-5, + "cache_creation_input_token_cost": 0.00001875, + "cache_creation_input_token_cost_above_1hr": 0.00003, + "cache_read_input_token_cost": 0.0000015, + "input_cost_per_token": 0.000015, "litellm_provider": "anthropic", "max_input_tokens": 200000, "max_output_tokens": 32000, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 7.5e-5, + "output_cost_per_token": 0.000075, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -5244,17 +6549,17 @@ "tool_use_system_prompt_tokens": 159 }, "claude-opus-4-1-20250805": { - "cache_creation_input_token_cost": 1.875e-5, - "cache_creation_input_token_cost_above_1hr": 3e-5, - "cache_read_input_token_cost": 1.5e-6, - "input_cost_per_token": 1.5e-5, + "cache_creation_input_token_cost": 0.00001875, + "cache_creation_input_token_cost_above_1hr": 0.00003, + "cache_read_input_token_cost": 0.0000015, + "input_cost_per_token": 0.000015, "deprecation_date": "2026-08-05", "litellm_provider": "anthropic", "max_input_tokens": 200000, "max_output_tokens": 32000, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 7.5e-5, + "output_cost_per_token": 0.000075, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -5272,17 +6577,17 @@ "tool_use_system_prompt_tokens": 159 }, "claude-opus-4-20250514": { - "cache_creation_input_token_cost": 1.875e-5, - "cache_creation_input_token_cost_above_1hr": 3e-5, - "cache_read_input_token_cost": 1.5e-6, - "input_cost_per_token": 1.5e-5, + "cache_creation_input_token_cost": 0.00001875, + "cache_creation_input_token_cost_above_1hr": 0.00003, + "cache_read_input_token_cost": 0.0000015, + "input_cost_per_token": 0.000015, "deprecation_date": "2026-05-14", "litellm_provider": "anthropic", "max_input_tokens": 200000, "max_output_tokens": 32000, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 7.5e-5, + "output_cost_per_token": 0.000075, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -5301,20 +6606,20 @@ }, "claude-sonnet-4-20250514": { "deprecation_date": "2026-05-14", - "cache_creation_input_token_cost": 3.75e-6, - "cache_creation_input_token_cost_above_1hr": 6e-6, + "cache_creation_input_token_cost": 0.00000375, + "cache_creation_input_token_cost_above_1hr": 0.000006, "cache_read_input_token_cost": 3e-7, - "input_cost_per_token": 3e-6, - "input_cost_per_token_above_200k_tokens": 6e-6, - "output_cost_per_token_above_200k_tokens": 2.25e-5, - "cache_creation_input_token_cost_above_200k_tokens": 7.5e-6, + "input_cost_per_token": 0.000003, + "input_cost_per_token_above_200k_tokens": 0.000006, + "output_cost_per_token_above_200k_tokens": 0.0000225, + "cache_creation_input_token_cost_above_200k_tokens": 0.0000075, "cache_read_input_token_cost_above_200k_tokens": 6e-7, "litellm_provider": "anthropic", "max_input_tokens": 1000000, "max_output_tokens": 64000, "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -5332,40 +6637,40 @@ "tool_use_system_prompt_tokens": 159 }, "cloudflare/@cf/meta/llama-2-7b-chat-fp16": { - "input_cost_per_token": 1.923e-6, + "input_cost_per_token": 0.000001923, "litellm_provider": "cloudflare", "max_input_tokens": 3072, "max_output_tokens": 3072, "max_tokens": 3072, "mode": "chat", - "output_cost_per_token": 1.923e-6 + "output_cost_per_token": 0.000001923 }, "cloudflare/@cf/meta/llama-2-7b-chat-int8": { - "input_cost_per_token": 1.923e-6, + "input_cost_per_token": 0.000001923, "litellm_provider": "cloudflare", "max_input_tokens": 2048, "max_output_tokens": 2048, "max_tokens": 2048, "mode": "chat", - "output_cost_per_token": 1.923e-6 + "output_cost_per_token": 0.000001923 }, "cloudflare/@cf/mistral/mistral-7b-instruct-v0.1": { - "input_cost_per_token": 1.923e-6, + "input_cost_per_token": 0.000001923, "litellm_provider": "cloudflare", "max_input_tokens": 8192, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 1.923e-6 + "output_cost_per_token": 0.000001923 }, "cloudflare/@hf/thebloke/codellama-7b-instruct-awq": { - "input_cost_per_token": 1.923e-6, + "input_cost_per_token": 0.000001923, "litellm_provider": "cloudflare", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 1.923e-6 + "output_cost_per_token": 0.000001923 }, "code-bison": { "input_cost_per_character": 2.5e-7, @@ -5547,41 +6852,48 @@ "supports_tool_choice": true }, "codestral/codestral-2405": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "codestral", "max_input_tokens": 32000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "source": "https://docs.mistral.ai/capabilities/code_generation/", "supports_assistant_prefill": true, "supports_tool_choice": true }, "codestral/codestral-latest": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "codestral", "max_input_tokens": 32000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "source": "https://docs.mistral.ai/capabilities/code_generation/", "supports_assistant_prefill": true, "supports_tool_choice": true }, "codex-mini-latest": { "cache_read_input_token_cost": 3.75e-7, - "input_cost_per_token": 1.5e-6, + "input_cost_per_token": 0.0000015, "litellm_provider": "openai", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "responses", - "output_cost_per_token": 6e-6, - "supported_endpoints": ["/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "output_cost_per_token": 0.000006, + "supported_endpoints": [ + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_pdf_input": true, @@ -5603,13 +6915,13 @@ "supports_tool_choice": true }, "cohere.command-r-plus-v1:0": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "bedrock", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "supports_tool_choice": true }, "cohere.command-r-v1:0": { @@ -5619,17 +6931,17 @@ "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 1.5e-6, + "output_cost_per_token": 0.0000015, "supports_tool_choice": true }, "cohere.command-text-v14": { - "input_cost_per_token": 1.5e-6, + "input_cost_per_token": 0.0000015, "litellm_provider": "bedrock", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 2e-6, + "output_cost_per_token": 0.000002, "supports_tool_choice": true }, "cohere.embed-english-v3": { @@ -5638,7 +6950,7 @@ "max_input_tokens": 512, "max_tokens": 512, "mode": "embedding", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "supports_embedding_image_input": true }, "cohere.embed-multilingual-v3": { @@ -5647,7 +6959,7 @@ "max_input_tokens": 512, "max_tokens": 512, "mode": "embedding", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "supports_embedding_image_input": true }, "cohere.embed-v4:0": { @@ -5656,7 +6968,7 @@ "max_input_tokens": 128000, "max_tokens": 128000, "mode": "embedding", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "output_vector_size": 1536, "supports_embedding_image_input": true }, @@ -5666,13 +6978,13 @@ "max_input_tokens": 128000, "max_tokens": 128000, "mode": "embedding", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "output_vector_size": 1536, "supports_embedding_image_input": true }, "cohere.rerank-v3-5:0": { "input_cost_per_query": 0.002, - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "bedrock", "max_document_chunks_per_query": 100, "max_input_tokens": 32000, @@ -5681,25 +6993,25 @@ "max_tokens": 32000, "max_tokens_per_document_chunk": 512, "mode": "rerank", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "command": { - "input_cost_per_token": 1e-6, + "input_cost_per_token": 0.000001, "litellm_provider": "cohere", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "completion", - "output_cost_per_token": 2e-6 + "output_cost_per_token": 0.000002 }, "command-a-03-2025": { - "input_cost_per_token": 2.5e-6, + "input_cost_per_token": 0.0000025, "litellm_provider": "cohere_chat", "max_input_tokens": 256000, "max_output_tokens": 8000, "max_tokens": 8000, "mode": "chat", - "output_cost_per_token": 1e-5, + "output_cost_per_token": 0.00001, "supports_function_calling": true, "supports_tool_choice": true }, @@ -5714,13 +7026,13 @@ "supports_tool_choice": true }, "command-nightly": { - "input_cost_per_token": 1e-6, + "input_cost_per_token": 0.000001, "litellm_provider": "cohere", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "completion", - "output_cost_per_token": 2e-6 + "output_cost_per_token": 0.000002 }, "command-r": { "input_cost_per_token": 1.5e-7, @@ -5745,24 +7057,24 @@ "supports_tool_choice": true }, "command-r-plus": { - "input_cost_per_token": 2.5e-6, + "input_cost_per_token": 0.0000025, "litellm_provider": "cohere_chat", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 1e-5, + "output_cost_per_token": 0.00001, "supports_function_calling": true, "supports_tool_choice": true }, "command-r-plus-08-2024": { - "input_cost_per_token": 2.5e-6, + "input_cost_per_token": 0.0000025, "litellm_provider": "cohere_chat", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 1e-5, + "output_cost_per_token": 0.00001, "supports_function_calling": true, "supports_tool_choice": true }, @@ -5779,16 +7091,23 @@ "supports_tool_choice": true }, "computer-use-preview": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "azure", "max_input_tokens": 8192, "max_output_tokens": 1024, "max_tokens": 1024, "mode": "chat", - "output_cost_per_token": 1.2e-5, - "supported_endpoints": ["/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "output_cost_per_token": 0.000012, + "supported_endpoints": [ + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": false, @@ -5806,9 +7125,11 @@ "max_output_tokens": 8192, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 1.7e-6, + "output_cost_per_token": 0.0000017, "source": "https://api-docs.deepseek.com/quick_start/pricing", - "supported_endpoints": ["/v1/chat/completions"], + "supported_endpoints": [ + "/v1/chat/completions" + ], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -5825,9 +7146,11 @@ "max_output_tokens": 65536, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 1.7e-6, + "output_cost_per_token": 0.0000017, "source": "https://api-docs.deepseek.com/quick_start/pricing", - "supported_endpoints": ["/v1/chat/completions"], + "supported_endpoints": [ + "/v1/chat/completions" + ], "supports_function_calling": false, "supports_native_streaming": true, "supports_parallel_function_calling": false, @@ -5844,7 +7167,7 @@ "max_output_tokens": 16384, "max_tokens": 1000000, "mode": "chat", - "output_cost_per_token": 1.5e-6, + "output_cost_per_token": 0.0000015, "source": "https://www.alibabacloud.com/help/en/model-studio/models", "supports_function_calling": true, "supports_reasoning": true, @@ -5864,12 +7187,18 @@ { "input_cost_per_token": 5e-8, "output_cost_per_token": 4e-7, - "range": [0, 256000.0] + "range": [ + 0, + 256000 + ] }, { "input_cost_per_token": 2.5e-7, - "output_cost_per_token": 2e-6, - "range": [256000.0, 1000000.0] + "output_cost_per_token": 0.000002, + "range": [ + 256000, + 1000000 + ] } ] }, @@ -5887,23 +7216,29 @@ { "input_cost_per_token": 5e-8, "output_cost_per_token": 4e-7, - "range": [0, 256000.0] + "range": [ + 0, + 256000 + ] }, { "input_cost_per_token": 2.5e-7, - "output_cost_per_token": 2e-6, - "range": [256000.0, 1000000.0] + "output_cost_per_token": 0.000002, + "range": [ + 256000, + 1000000 + ] } ] }, "dashscope/qwen-max": { - "input_cost_per_token": 1.6e-6, + "input_cost_per_token": 0.0000016, "litellm_provider": "dashscope", "max_input_tokens": 30720, "max_output_tokens": 8192, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 6.4e-6, + "output_cost_per_token": 0.0000064, "source": "https://www.alibabacloud.com/help/en/model-studio/models", "supports_function_calling": true, "supports_reasoning": true, @@ -5916,7 +7251,7 @@ "max_output_tokens": 16384, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 1.2e-6, + "output_cost_per_token": 0.0000012, "source": "https://www.alibabacloud.com/help/en/model-studio/models", "supports_function_calling": true, "supports_reasoning": true, @@ -5929,7 +7264,7 @@ "max_output_tokens": 8192, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 1.2e-6, + "output_cost_per_token": 0.0000012, "source": "https://www.alibabacloud.com/help/en/model-studio/models", "supports_function_calling": true, "supports_reasoning": true, @@ -5942,8 +7277,8 @@ "max_output_tokens": 16384, "max_tokens": 131072, "mode": "chat", - "output_cost_per_reasoning_token": 4e-6, - "output_cost_per_token": 1.2e-6, + "output_cost_per_reasoning_token": 0.000004, + "output_cost_per_token": 0.0000012, "source": "https://www.alibabacloud.com/help/en/model-studio/models", "supports_function_calling": true, "supports_reasoning": true, @@ -5956,8 +7291,8 @@ "max_output_tokens": 16384, "max_tokens": 131072, "mode": "chat", - "output_cost_per_reasoning_token": 4e-6, - "output_cost_per_token": 1.2e-6, + "output_cost_per_reasoning_token": 0.000004, + "output_cost_per_token": 0.0000012, "source": "https://www.alibabacloud.com/help/en/model-studio/models", "supports_function_calling": true, "supports_reasoning": true, @@ -5976,15 +7311,21 @@ "tiered_pricing": [ { "input_cost_per_token": 4e-7, - "output_cost_per_reasoning_token": 4e-6, - "output_cost_per_token": 1.2e-6, - "range": [0, 256000.0] + "output_cost_per_reasoning_token": 0.000004, + "output_cost_per_token": 0.0000012, + "range": [ + 0, + 256000 + ] }, { - "input_cost_per_token": 1.2e-6, - "output_cost_per_reasoning_token": 1.2e-5, - "output_cost_per_token": 3.6e-6, - "range": [256000.0, 1000000.0] + "input_cost_per_token": 0.0000012, + "output_cost_per_reasoning_token": 0.000012, + "output_cost_per_token": 0.0000036, + "range": [ + 256000, + 1000000 + ] } ] }, @@ -6001,15 +7342,21 @@ "tiered_pricing": [ { "input_cost_per_token": 4e-7, - "output_cost_per_reasoning_token": 4e-6, - "output_cost_per_token": 1.2e-6, - "range": [0, 256000.0] + "output_cost_per_reasoning_token": 0.000004, + "output_cost_per_token": 0.0000012, + "range": [ + 0, + 256000 + ] }, { - "input_cost_per_token": 1.2e-6, - "output_cost_per_reasoning_token": 1.2e-5, - "output_cost_per_token": 3.6e-6, - "range": [256000.0, 1000000.0] + "input_cost_per_token": 0.0000012, + "output_cost_per_reasoning_token": 0.000012, + "output_cost_per_token": 0.0000036, + "range": [ + 256000, + 1000000 + ] } ] }, @@ -6026,15 +7373,21 @@ "tiered_pricing": [ { "input_cost_per_token": 4e-7, - "output_cost_per_reasoning_token": 4e-6, - "output_cost_per_token": 1.2e-6, - "range": [0, 256000.0] + "output_cost_per_reasoning_token": 0.000004, + "output_cost_per_token": 0.0000012, + "range": [ + 0, + 256000 + ] }, { - "input_cost_per_token": 1.2e-6, - "output_cost_per_reasoning_token": 1.2e-5, - "output_cost_per_token": 3.6e-6, - "range": [256000.0, 1000000.0] + "input_cost_per_token": 0.0000012, + "output_cost_per_reasoning_token": 0.000012, + "output_cost_per_token": 0.0000036, + "range": [ + 256000, + 1000000 + ] } ] }, @@ -6118,26 +7471,38 @@ { "cache_read_input_token_cost": 8e-8, "input_cost_per_token": 3e-7, - "output_cost_per_token": 1.5e-6, - "range": [0, 32000.0] + "output_cost_per_token": 0.0000015, + "range": [ + 0, + 32000 + ] }, { "cache_read_input_token_cost": 1.2e-7, "input_cost_per_token": 5e-7, - "output_cost_per_token": 2.5e-6, - "range": [32000.0, 128000.0] + "output_cost_per_token": 0.0000025, + "range": [ + 32000, + 128000 + ] }, { "cache_read_input_token_cost": 2e-7, "input_cost_per_token": 8e-7, - "output_cost_per_token": 4e-6, - "range": [128000.0, 256000.0] + "output_cost_per_token": 0.000004, + "range": [ + 128000, + 256000 + ] }, { "cache_read_input_token_cost": 4e-7, - "input_cost_per_token": 1.6e-6, - "output_cost_per_token": 9.6e-6, - "range": [256000.0, 1000000.0] + "input_cost_per_token": 0.0000016, + "output_cost_per_token": 0.0000096, + "range": [ + 256000, + 1000000 + ] } ] }, @@ -6154,23 +7519,35 @@ "tiered_pricing": [ { "input_cost_per_token": 3e-7, - "output_cost_per_token": 1.5e-6, - "range": [0, 32000.0] + "output_cost_per_token": 0.0000015, + "range": [ + 0, + 32000 + ] }, { "input_cost_per_token": 5e-7, - "output_cost_per_token": 2.5e-6, - "range": [32000.0, 128000.0] + "output_cost_per_token": 0.0000025, + "range": [ + 32000, + 128000 + ] }, { "input_cost_per_token": 8e-7, - "output_cost_per_token": 4e-6, - "range": [128000.0, 256000.0] + "output_cost_per_token": 0.000004, + "range": [ + 128000, + 256000 + ] }, { - "input_cost_per_token": 1.6e-6, - "output_cost_per_token": 9.6e-6, - "range": [256000.0, 1000000.0] + "input_cost_per_token": 0.0000016, + "output_cost_per_token": 0.0000096, + "range": [ + 256000, + 1000000 + ] } ] }, @@ -6187,27 +7564,39 @@ "tiered_pricing": [ { "cache_read_input_token_cost": 1e-7, - "input_cost_per_token": 1e-6, - "output_cost_per_token": 5e-6, - "range": [0, 32000.0] + "input_cost_per_token": 0.000001, + "output_cost_per_token": 0.000005, + "range": [ + 0, + 32000 + ] }, { "cache_read_input_token_cost": 1.8e-7, - "input_cost_per_token": 1.8e-6, - "output_cost_per_token": 9e-6, - "range": [32000.0, 128000.0] + "input_cost_per_token": 0.0000018, + "output_cost_per_token": 0.000009, + "range": [ + 32000, + 128000 + ] }, { "cache_read_input_token_cost": 3e-7, - "input_cost_per_token": 3e-6, - "output_cost_per_token": 1.5e-5, - "range": [128000.0, 256000.0] + "input_cost_per_token": 0.000003, + "output_cost_per_token": 0.000015, + "range": [ + 128000, + 256000 + ] }, { "cache_read_input_token_cost": 6e-7, - "input_cost_per_token": 6e-6, - "output_cost_per_token": 6e-5, - "range": [256000.0, 1000000.0] + "input_cost_per_token": 0.000006, + "output_cost_per_token": 0.00006, + "range": [ + 256000, + 1000000 + ] } ] }, @@ -6223,24 +7612,36 @@ "supports_tool_choice": true, "tiered_pricing": [ { - "input_cost_per_token": 1e-6, - "output_cost_per_token": 5e-6, - "range": [0, 32000.0] + "input_cost_per_token": 0.000001, + "output_cost_per_token": 0.000005, + "range": [ + 0, + 32000 + ] }, { - "input_cost_per_token": 1.8e-6, - "output_cost_per_token": 9e-6, - "range": [32000.0, 128000.0] + "input_cost_per_token": 0.0000018, + "output_cost_per_token": 0.000009, + "range": [ + 32000, + 128000 + ] }, { - "input_cost_per_token": 3e-6, - "output_cost_per_token": 1.5e-5, - "range": [128000.0, 256000.0] + "input_cost_per_token": 0.000003, + "output_cost_per_token": 0.000015, + "range": [ + 128000, + 256000 + ] }, { - "input_cost_per_token": 6e-6, - "output_cost_per_token": 6e-5, - "range": [256000.0, 1000000.0] + "input_cost_per_token": 0.000006, + "output_cost_per_token": 0.00006, + "range": [ + 256000, + 1000000 + ] } ] }, @@ -6256,19 +7657,28 @@ "supports_tool_choice": true, "tiered_pricing": [ { - "input_cost_per_token": 1.2e-6, - "output_cost_per_token": 6e-6, - "range": [0, 32000.0] + "input_cost_per_token": 0.0000012, + "output_cost_per_token": 0.000006, + "range": [ + 0, + 32000 + ] }, { - "input_cost_per_token": 2.4e-6, - "output_cost_per_token": 1.2e-5, - "range": [32000.0, 128000.0] + "input_cost_per_token": 0.0000024, + "output_cost_per_token": 0.000012, + "range": [ + 32000, + 128000 + ] }, { - "input_cost_per_token": 3e-6, - "output_cost_per_token": 1.5e-5, - "range": [128000.0, 252000.0] + "input_cost_per_token": 0.000003, + "output_cost_per_token": 0.000015, + "range": [ + 128000, + 252000 + ] } ] }, @@ -6279,7 +7689,7 @@ "max_output_tokens": 8192, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 2.4e-6, + "output_cost_per_token": 0.0000024, "source": "https://www.alibabacloud.com/help/en/model-studio/models", "supports_function_calling": true, "supports_reasoning": true, @@ -6287,7 +7697,7 @@ }, "databricks/databricks-bge-large-en": { "input_cost_per_token": 1.0003e-7, - "input_dbu_cost_per_token": 1.429e-6, + "input_dbu_cost_per_token": 0.000001429, "litellm_provider": "databricks", "max_input_tokens": 512, "max_tokens": 512, @@ -6295,14 +7705,14 @@ "notes": "Input/output cost per token is dbu cost * $0.070, based on databricks Llama 3.1 70B conversion. Number provided for reference, '*_dbu_cost_per_token' used in actual calculation." }, "mode": "embedding", - "output_cost_per_token": 0.0, - "output_dbu_cost_per_token": 0.0, + "output_cost_per_token": 0, + "output_dbu_cost_per_token": 0, "output_vector_size": 1024, "source": "https://www.databricks.com/product/pricing/foundation-model-serving" }, "databricks/databricks-claude-3-7-sonnet": { - "input_cost_per_token": 2.5e-6, - "input_dbu_cost_per_token": 3.571e-5, + "input_cost_per_token": 0.0000025, + "input_dbu_cost_per_token": 0.00003571, "litellm_provider": "databricks", "max_input_tokens": 200000, "max_output_tokens": 128000, @@ -6311,7 +7721,7 @@ "notes": "Input/output cost per token is dbu cost * $0.070, based on databricks Claude 3.7 conversion. Number provided for reference, '*_dbu_cost_per_token' used in actual calculation." }, "mode": "chat", - "output_cost_per_token": 1.7857e-5, + "output_cost_per_token": 0.000017857, "output_db_cost_per_token": 0.000214286, "source": "https://www.databricks.com/product/pricing/foundation-model-serving", "supports_assistant_prefill": true, @@ -6321,7 +7731,7 @@ }, "databricks/databricks-gte-large-en": { "input_cost_per_token": 1.2999e-7, - "input_dbu_cost_per_token": 1.857e-6, + "input_dbu_cost_per_token": 0.000001857, "litellm_provider": "databricks", "max_input_tokens": 8192, "max_tokens": 8192, @@ -6329,14 +7739,14 @@ "notes": "Input/output cost per token is dbu cost * $0.070, based on databricks Llama 3.1 70B conversion. Number provided for reference, '*_dbu_cost_per_token' used in actual calculation." }, "mode": "embedding", - "output_cost_per_token": 0.0, - "output_dbu_cost_per_token": 0.0, + "output_cost_per_token": 0, + "output_dbu_cost_per_token": 0, "output_vector_size": 1024, "source": "https://www.databricks.com/product/pricing/foundation-model-serving" }, "databricks/databricks-llama-2-70b-chat": { "input_cost_per_token": 5.0001e-7, - "input_dbu_cost_per_token": 7.143e-6, + "input_dbu_cost_per_token": 0.000007143, "litellm_provider": "databricks", "max_input_tokens": 4096, "max_output_tokens": 4096, @@ -6345,14 +7755,14 @@ "notes": "Input/output cost per token is dbu cost * $0.070, based on databricks Llama 3.1 70B conversion. Number provided for reference, '*_dbu_cost_per_token' used in actual calculation." }, "mode": "chat", - "output_cost_per_token": 1.5e-6, - "output_dbu_cost_per_token": 2.1429e-5, + "output_cost_per_token": 0.0000015, + "output_dbu_cost_per_token": 0.000021429, "source": "https://www.databricks.com/product/pricing/foundation-model-serving", "supports_tool_choice": true }, "databricks/databricks-llama-4-maverick": { - "input_cost_per_token": 5e-6, - "input_dbu_cost_per_token": 7.143e-5, + "input_cost_per_token": 0.000005, + "input_dbu_cost_per_token": 0.00007143, "litellm_provider": "databricks", "max_input_tokens": 128000, "max_output_tokens": 128000, @@ -6361,14 +7771,14 @@ "notes": "Databricks documentation now provides both DBU costs (_dbu_cost_per_token) and dollar costs(_cost_per_token)." }, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "output_dbu_cost_per_token": 0.00021429, "source": "https://www.databricks.com/product/pricing/foundation-model-serving", "supports_tool_choice": true }, "databricks/databricks-meta-llama-3-1-405b-instruct": { - "input_cost_per_token": 5e-6, - "input_dbu_cost_per_token": 7.1429e-5, + "input_cost_per_token": 0.000005, + "input_dbu_cost_per_token": 0.000071429, "litellm_provider": "databricks", "max_input_tokens": 128000, "max_output_tokens": 128000, @@ -6377,14 +7787,14 @@ "notes": "Input/output cost per token is dbu cost * $0.070, based on databricks Llama 3.1 70B conversion. Number provided for reference, '*_dbu_cost_per_token' used in actual calculation." }, "mode": "chat", - "output_cost_per_token": 1.500002e-5, + "output_cost_per_token": 0.00001500002, "output_db_cost_per_token": 0.000214286, "source": "https://www.databricks.com/product/pricing/foundation-model-serving", "supports_tool_choice": true }, "databricks/databricks-meta-llama-3-3-70b-instruct": { - "input_cost_per_token": 1.00002e-6, - "input_dbu_cost_per_token": 1.4286e-5, + "input_cost_per_token": 0.00000100002, + "input_dbu_cost_per_token": 0.000014286, "litellm_provider": "databricks", "max_input_tokens": 128000, "max_output_tokens": 128000, @@ -6393,14 +7803,14 @@ "notes": "Input/output cost per token is dbu cost * $0.070, based on databricks Llama 3.1 70B conversion. Number provided for reference, '*_dbu_cost_per_token' used in actual calculation." }, "mode": "chat", - "output_cost_per_token": 2.99999e-6, - "output_dbu_cost_per_token": 4.2857e-5, + "output_cost_per_token": 0.00000299999, + "output_dbu_cost_per_token": 0.000042857, "source": "https://www.databricks.com/product/pricing/foundation-model-serving", "supports_tool_choice": true }, "databricks/databricks-meta-llama-3-70b-instruct": { - "input_cost_per_token": 1.00002e-6, - "input_dbu_cost_per_token": 1.4286e-5, + "input_cost_per_token": 0.00000100002, + "input_dbu_cost_per_token": 0.000014286, "litellm_provider": "databricks", "max_input_tokens": 128000, "max_output_tokens": 128000, @@ -6409,14 +7819,14 @@ "notes": "Input/output cost per token is dbu cost * $0.070, based on databricks Llama 3.1 70B conversion. Number provided for reference, '*_dbu_cost_per_token' used in actual calculation." }, "mode": "chat", - "output_cost_per_token": 2.99999e-6, - "output_dbu_cost_per_token": 4.2857e-5, + "output_cost_per_token": 0.00000299999, + "output_dbu_cost_per_token": 0.000042857, "source": "https://www.databricks.com/product/pricing/foundation-model-serving", "supports_tool_choice": true }, "databricks/databricks-mixtral-8x7b-instruct": { "input_cost_per_token": 5.0001e-7, - "input_dbu_cost_per_token": 7.143e-6, + "input_dbu_cost_per_token": 0.000007143, "litellm_provider": "databricks", "max_input_tokens": 4096, "max_output_tokens": 4096, @@ -6426,13 +7836,13 @@ }, "mode": "chat", "output_cost_per_token": 9.9902e-7, - "output_dbu_cost_per_token": 1.4286e-5, + "output_dbu_cost_per_token": 0.000014286, "source": "https://www.databricks.com/product/pricing/foundation-model-serving", "supports_tool_choice": true }, "databricks/databricks-mpt-30b-instruct": { "input_cost_per_token": 9.9902e-7, - "input_dbu_cost_per_token": 1.4286e-5, + "input_dbu_cost_per_token": 0.000014286, "litellm_provider": "databricks", "max_input_tokens": 8192, "max_output_tokens": 8192, @@ -6442,13 +7852,13 @@ }, "mode": "chat", "output_cost_per_token": 9.9902e-7, - "output_dbu_cost_per_token": 1.4286e-5, + "output_dbu_cost_per_token": 0.000014286, "source": "https://www.databricks.com/product/pricing/foundation-model-serving", "supports_tool_choice": true }, "databricks/databricks-mpt-7b-instruct": { "input_cost_per_token": 5.0001e-7, - "input_dbu_cost_per_token": 7.143e-6, + "input_dbu_cost_per_token": 0.000007143, "litellm_provider": "databricks", "max_input_tokens": 8192, "max_output_tokens": 8192, @@ -6457,8 +7867,8 @@ "notes": "Input/output cost per token is dbu cost * $0.070, based on databricks Llama 3.1 70B conversion. Number provided for reference, '*_dbu_cost_per_token' used in actual calculation." }, "mode": "chat", - "output_cost_per_token": 0.0, - "output_dbu_cost_per_token": 0.0, + "output_cost_per_token": 0, + "output_dbu_cost_per_token": 0, "source": "https://www.databricks.com/product/pricing/foundation-model-serving", "supports_tool_choice": true }, @@ -6468,13 +7878,13 @@ "mode": "search" }, "davinci-002": { - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "text-completion-openai", "max_input_tokens": 16384, "max_output_tokens": 4096, "max_tokens": 16384, "mode": "completion", - "output_cost_per_token": 2e-6 + "output_cost_per_token": 0.000002 }, "deepgram/base": { "input_cost_per_second": 0.00020833, @@ -6484,9 +7894,11 @@ "original_pricing_per_minute": 0.0125 }, "mode": "audio_transcription", - "output_cost_per_second": 0.0, + "output_cost_per_second": 0, "source": "https://deepgram.com/pricing", - "supported_endpoints": ["/v1/audio/transcriptions"] + "supported_endpoints": [ + "/v1/audio/transcriptions" + ] }, "deepgram/base-conversationalai": { "input_cost_per_second": 0.00020833, @@ -6496,9 +7908,11 @@ "original_pricing_per_minute": 0.0125 }, "mode": "audio_transcription", - "output_cost_per_second": 0.0, + "output_cost_per_second": 0, "source": "https://deepgram.com/pricing", - "supported_endpoints": ["/v1/audio/transcriptions"] + "supported_endpoints": [ + "/v1/audio/transcriptions" + ] }, "deepgram/base-finance": { "input_cost_per_second": 0.00020833, @@ -6508,9 +7922,11 @@ "original_pricing_per_minute": 0.0125 }, "mode": "audio_transcription", - "output_cost_per_second": 0.0, + "output_cost_per_second": 0, "source": "https://deepgram.com/pricing", - "supported_endpoints": ["/v1/audio/transcriptions"] + "supported_endpoints": [ + "/v1/audio/transcriptions" + ] }, "deepgram/base-general": { "input_cost_per_second": 0.00020833, @@ -6520,9 +7936,11 @@ "original_pricing_per_minute": 0.0125 }, "mode": "audio_transcription", - "output_cost_per_second": 0.0, + "output_cost_per_second": 0, "source": "https://deepgram.com/pricing", - "supported_endpoints": ["/v1/audio/transcriptions"] + "supported_endpoints": [ + "/v1/audio/transcriptions" + ] }, "deepgram/base-meeting": { "input_cost_per_second": 0.00020833, @@ -6532,9 +7950,11 @@ "original_pricing_per_minute": 0.0125 }, "mode": "audio_transcription", - "output_cost_per_second": 0.0, + "output_cost_per_second": 0, "source": "https://deepgram.com/pricing", - "supported_endpoints": ["/v1/audio/transcriptions"] + "supported_endpoints": [ + "/v1/audio/transcriptions" + ] }, "deepgram/base-phonecall": { "input_cost_per_second": 0.00020833, @@ -6544,9 +7964,11 @@ "original_pricing_per_minute": 0.0125 }, "mode": "audio_transcription", - "output_cost_per_second": 0.0, + "output_cost_per_second": 0, "source": "https://deepgram.com/pricing", - "supported_endpoints": ["/v1/audio/transcriptions"] + "supported_endpoints": [ + "/v1/audio/transcriptions" + ] }, "deepgram/base-video": { "input_cost_per_second": 0.00020833, @@ -6556,9 +7978,11 @@ "original_pricing_per_minute": 0.0125 }, "mode": "audio_transcription", - "output_cost_per_second": 0.0, + "output_cost_per_second": 0, "source": "https://deepgram.com/pricing", - "supported_endpoints": ["/v1/audio/transcriptions"] + "supported_endpoints": [ + "/v1/audio/transcriptions" + ] }, "deepgram/base-voicemail": { "input_cost_per_second": 0.00020833, @@ -6568,9 +7992,11 @@ "original_pricing_per_minute": 0.0125 }, "mode": "audio_transcription", - "output_cost_per_second": 0.0, + "output_cost_per_second": 0, "source": "https://deepgram.com/pricing", - "supported_endpoints": ["/v1/audio/transcriptions"] + "supported_endpoints": [ + "/v1/audio/transcriptions" + ] }, "deepgram/enhanced": { "input_cost_per_second": 0.00024167, @@ -6580,9 +8006,11 @@ "original_pricing_per_minute": 0.0145 }, "mode": "audio_transcription", - "output_cost_per_second": 0.0, + "output_cost_per_second": 0, "source": "https://deepgram.com/pricing", - "supported_endpoints": ["/v1/audio/transcriptions"] + "supported_endpoints": [ + "/v1/audio/transcriptions" + ] }, "deepgram/enhanced-finance": { "input_cost_per_second": 0.00024167, @@ -6592,9 +8020,11 @@ "original_pricing_per_minute": 0.0145 }, "mode": "audio_transcription", - "output_cost_per_second": 0.0, + "output_cost_per_second": 0, "source": "https://deepgram.com/pricing", - "supported_endpoints": ["/v1/audio/transcriptions"] + "supported_endpoints": [ + "/v1/audio/transcriptions" + ] }, "deepgram/enhanced-general": { "input_cost_per_second": 0.00024167, @@ -6604,9 +8034,11 @@ "original_pricing_per_minute": 0.0145 }, "mode": "audio_transcription", - "output_cost_per_second": 0.0, + "output_cost_per_second": 0, "source": "https://deepgram.com/pricing", - "supported_endpoints": ["/v1/audio/transcriptions"] + "supported_endpoints": [ + "/v1/audio/transcriptions" + ] }, "deepgram/enhanced-meeting": { "input_cost_per_second": 0.00024167, @@ -6616,9 +8048,11 @@ "original_pricing_per_minute": 0.0145 }, "mode": "audio_transcription", - "output_cost_per_second": 0.0, + "output_cost_per_second": 0, "source": "https://deepgram.com/pricing", - "supported_endpoints": ["/v1/audio/transcriptions"] + "supported_endpoints": [ + "/v1/audio/transcriptions" + ] }, "deepgram/enhanced-phonecall": { "input_cost_per_second": 0.00024167, @@ -6628,213 +8062,249 @@ "original_pricing_per_minute": 0.0145 }, "mode": "audio_transcription", - "output_cost_per_second": 0.0, + "output_cost_per_second": 0, "source": "https://deepgram.com/pricing", - "supported_endpoints": ["/v1/audio/transcriptions"] + "supported_endpoints": [ + "/v1/audio/transcriptions" + ] }, "deepgram/nova": { - "input_cost_per_second": 7.167e-5, + "input_cost_per_second": 0.00007167, "litellm_provider": "deepgram", "metadata": { "calculation": "$0.0043/60 seconds = $0.00007167 per second", "original_pricing_per_minute": 0.0043 }, "mode": "audio_transcription", - "output_cost_per_second": 0.0, + "output_cost_per_second": 0, "source": "https://deepgram.com/pricing", - "supported_endpoints": ["/v1/audio/transcriptions"] + "supported_endpoints": [ + "/v1/audio/transcriptions" + ] }, "deepgram/nova-2": { - "input_cost_per_second": 7.167e-5, + "input_cost_per_second": 0.00007167, "litellm_provider": "deepgram", "metadata": { "calculation": "$0.0043/60 seconds = $0.00007167 per second", "original_pricing_per_minute": 0.0043 }, "mode": "audio_transcription", - "output_cost_per_second": 0.0, + "output_cost_per_second": 0, "source": "https://deepgram.com/pricing", - "supported_endpoints": ["/v1/audio/transcriptions"] + "supported_endpoints": [ + "/v1/audio/transcriptions" + ] }, "deepgram/nova-2-atc": { - "input_cost_per_second": 7.167e-5, + "input_cost_per_second": 0.00007167, "litellm_provider": "deepgram", "metadata": { "calculation": "$0.0043/60 seconds = $0.00007167 per second", "original_pricing_per_minute": 0.0043 }, "mode": "audio_transcription", - "output_cost_per_second": 0.0, + "output_cost_per_second": 0, "source": "https://deepgram.com/pricing", - "supported_endpoints": ["/v1/audio/transcriptions"] + "supported_endpoints": [ + "/v1/audio/transcriptions" + ] }, "deepgram/nova-2-automotive": { - "input_cost_per_second": 7.167e-5, + "input_cost_per_second": 0.00007167, "litellm_provider": "deepgram", "metadata": { "calculation": "$0.0043/60 seconds = $0.00007167 per second", "original_pricing_per_minute": 0.0043 }, "mode": "audio_transcription", - "output_cost_per_second": 0.0, + "output_cost_per_second": 0, "source": "https://deepgram.com/pricing", - "supported_endpoints": ["/v1/audio/transcriptions"] + "supported_endpoints": [ + "/v1/audio/transcriptions" + ] }, "deepgram/nova-2-conversationalai": { - "input_cost_per_second": 7.167e-5, + "input_cost_per_second": 0.00007167, "litellm_provider": "deepgram", "metadata": { "calculation": "$0.0043/60 seconds = $0.00007167 per second", "original_pricing_per_minute": 0.0043 }, "mode": "audio_transcription", - "output_cost_per_second": 0.0, + "output_cost_per_second": 0, "source": "https://deepgram.com/pricing", - "supported_endpoints": ["/v1/audio/transcriptions"] + "supported_endpoints": [ + "/v1/audio/transcriptions" + ] }, "deepgram/nova-2-drivethru": { - "input_cost_per_second": 7.167e-5, + "input_cost_per_second": 0.00007167, "litellm_provider": "deepgram", "metadata": { "calculation": "$0.0043/60 seconds = $0.00007167 per second", "original_pricing_per_minute": 0.0043 }, "mode": "audio_transcription", - "output_cost_per_second": 0.0, + "output_cost_per_second": 0, "source": "https://deepgram.com/pricing", - "supported_endpoints": ["/v1/audio/transcriptions"] + "supported_endpoints": [ + "/v1/audio/transcriptions" + ] }, "deepgram/nova-2-finance": { - "input_cost_per_second": 7.167e-5, + "input_cost_per_second": 0.00007167, "litellm_provider": "deepgram", "metadata": { "calculation": "$0.0043/60 seconds = $0.00007167 per second", "original_pricing_per_minute": 0.0043 }, "mode": "audio_transcription", - "output_cost_per_second": 0.0, + "output_cost_per_second": 0, "source": "https://deepgram.com/pricing", - "supported_endpoints": ["/v1/audio/transcriptions"] + "supported_endpoints": [ + "/v1/audio/transcriptions" + ] }, "deepgram/nova-2-general": { - "input_cost_per_second": 7.167e-5, + "input_cost_per_second": 0.00007167, "litellm_provider": "deepgram", "metadata": { "calculation": "$0.0043/60 seconds = $0.00007167 per second", "original_pricing_per_minute": 0.0043 }, "mode": "audio_transcription", - "output_cost_per_second": 0.0, + "output_cost_per_second": 0, "source": "https://deepgram.com/pricing", - "supported_endpoints": ["/v1/audio/transcriptions"] + "supported_endpoints": [ + "/v1/audio/transcriptions" + ] }, "deepgram/nova-2-meeting": { - "input_cost_per_second": 7.167e-5, + "input_cost_per_second": 0.00007167, "litellm_provider": "deepgram", "metadata": { "calculation": "$0.0043/60 seconds = $0.00007167 per second", "original_pricing_per_minute": 0.0043 }, "mode": "audio_transcription", - "output_cost_per_second": 0.0, + "output_cost_per_second": 0, "source": "https://deepgram.com/pricing", - "supported_endpoints": ["/v1/audio/transcriptions"] + "supported_endpoints": [ + "/v1/audio/transcriptions" + ] }, "deepgram/nova-2-phonecall": { - "input_cost_per_second": 7.167e-5, + "input_cost_per_second": 0.00007167, "litellm_provider": "deepgram", "metadata": { "calculation": "$0.0043/60 seconds = $0.00007167 per second", "original_pricing_per_minute": 0.0043 }, "mode": "audio_transcription", - "output_cost_per_second": 0.0, + "output_cost_per_second": 0, "source": "https://deepgram.com/pricing", - "supported_endpoints": ["/v1/audio/transcriptions"] + "supported_endpoints": [ + "/v1/audio/transcriptions" + ] }, "deepgram/nova-2-video": { - "input_cost_per_second": 7.167e-5, + "input_cost_per_second": 0.00007167, "litellm_provider": "deepgram", "metadata": { "calculation": "$0.0043/60 seconds = $0.00007167 per second", "original_pricing_per_minute": 0.0043 }, "mode": "audio_transcription", - "output_cost_per_second": 0.0, + "output_cost_per_second": 0, "source": "https://deepgram.com/pricing", - "supported_endpoints": ["/v1/audio/transcriptions"] + "supported_endpoints": [ + "/v1/audio/transcriptions" + ] }, "deepgram/nova-2-voicemail": { - "input_cost_per_second": 7.167e-5, + "input_cost_per_second": 0.00007167, "litellm_provider": "deepgram", "metadata": { "calculation": "$0.0043/60 seconds = $0.00007167 per second", "original_pricing_per_minute": 0.0043 }, "mode": "audio_transcription", - "output_cost_per_second": 0.0, + "output_cost_per_second": 0, "source": "https://deepgram.com/pricing", - "supported_endpoints": ["/v1/audio/transcriptions"] + "supported_endpoints": [ + "/v1/audio/transcriptions" + ] }, "deepgram/nova-3": { - "input_cost_per_second": 7.167e-5, + "input_cost_per_second": 0.00007167, "litellm_provider": "deepgram", "metadata": { "calculation": "$0.0043/60 seconds = $0.00007167 per second", "original_pricing_per_minute": 0.0043 }, "mode": "audio_transcription", - "output_cost_per_second": 0.0, + "output_cost_per_second": 0, "source": "https://deepgram.com/pricing", - "supported_endpoints": ["/v1/audio/transcriptions"] + "supported_endpoints": [ + "/v1/audio/transcriptions" + ] }, "deepgram/nova-3-general": { - "input_cost_per_second": 7.167e-5, + "input_cost_per_second": 0.00007167, "litellm_provider": "deepgram", "metadata": { "calculation": "$0.0043/60 seconds = $0.00007167 per second", "original_pricing_per_minute": 0.0043 }, "mode": "audio_transcription", - "output_cost_per_second": 0.0, + "output_cost_per_second": 0, "source": "https://deepgram.com/pricing", - "supported_endpoints": ["/v1/audio/transcriptions"] + "supported_endpoints": [ + "/v1/audio/transcriptions" + ] }, "deepgram/nova-3-medical": { - "input_cost_per_second": 8.667e-5, + "input_cost_per_second": 0.00008667, "litellm_provider": "deepgram", "metadata": { "calculation": "$0.0052/60 seconds = $0.00008667 per second (multilingual)", "original_pricing_per_minute": 0.0052 }, "mode": "audio_transcription", - "output_cost_per_second": 0.0, + "output_cost_per_second": 0, "source": "https://deepgram.com/pricing", - "supported_endpoints": ["/v1/audio/transcriptions"] + "supported_endpoints": [ + "/v1/audio/transcriptions" + ] }, "deepgram/nova-general": { - "input_cost_per_second": 7.167e-5, + "input_cost_per_second": 0.00007167, "litellm_provider": "deepgram", "metadata": { "calculation": "$0.0043/60 seconds = $0.00007167 per second", "original_pricing_per_minute": 0.0043 }, "mode": "audio_transcription", - "output_cost_per_second": 0.0, + "output_cost_per_second": 0, "source": "https://deepgram.com/pricing", - "supported_endpoints": ["/v1/audio/transcriptions"] + "supported_endpoints": [ + "/v1/audio/transcriptions" + ] }, "deepgram/nova-phonecall": { - "input_cost_per_second": 7.167e-5, + "input_cost_per_second": 0.00007167, "litellm_provider": "deepgram", "metadata": { "calculation": "$0.0043/60 seconds = $0.00007167 per second", "original_pricing_per_minute": 0.0043 }, "mode": "audio_transcription", - "output_cost_per_second": 0.0, + "output_cost_per_second": 0, "source": "https://deepgram.com/pricing", - "supported_endpoints": ["/v1/audio/transcriptions"] + "supported_endpoints": [ + "/v1/audio/transcriptions" + ] }, "deepgram/whisper": { "input_cost_per_second": 0.0001, @@ -6843,9 +8313,11 @@ "notes": "Deepgram's hosted OpenAI Whisper models - pricing may differ from native Deepgram models" }, "mode": "audio_transcription", - "output_cost_per_second": 0.0, + "output_cost_per_second": 0, "source": "https://deepgram.com/pricing", - "supported_endpoints": ["/v1/audio/transcriptions"] + "supported_endpoints": [ + "/v1/audio/transcriptions" + ] }, "deepgram/whisper-base": { "input_cost_per_second": 0.0001, @@ -6854,9 +8326,11 @@ "notes": "Deepgram's hosted OpenAI Whisper models - pricing may differ from native Deepgram models" }, "mode": "audio_transcription", - "output_cost_per_second": 0.0, + "output_cost_per_second": 0, "source": "https://deepgram.com/pricing", - "supported_endpoints": ["/v1/audio/transcriptions"] + "supported_endpoints": [ + "/v1/audio/transcriptions" + ] }, "deepgram/whisper-large": { "input_cost_per_second": 0.0001, @@ -6865,9 +8339,11 @@ "notes": "Deepgram's hosted OpenAI Whisper models - pricing may differ from native Deepgram models" }, "mode": "audio_transcription", - "output_cost_per_second": 0.0, + "output_cost_per_second": 0, "source": "https://deepgram.com/pricing", - "supported_endpoints": ["/v1/audio/transcriptions"] + "supported_endpoints": [ + "/v1/audio/transcriptions" + ] }, "deepgram/whisper-medium": { "input_cost_per_second": 0.0001, @@ -6876,9 +8352,11 @@ "notes": "Deepgram's hosted OpenAI Whisper models - pricing may differ from native Deepgram models" }, "mode": "audio_transcription", - "output_cost_per_second": 0.0, + "output_cost_per_second": 0, "source": "https://deepgram.com/pricing", - "supported_endpoints": ["/v1/audio/transcriptions"] + "supported_endpoints": [ + "/v1/audio/transcriptions" + ] }, "deepgram/whisper-small": { "input_cost_per_second": 0.0001, @@ -6887,9 +8365,11 @@ "notes": "Deepgram's hosted OpenAI Whisper models - pricing may differ from native Deepgram models" }, "mode": "audio_transcription", - "output_cost_per_second": 0.0, + "output_cost_per_second": 0, "source": "https://deepgram.com/pricing", - "supported_endpoints": ["/v1/audio/transcriptions"] + "supported_endpoints": [ + "/v1/audio/transcriptions" + ] }, "deepgram/whisper-tiny": { "input_cost_per_second": 0.0001, @@ -6898,9 +8378,11 @@ "notes": "Deepgram's hosted OpenAI Whisper models - pricing may differ from native Deepgram models" }, "mode": "audio_transcription", - "output_cost_per_second": 0.0, + "output_cost_per_second": 0, "source": "https://deepgram.com/pricing", - "supported_endpoints": ["/v1/audio/transcriptions"] + "supported_endpoints": [ + "/v1/audio/transcriptions" + ] }, "deepinfra/Gryphe/MythoMax-L2-13b": { "max_tokens": 4096, @@ -6916,8 +8398,8 @@ "max_tokens": 131072, "max_input_tokens": 131072, "max_output_tokens": 131072, - "input_cost_per_token": 1e-6, - "output_cost_per_token": 1e-6, + "input_cost_per_token": 0.000001, + "output_cost_per_token": 0.000001, "litellm_provider": "deepinfra", "mode": "chat", "supports_tool_choice": true @@ -7008,7 +8490,7 @@ "max_input_tokens": 262144, "max_output_tokens": 262144, "input_cost_per_token": 3e-7, - "output_cost_per_token": 2.9e-6, + "output_cost_per_token": 0.0000029, "litellm_provider": "deepinfra", "mode": "chat", "supports_tool_choice": true @@ -7038,7 +8520,7 @@ "max_input_tokens": 262144, "max_output_tokens": 262144, "input_cost_per_token": 4e-7, - "output_cost_per_token": 1.6e-6, + "output_cost_per_token": 0.0000016, "litellm_provider": "deepinfra", "mode": "chat", "supports_tool_choice": true @@ -7048,7 +8530,7 @@ "max_input_tokens": 262144, "max_output_tokens": 262144, "input_cost_per_token": 2.9e-7, - "output_cost_per_token": 1.2e-6, + "output_cost_per_token": 0.0000012, "litellm_provider": "deepinfra", "mode": "chat", "supports_tool_choice": true @@ -7058,7 +8540,7 @@ "max_input_tokens": 262144, "max_output_tokens": 262144, "input_cost_per_token": 1.4e-7, - "output_cost_per_token": 1.4e-6, + "output_cost_per_token": 0.0000014, "litellm_provider": "deepinfra", "mode": "chat", "supports_tool_choice": true @@ -7068,7 +8550,7 @@ "max_input_tokens": 262144, "max_output_tokens": 262144, "input_cost_per_token": 1.4e-7, - "output_cost_per_token": 1.4e-6, + "output_cost_per_token": 0.0000014, "litellm_provider": "deepinfra", "mode": "chat", "supports_tool_choice": true @@ -7108,7 +8590,7 @@ "max_input_tokens": 16384, "max_output_tokens": 16384, "input_cost_per_token": 2.7e-7, - "output_cost_per_token": 1.5e-6, + "output_cost_per_token": 0.0000015, "litellm_provider": "deepinfra", "mode": "chat", "supports_tool_choice": false @@ -7117,8 +8599,8 @@ "max_tokens": 200000, "max_input_tokens": 200000, "max_output_tokens": 200000, - "input_cost_per_token": 3.3e-6, - "output_cost_per_token": 1.65e-5, + "input_cost_per_token": 0.0000033, + "output_cost_per_token": 0.0000165, "cache_read_input_token_cost": 3.3e-7, "litellm_provider": "deepinfra", "mode": "chat", @@ -7128,8 +8610,8 @@ "max_tokens": 200000, "max_input_tokens": 200000, "max_output_tokens": 200000, - "input_cost_per_token": 1.65e-5, - "output_cost_per_token": 8.25e-5, + "input_cost_per_token": 0.0000165, + "output_cost_per_token": 0.0000825, "litellm_provider": "deepinfra", "mode": "chat", "supports_tool_choice": true @@ -7138,8 +8620,8 @@ "max_tokens": 200000, "max_input_tokens": 200000, "max_output_tokens": 200000, - "input_cost_per_token": 3.3e-6, - "output_cost_per_token": 1.65e-5, + "input_cost_per_token": 0.0000033, + "output_cost_per_token": 0.0000165, "litellm_provider": "deepinfra", "mode": "chat", "supports_tool_choice": true @@ -7149,7 +8631,7 @@ "max_input_tokens": 163840, "max_output_tokens": 163840, "input_cost_per_token": 7e-7, - "output_cost_per_token": 2.4e-6, + "output_cost_per_token": 0.0000024, "litellm_provider": "deepinfra", "mode": "chat", "supports_tool_choice": true @@ -7159,7 +8641,7 @@ "max_input_tokens": 163840, "max_output_tokens": 163840, "input_cost_per_token": 5e-7, - "output_cost_per_token": 2.15e-6, + "output_cost_per_token": 0.00000215, "cache_read_input_token_cost": 4e-7, "litellm_provider": "deepinfra", "mode": "chat", @@ -7169,8 +8651,8 @@ "max_tokens": 32768, "max_input_tokens": 32768, "max_output_tokens": 32768, - "input_cost_per_token": 1e-6, - "output_cost_per_token": 3e-6, + "input_cost_per_token": 0.000001, + "output_cost_per_token": 0.000003, "litellm_provider": "deepinfra", "mode": "chat", "supports_tool_choice": true @@ -7199,8 +8681,8 @@ "max_tokens": 40960, "max_input_tokens": 40960, "max_output_tokens": 40960, - "input_cost_per_token": 1e-6, - "output_cost_per_token": 3e-6, + "input_cost_per_token": 0.000001, + "output_cost_per_token": 0.000003, "litellm_provider": "deepinfra", "mode": "chat", "supports_tool_choice": true @@ -7230,7 +8712,7 @@ "max_input_tokens": 163840, "max_output_tokens": 163840, "input_cost_per_token": 2.7e-7, - "output_cost_per_token": 1e-6, + "output_cost_per_token": 0.000001, "cache_read_input_token_cost": 2.16e-7, "litellm_provider": "deepinfra", "mode": "chat", @@ -7242,7 +8724,7 @@ "max_input_tokens": 163840, "max_output_tokens": 163840, "input_cost_per_token": 2.7e-7, - "output_cost_per_token": 1e-6, + "output_cost_per_token": 0.000001, "cache_read_input_token_cost": 2.16e-7, "litellm_provider": "deepinfra", "mode": "chat", @@ -7263,7 +8745,7 @@ "max_input_tokens": 1000000, "max_output_tokens": 1000000, "input_cost_per_token": 3e-7, - "output_cost_per_token": 2.5e-6, + "output_cost_per_token": 0.0000025, "litellm_provider": "deepinfra", "mode": "chat", "supports_tool_choice": true @@ -7272,8 +8754,8 @@ "max_tokens": 1000000, "max_input_tokens": 1000000, "max_output_tokens": 1000000, - "input_cost_per_token": 1.25e-6, - "output_cost_per_token": 1e-5, + "input_cost_per_token": 0.00000125, + "output_cost_per_token": 0.00001, "litellm_provider": "deepinfra", "mode": "chat", "supports_tool_choice": true @@ -7503,7 +8985,7 @@ "max_input_tokens": 131072, "max_output_tokens": 131072, "input_cost_per_token": 5e-7, - "output_cost_per_token": 2e-6, + "output_cost_per_token": 0.000002, "litellm_provider": "deepinfra", "mode": "chat", "supports_tool_choice": true @@ -7513,7 +8995,7 @@ "max_input_tokens": 262144, "max_output_tokens": 262144, "input_cost_per_token": 5e-7, - "output_cost_per_token": 2e-6, + "output_cost_per_token": 0.000002, "cache_read_input_token_cost": 4e-7, "litellm_provider": "deepinfra", "mode": "chat", @@ -7574,13 +9056,13 @@ "max_input_tokens": 131072, "max_output_tokens": 131072, "input_cost_per_token": 4e-7, - "output_cost_per_token": 1.6e-6, + "output_cost_per_token": 0.0000016, "litellm_provider": "deepinfra", "mode": "chat", "supports_tool_choice": true }, "deepseek/deepseek-chat": { - "cache_creation_input_token_cost": 0.0, + "cache_creation_input_token_cost": 0, "cache_read_input_token_cost": 7e-8, "input_cost_per_token": 2.7e-7, "input_cost_per_token_cache_hit": 7e-8, @@ -7589,7 +9071,7 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 1.1e-6, + "output_cost_per_token": 0.0000011, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_prompt_caching": true, @@ -7617,7 +9099,7 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 2.19e-6, + "output_cost_per_token": 0.00000219, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_prompt_caching": true, @@ -7632,7 +9114,7 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 2.19e-6, + "output_cost_per_token": 0.00000219, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_prompt_caching": true, @@ -7640,7 +9122,7 @@ "supports_tool_choice": true }, "deepseek/deepseek-v3": { - "cache_creation_input_token_cost": 0.0, + "cache_creation_input_token_cost": 0, "cache_read_input_token_cost": 7e-8, "input_cost_per_token": 2.7e-7, "input_cost_per_token_cache_hit": 7e-8, @@ -7649,7 +9131,7 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 1.1e-6, + "output_cost_per_token": 0.0000011, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_prompt_caching": true, @@ -7662,7 +9144,7 @@ "max_output_tokens": 81920, "max_tokens": 163840, "mode": "chat", - "output_cost_per_token": 1.68e-6, + "output_cost_per_token": 0.00000168, "supports_function_calling": true, "supports_reasoning": true, "supports_tool_choice": true @@ -7677,7 +9159,7 @@ "output_cost_per_token": 5e-7 }, "doubao-embedding": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "volcengine", "max_input_tokens": 4096, "max_tokens": 4096, @@ -7685,11 +9167,11 @@ "notes": "Volcengine Doubao embedding model - standard version with 2560 dimensions" }, "mode": "embedding", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "output_vector_size": 2560 }, "doubao-embedding-large": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "volcengine", "max_input_tokens": 4096, "max_tokens": 4096, @@ -7697,11 +9179,11 @@ "notes": "Volcengine Doubao embedding model - large version with 2048 dimensions" }, "mode": "embedding", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "output_vector_size": 2048 }, "doubao-embedding-large-text-240915": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "volcengine", "max_input_tokens": 4096, "max_tokens": 4096, @@ -7709,11 +9191,11 @@ "notes": "Volcengine Doubao embedding model - text-240915 version with 4096 dimensions" }, "mode": "embedding", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "output_vector_size": 4096 }, "doubao-embedding-large-text-250515": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "volcengine", "max_input_tokens": 4096, "max_tokens": 4096, @@ -7721,11 +9203,11 @@ "notes": "Volcengine Doubao embedding model - text-250515 version with 2048 dimensions" }, "mode": "embedding", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "output_vector_size": 2048 }, "doubao-embedding-text-240715": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "volcengine", "max_input_tokens": 4096, "max_tokens": 4096, @@ -7733,7 +9215,7 @@ "notes": "Volcengine Doubao embedding model - text-240715 version with 2560 dimensions" }, "mode": "embedding", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "output_vector_size": 2560 }, "exa_ai/search": { @@ -7741,12 +9223,18 @@ "mode": "search", "tiered_pricing": [ { - "input_cost_per_query": 5e-3, - "max_results_range": [0, 25] + "input_cost_per_query": 0.005, + "max_results_range": [ + 0, + 25 + ] }, { - "input_cost_per_query": 25e-3, - "max_results_range": [26, 100] + "input_cost_per_query": 0.025, + "max_results_range": [ + 26, + 100 + ] } ] }, @@ -7755,44 +9243,74 @@ "mode": "search", "tiered_pricing": [ { - "input_cost_per_query": 1.66e-3, - "max_results_range": [1, 10] + "input_cost_per_query": 0.00166, + "max_results_range": [ + 1, + 10 + ] }, { - "input_cost_per_query": 3.32e-3, - "max_results_range": [11, 20] + "input_cost_per_query": 0.00332, + "max_results_range": [ + 11, + 20 + ] }, { - "input_cost_per_query": 4.98e-3, - "max_results_range": [21, 30] + "input_cost_per_query": 0.00498, + "max_results_range": [ + 21, + 30 + ] }, { - "input_cost_per_query": 6.64e-3, - "max_results_range": [31, 40] + "input_cost_per_query": 0.00664, + "max_results_range": [ + 31, + 40 + ] }, { - "input_cost_per_query": 8.3e-3, - "max_results_range": [41, 50] + "input_cost_per_query": 0.0083, + "max_results_range": [ + 41, + 50 + ] }, { - "input_cost_per_query": 9.96e-3, - "max_results_range": [51, 60] + "input_cost_per_query": 0.00996, + "max_results_range": [ + 51, + 60 + ] }, { - "input_cost_per_query": 11.62e-3, - "max_results_range": [61, 70] + "input_cost_per_query": 0.01162, + "max_results_range": [ + 61, + 70 + ] }, { - "input_cost_per_query": 13.28e-3, - "max_results_range": [71, 80] + "input_cost_per_query": 0.01328, + "max_results_range": [ + 71, + 80 + ] }, { - "input_cost_per_query": 14.94e-3, - "max_results_range": [81, 90] + "input_cost_per_query": 0.01494, + "max_results_range": [ + 81, + 90 + ] }, { - "input_cost_per_query": 16.6e-3, - "max_results_range": [91, 100] + "input_cost_per_query": 0.0166, + "max_results_range": [ + 91, + 100 + ] } ], "metadata": { @@ -7800,20 +9318,20 @@ } }, "perplexity/search": { - "input_cost_per_query": 5e-3, + "input_cost_per_query": 0.005, "litellm_provider": "perplexity", "mode": "search" }, "searxng/search": { "litellm_provider": "searxng", "mode": "search", - "input_cost_per_query": 0.0, + "input_cost_per_query": 0, "metadata": { "notes": "SearXNG is an open-source metasearch engine. Free to use when self-hosted or using public instances." } }, "elevenlabs/scribe_v1": { - "input_cost_per_second": 6.11e-5, + "input_cost_per_second": 0.0000611, "litellm_provider": "elevenlabs", "metadata": { "calculation": "$0.22/hour = $0.00366/minute = $0.0000611 per second (enterprise pricing)", @@ -7821,12 +9339,14 @@ "original_pricing_per_hour": 0.22 }, "mode": "audio_transcription", - "output_cost_per_second": 0.0, + "output_cost_per_second": 0, "source": "https://elevenlabs.io/pricing", - "supported_endpoints": ["/v1/audio/transcriptions"] + "supported_endpoints": [ + "/v1/audio/transcriptions" + ] }, "elevenlabs/scribe_v1_experimental": { - "input_cost_per_second": 6.11e-5, + "input_cost_per_second": 0.0000611, "litellm_provider": "elevenlabs", "metadata": { "calculation": "$0.22/hour = $0.00366/minute = $0.0000611 per second (enterprise pricing)", @@ -7834,9 +9354,11 @@ "original_pricing_per_hour": 0.22 }, "mode": "audio_transcription", - "output_cost_per_second": 0.0, + "output_cost_per_second": 0, "source": "https://elevenlabs.io/pricing", - "supported_endpoints": ["/v1/audio/transcriptions"] + "supported_endpoints": [ + "/v1/audio/transcriptions" + ] }, "embed-english-light-v2.0": { "input_cost_per_token": 1e-7, @@ -7844,7 +9366,7 @@ "max_input_tokens": 1024, "max_tokens": 1024, "mode": "embedding", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "embed-english-light-v3.0": { "input_cost_per_token": 1e-7, @@ -7852,7 +9374,7 @@ "max_input_tokens": 1024, "max_tokens": 1024, "mode": "embedding", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "embed-english-v2.0": { "input_cost_per_token": 1e-7, @@ -7860,7 +9382,7 @@ "max_input_tokens": 4096, "max_tokens": 4096, "mode": "embedding", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "embed-english-v3.0": { "input_cost_per_image": 0.0001, @@ -7872,7 +9394,7 @@ "notes": "'supports_image_input' is a deprecated field. Use 'supports_embedding_image_input' instead." }, "mode": "embedding", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "supports_embedding_image_input": true, "supports_image_input": true }, @@ -7882,7 +9404,7 @@ "max_input_tokens": 768, "max_tokens": 768, "mode": "embedding", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "embed-multilingual-v3.0": { "input_cost_per_token": 1e-7, @@ -7890,7 +9412,7 @@ "max_input_tokens": 1024, "max_tokens": 1024, "mode": "embedding", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "supports_embedding_image_input": true }, "eu.amazon.nova-lite-v1:0": { @@ -7920,13 +9442,13 @@ "supports_response_schema": true }, "eu.amazon.nova-pro-v1:0": { - "input_cost_per_token": 1.05e-6, + "input_cost_per_token": 0.00000105, "litellm_provider": "bedrock_converse", "max_input_tokens": 300000, "max_output_tokens": 10000, "max_tokens": 10000, "mode": "chat", - "output_cost_per_token": 4.2e-6, + "output_cost_per_token": 0.0000042, "source": "https://aws.amazon.com/bedrock/pricing/", "supports_function_calling": true, "supports_pdf_input": true, @@ -7941,7 +9463,7 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 1.25e-6, + "output_cost_per_token": 0.00000125, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_pdf_input": true, @@ -7950,33 +9472,36 @@ "supports_tool_choice": true }, "eu.anthropic.claude-haiku-4-5-20251001-v1:0": { - "cache_creation_input_token_cost": 1.375e-6, + "cache_creation_input_token_cost": 0.000001375, "cache_read_input_token_cost": 1.1e-7, - "input_cost_per_token": 1.1e-6, + "input_cost_per_token": 0.0000011, "deprecation_date": "2026-10-15", - "litellm_provider": "bedrock", + "litellm_provider": "bedrock_converse", "max_input_tokens": 200000, - "max_output_tokens": 8192, - "max_tokens": 8192, + "max_output_tokens": 64000, + "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 5.5e-6, + "output_cost_per_token": 0.0000055, "source": "https://aws.amazon.com/about-aws/whats-new/2025/10/claude-4-5-haiku-anthropic-amazon-bedrock", "supports_assistant_prefill": true, + "supports_computer_use": true, "supports_function_calling": true, "supports_pdf_input": true, "supports_prompt_caching": true, "supports_reasoning": true, "supports_response_schema": true, - "supports_tool_choice": true + "supports_tool_choice": true, + "supports_vision": true, + "tool_use_system_prompt_tokens": 346 }, "eu.anthropic.claude-3-5-sonnet-20240620-v1:0": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "bedrock", "max_input_tokens": 200000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "supports_function_calling": true, "supports_pdf_input": true, "supports_response_schema": true, @@ -7984,13 +9509,13 @@ "supports_vision": true }, "eu.anthropic.claude-3-5-sonnet-20241022-v2:0": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "bedrock", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "supports_assistant_prefill": true, "supports_computer_use": true, "supports_function_calling": true, @@ -8001,13 +9526,13 @@ "supports_vision": true }, "eu.anthropic.claude-3-7-sonnet-20250219-v1:0": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "bedrock", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "supports_assistant_prefill": true, "supports_computer_use": true, "supports_function_calling": true, @@ -8025,7 +9550,7 @@ "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 1.25e-6, + "output_cost_per_token": 0.00000125, "supports_function_calling": true, "supports_pdf_input": true, "supports_response_schema": true, @@ -8033,26 +9558,26 @@ "supports_vision": true }, "eu.anthropic.claude-3-opus-20240229-v1:0": { - "input_cost_per_token": 1.5e-5, + "input_cost_per_token": 0.000015, "litellm_provider": "bedrock", "max_input_tokens": 200000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 7.5e-5, + "output_cost_per_token": 0.000075, "supports_function_calling": true, "supports_response_schema": true, "supports_tool_choice": true, "supports_vision": true }, "eu.anthropic.claude-3-sonnet-20240229-v1:0": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "bedrock", "max_input_tokens": 200000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "supports_function_calling": true, "supports_pdf_input": true, "supports_response_schema": true, @@ -8060,15 +9585,15 @@ "supports_vision": true }, "eu.anthropic.claude-opus-4-1-20250805-v1:0": { - "cache_creation_input_token_cost": 1.875e-5, - "cache_read_input_token_cost": 1.5e-6, - "input_cost_per_token": 1.5e-5, + "cache_creation_input_token_cost": 0.00001875, + "cache_read_input_token_cost": 0.0000015, + "input_cost_per_token": 0.000015, "litellm_provider": "bedrock_converse", "max_input_tokens": 200000, "max_output_tokens": 32000, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 7.5e-5, + "output_cost_per_token": 0.000075, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -8086,15 +9611,15 @@ "tool_use_system_prompt_tokens": 159 }, "eu.anthropic.claude-opus-4-20250514-v1:0": { - "cache_creation_input_token_cost": 1.875e-5, - "cache_read_input_token_cost": 1.5e-6, - "input_cost_per_token": 1.5e-5, + "cache_creation_input_token_cost": 0.00001875, + "cache_read_input_token_cost": 0.0000015, + "input_cost_per_token": 0.000015, "litellm_provider": "bedrock_converse", "max_input_tokens": 200000, "max_output_tokens": 32000, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 7.5e-5, + "output_cost_per_token": 0.000075, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -8112,19 +9637,19 @@ "tool_use_system_prompt_tokens": 159 }, "eu.anthropic.claude-sonnet-4-20250514-v1:0": { - "cache_creation_input_token_cost": 3.75e-6, + "cache_creation_input_token_cost": 0.00000375, "cache_read_input_token_cost": 3e-7, - "input_cost_per_token": 3e-6, - "input_cost_per_token_above_200k_tokens": 6e-6, - "output_cost_per_token_above_200k_tokens": 2.25e-5, - "cache_creation_input_token_cost_above_200k_tokens": 7.5e-6, + "input_cost_per_token": 0.000003, + "input_cost_per_token_above_200k_tokens": 0.000006, + "output_cost_per_token_above_200k_tokens": 0.0000225, + "cache_creation_input_token_cost_above_200k_tokens": 0.0000075, "cache_read_input_token_cost_above_200k_tokens": 6e-7, "litellm_provider": "bedrock_converse", "max_input_tokens": 1000000, "max_output_tokens": 64000, "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -8142,19 +9667,19 @@ "tool_use_system_prompt_tokens": 159 }, "eu.anthropic.claude-sonnet-4-5-20250929-v1:0": { - "cache_creation_input_token_cost": 4.125e-6, + "cache_creation_input_token_cost": 0.000004125, "cache_read_input_token_cost": 3.3e-7, - "input_cost_per_token": 3.3e-6, - "input_cost_per_token_above_200k_tokens": 6.6e-6, - "output_cost_per_token_above_200k_tokens": 2.475e-5, - "cache_creation_input_token_cost_above_200k_tokens": 8.25e-6, + "input_cost_per_token": 0.0000033, + "input_cost_per_token_above_200k_tokens": 0.0000066, + "output_cost_per_token_above_200k_tokens": 0.00002475, + "cache_creation_input_token_cost_above_200k_tokens": 0.00000825, "cache_read_input_token_cost_above_200k_tokens": 6.6e-7, "litellm_provider": "bedrock_converse", "max_input_tokens": 200000, "max_output_tokens": 64000, "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 1.65e-5, + "output_cost_per_token": 0.0000165, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -8194,13 +9719,13 @@ "supports_tool_choice": false }, "eu.mistral.pixtral-large-2502-v1:0": { - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "bedrock_converse", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 6e-6, + "output_cost_per_token": 0.000006, "supports_function_calling": true, "supports_tool_choice": false }, @@ -8208,73 +9733,97 @@ "litellm_provider": "fal_ai", "mode": "image_generation", "output_cost_per_image": 0.0398, - "supported_endpoints": ["/v1/images/generations"] + "supported_endpoints": [ + "/v1/images/generations" + ] }, "fal_ai/fal-ai/flux-pro/v1.1": { "litellm_provider": "fal_ai", "mode": "image_generation", "output_cost_per_image": 0.04, - "supported_endpoints": ["/v1/images/generations"] + "supported_endpoints": [ + "/v1/images/generations" + ] }, "fal_ai/fal-ai/flux-pro/v1.1-ultra": { "litellm_provider": "fal_ai", "mode": "image_generation", "output_cost_per_image": 0.06, - "supported_endpoints": ["/v1/images/generations"] + "supported_endpoints": [ + "/v1/images/generations" + ] }, "fal_ai/fal-ai/flux/schnell": { "litellm_provider": "fal_ai", "mode": "image_generation", "output_cost_per_image": 0.003, - "supported_endpoints": ["/v1/images/generations"] + "supported_endpoints": [ + "/v1/images/generations" + ] }, "fal_ai/fal-ai/bytedance/seedream/v3/text-to-image": { "litellm_provider": "fal_ai", "mode": "image_generation", "output_cost_per_image": 0.03, - "supported_endpoints": ["/v1/images/generations"] + "supported_endpoints": [ + "/v1/images/generations" + ] }, "fal_ai/fal-ai/bytedance/dreamina/v3.1/text-to-image": { "litellm_provider": "fal_ai", "mode": "image_generation", "output_cost_per_image": 0.03, - "supported_endpoints": ["/v1/images/generations"] + "supported_endpoints": [ + "/v1/images/generations" + ] }, "fal_ai/fal-ai/ideogram/v3": { "litellm_provider": "fal_ai", "mode": "image_generation", "output_cost_per_image": 0.06, - "supported_endpoints": ["/v1/images/generations"] + "supported_endpoints": [ + "/v1/images/generations" + ] }, "fal_ai/fal-ai/imagen4/preview": { "litellm_provider": "fal_ai", "mode": "image_generation", "output_cost_per_image": 0.0398, - "supported_endpoints": ["/v1/images/generations"] + "supported_endpoints": [ + "/v1/images/generations" + ] }, "fal_ai/fal-ai/imagen4/preview/fast": { "litellm_provider": "fal_ai", "mode": "image_generation", "output_cost_per_image": 0.02, - "supported_endpoints": ["/v1/images/generations"] + "supported_endpoints": [ + "/v1/images/generations" + ] }, "fal_ai/fal-ai/imagen4/preview/ultra": { "litellm_provider": "fal_ai", "mode": "image_generation", "output_cost_per_image": 0.06, - "supported_endpoints": ["/v1/images/generations"] + "supported_endpoints": [ + "/v1/images/generations" + ] }, "fal_ai/fal-ai/recraft/v3/text-to-image": { "litellm_provider": "fal_ai", "mode": "image_generation", "output_cost_per_image": 0.0398, - "supported_endpoints": ["/v1/images/generations"] + "supported_endpoints": [ + "/v1/images/generations" + ] }, "fal_ai/fal-ai/stable-diffusion-v35-medium": { "litellm_provider": "fal_ai", "mode": "image_generation", "output_cost_per_image": 0.0398, - "supported_endpoints": ["/v1/images/generations"] + "supported_endpoints": [ + "/v1/images/generations" + ] }, "featherless_ai/featherless-ai/Qwerky-72B": { "litellm_provider": "featherless_ai", @@ -8296,9 +9845,9 @@ "output_cost_per_token": 2e-7 }, "fireworks-ai-56b-to-176b": { - "input_cost_per_token": 1.2e-6, + "input_cost_per_token": 0.0000012, "litellm_provider": "fireworks_ai", - "output_cost_per_token": 1.2e-6 + "output_cost_per_token": 0.0000012 }, "fireworks-ai-above-16b": { "input_cost_per_token": 9e-7, @@ -8306,19 +9855,19 @@ "output_cost_per_token": 9e-7 }, "fireworks-ai-default": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "fireworks_ai", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "fireworks-ai-embedding-150m-to-350m": { "input_cost_per_token": 1.6e-8, "litellm_provider": "fireworks_ai-embedding-models", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "fireworks-ai-embedding-up-to-150m": { "input_cost_per_token": 8e-9, "litellm_provider": "fireworks_ai-embedding-models", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "fireworks-ai-moe-up-to-56b": { "input_cost_per_token": 5e-7, @@ -8336,42 +9885,42 @@ "max_input_tokens": 512, "max_tokens": 512, "mode": "embedding", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "source": "https://fireworks.ai/pricing" }, "fireworks_ai/accounts/fireworks/models/deepseek-coder-v2-instruct": { - "input_cost_per_token": 1.2e-6, + "input_cost_per_token": 0.0000012, "litellm_provider": "fireworks_ai", "max_input_tokens": 65536, "max_output_tokens": 65536, "max_tokens": 65536, "mode": "chat", - "output_cost_per_token": 1.2e-6, + "output_cost_per_token": 0.0000012, "source": "https://fireworks.ai/pricing", "supports_function_calling": false, "supports_response_schema": true, "supports_tool_choice": false }, "fireworks_ai/accounts/fireworks/models/deepseek-r1": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "fireworks_ai", "max_input_tokens": 128000, "max_output_tokens": 20480, "max_tokens": 20480, "mode": "chat", - "output_cost_per_token": 8e-6, + "output_cost_per_token": 0.000008, "source": "https://fireworks.ai/pricing", "supports_response_schema": true, "supports_tool_choice": false }, "fireworks_ai/accounts/fireworks/models/deepseek-r1-0528": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "fireworks_ai", "max_input_tokens": 160000, "max_output_tokens": 160000, "max_tokens": 160000, "mode": "chat", - "output_cost_per_token": 8e-6, + "output_cost_per_token": 0.000008, "source": "https://fireworks.ai/pricing", "supports_response_schema": true, "supports_tool_choice": false @@ -8383,7 +9932,7 @@ "max_output_tokens": 20480, "max_tokens": 20480, "mode": "chat", - "output_cost_per_token": 2.19e-6, + "output_cost_per_token": 0.00000219, "source": "https://fireworks.ai/pricing", "supports_response_schema": true, "supports_tool_choice": false @@ -8419,7 +9968,19 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 1.68e-6, + "output_cost_per_token": 0.00000168, + "source": "https://fireworks.ai/pricing", + "supports_response_schema": true, + "supports_tool_choice": true + }, + "fireworks_ai/accounts/fireworks/models/deepseek-v3p1-terminus": { + "input_cost_per_token": 5.6e-7, + "litellm_provider": "fireworks_ai", + "max_input_tokens": 128000, + "max_output_tokens": 8192, + "max_tokens": 8192, + "mode": "chat", + "output_cost_per_token": 0.00000168, "source": "https://fireworks.ai/pricing", "supports_response_schema": true, "supports_tool_choice": true @@ -8444,7 +10005,7 @@ "max_output_tokens": 96000, "max_tokens": 96000, "mode": "chat", - "output_cost_per_token": 2.19e-6, + "output_cost_per_token": 0.00000219, "source": "https://fireworks.ai/models/fireworks/glm-4p5", "supports_function_calling": true, "supports_response_schema": true, @@ -8496,20 +10057,34 @@ "max_output_tokens": 16384, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 2.5e-6, + "output_cost_per_token": 0.0000025, "source": "https://fireworks.ai/models/fireworks/kimi-k2-instruct", "supports_function_calling": true, "supports_response_schema": true, "supports_tool_choice": true }, + "fireworks_ai/accounts/fireworks/models/kimi-k2-thinking": { + "input_cost_per_token": 6e-7, + "litellm_provider": "fireworks_ai", + "max_input_tokens": 262144, + "max_output_tokens": 262144, + "max_tokens": 262144, + "mode": "chat", + "output_cost_per_token": 0.0000025, + "source": "https://fireworks.ai/pricing", + "supports_function_calling": true, + "supports_response_schema": true, + "supports_tool_choice": true, + "supports_web_search": true + }, "fireworks_ai/accounts/fireworks/models/llama-v3p1-405b-instruct": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "fireworks_ai", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 3e-6, + "output_cost_per_token": 0.000003, "source": "https://fireworks.ai/pricing", "supports_function_calling": true, "supports_response_schema": true, @@ -8606,13 +10181,13 @@ "supports_tool_choice": false }, "fireworks_ai/accounts/fireworks/models/mixtral-8x22b-instruct-hf": { - "input_cost_per_token": 1.2e-6, + "input_cost_per_token": 0.0000012, "litellm_provider": "fireworks_ai", "max_input_tokens": 65536, "max_output_tokens": 65536, "max_tokens": 65536, "mode": "chat", - "output_cost_per_token": 1.2e-6, + "output_cost_per_token": 0.0000012, "source": "https://fireworks.ai/pricing", "supports_function_calling": true, "supports_response_schema": true, @@ -8645,13 +10220,13 @@ "supports_tool_choice": false }, "fireworks_ai/accounts/fireworks/models/yi-large": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "fireworks_ai", "max_input_tokens": 32768, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 3e-6, + "output_cost_per_token": 0.000003, "source": "https://fireworks.ai/pricing", "supports_function_calling": false, "supports_response_schema": true, @@ -8663,7 +10238,7 @@ "max_input_tokens": 8192, "max_tokens": 8192, "mode": "embedding", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "source": "https://fireworks.ai/pricing" }, "fireworks_ai/nomic-ai/nomic-embed-text-v1.5": { @@ -8672,7 +10247,7 @@ "max_input_tokens": 8192, "max_tokens": 8192, "mode": "embedding", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "source": "https://fireworks.ai/pricing" }, "fireworks_ai/thenlper/gte-base": { @@ -8681,7 +10256,7 @@ "max_input_tokens": 512, "max_tokens": 512, "mode": "embedding", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "source": "https://fireworks.ai/pricing" }, "fireworks_ai/thenlper/gte-large": { @@ -8690,7 +10265,7 @@ "max_input_tokens": 512, "max_tokens": 512, "mode": "embedding", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "source": "https://fireworks.ai/pricing" }, "friendliai/meta-llama-3.1-70b-instruct": { @@ -8733,85 +10308,85 @@ "output_cost_per_token_batches": 2e-7 }, "ft:davinci-002": { - "input_cost_per_token": 2e-6, - "input_cost_per_token_batches": 1e-6, + "input_cost_per_token": 0.000002, + "input_cost_per_token_batches": 0.000001, "litellm_provider": "text-completion-openai", "max_input_tokens": 16384, "max_output_tokens": 4096, "max_tokens": 16384, "mode": "completion", - "output_cost_per_token": 2e-6, - "output_cost_per_token_batches": 1e-6 + "output_cost_per_token": 0.000002, + "output_cost_per_token_batches": 0.000001 }, "ft:gpt-3.5-turbo": { - "input_cost_per_token": 3e-6, - "input_cost_per_token_batches": 1.5e-6, + "input_cost_per_token": 0.000003, + "input_cost_per_token_batches": 0.0000015, "litellm_provider": "openai", "max_input_tokens": 16385, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 6e-6, - "output_cost_per_token_batches": 3e-6, + "output_cost_per_token": 0.000006, + "output_cost_per_token_batches": 0.000003, "supports_system_messages": true, "supports_tool_choice": true }, "ft:gpt-3.5-turbo-0125": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "openai", "max_input_tokens": 16385, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 6e-6, + "output_cost_per_token": 0.000006, "supports_system_messages": true, "supports_tool_choice": true }, "ft:gpt-3.5-turbo-0613": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "openai", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 6e-6, + "output_cost_per_token": 0.000006, "supports_system_messages": true, "supports_tool_choice": true }, "ft:gpt-3.5-turbo-1106": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "openai", "max_input_tokens": 16385, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 6e-6, + "output_cost_per_token": 0.000006, "supports_system_messages": true, "supports_tool_choice": true }, "ft:gpt-4-0613": { - "input_cost_per_token": 3e-5, + "input_cost_per_token": 0.00003, "litellm_provider": "openai", "max_input_tokens": 8192, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 6e-5, + "output_cost_per_token": 0.00006, "source": "OpenAI needs to add pricing for this ft model, will be updated when added by OpenAI. Defaulting to base model pricing", "supports_function_calling": true, "supports_system_messages": true, "supports_tool_choice": true }, "ft:gpt-4o-2024-08-06": { - "input_cost_per_token": 3.75e-6, - "input_cost_per_token_batches": 1.875e-6, + "input_cost_per_token": 0.00000375, + "input_cost_per_token_batches": 0.000001875, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 1.5e-5, - "output_cost_per_token_batches": 7.5e-6, + "output_cost_per_token": 0.000015, + "output_cost_per_token_batches": 0.0000075, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_pdf_input": true, @@ -8821,14 +10396,14 @@ "supports_vision": true }, "ft:gpt-4o-2024-11-20": { - "cache_creation_input_token_cost": 1.875e-6, - "input_cost_per_token": 3.75e-6, + "cache_creation_input_token_cost": 0.000001875, + "input_cost_per_token": 0.00000375, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_pdf_input": true, @@ -8847,7 +10422,7 @@ "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 1.2e-6, + "output_cost_per_token": 0.0000012, "output_cost_per_token_batches": 6e-7, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -8869,7 +10444,7 @@ "max_tokens": 8192, "mode": "chat", "output_cost_per_character": 3.75e-7, - "output_cost_per_token": 1.5e-6, + "output_cost_per_token": 0.0000015, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing#google_models", "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -8887,7 +10462,7 @@ "max_tokens": 8192, "mode": "chat", "output_cost_per_character": 3.75e-7, - "output_cost_per_token": 1.5e-6, + "output_cost_per_token": 0.0000015, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models", "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -8905,7 +10480,7 @@ "max_tokens": 8192, "mode": "chat", "output_cost_per_character": 3.75e-7, - "output_cost_per_token": 1.5e-6, + "output_cost_per_token": 0.0000015, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models", "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -8922,7 +10497,7 @@ "max_video_length": 2, "max_videos_per_prompt": 1, "mode": "chat", - "output_cost_per_token": 1.5e-6, + "output_cost_per_token": 0.0000015, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models", "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -8941,7 +10516,7 @@ "max_video_length": 2, "max_videos_per_prompt": 1, "mode": "chat", - "output_cost_per_token": 1.5e-6, + "output_cost_per_token": 0.0000015, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models", "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -8959,7 +10534,7 @@ "max_tokens": 8192, "mode": "chat", "output_cost_per_character": 3.75e-7, - "output_cost_per_token": 1.5e-6, + "output_cost_per_token": 0.0000015, "source": "As of Jun, 2024. There is no available doc on vertex ai pricing gemini-1.0-ultra-001. Using gemini-1.0-pro pricing. Got max_tokens info here: https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models", "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -8976,23 +10551,23 @@ "max_tokens": 8192, "mode": "chat", "output_cost_per_character": 3.75e-7, - "output_cost_per_token": 1.5e-6, + "output_cost_per_token": 0.0000015, "source": "As of Jun, 2024. There is no available doc on vertex ai pricing gemini-1.0-ultra-001. Using gemini-1.0-pro pricing. Got max_tokens info here: https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models", "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_tool_choice": true }, "gemini-1.5-flash": { - "input_cost_per_audio_per_second": 2e-6, - "input_cost_per_audio_per_second_above_128k_tokens": 4e-6, + "input_cost_per_audio_per_second": 0.000002, + "input_cost_per_audio_per_second_above_128k_tokens": 0.000004, "input_cost_per_character": 1.875e-8, "input_cost_per_character_above_128k_tokens": 2.5e-7, - "input_cost_per_image": 2e-5, - "input_cost_per_image_above_128k_tokens": 4e-5, + "input_cost_per_image": 0.00002, + "input_cost_per_image_above_128k_tokens": 0.00004, "input_cost_per_token": 7.5e-8, - "input_cost_per_token_above_128k_tokens": 1e-6, - "input_cost_per_video_per_second": 2e-5, - "input_cost_per_video_per_second_above_128k_tokens": 4e-5, + "input_cost_per_token_above_128k_tokens": 0.000001, + "input_cost_per_video_per_second": 0.00002, + "input_cost_per_video_per_second_above_128k_tokens": 0.00004, "litellm_provider": "vertex_ai-language-models", "max_audio_length_hours": 8.4, "max_audio_per_prompt": 1, @@ -9018,16 +10593,16 @@ }, "gemini-1.5-flash-001": { "deprecation_date": "2025-05-24", - "input_cost_per_audio_per_second": 2e-6, - "input_cost_per_audio_per_second_above_128k_tokens": 4e-6, + "input_cost_per_audio_per_second": 0.000002, + "input_cost_per_audio_per_second_above_128k_tokens": 0.000004, "input_cost_per_character": 1.875e-8, "input_cost_per_character_above_128k_tokens": 2.5e-7, - "input_cost_per_image": 2e-5, - "input_cost_per_image_above_128k_tokens": 4e-5, + "input_cost_per_image": 0.00002, + "input_cost_per_image_above_128k_tokens": 0.00004, "input_cost_per_token": 7.5e-8, - "input_cost_per_token_above_128k_tokens": 1e-6, - "input_cost_per_video_per_second": 2e-5, - "input_cost_per_video_per_second_above_128k_tokens": 4e-5, + "input_cost_per_token_above_128k_tokens": 0.000001, + "input_cost_per_video_per_second": 0.00002, + "input_cost_per_video_per_second_above_128k_tokens": 0.00004, "litellm_provider": "vertex_ai-language-models", "max_audio_length_hours": 8.4, "max_audio_per_prompt": 1, @@ -9053,16 +10628,16 @@ }, "gemini-1.5-flash-002": { "deprecation_date": "2025-09-24", - "input_cost_per_audio_per_second": 2e-6, - "input_cost_per_audio_per_second_above_128k_tokens": 4e-6, + "input_cost_per_audio_per_second": 0.000002, + "input_cost_per_audio_per_second_above_128k_tokens": 0.000004, "input_cost_per_character": 1.875e-8, "input_cost_per_character_above_128k_tokens": 2.5e-7, - "input_cost_per_image": 2e-5, - "input_cost_per_image_above_128k_tokens": 4e-5, + "input_cost_per_image": 0.00002, + "input_cost_per_image_above_128k_tokens": 0.00004, "input_cost_per_token": 7.5e-8, - "input_cost_per_token_above_128k_tokens": 1e-6, - "input_cost_per_video_per_second": 2e-5, - "input_cost_per_video_per_second_above_128k_tokens": 4e-5, + "input_cost_per_token_above_128k_tokens": 0.000001, + "input_cost_per_video_per_second": 0.00002, + "input_cost_per_video_per_second_above_128k_tokens": 0.00004, "litellm_provider": "vertex_ai-language-models", "max_audio_length_hours": 8.4, "max_audio_per_prompt": 1, @@ -9087,16 +10662,16 @@ "supports_vision": true }, "gemini-1.5-flash-exp-0827": { - "input_cost_per_audio_per_second": 2e-6, - "input_cost_per_audio_per_second_above_128k_tokens": 4e-6, + "input_cost_per_audio_per_second": 0.000002, + "input_cost_per_audio_per_second_above_128k_tokens": 0.000004, "input_cost_per_character": 1.875e-8, "input_cost_per_character_above_128k_tokens": 2.5e-7, - "input_cost_per_image": 2e-5, - "input_cost_per_image_above_128k_tokens": 4e-5, + "input_cost_per_image": 0.00002, + "input_cost_per_image_above_128k_tokens": 0.00004, "input_cost_per_token": 4.688e-9, - "input_cost_per_token_above_128k_tokens": 1e-6, - "input_cost_per_video_per_second": 2e-5, - "input_cost_per_video_per_second_above_128k_tokens": 4e-5, + "input_cost_per_token_above_128k_tokens": 0.000001, + "input_cost_per_video_per_second": 0.00002, + "input_cost_per_video_per_second_above_128k_tokens": 0.00004, "litellm_provider": "vertex_ai-language-models", "max_audio_length_hours": 8.4, "max_audio_per_prompt": 1, @@ -9121,16 +10696,16 @@ "supports_vision": true }, "gemini-1.5-flash-preview-0514": { - "input_cost_per_audio_per_second": 2e-6, - "input_cost_per_audio_per_second_above_128k_tokens": 4e-6, + "input_cost_per_audio_per_second": 0.000002, + "input_cost_per_audio_per_second_above_128k_tokens": 0.000004, "input_cost_per_character": 1.875e-8, "input_cost_per_character_above_128k_tokens": 2.5e-7, - "input_cost_per_image": 2e-5, - "input_cost_per_image_above_128k_tokens": 4e-5, + "input_cost_per_image": 0.00002, + "input_cost_per_image_above_128k_tokens": 0.00004, "input_cost_per_token": 7.5e-8, - "input_cost_per_token_above_128k_tokens": 1e-6, - "input_cost_per_video_per_second": 2e-5, - "input_cost_per_video_per_second_above_128k_tokens": 4e-5, + "input_cost_per_token_above_128k_tokens": 0.000001, + "input_cost_per_video_per_second": 0.00002, + "input_cost_per_video_per_second_above_128k_tokens": 0.00004, "litellm_provider": "vertex_ai-language-models", "max_audio_length_hours": 8.4, "max_audio_per_prompt": 1, @@ -9154,14 +10729,14 @@ "supports_vision": true }, "gemini-1.5-pro": { - "input_cost_per_audio_per_second": 3.125e-5, - "input_cost_per_audio_per_second_above_128k_tokens": 6.25e-5, + "input_cost_per_audio_per_second": 0.00003125, + "input_cost_per_audio_per_second_above_128k_tokens": 0.0000625, "input_cost_per_character": 3.125e-7, "input_cost_per_character_above_128k_tokens": 6.25e-7, "input_cost_per_image": 0.00032875, "input_cost_per_image_above_128k_tokens": 0.0006575, - "input_cost_per_token": 1.25e-6, - "input_cost_per_token_above_128k_tokens": 2.5e-6, + "input_cost_per_token": 0.00000125, + "input_cost_per_token_above_128k_tokens": 0.0000025, "input_cost_per_video_per_second": 0.00032875, "input_cost_per_video_per_second_above_128k_tokens": 0.0006575, "litellm_provider": "vertex_ai-language-models", @@ -9169,10 +10744,10 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_character": 1.25e-6, - "output_cost_per_character_above_128k_tokens": 2.5e-6, - "output_cost_per_token": 5e-6, - "output_cost_per_token_above_128k_tokens": 1e-5, + "output_cost_per_character": 0.00000125, + "output_cost_per_character_above_128k_tokens": 0.0000025, + "output_cost_per_token": 0.000005, + "output_cost_per_token_above_128k_tokens": 0.00001, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models", "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -9184,14 +10759,14 @@ }, "gemini-1.5-pro-001": { "deprecation_date": "2025-05-24", - "input_cost_per_audio_per_second": 3.125e-5, - "input_cost_per_audio_per_second_above_128k_tokens": 6.25e-5, + "input_cost_per_audio_per_second": 0.00003125, + "input_cost_per_audio_per_second_above_128k_tokens": 0.0000625, "input_cost_per_character": 3.125e-7, "input_cost_per_character_above_128k_tokens": 6.25e-7, "input_cost_per_image": 0.00032875, "input_cost_per_image_above_128k_tokens": 0.0006575, - "input_cost_per_token": 1.25e-6, - "input_cost_per_token_above_128k_tokens": 2.5e-6, + "input_cost_per_token": 0.00000125, + "input_cost_per_token_above_128k_tokens": 0.0000025, "input_cost_per_video_per_second": 0.00032875, "input_cost_per_video_per_second_above_128k_tokens": 0.0006575, "litellm_provider": "vertex_ai-language-models", @@ -9199,10 +10774,10 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_character": 1.25e-6, - "output_cost_per_character_above_128k_tokens": 2.5e-6, - "output_cost_per_token": 5e-6, - "output_cost_per_token_above_128k_tokens": 1e-5, + "output_cost_per_character": 0.00000125, + "output_cost_per_character_above_128k_tokens": 0.0000025, + "output_cost_per_token": 0.000005, + "output_cost_per_token_above_128k_tokens": 0.00001, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models", "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -9213,14 +10788,14 @@ }, "gemini-1.5-pro-002": { "deprecation_date": "2025-09-24", - "input_cost_per_audio_per_second": 3.125e-5, - "input_cost_per_audio_per_second_above_128k_tokens": 6.25e-5, + "input_cost_per_audio_per_second": 0.00003125, + "input_cost_per_audio_per_second_above_128k_tokens": 0.0000625, "input_cost_per_character": 3.125e-7, "input_cost_per_character_above_128k_tokens": 6.25e-7, "input_cost_per_image": 0.00032875, "input_cost_per_image_above_128k_tokens": 0.0006575, - "input_cost_per_token": 1.25e-6, - "input_cost_per_token_above_128k_tokens": 2.5e-6, + "input_cost_per_token": 0.00000125, + "input_cost_per_token_above_128k_tokens": 0.0000025, "input_cost_per_video_per_second": 0.00032875, "input_cost_per_video_per_second_above_128k_tokens": 0.0006575, "litellm_provider": "vertex_ai-language-models", @@ -9228,10 +10803,10 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_character": 1.25e-6, - "output_cost_per_character_above_128k_tokens": 2.5e-6, - "output_cost_per_token": 5e-6, - "output_cost_per_token_above_128k_tokens": 1e-5, + "output_cost_per_character": 0.00000125, + "output_cost_per_character_above_128k_tokens": 0.0000025, + "output_cost_per_token": 0.000005, + "output_cost_per_token_above_128k_tokens": 0.00001, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#gemini-1.5-pro", "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -9241,8 +10816,8 @@ "supports_vision": true }, "gemini-1.5-pro-preview-0215": { - "input_cost_per_audio_per_second": 3.125e-5, - "input_cost_per_audio_per_second_above_128k_tokens": 6.25e-5, + "input_cost_per_audio_per_second": 0.00003125, + "input_cost_per_audio_per_second_above_128k_tokens": 0.0000625, "input_cost_per_character": 3.125e-7, "input_cost_per_character_above_128k_tokens": 6.25e-7, "input_cost_per_image": 0.00032875, @@ -9256,8 +10831,8 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_character": 1.25e-6, - "output_cost_per_character_above_128k_tokens": 2.5e-6, + "output_cost_per_character": 0.00000125, + "output_cost_per_character_above_128k_tokens": 0.0000025, "output_cost_per_token": 3.125e-7, "output_cost_per_token_above_128k_tokens": 6.25e-7, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models", @@ -9268,8 +10843,8 @@ "supports_tool_choice": true }, "gemini-1.5-pro-preview-0409": { - "input_cost_per_audio_per_second": 3.125e-5, - "input_cost_per_audio_per_second_above_128k_tokens": 6.25e-5, + "input_cost_per_audio_per_second": 0.00003125, + "input_cost_per_audio_per_second_above_128k_tokens": 0.0000625, "input_cost_per_character": 3.125e-7, "input_cost_per_character_above_128k_tokens": 6.25e-7, "input_cost_per_image": 0.00032875, @@ -9283,8 +10858,8 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_character": 1.25e-6, - "output_cost_per_character_above_128k_tokens": 2.5e-6, + "output_cost_per_character": 0.00000125, + "output_cost_per_character_above_128k_tokens": 0.0000025, "output_cost_per_token": 3.125e-7, "output_cost_per_token_above_128k_tokens": 6.25e-7, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models", @@ -9294,8 +10869,8 @@ "supports_tool_choice": true }, "gemini-1.5-pro-preview-0514": { - "input_cost_per_audio_per_second": 3.125e-5, - "input_cost_per_audio_per_second_above_128k_tokens": 6.25e-5, + "input_cost_per_audio_per_second": 0.00003125, + "input_cost_per_audio_per_second_above_128k_tokens": 0.0000625, "input_cost_per_character": 3.125e-7, "input_cost_per_character_above_128k_tokens": 6.25e-7, "input_cost_per_image": 0.00032875, @@ -9309,8 +10884,8 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_character": 1.25e-6, - "output_cost_per_character_above_128k_tokens": 2.5e-6, + "output_cost_per_character": 0.00000125, + "output_cost_per_character_above_128k_tokens": 0.0000025, "output_cost_per_token": 3.125e-7, "output_cost_per_token_above_128k_tokens": 6.25e-7, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models", @@ -9337,8 +10912,16 @@ "mode": "chat", "output_cost_per_token": 4e-7, "source": "https://ai.google.dev/pricing#2_0flash", - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text", "image"], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text", + "image" + ], "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -9354,7 +10937,7 @@ "gemini-2.0-flash-001": { "cache_read_input_token_cost": 3.75e-8, "deprecation_date": "2026-02-05", - "input_cost_per_audio_token": 1e-6, + "input_cost_per_audio_token": 0.000001, "input_cost_per_token": 1.5e-7, "litellm_provider": "vertex_ai-language-models", "max_audio_length_hours": 8.4, @@ -9369,8 +10952,16 @@ "mode": "chat", "output_cost_per_token": 6e-7, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing", - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text", "image"], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text", + "image" + ], "supports_audio_output": true, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -9409,8 +11000,16 @@ "output_cost_per_token": 6e-7, "output_cost_per_token_above_128k_tokens": 0, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing", - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text", "image"], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text", + "image" + ], "supports_audio_output": true, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -9437,8 +11036,15 @@ "mode": "chat", "output_cost_per_token": 3e-7, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#gemini-2.0-flash", - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text"], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text" + ], "supports_audio_output": true, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -9466,8 +11072,15 @@ "mode": "chat", "output_cost_per_token": 3e-7, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#gemini-2.0-flash", - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text"], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text" + ], "supports_audio_output": true, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -9480,10 +11093,10 @@ }, "gemini-2.0-flash-live-preview-04-09": { "cache_read_input_token_cost": 7.5e-8, - "input_cost_per_audio_token": 3e-6, - "input_cost_per_image": 3e-6, + "input_cost_per_audio_token": 0.000003, + "input_cost_per_image": 0.000003, "input_cost_per_token": 5e-7, - "input_cost_per_video_per_second": 3e-6, + "input_cost_per_video_per_second": 0.000003, "litellm_provider": "vertex_ai-language-models", "max_audio_length_hours": 8.4, "max_audio_per_prompt": 1, @@ -9495,13 +11108,24 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_audio_token": 1.2e-5, - "output_cost_per_token": 2e-6, + "output_cost_per_audio_token": 0.000012, + "output_cost_per_token": 0.000002, "rpm": 10, "source": "https://cloud.google.com/vertex-ai/docs/generative-ai/model-reference/gemini#gemini-2-0-flash-live-preview-04-09", - "supported_endpoints": ["/v1/chat/completions", "/v1/completions"], - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text", "audio"], + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/completions" + ], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text", + "audio" + ], "supports_audio_output": true, "supports_function_calling": true, "supports_pdf_input": true, @@ -9532,8 +11156,16 @@ "mode": "chat", "output_cost_per_token": 4e-7, "source": "https://ai.google.dev/pricing#2_0flash", - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text", "image"], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text", + "image" + ], "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -9546,7 +11178,7 @@ "supports_web_search": true }, "gemini-2.0-flash-thinking-exp": { - "cache_read_input_token_cost": 0.0, + "cache_read_input_token_cost": 0, "input_cost_per_audio_per_second": 0, "input_cost_per_audio_per_second_above_128k_tokens": 0, "input_cost_per_character": 0, @@ -9573,8 +11205,16 @@ "output_cost_per_token": 0, "output_cost_per_token_above_128k_tokens": 0, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#gemini-2.0-flash", - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text", "image"], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text", + "image" + ], "supports_audio_output": true, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -9586,7 +11226,7 @@ "supports_web_search": true }, "gemini-2.0-flash-thinking-exp-01-21": { - "cache_read_input_token_cost": 0.0, + "cache_read_input_token_cost": 0, "input_cost_per_audio_per_second": 0, "input_cost_per_audio_per_second_above_128k_tokens": 0, "input_cost_per_character": 0, @@ -9613,8 +11253,16 @@ "output_cost_per_token": 0, "output_cost_per_token_above_128k_tokens": 0, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#gemini-2.0-flash", - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text", "image"], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text", + "image" + ], "supports_audio_output": false, "supports_function_calling": false, "supports_parallel_function_calling": true, @@ -9628,8 +11276,8 @@ }, "gemini-2.0-pro-exp-02-05": { "cache_read_input_token_cost": 3.125e-7, - "input_cost_per_token": 1.25e-6, - "input_cost_per_token_above_200k_tokens": 2.5e-6, + "input_cost_per_token": 0.00000125, + "input_cost_per_token_above_200k_tokens": 0.0000025, "litellm_provider": "vertex_ai-language-models", "max_audio_length_hours": 8.4, "max_audio_per_prompt": 1, @@ -9641,12 +11289,22 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_token": 1e-5, - "output_cost_per_token_above_200k_tokens": 1.5e-5, + "output_cost_per_token": 0.00001, + "output_cost_per_token_above_200k_tokens": 0.000015, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing", - "supported_endpoints": ["/v1/chat/completions", "/v1/completions"], - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text"], + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/completions" + ], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text" + ], "supports_audio_input": true, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -9661,7 +11319,7 @@ }, "gemini-2.5-flash": { "cache_read_input_token_cost": 3e-8, - "input_cost_per_audio_token": 1e-6, + "input_cost_per_audio_token": 0.000001, "input_cost_per_token": 3e-7, "litellm_provider": "vertex_ai-language-models", "max_audio_length_hours": 8.4, @@ -9674,12 +11332,23 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_reasoning_token": 2.5e-6, - "output_cost_per_token": 2.5e-6, + "output_cost_per_reasoning_token": 0.0000025, + "output_cost_per_token": 0.0000025, "source": "https://ai.google.dev/gemini-api/docs/models#gemini-2.5-flash-preview", - "supported_endpoints": ["/v1/chat/completions", "/v1/completions", "/v1/batch"], - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text"], + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/completions", + "/v1/batch" + ], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text" + ], "supports_audio_output": false, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -9695,7 +11364,7 @@ }, "gemini-2.5-flash-image": { "cache_read_input_token_cost": 3e-8, - "input_cost_per_audio_token": 1e-6, + "input_cost_per_audio_token": 0.000001, "input_cost_per_token": 3e-7, "litellm_provider": "vertex_ai-language-models", "max_audio_length_hours": 8.4, @@ -9709,13 +11378,25 @@ "max_videos_per_prompt": 10, "mode": "image_generation", "output_cost_per_image": 0.039, - "output_cost_per_reasoning_token": 2.5e-6, - "output_cost_per_token": 2.5e-6, + "output_cost_per_reasoning_token": 0.0000025, + "output_cost_per_token": 0.0000025, "rpm": 100000, "source": "https://ai.google.dev/gemini-api/docs/pricing#gemini-2.5-flash-image", - "supported_endpoints": ["/v1/chat/completions", "/v1/completions", "/v1/batch"], - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text", "image"], + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/completions", + "/v1/batch" + ], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text", + "image" + ], "supports_audio_output": false, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -9731,7 +11412,7 @@ }, "gemini-2.5-flash-image-preview": { "cache_read_input_token_cost": 7.5e-8, - "input_cost_per_audio_token": 1e-6, + "input_cost_per_audio_token": 0.000001, "input_cost_per_token": 3e-7, "litellm_provider": "vertex_ai-language-models", "max_audio_length_hours": 8.4, @@ -9745,13 +11426,25 @@ "max_videos_per_prompt": 10, "mode": "image_generation", "output_cost_per_image": 0.039, - "output_cost_per_reasoning_token": 3e-5, - "output_cost_per_token": 3e-5, + "output_cost_per_reasoning_token": 0.00003, + "output_cost_per_token": 0.00003, "rpm": 100000, "source": "https://ai.google.dev/gemini-api/docs/models#gemini-2.5-flash-preview", - "supported_endpoints": ["/v1/chat/completions", "/v1/completions", "/v1/batch"], - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text", "image"], + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/completions", + "/v1/batch" + ], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text", + "image" + ], "supports_audio_output": false, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -9765,6 +11458,39 @@ "supports_web_search": true, "tpm": 8000000 }, + "gemini-3-pro-image-preview": { + "input_cost_per_image": 0.0011, + "input_cost_per_token": 0.000002, + "input_cost_per_token_batches": 0.000001, + "litellm_provider": "vertex_ai-language-models", + "max_input_tokens": 65536, + "max_output_tokens": 32768, + "max_tokens": 65536, + "mode": "image_generation", + "output_cost_per_image": 0.134, + "output_cost_per_token": 0.000012, + "output_cost_per_token_batches": 0.000006, + "source": "https://ai.google.dev/gemini-api/docs/pricing", + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/completions", + "/v1/batch" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text", + "image" + ], + "supports_function_calling": false, + "supports_prompt_caching": true, + "supports_response_schema": true, + "supports_system_messages": true, + "supports_vision": true, + "supports_web_search": true + }, "gemini-2.5-flash-lite": { "cache_read_input_token_cost": 2.5e-8, "input_cost_per_audio_token": 5e-7, @@ -9783,9 +11509,20 @@ "output_cost_per_reasoning_token": 4e-7, "output_cost_per_token": 4e-7, "source": "https://ai.google.dev/gemini-api/docs/models#gemini-2.5-flash-preview", - "supported_endpoints": ["/v1/chat/completions", "/v1/completions", "/v1/batch"], - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text"], + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/completions", + "/v1/batch" + ], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text" + ], "supports_audio_output": false, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -9817,9 +11554,20 @@ "output_cost_per_reasoning_token": 4e-7, "output_cost_per_token": 4e-7, "source": "https://developers.googleblog.com/en/continuing-to-bring-you-our-latest-models-with-an-improved-gemini-2-5-flash-and-flash-lite-release/", - "supported_endpoints": ["/v1/chat/completions", "/v1/completions", "/v1/batch"], - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text"], + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/completions", + "/v1/batch" + ], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text" + ], "supports_audio_output": false, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -9835,7 +11583,7 @@ }, "gemini-2.5-flash-preview-09-2025": { "cache_read_input_token_cost": 7.5e-8, - "input_cost_per_audio_token": 1e-6, + "input_cost_per_audio_token": 0.000001, "input_cost_per_token": 3e-7, "litellm_provider": "vertex_ai-language-models", "max_audio_length_hours": 8.4, @@ -9848,12 +11596,23 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_reasoning_token": 2.5e-6, - "output_cost_per_token": 2.5e-6, + "output_cost_per_reasoning_token": 0.0000025, + "output_cost_per_token": 0.0000025, "source": "https://developers.googleblog.com/en/continuing-to-bring-you-our-latest-models-with-an-improved-gemini-2-5-flash-and-flash-lite-release/", - "supported_endpoints": ["/v1/chat/completions", "/v1/completions", "/v1/batch"], - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text"], + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/completions", + "/v1/batch" + ], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text" + ], "supports_audio_output": false, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -9869,7 +11628,7 @@ }, "gemini-live-2.5-flash-preview-native-audio-09-2025": { "cache_read_input_token_cost": 7.5e-8, - "input_cost_per_audio_token": 3e-6, + "input_cost_per_audio_token": 0.000003, "input_cost_per_token": 3e-7, "litellm_provider": "vertex_ai-language-models", "max_audio_length_hours": 8.4, @@ -9882,12 +11641,23 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_audio_token": 1.2e-5, - "output_cost_per_token": 2e-6, + "output_cost_per_audio_token": 0.000012, + "output_cost_per_token": 0.000002, "source": "https://ai.google.dev/gemini-api/docs/pricing", - "supported_endpoints": ["/v1/chat/completions", "/v1/completions"], - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text", "audio"], + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/completions" + ], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text", + "audio" + ], "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -9903,7 +11673,7 @@ }, "gemini/gemini-live-2.5-flash-preview-native-audio-09-2025": { "cache_read_input_token_cost": 7.5e-8, - "input_cost_per_audio_token": 3e-6, + "input_cost_per_audio_token": 0.000003, "input_cost_per_token": 3e-7, "litellm_provider": "gemini", "max_audio_length_hours": 8.4, @@ -9916,13 +11686,24 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_audio_token": 1.2e-5, - "output_cost_per_token": 2e-6, + "output_cost_per_audio_token": 0.000012, + "output_cost_per_token": 0.000002, "rpm": 100000, "source": "https://ai.google.dev/gemini-api/docs/pricing", - "supported_endpoints": ["/v1/chat/completions", "/v1/completions"], - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text", "audio"], + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/completions" + ], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text", + "audio" + ], "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -9955,9 +11736,20 @@ "output_cost_per_reasoning_token": 4e-7, "output_cost_per_token": 4e-7, "source": "https://ai.google.dev/gemini-api/docs/models#gemini-2.5-flash-preview", - "supported_endpoints": ["/v1/chat/completions", "/v1/completions", "/v1/batch"], - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text"], + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/completions", + "/v1/batch" + ], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text" + ], "supports_audio_output": false, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -9973,7 +11765,7 @@ }, "gemini-2.5-flash-preview-04-17": { "cache_read_input_token_cost": 3.75e-8, - "input_cost_per_audio_token": 1e-6, + "input_cost_per_audio_token": 0.000001, "input_cost_per_token": 1.5e-7, "litellm_provider": "vertex_ai-language-models", "max_audio_length_hours": 8.4, @@ -9986,12 +11778,23 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_reasoning_token": 3.5e-6, + "output_cost_per_reasoning_token": 0.0000035, "output_cost_per_token": 6e-7, "source": "https://ai.google.dev/gemini-api/docs/models#gemini-2.5-flash-preview", - "supported_endpoints": ["/v1/chat/completions", "/v1/completions", "/v1/batch"], - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text"], + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/completions", + "/v1/batch" + ], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text" + ], "supports_audio_output": false, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -10006,7 +11809,7 @@ }, "gemini-2.5-flash-preview-05-20": { "cache_read_input_token_cost": 7.5e-8, - "input_cost_per_audio_token": 1e-6, + "input_cost_per_audio_token": 0.000001, "input_cost_per_token": 3e-7, "litellm_provider": "vertex_ai-language-models", "max_audio_length_hours": 8.4, @@ -10019,12 +11822,23 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_reasoning_token": 2.5e-6, - "output_cost_per_token": 2.5e-6, + "output_cost_per_reasoning_token": 0.0000025, + "output_cost_per_token": 0.0000025, "source": "https://ai.google.dev/gemini-api/docs/models#gemini-2.5-flash-preview", - "supported_endpoints": ["/v1/chat/completions", "/v1/completions", "/v1/batch"], - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text"], + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/completions", + "/v1/batch" + ], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text" + ], "supports_audio_output": false, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -10041,8 +11855,8 @@ "gemini-2.5-pro": { "cache_read_input_token_cost": 1.25e-7, "cache_creation_input_token_cost_above_200k_tokens": 2.5e-7, - "input_cost_per_token": 1.25e-6, - "input_cost_per_token_above_200k_tokens": 2.5e-6, + "input_cost_per_token": 0.00000125, + "input_cost_per_token_above_200k_tokens": 0.0000025, "litellm_provider": "vertex_ai-language-models", "max_audio_length_hours": 8.4, "max_audio_per_prompt": 1, @@ -10054,12 +11868,22 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_token": 1e-5, - "output_cost_per_token_above_200k_tokens": 1.5e-5, + "output_cost_per_token": 0.00001, + "output_cost_per_token_above_200k_tokens": 0.000015, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing", - "supported_endpoints": ["/v1/chat/completions", "/v1/completions"], - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text"], + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/completions" + ], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text" + ], "supports_audio_input": true, "supports_function_calling": true, "supports_pdf_input": true, @@ -10073,10 +11897,12 @@ "supports_web_search": true }, "gemini-3-pro-preview": { - "cache_read_input_token_cost": 1.25e-7, + "cache_read_input_token_cost": 2e-7, + "cache_read_input_token_cost_above_200k_tokens": 4e-7, "cache_creation_input_token_cost_above_200k_tokens": 2.5e-7, - "input_cost_per_token": 2e-6, - "input_cost_per_token_above_200k_tokens": 4e-6, + "input_cost_per_token": 0.000002, + "input_cost_per_token_above_200k_tokens": 0.000004, + "input_cost_per_token_batches": 0.000001, "litellm_provider": "vertex_ai-language-models", "max_audio_length_hours": 8.4, "max_audio_per_prompt": 1, @@ -10088,12 +11914,72 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_token": 1.2e-5, - "output_cost_per_token_above_200k_tokens": 1.8e-5, + "output_cost_per_token": 0.000012, + "output_cost_per_token_above_200k_tokens": 0.000018, + "output_cost_per_token_batches": 0.000006, + "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing", + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/completions", + "/v1/batch" + ], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text" + ], + "supports_audio_input": true, + "supports_function_calling": true, + "supports_pdf_input": true, + "supports_prompt_caching": true, + "supports_reasoning": true, + "supports_response_schema": true, + "supports_system_messages": true, + "supports_tool_choice": true, + "supports_video_input": true, + "supports_vision": true, + "supports_web_search": true + }, + "vertex_ai/gemini-3-pro-preview": { + "cache_read_input_token_cost": 2e-7, + "cache_read_input_token_cost_above_200k_tokens": 4e-7, + "cache_creation_input_token_cost_above_200k_tokens": 2.5e-7, + "input_cost_per_token": 0.000002, + "input_cost_per_token_above_200k_tokens": 0.000004, + "input_cost_per_token_batches": 0.000001, + "litellm_provider": "vertex_ai", + "max_audio_length_hours": 8.4, + "max_audio_per_prompt": 1, + "max_images_per_prompt": 3000, + "max_input_tokens": 1048576, + "max_output_tokens": 65535, + "max_pdf_size_mb": 30, + "max_tokens": 65535, + "max_video_length": 1, + "max_videos_per_prompt": 10, + "mode": "chat", + "output_cost_per_token": 0.000012, + "output_cost_per_token_above_200k_tokens": 0.000018, + "output_cost_per_token_batches": 0.000006, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing", - "supported_endpoints": ["/v1/chat/completions", "/v1/completions"], - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text"], + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/completions", + "/v1/batch" + ], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text" + ], "supports_audio_input": true, "supports_function_calling": true, "supports_pdf_input": true, @@ -10108,8 +11994,8 @@ }, "gemini-2.5-pro-exp-03-25": { "cache_read_input_token_cost": 3.125e-7, - "input_cost_per_token": 1.25e-6, - "input_cost_per_token_above_200k_tokens": 2.5e-6, + "input_cost_per_token": 0.00000125, + "input_cost_per_token_above_200k_tokens": 0.0000025, "litellm_provider": "vertex_ai-language-models", "max_audio_length_hours": 8.4, "max_audio_per_prompt": 1, @@ -10121,12 +12007,22 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_token": 1e-5, - "output_cost_per_token_above_200k_tokens": 1.5e-5, + "output_cost_per_token": 0.00001, + "output_cost_per_token_above_200k_tokens": 0.000015, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing", - "supported_endpoints": ["/v1/chat/completions", "/v1/completions"], - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text"], + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/completions" + ], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text" + ], "supports_audio_input": true, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -10141,9 +12037,9 @@ }, "gemini-2.5-pro-preview-03-25": { "cache_read_input_token_cost": 3.125e-7, - "input_cost_per_audio_token": 1.25e-6, - "input_cost_per_token": 1.25e-6, - "input_cost_per_token_above_200k_tokens": 2.5e-6, + "input_cost_per_audio_token": 0.00000125, + "input_cost_per_token": 0.00000125, + "input_cost_per_token_above_200k_tokens": 0.0000025, "litellm_provider": "vertex_ai-language-models", "max_audio_length_hours": 8.4, "max_audio_per_prompt": 1, @@ -10155,12 +12051,23 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_token": 1e-5, - "output_cost_per_token_above_200k_tokens": 1.5e-5, + "output_cost_per_token": 0.00001, + "output_cost_per_token_above_200k_tokens": 0.000015, "source": "https://ai.google.dev/gemini-api/docs/models#gemini-2.5-flash-preview", - "supported_endpoints": ["/v1/chat/completions", "/v1/completions", "/v1/batch"], - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text"], + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/completions", + "/v1/batch" + ], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text" + ], "supports_audio_output": false, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -10175,9 +12082,9 @@ }, "gemini-2.5-pro-preview-05-06": { "cache_read_input_token_cost": 3.125e-7, - "input_cost_per_audio_token": 1.25e-6, - "input_cost_per_token": 1.25e-6, - "input_cost_per_token_above_200k_tokens": 2.5e-6, + "input_cost_per_audio_token": 0.00000125, + "input_cost_per_token": 0.00000125, + "input_cost_per_token_above_200k_tokens": 0.0000025, "litellm_provider": "vertex_ai-language-models", "max_audio_length_hours": 8.4, "max_audio_per_prompt": 1, @@ -10189,13 +12096,26 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_token": 1e-5, - "output_cost_per_token_above_200k_tokens": 1.5e-5, + "output_cost_per_token": 0.00001, + "output_cost_per_token_above_200k_tokens": 0.000015, "source": "https://ai.google.dev/gemini-api/docs/models#gemini-2.5-flash-preview", - "supported_endpoints": ["/v1/chat/completions", "/v1/completions", "/v1/batch"], - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text"], - "supported_regions": ["global"], + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/completions", + "/v1/batch" + ], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text" + ], + "supported_regions": [ + "global" + ], "supports_audio_output": false, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -10210,9 +12130,9 @@ }, "gemini-2.5-pro-preview-06-05": { "cache_read_input_token_cost": 3.125e-7, - "input_cost_per_audio_token": 1.25e-6, - "input_cost_per_token": 1.25e-6, - "input_cost_per_token_above_200k_tokens": 2.5e-6, + "input_cost_per_audio_token": 0.00000125, + "input_cost_per_token": 0.00000125, + "input_cost_per_token_above_200k_tokens": 0.0000025, "litellm_provider": "vertex_ai-language-models", "max_audio_length_hours": 8.4, "max_audio_per_prompt": 1, @@ -10224,12 +12144,23 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_token": 1e-5, - "output_cost_per_token_above_200k_tokens": 1.5e-5, + "output_cost_per_token": 0.00001, + "output_cost_per_token_above_200k_tokens": 0.000015, "source": "https://ai.google.dev/gemini-api/docs/models#gemini-2.5-flash-preview", - "supported_endpoints": ["/v1/chat/completions", "/v1/completions", "/v1/batch"], - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text"], + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/completions", + "/v1/batch" + ], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text" + ], "supports_audio_output": false, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -10245,8 +12176,8 @@ "gemini-2.5-pro-preview-tts": { "cache_read_input_token_cost": 3.125e-7, "input_cost_per_audio_token": 7e-7, - "input_cost_per_token": 1.25e-6, - "input_cost_per_token_above_200k_tokens": 2.5e-6, + "input_cost_per_token": 0.00000125, + "input_cost_per_token_above_200k_tokens": 0.0000025, "litellm_provider": "vertex_ai-language-models", "max_audio_length_hours": 8.4, "max_audio_per_prompt": 1, @@ -10258,11 +12189,15 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_token": 1e-5, - "output_cost_per_token_above_200k_tokens": 1.5e-5, + "output_cost_per_token": 0.00001, + "output_cost_per_token_above_200k_tokens": 0.000015, "source": "https://ai.google.dev/gemini-api/docs/pricing#gemini-2.5-pro-preview", - "supported_modalities": ["text"], - "supported_output_modalities": ["audio"], + "supported_modalities": [ + "text" + ], + "supported_output_modalities": [ + "audio" + ], "supports_audio_output": false, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -10309,7 +12244,7 @@ "max_tokens": 8192, "mode": "chat", "output_cost_per_character": 3.75e-7, - "output_cost_per_token": 1.5e-6, + "output_cost_per_token": 0.0000015, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing", "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -10341,7 +12276,7 @@ "max_video_length": 2, "max_videos_per_prompt": 1, "mode": "chat", - "output_cost_per_token": 1.5e-6, + "output_cost_per_token": 0.0000015, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models", "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -10386,7 +12321,7 @@ "tpm": 4000000 }, "gemini/gemini-1.5-flash-001": { - "cache_creation_input_token_cost": 1e-6, + "cache_creation_input_token_cost": 0.000001, "cache_read_input_token_cost": 1.875e-8, "deprecation_date": "2025-05-24", "input_cost_per_token": 7.5e-8, @@ -10415,7 +12350,7 @@ "tpm": 4000000 }, "gemini/gemini-1.5-flash-002": { - "cache_creation_input_token_cost": 1e-6, + "cache_creation_input_token_cost": 0.000001, "cache_read_input_token_cost": 1.875e-8, "deprecation_date": "2025-09-24", "input_cost_per_token": 7.5e-8, @@ -10572,15 +12507,15 @@ "tpm": 4000000 }, "gemini/gemini-1.5-pro": { - "input_cost_per_token": 3.5e-6, - "input_cost_per_token_above_128k_tokens": 7e-6, + "input_cost_per_token": 0.0000035, + "input_cost_per_token_above_128k_tokens": 0.000007, "litellm_provider": "gemini", "max_input_tokens": 2097152, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 1.05e-5, - "output_cost_per_token_above_128k_tokens": 2.1e-5, + "output_cost_per_token": 0.0000105, + "output_cost_per_token_above_128k_tokens": 0.000021, "rpm": 1000, "source": "https://ai.google.dev/pricing", "supports_function_calling": true, @@ -10592,15 +12527,15 @@ }, "gemini/gemini-1.5-pro-001": { "deprecation_date": "2025-05-24", - "input_cost_per_token": 3.5e-6, - "input_cost_per_token_above_128k_tokens": 7e-6, + "input_cost_per_token": 0.0000035, + "input_cost_per_token_above_128k_tokens": 0.000007, "litellm_provider": "gemini", "max_input_tokens": 2097152, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 1.05e-5, - "output_cost_per_token_above_128k_tokens": 2.1e-5, + "output_cost_per_token": 0.0000105, + "output_cost_per_token_above_128k_tokens": 0.000021, "rpm": 1000, "source": "https://ai.google.dev/pricing", "supports_function_calling": true, @@ -10613,15 +12548,15 @@ }, "gemini/gemini-1.5-pro-002": { "deprecation_date": "2025-09-24", - "input_cost_per_token": 3.5e-6, - "input_cost_per_token_above_128k_tokens": 7e-6, + "input_cost_per_token": 0.0000035, + "input_cost_per_token_above_128k_tokens": 0.000007, "litellm_provider": "gemini", "max_input_tokens": 2097152, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 1.05e-5, - "output_cost_per_token_above_128k_tokens": 2.1e-5, + "output_cost_per_token": 0.0000105, + "output_cost_per_token_above_128k_tokens": 0.000021, "rpm": 1000, "source": "https://ai.google.dev/pricing", "supports_function_calling": true, @@ -10633,15 +12568,15 @@ "tpm": 4000000 }, "gemini/gemini-1.5-pro-exp-0801": { - "input_cost_per_token": 3.5e-6, - "input_cost_per_token_above_128k_tokens": 7e-6, + "input_cost_per_token": 0.0000035, + "input_cost_per_token_above_128k_tokens": 0.000007, "litellm_provider": "gemini", "max_input_tokens": 2097152, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 1.05e-5, - "output_cost_per_token_above_128k_tokens": 2.1e-5, + "output_cost_per_token": 0.0000105, + "output_cost_per_token_above_128k_tokens": 0.000021, "rpm": 1000, "source": "https://ai.google.dev/pricing", "supports_function_calling": true, @@ -10671,15 +12606,15 @@ "tpm": 4000000 }, "gemini/gemini-1.5-pro-latest": { - "input_cost_per_token": 3.5e-6, - "input_cost_per_token_above_128k_tokens": 7e-6, + "input_cost_per_token": 0.0000035, + "input_cost_per_token_above_128k_tokens": 0.000007, "litellm_provider": "gemini", "max_input_tokens": 1048576, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 1.05e-6, - "output_cost_per_token_above_128k_tokens": 2.1e-5, + "output_cost_per_token": 0.00000105, + "output_cost_per_token_above_128k_tokens": 0.000021, "rpm": 1000, "source": "https://ai.google.dev/pricing", "supports_function_calling": true, @@ -10707,8 +12642,16 @@ "output_cost_per_token": 4e-7, "rpm": 10000, "source": "https://ai.google.dev/pricing#2_0flash", - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text", "image"], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text", + "image" + ], "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -10739,8 +12682,16 @@ "output_cost_per_token": 4e-7, "rpm": 10000, "source": "https://ai.google.dev/pricing#2_0flash", - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text", "image"], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text", + "image" + ], "supports_audio_output": false, "supports_function_calling": true, "supports_prompt_caching": true, @@ -10752,7 +12703,7 @@ "tpm": 10000000 }, "gemini/gemini-2.0-flash-exp": { - "cache_read_input_token_cost": 0.0, + "cache_read_input_token_cost": 0, "input_cost_per_audio_per_second": 0, "input_cost_per_audio_per_second_above_128k_tokens": 0, "input_cost_per_character": 0, @@ -10780,8 +12731,16 @@ "output_cost_per_token_above_128k_tokens": 0, "rpm": 10, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#gemini-2.0-flash", - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text", "image"], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text", + "image" + ], "supports_audio_output": true, "supports_function_calling": true, "supports_prompt_caching": true, @@ -10809,8 +12768,15 @@ "output_cost_per_token": 3e-7, "rpm": 4000, "source": "https://ai.google.dev/gemini-api/docs/pricing#gemini-2.0-flash-lite", - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text"], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text" + ], "supports_audio_output": true, "supports_function_calling": true, "supports_prompt_caching": true, @@ -10839,8 +12805,15 @@ "output_cost_per_token": 3e-7, "rpm": 60000, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#gemini-2.0-flash-lite", - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text"], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text" + ], "supports_audio_output": false, "supports_function_calling": true, "supports_prompt_caching": true, @@ -10853,10 +12826,10 @@ }, "gemini/gemini-2.0-flash-live-001": { "cache_read_input_token_cost": 7.5e-8, - "input_cost_per_audio_token": 2.1e-6, - "input_cost_per_image": 2.1e-6, + "input_cost_per_audio_token": 0.0000021, + "input_cost_per_image": 0.0000021, "input_cost_per_token": 3.5e-7, - "input_cost_per_video_per_second": 2.1e-6, + "input_cost_per_video_per_second": 0.0000021, "litellm_provider": "gemini", "max_audio_length_hours": 8.4, "max_audio_per_prompt": 1, @@ -10868,13 +12841,24 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_audio_token": 8.5e-6, - "output_cost_per_token": 1.5e-6, + "output_cost_per_audio_token": 0.0000085, + "output_cost_per_token": 0.0000015, "rpm": 10, "source": "https://ai.google.dev/gemini-api/docs/models#gemini-2-0-flash-live-001", - "supported_endpoints": ["/v1/chat/completions", "/v1/completions"], - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text", "audio"], + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/completions" + ], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text", + "audio" + ], "supports_audio_output": true, "supports_function_calling": true, "supports_pdf_input": true, @@ -10906,8 +12890,16 @@ "output_cost_per_token": 4e-7, "rpm": 10000, "source": "https://ai.google.dev/pricing#2_0flash", - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text", "image"], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text", + "image" + ], "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -10920,7 +12912,7 @@ "tpm": 10000000 }, "gemini/gemini-2.0-flash-thinking-exp": { - "cache_read_input_token_cost": 0.0, + "cache_read_input_token_cost": 0, "input_cost_per_audio_per_second": 0, "input_cost_per_audio_per_second_above_128k_tokens": 0, "input_cost_per_character": 0, @@ -10948,8 +12940,16 @@ "output_cost_per_token_above_128k_tokens": 0, "rpm": 10, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#gemini-2.0-flash", - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text", "image"], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text", + "image" + ], "supports_audio_output": true, "supports_function_calling": true, "supports_prompt_caching": true, @@ -10961,7 +12961,7 @@ "tpm": 4000000 }, "gemini/gemini-2.0-flash-thinking-exp-01-21": { - "cache_read_input_token_cost": 0.0, + "cache_read_input_token_cost": 0, "input_cost_per_audio_per_second": 0, "input_cost_per_audio_per_second_above_128k_tokens": 0, "input_cost_per_character": 0, @@ -10989,8 +12989,16 @@ "output_cost_per_token_above_128k_tokens": 0, "rpm": 10, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#gemini-2.0-flash", - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text", "image"], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text", + "image" + ], "supports_audio_output": true, "supports_function_calling": true, "supports_prompt_caching": true, @@ -11003,7 +13011,7 @@ "tpm": 4000000 }, "gemini/gemini-2.0-pro-exp-02-05": { - "cache_read_input_token_cost": 0.0, + "cache_read_input_token_cost": 0, "input_cost_per_audio_per_second": 0, "input_cost_per_audio_per_second_above_128k_tokens": 0, "input_cost_per_character": 0, @@ -11045,7 +13053,7 @@ }, "gemini/gemini-2.5-flash": { "cache_read_input_token_cost": 3e-8, - "input_cost_per_audio_token": 1e-6, + "input_cost_per_audio_token": 0.000001, "input_cost_per_token": 3e-7, "litellm_provider": "gemini", "max_audio_length_hours": 8.4, @@ -11058,13 +13066,24 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_reasoning_token": 2.5e-6, - "output_cost_per_token": 2.5e-6, + "output_cost_per_reasoning_token": 0.0000025, + "output_cost_per_token": 0.0000025, "rpm": 100000, "source": "https://ai.google.dev/gemini-api/docs/models#gemini-2.5-flash-preview", - "supported_endpoints": ["/v1/chat/completions", "/v1/completions", "/v1/batch"], - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text"], + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/completions", + "/v1/batch" + ], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text" + ], "supports_audio_output": false, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -11081,7 +13100,7 @@ }, "gemini/gemini-2.5-flash-image": { "cache_read_input_token_cost": 3e-8, - "input_cost_per_audio_token": 1e-6, + "input_cost_per_audio_token": 0.000001, "input_cost_per_token": 3e-7, "litellm_provider": "vertex_ai-language-models", "max_audio_length_hours": 8.4, @@ -11096,13 +13115,25 @@ "max_videos_per_prompt": 10, "mode": "image_generation", "output_cost_per_image": 0.039, - "output_cost_per_reasoning_token": 2.5e-6, - "output_cost_per_token": 2.5e-6, + "output_cost_per_reasoning_token": 0.0000025, + "output_cost_per_token": 0.0000025, "rpm": 100000, "source": "https://ai.google.dev/gemini-api/docs/pricing#gemini-2.5-flash-image", - "supported_endpoints": ["/v1/chat/completions", "/v1/completions", "/v1/batch"], - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text", "image"], + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/completions", + "/v1/batch" + ], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text", + "image" + ], "supports_audio_output": false, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -11118,7 +13149,7 @@ }, "gemini/gemini-2.5-flash-image-preview": { "cache_read_input_token_cost": 7.5e-8, - "input_cost_per_audio_token": 1e-6, + "input_cost_per_audio_token": 0.000001, "input_cost_per_token": 3e-7, "litellm_provider": "gemini", "max_audio_length_hours": 8.4, @@ -11132,13 +13163,25 @@ "max_videos_per_prompt": 10, "mode": "image_generation", "output_cost_per_image": 0.039, - "output_cost_per_reasoning_token": 3e-5, - "output_cost_per_token": 3e-5, + "output_cost_per_reasoning_token": 0.00003, + "output_cost_per_token": 0.00003, "rpm": 100000, "source": "https://ai.google.dev/gemini-api/docs/models#gemini-2.5-flash-preview", - "supported_endpoints": ["/v1/chat/completions", "/v1/completions", "/v1/batch"], - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text", "image"], + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/completions", + "/v1/batch" + ], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text", + "image" + ], "supports_audio_output": false, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -11152,6 +13195,41 @@ "supports_web_search": true, "tpm": 8000000 }, + "gemini/gemini-3-pro-image-preview": { + "input_cost_per_image": 0.0011, + "input_cost_per_token": 0.000002, + "input_cost_per_token_batches": 0.000001, + "litellm_provider": "gemini", + "max_input_tokens": 65536, + "max_output_tokens": 32768, + "max_tokens": 65536, + "mode": "image_generation", + "output_cost_per_image": 0.134, + "output_cost_per_token": 0.000012, + "rpm": 1000, + "tpm": 4000000, + "output_cost_per_token_batches": 0.000006, + "source": "https://ai.google.dev/gemini-api/docs/pricing", + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/completions", + "/v1/batch" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text", + "image" + ], + "supports_function_calling": false, + "supports_prompt_caching": true, + "supports_response_schema": true, + "supports_system_messages": true, + "supports_vision": true, + "supports_web_search": true + }, "gemini/gemini-2.5-flash-lite": { "cache_read_input_token_cost": 2.5e-8, "input_cost_per_audio_token": 5e-7, @@ -11171,9 +13249,20 @@ "output_cost_per_token": 4e-7, "rpm": 15, "source": "https://ai.google.dev/gemini-api/docs/models#gemini-2.5-flash-lite", - "supported_endpoints": ["/v1/chat/completions", "/v1/completions", "/v1/batch"], - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text"], + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/completions", + "/v1/batch" + ], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text" + ], "supports_audio_output": false, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -11207,9 +13296,20 @@ "output_cost_per_token": 4e-7, "rpm": 15, "source": "https://developers.googleblog.com/en/continuing-to-bring-you-our-latest-models-with-an-improved-gemini-2-5-flash-and-flash-lite-release/", - "supported_endpoints": ["/v1/chat/completions", "/v1/completions", "/v1/batch"], - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text"], + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/completions", + "/v1/batch" + ], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text" + ], "supports_audio_output": false, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -11226,7 +13326,7 @@ }, "gemini/gemini-2.5-flash-preview-09-2025": { "cache_read_input_token_cost": 7.5e-8, - "input_cost_per_audio_token": 1e-6, + "input_cost_per_audio_token": 0.000001, "input_cost_per_token": 3e-7, "litellm_provider": "gemini", "max_audio_length_hours": 8.4, @@ -11239,13 +13339,24 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_reasoning_token": 2.5e-6, - "output_cost_per_token": 2.5e-6, + "output_cost_per_reasoning_token": 0.0000025, + "output_cost_per_token": 0.0000025, "rpm": 15, "source": "https://developers.googleblog.com/en/continuing-to-bring-you-our-latest-models-with-an-improved-gemini-2-5-flash-and-flash-lite-release/", - "supported_endpoints": ["/v1/chat/completions", "/v1/completions", "/v1/batch"], - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text"], + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/completions", + "/v1/batch" + ], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text" + ], "supports_audio_output": false, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -11262,7 +13373,7 @@ }, "gemini/gemini-flash-latest": { "cache_read_input_token_cost": 7.5e-8, - "input_cost_per_audio_token": 1e-6, + "input_cost_per_audio_token": 0.000001, "input_cost_per_token": 3e-7, "litellm_provider": "gemini", "max_audio_length_hours": 8.4, @@ -11275,13 +13386,24 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_reasoning_token": 2.5e-6, - "output_cost_per_token": 2.5e-6, + "output_cost_per_reasoning_token": 0.0000025, + "output_cost_per_token": 0.0000025, "rpm": 15, "source": "https://developers.googleblog.com/en/continuing-to-bring-you-our-latest-models-with-an-improved-gemini-2-5-flash-and-flash-lite-release/", - "supported_endpoints": ["/v1/chat/completions", "/v1/completions", "/v1/batch"], - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text"], + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/completions", + "/v1/batch" + ], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text" + ], "supports_audio_output": false, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -11315,9 +13437,20 @@ "output_cost_per_token": 4e-7, "rpm": 15, "source": "https://developers.googleblog.com/en/continuing-to-bring-you-our-latest-models-with-an-improved-gemini-2-5-flash-and-flash-lite-release/", - "supported_endpoints": ["/v1/chat/completions", "/v1/completions", "/v1/batch"], - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text"], + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/completions", + "/v1/batch" + ], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text" + ], "supports_audio_output": false, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -11351,9 +13484,20 @@ "output_cost_per_token": 4e-7, "rpm": 15, "source": "https://ai.google.dev/gemini-api/docs/models#gemini-2.5-flash-lite", - "supported_endpoints": ["/v1/chat/completions", "/v1/completions", "/v1/batch"], - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text"], + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/completions", + "/v1/batch" + ], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text" + ], "supports_audio_output": false, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -11370,7 +13514,7 @@ }, "gemini/gemini-2.5-flash-preview-04-17": { "cache_read_input_token_cost": 3.75e-8, - "input_cost_per_audio_token": 1e-6, + "input_cost_per_audio_token": 0.000001, "input_cost_per_token": 1.5e-7, "litellm_provider": "gemini", "max_audio_length_hours": 8.4, @@ -11383,13 +13527,23 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_reasoning_token": 3.5e-6, + "output_cost_per_reasoning_token": 0.0000035, "output_cost_per_token": 6e-7, "rpm": 10, "source": "https://ai.google.dev/gemini-api/docs/models#gemini-2.5-flash-preview", - "supported_endpoints": ["/v1/chat/completions", "/v1/completions"], - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text"], + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/completions" + ], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text" + ], "supports_audio_output": false, "supports_function_calling": true, "supports_pdf_input": true, @@ -11404,7 +13558,7 @@ }, "gemini/gemini-2.5-flash-preview-05-20": { "cache_read_input_token_cost": 7.5e-8, - "input_cost_per_audio_token": 1e-6, + "input_cost_per_audio_token": 0.000001, "input_cost_per_token": 3e-7, "litellm_provider": "gemini", "max_audio_length_hours": 8.4, @@ -11417,13 +13571,23 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_reasoning_token": 2.5e-6, - "output_cost_per_token": 2.5e-6, + "output_cost_per_reasoning_token": 0.0000025, + "output_cost_per_token": 0.0000025, "rpm": 10, "source": "https://ai.google.dev/gemini-api/docs/models#gemini-2.5-flash-preview", - "supported_endpoints": ["/v1/chat/completions", "/v1/completions"], - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text"], + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/completions" + ], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text" + ], "supports_audio_output": false, "supports_function_calling": true, "supports_pdf_input": true, @@ -11439,7 +13603,7 @@ }, "gemini/gemini-2.5-flash-preview-tts": { "cache_read_input_token_cost": 3.75e-8, - "input_cost_per_audio_token": 1e-6, + "input_cost_per_audio_token": 0.000001, "input_cost_per_token": 1.5e-7, "litellm_provider": "gemini", "max_audio_length_hours": 8.4, @@ -11452,13 +13616,20 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_reasoning_token": 3.5e-6, + "output_cost_per_reasoning_token": 0.0000035, "output_cost_per_token": 6e-7, "rpm": 10, "source": "https://ai.google.dev/gemini-api/docs/models#gemini-2.5-flash-preview", - "supported_endpoints": ["/v1/chat/completions", "/v1/completions"], - "supported_modalities": ["text"], - "supported_output_modalities": ["audio"], + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/completions" + ], + "supported_modalities": [ + "text" + ], + "supported_output_modalities": [ + "audio" + ], "supports_audio_output": false, "supports_function_calling": true, "supports_prompt_caching": true, @@ -11472,8 +13643,8 @@ }, "gemini/gemini-2.5-pro": { "cache_read_input_token_cost": 3.125e-7, - "input_cost_per_token": 1.25e-6, - "input_cost_per_token_above_200k_tokens": 2.5e-6, + "input_cost_per_token": 0.00000125, + "input_cost_per_token_above_200k_tokens": 0.0000025, "litellm_provider": "gemini", "max_audio_length_hours": 8.4, "max_audio_per_prompt": 1, @@ -11485,13 +13656,23 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_token": 1e-5, - "output_cost_per_token_above_200k_tokens": 1.5e-5, + "output_cost_per_token": 0.00001, + "output_cost_per_token_above_200k_tokens": 0.000015, "rpm": 2000, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing", - "supported_endpoints": ["/v1/chat/completions", "/v1/completions"], - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text"], + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/completions" + ], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text" + ], "supports_audio_input": true, "supports_function_calling": true, "supports_pdf_input": true, @@ -11506,9 +13687,11 @@ "tpm": 800000 }, "gemini/gemini-3-pro-preview": { - "cache_read_input_token_cost": 3.125e-7, - "input_cost_per_token": 2e-6, - "input_cost_per_token_above_200k_tokens": 4e-6, + "cache_read_input_token_cost": 2e-7, + "cache_read_input_token_cost_above_200k_tokens": 4e-7, + "input_cost_per_token": 0.000002, + "input_cost_per_token_above_200k_tokens": 0.000004, + "input_cost_per_token_batches": 0.000001, "litellm_provider": "gemini", "max_audio_length_hours": 8.4, "max_audio_per_prompt": 1, @@ -11520,13 +13703,25 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_token": 1.2e-5, - "output_cost_per_token_above_200k_tokens": 1.8e-5, + "output_cost_per_token": 0.000012, + "output_cost_per_token_above_200k_tokens": 0.000018, + "output_cost_per_token_batches": 0.000006, "rpm": 2000, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing", - "supported_endpoints": ["/v1/chat/completions", "/v1/completions"], - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text"], + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/completions", + "/v1/batch" + ], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text" + ], "supports_audio_input": true, "supports_function_calling": true, "supports_pdf_input": true, @@ -11541,9 +13736,9 @@ "tpm": 800000 }, "gemini/gemini-2.5-pro-exp-03-25": { - "cache_read_input_token_cost": 0.0, - "input_cost_per_token": 0.0, - "input_cost_per_token_above_200k_tokens": 0.0, + "cache_read_input_token_cost": 0, + "input_cost_per_token": 0, + "input_cost_per_token_above_200k_tokens": 0, "litellm_provider": "gemini", "max_audio_length_hours": 8.4, "max_audio_per_prompt": 1, @@ -11555,13 +13750,23 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_token": 0.0, - "output_cost_per_token_above_200k_tokens": 0.0, + "output_cost_per_token": 0, + "output_cost_per_token_above_200k_tokens": 0, "rpm": 5, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing", - "supported_endpoints": ["/v1/chat/completions", "/v1/completions"], - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text"], + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/completions" + ], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text" + ], "supports_audio_input": true, "supports_function_calling": true, "supports_pdf_input": true, @@ -11577,8 +13782,8 @@ "gemini/gemini-2.5-pro-preview-03-25": { "cache_read_input_token_cost": 3.125e-7, "input_cost_per_audio_token": 7e-7, - "input_cost_per_token": 1.25e-6, - "input_cost_per_token_above_200k_tokens": 2.5e-6, + "input_cost_per_token": 0.00000125, + "input_cost_per_token_above_200k_tokens": 0.0000025, "litellm_provider": "gemini", "max_audio_length_hours": 8.4, "max_audio_per_prompt": 1, @@ -11590,12 +13795,19 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_token": 1e-5, - "output_cost_per_token_above_200k_tokens": 1.5e-5, + "output_cost_per_token": 0.00001, + "output_cost_per_token_above_200k_tokens": 0.000015, "rpm": 10000, "source": "https://ai.google.dev/gemini-api/docs/pricing#gemini-2.5-pro-preview", - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text"], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text" + ], "supports_audio_output": false, "supports_function_calling": true, "supports_pdf_input": true, @@ -11610,8 +13822,8 @@ "gemini/gemini-2.5-pro-preview-05-06": { "cache_read_input_token_cost": 3.125e-7, "input_cost_per_audio_token": 7e-7, - "input_cost_per_token": 1.25e-6, - "input_cost_per_token_above_200k_tokens": 2.5e-6, + "input_cost_per_token": 0.00000125, + "input_cost_per_token_above_200k_tokens": 0.0000025, "litellm_provider": "gemini", "max_audio_length_hours": 8.4, "max_audio_per_prompt": 1, @@ -11623,12 +13835,19 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_token": 1e-5, - "output_cost_per_token_above_200k_tokens": 1.5e-5, + "output_cost_per_token": 0.00001, + "output_cost_per_token_above_200k_tokens": 0.000015, "rpm": 10000, "source": "https://ai.google.dev/gemini-api/docs/pricing#gemini-2.5-pro-preview", - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text"], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text" + ], "supports_audio_output": false, "supports_function_calling": true, "supports_pdf_input": true, @@ -11644,8 +13863,8 @@ "gemini/gemini-2.5-pro-preview-06-05": { "cache_read_input_token_cost": 3.125e-7, "input_cost_per_audio_token": 7e-7, - "input_cost_per_token": 1.25e-6, - "input_cost_per_token_above_200k_tokens": 2.5e-6, + "input_cost_per_token": 0.00000125, + "input_cost_per_token_above_200k_tokens": 0.0000025, "litellm_provider": "gemini", "max_audio_length_hours": 8.4, "max_audio_per_prompt": 1, @@ -11657,12 +13876,19 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_token": 1e-5, - "output_cost_per_token_above_200k_tokens": 1.5e-5, + "output_cost_per_token": 0.00001, + "output_cost_per_token_above_200k_tokens": 0.000015, "rpm": 10000, "source": "https://ai.google.dev/gemini-api/docs/pricing#gemini-2.5-pro-preview", - "supported_modalities": ["text", "image", "audio", "video"], - "supported_output_modalities": ["text"], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text" + ], "supports_audio_output": false, "supports_function_calling": true, "supports_pdf_input": true, @@ -11678,8 +13904,8 @@ "gemini/gemini-2.5-pro-preview-tts": { "cache_read_input_token_cost": 3.125e-7, "input_cost_per_audio_token": 7e-7, - "input_cost_per_token": 1.25e-6, - "input_cost_per_token_above_200k_tokens": 2.5e-6, + "input_cost_per_token": 0.00000125, + "input_cost_per_token_above_200k_tokens": 0.0000025, "litellm_provider": "gemini", "max_audio_length_hours": 8.4, "max_audio_per_prompt": 1, @@ -11691,12 +13917,16 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_token": 1e-5, - "output_cost_per_token_above_200k_tokens": 1.5e-5, + "output_cost_per_token": 0.00001, + "output_cost_per_token_above_200k_tokens": 0.000015, "rpm": 10000, "source": "https://ai.google.dev/gemini-api/docs/pricing#gemini-2.5-pro-preview", - "supported_modalities": ["text"], - "supported_output_modalities": ["audio"], + "supported_modalities": [ + "text" + ], + "supported_output_modalities": [ + "audio" + ], "supports_audio_output": false, "supports_function_calling": true, "supports_prompt_caching": true, @@ -11771,7 +14001,7 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 1.05e-6, + "output_cost_per_token": 0.00000105, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models", "supports_function_calling": true, "supports_tool_choice": true, @@ -11783,7 +14013,7 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 1.05e-6, + "output_cost_per_token": 0.00000105, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models", "supports_function_calling": true, "supports_tool_choice": true, @@ -11797,8 +14027,8 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 1.05e-6, - "output_cost_per_token_above_128k_tokens": 2.1e-6, + "output_cost_per_token": 0.00000105, + "output_cost_per_token_above_128k_tokens": 0.0000021, "rpd": 30000, "rpm": 360, "source": "https://ai.google.dev/gemini-api/docs/models/gemini", @@ -11814,8 +14044,8 @@ "max_output_tokens": 2048, "max_tokens": 2048, "mode": "chat", - "output_cost_per_token": 1.05e-6, - "output_cost_per_token_above_128k_tokens": 2.1e-6, + "output_cost_per_token": 0.00000105, + "output_cost_per_token_above_128k_tokens": 0.0000021, "rpd": 30000, "rpm": 360, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models", @@ -11848,7 +14078,7 @@ "supports_audio_output": false, "supports_function_calling": true, "supports_response_schema": true, - "supports_system_messages": true, + "supports_system_messages": false, "supports_tool_choice": true, "supports_vision": true }, @@ -11923,8 +14153,12 @@ "mode": "video_generation", "output_cost_per_second": 0.35, "source": "https://ai.google.dev/gemini-api/docs/video", - "supported_modalities": ["text"], - "supported_output_modalities": ["video"] + "supported_modalities": [ + "text" + ], + "supported_output_modalities": [ + "video" + ] }, "gemini/veo-3.0-fast-generate-preview": { "litellm_provider": "gemini", @@ -11933,8 +14167,12 @@ "mode": "video_generation", "output_cost_per_second": 0.4, "source": "https://ai.google.dev/gemini-api/docs/video", - "supported_modalities": ["text"], - "supported_output_modalities": ["video"] + "supported_modalities": [ + "text" + ], + "supported_output_modalities": [ + "video" + ] }, "gemini/veo-3.0-generate-preview": { "litellm_provider": "gemini", @@ -11943,8 +14181,12 @@ "mode": "video_generation", "output_cost_per_second": 0.75, "source": "https://ai.google.dev/gemini-api/docs/video", - "supported_modalities": ["text"], - "supported_output_modalities": ["video"] + "supported_modalities": [ + "text" + ], + "supported_output_modalities": [ + "video" + ] }, "gemini/veo-3.1-fast-generate-preview": { "litellm_provider": "gemini", @@ -11953,8 +14195,12 @@ "mode": "video_generation", "output_cost_per_second": 0.15, "source": "https://ai.google.dev/gemini-api/docs/video", - "supported_modalities": ["text"], - "supported_output_modalities": ["video"] + "supported_modalities": [ + "text" + ], + "supported_output_modalities": [ + "video" + ] }, "gemini/veo-3.1-generate-preview": { "litellm_provider": "gemini", @@ -11963,8 +14209,12 @@ "mode": "video_generation", "output_cost_per_second": 0.4, "source": "https://ai.google.dev/gemini-api/docs/video", - "supported_modalities": ["text"], - "supported_output_modalities": ["video"] + "supported_modalities": [ + "text" + ], + "supported_output_modalities": [ + "video" + ] }, "google_pse/search": { "input_cost_per_query": 0.005, @@ -11972,19 +14222,19 @@ "mode": "search" }, "global.anthropic.claude-sonnet-4-5-20250929-v1:0": { - "cache_creation_input_token_cost": 3.75e-6, + "cache_creation_input_token_cost": 0.00000375, "cache_read_input_token_cost": 3e-7, - "input_cost_per_token": 3e-6, - "input_cost_per_token_above_200k_tokens": 6e-6, - "output_cost_per_token_above_200k_tokens": 2.25e-5, - "cache_creation_input_token_cost_above_200k_tokens": 7.5e-6, + "input_cost_per_token": 0.000003, + "input_cost_per_token_above_200k_tokens": 0.000006, + "output_cost_per_token_above_200k_tokens": 0.0000225, + "cache_creation_input_token_cost_above_200k_tokens": 0.0000075, "cache_read_input_token_cost_above_200k_tokens": 6e-7, "litellm_provider": "bedrock_converse", "max_input_tokens": 200000, "max_output_tokens": 64000, "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -12002,19 +14252,19 @@ "tool_use_system_prompt_tokens": 346 }, "global.anthropic.claude-sonnet-4-20250514-v1:0": { - "cache_creation_input_token_cost": 3.75e-6, + "cache_creation_input_token_cost": 0.00000375, "cache_read_input_token_cost": 3e-7, - "input_cost_per_token": 3e-6, - "input_cost_per_token_above_200k_tokens": 6e-6, - "output_cost_per_token_above_200k_tokens": 2.25e-5, - "cache_creation_input_token_cost_above_200k_tokens": 7.5e-6, + "input_cost_per_token": 0.000003, + "input_cost_per_token_above_200k_tokens": 0.000006, + "output_cost_per_token_above_200k_tokens": 0.0000225, + "cache_creation_input_token_cost_above_200k_tokens": 0.0000075, "cache_read_input_token_cost_above_200k_tokens": 6e-7, "litellm_provider": "bedrock_converse", "max_input_tokens": 1000000, "max_output_tokens": 64000, "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -12032,17 +14282,18 @@ "tool_use_system_prompt_tokens": 159 }, "global.anthropic.claude-haiku-4-5-20251001-v1:0": { - "cache_creation_input_token_cost": 1.375e-6, + "cache_creation_input_token_cost": 0.000001375, "cache_read_input_token_cost": 1.1e-7, - "input_cost_per_token": 1.1e-6, + "input_cost_per_token": 0.0000011, "litellm_provider": "bedrock_converse", "max_input_tokens": 200000, - "max_output_tokens": 8192, - "max_tokens": 8192, + "max_output_tokens": 64000, + "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 5.5e-6, + "output_cost_per_token": 0.0000055, "source": "https://aws.amazon.com/about-aws/whats-new/2025/10/claude-4-5-haiku-anthropic-amazon-bedrock", "supports_assistant_prefill": true, + "supports_computer_use": true, "supports_function_calling": true, "supports_pdf_input": true, "supports_prompt_caching": true, @@ -12053,13 +14304,13 @@ "tool_use_system_prompt_tokens": 346 }, "gpt-3.5-turbo": { - "input_cost_per_token": 0.5e-6, + "input_cost_per_token": 5e-7, "litellm_provider": "openai", "max_input_tokens": 16385, "max_output_tokens": 4096, "max_tokens": 4097, "mode": "chat", - "output_cost_per_token": 1.5e-6, + "output_cost_per_token": 0.0000015, "supports_function_calling": true, "supports_prompt_caching": true, "supports_system_messages": true, @@ -12072,7 +14323,7 @@ "max_output_tokens": 4096, "max_tokens": 16385, "mode": "chat", - "output_cost_per_token": 1.5e-6, + "output_cost_per_token": 0.0000015, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -12080,25 +14331,25 @@ "supports_tool_choice": true }, "gpt-3.5-turbo-0301": { - "input_cost_per_token": 1.5e-6, + "input_cost_per_token": 0.0000015, "litellm_provider": "openai", "max_input_tokens": 4097, "max_output_tokens": 4096, "max_tokens": 4097, "mode": "chat", - "output_cost_per_token": 2e-6, + "output_cost_per_token": 0.000002, "supports_prompt_caching": true, "supports_system_messages": true, "supports_tool_choice": true }, "gpt-3.5-turbo-0613": { - "input_cost_per_token": 1.5e-6, + "input_cost_per_token": 0.0000015, "litellm_provider": "openai", "max_input_tokens": 4097, "max_output_tokens": 4096, "max_tokens": 4097, "mode": "chat", - "output_cost_per_token": 2e-6, + "output_cost_per_token": 0.000002, "supports_function_calling": true, "supports_prompt_caching": true, "supports_system_messages": true, @@ -12106,13 +14357,13 @@ }, "gpt-3.5-turbo-1106": { "deprecation_date": "2026-09-28", - "input_cost_per_token": 1e-6, + "input_cost_per_token": 0.000001, "litellm_provider": "openai", "max_input_tokens": 16385, "max_output_tokens": 4096, "max_tokens": 16385, "mode": "chat", - "output_cost_per_token": 2e-6, + "output_cost_per_token": 0.000002, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -12120,55 +14371,55 @@ "supports_tool_choice": true }, "gpt-3.5-turbo-16k": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "openai", "max_input_tokens": 16385, "max_output_tokens": 4096, "max_tokens": 16385, "mode": "chat", - "output_cost_per_token": 4e-6, + "output_cost_per_token": 0.000004, "supports_prompt_caching": true, "supports_system_messages": true, "supports_tool_choice": true }, "gpt-3.5-turbo-16k-0613": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "openai", "max_input_tokens": 16385, "max_output_tokens": 4096, "max_tokens": 16385, "mode": "chat", - "output_cost_per_token": 4e-6, + "output_cost_per_token": 0.000004, "supports_prompt_caching": true, "supports_system_messages": true, "supports_tool_choice": true }, "gpt-3.5-turbo-instruct": { - "input_cost_per_token": 1.5e-6, + "input_cost_per_token": 0.0000015, "litellm_provider": "text-completion-openai", "max_input_tokens": 8192, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "completion", - "output_cost_per_token": 2e-6 + "output_cost_per_token": 0.000002 }, "gpt-3.5-turbo-instruct-0914": { - "input_cost_per_token": 1.5e-6, + "input_cost_per_token": 0.0000015, "litellm_provider": "text-completion-openai", "max_input_tokens": 8192, "max_output_tokens": 4097, "max_tokens": 4097, "mode": "completion", - "output_cost_per_token": 2e-6 + "output_cost_per_token": 0.000002 }, "gpt-4": { - "input_cost_per_token": 3e-5, + "input_cost_per_token": 0.00003, "litellm_provider": "openai", "max_input_tokens": 8192, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 6e-5, + "output_cost_per_token": 0.00006, "supports_function_calling": true, "supports_prompt_caching": true, "supports_system_messages": true, @@ -12176,13 +14427,13 @@ }, "gpt-4-0125-preview": { "deprecation_date": "2026-03-26", - "input_cost_per_token": 1e-5, + "input_cost_per_token": 0.00001, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 3e-5, + "output_cost_per_token": 0.00003, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -12190,26 +14441,26 @@ "supports_tool_choice": true }, "gpt-4-0314": { - "input_cost_per_token": 3e-5, + "input_cost_per_token": 0.00003, "litellm_provider": "openai", "max_input_tokens": 8192, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 6e-5, + "output_cost_per_token": 0.00006, "supports_prompt_caching": true, "supports_system_messages": true, "supports_tool_choice": true }, "gpt-4-0613": { "deprecation_date": "2025-06-06", - "input_cost_per_token": 3e-5, + "input_cost_per_token": 0.00003, "litellm_provider": "openai", "max_input_tokens": 8192, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 6e-5, + "output_cost_per_token": 0.00006, "supports_function_calling": true, "supports_prompt_caching": true, "supports_system_messages": true, @@ -12217,13 +14468,13 @@ }, "gpt-4-1106-preview": { "deprecation_date": "2026-03-26", - "input_cost_per_token": 1e-5, + "input_cost_per_token": 0.00001, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 3e-5, + "output_cost_per_token": 0.00003, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -12232,13 +14483,13 @@ }, "gpt-4-1106-vision-preview": { "deprecation_date": "2024-12-06", - "input_cost_per_token": 1e-5, + "input_cost_per_token": 0.00001, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 3e-5, + "output_cost_per_token": 0.00003, "supports_pdf_input": true, "supports_prompt_caching": true, "supports_system_messages": true, @@ -12246,7 +14497,7 @@ "supports_vision": true }, "gpt-4-32k": { - "input_cost_per_token": 6e-5, + "input_cost_per_token": 0.00006, "litellm_provider": "openai", "max_input_tokens": 32768, "max_output_tokens": 4096, @@ -12258,7 +14509,7 @@ "supports_tool_choice": true }, "gpt-4-32k-0314": { - "input_cost_per_token": 6e-5, + "input_cost_per_token": 0.00006, "litellm_provider": "openai", "max_input_tokens": 32768, "max_output_tokens": 4096, @@ -12270,7 +14521,7 @@ "supports_tool_choice": true }, "gpt-4-32k-0613": { - "input_cost_per_token": 6e-5, + "input_cost_per_token": 0.00006, "litellm_provider": "openai", "max_input_tokens": 32768, "max_output_tokens": 4096, @@ -12282,13 +14533,13 @@ "supports_tool_choice": true }, "gpt-4-turbo": { - "input_cost_per_token": 1e-5, + "input_cost_per_token": 0.00001, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 3e-5, + "output_cost_per_token": 0.00003, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_pdf_input": true, @@ -12298,13 +14549,13 @@ "supports_vision": true }, "gpt-4-turbo-2024-04-09": { - "input_cost_per_token": 1e-5, + "input_cost_per_token": 0.00001, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 3e-5, + "output_cost_per_token": 0.00003, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_pdf_input": true, @@ -12314,13 +14565,13 @@ "supports_vision": true }, "gpt-4-turbo-preview": { - "input_cost_per_token": 1e-5, + "input_cost_per_token": 0.00001, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 3e-5, + "output_cost_per_token": 0.00003, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_pdf_input": true, @@ -12330,13 +14581,13 @@ }, "gpt-4-vision-preview": { "deprecation_date": "2024-12-06", - "input_cost_per_token": 1e-5, + "input_cost_per_token": 0.00001, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 3e-5, + "output_cost_per_token": 0.00003, "supports_pdf_input": true, "supports_prompt_caching": true, "supports_system_messages": true, @@ -12346,20 +14597,29 @@ "gpt-4.1": { "cache_read_input_token_cost": 5e-7, "cache_read_input_token_cost_priority": 8.75e-7, - "input_cost_per_token": 2e-6, - "input_cost_per_token_batches": 1e-6, - "input_cost_per_token_priority": 3.5e-6, + "input_cost_per_token": 0.000002, + "input_cost_per_token_batches": 0.000001, + "input_cost_per_token_priority": 0.0000035, "litellm_provider": "openai", "max_input_tokens": 1047576, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 8e-6, - "output_cost_per_token_batches": 4e-6, - "output_cost_per_token_priority": 1.4e-5, - "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "output_cost_per_token": 0.000008, + "output_cost_per_token_batches": 0.000004, + "output_cost_per_token_priority": 0.000014, + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -12373,18 +14633,27 @@ }, "gpt-4.1-2025-04-14": { "cache_read_input_token_cost": 5e-7, - "input_cost_per_token": 2e-6, - "input_cost_per_token_batches": 1e-6, + "input_cost_per_token": 0.000002, + "input_cost_per_token_batches": 0.000001, "litellm_provider": "openai", "max_input_tokens": 1047576, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 8e-6, - "output_cost_per_token_batches": 4e-6, - "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "output_cost_per_token": 0.000008, + "output_cost_per_token_batches": 0.000004, + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -12407,12 +14676,21 @@ "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 1.6e-6, + "output_cost_per_token": 0.0000016, "output_cost_per_token_batches": 8e-7, - "output_cost_per_token_priority": 2.8e-6, - "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "output_cost_per_token_priority": 0.0000028, + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -12433,11 +14711,20 @@ "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 1.6e-6, + "output_cost_per_token": 0.0000016, "output_cost_per_token_batches": 8e-7, - "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -12463,9 +14750,18 @@ "output_cost_per_token": 4e-7, "output_cost_per_token_batches": 2e-7, "output_cost_per_token_priority": 8e-7, - "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -12488,9 +14784,18 @@ "mode": "chat", "output_cost_per_token": 4e-7, "output_cost_per_token_batches": 2e-7, - "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -12503,16 +14808,16 @@ "supports_vision": true }, "gpt-4.5-preview": { - "cache_read_input_token_cost": 3.75e-5, - "input_cost_per_token": 7.5e-5, - "input_cost_per_token_batches": 3.75e-5, + "cache_read_input_token_cost": 0.0000375, + "input_cost_per_token": 0.000075, + "input_cost_per_token_batches": 0.0000375, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", "output_cost_per_token": 0.00015, - "output_cost_per_token_batches": 7.5e-5, + "output_cost_per_token_batches": 0.000075, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_pdf_input": true, @@ -12523,17 +14828,17 @@ "supports_vision": true }, "gpt-4.5-preview-2025-02-27": { - "cache_read_input_token_cost": 3.75e-5, + "cache_read_input_token_cost": 0.0000375, "deprecation_date": "2025-07-14", - "input_cost_per_token": 7.5e-5, - "input_cost_per_token_batches": 3.75e-5, + "input_cost_per_token": 0.000075, + "input_cost_per_token_batches": 0.0000375, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", "output_cost_per_token": 0.00015, - "output_cost_per_token_batches": 7.5e-5, + "output_cost_per_token_batches": 0.000075, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_pdf_input": true, @@ -12544,19 +14849,19 @@ "supports_vision": true }, "gpt-4o": { - "cache_read_input_token_cost": 1.25e-6, - "cache_read_input_token_cost_priority": 2.125e-6, - "input_cost_per_token": 2.5e-6, - "input_cost_per_token_batches": 1.25e-6, - "input_cost_per_token_priority": 4.25e-6, + "cache_read_input_token_cost": 0.00000125, + "cache_read_input_token_cost_priority": 0.000002125, + "input_cost_per_token": 0.0000025, + "input_cost_per_token_batches": 0.00000125, + "input_cost_per_token_priority": 0.00000425, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 1e-5, - "output_cost_per_token_batches": 5e-6, - "output_cost_per_token_priority": 1.7e-5, + "output_cost_per_token": 0.00001, + "output_cost_per_token_batches": 0.000005, + "output_cost_per_token_priority": 0.000017, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_pdf_input": true, @@ -12568,17 +14873,17 @@ "supports_vision": true }, "gpt-4o-2024-05-13": { - "input_cost_per_token": 5e-6, - "input_cost_per_token_batches": 2.5e-6, - "input_cost_per_token_priority": 8.75e-6, + "input_cost_per_token": 0.000005, + "input_cost_per_token_batches": 0.0000025, + "input_cost_per_token_priority": 0.00000875, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 1.5e-5, - "output_cost_per_token_batches": 7.5e-6, - "output_cost_per_token_priority": 2.625e-5, + "output_cost_per_token": 0.000015, + "output_cost_per_token_batches": 0.0000075, + "output_cost_per_token_priority": 0.00002625, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_pdf_input": true, @@ -12588,16 +14893,16 @@ "supports_vision": true }, "gpt-4o-2024-08-06": { - "cache_read_input_token_cost": 1.25e-6, - "input_cost_per_token": 2.5e-6, - "input_cost_per_token_batches": 1.25e-6, + "cache_read_input_token_cost": 0.00000125, + "input_cost_per_token": 0.0000025, + "input_cost_per_token_batches": 0.00000125, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 1e-5, - "output_cost_per_token_batches": 5e-6, + "output_cost_per_token": 0.00001, + "output_cost_per_token_batches": 0.000005, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_pdf_input": true, @@ -12609,16 +14914,16 @@ "supports_vision": true }, "gpt-4o-2024-11-20": { - "cache_read_input_token_cost": 1.25e-6, - "input_cost_per_token": 2.5e-6, - "input_cost_per_token_batches": 1.25e-6, + "cache_read_input_token_cost": 0.00000125, + "input_cost_per_token": 0.0000025, + "input_cost_per_token_batches": 0.00000125, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 1e-5, - "output_cost_per_token_batches": 5e-6, + "output_cost_per_token": 0.00001, + "output_cost_per_token_batches": 0.000005, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_pdf_input": true, @@ -12631,14 +14936,14 @@ }, "gpt-4o-audio-preview": { "input_cost_per_audio_token": 0.0001, - "input_cost_per_token": 2.5e-6, + "input_cost_per_token": 0.0000025, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", "output_cost_per_audio_token": 0.0002, - "output_cost_per_token": 1e-5, + "output_cost_per_token": 0.00001, "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -12648,14 +14953,14 @@ }, "gpt-4o-audio-preview-2024-10-01": { "input_cost_per_audio_token": 0.0001, - "input_cost_per_token": 2.5e-6, + "input_cost_per_token": 0.0000025, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", "output_cost_per_audio_token": 0.0002, - "output_cost_per_token": 1e-5, + "output_cost_per_token": 0.00001, "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -12664,15 +14969,15 @@ "supports_tool_choice": true }, "gpt-4o-audio-preview-2024-12-17": { - "input_cost_per_audio_token": 4e-5, - "input_cost_per_token": 2.5e-6, + "input_cost_per_audio_token": 0.00004, + "input_cost_per_token": 0.0000025, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_audio_token": 8e-5, - "output_cost_per_token": 1e-5, + "output_cost_per_audio_token": 0.00008, + "output_cost_per_token": 0.00001, "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -12681,15 +14986,15 @@ "supports_tool_choice": true }, "gpt-4o-audio-preview-2025-06-03": { - "input_cost_per_audio_token": 4e-5, - "input_cost_per_token": 2.5e-6, + "input_cost_per_audio_token": 0.00004, + "input_cost_per_token": 0.0000025, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_audio_token": 8e-5, - "output_cost_per_token": 1e-5, + "output_cost_per_audio_token": 0.00008, + "output_cost_per_token": 0.00001, "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -12710,7 +15015,7 @@ "mode": "chat", "output_cost_per_token": 6e-7, "output_cost_per_token_batches": 3e-7, - "output_cost_per_token_priority": 1e-6, + "output_cost_per_token_priority": 0.000001, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_pdf_input": true, @@ -12748,14 +15053,14 @@ "supports_vision": true }, "gpt-4o-mini-audio-preview": { - "input_cost_per_audio_token": 1e-5, + "input_cost_per_audio_token": 0.00001, "input_cost_per_token": 1.5e-7, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_audio_token": 2e-5, + "output_cost_per_audio_token": 0.00002, "output_cost_per_token": 6e-7, "supports_audio_input": true, "supports_audio_output": true, @@ -12765,14 +15070,14 @@ "supports_tool_choice": true }, "gpt-4o-mini-audio-preview-2024-12-17": { - "input_cost_per_audio_token": 1e-5, + "input_cost_per_audio_token": 0.00001, "input_cost_per_token": 1.5e-7, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_audio_token": 2e-5, + "output_cost_per_audio_token": 0.00002, "output_cost_per_token": 6e-7, "supports_audio_input": true, "supports_audio_output": true, @@ -12784,15 +15089,15 @@ "gpt-4o-mini-realtime-preview": { "cache_creation_input_audio_token_cost": 3e-7, "cache_read_input_token_cost": 3e-7, - "input_cost_per_audio_token": 1e-5, + "input_cost_per_audio_token": 0.00001, "input_cost_per_token": 6e-7, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_audio_token": 2e-5, - "output_cost_per_token": 2.4e-6, + "output_cost_per_audio_token": 0.00002, + "output_cost_per_token": 0.0000024, "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -12803,15 +15108,15 @@ "gpt-4o-mini-realtime-preview-2024-12-17": { "cache_creation_input_audio_token_cost": 3e-7, "cache_read_input_token_cost": 3e-7, - "input_cost_per_audio_token": 1e-5, + "input_cost_per_audio_token": 0.00001, "input_cost_per_token": 6e-7, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_audio_token": 2e-5, - "output_cost_per_token": 2.4e-6, + "output_cost_per_audio_token": 0.00002, + "output_cost_per_token": 0.0000024, "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -12866,37 +15171,46 @@ "supports_vision": true }, "gpt-4o-mini-transcribe": { - "input_cost_per_audio_token": 3e-6, - "input_cost_per_token": 1.25e-6, + "input_cost_per_audio_token": 0.000003, + "input_cost_per_token": 0.00000125, "litellm_provider": "openai", "max_input_tokens": 16000, "max_output_tokens": 2000, "mode": "audio_transcription", - "output_cost_per_token": 5e-6, - "supported_endpoints": ["/v1/audio/transcriptions"] + "output_cost_per_token": 0.000005, + "supported_endpoints": [ + "/v1/audio/transcriptions" + ] }, "gpt-4o-mini-tts": { - "input_cost_per_token": 2.5e-6, + "input_cost_per_token": 0.0000025, "litellm_provider": "openai", "mode": "audio_speech", - "output_cost_per_audio_token": 1.2e-5, + "output_cost_per_audio_token": 0.000012, "output_cost_per_second": 0.00025, - "output_cost_per_token": 1e-5, - "supported_endpoints": ["/v1/audio/speech"], - "supported_modalities": ["text", "audio"], - "supported_output_modalities": ["audio"] + "output_cost_per_token": 0.00001, + "supported_endpoints": [ + "/v1/audio/speech" + ], + "supported_modalities": [ + "text", + "audio" + ], + "supported_output_modalities": [ + "audio" + ] }, "gpt-4o-realtime-preview": { - "cache_read_input_token_cost": 2.5e-6, - "input_cost_per_audio_token": 4e-5, - "input_cost_per_token": 5e-6, + "cache_read_input_token_cost": 0.0000025, + "input_cost_per_audio_token": 0.00004, + "input_cost_per_token": 0.000005, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_audio_token": 8e-5, - "output_cost_per_token": 2e-5, + "output_cost_per_audio_token": 0.00008, + "output_cost_per_token": 0.00002, "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -12905,17 +15219,17 @@ "supports_tool_choice": true }, "gpt-4o-realtime-preview-2024-10-01": { - "cache_creation_input_audio_token_cost": 2e-5, - "cache_read_input_token_cost": 2.5e-6, + "cache_creation_input_audio_token_cost": 0.00002, + "cache_read_input_token_cost": 0.0000025, "input_cost_per_audio_token": 0.0001, - "input_cost_per_token": 5e-6, + "input_cost_per_token": 0.000005, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", "output_cost_per_audio_token": 0.0002, - "output_cost_per_token": 2e-5, + "output_cost_per_token": 0.00002, "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -12924,16 +15238,16 @@ "supports_tool_choice": true }, "gpt-4o-realtime-preview-2024-12-17": { - "cache_read_input_token_cost": 2.5e-6, - "input_cost_per_audio_token": 4e-5, - "input_cost_per_token": 5e-6, + "cache_read_input_token_cost": 0.0000025, + "input_cost_per_audio_token": 0.00004, + "input_cost_per_token": 0.000005, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_audio_token": 8e-5, - "output_cost_per_token": 2e-5, + "output_cost_per_audio_token": 0.00008, + "output_cost_per_token": 0.00002, "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -12942,16 +15256,16 @@ "supports_tool_choice": true }, "gpt-4o-realtime-preview-2025-06-03": { - "cache_read_input_token_cost": 2.5e-6, - "input_cost_per_audio_token": 4e-5, - "input_cost_per_token": 5e-6, + "cache_read_input_token_cost": 0.0000025, + "input_cost_per_audio_token": 0.00004, + "input_cost_per_token": 0.000005, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_audio_token": 8e-5, - "output_cost_per_token": 2e-5, + "output_cost_per_audio_token": 0.00008, + "output_cost_per_token": 0.00002, "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -12960,16 +15274,16 @@ "supports_tool_choice": true }, "gpt-4o-search-preview": { - "cache_read_input_token_cost": 1.25e-6, - "input_cost_per_token": 2.5e-6, - "input_cost_per_token_batches": 1.25e-6, + "cache_read_input_token_cost": 0.00000125, + "input_cost_per_token": 0.0000025, + "input_cost_per_token_batches": 0.00000125, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 1e-5, - "output_cost_per_token_batches": 5e-6, + "output_cost_per_token": 0.00001, + "output_cost_per_token_batches": 0.000005, "search_context_cost_per_query": { "search_context_size_high": 0.05, "search_context_size_low": 0.03, @@ -12986,16 +15300,16 @@ "supports_web_search": true }, "gpt-4o-search-preview-2025-03-11": { - "cache_read_input_token_cost": 1.25e-6, - "input_cost_per_token": 2.5e-6, - "input_cost_per_token_batches": 1.25e-6, + "cache_read_input_token_cost": 0.00000125, + "input_cost_per_token": 0.0000025, + "input_cost_per_token_batches": 0.00000125, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 1e-5, - "output_cost_per_token_batches": 5e-6, + "output_cost_per_token": 0.00001, + "output_cost_per_token_batches": 0.000005, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_pdf_input": true, @@ -13006,33 +15320,44 @@ "supports_vision": true }, "gpt-4o-transcribe": { - "input_cost_per_audio_token": 6e-6, - "input_cost_per_token": 2.5e-6, + "input_cost_per_audio_token": 0.000006, + "input_cost_per_token": 0.0000025, "litellm_provider": "openai", "max_input_tokens": 16000, "max_output_tokens": 2000, "mode": "audio_transcription", - "output_cost_per_token": 1e-5, - "supported_endpoints": ["/v1/audio/transcriptions"] + "output_cost_per_token": 0.00001, + "supported_endpoints": [ + "/v1/audio/transcriptions" + ] }, "gpt-5": { "cache_read_input_token_cost": 1.25e-7, "cache_read_input_token_cost_flex": 6.25e-8, "cache_read_input_token_cost_priority": 2.5e-7, - "input_cost_per_token": 1.25e-6, + "input_cost_per_token": 0.00000125, "input_cost_per_token_flex": 6.25e-7, - "input_cost_per_token_priority": 2.5e-6, + "input_cost_per_token_priority": 0.0000025, "litellm_provider": "openai", "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 1e-5, - "output_cost_per_token_flex": 5e-6, - "output_cost_per_token_priority": 2e-5, - "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "output_cost_per_token": 0.00001, + "output_cost_per_token_flex": 0.000005, + "output_cost_per_token_priority": 0.00002, + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -13048,18 +15373,27 @@ "gpt-5.1": { "cache_read_input_token_cost": 1.25e-7, "cache_read_input_token_cost_priority": 2.5e-7, - "input_cost_per_token": 1.25e-6, - "input_cost_per_token_priority": 2.5e-6, + "input_cost_per_token": 0.00000125, + "input_cost_per_token_priority": 0.0000025, "litellm_provider": "openai", "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 1e-5, - "output_cost_per_token_priority": 2e-5, - "supported_endpoints": ["/v1/chat/completions", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text", "image"], + "output_cost_per_token": 0.00001, + "output_cost_per_token_priority": 0.00002, + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text", + "image" + ], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -13075,18 +15409,27 @@ "gpt-5.1-2025-11-13": { "cache_read_input_token_cost": 1.25e-7, "cache_read_input_token_cost_priority": 2.5e-7, - "input_cost_per_token": 1.25e-6, - "input_cost_per_token_priority": 2.5e-6, + "input_cost_per_token": 0.00000125, + "input_cost_per_token_priority": 0.0000025, "litellm_provider": "openai", "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 1e-5, - "output_cost_per_token_priority": 2e-5, - "supported_endpoints": ["/v1/chat/completions", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text", "image"], + "output_cost_per_token": 0.00001, + "output_cost_per_token_priority": 0.00002, + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text", + "image" + ], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -13102,18 +15445,27 @@ "gpt-5.1-chat-latest": { "cache_read_input_token_cost": 1.25e-7, "cache_read_input_token_cost_priority": 2.5e-7, - "input_cost_per_token": 1.25e-6, - "input_cost_per_token_priority": 2.5e-6, + "input_cost_per_token": 0.00000125, + "input_cost_per_token_priority": 0.0000025, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 1e-5, - "output_cost_per_token_priority": 2e-5, - "supported_endpoints": ["/v1/chat/completions", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text", "image"], + "output_cost_per_token": 0.00001, + "output_cost_per_token_priority": 0.00002, + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text", + "image" + ], "supports_function_calling": false, "supports_native_streaming": true, "supports_parallel_function_calling": false, @@ -13126,18 +15478,26 @@ "supports_vision": true }, "gpt-5-pro": { - "input_cost_per_token": 1.5e-5, - "input_cost_per_token_batches": 7.5e-6, + "input_cost_per_token": 0.000015, + "input_cost_per_token_batches": 0.0000075, "litellm_provider": "openai", "max_input_tokens": 400000, "max_output_tokens": 272000, "max_tokens": 272000, "mode": "responses", - "output_cost_per_token": 1.2e-4, - "output_cost_per_token_batches": 6e-5, - "supported_endpoints": ["/v1/batch", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "output_cost_per_token": 0.00012, + "output_cost_per_token_batches": 0.00006, + "supported_endpoints": [ + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_native_streaming": false, "supports_parallel_function_calling": true, @@ -13151,18 +15511,26 @@ "supports_web_search": true }, "gpt-5-pro-2025-10-06": { - "input_cost_per_token": 1.5e-5, - "input_cost_per_token_batches": 7.5e-6, + "input_cost_per_token": 0.000015, + "input_cost_per_token_batches": 0.0000075, "litellm_provider": "openai", "max_input_tokens": 400000, "max_output_tokens": 272000, "max_tokens": 272000, "mode": "responses", - "output_cost_per_token": 1.2e-4, - "output_cost_per_token_batches": 6e-5, - "supported_endpoints": ["/v1/batch", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "output_cost_per_token": 0.00012, + "output_cost_per_token_batches": 0.00006, + "supported_endpoints": [ + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_native_streaming": false, "supports_parallel_function_calling": true, @@ -13179,20 +15547,29 @@ "cache_read_input_token_cost": 1.25e-7, "cache_read_input_token_cost_flex": 6.25e-8, "cache_read_input_token_cost_priority": 2.5e-7, - "input_cost_per_token": 1.25e-6, + "input_cost_per_token": 0.00000125, "input_cost_per_token_flex": 6.25e-7, - "input_cost_per_token_priority": 2.5e-6, + "input_cost_per_token_priority": 0.0000025, "litellm_provider": "openai", "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 1e-5, - "output_cost_per_token_flex": 5e-6, - "output_cost_per_token_priority": 2e-5, - "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "output_cost_per_token": 0.00001, + "output_cost_per_token_flex": 0.000005, + "output_cost_per_token_priority": 0.00002, + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -13207,16 +15584,25 @@ }, "gpt-5-chat": { "cache_read_input_token_cost": 1.25e-7, - "input_cost_per_token": 1.25e-6, + "input_cost_per_token": 0.00000125, "litellm_provider": "openai", "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 1e-5, - "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "output_cost_per_token": 0.00001, + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": false, "supports_native_streaming": true, "supports_parallel_function_calling": false, @@ -13230,16 +15616,25 @@ }, "gpt-5-chat-latest": { "cache_read_input_token_cost": 1.25e-7, - "input_cost_per_token": 1.25e-6, + "input_cost_per_token": 0.00000125, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 1e-5, - "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "output_cost_per_token": 0.00001, + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": false, "supports_native_streaming": true, "supports_parallel_function_calling": false, @@ -13253,16 +15648,23 @@ }, "gpt-5-codex": { "cache_read_input_token_cost": 1.25e-7, - "input_cost_per_token": 1.25e-6, + "input_cost_per_token": 0.00000125, "litellm_provider": "openai", "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "responses", - "output_cost_per_token": 1e-5, - "supported_endpoints": ["/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "output_cost_per_token": 0.00001, + "supported_endpoints": [ + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -13277,18 +15679,25 @@ "gpt-5.1-codex": { "cache_read_input_token_cost": 1.25e-7, "cache_read_input_token_cost_priority": 2.5e-7, - "input_cost_per_token": 1.25e-6, - "input_cost_per_token_priority": 2.5e-6, + "input_cost_per_token": 0.00000125, + "input_cost_per_token_priority": 0.0000025, "litellm_provider": "openai", "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "responses", - "output_cost_per_token": 1e-5, - "output_cost_per_token_priority": 2e-5, - "supported_endpoints": ["/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "output_cost_per_token": 0.00001, + "output_cost_per_token_priority": 0.00002, + "supported_endpoints": [ + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -13310,11 +15719,18 @@ "max_output_tokens": 128000, "max_tokens": 128000, "mode": "responses", - "output_cost_per_token": 2e-6, - "output_cost_per_token_priority": 3.6e-6, - "supported_endpoints": ["/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "output_cost_per_token": 0.000002, + "output_cost_per_token_priority": 0.0000036, + "supported_endpoints": [ + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -13338,12 +15754,21 @@ "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 2e-6, - "output_cost_per_token_flex": 1e-6, - "output_cost_per_token_priority": 3.6e-6, - "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "output_cost_per_token": 0.000002, + "output_cost_per_token_flex": 0.000001, + "output_cost_per_token_priority": 0.0000036, + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -13368,12 +15793,21 @@ "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 2e-6, - "output_cost_per_token_flex": 1e-6, - "output_cost_per_token_priority": 3.6e-6, - "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "output_cost_per_token": 0.000002, + "output_cost_per_token_flex": 0.000001, + "output_cost_per_token_priority": 0.0000036, + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -13391,7 +15825,7 @@ "cache_read_input_token_cost_flex": 2.5e-9, "input_cost_per_token": 5e-8, "input_cost_per_token_flex": 2.5e-8, - "input_cost_per_token_priority": 2.5e-6, + "input_cost_per_token_priority": 0.0000025, "litellm_provider": "openai", "max_input_tokens": 272000, "max_output_tokens": 128000, @@ -13399,9 +15833,18 @@ "mode": "chat", "output_cost_per_token": 4e-7, "output_cost_per_token_flex": 2e-7, - "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -13425,9 +15868,18 @@ "mode": "chat", "output_cost_per_token": 4e-7, "output_cost_per_token_flex": 2e-7, - "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -13443,35 +15895,49 @@ "input_cost_per_pixel": 4.0054321e-8, "litellm_provider": "openai", "mode": "image_generation", - "output_cost_per_pixel": 0.0, - "supported_endpoints": ["/v1/images/generations"] + "output_cost_per_pixel": 0, + "supported_endpoints": [ + "/v1/images/generations" + ] }, "gpt-image-1-mini": { "cache_read_input_image_token_cost": 2.5e-7, "cache_read_input_token_cost": 2e-7, - "input_cost_per_image_token": 2.5e-6, - "input_cost_per_token": 2e-6, + "input_cost_per_image_token": 0.0000025, + "input_cost_per_token": 0.000002, "litellm_provider": "openai", "mode": "chat", - "output_cost_per_image_token": 8e-6, - "supported_endpoints": ["/v1/images/generations", "/v1/images/edits"] + "output_cost_per_image_token": 0.000008, + "supported_endpoints": [ + "/v1/images/generations", + "/v1/images/edits" + ] }, "gpt-realtime": { "cache_creation_input_audio_token_cost": 4e-7, "cache_read_input_token_cost": 4e-7, - "input_cost_per_audio_token": 3.2e-5, - "input_cost_per_image": 5e-6, - "input_cost_per_token": 4e-6, + "input_cost_per_audio_token": 0.000032, + "input_cost_per_image": 0.000005, + "input_cost_per_token": 0.000004, "litellm_provider": "openai", "max_input_tokens": 32000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_audio_token": 6.4e-5, - "output_cost_per_token": 1.6e-5, - "supported_endpoints": ["/v1/realtime"], - "supported_modalities": ["text", "image", "audio"], - "supported_output_modalities": ["text", "audio"], + "output_cost_per_audio_token": 0.000064, + "output_cost_per_token": 0.000016, + "supported_endpoints": [ + "/v1/realtime" + ], + "supported_modalities": [ + "text", + "image", + "audio" + ], + "supported_output_modalities": [ + "text", + "audio" + ], "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -13482,18 +15948,27 @@ "gpt-realtime-mini": { "cache_creation_input_audio_token_cost": 3e-7, "cache_read_input_audio_token_cost": 3e-7, - "input_cost_per_audio_token": 1e-5, + "input_cost_per_audio_token": 0.00001, "input_cost_per_token": 6e-7, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_audio_token": 2e-5, - "output_cost_per_token": 2.4e-6, - "supported_endpoints": ["/v1/realtime"], - "supported_modalities": ["text", "image", "audio"], - "supported_output_modalities": ["text", "audio"], + "output_cost_per_audio_token": 0.00002, + "output_cost_per_token": 0.0000024, + "supported_endpoints": [ + "/v1/realtime" + ], + "supported_modalities": [ + "text", + "image", + "audio" + ], + "supported_output_modalities": [ + "text", + "audio" + ], "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -13504,19 +15979,28 @@ "gpt-realtime-2025-08-28": { "cache_creation_input_audio_token_cost": 4e-7, "cache_read_input_token_cost": 4e-7, - "input_cost_per_audio_token": 3.2e-5, - "input_cost_per_image": 5e-6, - "input_cost_per_token": 4e-6, + "input_cost_per_audio_token": 0.000032, + "input_cost_per_image": 0.000005, + "input_cost_per_token": 0.000004, "litellm_provider": "openai", "max_input_tokens": 32000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_audio_token": 6.4e-5, - "output_cost_per_token": 1.6e-5, - "supported_endpoints": ["/v1/realtime"], - "supported_modalities": ["text", "image", "audio"], - "supported_output_modalities": ["text", "audio"], + "output_cost_per_audio_token": 0.000064, + "output_cost_per_token": 0.000016, + "supported_endpoints": [ + "/v1/realtime" + ], + "supported_modalities": [ + "text", + "image", + "audio" + ], + "supported_output_modalities": [ + "text", + "audio" + ], "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -13528,18 +16012,26 @@ "litellm_provider": "gradient_ai", "max_tokens": 2048, "mode": "chat", - "supported_endpoints": ["/v1/chat/completions"], - "supported_modalities": ["text"], + "supported_endpoints": [ + "/v1/chat/completions" + ], + "supported_modalities": [ + "text" + ], "supports_tool_choice": false }, "gradient_ai/anthropic-claude-3-opus": { - "input_cost_per_token": 1.5e-5, + "input_cost_per_token": 0.000015, "litellm_provider": "gradient_ai", "max_tokens": 1024, "mode": "chat", - "output_cost_per_token": 7.5e-5, - "supported_endpoints": ["/v1/chat/completions"], - "supported_modalities": ["text"], + "output_cost_per_token": 0.000075, + "supported_endpoints": [ + "/v1/chat/completions" + ], + "supported_modalities": [ + "text" + ], "supports_tool_choice": false }, "gradient_ai/anthropic-claude-3.5-haiku": { @@ -13547,29 +16039,41 @@ "litellm_provider": "gradient_ai", "max_tokens": 1024, "mode": "chat", - "output_cost_per_token": 4e-6, - "supported_endpoints": ["/v1/chat/completions"], - "supported_modalities": ["text"], + "output_cost_per_token": 0.000004, + "supported_endpoints": [ + "/v1/chat/completions" + ], + "supported_modalities": [ + "text" + ], "supports_tool_choice": false }, "gradient_ai/anthropic-claude-3.5-sonnet": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "gradient_ai", "max_tokens": 1024, "mode": "chat", - "output_cost_per_token": 1.5e-5, - "supported_endpoints": ["/v1/chat/completions"], - "supported_modalities": ["text"], + "output_cost_per_token": 0.000015, + "supported_endpoints": [ + "/v1/chat/completions" + ], + "supported_modalities": [ + "text" + ], "supports_tool_choice": false }, "gradient_ai/anthropic-claude-3.7-sonnet": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "gradient_ai", "max_tokens": 1024, "mode": "chat", - "output_cost_per_token": 1.5e-5, - "supported_endpoints": ["/v1/chat/completions"], - "supported_modalities": ["text"], + "output_cost_per_token": 0.000015, + "supported_endpoints": [ + "/v1/chat/completions" + ], + "supported_modalities": [ + "text" + ], "supports_tool_choice": false }, "gradient_ai/deepseek-r1-distill-llama-70b": { @@ -13578,8 +16082,12 @@ "max_tokens": 8000, "mode": "chat", "output_cost_per_token": 9.9e-7, - "supported_endpoints": ["/v1/chat/completions"], - "supported_modalities": ["text"], + "supported_endpoints": [ + "/v1/chat/completions" + ], + "supported_modalities": [ + "text" + ], "supports_tool_choice": false }, "gradient_ai/llama3-8b-instruct": { @@ -13588,8 +16096,12 @@ "max_tokens": 512, "mode": "chat", "output_cost_per_token": 2e-7, - "supported_endpoints": ["/v1/chat/completions"], - "supported_modalities": ["text"], + "supported_endpoints": [ + "/v1/chat/completions" + ], + "supported_modalities": [ + "text" + ], "supports_tool_choice": false }, "gradient_ai/llama3.3-70b-instruct": { @@ -13598,8 +16110,12 @@ "max_tokens": 2048, "mode": "chat", "output_cost_per_token": 6.5e-7, - "supported_endpoints": ["/v1/chat/completions"], - "supported_modalities": ["text"], + "supported_endpoints": [ + "/v1/chat/completions" + ], + "supported_modalities": [ + "text" + ], "supports_tool_choice": false }, "gradient_ai/mistral-nemo-instruct-2407": { @@ -13608,44 +16124,64 @@ "max_tokens": 512, "mode": "chat", "output_cost_per_token": 3e-7, - "supported_endpoints": ["/v1/chat/completions"], - "supported_modalities": ["text"], + "supported_endpoints": [ + "/v1/chat/completions" + ], + "supported_modalities": [ + "text" + ], "supports_tool_choice": false }, "gradient_ai/openai-gpt-4o": { "litellm_provider": "gradient_ai", "max_tokens": 16384, "mode": "chat", - "supported_endpoints": ["/v1/chat/completions"], - "supported_modalities": ["text"], + "supported_endpoints": [ + "/v1/chat/completions" + ], + "supported_modalities": [ + "text" + ], "supports_tool_choice": false }, "gradient_ai/openai-gpt-4o-mini": { "litellm_provider": "gradient_ai", "max_tokens": 16384, "mode": "chat", - "supported_endpoints": ["/v1/chat/completions"], - "supported_modalities": ["text"], + "supported_endpoints": [ + "/v1/chat/completions" + ], + "supported_modalities": [ + "text" + ], "supports_tool_choice": false }, "gradient_ai/openai-o3": { - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "gradient_ai", "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 8e-6, - "supported_endpoints": ["/v1/chat/completions"], - "supported_modalities": ["text"], + "output_cost_per_token": 0.000008, + "supported_endpoints": [ + "/v1/chat/completions" + ], + "supported_modalities": [ + "text" + ], "supports_tool_choice": false }, "gradient_ai/openai-o3-mini": { - "input_cost_per_token": 1.1e-6, + "input_cost_per_token": 0.0000011, "litellm_provider": "gradient_ai", "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 4.4e-6, - "supported_endpoints": ["/v1/chat/completions"], - "supported_modalities": ["text"], + "output_cost_per_token": 0.0000044, + "supported_endpoints": [ + "/v1/chat/completions" + ], + "supported_modalities": [ + "text" + ], "supports_tool_choice": false }, "lemonade/Qwen3-Coder-30B-A3B-Instruct-GGUF": { @@ -13722,10 +16258,10 @@ "supports_tool_choice": true }, "groq/distil-whisper-large-v3-en": { - "input_cost_per_second": 5.56e-6, + "input_cost_per_second": 0.00000556, "litellm_provider": "groq", "mode": "audio_transcription", - "output_cost_per_second": 0.0 + "output_cost_per_second": 0 }, "groq/gemma-7b-it": { "deprecation_date": "2024-12-18", @@ -13986,21 +16522,21 @@ "supports_tool_choice": true }, "groq/moonshotai/kimi-k2-instruct": { - "input_cost_per_token": 1e-6, + "input_cost_per_token": 0.000001, "litellm_provider": "groq", "max_input_tokens": 131072, "max_output_tokens": 16384, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 3e-6, + "output_cost_per_token": 0.000003, "supports_function_calling": true, "supports_response_schema": true, "supports_tool_choice": true }, "groq/moonshotai/kimi-k2-instruct-0905": { - "input_cost_per_token": 1e-6, - "output_cost_per_token": 3e-6, - "cache_read_input_token_cost": 0.5e-6, + "input_cost_per_token": 0.000001, + "output_cost_per_token": 0.000003, + "cache_read_input_token_cost": 5e-7, "litellm_provider": "groq", "max_input_tokens": 262144, "max_output_tokens": 16384, @@ -14041,7 +16577,7 @@ "supports_web_search": true }, "groq/playai-tts": { - "input_cost_per_character": 5e-5, + "input_cost_per_character": 0.00005, "litellm_provider": "groq", "max_input_tokens": 10000, "max_output_tokens": 10000, @@ -14062,34 +16598,34 @@ "supports_tool_choice": true }, "groq/whisper-large-v3": { - "input_cost_per_second": 3.083e-5, + "input_cost_per_second": 0.00003083, "litellm_provider": "groq", "mode": "audio_transcription", - "output_cost_per_second": 0.0 + "output_cost_per_second": 0 }, "groq/whisper-large-v3-turbo": { - "input_cost_per_second": 1.111e-5, + "input_cost_per_second": 0.00001111, "litellm_provider": "groq", "mode": "audio_transcription", - "output_cost_per_second": 0.0 + "output_cost_per_second": 0 }, "hd/1024-x-1024/dall-e-3": { "input_cost_per_pixel": 7.629e-8, "litellm_provider": "openai", "mode": "image_generation", - "output_cost_per_pixel": 0.0 + "output_cost_per_pixel": 0 }, "hd/1024-x-1792/dall-e-3": { "input_cost_per_pixel": 6.539e-8, "litellm_provider": "openai", "mode": "image_generation", - "output_cost_per_pixel": 0.0 + "output_cost_per_pixel": 0 }, "hd/1792-x-1024/dall-e-3": { "input_cost_per_pixel": 6.539e-8, "litellm_provider": "openai", "mode": "image_generation", - "output_cost_per_pixel": 0.0 + "output_cost_per_pixel": 0 }, "heroku/claude-3-5-haiku": { "litellm_provider": "heroku", @@ -14127,22 +16663,28 @@ "input_cost_per_pixel": 1.59263611e-7, "litellm_provider": "openai", "mode": "image_generation", - "output_cost_per_pixel": 0.0, - "supported_endpoints": ["/v1/images/generations"] + "output_cost_per_pixel": 0, + "supported_endpoints": [ + "/v1/images/generations" + ] }, "high/1024-x-1536/gpt-image-1": { "input_cost_per_pixel": 1.58945719e-7, "litellm_provider": "openai", "mode": "image_generation", - "output_cost_per_pixel": 0.0, - "supported_endpoints": ["/v1/images/generations"] + "output_cost_per_pixel": 0, + "supported_endpoints": [ + "/v1/images/generations" + ] }, "high/1536-x-1024/gpt-image-1": { "input_cost_per_pixel": 1.58945719e-7, "litellm_provider": "openai", "mode": "image_generation", - "output_cost_per_pixel": 0.0, - "supported_endpoints": ["/v1/images/generations"] + "output_cost_per_pixel": 0, + "supported_endpoints": [ + "/v1/images/generations" + ] }, "hyperbolic/NousResearch/Hermes-3-Llama-3.1-70B": { "input_cost_per_token": 1.2e-7, @@ -14197,13 +16739,13 @@ "supports_tool_choice": true }, "hyperbolic/Qwen/Qwen3-235B-A22B": { - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "hyperbolic", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 2e-6, + "output_cost_per_token": 0.000002, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_system_messages": true, @@ -14340,44 +16882,44 @@ "supports_tool_choice": true }, "hyperbolic/moonshotai/Kimi-K2-Instruct": { - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "hyperbolic", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 2e-6, + "output_cost_per_token": 0.000002, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_system_messages": true, "supports_tool_choice": true }, "j2-light": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "ai21", "max_input_tokens": 8192, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "completion", - "output_cost_per_token": 3e-6 + "output_cost_per_token": 0.000003 }, "j2-mid": { - "input_cost_per_token": 1e-5, + "input_cost_per_token": 0.00001, "litellm_provider": "ai21", "max_input_tokens": 8192, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "completion", - "output_cost_per_token": 1e-5 + "output_cost_per_token": 0.00001 }, "j2-ultra": { - "input_cost_per_token": 1.5e-5, + "input_cost_per_token": 0.000015, "litellm_provider": "ai21", "max_input_tokens": 8192, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "completion", - "output_cost_per_token": 1.5e-5 + "output_cost_per_token": 0.000015 }, "jamba-1.5": { "input_cost_per_token": 2e-7, @@ -14390,23 +16932,23 @@ "supports_tool_choice": true }, "jamba-1.5-large": { - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "ai21", "max_input_tokens": 256000, "max_output_tokens": 256000, "max_tokens": 256000, "mode": "chat", - "output_cost_per_token": 8e-6, + "output_cost_per_token": 0.000008, "supports_tool_choice": true }, "jamba-1.5-large@001": { - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "ai21", "max_input_tokens": 256000, "max_output_tokens": 256000, "max_tokens": 256000, "mode": "chat", - "output_cost_per_token": 8e-6, + "output_cost_per_token": 0.000008, "supports_tool_choice": true }, "jamba-1.5-mini": { @@ -14430,23 +16972,23 @@ "supports_tool_choice": true }, "jamba-large-1.6": { - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "ai21", "max_input_tokens": 256000, "max_output_tokens": 256000, "max_tokens": 256000, "mode": "chat", - "output_cost_per_token": 8e-6, + "output_cost_per_token": 0.000008, "supports_tool_choice": true }, "jamba-large-1.7": { - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "ai21", "max_input_tokens": 256000, "max_output_tokens": 256000, "max_tokens": 256000, "mode": "chat", - "output_cost_per_token": 8e-6, + "output_cost_per_token": 0.000008, "supports_tool_choice": true }, "jamba-mini-1.6": { @@ -14480,19 +17022,19 @@ "output_cost_per_token": 1.8e-8 }, "jp.anthropic.claude-sonnet-4-5-20250929-v1:0": { - "cache_creation_input_token_cost": 4.125e-6, + "cache_creation_input_token_cost": 0.000004125, "cache_read_input_token_cost": 3.3e-7, - "input_cost_per_token": 3.3e-6, - "input_cost_per_token_above_200k_tokens": 6.6e-6, - "output_cost_per_token_above_200k_tokens": 2.475e-5, - "cache_creation_input_token_cost_above_200k_tokens": 8.25e-6, + "input_cost_per_token": 0.0000033, + "input_cost_per_token_above_200k_tokens": 0.0000066, + "output_cost_per_token_above_200k_tokens": 0.00002475, + "cache_creation_input_token_cost_above_200k_tokens": 0.00000825, "cache_read_input_token_cost_above_200k_tokens": 6.6e-7, "litellm_provider": "bedrock_converse", "max_input_tokens": 200000, "max_output_tokens": 64000, "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 1.65e-5, + "output_cost_per_token": 0.0000165, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -14510,23 +17052,26 @@ "tool_use_system_prompt_tokens": 346 }, "jp.anthropic.claude-haiku-4-5-20251001-v1:0": { - "cache_creation_input_token_cost": 1.375e-6, + "cache_creation_input_token_cost": 0.000001375, "cache_read_input_token_cost": 1.1e-7, - "input_cost_per_token": 1.1e-6, - "litellm_provider": "bedrock", + "input_cost_per_token": 0.0000011, + "litellm_provider": "bedrock_converse", "max_input_tokens": 200000, - "max_output_tokens": 8192, - "max_tokens": 8192, + "max_output_tokens": 64000, + "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 5.5e-6, + "output_cost_per_token": 0.0000055, "source": "https://aws.amazon.com/about-aws/whats-new/2025/10/claude-4-5-haiku-anthropic-amazon-bedrock", "supports_assistant_prefill": true, + "supports_computer_use": true, "supports_function_calling": true, "supports_pdf_input": true, "supports_prompt_caching": true, "supports_reasoning": true, "supports_response_schema": true, - "supports_tool_choice": true + "supports_tool_choice": true, + "supports_vision": true, + "tool_use_system_prompt_tokens": 346 }, "lambda_ai/deepseek-llama3.3-70b": { "input_cost_per_token": 2e-7, @@ -14797,50 +17342,56 @@ "input_cost_per_pixel": 1.0490417e-8, "litellm_provider": "openai", "mode": "image_generation", - "output_cost_per_pixel": 0.0, - "supported_endpoints": ["/v1/images/generations"] + "output_cost_per_pixel": 0, + "supported_endpoints": [ + "/v1/images/generations" + ] }, "low/1024-x-1536/gpt-image-1": { "input_cost_per_pixel": 1.0172526e-8, "litellm_provider": "openai", "mode": "image_generation", - "output_cost_per_pixel": 0.0, - "supported_endpoints": ["/v1/images/generations"] + "output_cost_per_pixel": 0, + "supported_endpoints": [ + "/v1/images/generations" + ] }, "low/1536-x-1024/gpt-image-1": { "input_cost_per_pixel": 1.0172526e-8, "litellm_provider": "openai", "mode": "image_generation", - "output_cost_per_pixel": 0.0, - "supported_endpoints": ["/v1/images/generations"] + "output_cost_per_pixel": 0, + "supported_endpoints": [ + "/v1/images/generations" + ] }, "luminous-base": { - "input_cost_per_token": 3e-5, + "input_cost_per_token": 0.00003, "litellm_provider": "aleph_alpha", "max_tokens": 2048, "mode": "completion", - "output_cost_per_token": 3.3e-5 + "output_cost_per_token": 0.000033 }, "luminous-base-control": { - "input_cost_per_token": 3.75e-5, + "input_cost_per_token": 0.0000375, "litellm_provider": "aleph_alpha", "max_tokens": 2048, "mode": "chat", - "output_cost_per_token": 4.125e-5 + "output_cost_per_token": 0.00004125 }, "luminous-extended": { - "input_cost_per_token": 4.5e-5, + "input_cost_per_token": 0.000045, "litellm_provider": "aleph_alpha", "max_tokens": 2048, "mode": "completion", - "output_cost_per_token": 4.95e-5 + "output_cost_per_token": 0.0000495 }, "luminous-extended-control": { - "input_cost_per_token": 5.625e-5, + "input_cost_per_token": 0.00005625, "litellm_provider": "aleph_alpha", "max_tokens": 2048, "mode": "chat", - "output_cost_per_token": 6.1875e-5 + "output_cost_per_token": 0.000061875 }, "luminous-supreme": { "input_cost_per_token": 0.000175, @@ -14874,67 +17425,85 @@ "input_cost_per_pixel": 4.0054321e-8, "litellm_provider": "openai", "mode": "image_generation", - "output_cost_per_pixel": 0.0, - "supported_endpoints": ["/v1/images/generations"] + "output_cost_per_pixel": 0, + "supported_endpoints": [ + "/v1/images/generations" + ] }, "medium/1024-x-1536/gpt-image-1": { "input_cost_per_pixel": 4.0054321e-8, "litellm_provider": "openai", "mode": "image_generation", - "output_cost_per_pixel": 0.0, - "supported_endpoints": ["/v1/images/generations"] + "output_cost_per_pixel": 0, + "supported_endpoints": [ + "/v1/images/generations" + ] }, "medium/1536-x-1024/gpt-image-1": { "input_cost_per_pixel": 4.0054321e-8, "litellm_provider": "openai", "mode": "image_generation", - "output_cost_per_pixel": 0.0, - "supported_endpoints": ["/v1/images/generations"] + "output_cost_per_pixel": 0, + "supported_endpoints": [ + "/v1/images/generations" + ] }, "low/1024-x-1024/gpt-image-1-mini": { "input_cost_per_image": 0.005, "litellm_provider": "openai", "mode": "image_generation", - "supported_endpoints": ["/v1/images/generations"] + "supported_endpoints": [ + "/v1/images/generations" + ] }, "low/1024-x-1536/gpt-image-1-mini": { "input_cost_per_image": 0.006, "litellm_provider": "openai", "mode": "image_generation", - "supported_endpoints": ["/v1/images/generations"] + "supported_endpoints": [ + "/v1/images/generations" + ] }, "low/1536-x-1024/gpt-image-1-mini": { "input_cost_per_image": 0.006, "litellm_provider": "openai", "mode": "image_generation", - "supported_endpoints": ["/v1/images/generations"] + "supported_endpoints": [ + "/v1/images/generations" + ] }, "medium/1024-x-1024/gpt-image-1-mini": { "input_cost_per_image": 0.011, "litellm_provider": "openai", "mode": "image_generation", - "supported_endpoints": ["/v1/images/generations"] + "supported_endpoints": [ + "/v1/images/generations" + ] }, "medium/1024-x-1536/gpt-image-1-mini": { "input_cost_per_image": 0.015, "litellm_provider": "openai", "mode": "image_generation", - "supported_endpoints": ["/v1/images/generations"] + "supported_endpoints": [ + "/v1/images/generations" + ] }, "medium/1536-x-1024/gpt-image-1-mini": { "input_cost_per_image": 0.015, "litellm_provider": "openai", "mode": "image_generation", - "supported_endpoints": ["/v1/images/generations"] + "supported_endpoints": [ + "/v1/images/generations" + ] }, "medlm-large": { - "input_cost_per_character": 5e-6, + "input_cost_per_character": 0.000005, "litellm_provider": "vertex_ai-language-models", "max_input_tokens": 8192, "max_output_tokens": 1024, "max_tokens": 1024, "mode": "chat", - "output_cost_per_character": 1.5e-5, + "output_cost_per_character": 0.000015, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models", "supports_tool_choice": true }, @@ -14945,7 +17514,7 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_character": 1e-6, + "output_cost_per_character": 0.000001, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models", "supports_tool_choice": true }, @@ -14956,25 +17525,25 @@ "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 1e-6 + "output_cost_per_token": 0.000001 }, "meta.llama2-70b-chat-v1": { - "input_cost_per_token": 1.95e-6, + "input_cost_per_token": 0.00000195, "litellm_provider": "bedrock", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 2.56e-6 + "output_cost_per_token": 0.00000256 }, "meta.llama3-1-405b-instruct-v1:0": { - "input_cost_per_token": 5.32e-6, + "input_cost_per_token": 0.00000532, "litellm_provider": "bedrock", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 1.6e-5, + "output_cost_per_token": 0.000016, "supports_function_calling": true, "supports_tool_choice": false }, @@ -15035,13 +17604,13 @@ "supports_tool_choice": false }, "meta.llama3-2-90b-instruct-v1:0": { - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "bedrock", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 2e-6, + "output_cost_per_token": 0.000002, "supports_function_calling": true, "supports_tool_choice": false, "supports_vision": true @@ -15058,13 +17627,13 @@ "supports_tool_choice": false }, "meta.llama3-70b-instruct-v1:0": { - "input_cost_per_token": 2.65e-6, + "input_cost_per_token": 0.00000265, "litellm_provider": "bedrock", "max_input_tokens": 8192, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 3.5e-6 + "output_cost_per_token": 0.0000035 }, "meta.llama3-8b-instruct-v1:0": { "input_cost_per_token": 3e-7, @@ -15085,8 +17654,14 @@ "mode": "chat", "output_cost_per_token": 9.7e-7, "output_cost_per_token_batches": 4.85e-7, - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text", "code"], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text", + "code" + ], "supports_function_calling": true, "supports_tool_choice": false }, @@ -15100,8 +17675,14 @@ "mode": "chat", "output_cost_per_token": 6.6e-7, "output_cost_per_token_batches": 3.3e-7, - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text", "code"], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text", + "code" + ], "supports_function_calling": true, "supports_tool_choice": false }, @@ -15112,8 +17693,12 @@ "max_tokens": 128000, "mode": "chat", "source": "https://llama.developer.meta.com/docs/models", - "supported_modalities": ["text"], - "supported_output_modalities": ["text"], + "supported_modalities": [ + "text" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_tool_choice": true }, @@ -15124,8 +17709,12 @@ "max_tokens": 128000, "mode": "chat", "source": "https://llama.developer.meta.com/docs/models", - "supported_modalities": ["text"], - "supported_output_modalities": ["text"], + "supported_modalities": [ + "text" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_tool_choice": true }, @@ -15136,8 +17725,13 @@ "max_tokens": 128000, "mode": "chat", "source": "https://llama.developer.meta.com/docs/models", - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_tool_choice": true }, @@ -15148,8 +17742,13 @@ "max_tokens": 128000, "mode": "chat", "source": "https://llama.developer.meta.com/docs/models", - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_tool_choice": true }, @@ -15164,34 +17763,34 @@ "supports_tool_choice": true }, "mistral.mistral-large-2402-v1:0": { - "input_cost_per_token": 8e-6, + "input_cost_per_token": 0.000008, "litellm_provider": "bedrock", "max_input_tokens": 32000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 2.4e-5, + "output_cost_per_token": 0.000024, "supports_function_calling": true }, "mistral.mistral-large-2407-v1:0": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "bedrock", "max_input_tokens": 128000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 9e-6, + "output_cost_per_token": 0.000009, "supports_function_calling": true, "supports_tool_choice": true }, "mistral.mistral-small-2402-v1:0": { - "input_cost_per_token": 1e-6, + "input_cost_per_token": 0.000001, "litellm_provider": "bedrock", "max_input_tokens": 32000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 3e-6, + "output_cost_per_token": 0.000003, "supports_function_calling": true }, "mistral.mixtral-8x7b-instruct-v0:1": { @@ -15205,25 +17804,25 @@ "supports_tool_choice": true }, "mistral/codestral-2405": { - "input_cost_per_token": 1e-6, + "input_cost_per_token": 0.000001, "litellm_provider": "mistral", "max_input_tokens": 32000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 3e-6, + "output_cost_per_token": 0.000003, "supports_assistant_prefill": true, "supports_response_schema": true, "supports_tool_choice": true }, "mistral/codestral-latest": { - "input_cost_per_token": 1e-6, + "input_cost_per_token": 0.000001, "litellm_provider": "mistral", "max_input_tokens": 32000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 3e-6, + "output_cost_per_token": 0.000003, "supports_assistant_prefill": true, "supports_response_schema": true, "supports_tool_choice": true @@ -15247,7 +17846,7 @@ "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 2e-6, + "output_cost_per_token": 0.000002, "source": "https://mistral.ai/news/devstral", "supports_assistant_prefill": true, "supports_function_calling": true, @@ -15283,13 +17882,13 @@ "supports_tool_choice": true }, "mistral/magistral-medium-2506": { - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "mistral", "max_input_tokens": 40000, "max_output_tokens": 40000, "max_tokens": 40000, "mode": "chat", - "output_cost_per_token": 5e-6, + "output_cost_per_token": 0.000005, "source": "https://mistral.ai/news/magistral", "supports_assistant_prefill": true, "supports_function_calling": true, @@ -15298,13 +17897,13 @@ "supports_tool_choice": true }, "mistral/magistral-medium-2509": { - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "mistral", "max_input_tokens": 40000, "max_output_tokens": 40000, "max_tokens": 40000, "mode": "chat", - "output_cost_per_token": 5e-6, + "output_cost_per_token": 0.000005, "source": "https://mistral.ai/news/magistral", "supports_assistant_prefill": true, "supports_function_calling": true, @@ -15314,28 +17913,32 @@ }, "mistral/mistral-ocr-latest": { "litellm_provider": "mistral", - "ocr_cost_per_page": 1e-3, - "annotation_cost_per_page": 3e-3, + "ocr_cost_per_page": 0.001, + "annotation_cost_per_page": 0.003, "mode": "ocr", - "supported_endpoints": ["/v1/ocr"], + "supported_endpoints": [ + "/v1/ocr" + ], "source": "https://mistral.ai/pricing#api-pricing" }, "mistral/mistral-ocr-2505-completion": { "litellm_provider": "mistral", - "ocr_cost_per_page": 1e-3, - "annotation_cost_per_page": 3e-3, + "ocr_cost_per_page": 0.001, + "annotation_cost_per_page": 0.003, "mode": "ocr", - "supported_endpoints": ["/v1/ocr"], + "supported_endpoints": [ + "/v1/ocr" + ], "source": "https://mistral.ai/pricing#api-pricing" }, "mistral/magistral-medium-latest": { - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "mistral", "max_input_tokens": 40000, "max_output_tokens": 40000, "max_tokens": 40000, "mode": "chat", - "output_cost_per_token": 5e-6, + "output_cost_per_token": 0.000005, "source": "https://mistral.ai/news/magistral", "supports_assistant_prefill": true, "supports_function_calling": true, @@ -15350,7 +17953,7 @@ "max_output_tokens": 40000, "max_tokens": 40000, "mode": "chat", - "output_cost_per_token": 1.5e-6, + "output_cost_per_token": 0.0000015, "source": "https://mistral.ai/pricing#api-pricing", "supports_assistant_prefill": true, "supports_function_calling": true, @@ -15365,7 +17968,7 @@ "max_output_tokens": 40000, "max_tokens": 40000, "mode": "chat", - "output_cost_per_token": 1.5e-6, + "output_cost_per_token": 0.0000015, "source": "https://mistral.ai/pricing#api-pricing", "supports_assistant_prefill": true, "supports_function_calling": true, @@ -15381,91 +17984,91 @@ "mode": "embedding" }, "mistral/codestral-embed": { - "input_cost_per_token": 0.15e-6, + "input_cost_per_token": 1.5e-7, "litellm_provider": "mistral", "max_input_tokens": 8192, "max_tokens": 8192, "mode": "embedding" }, "mistral/codestral-embed-2505": { - "input_cost_per_token": 0.15e-6, + "input_cost_per_token": 1.5e-7, "litellm_provider": "mistral", "max_input_tokens": 8192, "max_tokens": 8192, "mode": "embedding" }, "mistral/mistral-large-2402": { - "input_cost_per_token": 4e-6, + "input_cost_per_token": 0.000004, "litellm_provider": "mistral", "max_input_tokens": 32000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 1.2e-5, + "output_cost_per_token": 0.000012, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_response_schema": true, "supports_tool_choice": true }, "mistral/mistral-large-2407": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "mistral", "max_input_tokens": 128000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 9e-6, + "output_cost_per_token": 0.000009, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_response_schema": true, "supports_tool_choice": true }, "mistral/mistral-large-2411": { - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "mistral", "max_input_tokens": 128000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 6e-6, + "output_cost_per_token": 0.000006, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_response_schema": true, "supports_tool_choice": true }, "mistral/mistral-large-latest": { - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "mistral", "max_input_tokens": 128000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 6e-6, + "output_cost_per_token": 0.000006, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_response_schema": true, "supports_tool_choice": true }, "mistral/mistral-medium": { - "input_cost_per_token": 2.7e-6, + "input_cost_per_token": 0.0000027, "litellm_provider": "mistral", "max_input_tokens": 32000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 8.1e-6, + "output_cost_per_token": 0.0000081, "supports_assistant_prefill": true, "supports_response_schema": true, "supports_tool_choice": true }, "mistral/mistral-medium-2312": { - "input_cost_per_token": 2.7e-6, + "input_cost_per_token": 0.0000027, "litellm_provider": "mistral", "max_input_tokens": 32000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 8.1e-6, + "output_cost_per_token": 0.0000081, "supports_assistant_prefill": true, "supports_response_schema": true, "supports_tool_choice": true @@ -15477,7 +18080,7 @@ "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 2e-6, + "output_cost_per_token": 0.000002, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_response_schema": true, @@ -15490,7 +18093,7 @@ "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 2e-6, + "output_cost_per_token": 0.000002, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_response_schema": true, @@ -15585,13 +18188,13 @@ "supports_tool_choice": true }, "mistral/open-mixtral-8x22b": { - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "mistral", "max_input_tokens": 65336, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 6e-6, + "output_cost_per_token": 0.000006, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_response_schema": true, @@ -15625,13 +18228,13 @@ "supports_vision": true }, "mistral/pixtral-large-2411": { - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "mistral", "max_input_tokens": 128000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 6e-6, + "output_cost_per_token": 0.000006, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_response_schema": true, @@ -15639,13 +18242,13 @@ "supports_vision": true }, "mistral/pixtral-large-latest": { - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "mistral", "max_input_tokens": 128000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 6e-6, + "output_cost_per_token": 0.000006, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_response_schema": true, @@ -15660,7 +18263,7 @@ "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 2.5e-6, + "output_cost_per_token": 0.0000025, "source": "https://platform.moonshot.ai/docs/pricing/chat#generation-model-kimi-k2", "supports_function_calling": true, "supports_tool_choice": true, @@ -15668,13 +18271,13 @@ }, "moonshot/kimi-latest": { "cache_read_input_token_cost": 1.5e-7, - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "moonshot", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 5e-6, + "output_cost_per_token": 0.000005, "source": "https://platform.moonshot.ai/docs/pricing", "supports_function_calling": true, "supports_tool_choice": true, @@ -15682,13 +18285,13 @@ }, "moonshot/kimi-latest-128k": { "cache_read_input_token_cost": 1.5e-7, - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "moonshot", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 5e-6, + "output_cost_per_token": 0.000005, "source": "https://platform.moonshot.ai/docs/pricing", "supports_function_calling": true, "supports_tool_choice": true, @@ -15696,13 +18299,13 @@ }, "moonshot/kimi-latest-32k": { "cache_read_input_token_cost": 1.5e-7, - "input_cost_per_token": 1e-6, + "input_cost_per_token": 0.000001, "litellm_provider": "moonshot", "max_input_tokens": 32768, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 3e-6, + "output_cost_per_token": 0.000003, "source": "https://platform.moonshot.ai/docs/pricing", "supports_function_calling": true, "supports_tool_choice": true, @@ -15716,20 +18319,20 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 2e-6, + "output_cost_per_token": 0.000002, "source": "https://platform.moonshot.ai/docs/pricing", "supports_function_calling": true, "supports_tool_choice": true, "supports_vision": true }, "moonshot/kimi-thinking-preview": { - "input_cost_per_token": 3e-5, + "input_cost_per_token": 0.00003, "litellm_provider": "moonshot", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 3e-5, + "output_cost_per_token": 0.00003, "source": "https://platform.moonshot.ai/docs/pricing", "supports_vision": true }, @@ -15741,81 +18344,81 @@ "max_output_tokens": 262144, "max_tokens": 262144, "mode": "chat", - "output_cost_per_token": 2.5e-6, + "output_cost_per_token": 0.0000025, "source": "https://platform.moonshot.ai/docs/pricing/chat#generation-model-kimi-k2", "supports_function_calling": true, "supports_tool_choice": true, "supports_web_search": true }, "moonshot/moonshot-v1-128k": { - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "moonshot", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 5e-6, + "output_cost_per_token": 0.000005, "source": "https://platform.moonshot.ai/docs/pricing", "supports_function_calling": true, "supports_tool_choice": true }, "moonshot/moonshot-v1-128k-0430": { - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "moonshot", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 5e-6, + "output_cost_per_token": 0.000005, "source": "https://platform.moonshot.ai/docs/pricing", "supports_function_calling": true, "supports_tool_choice": true }, "moonshot/moonshot-v1-128k-vision-preview": { - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "moonshot", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 5e-6, + "output_cost_per_token": 0.000005, "source": "https://platform.moonshot.ai/docs/pricing", "supports_function_calling": true, "supports_tool_choice": true, "supports_vision": true }, "moonshot/moonshot-v1-32k": { - "input_cost_per_token": 1e-6, + "input_cost_per_token": 0.000001, "litellm_provider": "moonshot", "max_input_tokens": 32768, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 3e-6, + "output_cost_per_token": 0.000003, "source": "https://platform.moonshot.ai/docs/pricing", "supports_function_calling": true, "supports_tool_choice": true }, "moonshot/moonshot-v1-32k-0430": { - "input_cost_per_token": 1e-6, + "input_cost_per_token": 0.000001, "litellm_provider": "moonshot", "max_input_tokens": 32768, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 3e-6, + "output_cost_per_token": 0.000003, "source": "https://platform.moonshot.ai/docs/pricing", "supports_function_calling": true, "supports_tool_choice": true }, "moonshot/moonshot-v1-32k-vision-preview": { - "input_cost_per_token": 1e-6, + "input_cost_per_token": 0.000001, "litellm_provider": "moonshot", "max_input_tokens": 32768, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 3e-6, + "output_cost_per_token": 0.000003, "source": "https://platform.moonshot.ai/docs/pricing", "supports_function_calling": true, "supports_tool_choice": true, @@ -15828,7 +18431,7 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 2e-6, + "output_cost_per_token": 0.000002, "source": "https://platform.moonshot.ai/docs/pricing", "supports_function_calling": true, "supports_tool_choice": true @@ -15840,7 +18443,7 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 2e-6, + "output_cost_per_token": 0.000002, "source": "https://platform.moonshot.ai/docs/pricing", "supports_function_calling": true, "supports_tool_choice": true @@ -15852,20 +18455,20 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 2e-6, + "output_cost_per_token": 0.000002, "source": "https://platform.moonshot.ai/docs/pricing", "supports_function_calling": true, "supports_tool_choice": true, "supports_vision": true }, "moonshot/moonshot-v1-auto": { - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "moonshot", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 5e-6, + "output_cost_per_token": 0.000005, "source": "https://platform.moonshot.ai/docs/pricing", "supports_function_calling": true, "supports_tool_choice": true @@ -15877,7 +18480,7 @@ "max_output_tokens": 16000, "max_tokens": 16000, "mode": "chat", - "output_cost_per_token": 1.2e-6, + "output_cost_per_token": 0.0000012, "supports_function_calling": false, "supports_parallel_function_calling": false, "supports_system_messages": true, @@ -15891,7 +18494,7 @@ "max_output_tokens": 16000, "max_tokens": 16000, "mode": "chat", - "output_cost_per_token": 1.9e-6, + "output_cost_per_token": 0.0000019, "supports_function_calling": false, "supports_parallel_function_calling": false, "supports_system_messages": true, @@ -15912,8 +18515,14 @@ "output_cost_per_token": 0, "output_vector_size": 768, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models", - "supported_endpoints": ["/v1/embeddings"], - "supported_modalities": ["text", "image", "video"] + "supported_endpoints": [ + "/v1/embeddings" + ], + "supported_modalities": [ + "text", + "image", + "video" + ] }, "multimodalembedding@001": { "input_cost_per_character": 2e-7, @@ -15929,8 +18538,14 @@ "output_cost_per_token": 0, "output_vector_size": 768, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models", - "supported_endpoints": ["/v1/embeddings"], - "supported_modalities": ["text", "image", "video"] + "supported_endpoints": [ + "/v1/embeddings" + ], + "supported_modalities": [ + "text", + "image", + "video" + ] }, "nscale/Qwen/QwQ-32B": { "input_cost_per_token": 1.8e-7, @@ -15964,9 +18579,11 @@ "input_cost_per_pixel": 1.3e-9, "litellm_provider": "nscale", "mode": "image_generation", - "output_cost_per_pixel": 0.0, + "output_cost_per_pixel": 0, "source": "https://docs.nscale.com/docs/inference/serverless-models/current#image-models", - "supported_endpoints": ["/v1/images/generations"] + "supported_endpoints": [ + "/v1/images/generations" + ] }, "nscale/deepseek-ai/DeepSeek-R1-Distill-Llama-70B": { "input_cost_per_token": 3.75e-7, @@ -16069,19 +18686,21 @@ "input_cost_per_pixel": 3e-9, "litellm_provider": "nscale", "mode": "image_generation", - "output_cost_per_pixel": 0.0, + "output_cost_per_pixel": 0, "source": "https://docs.nscale.com/docs/inference/serverless-models/current#image-models", - "supported_endpoints": ["/v1/images/generations"] + "supported_endpoints": [ + "/v1/images/generations" + ] }, "o1": { - "cache_read_input_token_cost": 7.5e-6, - "input_cost_per_token": 1.5e-5, + "cache_read_input_token_cost": 0.0000075, + "input_cost_per_token": 0.000015, "litellm_provider": "openai", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 6e-5, + "output_cost_per_token": 0.00006, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_pdf_input": true, @@ -16093,14 +18712,14 @@ "supports_vision": true }, "o1-2024-12-17": { - "cache_read_input_token_cost": 7.5e-6, - "input_cost_per_token": 1.5e-5, + "cache_read_input_token_cost": 0.0000075, + "input_cost_per_token": 0.000015, "litellm_provider": "openai", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 6e-5, + "output_cost_per_token": 0.00006, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_pdf_input": true, @@ -16113,55 +18732,55 @@ }, "o1-mini": { "cache_read_input_token_cost": 5.5e-7, - "input_cost_per_token": 1.1e-6, + "input_cost_per_token": 0.0000011, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 65536, "max_tokens": 65536, "mode": "chat", - "output_cost_per_token": 4.4e-6, + "output_cost_per_token": 0.0000044, "supports_pdf_input": true, "supports_prompt_caching": true, "supports_vision": true }, "o1-mini-2024-09-12": { "deprecation_date": "2025-10-27", - "cache_read_input_token_cost": 1.5e-6, - "input_cost_per_token": 3e-6, + "cache_read_input_token_cost": 0.0000015, + "input_cost_per_token": 0.000003, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 65536, "max_tokens": 65536, "mode": "chat", - "output_cost_per_token": 1.2e-5, + "output_cost_per_token": 0.000012, "supports_pdf_input": true, "supports_prompt_caching": true, "supports_reasoning": true, "supports_vision": true }, "o1-preview": { - "cache_read_input_token_cost": 7.5e-6, - "input_cost_per_token": 1.5e-5, + "cache_read_input_token_cost": 0.0000075, + "input_cost_per_token": 0.000015, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 6e-5, + "output_cost_per_token": 0.00006, "supports_pdf_input": true, "supports_prompt_caching": true, "supports_reasoning": true, "supports_vision": true }, "o1-preview-2024-09-12": { - "cache_read_input_token_cost": 7.5e-6, - "input_cost_per_token": 1.5e-5, + "cache_read_input_token_cost": 0.0000075, + "input_cost_per_token": 0.000015, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 6e-5, + "output_cost_per_token": 0.00006, "supports_pdf_input": true, "supports_prompt_caching": true, "supports_reasoning": true, @@ -16169,7 +18788,7 @@ }, "o1-pro": { "input_cost_per_token": 0.00015, - "input_cost_per_token_batches": 7.5e-5, + "input_cost_per_token_batches": 0.000075, "litellm_provider": "openai", "max_input_tokens": 200000, "max_output_tokens": 100000, @@ -16177,9 +18796,17 @@ "mode": "responses", "output_cost_per_token": 0.0006, "output_cost_per_token_batches": 0.0003, - "supported_endpoints": ["/v1/responses", "/v1/batch"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "supported_endpoints": [ + "/v1/responses", + "/v1/batch" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_native_streaming": false, "supports_parallel_function_calling": true, @@ -16193,7 +18820,7 @@ }, "o1-pro-2025-03-19": { "input_cost_per_token": 0.00015, - "input_cost_per_token_batches": 7.5e-5, + "input_cost_per_token_batches": 0.000075, "litellm_provider": "openai", "max_input_tokens": 200000, "max_output_tokens": 100000, @@ -16201,9 +18828,17 @@ "mode": "responses", "output_cost_per_token": 0.0006, "output_cost_per_token_batches": 0.0003, - "supported_endpoints": ["/v1/responses", "/v1/batch"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "supported_endpoints": [ + "/v1/responses", + "/v1/batch" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_native_streaming": false, "supports_parallel_function_calling": true, @@ -16219,25 +18854,30 @@ "cache_read_input_token_cost": 5e-7, "cache_read_input_token_cost_flex": 2.5e-7, "cache_read_input_token_cost_priority": 8.75e-7, - "input_cost_per_token": 2e-6, - "input_cost_per_token_flex": 1e-6, - "input_cost_per_token_priority": 3.5e-6, + "input_cost_per_token": 0.000002, + "input_cost_per_token_flex": 0.000001, + "input_cost_per_token_priority": 0.0000035, "litellm_provider": "openai", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 8e-6, - "output_cost_per_token_flex": 4e-6, - "output_cost_per_token_priority": 1.4e-5, + "output_cost_per_token": 0.000008, + "output_cost_per_token_flex": 0.000004, + "output_cost_per_token_priority": 0.000014, "supported_endpoints": [ "/v1/responses", "/v1/chat/completions", "/v1/completions", "/v1/batch" ], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_parallel_function_calling": false, "supports_pdf_input": true, @@ -16250,21 +18890,26 @@ }, "o3-2025-04-16": { "cache_read_input_token_cost": 5e-7, - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "openai", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 8e-6, + "output_cost_per_token": 0.000008, "supported_endpoints": [ "/v1/responses", "/v1/chat/completions", "/v1/completions", "/v1/batch" ], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_parallel_function_calling": false, "supports_pdf_input": true, @@ -16276,19 +18921,28 @@ "supports_vision": true }, "o3-deep-research": { - "cache_read_input_token_cost": 2.5e-6, - "input_cost_per_token": 1e-5, - "input_cost_per_token_batches": 5e-6, + "cache_read_input_token_cost": 0.0000025, + "input_cost_per_token": 0.00001, + "input_cost_per_token_batches": 0.000005, "litellm_provider": "openai", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "responses", - "output_cost_per_token": 4e-5, - "output_cost_per_token_batches": 2e-5, - "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "output_cost_per_token": 0.00004, + "output_cost_per_token_batches": 0.00002, + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -16300,19 +18954,28 @@ "supports_vision": true }, "o3-deep-research-2025-06-26": { - "cache_read_input_token_cost": 2.5e-6, - "input_cost_per_token": 1e-5, - "input_cost_per_token_batches": 5e-6, + "cache_read_input_token_cost": 0.0000025, + "input_cost_per_token": 0.00001, + "input_cost_per_token_batches": 0.000005, "litellm_provider": "openai", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "responses", - "output_cost_per_token": 4e-5, - "output_cost_per_token_batches": 2e-5, - "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "output_cost_per_token": 0.00004, + "output_cost_per_token_batches": 0.00002, + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -16325,13 +18988,13 @@ }, "o3-mini": { "cache_read_input_token_cost": 5.5e-7, - "input_cost_per_token": 1.1e-6, + "input_cost_per_token": 0.0000011, "litellm_provider": "openai", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 4.4e-6, + "output_cost_per_token": 0.0000044, "supports_function_calling": true, "supports_parallel_function_calling": false, "supports_prompt_caching": true, @@ -16342,13 +19005,13 @@ }, "o3-mini-2025-01-31": { "cache_read_input_token_cost": 5.5e-7, - "input_cost_per_token": 1.1e-6, + "input_cost_per_token": 0.0000011, "litellm_provider": "openai", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 4.4e-6, + "output_cost_per_token": 0.0000044, "supports_function_calling": true, "supports_parallel_function_calling": false, "supports_prompt_caching": true, @@ -16358,18 +19021,26 @@ "supports_vision": false }, "o3-pro": { - "input_cost_per_token": 2e-5, - "input_cost_per_token_batches": 1e-5, + "input_cost_per_token": 0.00002, + "input_cost_per_token_batches": 0.00001, "litellm_provider": "openai", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "responses", - "output_cost_per_token": 8e-5, - "output_cost_per_token_batches": 4e-5, - "supported_endpoints": ["/v1/responses", "/v1/batch"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "output_cost_per_token": 0.00008, + "output_cost_per_token_batches": 0.00004, + "supported_endpoints": [ + "/v1/responses", + "/v1/batch" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_parallel_function_calling": false, "supports_pdf_input": true, @@ -16380,18 +19051,26 @@ "supports_vision": true }, "o3-pro-2025-06-10": { - "input_cost_per_token": 2e-5, - "input_cost_per_token_batches": 1e-5, + "input_cost_per_token": 0.00002, + "input_cost_per_token_batches": 0.00001, "litellm_provider": "openai", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "responses", - "output_cost_per_token": 8e-5, - "output_cost_per_token_batches": 4e-5, - "supported_endpoints": ["/v1/responses", "/v1/batch"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "output_cost_per_token": 0.00008, + "output_cost_per_token_batches": 0.00004, + "supported_endpoints": [ + "/v1/responses", + "/v1/batch" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_parallel_function_calling": false, "supports_pdf_input": true, @@ -16405,17 +19084,17 @@ "cache_read_input_token_cost": 2.75e-7, "cache_read_input_token_cost_flex": 1.375e-7, "cache_read_input_token_cost_priority": 5e-7, - "input_cost_per_token": 1.1e-6, + "input_cost_per_token": 0.0000011, "input_cost_per_token_flex": 5.5e-7, - "input_cost_per_token_priority": 2e-6, + "input_cost_per_token_priority": 0.000002, "litellm_provider": "openai", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 4.4e-6, - "output_cost_per_token_flex": 2.2e-6, - "output_cost_per_token_priority": 8e-6, + "output_cost_per_token": 0.0000044, + "output_cost_per_token_flex": 0.0000022, + "output_cost_per_token_priority": 0.000008, "supports_function_calling": true, "supports_parallel_function_calling": false, "supports_pdf_input": true, @@ -16428,13 +19107,13 @@ }, "o4-mini-2025-04-16": { "cache_read_input_token_cost": 2.75e-7, - "input_cost_per_token": 1.1e-6, + "input_cost_per_token": 0.0000011, "litellm_provider": "openai", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 4.4e-6, + "output_cost_per_token": 0.0000044, "supports_function_calling": true, "supports_parallel_function_calling": false, "supports_pdf_input": true, @@ -16447,18 +19126,27 @@ }, "o4-mini-deep-research": { "cache_read_input_token_cost": 5e-7, - "input_cost_per_token": 2e-6, - "input_cost_per_token_batches": 1e-6, + "input_cost_per_token": 0.000002, + "input_cost_per_token_batches": 0.000001, "litellm_provider": "openai", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "responses", - "output_cost_per_token": 8e-6, - "output_cost_per_token_batches": 4e-6, - "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "output_cost_per_token": 0.000008, + "output_cost_per_token_batches": 0.000004, + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -16471,18 +19159,27 @@ }, "o4-mini-deep-research-2025-06-26": { "cache_read_input_token_cost": 5e-7, - "input_cost_per_token": 2e-6, - "input_cost_per_token_batches": 1e-6, + "input_cost_per_token": 0.000002, + "input_cost_per_token_batches": 0.000001, "litellm_provider": "openai", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "responses", - "output_cost_per_token": 8e-6, - "output_cost_per_token_batches": 4e-6, - "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "output_cost_per_token": 0.000008, + "output_cost_per_token_batches": 0.000004, + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/batch", + "/v1/responses" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -16494,25 +19191,25 @@ "supports_vision": true }, "oci/meta.llama-3.1-405b-instruct": { - "input_cost_per_token": 1.068e-5, + "input_cost_per_token": 0.00001068, "litellm_provider": "oci", "max_input_tokens": 128000, "max_output_tokens": 4000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 1.068e-5, + "output_cost_per_token": 0.00001068, "source": "https://www.oracle.com/artificial-intelligence/generative-ai/generative-ai-service/pricing", "supports_function_calling": true, "supports_response_schema": false }, "oci/meta.llama-3.2-90b-vision-instruct": { - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "oci", "max_input_tokens": 128000, "max_output_tokens": 4000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 2e-6, + "output_cost_per_token": 0.000002, "source": "https://www.oracle.com/artificial-intelligence/generative-ai/generative-ai-service/pricing", "supports_function_calling": true, "supports_response_schema": false @@ -16554,7 +19251,7 @@ "supports_response_schema": false }, "oci/xai.grok-3": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "oci", "max_input_tokens": 131072, "max_output_tokens": 131072, @@ -16566,13 +19263,13 @@ "supports_response_schema": false }, "oci/xai.grok-3-fast": { - "input_cost_per_token": 5e-6, + "input_cost_per_token": 0.000005, "litellm_provider": "oci", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 2.5e-5, + "output_cost_per_token": 0.000025, "source": "https://www.oracle.com/artificial-intelligence/generative-ai/generative-ai-service/pricing", "supports_function_calling": true, "supports_response_schema": false @@ -16596,13 +19293,13 @@ "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 4e-6, + "output_cost_per_token": 0.000004, "source": "https://www.oracle.com/artificial-intelligence/generative-ai/generative-ai-service/pricing", "supports_function_calling": true, "supports_response_schema": false }, "oci/xai.grok-4": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "oci", "max_input_tokens": 128000, "max_output_tokens": 128000, @@ -16614,345 +19311,345 @@ "supports_response_schema": false }, "oci/cohere.command-latest": { - "input_cost_per_token": 1.56e-6, + "input_cost_per_token": 0.00000156, "litellm_provider": "oci", "max_input_tokens": 128000, "max_output_tokens": 4000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 1.56e-6, + "output_cost_per_token": 0.00000156, "source": "https://www.oracle.com/cloud/ai/generative-ai/pricing/", "supports_function_calling": true, "supports_response_schema": false }, "oci/cohere.command-a-03-2025": { - "input_cost_per_token": 1.56e-6, + "input_cost_per_token": 0.00000156, "litellm_provider": "oci", "max_input_tokens": 256000, "max_output_tokens": 4000, "max_tokens": 256000, "mode": "chat", - "output_cost_per_token": 1.56e-6, + "output_cost_per_token": 0.00000156, "source": "https://www.oracle.com/cloud/ai/generative-ai/pricing/", "supports_function_calling": true, "supports_response_schema": false }, "oci/cohere.command-plus-latest": { - "input_cost_per_token": 1.56e-6, + "input_cost_per_token": 0.00000156, "litellm_provider": "oci", "max_input_tokens": 128000, "max_output_tokens": 4000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 1.56e-6, + "output_cost_per_token": 0.00000156, "source": "https://www.oracle.com/cloud/ai/generative-ai/pricing/", "supports_function_calling": true, "supports_response_schema": false }, "ollama/codegeex4": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "ollama", "max_input_tokens": 32768, "max_output_tokens": 8192, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "supports_function_calling": false }, "ollama/codegemma": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "ollama", "max_input_tokens": 8192, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "completion", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "ollama/codellama": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "ollama", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "completion", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "ollama/deepseek-coder-v2-base": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "ollama", "max_input_tokens": 8192, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "completion", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "supports_function_calling": true }, "ollama/deepseek-coder-v2-instruct": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "ollama", "max_input_tokens": 32768, "max_output_tokens": 8192, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "supports_function_calling": true }, "ollama/deepseek-coder-v2-lite-base": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "ollama", "max_input_tokens": 8192, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "completion", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "supports_function_calling": true }, "ollama/deepseek-coder-v2-lite-instruct": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "ollama", "max_input_tokens": 32768, "max_output_tokens": 8192, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "supports_function_calling": true }, "ollama/deepseek-v3.1:671b-cloud": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "ollama", "max_input_tokens": 163840, "max_output_tokens": 163840, "max_tokens": 163840, "mode": "chat", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "supports_function_calling": true }, "ollama/gpt-oss:120b-cloud": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "ollama", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "supports_function_calling": true }, "ollama/gpt-oss:20b-cloud": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "ollama", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "supports_function_calling": true }, "ollama/internlm2_5-20b-chat": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "ollama", "max_input_tokens": 32768, "max_output_tokens": 8192, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "supports_function_calling": true }, "ollama/llama2": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "ollama", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "ollama/llama2-uncensored": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "ollama", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "completion", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "ollama/llama2:13b": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "ollama", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "ollama/llama2:70b": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "ollama", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "ollama/llama2:7b": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "ollama", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "ollama/llama3": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "ollama", "max_input_tokens": 8192, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "ollama/llama3.1": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "ollama", "max_input_tokens": 8192, "max_output_tokens": 8192, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "supports_function_calling": true }, "ollama/llama3:70b": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "ollama", "max_input_tokens": 8192, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "ollama/llama3:8b": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "ollama", "max_input_tokens": 8192, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "ollama/mistral": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "ollama", "max_input_tokens": 8192, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "completion", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "supports_function_calling": true }, "ollama/mistral-7B-Instruct-v0.1": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "ollama", "max_input_tokens": 8192, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "supports_function_calling": true }, "ollama/mistral-7B-Instruct-v0.2": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "ollama", "max_input_tokens": 32768, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "supports_function_calling": true }, "ollama/mistral-large-instruct-2407": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "ollama", "max_input_tokens": 65536, "max_output_tokens": 8192, "max_tokens": 65536, "mode": "chat", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "supports_function_calling": true }, "ollama/mixtral-8x22B-Instruct-v0.1": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "ollama", "max_input_tokens": 65536, "max_output_tokens": 65536, "max_tokens": 65536, "mode": "chat", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "supports_function_calling": true }, "ollama/mixtral-8x7B-Instruct-v0.1": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "ollama", "max_input_tokens": 32768, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "supports_function_calling": true }, "ollama/orca-mini": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "ollama", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "completion", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "ollama/qwen3-coder:480b-cloud": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "ollama", "max_input_tokens": 262144, "max_output_tokens": 262144, "max_tokens": 262144, "mode": "chat", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "supports_function_calling": true }, "ollama/vicuna": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "ollama", "max_input_tokens": 2048, "max_output_tokens": 2048, "max_tokens": 2048, "mode": "completion", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "omni-moderation-2024-09-26": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "openai", "max_input_tokens": 32768, "max_output_tokens": 0, "max_tokens": 32768, "mode": "moderation", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "omni-moderation-latest": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "openai", "max_input_tokens": 32768, "max_output_tokens": 0, "max_tokens": 32768, "mode": "moderation", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "omni-moderation-latest-intents": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "openai", "max_input_tokens": 32768, "max_output_tokens": 0, "max_tokens": 32768, "mode": "moderation", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "openai.gpt-oss-120b-1:0": { "input_cost_per_token": 1.5e-7, @@ -16981,31 +19678,31 @@ "supports_tool_choice": true }, "openrouter/anthropic/claude-2": { - "input_cost_per_token": 1.102e-5, + "input_cost_per_token": 0.00001102, "litellm_provider": "openrouter", "max_output_tokens": 8191, "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 3.268e-5, + "output_cost_per_token": 0.00003268, "supports_tool_choice": true }, "openrouter/anthropic/claude-3-5-haiku": { - "input_cost_per_token": 1e-6, + "input_cost_per_token": 0.000001, "litellm_provider": "openrouter", "max_tokens": 200000, "mode": "chat", - "output_cost_per_token": 5e-6, + "output_cost_per_token": 0.000005, "supports_function_calling": true, "supports_tool_choice": true }, "openrouter/anthropic/claude-3-5-haiku-20241022": { - "input_cost_per_token": 1e-6, + "input_cost_per_token": 0.000001, "litellm_provider": "openrouter", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 5e-6, + "output_cost_per_token": 0.000005, "supports_function_calling": true, "supports_tool_choice": true, "tool_use_system_prompt_tokens": 264 @@ -17016,7 +19713,7 @@ "litellm_provider": "openrouter", "max_tokens": 200000, "mode": "chat", - "output_cost_per_token": 1.25e-6, + "output_cost_per_token": 0.00000125, "supports_function_calling": true, "supports_tool_choice": true, "supports_vision": true @@ -17028,20 +19725,20 @@ "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 1.25e-6, + "output_cost_per_token": 0.00000125, "supports_function_calling": true, "supports_tool_choice": true, "supports_vision": true, "tool_use_system_prompt_tokens": 264 }, "openrouter/anthropic/claude-3-opus": { - "input_cost_per_token": 1.5e-5, + "input_cost_per_token": 0.000015, "litellm_provider": "openrouter", "max_input_tokens": 200000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 7.5e-5, + "output_cost_per_token": 0.000075, "supports_function_calling": true, "supports_tool_choice": true, "supports_vision": true, @@ -17049,23 +19746,23 @@ }, "openrouter/anthropic/claude-3-sonnet": { "input_cost_per_image": 0.0048, - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "openrouter", "max_tokens": 200000, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "supports_function_calling": true, "supports_tool_choice": true, "supports_vision": true }, "openrouter/anthropic/claude-3.5-sonnet": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "openrouter", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "supports_assistant_prefill": true, "supports_computer_use": true, "supports_function_calling": true, @@ -17074,13 +19771,13 @@ "tool_use_system_prompt_tokens": 159 }, "openrouter/anthropic/claude-3.5-sonnet:beta": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "openrouter", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "supports_computer_use": true, "supports_function_calling": true, "supports_tool_choice": true, @@ -17089,13 +19786,13 @@ }, "openrouter/anthropic/claude-3.7-sonnet": { "input_cost_per_image": 0.0048, - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "openrouter", "max_input_tokens": 200000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "supports_assistant_prefill": true, "supports_computer_use": true, "supports_function_calling": true, @@ -17106,13 +19803,13 @@ }, "openrouter/anthropic/claude-3.7-sonnet:beta": { "input_cost_per_image": 0.0048, - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "openrouter", "max_input_tokens": 200000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "supports_computer_use": true, "supports_function_calling": true, "supports_reasoning": true, @@ -17121,25 +19818,25 @@ "tool_use_system_prompt_tokens": 159 }, "openrouter/anthropic/claude-instant-v1": { - "input_cost_per_token": 1.63e-6, + "input_cost_per_token": 0.00000163, "litellm_provider": "openrouter", "max_output_tokens": 8191, "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 5.51e-6, + "output_cost_per_token": 0.00000551, "supports_tool_choice": true }, "openrouter/anthropic/claude-opus-4": { "input_cost_per_image": 0.0048, - "cache_creation_input_token_cost": 1.875e-5, - "cache_read_input_token_cost": 1.5e-6, - "input_cost_per_token": 1.5e-5, + "cache_creation_input_token_cost": 0.00001875, + "cache_read_input_token_cost": 0.0000015, + "input_cost_per_token": 0.000015, "litellm_provider": "openrouter", "max_input_tokens": 200000, "max_output_tokens": 32000, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 7.5e-5, + "output_cost_per_token": 0.000075, "supports_assistant_prefill": true, "supports_computer_use": true, "supports_function_calling": true, @@ -17151,16 +19848,16 @@ }, "openrouter/anthropic/claude-opus-4.1": { "input_cost_per_image": 0.0048, - "cache_creation_input_token_cost": 1.875e-5, - "cache_creation_input_token_cost_above_1hr": 3e-5, - "cache_read_input_token_cost": 1.5e-6, - "input_cost_per_token": 1.5e-5, + "cache_creation_input_token_cost": 0.00001875, + "cache_creation_input_token_cost_above_1hr": 0.00003, + "cache_read_input_token_cost": 0.0000015, + "input_cost_per_token": 0.000015, "litellm_provider": "openrouter", "max_input_tokens": 200000, "max_output_tokens": 32000, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 7.5e-5, + "output_cost_per_token": 0.000075, "supports_assistant_prefill": true, "supports_computer_use": true, "supports_function_calling": true, @@ -17172,19 +19869,19 @@ }, "openrouter/anthropic/claude-sonnet-4": { "input_cost_per_image": 0.0048, - "cache_creation_input_token_cost": 3.75e-6, - "cache_creation_input_token_cost_above_200k_tokens": 7.5e-6, + "cache_creation_input_token_cost": 0.00000375, + "cache_creation_input_token_cost_above_200k_tokens": 0.0000075, "cache_read_input_token_cost": 3e-7, "cache_read_input_token_cost_above_200k_tokens": 6e-7, - "input_cost_per_token": 3e-6, - "input_cost_per_token_above_200k_tokens": 6e-6, - "output_cost_per_token_above_200k_tokens": 2.25e-5, + "input_cost_per_token": 0.000003, + "input_cost_per_token_above_200k_tokens": 0.000006, + "output_cost_per_token_above_200k_tokens": 0.0000225, "litellm_provider": "openrouter", "max_input_tokens": 1000000, "max_output_tokens": 64000, "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "supports_assistant_prefill": true, "supports_computer_use": true, "supports_function_calling": true, @@ -17196,19 +19893,19 @@ }, "openrouter/anthropic/claude-sonnet-4.5": { "input_cost_per_image": 0.0048, - "cache_creation_input_token_cost": 3.75e-6, + "cache_creation_input_token_cost": 0.00000375, "cache_read_input_token_cost": 3e-7, - "input_cost_per_token": 3e-6, - "input_cost_per_token_above_200k_tokens": 6e-6, - "output_cost_per_token_above_200k_tokens": 2.25e-5, - "cache_creation_input_token_cost_above_200k_tokens": 7.5e-6, + "input_cost_per_token": 0.000003, + "input_cost_per_token_above_200k_tokens": 0.000006, + "output_cost_per_token_above_200k_tokens": 0.0000225, + "cache_creation_input_token_cost_above_200k_tokens": 0.0000075, "cache_read_input_token_cost_above_200k_tokens": 6e-7, "litellm_provider": "openrouter", "max_input_tokens": 1000000, "max_output_tokens": 1000000, "max_tokens": 1000000, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "supports_assistant_prefill": true, "supports_computer_use": true, "supports_function_calling": true, @@ -17219,15 +19916,15 @@ "tool_use_system_prompt_tokens": 159 }, "openrouter/anthropic/claude-haiku-4.5": { - "cache_creation_input_token_cost": 1.25e-6, + "cache_creation_input_token_cost": 0.00000125, "cache_read_input_token_cost": 1e-7, - "input_cost_per_token": 1e-6, + "input_cost_per_token": 0.000001, "litellm_provider": "openrouter", "max_input_tokens": 200000, "max_output_tokens": 200000, "max_tokens": 200000, "mode": "chat", - "output_cost_per_token": 5e-6, + "output_cost_per_token": 0.000005, "supports_assistant_prefill": true, "supports_computer_use": true, "supports_function_calling": true, @@ -17257,11 +19954,11 @@ "supports_tool_choice": true }, "openrouter/cohere/command-r-plus": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "openrouter", "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "supports_tool_choice": true }, "openrouter/databricks/dbrx-instruct": { @@ -17343,7 +20040,7 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 2.19e-6, + "output_cost_per_token": 0.00000219, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_prompt_caching": true, @@ -17358,7 +20055,7 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 2.15e-6, + "output_cost_per_token": 0.00000215, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_prompt_caching": true, @@ -17409,7 +20106,7 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_token": 2.5e-6, + "output_cost_per_token": 0.0000025, "supports_audio_output": true, "supports_function_calling": true, "supports_response_schema": true, @@ -17419,7 +20116,7 @@ }, "openrouter/google/gemini-2.5-pro": { "input_cost_per_audio_token": 7e-7, - "input_cost_per_token": 1.25e-6, + "input_cost_per_token": 0.00000125, "litellm_provider": "openrouter", "max_audio_length_hours": 8.4, "max_audio_per_prompt": 1, @@ -17431,7 +20128,7 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_token": 1e-5, + "output_cost_per_token": 0.00001, "supports_audio_output": true, "supports_function_calling": true, "supports_response_schema": true, @@ -17439,15 +20136,62 @@ "supports_tool_choice": true, "supports_vision": true }, + "openrouter/google/gemini-3-pro-preview": { + "cache_read_input_token_cost": 2e-7, + "cache_read_input_token_cost_above_200k_tokens": 4e-7, + "cache_creation_input_token_cost_above_200k_tokens": 2.5e-7, + "input_cost_per_token": 0.000002, + "input_cost_per_token_above_200k_tokens": 0.000004, + "input_cost_per_token_batches": 0.000001, + "litellm_provider": "openrouter", + "max_audio_length_hours": 8.4, + "max_audio_per_prompt": 1, + "max_images_per_prompt": 3000, + "max_input_tokens": 1048576, + "max_output_tokens": 65535, + "max_pdf_size_mb": 30, + "max_tokens": 65535, + "max_video_length": 1, + "max_videos_per_prompt": 10, + "mode": "chat", + "output_cost_per_token": 0.000012, + "output_cost_per_token_above_200k_tokens": 0.000018, + "output_cost_per_token_batches": 0.000006, + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/completions", + "/v1/batch" + ], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text" + ], + "supports_audio_input": true, + "supports_function_calling": true, + "supports_pdf_input": true, + "supports_prompt_caching": true, + "supports_reasoning": true, + "supports_response_schema": true, + "supports_system_messages": true, + "supports_tool_choice": true, + "supports_video_input": true, + "supports_vision": true, + "supports_web_search": true + }, "openrouter/google/gemini-pro-1.5": { "input_cost_per_image": 0.00265, - "input_cost_per_token": 2.5e-6, + "input_cost_per_token": 0.0000025, "litellm_provider": "openrouter", "max_input_tokens": 1000000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 7.5e-6, + "output_cost_per_token": 0.0000075, "supports_function_calling": true, "supports_tool_choice": true, "supports_vision": true @@ -17480,27 +20224,27 @@ "supports_tool_choice": true }, "openrouter/gryphe/mythomax-l2-13b": { - "input_cost_per_token": 1.875e-6, + "input_cost_per_token": 0.000001875, "litellm_provider": "openrouter", "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 1.875e-6, + "output_cost_per_token": 0.000001875, "supports_tool_choice": true }, "openrouter/jondurbin/airoboros-l2-70b-2.1": { - "input_cost_per_token": 1.3875e-5, + "input_cost_per_token": 0.000013875, "litellm_provider": "openrouter", "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 1.3875e-5, + "output_cost_per_token": 0.000013875, "supports_tool_choice": true }, "openrouter/mancer/weaver": { - "input_cost_per_token": 5.625e-6, + "input_cost_per_token": 0.000005625, "litellm_provider": "openrouter", "max_tokens": 8000, "mode": "chat", - "output_cost_per_token": 5.625e-6, + "output_cost_per_token": 0.000005625, "supports_tool_choice": true }, "openrouter/meta-llama/codellama-34b-instruct": { @@ -17520,11 +20264,11 @@ "supports_tool_choice": true }, "openrouter/meta-llama/llama-2-70b-chat": { - "input_cost_per_token": 1.5e-6, + "input_cost_per_token": 0.0000015, "litellm_provider": "openrouter", "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 1.5e-6, + "output_cost_per_token": 0.0000015, "supports_tool_choice": true }, "openrouter/meta-llama/llama-3-70b-instruct": { @@ -17548,23 +20292,23 @@ "litellm_provider": "openrouter", "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 2.25e-6, + "output_cost_per_token": 0.00000225, "supports_tool_choice": true }, "openrouter/meta-llama/llama-3-8b-instruct:free": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "openrouter", "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "supports_tool_choice": true }, "openrouter/microsoft/wizardlm-2-8x22b:nitro": { - "input_cost_per_token": 1e-6, + "input_cost_per_token": 0.000001, "litellm_provider": "openrouter", "max_tokens": 65536, "mode": "chat", - "output_cost_per_token": 1e-6, + "output_cost_per_token": 0.000001, "supports_tool_choice": true }, "openrouter/minimax/minimax-m2": { @@ -17574,7 +20318,7 @@ "max_output_tokens": 204800, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 1.02e-6, + "output_cost_per_token": 0.00000102, "supports_function_calling": true, "supports_prompt_caching": false, "supports_reasoning": true, @@ -17589,19 +20333,19 @@ "supports_tool_choice": true }, "openrouter/mistralai/mistral-7b-instruct:free": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "openrouter", "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "supports_tool_choice": true }, "openrouter/mistralai/mistral-large": { - "input_cost_per_token": 8e-6, + "input_cost_per_token": 0.000008, "litellm_provider": "openrouter", "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 2.4e-5, + "output_cost_per_token": 0.000024, "supports_tool_choice": true }, "openrouter/mistralai/mistral-small-3.1-24b-instruct": { @@ -17637,49 +20381,49 @@ "supports_tool_choice": true }, "openrouter/openai/gpt-3.5-turbo": { - "input_cost_per_token": 1.5e-6, + "input_cost_per_token": 0.0000015, "litellm_provider": "openrouter", "max_tokens": 4095, "mode": "chat", - "output_cost_per_token": 2e-6, + "output_cost_per_token": 0.000002, "supports_tool_choice": true }, "openrouter/openai/gpt-3.5-turbo-16k": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "openrouter", "max_tokens": 16383, "mode": "chat", - "output_cost_per_token": 4e-6, + "output_cost_per_token": 0.000004, "supports_tool_choice": true }, "openrouter/openai/gpt-4": { - "input_cost_per_token": 3e-5, + "input_cost_per_token": 0.00003, "litellm_provider": "openrouter", "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 6e-5, + "output_cost_per_token": 0.00006, "supports_tool_choice": true }, "openrouter/openai/gpt-4-vision-preview": { "input_cost_per_image": 0.01445, - "input_cost_per_token": 1e-5, + "input_cost_per_token": 0.00001, "litellm_provider": "openrouter", "max_tokens": 130000, "mode": "chat", - "output_cost_per_token": 3e-5, + "output_cost_per_token": 0.00003, "supports_function_calling": true, "supports_tool_choice": true, "supports_vision": true }, "openrouter/openai/gpt-4.1": { "cache_read_input_token_cost": 5e-7, - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "openrouter", "max_input_tokens": 1047576, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 8e-6, + "output_cost_per_token": 0.000008, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -17690,13 +20434,13 @@ }, "openrouter/openai/gpt-4.1-2025-04-14": { "cache_read_input_token_cost": 5e-7, - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "openrouter", "max_input_tokens": 1047576, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 8e-6, + "output_cost_per_token": 0.000008, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -17713,7 +20457,7 @@ "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 1.6e-6, + "output_cost_per_token": 0.0000016, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -17730,7 +20474,7 @@ "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 1.6e-6, + "output_cost_per_token": 0.0000016, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -17774,26 +20518,26 @@ "supports_vision": true }, "openrouter/openai/gpt-4o": { - "input_cost_per_token": 2.5e-6, + "input_cost_per_token": 0.0000025, "litellm_provider": "openrouter", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 1e-5, + "output_cost_per_token": 0.00001, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_tool_choice": true, "supports_vision": true }, "openrouter/openai/gpt-4o-2024-05-13": { - "input_cost_per_token": 5e-6, + "input_cost_per_token": 0.000005, "litellm_provider": "openrouter", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_tool_choice": true, @@ -17801,43 +20545,58 @@ }, "openrouter/openai/gpt-5-chat": { "cache_read_input_token_cost": 1.25e-7, - "input_cost_per_token": 1.25e-6, + "input_cost_per_token": 0.00000125, "litellm_provider": "openrouter", "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 1e-5, - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "output_cost_per_token": 0.00001, + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_reasoning": true, "supports_tool_choice": true }, "openrouter/openai/gpt-5-codex": { "cache_read_input_token_cost": 1.25e-7, - "input_cost_per_token": 1.25e-6, + "input_cost_per_token": 0.00000125, "litellm_provider": "openrouter", "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 1e-5, - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "output_cost_per_token": 0.00001, + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_reasoning": true, "supports_tool_choice": true }, "openrouter/openai/gpt-5": { "cache_read_input_token_cost": 1.25e-7, - "input_cost_per_token": 1.25e-6, + "input_cost_per_token": 0.00000125, "litellm_provider": "openrouter", "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 1e-5, - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "output_cost_per_token": 0.00001, + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_reasoning": true, "supports_tool_choice": true }, @@ -17849,9 +20608,14 @@ "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 2e-6, - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "output_cost_per_token": 0.000002, + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_reasoning": true, "supports_tool_choice": true }, @@ -17864,8 +20628,13 @@ "max_tokens": 128000, "mode": "chat", "output_cost_per_token": 4e-7, - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text"], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ], "supports_reasoning": true, "supports_tool_choice": true }, @@ -17900,14 +20669,14 @@ "supports_tool_choice": true }, "openrouter/openai/o1": { - "cache_read_input_token_cost": 7.5e-6, - "input_cost_per_token": 1.5e-5, + "cache_read_input_token_cost": 0.0000075, + "input_cost_per_token": 0.000015, "litellm_provider": "openrouter", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 6e-5, + "output_cost_per_token": 0.00006, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -17917,65 +20686,65 @@ "supports_vision": true }, "openrouter/openai/o1-mini": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "openrouter", "max_input_tokens": 128000, "max_output_tokens": 65536, "max_tokens": 65536, "mode": "chat", - "output_cost_per_token": 1.2e-5, + "output_cost_per_token": 0.000012, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_tool_choice": true, "supports_vision": false }, "openrouter/openai/o1-mini-2024-09-12": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "openrouter", "max_input_tokens": 128000, "max_output_tokens": 65536, "max_tokens": 65536, "mode": "chat", - "output_cost_per_token": 1.2e-5, + "output_cost_per_token": 0.000012, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_tool_choice": true, "supports_vision": false }, "openrouter/openai/o1-preview": { - "input_cost_per_token": 1.5e-5, + "input_cost_per_token": 0.000015, "litellm_provider": "openrouter", "max_input_tokens": 128000, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 6e-5, + "output_cost_per_token": 0.00006, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_tool_choice": true, "supports_vision": false }, "openrouter/openai/o1-preview-2024-09-12": { - "input_cost_per_token": 1.5e-5, + "input_cost_per_token": 0.000015, "litellm_provider": "openrouter", "max_input_tokens": 128000, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 6e-5, + "output_cost_per_token": 0.00006, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_tool_choice": true, "supports_vision": false }, "openrouter/openai/o3-mini": { - "input_cost_per_token": 1.1e-6, + "input_cost_per_token": 0.0000011, "litellm_provider": "openrouter", "max_input_tokens": 128000, "max_output_tokens": 65536, "max_tokens": 65536, "mode": "chat", - "output_cost_per_token": 4.4e-6, + "output_cost_per_token": 0.0000044, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_reasoning": true, @@ -17983,13 +20752,13 @@ "supports_vision": false }, "openrouter/openai/o3-mini-high": { - "input_cost_per_token": 1.1e-6, + "input_cost_per_token": 0.0000011, "litellm_provider": "openrouter", "max_input_tokens": 128000, "max_output_tokens": 65536, "max_tokens": 65536, "mode": "chat", - "output_cost_per_token": 4.4e-6, + "output_cost_per_token": 0.0000044, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_reasoning": true, @@ -17997,11 +20766,11 @@ "supports_vision": false }, "openrouter/pygmalionai/mythalion-13b": { - "input_cost_per_token": 1.875e-6, + "input_cost_per_token": 0.000001875, "litellm_provider": "openrouter", "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 1.875e-6, + "output_cost_per_token": 0.000001875, "supports_tool_choice": true }, "openrouter/qwen/qwen-2.5-coder-32b-instruct": { @@ -18044,26 +20813,26 @@ "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 3.4e-6, + "output_cost_per_token": 0.0000034, "source": "https://openrouter.ai/switchpoint/router", "supports_tool_choice": true }, "openrouter/undi95/remm-slerp-l2-13b": { - "input_cost_per_token": 1.875e-6, + "input_cost_per_token": 0.000001875, "litellm_provider": "openrouter", "max_tokens": 6144, "mode": "chat", - "output_cost_per_token": 1.875e-6, + "output_cost_per_token": 0.000001875, "supports_tool_choice": true }, "openrouter/x-ai/grok-4": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "openrouter", "max_input_tokens": 256000, "max_output_tokens": 256000, "max_tokens": 256000, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "source": "https://openrouter.ai/x-ai/grok-4", "supports_function_calling": true, "supports_reasoning": true, @@ -18085,13 +20854,13 @@ "supports_web_search": false }, "openrouter/z-ai/glm-4.6": { - "input_cost_per_token": 4.0e-7, + "input_cost_per_token": 4e-7, "litellm_provider": "openrouter", "max_input_tokens": 202800, "max_output_tokens": 131000, "max_tokens": 202800, "mode": "chat", - "output_cost_per_token": 1.75e-6, + "output_cost_per_token": 0.00000175, "source": "https://openrouter.ai/z-ai/glm-4.6", "supports_function_calling": true, "supports_reasoning": true, @@ -18104,7 +20873,7 @@ "max_output_tokens": 131000, "max_tokens": 202800, "mode": "chat", - "output_cost_per_token": 1.9e-6, + "output_cost_per_token": 0.0000019, "source": "https://openrouter.ai/z-ai/glm-4.6:exacto", "supports_function_calling": true, "supports_reasoning": true, @@ -18389,7 +21158,7 @@ "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 1.4e-6 + "output_cost_per_token": 0.0000014 }, "perplexity/codellama-70b-instruct": { "input_cost_per_token": 7e-7, @@ -18398,7 +21167,7 @@ "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 2.8e-6 + "output_cost_per_token": 0.0000028 }, "perplexity/llama-2-70b-chat": { "input_cost_per_token": 7e-7, @@ -18407,16 +21176,16 @@ "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 2.8e-6 + "output_cost_per_token": 0.0000028 }, "perplexity/llama-3.1-70b-instruct": { - "input_cost_per_token": 1e-6, + "input_cost_per_token": 0.000001, "litellm_provider": "perplexity", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 1e-6 + "output_cost_per_token": 0.000001 }, "perplexity/llama-3.1-8b-instruct": { "input_cost_per_token": 2e-7, @@ -18429,33 +21198,33 @@ }, "perplexity/llama-3.1-sonar-huge-128k-online": { "deprecation_date": "2025-02-22", - "input_cost_per_token": 5e-6, + "input_cost_per_token": 0.000005, "litellm_provider": "perplexity", "max_input_tokens": 127072, "max_output_tokens": 127072, "max_tokens": 127072, "mode": "chat", - "output_cost_per_token": 5e-6 + "output_cost_per_token": 0.000005 }, "perplexity/llama-3.1-sonar-large-128k-chat": { "deprecation_date": "2025-02-22", - "input_cost_per_token": 1e-6, + "input_cost_per_token": 0.000001, "litellm_provider": "perplexity", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 1e-6 + "output_cost_per_token": 0.000001 }, "perplexity/llama-3.1-sonar-large-128k-online": { "deprecation_date": "2025-02-22", - "input_cost_per_token": 1e-6, + "input_cost_per_token": 0.000001, "litellm_provider": "perplexity", "max_input_tokens": 127072, "max_output_tokens": 127072, "max_tokens": 127072, "mode": "chat", - "output_cost_per_token": 1e-6 + "output_cost_per_token": 0.000001 }, "perplexity/llama-3.1-sonar-small-128k-chat": { "deprecation_date": "2025-02-22", @@ -18502,17 +21271,17 @@ "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 2.8e-6 + "output_cost_per_token": 0.0000028 }, "perplexity/pplx-70b-online": { "input_cost_per_request": 0.005, - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "perplexity", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 2.8e-6 + "output_cost_per_token": 0.0000028 }, "perplexity/pplx-7b-chat": { "input_cost_per_token": 7e-8, @@ -18525,7 +21294,7 @@ }, "perplexity/pplx-7b-online": { "input_cost_per_request": 0.005, - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "perplexity", "max_input_tokens": 4096, "max_output_tokens": 4096, @@ -18534,12 +21303,12 @@ "output_cost_per_token": 2.8e-7 }, "perplexity/sonar": { - "input_cost_per_token": 1e-6, + "input_cost_per_token": 0.000001, "litellm_provider": "perplexity", "max_input_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 1e-6, + "output_cost_per_token": 0.000001, "search_context_cost_per_query": { "search_context_size_high": 0.012, "search_context_size_low": 0.005, @@ -18548,14 +21317,14 @@ "supports_web_search": true }, "perplexity/sonar-deep-research": { - "citation_cost_per_token": 2e-6, - "input_cost_per_token": 2e-6, + "citation_cost_per_token": 0.000002, + "input_cost_per_token": 0.000002, "litellm_provider": "perplexity", "max_input_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_reasoning_token": 3e-6, - "output_cost_per_token": 8e-6, + "output_cost_per_reasoning_token": 0.000003, + "output_cost_per_token": 0.000008, "search_context_cost_per_query": { "search_context_size_high": 0.005, "search_context_size_low": 0.005, @@ -18571,7 +21340,7 @@ "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 1.8e-6 + "output_cost_per_token": 0.0000018 }, "perplexity/sonar-medium-online": { "input_cost_per_request": 0.005, @@ -18581,16 +21350,16 @@ "max_output_tokens": 12000, "max_tokens": 12000, "mode": "chat", - "output_cost_per_token": 1.8e-6 + "output_cost_per_token": 0.0000018 }, "perplexity/sonar-pro": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "perplexity", "max_input_tokens": 200000, "max_output_tokens": 8000, "max_tokens": 8000, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "search_context_cost_per_query": { "search_context_size_high": 0.014, "search_context_size_low": 0.006, @@ -18599,12 +21368,12 @@ "supports_web_search": true }, "perplexity/sonar-reasoning": { - "input_cost_per_token": 1e-6, + "input_cost_per_token": 0.000001, "litellm_provider": "perplexity", "max_input_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 5e-6, + "output_cost_per_token": 0.000005, "search_context_cost_per_query": { "search_context_size_high": 0.014, "search_context_size_low": 0.005, @@ -18614,12 +21383,12 @@ "supports_web_search": true }, "perplexity/sonar-reasoning-pro": { - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "perplexity", "max_input_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 8e-6, + "output_cost_per_token": 0.000008, "search_context_cost_per_query": { "search_context_size_high": 0.014, "search_context_size_low": 0.006, @@ -18654,7 +21423,7 @@ "max_output_tokens": 65536, "max_tokens": 262144, "mode": "chat", - "output_cost_per_token": 1.8e-6, + "output_cost_per_token": 0.0000018, "supports_function_calling": true, "supports_reasoning": true, "supports_tool_choice": true @@ -18678,7 +21447,7 @@ "max_output_tokens": 131072, "max_tokens": 262144, "mode": "chat", - "output_cost_per_token": 6.0e-7, + "output_cost_per_token": 6e-7, "supports_function_calling": true, "supports_reasoning": true, "supports_tool_choice": true @@ -18690,7 +21459,7 @@ "max_output_tokens": 16384, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 6.0e-7, + "output_cost_per_token": 6e-7, "supports_function_calling": true, "supports_reasoning": true, "supports_tool_choice": true @@ -18700,14 +21469,18 @@ "mode": "image_generation", "output_cost_per_image": 0.022, "source": "https://www.recraft.ai/docs#pricing", - "supported_endpoints": ["/v1/images/generations"] + "supported_endpoints": [ + "/v1/images/generations" + ] }, "recraft/recraftv3": { "litellm_provider": "recraft", "mode": "image_generation", "output_cost_per_image": 0.04, "source": "https://www.recraft.ai/docs#pricing", - "supported_endpoints": ["/v1/images/generations"] + "supported_endpoints": [ + "/v1/images/generations" + ] }, "replicate/meta/llama-2-13b": { "input_cost_per_token": 1e-7, @@ -18736,7 +21509,7 @@ "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 2.75e-6, + "output_cost_per_token": 0.00000275, "supports_tool_choice": true }, "replicate/meta/llama-2-70b-chat": { @@ -18746,7 +21519,7 @@ "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 2.75e-6, + "output_cost_per_token": 0.00000275, "supports_tool_choice": true }, "replicate/meta/llama-2-7b": { @@ -18776,7 +21549,7 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 2.75e-6, + "output_cost_per_token": 0.00000275, "supports_tool_choice": true }, "replicate/meta/llama-3-70b-instruct": { @@ -18786,7 +21559,7 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 2.75e-6, + "output_cost_per_token": 0.00000275, "supports_tool_choice": true }, "replicate/meta/llama-3-8b": { @@ -18836,140 +21609,140 @@ "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 1e-6, + "output_cost_per_token": 0.000001, "supports_tool_choice": true }, "rerank-english-v2.0": { "input_cost_per_query": 0.002, - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "cohere", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_query_tokens": 2048, "max_tokens": 4096, "mode": "rerank", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "rerank-english-v3.0": { "input_cost_per_query": 0.002, - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "cohere", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_query_tokens": 2048, "max_tokens": 4096, "mode": "rerank", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "rerank-multilingual-v2.0": { "input_cost_per_query": 0.002, - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "cohere", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_query_tokens": 2048, "max_tokens": 4096, "mode": "rerank", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "rerank-multilingual-v3.0": { "input_cost_per_query": 0.002, - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "cohere", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_query_tokens": 2048, "max_tokens": 4096, "mode": "rerank", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "rerank-v3.5": { "input_cost_per_query": 0.002, - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "cohere", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_query_tokens": 2048, "max_tokens": 4096, "mode": "rerank", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "nvidia_nim/nvidia/nv-rerankqa-mistral-4b-v3": { - "input_cost_per_query": 0.0, - "input_cost_per_token": 0.0, + "input_cost_per_query": 0, + "input_cost_per_token": 0, "litellm_provider": "nvidia_nim", "mode": "rerank", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "nvidia_nim/nvidia/llama-3_2-nv-rerankqa-1b-v2": { - "input_cost_per_query": 0.0, - "input_cost_per_token": 0.0, + "input_cost_per_query": 0, + "input_cost_per_token": 0, "litellm_provider": "nvidia_nim", "mode": "rerank", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "sagemaker/meta-textgeneration-llama-2-13b": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "sagemaker", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "completion", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "sagemaker/meta-textgeneration-llama-2-13b-f": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "sagemaker", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "sagemaker/meta-textgeneration-llama-2-70b": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "sagemaker", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "completion", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "sagemaker/meta-textgeneration-llama-2-70b-b-f": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "sagemaker", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "sagemaker/meta-textgeneration-llama-2-7b": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "sagemaker", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "completion", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "sagemaker/meta-textgeneration-llama-2-7b-f": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "sagemaker", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "sambanova/DeepSeek-R1": { - "input_cost_per_token": 5e-6, + "input_cost_per_token": 0.000005, "litellm_provider": "sambanova", "max_input_tokens": 32768, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 7e-6, + "output_cost_per_token": 0.000007, "source": "https://cloud.sambanova.ai/plans/pricing" }, "sambanova/DeepSeek-R1-Distill-Llama-70B": { @@ -18979,17 +21752,17 @@ "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 1.4e-6, + "output_cost_per_token": 0.0000014, "source": "https://cloud.sambanova.ai/plans/pricing" }, "sambanova/DeepSeek-V3-0324": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "sambanova", "max_input_tokens": 32768, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 4.5e-6, + "output_cost_per_token": 0.0000045, "source": "https://cloud.sambanova.ai/plans/pricing", "supports_function_calling": true, "supports_reasoning": true, @@ -19005,7 +21778,7 @@ "notes": "For vision models, images are converted to 6432 input tokens and are billed at that amount" }, "mode": "chat", - "output_cost_per_token": 1.8e-6, + "output_cost_per_token": 0.0000018, "source": "https://cloud.sambanova.ai/plans/pricing", "supports_function_calling": true, "supports_response_schema": true, @@ -19029,13 +21802,13 @@ "supports_tool_choice": true }, "sambanova/Meta-Llama-3.1-405B-Instruct": { - "input_cost_per_token": 5e-6, + "input_cost_per_token": 0.000005, "litellm_provider": "sambanova", "max_input_tokens": 16384, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 1e-5, + "output_cost_per_token": 0.00001, "source": "https://cloud.sambanova.ai/plans/pricing", "supports_function_calling": true, "supports_response_schema": true, @@ -19081,7 +21854,7 @@ "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 1.2e-6, + "output_cost_per_token": 0.0000012, "source": "https://cloud.sambanova.ai/plans/pricing", "supports_function_calling": true, "supports_response_schema": true, @@ -19104,7 +21877,7 @@ "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 1e-6, + "output_cost_per_token": 0.000001, "source": "https://cloud.sambanova.ai/plans/pricing" }, "sambanova/Qwen2-Audio-7B-Instruct": { @@ -19135,8 +21908,8 @@ "max_tokens": 32768, "max_input_tokens": 32768, "max_output_tokens": 32768, - "input_cost_per_token": 3e-6, - "output_cost_per_token": 4.5e-6, + "input_cost_per_token": 0.000003, + "output_cost_per_token": 0.0000045, "litellm_provider": "sambanova", "mode": "chat", "supports_function_calling": true, @@ -19148,8 +21921,8 @@ "max_tokens": 131072, "max_input_tokens": 131072, "max_output_tokens": 131072, - "input_cost_per_token": 3e-6, - "output_cost_per_token": 4.5e-6, + "input_cost_per_token": 0.000003, + "output_cost_per_token": 0.0000045, "litellm_provider": "sambanova", "mode": "chat", "supports_function_calling": true, @@ -19373,19 +22146,19 @@ "input_cost_per_pixel": 3.81469e-8, "litellm_provider": "openai", "mode": "image_generation", - "output_cost_per_pixel": 0.0 + "output_cost_per_pixel": 0 }, "standard/1024-x-1792/dall-e-3": { "input_cost_per_pixel": 4.359e-8, "litellm_provider": "openai", "mode": "image_generation", - "output_cost_per_pixel": 0.0 + "output_cost_per_pixel": 0 }, "standard/1792-x-1024/dall-e-3": { "input_cost_per_pixel": 4.359e-8, "litellm_provider": "openai", "mode": "image_generation", - "output_cost_per_pixel": 0.0 + "output_cost_per_pixel": 0 }, "tavily/search": { "input_cost_per_query": 0.008, @@ -19452,23 +22225,23 @@ "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models" }, "text-completion-codestral/codestral-2405": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "text-completion-codestral", "max_input_tokens": 32000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "completion", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "source": "https://docs.mistral.ai/capabilities/code_generation/" }, "text-completion-codestral/codestral-latest": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "text-completion-codestral", "max_input_tokens": 32000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "completion", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "source": "https://docs.mistral.ai/capabilities/code_generation/" }, "text-embedding-004": { @@ -19500,8 +22273,8 @@ "max_input_tokens": 8191, "max_tokens": 8191, "mode": "embedding", - "output_cost_per_token": 0.0, - "output_cost_per_token_batches": 0.0, + "output_cost_per_token": 0, + "output_cost_per_token_batches": 0, "output_vector_size": 3072 }, "text-embedding-3-small": { @@ -19511,8 +22284,8 @@ "max_input_tokens": 8191, "max_tokens": 8191, "mode": "embedding", - "output_cost_per_token": 0.0, - "output_cost_per_token_batches": 0.0, + "output_cost_per_token": 0, + "output_cost_per_token_batches": 0, "output_vector_size": 1536 }, "text-embedding-ada-002": { @@ -19521,7 +22294,7 @@ "max_input_tokens": 8191, "max_tokens": 8191, "mode": "embedding", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "output_vector_size": 1536 }, "text-embedding-ada-002-v2": { @@ -19531,8 +22304,8 @@ "max_input_tokens": 8191, "max_tokens": 8191, "mode": "embedding", - "output_cost_per_token": 0.0, - "output_cost_per_token_batches": 0.0 + "output_cost_per_token": 0, + "output_cost_per_token_batches": 0 }, "text-embedding-large-exp-03-07": { "input_cost_per_character": 2.5e-8, @@ -19557,31 +22330,31 @@ "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing" }, "text-moderation-007": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "openai", "max_input_tokens": 32768, "max_output_tokens": 0, "max_tokens": 32768, "mode": "moderation", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "text-moderation-latest": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "openai", "max_input_tokens": 32768, "max_output_tokens": 0, "max_tokens": 32768, "mode": "moderation", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "text-moderation-stable": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "openai", "max_input_tokens": 32768, "max_output_tokens": 0, "max_tokens": 32768, "mode": "moderation", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "text-multilingual-embedding-002": { "input_cost_per_character": 2.5e-8, @@ -19605,23 +22378,23 @@ "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models" }, "text-unicorn": { - "input_cost_per_token": 1e-5, + "input_cost_per_token": 0.00001, "litellm_provider": "vertex_ai-text-models", "max_input_tokens": 8192, "max_output_tokens": 1024, "max_tokens": 1024, "mode": "completion", - "output_cost_per_token": 2.8e-5, + "output_cost_per_token": 0.000028, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models" }, "text-unicorn@001": { - "input_cost_per_token": 1e-5, + "input_cost_per_token": 0.00001, "litellm_provider": "vertex_ai-text-models", "max_input_tokens": 8192, "max_output_tokens": 1024, "max_tokens": 1024, "mode": "completion", - "output_cost_per_token": 2.8e-5, + "output_cost_per_token": 0.000028, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models" }, "textembedding-gecko": { @@ -19705,29 +22478,29 @@ "output_cost_per_token": 3e-7 }, "together-ai-81.1b-110b": { - "input_cost_per_token": 1.8e-6, + "input_cost_per_token": 0.0000018, "litellm_provider": "together_ai", "mode": "chat", - "output_cost_per_token": 1.8e-6 + "output_cost_per_token": 0.0000018 }, "together-ai-embedding-151m-to-350m": { "input_cost_per_token": 1.6e-8, "litellm_provider": "together_ai", "mode": "embedding", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "together-ai-embedding-up-to-150m": { "input_cost_per_token": 8e-9, "litellm_provider": "together_ai", "mode": "embedding", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "together_ai/baai/bge-base-en-v1.5": { "input_cost_per_token": 8e-9, "litellm_provider": "together_ai", "max_input_tokens": 512, "mode": "embedding", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "output_vector_size": 768 }, "together_ai/BAAI/bge-base-en-v1.5": { @@ -19735,7 +22508,7 @@ "litellm_provider": "together_ai", "max_input_tokens": 512, "mode": "embedding", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "output_vector_size": 768 }, "together-ai-up-to-4b": { @@ -19763,7 +22536,7 @@ "litellm_provider": "together_ai", "max_input_tokens": 262000, "mode": "chat", - "output_cost_per_token": 6e-6, + "output_cost_per_token": 0.000006, "source": "https://www.together.ai/models/qwen3-235b-a22b-instruct-2507-fp8", "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -19774,7 +22547,7 @@ "litellm_provider": "together_ai", "max_input_tokens": 256000, "mode": "chat", - "output_cost_per_token": 3e-6, + "output_cost_per_token": 0.000003, "source": "https://www.together.ai/models/qwen3-235b-a22b-thinking-2507", "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -19792,24 +22565,24 @@ "supports_tool_choice": false }, "together_ai/Qwen/Qwen3-Coder-480B-A35B-Instruct-FP8": { - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "together_ai", "max_input_tokens": 256000, "mode": "chat", - "output_cost_per_token": 2e-6, + "output_cost_per_token": 0.000002, "source": "https://www.together.ai/models/qwen3-coder-480b-a35b-instruct", "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_tool_choice": true }, "together_ai/deepseek-ai/DeepSeek-R1": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "together_ai", "max_input_tokens": 128000, "max_output_tokens": 20480, "max_tokens": 20480, "mode": "chat", - "output_cost_per_token": 7e-6, + "output_cost_per_token": 0.000007, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_tool_choice": true @@ -19819,20 +22592,20 @@ "litellm_provider": "together_ai", "max_input_tokens": 128000, "mode": "chat", - "output_cost_per_token": 2.19e-6, + "output_cost_per_token": 0.00000219, "source": "https://www.together.ai/models/deepseek-r1-0528-throughput", "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_tool_choice": true }, "together_ai/deepseek-ai/DeepSeek-V3": { - "input_cost_per_token": 1.25e-6, + "input_cost_per_token": 0.00000125, "litellm_provider": "together_ai", "max_input_tokens": 65536, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 1.25e-6, + "output_cost_per_token": 0.00000125, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_tool_choice": true @@ -19842,7 +22615,7 @@ "litellm_provider": "together_ai", "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 1.7e-6, + "output_cost_per_token": 0.0000017, "source": "https://www.together.ai/models/deepseek-v3-1", "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -19895,10 +22668,10 @@ "supports_tool_choice": true }, "together_ai/meta-llama/Meta-Llama-3.1-405B-Instruct-Turbo": { - "input_cost_per_token": 3.5e-6, + "input_cost_per_token": 0.0000035, "litellm_provider": "together_ai", "mode": "chat", - "output_cost_per_token": 3.5e-6, + "output_cost_per_token": 0.0000035, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_tool_choice": true @@ -19949,10 +22722,10 @@ "supports_tool_choice": true }, "together_ai/moonshotai/Kimi-K2-Instruct": { - "input_cost_per_token": 1e-6, + "input_cost_per_token": 0.000001, "litellm_provider": "together_ai", "mode": "chat", - "output_cost_per_token": 3e-6, + "output_cost_per_token": 0.000003, "source": "https://www.together.ai/models/kimi-k2-instruct", "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -19992,18 +22765,32 @@ "litellm_provider": "together_ai", "max_input_tokens": 128000, "mode": "chat", - "output_cost_per_token": 1.1e-6, + "output_cost_per_token": 0.0000011, "source": "https://www.together.ai/models/glm-4-5-air", "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_tool_choice": true }, + "together_ai/zai-org/GLM-4.6": { + "input_cost_per_token": 6e-7, + "litellm_provider": "together_ai", + "max_input_tokens": 200000, + "max_output_tokens": 200000, + "max_tokens": 200000, + "mode": "chat", + "output_cost_per_token": 0.0000022, + "source": "https://www.together.ai/models/glm-4-6", + "supports_function_calling": true, + "supports_parallel_function_calling": true, + "supports_reasoning": true, + "supports_tool_choice": true + }, "together_ai/moonshotai/Kimi-K2-Instruct-0905": { - "input_cost_per_token": 1e-6, + "input_cost_per_token": 0.000001, "litellm_provider": "together_ai", "max_input_tokens": 262144, "mode": "chat", - "output_cost_per_token": 3e-6, + "output_cost_per_token": 0.000003, "source": "https://www.together.ai/models/kimi-k2-0905", "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -20014,7 +22801,7 @@ "litellm_provider": "together_ai", "max_input_tokens": 262144, "mode": "chat", - "output_cost_per_token": 1.5e-6, + "output_cost_per_token": 0.0000015, "source": "https://www.together.ai/models/qwen3-next-80b-a3b-instruct", "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -20025,23 +22812,27 @@ "litellm_provider": "together_ai", "max_input_tokens": 262144, "mode": "chat", - "output_cost_per_token": 1.5e-6, + "output_cost_per_token": 0.0000015, "source": "https://www.together.ai/models/qwen3-next-80b-a3b-thinking", "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_tool_choice": true }, "tts-1": { - "input_cost_per_character": 1.5e-5, + "input_cost_per_character": 0.000015, "litellm_provider": "openai", "mode": "audio_speech", - "supported_endpoints": ["/v1/audio/speech"] + "supported_endpoints": [ + "/v1/audio/speech" + ] }, "tts-1-hd": { - "input_cost_per_character": 3e-5, + "input_cost_per_character": 0.00003, "litellm_provider": "openai", "mode": "audio_speech", - "supported_endpoints": ["/v1/audio/speech"] + "supported_endpoints": [ + "/v1/audio/speech" + ] }, "us.amazon.nova-lite-v1:0": { "input_cost_per_token": 6e-8, @@ -20070,13 +22861,13 @@ "supports_response_schema": true }, "us.amazon.nova-premier-v1:0": { - "input_cost_per_token": 2.5e-6, + "input_cost_per_token": 0.0000025, "litellm_provider": "bedrock_converse", "max_input_tokens": 1000000, "max_output_tokens": 10000, "max_tokens": 10000, "mode": "chat", - "output_cost_per_token": 1.25e-5, + "output_cost_per_token": 0.0000125, "supports_function_calling": true, "supports_pdf_input": true, "supports_prompt_caching": false, @@ -20090,7 +22881,7 @@ "max_output_tokens": 10000, "max_tokens": 10000, "mode": "chat", - "output_cost_per_token": 3.2e-6, + "output_cost_per_token": 0.0000032, "supports_function_calling": true, "supports_pdf_input": true, "supports_prompt_caching": true, @@ -20098,7 +22889,7 @@ "supports_vision": true }, "us.anthropic.claude-3-5-haiku-20241022-v1:0": { - "cache_creation_input_token_cost": 1e-6, + "cache_creation_input_token_cost": 0.000001, "cache_read_input_token_cost": 8e-8, "input_cost_per_token": 8e-7, "litellm_provider": "bedrock", @@ -20106,7 +22897,7 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 4e-6, + "output_cost_per_token": 0.000004, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_pdf_input": true, @@ -20115,32 +22906,35 @@ "supports_tool_choice": true }, "us.anthropic.claude-haiku-4-5-20251001-v1:0": { - "cache_creation_input_token_cost": 1.375e-6, + "cache_creation_input_token_cost": 0.000001375, "cache_read_input_token_cost": 1.1e-7, - "input_cost_per_token": 1.1e-6, - "litellm_provider": "bedrock", + "input_cost_per_token": 0.0000011, + "litellm_provider": "bedrock_converse", "max_input_tokens": 200000, - "max_output_tokens": 8192, - "max_tokens": 8192, + "max_output_tokens": 64000, + "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 5.5e-6, + "output_cost_per_token": 0.0000055, "source": "https://aws.amazon.com/about-aws/whats-new/2025/10/claude-4-5-haiku-anthropic-amazon-bedrock", "supports_assistant_prefill": true, + "supports_computer_use": true, "supports_function_calling": true, "supports_pdf_input": true, "supports_prompt_caching": true, "supports_reasoning": true, "supports_response_schema": true, - "supports_tool_choice": true + "supports_tool_choice": true, + "supports_vision": true, + "tool_use_system_prompt_tokens": 346 }, "us.anthropic.claude-3-5-sonnet-20240620-v1:0": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "bedrock", "max_input_tokens": 200000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "supports_function_calling": true, "supports_pdf_input": true, "supports_response_schema": true, @@ -20148,15 +22942,15 @@ "supports_vision": true }, "us.anthropic.claude-3-5-sonnet-20241022-v2:0": { - "cache_creation_input_token_cost": 3.75e-6, + "cache_creation_input_token_cost": 0.00000375, "cache_read_input_token_cost": 3e-7, - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "bedrock", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "supports_assistant_prefill": true, "supports_computer_use": true, "supports_function_calling": true, @@ -20167,15 +22961,15 @@ "supports_vision": true }, "us.anthropic.claude-3-7-sonnet-20250219-v1:0": { - "cache_creation_input_token_cost": 3.75e-6, + "cache_creation_input_token_cost": 0.00000375, "cache_read_input_token_cost": 3e-7, - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "bedrock_converse", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "supports_assistant_prefill": true, "supports_computer_use": true, "supports_function_calling": true, @@ -20193,7 +22987,7 @@ "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 1.25e-6, + "output_cost_per_token": 0.00000125, "supports_function_calling": true, "supports_pdf_input": true, "supports_response_schema": true, @@ -20201,26 +22995,26 @@ "supports_vision": true }, "us.anthropic.claude-3-opus-20240229-v1:0": { - "input_cost_per_token": 1.5e-5, + "input_cost_per_token": 0.000015, "litellm_provider": "bedrock", "max_input_tokens": 200000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 7.5e-5, + "output_cost_per_token": 0.000075, "supports_function_calling": true, "supports_response_schema": true, "supports_tool_choice": true, "supports_vision": true }, "us.anthropic.claude-3-sonnet-20240229-v1:0": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "bedrock", "max_input_tokens": 200000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "supports_function_calling": true, "supports_pdf_input": true, "supports_response_schema": true, @@ -20228,15 +23022,15 @@ "supports_vision": true }, "us.anthropic.claude-opus-4-1-20250805-v1:0": { - "cache_creation_input_token_cost": 1.875e-5, - "cache_read_input_token_cost": 1.5e-6, - "input_cost_per_token": 1.5e-5, + "cache_creation_input_token_cost": 0.00001875, + "cache_read_input_token_cost": 0.0000015, + "input_cost_per_token": 0.000015, "litellm_provider": "bedrock_converse", "max_input_tokens": 200000, "max_output_tokens": 32000, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 7.5e-5, + "output_cost_per_token": 0.000075, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -20254,19 +23048,19 @@ "tool_use_system_prompt_tokens": 159 }, "us.anthropic.claude-sonnet-4-5-20250929-v1:0": { - "cache_creation_input_token_cost": 4.125e-6, + "cache_creation_input_token_cost": 0.000004125, "cache_read_input_token_cost": 3.3e-7, - "input_cost_per_token": 3.3e-6, - "input_cost_per_token_above_200k_tokens": 6.6e-6, - "output_cost_per_token_above_200k_tokens": 2.475e-5, - "cache_creation_input_token_cost_above_200k_tokens": 8.25e-6, + "input_cost_per_token": 0.0000033, + "input_cost_per_token_above_200k_tokens": 0.0000066, + "output_cost_per_token_above_200k_tokens": 0.00002475, + "cache_creation_input_token_cost_above_200k_tokens": 0.00000825, "cache_read_input_token_cost_above_200k_tokens": 6.6e-7, "litellm_provider": "bedrock_converse", "max_input_tokens": 200000, "max_output_tokens": 64000, "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 1.65e-5, + "output_cost_per_token": 0.0000165, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -20284,34 +23078,36 @@ "tool_use_system_prompt_tokens": 346 }, "au.anthropic.claude-haiku-4-5-20251001-v1:0": { - "cache_creation_input_token_cost": 1.375e-6, + "cache_creation_input_token_cost": 0.000001375, "cache_read_input_token_cost": 1.1e-7, - "input_cost_per_token": 1.1e-6, + "input_cost_per_token": 0.0000011, "litellm_provider": "bedrock_converse", "max_input_tokens": 200000, - "max_output_tokens": 8192, - "max_tokens": 8192, + "max_output_tokens": 64000, + "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 5.5e-6, + "output_cost_per_token": 0.0000055, "supports_assistant_prefill": true, + "supports_computer_use": true, "supports_function_calling": true, "supports_pdf_input": true, "supports_prompt_caching": true, + "supports_reasoning": true, "supports_response_schema": true, "supports_tool_choice": true, "supports_vision": true, "tool_use_system_prompt_tokens": 346 }, "us.anthropic.claude-opus-4-20250514-v1:0": { - "cache_creation_input_token_cost": 1.875e-5, - "cache_read_input_token_cost": 1.5e-6, - "input_cost_per_token": 1.5e-5, + "cache_creation_input_token_cost": 0.00001875, + "cache_read_input_token_cost": 0.0000015, + "input_cost_per_token": 0.000015, "litellm_provider": "bedrock_converse", "max_input_tokens": 200000, "max_output_tokens": 32000, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 7.5e-5, + "output_cost_per_token": 0.000075, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -20329,19 +23125,19 @@ "tool_use_system_prompt_tokens": 159 }, "us.anthropic.claude-sonnet-4-20250514-v1:0": { - "cache_creation_input_token_cost": 3.75e-6, + "cache_creation_input_token_cost": 0.00000375, "cache_read_input_token_cost": 3e-7, - "input_cost_per_token": 3e-6, - "input_cost_per_token_above_200k_tokens": 6e-6, - "output_cost_per_token_above_200k_tokens": 2.25e-5, - "cache_creation_input_token_cost_above_200k_tokens": 7.5e-6, + "input_cost_per_token": 0.000003, + "input_cost_per_token_above_200k_tokens": 0.000006, + "output_cost_per_token_above_200k_tokens": 0.0000225, + "cache_creation_input_token_cost_above_200k_tokens": 0.0000075, "cache_read_input_token_cost_above_200k_tokens": 6e-7, "litellm_provider": "bedrock_converse", "max_input_tokens": 1000000, "max_output_tokens": 64000, "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -20359,25 +23155,25 @@ "tool_use_system_prompt_tokens": 159 }, "us.deepseek.r1-v1:0": { - "input_cost_per_token": 1.35e-6, + "input_cost_per_token": 0.00000135, "litellm_provider": "bedrock_converse", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 5.4e-6, + "output_cost_per_token": 0.0000054, "supports_function_calling": false, "supports_reasoning": true, "supports_tool_choice": false }, "us.meta.llama3-1-405b-instruct-v1:0": { - "input_cost_per_token": 5.32e-6, + "input_cost_per_token": 0.00000532, "litellm_provider": "bedrock", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 1.6e-5, + "output_cost_per_token": 0.000016, "supports_function_calling": true, "supports_tool_choice": false }, @@ -20438,13 +23234,13 @@ "supports_tool_choice": false }, "us.meta.llama3-2-90b-instruct-v1:0": { - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "bedrock", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 2e-6, + "output_cost_per_token": 0.000002, "supports_function_calling": true, "supports_tool_choice": false, "supports_vision": true @@ -20470,8 +23266,14 @@ "mode": "chat", "output_cost_per_token": 9.7e-7, "output_cost_per_token_batches": 4.85e-7, - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text", "code"], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text", + "code" + ], "supports_function_calling": true, "supports_tool_choice": false }, @@ -20485,30 +23287,36 @@ "mode": "chat", "output_cost_per_token": 6.6e-7, "output_cost_per_token_batches": 3.3e-7, - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text", "code"], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text", + "code" + ], "supports_function_calling": true, "supports_tool_choice": false }, "us.mistral.pixtral-large-2502-v1:0": { - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "bedrock_converse", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 6e-6, + "output_cost_per_token": 0.000006, "supports_function_calling": true, "supports_tool_choice": false }, "v0/v0-1.0-md": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "v0", "max_input_tokens": 128000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_system_messages": true, @@ -20516,13 +23324,13 @@ "supports_vision": true }, "v0/v0-1.5-lg": { - "input_cost_per_token": 1.5e-5, + "input_cost_per_token": 0.000015, "litellm_provider": "v0", "max_input_tokens": 512000, "max_output_tokens": 512000, "max_tokens": 512000, "mode": "chat", - "output_cost_per_token": 7.5e-5, + "output_cost_per_token": 0.000075, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_system_messages": true, @@ -20530,13 +23338,13 @@ "supports_vision": true }, "v0/v0-1.5-md": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "v0", "max_input_tokens": 128000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_system_messages": true, @@ -20586,7 +23394,7 @@ "max_output_tokens": 66536, "max_tokens": 262144, "mode": "chat", - "output_cost_per_token": 1.6e-6 + "output_cost_per_token": 0.0000016 }, "vercel_ai_gateway/amazon/nova-lite": { "input_cost_per_token": 6e-8, @@ -20613,7 +23421,7 @@ "max_output_tokens": 8192, "max_tokens": 300000, "mode": "chat", - "output_cost_per_token": 3.2e-6 + "output_cost_per_token": 0.0000032 }, "vercel_ai_gateway/amazon/titan-embed-text-v2": { "input_cost_per_token": 2e-8, @@ -20622,7 +23430,7 @@ "max_output_tokens": 0, "max_tokens": 0, "mode": "chat", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "vercel_ai_gateway/anthropic/claude-3-haiku": { "cache_creation_input_token_cost": 3e-7, @@ -20633,21 +23441,21 @@ "max_output_tokens": 4096, "max_tokens": 200000, "mode": "chat", - "output_cost_per_token": 1.25e-6 + "output_cost_per_token": 0.00000125 }, "vercel_ai_gateway/anthropic/claude-3-opus": { - "cache_creation_input_token_cost": 1.875e-5, - "cache_read_input_token_cost": 1.5e-6, - "input_cost_per_token": 1.5e-5, + "cache_creation_input_token_cost": 0.00001875, + "cache_read_input_token_cost": 0.0000015, + "input_cost_per_token": 0.000015, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 200000, "max_output_tokens": 4096, "max_tokens": 200000, "mode": "chat", - "output_cost_per_token": 7.5e-5 + "output_cost_per_token": 0.000075 }, "vercel_ai_gateway/anthropic/claude-3.5-haiku": { - "cache_creation_input_token_cost": 1e-6, + "cache_creation_input_token_cost": 0.000001, "cache_read_input_token_cost": 8e-8, "input_cost_per_token": 8e-7, "litellm_provider": "vercel_ai_gateway", @@ -20655,60 +23463,60 @@ "max_output_tokens": 8192, "max_tokens": 200000, "mode": "chat", - "output_cost_per_token": 4e-6 + "output_cost_per_token": 0.000004 }, "vercel_ai_gateway/anthropic/claude-3.5-sonnet": { - "cache_creation_input_token_cost": 3.75e-6, + "cache_creation_input_token_cost": 0.00000375, "cache_read_input_token_cost": 3e-7, - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 200000, "mode": "chat", - "output_cost_per_token": 1.5e-5 + "output_cost_per_token": 0.000015 }, "vercel_ai_gateway/anthropic/claude-3.7-sonnet": { - "cache_creation_input_token_cost": 3.75e-6, + "cache_creation_input_token_cost": 0.00000375, "cache_read_input_token_cost": 3e-7, - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 200000, "max_output_tokens": 64000, "max_tokens": 200000, "mode": "chat", - "output_cost_per_token": 1.5e-5 + "output_cost_per_token": 0.000015 }, "vercel_ai_gateway/anthropic/claude-4-opus": { - "cache_creation_input_token_cost": 1.875e-5, - "cache_read_input_token_cost": 1.5e-6, - "input_cost_per_token": 1.5e-5, + "cache_creation_input_token_cost": 0.00001875, + "cache_read_input_token_cost": 0.0000015, + "input_cost_per_token": 0.000015, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 200000, "max_output_tokens": 32000, "max_tokens": 200000, "mode": "chat", - "output_cost_per_token": 7.5e-5 + "output_cost_per_token": 0.000075 }, "vercel_ai_gateway/anthropic/claude-4-sonnet": { - "cache_creation_input_token_cost": 3.75e-6, + "cache_creation_input_token_cost": 0.00000375, "cache_read_input_token_cost": 3e-7, - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 200000, "max_output_tokens": 64000, "max_tokens": 200000, "mode": "chat", - "output_cost_per_token": 1.5e-5 + "output_cost_per_token": 0.000015 }, "vercel_ai_gateway/cohere/command-a": { - "input_cost_per_token": 2.5e-6, + "input_cost_per_token": 0.0000025, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 256000, "max_output_tokens": 8000, "max_tokens": 256000, "mode": "chat", - "output_cost_per_token": 1e-5 + "output_cost_per_token": 0.00001 }, "vercel_ai_gateway/cohere/command-r": { "input_cost_per_token": 1.5e-7, @@ -20720,13 +23528,13 @@ "output_cost_per_token": 6e-7 }, "vercel_ai_gateway/cohere/command-r-plus": { - "input_cost_per_token": 2.5e-6, + "input_cost_per_token": 0.0000025, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 1e-5 + "output_cost_per_token": 0.00001 }, "vercel_ai_gateway/cohere/embed-v4.0": { "input_cost_per_token": 1.2e-7, @@ -20735,7 +23543,7 @@ "max_output_tokens": 0, "max_tokens": 0, "mode": "chat", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "vercel_ai_gateway/deepseek/deepseek-r1": { "input_cost_per_token": 5.5e-7, @@ -20744,7 +23552,7 @@ "max_output_tokens": 8192, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 2.19e-6 + "output_cost_per_token": 0.00000219 }, "vercel_ai_gateway/deepseek/deepseek-r1-distill-llama-70b": { "input_cost_per_token": 7.5e-7, @@ -20789,16 +23597,16 @@ "max_output_tokens": 65536, "max_tokens": 1000000, "mode": "chat", - "output_cost_per_token": 2.5e-6 + "output_cost_per_token": 0.0000025 }, "vercel_ai_gateway/google/gemini-2.5-pro": { - "input_cost_per_token": 2.5e-6, + "input_cost_per_token": 0.0000025, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 1048576, "max_output_tokens": 65536, "max_tokens": 1048576, "mode": "chat", - "output_cost_per_token": 1e-5 + "output_cost_per_token": 0.00001 }, "vercel_ai_gateway/google/gemini-embedding-001": { "input_cost_per_token": 1.5e-7, @@ -20807,7 +23615,7 @@ "max_output_tokens": 0, "max_tokens": 0, "mode": "embedding", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "vercel_ai_gateway/google/gemma-2-9b": { "input_cost_per_token": 2e-7, @@ -20825,7 +23633,7 @@ "max_output_tokens": 0, "max_tokens": 0, "mode": "embedding", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "vercel_ai_gateway/google/text-multilingual-embedding-002": { "input_cost_per_token": 2.5e-8, @@ -20834,7 +23642,7 @@ "max_output_tokens": 0, "max_tokens": 0, "mode": "embedding", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "vercel_ai_gateway/inception/mercury-coder-small": { "input_cost_per_token": 2.5e-7, @@ -20843,7 +23651,7 @@ "max_output_tokens": 16384, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 1e-6 + "output_cost_per_token": 0.000001 }, "vercel_ai_gateway/meta/llama-3-70b": { "input_cost_per_token": 5.9e-7, @@ -20960,7 +23768,7 @@ "max_output_tokens": 0, "max_tokens": 0, "mode": "chat", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "vercel_ai_gateway/mistral/devstral-small": { "input_cost_per_token": 7e-8, @@ -20972,13 +23780,13 @@ "output_cost_per_token": 2.8e-7 }, "vercel_ai_gateway/mistral/magistral-medium": { - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 128000, "max_output_tokens": 64000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 5e-6 + "output_cost_per_token": 0.000005 }, "vercel_ai_gateway/mistral/magistral-small": { "input_cost_per_token": 5e-7, @@ -20987,7 +23795,7 @@ "max_output_tokens": 64000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 1.5e-6 + "output_cost_per_token": 0.0000015 }, "vercel_ai_gateway/mistral/ministral-3b": { "input_cost_per_token": 4e-8, @@ -21014,16 +23822,16 @@ "max_output_tokens": 0, "max_tokens": 0, "mode": "chat", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "vercel_ai_gateway/mistral/mistral-large": { - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 32000, "max_output_tokens": 4000, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 6e-6 + "output_cost_per_token": 0.000006 }, "vercel_ai_gateway/mistral/mistral-saba-24b": { "input_cost_per_token": 7.9e-7, @@ -21044,13 +23852,13 @@ "output_cost_per_token": 3e-7 }, "vercel_ai_gateway/mistral/mixtral-8x22b-instruct": { - "input_cost_per_token": 1.2e-6, + "input_cost_per_token": 0.0000012, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 65536, "max_output_tokens": 2048, "max_tokens": 65536, "mode": "chat", - "output_cost_per_token": 1.2e-6 + "output_cost_per_token": 0.0000012 }, "vercel_ai_gateway/mistral/pixtral-12b": { "input_cost_per_token": 1.5e-7, @@ -21062,13 +23870,13 @@ "output_cost_per_token": 1.5e-7 }, "vercel_ai_gateway/mistral/pixtral-large": { - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 128000, "max_output_tokens": 4000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 6e-6 + "output_cost_per_token": 0.000006 }, "vercel_ai_gateway/moonshotai/kimi-k2": { "input_cost_per_token": 5.5e-7, @@ -21077,7 +23885,7 @@ "max_output_tokens": 16384, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 2.2e-6 + "output_cost_per_token": 0.0000022 }, "vercel_ai_gateway/morph/morph-v3-fast": { "input_cost_per_token": 8e-7, @@ -21086,7 +23894,7 @@ "max_output_tokens": 16384, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 1.2e-6 + "output_cost_per_token": 0.0000012 }, "vercel_ai_gateway/morph/morph-v3-large": { "input_cost_per_token": 9e-7, @@ -21095,7 +23903,7 @@ "max_output_tokens": 16384, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 1.9e-6 + "output_cost_per_token": 0.0000019 }, "vercel_ai_gateway/openai/gpt-3.5-turbo": { "input_cost_per_token": 5e-7, @@ -21104,39 +23912,39 @@ "max_output_tokens": 4096, "max_tokens": 16385, "mode": "chat", - "output_cost_per_token": 1.5e-6 + "output_cost_per_token": 0.0000015 }, "vercel_ai_gateway/openai/gpt-3.5-turbo-instruct": { - "input_cost_per_token": 1.5e-6, + "input_cost_per_token": 0.0000015, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 8192, "max_output_tokens": 4096, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 2e-6 + "output_cost_per_token": 0.000002 }, "vercel_ai_gateway/openai/gpt-4-turbo": { - "input_cost_per_token": 1e-5, + "input_cost_per_token": 0.00001, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 3e-5 + "output_cost_per_token": 0.00003 }, "vercel_ai_gateway/openai/gpt-4.1": { - "cache_creation_input_token_cost": 0.0, + "cache_creation_input_token_cost": 0, "cache_read_input_token_cost": 5e-7, - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 1047576, "max_output_tokens": 32768, "max_tokens": 1047576, "mode": "chat", - "output_cost_per_token": 8e-6 + "output_cost_per_token": 0.000008 }, "vercel_ai_gateway/openai/gpt-4.1-mini": { - "cache_creation_input_token_cost": 0.0, + "cache_creation_input_token_cost": 0, "cache_read_input_token_cost": 1e-7, "input_cost_per_token": 4e-7, "litellm_provider": "vercel_ai_gateway", @@ -21144,10 +23952,10 @@ "max_output_tokens": 32768, "max_tokens": 1047576, "mode": "chat", - "output_cost_per_token": 1.6e-6 + "output_cost_per_token": 0.0000016 }, "vercel_ai_gateway/openai/gpt-4.1-nano": { - "cache_creation_input_token_cost": 0.0, + "cache_creation_input_token_cost": 0, "cache_read_input_token_cost": 2.5e-8, "input_cost_per_token": 1e-7, "litellm_provider": "vercel_ai_gateway", @@ -21158,18 +23966,18 @@ "output_cost_per_token": 4e-7 }, "vercel_ai_gateway/openai/gpt-4o": { - "cache_creation_input_token_cost": 0.0, - "cache_read_input_token_cost": 1.25e-6, - "input_cost_per_token": 2.5e-6, + "cache_creation_input_token_cost": 0, + "cache_read_input_token_cost": 0.00000125, + "input_cost_per_token": 0.0000025, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 1e-5 + "output_cost_per_token": 0.00001 }, "vercel_ai_gateway/openai/gpt-4o-mini": { - "cache_creation_input_token_cost": 0.0, + "cache_creation_input_token_cost": 0, "cache_read_input_token_cost": 7.5e-8, "input_cost_per_token": 1.5e-7, "litellm_provider": "vercel_ai_gateway", @@ -21180,48 +23988,48 @@ "output_cost_per_token": 6e-7 }, "vercel_ai_gateway/openai/o1": { - "cache_creation_input_token_cost": 0.0, - "cache_read_input_token_cost": 7.5e-6, - "input_cost_per_token": 1.5e-5, + "cache_creation_input_token_cost": 0, + "cache_read_input_token_cost": 0.0000075, + "input_cost_per_token": 0.000015, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 200000, "mode": "chat", - "output_cost_per_token": 6e-5 + "output_cost_per_token": 0.00006 }, "vercel_ai_gateway/openai/o3": { - "cache_creation_input_token_cost": 0.0, + "cache_creation_input_token_cost": 0, "cache_read_input_token_cost": 5e-7, - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 200000, "mode": "chat", - "output_cost_per_token": 8e-6 + "output_cost_per_token": 0.000008 }, "vercel_ai_gateway/openai/o3-mini": { - "cache_creation_input_token_cost": 0.0, + "cache_creation_input_token_cost": 0, "cache_read_input_token_cost": 5.5e-7, - "input_cost_per_token": 1.1e-6, + "input_cost_per_token": 0.0000011, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 200000, "mode": "chat", - "output_cost_per_token": 4.4e-6 + "output_cost_per_token": 0.0000044 }, "vercel_ai_gateway/openai/o4-mini": { - "cache_creation_input_token_cost": 0.0, + "cache_creation_input_token_cost": 0, "cache_read_input_token_cost": 2.75e-7, - "input_cost_per_token": 1.1e-6, + "input_cost_per_token": 0.0000011, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 200000, "mode": "chat", - "output_cost_per_token": 4.4e-6 + "output_cost_per_token": 0.0000044 }, "vercel_ai_gateway/openai/text-embedding-3-large": { "input_cost_per_token": 1.3e-7, @@ -21230,7 +24038,7 @@ "max_output_tokens": 0, "max_tokens": 0, "mode": "embedding", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "vercel_ai_gateway/openai/text-embedding-3-small": { "input_cost_per_token": 2e-8, @@ -21239,7 +24047,7 @@ "max_output_tokens": 0, "max_tokens": 0, "mode": "embedding", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "vercel_ai_gateway/openai/text-embedding-ada-002": { "input_cost_per_token": 1e-7, @@ -21248,97 +24056,97 @@ "max_output_tokens": 0, "max_tokens": 0, "mode": "embedding", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "vercel_ai_gateway/perplexity/sonar": { - "input_cost_per_token": 1e-6, + "input_cost_per_token": 0.000001, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 127000, "max_output_tokens": 8000, "max_tokens": 127000, "mode": "chat", - "output_cost_per_token": 1e-6 + "output_cost_per_token": 0.000001 }, "vercel_ai_gateway/perplexity/sonar-pro": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 200000, "max_output_tokens": 8000, "max_tokens": 200000, "mode": "chat", - "output_cost_per_token": 1.5e-5 + "output_cost_per_token": 0.000015 }, "vercel_ai_gateway/perplexity/sonar-reasoning": { - "input_cost_per_token": 1e-6, + "input_cost_per_token": 0.000001, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 127000, "max_output_tokens": 8000, "max_tokens": 127000, "mode": "chat", - "output_cost_per_token": 5e-6 + "output_cost_per_token": 0.000005 }, "vercel_ai_gateway/perplexity/sonar-reasoning-pro": { - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 127000, "max_output_tokens": 8000, "max_tokens": 127000, "mode": "chat", - "output_cost_per_token": 8e-6 + "output_cost_per_token": 0.000008 }, "vercel_ai_gateway/vercel/v0-1.0-md": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 128000, "max_output_tokens": 32000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 1.5e-5 + "output_cost_per_token": 0.000015 }, "vercel_ai_gateway/vercel/v0-1.5-md": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 128000, "max_output_tokens": 32768, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 1.5e-5 + "output_cost_per_token": 0.000015 }, "vercel_ai_gateway/xai/grok-2": { - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 131072, "max_output_tokens": 4000, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 1e-5 + "output_cost_per_token": 0.00001 }, "vercel_ai_gateway/xai/grok-2-vision": { - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 32768, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 1e-5 + "output_cost_per_token": 0.00001 }, "vercel_ai_gateway/xai/grok-3": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 1.5e-5 + "output_cost_per_token": 0.000015 }, "vercel_ai_gateway/xai/grok-3-fast": { - "input_cost_per_token": 5e-6, + "input_cost_per_token": 0.000005, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 2.5e-5 + "output_cost_per_token": 0.000025 }, "vercel_ai_gateway/xai/grok-3-mini": { "input_cost_per_token": 3e-7, @@ -21356,16 +24164,16 @@ "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 4e-6 + "output_cost_per_token": 0.000004 }, "vercel_ai_gateway/xai/grok-4": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 256000, "max_output_tokens": 256000, "max_tokens": 256000, "mode": "chat", - "output_cost_per_token": 1.5e-5 + "output_cost_per_token": 0.000015 }, "vercel_ai_gateway/zai/glm-4.5": { "input_cost_per_token": 6e-7, @@ -21374,7 +24182,7 @@ "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 2.2e-6 + "output_cost_per_token": 0.0000022 }, "vercel_ai_gateway/zai/glm-4.5-air": { "input_cost_per_token": 2e-7, @@ -21383,7 +24191,7 @@ "max_output_tokens": 96000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 1.1e-6 + "output_cost_per_token": 0.0000011 }, "vercel_ai_gateway/zai/glm-4.6": { "litellm_provider": "vercel_ai_gateway", @@ -21393,48 +24201,48 @@ "max_output_tokens": 200000, "max_tokens": 200000, "mode": "chat", - "output_cost_per_token": 1.8e-6, + "output_cost_per_token": 0.0000018, "source": "https://vercel.com/ai-gateway/models/glm-4.6", "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_tool_choice": true }, "vertex_ai/claude-3-5-haiku": { - "input_cost_per_token": 1e-6, + "input_cost_per_token": 0.000001, "litellm_provider": "vertex_ai-anthropic_models", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 5e-6, + "output_cost_per_token": 0.000005, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_pdf_input": true, "supports_tool_choice": true }, "vertex_ai/claude-3-5-haiku@20241022": { - "input_cost_per_token": 1e-6, + "input_cost_per_token": 0.000001, "litellm_provider": "vertex_ai-anthropic_models", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 5e-6, + "output_cost_per_token": 0.000005, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_pdf_input": true, "supports_tool_choice": true }, "vertex_ai/claude-haiku-4-5@20251001": { - "cache_creation_input_token_cost": 1.25e-6, + "cache_creation_input_token_cost": 0.00000125, "cache_read_input_token_cost": 1e-7, - "input_cost_per_token": 1e-6, + "input_cost_per_token": 0.000001, "litellm_provider": "vertex_ai-anthropic_models", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 5e-6, + "output_cost_per_token": 0.000005, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/partner-models/claude/haiku-4-5", "supports_assistant_prefill": true, "supports_function_calling": true, @@ -21445,13 +24253,13 @@ "supports_tool_choice": true }, "vertex_ai/claude-3-5-sonnet": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "vertex_ai-anthropic_models", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "supports_assistant_prefill": true, "supports_computer_use": true, "supports_function_calling": true, @@ -21460,13 +24268,13 @@ "supports_vision": true }, "vertex_ai/claude-3-5-sonnet-v2": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "vertex_ai-anthropic_models", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "supports_assistant_prefill": true, "supports_computer_use": true, "supports_function_calling": true, @@ -21475,13 +24283,13 @@ "supports_vision": true }, "vertex_ai/claude-3-5-sonnet-v2@20241022": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "vertex_ai-anthropic_models", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "supports_assistant_prefill": true, "supports_computer_use": true, "supports_function_calling": true, @@ -21490,13 +24298,13 @@ "supports_vision": true }, "vertex_ai/claude-3-5-sonnet@20240620": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "vertex_ai-anthropic_models", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_pdf_input": true, @@ -21504,16 +24312,16 @@ "supports_vision": true }, "vertex_ai/claude-3-7-sonnet@20250219": { - "cache_creation_input_token_cost": 3.75e-6, + "cache_creation_input_token_cost": 0.00000375, "cache_read_input_token_cost": 3e-7, "deprecation_date": "2025-06-01", - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "vertex_ai-anthropic_models", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "supports_assistant_prefill": true, "supports_computer_use": true, "supports_function_calling": true, @@ -21532,7 +24340,7 @@ "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 1.25e-6, + "output_cost_per_token": 0.00000125, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_tool_choice": true, @@ -21545,74 +24353,74 @@ "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 1.25e-6, + "output_cost_per_token": 0.00000125, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_tool_choice": true, "supports_vision": true }, "vertex_ai/claude-3-opus": { - "input_cost_per_token": 1.5e-5, + "input_cost_per_token": 0.000015, "litellm_provider": "vertex_ai-anthropic_models", "max_input_tokens": 200000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 7.5e-5, + "output_cost_per_token": 0.000075, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_tool_choice": true, "supports_vision": true }, "vertex_ai/claude-3-opus@20240229": { - "input_cost_per_token": 1.5e-5, + "input_cost_per_token": 0.000015, "litellm_provider": "vertex_ai-anthropic_models", "max_input_tokens": 200000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 7.5e-5, + "output_cost_per_token": 0.000075, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_tool_choice": true, "supports_vision": true }, "vertex_ai/claude-3-sonnet": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "vertex_ai-anthropic_models", "max_input_tokens": 200000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_tool_choice": true, "supports_vision": true }, "vertex_ai/claude-3-sonnet@20240229": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "vertex_ai-anthropic_models", "max_input_tokens": 200000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_tool_choice": true, "supports_vision": true }, "vertex_ai/claude-opus-4": { - "cache_creation_input_token_cost": 1.875e-5, - "cache_read_input_token_cost": 1.5e-6, - "input_cost_per_token": 1.5e-5, + "cache_creation_input_token_cost": 0.00001875, + "cache_read_input_token_cost": 0.0000015, + "input_cost_per_token": 0.000015, "litellm_provider": "vertex_ai-anthropic_models", "max_input_tokens": 200000, "max_output_tokens": 32000, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 7.5e-5, + "output_cost_per_token": 0.000075, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -21630,55 +24438,55 @@ "tool_use_system_prompt_tokens": 159 }, "vertex_ai/claude-opus-4-1": { - "cache_creation_input_token_cost": 1.875e-5, - "cache_read_input_token_cost": 1.5e-6, - "input_cost_per_token": 1.5e-5, - "input_cost_per_token_batches": 7.5e-6, + "cache_creation_input_token_cost": 0.00001875, + "cache_read_input_token_cost": 0.0000015, + "input_cost_per_token": 0.000015, + "input_cost_per_token_batches": 0.0000075, "litellm_provider": "vertex_ai-anthropic_models", "max_input_tokens": 200000, "max_output_tokens": 32000, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 7.5e-5, - "output_cost_per_token_batches": 3.75e-5, + "output_cost_per_token": 0.000075, + "output_cost_per_token_batches": 0.0000375, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_tool_choice": true, "supports_vision": true }, "vertex_ai/claude-opus-4-1@20250805": { - "cache_creation_input_token_cost": 1.875e-5, - "cache_read_input_token_cost": 1.5e-6, - "input_cost_per_token": 1.5e-5, - "input_cost_per_token_batches": 7.5e-6, + "cache_creation_input_token_cost": 0.00001875, + "cache_read_input_token_cost": 0.0000015, + "input_cost_per_token": 0.000015, + "input_cost_per_token_batches": 0.0000075, "litellm_provider": "vertex_ai-anthropic_models", "max_input_tokens": 200000, "max_output_tokens": 32000, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 7.5e-5, - "output_cost_per_token_batches": 3.75e-5, + "output_cost_per_token": 0.000075, + "output_cost_per_token_batches": 0.0000375, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_tool_choice": true, "supports_vision": true }, "vertex_ai/claude-sonnet-4-5": { - "cache_creation_input_token_cost": 3.75e-6, + "cache_creation_input_token_cost": 0.00000375, "cache_read_input_token_cost": 3e-7, - "input_cost_per_token": 3e-6, - "input_cost_per_token_above_200k_tokens": 6e-6, - "output_cost_per_token_above_200k_tokens": 2.25e-5, - "cache_creation_input_token_cost_above_200k_tokens": 7.5e-6, + "input_cost_per_token": 0.000003, + "input_cost_per_token_above_200k_tokens": 0.000006, + "output_cost_per_token_above_200k_tokens": 0.0000225, + "cache_creation_input_token_cost_above_200k_tokens": 0.0000075, "cache_read_input_token_cost_above_200k_tokens": 6e-7, - "input_cost_per_token_batches": 1.5e-6, + "input_cost_per_token_batches": 0.0000015, "litellm_provider": "vertex_ai-anthropic_models", "max_input_tokens": 200000, "max_output_tokens": 64000, "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 1.5e-5, - "output_cost_per_token_batches": 7.5e-6, + "output_cost_per_token": 0.000015, + "output_cost_per_token_batches": 0.0000075, "supports_assistant_prefill": true, "supports_computer_use": true, "supports_function_calling": true, @@ -21690,21 +24498,21 @@ "supports_vision": true }, "vertex_ai/claude-sonnet-4-5@20250929": { - "cache_creation_input_token_cost": 3.75e-6, + "cache_creation_input_token_cost": 0.00000375, "cache_read_input_token_cost": 3e-7, - "input_cost_per_token": 3e-6, - "input_cost_per_token_above_200k_tokens": 6e-6, - "output_cost_per_token_above_200k_tokens": 2.25e-5, - "cache_creation_input_token_cost_above_200k_tokens": 7.5e-6, + "input_cost_per_token": 0.000003, + "input_cost_per_token_above_200k_tokens": 0.000006, + "output_cost_per_token_above_200k_tokens": 0.0000225, + "cache_creation_input_token_cost_above_200k_tokens": 0.0000075, "cache_read_input_token_cost_above_200k_tokens": 6e-7, - "input_cost_per_token_batches": 1.5e-6, + "input_cost_per_token_batches": 0.0000015, "litellm_provider": "vertex_ai-anthropic_models", "max_input_tokens": 200000, "max_output_tokens": 64000, "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 1.5e-5, - "output_cost_per_token_batches": 7.5e-6, + "output_cost_per_token": 0.000015, + "output_cost_per_token_batches": 0.0000075, "supports_assistant_prefill": true, "supports_computer_use": true, "supports_function_calling": true, @@ -21716,15 +24524,15 @@ "supports_vision": true }, "vertex_ai/claude-opus-4@20250514": { - "cache_creation_input_token_cost": 1.875e-5, - "cache_read_input_token_cost": 1.5e-6, - "input_cost_per_token": 1.5e-5, + "cache_creation_input_token_cost": 0.00001875, + "cache_read_input_token_cost": 0.0000015, + "input_cost_per_token": 0.000015, "litellm_provider": "vertex_ai-anthropic_models", "max_input_tokens": 200000, "max_output_tokens": 32000, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 7.5e-5, + "output_cost_per_token": 0.000075, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -21742,19 +24550,19 @@ "tool_use_system_prompt_tokens": 159 }, "vertex_ai/claude-sonnet-4": { - "cache_creation_input_token_cost": 3.75e-6, + "cache_creation_input_token_cost": 0.00000375, "cache_read_input_token_cost": 3e-7, - "input_cost_per_token": 3e-6, - "input_cost_per_token_above_200k_tokens": 6e-6, - "output_cost_per_token_above_200k_tokens": 2.25e-5, - "cache_creation_input_token_cost_above_200k_tokens": 7.5e-6, + "input_cost_per_token": 0.000003, + "input_cost_per_token_above_200k_tokens": 0.000006, + "output_cost_per_token_above_200k_tokens": 0.0000225, + "cache_creation_input_token_cost_above_200k_tokens": 0.0000075, "cache_read_input_token_cost_above_200k_tokens": 6e-7, "litellm_provider": "vertex_ai-anthropic_models", "max_input_tokens": 1000000, "max_output_tokens": 64000, "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -21772,19 +24580,19 @@ "tool_use_system_prompt_tokens": 159 }, "vertex_ai/claude-sonnet-4@20250514": { - "cache_creation_input_token_cost": 3.75e-6, + "cache_creation_input_token_cost": 0.00000375, "cache_read_input_token_cost": 3e-7, - "input_cost_per_token": 3e-6, - "input_cost_per_token_above_200k_tokens": 6e-6, - "output_cost_per_token_above_200k_tokens": 2.25e-5, - "cache_creation_input_token_cost_above_200k_tokens": 7.5e-6, + "input_cost_per_token": 0.000003, + "input_cost_per_token_above_200k_tokens": 0.000006, + "output_cost_per_token_above_200k_tokens": 0.0000225, + "cache_creation_input_token_cost_above_200k_tokens": 0.0000075, "cache_read_input_token_cost_above_200k_tokens": 6e-7, "litellm_provider": "vertex_ai-anthropic_models", "max_input_tokens": 1000000, "max_output_tokens": 64000, "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -21879,35 +24687,99 @@ "supports_tool_choice": true }, "vertex_ai/deepseek-ai/deepseek-v3.1-maas": { - "input_cost_per_token": 1.35e-6, + "input_cost_per_token": 0.00000135, "litellm_provider": "vertex_ai-deepseek_models", "max_input_tokens": 163840, "max_output_tokens": 32768, "max_tokens": 163840, "mode": "chat", - "output_cost_per_token": 5.4e-6, + "output_cost_per_token": 0.0000054, + "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing#partner-models", + "supported_regions": [ + "us-west2" + ], + "supports_assistant_prefill": true, + "supports_function_calling": true, + "supports_prompt_caching": true, + "supports_reasoning": true, + "supports_tool_choice": true + }, + "vertex_ai/deepseek-ai/deepseek-r1-0528-maas": { + "input_cost_per_token": 0.00000135, + "litellm_provider": "vertex_ai-deepseek_models", + "max_input_tokens": 65336, + "max_output_tokens": 8192, + "max_tokens": 8192, + "mode": "chat", + "output_cost_per_token": 0.0000054, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing#partner-models", - "supported_regions": ["us-west2"], "supports_assistant_prefill": true, "supports_function_calling": true, "supports_prompt_caching": true, - "supports_reasoning": true, - "supports_tool_choice": true + "supports_reasoning": true, + "supports_tool_choice": true + }, + "vertex_ai/gemini-2.5-flash-image": { + "cache_read_input_token_cost": 3e-8, + "input_cost_per_audio_token": 0.000001, + "input_cost_per_token": 3e-7, + "litellm_provider": "vertex_ai-language-models", + "max_audio_length_hours": 8.4, + "max_audio_per_prompt": 1, + "max_images_per_prompt": 3000, + "max_input_tokens": 32768, + "max_output_tokens": 32768, + "max_tokens": 32768, + "max_pdf_size_mb": 30, + "max_video_length": 1, + "max_videos_per_prompt": 10, + "mode": "image_generation", + "output_cost_per_image": 0.039, + "output_cost_per_reasoning_token": 0.0000025, + "output_cost_per_token": 0.0000025, + "rpm": 100000, + "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/multimodal/image-generation#edit-an-image", + "supported_endpoints": [ + "/v1/chat/completions", + "/v1/completions", + "/v1/batch" + ], + "supported_modalities": [ + "text", + "image", + "audio", + "video" + ], + "supported_output_modalities": [ + "text", + "image" + ], + "supports_audio_output": false, + "supports_function_calling": true, + "supports_parallel_function_calling": true, + "supports_pdf_input": true, + "supports_prompt_caching": true, + "supports_response_schema": true, + "supports_system_messages": true, + "supports_tool_choice": true, + "supports_url_context": true, + "supports_vision": true, + "supports_web_search": false, + "tpm": 8000000 }, - "vertex_ai/deepseek-ai/deepseek-r1-0528-maas": { - "input_cost_per_token": 1.35e-6, - "litellm_provider": "vertex_ai-deepseek_models", - "max_input_tokens": 65336, - "max_output_tokens": 8192, - "max_tokens": 8192, - "mode": "chat", - "output_cost_per_token": 5.4e-6, - "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing#partner-models", - "supports_assistant_prefill": true, - "supports_function_calling": true, - "supports_prompt_caching": true, - "supports_reasoning": true, - "supports_tool_choice": true + "vertex_ai/gemini-3-pro-image-preview": { + "input_cost_per_image": 0.0011, + "input_cost_per_token": 0.000002, + "input_cost_per_token_batches": 0.000001, + "litellm_provider": "vertex_ai-language-models", + "max_input_tokens": 65536, + "max_output_tokens": 32768, + "max_tokens": 65536, + "mode": "image_generation", + "output_cost_per_image": 0.134, + "output_cost_per_token": 0.000012, + "output_cost_per_token_batches": 0.000006, + "source": "https://docs.cloud.google.com/vertex-ai/generative-ai/docs/models/gemini/3-pro-image" }, "vertex_ai/imagegeneration@006": { "litellm_provider": "vertex_ai-image-models", @@ -21933,6 +24805,12 @@ "output_cost_per_image": 0.04, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing" }, + "vertex_ai/imagen-3.0-capability-001": { + "litellm_provider": "vertex_ai-image-models", + "mode": "image_generation", + "output_cost_per_image": 0.04, + "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/image/edit-insert-objects" + }, "vertex_ai/imagen-4.0-fast-generate-001": { "litellm_provider": "vertex_ai-image-models", "mode": "image_generation", @@ -21962,23 +24840,23 @@ "supports_tool_choice": true }, "vertex_ai/jamba-1.5-large": { - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "vertex_ai-ai21_models", "max_input_tokens": 256000, "max_output_tokens": 256000, "max_tokens": 256000, "mode": "chat", - "output_cost_per_token": 8e-6, + "output_cost_per_token": 0.000008, "supports_tool_choice": true }, "vertex_ai/jamba-1.5-large@001": { - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "vertex_ai-ai21_models", "max_input_tokens": 256000, "max_output_tokens": 256000, "max_tokens": 256000, "mode": "chat", - "output_cost_per_token": 8e-6, + "output_cost_per_token": 0.000008, "supports_tool_choice": true }, "vertex_ai/jamba-1.5-mini": { @@ -22002,33 +24880,33 @@ "supports_tool_choice": true }, "vertex_ai/meta/llama-3.1-405b-instruct-maas": { - "input_cost_per_token": 5e-6, + "input_cost_per_token": 0.000005, "litellm_provider": "vertex_ai-llama_models", "max_input_tokens": 128000, "max_output_tokens": 2048, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 1.6e-5, + "output_cost_per_token": 0.000016, "source": "https://console.cloud.google.com/vertex-ai/publishers/meta/model-garden/llama-3.2-90b-vision-instruct-maas", "supports_system_messages": true, "supports_tool_choice": true, "supports_vision": true }, "vertex_ai/meta/llama-3.1-70b-instruct-maas": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "vertex_ai-llama_models", "max_input_tokens": 128000, "max_output_tokens": 2048, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "source": "https://console.cloud.google.com/vertex-ai/publishers/meta/model-garden/llama-3.2-90b-vision-instruct-maas", "supports_system_messages": true, "supports_tool_choice": true, "supports_vision": true }, "vertex_ai/meta/llama-3.1-8b-instruct-maas": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "vertex_ai-llama_models", "max_input_tokens": 128000, "max_output_tokens": 2048, @@ -22037,14 +24915,14 @@ "notes": "VertexAI states that The Llama 3.1 API service for llama-3.1-70b-instruct-maas and llama-3.1-8b-instruct-maas are in public preview and at no cost." }, "mode": "chat", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "source": "https://console.cloud.google.com/vertex-ai/publishers/meta/model-garden/llama-3.2-90b-vision-instruct-maas", "supports_system_messages": true, "supports_tool_choice": true, "supports_vision": true }, "vertex_ai/meta/llama-3.2-90b-vision-instruct-maas": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "vertex_ai-llama_models", "max_input_tokens": 128000, "max_output_tokens": 2048, @@ -22053,7 +24931,7 @@ "notes": "VertexAI states that The Llama 3.2 API service is at no cost during public preview, and will be priced as per dollar-per-1M-tokens at GA." }, "mode": "chat", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "source": "https://console.cloud.google.com/vertex-ai/publishers/meta/model-garden/llama-3.2-90b-vision-instruct-maas", "supports_system_messages": true, "supports_tool_choice": true, @@ -22066,10 +24944,16 @@ "max_output_tokens": 1000000, "max_tokens": 1000000, "mode": "chat", - "output_cost_per_token": 1.15e-6, + "output_cost_per_token": 0.00000115, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing#partner-models", - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text", "code"], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text", + "code" + ], "supports_function_calling": true, "supports_tool_choice": true }, @@ -22080,10 +24964,16 @@ "max_output_tokens": 1000000, "max_tokens": 1000000, "mode": "chat", - "output_cost_per_token": 1.15e-6, + "output_cost_per_token": 0.00000115, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing#partner-models", - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text", "code"], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text", + "code" + ], "supports_function_calling": true, "supports_tool_choice": true }, @@ -22096,8 +24986,14 @@ "mode": "chat", "output_cost_per_token": 7e-7, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing#partner-models", - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text", "code"], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text", + "code" + ], "supports_function_calling": true, "supports_tool_choice": true }, @@ -22110,41 +25006,47 @@ "mode": "chat", "output_cost_per_token": 7e-7, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing#partner-models", - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["text", "code"], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text", + "code" + ], "supports_function_calling": true, "supports_tool_choice": true }, "vertex_ai/meta/llama3-405b-instruct-maas": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "vertex_ai-llama_models", "max_input_tokens": 32000, "max_output_tokens": 32000, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing#partner-models", "supports_tool_choice": true }, "vertex_ai/meta/llama3-70b-instruct-maas": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "vertex_ai-llama_models", "max_input_tokens": 32000, "max_output_tokens": 32000, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing#partner-models", "supports_tool_choice": true }, "vertex_ai/meta/llama3-8b-instruct-maas": { - "input_cost_per_token": 0.0, + "input_cost_per_token": 0, "litellm_provider": "vertex_ai-llama_models", "max_input_tokens": 32000, "max_output_tokens": 32000, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 0.0, + "output_cost_per_token": 0, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing#partner-models", "supports_tool_choice": true }, @@ -22155,7 +25057,7 @@ "max_output_tokens": 196608, "max_tokens": 196608, "mode": "chat", - "output_cost_per_token": 1.2e-6, + "output_cost_per_token": 0.0000012, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing#partner-models", "supports_function_calling": true, "supports_tool_choice": true @@ -22167,7 +25069,7 @@ "max_output_tokens": 256000, "max_tokens": 256000, "mode": "chat", - "output_cost_per_token": 2.5e-6, + "output_cost_per_token": 0.0000025, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing#partner-models", "supports_function_calling": true, "supports_tool_choice": true, @@ -22180,7 +25082,7 @@ "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 2e-6, + "output_cost_per_token": 0.000002, "supports_function_calling": true, "supports_tool_choice": true }, @@ -22191,7 +25093,7 @@ "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 2e-6, + "output_cost_per_token": 0.000002, "supports_function_calling": true, "supports_tool_choice": true }, @@ -22202,7 +25104,7 @@ "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 2e-6, + "output_cost_per_token": 0.000002, "supports_function_calling": true, "supports_tool_choice": true }, @@ -22213,62 +25115,62 @@ "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 2e-6, + "output_cost_per_token": 0.000002, "supports_function_calling": true, "supports_tool_choice": true }, "vertex_ai/mistral-large-2411": { - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "vertex_ai-mistral_models", "max_input_tokens": 128000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 6e-6, + "output_cost_per_token": 0.000006, "supports_function_calling": true, "supports_tool_choice": true }, "vertex_ai/mistral-large@2407": { - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "vertex_ai-mistral_models", "max_input_tokens": 128000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 6e-6, + "output_cost_per_token": 0.000006, "supports_function_calling": true, "supports_tool_choice": true }, "vertex_ai/mistral-large@2411-001": { - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "vertex_ai-mistral_models", "max_input_tokens": 128000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 6e-6, + "output_cost_per_token": 0.000006, "supports_function_calling": true, "supports_tool_choice": true }, "vertex_ai/mistral-large@latest": { - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "vertex_ai-mistral_models", "max_input_tokens": 128000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 6e-6, + "output_cost_per_token": 0.000006, "supports_function_calling": true, "supports_tool_choice": true }, "vertex_ai/mistral-nemo@2407": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "vertex_ai-mistral_models", "max_input_tokens": 128000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 3e-6, + "output_cost_per_token": 0.000003, "supports_function_calling": true, "supports_tool_choice": true }, @@ -22284,33 +25186,35 @@ "supports_tool_choice": true }, "vertex_ai/mistral-small-2503": { - "input_cost_per_token": 1e-6, + "input_cost_per_token": 0.000001, "litellm_provider": "vertex_ai-mistral_models", "max_input_tokens": 128000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 3e-6, + "output_cost_per_token": 0.000003, "supports_function_calling": true, "supports_tool_choice": true, "supports_vision": true }, "vertex_ai/mistral-small-2503@001": { - "input_cost_per_token": 1e-6, + "input_cost_per_token": 0.000001, "litellm_provider": "vertex_ai-mistral_models", "max_input_tokens": 32000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 3e-6, + "output_cost_per_token": 0.000003, "supports_function_calling": true, "supports_tool_choice": true }, "vertex_ai/mistral-ocr-2505": { "litellm_provider": "vertex_ai", "mode": "ocr", - "ocr_cost_per_page": 5e-4, - "supported_endpoints": ["/v1/ocr"], + "ocr_cost_per_page": 0.0005, + "supported_endpoints": [ + "/v1/ocr" + ], "source": "https://cloud.google.com/generative-ai-app-builder/pricing" }, "vertex_ai/openai/gpt-oss-120b-maas": { @@ -22342,19 +25246,19 @@ "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 1e-6, + "output_cost_per_token": 0.000001, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing", "supports_function_calling": true, "supports_tool_choice": true }, "vertex_ai/qwen/qwen3-coder-480b-a35b-instruct-maas": { - "input_cost_per_token": 1e-6, + "input_cost_per_token": 0.000001, "litellm_provider": "vertex_ai-qwen_models", "max_input_tokens": 262144, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 4e-6, + "output_cost_per_token": 0.000004, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing", "supports_function_calling": true, "supports_tool_choice": true @@ -22366,7 +25270,7 @@ "max_output_tokens": 262144, "max_tokens": 262144, "mode": "chat", - "output_cost_per_token": 1.2e-6, + "output_cost_per_token": 0.0000012, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing", "supports_function_calling": true, "supports_tool_choice": true @@ -22378,7 +25282,7 @@ "max_output_tokens": 262144, "max_tokens": 262144, "mode": "chat", - "output_cost_per_token": 1.2e-6, + "output_cost_per_token": 0.0000012, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing", "supports_function_calling": true, "supports_tool_choice": true @@ -22390,28 +25294,68 @@ "mode": "video_generation", "output_cost_per_second": 0.35, "source": "https://ai.google.dev/gemini-api/docs/video", - "supported_modalities": ["text"], - "supported_output_modalities": ["video"] + "supported_modalities": [ + "text" + ], + "supported_output_modalities": [ + "video" + ] }, "vertex_ai/veo-3.0-fast-generate-preview": { "litellm_provider": "vertex_ai-video-models", "max_input_tokens": 1024, "max_tokens": 1024, "mode": "video_generation", - "output_cost_per_second": 0.4, + "output_cost_per_second": 0.15, "source": "https://ai.google.dev/gemini-api/docs/video", - "supported_modalities": ["text"], - "supported_output_modalities": ["video"] + "supported_modalities": [ + "text" + ], + "supported_output_modalities": [ + "video" + ] }, "vertex_ai/veo-3.0-generate-preview": { "litellm_provider": "vertex_ai-video-models", "max_input_tokens": 1024, "max_tokens": 1024, "mode": "video_generation", - "output_cost_per_second": 0.75, + "output_cost_per_second": 0.4, + "source": "https://ai.google.dev/gemini-api/docs/video", + "supported_modalities": [ + "text" + ], + "supported_output_modalities": [ + "video" + ] + }, + "vertex_ai/veo-3.0-fast-generate-001": { + "litellm_provider": "vertex_ai-video-models", + "max_input_tokens": 1024, + "max_tokens": 1024, + "mode": "video_generation", + "output_cost_per_second": 0.15, + "source": "https://ai.google.dev/gemini-api/docs/video", + "supported_modalities": [ + "text" + ], + "supported_output_modalities": [ + "video" + ] + }, + "vertex_ai/veo-3.0-generate-001": { + "litellm_provider": "vertex_ai-video-models", + "max_input_tokens": 1024, + "max_tokens": 1024, + "mode": "video_generation", + "output_cost_per_second": 0.4, "source": "https://ai.google.dev/gemini-api/docs/video", - "supported_modalities": ["text"], - "supported_output_modalities": ["video"] + "supported_modalities": [ + "text" + ], + "supported_output_modalities": [ + "video" + ] }, "vertex_ai/veo-3.1-generate-preview": { "litellm_provider": "vertex_ai-video-models", @@ -22420,8 +25364,12 @@ "mode": "video_generation", "output_cost_per_second": 0.4, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/model-reference/veo", - "supported_modalities": ["text"], - "supported_output_modalities": ["video"] + "supported_modalities": [ + "text" + ], + "supported_output_modalities": [ + "video" + ] }, "vertex_ai/veo-3.1-fast-generate-preview": { "litellm_provider": "vertex_ai-video-models", @@ -22430,8 +25378,12 @@ "mode": "video_generation", "output_cost_per_second": 0.15, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/model-reference/veo", - "supported_modalities": ["text"], - "supported_output_modalities": ["video"] + "supported_modalities": [ + "text" + ], + "supported_output_modalities": [ + "video" + ] }, "voyage/rerank-2": { "input_cost_per_query": 5e-8, @@ -22442,7 +25394,7 @@ "max_query_tokens": 16000, "max_tokens": 16000, "mode": "rerank", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "voyage/rerank-2-lite": { "input_cost_per_query": 2e-8, @@ -22453,7 +25405,7 @@ "max_query_tokens": 8000, "max_tokens": 8000, "mode": "rerank", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "voyage/voyage-2": { "input_cost_per_token": 1e-7, @@ -22461,7 +25413,7 @@ "max_input_tokens": 4000, "max_tokens": 4000, "mode": "embedding", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "voyage/voyage-3": { "input_cost_per_token": 6e-8, @@ -22469,7 +25421,7 @@ "max_input_tokens": 32000, "max_tokens": 32000, "mode": "embedding", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "voyage/voyage-3-large": { "input_cost_per_token": 1.8e-7, @@ -22477,7 +25429,7 @@ "max_input_tokens": 32000, "max_tokens": 32000, "mode": "embedding", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "voyage/voyage-3-lite": { "input_cost_per_token": 2e-8, @@ -22485,7 +25437,7 @@ "max_input_tokens": 32000, "max_tokens": 32000, "mode": "embedding", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "voyage/voyage-3.5": { "input_cost_per_token": 6e-8, @@ -22493,7 +25445,7 @@ "max_input_tokens": 32000, "max_tokens": 32000, "mode": "embedding", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "voyage/voyage-3.5-lite": { "input_cost_per_token": 2e-8, @@ -22501,7 +25453,7 @@ "max_input_tokens": 32000, "max_tokens": 32000, "mode": "embedding", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "voyage/voyage-code-2": { "input_cost_per_token": 1.2e-7, @@ -22509,7 +25461,7 @@ "max_input_tokens": 16000, "max_tokens": 16000, "mode": "embedding", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "voyage/voyage-code-3": { "input_cost_per_token": 1.8e-7, @@ -22517,7 +25469,7 @@ "max_input_tokens": 32000, "max_tokens": 32000, "mode": "embedding", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "voyage/voyage-context-3": { "input_cost_per_token": 1.8e-7, @@ -22525,7 +25477,7 @@ "max_input_tokens": 120000, "max_tokens": 120000, "mode": "embedding", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "voyage/voyage-finance-2": { "input_cost_per_token": 1.2e-7, @@ -22533,7 +25485,7 @@ "max_input_tokens": 32000, "max_tokens": 32000, "mode": "embedding", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "voyage/voyage-large-2": { "input_cost_per_token": 1.2e-7, @@ -22541,7 +25493,7 @@ "max_input_tokens": 16000, "max_tokens": 16000, "mode": "embedding", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "voyage/voyage-law-2": { "input_cost_per_token": 1.2e-7, @@ -22549,7 +25501,7 @@ "max_input_tokens": 16000, "max_tokens": 16000, "mode": "embedding", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "voyage/voyage-lite-01": { "input_cost_per_token": 1e-7, @@ -22557,7 +25509,7 @@ "max_input_tokens": 4096, "max_tokens": 4096, "mode": "embedding", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "voyage/voyage-lite-02-instruct": { "input_cost_per_token": 1e-7, @@ -22565,7 +25517,7 @@ "max_input_tokens": 4000, "max_tokens": 4000, "mode": "embedding", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "voyage/voyage-multimodal-3": { "input_cost_per_token": 1.2e-7, @@ -22573,7 +25525,7 @@ "max_input_tokens": 32000, "max_tokens": 32000, "mode": "embedding", - "output_cost_per_token": 0.0 + "output_cost_per_token": 0 }, "wandb/openai/gpt-oss-120b": { "max_tokens": 131072, @@ -22702,13 +25654,13 @@ "mode": "chat" }, "watsonx/ibm/granite-3-8b-instruct": { - "input_cost_per_token": 0.2e-6, + "input_cost_per_token": 2e-7, "litellm_provider": "watsonx", "max_input_tokens": 8192, "max_output_tokens": 1024, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.2e-6, + "output_cost_per_token": 2e-7, "supports_audio_input": false, "supports_audio_output": false, "supports_function_calling": true, @@ -22720,13 +25672,13 @@ "supports_vision": false }, "watsonx/mistralai/mistral-large": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "watsonx", "max_input_tokens": 131072, "max_output_tokens": 16384, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 10e-6, + "output_cost_per_token": 0.00001, "supports_audio_input": false, "supports_audio_output": false, "supports_function_calling": true, @@ -22765,8 +25717,8 @@ "max_tokens": 8192, "max_input_tokens": 8192, "max_output_tokens": 8192, - "input_cost_per_token": 0.6e-6, - "output_cost_per_token": 0.6e-6, + "input_cost_per_token": 6e-7, + "output_cost_per_token": 6e-7, "litellm_provider": "watsonx", "mode": "chat", "supports_function_calling": false, @@ -22777,8 +25729,8 @@ "max_tokens": 8192, "max_input_tokens": 8192, "max_output_tokens": 8192, - "input_cost_per_token": 0.6e-6, - "output_cost_per_token": 0.6e-6, + "input_cost_per_token": 6e-7, + "output_cost_per_token": 6e-7, "litellm_provider": "watsonx", "mode": "chat", "supports_function_calling": false, @@ -22789,8 +25741,8 @@ "max_tokens": 8192, "max_input_tokens": 8192, "max_output_tokens": 8192, - "input_cost_per_token": 0.6e-6, - "output_cost_per_token": 0.6e-6, + "input_cost_per_token": 6e-7, + "output_cost_per_token": 6e-7, "litellm_provider": "watsonx", "mode": "chat", "supports_function_calling": false, @@ -22801,8 +25753,8 @@ "max_tokens": 8192, "max_input_tokens": 8192, "max_output_tokens": 8192, - "input_cost_per_token": 0.2e-6, - "output_cost_per_token": 0.2e-6, + "input_cost_per_token": 2e-7, + "output_cost_per_token": 2e-7, "litellm_provider": "watsonx", "mode": "chat", "supports_function_calling": true, @@ -22813,8 +25765,8 @@ "max_tokens": 20480, "max_input_tokens": 20480, "max_output_tokens": 20480, - "input_cost_per_token": 0.06e-6, - "output_cost_per_token": 0.25e-6, + "input_cost_per_token": 6e-8, + "output_cost_per_token": 2.5e-7, "litellm_provider": "watsonx", "mode": "chat", "supports_function_calling": true, @@ -22825,8 +25777,8 @@ "max_tokens": 8192, "max_input_tokens": 8192, "max_output_tokens": 8192, - "input_cost_per_token": 0.1e-6, - "output_cost_per_token": 0.1e-6, + "input_cost_per_token": 1e-7, + "output_cost_per_token": 1e-7, "litellm_provider": "watsonx", "mode": "chat", "supports_function_calling": false, @@ -22837,8 +25789,8 @@ "max_tokens": 8192, "max_input_tokens": 8192, "max_output_tokens": 8192, - "input_cost_per_token": 0.2e-6, - "output_cost_per_token": 0.2e-6, + "input_cost_per_token": 2e-7, + "output_cost_per_token": 2e-7, "litellm_provider": "watsonx", "mode": "chat", "supports_function_calling": false, @@ -22849,8 +25801,8 @@ "max_tokens": 512, "max_input_tokens": 512, "max_output_tokens": 512, - "input_cost_per_token": 0.38e-6, - "output_cost_per_token": 0.38e-6, + "input_cost_per_token": 3.8e-7, + "output_cost_per_token": 3.8e-7, "litellm_provider": "watsonx", "mode": "chat", "supports_function_calling": false, @@ -22861,8 +25813,8 @@ "max_tokens": 512, "max_input_tokens": 512, "max_output_tokens": 512, - "input_cost_per_token": 0.38e-6, - "output_cost_per_token": 0.38e-6, + "input_cost_per_token": 3.8e-7, + "output_cost_per_token": 3.8e-7, "litellm_provider": "watsonx", "mode": "chat", "supports_function_calling": false, @@ -22873,8 +25825,8 @@ "max_tokens": 512, "max_input_tokens": 512, "max_output_tokens": 512, - "input_cost_per_token": 0.38e-6, - "output_cost_per_token": 0.38e-6, + "input_cost_per_token": 3.8e-7, + "output_cost_per_token": 3.8e-7, "litellm_provider": "watsonx", "mode": "chat", "supports_function_calling": false, @@ -22885,8 +25837,8 @@ "max_tokens": 8192, "max_input_tokens": 8192, "max_output_tokens": 8192, - "input_cost_per_token": 0.1e-6, - "output_cost_per_token": 0.1e-6, + "input_cost_per_token": 1e-7, + "output_cost_per_token": 1e-7, "litellm_provider": "watsonx", "mode": "chat", "supports_function_calling": false, @@ -22897,8 +25849,8 @@ "max_tokens": 128000, "max_input_tokens": 128000, "max_output_tokens": 128000, - "input_cost_per_token": 0.35e-6, - "output_cost_per_token": 0.35e-6, + "input_cost_per_token": 3.5e-7, + "output_cost_per_token": 3.5e-7, "litellm_provider": "watsonx", "mode": "chat", "supports_function_calling": true, @@ -22909,8 +25861,8 @@ "max_tokens": 128000, "max_input_tokens": 128000, "max_output_tokens": 128000, - "input_cost_per_token": 0.1e-6, - "output_cost_per_token": 0.1e-6, + "input_cost_per_token": 1e-7, + "output_cost_per_token": 1e-7, "litellm_provider": "watsonx", "mode": "chat", "supports_function_calling": true, @@ -22921,8 +25873,8 @@ "max_tokens": 128000, "max_input_tokens": 128000, "max_output_tokens": 128000, - "input_cost_per_token": 0.15e-6, - "output_cost_per_token": 0.15e-6, + "input_cost_per_token": 1.5e-7, + "output_cost_per_token": 1.5e-7, "litellm_provider": "watsonx", "mode": "chat", "supports_function_calling": true, @@ -22933,8 +25885,8 @@ "max_tokens": 128000, "max_input_tokens": 128000, "max_output_tokens": 128000, - "input_cost_per_token": 2e-6, - "output_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, + "output_cost_per_token": 0.000002, "litellm_provider": "watsonx", "mode": "chat", "supports_function_calling": true, @@ -22945,8 +25897,8 @@ "max_tokens": 128000, "max_input_tokens": 128000, "max_output_tokens": 128000, - "input_cost_per_token": 0.71e-6, - "output_cost_per_token": 0.71e-6, + "input_cost_per_token": 7.1e-7, + "output_cost_per_token": 7.1e-7, "litellm_provider": "watsonx", "mode": "chat", "supports_function_calling": true, @@ -22957,8 +25909,8 @@ "max_tokens": 128000, "max_input_tokens": 128000, "max_output_tokens": 128000, - "input_cost_per_token": 0.35e-6, - "output_cost_per_token": 1.4e-6, + "input_cost_per_token": 3.5e-7, + "output_cost_per_token": 0.0000014, "litellm_provider": "watsonx", "mode": "chat", "supports_function_calling": true, @@ -22969,8 +25921,8 @@ "max_tokens": 128000, "max_input_tokens": 128000, "max_output_tokens": 128000, - "input_cost_per_token": 0.35e-6, - "output_cost_per_token": 0.35e-6, + "input_cost_per_token": 3.5e-7, + "output_cost_per_token": 3.5e-7, "litellm_provider": "watsonx", "mode": "chat", "supports_function_calling": false, @@ -22981,8 +25933,8 @@ "max_tokens": 128000, "max_input_tokens": 128000, "max_output_tokens": 128000, - "input_cost_per_token": 3e-6, - "output_cost_per_token": 10e-6, + "input_cost_per_token": 0.000003, + "output_cost_per_token": 0.00001, "litellm_provider": "watsonx", "mode": "chat", "supports_function_calling": true, @@ -22993,8 +25945,8 @@ "max_tokens": 32000, "max_input_tokens": 32000, "max_output_tokens": 32000, - "input_cost_per_token": 0.1e-6, - "output_cost_per_token": 0.3e-6, + "input_cost_per_token": 1e-7, + "output_cost_per_token": 3e-7, "litellm_provider": "watsonx", "mode": "chat", "supports_function_calling": true, @@ -23005,8 +25957,8 @@ "max_tokens": 32000, "max_input_tokens": 32000, "max_output_tokens": 32000, - "input_cost_per_token": 0.1e-6, - "output_cost_per_token": 0.3e-6, + "input_cost_per_token": 1e-7, + "output_cost_per_token": 3e-7, "litellm_provider": "watsonx", "mode": "chat", "supports_function_calling": true, @@ -23017,8 +25969,8 @@ "max_tokens": 128000, "max_input_tokens": 128000, "max_output_tokens": 128000, - "input_cost_per_token": 0.35e-6, - "output_cost_per_token": 0.35e-6, + "input_cost_per_token": 3.5e-7, + "output_cost_per_token": 3.5e-7, "litellm_provider": "watsonx", "mode": "chat", "supports_function_calling": false, @@ -23029,8 +25981,8 @@ "max_tokens": 8192, "max_input_tokens": 8192, "max_output_tokens": 8192, - "input_cost_per_token": 0.15e-6, - "output_cost_per_token": 0.6e-6, + "input_cost_per_token": 1.5e-7, + "output_cost_per_token": 6e-7, "litellm_provider": "watsonx", "mode": "chat", "supports_function_calling": false, @@ -23041,8 +25993,8 @@ "max_tokens": 8192, "max_input_tokens": 8192, "max_output_tokens": 8192, - "input_cost_per_token": 1.8e-6, - "output_cost_per_token": 1.8e-6, + "input_cost_per_token": 0.0000018, + "output_cost_per_token": 0.0000018, "litellm_provider": "watsonx", "mode": "chat", "supports_function_calling": false, @@ -23054,94 +26006,96 @@ "litellm_provider": "openai", "mode": "audio_transcription", "output_cost_per_second": 0.0001, - "supported_endpoints": ["/v1/audio/transcriptions"] + "supported_endpoints": [ + "/v1/audio/transcriptions" + ] }, "xai/grok-2": { - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "xai", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 1e-5, + "output_cost_per_token": 0.00001, "supports_function_calling": true, "supports_tool_choice": true, "supports_web_search": true }, "xai/grok-2-1212": { - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "xai", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 1e-5, + "output_cost_per_token": 0.00001, "supports_function_calling": true, "supports_tool_choice": true, "supports_web_search": true }, "xai/grok-2-latest": { - "input_cost_per_token": 2e-6, + "input_cost_per_token": 0.000002, "litellm_provider": "xai", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 1e-5, + "output_cost_per_token": 0.00001, "supports_function_calling": true, "supports_tool_choice": true, "supports_web_search": true }, "xai/grok-2-vision": { - "input_cost_per_image": 2e-6, - "input_cost_per_token": 2e-6, + "input_cost_per_image": 0.000002, + "input_cost_per_token": 0.000002, "litellm_provider": "xai", "max_input_tokens": 32768, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 1e-5, + "output_cost_per_token": 0.00001, "supports_function_calling": true, "supports_tool_choice": true, "supports_vision": true, "supports_web_search": true }, "xai/grok-2-vision-1212": { - "input_cost_per_image": 2e-6, - "input_cost_per_token": 2e-6, + "input_cost_per_image": 0.000002, + "input_cost_per_token": 0.000002, "litellm_provider": "xai", "max_input_tokens": 32768, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 1e-5, + "output_cost_per_token": 0.00001, "supports_function_calling": true, "supports_tool_choice": true, "supports_vision": true, "supports_web_search": true }, "xai/grok-2-vision-latest": { - "input_cost_per_image": 2e-6, - "input_cost_per_token": 2e-6, + "input_cost_per_image": 0.000002, + "input_cost_per_token": 0.000002, "litellm_provider": "xai", "max_input_tokens": 32768, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 1e-5, + "output_cost_per_token": 0.00001, "supports_function_calling": true, "supports_tool_choice": true, "supports_vision": true, "supports_web_search": true }, "xai/grok-3": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "xai", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "source": "https://x.ai/api#pricing", "supports_function_calling": true, "supports_response_schema": false, @@ -23149,13 +26103,13 @@ "supports_web_search": true }, "xai/grok-3-beta": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "xai", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "source": "https://x.ai/api#pricing", "supports_function_calling": true, "supports_response_schema": false, @@ -23163,13 +26117,13 @@ "supports_web_search": true }, "xai/grok-3-fast-beta": { - "input_cost_per_token": 5e-6, + "input_cost_per_token": 0.000005, "litellm_provider": "xai", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 2.5e-5, + "output_cost_per_token": 0.000025, "source": "https://x.ai/api#pricing", "supports_function_calling": true, "supports_response_schema": false, @@ -23177,13 +26131,13 @@ "supports_web_search": true }, "xai/grok-3-fast-latest": { - "input_cost_per_token": 5e-6, + "input_cost_per_token": 0.000005, "litellm_provider": "xai", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 2.5e-5, + "output_cost_per_token": 0.000025, "source": "https://x.ai/api#pricing", "supports_function_calling": true, "supports_response_schema": false, @@ -23191,13 +26145,13 @@ "supports_web_search": true }, "xai/grok-3-latest": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "xai", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "source": "https://x.ai/api#pricing", "supports_function_calling": true, "supports_response_schema": false, @@ -23241,7 +26195,7 @@ "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 4e-6, + "output_cost_per_token": 0.000004, "source": "https://x.ai/api#pricing", "supports_function_calling": true, "supports_reasoning": true, @@ -23256,7 +26210,7 @@ "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 4e-6, + "output_cost_per_token": 0.000004, "source": "https://x.ai/api#pricing", "supports_function_calling": true, "supports_reasoning": true, @@ -23271,7 +26225,7 @@ "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 4e-6, + "output_cost_per_token": 0.000004, "source": "https://x.ai/api#pricing", "supports_function_calling": true, "supports_reasoning": true, @@ -23295,13 +26249,13 @@ "supports_web_search": true }, "xai/grok-4": { - "input_cost_per_token": 3e-6, + "input_cost_per_token": 0.000003, "litellm_provider": "xai", "max_input_tokens": 256000, "max_output_tokens": 256000, "max_tokens": 256000, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "source": "https://docs.x.ai/docs/models", "supports_function_calling": true, "supports_tool_choice": true, @@ -23309,107 +26263,172 @@ }, "xai/grok-4-fast-reasoning": { "litellm_provider": "xai", - "max_input_tokens": 2e6, - "max_output_tokens": 2e6, - "max_tokens": 2e6, - "mode": "chat", - "input_cost_per_token": 0.2e-6, - "input_cost_per_token_above_128k_tokens": 0.4e-6, - "output_cost_per_token": 0.5e-6, - "output_cost_per_token_above_128k_tokens": 1e-6, - "cache_read_input_token_cost": 0.05e-6, + "max_input_tokens": 2000000, + "max_output_tokens": 2000000, + "max_tokens": 2000000, + "mode": "chat", + "input_cost_per_token": 2e-7, + "input_cost_per_token_above_128k_tokens": 4e-7, + "output_cost_per_token": 5e-7, + "output_cost_per_token_above_128k_tokens": 0.000001, + "cache_read_input_token_cost": 5e-8, "source": "https://docs.x.ai/docs/models", "supports_function_calling": true, "supports_tool_choice": true, "supports_web_search": true }, - - "xai/grok-4-1-fast-reasoning": { - "litellm_provider": "xai", - "max_input_tokens": 2e6, - "max_output_tokens": 2e6, - "max_tokens": 2e6, - "mode": "chat", - "input_cost_per_token": 0.2e-6, - "input_cost_per_token_above_128k_tokens": 0.4e-6, - "output_cost_per_token": 0.5e-6, - "output_cost_per_token_above_128k_tokens": 1e-6, - "cache_read_input_token_cost": 0.05e-6, - "source": "https://docs.x.ai/docs/models/grok-4-1-fast-reasoning", - "supports_function_calling": true, - "supports_tool_choice": true, - "supports_web_search": true - }, - "xai/grok-4-1-fast": { - "litellm_provider": "xai", - "max_input_tokens": 2e6, - "max_output_tokens": 2e6, - "cache_read_input_token_cost": 0.05e-6, - "max_tokens": 2e6, - "mode": "chat", - "input_cost_per_token": 0.2e-6, - "input_cost_per_token_above_128k_tokens": 0.4e-6, - "output_cost_per_token": 0.5e-6, - "output_cost_per_token_above_128k_tokens": 1e-6, - "source": "https://docs.x.ai/docs/models/grok-4-1-fast-reasoning", - "supports_function_calling": true, - "supports_tool_choice": true, - "supports_web_search": true - }, "xai/grok-4-fast-non-reasoning": { "litellm_provider": "xai", - "max_input_tokens": 2e6, - "max_output_tokens": 2e6, - "cache_read_input_token_cost": 0.05e-6, - "max_tokens": 2e6, - "mode": "chat", - "input_cost_per_token": 0.2e-6, - "input_cost_per_token_above_128k_tokens": 0.4e-6, - "output_cost_per_token": 0.5e-6, - "output_cost_per_token_above_128k_tokens": 1e-6, + "max_input_tokens": 2000000, + "max_output_tokens": 2000000, + "cache_read_input_token_cost": 5e-8, + "max_tokens": 2000000, + "mode": "chat", + "input_cost_per_token": 2e-7, + "input_cost_per_token_above_128k_tokens": 4e-7, + "output_cost_per_token": 5e-7, + "output_cost_per_token_above_128k_tokens": 0.000001, "source": "https://docs.x.ai/docs/models", "supports_function_calling": true, "supports_tool_choice": true, "supports_web_search": true }, "xai/grok-4-0709": { - "input_cost_per_token": 3e-6, - "input_cost_per_token_above_128k_tokens": 6e-6, + "input_cost_per_token": 0.000003, + "input_cost_per_token_above_128k_tokens": 0.000006, "litellm_provider": "xai", "max_input_tokens": 256000, "max_output_tokens": 256000, "max_tokens": 256000, "mode": "chat", - "output_cost_per_token": 1.5e-5, - "output_cost_per_token_above_128k_tokens": 30e-6, + "output_cost_per_token": 0.000015, + "output_cost_per_token_above_128k_tokens": 0.00003, "source": "https://docs.x.ai/docs/models", "supports_function_calling": true, "supports_tool_choice": true, "supports_web_search": true }, "xai/grok-4-latest": { - "input_cost_per_token": 3e-6, - "input_cost_per_token_above_128k_tokens": 6e-6, + "input_cost_per_token": 0.000003, + "input_cost_per_token_above_128k_tokens": 0.000006, "litellm_provider": "xai", "max_input_tokens": 256000, "max_output_tokens": 256000, "max_tokens": 256000, "mode": "chat", - "output_cost_per_token": 1.5e-5, - "output_cost_per_token_above_128k_tokens": 30e-6, + "output_cost_per_token": 0.000015, + "output_cost_per_token_above_128k_tokens": 0.00003, "source": "https://docs.x.ai/docs/models", "supports_function_calling": true, "supports_tool_choice": true, "supports_web_search": true }, + "xai/grok-4-1-fast": { + "cache_read_input_token_cost": 5e-8, + "input_cost_per_token": 2e-7, + "input_cost_per_token_above_128k_tokens": 4e-7, + "litellm_provider": "xai", + "max_input_tokens": 2000000, + "max_output_tokens": 2000000, + "max_tokens": 2000000, + "mode": "chat", + "output_cost_per_token": 5e-7, + "output_cost_per_token_above_128k_tokens": 0.000001, + "source": "https://docs.x.ai/docs/models/grok-4-1-fast-reasoning", + "supports_audio_input": true, + "supports_function_calling": true, + "supports_reasoning": true, + "supports_response_schema": true, + "supports_tool_choice": true, + "supports_vision": true, + "supports_web_search": true + }, + "xai/grok-4-1-fast-reasoning": { + "cache_read_input_token_cost": 5e-8, + "input_cost_per_token": 2e-7, + "input_cost_per_token_above_128k_tokens": 4e-7, + "litellm_provider": "xai", + "max_input_tokens": 2000000, + "max_output_tokens": 2000000, + "max_tokens": 2000000, + "mode": "chat", + "output_cost_per_token": 5e-7, + "output_cost_per_token_above_128k_tokens": 0.000001, + "source": "https://docs.x.ai/docs/models/grok-4-1-fast-reasoning", + "supports_audio_input": true, + "supports_function_calling": true, + "supports_reasoning": true, + "supports_response_schema": true, + "supports_tool_choice": true, + "supports_vision": true, + "supports_web_search": true + }, + "xai/grok-4-1-fast-reasoning-latest": { + "cache_read_input_token_cost": 5e-8, + "input_cost_per_token": 2e-7, + "input_cost_per_token_above_128k_tokens": 4e-7, + "litellm_provider": "xai", + "max_input_tokens": 2000000, + "max_output_tokens": 2000000, + "max_tokens": 2000000, + "mode": "chat", + "output_cost_per_token": 5e-7, + "output_cost_per_token_above_128k_tokens": 0.000001, + "source": "https://docs.x.ai/docs/models/grok-4-1-fast-reasoning", + "supports_audio_input": true, + "supports_function_calling": true, + "supports_reasoning": true, + "supports_response_schema": true, + "supports_tool_choice": true, + "supports_vision": true, + "supports_web_search": true + }, + "xai/grok-4-1-fast-non-reasoning": { + "cache_read_input_token_cost": 5e-8, + "input_cost_per_token": 2e-7, + "input_cost_per_token_above_128k_tokens": 4e-7, + "litellm_provider": "xai", + "max_input_tokens": 2000000, + "max_output_tokens": 2000000, + "max_tokens": 2000000, + "mode": "chat", + "output_cost_per_token": 5e-7, + "output_cost_per_token_above_128k_tokens": 0.000001, + "source": "https://docs.x.ai/docs/models/grok-4-1-fast-non-reasoning", + "supports_audio_input": true, + "supports_function_calling": true, + "supports_response_schema": true, + "supports_tool_choice": true, + "supports_vision": true, + "supports_web_search": true + }, + "xai/grok-4-1-fast-non-reasoning-latest": { + "cache_read_input_token_cost": 5e-8, + "input_cost_per_token": 2e-7, + "input_cost_per_token_above_128k_tokens": 4e-7, + "litellm_provider": "xai", + "max_input_tokens": 2000000, + "max_output_tokens": 2000000, + "max_tokens": 2000000, + "mode": "chat", + "output_cost_per_token": 5e-7, + "output_cost_per_token_above_128k_tokens": 0.000001, + "source": "https://docs.x.ai/docs/models/grok-4-1-fast-non-reasoning", + "supports_audio_input": true, + "supports_function_calling": true, + "supports_response_schema": true, + "supports_tool_choice": true, + "supports_vision": true, + "supports_web_search": true + }, "xai/grok-beta": { - "input_cost_per_token": 5e-6, + "input_cost_per_token": 0.000005, "litellm_provider": "xai", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "supports_function_calling": true, "supports_tool_choice": true, "supports_vision": true, @@ -23423,7 +26442,7 @@ "max_output_tokens": 256000, "max_tokens": 256000, "mode": "chat", - "output_cost_per_token": 1.5e-6, + "output_cost_per_token": 0.0000015, "source": "https://docs.x.ai/docs/models", "supports_function_calling": true, "supports_reasoning": true, @@ -23437,7 +26456,7 @@ "max_output_tokens": 256000, "max_tokens": 256000, "mode": "chat", - "output_cost_per_token": 1.5e-6, + "output_cost_per_token": 0.0000015, "source": "https://docs.x.ai/docs/models", "supports_function_calling": true, "supports_reasoning": true, @@ -23451,28 +26470,28 @@ "max_output_tokens": 256000, "max_tokens": 256000, "mode": "chat", - "output_cost_per_token": 1.5e-6, + "output_cost_per_token": 0.0000015, "source": "https://docs.x.ai/docs/models", "supports_function_calling": true, "supports_reasoning": true, "supports_tool_choice": true }, "xai/grok-vision-beta": { - "input_cost_per_image": 5e-6, - "input_cost_per_token": 5e-6, + "input_cost_per_image": 0.000005, + "input_cost_per_token": 0.000005, "litellm_provider": "xai", "max_input_tokens": 8192, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 1.5e-5, + "output_cost_per_token": 0.000015, "supports_function_calling": true, "supports_tool_choice": true, "supports_vision": true, "supports_web_search": true }, "vertex_ai/search_api": { - "input_cost_per_query": 1.5e-3, + "input_cost_per_query": 0.0015, "litellm_provider": "vertex_ai", "mode": "vector_store" }, @@ -23486,54 +26505,99 @@ "mode": "video_generation", "output_cost_per_video_per_second": 0.1, "source": "https://platform.openai.com/docs/api-reference/videos", - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["video"], - "supported_resolutions": ["720x1280", "1280x720"] + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "video" + ], + "supported_resolutions": [ + "720x1280", + "1280x720" + ] }, "openai/sora-2-pro": { "litellm_provider": "openai", "mode": "video_generation", "output_cost_per_video_per_second": 0.3, "source": "https://platform.openai.com/docs/api-reference/videos", - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["video"], - "supported_resolutions": ["720x1280", "1280x720"] + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "video" + ], + "supported_resolutions": [ + "720x1280", + "1280x720" + ] }, "azure/sora-2": { "litellm_provider": "azure", "mode": "video_generation", "output_cost_per_video_per_second": 0.1, "source": "https://azure.microsoft.com/en-us/products/ai-services/video-generation", - "supported_modalities": ["text"], - "supported_output_modalities": ["video"], - "supported_resolutions": ["720x1280", "1280x720"] + "supported_modalities": [ + "text" + ], + "supported_output_modalities": [ + "video" + ], + "supported_resolutions": [ + "720x1280", + "1280x720" + ] }, "azure/sora-2-pro": { "litellm_provider": "azure", "mode": "video_generation", "output_cost_per_video_per_second": 0.3, "source": "https://azure.microsoft.com/en-us/products/ai-services/video-generation", - "supported_modalities": ["text"], - "supported_output_modalities": ["video"], - "supported_resolutions": ["720x1280", "1280x720"] + "supported_modalities": [ + "text" + ], + "supported_output_modalities": [ + "video" + ], + "supported_resolutions": [ + "720x1280", + "1280x720" + ] }, "azure/sora-2-pro-high-res": { "litellm_provider": "azure", "mode": "video_generation", "output_cost_per_video_per_second": 0.5, "source": "https://azure.microsoft.com/en-us/products/ai-services/video-generation", - "supported_modalities": ["text"], - "supported_output_modalities": ["video"], - "supported_resolutions": ["1024x1792", "1792x1024"] + "supported_modalities": [ + "text" + ], + "supported_output_modalities": [ + "video" + ], + "supported_resolutions": [ + "1024x1792", + "1792x1024" + ] }, "runwayml/gen4_turbo": { "litellm_provider": "runwayml", "mode": "video_generation", "output_cost_per_video_per_second": 0.05, "source": "https://docs.dev.runwayml.com/guides/pricing/", - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["video"], - "supported_resolutions": ["1280x720", "720x1280"], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "video" + ], + "supported_resolutions": [ + "1280x720", + "720x1280" + ], "metadata": { "comment": "5 credits per second @ $0.01 per credit = $0.05 per second" } @@ -23543,9 +26607,17 @@ "mode": "video_generation", "output_cost_per_video_per_second": 0.15, "source": "https://docs.dev.runwayml.com/guides/pricing/", - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["video"], - "supported_resolutions": ["1280x720", "720x1280"], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "video" + ], + "supported_resolutions": [ + "1280x720", + "720x1280" + ], "metadata": { "comment": "15 credits per second @ $0.01 per credit = $0.15 per second" } @@ -23555,9 +26627,17 @@ "mode": "video_generation", "output_cost_per_video_per_second": 0.05, "source": "https://docs.dev.runwayml.com/guides/pricing/", - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["video"], - "supported_resolutions": ["1280x720", "720x1280"], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "video" + ], + "supported_resolutions": [ + "1280x720", + "720x1280" + ], "metadata": { "comment": "5 credits per second @ $0.01 per credit = $0.05 per second" } @@ -23568,9 +26648,17 @@ "input_cost_per_image": 0.05, "output_cost_per_image": 0.05, "source": "https://docs.dev.runwayml.com/guides/pricing/", - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["image"], - "supported_resolutions": ["1280x720", "1920x1080"], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "image" + ], + "supported_resolutions": [ + "1280x720", + "1920x1080" + ], "metadata": { "comment": "5 credits per 720p image or 8 credits per 1080p image @ $0.01 per credit. Using 5 credits ($0.05) as base cost" } @@ -23581,9 +26669,17 @@ "input_cost_per_image": 0.02, "output_cost_per_image": 0.02, "source": "https://docs.dev.runwayml.com/guides/pricing/", - "supported_modalities": ["text", "image"], - "supported_output_modalities": ["image"], - "supported_resolutions": ["1280x720", "1920x1080"], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "image" + ], + "supported_resolutions": [ + "1280x720", + "1920x1080" + ], "metadata": { "comment": "2 credits per image (any resolution) @ $0.01 per credit = $0.02 per image" } @@ -23597,4 +26693,4 @@ "comment": "Estimated cost based on standard TTS pricing. RunwayML uses ElevenLabs models." } } -} +} \ No newline at end of file From ce3b23e5e9600f654127700983010da063545c71 Mon Sep 17 00:00:00 2001 From: Ammar Date: Mon, 24 Nov 2025 14:35:55 -0600 Subject: [PATCH 2/2] =?UTF-8?q?=F0=9F=A4=96=20chore:=20update=20models.jso?= =?UTF-8?q?n=20and=20fix=20Opus=204.5=20specs?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit - Update models.json from LiteLLM - Fix claude-opus-4-5 max_output_tokens: 64k (matches AI SDK) - Update claude-opus-4-5 pricing: $5/$25 per million tokens (price drop) - This fixes the SDK warning about maxOutputTokens + thinkingBudget > max _Generated with `mux`_ --- bun.lock | 29 +- package.json | 12 +- src/common/utils/ai/providerOptions.ts | 75 + src/common/utils/tokens/modelStats.ts | 4 + src/common/utils/tokens/models-extra.ts | 15 +- src/common/utils/tokens/models.json | 8722 +++++++++-------------- src/node/services/aiService.ts | 16 +- 7 files changed, 3432 insertions(+), 5441 deletions(-) diff --git a/bun.lock b/bun.lock index 8cfd4bb451..7b3e933865 100644 --- a/bun.lock +++ b/bun.lock @@ -1,16 +1,15 @@ { "lockfileVersion": 1, - "configVersion": 0, "workspaces": { "": { "name": "@coder/cmux", "dependencies": { - "@ai-sdk/anthropic": "^2.0.44", - "@ai-sdk/google": "^2.0.38", - "@ai-sdk/openai": "^2.0.66", - "@ai-sdk/xai": "^2.0.33", + "@ai-sdk/anthropic": "^2.0.47", + "@ai-sdk/google": "^2.0.43", + "@ai-sdk/openai": "^2.0.72", + "@ai-sdk/xai": "^2.0.36", "@lydell/node-pty": "1.1.0", - "@openrouter/ai-sdk-provider": "^1.2.2", + "@openrouter/ai-sdk-provider": "^1.2.5", "@radix-ui/react-checkbox": "^1.3.3", "@radix-ui/react-dialog": "^1.1.15", "@radix-ui/react-dropdown-menu": "^2.1.16", @@ -22,7 +21,7 @@ "@radix-ui/react-tabs": "^1.1.13", "@radix-ui/react-toggle-group": "^1.1.11", "@radix-ui/react-tooltip": "^1.2.8", - "ai": "^5.0.93", + "ai": "^5.0.101", "ai-tokenizer": "^1.0.4", "chalk": "^5.6.2", "cors": "^2.8.5", @@ -137,13 +136,13 @@ "@adobe/css-tools": ["@adobe/css-tools@4.4.4", "", {}, "sha512-Elp+iwUx5rN5+Y8xLt5/GRoG20WGoDCQ/1Fb+1LiGtvwbDavuSk0jhD/eZdckHAuzcDzccnkv+rEjyWfRx18gg=="], - "@ai-sdk/anthropic": ["@ai-sdk/anthropic@2.0.44", "", { "dependencies": { "@ai-sdk/provider": "2.0.0", "@ai-sdk/provider-utils": "3.0.17" }, "peerDependencies": { "zod": "^3.25.76 || ^4.1.8" } }, "sha512-o8TfNXRzO/KZkBrcx+CL9LQsPhx7PHyqzUGjza3TJaF9WxfH1S5UQLAmEw8F7lQoHNLU0IX03WT8o8R/4JbUxQ=="], + "@ai-sdk/anthropic": ["@ai-sdk/anthropic@2.0.47", "", { "dependencies": { "@ai-sdk/provider": "2.0.0", "@ai-sdk/provider-utils": "3.0.17" }, "peerDependencies": { "zod": "^3.25.76 || ^4.1.8" } }, "sha512-YioBDTTQ6z2fijcOByG6Gj7me0ITqaJACprHROis7fXFzYIBzyAwxhsCnOrXO+oXv+9Ixddgy/Cahdmu84uRvQ=="], - "@ai-sdk/gateway": ["@ai-sdk/gateway@2.0.10", "", { "dependencies": { "@ai-sdk/provider": "2.0.0", "@ai-sdk/provider-utils": "3.0.17", "@vercel/oidc": "3.0.3" }, "peerDependencies": { "zod": "^3.25.76 || ^4.1.8" } }, "sha512-c++qOKfjKokTPAJ+vP9UXXNuTQ819yEDCZVXBhpZbgRly1P4fHTJbIAwuh+Qxxe9Bmtu8PEta0JGYZxc+hm7/Q=="], + "@ai-sdk/gateway": ["@ai-sdk/gateway@2.0.15", "", { "dependencies": { "@ai-sdk/provider": "2.0.0", "@ai-sdk/provider-utils": "3.0.17", "@vercel/oidc": "3.0.5" }, "peerDependencies": { "zod": "^3.25.76 || ^4.1.8" } }, "sha512-i1YVKzC1dg9LGvt+GthhD7NlRhz9J4+ZRj3KELU14IZ/MHPsOBiFeEoCCIDLR+3tqT8/+5nIsK3eZ7DFRfMfdw=="], - "@ai-sdk/google": ["@ai-sdk/google@2.0.38", "", { "dependencies": { "@ai-sdk/provider": "2.0.0", "@ai-sdk/provider-utils": "3.0.17" }, "peerDependencies": { "zod": "^3.25.76 || ^4.1.8" } }, "sha512-z+RFCxRA/dSd3eCkGBlnk79nz3jv8vwaW42gVc+qDuMofNfvjRz19rjnkFNuYQ6cEUcPKCo0P1rD/JLeTN2Z5A=="], + "@ai-sdk/google": ["@ai-sdk/google@2.0.43", "", { "dependencies": { "@ai-sdk/provider": "2.0.0", "@ai-sdk/provider-utils": "3.0.17" }, "peerDependencies": { "zod": "^3.25.76 || ^4.1.8" } }, "sha512-qO6giuoYCX/SdZScP/3VO5Xnbd392zm3HrTkhab/efocZU8J/VVEAcAUE1KJh0qOIAYllofRtpJIUGkRK8Q5rw=="], - "@ai-sdk/openai": ["@ai-sdk/openai@2.0.68", "", { "dependencies": { "@ai-sdk/provider": "2.0.0", "@ai-sdk/provider-utils": "3.0.17" }, "peerDependencies": { "zod": "^3.25.76 || ^4.1.8" } }, "sha512-qUSLFkqgUoFArzBwttu0KWVAZYjbsdZGOklSJXpfZ2nDC61yseHxtcnuG8u6tqKnGXDh4eakEgREDWU2sRht7A=="], + "@ai-sdk/openai": ["@ai-sdk/openai@2.0.72", "", { "dependencies": { "@ai-sdk/provider": "2.0.0", "@ai-sdk/provider-utils": "3.0.17" }, "peerDependencies": { "zod": "^3.25.76 || ^4.1.8" } }, "sha512-9j8Gdt9gFiUGFdQIjjynbC7+w8YQxkXje6dwAq1v2Pj17wmB3U0Td3lnEe/a+EnEysY3mdkc8dHPYc5BNev9NQ=="], "@ai-sdk/openai-compatible": ["@ai-sdk/openai-compatible@1.0.27", "", { "dependencies": { "@ai-sdk/provider": "2.0.0", "@ai-sdk/provider-utils": "3.0.17" }, "peerDependencies": { "zod": "^3.25.76 || ^4.1.8" } }, "sha512-bpYruxVLhrTbVH6CCq48zMJNeHu6FmHtEedl9FXckEgcIEAi036idFhJlcRwC1jNCwlacbzb8dPD7OAH1EKJaQ=="], @@ -151,7 +150,7 @@ "@ai-sdk/provider-utils": ["@ai-sdk/provider-utils@3.0.17", "", { "dependencies": { "@ai-sdk/provider": "2.0.0", "@standard-schema/spec": "^1.0.0", "eventsource-parser": "^3.0.6" }, "peerDependencies": { "zod": "^3.25.76 || ^4.1.8" } }, "sha512-TR3Gs4I3Tym4Ll+EPdzRdvo/rc8Js6c4nVhFLuvGLX/Y4V9ZcQMa/HTiYsHEgmYrf1zVi6Q145UEZUfleOwOjw=="], - "@ai-sdk/xai": ["@ai-sdk/xai@2.0.33", "", { "dependencies": { "@ai-sdk/openai-compatible": "1.0.27", "@ai-sdk/provider": "2.0.0", "@ai-sdk/provider-utils": "3.0.17" }, "peerDependencies": { "zod": "^3.25.76 || ^4.1.8" } }, "sha512-0+S+hxbAj8dA8/3dYQsmgkVkPcs8yptO1ueLWtJpa6PYjrdyliDcPSCZREL8aE76vHGvFsYlRABFfH9Ps2M8tg=="], + "@ai-sdk/xai": ["@ai-sdk/xai@2.0.36", "", { "dependencies": { "@ai-sdk/openai-compatible": "1.0.27", "@ai-sdk/provider": "2.0.0", "@ai-sdk/provider-utils": "3.0.17" }, "peerDependencies": { "zod": "^3.25.76 || ^4.1.8" } }, "sha512-tQuCDVNK4W4fiom59r2UnU7u9SAz58fpl5yKYoS9IbMOrDRO3fzQGWmj2p8MUvz9LzXf6hiyUkVNFGzzx+uZcw=="], "@antfu/install-pkg": ["@antfu/install-pkg@1.1.0", "", { "dependencies": { "package-manager-detector": "^1.3.0", "tinyexec": "^1.0.1" } }, "sha512-MGQsmw10ZyI+EJo45CdSER4zEb+p31LpDAFp2Z3gkSd1yqVZGi0Ebx++YTEMonJy4oChEMLsxZ64j8FH6sSqtQ=="], @@ -523,7 +522,7 @@ "@npmcli/move-file": ["@npmcli/move-file@2.0.1", "", { "dependencies": { "mkdirp": "^1.0.4", "rimraf": "^3.0.2" } }, "sha512-mJd2Z5TjYWq/ttPLLGqArdtnC74J6bOzg4rMDnN+p1xTacZ2yPRCk2y0oSWQtygLR9YVQXgOcONrwtnk3JupxQ=="], - "@openrouter/ai-sdk-provider": ["@openrouter/ai-sdk-provider@1.2.3", "", { "dependencies": { "@openrouter/sdk": "^0.1.8" }, "peerDependencies": { "ai": "^5.0.0", "zod": "^3.24.1 || ^v4" } }, "sha512-a6Nc8dPRHakRH9966YJ/HZJhLOds7DuPTscNZDoAr+Aw+tEFUlacSJMvb/b3gukn74mgbuaJRji9YOn62ipfVg=="], + "@openrouter/ai-sdk-provider": ["@openrouter/ai-sdk-provider@1.2.5", "", { "dependencies": { "@openrouter/sdk": "^0.1.8" }, "peerDependencies": { "ai": "^5.0.0", "zod": "^3.24.1 || ^v4" } }, "sha512-NrvJFPvdEUo6DYUQIVWPGfhafuZ2PAIX7+CUMKGknv8TcTNVo0TyP1y5SU7Bgjf/Wup9/74UFKUB07icOhVZjQ=="], "@openrouter/sdk": ["@openrouter/sdk@0.1.11", "", { "dependencies": { "zod": "^3.25.0 || ^4.0.0" }, "peerDependencies": { "@tanstack/react-query": "^5", "react": "^18 || ^19", "react-dom": "^18 || ^19" }, "optionalPeers": ["@tanstack/react-query", "react", "react-dom"] }, "sha512-OuPc8qqidL/PUM8+9WgrOfSR9+b6rKIWiezGcUJ54iPTdh+Gye5Qjut6hrLWlOCMZE7Z853gN90r1ft4iChj7Q=="], @@ -1085,7 +1084,7 @@ "@unrs/resolver-binding-win32-x64-msvc": ["@unrs/resolver-binding-win32-x64-msvc@1.11.1", "", { "os": "win32", "cpu": "x64" }, "sha512-lrW200hZdbfRtztbygyaq/6jP6AKE8qQN2KvPcJ+x7wiD038YtnYtZ82IMNJ69GJibV7bwL3y9FgK+5w/pYt6g=="], - "@vercel/oidc": ["@vercel/oidc@3.0.3", "", {}, "sha512-yNEQvPcVrK9sIe637+I0jD6leluPxzwJKx/Haw6F4H77CdDsszUn5V3o96LPziXkSNE2B83+Z3mjqGKBK/R6Gg=="], + "@vercel/oidc": ["@vercel/oidc@3.0.5", "", {}, "sha512-fnYhv671l+eTTp48gB4zEsTW/YtRgRPnkI2nT7x6qw5rkI1Lq2hTmQIpHPgyThI0znLK+vX2n9XxKdXZ7BUbbw=="], "@vitejs/plugin-react": ["@vitejs/plugin-react@4.7.0", "", { "dependencies": { "@babel/core": "^7.28.0", "@babel/plugin-transform-react-jsx-self": "^7.27.1", "@babel/plugin-transform-react-jsx-source": "^7.27.1", "@rolldown/pluginutils": "1.0.0-beta.27", "@types/babel__core": "^7.20.5", "react-refresh": "^0.17.0" }, "peerDependencies": { "vite": "^4.2.0 || ^5.0.0 || ^6.0.0 || ^7.0.0" } }, "sha512-gUu9hwfWvvEDBBmgtAowQCojwZmJ5mcLn3aufeCsitijs3+f2NsrPtlAWIR6OPiqljl96GVCUbLe0HyqIpVaoA=="], @@ -1115,7 +1114,7 @@ "aggregate-error": ["aggregate-error@3.1.0", "", { "dependencies": { "clean-stack": "^2.0.0", "indent-string": "^4.0.0" } }, "sha512-4I7Td01quW/RpocfNayFdFVk1qSuoh0E7JrbRJ16nH01HhKFQ88INq9Sd+nd72zqRySlr9BmDA8xlEJ6vJMrYA=="], - "ai": ["ai@5.0.94", "", { "dependencies": { "@ai-sdk/gateway": "2.0.10", "@ai-sdk/provider": "2.0.0", "@ai-sdk/provider-utils": "3.0.17", "@opentelemetry/api": "1.9.0" }, "peerDependencies": { "zod": "^3.25.76 || ^4.1.8" } }, "sha512-PQs3H8xDhORG/L6eyR+SxAKUsa0ORO4ENvRovzGgPmPGCIlwle6UD5VIMQFtj1gvZk+BKBUVEFFwtkTeJTAURw=="], + "ai": ["ai@5.0.101", "", { "dependencies": { "@ai-sdk/gateway": "2.0.15", "@ai-sdk/provider": "2.0.0", "@ai-sdk/provider-utils": "3.0.17", "@opentelemetry/api": "1.9.0" }, "peerDependencies": { "zod": "^3.25.76 || ^4.1.8" } }, "sha512-/P4fgs2PGYTBaZi192YkPikOudsl9vccA65F7J7LvoNTOoP5kh1yAsJPsKAy6FXU32bAngai7ft1UDyC3u7z5g=="], "ai-tokenizer": ["ai-tokenizer@1.0.4", "", { "peerDependencies": { "ai": "^5.0.0" }, "optionalPeers": ["ai"] }, "sha512-BHOUljsmH0SEO9bULQL3sz6pJ4jv00r+NHxX3kR6tn1suAAj6DDN4njSk+sqCOI5Cm6FqizUhDfoYZ0R+5/WVQ=="], diff --git a/package.json b/package.json index b6e01fd514..191c49eaf8 100644 --- a/package.json +++ b/package.json @@ -45,12 +45,12 @@ "postinstall": "sh scripts/postinstall.sh" }, "dependencies": { - "@ai-sdk/anthropic": "^2.0.44", - "@ai-sdk/google": "^2.0.38", - "@ai-sdk/openai": "^2.0.66", - "@ai-sdk/xai": "^2.0.33", + "@ai-sdk/anthropic": "^2.0.47", + "@ai-sdk/google": "^2.0.43", + "@ai-sdk/openai": "^2.0.72", + "@ai-sdk/xai": "^2.0.36", "@lydell/node-pty": "1.1.0", - "@openrouter/ai-sdk-provider": "^1.2.2", + "@openrouter/ai-sdk-provider": "^1.2.5", "@radix-ui/react-checkbox": "^1.3.3", "@radix-ui/react-dialog": "^1.1.15", "@radix-ui/react-dropdown-menu": "^2.1.16", @@ -62,7 +62,7 @@ "@radix-ui/react-tabs": "^1.1.13", "@radix-ui/react-toggle-group": "^1.1.11", "@radix-ui/react-tooltip": "^1.2.8", - "ai": "^5.0.93", + "ai": "^5.0.101", "ai-tokenizer": "^1.0.4", "chalk": "^5.6.2", "cors": "^2.8.5", diff --git a/src/common/utils/ai/providerOptions.ts b/src/common/utils/ai/providerOptions.ts index b1e9f55dbc..d343c49499 100644 --- a/src/common/utils/ai/providerOptions.ts +++ b/src/common/utils/ai/providerOptions.ts @@ -19,6 +19,7 @@ import { import { log } from "@/node/services/log"; import type { MuxMessage } from "@/common/types/message"; import { enforceThinkingPolicy } from "@/browser/utils/thinking/policy"; +import { getModelStats } from "@/common/utils/tokens/modelStats"; /** * OpenRouter reasoning options @@ -277,3 +278,77 @@ export function buildProviderOptions( log.debug("buildProviderOptions: Unsupported provider", provider); return {}; } + +/** + * Calculate the effective maxOutputTokens for a model based on its limits and thinking budget + * + * For Anthropic models with extended thinking, the AI SDK adds thinkingBudget to maxOutputTokens + * internally. We need to ensure the sum doesn't exceed the model's max_output_tokens limit. + * + * For example, Claude Opus 4 has max_output_tokens=32000. If we use: + * - thinkingBudget=20000 (high) + * - maxOutputTokens=32000 + * Then total=52000 which exceeds 32000 → SDK shows warning and caps output + * + * Solution: Reduce maxOutputTokens so that maxOutputTokens + thinkingBudget <= model limit + * + * @param modelString - Full model string (e.g., "anthropic:claude-opus-4-1") + * @param thinkingLevel - Current thinking level + * @param requestedMaxOutputTokens - Optional user-requested maxOutputTokens + * @returns Effective maxOutputTokens that respects model limits with thinking budget + */ +export function calculateEffectiveMaxOutputTokens( + modelString: string, + thinkingLevel: ThinkingLevel, + requestedMaxOutputTokens?: number +): number | undefined { + const [provider] = modelString.split(":"); + + // Only apply this adjustment for Anthropic models + if (provider !== "anthropic") { + return requestedMaxOutputTokens; + } + + // Get the actual thinking level after policy enforcement + const effectiveThinking = enforceThinkingPolicy(modelString, thinkingLevel); + const thinkingBudget = ANTHROPIC_THINKING_BUDGETS[effectiveThinking]; + + // Get model's max output tokens from models.json + const modelStats = getModelStats(modelString); + const modelMaxOutput = modelStats?.max_output_tokens; + + // If we don't know the model's max output, return requested value + if (!modelMaxOutput) { + log.debug("calculateEffectiveMaxOutputTokens: Unknown model max output, using requested", { + modelString, + requestedMaxOutputTokens, + }); + return requestedMaxOutputTokens; + } + + // Calculate the maximum safe maxOutputTokens + // The SDK will add thinkingBudget to maxOutputTokens, so we need room for both + const maxSafeOutput = modelMaxOutput - thinkingBudget; + + // If user didn't request specific tokens, use the max safe value + const targetOutput = requestedMaxOutputTokens ?? modelMaxOutput; + + // Cap at the safe maximum + const effectiveOutput = Math.min(targetOutput, maxSafeOutput); + + // Ensure we don't go below a reasonable minimum (1000 tokens) + const finalOutput = Math.max(effectiveOutput, 1000); + + log.debug("calculateEffectiveMaxOutputTokens", { + modelString, + thinkingLevel, + effectiveThinking, + thinkingBudget, + modelMaxOutput, + requestedMaxOutputTokens, + maxSafeOutput, + finalOutput, + }); + + return finalOutput; +} diff --git a/src/common/utils/tokens/modelStats.ts b/src/common/utils/tokens/modelStats.ts index 664b7db593..4e0dd6bc38 100644 --- a/src/common/utils/tokens/modelStats.ts +++ b/src/common/utils/tokens/modelStats.ts @@ -3,6 +3,7 @@ import { modelsExtra } from "./models-extra"; export interface ModelStats { max_input_tokens: number; + max_output_tokens?: number; input_cost_per_token: number; output_cost_per_token: number; cache_creation_input_token_cost?: number; @@ -11,6 +12,7 @@ export interface ModelStats { interface RawModelData { max_input_tokens?: number | string; + max_output_tokens?: number | string; input_cost_per_token?: number; output_cost_per_token?: number; cache_creation_input_token_cost?: number; @@ -37,6 +39,8 @@ function extractModelStats(data: RawModelData): ModelStats { /* eslint-disable @typescript-eslint/non-nullable-type-assertion-style */ return { max_input_tokens: data.max_input_tokens as number, + max_output_tokens: + typeof data.max_output_tokens === "number" ? data.max_output_tokens : undefined, input_cost_per_token: data.input_cost_per_token as number, output_cost_per_token: data.output_cost_per_token as number, cache_creation_input_token_cost: diff --git a/src/common/utils/tokens/models-extra.ts b/src/common/utils/tokens/models-extra.ts index edefced357..6e496c6ced 100644 --- a/src/common/utils/tokens/models-extra.ts +++ b/src/common/utils/tokens/models-extra.ts @@ -22,15 +22,16 @@ interface ModelData { } export const modelsExtra: Record = { - // Claude Opus 4.5 - Released November 2025 - // $15/M input, $75/M output (same pricing as Opus 4.1) + // Claude Opus 4.5 - Released November 24, 2025 + // $5/M input, $25/M output (price drop from Opus 4.1's $15/$75) + // 64K max output tokens (matches Sonnet 4.5) "claude-opus-4-5": { max_input_tokens: 200000, - max_output_tokens: 32000, - input_cost_per_token: 0.000015, // $15 per million input tokens - output_cost_per_token: 0.000075, // $75 per million output tokens - cache_creation_input_token_cost: 0.00001875, // $18.75 per million tokens - cache_read_input_token_cost: 0.0000015, // $1.50 per million tokens + max_output_tokens: 64000, + input_cost_per_token: 0.000005, // $5 per million input tokens + output_cost_per_token: 0.000025, // $25 per million output tokens + cache_creation_input_token_cost: 0.00000625, // $6.25 per million tokens (estimated) + cache_read_input_token_cost: 0.0000005, // $0.50 per million tokens (estimated) litellm_provider: "anthropic", mode: "chat", supports_function_calling: true, diff --git a/src/common/utils/tokens/models.json b/src/common/utils/tokens/models.json index 5400a23151..855e0ae542 100644 --- a/src/common/utils/tokens/models.json +++ b/src/common/utils/tokens/models.json @@ -1,32 +1,26 @@ { "sample_spec": { - "code_interpreter_cost_per_session": 0, - "computer_use_input_cost_per_1k_tokens": 0, - "computer_use_output_cost_per_1k_tokens": 0, + "code_interpreter_cost_per_session": 0.0, + "computer_use_input_cost_per_1k_tokens": 0.0, + "computer_use_output_cost_per_1k_tokens": 0.0, "deprecation_date": "date when the model becomes deprecated in the format YYYY-MM-DD", - "file_search_cost_per_1k_calls": 0, - "file_search_cost_per_gb_per_day": 0, - "input_cost_per_audio_token": 0, - "input_cost_per_token": 0, + "file_search_cost_per_1k_calls": 0.0, + "file_search_cost_per_gb_per_day": 0.0, + "input_cost_per_audio_token": 0.0, + "input_cost_per_token": 0.0, "litellm_provider": "one of https://docs.litellm.ai/docs/providers", "max_input_tokens": "max input tokens, if the provider specifies it. if not default to max_tokens", "max_output_tokens": "max output tokens, if the provider specifies it. if not default to max_tokens", "max_tokens": "LEGACY parameter. set to max_output_tokens if provider specifies it. IF not set to max_input_tokens, if provider specifies it.", "mode": "one of: chat, embedding, completion, image_generation, audio_transcription, audio_speech, image_generation, moderation, rerank, search", - "output_cost_per_reasoning_token": 0, - "output_cost_per_token": 0, + "output_cost_per_reasoning_token": 0.0, + "output_cost_per_token": 0.0, "search_context_cost_per_query": { - "search_context_size_high": 0, - "search_context_size_low": 0, - "search_context_size_medium": 0 + "search_context_size_high": 0.0, + "search_context_size_low": 0.0, + "search_context_size_medium": 0.0 }, - "supported_regions": [ - "global", - "us-west-2", - "eu-west-1", - "ap-southeast-1", - "ap-northeast-1" - ], + "supported_regions": ["global", "us-west-2", "eu-west-1", "ap-southeast-1", "ap-northeast-1"], "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -37,7 +31,7 @@ "supports_system_messages": true, "supports_vision": true, "supports_web_search": true, - "vector_store_cost_per_gb_per_day": 0 + "vector_store_cost_per_gb_per_day": 0.0 }, "1024-x-1024/50-steps/bedrock/amazon.nova-canvas-v1:0": { "litellm_provider": "bedrock", @@ -56,7 +50,7 @@ "input_cost_per_pixel": 1.9e-8, "litellm_provider": "openai", "mode": "image_generation", - "output_cost_per_pixel": 0 + "output_cost_per_pixel": 0.0 }, "1024-x-1024/max-steps/stability.stable-diffusion-xl-v1": { "litellm_provider": "bedrock", @@ -69,7 +63,7 @@ "input_cost_per_pixel": 2.4414e-7, "litellm_provider": "openai", "mode": "image_generation", - "output_cost_per_pixel": 0 + "output_cost_per_pixel": 0.0 }, "512-x-512/50-steps/stability.stable-diffusion-xl-v0": { "litellm_provider": "bedrock", @@ -82,7 +76,7 @@ "input_cost_per_pixel": 6.86e-8, "litellm_provider": "openai", "mode": "image_generation", - "output_cost_per_pixel": 0 + "output_cost_per_pixel": 0.0 }, "512-x-512/max-steps/stability.stable-diffusion-xl-v0": { "litellm_provider": "bedrock", @@ -92,31 +86,31 @@ "output_cost_per_image": 0.036 }, "ai21.j2-mid-v1": { - "input_cost_per_token": 0.0000125, + "input_cost_per_token": 1.25e-5, "litellm_provider": "bedrock", "max_input_tokens": 8191, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 0.0000125 + "output_cost_per_token": 1.25e-5 }, "ai21.j2-ultra-v1": { - "input_cost_per_token": 0.0000188, + "input_cost_per_token": 1.88e-5, "litellm_provider": "bedrock", "max_input_tokens": 8191, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 0.0000188 + "output_cost_per_token": 1.88e-5 }, "ai21.jamba-1-5-large-v1:0": { - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "bedrock", "max_input_tokens": 256000, "max_output_tokens": 256000, "max_tokens": 256000, "mode": "chat", - "output_cost_per_token": 0.000008 + "output_cost_per_token": 8e-6 }, "ai21.jamba-1-5-mini-v1:0": { "input_cost_per_token": 2e-7, @@ -145,9 +139,7 @@ "mode": "image_generation", "output_cost_per_image": 0.021, "source": "https://docs.aimlapi.com/", - "supported_endpoints": [ - "/v1/images/generations" - ] + "supported_endpoints": ["/v1/images/generations"] }, "aiml/dall-e-3": { "litellm_provider": "aiml", @@ -157,9 +149,7 @@ "mode": "image_generation", "output_cost_per_image": 0.042, "source": "https://docs.aimlapi.com/", - "supported_endpoints": [ - "/v1/images/generations" - ] + "supported_endpoints": ["/v1/images/generations"] }, "aiml/flux-pro": { "litellm_provider": "aiml", @@ -169,25 +159,19 @@ "mode": "image_generation", "output_cost_per_image": 0.053, "source": "https://docs.aimlapi.com/", - "supported_endpoints": [ - "/v1/images/generations" - ] + "supported_endpoints": ["/v1/images/generations"] }, "aiml/flux-pro/v1.1": { "litellm_provider": "aiml", "mode": "image_generation", "output_cost_per_image": 0.042, - "supported_endpoints": [ - "/v1/images/generations" - ] + "supported_endpoints": ["/v1/images/generations"] }, "aiml/flux-pro/v1.1-ultra": { "litellm_provider": "aiml", "mode": "image_generation", "output_cost_per_image": 0.063, - "supported_endpoints": [ - "/v1/images/generations" - ] + "supported_endpoints": ["/v1/images/generations"] }, "aiml/flux-realism": { "litellm_provider": "aiml", @@ -197,9 +181,7 @@ "mode": "image_generation", "output_cost_per_image": 0.037, "source": "https://docs.aimlapi.com/", - "supported_endpoints": [ - "/v1/images/generations" - ] + "supported_endpoints": ["/v1/images/generations"] }, "aiml/flux/dev": { "litellm_provider": "aiml", @@ -209,9 +191,7 @@ "mode": "image_generation", "output_cost_per_image": 0.026, "source": "https://docs.aimlapi.com/", - "supported_endpoints": [ - "/v1/images/generations" - ] + "supported_endpoints": ["/v1/images/generations"] }, "aiml/flux/kontext-max/text-to-image": { "litellm_provider": "aiml", @@ -221,9 +201,7 @@ "mode": "image_generation", "output_cost_per_image": 0.084, "source": "https://docs.aimlapi.com/", - "supported_endpoints": [ - "/v1/images/generations" - ] + "supported_endpoints": ["/v1/images/generations"] }, "aiml/flux/kontext-pro/text-to-image": { "litellm_provider": "aiml", @@ -233,9 +211,7 @@ "mode": "image_generation", "output_cost_per_image": 0.042, "source": "https://docs.aimlapi.com/", - "supported_endpoints": [ - "/v1/images/generations" - ] + "supported_endpoints": ["/v1/images/generations"] }, "aiml/flux/schnell": { "litellm_provider": "aiml", @@ -245,9 +221,7 @@ "mode": "image_generation", "output_cost_per_image": 0.003, "source": "https://docs.aimlapi.com/", - "supported_endpoints": [ - "/v1/images/generations" - ] + "supported_endpoints": ["/v1/images/generations"] }, "amazon.nova-lite-v1:0": { "input_cost_per_token": 6e-8, @@ -282,7 +256,7 @@ "max_output_tokens": 10000, "max_tokens": 10000, "mode": "chat", - "output_cost_per_token": 0.0000032, + "output_cost_per_token": 3.2e-6, "supports_function_calling": true, "supports_pdf_input": true, "supports_prompt_caching": true, @@ -291,7 +265,7 @@ }, "amazon.rerank-v1:0": { "input_cost_per_query": 0.001, - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "bedrock", "max_document_chunks_per_query": 100, "max_input_tokens": 32000, @@ -300,10 +274,10 @@ "max_tokens": 32000, "max_tokens_per_document_chunk": 512, "mode": "rerank", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "amazon.titan-embed-image-v1": { - "input_cost_per_image": 0.00006, + "input_cost_per_image": 6e-5, "input_cost_per_token": 8e-7, "litellm_provider": "bedrock", "max_input_tokens": 128, @@ -312,7 +286,7 @@ "notes": "'supports_image_input' is a deprecated field. Use 'supports_embedding_image_input' instead." }, "mode": "embedding", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "output_vector_size": 1024, "source": "https://us-east-1.console.aws.amazon.com/bedrock/home?region=us-east-1#/providers?model=amazon.titan-image-generator-v1", "supports_embedding_image_input": true, @@ -324,7 +298,7 @@ "max_input_tokens": 8192, "max_tokens": 8192, "mode": "embedding", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "output_vector_size": 1536 }, "amazon.titan-embed-text-v2:0": { @@ -333,11 +307,11 @@ "max_input_tokens": 8192, "max_tokens": 8192, "mode": "embedding", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "output_vector_size": 1024 }, "amazon.titan-image-generator-v1": { - "input_cost_per_image": 0, + "input_cost_per_image": 0.0, "output_cost_per_image": 0.008, "output_cost_per_image_premium_image": 0.01, "output_cost_per_image_above_512_and_512_pixels": 0.01, @@ -346,7 +320,7 @@ "mode": "image_generation" }, "amazon.titan-image-generator-v2": { - "input_cost_per_image": 0, + "input_cost_per_image": 0.0, "output_cost_per_image": 0.008, "output_cost_per_image_premium_image": 0.01, "output_cost_per_image_above_1024_and_1024_pixels": 0.01, @@ -355,18 +329,18 @@ "mode": "image_generation" }, "twelvelabs.marengo-embed-2-7-v1:0": { - "input_cost_per_token": 0.00007, + "input_cost_per_token": 7e-5, "litellm_provider": "bedrock", "max_input_tokens": 77, "max_tokens": 77, "mode": "embedding", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "output_vector_size": 1024, "supports_embedding_image_input": true, "supports_image_input": true }, "us.twelvelabs.marengo-embed-2-7-v1:0": { - "input_cost_per_token": 0.00007, + "input_cost_per_token": 7e-5, "input_cost_per_video_per_second": 0.0007, "input_cost_per_audio_per_second": 0.00014, "input_cost_per_image": 0.0001, @@ -374,13 +348,13 @@ "max_input_tokens": 77, "max_tokens": 77, "mode": "embedding", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "output_vector_size": 1024, "supports_embedding_image_input": true, "supports_image_input": true }, "eu.twelvelabs.marengo-embed-2-7-v1:0": { - "input_cost_per_token": 0.00007, + "input_cost_per_token": 7e-5, "input_cost_per_video_per_second": 0.0007, "input_cost_per_audio_per_second": 0.00014, "input_cost_per_image": 0.0001, @@ -388,40 +362,40 @@ "max_input_tokens": 77, "max_tokens": 77, "mode": "embedding", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "output_vector_size": 1024, "supports_embedding_image_input": true, "supports_image_input": true }, "twelvelabs.pegasus-1-2-v1:0": { "input_cost_per_video_per_second": 0.00049, - "output_cost_per_token": 0.0000075, + "output_cost_per_token": 7.5e-6, "litellm_provider": "bedrock", "mode": "chat", "supports_video_input": true }, "us.twelvelabs.pegasus-1-2-v1:0": { "input_cost_per_video_per_second": 0.00049, - "output_cost_per_token": 0.0000075, + "output_cost_per_token": 7.5e-6, "litellm_provider": "bedrock", "mode": "chat", "supports_video_input": true }, "eu.twelvelabs.pegasus-1-2-v1:0": { "input_cost_per_video_per_second": 0.00049, - "output_cost_per_token": 0.0000075, + "output_cost_per_token": 7.5e-6, "litellm_provider": "bedrock", "mode": "chat", "supports_video_input": true }, "amazon.titan-text-express-v1": { - "input_cost_per_token": 0.0000013, + "input_cost_per_token": 1.3e-6, "litellm_provider": "bedrock", "max_input_tokens": 42000, "max_output_tokens": 8000, "max_tokens": 8000, "mode": "chat", - "output_cost_per_token": 0.0000017 + "output_cost_per_token": 1.7e-6 }, "amazon.titan-text-lite-v1": { "input_cost_per_token": 3e-7, @@ -439,10 +413,10 @@ "max_output_tokens": 32000, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 0.0000015 + "output_cost_per_token": 1.5e-6 }, "anthropic.claude-3-5-haiku-20241022-v1:0": { - "cache_creation_input_token_cost": 0.000001, + "cache_creation_input_token_cost": 1e-6, "cache_read_input_token_cost": 8e-8, "input_cost_per_token": 8e-7, "litellm_provider": "bedrock", @@ -450,7 +424,7 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.000004, + "output_cost_per_token": 4e-6, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_pdf_input": true, @@ -459,15 +433,15 @@ "supports_tool_choice": true }, "anthropic.claude-haiku-4-5-20251001-v1:0": { - "cache_creation_input_token_cost": 0.00000125, + "cache_creation_input_token_cost": 1.25e-6, "cache_read_input_token_cost": 1e-7, - "input_cost_per_token": 0.000001, + "input_cost_per_token": 1e-6, "litellm_provider": "bedrock_converse", "max_input_tokens": 200000, "max_output_tokens": 64000, "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 0.000005, + "output_cost_per_token": 5e-6, "source": "https://aws.amazon.com/about-aws/whats-new/2025/10/claude-4-5-haiku-anthropic-amazon-bedrock", "supports_assistant_prefill": true, "supports_computer_use": true, @@ -481,15 +455,15 @@ "tool_use_system_prompt_tokens": 346 }, "anthropic.claude-haiku-4-5@20251001": { - "cache_creation_input_token_cost": 0.00000125, + "cache_creation_input_token_cost": 1.25e-6, "cache_read_input_token_cost": 1e-7, - "input_cost_per_token": 0.000001, + "input_cost_per_token": 1e-6, "litellm_provider": "bedrock_converse", "max_input_tokens": 200000, "max_output_tokens": 64000, "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 0.000005, + "output_cost_per_token": 5e-6, "source": "https://aws.amazon.com/about-aws/whats-new/2025/10/claude-4-5-haiku-anthropic-amazon-bedrock", "supports_assistant_prefill": true, "supports_computer_use": true, @@ -503,13 +477,13 @@ "tool_use_system_prompt_tokens": 346 }, "anthropic.claude-3-5-sonnet-20240620-v1:0": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "bedrock", "max_input_tokens": 200000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "supports_function_calling": true, "supports_pdf_input": true, "supports_response_schema": true, @@ -517,15 +491,15 @@ "supports_vision": true }, "anthropic.claude-3-5-sonnet-20241022-v2:0": { - "cache_creation_input_token_cost": 0.00000375, + "cache_creation_input_token_cost": 3.75e-6, "cache_read_input_token_cost": 3e-7, - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "bedrock", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "supports_assistant_prefill": true, "supports_computer_use": true, "supports_function_calling": true, @@ -536,15 +510,15 @@ "supports_vision": true }, "anthropic.claude-3-7-sonnet-20240620-v1:0": { - "cache_creation_input_token_cost": 0.0000045, + "cache_creation_input_token_cost": 4.5e-6, "cache_read_input_token_cost": 3.6e-7, - "input_cost_per_token": 0.0000036, + "input_cost_per_token": 3.6e-6, "litellm_provider": "bedrock", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.000018, + "output_cost_per_token": 1.8e-5, "supports_assistant_prefill": true, "supports_computer_use": true, "supports_function_calling": true, @@ -556,15 +530,15 @@ "supports_vision": true }, "anthropic.claude-3-7-sonnet-20250219-v1:0": { - "cache_creation_input_token_cost": 0.00000375, + "cache_creation_input_token_cost": 3.75e-6, "cache_read_input_token_cost": 3e-7, - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "bedrock_converse", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "supports_assistant_prefill": true, "supports_computer_use": true, "supports_function_calling": true, @@ -582,7 +556,7 @@ "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.00000125, + "output_cost_per_token": 1.25e-6, "supports_function_calling": true, "supports_pdf_input": true, "supports_response_schema": true, @@ -590,26 +564,26 @@ "supports_vision": true }, "anthropic.claude-3-opus-20240229-v1:0": { - "input_cost_per_token": 0.000015, + "input_cost_per_token": 1.5e-5, "litellm_provider": "bedrock", "max_input_tokens": 200000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.000075, + "output_cost_per_token": 7.5e-5, "supports_function_calling": true, "supports_response_schema": true, "supports_tool_choice": true, "supports_vision": true }, "anthropic.claude-3-sonnet-20240229-v1:0": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "bedrock", "max_input_tokens": 200000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "supports_function_calling": true, "supports_pdf_input": true, "supports_response_schema": true, @@ -623,19 +597,19 @@ "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 0.0000024, + "output_cost_per_token": 2.4e-6, "supports_tool_choice": true }, "anthropic.claude-opus-4-1-20250805-v1:0": { - "cache_creation_input_token_cost": 0.00001875, - "cache_read_input_token_cost": 0.0000015, - "input_cost_per_token": 0.000015, + "cache_creation_input_token_cost": 1.875e-5, + "cache_read_input_token_cost": 1.5e-6, + "input_cost_per_token": 1.5e-5, "litellm_provider": "bedrock_converse", "max_input_tokens": 200000, "max_output_tokens": 32000, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 0.000075, + "output_cost_per_token": 7.5e-5, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -653,15 +627,15 @@ "tool_use_system_prompt_tokens": 159 }, "anthropic.claude-opus-4-20250514-v1:0": { - "cache_creation_input_token_cost": 0.00001875, - "cache_read_input_token_cost": 0.0000015, - "input_cost_per_token": 0.000015, + "cache_creation_input_token_cost": 1.875e-5, + "cache_read_input_token_cost": 1.5e-6, + "input_cost_per_token": 1.5e-5, "litellm_provider": "bedrock_converse", "max_input_tokens": 200000, "max_output_tokens": 32000, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 0.000075, + "output_cost_per_token": 7.5e-5, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -679,19 +653,19 @@ "tool_use_system_prompt_tokens": 159 }, "anthropic.claude-sonnet-4-20250514-v1:0": { - "cache_creation_input_token_cost": 0.00000375, + "cache_creation_input_token_cost": 3.75e-6, "cache_read_input_token_cost": 3e-7, - "input_cost_per_token": 0.000003, - "input_cost_per_token_above_200k_tokens": 0.000006, - "output_cost_per_token_above_200k_tokens": 0.0000225, - "cache_creation_input_token_cost_above_200k_tokens": 0.0000075, + "input_cost_per_token": 3e-6, + "input_cost_per_token_above_200k_tokens": 6e-6, + "output_cost_per_token_above_200k_tokens": 2.25e-5, + "cache_creation_input_token_cost_above_200k_tokens": 7.5e-6, "cache_read_input_token_cost_above_200k_tokens": 6e-7, "litellm_provider": "bedrock_converse", "max_input_tokens": 1000000, "max_output_tokens": 64000, "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -709,19 +683,19 @@ "tool_use_system_prompt_tokens": 159 }, "anthropic.claude-sonnet-4-5-20250929-v1:0": { - "cache_creation_input_token_cost": 0.00000375, + "cache_creation_input_token_cost": 3.75e-6, "cache_read_input_token_cost": 3e-7, - "input_cost_per_token": 0.000003, - "input_cost_per_token_above_200k_tokens": 0.000006, - "output_cost_per_token_above_200k_tokens": 0.0000225, - "cache_creation_input_token_cost_above_200k_tokens": 0.0000075, + "input_cost_per_token": 3e-6, + "input_cost_per_token_above_200k_tokens": 6e-6, + "output_cost_per_token_above_200k_tokens": 2.25e-5, + "cache_creation_input_token_cost_above_200k_tokens": 7.5e-6, "cache_read_input_token_cost_above_200k_tokens": 6e-7, "litellm_provider": "bedrock_converse", "max_input_tokens": 200000, "max_output_tokens": 64000, "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -739,22 +713,22 @@ "tool_use_system_prompt_tokens": 159 }, "anthropic.claude-v1": { - "input_cost_per_token": 0.000008, + "input_cost_per_token": 8e-6, "litellm_provider": "bedrock", "max_input_tokens": 100000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 0.000024 + "output_cost_per_token": 2.4e-5 }, "anthropic.claude-v2:1": { - "input_cost_per_token": 0.000008, + "input_cost_per_token": 8e-6, "litellm_provider": "bedrock", "max_input_tokens": 100000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 0.000024, + "output_cost_per_token": 2.4e-5, "supports_tool_choice": true }, "anyscale/HuggingFaceH4/zephyr-7b-beta": { @@ -767,22 +741,22 @@ "output_cost_per_token": 1.5e-7 }, "anyscale/codellama/CodeLlama-34b-Instruct-hf": { - "input_cost_per_token": 0.000001, + "input_cost_per_token": 1e-6, "litellm_provider": "anyscale", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.000001 + "output_cost_per_token": 1e-6 }, "anyscale/codellama/CodeLlama-70b-Instruct-hf": { - "input_cost_per_token": 0.000001, + "input_cost_per_token": 1e-6, "litellm_provider": "anyscale", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.000001, + "output_cost_per_token": 1e-6, "source": "https://docs.anyscale.com/preview/endpoints/text-generation/supported-models/codellama-CodeLlama-70b-Instruct-hf" }, "anyscale/google/gemma-7b-it": { @@ -805,13 +779,13 @@ "output_cost_per_token": 2.5e-7 }, "anyscale/meta-llama/Llama-2-70b-chat-hf": { - "input_cost_per_token": 0.000001, + "input_cost_per_token": 1e-6, "litellm_provider": "anyscale", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.000001 + "output_cost_per_token": 1e-6 }, "anyscale/meta-llama/Llama-2-7b-chat-hf": { "input_cost_per_token": 1.5e-7, @@ -823,13 +797,13 @@ "output_cost_per_token": 1.5e-7 }, "anyscale/meta-llama/Meta-Llama-3-70B-Instruct": { - "input_cost_per_token": 0.000001, + "input_cost_per_token": 1e-6, "litellm_provider": "anyscale", "max_input_tokens": 8192, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.000001, + "output_cost_per_token": 1e-6, "source": "https://docs.anyscale.com/preview/endpoints/text-generation/supported-models/meta-llama-Meta-Llama-3-70B-Instruct" }, "anyscale/meta-llama/Meta-Llama-3-8B-Instruct": { @@ -908,7 +882,7 @@ "max_output_tokens": 10000, "max_tokens": 10000, "mode": "chat", - "output_cost_per_token": 0.00000336, + "output_cost_per_token": 3.36e-6, "supports_function_calling": true, "supports_pdf_input": true, "supports_prompt_caching": true, @@ -916,13 +890,13 @@ "supports_vision": true }, "apac.anthropic.claude-3-5-sonnet-20240620-v1:0": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "bedrock", "max_input_tokens": 200000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "supports_function_calling": true, "supports_pdf_input": true, "supports_response_schema": true, @@ -930,15 +904,15 @@ "supports_vision": true }, "apac.anthropic.claude-3-5-sonnet-20241022-v2:0": { - "cache_creation_input_token_cost": 0.00000375, + "cache_creation_input_token_cost": 3.75e-6, "cache_read_input_token_cost": 3e-7, - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "bedrock", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "supports_assistant_prefill": true, "supports_computer_use": true, "supports_function_calling": true, @@ -955,7 +929,7 @@ "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.00000125, + "output_cost_per_token": 1.25e-6, "supports_function_calling": true, "supports_pdf_input": true, "supports_response_schema": true, @@ -963,15 +937,15 @@ "supports_vision": true }, "apac.anthropic.claude-haiku-4-5-20251001-v1:0": { - "cache_creation_input_token_cost": 0.000001375, + "cache_creation_input_token_cost": 1.375e-6, "cache_read_input_token_cost": 1.1e-7, - "input_cost_per_token": 0.0000011, + "input_cost_per_token": 1.1e-6, "litellm_provider": "bedrock_converse", "max_input_tokens": 200000, "max_output_tokens": 64000, "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 0.0000055, + "output_cost_per_token": 5.5e-6, "source": "https://aws.amazon.com/about-aws/whats-new/2025/10/claude-4-5-haiku-anthropic-amazon-bedrock", "supports_assistant_prefill": true, "supports_computer_use": true, @@ -985,13 +959,13 @@ "tool_use_system_prompt_tokens": 346 }, "apac.anthropic.claude-3-sonnet-20240229-v1:0": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "bedrock", "max_input_tokens": 200000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "supports_function_calling": true, "supports_pdf_input": true, "supports_response_schema": true, @@ -999,19 +973,19 @@ "supports_vision": true }, "apac.anthropic.claude-sonnet-4-20250514-v1:0": { - "cache_creation_input_token_cost": 0.00000375, + "cache_creation_input_token_cost": 3.75e-6, "cache_read_input_token_cost": 3e-7, - "input_cost_per_token": 0.000003, - "input_cost_per_token_above_200k_tokens": 0.000006, - "output_cost_per_token_above_200k_tokens": 0.0000225, - "cache_creation_input_token_cost_above_200k_tokens": 0.0000075, + "input_cost_per_token": 3e-6, + "input_cost_per_token_above_200k_tokens": 6e-6, + "output_cost_per_token_above_200k_tokens": 2.25e-5, + "cache_creation_input_token_cost_above_200k_tokens": 7.5e-6, "cache_read_input_token_cost_above_200k_tokens": 6e-7, "litellm_provider": "bedrock_converse", "max_input_tokens": 1000000, "max_output_tokens": 64000, "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -1029,31 +1003,31 @@ "tool_use_system_prompt_tokens": 159 }, "assemblyai/best": { - "input_cost_per_second": 0.00003333, + "input_cost_per_second": 3.333e-5, "litellm_provider": "assemblyai", "mode": "audio_transcription", - "output_cost_per_second": 0 + "output_cost_per_second": 0.0 }, "assemblyai/nano": { "input_cost_per_second": 0.00010278, "litellm_provider": "assemblyai", "mode": "audio_transcription", - "output_cost_per_second": 0 + "output_cost_per_second": 0.0 }, "au.anthropic.claude-sonnet-4-5-20250929-v1:0": { - "cache_creation_input_token_cost": 0.000004125, + "cache_creation_input_token_cost": 4.125e-6, "cache_read_input_token_cost": 3.3e-7, - "input_cost_per_token": 0.0000033, - "input_cost_per_token_above_200k_tokens": 0.0000066, - "output_cost_per_token_above_200k_tokens": 0.00002475, - "cache_creation_input_token_cost_above_200k_tokens": 0.00000825, + "input_cost_per_token": 3.3e-6, + "input_cost_per_token_above_200k_tokens": 6.6e-6, + "output_cost_per_token_above_200k_tokens": 2.475e-5, + "cache_creation_input_token_cost_above_200k_tokens": 8.25e-6, "cache_read_input_token_cost_above_200k_tokens": 6.6e-7, "litellm_provider": "bedrock_converse", "max_input_tokens": 200000, "max_output_tokens": 64000, "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 0.0000165, + "output_cost_per_token": 1.65e-5, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -1076,27 +1050,20 @@ "max_input_tokens": 8191, "max_tokens": 8191, "mode": "embedding", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "azure/codex-mini": { "cache_read_input_token_cost": 3.75e-7, - "input_cost_per_token": 0.0000015, + "input_cost_per_token": 1.5e-6, "litellm_provider": "azure", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "responses", - "output_cost_per_token": 0.000006, - "supported_endpoints": [ - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 6e-6, + "supported_endpoints": ["/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_pdf_input": true, @@ -1108,33 +1075,26 @@ "supports_vision": true }, "azure/command-r-plus": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "supports_function_calling": true }, "azure/computer-use-preview": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "azure", "max_input_tokens": 8192, "max_output_tokens": 1024, "max_tokens": 1024, "mode": "chat", - "output_cost_per_token": 0.000012, - "supported_endpoints": [ - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 1.2e-5, + "supported_endpoints": ["/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": false, @@ -1151,14 +1111,14 @@ }, "azure/eu/gpt-4o-2024-08-06": { "deprecation_date": "2026-02-27", - "cache_read_input_token_cost": 0.000001375, - "input_cost_per_token": 0.00000275, + "cache_read_input_token_cost": 1.375e-6, + "input_cost_per_token": 2.75e-6, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 0.000011, + "output_cost_per_token": 1.1e-5, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -1168,14 +1128,14 @@ }, "azure/eu/gpt-4o-2024-11-20": { "deprecation_date": "2026-03-01", - "cache_creation_input_token_cost": 0.00000138, - "input_cost_per_token": 0.00000275, + "cache_creation_input_token_cost": 1.38e-6, + "input_cost_per_token": 2.75e-6, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 0.000011, + "output_cost_per_token": 1.1e-5, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_response_schema": true, @@ -1201,15 +1161,15 @@ "azure/eu/gpt-4o-mini-realtime-preview-2024-12-17": { "cache_creation_input_audio_token_cost": 3.3e-7, "cache_read_input_token_cost": 3.3e-7, - "input_cost_per_audio_token": 0.000011, + "input_cost_per_audio_token": 1.1e-5, "input_cost_per_token": 6.6e-7, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_audio_token": 0.000022, - "output_cost_per_token": 0.00000264, + "output_cost_per_audio_token": 2.2e-5, + "output_cost_per_token": 2.64e-6, "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -1218,17 +1178,17 @@ "supports_tool_choice": true }, "azure/eu/gpt-4o-realtime-preview-2024-10-01": { - "cache_creation_input_audio_token_cost": 0.000022, - "cache_read_input_token_cost": 0.00000275, + "cache_creation_input_audio_token_cost": 2.2e-5, + "cache_read_input_token_cost": 2.75e-6, "input_cost_per_audio_token": 0.00011, - "input_cost_per_token": 0.0000055, + "input_cost_per_token": 5.5e-6, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", "output_cost_per_audio_token": 0.00022, - "output_cost_per_token": 0.000022, + "output_cost_per_token": 2.2e-5, "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -1237,25 +1197,19 @@ "supports_tool_choice": true }, "azure/eu/gpt-4o-realtime-preview-2024-12-17": { - "cache_read_input_audio_token_cost": 0.0000025, - "cache_read_input_token_cost": 0.00000275, - "input_cost_per_audio_token": 0.000044, - "input_cost_per_token": 0.0000055, + "cache_read_input_audio_token_cost": 2.5e-6, + "cache_read_input_token_cost": 2.75e-6, + "input_cost_per_audio_token": 4.4e-5, + "input_cost_per_token": 5.5e-6, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_audio_token": 0.00008, - "output_cost_per_token": 0.000022, - "supported_modalities": [ - "text", - "audio" - ], - "supported_output_modalities": [ - "text", - "audio" - ], + "output_cost_per_audio_token": 8e-5, + "output_cost_per_token": 2.2e-5, + "supported_modalities": ["text", "audio"], + "supported_output_modalities": ["text", "audio"], "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -1265,25 +1219,16 @@ }, "azure/eu/gpt-5-2025-08-07": { "cache_read_input_token_cost": 1.375e-7, - "input_cost_per_token": 0.000001375, + "input_cost_per_token": 1.375e-6, "litellm_provider": "azure", "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.000011, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 1.1e-5, + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -1303,19 +1248,10 @@ "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.0000022, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 2.2e-6, + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -1329,26 +1265,16 @@ }, "azure/eu/gpt-5.1": { "cache_read_input_token_cost": 1.4e-7, - "input_cost_per_token": 0.00000138, + "input_cost_per_token": 1.38e-6, "litellm_provider": "azure", "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.000011, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text", - "image" - ], + "output_cost_per_token": 1.1e-5, + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text", "image"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -1362,26 +1288,16 @@ }, "azure/eu/gpt-5.1-chat": { "cache_read_input_token_cost": 1.4e-7, - "input_cost_per_token": 0.00000138, + "input_cost_per_token": 1.38e-6, "litellm_provider": "azure", "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.000011, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text", - "image" - ], + "output_cost_per_token": 1.1e-5, + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text", "image"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -1395,23 +1311,16 @@ }, "azure/eu/gpt-5.1-codex": { "cache_read_input_token_cost": 1.4e-7, - "input_cost_per_token": 0.00000138, + "input_cost_per_token": 1.38e-6, "litellm_provider": "azure", "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "responses", - "output_cost_per_token": 0.000011, - "supported_endpoints": [ - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 1.1e-5, + "supported_endpoints": ["/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -1431,17 +1340,10 @@ "max_output_tokens": 128000, "max_tokens": 128000, "mode": "responses", - "output_cost_per_token": 0.0000022, - "supported_endpoints": [ - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 2.2e-6, + "supported_endpoints": ["/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -1462,18 +1364,9 @@ "max_tokens": 128000, "mode": "chat", "output_cost_per_token": 4.4e-7, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -1486,14 +1379,14 @@ "supports_vision": true }, "azure/eu/o1-2024-12-17": { - "cache_read_input_token_cost": 0.00000825, - "input_cost_per_token": 0.0000165, + "cache_read_input_token_cost": 8.25e-6, + "input_cost_per_token": 1.65e-5, "litellm_provider": "azure", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 0.000066, + "output_cost_per_token": 6.6e-5, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -1502,29 +1395,29 @@ }, "azure/eu/o1-mini-2024-09-12": { "cache_read_input_token_cost": 6.05e-7, - "input_cost_per_token": 0.00000121, + "input_cost_per_token": 1.21e-6, "input_cost_per_token_batches": 6.05e-7, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 65536, "max_tokens": 65536, "mode": "chat", - "output_cost_per_token": 0.00000484, - "output_cost_per_token_batches": 0.00000242, + "output_cost_per_token": 4.84e-6, + "output_cost_per_token_batches": 2.42e-6, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, "supports_vision": false }, "azure/eu/o1-preview-2024-09-12": { - "cache_read_input_token_cost": 0.00000825, - "input_cost_per_token": 0.0000165, + "cache_read_input_token_cost": 8.25e-6, + "input_cost_per_token": 1.65e-5, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0.000066, + "output_cost_per_token": 6.6e-5, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -1532,30 +1425,30 @@ }, "azure/eu/o3-mini-2025-01-31": { "cache_read_input_token_cost": 6.05e-7, - "input_cost_per_token": 0.00000121, + "input_cost_per_token": 1.21e-6, "input_cost_per_token_batches": 6.05e-7, "litellm_provider": "azure", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 0.00000484, - "output_cost_per_token_batches": 0.00000242, + "output_cost_per_token": 4.84e-6, + "output_cost_per_token_batches": 2.42e-6, "supports_prompt_caching": true, "supports_reasoning": true, "supports_tool_choice": true, "supports_vision": false }, "azure/global-standard/gpt-4o-2024-08-06": { - "cache_read_input_token_cost": 0.00000125, + "cache_read_input_token_cost": 1.25e-6, "deprecation_date": "2026-02-27", - "input_cost_per_token": 0.0000025, + "input_cost_per_token": 2.5e-6, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 0.00001, + "output_cost_per_token": 1e-5, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -1564,15 +1457,15 @@ "supports_vision": true }, "azure/global-standard/gpt-4o-2024-11-20": { - "cache_read_input_token_cost": 0.00000125, + "cache_read_input_token_cost": 1.25e-6, "deprecation_date": "2026-03-01", - "input_cost_per_token": 0.0000025, + "input_cost_per_token": 2.5e-6, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 0.00001, + "output_cost_per_token": 1e-5, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_response_schema": true, @@ -1595,14 +1488,14 @@ }, "azure/global/gpt-4o-2024-08-06": { "deprecation_date": "2026-02-27", - "cache_read_input_token_cost": 0.00000125, - "input_cost_per_token": 0.0000025, + "cache_read_input_token_cost": 1.25e-6, + "input_cost_per_token": 2.5e-6, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 0.00001, + "output_cost_per_token": 1e-5, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -1612,14 +1505,14 @@ }, "azure/global/gpt-4o-2024-11-20": { "deprecation_date": "2026-03-01", - "cache_read_input_token_cost": 0.00000125, - "input_cost_per_token": 0.0000025, + "cache_read_input_token_cost": 1.25e-6, + "input_cost_per_token": 2.5e-6, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 0.00001, + "output_cost_per_token": 1e-5, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -1629,26 +1522,16 @@ }, "azure/global/gpt-5.1": { "cache_read_input_token_cost": 1.25e-7, - "input_cost_per_token": 0.00000125, + "input_cost_per_token": 1.25e-6, "litellm_provider": "azure", "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.00001, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text", - "image" - ], + "output_cost_per_token": 1e-5, + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text", "image"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -1662,26 +1545,16 @@ }, "azure/global/gpt-5.1-chat": { "cache_read_input_token_cost": 1.25e-7, - "input_cost_per_token": 0.00000125, + "input_cost_per_token": 1.25e-6, "litellm_provider": "azure", "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.00001, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text", - "image" - ], + "output_cost_per_token": 1e-5, + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text", "image"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -1695,23 +1568,16 @@ }, "azure/global/gpt-5.1-codex": { "cache_read_input_token_cost": 1.25e-7, - "input_cost_per_token": 0.00000125, + "input_cost_per_token": 1.25e-6, "litellm_provider": "azure", "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "responses", - "output_cost_per_token": 0.00001, - "supported_endpoints": [ - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 1e-5, + "supported_endpoints": ["/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -1731,17 +1597,10 @@ "max_output_tokens": 128000, "max_tokens": 128000, "mode": "responses", - "output_cost_per_token": 0.000002, - "supported_endpoints": [ - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 2e-6, + "supported_endpoints": ["/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -1760,7 +1619,7 @@ "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.0000015, + "output_cost_per_token": 1.5e-6, "supports_function_calling": true, "supports_tool_choice": true }, @@ -1772,18 +1631,18 @@ "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.0000015, + "output_cost_per_token": 1.5e-6, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_tool_choice": true }, "azure/gpt-3.5-turbo-instruct-0914": { - "input_cost_per_token": 0.0000015, + "input_cost_per_token": 1.5e-6, "litellm_provider": "azure_text", "max_input_tokens": 4097, "max_tokens": 4097, "mode": "completion", - "output_cost_per_token": 0.000002 + "output_cost_per_token": 2e-6 }, "azure/gpt-35-turbo": { "input_cost_per_token": 5e-7, @@ -1792,7 +1651,7 @@ "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.0000015, + "output_cost_per_token": 1.5e-6, "supports_function_calling": true, "supports_tool_choice": true }, @@ -1804,7 +1663,7 @@ "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.0000015, + "output_cost_per_token": 1.5e-6, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_tool_choice": true @@ -1817,122 +1676,122 @@ "max_output_tokens": 4096, "max_tokens": 4097, "mode": "chat", - "output_cost_per_token": 0.000002, + "output_cost_per_token": 2e-6, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_tool_choice": true }, "azure/gpt-35-turbo-0613": { "deprecation_date": "2025-02-13", - "input_cost_per_token": 0.0000015, + "input_cost_per_token": 1.5e-6, "litellm_provider": "azure", "max_input_tokens": 4097, "max_output_tokens": 4096, "max_tokens": 4097, "mode": "chat", - "output_cost_per_token": 0.000002, + "output_cost_per_token": 2e-6, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_tool_choice": true }, "azure/gpt-35-turbo-1106": { "deprecation_date": "2025-03-31", - "input_cost_per_token": 0.000001, + "input_cost_per_token": 1e-6, "litellm_provider": "azure", "max_input_tokens": 16384, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.000002, + "output_cost_per_token": 2e-6, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_tool_choice": true }, "azure/gpt-35-turbo-16k": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "azure", "max_input_tokens": 16385, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.000004, + "output_cost_per_token": 4e-6, "supports_tool_choice": true }, "azure/gpt-35-turbo-16k-0613": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "azure", "max_input_tokens": 16385, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.000004, + "output_cost_per_token": 4e-6, "supports_function_calling": true, "supports_tool_choice": true }, "azure/gpt-35-turbo-instruct": { - "input_cost_per_token": 0.0000015, + "input_cost_per_token": 1.5e-6, "litellm_provider": "azure_text", "max_input_tokens": 4097, "max_tokens": 4097, "mode": "completion", - "output_cost_per_token": 0.000002 + "output_cost_per_token": 2e-6 }, "azure/gpt-35-turbo-instruct-0914": { - "input_cost_per_token": 0.0000015, + "input_cost_per_token": 1.5e-6, "litellm_provider": "azure_text", "max_input_tokens": 4097, "max_tokens": 4097, "mode": "completion", - "output_cost_per_token": 0.000002 + "output_cost_per_token": 2e-6 }, "azure/gpt-4": { - "input_cost_per_token": 0.00003, + "input_cost_per_token": 3e-5, "litellm_provider": "azure", "max_input_tokens": 8192, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.00006, + "output_cost_per_token": 6e-5, "supports_function_calling": true, "supports_tool_choice": true }, "azure/gpt-4-0125-preview": { - "input_cost_per_token": 0.00001, + "input_cost_per_token": 1e-5, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.00003, + "output_cost_per_token": 3e-5, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_tool_choice": true }, "azure/gpt-4-0613": { - "input_cost_per_token": 0.00003, + "input_cost_per_token": 3e-5, "litellm_provider": "azure", "max_input_tokens": 8192, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.00006, + "output_cost_per_token": 6e-5, "supports_function_calling": true, "supports_tool_choice": true }, "azure/gpt-4-1106-preview": { - "input_cost_per_token": 0.00001, + "input_cost_per_token": 1e-5, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.00003, + "output_cost_per_token": 3e-5, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_tool_choice": true }, "azure/gpt-4-32k": { - "input_cost_per_token": 0.00006, + "input_cost_per_token": 6e-5, "litellm_provider": "azure", "max_input_tokens": 32768, "max_output_tokens": 4096, @@ -1942,7 +1801,7 @@ "supports_tool_choice": true }, "azure/gpt-4-32k-0613": { - "input_cost_per_token": 0.00006, + "input_cost_per_token": 6e-5, "litellm_provider": "azure", "max_input_tokens": 32768, "max_output_tokens": 4096, @@ -1952,64 +1811,55 @@ "supports_tool_choice": true }, "azure/gpt-4-turbo": { - "input_cost_per_token": 0.00001, + "input_cost_per_token": 1e-5, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.00003, + "output_cost_per_token": 3e-5, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_tool_choice": true }, "azure/gpt-4-turbo-2024-04-09": { - "input_cost_per_token": 0.00001, + "input_cost_per_token": 1e-5, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.00003, + "output_cost_per_token": 3e-5, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_tool_choice": true, "supports_vision": true }, "azure/gpt-4-turbo-vision-preview": { - "input_cost_per_token": 0.00001, + "input_cost_per_token": 1e-5, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.00003, + "output_cost_per_token": 3e-5, "supports_tool_choice": true, "supports_vision": true }, "azure/gpt-4.1": { "cache_read_input_token_cost": 5e-7, - "input_cost_per_token": 0.000002, - "input_cost_per_token_batches": 0.000001, + "input_cost_per_token": 2e-6, + "input_cost_per_token_batches": 1e-6, "litellm_provider": "azure", "max_input_tokens": 1047576, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0.000008, - "output_cost_per_token_batches": 0.000004, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 8e-6, + "output_cost_per_token_batches": 4e-6, + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -2023,27 +1873,18 @@ "azure/gpt-4.1-2025-04-14": { "deprecation_date": "2026-11-04", "cache_read_input_token_cost": 5e-7, - "input_cost_per_token": 0.000002, - "input_cost_per_token_batches": 0.000001, + "input_cost_per_token": 2e-6, + "input_cost_per_token_batches": 1e-6, "litellm_provider": "azure", "max_input_tokens": 1047576, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0.000008, - "output_cost_per_token_batches": 0.000004, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 8e-6, + "output_cost_per_token_batches": 4e-6, + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -2063,20 +1904,11 @@ "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0.0000016, + "output_cost_per_token": 1.6e-6, "output_cost_per_token_batches": 8e-7, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -2097,20 +1929,11 @@ "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0.0000016, + "output_cost_per_token": 1.6e-6, "output_cost_per_token_batches": 8e-7, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -2132,18 +1955,9 @@ "mode": "chat", "output_cost_per_token": 4e-7, "output_cost_per_token_batches": 2e-7, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -2165,18 +1979,9 @@ "mode": "chat", "output_cost_per_token": 4e-7, "output_cost_per_token_batches": 2e-7, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -2187,16 +1992,16 @@ "supports_vision": true }, "azure/gpt-4.5-preview": { - "cache_read_input_token_cost": 0.0000375, - "input_cost_per_token": 0.000075, - "input_cost_per_token_batches": 0.0000375, + "cache_read_input_token_cost": 3.75e-5, + "input_cost_per_token": 7.5e-5, + "input_cost_per_token_batches": 3.75e-5, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", "output_cost_per_token": 0.00015, - "output_cost_per_token_batches": 0.000075, + "output_cost_per_token_batches": 7.5e-5, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -2206,14 +2011,14 @@ "supports_vision": true }, "azure/gpt-4o": { - "cache_read_input_token_cost": 0.00000125, - "input_cost_per_token": 0.0000025, + "cache_read_input_token_cost": 1.25e-6, + "input_cost_per_token": 2.5e-6, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 0.00001, + "output_cost_per_token": 1e-5, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -2222,13 +2027,13 @@ "supports_vision": true }, "azure/gpt-4o-2024-05-13": { - "input_cost_per_token": 0.000005, + "input_cost_per_token": 5e-6, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -2237,14 +2042,14 @@ }, "azure/gpt-4o-2024-08-06": { "deprecation_date": "2026-02-27", - "cache_read_input_token_cost": 0.00000125, - "input_cost_per_token": 0.0000025, + "cache_read_input_token_cost": 1.25e-6, + "input_cost_per_token": 2.5e-6, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 0.00001, + "output_cost_per_token": 1e-5, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -2254,14 +2059,14 @@ }, "azure/gpt-4o-2024-11-20": { "deprecation_date": "2026-03-01", - "cache_read_input_token_cost": 0.00000125, - "input_cost_per_token": 0.00000275, + "cache_read_input_token_cost": 1.25e-6, + "input_cost_per_token": 2.75e-6, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 0.000011, + "output_cost_per_token": 1.1e-5, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -2270,26 +2075,18 @@ "supports_vision": true }, "azure/gpt-audio-2025-08-28": { - "input_cost_per_audio_token": 0.00004, - "input_cost_per_token": 0.0000025, + "input_cost_per_audio_token": 4e-5, + "input_cost_per_token": 2.5e-6, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_audio_token": 0.00008, - "output_cost_per_token": 0.00001, - "supported_endpoints": [ - "/v1/chat/completions" - ], - "supported_modalities": [ - "text", - "audio" - ], - "supported_output_modalities": [ - "text", - "audio" - ], + "output_cost_per_audio_token": 8e-5, + "output_cost_per_token": 1e-5, + "supported_endpoints": ["/v1/chat/completions"], + "supported_modalities": ["text", "audio"], + "supported_output_modalities": ["text", "audio"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -2301,26 +2098,18 @@ "supports_vision": false }, "azure/gpt-audio-mini-2025-10-06": { - "input_cost_per_audio_token": 0.00001, + "input_cost_per_audio_token": 1e-5, "input_cost_per_token": 6e-7, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_audio_token": 0.00002, - "output_cost_per_token": 0.0000024, - "supported_endpoints": [ - "/v1/chat/completions" - ], - "supported_modalities": [ - "text", - "audio" - ], - "supported_output_modalities": [ - "text", - "audio" - ], + "output_cost_per_audio_token": 2e-5, + "output_cost_per_token": 2.4e-6, + "supported_endpoints": ["/v1/chat/completions"], + "supported_modalities": ["text", "audio"], + "supported_output_modalities": ["text", "audio"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -2332,26 +2121,18 @@ "supports_vision": false }, "azure/gpt-4o-audio-preview-2024-12-17": { - "input_cost_per_audio_token": 0.00004, - "input_cost_per_token": 0.0000025, + "input_cost_per_audio_token": 4e-5, + "input_cost_per_token": 2.5e-6, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_audio_token": 0.00008, - "output_cost_per_token": 0.00001, - "supported_endpoints": [ - "/v1/chat/completions" - ], - "supported_modalities": [ - "text", - "audio" - ], - "supported_output_modalities": [ - "text", - "audio" - ], + "output_cost_per_audio_token": 8e-5, + "output_cost_per_token": 1e-5, + "supported_endpoints": ["/v1/chat/completions"], + "supported_modalities": ["text", "audio"], + "supported_output_modalities": ["text", "audio"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -2395,26 +2176,18 @@ "supports_vision": true }, "azure/gpt-4o-mini-audio-preview-2024-12-17": { - "input_cost_per_audio_token": 0.00004, - "input_cost_per_token": 0.0000025, + "input_cost_per_audio_token": 4e-5, + "input_cost_per_token": 2.5e-6, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_audio_token": 0.00008, - "output_cost_per_token": 0.00001, - "supported_endpoints": [ - "/v1/chat/completions" - ], - "supported_modalities": [ - "text", - "audio" - ], - "supported_output_modalities": [ - "text", - "audio" - ], + "output_cost_per_audio_token": 8e-5, + "output_cost_per_token": 1e-5, + "supported_endpoints": ["/v1/chat/completions"], + "supported_modalities": ["text", "audio"], + "supported_output_modalities": ["text", "audio"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -2428,15 +2201,15 @@ "azure/gpt-4o-mini-realtime-preview-2024-12-17": { "cache_creation_input_audio_token_cost": 3e-7, "cache_read_input_token_cost": 3e-7, - "input_cost_per_audio_token": 0.00001, + "input_cost_per_audio_token": 1e-5, "input_cost_per_token": 6e-7, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_audio_token": 0.00002, - "output_cost_per_token": 0.0000024, + "output_cost_per_audio_token": 2e-5, + "output_cost_per_token": 2.4e-6, "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -2445,30 +2218,21 @@ "supports_tool_choice": true }, "azure/gpt-realtime-2025-08-28": { - "cache_creation_input_audio_token_cost": 0.000004, - "cache_read_input_token_cost": 0.000004, - "input_cost_per_audio_token": 0.000032, - "input_cost_per_image": 0.000005, - "input_cost_per_token": 0.000004, + "cache_creation_input_audio_token_cost": 4e-6, + "cache_read_input_token_cost": 4e-6, + "input_cost_per_audio_token": 3.2e-5, + "input_cost_per_image": 5e-6, + "input_cost_per_token": 4e-6, "litellm_provider": "azure", "max_input_tokens": 32000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_audio_token": 0.000064, - "output_cost_per_token": 0.000016, - "supported_endpoints": [ - "/v1/realtime" - ], - "supported_modalities": [ - "text", - "image", - "audio" - ], - "supported_output_modalities": [ - "text", - "audio" - ], + "output_cost_per_audio_token": 6.4e-5, + "output_cost_per_token": 1.6e-5, + "supported_endpoints": ["/v1/realtime"], + "supported_modalities": ["text", "image", "audio"], + "supported_output_modalities": ["text", "audio"], "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -2479,7 +2243,7 @@ "azure/gpt-realtime-mini-2025-10-06": { "cache_creation_input_audio_token_cost": 3e-7, "cache_read_input_token_cost": 6e-8, - "input_cost_per_audio_token": 0.00001, + "input_cost_per_audio_token": 1e-5, "input_cost_per_image": 8e-7, "input_cost_per_token": 6e-7, "litellm_provider": "azure", @@ -2487,20 +2251,11 @@ "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_audio_token": 0.00002, - "output_cost_per_token": 0.0000024, - "supported_endpoints": [ - "/v1/realtime" - ], - "supported_modalities": [ - "text", - "image", - "audio" - ], - "supported_output_modalities": [ - "text", - "audio" - ], + "output_cost_per_audio_token": 2e-5, + "output_cost_per_token": 2.4e-6, + "supported_endpoints": ["/v1/realtime"], + "supported_modalities": ["text", "image", "audio"], + "supported_output_modalities": ["text", "audio"], "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -2509,47 +2264,38 @@ "supports_tool_choice": true }, "azure/gpt-4o-mini-transcribe": { - "input_cost_per_audio_token": 0.000003, - "input_cost_per_token": 0.00000125, + "input_cost_per_audio_token": 3e-6, + "input_cost_per_token": 1.25e-6, "litellm_provider": "azure", "max_input_tokens": 16000, "max_output_tokens": 2000, "mode": "audio_transcription", - "output_cost_per_token": 0.000005, - "supported_endpoints": [ - "/v1/audio/transcriptions" - ] + "output_cost_per_token": 5e-6, + "supported_endpoints": ["/v1/audio/transcriptions"] }, "azure/gpt-4o-mini-tts": { - "input_cost_per_token": 0.0000025, + "input_cost_per_token": 2.5e-6, "litellm_provider": "azure", "mode": "audio_speech", - "output_cost_per_audio_token": 0.000012, + "output_cost_per_audio_token": 1.2e-5, "output_cost_per_second": 0.00025, - "output_cost_per_token": 0.00001, - "supported_endpoints": [ - "/v1/audio/speech" - ], - "supported_modalities": [ - "text", - "audio" - ], - "supported_output_modalities": [ - "audio" - ] + "output_cost_per_token": 1e-5, + "supported_endpoints": ["/v1/audio/speech"], + "supported_modalities": ["text", "audio"], + "supported_output_modalities": ["audio"] }, "azure/gpt-4o-realtime-preview-2024-10-01": { - "cache_creation_input_audio_token_cost": 0.00002, - "cache_read_input_token_cost": 0.0000025, + "cache_creation_input_audio_token_cost": 2e-5, + "cache_read_input_token_cost": 2.5e-6, "input_cost_per_audio_token": 0.0001, - "input_cost_per_token": 0.000005, + "input_cost_per_token": 5e-6, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", "output_cost_per_audio_token": 0.0002, - "output_cost_per_token": 0.00002, + "output_cost_per_token": 2e-5, "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -2558,24 +2304,18 @@ "supports_tool_choice": true }, "azure/gpt-4o-realtime-preview-2024-12-17": { - "cache_read_input_token_cost": 0.0000025, - "input_cost_per_audio_token": 0.00004, - "input_cost_per_token": 0.000005, + "cache_read_input_token_cost": 2.5e-6, + "input_cost_per_audio_token": 4e-5, + "input_cost_per_token": 5e-6, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_audio_token": 0.00008, - "output_cost_per_token": 0.00002, - "supported_modalities": [ - "text", - "audio" - ], - "supported_output_modalities": [ - "text", - "audio" - ], + "output_cost_per_audio_token": 8e-5, + "output_cost_per_token": 2e-5, + "supported_modalities": ["text", "audio"], + "supported_output_modalities": ["text", "audio"], "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -2584,53 +2324,40 @@ "supports_tool_choice": true }, "azure/gpt-4o-transcribe": { - "input_cost_per_audio_token": 0.000006, - "input_cost_per_token": 0.0000025, + "input_cost_per_audio_token": 6e-6, + "input_cost_per_token": 2.5e-6, "litellm_provider": "azure", "max_input_tokens": 16000, "max_output_tokens": 2000, "mode": "audio_transcription", - "output_cost_per_token": 0.00001, - "supported_endpoints": [ - "/v1/audio/transcriptions" - ] + "output_cost_per_token": 1e-5, + "supported_endpoints": ["/v1/audio/transcriptions"] }, "azure/gpt-4o-transcribe-diarize": { - "input_cost_per_audio_token": 0.000006, - "input_cost_per_token": 0.0000025, + "input_cost_per_audio_token": 6e-6, + "input_cost_per_token": 2.5e-6, "litellm_provider": "azure", "max_input_tokens": 16000, "max_output_tokens": 2000, "mode": "audio_transcription", - "output_cost_per_token": 0.00001, - "supported_endpoints": [ - "/v1/audio/transcriptions" - ] + "output_cost_per_token": 1e-5, + "supported_endpoints": ["/v1/audio/transcriptions"] }, "azure/gpt-5.1-2025-11-13": { "cache_read_input_token_cost": 1.25e-7, "cache_read_input_token_cost_priority": 2.5e-7, - "input_cost_per_token": 0.00000125, - "input_cost_per_token_priority": 0.0000025, + "input_cost_per_token": 1.25e-6, + "input_cost_per_token_priority": 2.5e-6, "litellm_provider": "azure", "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.00001, - "output_cost_per_token_priority": 0.00002, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text", - "image" - ], + "output_cost_per_token": 1e-5, + "output_cost_per_token_priority": 2e-5, + "supported_endpoints": ["/v1/chat/completions", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text", "image"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -2646,27 +2373,18 @@ "azure/gpt-5.1-chat-2025-11-13": { "cache_read_input_token_cost": 1.25e-7, "cache_read_input_token_cost_priority": 2.5e-7, - "input_cost_per_token": 0.00000125, - "input_cost_per_token_priority": 0.0000025, + "input_cost_per_token": 1.25e-6, + "input_cost_per_token_priority": 2.5e-6, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 0.00001, - "output_cost_per_token_priority": 0.00002, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text", - "image" - ], + "output_cost_per_token": 1e-5, + "output_cost_per_token_priority": 2e-5, + "supported_endpoints": ["/v1/chat/completions", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text", "image"], "supports_function_calling": false, "supports_native_streaming": true, "supports_parallel_function_calling": false, @@ -2681,25 +2399,18 @@ "azure/gpt-5.1-codex-2025-11-13": { "cache_read_input_token_cost": 1.25e-7, "cache_read_input_token_cost_priority": 2.5e-7, - "input_cost_per_token": 0.00000125, - "input_cost_per_token_priority": 0.0000025, + "input_cost_per_token": 1.25e-6, + "input_cost_per_token_priority": 2.5e-6, "litellm_provider": "azure", "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "responses", - "output_cost_per_token": 0.00001, - "output_cost_per_token_priority": 0.00002, - "supported_endpoints": [ - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 1e-5, + "output_cost_per_token_priority": 2e-5, + "supported_endpoints": ["/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -2721,18 +2432,11 @@ "max_output_tokens": 128000, "max_tokens": 128000, "mode": "responses", - "output_cost_per_token": 0.000002, - "output_cost_per_token_priority": 0.0000036, - "supported_endpoints": [ - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 2e-6, + "output_cost_per_token_priority": 3.6e-6, + "supported_endpoints": ["/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -2746,25 +2450,16 @@ }, "azure/gpt-5": { "cache_read_input_token_cost": 1.25e-7, - "input_cost_per_token": 0.00000125, + "input_cost_per_token": 1.25e-6, "litellm_provider": "azure", "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.00001, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 1e-5, + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -2778,25 +2473,16 @@ }, "azure/gpt-5-2025-08-07": { "cache_read_input_token_cost": 1.25e-7, - "input_cost_per_token": 0.00000125, + "input_cost_per_token": 1.25e-6, "litellm_provider": "azure", "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.00001, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 1e-5, + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -2810,26 +2496,17 @@ }, "azure/gpt-5-chat": { "cache_read_input_token_cost": 1.25e-7, - "input_cost_per_token": 0.00000125, + "input_cost_per_token": 1.25e-6, "litellm_provider": "azure", "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.00001, + "output_cost_per_token": 1e-5, "source": "https://azure.microsoft.com/en-us/blog/gpt-5-in-azure-ai-foundry-the-future-of-ai-apps-and-agents-starts-here/", - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -2843,25 +2520,16 @@ }, "azure/gpt-5-chat-latest": { "cache_read_input_token_cost": 1.25e-7, - "input_cost_per_token": 0.00000125, + "input_cost_per_token": 1.25e-6, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 0.00001, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 1e-5, + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -2875,23 +2543,16 @@ }, "azure/gpt-5-codex": { "cache_read_input_token_cost": 1.25e-7, - "input_cost_per_token": 0.00000125, + "input_cost_per_token": 1.25e-6, "litellm_provider": "azure", "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "responses", - "output_cost_per_token": 0.00001, - "supported_endpoints": [ - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 1e-5, + "supported_endpoints": ["/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -2911,19 +2572,10 @@ "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.000002, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 2e-6, + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -2943,19 +2595,10 @@ "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.000002, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 2e-6, + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -2976,18 +2619,9 @@ "max_tokens": 128000, "mode": "chat", "output_cost_per_token": 4e-7, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -3008,18 +2642,9 @@ "max_tokens": 128000, "mode": "chat", "output_cost_per_token": 4e-7, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -3032,7 +2657,7 @@ "supports_vision": true }, "azure/gpt-5-pro": { - "input_cost_per_token": 0.000015, + "input_cost_per_token": 1.5e-5, "litellm_provider": "azure", "max_input_tokens": 272000, "max_output_tokens": 128000, @@ -3040,16 +2665,9 @@ "mode": "responses", "output_cost_per_token": 0.00012, "source": "https://learn.microsoft.com/en-us/azure/ai-foundry/foundry-models/concepts/models-sold-directly-by-azure?pivots=azure-openai&tabs=global-standard-aoai%2Cstandard-chat-completions%2Cglobal-standard#gpt-5", - "supported_endpoints": [ - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "supported_endpoints": ["/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_pdf_input": true, @@ -3062,26 +2680,16 @@ }, "azure/gpt-5.1": { "cache_read_input_token_cost": 1.25e-7, - "input_cost_per_token": 0.00000125, + "input_cost_per_token": 1.25e-6, "litellm_provider": "azure", "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.00001, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text", - "image" - ], + "output_cost_per_token": 1e-5, + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text", "image"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -3095,26 +2703,16 @@ }, "azure/gpt-5.1-chat": { "cache_read_input_token_cost": 1.25e-7, - "input_cost_per_token": 0.00000125, + "input_cost_per_token": 1.25e-6, "litellm_provider": "azure", "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.00001, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text", - "image" - ], + "output_cost_per_token": 1e-5, + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text", "image"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -3128,23 +2726,16 @@ }, "azure/gpt-5.1-codex": { "cache_read_input_token_cost": 1.25e-7, - "input_cost_per_token": 0.00000125, + "input_cost_per_token": 1.25e-6, "litellm_provider": "azure", "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "responses", - "output_cost_per_token": 0.00001, - "supported_endpoints": [ - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 1e-5, + "supported_endpoints": ["/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -3164,17 +2755,10 @@ "max_output_tokens": 128000, "max_tokens": 128000, "mode": "responses", - "output_cost_per_token": 0.000002, - "supported_endpoints": [ - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 2e-6, + "supported_endpoints": ["/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -3190,227 +2774,187 @@ "input_cost_per_pixel": 4.0054321e-8, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_pixel": 0, - "supported_endpoints": [ - "/v1/images/generations" - ] + "output_cost_per_pixel": 0.0, + "supported_endpoints": ["/v1/images/generations"] }, "azure/hd/1024-x-1024/dall-e-3": { "input_cost_per_pixel": 7.629e-8, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "azure/hd/1024-x-1792/dall-e-3": { "input_cost_per_pixel": 6.539e-8, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "azure/hd/1792-x-1024/dall-e-3": { "input_cost_per_pixel": 6.539e-8, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "azure/high/1024-x-1024/gpt-image-1": { "input_cost_per_pixel": 1.59263611e-7, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_pixel": 0, - "supported_endpoints": [ - "/v1/images/generations" - ] + "output_cost_per_pixel": 0.0, + "supported_endpoints": ["/v1/images/generations"] }, "azure/high/1024-x-1536/gpt-image-1": { "input_cost_per_pixel": 1.58945719e-7, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_pixel": 0, - "supported_endpoints": [ - "/v1/images/generations" - ] + "output_cost_per_pixel": 0.0, + "supported_endpoints": ["/v1/images/generations"] }, "azure/high/1536-x-1024/gpt-image-1": { "input_cost_per_pixel": 1.58945719e-7, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_pixel": 0, - "supported_endpoints": [ - "/v1/images/generations" - ] + "output_cost_per_pixel": 0.0, + "supported_endpoints": ["/v1/images/generations"] }, "azure/low/1024-x-1024/gpt-image-1": { "input_cost_per_pixel": 1.0490417e-8, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_pixel": 0, - "supported_endpoints": [ - "/v1/images/generations" - ] + "output_cost_per_pixel": 0.0, + "supported_endpoints": ["/v1/images/generations"] }, "azure/low/1024-x-1536/gpt-image-1": { "input_cost_per_pixel": 1.0172526e-8, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_pixel": 0, - "supported_endpoints": [ - "/v1/images/generations" - ] + "output_cost_per_pixel": 0.0, + "supported_endpoints": ["/v1/images/generations"] }, "azure/low/1536-x-1024/gpt-image-1": { "input_cost_per_pixel": 1.0172526e-8, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_pixel": 0, - "supported_endpoints": [ - "/v1/images/generations" - ] + "output_cost_per_pixel": 0.0, + "supported_endpoints": ["/v1/images/generations"] }, "azure/medium/1024-x-1024/gpt-image-1": { "input_cost_per_pixel": 4.0054321e-8, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_pixel": 0, - "supported_endpoints": [ - "/v1/images/generations" - ] + "output_cost_per_pixel": 0.0, + "supported_endpoints": ["/v1/images/generations"] }, "azure/medium/1024-x-1536/gpt-image-1": { "input_cost_per_pixel": 4.0054321e-8, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_pixel": 0, - "supported_endpoints": [ - "/v1/images/generations" - ] + "output_cost_per_pixel": 0.0, + "supported_endpoints": ["/v1/images/generations"] }, "azure/medium/1536-x-1024/gpt-image-1": { "input_cost_per_pixel": 4.0054321e-8, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_pixel": 0, - "supported_endpoints": [ - "/v1/images/generations" - ] + "output_cost_per_pixel": 0.0, + "supported_endpoints": ["/v1/images/generations"] }, "azure/gpt-image-1-mini": { "input_cost_per_pixel": 8.0566406e-9, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_pixel": 0, - "supported_endpoints": [ - "/v1/images/generations" - ] + "output_cost_per_pixel": 0.0, + "supported_endpoints": ["/v1/images/generations"] }, "azure/low/1024-x-1024/gpt-image-1-mini": { "input_cost_per_pixel": 2.0751953125e-9, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_pixel": 0, - "supported_endpoints": [ - "/v1/images/generations" - ] + "output_cost_per_pixel": 0.0, + "supported_endpoints": ["/v1/images/generations"] }, "azure/low/1024-x-1536/gpt-image-1-mini": { "input_cost_per_pixel": 2.0751953125e-9, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_pixel": 0, - "supported_endpoints": [ - "/v1/images/generations" - ] + "output_cost_per_pixel": 0.0, + "supported_endpoints": ["/v1/images/generations"] }, "azure/low/1536-x-1024/gpt-image-1-mini": { "input_cost_per_pixel": 2.0345052083e-9, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_pixel": 0, - "supported_endpoints": [ - "/v1/images/generations" - ] + "output_cost_per_pixel": 0.0, + "supported_endpoints": ["/v1/images/generations"] }, "azure/medium/1024-x-1024/gpt-image-1-mini": { "input_cost_per_pixel": 8.056640625e-9, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_pixel": 0, - "supported_endpoints": [ - "/v1/images/generations" - ] + "output_cost_per_pixel": 0.0, + "supported_endpoints": ["/v1/images/generations"] }, "azure/medium/1024-x-1536/gpt-image-1-mini": { "input_cost_per_pixel": 8.056640625e-9, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_pixel": 0, - "supported_endpoints": [ - "/v1/images/generations" - ] + "output_cost_per_pixel": 0.0, + "supported_endpoints": ["/v1/images/generations"] }, "azure/medium/1536-x-1024/gpt-image-1-mini": { "input_cost_per_pixel": 7.9752604167e-9, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_pixel": 0, - "supported_endpoints": [ - "/v1/images/generations" - ] + "output_cost_per_pixel": 0.0, + "supported_endpoints": ["/v1/images/generations"] }, "azure/high/1024-x-1024/gpt-image-1-mini": { "input_cost_per_pixel": 3.173828125e-8, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_pixel": 0, - "supported_endpoints": [ - "/v1/images/generations" - ] + "output_cost_per_pixel": 0.0, + "supported_endpoints": ["/v1/images/generations"] }, "azure/high/1024-x-1536/gpt-image-1-mini": { "input_cost_per_pixel": 3.173828125e-8, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_pixel": 0, - "supported_endpoints": [ - "/v1/images/generations" - ] + "output_cost_per_pixel": 0.0, + "supported_endpoints": ["/v1/images/generations"] }, "azure/high/1536-x-1024/gpt-image-1-mini": { "input_cost_per_pixel": 3.1575520833e-8, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_pixel": 0, - "supported_endpoints": [ - "/v1/images/generations" - ] + "output_cost_per_pixel": 0.0, + "supported_endpoints": ["/v1/images/generations"] }, "azure/mistral-large-2402": { - "input_cost_per_token": 0.000008, + "input_cost_per_token": 8e-6, "litellm_provider": "azure", "max_input_tokens": 32000, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 0.000024, + "output_cost_per_token": 2.4e-5, "supports_function_calling": true }, "azure/mistral-large-latest": { - "input_cost_per_token": 0.000008, + "input_cost_per_token": 8e-6, "litellm_provider": "azure", "max_input_tokens": 32000, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 0.000024, + "output_cost_per_token": 2.4e-5, "supports_function_calling": true }, "azure/o1": { - "cache_read_input_token_cost": 0.0000075, - "input_cost_per_token": 0.000015, + "cache_read_input_token_cost": 7.5e-6, + "input_cost_per_token": 1.5e-5, "litellm_provider": "azure", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 0.00006, + "output_cost_per_token": 6e-5, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -3419,14 +2963,14 @@ "supports_vision": true }, "azure/o1-2024-12-17": { - "cache_read_input_token_cost": 0.0000075, - "input_cost_per_token": 0.000015, + "cache_read_input_token_cost": 7.5e-6, + "input_cost_per_token": 1.5e-5, "litellm_provider": "azure", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 0.00006, + "output_cost_per_token": 6e-5, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -3436,13 +2980,13 @@ }, "azure/o1-mini": { "cache_read_input_token_cost": 6.05e-7, - "input_cost_per_token": 0.00000121, + "input_cost_per_token": 1.21e-6, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 65536, "max_tokens": 65536, "mode": "chat", - "output_cost_per_token": 0.00000484, + "output_cost_per_token": 4.84e-6, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -3451,13 +2995,13 @@ }, "azure/o1-mini-2024-09-12": { "cache_read_input_token_cost": 5.5e-7, - "input_cost_per_token": 0.0000011, + "input_cost_per_token": 1.1e-6, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 65536, "max_tokens": 65536, "mode": "chat", - "output_cost_per_token": 0.0000044, + "output_cost_per_token": 4.4e-6, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -3465,14 +3009,14 @@ "supports_vision": false }, "azure/o1-preview": { - "cache_read_input_token_cost": 0.0000075, - "input_cost_per_token": 0.000015, + "cache_read_input_token_cost": 7.5e-6, + "input_cost_per_token": 1.5e-5, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0.00006, + "output_cost_per_token": 6e-5, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -3480,14 +3024,14 @@ "supports_vision": false }, "azure/o1-preview-2024-09-12": { - "cache_read_input_token_cost": 0.0000075, - "input_cost_per_token": 0.000015, + "cache_read_input_token_cost": 7.5e-6, + "input_cost_per_token": 1.5e-5, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0.00006, + "output_cost_per_token": 6e-5, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_pdf_input": true, @@ -3497,25 +3041,16 @@ }, "azure/o3": { "cache_read_input_token_cost": 5e-7, - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "azure", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 0.000008, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 8e-6, + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_parallel_function_calling": false, "supports_prompt_caching": true, @@ -3527,25 +3062,16 @@ "azure/o3-2025-04-16": { "deprecation_date": "2026-04-16", "cache_read_input_token_cost": 5e-7, - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "azure", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 0.000008, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 8e-6, + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_parallel_function_calling": false, "supports_prompt_caching": true, @@ -3555,26 +3081,17 @@ "supports_vision": true }, "azure/o3-deep-research": { - "cache_read_input_token_cost": 0.0000025, - "input_cost_per_token": 0.00001, + "cache_read_input_token_cost": 2.5e-6, + "input_cost_per_token": 1e-5, "litellm_provider": "azure", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "responses", - "output_cost_per_token": 0.00004, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 4e-5, + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_pdf_input": true, @@ -3588,13 +3105,13 @@ }, "azure/o3-mini": { "cache_read_input_token_cost": 5.5e-7, - "input_cost_per_token": 0.0000011, + "input_cost_per_token": 1.1e-6, "litellm_provider": "azure", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 0.0000044, + "output_cost_per_token": 4.4e-6, "supports_prompt_caching": true, "supports_reasoning": true, "supports_response_schema": true, @@ -3603,40 +3120,31 @@ }, "azure/o3-mini-2025-01-31": { "cache_read_input_token_cost": 5.5e-7, - "input_cost_per_token": 0.0000011, + "input_cost_per_token": 1.1e-6, "litellm_provider": "azure", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 0.0000044, + "output_cost_per_token": 4.4e-6, "supports_prompt_caching": true, "supports_reasoning": true, "supports_tool_choice": true, "supports_vision": false }, "azure/o3-pro": { - "input_cost_per_token": 0.00002, - "input_cost_per_token_batches": 0.00001, + "input_cost_per_token": 2e-5, + "input_cost_per_token_batches": 1e-5, "litellm_provider": "azure", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "responses", - "output_cost_per_token": 0.00008, - "output_cost_per_token_batches": 0.00004, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 8e-5, + "output_cost_per_token_batches": 4e-5, + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_parallel_function_calling": false, "supports_prompt_caching": false, @@ -3646,27 +3154,18 @@ "supports_vision": true }, "azure/o3-pro-2025-06-10": { - "input_cost_per_token": 0.00002, - "input_cost_per_token_batches": 0.00001, + "input_cost_per_token": 2e-5, + "input_cost_per_token_batches": 1e-5, "litellm_provider": "azure", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "responses", - "output_cost_per_token": 0.00008, - "output_cost_per_token_batches": 0.00004, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 8e-5, + "output_cost_per_token_batches": 4e-5, + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_parallel_function_calling": false, "supports_prompt_caching": false, @@ -3677,25 +3176,16 @@ }, "azure/o4-mini": { "cache_read_input_token_cost": 2.75e-7, - "input_cost_per_token": 0.0000011, + "input_cost_per_token": 1.1e-6, "litellm_provider": "azure", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 0.0000044, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 4.4e-6, + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_parallel_function_calling": false, "supports_prompt_caching": true, @@ -3706,13 +3196,13 @@ }, "azure/o4-mini-2025-04-16": { "cache_read_input_token_cost": 2.75e-7, - "input_cost_per_token": 0.0000011, + "input_cost_per_token": 1.1e-6, "litellm_provider": "azure", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 0.0000044, + "output_cost_per_token": 4.4e-6, "supports_function_calling": true, "supports_parallel_function_calling": false, "supports_prompt_caching": true, @@ -3722,28 +3212,28 @@ "supports_vision": true }, "azure/standard/1024-x-1024/dall-e-2": { - "input_cost_per_pixel": 0, + "input_cost_per_pixel": 0.0, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "azure/standard/1024-x-1024/dall-e-3": { "input_cost_per_pixel": 3.81469e-8, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "azure/standard/1024-x-1792/dall-e-3": { "input_cost_per_pixel": 4.359e-8, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "azure/standard/1792-x-1024/dall-e-3": { "input_cost_per_pixel": 4.359e-8, "litellm_provider": "azure", "mode": "image_generation", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "azure/text-embedding-3-large": { "input_cost_per_token": 1.3e-7, @@ -3751,7 +3241,7 @@ "max_input_tokens": 8191, "max_tokens": 8191, "mode": "embedding", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "azure/text-embedding-3-small": { "deprecation_date": "2026-04-30", @@ -3760,7 +3250,7 @@ "max_input_tokens": 8191, "max_tokens": 8191, "mode": "embedding", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "azure/text-embedding-ada-002": { "input_cost_per_token": 1e-7, @@ -3768,54 +3258,45 @@ "max_input_tokens": 8191, "max_tokens": 8191, "mode": "embedding", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "azure/speech/azure-tts": { - "input_cost_per_character": 0.000015, + "input_cost_per_character": 15e-6, "litellm_provider": "azure", "mode": "audio_speech", "source": "https://azure.microsoft.com/en-us/pricing/calculator/" }, "azure/speech/azure-tts-hd": { - "input_cost_per_character": 0.00003, + "input_cost_per_character": 30e-6, "litellm_provider": "azure", "mode": "audio_speech", "source": "https://azure.microsoft.com/en-us/pricing/calculator/" }, "azure/tts-1": { - "input_cost_per_character": 0.000015, + "input_cost_per_character": 1.5e-5, "litellm_provider": "azure", "mode": "audio_speech" }, "azure/tts-1-hd": { - "input_cost_per_character": 0.00003, + "input_cost_per_character": 3e-5, "litellm_provider": "azure", "mode": "audio_speech" }, "azure/us/gpt-4.1-2025-04-14": { "deprecation_date": "2026-11-04", "cache_read_input_token_cost": 5.5e-7, - "input_cost_per_token": 0.0000022, - "input_cost_per_token_batches": 0.0000011, + "input_cost_per_token": 2.2e-6, + "input_cost_per_token_batches": 1.1e-6, "litellm_provider": "azure", "max_input_tokens": 1047576, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0.0000088, - "output_cost_per_token_batches": 0.0000044, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 8.8e-6, + "output_cost_per_token_batches": 4.4e-6, + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -3836,20 +3317,11 @@ "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0.00000176, + "output_cost_per_token": 1.76e-6, "output_cost_per_token_batches": 8.8e-7, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -3872,18 +3344,9 @@ "mode": "chat", "output_cost_per_token": 4.4e-7, "output_cost_per_token_batches": 2.2e-7, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -3895,14 +3358,14 @@ }, "azure/us/gpt-4o-2024-08-06": { "deprecation_date": "2026-02-27", - "cache_read_input_token_cost": 0.000001375, - "input_cost_per_token": 0.00000275, + "cache_read_input_token_cost": 1.375e-6, + "input_cost_per_token": 2.75e-6, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 0.000011, + "output_cost_per_token": 1.1e-5, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -3912,14 +3375,14 @@ }, "azure/us/gpt-4o-2024-11-20": { "deprecation_date": "2026-03-01", - "cache_creation_input_token_cost": 0.00000138, - "input_cost_per_token": 0.00000275, + "cache_creation_input_token_cost": 1.38e-6, + "input_cost_per_token": 2.75e-6, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 0.000011, + "output_cost_per_token": 1.1e-5, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_response_schema": true, @@ -3945,15 +3408,15 @@ "azure/us/gpt-4o-mini-realtime-preview-2024-12-17": { "cache_creation_input_audio_token_cost": 3.3e-7, "cache_read_input_token_cost": 3.3e-7, - "input_cost_per_audio_token": 0.000011, + "input_cost_per_audio_token": 1.1e-5, "input_cost_per_token": 6.6e-7, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_audio_token": 0.000022, - "output_cost_per_token": 0.00000264, + "output_cost_per_audio_token": 2.2e-5, + "output_cost_per_token": 2.64e-6, "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -3962,17 +3425,17 @@ "supports_tool_choice": true }, "azure/us/gpt-4o-realtime-preview-2024-10-01": { - "cache_creation_input_audio_token_cost": 0.000022, - "cache_read_input_token_cost": 0.00000275, + "cache_creation_input_audio_token_cost": 2.2e-5, + "cache_read_input_token_cost": 2.75e-6, "input_cost_per_audio_token": 0.00011, - "input_cost_per_token": 0.0000055, + "input_cost_per_token": 5.5e-6, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", "output_cost_per_audio_token": 0.00022, - "output_cost_per_token": 0.000022, + "output_cost_per_token": 2.2e-5, "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -3981,25 +3444,19 @@ "supports_tool_choice": true }, "azure/us/gpt-4o-realtime-preview-2024-12-17": { - "cache_read_input_audio_token_cost": 0.0000025, - "cache_read_input_token_cost": 0.00000275, - "input_cost_per_audio_token": 0.000044, - "input_cost_per_token": 0.0000055, + "cache_read_input_audio_token_cost": 2.5e-6, + "cache_read_input_token_cost": 2.75e-6, + "input_cost_per_audio_token": 4.4e-5, + "input_cost_per_token": 5.5e-6, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_audio_token": 0.00008, - "output_cost_per_token": 0.000022, - "supported_modalities": [ - "text", - "audio" - ], - "supported_output_modalities": [ - "text", - "audio" - ], + "output_cost_per_audio_token": 8e-5, + "output_cost_per_token": 2.2e-5, + "supported_modalities": ["text", "audio"], + "supported_output_modalities": ["text", "audio"], "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -4009,25 +3466,16 @@ }, "azure/us/gpt-5-2025-08-07": { "cache_read_input_token_cost": 1.375e-7, - "input_cost_per_token": 0.000001375, + "input_cost_per_token": 1.375e-6, "litellm_provider": "azure", "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.000011, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 1.1e-5, + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -4047,19 +3495,10 @@ "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.0000022, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 2.2e-6, + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -4080,18 +3519,9 @@ "max_tokens": 128000, "mode": "chat", "output_cost_per_token": 4.4e-7, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -4105,26 +3535,16 @@ }, "azure/us/gpt-5.1": { "cache_read_input_token_cost": 1.4e-7, - "input_cost_per_token": 0.00000138, + "input_cost_per_token": 1.38e-6, "litellm_provider": "azure", "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, - "mode": "chat", - "output_cost_per_token": 0.000011, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text", - "image" - ], + "mode": "chat", + "output_cost_per_token": 1.1e-5, + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text", "image"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -4138,26 +3558,16 @@ }, "azure/us/gpt-5.1-chat": { "cache_read_input_token_cost": 1.4e-7, - "input_cost_per_token": 0.00000138, + "input_cost_per_token": 1.38e-6, "litellm_provider": "azure", "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.000011, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text", - "image" - ], + "output_cost_per_token": 1.1e-5, + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text", "image"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -4171,23 +3581,16 @@ }, "azure/us/gpt-5.1-codex": { "cache_read_input_token_cost": 1.4e-7, - "input_cost_per_token": 0.00000138, + "input_cost_per_token": 1.38e-6, "litellm_provider": "azure", "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "responses", - "output_cost_per_token": 0.000011, - "supported_endpoints": [ - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 1.1e-5, + "supported_endpoints": ["/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -4207,17 +3610,10 @@ "max_output_tokens": 128000, "max_tokens": 128000, "mode": "responses", - "output_cost_per_token": 0.0000022, - "supported_endpoints": [ - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 2.2e-6, + "supported_endpoints": ["/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -4230,14 +3626,14 @@ "supports_vision": true }, "azure/us/o1-2024-12-17": { - "cache_read_input_token_cost": 0.00000825, - "input_cost_per_token": 0.0000165, + "cache_read_input_token_cost": 8.25e-6, + "input_cost_per_token": 1.65e-5, "litellm_provider": "azure", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 0.000066, + "output_cost_per_token": 6.6e-5, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -4246,29 +3642,29 @@ }, "azure/us/o1-mini-2024-09-12": { "cache_read_input_token_cost": 6.05e-7, - "input_cost_per_token": 0.00000121, + "input_cost_per_token": 1.21e-6, "input_cost_per_token_batches": 6.05e-7, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 65536, "max_tokens": 65536, "mode": "chat", - "output_cost_per_token": 0.00000484, - "output_cost_per_token_batches": 0.00000242, + "output_cost_per_token": 4.84e-6, + "output_cost_per_token_batches": 2.42e-6, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, "supports_vision": false }, "azure/us/o1-preview-2024-09-12": { - "cache_read_input_token_cost": 0.00000825, - "input_cost_per_token": 0.0000165, + "cache_read_input_token_cost": 8.25e-6, + "input_cost_per_token": 1.65e-5, "litellm_provider": "azure", "max_input_tokens": 128000, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0.000066, + "output_cost_per_token": 6.6e-5, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -4277,25 +3673,16 @@ "azure/us/o3-2025-04-16": { "deprecation_date": "2026-04-16", "cache_read_input_token_cost": 5.5e-7, - "input_cost_per_token": 0.0000022, + "input_cost_per_token": 2.2e-6, "litellm_provider": "azure", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 0.0000088, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 8.8e-6, + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_parallel_function_calling": false, "supports_prompt_caching": true, @@ -4306,15 +3693,15 @@ }, "azure/us/o3-mini-2025-01-31": { "cache_read_input_token_cost": 6.05e-7, - "input_cost_per_token": 0.00000121, + "input_cost_per_token": 1.21e-6, "input_cost_per_token_batches": 6.05e-7, "litellm_provider": "azure", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 0.00000484, - "output_cost_per_token_batches": 0.00000242, + "output_cost_per_token": 4.84e-6, + "output_cost_per_token_batches": 2.42e-6, "supports_prompt_caching": true, "supports_reasoning": true, "supports_tool_choice": true, @@ -4322,13 +3709,13 @@ }, "azure/us/o4-mini-2025-04-16": { "cache_read_input_token_cost": 3.1e-7, - "input_cost_per_token": 0.00000121, + "input_cost_per_token": 1.21e-6, "litellm_provider": "azure", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 0.00000484, + "output_cost_per_token": 4.84e-6, "supports_function_calling": true, "supports_parallel_function_calling": false, "supports_prompt_caching": true, @@ -4349,7 +3736,7 @@ "max_input_tokens": 512, "max_tokens": 512, "mode": "embedding", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "output_vector_size": 1024, "source": "https://azuremarketplace.microsoft.com/en-us/marketplace/apps/cohere.cohere-embed-v3-english-offer?tab=PlansAndPrice", "supports_embedding_image_input": true @@ -4360,7 +3747,7 @@ "max_input_tokens": 512, "max_tokens": 512, "mode": "embedding", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "output_vector_size": 1024, "source": "https://azuremarketplace.microsoft.com/en-us/marketplace/apps/cohere.cohere-embed-v3-english-offer?tab=PlansAndPrice", "supports_embedding_image_input": true @@ -4370,18 +3757,14 @@ "mode": "image_generation", "output_cost_per_image": 0.04, "source": "https://techcommunity.microsoft.com/blog/azure-ai-foundry-blog/black-forest-labs-flux-1-kontext-pro-and-flux1-1-pro-now-available-in-azure-ai-f/4434659", - "supported_endpoints": [ - "/v1/images/generations" - ] + "supported_endpoints": ["/v1/images/generations"] }, "azure_ai/FLUX.1-Kontext-pro": { "litellm_provider": "azure_ai", "mode": "image_generation", "output_cost_per_image": 0.04, "source": "https://azuremarketplace.microsoft.com/pt-br/marketplace/apps/cohere.cohere-embed-4-offer?tab=PlansAndPrice", - "supported_endpoints": [ - "/v1/images/generations" - ] + "supported_endpoints": ["/v1/images/generations"] }, "azure_ai/Llama-3.2-11B-Vision-Instruct": { "input_cost_per_token": 3.7e-7, @@ -4397,13 +3780,13 @@ "supports_vision": true }, "azure_ai/Llama-3.2-90B-Vision-Instruct": { - "input_cost_per_token": 0.00000204, + "input_cost_per_token": 2.04e-6, "litellm_provider": "azure_ai", "max_input_tokens": 128000, "max_output_tokens": 2048, "max_tokens": 2048, "mode": "chat", - "output_cost_per_token": 0.00000204, + "output_cost_per_token": 2.04e-6, "source": "https://azuremarketplace.microsoft.com/en/marketplace/apps/metagenai.meta-llama-3-2-90b-vision-instruct-offer?tab=Overview", "supports_function_calling": true, "supports_tool_choice": true, @@ -4422,7 +3805,7 @@ "supports_tool_choice": true }, "azure_ai/Llama-4-Maverick-17B-128E-Instruct-FP8": { - "input_cost_per_token": 0.00000141, + "input_cost_per_token": 1.41e-6, "litellm_provider": "azure_ai", "max_input_tokens": 1000000, "max_output_tokens": 16384, @@ -4448,7 +3831,7 @@ "supports_vision": true }, "azure_ai/Meta-Llama-3-70B-Instruct": { - "input_cost_per_token": 0.0000011, + "input_cost_per_token": 1.1e-6, "litellm_provider": "azure_ai", "max_input_tokens": 8192, "max_output_tokens": 2048, @@ -4458,24 +3841,24 @@ "supports_tool_choice": true }, "azure_ai/Meta-Llama-3.1-405B-Instruct": { - "input_cost_per_token": 0.00000533, + "input_cost_per_token": 5.33e-6, "litellm_provider": "azure_ai", "max_input_tokens": 128000, "max_output_tokens": 2048, "max_tokens": 2048, "mode": "chat", - "output_cost_per_token": 0.000016, + "output_cost_per_token": 1.6e-5, "source": "https://azuremarketplace.microsoft.com/en-us/marketplace/apps/metagenai.meta-llama-3-1-405b-instruct-offer?tab=PlansAndPrice", "supports_tool_choice": true }, "azure_ai/Meta-Llama-3.1-70B-Instruct": { - "input_cost_per_token": 0.00000268, + "input_cost_per_token": 2.68e-6, "litellm_provider": "azure_ai", "max_input_tokens": 128000, "max_output_tokens": 2048, "max_tokens": 2048, "mode": "chat", - "output_cost_per_token": 0.00000354, + "output_cost_per_token": 3.54e-6, "source": "https://azuremarketplace.microsoft.com/en-us/marketplace/apps/metagenai.meta-llama-3-1-70b-instruct-offer?tab=PlansAndPrice", "supports_tool_choice": true }, @@ -4623,7 +4006,7 @@ "supports_function_calling": true }, "azure_ai/Phi-4-multimodal-instruct": { - "input_cost_per_audio_token": 0.000004, + "input_cost_per_audio_token": 4e-6, "input_cost_per_token": 8e-8, "litellm_provider": "azure_ai", "max_input_tokens": 131072, @@ -4662,116 +4045,108 @@ }, "azure_ai/mistral-document-ai-2505": { "litellm_provider": "azure_ai", - "ocr_cost_per_page": 0.003, + "ocr_cost_per_page": 3e-3, "mode": "ocr", - "supported_endpoints": [ - "/v1/ocr" - ], + "supported_endpoints": ["/v1/ocr"], "source": "https://devblogs.microsoft.com/foundry/whats-new-in-azure-ai-foundry-august-2025/#mistral-document-ai-(ocr)-%E2%80%94-serverless-in-foundry" }, "azure_ai/doc-intelligence/prebuilt-read": { "litellm_provider": "azure_ai", - "ocr_cost_per_page": 0.0015, + "ocr_cost_per_page": 1.5e-3, "mode": "ocr", - "supported_endpoints": [ - "/v1/ocr" - ], + "supported_endpoints": ["/v1/ocr"], "source": "https://azure.microsoft.com/en-us/pricing/details/ai-document-intelligence/" }, "azure_ai/doc-intelligence/prebuilt-layout": { "litellm_provider": "azure_ai", - "ocr_cost_per_page": 0.01, + "ocr_cost_per_page": 1e-2, "mode": "ocr", - "supported_endpoints": [ - "/v1/ocr" - ], + "supported_endpoints": ["/v1/ocr"], "source": "https://azure.microsoft.com/en-us/pricing/details/ai-document-intelligence/" }, "azure_ai/doc-intelligence/prebuilt-document": { "litellm_provider": "azure_ai", - "ocr_cost_per_page": 0.01, + "ocr_cost_per_page": 1e-2, "mode": "ocr", - "supported_endpoints": [ - "/v1/ocr" - ], + "supported_endpoints": ["/v1/ocr"], "source": "https://azure.microsoft.com/en-us/pricing/details/ai-document-intelligence/" }, "azure_ai/MAI-DS-R1": { - "input_cost_per_token": 0.00000135, + "input_cost_per_token": 1.35e-6, "litellm_provider": "azure_ai", "max_input_tokens": 128000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.0000054, + "output_cost_per_token": 5.4e-6, "source": "https://azure.microsoft.com/en-us/pricing/details/ai-foundry-models/microsoft/", "supports_reasoning": true, "supports_tool_choice": true }, "azure_ai/cohere-rerank-v3-english": { "input_cost_per_query": 0.002, - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "azure_ai", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_query_tokens": 2048, "max_tokens": 4096, "mode": "rerank", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "azure_ai/cohere-rerank-v3-multilingual": { "input_cost_per_query": 0.002, - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "azure_ai", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_query_tokens": 2048, "max_tokens": 4096, "mode": "rerank", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "azure_ai/cohere-rerank-v3.5": { "input_cost_per_query": 0.002, - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "azure_ai", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_query_tokens": 2048, "max_tokens": 4096, "mode": "rerank", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "azure_ai/deepseek-r1": { - "input_cost_per_token": 0.00000135, + "input_cost_per_token": 1.35e-6, "litellm_provider": "azure_ai", "max_input_tokens": 128000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.0000054, + "output_cost_per_token": 5.4e-6, "source": "https://techcommunity.microsoft.com/blog/machinelearningblog/deepseek-r1-improved-performance-higher-limits-and-transparent-pricing/4386367", "supports_reasoning": true, "supports_tool_choice": true }, "azure_ai/deepseek-v3": { - "input_cost_per_token": 0.00000114, + "input_cost_per_token": 1.14e-6, "litellm_provider": "azure_ai", "max_input_tokens": 128000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.00000456, + "output_cost_per_token": 4.56e-6, "source": "https://techcommunity.microsoft.com/blog/machinelearningblog/announcing-deepseek-v3-on-azure-ai-foundry-and-github/4390438", "supports_tool_choice": true }, "azure_ai/deepseek-v3-0324": { - "input_cost_per_token": 0.00000114, + "input_cost_per_token": 1.14e-6, "litellm_provider": "azure_ai", "max_input_tokens": 128000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.00000456, + "output_cost_per_token": 4.56e-6, "source": "https://techcommunity.microsoft.com/blog/machinelearningblog/announcing-deepseek-v3-on-azure-ai-foundry-and-github/4390438", "supports_function_calling": true, "supports_tool_choice": true @@ -4782,26 +4157,21 @@ "max_input_tokens": 128000, "max_tokens": 128000, "mode": "embedding", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "output_vector_size": 3072, "source": "https://azuremarketplace.microsoft.com/pt-br/marketplace/apps/cohere.cohere-embed-4-offer?tab=PlansAndPrice", - "supported_endpoints": [ - "/v1/embeddings" - ], - "supported_modalities": [ - "text", - "image" - ], + "supported_endpoints": ["/v1/embeddings"], + "supported_modalities": ["text", "image"], "supports_embedding_image_input": true }, "azure_ai/global/grok-3": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "azure_ai", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "source": "https://devblogs.microsoft.com/foundry/announcing-grok-3-and-grok-3-mini-on-azure-ai-foundry/", "supports_function_calling": true, "supports_response_schema": false, @@ -4815,7 +4185,7 @@ "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0.00000127, + "output_cost_per_token": 1.27e-6, "source": "https://devblogs.microsoft.com/foundry/announcing-grok-3-and-grok-3-mini-on-azure-ai-foundry/", "supports_function_calling": true, "supports_reasoning": true, @@ -4824,13 +4194,13 @@ "supports_web_search": true }, "azure_ai/grok-3": { - "input_cost_per_token": 0.0000033, + "input_cost_per_token": 3.3e-6, "litellm_provider": "azure_ai", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0.0000165, + "output_cost_per_token": 1.65e-5, "source": "https://devblogs.microsoft.com/foundry/announcing-grok-3-and-grok-3-mini-on-azure-ai-foundry/", "supports_function_calling": true, "supports_response_schema": false, @@ -4844,7 +4214,7 @@ "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0.00000138, + "output_cost_per_token": 1.38e-6, "source": "https://devblogs.microsoft.com/foundry/announcing-grok-3-and-grok-3-mini-on-azure-ai-foundry/", "supports_function_calling": true, "supports_reasoning": true, @@ -4853,13 +4223,13 @@ "supports_web_search": true }, "azure_ai/grok-4": { - "input_cost_per_token": 0.0000055, + "input_cost_per_token": 5.5e-6, "litellm_provider": "azure_ai", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0.0000275, + "output_cost_per_token": 2.75e-5, "source": "https://azure.microsoft.com/en-us/blog/grok-4-is-now-available-in-azure-ai-foundry-unlock-frontier-intelligence-and-business-ready-capabilities/", "supports_function_calling": true, "supports_response_schema": true, @@ -4867,8 +4237,8 @@ "supports_web_search": true }, "azure_ai/grok-4-fast-non-reasoning": { - "input_cost_per_token": 4.3e-7, - "output_cost_per_token": 0.00000173, + "input_cost_per_token": 0.43e-6, + "output_cost_per_token": 1.73e-6, "litellm_provider": "azure_ai", "max_input_tokens": 131072, "max_output_tokens": 131072, @@ -4880,8 +4250,8 @@ "supports_web_search": true }, "azure_ai/grok-4-fast-reasoning": { - "input_cost_per_token": 4.3e-7, - "output_cost_per_token": 0.00000173, + "input_cost_per_token": 0.43e-6, + "output_cost_per_token": 1.73e-6, "litellm_provider": "azure_ai", "max_input_tokens": 131072, "max_output_tokens": 131072, @@ -4894,13 +4264,13 @@ "supports_web_search": true }, "azure_ai/grok-code-fast-1": { - "input_cost_per_token": 0.0000035, + "input_cost_per_token": 3.5e-6, "litellm_provider": "azure_ai", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0.0000175, + "output_cost_per_token": 1.75e-5, "source": "https://azure.microsoft.com/en-us/blog/grok-4-is-now-available-in-azure-ai-foundry-unlock-frontier-intelligence-and-business-ready-capabilities/", "supports_function_calling": true, "supports_response_schema": true, @@ -4940,36 +4310,36 @@ "supports_tool_choice": true }, "azure_ai/mistral-large": { - "input_cost_per_token": 0.000004, + "input_cost_per_token": 4e-6, "litellm_provider": "azure_ai", "max_input_tokens": 32000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 0.000012, + "output_cost_per_token": 1.2e-5, "supports_function_calling": true, "supports_tool_choice": true }, "azure_ai/mistral-large-2407": { - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "azure_ai", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.000006, + "output_cost_per_token": 6e-6, "source": "https://azuremarketplace.microsoft.com/en/marketplace/apps/000-000.mistral-ai-large-2407-offer?tab=Overview", "supports_function_calling": true, "supports_tool_choice": true }, "azure_ai/mistral-large-latest": { - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "azure_ai", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.000006, + "output_cost_per_token": 6e-6, "source": "https://azuremarketplace.microsoft.com/en/marketplace/apps/000-000.mistral-ai-large-2407-offer?tab=Overview", "supports_function_calling": true, "supports_tool_choice": true @@ -4981,7 +4351,7 @@ "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 0.000002, + "output_cost_per_token": 2e-6, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_tool_choice": true @@ -4998,24 +4368,24 @@ "supports_function_calling": true }, "azure_ai/mistral-small": { - "input_cost_per_token": 0.000001, + "input_cost_per_token": 1e-6, "litellm_provider": "azure_ai", "max_input_tokens": 32000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 0.000003, + "output_cost_per_token": 3e-6, "supports_function_calling": true, "supports_tool_choice": true }, "azure_ai/mistral-small-2503": { - "input_cost_per_token": 0.000001, + "input_cost_per_token": 1e-6, "litellm_provider": "azure_ai", "max_input_tokens": 128000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.000003, + "output_cost_per_token": 3e-6, "supports_function_calling": true, "supports_tool_choice": true, "supports_vision": true @@ -5128,43 +4498,43 @@ "supports_tool_choice": true }, "bedrock/ap-northeast-1/anthropic.claude-instant-v1": { - "input_cost_per_token": 0.00000223, + "input_cost_per_token": 2.23e-6, "litellm_provider": "bedrock", "max_input_tokens": 100000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 0.00000755, + "output_cost_per_token": 7.55e-6, "supports_tool_choice": true }, "bedrock/ap-northeast-1/anthropic.claude-v1": { - "input_cost_per_token": 0.000008, + "input_cost_per_token": 8e-6, "litellm_provider": "bedrock", "max_input_tokens": 100000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 0.000024, + "output_cost_per_token": 2.4e-5, "supports_tool_choice": true }, "bedrock/ap-northeast-1/anthropic.claude-v2:1": { - "input_cost_per_token": 0.000008, + "input_cost_per_token": 8e-6, "litellm_provider": "bedrock", "max_input_tokens": 100000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 0.000024, + "output_cost_per_token": 2.4e-5, "supports_tool_choice": true }, "bedrock/ap-south-1/meta.llama3-70b-instruct-v1:0": { - "input_cost_per_token": 0.00000318, + "input_cost_per_token": 3.18e-6, "litellm_provider": "bedrock", "max_input_tokens": 8192, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.0000042 + "output_cost_per_token": 4.2e-6 }, "bedrock/ap-south-1/meta.llama3-8b-instruct-v1:0": { "input_cost_per_token": 3.6e-7, @@ -5176,13 +4546,13 @@ "output_cost_per_token": 7.2e-7 }, "bedrock/ca-central-1/meta.llama3-70b-instruct-v1:0": { - "input_cost_per_token": 0.00000305, + "input_cost_per_token": 3.05e-6, "litellm_provider": "bedrock", "max_input_tokens": 8192, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.00000403 + "output_cost_per_token": 4.03e-6 }, "bedrock/ca-central-1/meta.llama3-8b-instruct-v1:0": { "input_cost_per_token": 3.5e-7, @@ -5252,42 +4622,42 @@ "supports_tool_choice": true }, "bedrock/eu-central-1/anthropic.claude-instant-v1": { - "input_cost_per_token": 0.00000248, + "input_cost_per_token": 2.48e-6, "litellm_provider": "bedrock", "max_input_tokens": 100000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 0.00000838, + "output_cost_per_token": 8.38e-6, "supports_tool_choice": true }, "bedrock/eu-central-1/anthropic.claude-v1": { - "input_cost_per_token": 0.000008, + "input_cost_per_token": 8e-6, "litellm_provider": "bedrock", "max_input_tokens": 100000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 0.000024 + "output_cost_per_token": 2.4e-5 }, "bedrock/eu-central-1/anthropic.claude-v2:1": { - "input_cost_per_token": 0.000008, + "input_cost_per_token": 8e-6, "litellm_provider": "bedrock", "max_input_tokens": 100000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 0.000024, + "output_cost_per_token": 2.4e-5, "supports_tool_choice": true }, "bedrock/eu-west-1/meta.llama3-70b-instruct-v1:0": { - "input_cost_per_token": 0.00000286, + "input_cost_per_token": 2.86e-6, "litellm_provider": "bedrock", "max_input_tokens": 8192, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.00000378 + "output_cost_per_token": 3.78e-6 }, "bedrock/eu-west-1/meta.llama3-8b-instruct-v1:0": { "input_cost_per_token": 3.2e-7, @@ -5299,13 +4669,13 @@ "output_cost_per_token": 6.5e-7 }, "bedrock/eu-west-2/meta.llama3-70b-instruct-v1:0": { - "input_cost_per_token": 0.00000345, + "input_cost_per_token": 3.45e-6, "litellm_provider": "bedrock", "max_input_tokens": 8192, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.00000455 + "output_cost_per_token": 4.55e-6 }, "bedrock/eu-west-2/meta.llama3-8b-instruct-v1:0": { "input_cost_per_token": 3.9e-7, @@ -5327,13 +4697,13 @@ "supports_tool_choice": true }, "bedrock/eu-west-3/mistral.mistral-large-2402-v1:0": { - "input_cost_per_token": 0.0000104, + "input_cost_per_token": 1.04e-5, "litellm_provider": "bedrock", "max_input_tokens": 32000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 0.0000312, + "output_cost_per_token": 3.12e-5, "supports_function_calling": true }, "bedrock/eu-west-3/mistral.mixtral-8x7b-instruct-v0:1": { @@ -5347,7 +4717,7 @@ "supports_tool_choice": true }, "bedrock/invoke/anthropic.claude-3-5-sonnet-20240620-v1:0": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "bedrock", "max_input_tokens": 200000, "max_output_tokens": 4096, @@ -5356,20 +4726,20 @@ "notes": "Anthropic via Invoke route does not currently support pdf input." }, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "supports_function_calling": true, "supports_response_schema": true, "supports_tool_choice": true, "supports_vision": true }, "bedrock/sa-east-1/meta.llama3-70b-instruct-v1:0": { - "input_cost_per_token": 0.00000445, + "input_cost_per_token": 4.45e-6, "litellm_provider": "bedrock", "max_input_tokens": 8192, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.00000588 + "output_cost_per_token": 5.88e-6 }, "bedrock/sa-east-1/meta.llama3-8b-instruct-v1:0": { "input_cost_per_token": 5e-7, @@ -5378,7 +4748,7 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.00000101 + "output_cost_per_token": 1.01e-6 }, "bedrock/us-east-1/1-month-commitment/anthropic.claude-instant-v1": { "input_cost_per_second": 0.011, @@ -5445,37 +4815,37 @@ "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 0.0000024, + "output_cost_per_token": 2.4e-6, "supports_tool_choice": true }, "bedrock/us-east-1/anthropic.claude-v1": { - "input_cost_per_token": 0.000008, + "input_cost_per_token": 8e-6, "litellm_provider": "bedrock", "max_input_tokens": 100000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 0.000024, + "output_cost_per_token": 2.4e-5, "supports_tool_choice": true }, "bedrock/us-east-1/anthropic.claude-v2:1": { - "input_cost_per_token": 0.000008, + "input_cost_per_token": 8e-6, "litellm_provider": "bedrock", "max_input_tokens": 100000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 0.000024, + "output_cost_per_token": 2.4e-5, "supports_tool_choice": true }, "bedrock/us-east-1/meta.llama3-70b-instruct-v1:0": { - "input_cost_per_token": 0.00000265, + "input_cost_per_token": 2.65e-6, "litellm_provider": "bedrock", "max_input_tokens": 8192, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.0000035 + "output_cost_per_token": 3.5e-6 }, "bedrock/us-east-1/meta.llama3-8b-instruct-v1:0": { "input_cost_per_token": 3e-7, @@ -5497,13 +4867,13 @@ "supports_tool_choice": true }, "bedrock/us-east-1/mistral.mistral-large-2402-v1:0": { - "input_cost_per_token": 0.000008, + "input_cost_per_token": 8e-6, "litellm_provider": "bedrock", "max_input_tokens": 32000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 0.000024, + "output_cost_per_token": 2.4e-5, "supports_function_calling": true }, "bedrock/us-east-1/mistral.mixtral-8x7b-instruct-v0:1": { @@ -5523,7 +4893,7 @@ "max_output_tokens": 10000, "max_tokens": 10000, "mode": "chat", - "output_cost_per_token": 0.00000384, + "output_cost_per_token": 3.84e-6, "supports_function_calling": true, "supports_pdf_input": true, "supports_prompt_caching": true, @@ -5536,7 +4906,7 @@ "max_input_tokens": 8192, "max_tokens": 8192, "mode": "embedding", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "output_vector_size": 1536 }, "bedrock/us-gov-east-1/amazon.titan-embed-text-v2:0": { @@ -5545,17 +4915,17 @@ "max_input_tokens": 8192, "max_tokens": 8192, "mode": "embedding", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "output_vector_size": 1024 }, "bedrock/us-gov-east-1/amazon.titan-text-express-v1": { - "input_cost_per_token": 0.0000013, + "input_cost_per_token": 1.3e-6, "litellm_provider": "bedrock", "max_input_tokens": 42000, "max_output_tokens": 8000, "max_tokens": 8000, "mode": "chat", - "output_cost_per_token": 0.0000017 + "output_cost_per_token": 1.7e-6 }, "bedrock/us-gov-east-1/amazon.titan-text-lite-v1": { "input_cost_per_token": 3e-7, @@ -5573,16 +4943,16 @@ "max_output_tokens": 32000, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 0.0000015 + "output_cost_per_token": 1.5e-6 }, "bedrock/us-gov-east-1/anthropic.claude-3-5-sonnet-20240620-v1:0": { - "input_cost_per_token": 0.0000036, + "input_cost_per_token": 3.6e-6, "litellm_provider": "bedrock", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.000018, + "output_cost_per_token": 1.8e-5, "supports_function_calling": true, "supports_pdf_input": true, "supports_response_schema": true, @@ -5596,7 +4966,7 @@ "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.0000015, + "output_cost_per_token": 1.5e-6, "supports_function_calling": true, "supports_pdf_input": true, "supports_response_schema": true, @@ -5604,13 +4974,13 @@ "supports_vision": true }, "bedrock/us-gov-east-1/claude-sonnet-4-5-20250929-v1:0": { - "input_cost_per_token": 0.0000033, + "input_cost_per_token": 3.3e-6, "litellm_provider": "bedrock", "max_input_tokens": 200000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.0000165, + "output_cost_per_token": 1.65e-5, "supports_assistant_prefill": true, "supports_computer_use": true, "supports_function_calling": true, @@ -5622,13 +4992,13 @@ "supports_vision": true }, "bedrock/us-gov-east-1/meta.llama3-70b-instruct-v1:0": { - "input_cost_per_token": 0.00000265, + "input_cost_per_token": 2.65e-6, "litellm_provider": "bedrock", "max_input_tokens": 8000, "max_output_tokens": 2048, "max_tokens": 2048, "mode": "chat", - "output_cost_per_token": 0.0000035, + "output_cost_per_token": 3.5e-6, "supports_pdf_input": true }, "bedrock/us-gov-east-1/meta.llama3-8b-instruct-v1:0": { @@ -5638,7 +5008,7 @@ "max_output_tokens": 2048, "max_tokens": 2048, "mode": "chat", - "output_cost_per_token": 0.00000265, + "output_cost_per_token": 2.65e-6, "supports_pdf_input": true }, "bedrock/us-gov-west-1/amazon.nova-pro-v1:0": { @@ -5648,7 +5018,7 @@ "max_output_tokens": 10000, "max_tokens": 10000, "mode": "chat", - "output_cost_per_token": 0.00000384, + "output_cost_per_token": 3.84e-6, "supports_function_calling": true, "supports_pdf_input": true, "supports_prompt_caching": true, @@ -5661,7 +5031,7 @@ "max_input_tokens": 8192, "max_tokens": 8192, "mode": "embedding", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "output_vector_size": 1536 }, "bedrock/us-gov-west-1/amazon.titan-embed-text-v2:0": { @@ -5670,17 +5040,17 @@ "max_input_tokens": 8192, "max_tokens": 8192, "mode": "embedding", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "output_vector_size": 1024 }, "bedrock/us-gov-west-1/amazon.titan-text-express-v1": { - "input_cost_per_token": 0.0000013, + "input_cost_per_token": 1.3e-6, "litellm_provider": "bedrock", "max_input_tokens": 42000, "max_output_tokens": 8000, "max_tokens": 8000, "mode": "chat", - "output_cost_per_token": 0.0000017 + "output_cost_per_token": 1.7e-6 }, "bedrock/us-gov-west-1/amazon.titan-text-lite-v1": { "input_cost_per_token": 3e-7, @@ -5698,18 +5068,18 @@ "max_output_tokens": 32000, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 0.0000015 + "output_cost_per_token": 1.5e-6 }, "bedrock/us-gov-west-1/anthropic.claude-3-7-sonnet-20250219-v1:0": { - "cache_creation_input_token_cost": 0.0000045, + "cache_creation_input_token_cost": 4.5e-6, "cache_read_input_token_cost": 3.6e-7, - "input_cost_per_token": 0.0000036, + "input_cost_per_token": 3.6e-6, "litellm_provider": "bedrock", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.000018, + "output_cost_per_token": 1.8e-5, "supports_assistant_prefill": true, "supports_computer_use": true, "supports_function_calling": true, @@ -5721,13 +5091,13 @@ "supports_vision": true }, "bedrock/us-gov-west-1/anthropic.claude-3-5-sonnet-20240620-v1:0": { - "input_cost_per_token": 0.0000036, + "input_cost_per_token": 3.6e-6, "litellm_provider": "bedrock", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.000018, + "output_cost_per_token": 1.8e-5, "supports_function_calling": true, "supports_pdf_input": true, "supports_response_schema": true, @@ -5741,7 +5111,7 @@ "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.0000015, + "output_cost_per_token": 1.5e-6, "supports_function_calling": true, "supports_pdf_input": true, "supports_response_schema": true, @@ -5749,13 +5119,13 @@ "supports_vision": true }, "bedrock/us-gov-west-1/claude-sonnet-4-5-20250929-v1:0": { - "input_cost_per_token": 0.0000033, + "input_cost_per_token": 3.3e-6, "litellm_provider": "bedrock", "max_input_tokens": 200000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.0000165, + "output_cost_per_token": 1.65e-5, "supports_assistant_prefill": true, "supports_computer_use": true, "supports_function_calling": true, @@ -5767,13 +5137,13 @@ "supports_vision": true }, "bedrock/us-gov-west-1/meta.llama3-70b-instruct-v1:0": { - "input_cost_per_token": 0.00000265, + "input_cost_per_token": 2.65e-6, "litellm_provider": "bedrock", "max_input_tokens": 8000, "max_output_tokens": 2048, "max_tokens": 2048, "mode": "chat", - "output_cost_per_token": 0.0000035, + "output_cost_per_token": 3.5e-6, "supports_pdf_input": true }, "bedrock/us-gov-west-1/meta.llama3-8b-instruct-v1:0": { @@ -5783,17 +5153,17 @@ "max_output_tokens": 2048, "max_tokens": 2048, "mode": "chat", - "output_cost_per_token": 0.00000265, + "output_cost_per_token": 2.65e-6, "supports_pdf_input": true }, "bedrock/us-west-1/meta.llama3-70b-instruct-v1:0": { - "input_cost_per_token": 0.00000265, + "input_cost_per_token": 2.65e-6, "litellm_provider": "bedrock", "max_input_tokens": 8192, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.0000035 + "output_cost_per_token": 3.5e-6 }, "bedrock/us-west-1/meta.llama3-8b-instruct-v1:0": { "input_cost_per_token": 3e-7, @@ -5869,27 +5239,27 @@ "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 0.0000024, + "output_cost_per_token": 2.4e-6, "supports_tool_choice": true }, "bedrock/us-west-2/anthropic.claude-v1": { - "input_cost_per_token": 0.000008, + "input_cost_per_token": 8e-6, "litellm_provider": "bedrock", "max_input_tokens": 100000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 0.000024, + "output_cost_per_token": 2.4e-5, "supports_tool_choice": true }, "bedrock/us-west-2/anthropic.claude-v2:1": { - "input_cost_per_token": 0.000008, + "input_cost_per_token": 8e-6, "litellm_provider": "bedrock", "max_input_tokens": 100000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 0.000024, + "output_cost_per_token": 2.4e-5, "supports_tool_choice": true }, "bedrock/us-west-2/mistral.mistral-7b-instruct-v0:2": { @@ -5903,13 +5273,13 @@ "supports_tool_choice": true }, "bedrock/us-west-2/mistral.mistral-large-2402-v1:0": { - "input_cost_per_token": 0.000008, + "input_cost_per_token": 8e-6, "litellm_provider": "bedrock", "max_input_tokens": 32000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 0.000024, + "output_cost_per_token": 2.4e-5, "supports_function_calling": true }, "bedrock/us-west-2/mistral.mixtral-8x7b-instruct-v0:1": { @@ -5923,7 +5293,7 @@ "supports_tool_choice": true }, "bedrock/us.anthropic.claude-3-5-haiku-20241022-v1:0": { - "cache_creation_input_token_cost": 0.000001, + "cache_creation_input_token_cost": 1e-6, "cache_read_input_token_cost": 8e-8, "input_cost_per_token": 8e-7, "litellm_provider": "bedrock", @@ -5931,7 +5301,7 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.000004, + "output_cost_per_token": 4e-6, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_pdf_input": true, @@ -5946,7 +5316,7 @@ "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.0000012, + "output_cost_per_token": 1.2e-6, "supports_function_calling": true, "supports_tool_choice": true }, @@ -6075,13 +5445,13 @@ "output_cost_per_token": 5e-7 }, "chatgpt-4o-latest": { - "input_cost_per_token": 0.000005, + "input_cost_per_token": 5e-6, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_pdf_input": true, @@ -6091,8 +5461,8 @@ "supports_vision": true }, "claude-3-5-haiku-20241022": { - "cache_creation_input_token_cost": 0.000001, - "cache_creation_input_token_cost_above_1hr": 0.000006, + "cache_creation_input_token_cost": 1e-6, + "cache_creation_input_token_cost_above_1hr": 6e-6, "cache_read_input_token_cost": 8e-8, "deprecation_date": "2025-10-01", "input_cost_per_token": 8e-7, @@ -6101,7 +5471,7 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.000004, + "output_cost_per_token": 4e-6, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -6118,17 +5488,17 @@ "tool_use_system_prompt_tokens": 264 }, "claude-3-5-haiku-latest": { - "cache_creation_input_token_cost": 0.00000125, - "cache_creation_input_token_cost_above_1hr": 0.000006, + "cache_creation_input_token_cost": 1.25e-6, + "cache_creation_input_token_cost_above_1hr": 6e-6, "cache_read_input_token_cost": 1e-7, "deprecation_date": "2025-10-01", - "input_cost_per_token": 0.000001, + "input_cost_per_token": 1e-6, "litellm_provider": "anthropic", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.000005, + "output_cost_per_token": 5e-6, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -6145,16 +5515,16 @@ "tool_use_system_prompt_tokens": 264 }, "claude-haiku-4-5-20251001": { - "cache_creation_input_token_cost": 0.00000125, - "cache_creation_input_token_cost_above_1hr": 0.000002, + "cache_creation_input_token_cost": 1.25e-6, + "cache_creation_input_token_cost_above_1hr": 2e-6, "cache_read_input_token_cost": 1e-7, - "input_cost_per_token": 0.000001, + "input_cost_per_token": 1e-6, "litellm_provider": "anthropic", "max_input_tokens": 200000, "max_output_tokens": 64000, "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 0.000005, + "output_cost_per_token": 5e-6, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_computer_use": true, @@ -6166,16 +5536,16 @@ "supports_vision": true }, "claude-haiku-4-5": { - "cache_creation_input_token_cost": 0.00000125, - "cache_creation_input_token_cost_above_1hr": 0.000002, + "cache_creation_input_token_cost": 1.25e-6, + "cache_creation_input_token_cost_above_1hr": 2e-6, "cache_read_input_token_cost": 1e-7, - "input_cost_per_token": 0.000001, + "input_cost_per_token": 1e-6, "litellm_provider": "anthropic", "max_input_tokens": 200000, "max_output_tokens": 64000, "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 0.000005, + "output_cost_per_token": 5e-6, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_computer_use": true, @@ -6187,17 +5557,17 @@ "supports_vision": true }, "claude-3-5-sonnet-20240620": { - "cache_creation_input_token_cost": 0.00000375, - "cache_creation_input_token_cost_above_1hr": 0.000006, + "cache_creation_input_token_cost": 3.75e-6, + "cache_creation_input_token_cost_above_1hr": 6e-6, "cache_read_input_token_cost": 3e-7, "deprecation_date": "2025-06-01", - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "anthropic", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_pdf_input": true, @@ -6208,17 +5578,17 @@ "tool_use_system_prompt_tokens": 159 }, "claude-3-5-sonnet-20241022": { - "cache_creation_input_token_cost": 0.00000375, - "cache_creation_input_token_cost_above_1hr": 0.000006, + "cache_creation_input_token_cost": 3.75e-6, + "cache_creation_input_token_cost_above_1hr": 6e-6, "cache_read_input_token_cost": 3e-7, "deprecation_date": "2025-10-01", - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "anthropic", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -6236,17 +5606,17 @@ "tool_use_system_prompt_tokens": 159 }, "claude-3-5-sonnet-latest": { - "cache_creation_input_token_cost": 0.00000375, - "cache_creation_input_token_cost_above_1hr": 0.000006, + "cache_creation_input_token_cost": 3.75e-6, + "cache_creation_input_token_cost_above_1hr": 6e-6, "cache_read_input_token_cost": 3e-7, "deprecation_date": "2025-06-01", - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "anthropic", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -6264,17 +5634,17 @@ "tool_use_system_prompt_tokens": 159 }, "claude-3-7-sonnet-20250219": { - "cache_creation_input_token_cost": 0.00000375, - "cache_creation_input_token_cost_above_1hr": 0.000006, + "cache_creation_input_token_cost": 3.75e-6, + "cache_creation_input_token_cost_above_1hr": 6e-6, "cache_read_input_token_cost": 3e-7, "deprecation_date": "2026-02-19", - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "anthropic", "max_input_tokens": 200000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -6293,17 +5663,17 @@ "tool_use_system_prompt_tokens": 159 }, "claude-3-7-sonnet-latest": { - "cache_creation_input_token_cost": 0.00000375, - "cache_creation_input_token_cost_above_1hr": 0.000006, + "cache_creation_input_token_cost": 3.75e-6, + "cache_creation_input_token_cost_above_1hr": 6e-6, "cache_read_input_token_cost": 3e-7, "deprecation_date": "2025-06-01", - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "anthropic", "max_input_tokens": 200000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -6322,7 +5692,7 @@ }, "claude-3-haiku-20240307": { "cache_creation_input_token_cost": 3e-7, - "cache_creation_input_token_cost_above_1hr": 0.000006, + "cache_creation_input_token_cost_above_1hr": 6e-6, "cache_read_input_token_cost": 3e-8, "input_cost_per_token": 2.5e-7, "litellm_provider": "anthropic", @@ -6330,7 +5700,7 @@ "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.00000125, + "output_cost_per_token": 1.25e-6, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_prompt_caching": true, @@ -6340,17 +5710,17 @@ "tool_use_system_prompt_tokens": 264 }, "claude-3-opus-20240229": { - "cache_creation_input_token_cost": 0.00001875, - "cache_creation_input_token_cost_above_1hr": 0.000006, - "cache_read_input_token_cost": 0.0000015, + "cache_creation_input_token_cost": 1.875e-5, + "cache_creation_input_token_cost_above_1hr": 6e-6, + "cache_read_input_token_cost": 1.5e-6, "deprecation_date": "2026-05-01", - "input_cost_per_token": 0.000015, + "input_cost_per_token": 1.5e-5, "litellm_provider": "anthropic", "max_input_tokens": 200000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.000075, + "output_cost_per_token": 7.5e-5, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_prompt_caching": true, @@ -6360,17 +5730,17 @@ "tool_use_system_prompt_tokens": 395 }, "claude-3-opus-latest": { - "cache_creation_input_token_cost": 0.00001875, - "cache_creation_input_token_cost_above_1hr": 0.000006, - "cache_read_input_token_cost": 0.0000015, + "cache_creation_input_token_cost": 1.875e-5, + "cache_creation_input_token_cost_above_1hr": 6e-6, + "cache_read_input_token_cost": 1.5e-6, "deprecation_date": "2025-03-01", - "input_cost_per_token": 0.000015, + "input_cost_per_token": 1.5e-5, "litellm_provider": "anthropic", "max_input_tokens": 200000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.000075, + "output_cost_per_token": 7.5e-5, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_prompt_caching": true, @@ -6380,15 +5750,15 @@ "tool_use_system_prompt_tokens": 395 }, "claude-4-opus-20250514": { - "cache_creation_input_token_cost": 0.00001875, - "cache_read_input_token_cost": 0.0000015, - "input_cost_per_token": 0.000015, + "cache_creation_input_token_cost": 1.875e-5, + "cache_read_input_token_cost": 1.5e-6, + "input_cost_per_token": 1.5e-5, "litellm_provider": "anthropic", "max_input_tokens": 200000, "max_output_tokens": 32000, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 0.000075, + "output_cost_per_token": 7.5e-5, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -6406,19 +5776,19 @@ "tool_use_system_prompt_tokens": 159 }, "claude-4-sonnet-20250514": { - "cache_creation_input_token_cost": 0.00000375, - "cache_creation_input_token_cost_above_200k_tokens": 0.0000075, + "cache_creation_input_token_cost": 3.75e-6, + "cache_creation_input_token_cost_above_200k_tokens": 7.5e-6, "cache_read_input_token_cost": 3e-7, "cache_read_input_token_cost_above_200k_tokens": 6e-7, - "input_cost_per_token": 0.000003, - "input_cost_per_token_above_200k_tokens": 0.000006, + "input_cost_per_token": 3e-6, + "input_cost_per_token_above_200k_tokens": 6e-6, "litellm_provider": "anthropic", "max_input_tokens": 1000000, "max_output_tokens": 64000, "max_tokens": 1000000, "mode": "chat", - "output_cost_per_token": 0.000015, - "output_cost_per_token_above_200k_tokens": 0.0000225, + "output_cost_per_token": 1.5e-5, + "output_cost_per_token_above_200k_tokens": 2.25e-5, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -6436,19 +5806,19 @@ "tool_use_system_prompt_tokens": 159 }, "claude-sonnet-4-5": { - "cache_creation_input_token_cost": 0.00000375, + "cache_creation_input_token_cost": 3.75e-6, "cache_read_input_token_cost": 3e-7, - "input_cost_per_token": 0.000003, - "input_cost_per_token_above_200k_tokens": 0.000006, - "output_cost_per_token_above_200k_tokens": 0.0000225, - "cache_creation_input_token_cost_above_200k_tokens": 0.0000075, + "input_cost_per_token": 3e-6, + "input_cost_per_token_above_200k_tokens": 6e-6, + "output_cost_per_token_above_200k_tokens": 2.25e-5, + "cache_creation_input_token_cost_above_200k_tokens": 7.5e-6, "cache_read_input_token_cost_above_200k_tokens": 6e-7, "litellm_provider": "anthropic", "max_input_tokens": 200000, "max_output_tokens": 64000, "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -6466,19 +5836,19 @@ "tool_use_system_prompt_tokens": 346 }, "claude-sonnet-4-5-20250929": { - "cache_creation_input_token_cost": 0.00000375, + "cache_creation_input_token_cost": 3.75e-6, "cache_read_input_token_cost": 3e-7, - "input_cost_per_token": 0.000003, - "input_cost_per_token_above_200k_tokens": 0.000006, - "output_cost_per_token_above_200k_tokens": 0.0000225, - "cache_creation_input_token_cost_above_200k_tokens": 0.0000075, + "input_cost_per_token": 3e-6, + "input_cost_per_token_above_200k_tokens": 6e-6, + "output_cost_per_token_above_200k_tokens": 2.25e-5, + "cache_creation_input_token_cost_above_200k_tokens": 7.5e-6, "cache_read_input_token_cost_above_200k_tokens": 6e-7, "litellm_provider": "anthropic", "max_input_tokens": 200000, "max_output_tokens": 64000, "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -6497,19 +5867,19 @@ "tool_use_system_prompt_tokens": 346 }, "claude-sonnet-4-5-20250929-v1:0": { - "cache_creation_input_token_cost": 0.00000375, + "cache_creation_input_token_cost": 3.75e-6, "cache_read_input_token_cost": 3e-7, - "input_cost_per_token": 0.000003, - "input_cost_per_token_above_200k_tokens": 0.000006, - "output_cost_per_token_above_200k_tokens": 0.0000225, - "cache_creation_input_token_cost_above_200k_tokens": 0.0000075, + "input_cost_per_token": 3e-6, + "input_cost_per_token_above_200k_tokens": 6e-6, + "output_cost_per_token_above_200k_tokens": 2.25e-5, + "cache_creation_input_token_cost_above_200k_tokens": 7.5e-6, "cache_read_input_token_cost_above_200k_tokens": 6e-7, "litellm_provider": "bedrock", "max_input_tokens": 200000, "max_output_tokens": 64000, "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "supports_assistant_prefill": true, "supports_computer_use": true, "supports_function_calling": true, @@ -6522,16 +5892,16 @@ "tool_use_system_prompt_tokens": 159 }, "claude-opus-4-1": { - "cache_creation_input_token_cost": 0.00001875, - "cache_creation_input_token_cost_above_1hr": 0.00003, - "cache_read_input_token_cost": 0.0000015, - "input_cost_per_token": 0.000015, + "cache_creation_input_token_cost": 1.875e-5, + "cache_creation_input_token_cost_above_1hr": 3e-5, + "cache_read_input_token_cost": 1.5e-6, + "input_cost_per_token": 1.5e-5, "litellm_provider": "anthropic", "max_input_tokens": 200000, "max_output_tokens": 32000, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 0.000075, + "output_cost_per_token": 7.5e-5, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -6549,17 +5919,17 @@ "tool_use_system_prompt_tokens": 159 }, "claude-opus-4-1-20250805": { - "cache_creation_input_token_cost": 0.00001875, - "cache_creation_input_token_cost_above_1hr": 0.00003, - "cache_read_input_token_cost": 0.0000015, - "input_cost_per_token": 0.000015, + "cache_creation_input_token_cost": 1.875e-5, + "cache_creation_input_token_cost_above_1hr": 3e-5, + "cache_read_input_token_cost": 1.5e-6, + "input_cost_per_token": 1.5e-5, "deprecation_date": "2026-08-05", "litellm_provider": "anthropic", "max_input_tokens": 200000, "max_output_tokens": 32000, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 0.000075, + "output_cost_per_token": 7.5e-5, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -6577,17 +5947,17 @@ "tool_use_system_prompt_tokens": 159 }, "claude-opus-4-20250514": { - "cache_creation_input_token_cost": 0.00001875, - "cache_creation_input_token_cost_above_1hr": 0.00003, - "cache_read_input_token_cost": 0.0000015, - "input_cost_per_token": 0.000015, + "cache_creation_input_token_cost": 1.875e-5, + "cache_creation_input_token_cost_above_1hr": 3e-5, + "cache_read_input_token_cost": 1.5e-6, + "input_cost_per_token": 1.5e-5, "deprecation_date": "2026-05-14", "litellm_provider": "anthropic", "max_input_tokens": 200000, "max_output_tokens": 32000, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 0.000075, + "output_cost_per_token": 7.5e-5, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -6606,20 +5976,20 @@ }, "claude-sonnet-4-20250514": { "deprecation_date": "2026-05-14", - "cache_creation_input_token_cost": 0.00000375, - "cache_creation_input_token_cost_above_1hr": 0.000006, + "cache_creation_input_token_cost": 3.75e-6, + "cache_creation_input_token_cost_above_1hr": 6e-6, "cache_read_input_token_cost": 3e-7, - "input_cost_per_token": 0.000003, - "input_cost_per_token_above_200k_tokens": 0.000006, - "output_cost_per_token_above_200k_tokens": 0.0000225, - "cache_creation_input_token_cost_above_200k_tokens": 0.0000075, + "input_cost_per_token": 3e-6, + "input_cost_per_token_above_200k_tokens": 6e-6, + "output_cost_per_token_above_200k_tokens": 2.25e-5, + "cache_creation_input_token_cost_above_200k_tokens": 7.5e-6, "cache_read_input_token_cost_above_200k_tokens": 6e-7, "litellm_provider": "anthropic", "max_input_tokens": 1000000, "max_output_tokens": 64000, "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -6637,40 +6007,40 @@ "tool_use_system_prompt_tokens": 159 }, "cloudflare/@cf/meta/llama-2-7b-chat-fp16": { - "input_cost_per_token": 0.000001923, + "input_cost_per_token": 1.923e-6, "litellm_provider": "cloudflare", "max_input_tokens": 3072, "max_output_tokens": 3072, "max_tokens": 3072, "mode": "chat", - "output_cost_per_token": 0.000001923 + "output_cost_per_token": 1.923e-6 }, "cloudflare/@cf/meta/llama-2-7b-chat-int8": { - "input_cost_per_token": 0.000001923, + "input_cost_per_token": 1.923e-6, "litellm_provider": "cloudflare", "max_input_tokens": 2048, "max_output_tokens": 2048, "max_tokens": 2048, "mode": "chat", - "output_cost_per_token": 0.000001923 + "output_cost_per_token": 1.923e-6 }, "cloudflare/@cf/mistral/mistral-7b-instruct-v0.1": { - "input_cost_per_token": 0.000001923, + "input_cost_per_token": 1.923e-6, "litellm_provider": "cloudflare", "max_input_tokens": 8192, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.000001923 + "output_cost_per_token": 1.923e-6 }, "cloudflare/@hf/thebloke/codellama-7b-instruct-awq": { - "input_cost_per_token": 0.000001923, + "input_cost_per_token": 1.923e-6, "litellm_provider": "cloudflare", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.000001923 + "output_cost_per_token": 1.923e-6 }, "code-bison": { "input_cost_per_character": 2.5e-7, @@ -6852,48 +6222,41 @@ "supports_tool_choice": true }, "codestral/codestral-2405": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "codestral", "max_input_tokens": 32000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "source": "https://docs.mistral.ai/capabilities/code_generation/", "supports_assistant_prefill": true, "supports_tool_choice": true }, "codestral/codestral-latest": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "codestral", "max_input_tokens": 32000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "source": "https://docs.mistral.ai/capabilities/code_generation/", "supports_assistant_prefill": true, "supports_tool_choice": true }, "codex-mini-latest": { "cache_read_input_token_cost": 3.75e-7, - "input_cost_per_token": 0.0000015, + "input_cost_per_token": 1.5e-6, "litellm_provider": "openai", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "responses", - "output_cost_per_token": 0.000006, - "supported_endpoints": [ - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 6e-6, + "supported_endpoints": ["/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_pdf_input": true, @@ -6915,13 +6278,13 @@ "supports_tool_choice": true }, "cohere.command-r-plus-v1:0": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "bedrock", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "supports_tool_choice": true }, "cohere.command-r-v1:0": { @@ -6931,17 +6294,17 @@ "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.0000015, + "output_cost_per_token": 1.5e-6, "supports_tool_choice": true }, "cohere.command-text-v14": { - "input_cost_per_token": 0.0000015, + "input_cost_per_token": 1.5e-6, "litellm_provider": "bedrock", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.000002, + "output_cost_per_token": 2e-6, "supports_tool_choice": true }, "cohere.embed-english-v3": { @@ -6950,7 +6313,7 @@ "max_input_tokens": 512, "max_tokens": 512, "mode": "embedding", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "supports_embedding_image_input": true }, "cohere.embed-multilingual-v3": { @@ -6959,7 +6322,7 @@ "max_input_tokens": 512, "max_tokens": 512, "mode": "embedding", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "supports_embedding_image_input": true }, "cohere.embed-v4:0": { @@ -6968,7 +6331,7 @@ "max_input_tokens": 128000, "max_tokens": 128000, "mode": "embedding", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "output_vector_size": 1536, "supports_embedding_image_input": true }, @@ -6978,13 +6341,13 @@ "max_input_tokens": 128000, "max_tokens": 128000, "mode": "embedding", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "output_vector_size": 1536, "supports_embedding_image_input": true }, "cohere.rerank-v3-5:0": { "input_cost_per_query": 0.002, - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "bedrock", "max_document_chunks_per_query": 100, "max_input_tokens": 32000, @@ -6993,25 +6356,25 @@ "max_tokens": 32000, "max_tokens_per_document_chunk": 512, "mode": "rerank", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "command": { - "input_cost_per_token": 0.000001, + "input_cost_per_token": 1e-6, "litellm_provider": "cohere", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "completion", - "output_cost_per_token": 0.000002 + "output_cost_per_token": 2e-6 }, "command-a-03-2025": { - "input_cost_per_token": 0.0000025, + "input_cost_per_token": 2.5e-6, "litellm_provider": "cohere_chat", "max_input_tokens": 256000, "max_output_tokens": 8000, "max_tokens": 8000, "mode": "chat", - "output_cost_per_token": 0.00001, + "output_cost_per_token": 1e-5, "supports_function_calling": true, "supports_tool_choice": true }, @@ -7026,13 +6389,13 @@ "supports_tool_choice": true }, "command-nightly": { - "input_cost_per_token": 0.000001, + "input_cost_per_token": 1e-6, "litellm_provider": "cohere", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "completion", - "output_cost_per_token": 0.000002 + "output_cost_per_token": 2e-6 }, "command-r": { "input_cost_per_token": 1.5e-7, @@ -7057,24 +6420,24 @@ "supports_tool_choice": true }, "command-r-plus": { - "input_cost_per_token": 0.0000025, + "input_cost_per_token": 2.5e-6, "litellm_provider": "cohere_chat", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.00001, + "output_cost_per_token": 1e-5, "supports_function_calling": true, "supports_tool_choice": true }, "command-r-plus-08-2024": { - "input_cost_per_token": 0.0000025, + "input_cost_per_token": 2.5e-6, "litellm_provider": "cohere_chat", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.00001, + "output_cost_per_token": 1e-5, "supports_function_calling": true, "supports_tool_choice": true }, @@ -7091,23 +6454,16 @@ "supports_tool_choice": true }, "computer-use-preview": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "azure", "max_input_tokens": 8192, "max_output_tokens": 1024, "max_tokens": 1024, "mode": "chat", - "output_cost_per_token": 0.000012, - "supported_endpoints": [ - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 1.2e-5, + "supported_endpoints": ["/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": false, @@ -7125,11 +6481,9 @@ "max_output_tokens": 8192, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0.0000017, + "output_cost_per_token": 1.7e-6, "source": "https://api-docs.deepseek.com/quick_start/pricing", - "supported_endpoints": [ - "/v1/chat/completions" - ], + "supported_endpoints": ["/v1/chat/completions"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -7146,11 +6500,9 @@ "max_output_tokens": 65536, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0.0000017, + "output_cost_per_token": 1.7e-6, "source": "https://api-docs.deepseek.com/quick_start/pricing", - "supported_endpoints": [ - "/v1/chat/completions" - ], + "supported_endpoints": ["/v1/chat/completions"], "supports_function_calling": false, "supports_native_streaming": true, "supports_parallel_function_calling": false, @@ -7167,7 +6519,7 @@ "max_output_tokens": 16384, "max_tokens": 1000000, "mode": "chat", - "output_cost_per_token": 0.0000015, + "output_cost_per_token": 1.5e-6, "source": "https://www.alibabacloud.com/help/en/model-studio/models", "supports_function_calling": true, "supports_reasoning": true, @@ -7187,18 +6539,12 @@ { "input_cost_per_token": 5e-8, "output_cost_per_token": 4e-7, - "range": [ - 0, - 256000 - ] + "range": [0, 256000.0] }, { "input_cost_per_token": 2.5e-7, - "output_cost_per_token": 0.000002, - "range": [ - 256000, - 1000000 - ] + "output_cost_per_token": 2e-6, + "range": [256000.0, 1000000.0] } ] }, @@ -7216,29 +6562,23 @@ { "input_cost_per_token": 5e-8, "output_cost_per_token": 4e-7, - "range": [ - 0, - 256000 - ] + "range": [0, 256000.0] }, { "input_cost_per_token": 2.5e-7, - "output_cost_per_token": 0.000002, - "range": [ - 256000, - 1000000 - ] + "output_cost_per_token": 2e-6, + "range": [256000.0, 1000000.0] } ] }, "dashscope/qwen-max": { - "input_cost_per_token": 0.0000016, + "input_cost_per_token": 1.6e-6, "litellm_provider": "dashscope", "max_input_tokens": 30720, "max_output_tokens": 8192, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0.0000064, + "output_cost_per_token": 6.4e-6, "source": "https://www.alibabacloud.com/help/en/model-studio/models", "supports_function_calling": true, "supports_reasoning": true, @@ -7251,7 +6591,7 @@ "max_output_tokens": 16384, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0.0000012, + "output_cost_per_token": 1.2e-6, "source": "https://www.alibabacloud.com/help/en/model-studio/models", "supports_function_calling": true, "supports_reasoning": true, @@ -7264,7 +6604,7 @@ "max_output_tokens": 8192, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0.0000012, + "output_cost_per_token": 1.2e-6, "source": "https://www.alibabacloud.com/help/en/model-studio/models", "supports_function_calling": true, "supports_reasoning": true, @@ -7277,8 +6617,8 @@ "max_output_tokens": 16384, "max_tokens": 131072, "mode": "chat", - "output_cost_per_reasoning_token": 0.000004, - "output_cost_per_token": 0.0000012, + "output_cost_per_reasoning_token": 4e-6, + "output_cost_per_token": 1.2e-6, "source": "https://www.alibabacloud.com/help/en/model-studio/models", "supports_function_calling": true, "supports_reasoning": true, @@ -7291,8 +6631,8 @@ "max_output_tokens": 16384, "max_tokens": 131072, "mode": "chat", - "output_cost_per_reasoning_token": 0.000004, - "output_cost_per_token": 0.0000012, + "output_cost_per_reasoning_token": 4e-6, + "output_cost_per_token": 1.2e-6, "source": "https://www.alibabacloud.com/help/en/model-studio/models", "supports_function_calling": true, "supports_reasoning": true, @@ -7311,21 +6651,15 @@ "tiered_pricing": [ { "input_cost_per_token": 4e-7, - "output_cost_per_reasoning_token": 0.000004, - "output_cost_per_token": 0.0000012, - "range": [ - 0, - 256000 - ] + "output_cost_per_reasoning_token": 4e-6, + "output_cost_per_token": 1.2e-6, + "range": [0, 256000.0] }, { - "input_cost_per_token": 0.0000012, - "output_cost_per_reasoning_token": 0.000012, - "output_cost_per_token": 0.0000036, - "range": [ - 256000, - 1000000 - ] + "input_cost_per_token": 1.2e-6, + "output_cost_per_reasoning_token": 1.2e-5, + "output_cost_per_token": 3.6e-6, + "range": [256000.0, 1000000.0] } ] }, @@ -7342,21 +6676,15 @@ "tiered_pricing": [ { "input_cost_per_token": 4e-7, - "output_cost_per_reasoning_token": 0.000004, - "output_cost_per_token": 0.0000012, - "range": [ - 0, - 256000 - ] + "output_cost_per_reasoning_token": 4e-6, + "output_cost_per_token": 1.2e-6, + "range": [0, 256000.0] }, { - "input_cost_per_token": 0.0000012, - "output_cost_per_reasoning_token": 0.000012, - "output_cost_per_token": 0.0000036, - "range": [ - 256000, - 1000000 - ] + "input_cost_per_token": 1.2e-6, + "output_cost_per_reasoning_token": 1.2e-5, + "output_cost_per_token": 3.6e-6, + "range": [256000.0, 1000000.0] } ] }, @@ -7373,21 +6701,15 @@ "tiered_pricing": [ { "input_cost_per_token": 4e-7, - "output_cost_per_reasoning_token": 0.000004, - "output_cost_per_token": 0.0000012, - "range": [ - 0, - 256000 - ] + "output_cost_per_reasoning_token": 4e-6, + "output_cost_per_token": 1.2e-6, + "range": [0, 256000.0] }, { - "input_cost_per_token": 0.0000012, - "output_cost_per_reasoning_token": 0.000012, - "output_cost_per_token": 0.0000036, - "range": [ - 256000, - 1000000 - ] + "input_cost_per_token": 1.2e-6, + "output_cost_per_reasoning_token": 1.2e-5, + "output_cost_per_token": 3.6e-6, + "range": [256000.0, 1000000.0] } ] }, @@ -7471,38 +6793,26 @@ { "cache_read_input_token_cost": 8e-8, "input_cost_per_token": 3e-7, - "output_cost_per_token": 0.0000015, - "range": [ - 0, - 32000 - ] + "output_cost_per_token": 1.5e-6, + "range": [0, 32000.0] }, { "cache_read_input_token_cost": 1.2e-7, "input_cost_per_token": 5e-7, - "output_cost_per_token": 0.0000025, - "range": [ - 32000, - 128000 - ] + "output_cost_per_token": 2.5e-6, + "range": [32000.0, 128000.0] }, { "cache_read_input_token_cost": 2e-7, "input_cost_per_token": 8e-7, - "output_cost_per_token": 0.000004, - "range": [ - 128000, - 256000 - ] + "output_cost_per_token": 4e-6, + "range": [128000.0, 256000.0] }, { "cache_read_input_token_cost": 4e-7, - "input_cost_per_token": 0.0000016, - "output_cost_per_token": 0.0000096, - "range": [ - 256000, - 1000000 - ] + "input_cost_per_token": 1.6e-6, + "output_cost_per_token": 9.6e-6, + "range": [256000.0, 1000000.0] } ] }, @@ -7519,35 +6829,23 @@ "tiered_pricing": [ { "input_cost_per_token": 3e-7, - "output_cost_per_token": 0.0000015, - "range": [ - 0, - 32000 - ] + "output_cost_per_token": 1.5e-6, + "range": [0, 32000.0] }, { "input_cost_per_token": 5e-7, - "output_cost_per_token": 0.0000025, - "range": [ - 32000, - 128000 - ] + "output_cost_per_token": 2.5e-6, + "range": [32000.0, 128000.0] }, { "input_cost_per_token": 8e-7, - "output_cost_per_token": 0.000004, - "range": [ - 128000, - 256000 - ] + "output_cost_per_token": 4e-6, + "range": [128000.0, 256000.0] }, { - "input_cost_per_token": 0.0000016, - "output_cost_per_token": 0.0000096, - "range": [ - 256000, - 1000000 - ] + "input_cost_per_token": 1.6e-6, + "output_cost_per_token": 9.6e-6, + "range": [256000.0, 1000000.0] } ] }, @@ -7564,39 +6862,27 @@ "tiered_pricing": [ { "cache_read_input_token_cost": 1e-7, - "input_cost_per_token": 0.000001, - "output_cost_per_token": 0.000005, - "range": [ - 0, - 32000 - ] + "input_cost_per_token": 1e-6, + "output_cost_per_token": 5e-6, + "range": [0, 32000.0] }, { "cache_read_input_token_cost": 1.8e-7, - "input_cost_per_token": 0.0000018, - "output_cost_per_token": 0.000009, - "range": [ - 32000, - 128000 - ] + "input_cost_per_token": 1.8e-6, + "output_cost_per_token": 9e-6, + "range": [32000.0, 128000.0] }, { "cache_read_input_token_cost": 3e-7, - "input_cost_per_token": 0.000003, - "output_cost_per_token": 0.000015, - "range": [ - 128000, - 256000 - ] + "input_cost_per_token": 3e-6, + "output_cost_per_token": 1.5e-5, + "range": [128000.0, 256000.0] }, { "cache_read_input_token_cost": 6e-7, - "input_cost_per_token": 0.000006, - "output_cost_per_token": 0.00006, - "range": [ - 256000, - 1000000 - ] + "input_cost_per_token": 6e-6, + "output_cost_per_token": 6e-5, + "range": [256000.0, 1000000.0] } ] }, @@ -7612,36 +6898,24 @@ "supports_tool_choice": true, "tiered_pricing": [ { - "input_cost_per_token": 0.000001, - "output_cost_per_token": 0.000005, - "range": [ - 0, - 32000 - ] + "input_cost_per_token": 1e-6, + "output_cost_per_token": 5e-6, + "range": [0, 32000.0] }, { - "input_cost_per_token": 0.0000018, - "output_cost_per_token": 0.000009, - "range": [ - 32000, - 128000 - ] + "input_cost_per_token": 1.8e-6, + "output_cost_per_token": 9e-6, + "range": [32000.0, 128000.0] }, { - "input_cost_per_token": 0.000003, - "output_cost_per_token": 0.000015, - "range": [ - 128000, - 256000 - ] + "input_cost_per_token": 3e-6, + "output_cost_per_token": 1.5e-5, + "range": [128000.0, 256000.0] }, { - "input_cost_per_token": 0.000006, - "output_cost_per_token": 0.00006, - "range": [ - 256000, - 1000000 - ] + "input_cost_per_token": 6e-6, + "output_cost_per_token": 6e-5, + "range": [256000.0, 1000000.0] } ] }, @@ -7657,28 +6931,19 @@ "supports_tool_choice": true, "tiered_pricing": [ { - "input_cost_per_token": 0.0000012, - "output_cost_per_token": 0.000006, - "range": [ - 0, - 32000 - ] + "input_cost_per_token": 1.2e-6, + "output_cost_per_token": 6e-6, + "range": [0, 32000.0] }, { - "input_cost_per_token": 0.0000024, - "output_cost_per_token": 0.000012, - "range": [ - 32000, - 128000 - ] + "input_cost_per_token": 2.4e-6, + "output_cost_per_token": 1.2e-5, + "range": [32000.0, 128000.0] }, { - "input_cost_per_token": 0.000003, - "output_cost_per_token": 0.000015, - "range": [ - 128000, - 252000 - ] + "input_cost_per_token": 3e-6, + "output_cost_per_token": 1.5e-5, + "range": [128000.0, 252000.0] } ] }, @@ -7689,7 +6954,7 @@ "max_output_tokens": 8192, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0.0000024, + "output_cost_per_token": 2.4e-6, "source": "https://www.alibabacloud.com/help/en/model-studio/models", "supports_function_calling": true, "supports_reasoning": true, @@ -7697,7 +6962,7 @@ }, "databricks/databricks-bge-large-en": { "input_cost_per_token": 1.0003e-7, - "input_dbu_cost_per_token": 0.000001429, + "input_dbu_cost_per_token": 1.429e-6, "litellm_provider": "databricks", "max_input_tokens": 512, "max_tokens": 512, @@ -7705,14 +6970,14 @@ "notes": "Input/output cost per token is dbu cost * $0.070, based on databricks Llama 3.1 70B conversion. Number provided for reference, '*_dbu_cost_per_token' used in actual calculation." }, "mode": "embedding", - "output_cost_per_token": 0, - "output_dbu_cost_per_token": 0, + "output_cost_per_token": 0.0, + "output_dbu_cost_per_token": 0.0, "output_vector_size": 1024, "source": "https://www.databricks.com/product/pricing/foundation-model-serving" }, "databricks/databricks-claude-3-7-sonnet": { - "input_cost_per_token": 0.0000025, - "input_dbu_cost_per_token": 0.00003571, + "input_cost_per_token": 2.5e-6, + "input_dbu_cost_per_token": 3.571e-5, "litellm_provider": "databricks", "max_input_tokens": 200000, "max_output_tokens": 128000, @@ -7721,7 +6986,7 @@ "notes": "Input/output cost per token is dbu cost * $0.070, based on databricks Claude 3.7 conversion. Number provided for reference, '*_dbu_cost_per_token' used in actual calculation." }, "mode": "chat", - "output_cost_per_token": 0.000017857, + "output_cost_per_token": 1.7857e-5, "output_db_cost_per_token": 0.000214286, "source": "https://www.databricks.com/product/pricing/foundation-model-serving", "supports_assistant_prefill": true, @@ -7731,7 +6996,7 @@ }, "databricks/databricks-gte-large-en": { "input_cost_per_token": 1.2999e-7, - "input_dbu_cost_per_token": 0.000001857, + "input_dbu_cost_per_token": 1.857e-6, "litellm_provider": "databricks", "max_input_tokens": 8192, "max_tokens": 8192, @@ -7739,14 +7004,14 @@ "notes": "Input/output cost per token is dbu cost * $0.070, based on databricks Llama 3.1 70B conversion. Number provided for reference, '*_dbu_cost_per_token' used in actual calculation." }, "mode": "embedding", - "output_cost_per_token": 0, - "output_dbu_cost_per_token": 0, + "output_cost_per_token": 0.0, + "output_dbu_cost_per_token": 0.0, "output_vector_size": 1024, "source": "https://www.databricks.com/product/pricing/foundation-model-serving" }, "databricks/databricks-llama-2-70b-chat": { "input_cost_per_token": 5.0001e-7, - "input_dbu_cost_per_token": 0.000007143, + "input_dbu_cost_per_token": 7.143e-6, "litellm_provider": "databricks", "max_input_tokens": 4096, "max_output_tokens": 4096, @@ -7755,14 +7020,14 @@ "notes": "Input/output cost per token is dbu cost * $0.070, based on databricks Llama 3.1 70B conversion. Number provided for reference, '*_dbu_cost_per_token' used in actual calculation." }, "mode": "chat", - "output_cost_per_token": 0.0000015, - "output_dbu_cost_per_token": 0.000021429, + "output_cost_per_token": 1.5e-6, + "output_dbu_cost_per_token": 2.1429e-5, "source": "https://www.databricks.com/product/pricing/foundation-model-serving", "supports_tool_choice": true }, "databricks/databricks-llama-4-maverick": { - "input_cost_per_token": 0.000005, - "input_dbu_cost_per_token": 0.00007143, + "input_cost_per_token": 5e-6, + "input_dbu_cost_per_token": 7.143e-5, "litellm_provider": "databricks", "max_input_tokens": 128000, "max_output_tokens": 128000, @@ -7771,14 +7036,14 @@ "notes": "Databricks documentation now provides both DBU costs (_dbu_cost_per_token) and dollar costs(_cost_per_token)." }, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "output_dbu_cost_per_token": 0.00021429, "source": "https://www.databricks.com/product/pricing/foundation-model-serving", "supports_tool_choice": true }, "databricks/databricks-meta-llama-3-1-405b-instruct": { - "input_cost_per_token": 0.000005, - "input_dbu_cost_per_token": 0.000071429, + "input_cost_per_token": 5e-6, + "input_dbu_cost_per_token": 7.1429e-5, "litellm_provider": "databricks", "max_input_tokens": 128000, "max_output_tokens": 128000, @@ -7787,14 +7052,14 @@ "notes": "Input/output cost per token is dbu cost * $0.070, based on databricks Llama 3.1 70B conversion. Number provided for reference, '*_dbu_cost_per_token' used in actual calculation." }, "mode": "chat", - "output_cost_per_token": 0.00001500002, + "output_cost_per_token": 1.500002e-5, "output_db_cost_per_token": 0.000214286, "source": "https://www.databricks.com/product/pricing/foundation-model-serving", "supports_tool_choice": true }, "databricks/databricks-meta-llama-3-3-70b-instruct": { - "input_cost_per_token": 0.00000100002, - "input_dbu_cost_per_token": 0.000014286, + "input_cost_per_token": 1.00002e-6, + "input_dbu_cost_per_token": 1.4286e-5, "litellm_provider": "databricks", "max_input_tokens": 128000, "max_output_tokens": 128000, @@ -7803,14 +7068,14 @@ "notes": "Input/output cost per token is dbu cost * $0.070, based on databricks Llama 3.1 70B conversion. Number provided for reference, '*_dbu_cost_per_token' used in actual calculation." }, "mode": "chat", - "output_cost_per_token": 0.00000299999, - "output_dbu_cost_per_token": 0.000042857, + "output_cost_per_token": 2.99999e-6, + "output_dbu_cost_per_token": 4.2857e-5, "source": "https://www.databricks.com/product/pricing/foundation-model-serving", "supports_tool_choice": true }, "databricks/databricks-meta-llama-3-70b-instruct": { - "input_cost_per_token": 0.00000100002, - "input_dbu_cost_per_token": 0.000014286, + "input_cost_per_token": 1.00002e-6, + "input_dbu_cost_per_token": 1.4286e-5, "litellm_provider": "databricks", "max_input_tokens": 128000, "max_output_tokens": 128000, @@ -7819,14 +7084,14 @@ "notes": "Input/output cost per token is dbu cost * $0.070, based on databricks Llama 3.1 70B conversion. Number provided for reference, '*_dbu_cost_per_token' used in actual calculation." }, "mode": "chat", - "output_cost_per_token": 0.00000299999, - "output_dbu_cost_per_token": 0.000042857, + "output_cost_per_token": 2.99999e-6, + "output_dbu_cost_per_token": 4.2857e-5, "source": "https://www.databricks.com/product/pricing/foundation-model-serving", "supports_tool_choice": true }, "databricks/databricks-mixtral-8x7b-instruct": { "input_cost_per_token": 5.0001e-7, - "input_dbu_cost_per_token": 0.000007143, + "input_dbu_cost_per_token": 7.143e-6, "litellm_provider": "databricks", "max_input_tokens": 4096, "max_output_tokens": 4096, @@ -7836,13 +7101,13 @@ }, "mode": "chat", "output_cost_per_token": 9.9902e-7, - "output_dbu_cost_per_token": 0.000014286, + "output_dbu_cost_per_token": 1.4286e-5, "source": "https://www.databricks.com/product/pricing/foundation-model-serving", "supports_tool_choice": true }, "databricks/databricks-mpt-30b-instruct": { "input_cost_per_token": 9.9902e-7, - "input_dbu_cost_per_token": 0.000014286, + "input_dbu_cost_per_token": 1.4286e-5, "litellm_provider": "databricks", "max_input_tokens": 8192, "max_output_tokens": 8192, @@ -7852,13 +7117,13 @@ }, "mode": "chat", "output_cost_per_token": 9.9902e-7, - "output_dbu_cost_per_token": 0.000014286, + "output_dbu_cost_per_token": 1.4286e-5, "source": "https://www.databricks.com/product/pricing/foundation-model-serving", "supports_tool_choice": true }, "databricks/databricks-mpt-7b-instruct": { "input_cost_per_token": 5.0001e-7, - "input_dbu_cost_per_token": 0.000007143, + "input_dbu_cost_per_token": 7.143e-6, "litellm_provider": "databricks", "max_input_tokens": 8192, "max_output_tokens": 8192, @@ -7867,8 +7132,8 @@ "notes": "Input/output cost per token is dbu cost * $0.070, based on databricks Llama 3.1 70B conversion. Number provided for reference, '*_dbu_cost_per_token' used in actual calculation." }, "mode": "chat", - "output_cost_per_token": 0, - "output_dbu_cost_per_token": 0, + "output_cost_per_token": 0.0, + "output_dbu_cost_per_token": 0.0, "source": "https://www.databricks.com/product/pricing/foundation-model-serving", "supports_tool_choice": true }, @@ -7878,13 +7143,13 @@ "mode": "search" }, "davinci-002": { - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "text-completion-openai", "max_input_tokens": 16384, "max_output_tokens": 4096, "max_tokens": 16384, "mode": "completion", - "output_cost_per_token": 0.000002 + "output_cost_per_token": 2e-6 }, "deepgram/base": { "input_cost_per_second": 0.00020833, @@ -7894,11 +7159,9 @@ "original_pricing_per_minute": 0.0125 }, "mode": "audio_transcription", - "output_cost_per_second": 0, + "output_cost_per_second": 0.0, "source": "https://deepgram.com/pricing", - "supported_endpoints": [ - "/v1/audio/transcriptions" - ] + "supported_endpoints": ["/v1/audio/transcriptions"] }, "deepgram/base-conversationalai": { "input_cost_per_second": 0.00020833, @@ -7908,11 +7171,9 @@ "original_pricing_per_minute": 0.0125 }, "mode": "audio_transcription", - "output_cost_per_second": 0, + "output_cost_per_second": 0.0, "source": "https://deepgram.com/pricing", - "supported_endpoints": [ - "/v1/audio/transcriptions" - ] + "supported_endpoints": ["/v1/audio/transcriptions"] }, "deepgram/base-finance": { "input_cost_per_second": 0.00020833, @@ -7922,11 +7183,9 @@ "original_pricing_per_minute": 0.0125 }, "mode": "audio_transcription", - "output_cost_per_second": 0, + "output_cost_per_second": 0.0, "source": "https://deepgram.com/pricing", - "supported_endpoints": [ - "/v1/audio/transcriptions" - ] + "supported_endpoints": ["/v1/audio/transcriptions"] }, "deepgram/base-general": { "input_cost_per_second": 0.00020833, @@ -7936,11 +7195,9 @@ "original_pricing_per_minute": 0.0125 }, "mode": "audio_transcription", - "output_cost_per_second": 0, + "output_cost_per_second": 0.0, "source": "https://deepgram.com/pricing", - "supported_endpoints": [ - "/v1/audio/transcriptions" - ] + "supported_endpoints": ["/v1/audio/transcriptions"] }, "deepgram/base-meeting": { "input_cost_per_second": 0.00020833, @@ -7950,11 +7207,9 @@ "original_pricing_per_minute": 0.0125 }, "mode": "audio_transcription", - "output_cost_per_second": 0, + "output_cost_per_second": 0.0, "source": "https://deepgram.com/pricing", - "supported_endpoints": [ - "/v1/audio/transcriptions" - ] + "supported_endpoints": ["/v1/audio/transcriptions"] }, "deepgram/base-phonecall": { "input_cost_per_second": 0.00020833, @@ -7964,11 +7219,9 @@ "original_pricing_per_minute": 0.0125 }, "mode": "audio_transcription", - "output_cost_per_second": 0, + "output_cost_per_second": 0.0, "source": "https://deepgram.com/pricing", - "supported_endpoints": [ - "/v1/audio/transcriptions" - ] + "supported_endpoints": ["/v1/audio/transcriptions"] }, "deepgram/base-video": { "input_cost_per_second": 0.00020833, @@ -7978,11 +7231,9 @@ "original_pricing_per_minute": 0.0125 }, "mode": "audio_transcription", - "output_cost_per_second": 0, + "output_cost_per_second": 0.0, "source": "https://deepgram.com/pricing", - "supported_endpoints": [ - "/v1/audio/transcriptions" - ] + "supported_endpoints": ["/v1/audio/transcriptions"] }, "deepgram/base-voicemail": { "input_cost_per_second": 0.00020833, @@ -7992,11 +7243,9 @@ "original_pricing_per_minute": 0.0125 }, "mode": "audio_transcription", - "output_cost_per_second": 0, + "output_cost_per_second": 0.0, "source": "https://deepgram.com/pricing", - "supported_endpoints": [ - "/v1/audio/transcriptions" - ] + "supported_endpoints": ["/v1/audio/transcriptions"] }, "deepgram/enhanced": { "input_cost_per_second": 0.00024167, @@ -8006,11 +7255,9 @@ "original_pricing_per_minute": 0.0145 }, "mode": "audio_transcription", - "output_cost_per_second": 0, + "output_cost_per_second": 0.0, "source": "https://deepgram.com/pricing", - "supported_endpoints": [ - "/v1/audio/transcriptions" - ] + "supported_endpoints": ["/v1/audio/transcriptions"] }, "deepgram/enhanced-finance": { "input_cost_per_second": 0.00024167, @@ -8020,11 +7267,9 @@ "original_pricing_per_minute": 0.0145 }, "mode": "audio_transcription", - "output_cost_per_second": 0, + "output_cost_per_second": 0.0, "source": "https://deepgram.com/pricing", - "supported_endpoints": [ - "/v1/audio/transcriptions" - ] + "supported_endpoints": ["/v1/audio/transcriptions"] }, "deepgram/enhanced-general": { "input_cost_per_second": 0.00024167, @@ -8034,11 +7279,9 @@ "original_pricing_per_minute": 0.0145 }, "mode": "audio_transcription", - "output_cost_per_second": 0, + "output_cost_per_second": 0.0, "source": "https://deepgram.com/pricing", - "supported_endpoints": [ - "/v1/audio/transcriptions" - ] + "supported_endpoints": ["/v1/audio/transcriptions"] }, "deepgram/enhanced-meeting": { "input_cost_per_second": 0.00024167, @@ -8048,11 +7291,9 @@ "original_pricing_per_minute": 0.0145 }, "mode": "audio_transcription", - "output_cost_per_second": 0, + "output_cost_per_second": 0.0, "source": "https://deepgram.com/pricing", - "supported_endpoints": [ - "/v1/audio/transcriptions" - ] + "supported_endpoints": ["/v1/audio/transcriptions"] }, "deepgram/enhanced-phonecall": { "input_cost_per_second": 0.00024167, @@ -8062,249 +7303,213 @@ "original_pricing_per_minute": 0.0145 }, "mode": "audio_transcription", - "output_cost_per_second": 0, + "output_cost_per_second": 0.0, "source": "https://deepgram.com/pricing", - "supported_endpoints": [ - "/v1/audio/transcriptions" - ] + "supported_endpoints": ["/v1/audio/transcriptions"] }, "deepgram/nova": { - "input_cost_per_second": 0.00007167, + "input_cost_per_second": 7.167e-5, "litellm_provider": "deepgram", "metadata": { "calculation": "$0.0043/60 seconds = $0.00007167 per second", "original_pricing_per_minute": 0.0043 }, "mode": "audio_transcription", - "output_cost_per_second": 0, + "output_cost_per_second": 0.0, "source": "https://deepgram.com/pricing", - "supported_endpoints": [ - "/v1/audio/transcriptions" - ] + "supported_endpoints": ["/v1/audio/transcriptions"] }, "deepgram/nova-2": { - "input_cost_per_second": 0.00007167, + "input_cost_per_second": 7.167e-5, "litellm_provider": "deepgram", "metadata": { "calculation": "$0.0043/60 seconds = $0.00007167 per second", "original_pricing_per_minute": 0.0043 }, "mode": "audio_transcription", - "output_cost_per_second": 0, + "output_cost_per_second": 0.0, "source": "https://deepgram.com/pricing", - "supported_endpoints": [ - "/v1/audio/transcriptions" - ] + "supported_endpoints": ["/v1/audio/transcriptions"] }, "deepgram/nova-2-atc": { - "input_cost_per_second": 0.00007167, + "input_cost_per_second": 7.167e-5, "litellm_provider": "deepgram", "metadata": { "calculation": "$0.0043/60 seconds = $0.00007167 per second", "original_pricing_per_minute": 0.0043 }, "mode": "audio_transcription", - "output_cost_per_second": 0, + "output_cost_per_second": 0.0, "source": "https://deepgram.com/pricing", - "supported_endpoints": [ - "/v1/audio/transcriptions" - ] + "supported_endpoints": ["/v1/audio/transcriptions"] }, "deepgram/nova-2-automotive": { - "input_cost_per_second": 0.00007167, + "input_cost_per_second": 7.167e-5, "litellm_provider": "deepgram", "metadata": { "calculation": "$0.0043/60 seconds = $0.00007167 per second", "original_pricing_per_minute": 0.0043 }, "mode": "audio_transcription", - "output_cost_per_second": 0, + "output_cost_per_second": 0.0, "source": "https://deepgram.com/pricing", - "supported_endpoints": [ - "/v1/audio/transcriptions" - ] + "supported_endpoints": ["/v1/audio/transcriptions"] }, "deepgram/nova-2-conversationalai": { - "input_cost_per_second": 0.00007167, + "input_cost_per_second": 7.167e-5, "litellm_provider": "deepgram", "metadata": { "calculation": "$0.0043/60 seconds = $0.00007167 per second", "original_pricing_per_minute": 0.0043 }, "mode": "audio_transcription", - "output_cost_per_second": 0, + "output_cost_per_second": 0.0, "source": "https://deepgram.com/pricing", - "supported_endpoints": [ - "/v1/audio/transcriptions" - ] + "supported_endpoints": ["/v1/audio/transcriptions"] }, "deepgram/nova-2-drivethru": { - "input_cost_per_second": 0.00007167, + "input_cost_per_second": 7.167e-5, "litellm_provider": "deepgram", "metadata": { "calculation": "$0.0043/60 seconds = $0.00007167 per second", "original_pricing_per_minute": 0.0043 }, "mode": "audio_transcription", - "output_cost_per_second": 0, + "output_cost_per_second": 0.0, "source": "https://deepgram.com/pricing", - "supported_endpoints": [ - "/v1/audio/transcriptions" - ] + "supported_endpoints": ["/v1/audio/transcriptions"] }, "deepgram/nova-2-finance": { - "input_cost_per_second": 0.00007167, + "input_cost_per_second": 7.167e-5, "litellm_provider": "deepgram", "metadata": { "calculation": "$0.0043/60 seconds = $0.00007167 per second", "original_pricing_per_minute": 0.0043 }, "mode": "audio_transcription", - "output_cost_per_second": 0, + "output_cost_per_second": 0.0, "source": "https://deepgram.com/pricing", - "supported_endpoints": [ - "/v1/audio/transcriptions" - ] + "supported_endpoints": ["/v1/audio/transcriptions"] }, "deepgram/nova-2-general": { - "input_cost_per_second": 0.00007167, + "input_cost_per_second": 7.167e-5, "litellm_provider": "deepgram", "metadata": { "calculation": "$0.0043/60 seconds = $0.00007167 per second", "original_pricing_per_minute": 0.0043 }, "mode": "audio_transcription", - "output_cost_per_second": 0, + "output_cost_per_second": 0.0, "source": "https://deepgram.com/pricing", - "supported_endpoints": [ - "/v1/audio/transcriptions" - ] + "supported_endpoints": ["/v1/audio/transcriptions"] }, "deepgram/nova-2-meeting": { - "input_cost_per_second": 0.00007167, + "input_cost_per_second": 7.167e-5, "litellm_provider": "deepgram", "metadata": { "calculation": "$0.0043/60 seconds = $0.00007167 per second", "original_pricing_per_minute": 0.0043 }, "mode": "audio_transcription", - "output_cost_per_second": 0, + "output_cost_per_second": 0.0, "source": "https://deepgram.com/pricing", - "supported_endpoints": [ - "/v1/audio/transcriptions" - ] + "supported_endpoints": ["/v1/audio/transcriptions"] }, "deepgram/nova-2-phonecall": { - "input_cost_per_second": 0.00007167, + "input_cost_per_second": 7.167e-5, "litellm_provider": "deepgram", "metadata": { "calculation": "$0.0043/60 seconds = $0.00007167 per second", "original_pricing_per_minute": 0.0043 }, "mode": "audio_transcription", - "output_cost_per_second": 0, + "output_cost_per_second": 0.0, "source": "https://deepgram.com/pricing", - "supported_endpoints": [ - "/v1/audio/transcriptions" - ] + "supported_endpoints": ["/v1/audio/transcriptions"] }, "deepgram/nova-2-video": { - "input_cost_per_second": 0.00007167, + "input_cost_per_second": 7.167e-5, "litellm_provider": "deepgram", "metadata": { "calculation": "$0.0043/60 seconds = $0.00007167 per second", "original_pricing_per_minute": 0.0043 }, "mode": "audio_transcription", - "output_cost_per_second": 0, + "output_cost_per_second": 0.0, "source": "https://deepgram.com/pricing", - "supported_endpoints": [ - "/v1/audio/transcriptions" - ] + "supported_endpoints": ["/v1/audio/transcriptions"] }, "deepgram/nova-2-voicemail": { - "input_cost_per_second": 0.00007167, + "input_cost_per_second": 7.167e-5, "litellm_provider": "deepgram", "metadata": { "calculation": "$0.0043/60 seconds = $0.00007167 per second", "original_pricing_per_minute": 0.0043 }, "mode": "audio_transcription", - "output_cost_per_second": 0, + "output_cost_per_second": 0.0, "source": "https://deepgram.com/pricing", - "supported_endpoints": [ - "/v1/audio/transcriptions" - ] + "supported_endpoints": ["/v1/audio/transcriptions"] }, "deepgram/nova-3": { - "input_cost_per_second": 0.00007167, + "input_cost_per_second": 7.167e-5, "litellm_provider": "deepgram", "metadata": { "calculation": "$0.0043/60 seconds = $0.00007167 per second", "original_pricing_per_minute": 0.0043 }, "mode": "audio_transcription", - "output_cost_per_second": 0, + "output_cost_per_second": 0.0, "source": "https://deepgram.com/pricing", - "supported_endpoints": [ - "/v1/audio/transcriptions" - ] + "supported_endpoints": ["/v1/audio/transcriptions"] }, "deepgram/nova-3-general": { - "input_cost_per_second": 0.00007167, + "input_cost_per_second": 7.167e-5, "litellm_provider": "deepgram", "metadata": { "calculation": "$0.0043/60 seconds = $0.00007167 per second", "original_pricing_per_minute": 0.0043 }, "mode": "audio_transcription", - "output_cost_per_second": 0, + "output_cost_per_second": 0.0, "source": "https://deepgram.com/pricing", - "supported_endpoints": [ - "/v1/audio/transcriptions" - ] + "supported_endpoints": ["/v1/audio/transcriptions"] }, "deepgram/nova-3-medical": { - "input_cost_per_second": 0.00008667, + "input_cost_per_second": 8.667e-5, "litellm_provider": "deepgram", "metadata": { "calculation": "$0.0052/60 seconds = $0.00008667 per second (multilingual)", "original_pricing_per_minute": 0.0052 }, "mode": "audio_transcription", - "output_cost_per_second": 0, + "output_cost_per_second": 0.0, "source": "https://deepgram.com/pricing", - "supported_endpoints": [ - "/v1/audio/transcriptions" - ] + "supported_endpoints": ["/v1/audio/transcriptions"] }, "deepgram/nova-general": { - "input_cost_per_second": 0.00007167, + "input_cost_per_second": 7.167e-5, "litellm_provider": "deepgram", "metadata": { "calculation": "$0.0043/60 seconds = $0.00007167 per second", "original_pricing_per_minute": 0.0043 }, "mode": "audio_transcription", - "output_cost_per_second": 0, + "output_cost_per_second": 0.0, "source": "https://deepgram.com/pricing", - "supported_endpoints": [ - "/v1/audio/transcriptions" - ] + "supported_endpoints": ["/v1/audio/transcriptions"] }, "deepgram/nova-phonecall": { - "input_cost_per_second": 0.00007167, + "input_cost_per_second": 7.167e-5, "litellm_provider": "deepgram", "metadata": { "calculation": "$0.0043/60 seconds = $0.00007167 per second", "original_pricing_per_minute": 0.0043 }, "mode": "audio_transcription", - "output_cost_per_second": 0, + "output_cost_per_second": 0.0, "source": "https://deepgram.com/pricing", - "supported_endpoints": [ - "/v1/audio/transcriptions" - ] + "supported_endpoints": ["/v1/audio/transcriptions"] }, "deepgram/whisper": { "input_cost_per_second": 0.0001, @@ -8313,11 +7518,9 @@ "notes": "Deepgram's hosted OpenAI Whisper models - pricing may differ from native Deepgram models" }, "mode": "audio_transcription", - "output_cost_per_second": 0, + "output_cost_per_second": 0.0, "source": "https://deepgram.com/pricing", - "supported_endpoints": [ - "/v1/audio/transcriptions" - ] + "supported_endpoints": ["/v1/audio/transcriptions"] }, "deepgram/whisper-base": { "input_cost_per_second": 0.0001, @@ -8326,11 +7529,9 @@ "notes": "Deepgram's hosted OpenAI Whisper models - pricing may differ from native Deepgram models" }, "mode": "audio_transcription", - "output_cost_per_second": 0, + "output_cost_per_second": 0.0, "source": "https://deepgram.com/pricing", - "supported_endpoints": [ - "/v1/audio/transcriptions" - ] + "supported_endpoints": ["/v1/audio/transcriptions"] }, "deepgram/whisper-large": { "input_cost_per_second": 0.0001, @@ -8339,11 +7540,9 @@ "notes": "Deepgram's hosted OpenAI Whisper models - pricing may differ from native Deepgram models" }, "mode": "audio_transcription", - "output_cost_per_second": 0, + "output_cost_per_second": 0.0, "source": "https://deepgram.com/pricing", - "supported_endpoints": [ - "/v1/audio/transcriptions" - ] + "supported_endpoints": ["/v1/audio/transcriptions"] }, "deepgram/whisper-medium": { "input_cost_per_second": 0.0001, @@ -8352,11 +7551,9 @@ "notes": "Deepgram's hosted OpenAI Whisper models - pricing may differ from native Deepgram models" }, "mode": "audio_transcription", - "output_cost_per_second": 0, + "output_cost_per_second": 0.0, "source": "https://deepgram.com/pricing", - "supported_endpoints": [ - "/v1/audio/transcriptions" - ] + "supported_endpoints": ["/v1/audio/transcriptions"] }, "deepgram/whisper-small": { "input_cost_per_second": 0.0001, @@ -8365,11 +7562,9 @@ "notes": "Deepgram's hosted OpenAI Whisper models - pricing may differ from native Deepgram models" }, "mode": "audio_transcription", - "output_cost_per_second": 0, + "output_cost_per_second": 0.0, "source": "https://deepgram.com/pricing", - "supported_endpoints": [ - "/v1/audio/transcriptions" - ] + "supported_endpoints": ["/v1/audio/transcriptions"] }, "deepgram/whisper-tiny": { "input_cost_per_second": 0.0001, @@ -8378,11 +7573,9 @@ "notes": "Deepgram's hosted OpenAI Whisper models - pricing may differ from native Deepgram models" }, "mode": "audio_transcription", - "output_cost_per_second": 0, + "output_cost_per_second": 0.0, "source": "https://deepgram.com/pricing", - "supported_endpoints": [ - "/v1/audio/transcriptions" - ] + "supported_endpoints": ["/v1/audio/transcriptions"] }, "deepinfra/Gryphe/MythoMax-L2-13b": { "max_tokens": 4096, @@ -8398,8 +7591,8 @@ "max_tokens": 131072, "max_input_tokens": 131072, "max_output_tokens": 131072, - "input_cost_per_token": 0.000001, - "output_cost_per_token": 0.000001, + "input_cost_per_token": 1e-6, + "output_cost_per_token": 1e-6, "litellm_provider": "deepinfra", "mode": "chat", "supports_tool_choice": true @@ -8490,7 +7683,7 @@ "max_input_tokens": 262144, "max_output_tokens": 262144, "input_cost_per_token": 3e-7, - "output_cost_per_token": 0.0000029, + "output_cost_per_token": 2.9e-6, "litellm_provider": "deepinfra", "mode": "chat", "supports_tool_choice": true @@ -8520,7 +7713,7 @@ "max_input_tokens": 262144, "max_output_tokens": 262144, "input_cost_per_token": 4e-7, - "output_cost_per_token": 0.0000016, + "output_cost_per_token": 1.6e-6, "litellm_provider": "deepinfra", "mode": "chat", "supports_tool_choice": true @@ -8530,7 +7723,7 @@ "max_input_tokens": 262144, "max_output_tokens": 262144, "input_cost_per_token": 2.9e-7, - "output_cost_per_token": 0.0000012, + "output_cost_per_token": 1.2e-6, "litellm_provider": "deepinfra", "mode": "chat", "supports_tool_choice": true @@ -8540,7 +7733,7 @@ "max_input_tokens": 262144, "max_output_tokens": 262144, "input_cost_per_token": 1.4e-7, - "output_cost_per_token": 0.0000014, + "output_cost_per_token": 1.4e-6, "litellm_provider": "deepinfra", "mode": "chat", "supports_tool_choice": true @@ -8550,7 +7743,7 @@ "max_input_tokens": 262144, "max_output_tokens": 262144, "input_cost_per_token": 1.4e-7, - "output_cost_per_token": 0.0000014, + "output_cost_per_token": 1.4e-6, "litellm_provider": "deepinfra", "mode": "chat", "supports_tool_choice": true @@ -8590,7 +7783,7 @@ "max_input_tokens": 16384, "max_output_tokens": 16384, "input_cost_per_token": 2.7e-7, - "output_cost_per_token": 0.0000015, + "output_cost_per_token": 1.5e-6, "litellm_provider": "deepinfra", "mode": "chat", "supports_tool_choice": false @@ -8599,8 +7792,8 @@ "max_tokens": 200000, "max_input_tokens": 200000, "max_output_tokens": 200000, - "input_cost_per_token": 0.0000033, - "output_cost_per_token": 0.0000165, + "input_cost_per_token": 3.3e-6, + "output_cost_per_token": 1.65e-5, "cache_read_input_token_cost": 3.3e-7, "litellm_provider": "deepinfra", "mode": "chat", @@ -8610,8 +7803,8 @@ "max_tokens": 200000, "max_input_tokens": 200000, "max_output_tokens": 200000, - "input_cost_per_token": 0.0000165, - "output_cost_per_token": 0.0000825, + "input_cost_per_token": 1.65e-5, + "output_cost_per_token": 8.25e-5, "litellm_provider": "deepinfra", "mode": "chat", "supports_tool_choice": true @@ -8620,8 +7813,8 @@ "max_tokens": 200000, "max_input_tokens": 200000, "max_output_tokens": 200000, - "input_cost_per_token": 0.0000033, - "output_cost_per_token": 0.0000165, + "input_cost_per_token": 3.3e-6, + "output_cost_per_token": 1.65e-5, "litellm_provider": "deepinfra", "mode": "chat", "supports_tool_choice": true @@ -8631,7 +7824,7 @@ "max_input_tokens": 163840, "max_output_tokens": 163840, "input_cost_per_token": 7e-7, - "output_cost_per_token": 0.0000024, + "output_cost_per_token": 2.4e-6, "litellm_provider": "deepinfra", "mode": "chat", "supports_tool_choice": true @@ -8641,7 +7834,7 @@ "max_input_tokens": 163840, "max_output_tokens": 163840, "input_cost_per_token": 5e-7, - "output_cost_per_token": 0.00000215, + "output_cost_per_token": 2.15e-6, "cache_read_input_token_cost": 4e-7, "litellm_provider": "deepinfra", "mode": "chat", @@ -8651,8 +7844,8 @@ "max_tokens": 32768, "max_input_tokens": 32768, "max_output_tokens": 32768, - "input_cost_per_token": 0.000001, - "output_cost_per_token": 0.000003, + "input_cost_per_token": 1e-6, + "output_cost_per_token": 3e-6, "litellm_provider": "deepinfra", "mode": "chat", "supports_tool_choice": true @@ -8681,8 +7874,8 @@ "max_tokens": 40960, "max_input_tokens": 40960, "max_output_tokens": 40960, - "input_cost_per_token": 0.000001, - "output_cost_per_token": 0.000003, + "input_cost_per_token": 1e-6, + "output_cost_per_token": 3e-6, "litellm_provider": "deepinfra", "mode": "chat", "supports_tool_choice": true @@ -8712,7 +7905,7 @@ "max_input_tokens": 163840, "max_output_tokens": 163840, "input_cost_per_token": 2.7e-7, - "output_cost_per_token": 0.000001, + "output_cost_per_token": 1e-6, "cache_read_input_token_cost": 2.16e-7, "litellm_provider": "deepinfra", "mode": "chat", @@ -8724,7 +7917,7 @@ "max_input_tokens": 163840, "max_output_tokens": 163840, "input_cost_per_token": 2.7e-7, - "output_cost_per_token": 0.000001, + "output_cost_per_token": 1e-6, "cache_read_input_token_cost": 2.16e-7, "litellm_provider": "deepinfra", "mode": "chat", @@ -8745,7 +7938,7 @@ "max_input_tokens": 1000000, "max_output_tokens": 1000000, "input_cost_per_token": 3e-7, - "output_cost_per_token": 0.0000025, + "output_cost_per_token": 2.5e-6, "litellm_provider": "deepinfra", "mode": "chat", "supports_tool_choice": true @@ -8754,8 +7947,8 @@ "max_tokens": 1000000, "max_input_tokens": 1000000, "max_output_tokens": 1000000, - "input_cost_per_token": 0.00000125, - "output_cost_per_token": 0.00001, + "input_cost_per_token": 1.25e-6, + "output_cost_per_token": 1e-5, "litellm_provider": "deepinfra", "mode": "chat", "supports_tool_choice": true @@ -8985,7 +8178,7 @@ "max_input_tokens": 131072, "max_output_tokens": 131072, "input_cost_per_token": 5e-7, - "output_cost_per_token": 0.000002, + "output_cost_per_token": 2e-6, "litellm_provider": "deepinfra", "mode": "chat", "supports_tool_choice": true @@ -8995,7 +8188,7 @@ "max_input_tokens": 262144, "max_output_tokens": 262144, "input_cost_per_token": 5e-7, - "output_cost_per_token": 0.000002, + "output_cost_per_token": 2e-6, "cache_read_input_token_cost": 4e-7, "litellm_provider": "deepinfra", "mode": "chat", @@ -9056,13 +8249,13 @@ "max_input_tokens": 131072, "max_output_tokens": 131072, "input_cost_per_token": 4e-7, - "output_cost_per_token": 0.0000016, + "output_cost_per_token": 1.6e-6, "litellm_provider": "deepinfra", "mode": "chat", "supports_tool_choice": true }, "deepseek/deepseek-chat": { - "cache_creation_input_token_cost": 0, + "cache_creation_input_token_cost": 0.0, "cache_read_input_token_cost": 7e-8, "input_cost_per_token": 2.7e-7, "input_cost_per_token_cache_hit": 7e-8, @@ -9071,7 +8264,7 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.0000011, + "output_cost_per_token": 1.1e-6, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_prompt_caching": true, @@ -9099,7 +8292,7 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.00000219, + "output_cost_per_token": 2.19e-6, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_prompt_caching": true, @@ -9114,7 +8307,7 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.00000219, + "output_cost_per_token": 2.19e-6, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_prompt_caching": true, @@ -9122,7 +8315,7 @@ "supports_tool_choice": true }, "deepseek/deepseek-v3": { - "cache_creation_input_token_cost": 0, + "cache_creation_input_token_cost": 0.0, "cache_read_input_token_cost": 7e-8, "input_cost_per_token": 2.7e-7, "input_cost_per_token_cache_hit": 7e-8, @@ -9131,7 +8324,7 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.0000011, + "output_cost_per_token": 1.1e-6, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_prompt_caching": true, @@ -9144,7 +8337,7 @@ "max_output_tokens": 81920, "max_tokens": 163840, "mode": "chat", - "output_cost_per_token": 0.00000168, + "output_cost_per_token": 1.68e-6, "supports_function_calling": true, "supports_reasoning": true, "supports_tool_choice": true @@ -9159,7 +8352,7 @@ "output_cost_per_token": 5e-7 }, "doubao-embedding": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "volcengine", "max_input_tokens": 4096, "max_tokens": 4096, @@ -9167,11 +8360,11 @@ "notes": "Volcengine Doubao embedding model - standard version with 2560 dimensions" }, "mode": "embedding", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "output_vector_size": 2560 }, "doubao-embedding-large": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "volcengine", "max_input_tokens": 4096, "max_tokens": 4096, @@ -9179,11 +8372,11 @@ "notes": "Volcengine Doubao embedding model - large version with 2048 dimensions" }, "mode": "embedding", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "output_vector_size": 2048 }, "doubao-embedding-large-text-240915": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "volcengine", "max_input_tokens": 4096, "max_tokens": 4096, @@ -9191,11 +8384,11 @@ "notes": "Volcengine Doubao embedding model - text-240915 version with 4096 dimensions" }, "mode": "embedding", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "output_vector_size": 4096 }, "doubao-embedding-large-text-250515": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "volcengine", "max_input_tokens": 4096, "max_tokens": 4096, @@ -9203,11 +8396,11 @@ "notes": "Volcengine Doubao embedding model - text-250515 version with 2048 dimensions" }, "mode": "embedding", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "output_vector_size": 2048 }, "doubao-embedding-text-240715": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "volcengine", "max_input_tokens": 4096, "max_tokens": 4096, @@ -9215,7 +8408,7 @@ "notes": "Volcengine Doubao embedding model - text-240715 version with 2560 dimensions" }, "mode": "embedding", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "output_vector_size": 2560 }, "exa_ai/search": { @@ -9223,18 +8416,12 @@ "mode": "search", "tiered_pricing": [ { - "input_cost_per_query": 0.005, - "max_results_range": [ - 0, - 25 - ] + "input_cost_per_query": 5e-3, + "max_results_range": [0, 25] }, { - "input_cost_per_query": 0.025, - "max_results_range": [ - 26, - 100 - ] + "input_cost_per_query": 25e-3, + "max_results_range": [26, 100] } ] }, @@ -9243,74 +8430,44 @@ "mode": "search", "tiered_pricing": [ { - "input_cost_per_query": 0.00166, - "max_results_range": [ - 1, - 10 - ] + "input_cost_per_query": 1.66e-3, + "max_results_range": [1, 10] }, { - "input_cost_per_query": 0.00332, - "max_results_range": [ - 11, - 20 - ] + "input_cost_per_query": 3.32e-3, + "max_results_range": [11, 20] }, { - "input_cost_per_query": 0.00498, - "max_results_range": [ - 21, - 30 - ] + "input_cost_per_query": 4.98e-3, + "max_results_range": [21, 30] }, { - "input_cost_per_query": 0.00664, - "max_results_range": [ - 31, - 40 - ] + "input_cost_per_query": 6.64e-3, + "max_results_range": [31, 40] }, { - "input_cost_per_query": 0.0083, - "max_results_range": [ - 41, - 50 - ] + "input_cost_per_query": 8.3e-3, + "max_results_range": [41, 50] }, { - "input_cost_per_query": 0.00996, - "max_results_range": [ - 51, - 60 - ] + "input_cost_per_query": 9.96e-3, + "max_results_range": [51, 60] }, { - "input_cost_per_query": 0.01162, - "max_results_range": [ - 61, - 70 - ] + "input_cost_per_query": 11.62e-3, + "max_results_range": [61, 70] }, { - "input_cost_per_query": 0.01328, - "max_results_range": [ - 71, - 80 - ] + "input_cost_per_query": 13.28e-3, + "max_results_range": [71, 80] }, { - "input_cost_per_query": 0.01494, - "max_results_range": [ - 81, - 90 - ] + "input_cost_per_query": 14.94e-3, + "max_results_range": [81, 90] }, { - "input_cost_per_query": 0.0166, - "max_results_range": [ - 91, - 100 - ] + "input_cost_per_query": 16.6e-3, + "max_results_range": [91, 100] } ], "metadata": { @@ -9318,20 +8475,20 @@ } }, "perplexity/search": { - "input_cost_per_query": 0.005, + "input_cost_per_query": 5e-3, "litellm_provider": "perplexity", "mode": "search" }, "searxng/search": { "litellm_provider": "searxng", "mode": "search", - "input_cost_per_query": 0, + "input_cost_per_query": 0.0, "metadata": { "notes": "SearXNG is an open-source metasearch engine. Free to use when self-hosted or using public instances." } }, "elevenlabs/scribe_v1": { - "input_cost_per_second": 0.0000611, + "input_cost_per_second": 6.11e-5, "litellm_provider": "elevenlabs", "metadata": { "calculation": "$0.22/hour = $0.00366/minute = $0.0000611 per second (enterprise pricing)", @@ -9339,14 +8496,12 @@ "original_pricing_per_hour": 0.22 }, "mode": "audio_transcription", - "output_cost_per_second": 0, + "output_cost_per_second": 0.0, "source": "https://elevenlabs.io/pricing", - "supported_endpoints": [ - "/v1/audio/transcriptions" - ] + "supported_endpoints": ["/v1/audio/transcriptions"] }, "elevenlabs/scribe_v1_experimental": { - "input_cost_per_second": 0.0000611, + "input_cost_per_second": 6.11e-5, "litellm_provider": "elevenlabs", "metadata": { "calculation": "$0.22/hour = $0.00366/minute = $0.0000611 per second (enterprise pricing)", @@ -9354,11 +8509,9 @@ "original_pricing_per_hour": 0.22 }, "mode": "audio_transcription", - "output_cost_per_second": 0, + "output_cost_per_second": 0.0, "source": "https://elevenlabs.io/pricing", - "supported_endpoints": [ - "/v1/audio/transcriptions" - ] + "supported_endpoints": ["/v1/audio/transcriptions"] }, "embed-english-light-v2.0": { "input_cost_per_token": 1e-7, @@ -9366,7 +8519,7 @@ "max_input_tokens": 1024, "max_tokens": 1024, "mode": "embedding", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "embed-english-light-v3.0": { "input_cost_per_token": 1e-7, @@ -9374,7 +8527,7 @@ "max_input_tokens": 1024, "max_tokens": 1024, "mode": "embedding", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "embed-english-v2.0": { "input_cost_per_token": 1e-7, @@ -9382,7 +8535,7 @@ "max_input_tokens": 4096, "max_tokens": 4096, "mode": "embedding", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "embed-english-v3.0": { "input_cost_per_image": 0.0001, @@ -9394,7 +8547,7 @@ "notes": "'supports_image_input' is a deprecated field. Use 'supports_embedding_image_input' instead." }, "mode": "embedding", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "supports_embedding_image_input": true, "supports_image_input": true }, @@ -9404,7 +8557,7 @@ "max_input_tokens": 768, "max_tokens": 768, "mode": "embedding", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "embed-multilingual-v3.0": { "input_cost_per_token": 1e-7, @@ -9412,7 +8565,7 @@ "max_input_tokens": 1024, "max_tokens": 1024, "mode": "embedding", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "supports_embedding_image_input": true }, "eu.amazon.nova-lite-v1:0": { @@ -9442,13 +8595,13 @@ "supports_response_schema": true }, "eu.amazon.nova-pro-v1:0": { - "input_cost_per_token": 0.00000105, + "input_cost_per_token": 1.05e-6, "litellm_provider": "bedrock_converse", "max_input_tokens": 300000, "max_output_tokens": 10000, "max_tokens": 10000, "mode": "chat", - "output_cost_per_token": 0.0000042, + "output_cost_per_token": 4.2e-6, "source": "https://aws.amazon.com/bedrock/pricing/", "supports_function_calling": true, "supports_pdf_input": true, @@ -9463,7 +8616,7 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.00000125, + "output_cost_per_token": 1.25e-6, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_pdf_input": true, @@ -9472,16 +8625,16 @@ "supports_tool_choice": true }, "eu.anthropic.claude-haiku-4-5-20251001-v1:0": { - "cache_creation_input_token_cost": 0.000001375, + "cache_creation_input_token_cost": 1.375e-6, "cache_read_input_token_cost": 1.1e-7, - "input_cost_per_token": 0.0000011, + "input_cost_per_token": 1.1e-6, "deprecation_date": "2026-10-15", "litellm_provider": "bedrock_converse", "max_input_tokens": 200000, "max_output_tokens": 64000, "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 0.0000055, + "output_cost_per_token": 5.5e-6, "source": "https://aws.amazon.com/about-aws/whats-new/2025/10/claude-4-5-haiku-anthropic-amazon-bedrock", "supports_assistant_prefill": true, "supports_computer_use": true, @@ -9495,13 +8648,13 @@ "tool_use_system_prompt_tokens": 346 }, "eu.anthropic.claude-3-5-sonnet-20240620-v1:0": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "bedrock", "max_input_tokens": 200000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "supports_function_calling": true, "supports_pdf_input": true, "supports_response_schema": true, @@ -9509,13 +8662,13 @@ "supports_vision": true }, "eu.anthropic.claude-3-5-sonnet-20241022-v2:0": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "bedrock", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "supports_assistant_prefill": true, "supports_computer_use": true, "supports_function_calling": true, @@ -9526,13 +8679,13 @@ "supports_vision": true }, "eu.anthropic.claude-3-7-sonnet-20250219-v1:0": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "bedrock", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "supports_assistant_prefill": true, "supports_computer_use": true, "supports_function_calling": true, @@ -9550,7 +8703,7 @@ "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.00000125, + "output_cost_per_token": 1.25e-6, "supports_function_calling": true, "supports_pdf_input": true, "supports_response_schema": true, @@ -9558,26 +8711,26 @@ "supports_vision": true }, "eu.anthropic.claude-3-opus-20240229-v1:0": { - "input_cost_per_token": 0.000015, + "input_cost_per_token": 1.5e-5, "litellm_provider": "bedrock", "max_input_tokens": 200000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.000075, + "output_cost_per_token": 7.5e-5, "supports_function_calling": true, "supports_response_schema": true, "supports_tool_choice": true, "supports_vision": true }, "eu.anthropic.claude-3-sonnet-20240229-v1:0": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "bedrock", "max_input_tokens": 200000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "supports_function_calling": true, "supports_pdf_input": true, "supports_response_schema": true, @@ -9585,15 +8738,15 @@ "supports_vision": true }, "eu.anthropic.claude-opus-4-1-20250805-v1:0": { - "cache_creation_input_token_cost": 0.00001875, - "cache_read_input_token_cost": 0.0000015, - "input_cost_per_token": 0.000015, + "cache_creation_input_token_cost": 1.875e-5, + "cache_read_input_token_cost": 1.5e-6, + "input_cost_per_token": 1.5e-5, "litellm_provider": "bedrock_converse", "max_input_tokens": 200000, "max_output_tokens": 32000, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 0.000075, + "output_cost_per_token": 7.5e-5, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -9611,15 +8764,15 @@ "tool_use_system_prompt_tokens": 159 }, "eu.anthropic.claude-opus-4-20250514-v1:0": { - "cache_creation_input_token_cost": 0.00001875, - "cache_read_input_token_cost": 0.0000015, - "input_cost_per_token": 0.000015, + "cache_creation_input_token_cost": 1.875e-5, + "cache_read_input_token_cost": 1.5e-6, + "input_cost_per_token": 1.5e-5, "litellm_provider": "bedrock_converse", "max_input_tokens": 200000, "max_output_tokens": 32000, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 0.000075, + "output_cost_per_token": 7.5e-5, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -9637,19 +8790,19 @@ "tool_use_system_prompt_tokens": 159 }, "eu.anthropic.claude-sonnet-4-20250514-v1:0": { - "cache_creation_input_token_cost": 0.00000375, + "cache_creation_input_token_cost": 3.75e-6, "cache_read_input_token_cost": 3e-7, - "input_cost_per_token": 0.000003, - "input_cost_per_token_above_200k_tokens": 0.000006, - "output_cost_per_token_above_200k_tokens": 0.0000225, - "cache_creation_input_token_cost_above_200k_tokens": 0.0000075, + "input_cost_per_token": 3e-6, + "input_cost_per_token_above_200k_tokens": 6e-6, + "output_cost_per_token_above_200k_tokens": 2.25e-5, + "cache_creation_input_token_cost_above_200k_tokens": 7.5e-6, "cache_read_input_token_cost_above_200k_tokens": 6e-7, "litellm_provider": "bedrock_converse", "max_input_tokens": 1000000, "max_output_tokens": 64000, "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -9667,19 +8820,19 @@ "tool_use_system_prompt_tokens": 159 }, "eu.anthropic.claude-sonnet-4-5-20250929-v1:0": { - "cache_creation_input_token_cost": 0.000004125, + "cache_creation_input_token_cost": 4.125e-6, "cache_read_input_token_cost": 3.3e-7, - "input_cost_per_token": 0.0000033, - "input_cost_per_token_above_200k_tokens": 0.0000066, - "output_cost_per_token_above_200k_tokens": 0.00002475, - "cache_creation_input_token_cost_above_200k_tokens": 0.00000825, + "input_cost_per_token": 3.3e-6, + "input_cost_per_token_above_200k_tokens": 6.6e-6, + "output_cost_per_token_above_200k_tokens": 2.475e-5, + "cache_creation_input_token_cost_above_200k_tokens": 8.25e-6, "cache_read_input_token_cost_above_200k_tokens": 6.6e-7, "litellm_provider": "bedrock_converse", "max_input_tokens": 200000, "max_output_tokens": 64000, "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 0.0000165, + "output_cost_per_token": 1.65e-5, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -9719,13 +8872,13 @@ "supports_tool_choice": false }, "eu.mistral.pixtral-large-2502-v1:0": { - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "bedrock_converse", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.000006, + "output_cost_per_token": 6e-6, "supports_function_calling": true, "supports_tool_choice": false }, @@ -9733,97 +8886,73 @@ "litellm_provider": "fal_ai", "mode": "image_generation", "output_cost_per_image": 0.0398, - "supported_endpoints": [ - "/v1/images/generations" - ] + "supported_endpoints": ["/v1/images/generations"] }, "fal_ai/fal-ai/flux-pro/v1.1": { "litellm_provider": "fal_ai", "mode": "image_generation", "output_cost_per_image": 0.04, - "supported_endpoints": [ - "/v1/images/generations" - ] + "supported_endpoints": ["/v1/images/generations"] }, "fal_ai/fal-ai/flux-pro/v1.1-ultra": { "litellm_provider": "fal_ai", "mode": "image_generation", "output_cost_per_image": 0.06, - "supported_endpoints": [ - "/v1/images/generations" - ] + "supported_endpoints": ["/v1/images/generations"] }, "fal_ai/fal-ai/flux/schnell": { "litellm_provider": "fal_ai", "mode": "image_generation", "output_cost_per_image": 0.003, - "supported_endpoints": [ - "/v1/images/generations" - ] + "supported_endpoints": ["/v1/images/generations"] }, "fal_ai/fal-ai/bytedance/seedream/v3/text-to-image": { "litellm_provider": "fal_ai", "mode": "image_generation", "output_cost_per_image": 0.03, - "supported_endpoints": [ - "/v1/images/generations" - ] + "supported_endpoints": ["/v1/images/generations"] }, "fal_ai/fal-ai/bytedance/dreamina/v3.1/text-to-image": { "litellm_provider": "fal_ai", "mode": "image_generation", "output_cost_per_image": 0.03, - "supported_endpoints": [ - "/v1/images/generations" - ] + "supported_endpoints": ["/v1/images/generations"] }, "fal_ai/fal-ai/ideogram/v3": { "litellm_provider": "fal_ai", "mode": "image_generation", "output_cost_per_image": 0.06, - "supported_endpoints": [ - "/v1/images/generations" - ] + "supported_endpoints": ["/v1/images/generations"] }, "fal_ai/fal-ai/imagen4/preview": { "litellm_provider": "fal_ai", "mode": "image_generation", "output_cost_per_image": 0.0398, - "supported_endpoints": [ - "/v1/images/generations" - ] + "supported_endpoints": ["/v1/images/generations"] }, "fal_ai/fal-ai/imagen4/preview/fast": { "litellm_provider": "fal_ai", "mode": "image_generation", "output_cost_per_image": 0.02, - "supported_endpoints": [ - "/v1/images/generations" - ] + "supported_endpoints": ["/v1/images/generations"] }, "fal_ai/fal-ai/imagen4/preview/ultra": { "litellm_provider": "fal_ai", "mode": "image_generation", "output_cost_per_image": 0.06, - "supported_endpoints": [ - "/v1/images/generations" - ] + "supported_endpoints": ["/v1/images/generations"] }, "fal_ai/fal-ai/recraft/v3/text-to-image": { "litellm_provider": "fal_ai", "mode": "image_generation", "output_cost_per_image": 0.0398, - "supported_endpoints": [ - "/v1/images/generations" - ] + "supported_endpoints": ["/v1/images/generations"] }, "fal_ai/fal-ai/stable-diffusion-v35-medium": { "litellm_provider": "fal_ai", "mode": "image_generation", "output_cost_per_image": 0.0398, - "supported_endpoints": [ - "/v1/images/generations" - ] + "supported_endpoints": ["/v1/images/generations"] }, "featherless_ai/featherless-ai/Qwerky-72B": { "litellm_provider": "featherless_ai", @@ -9845,9 +8974,9 @@ "output_cost_per_token": 2e-7 }, "fireworks-ai-56b-to-176b": { - "input_cost_per_token": 0.0000012, + "input_cost_per_token": 1.2e-6, "litellm_provider": "fireworks_ai", - "output_cost_per_token": 0.0000012 + "output_cost_per_token": 1.2e-6 }, "fireworks-ai-above-16b": { "input_cost_per_token": 9e-7, @@ -9855,19 +8984,19 @@ "output_cost_per_token": 9e-7 }, "fireworks-ai-default": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "fireworks_ai", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "fireworks-ai-embedding-150m-to-350m": { "input_cost_per_token": 1.6e-8, "litellm_provider": "fireworks_ai-embedding-models", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "fireworks-ai-embedding-up-to-150m": { "input_cost_per_token": 8e-9, "litellm_provider": "fireworks_ai-embedding-models", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "fireworks-ai-moe-up-to-56b": { "input_cost_per_token": 5e-7, @@ -9885,42 +9014,42 @@ "max_input_tokens": 512, "max_tokens": 512, "mode": "embedding", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "source": "https://fireworks.ai/pricing" }, "fireworks_ai/accounts/fireworks/models/deepseek-coder-v2-instruct": { - "input_cost_per_token": 0.0000012, + "input_cost_per_token": 1.2e-6, "litellm_provider": "fireworks_ai", "max_input_tokens": 65536, "max_output_tokens": 65536, "max_tokens": 65536, "mode": "chat", - "output_cost_per_token": 0.0000012, + "output_cost_per_token": 1.2e-6, "source": "https://fireworks.ai/pricing", "supports_function_calling": false, "supports_response_schema": true, "supports_tool_choice": false }, "fireworks_ai/accounts/fireworks/models/deepseek-r1": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "fireworks_ai", "max_input_tokens": 128000, "max_output_tokens": 20480, "max_tokens": 20480, "mode": "chat", - "output_cost_per_token": 0.000008, + "output_cost_per_token": 8e-6, "source": "https://fireworks.ai/pricing", "supports_response_schema": true, "supports_tool_choice": false }, "fireworks_ai/accounts/fireworks/models/deepseek-r1-0528": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "fireworks_ai", "max_input_tokens": 160000, "max_output_tokens": 160000, "max_tokens": 160000, "mode": "chat", - "output_cost_per_token": 0.000008, + "output_cost_per_token": 8e-6, "source": "https://fireworks.ai/pricing", "supports_response_schema": true, "supports_tool_choice": false @@ -9932,7 +9061,7 @@ "max_output_tokens": 20480, "max_tokens": 20480, "mode": "chat", - "output_cost_per_token": 0.00000219, + "output_cost_per_token": 2.19e-6, "source": "https://fireworks.ai/pricing", "supports_response_schema": true, "supports_tool_choice": false @@ -9968,7 +9097,7 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.00000168, + "output_cost_per_token": 1.68e-6, "source": "https://fireworks.ai/pricing", "supports_response_schema": true, "supports_tool_choice": true @@ -9980,7 +9109,7 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.00000168, + "output_cost_per_token": 1.68e-6, "source": "https://fireworks.ai/pricing", "supports_response_schema": true, "supports_tool_choice": true @@ -10005,7 +9134,7 @@ "max_output_tokens": 96000, "max_tokens": 96000, "mode": "chat", - "output_cost_per_token": 0.00000219, + "output_cost_per_token": 2.19e-6, "source": "https://fireworks.ai/models/fireworks/glm-4p5", "supports_function_calling": true, "supports_response_schema": true, @@ -10057,7 +9186,7 @@ "max_output_tokens": 16384, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0.0000025, + "output_cost_per_token": 2.5e-6, "source": "https://fireworks.ai/models/fireworks/kimi-k2-instruct", "supports_function_calling": true, "supports_response_schema": true, @@ -10070,7 +9199,7 @@ "max_output_tokens": 262144, "max_tokens": 262144, "mode": "chat", - "output_cost_per_token": 0.0000025, + "output_cost_per_token": 2.5e-6, "source": "https://fireworks.ai/pricing", "supports_function_calling": true, "supports_response_schema": true, @@ -10078,13 +9207,13 @@ "supports_web_search": true }, "fireworks_ai/accounts/fireworks/models/llama-v3p1-405b-instruct": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "fireworks_ai", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 0.000003, + "output_cost_per_token": 3e-6, "source": "https://fireworks.ai/pricing", "supports_function_calling": true, "supports_response_schema": true, @@ -10181,13 +9310,13 @@ "supports_tool_choice": false }, "fireworks_ai/accounts/fireworks/models/mixtral-8x22b-instruct-hf": { - "input_cost_per_token": 0.0000012, + "input_cost_per_token": 1.2e-6, "litellm_provider": "fireworks_ai", "max_input_tokens": 65536, "max_output_tokens": 65536, "max_tokens": 65536, "mode": "chat", - "output_cost_per_token": 0.0000012, + "output_cost_per_token": 1.2e-6, "source": "https://fireworks.ai/pricing", "supports_function_calling": true, "supports_response_schema": true, @@ -10220,13 +9349,13 @@ "supports_tool_choice": false }, "fireworks_ai/accounts/fireworks/models/yi-large": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "fireworks_ai", "max_input_tokens": 32768, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0.000003, + "output_cost_per_token": 3e-6, "source": "https://fireworks.ai/pricing", "supports_function_calling": false, "supports_response_schema": true, @@ -10238,7 +9367,7 @@ "max_input_tokens": 8192, "max_tokens": 8192, "mode": "embedding", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "source": "https://fireworks.ai/pricing" }, "fireworks_ai/nomic-ai/nomic-embed-text-v1.5": { @@ -10247,7 +9376,7 @@ "max_input_tokens": 8192, "max_tokens": 8192, "mode": "embedding", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "source": "https://fireworks.ai/pricing" }, "fireworks_ai/thenlper/gte-base": { @@ -10256,7 +9385,7 @@ "max_input_tokens": 512, "max_tokens": 512, "mode": "embedding", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "source": "https://fireworks.ai/pricing" }, "fireworks_ai/thenlper/gte-large": { @@ -10265,7 +9394,7 @@ "max_input_tokens": 512, "max_tokens": 512, "mode": "embedding", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "source": "https://fireworks.ai/pricing" }, "friendliai/meta-llama-3.1-70b-instruct": { @@ -10308,85 +9437,85 @@ "output_cost_per_token_batches": 2e-7 }, "ft:davinci-002": { - "input_cost_per_token": 0.000002, - "input_cost_per_token_batches": 0.000001, + "input_cost_per_token": 2e-6, + "input_cost_per_token_batches": 1e-6, "litellm_provider": "text-completion-openai", "max_input_tokens": 16384, "max_output_tokens": 4096, "max_tokens": 16384, "mode": "completion", - "output_cost_per_token": 0.000002, - "output_cost_per_token_batches": 0.000001 + "output_cost_per_token": 2e-6, + "output_cost_per_token_batches": 1e-6 }, "ft:gpt-3.5-turbo": { - "input_cost_per_token": 0.000003, - "input_cost_per_token_batches": 0.0000015, + "input_cost_per_token": 3e-6, + "input_cost_per_token_batches": 1.5e-6, "litellm_provider": "openai", "max_input_tokens": 16385, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.000006, - "output_cost_per_token_batches": 0.000003, + "output_cost_per_token": 6e-6, + "output_cost_per_token_batches": 3e-6, "supports_system_messages": true, "supports_tool_choice": true }, "ft:gpt-3.5-turbo-0125": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "openai", "max_input_tokens": 16385, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.000006, + "output_cost_per_token": 6e-6, "supports_system_messages": true, "supports_tool_choice": true }, "ft:gpt-3.5-turbo-0613": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "openai", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.000006, + "output_cost_per_token": 6e-6, "supports_system_messages": true, "supports_tool_choice": true }, "ft:gpt-3.5-turbo-1106": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "openai", "max_input_tokens": 16385, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.000006, + "output_cost_per_token": 6e-6, "supports_system_messages": true, "supports_tool_choice": true }, "ft:gpt-4-0613": { - "input_cost_per_token": 0.00003, + "input_cost_per_token": 3e-5, "litellm_provider": "openai", "max_input_tokens": 8192, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.00006, + "output_cost_per_token": 6e-5, "source": "OpenAI needs to add pricing for this ft model, will be updated when added by OpenAI. Defaulting to base model pricing", "supports_function_calling": true, "supports_system_messages": true, "supports_tool_choice": true }, "ft:gpt-4o-2024-08-06": { - "input_cost_per_token": 0.00000375, - "input_cost_per_token_batches": 0.000001875, + "input_cost_per_token": 3.75e-6, + "input_cost_per_token_batches": 1.875e-6, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 0.000015, - "output_cost_per_token_batches": 0.0000075, + "output_cost_per_token": 1.5e-5, + "output_cost_per_token_batches": 7.5e-6, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_pdf_input": true, @@ -10396,14 +9525,14 @@ "supports_vision": true }, "ft:gpt-4o-2024-11-20": { - "cache_creation_input_token_cost": 0.000001875, - "input_cost_per_token": 0.00000375, + "cache_creation_input_token_cost": 1.875e-6, + "input_cost_per_token": 3.75e-6, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_pdf_input": true, @@ -10422,7 +9551,7 @@ "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 0.0000012, + "output_cost_per_token": 1.2e-6, "output_cost_per_token_batches": 6e-7, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -10444,7 +9573,7 @@ "max_tokens": 8192, "mode": "chat", "output_cost_per_character": 3.75e-7, - "output_cost_per_token": 0.0000015, + "output_cost_per_token": 1.5e-6, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing#google_models", "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -10462,7 +9591,7 @@ "max_tokens": 8192, "mode": "chat", "output_cost_per_character": 3.75e-7, - "output_cost_per_token": 0.0000015, + "output_cost_per_token": 1.5e-6, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models", "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -10480,7 +9609,7 @@ "max_tokens": 8192, "mode": "chat", "output_cost_per_character": 3.75e-7, - "output_cost_per_token": 0.0000015, + "output_cost_per_token": 1.5e-6, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models", "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -10497,7 +9626,7 @@ "max_video_length": 2, "max_videos_per_prompt": 1, "mode": "chat", - "output_cost_per_token": 0.0000015, + "output_cost_per_token": 1.5e-6, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models", "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -10516,7 +9645,7 @@ "max_video_length": 2, "max_videos_per_prompt": 1, "mode": "chat", - "output_cost_per_token": 0.0000015, + "output_cost_per_token": 1.5e-6, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models", "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -10534,7 +9663,7 @@ "max_tokens": 8192, "mode": "chat", "output_cost_per_character": 3.75e-7, - "output_cost_per_token": 0.0000015, + "output_cost_per_token": 1.5e-6, "source": "As of Jun, 2024. There is no available doc on vertex ai pricing gemini-1.0-ultra-001. Using gemini-1.0-pro pricing. Got max_tokens info here: https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models", "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -10551,23 +9680,23 @@ "max_tokens": 8192, "mode": "chat", "output_cost_per_character": 3.75e-7, - "output_cost_per_token": 0.0000015, + "output_cost_per_token": 1.5e-6, "source": "As of Jun, 2024. There is no available doc on vertex ai pricing gemini-1.0-ultra-001. Using gemini-1.0-pro pricing. Got max_tokens info here: https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models", "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_tool_choice": true }, "gemini-1.5-flash": { - "input_cost_per_audio_per_second": 0.000002, - "input_cost_per_audio_per_second_above_128k_tokens": 0.000004, + "input_cost_per_audio_per_second": 2e-6, + "input_cost_per_audio_per_second_above_128k_tokens": 4e-6, "input_cost_per_character": 1.875e-8, "input_cost_per_character_above_128k_tokens": 2.5e-7, - "input_cost_per_image": 0.00002, - "input_cost_per_image_above_128k_tokens": 0.00004, + "input_cost_per_image": 2e-5, + "input_cost_per_image_above_128k_tokens": 4e-5, "input_cost_per_token": 7.5e-8, - "input_cost_per_token_above_128k_tokens": 0.000001, - "input_cost_per_video_per_second": 0.00002, - "input_cost_per_video_per_second_above_128k_tokens": 0.00004, + "input_cost_per_token_above_128k_tokens": 1e-6, + "input_cost_per_video_per_second": 2e-5, + "input_cost_per_video_per_second_above_128k_tokens": 4e-5, "litellm_provider": "vertex_ai-language-models", "max_audio_length_hours": 8.4, "max_audio_per_prompt": 1, @@ -10593,16 +9722,16 @@ }, "gemini-1.5-flash-001": { "deprecation_date": "2025-05-24", - "input_cost_per_audio_per_second": 0.000002, - "input_cost_per_audio_per_second_above_128k_tokens": 0.000004, + "input_cost_per_audio_per_second": 2e-6, + "input_cost_per_audio_per_second_above_128k_tokens": 4e-6, "input_cost_per_character": 1.875e-8, "input_cost_per_character_above_128k_tokens": 2.5e-7, - "input_cost_per_image": 0.00002, - "input_cost_per_image_above_128k_tokens": 0.00004, + "input_cost_per_image": 2e-5, + "input_cost_per_image_above_128k_tokens": 4e-5, "input_cost_per_token": 7.5e-8, - "input_cost_per_token_above_128k_tokens": 0.000001, - "input_cost_per_video_per_second": 0.00002, - "input_cost_per_video_per_second_above_128k_tokens": 0.00004, + "input_cost_per_token_above_128k_tokens": 1e-6, + "input_cost_per_video_per_second": 2e-5, + "input_cost_per_video_per_second_above_128k_tokens": 4e-5, "litellm_provider": "vertex_ai-language-models", "max_audio_length_hours": 8.4, "max_audio_per_prompt": 1, @@ -10628,16 +9757,16 @@ }, "gemini-1.5-flash-002": { "deprecation_date": "2025-09-24", - "input_cost_per_audio_per_second": 0.000002, - "input_cost_per_audio_per_second_above_128k_tokens": 0.000004, + "input_cost_per_audio_per_second": 2e-6, + "input_cost_per_audio_per_second_above_128k_tokens": 4e-6, "input_cost_per_character": 1.875e-8, "input_cost_per_character_above_128k_tokens": 2.5e-7, - "input_cost_per_image": 0.00002, - "input_cost_per_image_above_128k_tokens": 0.00004, + "input_cost_per_image": 2e-5, + "input_cost_per_image_above_128k_tokens": 4e-5, "input_cost_per_token": 7.5e-8, - "input_cost_per_token_above_128k_tokens": 0.000001, - "input_cost_per_video_per_second": 0.00002, - "input_cost_per_video_per_second_above_128k_tokens": 0.00004, + "input_cost_per_token_above_128k_tokens": 1e-6, + "input_cost_per_video_per_second": 2e-5, + "input_cost_per_video_per_second_above_128k_tokens": 4e-5, "litellm_provider": "vertex_ai-language-models", "max_audio_length_hours": 8.4, "max_audio_per_prompt": 1, @@ -10662,16 +9791,16 @@ "supports_vision": true }, "gemini-1.5-flash-exp-0827": { - "input_cost_per_audio_per_second": 0.000002, - "input_cost_per_audio_per_second_above_128k_tokens": 0.000004, + "input_cost_per_audio_per_second": 2e-6, + "input_cost_per_audio_per_second_above_128k_tokens": 4e-6, "input_cost_per_character": 1.875e-8, "input_cost_per_character_above_128k_tokens": 2.5e-7, - "input_cost_per_image": 0.00002, - "input_cost_per_image_above_128k_tokens": 0.00004, + "input_cost_per_image": 2e-5, + "input_cost_per_image_above_128k_tokens": 4e-5, "input_cost_per_token": 4.688e-9, - "input_cost_per_token_above_128k_tokens": 0.000001, - "input_cost_per_video_per_second": 0.00002, - "input_cost_per_video_per_second_above_128k_tokens": 0.00004, + "input_cost_per_token_above_128k_tokens": 1e-6, + "input_cost_per_video_per_second": 2e-5, + "input_cost_per_video_per_second_above_128k_tokens": 4e-5, "litellm_provider": "vertex_ai-language-models", "max_audio_length_hours": 8.4, "max_audio_per_prompt": 1, @@ -10696,16 +9825,16 @@ "supports_vision": true }, "gemini-1.5-flash-preview-0514": { - "input_cost_per_audio_per_second": 0.000002, - "input_cost_per_audio_per_second_above_128k_tokens": 0.000004, + "input_cost_per_audio_per_second": 2e-6, + "input_cost_per_audio_per_second_above_128k_tokens": 4e-6, "input_cost_per_character": 1.875e-8, "input_cost_per_character_above_128k_tokens": 2.5e-7, - "input_cost_per_image": 0.00002, - "input_cost_per_image_above_128k_tokens": 0.00004, + "input_cost_per_image": 2e-5, + "input_cost_per_image_above_128k_tokens": 4e-5, "input_cost_per_token": 7.5e-8, - "input_cost_per_token_above_128k_tokens": 0.000001, - "input_cost_per_video_per_second": 0.00002, - "input_cost_per_video_per_second_above_128k_tokens": 0.00004, + "input_cost_per_token_above_128k_tokens": 1e-6, + "input_cost_per_video_per_second": 2e-5, + "input_cost_per_video_per_second_above_128k_tokens": 4e-5, "litellm_provider": "vertex_ai-language-models", "max_audio_length_hours": 8.4, "max_audio_per_prompt": 1, @@ -10729,14 +9858,14 @@ "supports_vision": true }, "gemini-1.5-pro": { - "input_cost_per_audio_per_second": 0.00003125, - "input_cost_per_audio_per_second_above_128k_tokens": 0.0000625, + "input_cost_per_audio_per_second": 3.125e-5, + "input_cost_per_audio_per_second_above_128k_tokens": 6.25e-5, "input_cost_per_character": 3.125e-7, "input_cost_per_character_above_128k_tokens": 6.25e-7, "input_cost_per_image": 0.00032875, "input_cost_per_image_above_128k_tokens": 0.0006575, - "input_cost_per_token": 0.00000125, - "input_cost_per_token_above_128k_tokens": 0.0000025, + "input_cost_per_token": 1.25e-6, + "input_cost_per_token_above_128k_tokens": 2.5e-6, "input_cost_per_video_per_second": 0.00032875, "input_cost_per_video_per_second_above_128k_tokens": 0.0006575, "litellm_provider": "vertex_ai-language-models", @@ -10744,10 +9873,10 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_character": 0.00000125, - "output_cost_per_character_above_128k_tokens": 0.0000025, - "output_cost_per_token": 0.000005, - "output_cost_per_token_above_128k_tokens": 0.00001, + "output_cost_per_character": 1.25e-6, + "output_cost_per_character_above_128k_tokens": 2.5e-6, + "output_cost_per_token": 5e-6, + "output_cost_per_token_above_128k_tokens": 1e-5, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models", "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -10759,14 +9888,14 @@ }, "gemini-1.5-pro-001": { "deprecation_date": "2025-05-24", - "input_cost_per_audio_per_second": 0.00003125, - "input_cost_per_audio_per_second_above_128k_tokens": 0.0000625, + "input_cost_per_audio_per_second": 3.125e-5, + "input_cost_per_audio_per_second_above_128k_tokens": 6.25e-5, "input_cost_per_character": 3.125e-7, "input_cost_per_character_above_128k_tokens": 6.25e-7, "input_cost_per_image": 0.00032875, "input_cost_per_image_above_128k_tokens": 0.0006575, - "input_cost_per_token": 0.00000125, - "input_cost_per_token_above_128k_tokens": 0.0000025, + "input_cost_per_token": 1.25e-6, + "input_cost_per_token_above_128k_tokens": 2.5e-6, "input_cost_per_video_per_second": 0.00032875, "input_cost_per_video_per_second_above_128k_tokens": 0.0006575, "litellm_provider": "vertex_ai-language-models", @@ -10774,10 +9903,10 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_character": 0.00000125, - "output_cost_per_character_above_128k_tokens": 0.0000025, - "output_cost_per_token": 0.000005, - "output_cost_per_token_above_128k_tokens": 0.00001, + "output_cost_per_character": 1.25e-6, + "output_cost_per_character_above_128k_tokens": 2.5e-6, + "output_cost_per_token": 5e-6, + "output_cost_per_token_above_128k_tokens": 1e-5, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models", "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -10788,14 +9917,14 @@ }, "gemini-1.5-pro-002": { "deprecation_date": "2025-09-24", - "input_cost_per_audio_per_second": 0.00003125, - "input_cost_per_audio_per_second_above_128k_tokens": 0.0000625, + "input_cost_per_audio_per_second": 3.125e-5, + "input_cost_per_audio_per_second_above_128k_tokens": 6.25e-5, "input_cost_per_character": 3.125e-7, "input_cost_per_character_above_128k_tokens": 6.25e-7, "input_cost_per_image": 0.00032875, "input_cost_per_image_above_128k_tokens": 0.0006575, - "input_cost_per_token": 0.00000125, - "input_cost_per_token_above_128k_tokens": 0.0000025, + "input_cost_per_token": 1.25e-6, + "input_cost_per_token_above_128k_tokens": 2.5e-6, "input_cost_per_video_per_second": 0.00032875, "input_cost_per_video_per_second_above_128k_tokens": 0.0006575, "litellm_provider": "vertex_ai-language-models", @@ -10803,10 +9932,10 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_character": 0.00000125, - "output_cost_per_character_above_128k_tokens": 0.0000025, - "output_cost_per_token": 0.000005, - "output_cost_per_token_above_128k_tokens": 0.00001, + "output_cost_per_character": 1.25e-6, + "output_cost_per_character_above_128k_tokens": 2.5e-6, + "output_cost_per_token": 5e-6, + "output_cost_per_token_above_128k_tokens": 1e-5, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#gemini-1.5-pro", "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -10816,8 +9945,8 @@ "supports_vision": true }, "gemini-1.5-pro-preview-0215": { - "input_cost_per_audio_per_second": 0.00003125, - "input_cost_per_audio_per_second_above_128k_tokens": 0.0000625, + "input_cost_per_audio_per_second": 3.125e-5, + "input_cost_per_audio_per_second_above_128k_tokens": 6.25e-5, "input_cost_per_character": 3.125e-7, "input_cost_per_character_above_128k_tokens": 6.25e-7, "input_cost_per_image": 0.00032875, @@ -10831,8 +9960,8 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_character": 0.00000125, - "output_cost_per_character_above_128k_tokens": 0.0000025, + "output_cost_per_character": 1.25e-6, + "output_cost_per_character_above_128k_tokens": 2.5e-6, "output_cost_per_token": 3.125e-7, "output_cost_per_token_above_128k_tokens": 6.25e-7, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models", @@ -10843,8 +9972,8 @@ "supports_tool_choice": true }, "gemini-1.5-pro-preview-0409": { - "input_cost_per_audio_per_second": 0.00003125, - "input_cost_per_audio_per_second_above_128k_tokens": 0.0000625, + "input_cost_per_audio_per_second": 3.125e-5, + "input_cost_per_audio_per_second_above_128k_tokens": 6.25e-5, "input_cost_per_character": 3.125e-7, "input_cost_per_character_above_128k_tokens": 6.25e-7, "input_cost_per_image": 0.00032875, @@ -10858,8 +9987,8 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_character": 0.00000125, - "output_cost_per_character_above_128k_tokens": 0.0000025, + "output_cost_per_character": 1.25e-6, + "output_cost_per_character_above_128k_tokens": 2.5e-6, "output_cost_per_token": 3.125e-7, "output_cost_per_token_above_128k_tokens": 6.25e-7, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models", @@ -10869,8 +9998,8 @@ "supports_tool_choice": true }, "gemini-1.5-pro-preview-0514": { - "input_cost_per_audio_per_second": 0.00003125, - "input_cost_per_audio_per_second_above_128k_tokens": 0.0000625, + "input_cost_per_audio_per_second": 3.125e-5, + "input_cost_per_audio_per_second_above_128k_tokens": 6.25e-5, "input_cost_per_character": 3.125e-7, "input_cost_per_character_above_128k_tokens": 6.25e-7, "input_cost_per_image": 0.00032875, @@ -10884,8 +10013,8 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_character": 0.00000125, - "output_cost_per_character_above_128k_tokens": 0.0000025, + "output_cost_per_character": 1.25e-6, + "output_cost_per_character_above_128k_tokens": 2.5e-6, "output_cost_per_token": 3.125e-7, "output_cost_per_token_above_128k_tokens": 6.25e-7, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models", @@ -10912,16 +10041,8 @@ "mode": "chat", "output_cost_per_token": 4e-7, "source": "https://ai.google.dev/pricing#2_0flash", - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text", - "image" - ], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text", "image"], "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -10937,7 +10058,7 @@ "gemini-2.0-flash-001": { "cache_read_input_token_cost": 3.75e-8, "deprecation_date": "2026-02-05", - "input_cost_per_audio_token": 0.000001, + "input_cost_per_audio_token": 1e-6, "input_cost_per_token": 1.5e-7, "litellm_provider": "vertex_ai-language-models", "max_audio_length_hours": 8.4, @@ -10952,16 +10073,8 @@ "mode": "chat", "output_cost_per_token": 6e-7, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing", - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text", - "image" - ], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text", "image"], "supports_audio_output": true, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -11000,16 +10113,8 @@ "output_cost_per_token": 6e-7, "output_cost_per_token_above_128k_tokens": 0, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing", - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text", - "image" - ], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text", "image"], "supports_audio_output": true, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -11036,15 +10141,8 @@ "mode": "chat", "output_cost_per_token": 3e-7, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#gemini-2.0-flash", - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text" - ], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text"], "supports_audio_output": true, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -11072,15 +10170,8 @@ "mode": "chat", "output_cost_per_token": 3e-7, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#gemini-2.0-flash", - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text" - ], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text"], "supports_audio_output": true, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -11093,10 +10184,10 @@ }, "gemini-2.0-flash-live-preview-04-09": { "cache_read_input_token_cost": 7.5e-8, - "input_cost_per_audio_token": 0.000003, - "input_cost_per_image": 0.000003, + "input_cost_per_audio_token": 3e-6, + "input_cost_per_image": 3e-6, "input_cost_per_token": 5e-7, - "input_cost_per_video_per_second": 0.000003, + "input_cost_per_video_per_second": 3e-6, "litellm_provider": "vertex_ai-language-models", "max_audio_length_hours": 8.4, "max_audio_per_prompt": 1, @@ -11108,24 +10199,13 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_audio_token": 0.000012, - "output_cost_per_token": 0.000002, + "output_cost_per_audio_token": 1.2e-5, + "output_cost_per_token": 2e-6, "rpm": 10, "source": "https://cloud.google.com/vertex-ai/docs/generative-ai/model-reference/gemini#gemini-2-0-flash-live-preview-04-09", - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/completions" - ], - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text", - "audio" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/completions"], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text", "audio"], "supports_audio_output": true, "supports_function_calling": true, "supports_pdf_input": true, @@ -11156,16 +10236,8 @@ "mode": "chat", "output_cost_per_token": 4e-7, "source": "https://ai.google.dev/pricing#2_0flash", - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text", - "image" - ], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text", "image"], "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -11178,7 +10250,7 @@ "supports_web_search": true }, "gemini-2.0-flash-thinking-exp": { - "cache_read_input_token_cost": 0, + "cache_read_input_token_cost": 0.0, "input_cost_per_audio_per_second": 0, "input_cost_per_audio_per_second_above_128k_tokens": 0, "input_cost_per_character": 0, @@ -11205,16 +10277,8 @@ "output_cost_per_token": 0, "output_cost_per_token_above_128k_tokens": 0, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#gemini-2.0-flash", - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text", - "image" - ], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text", "image"], "supports_audio_output": true, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -11226,7 +10290,7 @@ "supports_web_search": true }, "gemini-2.0-flash-thinking-exp-01-21": { - "cache_read_input_token_cost": 0, + "cache_read_input_token_cost": 0.0, "input_cost_per_audio_per_second": 0, "input_cost_per_audio_per_second_above_128k_tokens": 0, "input_cost_per_character": 0, @@ -11253,16 +10317,8 @@ "output_cost_per_token": 0, "output_cost_per_token_above_128k_tokens": 0, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#gemini-2.0-flash", - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text", - "image" - ], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text", "image"], "supports_audio_output": false, "supports_function_calling": false, "supports_parallel_function_calling": true, @@ -11276,8 +10332,8 @@ }, "gemini-2.0-pro-exp-02-05": { "cache_read_input_token_cost": 3.125e-7, - "input_cost_per_token": 0.00000125, - "input_cost_per_token_above_200k_tokens": 0.0000025, + "input_cost_per_token": 1.25e-6, + "input_cost_per_token_above_200k_tokens": 2.5e-6, "litellm_provider": "vertex_ai-language-models", "max_audio_length_hours": 8.4, "max_audio_per_prompt": 1, @@ -11289,22 +10345,12 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_token": 0.00001, - "output_cost_per_token_above_200k_tokens": 0.000015, + "output_cost_per_token": 1e-5, + "output_cost_per_token_above_200k_tokens": 1.5e-5, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing", - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/completions" - ], - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/completions"], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text"], "supports_audio_input": true, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -11319,7 +10365,7 @@ }, "gemini-2.5-flash": { "cache_read_input_token_cost": 3e-8, - "input_cost_per_audio_token": 0.000001, + "input_cost_per_audio_token": 1e-6, "input_cost_per_token": 3e-7, "litellm_provider": "vertex_ai-language-models", "max_audio_length_hours": 8.4, @@ -11332,23 +10378,12 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_reasoning_token": 0.0000025, - "output_cost_per_token": 0.0000025, + "output_cost_per_reasoning_token": 2.5e-6, + "output_cost_per_token": 2.5e-6, "source": "https://ai.google.dev/gemini-api/docs/models#gemini-2.5-flash-preview", - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/completions", - "/v1/batch" - ], - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/completions", "/v1/batch"], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text"], "supports_audio_output": false, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -11364,7 +10399,7 @@ }, "gemini-2.5-flash-image": { "cache_read_input_token_cost": 3e-8, - "input_cost_per_audio_token": 0.000001, + "input_cost_per_audio_token": 1e-6, "input_cost_per_token": 3e-7, "litellm_provider": "vertex_ai-language-models", "max_audio_length_hours": 8.4, @@ -11378,25 +10413,13 @@ "max_videos_per_prompt": 10, "mode": "image_generation", "output_cost_per_image": 0.039, - "output_cost_per_reasoning_token": 0.0000025, - "output_cost_per_token": 0.0000025, + "output_cost_per_reasoning_token": 2.5e-6, + "output_cost_per_token": 2.5e-6, "rpm": 100000, "source": "https://ai.google.dev/gemini-api/docs/pricing#gemini-2.5-flash-image", - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/completions", - "/v1/batch" - ], - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text", - "image" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/completions", "/v1/batch"], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text", "image"], "supports_audio_output": false, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -11412,7 +10435,7 @@ }, "gemini-2.5-flash-image-preview": { "cache_read_input_token_cost": 7.5e-8, - "input_cost_per_audio_token": 0.000001, + "input_cost_per_audio_token": 1e-6, "input_cost_per_token": 3e-7, "litellm_provider": "vertex_ai-language-models", "max_audio_length_hours": 8.4, @@ -11426,25 +10449,13 @@ "max_videos_per_prompt": 10, "mode": "image_generation", "output_cost_per_image": 0.039, - "output_cost_per_reasoning_token": 0.00003, - "output_cost_per_token": 0.00003, + "output_cost_per_reasoning_token": 3e-5, + "output_cost_per_token": 3e-5, "rpm": 100000, "source": "https://ai.google.dev/gemini-api/docs/models#gemini-2.5-flash-preview", - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/completions", - "/v1/batch" - ], - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text", - "image" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/completions", "/v1/batch"], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text", "image"], "supports_audio_output": false, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -11460,30 +10471,20 @@ }, "gemini-3-pro-image-preview": { "input_cost_per_image": 0.0011, - "input_cost_per_token": 0.000002, - "input_cost_per_token_batches": 0.000001, + "input_cost_per_token": 2e-6, + "input_cost_per_token_batches": 1e-6, "litellm_provider": "vertex_ai-language-models", "max_input_tokens": 65536, "max_output_tokens": 32768, "max_tokens": 65536, "mode": "image_generation", "output_cost_per_image": 0.134, - "output_cost_per_token": 0.000012, - "output_cost_per_token_batches": 0.000006, + "output_cost_per_token": 1.2e-5, + "output_cost_per_token_batches": 6e-6, "source": "https://ai.google.dev/gemini-api/docs/pricing", - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/completions", - "/v1/batch" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text", - "image" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/completions", "/v1/batch"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text", "image"], "supports_function_calling": false, "supports_prompt_caching": true, "supports_response_schema": true, @@ -11509,20 +10510,9 @@ "output_cost_per_reasoning_token": 4e-7, "output_cost_per_token": 4e-7, "source": "https://ai.google.dev/gemini-api/docs/models#gemini-2.5-flash-preview", - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/completions", - "/v1/batch" - ], - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/completions", "/v1/batch"], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text"], "supports_audio_output": false, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -11553,21 +10543,10 @@ "mode": "chat", "output_cost_per_reasoning_token": 4e-7, "output_cost_per_token": 4e-7, - "source": "https://developers.googleblog.com/en/continuing-to-bring-you-our-latest-models-with-an-improved-gemini-2-5-flash-and-flash-lite-release/", - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/completions", - "/v1/batch" - ], - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text" - ], + "source": "https://developers.googleblog.com/en/continuing-to-bring-you-our-latest-models-with-an-improved-gemini-2-5-flash-and-flash-lite-release/", + "supported_endpoints": ["/v1/chat/completions", "/v1/completions", "/v1/batch"], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text"], "supports_audio_output": false, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -11583,7 +10562,7 @@ }, "gemini-2.5-flash-preview-09-2025": { "cache_read_input_token_cost": 7.5e-8, - "input_cost_per_audio_token": 0.000001, + "input_cost_per_audio_token": 1e-6, "input_cost_per_token": 3e-7, "litellm_provider": "vertex_ai-language-models", "max_audio_length_hours": 8.4, @@ -11596,23 +10575,12 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_reasoning_token": 0.0000025, - "output_cost_per_token": 0.0000025, + "output_cost_per_reasoning_token": 2.5e-6, + "output_cost_per_token": 2.5e-6, "source": "https://developers.googleblog.com/en/continuing-to-bring-you-our-latest-models-with-an-improved-gemini-2-5-flash-and-flash-lite-release/", - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/completions", - "/v1/batch" - ], - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/completions", "/v1/batch"], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text"], "supports_audio_output": false, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -11628,7 +10596,7 @@ }, "gemini-live-2.5-flash-preview-native-audio-09-2025": { "cache_read_input_token_cost": 7.5e-8, - "input_cost_per_audio_token": 0.000003, + "input_cost_per_audio_token": 3e-6, "input_cost_per_token": 3e-7, "litellm_provider": "vertex_ai-language-models", "max_audio_length_hours": 8.4, @@ -11641,23 +10609,12 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_audio_token": 0.000012, - "output_cost_per_token": 0.000002, + "output_cost_per_audio_token": 1.2e-5, + "output_cost_per_token": 2e-6, "source": "https://ai.google.dev/gemini-api/docs/pricing", - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/completions" - ], - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text", - "audio" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/completions"], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text", "audio"], "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -11673,7 +10630,7 @@ }, "gemini/gemini-live-2.5-flash-preview-native-audio-09-2025": { "cache_read_input_token_cost": 7.5e-8, - "input_cost_per_audio_token": 0.000003, + "input_cost_per_audio_token": 3e-6, "input_cost_per_token": 3e-7, "litellm_provider": "gemini", "max_audio_length_hours": 8.4, @@ -11686,24 +10643,13 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_audio_token": 0.000012, - "output_cost_per_token": 0.000002, + "output_cost_per_audio_token": 1.2e-5, + "output_cost_per_token": 2e-6, "rpm": 100000, "source": "https://ai.google.dev/gemini-api/docs/pricing", - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/completions" - ], - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text", - "audio" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/completions"], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text", "audio"], "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -11736,20 +10682,9 @@ "output_cost_per_reasoning_token": 4e-7, "output_cost_per_token": 4e-7, "source": "https://ai.google.dev/gemini-api/docs/models#gemini-2.5-flash-preview", - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/completions", - "/v1/batch" - ], - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/completions", "/v1/batch"], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text"], "supports_audio_output": false, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -11765,7 +10700,7 @@ }, "gemini-2.5-flash-preview-04-17": { "cache_read_input_token_cost": 3.75e-8, - "input_cost_per_audio_token": 0.000001, + "input_cost_per_audio_token": 1e-6, "input_cost_per_token": 1.5e-7, "litellm_provider": "vertex_ai-language-models", "max_audio_length_hours": 8.4, @@ -11778,23 +10713,12 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_reasoning_token": 0.0000035, + "output_cost_per_reasoning_token": 3.5e-6, "output_cost_per_token": 6e-7, "source": "https://ai.google.dev/gemini-api/docs/models#gemini-2.5-flash-preview", - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/completions", - "/v1/batch" - ], - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/completions", "/v1/batch"], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text"], "supports_audio_output": false, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -11809,7 +10733,7 @@ }, "gemini-2.5-flash-preview-05-20": { "cache_read_input_token_cost": 7.5e-8, - "input_cost_per_audio_token": 0.000001, + "input_cost_per_audio_token": 1e-6, "input_cost_per_token": 3e-7, "litellm_provider": "vertex_ai-language-models", "max_audio_length_hours": 8.4, @@ -11822,23 +10746,12 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_reasoning_token": 0.0000025, - "output_cost_per_token": 0.0000025, + "output_cost_per_reasoning_token": 2.5e-6, + "output_cost_per_token": 2.5e-6, "source": "https://ai.google.dev/gemini-api/docs/models#gemini-2.5-flash-preview", - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/completions", - "/v1/batch" - ], - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/completions", "/v1/batch"], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text"], "supports_audio_output": false, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -11855,8 +10768,8 @@ "gemini-2.5-pro": { "cache_read_input_token_cost": 1.25e-7, "cache_creation_input_token_cost_above_200k_tokens": 2.5e-7, - "input_cost_per_token": 0.00000125, - "input_cost_per_token_above_200k_tokens": 0.0000025, + "input_cost_per_token": 1.25e-6, + "input_cost_per_token_above_200k_tokens": 2.5e-6, "litellm_provider": "vertex_ai-language-models", "max_audio_length_hours": 8.4, "max_audio_per_prompt": 1, @@ -11868,22 +10781,12 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_token": 0.00001, - "output_cost_per_token_above_200k_tokens": 0.000015, + "output_cost_per_token": 1e-5, + "output_cost_per_token_above_200k_tokens": 1.5e-5, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing", - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/completions" - ], - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/completions"], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text"], "supports_audio_input": true, "supports_function_calling": true, "supports_pdf_input": true, @@ -11900,9 +10803,9 @@ "cache_read_input_token_cost": 2e-7, "cache_read_input_token_cost_above_200k_tokens": 4e-7, "cache_creation_input_token_cost_above_200k_tokens": 2.5e-7, - "input_cost_per_token": 0.000002, - "input_cost_per_token_above_200k_tokens": 0.000004, - "input_cost_per_token_batches": 0.000001, + "input_cost_per_token": 2e-6, + "input_cost_per_token_above_200k_tokens": 4e-6, + "input_cost_per_token_batches": 1e-6, "litellm_provider": "vertex_ai-language-models", "max_audio_length_hours": 8.4, "max_audio_per_prompt": 1, @@ -11914,24 +10817,13 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_token": 0.000012, - "output_cost_per_token_above_200k_tokens": 0.000018, - "output_cost_per_token_batches": 0.000006, + "output_cost_per_token": 1.2e-5, + "output_cost_per_token_above_200k_tokens": 1.8e-5, + "output_cost_per_token_batches": 6e-6, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing", - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/completions", - "/v1/batch" - ], - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/completions", "/v1/batch"], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text"], "supports_audio_input": true, "supports_function_calling": true, "supports_pdf_input": true, @@ -11948,9 +10840,9 @@ "cache_read_input_token_cost": 2e-7, "cache_read_input_token_cost_above_200k_tokens": 4e-7, "cache_creation_input_token_cost_above_200k_tokens": 2.5e-7, - "input_cost_per_token": 0.000002, - "input_cost_per_token_above_200k_tokens": 0.000004, - "input_cost_per_token_batches": 0.000001, + "input_cost_per_token": 2e-6, + "input_cost_per_token_above_200k_tokens": 4e-6, + "input_cost_per_token_batches": 1e-6, "litellm_provider": "vertex_ai", "max_audio_length_hours": 8.4, "max_audio_per_prompt": 1, @@ -11962,24 +10854,13 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_token": 0.000012, - "output_cost_per_token_above_200k_tokens": 0.000018, - "output_cost_per_token_batches": 0.000006, + "output_cost_per_token": 1.2e-5, + "output_cost_per_token_above_200k_tokens": 1.8e-5, + "output_cost_per_token_batches": 6e-6, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing", - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/completions", - "/v1/batch" - ], - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/completions", "/v1/batch"], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text"], "supports_audio_input": true, "supports_function_calling": true, "supports_pdf_input": true, @@ -11994,8 +10875,8 @@ }, "gemini-2.5-pro-exp-03-25": { "cache_read_input_token_cost": 3.125e-7, - "input_cost_per_token": 0.00000125, - "input_cost_per_token_above_200k_tokens": 0.0000025, + "input_cost_per_token": 1.25e-6, + "input_cost_per_token_above_200k_tokens": 2.5e-6, "litellm_provider": "vertex_ai-language-models", "max_audio_length_hours": 8.4, "max_audio_per_prompt": 1, @@ -12007,22 +10888,12 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_token": 0.00001, - "output_cost_per_token_above_200k_tokens": 0.000015, + "output_cost_per_token": 1e-5, + "output_cost_per_token_above_200k_tokens": 1.5e-5, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing", - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/completions" - ], - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/completions"], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text"], "supports_audio_input": true, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -12037,9 +10908,9 @@ }, "gemini-2.5-pro-preview-03-25": { "cache_read_input_token_cost": 3.125e-7, - "input_cost_per_audio_token": 0.00000125, - "input_cost_per_token": 0.00000125, - "input_cost_per_token_above_200k_tokens": 0.0000025, + "input_cost_per_audio_token": 1.25e-6, + "input_cost_per_token": 1.25e-6, + "input_cost_per_token_above_200k_tokens": 2.5e-6, "litellm_provider": "vertex_ai-language-models", "max_audio_length_hours": 8.4, "max_audio_per_prompt": 1, @@ -12051,23 +10922,12 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_token": 0.00001, - "output_cost_per_token_above_200k_tokens": 0.000015, + "output_cost_per_token": 1e-5, + "output_cost_per_token_above_200k_tokens": 1.5e-5, "source": "https://ai.google.dev/gemini-api/docs/models#gemini-2.5-flash-preview", - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/completions", - "/v1/batch" - ], - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/completions", "/v1/batch"], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text"], "supports_audio_output": false, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -12082,9 +10942,9 @@ }, "gemini-2.5-pro-preview-05-06": { "cache_read_input_token_cost": 3.125e-7, - "input_cost_per_audio_token": 0.00000125, - "input_cost_per_token": 0.00000125, - "input_cost_per_token_above_200k_tokens": 0.0000025, + "input_cost_per_audio_token": 1.25e-6, + "input_cost_per_token": 1.25e-6, + "input_cost_per_token_above_200k_tokens": 2.5e-6, "litellm_provider": "vertex_ai-language-models", "max_audio_length_hours": 8.4, "max_audio_per_prompt": 1, @@ -12096,26 +10956,13 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_token": 0.00001, - "output_cost_per_token_above_200k_tokens": 0.000015, + "output_cost_per_token": 1e-5, + "output_cost_per_token_above_200k_tokens": 1.5e-5, "source": "https://ai.google.dev/gemini-api/docs/models#gemini-2.5-flash-preview", - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/completions", - "/v1/batch" - ], - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text" - ], - "supported_regions": [ - "global" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/completions", "/v1/batch"], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text"], + "supported_regions": ["global"], "supports_audio_output": false, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -12130,9 +10977,9 @@ }, "gemini-2.5-pro-preview-06-05": { "cache_read_input_token_cost": 3.125e-7, - "input_cost_per_audio_token": 0.00000125, - "input_cost_per_token": 0.00000125, - "input_cost_per_token_above_200k_tokens": 0.0000025, + "input_cost_per_audio_token": 1.25e-6, + "input_cost_per_token": 1.25e-6, + "input_cost_per_token_above_200k_tokens": 2.5e-6, "litellm_provider": "vertex_ai-language-models", "max_audio_length_hours": 8.4, "max_audio_per_prompt": 1, @@ -12144,23 +10991,12 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_token": 0.00001, - "output_cost_per_token_above_200k_tokens": 0.000015, + "output_cost_per_token": 1e-5, + "output_cost_per_token_above_200k_tokens": 1.5e-5, "source": "https://ai.google.dev/gemini-api/docs/models#gemini-2.5-flash-preview", - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/completions", - "/v1/batch" - ], - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/completions", "/v1/batch"], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text"], "supports_audio_output": false, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -12176,8 +11012,8 @@ "gemini-2.5-pro-preview-tts": { "cache_read_input_token_cost": 3.125e-7, "input_cost_per_audio_token": 7e-7, - "input_cost_per_token": 0.00000125, - "input_cost_per_token_above_200k_tokens": 0.0000025, + "input_cost_per_token": 1.25e-6, + "input_cost_per_token_above_200k_tokens": 2.5e-6, "litellm_provider": "vertex_ai-language-models", "max_audio_length_hours": 8.4, "max_audio_per_prompt": 1, @@ -12189,15 +11025,11 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_token": 0.00001, - "output_cost_per_token_above_200k_tokens": 0.000015, + "output_cost_per_token": 1e-5, + "output_cost_per_token_above_200k_tokens": 1.5e-5, "source": "https://ai.google.dev/gemini-api/docs/pricing#gemini-2.5-pro-preview", - "supported_modalities": [ - "text" - ], - "supported_output_modalities": [ - "audio" - ], + "supported_modalities": ["text"], + "supported_output_modalities": ["audio"], "supports_audio_output": false, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -12244,7 +11076,7 @@ "max_tokens": 8192, "mode": "chat", "output_cost_per_character": 3.75e-7, - "output_cost_per_token": 0.0000015, + "output_cost_per_token": 1.5e-6, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing", "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -12276,7 +11108,7 @@ "max_video_length": 2, "max_videos_per_prompt": 1, "mode": "chat", - "output_cost_per_token": 0.0000015, + "output_cost_per_token": 1.5e-6, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models", "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -12321,7 +11153,7 @@ "tpm": 4000000 }, "gemini/gemini-1.5-flash-001": { - "cache_creation_input_token_cost": 0.000001, + "cache_creation_input_token_cost": 1e-6, "cache_read_input_token_cost": 1.875e-8, "deprecation_date": "2025-05-24", "input_cost_per_token": 7.5e-8, @@ -12350,7 +11182,7 @@ "tpm": 4000000 }, "gemini/gemini-1.5-flash-002": { - "cache_creation_input_token_cost": 0.000001, + "cache_creation_input_token_cost": 1e-6, "cache_read_input_token_cost": 1.875e-8, "deprecation_date": "2025-09-24", "input_cost_per_token": 7.5e-8, @@ -12507,15 +11339,15 @@ "tpm": 4000000 }, "gemini/gemini-1.5-pro": { - "input_cost_per_token": 0.0000035, - "input_cost_per_token_above_128k_tokens": 0.000007, + "input_cost_per_token": 3.5e-6, + "input_cost_per_token_above_128k_tokens": 7e-6, "litellm_provider": "gemini", "max_input_tokens": 2097152, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.0000105, - "output_cost_per_token_above_128k_tokens": 0.000021, + "output_cost_per_token": 1.05e-5, + "output_cost_per_token_above_128k_tokens": 2.1e-5, "rpm": 1000, "source": "https://ai.google.dev/pricing", "supports_function_calling": true, @@ -12527,15 +11359,15 @@ }, "gemini/gemini-1.5-pro-001": { "deprecation_date": "2025-05-24", - "input_cost_per_token": 0.0000035, - "input_cost_per_token_above_128k_tokens": 0.000007, + "input_cost_per_token": 3.5e-6, + "input_cost_per_token_above_128k_tokens": 7e-6, "litellm_provider": "gemini", "max_input_tokens": 2097152, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.0000105, - "output_cost_per_token_above_128k_tokens": 0.000021, + "output_cost_per_token": 1.05e-5, + "output_cost_per_token_above_128k_tokens": 2.1e-5, "rpm": 1000, "source": "https://ai.google.dev/pricing", "supports_function_calling": true, @@ -12548,15 +11380,15 @@ }, "gemini/gemini-1.5-pro-002": { "deprecation_date": "2025-09-24", - "input_cost_per_token": 0.0000035, - "input_cost_per_token_above_128k_tokens": 0.000007, + "input_cost_per_token": 3.5e-6, + "input_cost_per_token_above_128k_tokens": 7e-6, "litellm_provider": "gemini", "max_input_tokens": 2097152, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.0000105, - "output_cost_per_token_above_128k_tokens": 0.000021, + "output_cost_per_token": 1.05e-5, + "output_cost_per_token_above_128k_tokens": 2.1e-5, "rpm": 1000, "source": "https://ai.google.dev/pricing", "supports_function_calling": true, @@ -12568,15 +11400,15 @@ "tpm": 4000000 }, "gemini/gemini-1.5-pro-exp-0801": { - "input_cost_per_token": 0.0000035, - "input_cost_per_token_above_128k_tokens": 0.000007, + "input_cost_per_token": 3.5e-6, + "input_cost_per_token_above_128k_tokens": 7e-6, "litellm_provider": "gemini", "max_input_tokens": 2097152, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.0000105, - "output_cost_per_token_above_128k_tokens": 0.000021, + "output_cost_per_token": 1.05e-5, + "output_cost_per_token_above_128k_tokens": 2.1e-5, "rpm": 1000, "source": "https://ai.google.dev/pricing", "supports_function_calling": true, @@ -12606,15 +11438,15 @@ "tpm": 4000000 }, "gemini/gemini-1.5-pro-latest": { - "input_cost_per_token": 0.0000035, - "input_cost_per_token_above_128k_tokens": 0.000007, + "input_cost_per_token": 3.5e-6, + "input_cost_per_token_above_128k_tokens": 7e-6, "litellm_provider": "gemini", "max_input_tokens": 1048576, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.00000105, - "output_cost_per_token_above_128k_tokens": 0.000021, + "output_cost_per_token": 1.05e-6, + "output_cost_per_token_above_128k_tokens": 2.1e-5, "rpm": 1000, "source": "https://ai.google.dev/pricing", "supports_function_calling": true, @@ -12642,16 +11474,8 @@ "output_cost_per_token": 4e-7, "rpm": 10000, "source": "https://ai.google.dev/pricing#2_0flash", - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text", - "image" - ], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text", "image"], "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -12682,16 +11506,8 @@ "output_cost_per_token": 4e-7, "rpm": 10000, "source": "https://ai.google.dev/pricing#2_0flash", - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text", - "image" - ], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text", "image"], "supports_audio_output": false, "supports_function_calling": true, "supports_prompt_caching": true, @@ -12703,7 +11519,7 @@ "tpm": 10000000 }, "gemini/gemini-2.0-flash-exp": { - "cache_read_input_token_cost": 0, + "cache_read_input_token_cost": 0.0, "input_cost_per_audio_per_second": 0, "input_cost_per_audio_per_second_above_128k_tokens": 0, "input_cost_per_character": 0, @@ -12731,16 +11547,8 @@ "output_cost_per_token_above_128k_tokens": 0, "rpm": 10, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#gemini-2.0-flash", - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text", - "image" - ], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text", "image"], "supports_audio_output": true, "supports_function_calling": true, "supports_prompt_caching": true, @@ -12768,15 +11576,8 @@ "output_cost_per_token": 3e-7, "rpm": 4000, "source": "https://ai.google.dev/gemini-api/docs/pricing#gemini-2.0-flash-lite", - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text" - ], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text"], "supports_audio_output": true, "supports_function_calling": true, "supports_prompt_caching": true, @@ -12805,15 +11606,8 @@ "output_cost_per_token": 3e-7, "rpm": 60000, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#gemini-2.0-flash-lite", - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text" - ], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text"], "supports_audio_output": false, "supports_function_calling": true, "supports_prompt_caching": true, @@ -12826,10 +11620,10 @@ }, "gemini/gemini-2.0-flash-live-001": { "cache_read_input_token_cost": 7.5e-8, - "input_cost_per_audio_token": 0.0000021, - "input_cost_per_image": 0.0000021, + "input_cost_per_audio_token": 2.1e-6, + "input_cost_per_image": 2.1e-6, "input_cost_per_token": 3.5e-7, - "input_cost_per_video_per_second": 0.0000021, + "input_cost_per_video_per_second": 2.1e-6, "litellm_provider": "gemini", "max_audio_length_hours": 8.4, "max_audio_per_prompt": 1, @@ -12841,24 +11635,13 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_audio_token": 0.0000085, - "output_cost_per_token": 0.0000015, + "output_cost_per_audio_token": 8.5e-6, + "output_cost_per_token": 1.5e-6, "rpm": 10, "source": "https://ai.google.dev/gemini-api/docs/models#gemini-2-0-flash-live-001", - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/completions" - ], - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text", - "audio" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/completions"], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text", "audio"], "supports_audio_output": true, "supports_function_calling": true, "supports_pdf_input": true, @@ -12890,16 +11673,8 @@ "output_cost_per_token": 4e-7, "rpm": 10000, "source": "https://ai.google.dev/pricing#2_0flash", - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text", - "image" - ], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text", "image"], "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -12912,7 +11687,7 @@ "tpm": 10000000 }, "gemini/gemini-2.0-flash-thinking-exp": { - "cache_read_input_token_cost": 0, + "cache_read_input_token_cost": 0.0, "input_cost_per_audio_per_second": 0, "input_cost_per_audio_per_second_above_128k_tokens": 0, "input_cost_per_character": 0, @@ -12940,16 +11715,8 @@ "output_cost_per_token_above_128k_tokens": 0, "rpm": 10, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#gemini-2.0-flash", - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text", - "image" - ], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text", "image"], "supports_audio_output": true, "supports_function_calling": true, "supports_prompt_caching": true, @@ -12961,7 +11728,7 @@ "tpm": 4000000 }, "gemini/gemini-2.0-flash-thinking-exp-01-21": { - "cache_read_input_token_cost": 0, + "cache_read_input_token_cost": 0.0, "input_cost_per_audio_per_second": 0, "input_cost_per_audio_per_second_above_128k_tokens": 0, "input_cost_per_character": 0, @@ -12989,16 +11756,8 @@ "output_cost_per_token_above_128k_tokens": 0, "rpm": 10, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#gemini-2.0-flash", - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text", - "image" - ], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text", "image"], "supports_audio_output": true, "supports_function_calling": true, "supports_prompt_caching": true, @@ -13011,7 +11770,7 @@ "tpm": 4000000 }, "gemini/gemini-2.0-pro-exp-02-05": { - "cache_read_input_token_cost": 0, + "cache_read_input_token_cost": 0.0, "input_cost_per_audio_per_second": 0, "input_cost_per_audio_per_second_above_128k_tokens": 0, "input_cost_per_character": 0, @@ -13053,7 +11812,7 @@ }, "gemini/gemini-2.5-flash": { "cache_read_input_token_cost": 3e-8, - "input_cost_per_audio_token": 0.000001, + "input_cost_per_audio_token": 1e-6, "input_cost_per_token": 3e-7, "litellm_provider": "gemini", "max_audio_length_hours": 8.4, @@ -13066,24 +11825,13 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_reasoning_token": 0.0000025, - "output_cost_per_token": 0.0000025, + "output_cost_per_reasoning_token": 2.5e-6, + "output_cost_per_token": 2.5e-6, "rpm": 100000, "source": "https://ai.google.dev/gemini-api/docs/models#gemini-2.5-flash-preview", - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/completions", - "/v1/batch" - ], - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/completions", "/v1/batch"], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text"], "supports_audio_output": false, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -13100,7 +11848,7 @@ }, "gemini/gemini-2.5-flash-image": { "cache_read_input_token_cost": 3e-8, - "input_cost_per_audio_token": 0.000001, + "input_cost_per_audio_token": 1e-6, "input_cost_per_token": 3e-7, "litellm_provider": "vertex_ai-language-models", "max_audio_length_hours": 8.4, @@ -13115,25 +11863,13 @@ "max_videos_per_prompt": 10, "mode": "image_generation", "output_cost_per_image": 0.039, - "output_cost_per_reasoning_token": 0.0000025, - "output_cost_per_token": 0.0000025, + "output_cost_per_reasoning_token": 2.5e-6, + "output_cost_per_token": 2.5e-6, "rpm": 100000, "source": "https://ai.google.dev/gemini-api/docs/pricing#gemini-2.5-flash-image", - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/completions", - "/v1/batch" - ], - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text", - "image" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/completions", "/v1/batch"], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text", "image"], "supports_audio_output": false, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -13149,7 +11885,7 @@ }, "gemini/gemini-2.5-flash-image-preview": { "cache_read_input_token_cost": 7.5e-8, - "input_cost_per_audio_token": 0.000001, + "input_cost_per_audio_token": 1e-6, "input_cost_per_token": 3e-7, "litellm_provider": "gemini", "max_audio_length_hours": 8.4, @@ -13163,25 +11899,13 @@ "max_videos_per_prompt": 10, "mode": "image_generation", "output_cost_per_image": 0.039, - "output_cost_per_reasoning_token": 0.00003, - "output_cost_per_token": 0.00003, + "output_cost_per_reasoning_token": 3e-5, + "output_cost_per_token": 3e-5, "rpm": 100000, "source": "https://ai.google.dev/gemini-api/docs/models#gemini-2.5-flash-preview", - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/completions", - "/v1/batch" - ], - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text", - "image" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/completions", "/v1/batch"], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text", "image"], "supports_audio_output": false, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -13197,32 +11921,22 @@ }, "gemini/gemini-3-pro-image-preview": { "input_cost_per_image": 0.0011, - "input_cost_per_token": 0.000002, - "input_cost_per_token_batches": 0.000001, + "input_cost_per_token": 2e-6, + "input_cost_per_token_batches": 1e-6, "litellm_provider": "gemini", "max_input_tokens": 65536, "max_output_tokens": 32768, "max_tokens": 65536, "mode": "image_generation", "output_cost_per_image": 0.134, - "output_cost_per_token": 0.000012, + "output_cost_per_token": 1.2e-5, "rpm": 1000, "tpm": 4000000, - "output_cost_per_token_batches": 0.000006, + "output_cost_per_token_batches": 6e-6, "source": "https://ai.google.dev/gemini-api/docs/pricing", - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/completions", - "/v1/batch" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text", - "image" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/completions", "/v1/batch"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text", "image"], "supports_function_calling": false, "supports_prompt_caching": true, "supports_response_schema": true, @@ -13249,20 +11963,9 @@ "output_cost_per_token": 4e-7, "rpm": 15, "source": "https://ai.google.dev/gemini-api/docs/models#gemini-2.5-flash-lite", - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/completions", - "/v1/batch" - ], - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/completions", "/v1/batch"], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text"], "supports_audio_output": false, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -13296,20 +11999,9 @@ "output_cost_per_token": 4e-7, "rpm": 15, "source": "https://developers.googleblog.com/en/continuing-to-bring-you-our-latest-models-with-an-improved-gemini-2-5-flash-and-flash-lite-release/", - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/completions", - "/v1/batch" - ], - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/completions", "/v1/batch"], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text"], "supports_audio_output": false, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -13326,7 +12018,7 @@ }, "gemini/gemini-2.5-flash-preview-09-2025": { "cache_read_input_token_cost": 7.5e-8, - "input_cost_per_audio_token": 0.000001, + "input_cost_per_audio_token": 1e-6, "input_cost_per_token": 3e-7, "litellm_provider": "gemini", "max_audio_length_hours": 8.4, @@ -13339,24 +12031,13 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_reasoning_token": 0.0000025, - "output_cost_per_token": 0.0000025, + "output_cost_per_reasoning_token": 2.5e-6, + "output_cost_per_token": 2.5e-6, "rpm": 15, "source": "https://developers.googleblog.com/en/continuing-to-bring-you-our-latest-models-with-an-improved-gemini-2-5-flash-and-flash-lite-release/", - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/completions", - "/v1/batch" - ], - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/completions", "/v1/batch"], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text"], "supports_audio_output": false, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -13373,7 +12054,7 @@ }, "gemini/gemini-flash-latest": { "cache_read_input_token_cost": 7.5e-8, - "input_cost_per_audio_token": 0.000001, + "input_cost_per_audio_token": 1e-6, "input_cost_per_token": 3e-7, "litellm_provider": "gemini", "max_audio_length_hours": 8.4, @@ -13386,24 +12067,13 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_reasoning_token": 0.0000025, - "output_cost_per_token": 0.0000025, + "output_cost_per_reasoning_token": 2.5e-6, + "output_cost_per_token": 2.5e-6, "rpm": 15, "source": "https://developers.googleblog.com/en/continuing-to-bring-you-our-latest-models-with-an-improved-gemini-2-5-flash-and-flash-lite-release/", - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/completions", - "/v1/batch" - ], - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/completions", "/v1/batch"], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text"], "supports_audio_output": false, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -13437,20 +12107,9 @@ "output_cost_per_token": 4e-7, "rpm": 15, "source": "https://developers.googleblog.com/en/continuing-to-bring-you-our-latest-models-with-an-improved-gemini-2-5-flash-and-flash-lite-release/", - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/completions", - "/v1/batch" - ], - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/completions", "/v1/batch"], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text"], "supports_audio_output": false, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -13484,20 +12143,9 @@ "output_cost_per_token": 4e-7, "rpm": 15, "source": "https://ai.google.dev/gemini-api/docs/models#gemini-2.5-flash-lite", - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/completions", - "/v1/batch" - ], - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/completions", "/v1/batch"], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text"], "supports_audio_output": false, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -13514,7 +12162,7 @@ }, "gemini/gemini-2.5-flash-preview-04-17": { "cache_read_input_token_cost": 3.75e-8, - "input_cost_per_audio_token": 0.000001, + "input_cost_per_audio_token": 1e-6, "input_cost_per_token": 1.5e-7, "litellm_provider": "gemini", "max_audio_length_hours": 8.4, @@ -13527,23 +12175,13 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_reasoning_token": 0.0000035, + "output_cost_per_reasoning_token": 3.5e-6, "output_cost_per_token": 6e-7, "rpm": 10, "source": "https://ai.google.dev/gemini-api/docs/models#gemini-2.5-flash-preview", - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/completions" - ], - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/completions"], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text"], "supports_audio_output": false, "supports_function_calling": true, "supports_pdf_input": true, @@ -13558,7 +12196,7 @@ }, "gemini/gemini-2.5-flash-preview-05-20": { "cache_read_input_token_cost": 7.5e-8, - "input_cost_per_audio_token": 0.000001, + "input_cost_per_audio_token": 1e-6, "input_cost_per_token": 3e-7, "litellm_provider": "gemini", "max_audio_length_hours": 8.4, @@ -13571,23 +12209,13 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_reasoning_token": 0.0000025, - "output_cost_per_token": 0.0000025, + "output_cost_per_reasoning_token": 2.5e-6, + "output_cost_per_token": 2.5e-6, "rpm": 10, "source": "https://ai.google.dev/gemini-api/docs/models#gemini-2.5-flash-preview", - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/completions" - ], - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/completions"], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text"], "supports_audio_output": false, "supports_function_calling": true, "supports_pdf_input": true, @@ -13603,7 +12231,7 @@ }, "gemini/gemini-2.5-flash-preview-tts": { "cache_read_input_token_cost": 3.75e-8, - "input_cost_per_audio_token": 0.000001, + "input_cost_per_audio_token": 1e-6, "input_cost_per_token": 1.5e-7, "litellm_provider": "gemini", "max_audio_length_hours": 8.4, @@ -13616,20 +12244,13 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_reasoning_token": 0.0000035, + "output_cost_per_reasoning_token": 3.5e-6, "output_cost_per_token": 6e-7, "rpm": 10, "source": "https://ai.google.dev/gemini-api/docs/models#gemini-2.5-flash-preview", - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/completions" - ], - "supported_modalities": [ - "text" - ], - "supported_output_modalities": [ - "audio" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/completions"], + "supported_modalities": ["text"], + "supported_output_modalities": ["audio"], "supports_audio_output": false, "supports_function_calling": true, "supports_prompt_caching": true, @@ -13643,8 +12264,8 @@ }, "gemini/gemini-2.5-pro": { "cache_read_input_token_cost": 3.125e-7, - "input_cost_per_token": 0.00000125, - "input_cost_per_token_above_200k_tokens": 0.0000025, + "input_cost_per_token": 1.25e-6, + "input_cost_per_token_above_200k_tokens": 2.5e-6, "litellm_provider": "gemini", "max_audio_length_hours": 8.4, "max_audio_per_prompt": 1, @@ -13656,23 +12277,13 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_token": 0.00001, - "output_cost_per_token_above_200k_tokens": 0.000015, + "output_cost_per_token": 1e-5, + "output_cost_per_token_above_200k_tokens": 1.5e-5, "rpm": 2000, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing", - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/completions" - ], - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/completions"], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text"], "supports_audio_input": true, "supports_function_calling": true, "supports_pdf_input": true, @@ -13689,9 +12300,9 @@ "gemini/gemini-3-pro-preview": { "cache_read_input_token_cost": 2e-7, "cache_read_input_token_cost_above_200k_tokens": 4e-7, - "input_cost_per_token": 0.000002, - "input_cost_per_token_above_200k_tokens": 0.000004, - "input_cost_per_token_batches": 0.000001, + "input_cost_per_token": 2e-6, + "input_cost_per_token_above_200k_tokens": 4e-6, + "input_cost_per_token_batches": 1e-6, "litellm_provider": "gemini", "max_audio_length_hours": 8.4, "max_audio_per_prompt": 1, @@ -13703,25 +12314,14 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_token": 0.000012, - "output_cost_per_token_above_200k_tokens": 0.000018, - "output_cost_per_token_batches": 0.000006, + "output_cost_per_token": 1.2e-5, + "output_cost_per_token_above_200k_tokens": 1.8e-5, + "output_cost_per_token_batches": 6e-6, "rpm": 2000, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing", - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/completions", - "/v1/batch" - ], - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/completions", "/v1/batch"], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text"], "supports_audio_input": true, "supports_function_calling": true, "supports_pdf_input": true, @@ -13736,9 +12336,9 @@ "tpm": 800000 }, "gemini/gemini-2.5-pro-exp-03-25": { - "cache_read_input_token_cost": 0, - "input_cost_per_token": 0, - "input_cost_per_token_above_200k_tokens": 0, + "cache_read_input_token_cost": 0.0, + "input_cost_per_token": 0.0, + "input_cost_per_token_above_200k_tokens": 0.0, "litellm_provider": "gemini", "max_audio_length_hours": 8.4, "max_audio_per_prompt": 1, @@ -13750,23 +12350,13 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_token": 0, - "output_cost_per_token_above_200k_tokens": 0, + "output_cost_per_token": 0.0, + "output_cost_per_token_above_200k_tokens": 0.0, "rpm": 5, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing", - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/completions" - ], - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/completions"], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text"], "supports_audio_input": true, "supports_function_calling": true, "supports_pdf_input": true, @@ -13782,8 +12372,8 @@ "gemini/gemini-2.5-pro-preview-03-25": { "cache_read_input_token_cost": 3.125e-7, "input_cost_per_audio_token": 7e-7, - "input_cost_per_token": 0.00000125, - "input_cost_per_token_above_200k_tokens": 0.0000025, + "input_cost_per_token": 1.25e-6, + "input_cost_per_token_above_200k_tokens": 2.5e-6, "litellm_provider": "gemini", "max_audio_length_hours": 8.4, "max_audio_per_prompt": 1, @@ -13795,19 +12385,12 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_token": 0.00001, - "output_cost_per_token_above_200k_tokens": 0.000015, + "output_cost_per_token": 1e-5, + "output_cost_per_token_above_200k_tokens": 1.5e-5, "rpm": 10000, "source": "https://ai.google.dev/gemini-api/docs/pricing#gemini-2.5-pro-preview", - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text" - ], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text"], "supports_audio_output": false, "supports_function_calling": true, "supports_pdf_input": true, @@ -13822,8 +12405,8 @@ "gemini/gemini-2.5-pro-preview-05-06": { "cache_read_input_token_cost": 3.125e-7, "input_cost_per_audio_token": 7e-7, - "input_cost_per_token": 0.00000125, - "input_cost_per_token_above_200k_tokens": 0.0000025, + "input_cost_per_token": 1.25e-6, + "input_cost_per_token_above_200k_tokens": 2.5e-6, "litellm_provider": "gemini", "max_audio_length_hours": 8.4, "max_audio_per_prompt": 1, @@ -13835,19 +12418,12 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_token": 0.00001, - "output_cost_per_token_above_200k_tokens": 0.000015, + "output_cost_per_token": 1e-5, + "output_cost_per_token_above_200k_tokens": 1.5e-5, "rpm": 10000, "source": "https://ai.google.dev/gemini-api/docs/pricing#gemini-2.5-pro-preview", - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text" - ], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text"], "supports_audio_output": false, "supports_function_calling": true, "supports_pdf_input": true, @@ -13863,8 +12439,8 @@ "gemini/gemini-2.5-pro-preview-06-05": { "cache_read_input_token_cost": 3.125e-7, "input_cost_per_audio_token": 7e-7, - "input_cost_per_token": 0.00000125, - "input_cost_per_token_above_200k_tokens": 0.0000025, + "input_cost_per_token": 1.25e-6, + "input_cost_per_token_above_200k_tokens": 2.5e-6, "litellm_provider": "gemini", "max_audio_length_hours": 8.4, "max_audio_per_prompt": 1, @@ -13876,19 +12452,12 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_token": 0.00001, - "output_cost_per_token_above_200k_tokens": 0.000015, + "output_cost_per_token": 1e-5, + "output_cost_per_token_above_200k_tokens": 1.5e-5, "rpm": 10000, "source": "https://ai.google.dev/gemini-api/docs/pricing#gemini-2.5-pro-preview", - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text" - ], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text"], "supports_audio_output": false, "supports_function_calling": true, "supports_pdf_input": true, @@ -13904,8 +12473,8 @@ "gemini/gemini-2.5-pro-preview-tts": { "cache_read_input_token_cost": 3.125e-7, "input_cost_per_audio_token": 7e-7, - "input_cost_per_token": 0.00000125, - "input_cost_per_token_above_200k_tokens": 0.0000025, + "input_cost_per_token": 1.25e-6, + "input_cost_per_token_above_200k_tokens": 2.5e-6, "litellm_provider": "gemini", "max_audio_length_hours": 8.4, "max_audio_per_prompt": 1, @@ -13917,16 +12486,12 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_token": 0.00001, - "output_cost_per_token_above_200k_tokens": 0.000015, + "output_cost_per_token": 1e-5, + "output_cost_per_token_above_200k_tokens": 1.5e-5, "rpm": 10000, "source": "https://ai.google.dev/gemini-api/docs/pricing#gemini-2.5-pro-preview", - "supported_modalities": [ - "text" - ], - "supported_output_modalities": [ - "audio" - ], + "supported_modalities": ["text"], + "supported_output_modalities": ["audio"], "supports_audio_output": false, "supports_function_calling": true, "supports_prompt_caching": true, @@ -14001,7 +12566,7 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.00000105, + "output_cost_per_token": 1.05e-6, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models", "supports_function_calling": true, "supports_tool_choice": true, @@ -14013,7 +12578,7 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.00000105, + "output_cost_per_token": 1.05e-6, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models", "supports_function_calling": true, "supports_tool_choice": true, @@ -14027,8 +12592,8 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.00000105, - "output_cost_per_token_above_128k_tokens": 0.0000021, + "output_cost_per_token": 1.05e-6, + "output_cost_per_token_above_128k_tokens": 2.1e-6, "rpd": 30000, "rpm": 360, "source": "https://ai.google.dev/gemini-api/docs/models/gemini", @@ -14044,8 +12609,8 @@ "max_output_tokens": 2048, "max_tokens": 2048, "mode": "chat", - "output_cost_per_token": 0.00000105, - "output_cost_per_token_above_128k_tokens": 0.0000021, + "output_cost_per_token": 1.05e-6, + "output_cost_per_token_above_128k_tokens": 2.1e-6, "rpd": 30000, "rpm": 360, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models", @@ -14153,12 +12718,8 @@ "mode": "video_generation", "output_cost_per_second": 0.35, "source": "https://ai.google.dev/gemini-api/docs/video", - "supported_modalities": [ - "text" - ], - "supported_output_modalities": [ - "video" - ] + "supported_modalities": ["text"], + "supported_output_modalities": ["video"] }, "gemini/veo-3.0-fast-generate-preview": { "litellm_provider": "gemini", @@ -14167,12 +12728,8 @@ "mode": "video_generation", "output_cost_per_second": 0.4, "source": "https://ai.google.dev/gemini-api/docs/video", - "supported_modalities": [ - "text" - ], - "supported_output_modalities": [ - "video" - ] + "supported_modalities": ["text"], + "supported_output_modalities": ["video"] }, "gemini/veo-3.0-generate-preview": { "litellm_provider": "gemini", @@ -14181,12 +12738,8 @@ "mode": "video_generation", "output_cost_per_second": 0.75, "source": "https://ai.google.dev/gemini-api/docs/video", - "supported_modalities": [ - "text" - ], - "supported_output_modalities": [ - "video" - ] + "supported_modalities": ["text"], + "supported_output_modalities": ["video"] }, "gemini/veo-3.1-fast-generate-preview": { "litellm_provider": "gemini", @@ -14195,12 +12748,8 @@ "mode": "video_generation", "output_cost_per_second": 0.15, "source": "https://ai.google.dev/gemini-api/docs/video", - "supported_modalities": [ - "text" - ], - "supported_output_modalities": [ - "video" - ] + "supported_modalities": ["text"], + "supported_output_modalities": ["video"] }, "gemini/veo-3.1-generate-preview": { "litellm_provider": "gemini", @@ -14209,12 +12758,8 @@ "mode": "video_generation", "output_cost_per_second": 0.4, "source": "https://ai.google.dev/gemini-api/docs/video", - "supported_modalities": [ - "text" - ], - "supported_output_modalities": [ - "video" - ] + "supported_modalities": ["text"], + "supported_output_modalities": ["video"] }, "google_pse/search": { "input_cost_per_query": 0.005, @@ -14222,19 +12767,19 @@ "mode": "search" }, "global.anthropic.claude-sonnet-4-5-20250929-v1:0": { - "cache_creation_input_token_cost": 0.00000375, + "cache_creation_input_token_cost": 3.75e-6, "cache_read_input_token_cost": 3e-7, - "input_cost_per_token": 0.000003, - "input_cost_per_token_above_200k_tokens": 0.000006, - "output_cost_per_token_above_200k_tokens": 0.0000225, - "cache_creation_input_token_cost_above_200k_tokens": 0.0000075, + "input_cost_per_token": 3e-6, + "input_cost_per_token_above_200k_tokens": 6e-6, + "output_cost_per_token_above_200k_tokens": 2.25e-5, + "cache_creation_input_token_cost_above_200k_tokens": 7.5e-6, "cache_read_input_token_cost_above_200k_tokens": 6e-7, "litellm_provider": "bedrock_converse", "max_input_tokens": 200000, "max_output_tokens": 64000, "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -14252,19 +12797,19 @@ "tool_use_system_prompt_tokens": 346 }, "global.anthropic.claude-sonnet-4-20250514-v1:0": { - "cache_creation_input_token_cost": 0.00000375, + "cache_creation_input_token_cost": 3.75e-6, "cache_read_input_token_cost": 3e-7, - "input_cost_per_token": 0.000003, - "input_cost_per_token_above_200k_tokens": 0.000006, - "output_cost_per_token_above_200k_tokens": 0.0000225, - "cache_creation_input_token_cost_above_200k_tokens": 0.0000075, + "input_cost_per_token": 3e-6, + "input_cost_per_token_above_200k_tokens": 6e-6, + "output_cost_per_token_above_200k_tokens": 2.25e-5, + "cache_creation_input_token_cost_above_200k_tokens": 7.5e-6, "cache_read_input_token_cost_above_200k_tokens": 6e-7, "litellm_provider": "bedrock_converse", "max_input_tokens": 1000000, "max_output_tokens": 64000, "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -14282,15 +12827,15 @@ "tool_use_system_prompt_tokens": 159 }, "global.anthropic.claude-haiku-4-5-20251001-v1:0": { - "cache_creation_input_token_cost": 0.000001375, + "cache_creation_input_token_cost": 1.375e-6, "cache_read_input_token_cost": 1.1e-7, - "input_cost_per_token": 0.0000011, + "input_cost_per_token": 1.1e-6, "litellm_provider": "bedrock_converse", "max_input_tokens": 200000, "max_output_tokens": 64000, "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 0.0000055, + "output_cost_per_token": 5.5e-6, "source": "https://aws.amazon.com/about-aws/whats-new/2025/10/claude-4-5-haiku-anthropic-amazon-bedrock", "supports_assistant_prefill": true, "supports_computer_use": true, @@ -14304,13 +12849,13 @@ "tool_use_system_prompt_tokens": 346 }, "gpt-3.5-turbo": { - "input_cost_per_token": 5e-7, + "input_cost_per_token": 0.5e-6, "litellm_provider": "openai", "max_input_tokens": 16385, "max_output_tokens": 4096, "max_tokens": 4097, "mode": "chat", - "output_cost_per_token": 0.0000015, + "output_cost_per_token": 1.5e-6, "supports_function_calling": true, "supports_prompt_caching": true, "supports_system_messages": true, @@ -14323,7 +12868,7 @@ "max_output_tokens": 4096, "max_tokens": 16385, "mode": "chat", - "output_cost_per_token": 0.0000015, + "output_cost_per_token": 1.5e-6, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -14331,25 +12876,25 @@ "supports_tool_choice": true }, "gpt-3.5-turbo-0301": { - "input_cost_per_token": 0.0000015, + "input_cost_per_token": 1.5e-6, "litellm_provider": "openai", "max_input_tokens": 4097, "max_output_tokens": 4096, "max_tokens": 4097, "mode": "chat", - "output_cost_per_token": 0.000002, + "output_cost_per_token": 2e-6, "supports_prompt_caching": true, "supports_system_messages": true, "supports_tool_choice": true }, "gpt-3.5-turbo-0613": { - "input_cost_per_token": 0.0000015, + "input_cost_per_token": 1.5e-6, "litellm_provider": "openai", "max_input_tokens": 4097, "max_output_tokens": 4096, "max_tokens": 4097, "mode": "chat", - "output_cost_per_token": 0.000002, + "output_cost_per_token": 2e-6, "supports_function_calling": true, "supports_prompt_caching": true, "supports_system_messages": true, @@ -14357,13 +12902,13 @@ }, "gpt-3.5-turbo-1106": { "deprecation_date": "2026-09-28", - "input_cost_per_token": 0.000001, + "input_cost_per_token": 1e-6, "litellm_provider": "openai", "max_input_tokens": 16385, "max_output_tokens": 4096, "max_tokens": 16385, "mode": "chat", - "output_cost_per_token": 0.000002, + "output_cost_per_token": 2e-6, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -14371,55 +12916,55 @@ "supports_tool_choice": true }, "gpt-3.5-turbo-16k": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "openai", "max_input_tokens": 16385, "max_output_tokens": 4096, "max_tokens": 16385, "mode": "chat", - "output_cost_per_token": 0.000004, + "output_cost_per_token": 4e-6, "supports_prompt_caching": true, "supports_system_messages": true, "supports_tool_choice": true }, "gpt-3.5-turbo-16k-0613": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "openai", "max_input_tokens": 16385, "max_output_tokens": 4096, "max_tokens": 16385, "mode": "chat", - "output_cost_per_token": 0.000004, + "output_cost_per_token": 4e-6, "supports_prompt_caching": true, "supports_system_messages": true, "supports_tool_choice": true }, "gpt-3.5-turbo-instruct": { - "input_cost_per_token": 0.0000015, + "input_cost_per_token": 1.5e-6, "litellm_provider": "text-completion-openai", "max_input_tokens": 8192, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "completion", - "output_cost_per_token": 0.000002 + "output_cost_per_token": 2e-6 }, "gpt-3.5-turbo-instruct-0914": { - "input_cost_per_token": 0.0000015, + "input_cost_per_token": 1.5e-6, "litellm_provider": "text-completion-openai", "max_input_tokens": 8192, "max_output_tokens": 4097, "max_tokens": 4097, "mode": "completion", - "output_cost_per_token": 0.000002 + "output_cost_per_token": 2e-6 }, "gpt-4": { - "input_cost_per_token": 0.00003, + "input_cost_per_token": 3e-5, "litellm_provider": "openai", "max_input_tokens": 8192, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.00006, + "output_cost_per_token": 6e-5, "supports_function_calling": true, "supports_prompt_caching": true, "supports_system_messages": true, @@ -14427,13 +12972,13 @@ }, "gpt-4-0125-preview": { "deprecation_date": "2026-03-26", - "input_cost_per_token": 0.00001, + "input_cost_per_token": 1e-5, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.00003, + "output_cost_per_token": 3e-5, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -14441,26 +12986,26 @@ "supports_tool_choice": true }, "gpt-4-0314": { - "input_cost_per_token": 0.00003, + "input_cost_per_token": 3e-5, "litellm_provider": "openai", "max_input_tokens": 8192, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.00006, + "output_cost_per_token": 6e-5, "supports_prompt_caching": true, "supports_system_messages": true, "supports_tool_choice": true }, "gpt-4-0613": { "deprecation_date": "2025-06-06", - "input_cost_per_token": 0.00003, + "input_cost_per_token": 3e-5, "litellm_provider": "openai", "max_input_tokens": 8192, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.00006, + "output_cost_per_token": 6e-5, "supports_function_calling": true, "supports_prompt_caching": true, "supports_system_messages": true, @@ -14468,13 +13013,13 @@ }, "gpt-4-1106-preview": { "deprecation_date": "2026-03-26", - "input_cost_per_token": 0.00001, + "input_cost_per_token": 1e-5, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.00003, + "output_cost_per_token": 3e-5, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -14483,13 +13028,13 @@ }, "gpt-4-1106-vision-preview": { "deprecation_date": "2024-12-06", - "input_cost_per_token": 0.00001, + "input_cost_per_token": 1e-5, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.00003, + "output_cost_per_token": 3e-5, "supports_pdf_input": true, "supports_prompt_caching": true, "supports_system_messages": true, @@ -14497,7 +13042,7 @@ "supports_vision": true }, "gpt-4-32k": { - "input_cost_per_token": 0.00006, + "input_cost_per_token": 6e-5, "litellm_provider": "openai", "max_input_tokens": 32768, "max_output_tokens": 4096, @@ -14509,7 +13054,7 @@ "supports_tool_choice": true }, "gpt-4-32k-0314": { - "input_cost_per_token": 0.00006, + "input_cost_per_token": 6e-5, "litellm_provider": "openai", "max_input_tokens": 32768, "max_output_tokens": 4096, @@ -14521,7 +13066,7 @@ "supports_tool_choice": true }, "gpt-4-32k-0613": { - "input_cost_per_token": 0.00006, + "input_cost_per_token": 6e-5, "litellm_provider": "openai", "max_input_tokens": 32768, "max_output_tokens": 4096, @@ -14533,13 +13078,13 @@ "supports_tool_choice": true }, "gpt-4-turbo": { - "input_cost_per_token": 0.00001, + "input_cost_per_token": 1e-5, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.00003, + "output_cost_per_token": 3e-5, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_pdf_input": true, @@ -14549,13 +13094,13 @@ "supports_vision": true }, "gpt-4-turbo-2024-04-09": { - "input_cost_per_token": 0.00001, + "input_cost_per_token": 1e-5, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.00003, + "output_cost_per_token": 3e-5, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_pdf_input": true, @@ -14565,13 +13110,13 @@ "supports_vision": true }, "gpt-4-turbo-preview": { - "input_cost_per_token": 0.00001, + "input_cost_per_token": 1e-5, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.00003, + "output_cost_per_token": 3e-5, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_pdf_input": true, @@ -14581,13 +13126,13 @@ }, "gpt-4-vision-preview": { "deprecation_date": "2024-12-06", - "input_cost_per_token": 0.00001, + "input_cost_per_token": 1e-5, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.00003, + "output_cost_per_token": 3e-5, "supports_pdf_input": true, "supports_prompt_caching": true, "supports_system_messages": true, @@ -14597,29 +13142,20 @@ "gpt-4.1": { "cache_read_input_token_cost": 5e-7, "cache_read_input_token_cost_priority": 8.75e-7, - "input_cost_per_token": 0.000002, - "input_cost_per_token_batches": 0.000001, - "input_cost_per_token_priority": 0.0000035, + "input_cost_per_token": 2e-6, + "input_cost_per_token_batches": 1e-6, + "input_cost_per_token_priority": 3.5e-6, "litellm_provider": "openai", "max_input_tokens": 1047576, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0.000008, - "output_cost_per_token_batches": 0.000004, - "output_cost_per_token_priority": 0.000014, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 8e-6, + "output_cost_per_token_batches": 4e-6, + "output_cost_per_token_priority": 1.4e-5, + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -14633,27 +13169,18 @@ }, "gpt-4.1-2025-04-14": { "cache_read_input_token_cost": 5e-7, - "input_cost_per_token": 0.000002, - "input_cost_per_token_batches": 0.000001, + "input_cost_per_token": 2e-6, + "input_cost_per_token_batches": 1e-6, "litellm_provider": "openai", "max_input_tokens": 1047576, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0.000008, - "output_cost_per_token_batches": 0.000004, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 8e-6, + "output_cost_per_token_batches": 4e-6, + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -14676,21 +13203,12 @@ "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0.0000016, + "output_cost_per_token": 1.6e-6, "output_cost_per_token_batches": 8e-7, - "output_cost_per_token_priority": 0.0000028, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token_priority": 2.8e-6, + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -14711,20 +13229,11 @@ "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0.0000016, + "output_cost_per_token": 1.6e-6, "output_cost_per_token_batches": 8e-7, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -14750,18 +13259,9 @@ "output_cost_per_token": 4e-7, "output_cost_per_token_batches": 2e-7, "output_cost_per_token_priority": 8e-7, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -14784,18 +13284,9 @@ "mode": "chat", "output_cost_per_token": 4e-7, "output_cost_per_token_batches": 2e-7, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -14808,16 +13299,16 @@ "supports_vision": true }, "gpt-4.5-preview": { - "cache_read_input_token_cost": 0.0000375, - "input_cost_per_token": 0.000075, - "input_cost_per_token_batches": 0.0000375, + "cache_read_input_token_cost": 3.75e-5, + "input_cost_per_token": 7.5e-5, + "input_cost_per_token_batches": 3.75e-5, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", "output_cost_per_token": 0.00015, - "output_cost_per_token_batches": 0.000075, + "output_cost_per_token_batches": 7.5e-5, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_pdf_input": true, @@ -14828,17 +13319,17 @@ "supports_vision": true }, "gpt-4.5-preview-2025-02-27": { - "cache_read_input_token_cost": 0.0000375, + "cache_read_input_token_cost": 3.75e-5, "deprecation_date": "2025-07-14", - "input_cost_per_token": 0.000075, - "input_cost_per_token_batches": 0.0000375, + "input_cost_per_token": 7.5e-5, + "input_cost_per_token_batches": 3.75e-5, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", "output_cost_per_token": 0.00015, - "output_cost_per_token_batches": 0.000075, + "output_cost_per_token_batches": 7.5e-5, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_pdf_input": true, @@ -14849,19 +13340,19 @@ "supports_vision": true }, "gpt-4o": { - "cache_read_input_token_cost": 0.00000125, - "cache_read_input_token_cost_priority": 0.000002125, - "input_cost_per_token": 0.0000025, - "input_cost_per_token_batches": 0.00000125, - "input_cost_per_token_priority": 0.00000425, + "cache_read_input_token_cost": 1.25e-6, + "cache_read_input_token_cost_priority": 2.125e-6, + "input_cost_per_token": 2.5e-6, + "input_cost_per_token_batches": 1.25e-6, + "input_cost_per_token_priority": 4.25e-6, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 0.00001, - "output_cost_per_token_batches": 0.000005, - "output_cost_per_token_priority": 0.000017, + "output_cost_per_token": 1e-5, + "output_cost_per_token_batches": 5e-6, + "output_cost_per_token_priority": 1.7e-5, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_pdf_input": true, @@ -14873,17 +13364,17 @@ "supports_vision": true }, "gpt-4o-2024-05-13": { - "input_cost_per_token": 0.000005, - "input_cost_per_token_batches": 0.0000025, - "input_cost_per_token_priority": 0.00000875, + "input_cost_per_token": 5e-6, + "input_cost_per_token_batches": 2.5e-6, + "input_cost_per_token_priority": 8.75e-6, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.000015, - "output_cost_per_token_batches": 0.0000075, - "output_cost_per_token_priority": 0.00002625, + "output_cost_per_token": 1.5e-5, + "output_cost_per_token_batches": 7.5e-6, + "output_cost_per_token_priority": 2.625e-5, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_pdf_input": true, @@ -14893,16 +13384,16 @@ "supports_vision": true }, "gpt-4o-2024-08-06": { - "cache_read_input_token_cost": 0.00000125, - "input_cost_per_token": 0.0000025, - "input_cost_per_token_batches": 0.00000125, + "cache_read_input_token_cost": 1.25e-6, + "input_cost_per_token": 2.5e-6, + "input_cost_per_token_batches": 1.25e-6, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 0.00001, - "output_cost_per_token_batches": 0.000005, + "output_cost_per_token": 1e-5, + "output_cost_per_token_batches": 5e-6, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_pdf_input": true, @@ -14914,16 +13405,16 @@ "supports_vision": true }, "gpt-4o-2024-11-20": { - "cache_read_input_token_cost": 0.00000125, - "input_cost_per_token": 0.0000025, - "input_cost_per_token_batches": 0.00000125, + "cache_read_input_token_cost": 1.25e-6, + "input_cost_per_token": 2.5e-6, + "input_cost_per_token_batches": 1.25e-6, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 0.00001, - "output_cost_per_token_batches": 0.000005, + "output_cost_per_token": 1e-5, + "output_cost_per_token_batches": 5e-6, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_pdf_input": true, @@ -14936,14 +13427,14 @@ }, "gpt-4o-audio-preview": { "input_cost_per_audio_token": 0.0001, - "input_cost_per_token": 0.0000025, + "input_cost_per_token": 2.5e-6, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", "output_cost_per_audio_token": 0.0002, - "output_cost_per_token": 0.00001, + "output_cost_per_token": 1e-5, "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -14953,14 +13444,14 @@ }, "gpt-4o-audio-preview-2024-10-01": { "input_cost_per_audio_token": 0.0001, - "input_cost_per_token": 0.0000025, + "input_cost_per_token": 2.5e-6, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", "output_cost_per_audio_token": 0.0002, - "output_cost_per_token": 0.00001, + "output_cost_per_token": 1e-5, "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -14969,15 +13460,15 @@ "supports_tool_choice": true }, "gpt-4o-audio-preview-2024-12-17": { - "input_cost_per_audio_token": 0.00004, - "input_cost_per_token": 0.0000025, + "input_cost_per_audio_token": 4e-5, + "input_cost_per_token": 2.5e-6, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_audio_token": 0.00008, - "output_cost_per_token": 0.00001, + "output_cost_per_audio_token": 8e-5, + "output_cost_per_token": 1e-5, "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -14986,15 +13477,15 @@ "supports_tool_choice": true }, "gpt-4o-audio-preview-2025-06-03": { - "input_cost_per_audio_token": 0.00004, - "input_cost_per_token": 0.0000025, + "input_cost_per_audio_token": 4e-5, + "input_cost_per_token": 2.5e-6, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_audio_token": 0.00008, - "output_cost_per_token": 0.00001, + "output_cost_per_audio_token": 8e-5, + "output_cost_per_token": 1e-5, "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -15015,7 +13506,7 @@ "mode": "chat", "output_cost_per_token": 6e-7, "output_cost_per_token_batches": 3e-7, - "output_cost_per_token_priority": 0.000001, + "output_cost_per_token_priority": 1e-6, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_pdf_input": true, @@ -15053,14 +13544,14 @@ "supports_vision": true }, "gpt-4o-mini-audio-preview": { - "input_cost_per_audio_token": 0.00001, + "input_cost_per_audio_token": 1e-5, "input_cost_per_token": 1.5e-7, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_audio_token": 0.00002, + "output_cost_per_audio_token": 2e-5, "output_cost_per_token": 6e-7, "supports_audio_input": true, "supports_audio_output": true, @@ -15070,14 +13561,14 @@ "supports_tool_choice": true }, "gpt-4o-mini-audio-preview-2024-12-17": { - "input_cost_per_audio_token": 0.00001, + "input_cost_per_audio_token": 1e-5, "input_cost_per_token": 1.5e-7, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_audio_token": 0.00002, + "output_cost_per_audio_token": 2e-5, "output_cost_per_token": 6e-7, "supports_audio_input": true, "supports_audio_output": true, @@ -15089,15 +13580,15 @@ "gpt-4o-mini-realtime-preview": { "cache_creation_input_audio_token_cost": 3e-7, "cache_read_input_token_cost": 3e-7, - "input_cost_per_audio_token": 0.00001, + "input_cost_per_audio_token": 1e-5, "input_cost_per_token": 6e-7, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_audio_token": 0.00002, - "output_cost_per_token": 0.0000024, + "output_cost_per_audio_token": 2e-5, + "output_cost_per_token": 2.4e-6, "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -15108,15 +13599,15 @@ "gpt-4o-mini-realtime-preview-2024-12-17": { "cache_creation_input_audio_token_cost": 3e-7, "cache_read_input_token_cost": 3e-7, - "input_cost_per_audio_token": 0.00001, + "input_cost_per_audio_token": 1e-5, "input_cost_per_token": 6e-7, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_audio_token": 0.00002, - "output_cost_per_token": 0.0000024, + "output_cost_per_audio_token": 2e-5, + "output_cost_per_token": 2.4e-6, "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -15171,46 +13662,37 @@ "supports_vision": true }, "gpt-4o-mini-transcribe": { - "input_cost_per_audio_token": 0.000003, - "input_cost_per_token": 0.00000125, + "input_cost_per_audio_token": 3e-6, + "input_cost_per_token": 1.25e-6, "litellm_provider": "openai", "max_input_tokens": 16000, "max_output_tokens": 2000, "mode": "audio_transcription", - "output_cost_per_token": 0.000005, - "supported_endpoints": [ - "/v1/audio/transcriptions" - ] + "output_cost_per_token": 5e-6, + "supported_endpoints": ["/v1/audio/transcriptions"] }, "gpt-4o-mini-tts": { - "input_cost_per_token": 0.0000025, + "input_cost_per_token": 2.5e-6, "litellm_provider": "openai", "mode": "audio_speech", - "output_cost_per_audio_token": 0.000012, + "output_cost_per_audio_token": 1.2e-5, "output_cost_per_second": 0.00025, - "output_cost_per_token": 0.00001, - "supported_endpoints": [ - "/v1/audio/speech" - ], - "supported_modalities": [ - "text", - "audio" - ], - "supported_output_modalities": [ - "audio" - ] + "output_cost_per_token": 1e-5, + "supported_endpoints": ["/v1/audio/speech"], + "supported_modalities": ["text", "audio"], + "supported_output_modalities": ["audio"] }, "gpt-4o-realtime-preview": { - "cache_read_input_token_cost": 0.0000025, - "input_cost_per_audio_token": 0.00004, - "input_cost_per_token": 0.000005, + "cache_read_input_token_cost": 2.5e-6, + "input_cost_per_audio_token": 4e-5, + "input_cost_per_token": 5e-6, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_audio_token": 0.00008, - "output_cost_per_token": 0.00002, + "output_cost_per_audio_token": 8e-5, + "output_cost_per_token": 2e-5, "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -15219,17 +13701,17 @@ "supports_tool_choice": true }, "gpt-4o-realtime-preview-2024-10-01": { - "cache_creation_input_audio_token_cost": 0.00002, - "cache_read_input_token_cost": 0.0000025, + "cache_creation_input_audio_token_cost": 2e-5, + "cache_read_input_token_cost": 2.5e-6, "input_cost_per_audio_token": 0.0001, - "input_cost_per_token": 0.000005, + "input_cost_per_token": 5e-6, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", "output_cost_per_audio_token": 0.0002, - "output_cost_per_token": 0.00002, + "output_cost_per_token": 2e-5, "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -15238,16 +13720,16 @@ "supports_tool_choice": true }, "gpt-4o-realtime-preview-2024-12-17": { - "cache_read_input_token_cost": 0.0000025, - "input_cost_per_audio_token": 0.00004, - "input_cost_per_token": 0.000005, + "cache_read_input_token_cost": 2.5e-6, + "input_cost_per_audio_token": 4e-5, + "input_cost_per_token": 5e-6, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_audio_token": 0.00008, - "output_cost_per_token": 0.00002, + "output_cost_per_audio_token": 8e-5, + "output_cost_per_token": 2e-5, "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -15256,16 +13738,16 @@ "supports_tool_choice": true }, "gpt-4o-realtime-preview-2025-06-03": { - "cache_read_input_token_cost": 0.0000025, - "input_cost_per_audio_token": 0.00004, - "input_cost_per_token": 0.000005, + "cache_read_input_token_cost": 2.5e-6, + "input_cost_per_audio_token": 4e-5, + "input_cost_per_token": 5e-6, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_audio_token": 0.00008, - "output_cost_per_token": 0.00002, + "output_cost_per_audio_token": 8e-5, + "output_cost_per_token": 2e-5, "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -15274,16 +13756,16 @@ "supports_tool_choice": true }, "gpt-4o-search-preview": { - "cache_read_input_token_cost": 0.00000125, - "input_cost_per_token": 0.0000025, - "input_cost_per_token_batches": 0.00000125, + "cache_read_input_token_cost": 1.25e-6, + "input_cost_per_token": 2.5e-6, + "input_cost_per_token_batches": 1.25e-6, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 0.00001, - "output_cost_per_token_batches": 0.000005, + "output_cost_per_token": 1e-5, + "output_cost_per_token_batches": 5e-6, "search_context_cost_per_query": { "search_context_size_high": 0.05, "search_context_size_low": 0.03, @@ -15300,16 +13782,16 @@ "supports_web_search": true }, "gpt-4o-search-preview-2025-03-11": { - "cache_read_input_token_cost": 0.00000125, - "input_cost_per_token": 0.0000025, - "input_cost_per_token_batches": 0.00000125, + "cache_read_input_token_cost": 1.25e-6, + "input_cost_per_token": 2.5e-6, + "input_cost_per_token_batches": 1.25e-6, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 0.00001, - "output_cost_per_token_batches": 0.000005, + "output_cost_per_token": 1e-5, + "output_cost_per_token_batches": 5e-6, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_pdf_input": true, @@ -15320,44 +13802,33 @@ "supports_vision": true }, "gpt-4o-transcribe": { - "input_cost_per_audio_token": 0.000006, - "input_cost_per_token": 0.0000025, + "input_cost_per_audio_token": 6e-6, + "input_cost_per_token": 2.5e-6, "litellm_provider": "openai", "max_input_tokens": 16000, "max_output_tokens": 2000, "mode": "audio_transcription", - "output_cost_per_token": 0.00001, - "supported_endpoints": [ - "/v1/audio/transcriptions" - ] + "output_cost_per_token": 1e-5, + "supported_endpoints": ["/v1/audio/transcriptions"] }, "gpt-5": { "cache_read_input_token_cost": 1.25e-7, "cache_read_input_token_cost_flex": 6.25e-8, "cache_read_input_token_cost_priority": 2.5e-7, - "input_cost_per_token": 0.00000125, + "input_cost_per_token": 1.25e-6, "input_cost_per_token_flex": 6.25e-7, - "input_cost_per_token_priority": 0.0000025, + "input_cost_per_token_priority": 2.5e-6, "litellm_provider": "openai", "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.00001, - "output_cost_per_token_flex": 0.000005, - "output_cost_per_token_priority": 0.00002, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 1e-5, + "output_cost_per_token_flex": 5e-6, + "output_cost_per_token_priority": 2e-5, + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -15373,27 +13844,18 @@ "gpt-5.1": { "cache_read_input_token_cost": 1.25e-7, "cache_read_input_token_cost_priority": 2.5e-7, - "input_cost_per_token": 0.00000125, - "input_cost_per_token_priority": 0.0000025, + "input_cost_per_token": 1.25e-6, + "input_cost_per_token_priority": 2.5e-6, "litellm_provider": "openai", "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.00001, - "output_cost_per_token_priority": 0.00002, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text", - "image" - ], + "output_cost_per_token": 1e-5, + "output_cost_per_token_priority": 2e-5, + "supported_endpoints": ["/v1/chat/completions", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text", "image"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -15409,27 +13871,18 @@ "gpt-5.1-2025-11-13": { "cache_read_input_token_cost": 1.25e-7, "cache_read_input_token_cost_priority": 2.5e-7, - "input_cost_per_token": 0.00000125, - "input_cost_per_token_priority": 0.0000025, + "input_cost_per_token": 1.25e-6, + "input_cost_per_token_priority": 2.5e-6, "litellm_provider": "openai", "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.00001, - "output_cost_per_token_priority": 0.00002, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text", - "image" - ], + "output_cost_per_token": 1e-5, + "output_cost_per_token_priority": 2e-5, + "supported_endpoints": ["/v1/chat/completions", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text", "image"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -15445,27 +13898,18 @@ "gpt-5.1-chat-latest": { "cache_read_input_token_cost": 1.25e-7, "cache_read_input_token_cost_priority": 2.5e-7, - "input_cost_per_token": 0.00000125, - "input_cost_per_token_priority": 0.0000025, + "input_cost_per_token": 1.25e-6, + "input_cost_per_token_priority": 2.5e-6, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 0.00001, - "output_cost_per_token_priority": 0.00002, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text", - "image" - ], + "output_cost_per_token": 1e-5, + "output_cost_per_token_priority": 2e-5, + "supported_endpoints": ["/v1/chat/completions", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text", "image"], "supports_function_calling": false, "supports_native_streaming": true, "supports_parallel_function_calling": false, @@ -15478,26 +13922,18 @@ "supports_vision": true }, "gpt-5-pro": { - "input_cost_per_token": 0.000015, - "input_cost_per_token_batches": 0.0000075, + "input_cost_per_token": 1.5e-5, + "input_cost_per_token_batches": 7.5e-6, "litellm_provider": "openai", "max_input_tokens": 400000, "max_output_tokens": 272000, "max_tokens": 272000, "mode": "responses", - "output_cost_per_token": 0.00012, - "output_cost_per_token_batches": 0.00006, - "supported_endpoints": [ - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 1.2e-4, + "output_cost_per_token_batches": 6e-5, + "supported_endpoints": ["/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": false, "supports_parallel_function_calling": true, @@ -15511,26 +13947,18 @@ "supports_web_search": true }, "gpt-5-pro-2025-10-06": { - "input_cost_per_token": 0.000015, - "input_cost_per_token_batches": 0.0000075, + "input_cost_per_token": 1.5e-5, + "input_cost_per_token_batches": 7.5e-6, "litellm_provider": "openai", "max_input_tokens": 400000, "max_output_tokens": 272000, "max_tokens": 272000, "mode": "responses", - "output_cost_per_token": 0.00012, - "output_cost_per_token_batches": 0.00006, - "supported_endpoints": [ - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 1.2e-4, + "output_cost_per_token_batches": 6e-5, + "supported_endpoints": ["/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": false, "supports_parallel_function_calling": true, @@ -15547,29 +13975,20 @@ "cache_read_input_token_cost": 1.25e-7, "cache_read_input_token_cost_flex": 6.25e-8, "cache_read_input_token_cost_priority": 2.5e-7, - "input_cost_per_token": 0.00000125, + "input_cost_per_token": 1.25e-6, "input_cost_per_token_flex": 6.25e-7, - "input_cost_per_token_priority": 0.0000025, + "input_cost_per_token_priority": 2.5e-6, "litellm_provider": "openai", "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.00001, - "output_cost_per_token_flex": 0.000005, - "output_cost_per_token_priority": 0.00002, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 1e-5, + "output_cost_per_token_flex": 5e-6, + "output_cost_per_token_priority": 2e-5, + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -15584,25 +14003,16 @@ }, "gpt-5-chat": { "cache_read_input_token_cost": 1.25e-7, - "input_cost_per_token": 0.00000125, + "input_cost_per_token": 1.25e-6, "litellm_provider": "openai", "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.00001, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 1e-5, + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": false, "supports_native_streaming": true, "supports_parallel_function_calling": false, @@ -15616,25 +14026,16 @@ }, "gpt-5-chat-latest": { "cache_read_input_token_cost": 1.25e-7, - "input_cost_per_token": 0.00000125, + "input_cost_per_token": 1.25e-6, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 0.00001, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 1e-5, + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": false, "supports_native_streaming": true, "supports_parallel_function_calling": false, @@ -15648,23 +14049,16 @@ }, "gpt-5-codex": { "cache_read_input_token_cost": 1.25e-7, - "input_cost_per_token": 0.00000125, + "input_cost_per_token": 1.25e-6, "litellm_provider": "openai", "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "responses", - "output_cost_per_token": 0.00001, - "supported_endpoints": [ - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 1e-5, + "supported_endpoints": ["/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -15679,25 +14073,18 @@ "gpt-5.1-codex": { "cache_read_input_token_cost": 1.25e-7, "cache_read_input_token_cost_priority": 2.5e-7, - "input_cost_per_token": 0.00000125, - "input_cost_per_token_priority": 0.0000025, + "input_cost_per_token": 1.25e-6, + "input_cost_per_token_priority": 2.5e-6, "litellm_provider": "openai", "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "responses", - "output_cost_per_token": 0.00001, - "output_cost_per_token_priority": 0.00002, - "supported_endpoints": [ - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 1e-5, + "output_cost_per_token_priority": 2e-5, + "supported_endpoints": ["/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -15719,18 +14106,11 @@ "max_output_tokens": 128000, "max_tokens": 128000, "mode": "responses", - "output_cost_per_token": 0.000002, - "output_cost_per_token_priority": 0.0000036, - "supported_endpoints": [ - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 2e-6, + "output_cost_per_token_priority": 3.6e-6, + "supported_endpoints": ["/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -15754,21 +14134,12 @@ "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.000002, - "output_cost_per_token_flex": 0.000001, - "output_cost_per_token_priority": 0.0000036, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 2e-6, + "output_cost_per_token_flex": 1e-6, + "output_cost_per_token_priority": 3.6e-6, + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -15793,21 +14164,12 @@ "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.000002, - "output_cost_per_token_flex": 0.000001, - "output_cost_per_token_priority": 0.0000036, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 2e-6, + "output_cost_per_token_flex": 1e-6, + "output_cost_per_token_priority": 3.6e-6, + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -15825,7 +14187,7 @@ "cache_read_input_token_cost_flex": 2.5e-9, "input_cost_per_token": 5e-8, "input_cost_per_token_flex": 2.5e-8, - "input_cost_per_token_priority": 0.0000025, + "input_cost_per_token_priority": 2.5e-6, "litellm_provider": "openai", "max_input_tokens": 272000, "max_output_tokens": 128000, @@ -15833,18 +14195,9 @@ "mode": "chat", "output_cost_per_token": 4e-7, "output_cost_per_token_flex": 2e-7, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -15868,18 +14221,9 @@ "mode": "chat", "output_cost_per_token": 4e-7, "output_cost_per_token_flex": 2e-7, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -15895,49 +14239,35 @@ "input_cost_per_pixel": 4.0054321e-8, "litellm_provider": "openai", "mode": "image_generation", - "output_cost_per_pixel": 0, - "supported_endpoints": [ - "/v1/images/generations" - ] + "output_cost_per_pixel": 0.0, + "supported_endpoints": ["/v1/images/generations"] }, "gpt-image-1-mini": { "cache_read_input_image_token_cost": 2.5e-7, "cache_read_input_token_cost": 2e-7, - "input_cost_per_image_token": 0.0000025, - "input_cost_per_token": 0.000002, + "input_cost_per_image_token": 2.5e-6, + "input_cost_per_token": 2e-6, "litellm_provider": "openai", "mode": "chat", - "output_cost_per_image_token": 0.000008, - "supported_endpoints": [ - "/v1/images/generations", - "/v1/images/edits" - ] + "output_cost_per_image_token": 8e-6, + "supported_endpoints": ["/v1/images/generations", "/v1/images/edits"] }, "gpt-realtime": { "cache_creation_input_audio_token_cost": 4e-7, "cache_read_input_token_cost": 4e-7, - "input_cost_per_audio_token": 0.000032, - "input_cost_per_image": 0.000005, - "input_cost_per_token": 0.000004, + "input_cost_per_audio_token": 3.2e-5, + "input_cost_per_image": 5e-6, + "input_cost_per_token": 4e-6, "litellm_provider": "openai", "max_input_tokens": 32000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_audio_token": 0.000064, - "output_cost_per_token": 0.000016, - "supported_endpoints": [ - "/v1/realtime" - ], - "supported_modalities": [ - "text", - "image", - "audio" - ], - "supported_output_modalities": [ - "text", - "audio" - ], + "output_cost_per_audio_token": 6.4e-5, + "output_cost_per_token": 1.6e-5, + "supported_endpoints": ["/v1/realtime"], + "supported_modalities": ["text", "image", "audio"], + "supported_output_modalities": ["text", "audio"], "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -15948,27 +14278,18 @@ "gpt-realtime-mini": { "cache_creation_input_audio_token_cost": 3e-7, "cache_read_input_audio_token_cost": 3e-7, - "input_cost_per_audio_token": 0.00001, + "input_cost_per_audio_token": 1e-5, "input_cost_per_token": 6e-7, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_audio_token": 0.00002, - "output_cost_per_token": 0.0000024, - "supported_endpoints": [ - "/v1/realtime" - ], - "supported_modalities": [ - "text", - "image", - "audio" - ], - "supported_output_modalities": [ - "text", - "audio" - ], + "output_cost_per_audio_token": 2e-5, + "output_cost_per_token": 2.4e-6, + "supported_endpoints": ["/v1/realtime"], + "supported_modalities": ["text", "image", "audio"], + "supported_output_modalities": ["text", "audio"], "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -15979,28 +14300,19 @@ "gpt-realtime-2025-08-28": { "cache_creation_input_audio_token_cost": 4e-7, "cache_read_input_token_cost": 4e-7, - "input_cost_per_audio_token": 0.000032, - "input_cost_per_image": 0.000005, - "input_cost_per_token": 0.000004, + "input_cost_per_audio_token": 3.2e-5, + "input_cost_per_image": 5e-6, + "input_cost_per_token": 4e-6, "litellm_provider": "openai", "max_input_tokens": 32000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_audio_token": 0.000064, - "output_cost_per_token": 0.000016, - "supported_endpoints": [ - "/v1/realtime" - ], - "supported_modalities": [ - "text", - "image", - "audio" - ], - "supported_output_modalities": [ - "text", - "audio" - ], + "output_cost_per_audio_token": 6.4e-5, + "output_cost_per_token": 1.6e-5, + "supported_endpoints": ["/v1/realtime"], + "supported_modalities": ["text", "image", "audio"], + "supported_output_modalities": ["text", "audio"], "supports_audio_input": true, "supports_audio_output": true, "supports_function_calling": true, @@ -16012,26 +14324,18 @@ "litellm_provider": "gradient_ai", "max_tokens": 2048, "mode": "chat", - "supported_endpoints": [ - "/v1/chat/completions" - ], - "supported_modalities": [ - "text" - ], + "supported_endpoints": ["/v1/chat/completions"], + "supported_modalities": ["text"], "supports_tool_choice": false }, "gradient_ai/anthropic-claude-3-opus": { - "input_cost_per_token": 0.000015, + "input_cost_per_token": 1.5e-5, "litellm_provider": "gradient_ai", "max_tokens": 1024, "mode": "chat", - "output_cost_per_token": 0.000075, - "supported_endpoints": [ - "/v1/chat/completions" - ], - "supported_modalities": [ - "text" - ], + "output_cost_per_token": 7.5e-5, + "supported_endpoints": ["/v1/chat/completions"], + "supported_modalities": ["text"], "supports_tool_choice": false }, "gradient_ai/anthropic-claude-3.5-haiku": { @@ -16039,41 +14343,29 @@ "litellm_provider": "gradient_ai", "max_tokens": 1024, "mode": "chat", - "output_cost_per_token": 0.000004, - "supported_endpoints": [ - "/v1/chat/completions" - ], - "supported_modalities": [ - "text" - ], + "output_cost_per_token": 4e-6, + "supported_endpoints": ["/v1/chat/completions"], + "supported_modalities": ["text"], "supports_tool_choice": false }, "gradient_ai/anthropic-claude-3.5-sonnet": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "gradient_ai", "max_tokens": 1024, "mode": "chat", - "output_cost_per_token": 0.000015, - "supported_endpoints": [ - "/v1/chat/completions" - ], - "supported_modalities": [ - "text" - ], + "output_cost_per_token": 1.5e-5, + "supported_endpoints": ["/v1/chat/completions"], + "supported_modalities": ["text"], "supports_tool_choice": false }, "gradient_ai/anthropic-claude-3.7-sonnet": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "gradient_ai", "max_tokens": 1024, "mode": "chat", - "output_cost_per_token": 0.000015, - "supported_endpoints": [ - "/v1/chat/completions" - ], - "supported_modalities": [ - "text" - ], + "output_cost_per_token": 1.5e-5, + "supported_endpoints": ["/v1/chat/completions"], + "supported_modalities": ["text"], "supports_tool_choice": false }, "gradient_ai/deepseek-r1-distill-llama-70b": { @@ -16082,12 +14374,8 @@ "max_tokens": 8000, "mode": "chat", "output_cost_per_token": 9.9e-7, - "supported_endpoints": [ - "/v1/chat/completions" - ], - "supported_modalities": [ - "text" - ], + "supported_endpoints": ["/v1/chat/completions"], + "supported_modalities": ["text"], "supports_tool_choice": false }, "gradient_ai/llama3-8b-instruct": { @@ -16096,12 +14384,8 @@ "max_tokens": 512, "mode": "chat", "output_cost_per_token": 2e-7, - "supported_endpoints": [ - "/v1/chat/completions" - ], - "supported_modalities": [ - "text" - ], + "supported_endpoints": ["/v1/chat/completions"], + "supported_modalities": ["text"], "supports_tool_choice": false }, "gradient_ai/llama3.3-70b-instruct": { @@ -16110,12 +14394,8 @@ "max_tokens": 2048, "mode": "chat", "output_cost_per_token": 6.5e-7, - "supported_endpoints": [ - "/v1/chat/completions" - ], - "supported_modalities": [ - "text" - ], + "supported_endpoints": ["/v1/chat/completions"], + "supported_modalities": ["text"], "supports_tool_choice": false }, "gradient_ai/mistral-nemo-instruct-2407": { @@ -16124,64 +14404,44 @@ "max_tokens": 512, "mode": "chat", "output_cost_per_token": 3e-7, - "supported_endpoints": [ - "/v1/chat/completions" - ], - "supported_modalities": [ - "text" - ], + "supported_endpoints": ["/v1/chat/completions"], + "supported_modalities": ["text"], "supports_tool_choice": false }, "gradient_ai/openai-gpt-4o": { "litellm_provider": "gradient_ai", "max_tokens": 16384, "mode": "chat", - "supported_endpoints": [ - "/v1/chat/completions" - ], - "supported_modalities": [ - "text" - ], + "supported_endpoints": ["/v1/chat/completions"], + "supported_modalities": ["text"], "supports_tool_choice": false }, "gradient_ai/openai-gpt-4o-mini": { "litellm_provider": "gradient_ai", "max_tokens": 16384, "mode": "chat", - "supported_endpoints": [ - "/v1/chat/completions" - ], - "supported_modalities": [ - "text" - ], + "supported_endpoints": ["/v1/chat/completions"], + "supported_modalities": ["text"], "supports_tool_choice": false }, "gradient_ai/openai-o3": { - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "gradient_ai", "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 0.000008, - "supported_endpoints": [ - "/v1/chat/completions" - ], - "supported_modalities": [ - "text" - ], + "output_cost_per_token": 8e-6, + "supported_endpoints": ["/v1/chat/completions"], + "supported_modalities": ["text"], "supports_tool_choice": false }, "gradient_ai/openai-o3-mini": { - "input_cost_per_token": 0.0000011, + "input_cost_per_token": 1.1e-6, "litellm_provider": "gradient_ai", "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 0.0000044, - "supported_endpoints": [ - "/v1/chat/completions" - ], - "supported_modalities": [ - "text" - ], + "output_cost_per_token": 4.4e-6, + "supported_endpoints": ["/v1/chat/completions"], + "supported_modalities": ["text"], "supports_tool_choice": false }, "lemonade/Qwen3-Coder-30B-A3B-Instruct-GGUF": { @@ -16258,10 +14518,10 @@ "supports_tool_choice": true }, "groq/distil-whisper-large-v3-en": { - "input_cost_per_second": 0.00000556, + "input_cost_per_second": 5.56e-6, "litellm_provider": "groq", "mode": "audio_transcription", - "output_cost_per_second": 0 + "output_cost_per_second": 0.0 }, "groq/gemma-7b-it": { "deprecation_date": "2024-12-18", @@ -16522,21 +14782,21 @@ "supports_tool_choice": true }, "groq/moonshotai/kimi-k2-instruct": { - "input_cost_per_token": 0.000001, + "input_cost_per_token": 1e-6, "litellm_provider": "groq", "max_input_tokens": 131072, "max_output_tokens": 16384, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0.000003, + "output_cost_per_token": 3e-6, "supports_function_calling": true, "supports_response_schema": true, "supports_tool_choice": true }, "groq/moonshotai/kimi-k2-instruct-0905": { - "input_cost_per_token": 0.000001, - "output_cost_per_token": 0.000003, - "cache_read_input_token_cost": 5e-7, + "input_cost_per_token": 1e-6, + "output_cost_per_token": 3e-6, + "cache_read_input_token_cost": 0.5e-6, "litellm_provider": "groq", "max_input_tokens": 262144, "max_output_tokens": 16384, @@ -16577,7 +14837,7 @@ "supports_web_search": true }, "groq/playai-tts": { - "input_cost_per_character": 0.00005, + "input_cost_per_character": 5e-5, "litellm_provider": "groq", "max_input_tokens": 10000, "max_output_tokens": 10000, @@ -16598,34 +14858,34 @@ "supports_tool_choice": true }, "groq/whisper-large-v3": { - "input_cost_per_second": 0.00003083, + "input_cost_per_second": 3.083e-5, "litellm_provider": "groq", "mode": "audio_transcription", - "output_cost_per_second": 0 + "output_cost_per_second": 0.0 }, "groq/whisper-large-v3-turbo": { - "input_cost_per_second": 0.00001111, + "input_cost_per_second": 1.111e-5, "litellm_provider": "groq", "mode": "audio_transcription", - "output_cost_per_second": 0 + "output_cost_per_second": 0.0 }, "hd/1024-x-1024/dall-e-3": { "input_cost_per_pixel": 7.629e-8, "litellm_provider": "openai", "mode": "image_generation", - "output_cost_per_pixel": 0 + "output_cost_per_pixel": 0.0 }, "hd/1024-x-1792/dall-e-3": { "input_cost_per_pixel": 6.539e-8, "litellm_provider": "openai", "mode": "image_generation", - "output_cost_per_pixel": 0 + "output_cost_per_pixel": 0.0 }, "hd/1792-x-1024/dall-e-3": { "input_cost_per_pixel": 6.539e-8, "litellm_provider": "openai", "mode": "image_generation", - "output_cost_per_pixel": 0 + "output_cost_per_pixel": 0.0 }, "heroku/claude-3-5-haiku": { "litellm_provider": "heroku", @@ -16663,28 +14923,22 @@ "input_cost_per_pixel": 1.59263611e-7, "litellm_provider": "openai", "mode": "image_generation", - "output_cost_per_pixel": 0, - "supported_endpoints": [ - "/v1/images/generations" - ] + "output_cost_per_pixel": 0.0, + "supported_endpoints": ["/v1/images/generations"] }, "high/1024-x-1536/gpt-image-1": { "input_cost_per_pixel": 1.58945719e-7, "litellm_provider": "openai", "mode": "image_generation", - "output_cost_per_pixel": 0, - "supported_endpoints": [ - "/v1/images/generations" - ] + "output_cost_per_pixel": 0.0, + "supported_endpoints": ["/v1/images/generations"] }, "high/1536-x-1024/gpt-image-1": { "input_cost_per_pixel": 1.58945719e-7, "litellm_provider": "openai", "mode": "image_generation", - "output_cost_per_pixel": 0, - "supported_endpoints": [ - "/v1/images/generations" - ] + "output_cost_per_pixel": 0.0, + "supported_endpoints": ["/v1/images/generations"] }, "hyperbolic/NousResearch/Hermes-3-Llama-3.1-70B": { "input_cost_per_token": 1.2e-7, @@ -16739,13 +14993,13 @@ "supports_tool_choice": true }, "hyperbolic/Qwen/Qwen3-235B-A22B": { - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "hyperbolic", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0.000002, + "output_cost_per_token": 2e-6, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_system_messages": true, @@ -16882,44 +15136,44 @@ "supports_tool_choice": true }, "hyperbolic/moonshotai/Kimi-K2-Instruct": { - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "hyperbolic", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0.000002, + "output_cost_per_token": 2e-6, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_system_messages": true, "supports_tool_choice": true }, "j2-light": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "ai21", "max_input_tokens": 8192, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "completion", - "output_cost_per_token": 0.000003 + "output_cost_per_token": 3e-6 }, "j2-mid": { - "input_cost_per_token": 0.00001, + "input_cost_per_token": 1e-5, "litellm_provider": "ai21", "max_input_tokens": 8192, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "completion", - "output_cost_per_token": 0.00001 + "output_cost_per_token": 1e-5 }, "j2-ultra": { - "input_cost_per_token": 0.000015, + "input_cost_per_token": 1.5e-5, "litellm_provider": "ai21", "max_input_tokens": 8192, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "completion", - "output_cost_per_token": 0.000015 + "output_cost_per_token": 1.5e-5 }, "jamba-1.5": { "input_cost_per_token": 2e-7, @@ -16932,23 +15186,23 @@ "supports_tool_choice": true }, "jamba-1.5-large": { - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "ai21", "max_input_tokens": 256000, "max_output_tokens": 256000, "max_tokens": 256000, "mode": "chat", - "output_cost_per_token": 0.000008, + "output_cost_per_token": 8e-6, "supports_tool_choice": true }, "jamba-1.5-large@001": { - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "ai21", "max_input_tokens": 256000, "max_output_tokens": 256000, "max_tokens": 256000, "mode": "chat", - "output_cost_per_token": 0.000008, + "output_cost_per_token": 8e-6, "supports_tool_choice": true }, "jamba-1.5-mini": { @@ -16972,23 +15226,23 @@ "supports_tool_choice": true }, "jamba-large-1.6": { - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "ai21", "max_input_tokens": 256000, "max_output_tokens": 256000, "max_tokens": 256000, "mode": "chat", - "output_cost_per_token": 0.000008, + "output_cost_per_token": 8e-6, "supports_tool_choice": true }, "jamba-large-1.7": { - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "ai21", "max_input_tokens": 256000, "max_output_tokens": 256000, "max_tokens": 256000, "mode": "chat", - "output_cost_per_token": 0.000008, + "output_cost_per_token": 8e-6, "supports_tool_choice": true }, "jamba-mini-1.6": { @@ -17022,19 +15276,19 @@ "output_cost_per_token": 1.8e-8 }, "jp.anthropic.claude-sonnet-4-5-20250929-v1:0": { - "cache_creation_input_token_cost": 0.000004125, + "cache_creation_input_token_cost": 4.125e-6, "cache_read_input_token_cost": 3.3e-7, - "input_cost_per_token": 0.0000033, - "input_cost_per_token_above_200k_tokens": 0.0000066, - "output_cost_per_token_above_200k_tokens": 0.00002475, - "cache_creation_input_token_cost_above_200k_tokens": 0.00000825, + "input_cost_per_token": 3.3e-6, + "input_cost_per_token_above_200k_tokens": 6.6e-6, + "output_cost_per_token_above_200k_tokens": 2.475e-5, + "cache_creation_input_token_cost_above_200k_tokens": 8.25e-6, "cache_read_input_token_cost_above_200k_tokens": 6.6e-7, "litellm_provider": "bedrock_converse", "max_input_tokens": 200000, "max_output_tokens": 64000, "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 0.0000165, + "output_cost_per_token": 1.65e-5, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -17052,15 +15306,15 @@ "tool_use_system_prompt_tokens": 346 }, "jp.anthropic.claude-haiku-4-5-20251001-v1:0": { - "cache_creation_input_token_cost": 0.000001375, + "cache_creation_input_token_cost": 1.375e-6, "cache_read_input_token_cost": 1.1e-7, - "input_cost_per_token": 0.0000011, + "input_cost_per_token": 1.1e-6, "litellm_provider": "bedrock_converse", "max_input_tokens": 200000, "max_output_tokens": 64000, "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 0.0000055, + "output_cost_per_token": 5.5e-6, "source": "https://aws.amazon.com/about-aws/whats-new/2025/10/claude-4-5-haiku-anthropic-amazon-bedrock", "supports_assistant_prefill": true, "supports_computer_use": true, @@ -17342,56 +15596,50 @@ "input_cost_per_pixel": 1.0490417e-8, "litellm_provider": "openai", "mode": "image_generation", - "output_cost_per_pixel": 0, - "supported_endpoints": [ - "/v1/images/generations" - ] + "output_cost_per_pixel": 0.0, + "supported_endpoints": ["/v1/images/generations"] }, "low/1024-x-1536/gpt-image-1": { "input_cost_per_pixel": 1.0172526e-8, "litellm_provider": "openai", "mode": "image_generation", - "output_cost_per_pixel": 0, - "supported_endpoints": [ - "/v1/images/generations" - ] + "output_cost_per_pixel": 0.0, + "supported_endpoints": ["/v1/images/generations"] }, "low/1536-x-1024/gpt-image-1": { "input_cost_per_pixel": 1.0172526e-8, "litellm_provider": "openai", "mode": "image_generation", - "output_cost_per_pixel": 0, - "supported_endpoints": [ - "/v1/images/generations" - ] + "output_cost_per_pixel": 0.0, + "supported_endpoints": ["/v1/images/generations"] }, "luminous-base": { - "input_cost_per_token": 0.00003, + "input_cost_per_token": 3e-5, "litellm_provider": "aleph_alpha", "max_tokens": 2048, "mode": "completion", - "output_cost_per_token": 0.000033 + "output_cost_per_token": 3.3e-5 }, "luminous-base-control": { - "input_cost_per_token": 0.0000375, + "input_cost_per_token": 3.75e-5, "litellm_provider": "aleph_alpha", "max_tokens": 2048, "mode": "chat", - "output_cost_per_token": 0.00004125 + "output_cost_per_token": 4.125e-5 }, "luminous-extended": { - "input_cost_per_token": 0.000045, + "input_cost_per_token": 4.5e-5, "litellm_provider": "aleph_alpha", "max_tokens": 2048, "mode": "completion", - "output_cost_per_token": 0.0000495 + "output_cost_per_token": 4.95e-5 }, "luminous-extended-control": { - "input_cost_per_token": 0.00005625, + "input_cost_per_token": 5.625e-5, "litellm_provider": "aleph_alpha", "max_tokens": 2048, "mode": "chat", - "output_cost_per_token": 0.000061875 + "output_cost_per_token": 6.1875e-5 }, "luminous-supreme": { "input_cost_per_token": 0.000175, @@ -17425,85 +15673,67 @@ "input_cost_per_pixel": 4.0054321e-8, "litellm_provider": "openai", "mode": "image_generation", - "output_cost_per_pixel": 0, - "supported_endpoints": [ - "/v1/images/generations" - ] + "output_cost_per_pixel": 0.0, + "supported_endpoints": ["/v1/images/generations"] }, "medium/1024-x-1536/gpt-image-1": { "input_cost_per_pixel": 4.0054321e-8, "litellm_provider": "openai", "mode": "image_generation", - "output_cost_per_pixel": 0, - "supported_endpoints": [ - "/v1/images/generations" - ] + "output_cost_per_pixel": 0.0, + "supported_endpoints": ["/v1/images/generations"] }, "medium/1536-x-1024/gpt-image-1": { "input_cost_per_pixel": 4.0054321e-8, "litellm_provider": "openai", "mode": "image_generation", - "output_cost_per_pixel": 0, - "supported_endpoints": [ - "/v1/images/generations" - ] + "output_cost_per_pixel": 0.0, + "supported_endpoints": ["/v1/images/generations"] }, "low/1024-x-1024/gpt-image-1-mini": { "input_cost_per_image": 0.005, "litellm_provider": "openai", "mode": "image_generation", - "supported_endpoints": [ - "/v1/images/generations" - ] + "supported_endpoints": ["/v1/images/generations"] }, "low/1024-x-1536/gpt-image-1-mini": { "input_cost_per_image": 0.006, "litellm_provider": "openai", "mode": "image_generation", - "supported_endpoints": [ - "/v1/images/generations" - ] + "supported_endpoints": ["/v1/images/generations"] }, "low/1536-x-1024/gpt-image-1-mini": { "input_cost_per_image": 0.006, "litellm_provider": "openai", "mode": "image_generation", - "supported_endpoints": [ - "/v1/images/generations" - ] + "supported_endpoints": ["/v1/images/generations"] }, "medium/1024-x-1024/gpt-image-1-mini": { "input_cost_per_image": 0.011, "litellm_provider": "openai", "mode": "image_generation", - "supported_endpoints": [ - "/v1/images/generations" - ] + "supported_endpoints": ["/v1/images/generations"] }, "medium/1024-x-1536/gpt-image-1-mini": { "input_cost_per_image": 0.015, "litellm_provider": "openai", "mode": "image_generation", - "supported_endpoints": [ - "/v1/images/generations" - ] + "supported_endpoints": ["/v1/images/generations"] }, "medium/1536-x-1024/gpt-image-1-mini": { "input_cost_per_image": 0.015, "litellm_provider": "openai", "mode": "image_generation", - "supported_endpoints": [ - "/v1/images/generations" - ] + "supported_endpoints": ["/v1/images/generations"] }, "medlm-large": { - "input_cost_per_character": 0.000005, + "input_cost_per_character": 5e-6, "litellm_provider": "vertex_ai-language-models", "max_input_tokens": 8192, "max_output_tokens": 1024, "max_tokens": 1024, "mode": "chat", - "output_cost_per_character": 0.000015, + "output_cost_per_character": 1.5e-5, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models", "supports_tool_choice": true }, @@ -17514,7 +15744,7 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_character": 0.000001, + "output_cost_per_character": 1e-6, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models", "supports_tool_choice": true }, @@ -17525,25 +15755,25 @@ "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.000001 + "output_cost_per_token": 1e-6 }, "meta.llama2-70b-chat-v1": { - "input_cost_per_token": 0.00000195, + "input_cost_per_token": 1.95e-6, "litellm_provider": "bedrock", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.00000256 + "output_cost_per_token": 2.56e-6 }, "meta.llama3-1-405b-instruct-v1:0": { - "input_cost_per_token": 0.00000532, + "input_cost_per_token": 5.32e-6, "litellm_provider": "bedrock", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.000016, + "output_cost_per_token": 1.6e-5, "supports_function_calling": true, "supports_tool_choice": false }, @@ -17604,13 +15834,13 @@ "supports_tool_choice": false }, "meta.llama3-2-90b-instruct-v1:0": { - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "bedrock", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.000002, + "output_cost_per_token": 2e-6, "supports_function_calling": true, "supports_tool_choice": false, "supports_vision": true @@ -17627,13 +15857,13 @@ "supports_tool_choice": false }, "meta.llama3-70b-instruct-v1:0": { - "input_cost_per_token": 0.00000265, + "input_cost_per_token": 2.65e-6, "litellm_provider": "bedrock", "max_input_tokens": 8192, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.0000035 + "output_cost_per_token": 3.5e-6 }, "meta.llama3-8b-instruct-v1:0": { "input_cost_per_token": 3e-7, @@ -17654,14 +15884,8 @@ "mode": "chat", "output_cost_per_token": 9.7e-7, "output_cost_per_token_batches": 4.85e-7, - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text", - "code" - ], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text", "code"], "supports_function_calling": true, "supports_tool_choice": false }, @@ -17675,14 +15899,8 @@ "mode": "chat", "output_cost_per_token": 6.6e-7, "output_cost_per_token_batches": 3.3e-7, - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text", - "code" - ], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text", "code"], "supports_function_calling": true, "supports_tool_choice": false }, @@ -17693,12 +15911,8 @@ "max_tokens": 128000, "mode": "chat", "source": "https://llama.developer.meta.com/docs/models", - "supported_modalities": [ - "text" - ], - "supported_output_modalities": [ - "text" - ], + "supported_modalities": ["text"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_tool_choice": true }, @@ -17709,12 +15923,8 @@ "max_tokens": 128000, "mode": "chat", "source": "https://llama.developer.meta.com/docs/models", - "supported_modalities": [ - "text" - ], - "supported_output_modalities": [ - "text" - ], + "supported_modalities": ["text"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_tool_choice": true }, @@ -17725,13 +15935,8 @@ "max_tokens": 128000, "mode": "chat", "source": "https://llama.developer.meta.com/docs/models", - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_tool_choice": true }, @@ -17742,13 +15947,8 @@ "max_tokens": 128000, "mode": "chat", "source": "https://llama.developer.meta.com/docs/models", - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_tool_choice": true }, @@ -17763,34 +15963,34 @@ "supports_tool_choice": true }, "mistral.mistral-large-2402-v1:0": { - "input_cost_per_token": 0.000008, + "input_cost_per_token": 8e-6, "litellm_provider": "bedrock", "max_input_tokens": 32000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 0.000024, + "output_cost_per_token": 2.4e-5, "supports_function_calling": true }, "mistral.mistral-large-2407-v1:0": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "bedrock", "max_input_tokens": 128000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 0.000009, + "output_cost_per_token": 9e-6, "supports_function_calling": true, "supports_tool_choice": true }, "mistral.mistral-small-2402-v1:0": { - "input_cost_per_token": 0.000001, + "input_cost_per_token": 1e-6, "litellm_provider": "bedrock", "max_input_tokens": 32000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 0.000003, + "output_cost_per_token": 3e-6, "supports_function_calling": true }, "mistral.mixtral-8x7b-instruct-v0:1": { @@ -17804,25 +16004,25 @@ "supports_tool_choice": true }, "mistral/codestral-2405": { - "input_cost_per_token": 0.000001, + "input_cost_per_token": 1e-6, "litellm_provider": "mistral", "max_input_tokens": 32000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 0.000003, + "output_cost_per_token": 3e-6, "supports_assistant_prefill": true, "supports_response_schema": true, "supports_tool_choice": true }, "mistral/codestral-latest": { - "input_cost_per_token": 0.000001, + "input_cost_per_token": 1e-6, "litellm_provider": "mistral", "max_input_tokens": 32000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 0.000003, + "output_cost_per_token": 3e-6, "supports_assistant_prefill": true, "supports_response_schema": true, "supports_tool_choice": true @@ -17846,7 +16046,7 @@ "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.000002, + "output_cost_per_token": 2e-6, "source": "https://mistral.ai/news/devstral", "supports_assistant_prefill": true, "supports_function_calling": true, @@ -17882,13 +16082,13 @@ "supports_tool_choice": true }, "mistral/magistral-medium-2506": { - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "mistral", "max_input_tokens": 40000, "max_output_tokens": 40000, "max_tokens": 40000, "mode": "chat", - "output_cost_per_token": 0.000005, + "output_cost_per_token": 5e-6, "source": "https://mistral.ai/news/magistral", "supports_assistant_prefill": true, "supports_function_calling": true, @@ -17897,13 +16097,13 @@ "supports_tool_choice": true }, "mistral/magistral-medium-2509": { - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "mistral", "max_input_tokens": 40000, "max_output_tokens": 40000, "max_tokens": 40000, "mode": "chat", - "output_cost_per_token": 0.000005, + "output_cost_per_token": 5e-6, "source": "https://mistral.ai/news/magistral", "supports_assistant_prefill": true, "supports_function_calling": true, @@ -17913,32 +16113,28 @@ }, "mistral/mistral-ocr-latest": { "litellm_provider": "mistral", - "ocr_cost_per_page": 0.001, - "annotation_cost_per_page": 0.003, + "ocr_cost_per_page": 1e-3, + "annotation_cost_per_page": 3e-3, "mode": "ocr", - "supported_endpoints": [ - "/v1/ocr" - ], + "supported_endpoints": ["/v1/ocr"], "source": "https://mistral.ai/pricing#api-pricing" }, "mistral/mistral-ocr-2505-completion": { "litellm_provider": "mistral", - "ocr_cost_per_page": 0.001, - "annotation_cost_per_page": 0.003, + "ocr_cost_per_page": 1e-3, + "annotation_cost_per_page": 3e-3, "mode": "ocr", - "supported_endpoints": [ - "/v1/ocr" - ], + "supported_endpoints": ["/v1/ocr"], "source": "https://mistral.ai/pricing#api-pricing" }, "mistral/magistral-medium-latest": { - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "mistral", "max_input_tokens": 40000, "max_output_tokens": 40000, "max_tokens": 40000, "mode": "chat", - "output_cost_per_token": 0.000005, + "output_cost_per_token": 5e-6, "source": "https://mistral.ai/news/magistral", "supports_assistant_prefill": true, "supports_function_calling": true, @@ -17953,7 +16149,7 @@ "max_output_tokens": 40000, "max_tokens": 40000, "mode": "chat", - "output_cost_per_token": 0.0000015, + "output_cost_per_token": 1.5e-6, "source": "https://mistral.ai/pricing#api-pricing", "supports_assistant_prefill": true, "supports_function_calling": true, @@ -17968,7 +16164,7 @@ "max_output_tokens": 40000, "max_tokens": 40000, "mode": "chat", - "output_cost_per_token": 0.0000015, + "output_cost_per_token": 1.5e-6, "source": "https://mistral.ai/pricing#api-pricing", "supports_assistant_prefill": true, "supports_function_calling": true, @@ -17984,91 +16180,91 @@ "mode": "embedding" }, "mistral/codestral-embed": { - "input_cost_per_token": 1.5e-7, + "input_cost_per_token": 0.15e-6, "litellm_provider": "mistral", "max_input_tokens": 8192, "max_tokens": 8192, "mode": "embedding" }, "mistral/codestral-embed-2505": { - "input_cost_per_token": 1.5e-7, + "input_cost_per_token": 0.15e-6, "litellm_provider": "mistral", "max_input_tokens": 8192, "max_tokens": 8192, "mode": "embedding" }, "mistral/mistral-large-2402": { - "input_cost_per_token": 0.000004, + "input_cost_per_token": 4e-6, "litellm_provider": "mistral", "max_input_tokens": 32000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 0.000012, + "output_cost_per_token": 1.2e-5, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_response_schema": true, "supports_tool_choice": true }, "mistral/mistral-large-2407": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "mistral", "max_input_tokens": 128000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.000009, + "output_cost_per_token": 9e-6, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_response_schema": true, "supports_tool_choice": true }, "mistral/mistral-large-2411": { - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "mistral", "max_input_tokens": 128000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.000006, + "output_cost_per_token": 6e-6, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_response_schema": true, "supports_tool_choice": true }, "mistral/mistral-large-latest": { - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "mistral", "max_input_tokens": 128000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.000006, + "output_cost_per_token": 6e-6, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_response_schema": true, "supports_tool_choice": true }, "mistral/mistral-medium": { - "input_cost_per_token": 0.0000027, + "input_cost_per_token": 2.7e-6, "litellm_provider": "mistral", "max_input_tokens": 32000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 0.0000081, + "output_cost_per_token": 8.1e-6, "supports_assistant_prefill": true, "supports_response_schema": true, "supports_tool_choice": true }, "mistral/mistral-medium-2312": { - "input_cost_per_token": 0.0000027, + "input_cost_per_token": 2.7e-6, "litellm_provider": "mistral", "max_input_tokens": 32000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 0.0000081, + "output_cost_per_token": 8.1e-6, "supports_assistant_prefill": true, "supports_response_schema": true, "supports_tool_choice": true @@ -18080,7 +16276,7 @@ "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 0.000002, + "output_cost_per_token": 2e-6, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_response_schema": true, @@ -18093,7 +16289,7 @@ "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 0.000002, + "output_cost_per_token": 2e-6, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_response_schema": true, @@ -18188,13 +16384,13 @@ "supports_tool_choice": true }, "mistral/open-mixtral-8x22b": { - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "mistral", "max_input_tokens": 65336, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 0.000006, + "output_cost_per_token": 6e-6, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_response_schema": true, @@ -18228,13 +16424,13 @@ "supports_vision": true }, "mistral/pixtral-large-2411": { - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "mistral", "max_input_tokens": 128000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.000006, + "output_cost_per_token": 6e-6, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_response_schema": true, @@ -18242,13 +16438,13 @@ "supports_vision": true }, "mistral/pixtral-large-latest": { - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "mistral", "max_input_tokens": 128000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.000006, + "output_cost_per_token": 6e-6, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_response_schema": true, @@ -18263,7 +16459,7 @@ "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0.0000025, + "output_cost_per_token": 2.5e-6, "source": "https://platform.moonshot.ai/docs/pricing/chat#generation-model-kimi-k2", "supports_function_calling": true, "supports_tool_choice": true, @@ -18271,13 +16467,13 @@ }, "moonshot/kimi-latest": { "cache_read_input_token_cost": 1.5e-7, - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "moonshot", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0.000005, + "output_cost_per_token": 5e-6, "source": "https://platform.moonshot.ai/docs/pricing", "supports_function_calling": true, "supports_tool_choice": true, @@ -18285,13 +16481,13 @@ }, "moonshot/kimi-latest-128k": { "cache_read_input_token_cost": 1.5e-7, - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "moonshot", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0.000005, + "output_cost_per_token": 5e-6, "source": "https://platform.moonshot.ai/docs/pricing", "supports_function_calling": true, "supports_tool_choice": true, @@ -18299,13 +16495,13 @@ }, "moonshot/kimi-latest-32k": { "cache_read_input_token_cost": 1.5e-7, - "input_cost_per_token": 0.000001, + "input_cost_per_token": 1e-6, "litellm_provider": "moonshot", "max_input_tokens": 32768, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0.000003, + "output_cost_per_token": 3e-6, "source": "https://platform.moonshot.ai/docs/pricing", "supports_function_calling": true, "supports_tool_choice": true, @@ -18319,20 +16515,20 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.000002, + "output_cost_per_token": 2e-6, "source": "https://platform.moonshot.ai/docs/pricing", "supports_function_calling": true, "supports_tool_choice": true, "supports_vision": true }, "moonshot/kimi-thinking-preview": { - "input_cost_per_token": 0.00003, + "input_cost_per_token": 3e-5, "litellm_provider": "moonshot", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0.00003, + "output_cost_per_token": 3e-5, "source": "https://platform.moonshot.ai/docs/pricing", "supports_vision": true }, @@ -18344,81 +16540,81 @@ "max_output_tokens": 262144, "max_tokens": 262144, "mode": "chat", - "output_cost_per_token": 0.0000025, + "output_cost_per_token": 2.5e-6, "source": "https://platform.moonshot.ai/docs/pricing/chat#generation-model-kimi-k2", "supports_function_calling": true, "supports_tool_choice": true, "supports_web_search": true }, "moonshot/moonshot-v1-128k": { - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "moonshot", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0.000005, + "output_cost_per_token": 5e-6, "source": "https://platform.moonshot.ai/docs/pricing", "supports_function_calling": true, "supports_tool_choice": true }, "moonshot/moonshot-v1-128k-0430": { - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "moonshot", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0.000005, + "output_cost_per_token": 5e-6, "source": "https://platform.moonshot.ai/docs/pricing", "supports_function_calling": true, "supports_tool_choice": true }, "moonshot/moonshot-v1-128k-vision-preview": { - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "moonshot", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0.000005, + "output_cost_per_token": 5e-6, "source": "https://platform.moonshot.ai/docs/pricing", "supports_function_calling": true, "supports_tool_choice": true, "supports_vision": true }, "moonshot/moonshot-v1-32k": { - "input_cost_per_token": 0.000001, + "input_cost_per_token": 1e-6, "litellm_provider": "moonshot", "max_input_tokens": 32768, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0.000003, + "output_cost_per_token": 3e-6, "source": "https://platform.moonshot.ai/docs/pricing", "supports_function_calling": true, "supports_tool_choice": true }, "moonshot/moonshot-v1-32k-0430": { - "input_cost_per_token": 0.000001, + "input_cost_per_token": 1e-6, "litellm_provider": "moonshot", "max_input_tokens": 32768, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0.000003, + "output_cost_per_token": 3e-6, "source": "https://platform.moonshot.ai/docs/pricing", "supports_function_calling": true, "supports_tool_choice": true }, "moonshot/moonshot-v1-32k-vision-preview": { - "input_cost_per_token": 0.000001, + "input_cost_per_token": 1e-6, "litellm_provider": "moonshot", "max_input_tokens": 32768, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0.000003, + "output_cost_per_token": 3e-6, "source": "https://platform.moonshot.ai/docs/pricing", "supports_function_calling": true, "supports_tool_choice": true, @@ -18431,7 +16627,7 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.000002, + "output_cost_per_token": 2e-6, "source": "https://platform.moonshot.ai/docs/pricing", "supports_function_calling": true, "supports_tool_choice": true @@ -18443,7 +16639,7 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.000002, + "output_cost_per_token": 2e-6, "source": "https://platform.moonshot.ai/docs/pricing", "supports_function_calling": true, "supports_tool_choice": true @@ -18455,20 +16651,20 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.000002, + "output_cost_per_token": 2e-6, "source": "https://platform.moonshot.ai/docs/pricing", "supports_function_calling": true, "supports_tool_choice": true, "supports_vision": true }, "moonshot/moonshot-v1-auto": { - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "moonshot", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0.000005, + "output_cost_per_token": 5e-6, "source": "https://platform.moonshot.ai/docs/pricing", "supports_function_calling": true, "supports_tool_choice": true @@ -18480,7 +16676,7 @@ "max_output_tokens": 16000, "max_tokens": 16000, "mode": "chat", - "output_cost_per_token": 0.0000012, + "output_cost_per_token": 1.2e-6, "supports_function_calling": false, "supports_parallel_function_calling": false, "supports_system_messages": true, @@ -18494,7 +16690,7 @@ "max_output_tokens": 16000, "max_tokens": 16000, "mode": "chat", - "output_cost_per_token": 0.0000019, + "output_cost_per_token": 1.9e-6, "supports_function_calling": false, "supports_parallel_function_calling": false, "supports_system_messages": true, @@ -18515,14 +16711,8 @@ "output_cost_per_token": 0, "output_vector_size": 768, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models", - "supported_endpoints": [ - "/v1/embeddings" - ], - "supported_modalities": [ - "text", - "image", - "video" - ] + "supported_endpoints": ["/v1/embeddings"], + "supported_modalities": ["text", "image", "video"] }, "multimodalembedding@001": { "input_cost_per_character": 2e-7, @@ -18538,14 +16728,8 @@ "output_cost_per_token": 0, "output_vector_size": 768, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models", - "supported_endpoints": [ - "/v1/embeddings" - ], - "supported_modalities": [ - "text", - "image", - "video" - ] + "supported_endpoints": ["/v1/embeddings"], + "supported_modalities": ["text", "image", "video"] }, "nscale/Qwen/QwQ-32B": { "input_cost_per_token": 1.8e-7, @@ -18579,11 +16763,9 @@ "input_cost_per_pixel": 1.3e-9, "litellm_provider": "nscale", "mode": "image_generation", - "output_cost_per_pixel": 0, + "output_cost_per_pixel": 0.0, "source": "https://docs.nscale.com/docs/inference/serverless-models/current#image-models", - "supported_endpoints": [ - "/v1/images/generations" - ] + "supported_endpoints": ["/v1/images/generations"] }, "nscale/deepseek-ai/DeepSeek-R1-Distill-Llama-70B": { "input_cost_per_token": 3.75e-7, @@ -18686,21 +16868,19 @@ "input_cost_per_pixel": 3e-9, "litellm_provider": "nscale", "mode": "image_generation", - "output_cost_per_pixel": 0, + "output_cost_per_pixel": 0.0, "source": "https://docs.nscale.com/docs/inference/serverless-models/current#image-models", - "supported_endpoints": [ - "/v1/images/generations" - ] + "supported_endpoints": ["/v1/images/generations"] }, "o1": { - "cache_read_input_token_cost": 0.0000075, - "input_cost_per_token": 0.000015, + "cache_read_input_token_cost": 7.5e-6, + "input_cost_per_token": 1.5e-5, "litellm_provider": "openai", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 0.00006, + "output_cost_per_token": 6e-5, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_pdf_input": true, @@ -18712,14 +16892,14 @@ "supports_vision": true }, "o1-2024-12-17": { - "cache_read_input_token_cost": 0.0000075, - "input_cost_per_token": 0.000015, + "cache_read_input_token_cost": 7.5e-6, + "input_cost_per_token": 1.5e-5, "litellm_provider": "openai", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 0.00006, + "output_cost_per_token": 6e-5, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_pdf_input": true, @@ -18732,55 +16912,55 @@ }, "o1-mini": { "cache_read_input_token_cost": 5.5e-7, - "input_cost_per_token": 0.0000011, + "input_cost_per_token": 1.1e-6, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 65536, "max_tokens": 65536, "mode": "chat", - "output_cost_per_token": 0.0000044, + "output_cost_per_token": 4.4e-6, "supports_pdf_input": true, "supports_prompt_caching": true, "supports_vision": true }, "o1-mini-2024-09-12": { "deprecation_date": "2025-10-27", - "cache_read_input_token_cost": 0.0000015, - "input_cost_per_token": 0.000003, + "cache_read_input_token_cost": 1.5e-6, + "input_cost_per_token": 3e-6, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 65536, "max_tokens": 65536, "mode": "chat", - "output_cost_per_token": 0.000012, + "output_cost_per_token": 1.2e-5, "supports_pdf_input": true, "supports_prompt_caching": true, "supports_reasoning": true, "supports_vision": true }, "o1-preview": { - "cache_read_input_token_cost": 0.0000075, - "input_cost_per_token": 0.000015, + "cache_read_input_token_cost": 7.5e-6, + "input_cost_per_token": 1.5e-5, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0.00006, + "output_cost_per_token": 6e-5, "supports_pdf_input": true, "supports_prompt_caching": true, "supports_reasoning": true, "supports_vision": true }, "o1-preview-2024-09-12": { - "cache_read_input_token_cost": 0.0000075, - "input_cost_per_token": 0.000015, + "cache_read_input_token_cost": 7.5e-6, + "input_cost_per_token": 1.5e-5, "litellm_provider": "openai", "max_input_tokens": 128000, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0.00006, + "output_cost_per_token": 6e-5, "supports_pdf_input": true, "supports_prompt_caching": true, "supports_reasoning": true, @@ -18788,7 +16968,7 @@ }, "o1-pro": { "input_cost_per_token": 0.00015, - "input_cost_per_token_batches": 0.000075, + "input_cost_per_token_batches": 7.5e-5, "litellm_provider": "openai", "max_input_tokens": 200000, "max_output_tokens": 100000, @@ -18796,17 +16976,9 @@ "mode": "responses", "output_cost_per_token": 0.0006, "output_cost_per_token_batches": 0.0003, - "supported_endpoints": [ - "/v1/responses", - "/v1/batch" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "supported_endpoints": ["/v1/responses", "/v1/batch"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": false, "supports_parallel_function_calling": true, @@ -18820,7 +16992,7 @@ }, "o1-pro-2025-03-19": { "input_cost_per_token": 0.00015, - "input_cost_per_token_batches": 0.000075, + "input_cost_per_token_batches": 7.5e-5, "litellm_provider": "openai", "max_input_tokens": 200000, "max_output_tokens": 100000, @@ -18828,17 +17000,9 @@ "mode": "responses", "output_cost_per_token": 0.0006, "output_cost_per_token_batches": 0.0003, - "supported_endpoints": [ - "/v1/responses", - "/v1/batch" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "supported_endpoints": ["/v1/responses", "/v1/batch"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": false, "supports_parallel_function_calling": true, @@ -18854,30 +17018,25 @@ "cache_read_input_token_cost": 5e-7, "cache_read_input_token_cost_flex": 2.5e-7, "cache_read_input_token_cost_priority": 8.75e-7, - "input_cost_per_token": 0.000002, - "input_cost_per_token_flex": 0.000001, - "input_cost_per_token_priority": 0.0000035, + "input_cost_per_token": 2e-6, + "input_cost_per_token_flex": 1e-6, + "input_cost_per_token_priority": 3.5e-6, "litellm_provider": "openai", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 0.000008, - "output_cost_per_token_flex": 0.000004, - "output_cost_per_token_priority": 0.000014, + "output_cost_per_token": 8e-6, + "output_cost_per_token_flex": 4e-6, + "output_cost_per_token_priority": 1.4e-5, "supported_endpoints": [ "/v1/responses", "/v1/chat/completions", "/v1/completions", "/v1/batch" ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_parallel_function_calling": false, "supports_pdf_input": true, @@ -18890,26 +17049,21 @@ }, "o3-2025-04-16": { "cache_read_input_token_cost": 5e-7, - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "openai", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 0.000008, + "output_cost_per_token": 8e-6, "supported_endpoints": [ "/v1/responses", "/v1/chat/completions", "/v1/completions", "/v1/batch" ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_parallel_function_calling": false, "supports_pdf_input": true, @@ -18921,28 +17075,19 @@ "supports_vision": true }, "o3-deep-research": { - "cache_read_input_token_cost": 0.0000025, - "input_cost_per_token": 0.00001, - "input_cost_per_token_batches": 0.000005, + "cache_read_input_token_cost": 2.5e-6, + "input_cost_per_token": 1e-5, + "input_cost_per_token_batches": 5e-6, "litellm_provider": "openai", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "responses", - "output_cost_per_token": 0.00004, - "output_cost_per_token_batches": 0.00002, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 4e-5, + "output_cost_per_token_batches": 2e-5, + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -18954,28 +17099,19 @@ "supports_vision": true }, "o3-deep-research-2025-06-26": { - "cache_read_input_token_cost": 0.0000025, - "input_cost_per_token": 0.00001, - "input_cost_per_token_batches": 0.000005, + "cache_read_input_token_cost": 2.5e-6, + "input_cost_per_token": 1e-5, + "input_cost_per_token_batches": 5e-6, "litellm_provider": "openai", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "responses", - "output_cost_per_token": 0.00004, - "output_cost_per_token_batches": 0.00002, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 4e-5, + "output_cost_per_token_batches": 2e-5, + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -18988,13 +17124,13 @@ }, "o3-mini": { "cache_read_input_token_cost": 5.5e-7, - "input_cost_per_token": 0.0000011, + "input_cost_per_token": 1.1e-6, "litellm_provider": "openai", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 0.0000044, + "output_cost_per_token": 4.4e-6, "supports_function_calling": true, "supports_parallel_function_calling": false, "supports_prompt_caching": true, @@ -19005,13 +17141,13 @@ }, "o3-mini-2025-01-31": { "cache_read_input_token_cost": 5.5e-7, - "input_cost_per_token": 0.0000011, + "input_cost_per_token": 1.1e-6, "litellm_provider": "openai", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 0.0000044, + "output_cost_per_token": 4.4e-6, "supports_function_calling": true, "supports_parallel_function_calling": false, "supports_prompt_caching": true, @@ -19021,26 +17157,18 @@ "supports_vision": false }, "o3-pro": { - "input_cost_per_token": 0.00002, - "input_cost_per_token_batches": 0.00001, + "input_cost_per_token": 2e-5, + "input_cost_per_token_batches": 1e-5, "litellm_provider": "openai", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "responses", - "output_cost_per_token": 0.00008, - "output_cost_per_token_batches": 0.00004, - "supported_endpoints": [ - "/v1/responses", - "/v1/batch" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 8e-5, + "output_cost_per_token_batches": 4e-5, + "supported_endpoints": ["/v1/responses", "/v1/batch"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_parallel_function_calling": false, "supports_pdf_input": true, @@ -19051,26 +17179,18 @@ "supports_vision": true }, "o3-pro-2025-06-10": { - "input_cost_per_token": 0.00002, - "input_cost_per_token_batches": 0.00001, + "input_cost_per_token": 2e-5, + "input_cost_per_token_batches": 1e-5, "litellm_provider": "openai", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "responses", - "output_cost_per_token": 0.00008, - "output_cost_per_token_batches": 0.00004, - "supported_endpoints": [ - "/v1/responses", - "/v1/batch" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 8e-5, + "output_cost_per_token_batches": 4e-5, + "supported_endpoints": ["/v1/responses", "/v1/batch"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_parallel_function_calling": false, "supports_pdf_input": true, @@ -19084,17 +17204,17 @@ "cache_read_input_token_cost": 2.75e-7, "cache_read_input_token_cost_flex": 1.375e-7, "cache_read_input_token_cost_priority": 5e-7, - "input_cost_per_token": 0.0000011, + "input_cost_per_token": 1.1e-6, "input_cost_per_token_flex": 5.5e-7, - "input_cost_per_token_priority": 0.000002, + "input_cost_per_token_priority": 2e-6, "litellm_provider": "openai", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 0.0000044, - "output_cost_per_token_flex": 0.0000022, - "output_cost_per_token_priority": 0.000008, + "output_cost_per_token": 4.4e-6, + "output_cost_per_token_flex": 2.2e-6, + "output_cost_per_token_priority": 8e-6, "supports_function_calling": true, "supports_parallel_function_calling": false, "supports_pdf_input": true, @@ -19107,13 +17227,13 @@ }, "o4-mini-2025-04-16": { "cache_read_input_token_cost": 2.75e-7, - "input_cost_per_token": 0.0000011, + "input_cost_per_token": 1.1e-6, "litellm_provider": "openai", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 0.0000044, + "output_cost_per_token": 4.4e-6, "supports_function_calling": true, "supports_parallel_function_calling": false, "supports_pdf_input": true, @@ -19126,27 +17246,18 @@ }, "o4-mini-deep-research": { "cache_read_input_token_cost": 5e-7, - "input_cost_per_token": 0.000002, - "input_cost_per_token_batches": 0.000001, + "input_cost_per_token": 2e-6, + "input_cost_per_token_batches": 1e-6, "litellm_provider": "openai", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "responses", - "output_cost_per_token": 0.000008, - "output_cost_per_token_batches": 0.000004, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 8e-6, + "output_cost_per_token_batches": 4e-6, + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -19159,27 +17270,18 @@ }, "o4-mini-deep-research-2025-06-26": { "cache_read_input_token_cost": 5e-7, - "input_cost_per_token": 0.000002, - "input_cost_per_token_batches": 0.000001, + "input_cost_per_token": 2e-6, + "input_cost_per_token_batches": 1e-6, "litellm_provider": "openai", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "responses", - "output_cost_per_token": 0.000008, - "output_cost_per_token_batches": 0.000004, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/batch", - "/v1/responses" - ], - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 8e-6, + "output_cost_per_token_batches": 4e-6, + "supported_endpoints": ["/v1/chat/completions", "/v1/batch", "/v1/responses"], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_function_calling": true, "supports_native_streaming": true, "supports_parallel_function_calling": true, @@ -19191,25 +17293,25 @@ "supports_vision": true }, "oci/meta.llama-3.1-405b-instruct": { - "input_cost_per_token": 0.00001068, + "input_cost_per_token": 1.068e-5, "litellm_provider": "oci", "max_input_tokens": 128000, "max_output_tokens": 4000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.00001068, + "output_cost_per_token": 1.068e-5, "source": "https://www.oracle.com/artificial-intelligence/generative-ai/generative-ai-service/pricing", "supports_function_calling": true, "supports_response_schema": false }, "oci/meta.llama-3.2-90b-vision-instruct": { - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "oci", "max_input_tokens": 128000, "max_output_tokens": 4000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.000002, + "output_cost_per_token": 2e-6, "source": "https://www.oracle.com/artificial-intelligence/generative-ai/generative-ai-service/pricing", "supports_function_calling": true, "supports_response_schema": false @@ -19251,7 +17353,7 @@ "supports_response_schema": false }, "oci/xai.grok-3": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "oci", "max_input_tokens": 131072, "max_output_tokens": 131072, @@ -19263,13 +17365,13 @@ "supports_response_schema": false }, "oci/xai.grok-3-fast": { - "input_cost_per_token": 0.000005, + "input_cost_per_token": 5e-6, "litellm_provider": "oci", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0.000025, + "output_cost_per_token": 2.5e-5, "source": "https://www.oracle.com/artificial-intelligence/generative-ai/generative-ai-service/pricing", "supports_function_calling": true, "supports_response_schema": false @@ -19293,13 +17395,13 @@ "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0.000004, + "output_cost_per_token": 4e-6, "source": "https://www.oracle.com/artificial-intelligence/generative-ai/generative-ai-service/pricing", "supports_function_calling": true, "supports_response_schema": false }, "oci/xai.grok-4": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "oci", "max_input_tokens": 128000, "max_output_tokens": 128000, @@ -19311,345 +17413,345 @@ "supports_response_schema": false }, "oci/cohere.command-latest": { - "input_cost_per_token": 0.00000156, + "input_cost_per_token": 1.56e-6, "litellm_provider": "oci", "max_input_tokens": 128000, "max_output_tokens": 4000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.00000156, + "output_cost_per_token": 1.56e-6, "source": "https://www.oracle.com/cloud/ai/generative-ai/pricing/", "supports_function_calling": true, "supports_response_schema": false }, "oci/cohere.command-a-03-2025": { - "input_cost_per_token": 0.00000156, + "input_cost_per_token": 1.56e-6, "litellm_provider": "oci", "max_input_tokens": 256000, "max_output_tokens": 4000, "max_tokens": 256000, "mode": "chat", - "output_cost_per_token": 0.00000156, + "output_cost_per_token": 1.56e-6, "source": "https://www.oracle.com/cloud/ai/generative-ai/pricing/", "supports_function_calling": true, "supports_response_schema": false }, "oci/cohere.command-plus-latest": { - "input_cost_per_token": 0.00000156, + "input_cost_per_token": 1.56e-6, "litellm_provider": "oci", "max_input_tokens": 128000, "max_output_tokens": 4000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.00000156, + "output_cost_per_token": 1.56e-6, "source": "https://www.oracle.com/cloud/ai/generative-ai/pricing/", "supports_function_calling": true, "supports_response_schema": false }, "ollama/codegeex4": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "ollama", "max_input_tokens": 32768, "max_output_tokens": 8192, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "supports_function_calling": false }, "ollama/codegemma": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "ollama", "max_input_tokens": 8192, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "completion", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "ollama/codellama": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "ollama", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "completion", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "ollama/deepseek-coder-v2-base": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "ollama", "max_input_tokens": 8192, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "completion", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "supports_function_calling": true }, "ollama/deepseek-coder-v2-instruct": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "ollama", "max_input_tokens": 32768, "max_output_tokens": 8192, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "supports_function_calling": true }, "ollama/deepseek-coder-v2-lite-base": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "ollama", "max_input_tokens": 8192, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "completion", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "supports_function_calling": true }, "ollama/deepseek-coder-v2-lite-instruct": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "ollama", "max_input_tokens": 32768, "max_output_tokens": 8192, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "supports_function_calling": true }, "ollama/deepseek-v3.1:671b-cloud": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "ollama", "max_input_tokens": 163840, "max_output_tokens": 163840, "max_tokens": 163840, "mode": "chat", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "supports_function_calling": true }, "ollama/gpt-oss:120b-cloud": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "ollama", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "supports_function_calling": true }, "ollama/gpt-oss:20b-cloud": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "ollama", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "supports_function_calling": true }, "ollama/internlm2_5-20b-chat": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "ollama", "max_input_tokens": 32768, "max_output_tokens": 8192, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "supports_function_calling": true }, "ollama/llama2": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "ollama", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "ollama/llama2-uncensored": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "ollama", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "completion", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "ollama/llama2:13b": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "ollama", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "ollama/llama2:70b": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "ollama", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "ollama/llama2:7b": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "ollama", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "ollama/llama3": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "ollama", "max_input_tokens": 8192, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "ollama/llama3.1": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "ollama", "max_input_tokens": 8192, "max_output_tokens": 8192, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "supports_function_calling": true }, "ollama/llama3:70b": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "ollama", "max_input_tokens": 8192, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "ollama/llama3:8b": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "ollama", "max_input_tokens": 8192, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "ollama/mistral": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "ollama", "max_input_tokens": 8192, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "completion", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "supports_function_calling": true }, "ollama/mistral-7B-Instruct-v0.1": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "ollama", "max_input_tokens": 8192, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "supports_function_calling": true }, "ollama/mistral-7B-Instruct-v0.2": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "ollama", "max_input_tokens": 32768, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "supports_function_calling": true }, "ollama/mistral-large-instruct-2407": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "ollama", "max_input_tokens": 65536, "max_output_tokens": 8192, "max_tokens": 65536, "mode": "chat", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "supports_function_calling": true }, "ollama/mixtral-8x22B-Instruct-v0.1": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "ollama", "max_input_tokens": 65536, "max_output_tokens": 65536, "max_tokens": 65536, "mode": "chat", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "supports_function_calling": true }, "ollama/mixtral-8x7B-Instruct-v0.1": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "ollama", "max_input_tokens": 32768, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "supports_function_calling": true }, "ollama/orca-mini": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "ollama", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "completion", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "ollama/qwen3-coder:480b-cloud": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "ollama", "max_input_tokens": 262144, "max_output_tokens": 262144, "max_tokens": 262144, "mode": "chat", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "supports_function_calling": true }, "ollama/vicuna": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "ollama", "max_input_tokens": 2048, "max_output_tokens": 2048, "max_tokens": 2048, "mode": "completion", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "omni-moderation-2024-09-26": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "openai", "max_input_tokens": 32768, "max_output_tokens": 0, "max_tokens": 32768, "mode": "moderation", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "omni-moderation-latest": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "openai", "max_input_tokens": 32768, "max_output_tokens": 0, "max_tokens": 32768, "mode": "moderation", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "omni-moderation-latest-intents": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "openai", "max_input_tokens": 32768, "max_output_tokens": 0, "max_tokens": 32768, "mode": "moderation", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "openai.gpt-oss-120b-1:0": { "input_cost_per_token": 1.5e-7, @@ -19678,31 +17780,31 @@ "supports_tool_choice": true }, "openrouter/anthropic/claude-2": { - "input_cost_per_token": 0.00001102, + "input_cost_per_token": 1.102e-5, "litellm_provider": "openrouter", "max_output_tokens": 8191, "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 0.00003268, + "output_cost_per_token": 3.268e-5, "supports_tool_choice": true }, "openrouter/anthropic/claude-3-5-haiku": { - "input_cost_per_token": 0.000001, + "input_cost_per_token": 1e-6, "litellm_provider": "openrouter", "max_tokens": 200000, "mode": "chat", - "output_cost_per_token": 0.000005, + "output_cost_per_token": 5e-6, "supports_function_calling": true, "supports_tool_choice": true }, "openrouter/anthropic/claude-3-5-haiku-20241022": { - "input_cost_per_token": 0.000001, + "input_cost_per_token": 1e-6, "litellm_provider": "openrouter", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.000005, + "output_cost_per_token": 5e-6, "supports_function_calling": true, "supports_tool_choice": true, "tool_use_system_prompt_tokens": 264 @@ -19713,7 +17815,7 @@ "litellm_provider": "openrouter", "max_tokens": 200000, "mode": "chat", - "output_cost_per_token": 0.00000125, + "output_cost_per_token": 1.25e-6, "supports_function_calling": true, "supports_tool_choice": true, "supports_vision": true @@ -19725,20 +17827,20 @@ "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.00000125, + "output_cost_per_token": 1.25e-6, "supports_function_calling": true, "supports_tool_choice": true, "supports_vision": true, "tool_use_system_prompt_tokens": 264 }, "openrouter/anthropic/claude-3-opus": { - "input_cost_per_token": 0.000015, + "input_cost_per_token": 1.5e-5, "litellm_provider": "openrouter", "max_input_tokens": 200000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.000075, + "output_cost_per_token": 7.5e-5, "supports_function_calling": true, "supports_tool_choice": true, "supports_vision": true, @@ -19746,23 +17848,23 @@ }, "openrouter/anthropic/claude-3-sonnet": { "input_cost_per_image": 0.0048, - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "openrouter", "max_tokens": 200000, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "supports_function_calling": true, "supports_tool_choice": true, "supports_vision": true }, "openrouter/anthropic/claude-3.5-sonnet": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "openrouter", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "supports_assistant_prefill": true, "supports_computer_use": true, "supports_function_calling": true, @@ -19771,13 +17873,13 @@ "tool_use_system_prompt_tokens": 159 }, "openrouter/anthropic/claude-3.5-sonnet:beta": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "openrouter", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "supports_computer_use": true, "supports_function_calling": true, "supports_tool_choice": true, @@ -19786,13 +17888,13 @@ }, "openrouter/anthropic/claude-3.7-sonnet": { "input_cost_per_image": 0.0048, - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "openrouter", "max_input_tokens": 200000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "supports_assistant_prefill": true, "supports_computer_use": true, "supports_function_calling": true, @@ -19803,13 +17905,13 @@ }, "openrouter/anthropic/claude-3.7-sonnet:beta": { "input_cost_per_image": 0.0048, - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "openrouter", "max_input_tokens": 200000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "supports_computer_use": true, "supports_function_calling": true, "supports_reasoning": true, @@ -19818,25 +17920,25 @@ "tool_use_system_prompt_tokens": 159 }, "openrouter/anthropic/claude-instant-v1": { - "input_cost_per_token": 0.00000163, + "input_cost_per_token": 1.63e-6, "litellm_provider": "openrouter", "max_output_tokens": 8191, "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 0.00000551, + "output_cost_per_token": 5.51e-6, "supports_tool_choice": true }, "openrouter/anthropic/claude-opus-4": { "input_cost_per_image": 0.0048, - "cache_creation_input_token_cost": 0.00001875, - "cache_read_input_token_cost": 0.0000015, - "input_cost_per_token": 0.000015, + "cache_creation_input_token_cost": 1.875e-5, + "cache_read_input_token_cost": 1.5e-6, + "input_cost_per_token": 1.5e-5, "litellm_provider": "openrouter", "max_input_tokens": 200000, "max_output_tokens": 32000, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 0.000075, + "output_cost_per_token": 7.5e-5, "supports_assistant_prefill": true, "supports_computer_use": true, "supports_function_calling": true, @@ -19848,16 +17950,16 @@ }, "openrouter/anthropic/claude-opus-4.1": { "input_cost_per_image": 0.0048, - "cache_creation_input_token_cost": 0.00001875, - "cache_creation_input_token_cost_above_1hr": 0.00003, - "cache_read_input_token_cost": 0.0000015, - "input_cost_per_token": 0.000015, + "cache_creation_input_token_cost": 1.875e-5, + "cache_creation_input_token_cost_above_1hr": 3e-5, + "cache_read_input_token_cost": 1.5e-6, + "input_cost_per_token": 1.5e-5, "litellm_provider": "openrouter", "max_input_tokens": 200000, "max_output_tokens": 32000, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 0.000075, + "output_cost_per_token": 7.5e-5, "supports_assistant_prefill": true, "supports_computer_use": true, "supports_function_calling": true, @@ -19869,19 +17971,19 @@ }, "openrouter/anthropic/claude-sonnet-4": { "input_cost_per_image": 0.0048, - "cache_creation_input_token_cost": 0.00000375, - "cache_creation_input_token_cost_above_200k_tokens": 0.0000075, + "cache_creation_input_token_cost": 3.75e-6, + "cache_creation_input_token_cost_above_200k_tokens": 7.5e-6, "cache_read_input_token_cost": 3e-7, "cache_read_input_token_cost_above_200k_tokens": 6e-7, - "input_cost_per_token": 0.000003, - "input_cost_per_token_above_200k_tokens": 0.000006, - "output_cost_per_token_above_200k_tokens": 0.0000225, + "input_cost_per_token": 3e-6, + "input_cost_per_token_above_200k_tokens": 6e-6, + "output_cost_per_token_above_200k_tokens": 2.25e-5, "litellm_provider": "openrouter", "max_input_tokens": 1000000, "max_output_tokens": 64000, "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "supports_assistant_prefill": true, "supports_computer_use": true, "supports_function_calling": true, @@ -19893,19 +17995,19 @@ }, "openrouter/anthropic/claude-sonnet-4.5": { "input_cost_per_image": 0.0048, - "cache_creation_input_token_cost": 0.00000375, + "cache_creation_input_token_cost": 3.75e-6, "cache_read_input_token_cost": 3e-7, - "input_cost_per_token": 0.000003, - "input_cost_per_token_above_200k_tokens": 0.000006, - "output_cost_per_token_above_200k_tokens": 0.0000225, - "cache_creation_input_token_cost_above_200k_tokens": 0.0000075, + "input_cost_per_token": 3e-6, + "input_cost_per_token_above_200k_tokens": 6e-6, + "output_cost_per_token_above_200k_tokens": 2.25e-5, + "cache_creation_input_token_cost_above_200k_tokens": 7.5e-6, "cache_read_input_token_cost_above_200k_tokens": 6e-7, "litellm_provider": "openrouter", "max_input_tokens": 1000000, "max_output_tokens": 1000000, "max_tokens": 1000000, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "supports_assistant_prefill": true, "supports_computer_use": true, "supports_function_calling": true, @@ -19916,15 +18018,15 @@ "tool_use_system_prompt_tokens": 159 }, "openrouter/anthropic/claude-haiku-4.5": { - "cache_creation_input_token_cost": 0.00000125, + "cache_creation_input_token_cost": 1.25e-6, "cache_read_input_token_cost": 1e-7, - "input_cost_per_token": 0.000001, + "input_cost_per_token": 1e-6, "litellm_provider": "openrouter", "max_input_tokens": 200000, "max_output_tokens": 200000, "max_tokens": 200000, "mode": "chat", - "output_cost_per_token": 0.000005, + "output_cost_per_token": 5e-6, "supports_assistant_prefill": true, "supports_computer_use": true, "supports_function_calling": true, @@ -19954,11 +18056,11 @@ "supports_tool_choice": true }, "openrouter/cohere/command-r-plus": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "openrouter", "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "supports_tool_choice": true }, "openrouter/databricks/dbrx-instruct": { @@ -20040,7 +18142,7 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.00000219, + "output_cost_per_token": 2.19e-6, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_prompt_caching": true, @@ -20055,7 +18157,7 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.00000215, + "output_cost_per_token": 2.15e-6, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_prompt_caching": true, @@ -20106,7 +18208,7 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_token": 0.0000025, + "output_cost_per_token": 2.5e-6, "supports_audio_output": true, "supports_function_calling": true, "supports_response_schema": true, @@ -20116,7 +18218,7 @@ }, "openrouter/google/gemini-2.5-pro": { "input_cost_per_audio_token": 7e-7, - "input_cost_per_token": 0.00000125, + "input_cost_per_token": 1.25e-6, "litellm_provider": "openrouter", "max_audio_length_hours": 8.4, "max_audio_per_prompt": 1, @@ -20128,7 +18230,7 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_token": 0.00001, + "output_cost_per_token": 1e-5, "supports_audio_output": true, "supports_function_calling": true, "supports_response_schema": true, @@ -20140,9 +18242,9 @@ "cache_read_input_token_cost": 2e-7, "cache_read_input_token_cost_above_200k_tokens": 4e-7, "cache_creation_input_token_cost_above_200k_tokens": 2.5e-7, - "input_cost_per_token": 0.000002, - "input_cost_per_token_above_200k_tokens": 0.000004, - "input_cost_per_token_batches": 0.000001, + "input_cost_per_token": 2e-6, + "input_cost_per_token_above_200k_tokens": 4e-6, + "input_cost_per_token_batches": 1e-6, "litellm_provider": "openrouter", "max_audio_length_hours": 8.4, "max_audio_per_prompt": 1, @@ -20154,23 +18256,12 @@ "max_video_length": 1, "max_videos_per_prompt": 10, "mode": "chat", - "output_cost_per_token": 0.000012, - "output_cost_per_token_above_200k_tokens": 0.000018, - "output_cost_per_token_batches": 0.000006, - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/completions", - "/v1/batch" - ], - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 1.2e-5, + "output_cost_per_token_above_200k_tokens": 1.8e-5, + "output_cost_per_token_batches": 6e-6, + "supported_endpoints": ["/v1/chat/completions", "/v1/completions", "/v1/batch"], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text"], "supports_audio_input": true, "supports_function_calling": true, "supports_pdf_input": true, @@ -20185,13 +18276,13 @@ }, "openrouter/google/gemini-pro-1.5": { "input_cost_per_image": 0.00265, - "input_cost_per_token": 0.0000025, + "input_cost_per_token": 2.5e-6, "litellm_provider": "openrouter", "max_input_tokens": 1000000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.0000075, + "output_cost_per_token": 7.5e-6, "supports_function_calling": true, "supports_tool_choice": true, "supports_vision": true @@ -20224,27 +18315,27 @@ "supports_tool_choice": true }, "openrouter/gryphe/mythomax-l2-13b": { - "input_cost_per_token": 0.000001875, + "input_cost_per_token": 1.875e-6, "litellm_provider": "openrouter", "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.000001875, + "output_cost_per_token": 1.875e-6, "supports_tool_choice": true }, "openrouter/jondurbin/airoboros-l2-70b-2.1": { - "input_cost_per_token": 0.000013875, + "input_cost_per_token": 1.3875e-5, "litellm_provider": "openrouter", "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.000013875, + "output_cost_per_token": 1.3875e-5, "supports_tool_choice": true }, "openrouter/mancer/weaver": { - "input_cost_per_token": 0.000005625, + "input_cost_per_token": 5.625e-6, "litellm_provider": "openrouter", "max_tokens": 8000, "mode": "chat", - "output_cost_per_token": 0.000005625, + "output_cost_per_token": 5.625e-6, "supports_tool_choice": true }, "openrouter/meta-llama/codellama-34b-instruct": { @@ -20264,11 +18355,11 @@ "supports_tool_choice": true }, "openrouter/meta-llama/llama-2-70b-chat": { - "input_cost_per_token": 0.0000015, + "input_cost_per_token": 1.5e-6, "litellm_provider": "openrouter", "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.0000015, + "output_cost_per_token": 1.5e-6, "supports_tool_choice": true }, "openrouter/meta-llama/llama-3-70b-instruct": { @@ -20292,23 +18383,23 @@ "litellm_provider": "openrouter", "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 0.00000225, + "output_cost_per_token": 2.25e-6, "supports_tool_choice": true }, "openrouter/meta-llama/llama-3-8b-instruct:free": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "openrouter", "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "supports_tool_choice": true }, "openrouter/microsoft/wizardlm-2-8x22b:nitro": { - "input_cost_per_token": 0.000001, + "input_cost_per_token": 1e-6, "litellm_provider": "openrouter", "max_tokens": 65536, "mode": "chat", - "output_cost_per_token": 0.000001, + "output_cost_per_token": 1e-6, "supports_tool_choice": true }, "openrouter/minimax/minimax-m2": { @@ -20318,7 +18409,7 @@ "max_output_tokens": 204800, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0.00000102, + "output_cost_per_token": 1.02e-6, "supports_function_calling": true, "supports_prompt_caching": false, "supports_reasoning": true, @@ -20333,19 +18424,19 @@ "supports_tool_choice": true }, "openrouter/mistralai/mistral-7b-instruct:free": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "openrouter", "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "supports_tool_choice": true }, "openrouter/mistralai/mistral-large": { - "input_cost_per_token": 0.000008, + "input_cost_per_token": 8e-6, "litellm_provider": "openrouter", "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 0.000024, + "output_cost_per_token": 2.4e-5, "supports_tool_choice": true }, "openrouter/mistralai/mistral-small-3.1-24b-instruct": { @@ -20381,49 +18472,49 @@ "supports_tool_choice": true }, "openrouter/openai/gpt-3.5-turbo": { - "input_cost_per_token": 0.0000015, + "input_cost_per_token": 1.5e-6, "litellm_provider": "openrouter", "max_tokens": 4095, "mode": "chat", - "output_cost_per_token": 0.000002, + "output_cost_per_token": 2e-6, "supports_tool_choice": true }, "openrouter/openai/gpt-3.5-turbo-16k": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "openrouter", "max_tokens": 16383, "mode": "chat", - "output_cost_per_token": 0.000004, + "output_cost_per_token": 4e-6, "supports_tool_choice": true }, "openrouter/openai/gpt-4": { - "input_cost_per_token": 0.00003, + "input_cost_per_token": 3e-5, "litellm_provider": "openrouter", "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.00006, + "output_cost_per_token": 6e-5, "supports_tool_choice": true }, "openrouter/openai/gpt-4-vision-preview": { "input_cost_per_image": 0.01445, - "input_cost_per_token": 0.00001, + "input_cost_per_token": 1e-5, "litellm_provider": "openrouter", "max_tokens": 130000, "mode": "chat", - "output_cost_per_token": 0.00003, + "output_cost_per_token": 3e-5, "supports_function_calling": true, "supports_tool_choice": true, "supports_vision": true }, "openrouter/openai/gpt-4.1": { "cache_read_input_token_cost": 5e-7, - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "openrouter", "max_input_tokens": 1047576, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0.000008, + "output_cost_per_token": 8e-6, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -20434,13 +18525,13 @@ }, "openrouter/openai/gpt-4.1-2025-04-14": { "cache_read_input_token_cost": 5e-7, - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "openrouter", "max_input_tokens": 1047576, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0.000008, + "output_cost_per_token": 8e-6, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -20457,7 +18548,7 @@ "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0.0000016, + "output_cost_per_token": 1.6e-6, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -20474,7 +18565,7 @@ "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0.0000016, + "output_cost_per_token": 1.6e-6, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -20518,26 +18609,26 @@ "supports_vision": true }, "openrouter/openai/gpt-4o": { - "input_cost_per_token": 0.0000025, + "input_cost_per_token": 2.5e-6, "litellm_provider": "openrouter", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.00001, + "output_cost_per_token": 1e-5, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_tool_choice": true, "supports_vision": true }, "openrouter/openai/gpt-4o-2024-05-13": { - "input_cost_per_token": 0.000005, + "input_cost_per_token": 5e-6, "litellm_provider": "openrouter", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_tool_choice": true, @@ -20545,58 +18636,43 @@ }, "openrouter/openai/gpt-5-chat": { "cache_read_input_token_cost": 1.25e-7, - "input_cost_per_token": 0.00000125, + "input_cost_per_token": 1.25e-6, "litellm_provider": "openrouter", "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.00001, - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 1e-5, + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_reasoning": true, "supports_tool_choice": true }, "openrouter/openai/gpt-5-codex": { "cache_read_input_token_cost": 1.25e-7, - "input_cost_per_token": 0.00000125, + "input_cost_per_token": 1.25e-6, "litellm_provider": "openrouter", "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.00001, - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 1e-5, + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_reasoning": true, "supports_tool_choice": true }, "openrouter/openai/gpt-5": { "cache_read_input_token_cost": 1.25e-7, - "input_cost_per_token": 0.00000125, + "input_cost_per_token": 1.25e-6, "litellm_provider": "openrouter", "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.00001, - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 1e-5, + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_reasoning": true, "supports_tool_choice": true }, @@ -20608,14 +18684,9 @@ "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.000002, - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "output_cost_per_token": 2e-6, + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_reasoning": true, "supports_tool_choice": true }, @@ -20628,13 +18699,8 @@ "max_tokens": 128000, "mode": "chat", "output_cost_per_token": 4e-7, - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text" - ], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text"], "supports_reasoning": true, "supports_tool_choice": true }, @@ -20669,14 +18735,14 @@ "supports_tool_choice": true }, "openrouter/openai/o1": { - "cache_read_input_token_cost": 0.0000075, - "input_cost_per_token": 0.000015, + "cache_read_input_token_cost": 7.5e-6, + "input_cost_per_token": 1.5e-5, "litellm_provider": "openrouter", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 100000, "mode": "chat", - "output_cost_per_token": 0.00006, + "output_cost_per_token": 6e-5, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_prompt_caching": true, @@ -20686,65 +18752,65 @@ "supports_vision": true }, "openrouter/openai/o1-mini": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "openrouter", "max_input_tokens": 128000, "max_output_tokens": 65536, "max_tokens": 65536, "mode": "chat", - "output_cost_per_token": 0.000012, + "output_cost_per_token": 1.2e-5, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_tool_choice": true, "supports_vision": false }, "openrouter/openai/o1-mini-2024-09-12": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "openrouter", "max_input_tokens": 128000, "max_output_tokens": 65536, "max_tokens": 65536, "mode": "chat", - "output_cost_per_token": 0.000012, + "output_cost_per_token": 1.2e-5, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_tool_choice": true, "supports_vision": false }, "openrouter/openai/o1-preview": { - "input_cost_per_token": 0.000015, + "input_cost_per_token": 1.5e-5, "litellm_provider": "openrouter", "max_input_tokens": 128000, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0.00006, + "output_cost_per_token": 6e-5, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_tool_choice": true, "supports_vision": false }, "openrouter/openai/o1-preview-2024-09-12": { - "input_cost_per_token": 0.000015, + "input_cost_per_token": 1.5e-5, "litellm_provider": "openrouter", "max_input_tokens": 128000, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0.00006, + "output_cost_per_token": 6e-5, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_tool_choice": true, "supports_vision": false }, "openrouter/openai/o3-mini": { - "input_cost_per_token": 0.0000011, + "input_cost_per_token": 1.1e-6, "litellm_provider": "openrouter", "max_input_tokens": 128000, "max_output_tokens": 65536, "max_tokens": 65536, "mode": "chat", - "output_cost_per_token": 0.0000044, + "output_cost_per_token": 4.4e-6, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_reasoning": true, @@ -20752,13 +18818,13 @@ "supports_vision": false }, "openrouter/openai/o3-mini-high": { - "input_cost_per_token": 0.0000011, + "input_cost_per_token": 1.1e-6, "litellm_provider": "openrouter", "max_input_tokens": 128000, "max_output_tokens": 65536, "max_tokens": 65536, "mode": "chat", - "output_cost_per_token": 0.0000044, + "output_cost_per_token": 4.4e-6, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_reasoning": true, @@ -20766,11 +18832,11 @@ "supports_vision": false }, "openrouter/pygmalionai/mythalion-13b": { - "input_cost_per_token": 0.000001875, + "input_cost_per_token": 1.875e-6, "litellm_provider": "openrouter", "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.000001875, + "output_cost_per_token": 1.875e-6, "supports_tool_choice": true }, "openrouter/qwen/qwen-2.5-coder-32b-instruct": { @@ -20813,26 +18879,26 @@ "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0.0000034, + "output_cost_per_token": 3.4e-6, "source": "https://openrouter.ai/switchpoint/router", "supports_tool_choice": true }, "openrouter/undi95/remm-slerp-l2-13b": { - "input_cost_per_token": 0.000001875, + "input_cost_per_token": 1.875e-6, "litellm_provider": "openrouter", "max_tokens": 6144, "mode": "chat", - "output_cost_per_token": 0.000001875, + "output_cost_per_token": 1.875e-6, "supports_tool_choice": true }, "openrouter/x-ai/grok-4": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "openrouter", "max_input_tokens": 256000, "max_output_tokens": 256000, "max_tokens": 256000, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "source": "https://openrouter.ai/x-ai/grok-4", "supports_function_calling": true, "supports_reasoning": true, @@ -20854,13 +18920,13 @@ "supports_web_search": false }, "openrouter/z-ai/glm-4.6": { - "input_cost_per_token": 4e-7, + "input_cost_per_token": 4.0e-7, "litellm_provider": "openrouter", "max_input_tokens": 202800, "max_output_tokens": 131000, "max_tokens": 202800, "mode": "chat", - "output_cost_per_token": 0.00000175, + "output_cost_per_token": 1.75e-6, "source": "https://openrouter.ai/z-ai/glm-4.6", "supports_function_calling": true, "supports_reasoning": true, @@ -20873,7 +18939,7 @@ "max_output_tokens": 131000, "max_tokens": 202800, "mode": "chat", - "output_cost_per_token": 0.0000019, + "output_cost_per_token": 1.9e-6, "source": "https://openrouter.ai/z-ai/glm-4.6:exacto", "supports_function_calling": true, "supports_reasoning": true, @@ -21158,7 +19224,7 @@ "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 0.0000014 + "output_cost_per_token": 1.4e-6 }, "perplexity/codellama-70b-instruct": { "input_cost_per_token": 7e-7, @@ -21167,7 +19233,7 @@ "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 0.0000028 + "output_cost_per_token": 2.8e-6 }, "perplexity/llama-2-70b-chat": { "input_cost_per_token": 7e-7, @@ -21176,16 +19242,16 @@ "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.0000028 + "output_cost_per_token": 2.8e-6 }, "perplexity/llama-3.1-70b-instruct": { - "input_cost_per_token": 0.000001, + "input_cost_per_token": 1e-6, "litellm_provider": "perplexity", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0.000001 + "output_cost_per_token": 1e-6 }, "perplexity/llama-3.1-8b-instruct": { "input_cost_per_token": 2e-7, @@ -21198,33 +19264,33 @@ }, "perplexity/llama-3.1-sonar-huge-128k-online": { "deprecation_date": "2025-02-22", - "input_cost_per_token": 0.000005, + "input_cost_per_token": 5e-6, "litellm_provider": "perplexity", "max_input_tokens": 127072, "max_output_tokens": 127072, "max_tokens": 127072, "mode": "chat", - "output_cost_per_token": 0.000005 + "output_cost_per_token": 5e-6 }, "perplexity/llama-3.1-sonar-large-128k-chat": { "deprecation_date": "2025-02-22", - "input_cost_per_token": 0.000001, + "input_cost_per_token": 1e-6, "litellm_provider": "perplexity", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0.000001 + "output_cost_per_token": 1e-6 }, "perplexity/llama-3.1-sonar-large-128k-online": { "deprecation_date": "2025-02-22", - "input_cost_per_token": 0.000001, + "input_cost_per_token": 1e-6, "litellm_provider": "perplexity", "max_input_tokens": 127072, "max_output_tokens": 127072, "max_tokens": 127072, "mode": "chat", - "output_cost_per_token": 0.000001 + "output_cost_per_token": 1e-6 }, "perplexity/llama-3.1-sonar-small-128k-chat": { "deprecation_date": "2025-02-22", @@ -21271,17 +19337,17 @@ "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.0000028 + "output_cost_per_token": 2.8e-6 }, "perplexity/pplx-70b-online": { "input_cost_per_request": 0.005, - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "perplexity", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.0000028 + "output_cost_per_token": 2.8e-6 }, "perplexity/pplx-7b-chat": { "input_cost_per_token": 7e-8, @@ -21294,7 +19360,7 @@ }, "perplexity/pplx-7b-online": { "input_cost_per_request": 0.005, - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "perplexity", "max_input_tokens": 4096, "max_output_tokens": 4096, @@ -21303,12 +19369,12 @@ "output_cost_per_token": 2.8e-7 }, "perplexity/sonar": { - "input_cost_per_token": 0.000001, + "input_cost_per_token": 1e-6, "litellm_provider": "perplexity", "max_input_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.000001, + "output_cost_per_token": 1e-6, "search_context_cost_per_query": { "search_context_size_high": 0.012, "search_context_size_low": 0.005, @@ -21317,14 +19383,14 @@ "supports_web_search": true }, "perplexity/sonar-deep-research": { - "citation_cost_per_token": 0.000002, - "input_cost_per_token": 0.000002, + "citation_cost_per_token": 2e-6, + "input_cost_per_token": 2e-6, "litellm_provider": "perplexity", "max_input_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_reasoning_token": 0.000003, - "output_cost_per_token": 0.000008, + "output_cost_per_reasoning_token": 3e-6, + "output_cost_per_token": 8e-6, "search_context_cost_per_query": { "search_context_size_high": 0.005, "search_context_size_low": 0.005, @@ -21340,7 +19406,7 @@ "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 0.0000018 + "output_cost_per_token": 1.8e-6 }, "perplexity/sonar-medium-online": { "input_cost_per_request": 0.005, @@ -21350,16 +19416,16 @@ "max_output_tokens": 12000, "max_tokens": 12000, "mode": "chat", - "output_cost_per_token": 0.0000018 + "output_cost_per_token": 1.8e-6 }, "perplexity/sonar-pro": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "perplexity", "max_input_tokens": 200000, "max_output_tokens": 8000, "max_tokens": 8000, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "search_context_cost_per_query": { "search_context_size_high": 0.014, "search_context_size_low": 0.006, @@ -21368,12 +19434,12 @@ "supports_web_search": true }, "perplexity/sonar-reasoning": { - "input_cost_per_token": 0.000001, + "input_cost_per_token": 1e-6, "litellm_provider": "perplexity", "max_input_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.000005, + "output_cost_per_token": 5e-6, "search_context_cost_per_query": { "search_context_size_high": 0.014, "search_context_size_low": 0.005, @@ -21383,12 +19449,12 @@ "supports_web_search": true }, "perplexity/sonar-reasoning-pro": { - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "perplexity", "max_input_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.000008, + "output_cost_per_token": 8e-6, "search_context_cost_per_query": { "search_context_size_high": 0.014, "search_context_size_low": 0.006, @@ -21423,7 +19489,7 @@ "max_output_tokens": 65536, "max_tokens": 262144, "mode": "chat", - "output_cost_per_token": 0.0000018, + "output_cost_per_token": 1.8e-6, "supports_function_calling": true, "supports_reasoning": true, "supports_tool_choice": true @@ -21447,7 +19513,7 @@ "max_output_tokens": 131072, "max_tokens": 262144, "mode": "chat", - "output_cost_per_token": 6e-7, + "output_cost_per_token": 6.0e-7, "supports_function_calling": true, "supports_reasoning": true, "supports_tool_choice": true @@ -21459,7 +19525,7 @@ "max_output_tokens": 16384, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 6e-7, + "output_cost_per_token": 6.0e-7, "supports_function_calling": true, "supports_reasoning": true, "supports_tool_choice": true @@ -21469,18 +19535,14 @@ "mode": "image_generation", "output_cost_per_image": 0.022, "source": "https://www.recraft.ai/docs#pricing", - "supported_endpoints": [ - "/v1/images/generations" - ] + "supported_endpoints": ["/v1/images/generations"] }, "recraft/recraftv3": { "litellm_provider": "recraft", "mode": "image_generation", "output_cost_per_image": 0.04, "source": "https://www.recraft.ai/docs#pricing", - "supported_endpoints": [ - "/v1/images/generations" - ] + "supported_endpoints": ["/v1/images/generations"] }, "replicate/meta/llama-2-13b": { "input_cost_per_token": 1e-7, @@ -21509,7 +19571,7 @@ "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.00000275, + "output_cost_per_token": 2.75e-6, "supports_tool_choice": true }, "replicate/meta/llama-2-70b-chat": { @@ -21519,7 +19581,7 @@ "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.00000275, + "output_cost_per_token": 2.75e-6, "supports_tool_choice": true }, "replicate/meta/llama-2-7b": { @@ -21549,7 +19611,7 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.00000275, + "output_cost_per_token": 2.75e-6, "supports_tool_choice": true }, "replicate/meta/llama-3-70b-instruct": { @@ -21559,7 +19621,7 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.00000275, + "output_cost_per_token": 2.75e-6, "supports_tool_choice": true }, "replicate/meta/llama-3-8b": { @@ -21609,140 +19671,140 @@ "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.000001, + "output_cost_per_token": 1e-6, "supports_tool_choice": true }, "rerank-english-v2.0": { "input_cost_per_query": 0.002, - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "cohere", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_query_tokens": 2048, "max_tokens": 4096, "mode": "rerank", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "rerank-english-v3.0": { "input_cost_per_query": 0.002, - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "cohere", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_query_tokens": 2048, "max_tokens": 4096, "mode": "rerank", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "rerank-multilingual-v2.0": { "input_cost_per_query": 0.002, - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "cohere", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_query_tokens": 2048, "max_tokens": 4096, "mode": "rerank", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "rerank-multilingual-v3.0": { "input_cost_per_query": 0.002, - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "cohere", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_query_tokens": 2048, "max_tokens": 4096, "mode": "rerank", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "rerank-v3.5": { "input_cost_per_query": 0.002, - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "cohere", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_query_tokens": 2048, "max_tokens": 4096, "mode": "rerank", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "nvidia_nim/nvidia/nv-rerankqa-mistral-4b-v3": { - "input_cost_per_query": 0, - "input_cost_per_token": 0, + "input_cost_per_query": 0.0, + "input_cost_per_token": 0.0, "litellm_provider": "nvidia_nim", "mode": "rerank", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "nvidia_nim/nvidia/llama-3_2-nv-rerankqa-1b-v2": { - "input_cost_per_query": 0, - "input_cost_per_token": 0, + "input_cost_per_query": 0.0, + "input_cost_per_token": 0.0, "litellm_provider": "nvidia_nim", "mode": "rerank", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "sagemaker/meta-textgeneration-llama-2-13b": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "sagemaker", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "completion", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "sagemaker/meta-textgeneration-llama-2-13b-f": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "sagemaker", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "sagemaker/meta-textgeneration-llama-2-70b": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "sagemaker", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "completion", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "sagemaker/meta-textgeneration-llama-2-70b-b-f": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "sagemaker", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "sagemaker/meta-textgeneration-llama-2-7b": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "sagemaker", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "completion", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "sagemaker/meta-textgeneration-llama-2-7b-f": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "sagemaker", "max_input_tokens": 4096, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "sambanova/DeepSeek-R1": { - "input_cost_per_token": 0.000005, + "input_cost_per_token": 5e-6, "litellm_provider": "sambanova", "max_input_tokens": 32768, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0.000007, + "output_cost_per_token": 7e-6, "source": "https://cloud.sambanova.ai/plans/pricing" }, "sambanova/DeepSeek-R1-Distill-Llama-70B": { @@ -21752,17 +19814,17 @@ "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0.0000014, + "output_cost_per_token": 1.4e-6, "source": "https://cloud.sambanova.ai/plans/pricing" }, "sambanova/DeepSeek-V3-0324": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "sambanova", "max_input_tokens": 32768, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0.0000045, + "output_cost_per_token": 4.5e-6, "source": "https://cloud.sambanova.ai/plans/pricing", "supports_function_calling": true, "supports_reasoning": true, @@ -21778,7 +19840,7 @@ "notes": "For vision models, images are converted to 6432 input tokens and are billed at that amount" }, "mode": "chat", - "output_cost_per_token": 0.0000018, + "output_cost_per_token": 1.8e-6, "source": "https://cloud.sambanova.ai/plans/pricing", "supports_function_calling": true, "supports_response_schema": true, @@ -21802,13 +19864,13 @@ "supports_tool_choice": true }, "sambanova/Meta-Llama-3.1-405B-Instruct": { - "input_cost_per_token": 0.000005, + "input_cost_per_token": 5e-6, "litellm_provider": "sambanova", "max_input_tokens": 16384, "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 0.00001, + "output_cost_per_token": 1e-5, "source": "https://cloud.sambanova.ai/plans/pricing", "supports_function_calling": true, "supports_response_schema": true, @@ -21854,7 +19916,7 @@ "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0.0000012, + "output_cost_per_token": 1.2e-6, "source": "https://cloud.sambanova.ai/plans/pricing", "supports_function_calling": true, "supports_response_schema": true, @@ -21877,7 +19939,7 @@ "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 0.000001, + "output_cost_per_token": 1e-6, "source": "https://cloud.sambanova.ai/plans/pricing" }, "sambanova/Qwen2-Audio-7B-Instruct": { @@ -21908,8 +19970,8 @@ "max_tokens": 32768, "max_input_tokens": 32768, "max_output_tokens": 32768, - "input_cost_per_token": 0.000003, - "output_cost_per_token": 0.0000045, + "input_cost_per_token": 3e-6, + "output_cost_per_token": 4.5e-6, "litellm_provider": "sambanova", "mode": "chat", "supports_function_calling": true, @@ -21921,8 +19983,8 @@ "max_tokens": 131072, "max_input_tokens": 131072, "max_output_tokens": 131072, - "input_cost_per_token": 0.000003, - "output_cost_per_token": 0.0000045, + "input_cost_per_token": 3e-6, + "output_cost_per_token": 4.5e-6, "litellm_provider": "sambanova", "mode": "chat", "supports_function_calling": true, @@ -21930,6 +19992,7 @@ "supports_reasoning": true, "source": "https://cloud.sambanova.ai/plans/pricing" }, + "snowflake/claude-3-5-sonnet": { "litellm_provider": "snowflake", "max_input_tokens": 18000, @@ -22146,19 +20209,19 @@ "input_cost_per_pixel": 3.81469e-8, "litellm_provider": "openai", "mode": "image_generation", - "output_cost_per_pixel": 0 + "output_cost_per_pixel": 0.0 }, "standard/1024-x-1792/dall-e-3": { "input_cost_per_pixel": 4.359e-8, "litellm_provider": "openai", "mode": "image_generation", - "output_cost_per_pixel": 0 + "output_cost_per_pixel": 0.0 }, "standard/1792-x-1024/dall-e-3": { "input_cost_per_pixel": 4.359e-8, "litellm_provider": "openai", "mode": "image_generation", - "output_cost_per_pixel": 0 + "output_cost_per_pixel": 0.0 }, "tavily/search": { "input_cost_per_query": 0.008, @@ -22225,23 +20288,23 @@ "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models" }, "text-completion-codestral/codestral-2405": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "text-completion-codestral", "max_input_tokens": 32000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "completion", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "source": "https://docs.mistral.ai/capabilities/code_generation/" }, "text-completion-codestral/codestral-latest": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "text-completion-codestral", "max_input_tokens": 32000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "completion", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "source": "https://docs.mistral.ai/capabilities/code_generation/" }, "text-embedding-004": { @@ -22273,8 +20336,8 @@ "max_input_tokens": 8191, "max_tokens": 8191, "mode": "embedding", - "output_cost_per_token": 0, - "output_cost_per_token_batches": 0, + "output_cost_per_token": 0.0, + "output_cost_per_token_batches": 0.0, "output_vector_size": 3072 }, "text-embedding-3-small": { @@ -22284,8 +20347,8 @@ "max_input_tokens": 8191, "max_tokens": 8191, "mode": "embedding", - "output_cost_per_token": 0, - "output_cost_per_token_batches": 0, + "output_cost_per_token": 0.0, + "output_cost_per_token_batches": 0.0, "output_vector_size": 1536 }, "text-embedding-ada-002": { @@ -22294,7 +20357,7 @@ "max_input_tokens": 8191, "max_tokens": 8191, "mode": "embedding", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "output_vector_size": 1536 }, "text-embedding-ada-002-v2": { @@ -22304,8 +20367,8 @@ "max_input_tokens": 8191, "max_tokens": 8191, "mode": "embedding", - "output_cost_per_token": 0, - "output_cost_per_token_batches": 0 + "output_cost_per_token": 0.0, + "output_cost_per_token_batches": 0.0 }, "text-embedding-large-exp-03-07": { "input_cost_per_character": 2.5e-8, @@ -22330,31 +20393,31 @@ "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing" }, "text-moderation-007": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "openai", "max_input_tokens": 32768, "max_output_tokens": 0, "max_tokens": 32768, "mode": "moderation", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "text-moderation-latest": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "openai", "max_input_tokens": 32768, "max_output_tokens": 0, "max_tokens": 32768, "mode": "moderation", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "text-moderation-stable": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "openai", "max_input_tokens": 32768, "max_output_tokens": 0, "max_tokens": 32768, "mode": "moderation", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "text-multilingual-embedding-002": { "input_cost_per_character": 2.5e-8, @@ -22378,23 +20441,23 @@ "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models" }, "text-unicorn": { - "input_cost_per_token": 0.00001, + "input_cost_per_token": 1e-5, "litellm_provider": "vertex_ai-text-models", "max_input_tokens": 8192, "max_output_tokens": 1024, "max_tokens": 1024, "mode": "completion", - "output_cost_per_token": 0.000028, + "output_cost_per_token": 2.8e-5, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models" }, "text-unicorn@001": { - "input_cost_per_token": 0.00001, + "input_cost_per_token": 1e-5, "litellm_provider": "vertex_ai-text-models", "max_input_tokens": 8192, "max_output_tokens": 1024, "max_tokens": 1024, "mode": "completion", - "output_cost_per_token": 0.000028, + "output_cost_per_token": 2.8e-5, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models" }, "textembedding-gecko": { @@ -22478,29 +20541,29 @@ "output_cost_per_token": 3e-7 }, "together-ai-81.1b-110b": { - "input_cost_per_token": 0.0000018, + "input_cost_per_token": 1.8e-6, "litellm_provider": "together_ai", "mode": "chat", - "output_cost_per_token": 0.0000018 + "output_cost_per_token": 1.8e-6 }, "together-ai-embedding-151m-to-350m": { "input_cost_per_token": 1.6e-8, "litellm_provider": "together_ai", "mode": "embedding", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "together-ai-embedding-up-to-150m": { "input_cost_per_token": 8e-9, "litellm_provider": "together_ai", "mode": "embedding", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "together_ai/baai/bge-base-en-v1.5": { "input_cost_per_token": 8e-9, "litellm_provider": "together_ai", "max_input_tokens": 512, "mode": "embedding", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "output_vector_size": 768 }, "together_ai/BAAI/bge-base-en-v1.5": { @@ -22508,7 +20571,7 @@ "litellm_provider": "together_ai", "max_input_tokens": 512, "mode": "embedding", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "output_vector_size": 768 }, "together-ai-up-to-4b": { @@ -22536,7 +20599,7 @@ "litellm_provider": "together_ai", "max_input_tokens": 262000, "mode": "chat", - "output_cost_per_token": 0.000006, + "output_cost_per_token": 6e-6, "source": "https://www.together.ai/models/qwen3-235b-a22b-instruct-2507-fp8", "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -22547,7 +20610,7 @@ "litellm_provider": "together_ai", "max_input_tokens": 256000, "mode": "chat", - "output_cost_per_token": 0.000003, + "output_cost_per_token": 3e-6, "source": "https://www.together.ai/models/qwen3-235b-a22b-thinking-2507", "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -22565,24 +20628,24 @@ "supports_tool_choice": false }, "together_ai/Qwen/Qwen3-Coder-480B-A35B-Instruct-FP8": { - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "together_ai", "max_input_tokens": 256000, "mode": "chat", - "output_cost_per_token": 0.000002, + "output_cost_per_token": 2e-6, "source": "https://www.together.ai/models/qwen3-coder-480b-a35b-instruct", "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_tool_choice": true }, "together_ai/deepseek-ai/DeepSeek-R1": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "together_ai", "max_input_tokens": 128000, "max_output_tokens": 20480, "max_tokens": 20480, "mode": "chat", - "output_cost_per_token": 0.000007, + "output_cost_per_token": 7e-6, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_tool_choice": true @@ -22592,20 +20655,20 @@ "litellm_provider": "together_ai", "max_input_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.00000219, + "output_cost_per_token": 2.19e-6, "source": "https://www.together.ai/models/deepseek-r1-0528-throughput", "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_tool_choice": true }, "together_ai/deepseek-ai/DeepSeek-V3": { - "input_cost_per_token": 0.00000125, + "input_cost_per_token": 1.25e-6, "litellm_provider": "together_ai", "max_input_tokens": 65536, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.00000125, + "output_cost_per_token": 1.25e-6, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_tool_choice": true @@ -22615,7 +20678,7 @@ "litellm_provider": "together_ai", "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.0000017, + "output_cost_per_token": 1.7e-6, "source": "https://www.together.ai/models/deepseek-v3-1", "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -22668,10 +20731,10 @@ "supports_tool_choice": true }, "together_ai/meta-llama/Meta-Llama-3.1-405B-Instruct-Turbo": { - "input_cost_per_token": 0.0000035, + "input_cost_per_token": 3.5e-6, "litellm_provider": "together_ai", "mode": "chat", - "output_cost_per_token": 0.0000035, + "output_cost_per_token": 3.5e-6, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_tool_choice": true @@ -22722,10 +20785,10 @@ "supports_tool_choice": true }, "together_ai/moonshotai/Kimi-K2-Instruct": { - "input_cost_per_token": 0.000001, + "input_cost_per_token": 1e-6, "litellm_provider": "together_ai", "mode": "chat", - "output_cost_per_token": 0.000003, + "output_cost_per_token": 3e-6, "source": "https://www.together.ai/models/kimi-k2-instruct", "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -22765,20 +20828,20 @@ "litellm_provider": "together_ai", "max_input_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.0000011, + "output_cost_per_token": 1.1e-6, "source": "https://www.together.ai/models/glm-4-5-air", "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_tool_choice": true }, "together_ai/zai-org/GLM-4.6": { - "input_cost_per_token": 6e-7, + "input_cost_per_token": 0.6e-6, "litellm_provider": "together_ai", "max_input_tokens": 200000, "max_output_tokens": 200000, "max_tokens": 200000, "mode": "chat", - "output_cost_per_token": 0.0000022, + "output_cost_per_token": 2.2e-6, "source": "https://www.together.ai/models/glm-4-6", "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -22786,11 +20849,11 @@ "supports_tool_choice": true }, "together_ai/moonshotai/Kimi-K2-Instruct-0905": { - "input_cost_per_token": 0.000001, + "input_cost_per_token": 1e-6, "litellm_provider": "together_ai", "max_input_tokens": 262144, "mode": "chat", - "output_cost_per_token": 0.000003, + "output_cost_per_token": 3e-6, "source": "https://www.together.ai/models/kimi-k2-0905", "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -22801,7 +20864,7 @@ "litellm_provider": "together_ai", "max_input_tokens": 262144, "mode": "chat", - "output_cost_per_token": 0.0000015, + "output_cost_per_token": 1.5e-6, "source": "https://www.together.ai/models/qwen3-next-80b-a3b-instruct", "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -22812,27 +20875,23 @@ "litellm_provider": "together_ai", "max_input_tokens": 262144, "mode": "chat", - "output_cost_per_token": 0.0000015, + "output_cost_per_token": 1.5e-6, "source": "https://www.together.ai/models/qwen3-next-80b-a3b-thinking", "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_tool_choice": true }, "tts-1": { - "input_cost_per_character": 0.000015, + "input_cost_per_character": 1.5e-5, "litellm_provider": "openai", "mode": "audio_speech", - "supported_endpoints": [ - "/v1/audio/speech" - ] + "supported_endpoints": ["/v1/audio/speech"] }, "tts-1-hd": { - "input_cost_per_character": 0.00003, + "input_cost_per_character": 3e-5, "litellm_provider": "openai", "mode": "audio_speech", - "supported_endpoints": [ - "/v1/audio/speech" - ] + "supported_endpoints": ["/v1/audio/speech"] }, "us.amazon.nova-lite-v1:0": { "input_cost_per_token": 6e-8, @@ -22861,13 +20920,13 @@ "supports_response_schema": true }, "us.amazon.nova-premier-v1:0": { - "input_cost_per_token": 0.0000025, + "input_cost_per_token": 2.5e-6, "litellm_provider": "bedrock_converse", "max_input_tokens": 1000000, "max_output_tokens": 10000, "max_tokens": 10000, "mode": "chat", - "output_cost_per_token": 0.0000125, + "output_cost_per_token": 1.25e-5, "supports_function_calling": true, "supports_pdf_input": true, "supports_prompt_caching": false, @@ -22881,7 +20940,7 @@ "max_output_tokens": 10000, "max_tokens": 10000, "mode": "chat", - "output_cost_per_token": 0.0000032, + "output_cost_per_token": 3.2e-6, "supports_function_calling": true, "supports_pdf_input": true, "supports_prompt_caching": true, @@ -22889,7 +20948,7 @@ "supports_vision": true }, "us.anthropic.claude-3-5-haiku-20241022-v1:0": { - "cache_creation_input_token_cost": 0.000001, + "cache_creation_input_token_cost": 1e-6, "cache_read_input_token_cost": 8e-8, "input_cost_per_token": 8e-7, "litellm_provider": "bedrock", @@ -22897,7 +20956,7 @@ "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.000004, + "output_cost_per_token": 4e-6, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_pdf_input": true, @@ -22906,15 +20965,15 @@ "supports_tool_choice": true }, "us.anthropic.claude-haiku-4-5-20251001-v1:0": { - "cache_creation_input_token_cost": 0.000001375, + "cache_creation_input_token_cost": 1.375e-6, "cache_read_input_token_cost": 1.1e-7, - "input_cost_per_token": 0.0000011, + "input_cost_per_token": 1.1e-6, "litellm_provider": "bedrock_converse", "max_input_tokens": 200000, "max_output_tokens": 64000, "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 0.0000055, + "output_cost_per_token": 5.5e-6, "source": "https://aws.amazon.com/about-aws/whats-new/2025/10/claude-4-5-haiku-anthropic-amazon-bedrock", "supports_assistant_prefill": true, "supports_computer_use": true, @@ -22928,13 +20987,13 @@ "tool_use_system_prompt_tokens": 346 }, "us.anthropic.claude-3-5-sonnet-20240620-v1:0": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "bedrock", "max_input_tokens": 200000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "supports_function_calling": true, "supports_pdf_input": true, "supports_response_schema": true, @@ -22942,15 +21001,15 @@ "supports_vision": true }, "us.anthropic.claude-3-5-sonnet-20241022-v2:0": { - "cache_creation_input_token_cost": 0.00000375, + "cache_creation_input_token_cost": 3.75e-6, "cache_read_input_token_cost": 3e-7, - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "bedrock", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "supports_assistant_prefill": true, "supports_computer_use": true, "supports_function_calling": true, @@ -22961,15 +21020,15 @@ "supports_vision": true }, "us.anthropic.claude-3-7-sonnet-20250219-v1:0": { - "cache_creation_input_token_cost": 0.00000375, + "cache_creation_input_token_cost": 3.75e-6, "cache_read_input_token_cost": 3e-7, - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "bedrock_converse", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "supports_assistant_prefill": true, "supports_computer_use": true, "supports_function_calling": true, @@ -22987,7 +21046,7 @@ "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.00000125, + "output_cost_per_token": 1.25e-6, "supports_function_calling": true, "supports_pdf_input": true, "supports_response_schema": true, @@ -22995,26 +21054,26 @@ "supports_vision": true }, "us.anthropic.claude-3-opus-20240229-v1:0": { - "input_cost_per_token": 0.000015, + "input_cost_per_token": 1.5e-5, "litellm_provider": "bedrock", "max_input_tokens": 200000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.000075, + "output_cost_per_token": 7.5e-5, "supports_function_calling": true, "supports_response_schema": true, "supports_tool_choice": true, "supports_vision": true }, "us.anthropic.claude-3-sonnet-20240229-v1:0": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "bedrock", "max_input_tokens": 200000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "supports_function_calling": true, "supports_pdf_input": true, "supports_response_schema": true, @@ -23022,15 +21081,15 @@ "supports_vision": true }, "us.anthropic.claude-opus-4-1-20250805-v1:0": { - "cache_creation_input_token_cost": 0.00001875, - "cache_read_input_token_cost": 0.0000015, - "input_cost_per_token": 0.000015, + "cache_creation_input_token_cost": 1.875e-5, + "cache_read_input_token_cost": 1.5e-6, + "input_cost_per_token": 1.5e-5, "litellm_provider": "bedrock_converse", "max_input_tokens": 200000, "max_output_tokens": 32000, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 0.000075, + "output_cost_per_token": 7.5e-5, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -23048,19 +21107,19 @@ "tool_use_system_prompt_tokens": 159 }, "us.anthropic.claude-sonnet-4-5-20250929-v1:0": { - "cache_creation_input_token_cost": 0.000004125, + "cache_creation_input_token_cost": 4.125e-6, "cache_read_input_token_cost": 3.3e-7, - "input_cost_per_token": 0.0000033, - "input_cost_per_token_above_200k_tokens": 0.0000066, - "output_cost_per_token_above_200k_tokens": 0.00002475, - "cache_creation_input_token_cost_above_200k_tokens": 0.00000825, + "input_cost_per_token": 3.3e-6, + "input_cost_per_token_above_200k_tokens": 6.6e-6, + "output_cost_per_token_above_200k_tokens": 2.475e-5, + "cache_creation_input_token_cost_above_200k_tokens": 8.25e-6, "cache_read_input_token_cost_above_200k_tokens": 6.6e-7, "litellm_provider": "bedrock_converse", "max_input_tokens": 200000, "max_output_tokens": 64000, "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 0.0000165, + "output_cost_per_token": 1.65e-5, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -23078,15 +21137,15 @@ "tool_use_system_prompt_tokens": 346 }, "au.anthropic.claude-haiku-4-5-20251001-v1:0": { - "cache_creation_input_token_cost": 0.000001375, + "cache_creation_input_token_cost": 1.375e-6, "cache_read_input_token_cost": 1.1e-7, - "input_cost_per_token": 0.0000011, + "input_cost_per_token": 1.1e-6, "litellm_provider": "bedrock_converse", "max_input_tokens": 200000, "max_output_tokens": 64000, "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 0.0000055, + "output_cost_per_token": 5.5e-6, "supports_assistant_prefill": true, "supports_computer_use": true, "supports_function_calling": true, @@ -23099,15 +21158,15 @@ "tool_use_system_prompt_tokens": 346 }, "us.anthropic.claude-opus-4-20250514-v1:0": { - "cache_creation_input_token_cost": 0.00001875, - "cache_read_input_token_cost": 0.0000015, - "input_cost_per_token": 0.000015, + "cache_creation_input_token_cost": 1.875e-5, + "cache_read_input_token_cost": 1.5e-6, + "input_cost_per_token": 1.5e-5, "litellm_provider": "bedrock_converse", "max_input_tokens": 200000, "max_output_tokens": 32000, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 0.000075, + "output_cost_per_token": 7.5e-5, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -23125,19 +21184,19 @@ "tool_use_system_prompt_tokens": 159 }, "us.anthropic.claude-sonnet-4-20250514-v1:0": { - "cache_creation_input_token_cost": 0.00000375, + "cache_creation_input_token_cost": 3.75e-6, "cache_read_input_token_cost": 3e-7, - "input_cost_per_token": 0.000003, - "input_cost_per_token_above_200k_tokens": 0.000006, - "output_cost_per_token_above_200k_tokens": 0.0000225, - "cache_creation_input_token_cost_above_200k_tokens": 0.0000075, + "input_cost_per_token": 3e-6, + "input_cost_per_token_above_200k_tokens": 6e-6, + "output_cost_per_token_above_200k_tokens": 2.25e-5, + "cache_creation_input_token_cost_above_200k_tokens": 7.5e-6, "cache_read_input_token_cost_above_200k_tokens": 6e-7, "litellm_provider": "bedrock_converse", "max_input_tokens": 1000000, "max_output_tokens": 64000, "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -23155,25 +21214,25 @@ "tool_use_system_prompt_tokens": 159 }, "us.deepseek.r1-v1:0": { - "input_cost_per_token": 0.00000135, + "input_cost_per_token": 1.35e-6, "litellm_provider": "bedrock_converse", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.0000054, + "output_cost_per_token": 5.4e-6, "supports_function_calling": false, "supports_reasoning": true, "supports_tool_choice": false }, "us.meta.llama3-1-405b-instruct-v1:0": { - "input_cost_per_token": 0.00000532, + "input_cost_per_token": 5.32e-6, "litellm_provider": "bedrock", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.000016, + "output_cost_per_token": 1.6e-5, "supports_function_calling": true, "supports_tool_choice": false }, @@ -23234,13 +21293,13 @@ "supports_tool_choice": false }, "us.meta.llama3-2-90b-instruct-v1:0": { - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "bedrock", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.000002, + "output_cost_per_token": 2e-6, "supports_function_calling": true, "supports_tool_choice": false, "supports_vision": true @@ -23266,14 +21325,8 @@ "mode": "chat", "output_cost_per_token": 9.7e-7, "output_cost_per_token_batches": 4.85e-7, - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text", - "code" - ], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text", "code"], "supports_function_calling": true, "supports_tool_choice": false }, @@ -23287,36 +21340,30 @@ "mode": "chat", "output_cost_per_token": 6.6e-7, "output_cost_per_token_batches": 3.3e-7, - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text", - "code" - ], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text", "code"], "supports_function_calling": true, "supports_tool_choice": false }, "us.mistral.pixtral-large-2502-v1:0": { - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "bedrock_converse", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.000006, + "output_cost_per_token": 6e-6, "supports_function_calling": true, "supports_tool_choice": false }, "v0/v0-1.0-md": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "v0", "max_input_tokens": 128000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_system_messages": true, @@ -23324,13 +21371,13 @@ "supports_vision": true }, "v0/v0-1.5-lg": { - "input_cost_per_token": 0.000015, + "input_cost_per_token": 1.5e-5, "litellm_provider": "v0", "max_input_tokens": 512000, "max_output_tokens": 512000, "max_tokens": 512000, "mode": "chat", - "output_cost_per_token": 0.000075, + "output_cost_per_token": 7.5e-5, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_system_messages": true, @@ -23338,13 +21385,13 @@ "supports_vision": true }, "v0/v0-1.5-md": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "v0", "max_input_tokens": 128000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_system_messages": true, @@ -23394,7 +21441,7 @@ "max_output_tokens": 66536, "max_tokens": 262144, "mode": "chat", - "output_cost_per_token": 0.0000016 + "output_cost_per_token": 1.6e-6 }, "vercel_ai_gateway/amazon/nova-lite": { "input_cost_per_token": 6e-8, @@ -23421,7 +21468,7 @@ "max_output_tokens": 8192, "max_tokens": 300000, "mode": "chat", - "output_cost_per_token": 0.0000032 + "output_cost_per_token": 3.2e-6 }, "vercel_ai_gateway/amazon/titan-embed-text-v2": { "input_cost_per_token": 2e-8, @@ -23430,7 +21477,7 @@ "max_output_tokens": 0, "max_tokens": 0, "mode": "chat", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "vercel_ai_gateway/anthropic/claude-3-haiku": { "cache_creation_input_token_cost": 3e-7, @@ -23441,21 +21488,21 @@ "max_output_tokens": 4096, "max_tokens": 200000, "mode": "chat", - "output_cost_per_token": 0.00000125 + "output_cost_per_token": 1.25e-6 }, "vercel_ai_gateway/anthropic/claude-3-opus": { - "cache_creation_input_token_cost": 0.00001875, - "cache_read_input_token_cost": 0.0000015, - "input_cost_per_token": 0.000015, + "cache_creation_input_token_cost": 1.875e-5, + "cache_read_input_token_cost": 1.5e-6, + "input_cost_per_token": 1.5e-5, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 200000, "max_output_tokens": 4096, "max_tokens": 200000, "mode": "chat", - "output_cost_per_token": 0.000075 + "output_cost_per_token": 7.5e-5 }, "vercel_ai_gateway/anthropic/claude-3.5-haiku": { - "cache_creation_input_token_cost": 0.000001, + "cache_creation_input_token_cost": 1e-6, "cache_read_input_token_cost": 8e-8, "input_cost_per_token": 8e-7, "litellm_provider": "vercel_ai_gateway", @@ -23463,60 +21510,60 @@ "max_output_tokens": 8192, "max_tokens": 200000, "mode": "chat", - "output_cost_per_token": 0.000004 + "output_cost_per_token": 4e-6 }, "vercel_ai_gateway/anthropic/claude-3.5-sonnet": { - "cache_creation_input_token_cost": 0.00000375, + "cache_creation_input_token_cost": 3.75e-6, "cache_read_input_token_cost": 3e-7, - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 200000, "mode": "chat", - "output_cost_per_token": 0.000015 + "output_cost_per_token": 1.5e-5 }, "vercel_ai_gateway/anthropic/claude-3.7-sonnet": { - "cache_creation_input_token_cost": 0.00000375, + "cache_creation_input_token_cost": 3.75e-6, "cache_read_input_token_cost": 3e-7, - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 200000, "max_output_tokens": 64000, "max_tokens": 200000, "mode": "chat", - "output_cost_per_token": 0.000015 + "output_cost_per_token": 1.5e-5 }, "vercel_ai_gateway/anthropic/claude-4-opus": { - "cache_creation_input_token_cost": 0.00001875, - "cache_read_input_token_cost": 0.0000015, - "input_cost_per_token": 0.000015, + "cache_creation_input_token_cost": 1.875e-5, + "cache_read_input_token_cost": 1.5e-6, + "input_cost_per_token": 1.5e-5, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 200000, "max_output_tokens": 32000, "max_tokens": 200000, "mode": "chat", - "output_cost_per_token": 0.000075 + "output_cost_per_token": 7.5e-5 }, "vercel_ai_gateway/anthropic/claude-4-sonnet": { - "cache_creation_input_token_cost": 0.00000375, + "cache_creation_input_token_cost": 3.75e-6, "cache_read_input_token_cost": 3e-7, - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 200000, "max_output_tokens": 64000, "max_tokens": 200000, "mode": "chat", - "output_cost_per_token": 0.000015 + "output_cost_per_token": 1.5e-5 }, "vercel_ai_gateway/cohere/command-a": { - "input_cost_per_token": 0.0000025, + "input_cost_per_token": 2.5e-6, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 256000, "max_output_tokens": 8000, "max_tokens": 256000, "mode": "chat", - "output_cost_per_token": 0.00001 + "output_cost_per_token": 1e-5 }, "vercel_ai_gateway/cohere/command-r": { "input_cost_per_token": 1.5e-7, @@ -23528,13 +21575,13 @@ "output_cost_per_token": 6e-7 }, "vercel_ai_gateway/cohere/command-r-plus": { - "input_cost_per_token": 0.0000025, + "input_cost_per_token": 2.5e-6, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.00001 + "output_cost_per_token": 1e-5 }, "vercel_ai_gateway/cohere/embed-v4.0": { "input_cost_per_token": 1.2e-7, @@ -23543,7 +21590,7 @@ "max_output_tokens": 0, "max_tokens": 0, "mode": "chat", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "vercel_ai_gateway/deepseek/deepseek-r1": { "input_cost_per_token": 5.5e-7, @@ -23552,7 +21599,7 @@ "max_output_tokens": 8192, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.00000219 + "output_cost_per_token": 2.19e-6 }, "vercel_ai_gateway/deepseek/deepseek-r1-distill-llama-70b": { "input_cost_per_token": 7.5e-7, @@ -23597,16 +21644,16 @@ "max_output_tokens": 65536, "max_tokens": 1000000, "mode": "chat", - "output_cost_per_token": 0.0000025 + "output_cost_per_token": 2.5e-6 }, "vercel_ai_gateway/google/gemini-2.5-pro": { - "input_cost_per_token": 0.0000025, + "input_cost_per_token": 2.5e-6, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 1048576, "max_output_tokens": 65536, "max_tokens": 1048576, "mode": "chat", - "output_cost_per_token": 0.00001 + "output_cost_per_token": 1e-5 }, "vercel_ai_gateway/google/gemini-embedding-001": { "input_cost_per_token": 1.5e-7, @@ -23615,7 +21662,7 @@ "max_output_tokens": 0, "max_tokens": 0, "mode": "embedding", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "vercel_ai_gateway/google/gemma-2-9b": { "input_cost_per_token": 2e-7, @@ -23633,7 +21680,7 @@ "max_output_tokens": 0, "max_tokens": 0, "mode": "embedding", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "vercel_ai_gateway/google/text-multilingual-embedding-002": { "input_cost_per_token": 2.5e-8, @@ -23642,7 +21689,7 @@ "max_output_tokens": 0, "max_tokens": 0, "mode": "embedding", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "vercel_ai_gateway/inception/mercury-coder-small": { "input_cost_per_token": 2.5e-7, @@ -23651,7 +21698,7 @@ "max_output_tokens": 16384, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 0.000001 + "output_cost_per_token": 1e-6 }, "vercel_ai_gateway/meta/llama-3-70b": { "input_cost_per_token": 5.9e-7, @@ -23768,7 +21815,7 @@ "max_output_tokens": 0, "max_tokens": 0, "mode": "chat", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "vercel_ai_gateway/mistral/devstral-small": { "input_cost_per_token": 7e-8, @@ -23780,13 +21827,13 @@ "output_cost_per_token": 2.8e-7 }, "vercel_ai_gateway/mistral/magistral-medium": { - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 128000, "max_output_tokens": 64000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.000005 + "output_cost_per_token": 5e-6 }, "vercel_ai_gateway/mistral/magistral-small": { "input_cost_per_token": 5e-7, @@ -23795,7 +21842,7 @@ "max_output_tokens": 64000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.0000015 + "output_cost_per_token": 1.5e-6 }, "vercel_ai_gateway/mistral/ministral-3b": { "input_cost_per_token": 4e-8, @@ -23822,16 +21869,16 @@ "max_output_tokens": 0, "max_tokens": 0, "mode": "chat", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "vercel_ai_gateway/mistral/mistral-large": { - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 32000, "max_output_tokens": 4000, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 0.000006 + "output_cost_per_token": 6e-6 }, "vercel_ai_gateway/mistral/mistral-saba-24b": { "input_cost_per_token": 7.9e-7, @@ -23852,13 +21899,13 @@ "output_cost_per_token": 3e-7 }, "vercel_ai_gateway/mistral/mixtral-8x22b-instruct": { - "input_cost_per_token": 0.0000012, + "input_cost_per_token": 1.2e-6, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 65536, "max_output_tokens": 2048, "max_tokens": 65536, "mode": "chat", - "output_cost_per_token": 0.0000012 + "output_cost_per_token": 1.2e-6 }, "vercel_ai_gateway/mistral/pixtral-12b": { "input_cost_per_token": 1.5e-7, @@ -23870,13 +21917,13 @@ "output_cost_per_token": 1.5e-7 }, "vercel_ai_gateway/mistral/pixtral-large": { - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 128000, "max_output_tokens": 4000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.000006 + "output_cost_per_token": 6e-6 }, "vercel_ai_gateway/moonshotai/kimi-k2": { "input_cost_per_token": 5.5e-7, @@ -23885,7 +21932,7 @@ "max_output_tokens": 16384, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0.0000022 + "output_cost_per_token": 2.2e-6 }, "vercel_ai_gateway/morph/morph-v3-fast": { "input_cost_per_token": 8e-7, @@ -23894,7 +21941,7 @@ "max_output_tokens": 16384, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0.0000012 + "output_cost_per_token": 1.2e-6 }, "vercel_ai_gateway/morph/morph-v3-large": { "input_cost_per_token": 9e-7, @@ -23903,7 +21950,7 @@ "max_output_tokens": 16384, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0.0000019 + "output_cost_per_token": 1.9e-6 }, "vercel_ai_gateway/openai/gpt-3.5-turbo": { "input_cost_per_token": 5e-7, @@ -23912,39 +21959,39 @@ "max_output_tokens": 4096, "max_tokens": 16385, "mode": "chat", - "output_cost_per_token": 0.0000015 + "output_cost_per_token": 1.5e-6 }, "vercel_ai_gateway/openai/gpt-3.5-turbo-instruct": { - "input_cost_per_token": 0.0000015, + "input_cost_per_token": 1.5e-6, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 8192, "max_output_tokens": 4096, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.000002 + "output_cost_per_token": 2e-6 }, "vercel_ai_gateway/openai/gpt-4-turbo": { - "input_cost_per_token": 0.00001, + "input_cost_per_token": 1e-5, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 128000, "max_output_tokens": 4096, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.00003 + "output_cost_per_token": 3e-5 }, "vercel_ai_gateway/openai/gpt-4.1": { - "cache_creation_input_token_cost": 0, + "cache_creation_input_token_cost": 0.0, "cache_read_input_token_cost": 5e-7, - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 1047576, "max_output_tokens": 32768, "max_tokens": 1047576, "mode": "chat", - "output_cost_per_token": 0.000008 + "output_cost_per_token": 8e-6 }, "vercel_ai_gateway/openai/gpt-4.1-mini": { - "cache_creation_input_token_cost": 0, + "cache_creation_input_token_cost": 0.0, "cache_read_input_token_cost": 1e-7, "input_cost_per_token": 4e-7, "litellm_provider": "vercel_ai_gateway", @@ -23952,10 +21999,10 @@ "max_output_tokens": 32768, "max_tokens": 1047576, "mode": "chat", - "output_cost_per_token": 0.0000016 + "output_cost_per_token": 1.6e-6 }, "vercel_ai_gateway/openai/gpt-4.1-nano": { - "cache_creation_input_token_cost": 0, + "cache_creation_input_token_cost": 0.0, "cache_read_input_token_cost": 2.5e-8, "input_cost_per_token": 1e-7, "litellm_provider": "vercel_ai_gateway", @@ -23966,18 +22013,18 @@ "output_cost_per_token": 4e-7 }, "vercel_ai_gateway/openai/gpt-4o": { - "cache_creation_input_token_cost": 0, - "cache_read_input_token_cost": 0.00000125, - "input_cost_per_token": 0.0000025, + "cache_creation_input_token_cost": 0.0, + "cache_read_input_token_cost": 1.25e-6, + "input_cost_per_token": 2.5e-6, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 128000, "max_output_tokens": 16384, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.00001 + "output_cost_per_token": 1e-5 }, "vercel_ai_gateway/openai/gpt-4o-mini": { - "cache_creation_input_token_cost": 0, + "cache_creation_input_token_cost": 0.0, "cache_read_input_token_cost": 7.5e-8, "input_cost_per_token": 1.5e-7, "litellm_provider": "vercel_ai_gateway", @@ -23988,48 +22035,48 @@ "output_cost_per_token": 6e-7 }, "vercel_ai_gateway/openai/o1": { - "cache_creation_input_token_cost": 0, - "cache_read_input_token_cost": 0.0000075, - "input_cost_per_token": 0.000015, + "cache_creation_input_token_cost": 0.0, + "cache_read_input_token_cost": 7.5e-6, + "input_cost_per_token": 1.5e-5, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 200000, "mode": "chat", - "output_cost_per_token": 0.00006 + "output_cost_per_token": 6e-5 }, "vercel_ai_gateway/openai/o3": { - "cache_creation_input_token_cost": 0, + "cache_creation_input_token_cost": 0.0, "cache_read_input_token_cost": 5e-7, - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 200000, "mode": "chat", - "output_cost_per_token": 0.000008 + "output_cost_per_token": 8e-6 }, "vercel_ai_gateway/openai/o3-mini": { - "cache_creation_input_token_cost": 0, + "cache_creation_input_token_cost": 0.0, "cache_read_input_token_cost": 5.5e-7, - "input_cost_per_token": 0.0000011, + "input_cost_per_token": 1.1e-6, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 200000, "mode": "chat", - "output_cost_per_token": 0.0000044 + "output_cost_per_token": 4.4e-6 }, "vercel_ai_gateway/openai/o4-mini": { - "cache_creation_input_token_cost": 0, + "cache_creation_input_token_cost": 0.0, "cache_read_input_token_cost": 2.75e-7, - "input_cost_per_token": 0.0000011, + "input_cost_per_token": 1.1e-6, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 200000, "max_output_tokens": 100000, "max_tokens": 200000, "mode": "chat", - "output_cost_per_token": 0.0000044 + "output_cost_per_token": 4.4e-6 }, "vercel_ai_gateway/openai/text-embedding-3-large": { "input_cost_per_token": 1.3e-7, @@ -24038,7 +22085,7 @@ "max_output_tokens": 0, "max_tokens": 0, "mode": "embedding", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "vercel_ai_gateway/openai/text-embedding-3-small": { "input_cost_per_token": 2e-8, @@ -24047,7 +22094,7 @@ "max_output_tokens": 0, "max_tokens": 0, "mode": "embedding", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "vercel_ai_gateway/openai/text-embedding-ada-002": { "input_cost_per_token": 1e-7, @@ -24056,97 +22103,97 @@ "max_output_tokens": 0, "max_tokens": 0, "mode": "embedding", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "vercel_ai_gateway/perplexity/sonar": { - "input_cost_per_token": 0.000001, + "input_cost_per_token": 1e-6, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 127000, "max_output_tokens": 8000, "max_tokens": 127000, "mode": "chat", - "output_cost_per_token": 0.000001 + "output_cost_per_token": 1e-6 }, "vercel_ai_gateway/perplexity/sonar-pro": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 200000, "max_output_tokens": 8000, "max_tokens": 200000, "mode": "chat", - "output_cost_per_token": 0.000015 + "output_cost_per_token": 1.5e-5 }, "vercel_ai_gateway/perplexity/sonar-reasoning": { - "input_cost_per_token": 0.000001, + "input_cost_per_token": 1e-6, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 127000, "max_output_tokens": 8000, "max_tokens": 127000, "mode": "chat", - "output_cost_per_token": 0.000005 + "output_cost_per_token": 5e-6 }, "vercel_ai_gateway/perplexity/sonar-reasoning-pro": { - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 127000, "max_output_tokens": 8000, "max_tokens": 127000, "mode": "chat", - "output_cost_per_token": 0.000008 + "output_cost_per_token": 8e-6 }, "vercel_ai_gateway/vercel/v0-1.0-md": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 128000, "max_output_tokens": 32000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.000015 + "output_cost_per_token": 1.5e-5 }, "vercel_ai_gateway/vercel/v0-1.5-md": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 128000, "max_output_tokens": 32768, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.000015 + "output_cost_per_token": 1.5e-5 }, "vercel_ai_gateway/xai/grok-2": { - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 131072, "max_output_tokens": 4000, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0.00001 + "output_cost_per_token": 1e-5 }, "vercel_ai_gateway/xai/grok-2-vision": { - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 32768, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0.00001 + "output_cost_per_token": 1e-5 }, "vercel_ai_gateway/xai/grok-3": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0.000015 + "output_cost_per_token": 1.5e-5 }, "vercel_ai_gateway/xai/grok-3-fast": { - "input_cost_per_token": 0.000005, + "input_cost_per_token": 5e-6, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0.000025 + "output_cost_per_token": 2.5e-5 }, "vercel_ai_gateway/xai/grok-3-mini": { "input_cost_per_token": 3e-7, @@ -24164,16 +22211,16 @@ "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0.000004 + "output_cost_per_token": 4e-6 }, "vercel_ai_gateway/xai/grok-4": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "vercel_ai_gateway", "max_input_tokens": 256000, "max_output_tokens": 256000, "max_tokens": 256000, "mode": "chat", - "output_cost_per_token": 0.000015 + "output_cost_per_token": 1.5e-5 }, "vercel_ai_gateway/zai/glm-4.5": { "input_cost_per_token": 6e-7, @@ -24182,7 +22229,7 @@ "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0.0000022 + "output_cost_per_token": 2.2e-6 }, "vercel_ai_gateway/zai/glm-4.5-air": { "input_cost_per_token": 2e-7, @@ -24191,7 +22238,7 @@ "max_output_tokens": 96000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.0000011 + "output_cost_per_token": 1.1e-6 }, "vercel_ai_gateway/zai/glm-4.6": { "litellm_provider": "vercel_ai_gateway", @@ -24201,48 +22248,48 @@ "max_output_tokens": 200000, "max_tokens": 200000, "mode": "chat", - "output_cost_per_token": 0.0000018, + "output_cost_per_token": 1.8e-6, "source": "https://vercel.com/ai-gateway/models/glm-4.6", "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_tool_choice": true }, "vertex_ai/claude-3-5-haiku": { - "input_cost_per_token": 0.000001, + "input_cost_per_token": 1e-6, "litellm_provider": "vertex_ai-anthropic_models", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.000005, + "output_cost_per_token": 5e-6, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_pdf_input": true, "supports_tool_choice": true }, "vertex_ai/claude-3-5-haiku@20241022": { - "input_cost_per_token": 0.000001, + "input_cost_per_token": 1e-6, "litellm_provider": "vertex_ai-anthropic_models", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.000005, + "output_cost_per_token": 5e-6, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_pdf_input": true, "supports_tool_choice": true }, "vertex_ai/claude-haiku-4-5@20251001": { - "cache_creation_input_token_cost": 0.00000125, + "cache_creation_input_token_cost": 1.25e-6, "cache_read_input_token_cost": 1e-7, - "input_cost_per_token": 0.000001, + "input_cost_per_token": 1e-6, "litellm_provider": "vertex_ai-anthropic_models", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.000005, + "output_cost_per_token": 5e-6, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/partner-models/claude/haiku-4-5", "supports_assistant_prefill": true, "supports_function_calling": true, @@ -24253,13 +22300,13 @@ "supports_tool_choice": true }, "vertex_ai/claude-3-5-sonnet": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "vertex_ai-anthropic_models", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "supports_assistant_prefill": true, "supports_computer_use": true, "supports_function_calling": true, @@ -24268,13 +22315,13 @@ "supports_vision": true }, "vertex_ai/claude-3-5-sonnet-v2": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "vertex_ai-anthropic_models", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "supports_assistant_prefill": true, "supports_computer_use": true, "supports_function_calling": true, @@ -24283,13 +22330,13 @@ "supports_vision": true }, "vertex_ai/claude-3-5-sonnet-v2@20241022": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "vertex_ai-anthropic_models", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "supports_assistant_prefill": true, "supports_computer_use": true, "supports_function_calling": true, @@ -24298,13 +22345,13 @@ "supports_vision": true }, "vertex_ai/claude-3-5-sonnet@20240620": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "vertex_ai-anthropic_models", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_pdf_input": true, @@ -24312,16 +22359,16 @@ "supports_vision": true }, "vertex_ai/claude-3-7-sonnet@20250219": { - "cache_creation_input_token_cost": 0.00000375, + "cache_creation_input_token_cost": 3.75e-6, "cache_read_input_token_cost": 3e-7, "deprecation_date": "2025-06-01", - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "vertex_ai-anthropic_models", "max_input_tokens": 200000, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "supports_assistant_prefill": true, "supports_computer_use": true, "supports_function_calling": true, @@ -24340,7 +22387,7 @@ "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.00000125, + "output_cost_per_token": 1.25e-6, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_tool_choice": true, @@ -24353,74 +22400,74 @@ "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.00000125, + "output_cost_per_token": 1.25e-6, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_tool_choice": true, "supports_vision": true }, "vertex_ai/claude-3-opus": { - "input_cost_per_token": 0.000015, + "input_cost_per_token": 1.5e-5, "litellm_provider": "vertex_ai-anthropic_models", "max_input_tokens": 200000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.000075, + "output_cost_per_token": 7.5e-5, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_tool_choice": true, "supports_vision": true }, "vertex_ai/claude-3-opus@20240229": { - "input_cost_per_token": 0.000015, + "input_cost_per_token": 1.5e-5, "litellm_provider": "vertex_ai-anthropic_models", "max_input_tokens": 200000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.000075, + "output_cost_per_token": 7.5e-5, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_tool_choice": true, "supports_vision": true }, "vertex_ai/claude-3-sonnet": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "vertex_ai-anthropic_models", "max_input_tokens": 200000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_tool_choice": true, "supports_vision": true }, "vertex_ai/claude-3-sonnet@20240229": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "vertex_ai-anthropic_models", "max_input_tokens": 200000, "max_output_tokens": 4096, "max_tokens": 4096, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_tool_choice": true, "supports_vision": true }, "vertex_ai/claude-opus-4": { - "cache_creation_input_token_cost": 0.00001875, - "cache_read_input_token_cost": 0.0000015, - "input_cost_per_token": 0.000015, + "cache_creation_input_token_cost": 1.875e-5, + "cache_read_input_token_cost": 1.5e-6, + "input_cost_per_token": 1.5e-5, "litellm_provider": "vertex_ai-anthropic_models", "max_input_tokens": 200000, "max_output_tokens": 32000, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 0.000075, + "output_cost_per_token": 7.5e-5, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -24438,55 +22485,55 @@ "tool_use_system_prompt_tokens": 159 }, "vertex_ai/claude-opus-4-1": { - "cache_creation_input_token_cost": 0.00001875, - "cache_read_input_token_cost": 0.0000015, - "input_cost_per_token": 0.000015, - "input_cost_per_token_batches": 0.0000075, + "cache_creation_input_token_cost": 1.875e-5, + "cache_read_input_token_cost": 1.5e-6, + "input_cost_per_token": 1.5e-5, + "input_cost_per_token_batches": 7.5e-6, "litellm_provider": "vertex_ai-anthropic_models", "max_input_tokens": 200000, "max_output_tokens": 32000, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 0.000075, - "output_cost_per_token_batches": 0.0000375, + "output_cost_per_token": 7.5e-5, + "output_cost_per_token_batches": 3.75e-5, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_tool_choice": true, "supports_vision": true }, "vertex_ai/claude-opus-4-1@20250805": { - "cache_creation_input_token_cost": 0.00001875, - "cache_read_input_token_cost": 0.0000015, - "input_cost_per_token": 0.000015, - "input_cost_per_token_batches": 0.0000075, + "cache_creation_input_token_cost": 1.875e-5, + "cache_read_input_token_cost": 1.5e-6, + "input_cost_per_token": 1.5e-5, + "input_cost_per_token_batches": 7.5e-6, "litellm_provider": "vertex_ai-anthropic_models", "max_input_tokens": 200000, "max_output_tokens": 32000, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 0.000075, - "output_cost_per_token_batches": 0.0000375, + "output_cost_per_token": 7.5e-5, + "output_cost_per_token_batches": 3.75e-5, "supports_assistant_prefill": true, "supports_function_calling": true, "supports_tool_choice": true, "supports_vision": true }, "vertex_ai/claude-sonnet-4-5": { - "cache_creation_input_token_cost": 0.00000375, + "cache_creation_input_token_cost": 3.75e-6, "cache_read_input_token_cost": 3e-7, - "input_cost_per_token": 0.000003, - "input_cost_per_token_above_200k_tokens": 0.000006, - "output_cost_per_token_above_200k_tokens": 0.0000225, - "cache_creation_input_token_cost_above_200k_tokens": 0.0000075, + "input_cost_per_token": 3e-6, + "input_cost_per_token_above_200k_tokens": 6e-6, + "output_cost_per_token_above_200k_tokens": 2.25e-5, + "cache_creation_input_token_cost_above_200k_tokens": 7.5e-6, "cache_read_input_token_cost_above_200k_tokens": 6e-7, - "input_cost_per_token_batches": 0.0000015, + "input_cost_per_token_batches": 1.5e-6, "litellm_provider": "vertex_ai-anthropic_models", "max_input_tokens": 200000, "max_output_tokens": 64000, "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 0.000015, - "output_cost_per_token_batches": 0.0000075, + "output_cost_per_token": 1.5e-5, + "output_cost_per_token_batches": 7.5e-6, "supports_assistant_prefill": true, "supports_computer_use": true, "supports_function_calling": true, @@ -24498,21 +22545,21 @@ "supports_vision": true }, "vertex_ai/claude-sonnet-4-5@20250929": { - "cache_creation_input_token_cost": 0.00000375, + "cache_creation_input_token_cost": 3.75e-6, "cache_read_input_token_cost": 3e-7, - "input_cost_per_token": 0.000003, - "input_cost_per_token_above_200k_tokens": 0.000006, - "output_cost_per_token_above_200k_tokens": 0.0000225, - "cache_creation_input_token_cost_above_200k_tokens": 0.0000075, + "input_cost_per_token": 3e-6, + "input_cost_per_token_above_200k_tokens": 6e-6, + "output_cost_per_token_above_200k_tokens": 2.25e-5, + "cache_creation_input_token_cost_above_200k_tokens": 7.5e-6, "cache_read_input_token_cost_above_200k_tokens": 6e-7, - "input_cost_per_token_batches": 0.0000015, + "input_cost_per_token_batches": 1.5e-6, "litellm_provider": "vertex_ai-anthropic_models", "max_input_tokens": 200000, "max_output_tokens": 64000, "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 0.000015, - "output_cost_per_token_batches": 0.0000075, + "output_cost_per_token": 1.5e-5, + "output_cost_per_token_batches": 7.5e-6, "supports_assistant_prefill": true, "supports_computer_use": true, "supports_function_calling": true, @@ -24524,15 +22571,15 @@ "supports_vision": true }, "vertex_ai/claude-opus-4@20250514": { - "cache_creation_input_token_cost": 0.00001875, - "cache_read_input_token_cost": 0.0000015, - "input_cost_per_token": 0.000015, + "cache_creation_input_token_cost": 1.875e-5, + "cache_read_input_token_cost": 1.5e-6, + "input_cost_per_token": 1.5e-5, "litellm_provider": "vertex_ai-anthropic_models", "max_input_tokens": 200000, "max_output_tokens": 32000, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 0.000075, + "output_cost_per_token": 7.5e-5, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -24550,19 +22597,19 @@ "tool_use_system_prompt_tokens": 159 }, "vertex_ai/claude-sonnet-4": { - "cache_creation_input_token_cost": 0.00000375, + "cache_creation_input_token_cost": 3.75e-6, "cache_read_input_token_cost": 3e-7, - "input_cost_per_token": 0.000003, - "input_cost_per_token_above_200k_tokens": 0.000006, - "output_cost_per_token_above_200k_tokens": 0.0000225, - "cache_creation_input_token_cost_above_200k_tokens": 0.0000075, + "input_cost_per_token": 3e-6, + "input_cost_per_token_above_200k_tokens": 6e-6, + "output_cost_per_token_above_200k_tokens": 2.25e-5, + "cache_creation_input_token_cost_above_200k_tokens": 7.5e-6, "cache_read_input_token_cost_above_200k_tokens": 6e-7, "litellm_provider": "vertex_ai-anthropic_models", "max_input_tokens": 1000000, "max_output_tokens": 64000, "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -24580,19 +22627,19 @@ "tool_use_system_prompt_tokens": 159 }, "vertex_ai/claude-sonnet-4@20250514": { - "cache_creation_input_token_cost": 0.00000375, + "cache_creation_input_token_cost": 3.75e-6, "cache_read_input_token_cost": 3e-7, - "input_cost_per_token": 0.000003, - "input_cost_per_token_above_200k_tokens": 0.000006, - "output_cost_per_token_above_200k_tokens": 0.0000225, - "cache_creation_input_token_cost_above_200k_tokens": 0.0000075, + "input_cost_per_token": 3e-6, + "input_cost_per_token_above_200k_tokens": 6e-6, + "output_cost_per_token_above_200k_tokens": 2.25e-5, + "cache_creation_input_token_cost_above_200k_tokens": 7.5e-6, "cache_read_input_token_cost_above_200k_tokens": 6e-7, "litellm_provider": "vertex_ai-anthropic_models", "max_input_tokens": 1000000, "max_output_tokens": 64000, "max_tokens": 64000, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "search_context_cost_per_query": { "search_context_size_high": 0.01, "search_context_size_low": 0.01, @@ -24687,17 +22734,15 @@ "supports_tool_choice": true }, "vertex_ai/deepseek-ai/deepseek-v3.1-maas": { - "input_cost_per_token": 0.00000135, + "input_cost_per_token": 1.35e-6, "litellm_provider": "vertex_ai-deepseek_models", "max_input_tokens": 163840, "max_output_tokens": 32768, "max_tokens": 163840, "mode": "chat", - "output_cost_per_token": 0.0000054, + "output_cost_per_token": 5.4e-6, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing#partner-models", - "supported_regions": [ - "us-west2" - ], + "supported_regions": ["us-west2"], "supports_assistant_prefill": true, "supports_function_calling": true, "supports_prompt_caching": true, @@ -24705,13 +22750,13 @@ "supports_tool_choice": true }, "vertex_ai/deepseek-ai/deepseek-r1-0528-maas": { - "input_cost_per_token": 0.00000135, + "input_cost_per_token": 1.35e-6, "litellm_provider": "vertex_ai-deepseek_models", "max_input_tokens": 65336, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.0000054, + "output_cost_per_token": 5.4e-6, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing#partner-models", "supports_assistant_prefill": true, "supports_function_calling": true, @@ -24721,7 +22766,7 @@ }, "vertex_ai/gemini-2.5-flash-image": { "cache_read_input_token_cost": 3e-8, - "input_cost_per_audio_token": 0.000001, + "input_cost_per_audio_token": 1e-6, "input_cost_per_token": 3e-7, "litellm_provider": "vertex_ai-language-models", "max_audio_length_hours": 8.4, @@ -24735,25 +22780,13 @@ "max_videos_per_prompt": 10, "mode": "image_generation", "output_cost_per_image": 0.039, - "output_cost_per_reasoning_token": 0.0000025, - "output_cost_per_token": 0.0000025, + "output_cost_per_reasoning_token": 2.5e-6, + "output_cost_per_token": 2.5e-6, "rpm": 100000, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/multimodal/image-generation#edit-an-image", - "supported_endpoints": [ - "/v1/chat/completions", - "/v1/completions", - "/v1/batch" - ], - "supported_modalities": [ - "text", - "image", - "audio", - "video" - ], - "supported_output_modalities": [ - "text", - "image" - ], + "supported_endpoints": ["/v1/chat/completions", "/v1/completions", "/v1/batch"], + "supported_modalities": ["text", "image", "audio", "video"], + "supported_output_modalities": ["text", "image"], "supports_audio_output": false, "supports_function_calling": true, "supports_parallel_function_calling": true, @@ -24769,16 +22802,16 @@ }, "vertex_ai/gemini-3-pro-image-preview": { "input_cost_per_image": 0.0011, - "input_cost_per_token": 0.000002, - "input_cost_per_token_batches": 0.000001, + "input_cost_per_token": 2e-6, + "input_cost_per_token_batches": 1e-6, "litellm_provider": "vertex_ai-language-models", "max_input_tokens": 65536, "max_output_tokens": 32768, "max_tokens": 65536, "mode": "image_generation", "output_cost_per_image": 0.134, - "output_cost_per_token": 0.000012, - "output_cost_per_token_batches": 0.000006, + "output_cost_per_token": 1.2e-5, + "output_cost_per_token_batches": 6e-6, "source": "https://docs.cloud.google.com/vertex-ai/generative-ai/docs/models/gemini/3-pro-image" }, "vertex_ai/imagegeneration@006": { @@ -24840,23 +22873,23 @@ "supports_tool_choice": true }, "vertex_ai/jamba-1.5-large": { - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "vertex_ai-ai21_models", "max_input_tokens": 256000, "max_output_tokens": 256000, "max_tokens": 256000, "mode": "chat", - "output_cost_per_token": 0.000008, + "output_cost_per_token": 8e-6, "supports_tool_choice": true }, "vertex_ai/jamba-1.5-large@001": { - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "vertex_ai-ai21_models", "max_input_tokens": 256000, "max_output_tokens": 256000, "max_tokens": 256000, "mode": "chat", - "output_cost_per_token": 0.000008, + "output_cost_per_token": 8e-6, "supports_tool_choice": true }, "vertex_ai/jamba-1.5-mini": { @@ -24880,33 +22913,33 @@ "supports_tool_choice": true }, "vertex_ai/meta/llama-3.1-405b-instruct-maas": { - "input_cost_per_token": 0.000005, + "input_cost_per_token": 5e-6, "litellm_provider": "vertex_ai-llama_models", "max_input_tokens": 128000, "max_output_tokens": 2048, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.000016, + "output_cost_per_token": 1.6e-5, "source": "https://console.cloud.google.com/vertex-ai/publishers/meta/model-garden/llama-3.2-90b-vision-instruct-maas", "supports_system_messages": true, "supports_tool_choice": true, "supports_vision": true }, "vertex_ai/meta/llama-3.1-70b-instruct-maas": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "vertex_ai-llama_models", "max_input_tokens": 128000, "max_output_tokens": 2048, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "source": "https://console.cloud.google.com/vertex-ai/publishers/meta/model-garden/llama-3.2-90b-vision-instruct-maas", "supports_system_messages": true, "supports_tool_choice": true, "supports_vision": true }, "vertex_ai/meta/llama-3.1-8b-instruct-maas": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "vertex_ai-llama_models", "max_input_tokens": 128000, "max_output_tokens": 2048, @@ -24915,14 +22948,14 @@ "notes": "VertexAI states that The Llama 3.1 API service for llama-3.1-70b-instruct-maas and llama-3.1-8b-instruct-maas are in public preview and at no cost." }, "mode": "chat", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "source": "https://console.cloud.google.com/vertex-ai/publishers/meta/model-garden/llama-3.2-90b-vision-instruct-maas", "supports_system_messages": true, "supports_tool_choice": true, "supports_vision": true }, "vertex_ai/meta/llama-3.2-90b-vision-instruct-maas": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "vertex_ai-llama_models", "max_input_tokens": 128000, "max_output_tokens": 2048, @@ -24931,7 +22964,7 @@ "notes": "VertexAI states that The Llama 3.2 API service is at no cost during public preview, and will be priced as per dollar-per-1M-tokens at GA." }, "mode": "chat", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "source": "https://console.cloud.google.com/vertex-ai/publishers/meta/model-garden/llama-3.2-90b-vision-instruct-maas", "supports_system_messages": true, "supports_tool_choice": true, @@ -24944,16 +22977,10 @@ "max_output_tokens": 1000000, "max_tokens": 1000000, "mode": "chat", - "output_cost_per_token": 0.00000115, + "output_cost_per_token": 1.15e-6, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing#partner-models", - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text", - "code" - ], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text", "code"], "supports_function_calling": true, "supports_tool_choice": true }, @@ -24964,16 +22991,10 @@ "max_output_tokens": 1000000, "max_tokens": 1000000, "mode": "chat", - "output_cost_per_token": 0.00000115, + "output_cost_per_token": 1.15e-6, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing#partner-models", - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text", - "code" - ], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text", "code"], "supports_function_calling": true, "supports_tool_choice": true }, @@ -24986,14 +23007,8 @@ "mode": "chat", "output_cost_per_token": 7e-7, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing#partner-models", - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text", - "code" - ], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text", "code"], "supports_function_calling": true, "supports_tool_choice": true }, @@ -25006,47 +23021,41 @@ "mode": "chat", "output_cost_per_token": 7e-7, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing#partner-models", - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "text", - "code" - ], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["text", "code"], "supports_function_calling": true, "supports_tool_choice": true }, "vertex_ai/meta/llama3-405b-instruct-maas": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "vertex_ai-llama_models", "max_input_tokens": 32000, "max_output_tokens": 32000, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing#partner-models", "supports_tool_choice": true }, "vertex_ai/meta/llama3-70b-instruct-maas": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "vertex_ai-llama_models", "max_input_tokens": 32000, "max_output_tokens": 32000, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing#partner-models", "supports_tool_choice": true }, "vertex_ai/meta/llama3-8b-instruct-maas": { - "input_cost_per_token": 0, + "input_cost_per_token": 0.0, "litellm_provider": "vertex_ai-llama_models", "max_input_tokens": 32000, "max_output_tokens": 32000, "max_tokens": 32000, "mode": "chat", - "output_cost_per_token": 0, + "output_cost_per_token": 0.0, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing#partner-models", "supports_tool_choice": true }, @@ -25057,7 +23066,7 @@ "max_output_tokens": 196608, "max_tokens": 196608, "mode": "chat", - "output_cost_per_token": 0.0000012, + "output_cost_per_token": 1.2e-6, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing#partner-models", "supports_function_calling": true, "supports_tool_choice": true @@ -25069,7 +23078,7 @@ "max_output_tokens": 256000, "max_tokens": 256000, "mode": "chat", - "output_cost_per_token": 0.0000025, + "output_cost_per_token": 2.5e-6, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing#partner-models", "supports_function_calling": true, "supports_tool_choice": true, @@ -25082,7 +23091,7 @@ "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 0.000002, + "output_cost_per_token": 2e-6, "supports_function_calling": true, "supports_tool_choice": true }, @@ -25093,7 +23102,7 @@ "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 0.000002, + "output_cost_per_token": 2e-6, "supports_function_calling": true, "supports_tool_choice": true }, @@ -25104,7 +23113,7 @@ "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 0.000002, + "output_cost_per_token": 2e-6, "supports_function_calling": true, "supports_tool_choice": true }, @@ -25115,62 +23124,62 @@ "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 0.000002, + "output_cost_per_token": 2e-6, "supports_function_calling": true, "supports_tool_choice": true }, "vertex_ai/mistral-large-2411": { - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "vertex_ai-mistral_models", "max_input_tokens": 128000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 0.000006, + "output_cost_per_token": 6e-6, "supports_function_calling": true, "supports_tool_choice": true }, "vertex_ai/mistral-large@2407": { - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "vertex_ai-mistral_models", "max_input_tokens": 128000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 0.000006, + "output_cost_per_token": 6e-6, "supports_function_calling": true, "supports_tool_choice": true }, "vertex_ai/mistral-large@2411-001": { - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "vertex_ai-mistral_models", "max_input_tokens": 128000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 0.000006, + "output_cost_per_token": 6e-6, "supports_function_calling": true, "supports_tool_choice": true }, "vertex_ai/mistral-large@latest": { - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "vertex_ai-mistral_models", "max_input_tokens": 128000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 0.000006, + "output_cost_per_token": 6e-6, "supports_function_calling": true, "supports_tool_choice": true }, "vertex_ai/mistral-nemo@2407": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "vertex_ai-mistral_models", "max_input_tokens": 128000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.000003, + "output_cost_per_token": 3e-6, "supports_function_calling": true, "supports_tool_choice": true }, @@ -25186,35 +23195,33 @@ "supports_tool_choice": true }, "vertex_ai/mistral-small-2503": { - "input_cost_per_token": 0.000001, + "input_cost_per_token": 1e-6, "litellm_provider": "vertex_ai-mistral_models", "max_input_tokens": 128000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", - "output_cost_per_token": 0.000003, + "output_cost_per_token": 3e-6, "supports_function_calling": true, "supports_tool_choice": true, "supports_vision": true }, "vertex_ai/mistral-small-2503@001": { - "input_cost_per_token": 0.000001, + "input_cost_per_token": 1e-6, "litellm_provider": "vertex_ai-mistral_models", "max_input_tokens": 32000, "max_output_tokens": 8191, "max_tokens": 8191, "mode": "chat", - "output_cost_per_token": 0.000003, + "output_cost_per_token": 3e-6, "supports_function_calling": true, "supports_tool_choice": true }, "vertex_ai/mistral-ocr-2505": { "litellm_provider": "vertex_ai", "mode": "ocr", - "ocr_cost_per_page": 0.0005, - "supported_endpoints": [ - "/v1/ocr" - ], + "ocr_cost_per_page": 5e-4, + "supported_endpoints": ["/v1/ocr"], "source": "https://cloud.google.com/generative-ai-app-builder/pricing" }, "vertex_ai/openai/gpt-oss-120b-maas": { @@ -25246,19 +23253,19 @@ "max_output_tokens": 16384, "max_tokens": 16384, "mode": "chat", - "output_cost_per_token": 0.000001, + "output_cost_per_token": 1e-6, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing", "supports_function_calling": true, "supports_tool_choice": true }, "vertex_ai/qwen/qwen3-coder-480b-a35b-instruct-maas": { - "input_cost_per_token": 0.000001, + "input_cost_per_token": 1e-6, "litellm_provider": "vertex_ai-qwen_models", "max_input_tokens": 262144, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0.000004, + "output_cost_per_token": 4e-6, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing", "supports_function_calling": true, "supports_tool_choice": true @@ -25270,7 +23277,7 @@ "max_output_tokens": 262144, "max_tokens": 262144, "mode": "chat", - "output_cost_per_token": 0.0000012, + "output_cost_per_token": 1.2e-6, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing", "supports_function_calling": true, "supports_tool_choice": true @@ -25282,7 +23289,7 @@ "max_output_tokens": 262144, "max_tokens": 262144, "mode": "chat", - "output_cost_per_token": 0.0000012, + "output_cost_per_token": 1.2e-6, "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing", "supports_function_calling": true, "supports_tool_choice": true @@ -25294,12 +23301,8 @@ "mode": "video_generation", "output_cost_per_second": 0.35, "source": "https://ai.google.dev/gemini-api/docs/video", - "supported_modalities": [ - "text" - ], - "supported_output_modalities": [ - "video" - ] + "supported_modalities": ["text"], + "supported_output_modalities": ["video"] }, "vertex_ai/veo-3.0-fast-generate-preview": { "litellm_provider": "vertex_ai-video-models", @@ -25308,12 +23311,8 @@ "mode": "video_generation", "output_cost_per_second": 0.15, "source": "https://ai.google.dev/gemini-api/docs/video", - "supported_modalities": [ - "text" - ], - "supported_output_modalities": [ - "video" - ] + "supported_modalities": ["text"], + "supported_output_modalities": ["video"] }, "vertex_ai/veo-3.0-generate-preview": { "litellm_provider": "vertex_ai-video-models", @@ -25322,12 +23321,8 @@ "mode": "video_generation", "output_cost_per_second": 0.4, "source": "https://ai.google.dev/gemini-api/docs/video", - "supported_modalities": [ - "text" - ], - "supported_output_modalities": [ - "video" - ] + "supported_modalities": ["text"], + "supported_output_modalities": ["video"] }, "vertex_ai/veo-3.0-fast-generate-001": { "litellm_provider": "vertex_ai-video-models", @@ -25336,12 +23331,8 @@ "mode": "video_generation", "output_cost_per_second": 0.15, "source": "https://ai.google.dev/gemini-api/docs/video", - "supported_modalities": [ - "text" - ], - "supported_output_modalities": [ - "video" - ] + "supported_modalities": ["text"], + "supported_output_modalities": ["video"] }, "vertex_ai/veo-3.0-generate-001": { "litellm_provider": "vertex_ai-video-models", @@ -25350,12 +23341,8 @@ "mode": "video_generation", "output_cost_per_second": 0.4, "source": "https://ai.google.dev/gemini-api/docs/video", - "supported_modalities": [ - "text" - ], - "supported_output_modalities": [ - "video" - ] + "supported_modalities": ["text"], + "supported_output_modalities": ["video"] }, "vertex_ai/veo-3.1-generate-preview": { "litellm_provider": "vertex_ai-video-models", @@ -25364,12 +23351,8 @@ "mode": "video_generation", "output_cost_per_second": 0.4, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/model-reference/veo", - "supported_modalities": [ - "text" - ], - "supported_output_modalities": [ - "video" - ] + "supported_modalities": ["text"], + "supported_output_modalities": ["video"] }, "vertex_ai/veo-3.1-fast-generate-preview": { "litellm_provider": "vertex_ai-video-models", @@ -25378,12 +23361,8 @@ "mode": "video_generation", "output_cost_per_second": 0.15, "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/model-reference/veo", - "supported_modalities": [ - "text" - ], - "supported_output_modalities": [ - "video" - ] + "supported_modalities": ["text"], + "supported_output_modalities": ["video"] }, "voyage/rerank-2": { "input_cost_per_query": 5e-8, @@ -25394,7 +23373,7 @@ "max_query_tokens": 16000, "max_tokens": 16000, "mode": "rerank", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "voyage/rerank-2-lite": { "input_cost_per_query": 2e-8, @@ -25405,7 +23384,7 @@ "max_query_tokens": 8000, "max_tokens": 8000, "mode": "rerank", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "voyage/voyage-2": { "input_cost_per_token": 1e-7, @@ -25413,7 +23392,7 @@ "max_input_tokens": 4000, "max_tokens": 4000, "mode": "embedding", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "voyage/voyage-3": { "input_cost_per_token": 6e-8, @@ -25421,7 +23400,7 @@ "max_input_tokens": 32000, "max_tokens": 32000, "mode": "embedding", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "voyage/voyage-3-large": { "input_cost_per_token": 1.8e-7, @@ -25429,7 +23408,7 @@ "max_input_tokens": 32000, "max_tokens": 32000, "mode": "embedding", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "voyage/voyage-3-lite": { "input_cost_per_token": 2e-8, @@ -25437,7 +23416,7 @@ "max_input_tokens": 32000, "max_tokens": 32000, "mode": "embedding", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "voyage/voyage-3.5": { "input_cost_per_token": 6e-8, @@ -25445,7 +23424,7 @@ "max_input_tokens": 32000, "max_tokens": 32000, "mode": "embedding", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "voyage/voyage-3.5-lite": { "input_cost_per_token": 2e-8, @@ -25453,7 +23432,7 @@ "max_input_tokens": 32000, "max_tokens": 32000, "mode": "embedding", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "voyage/voyage-code-2": { "input_cost_per_token": 1.2e-7, @@ -25461,7 +23440,7 @@ "max_input_tokens": 16000, "max_tokens": 16000, "mode": "embedding", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "voyage/voyage-code-3": { "input_cost_per_token": 1.8e-7, @@ -25469,7 +23448,7 @@ "max_input_tokens": 32000, "max_tokens": 32000, "mode": "embedding", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "voyage/voyage-context-3": { "input_cost_per_token": 1.8e-7, @@ -25477,7 +23456,7 @@ "max_input_tokens": 120000, "max_tokens": 120000, "mode": "embedding", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "voyage/voyage-finance-2": { "input_cost_per_token": 1.2e-7, @@ -25485,7 +23464,7 @@ "max_input_tokens": 32000, "max_tokens": 32000, "mode": "embedding", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "voyage/voyage-large-2": { "input_cost_per_token": 1.2e-7, @@ -25493,7 +23472,7 @@ "max_input_tokens": 16000, "max_tokens": 16000, "mode": "embedding", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "voyage/voyage-law-2": { "input_cost_per_token": 1.2e-7, @@ -25501,7 +23480,7 @@ "max_input_tokens": 16000, "max_tokens": 16000, "mode": "embedding", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "voyage/voyage-lite-01": { "input_cost_per_token": 1e-7, @@ -25509,7 +23488,7 @@ "max_input_tokens": 4096, "max_tokens": 4096, "mode": "embedding", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "voyage/voyage-lite-02-instruct": { "input_cost_per_token": 1e-7, @@ -25517,7 +23496,7 @@ "max_input_tokens": 4000, "max_tokens": 4000, "mode": "embedding", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "voyage/voyage-multimodal-3": { "input_cost_per_token": 1.2e-7, @@ -25525,7 +23504,7 @@ "max_input_tokens": 32000, "max_tokens": 32000, "mode": "embedding", - "output_cost_per_token": 0 + "output_cost_per_token": 0.0 }, "wandb/openai/gpt-oss-120b": { "max_tokens": 131072, @@ -25654,13 +23633,13 @@ "mode": "chat" }, "watsonx/ibm/granite-3-8b-instruct": { - "input_cost_per_token": 2e-7, + "input_cost_per_token": 0.2e-6, "litellm_provider": "watsonx", "max_input_tokens": 8192, "max_output_tokens": 1024, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 2e-7, + "output_cost_per_token": 0.2e-6, "supports_audio_input": false, "supports_audio_output": false, "supports_function_calling": true, @@ -25672,13 +23651,13 @@ "supports_vision": false }, "watsonx/mistralai/mistral-large": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "watsonx", "max_input_tokens": 131072, "max_output_tokens": 16384, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0.00001, + "output_cost_per_token": 10e-6, "supports_audio_input": false, "supports_audio_output": false, "supports_function_calling": true, @@ -25717,8 +23696,8 @@ "max_tokens": 8192, "max_input_tokens": 8192, "max_output_tokens": 8192, - "input_cost_per_token": 6e-7, - "output_cost_per_token": 6e-7, + "input_cost_per_token": 0.6e-6, + "output_cost_per_token": 0.6e-6, "litellm_provider": "watsonx", "mode": "chat", "supports_function_calling": false, @@ -25729,8 +23708,8 @@ "max_tokens": 8192, "max_input_tokens": 8192, "max_output_tokens": 8192, - "input_cost_per_token": 6e-7, - "output_cost_per_token": 6e-7, + "input_cost_per_token": 0.6e-6, + "output_cost_per_token": 0.6e-6, "litellm_provider": "watsonx", "mode": "chat", "supports_function_calling": false, @@ -25741,8 +23720,8 @@ "max_tokens": 8192, "max_input_tokens": 8192, "max_output_tokens": 8192, - "input_cost_per_token": 6e-7, - "output_cost_per_token": 6e-7, + "input_cost_per_token": 0.6e-6, + "output_cost_per_token": 0.6e-6, "litellm_provider": "watsonx", "mode": "chat", "supports_function_calling": false, @@ -25753,8 +23732,8 @@ "max_tokens": 8192, "max_input_tokens": 8192, "max_output_tokens": 8192, - "input_cost_per_token": 2e-7, - "output_cost_per_token": 2e-7, + "input_cost_per_token": 0.2e-6, + "output_cost_per_token": 0.2e-6, "litellm_provider": "watsonx", "mode": "chat", "supports_function_calling": true, @@ -25765,8 +23744,8 @@ "max_tokens": 20480, "max_input_tokens": 20480, "max_output_tokens": 20480, - "input_cost_per_token": 6e-8, - "output_cost_per_token": 2.5e-7, + "input_cost_per_token": 0.06e-6, + "output_cost_per_token": 0.25e-6, "litellm_provider": "watsonx", "mode": "chat", "supports_function_calling": true, @@ -25777,8 +23756,8 @@ "max_tokens": 8192, "max_input_tokens": 8192, "max_output_tokens": 8192, - "input_cost_per_token": 1e-7, - "output_cost_per_token": 1e-7, + "input_cost_per_token": 0.1e-6, + "output_cost_per_token": 0.1e-6, "litellm_provider": "watsonx", "mode": "chat", "supports_function_calling": false, @@ -25789,8 +23768,8 @@ "max_tokens": 8192, "max_input_tokens": 8192, "max_output_tokens": 8192, - "input_cost_per_token": 2e-7, - "output_cost_per_token": 2e-7, + "input_cost_per_token": 0.2e-6, + "output_cost_per_token": 0.2e-6, "litellm_provider": "watsonx", "mode": "chat", "supports_function_calling": false, @@ -25801,8 +23780,8 @@ "max_tokens": 512, "max_input_tokens": 512, "max_output_tokens": 512, - "input_cost_per_token": 3.8e-7, - "output_cost_per_token": 3.8e-7, + "input_cost_per_token": 0.38e-6, + "output_cost_per_token": 0.38e-6, "litellm_provider": "watsonx", "mode": "chat", "supports_function_calling": false, @@ -25813,8 +23792,8 @@ "max_tokens": 512, "max_input_tokens": 512, "max_output_tokens": 512, - "input_cost_per_token": 3.8e-7, - "output_cost_per_token": 3.8e-7, + "input_cost_per_token": 0.38e-6, + "output_cost_per_token": 0.38e-6, "litellm_provider": "watsonx", "mode": "chat", "supports_function_calling": false, @@ -25825,8 +23804,8 @@ "max_tokens": 512, "max_input_tokens": 512, "max_output_tokens": 512, - "input_cost_per_token": 3.8e-7, - "output_cost_per_token": 3.8e-7, + "input_cost_per_token": 0.38e-6, + "output_cost_per_token": 0.38e-6, "litellm_provider": "watsonx", "mode": "chat", "supports_function_calling": false, @@ -25837,8 +23816,8 @@ "max_tokens": 8192, "max_input_tokens": 8192, "max_output_tokens": 8192, - "input_cost_per_token": 1e-7, - "output_cost_per_token": 1e-7, + "input_cost_per_token": 0.1e-6, + "output_cost_per_token": 0.1e-6, "litellm_provider": "watsonx", "mode": "chat", "supports_function_calling": false, @@ -25849,8 +23828,8 @@ "max_tokens": 128000, "max_input_tokens": 128000, "max_output_tokens": 128000, - "input_cost_per_token": 3.5e-7, - "output_cost_per_token": 3.5e-7, + "input_cost_per_token": 0.35e-6, + "output_cost_per_token": 0.35e-6, "litellm_provider": "watsonx", "mode": "chat", "supports_function_calling": true, @@ -25861,8 +23840,8 @@ "max_tokens": 128000, "max_input_tokens": 128000, "max_output_tokens": 128000, - "input_cost_per_token": 1e-7, - "output_cost_per_token": 1e-7, + "input_cost_per_token": 0.1e-6, + "output_cost_per_token": 0.1e-6, "litellm_provider": "watsonx", "mode": "chat", "supports_function_calling": true, @@ -25873,8 +23852,8 @@ "max_tokens": 128000, "max_input_tokens": 128000, "max_output_tokens": 128000, - "input_cost_per_token": 1.5e-7, - "output_cost_per_token": 1.5e-7, + "input_cost_per_token": 0.15e-6, + "output_cost_per_token": 0.15e-6, "litellm_provider": "watsonx", "mode": "chat", "supports_function_calling": true, @@ -25885,8 +23864,8 @@ "max_tokens": 128000, "max_input_tokens": 128000, "max_output_tokens": 128000, - "input_cost_per_token": 0.000002, - "output_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, + "output_cost_per_token": 2e-6, "litellm_provider": "watsonx", "mode": "chat", "supports_function_calling": true, @@ -25897,8 +23876,8 @@ "max_tokens": 128000, "max_input_tokens": 128000, "max_output_tokens": 128000, - "input_cost_per_token": 7.1e-7, - "output_cost_per_token": 7.1e-7, + "input_cost_per_token": 0.71e-6, + "output_cost_per_token": 0.71e-6, "litellm_provider": "watsonx", "mode": "chat", "supports_function_calling": true, @@ -25909,8 +23888,8 @@ "max_tokens": 128000, "max_input_tokens": 128000, "max_output_tokens": 128000, - "input_cost_per_token": 3.5e-7, - "output_cost_per_token": 0.0000014, + "input_cost_per_token": 0.35e-6, + "output_cost_per_token": 1.4e-6, "litellm_provider": "watsonx", "mode": "chat", "supports_function_calling": true, @@ -25921,8 +23900,8 @@ "max_tokens": 128000, "max_input_tokens": 128000, "max_output_tokens": 128000, - "input_cost_per_token": 3.5e-7, - "output_cost_per_token": 3.5e-7, + "input_cost_per_token": 0.35e-6, + "output_cost_per_token": 0.35e-6, "litellm_provider": "watsonx", "mode": "chat", "supports_function_calling": false, @@ -25933,8 +23912,8 @@ "max_tokens": 128000, "max_input_tokens": 128000, "max_output_tokens": 128000, - "input_cost_per_token": 0.000003, - "output_cost_per_token": 0.00001, + "input_cost_per_token": 3e-6, + "output_cost_per_token": 10e-6, "litellm_provider": "watsonx", "mode": "chat", "supports_function_calling": true, @@ -25945,8 +23924,8 @@ "max_tokens": 32000, "max_input_tokens": 32000, "max_output_tokens": 32000, - "input_cost_per_token": 1e-7, - "output_cost_per_token": 3e-7, + "input_cost_per_token": 0.1e-6, + "output_cost_per_token": 0.3e-6, "litellm_provider": "watsonx", "mode": "chat", "supports_function_calling": true, @@ -25957,8 +23936,8 @@ "max_tokens": 32000, "max_input_tokens": 32000, "max_output_tokens": 32000, - "input_cost_per_token": 1e-7, - "output_cost_per_token": 3e-7, + "input_cost_per_token": 0.1e-6, + "output_cost_per_token": 0.3e-6, "litellm_provider": "watsonx", "mode": "chat", "supports_function_calling": true, @@ -25969,8 +23948,8 @@ "max_tokens": 128000, "max_input_tokens": 128000, "max_output_tokens": 128000, - "input_cost_per_token": 3.5e-7, - "output_cost_per_token": 3.5e-7, + "input_cost_per_token": 0.35e-6, + "output_cost_per_token": 0.35e-6, "litellm_provider": "watsonx", "mode": "chat", "supports_function_calling": false, @@ -25981,8 +23960,8 @@ "max_tokens": 8192, "max_input_tokens": 8192, "max_output_tokens": 8192, - "input_cost_per_token": 1.5e-7, - "output_cost_per_token": 6e-7, + "input_cost_per_token": 0.15e-6, + "output_cost_per_token": 0.6e-6, "litellm_provider": "watsonx", "mode": "chat", "supports_function_calling": false, @@ -25993,8 +23972,8 @@ "max_tokens": 8192, "max_input_tokens": 8192, "max_output_tokens": 8192, - "input_cost_per_token": 0.0000018, - "output_cost_per_token": 0.0000018, + "input_cost_per_token": 1.8e-6, + "output_cost_per_token": 1.8e-6, "litellm_provider": "watsonx", "mode": "chat", "supports_function_calling": false, @@ -26006,96 +23985,94 @@ "litellm_provider": "openai", "mode": "audio_transcription", "output_cost_per_second": 0.0001, - "supported_endpoints": [ - "/v1/audio/transcriptions" - ] + "supported_endpoints": ["/v1/audio/transcriptions"] }, "xai/grok-2": { - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "xai", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0.00001, + "output_cost_per_token": 1e-5, "supports_function_calling": true, "supports_tool_choice": true, "supports_web_search": true }, "xai/grok-2-1212": { - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "xai", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0.00001, + "output_cost_per_token": 1e-5, "supports_function_calling": true, "supports_tool_choice": true, "supports_web_search": true }, "xai/grok-2-latest": { - "input_cost_per_token": 0.000002, + "input_cost_per_token": 2e-6, "litellm_provider": "xai", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0.00001, + "output_cost_per_token": 1e-5, "supports_function_calling": true, "supports_tool_choice": true, "supports_web_search": true }, "xai/grok-2-vision": { - "input_cost_per_image": 0.000002, - "input_cost_per_token": 0.000002, + "input_cost_per_image": 2e-6, + "input_cost_per_token": 2e-6, "litellm_provider": "xai", "max_input_tokens": 32768, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0.00001, + "output_cost_per_token": 1e-5, "supports_function_calling": true, "supports_tool_choice": true, "supports_vision": true, "supports_web_search": true }, "xai/grok-2-vision-1212": { - "input_cost_per_image": 0.000002, - "input_cost_per_token": 0.000002, + "input_cost_per_image": 2e-6, + "input_cost_per_token": 2e-6, "litellm_provider": "xai", "max_input_tokens": 32768, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0.00001, + "output_cost_per_token": 1e-5, "supports_function_calling": true, "supports_tool_choice": true, "supports_vision": true, "supports_web_search": true }, "xai/grok-2-vision-latest": { - "input_cost_per_image": 0.000002, - "input_cost_per_token": 0.000002, + "input_cost_per_image": 2e-6, + "input_cost_per_token": 2e-6, "litellm_provider": "xai", "max_input_tokens": 32768, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", - "output_cost_per_token": 0.00001, + "output_cost_per_token": 1e-5, "supports_function_calling": true, "supports_tool_choice": true, "supports_vision": true, "supports_web_search": true }, "xai/grok-3": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "xai", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "source": "https://x.ai/api#pricing", "supports_function_calling": true, "supports_response_schema": false, @@ -26103,13 +24080,13 @@ "supports_web_search": true }, "xai/grok-3-beta": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "xai", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "source": "https://x.ai/api#pricing", "supports_function_calling": true, "supports_response_schema": false, @@ -26117,13 +24094,13 @@ "supports_web_search": true }, "xai/grok-3-fast-beta": { - "input_cost_per_token": 0.000005, + "input_cost_per_token": 5e-6, "litellm_provider": "xai", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0.000025, + "output_cost_per_token": 2.5e-5, "source": "https://x.ai/api#pricing", "supports_function_calling": true, "supports_response_schema": false, @@ -26131,13 +24108,13 @@ "supports_web_search": true }, "xai/grok-3-fast-latest": { - "input_cost_per_token": 0.000005, + "input_cost_per_token": 5e-6, "litellm_provider": "xai", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0.000025, + "output_cost_per_token": 2.5e-5, "source": "https://x.ai/api#pricing", "supports_function_calling": true, "supports_response_schema": false, @@ -26145,13 +24122,13 @@ "supports_web_search": true }, "xai/grok-3-latest": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "xai", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "source": "https://x.ai/api#pricing", "supports_function_calling": true, "supports_response_schema": false, @@ -26195,7 +24172,7 @@ "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0.000004, + "output_cost_per_token": 4e-6, "source": "https://x.ai/api#pricing", "supports_function_calling": true, "supports_reasoning": true, @@ -26210,7 +24187,7 @@ "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0.000004, + "output_cost_per_token": 4e-6, "source": "https://x.ai/api#pricing", "supports_function_calling": true, "supports_reasoning": true, @@ -26225,7 +24202,7 @@ "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0.000004, + "output_cost_per_token": 4e-6, "source": "https://x.ai/api#pricing", "supports_function_calling": true, "supports_reasoning": true, @@ -26249,13 +24226,13 @@ "supports_web_search": true }, "xai/grok-4": { - "input_cost_per_token": 0.000003, + "input_cost_per_token": 3e-6, "litellm_provider": "xai", "max_input_tokens": 256000, "max_output_tokens": 256000, "max_tokens": 256000, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "source": "https://docs.x.ai/docs/models", "supports_function_calling": true, "supports_tool_choice": true, @@ -26263,15 +24240,15 @@ }, "xai/grok-4-fast-reasoning": { "litellm_provider": "xai", - "max_input_tokens": 2000000, - "max_output_tokens": 2000000, - "max_tokens": 2000000, - "mode": "chat", - "input_cost_per_token": 2e-7, - "input_cost_per_token_above_128k_tokens": 4e-7, - "output_cost_per_token": 5e-7, - "output_cost_per_token_above_128k_tokens": 0.000001, - "cache_read_input_token_cost": 5e-8, + "max_input_tokens": 2e6, + "max_output_tokens": 2e6, + "max_tokens": 2e6, + "mode": "chat", + "input_cost_per_token": 0.2e-6, + "input_cost_per_token_above_128k_tokens": 0.4e-6, + "output_cost_per_token": 0.5e-6, + "output_cost_per_token_above_128k_tokens": 1e-6, + "cache_read_input_token_cost": 0.05e-6, "source": "https://docs.x.ai/docs/models", "supports_function_calling": true, "supports_tool_choice": true, @@ -26279,61 +24256,61 @@ }, "xai/grok-4-fast-non-reasoning": { "litellm_provider": "xai", - "max_input_tokens": 2000000, - "max_output_tokens": 2000000, - "cache_read_input_token_cost": 5e-8, - "max_tokens": 2000000, - "mode": "chat", - "input_cost_per_token": 2e-7, - "input_cost_per_token_above_128k_tokens": 4e-7, - "output_cost_per_token": 5e-7, - "output_cost_per_token_above_128k_tokens": 0.000001, + "max_input_tokens": 2e6, + "max_output_tokens": 2e6, + "cache_read_input_token_cost": 0.05e-6, + "max_tokens": 2e6, + "mode": "chat", + "input_cost_per_token": 0.2e-6, + "input_cost_per_token_above_128k_tokens": 0.4e-6, + "output_cost_per_token": 0.5e-6, + "output_cost_per_token_above_128k_tokens": 1e-6, "source": "https://docs.x.ai/docs/models", "supports_function_calling": true, "supports_tool_choice": true, "supports_web_search": true }, "xai/grok-4-0709": { - "input_cost_per_token": 0.000003, - "input_cost_per_token_above_128k_tokens": 0.000006, + "input_cost_per_token": 3e-6, + "input_cost_per_token_above_128k_tokens": 6e-6, "litellm_provider": "xai", "max_input_tokens": 256000, "max_output_tokens": 256000, "max_tokens": 256000, "mode": "chat", - "output_cost_per_token": 0.000015, - "output_cost_per_token_above_128k_tokens": 0.00003, + "output_cost_per_token": 1.5e-5, + "output_cost_per_token_above_128k_tokens": 30e-6, "source": "https://docs.x.ai/docs/models", "supports_function_calling": true, "supports_tool_choice": true, "supports_web_search": true }, "xai/grok-4-latest": { - "input_cost_per_token": 0.000003, - "input_cost_per_token_above_128k_tokens": 0.000006, + "input_cost_per_token": 3e-6, + "input_cost_per_token_above_128k_tokens": 6e-6, "litellm_provider": "xai", "max_input_tokens": 256000, "max_output_tokens": 256000, "max_tokens": 256000, "mode": "chat", - "output_cost_per_token": 0.000015, - "output_cost_per_token_above_128k_tokens": 0.00003, + "output_cost_per_token": 1.5e-5, + "output_cost_per_token_above_128k_tokens": 30e-6, "source": "https://docs.x.ai/docs/models", "supports_function_calling": true, "supports_tool_choice": true, "supports_web_search": true }, "xai/grok-4-1-fast": { - "cache_read_input_token_cost": 5e-8, - "input_cost_per_token": 2e-7, - "input_cost_per_token_above_128k_tokens": 4e-7, + "cache_read_input_token_cost": 0.05e-6, + "input_cost_per_token": 0.2e-6, + "input_cost_per_token_above_128k_tokens": 0.4e-6, "litellm_provider": "xai", - "max_input_tokens": 2000000, - "max_output_tokens": 2000000, - "max_tokens": 2000000, + "max_input_tokens": 2e6, + "max_output_tokens": 2e6, + "max_tokens": 2e6, "mode": "chat", - "output_cost_per_token": 5e-7, - "output_cost_per_token_above_128k_tokens": 0.000001, + "output_cost_per_token": 0.5e-6, + "output_cost_per_token_above_128k_tokens": 1e-6, "source": "https://docs.x.ai/docs/models/grok-4-1-fast-reasoning", "supports_audio_input": true, "supports_function_calling": true, @@ -26344,16 +24321,16 @@ "supports_web_search": true }, "xai/grok-4-1-fast-reasoning": { - "cache_read_input_token_cost": 5e-8, - "input_cost_per_token": 2e-7, - "input_cost_per_token_above_128k_tokens": 4e-7, + "cache_read_input_token_cost": 0.05e-6, + "input_cost_per_token": 0.2e-6, + "input_cost_per_token_above_128k_tokens": 0.4e-6, "litellm_provider": "xai", - "max_input_tokens": 2000000, - "max_output_tokens": 2000000, - "max_tokens": 2000000, + "max_input_tokens": 2e6, + "max_output_tokens": 2e6, + "max_tokens": 2e6, "mode": "chat", - "output_cost_per_token": 5e-7, - "output_cost_per_token_above_128k_tokens": 0.000001, + "output_cost_per_token": 0.5e-6, + "output_cost_per_token_above_128k_tokens": 1e-6, "source": "https://docs.x.ai/docs/models/grok-4-1-fast-reasoning", "supports_audio_input": true, "supports_function_calling": true, @@ -26364,16 +24341,16 @@ "supports_web_search": true }, "xai/grok-4-1-fast-reasoning-latest": { - "cache_read_input_token_cost": 5e-8, - "input_cost_per_token": 2e-7, - "input_cost_per_token_above_128k_tokens": 4e-7, + "cache_read_input_token_cost": 0.05e-6, + "input_cost_per_token": 0.2e-6, + "input_cost_per_token_above_128k_tokens": 0.4e-6, "litellm_provider": "xai", - "max_input_tokens": 2000000, - "max_output_tokens": 2000000, - "max_tokens": 2000000, + "max_input_tokens": 2e6, + "max_output_tokens": 2e6, + "max_tokens": 2e6, "mode": "chat", - "output_cost_per_token": 5e-7, - "output_cost_per_token_above_128k_tokens": 0.000001, + "output_cost_per_token": 0.5e-6, + "output_cost_per_token_above_128k_tokens": 1e-6, "source": "https://docs.x.ai/docs/models/grok-4-1-fast-reasoning", "supports_audio_input": true, "supports_function_calling": true, @@ -26384,16 +24361,16 @@ "supports_web_search": true }, "xai/grok-4-1-fast-non-reasoning": { - "cache_read_input_token_cost": 5e-8, - "input_cost_per_token": 2e-7, - "input_cost_per_token_above_128k_tokens": 4e-7, + "cache_read_input_token_cost": 0.05e-6, + "input_cost_per_token": 0.2e-6, + "input_cost_per_token_above_128k_tokens": 0.4e-6, "litellm_provider": "xai", - "max_input_tokens": 2000000, - "max_output_tokens": 2000000, - "max_tokens": 2000000, + "max_input_tokens": 2e6, + "max_output_tokens": 2e6, + "max_tokens": 2e6, "mode": "chat", - "output_cost_per_token": 5e-7, - "output_cost_per_token_above_128k_tokens": 0.000001, + "output_cost_per_token": 0.5e-6, + "output_cost_per_token_above_128k_tokens": 1e-6, "source": "https://docs.x.ai/docs/models/grok-4-1-fast-non-reasoning", "supports_audio_input": true, "supports_function_calling": true, @@ -26403,16 +24380,16 @@ "supports_web_search": true }, "xai/grok-4-1-fast-non-reasoning-latest": { - "cache_read_input_token_cost": 5e-8, - "input_cost_per_token": 2e-7, - "input_cost_per_token_above_128k_tokens": 4e-7, + "cache_read_input_token_cost": 0.05e-6, + "input_cost_per_token": 0.2e-6, + "input_cost_per_token_above_128k_tokens": 0.4e-6, "litellm_provider": "xai", - "max_input_tokens": 2000000, - "max_output_tokens": 2000000, - "max_tokens": 2000000, + "max_input_tokens": 2e6, + "max_output_tokens": 2e6, + "max_tokens": 2e6, "mode": "chat", - "output_cost_per_token": 5e-7, - "output_cost_per_token_above_128k_tokens": 0.000001, + "output_cost_per_token": 0.5e-6, + "output_cost_per_token_above_128k_tokens": 1e-6, "source": "https://docs.x.ai/docs/models/grok-4-1-fast-non-reasoning", "supports_audio_input": true, "supports_function_calling": true, @@ -26422,13 +24399,13 @@ "supports_web_search": true }, "xai/grok-beta": { - "input_cost_per_token": 0.000005, + "input_cost_per_token": 5e-6, "litellm_provider": "xai", "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "supports_function_calling": true, "supports_tool_choice": true, "supports_vision": true, @@ -26442,7 +24419,7 @@ "max_output_tokens": 256000, "max_tokens": 256000, "mode": "chat", - "output_cost_per_token": 0.0000015, + "output_cost_per_token": 1.5e-6, "source": "https://docs.x.ai/docs/models", "supports_function_calling": true, "supports_reasoning": true, @@ -26456,7 +24433,7 @@ "max_output_tokens": 256000, "max_tokens": 256000, "mode": "chat", - "output_cost_per_token": 0.0000015, + "output_cost_per_token": 1.5e-6, "source": "https://docs.x.ai/docs/models", "supports_function_calling": true, "supports_reasoning": true, @@ -26470,28 +24447,28 @@ "max_output_tokens": 256000, "max_tokens": 256000, "mode": "chat", - "output_cost_per_token": 0.0000015, + "output_cost_per_token": 1.5e-6, "source": "https://docs.x.ai/docs/models", "supports_function_calling": true, "supports_reasoning": true, "supports_tool_choice": true }, "xai/grok-vision-beta": { - "input_cost_per_image": 0.000005, - "input_cost_per_token": 0.000005, + "input_cost_per_image": 5e-6, + "input_cost_per_token": 5e-6, "litellm_provider": "xai", "max_input_tokens": 8192, "max_output_tokens": 8192, "max_tokens": 8192, "mode": "chat", - "output_cost_per_token": 0.000015, + "output_cost_per_token": 1.5e-5, "supports_function_calling": true, "supports_tool_choice": true, "supports_vision": true, "supports_web_search": true }, "vertex_ai/search_api": { - "input_cost_per_query": 0.0015, + "input_cost_per_query": 1.5e-3, "litellm_provider": "vertex_ai", "mode": "vector_store" }, @@ -26505,99 +24482,54 @@ "mode": "video_generation", "output_cost_per_video_per_second": 0.1, "source": "https://platform.openai.com/docs/api-reference/videos", - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "video" - ], - "supported_resolutions": [ - "720x1280", - "1280x720" - ] + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["video"], + "supported_resolutions": ["720x1280", "1280x720"] }, "openai/sora-2-pro": { "litellm_provider": "openai", "mode": "video_generation", "output_cost_per_video_per_second": 0.3, "source": "https://platform.openai.com/docs/api-reference/videos", - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "video" - ], - "supported_resolutions": [ - "720x1280", - "1280x720" - ] + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["video"], + "supported_resolutions": ["720x1280", "1280x720"] }, "azure/sora-2": { "litellm_provider": "azure", "mode": "video_generation", "output_cost_per_video_per_second": 0.1, "source": "https://azure.microsoft.com/en-us/products/ai-services/video-generation", - "supported_modalities": [ - "text" - ], - "supported_output_modalities": [ - "video" - ], - "supported_resolutions": [ - "720x1280", - "1280x720" - ] + "supported_modalities": ["text"], + "supported_output_modalities": ["video"], + "supported_resolutions": ["720x1280", "1280x720"] }, "azure/sora-2-pro": { "litellm_provider": "azure", "mode": "video_generation", "output_cost_per_video_per_second": 0.3, "source": "https://azure.microsoft.com/en-us/products/ai-services/video-generation", - "supported_modalities": [ - "text" - ], - "supported_output_modalities": [ - "video" - ], - "supported_resolutions": [ - "720x1280", - "1280x720" - ] + "supported_modalities": ["text"], + "supported_output_modalities": ["video"], + "supported_resolutions": ["720x1280", "1280x720"] }, "azure/sora-2-pro-high-res": { "litellm_provider": "azure", "mode": "video_generation", "output_cost_per_video_per_second": 0.5, "source": "https://azure.microsoft.com/en-us/products/ai-services/video-generation", - "supported_modalities": [ - "text" - ], - "supported_output_modalities": [ - "video" - ], - "supported_resolutions": [ - "1024x1792", - "1792x1024" - ] + "supported_modalities": ["text"], + "supported_output_modalities": ["video"], + "supported_resolutions": ["1024x1792", "1792x1024"] }, "runwayml/gen4_turbo": { "litellm_provider": "runwayml", "mode": "video_generation", "output_cost_per_video_per_second": 0.05, "source": "https://docs.dev.runwayml.com/guides/pricing/", - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "video" - ], - "supported_resolutions": [ - "1280x720", - "720x1280" - ], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["video"], + "supported_resolutions": ["1280x720", "720x1280"], "metadata": { "comment": "5 credits per second @ $0.01 per credit = $0.05 per second" } @@ -26607,17 +24539,9 @@ "mode": "video_generation", "output_cost_per_video_per_second": 0.15, "source": "https://docs.dev.runwayml.com/guides/pricing/", - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "video" - ], - "supported_resolutions": [ - "1280x720", - "720x1280" - ], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["video"], + "supported_resolutions": ["1280x720", "720x1280"], "metadata": { "comment": "15 credits per second @ $0.01 per credit = $0.15 per second" } @@ -26627,17 +24551,9 @@ "mode": "video_generation", "output_cost_per_video_per_second": 0.05, "source": "https://docs.dev.runwayml.com/guides/pricing/", - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "video" - ], - "supported_resolutions": [ - "1280x720", - "720x1280" - ], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["video"], + "supported_resolutions": ["1280x720", "720x1280"], "metadata": { "comment": "5 credits per second @ $0.01 per credit = $0.05 per second" } @@ -26648,17 +24564,9 @@ "input_cost_per_image": 0.05, "output_cost_per_image": 0.05, "source": "https://docs.dev.runwayml.com/guides/pricing/", - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "image" - ], - "supported_resolutions": [ - "1280x720", - "1920x1080" - ], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["image"], + "supported_resolutions": ["1280x720", "1920x1080"], "metadata": { "comment": "5 credits per 720p image or 8 credits per 1080p image @ $0.01 per credit. Using 5 credits ($0.05) as base cost" } @@ -26669,17 +24577,9 @@ "input_cost_per_image": 0.02, "output_cost_per_image": 0.02, "source": "https://docs.dev.runwayml.com/guides/pricing/", - "supported_modalities": [ - "text", - "image" - ], - "supported_output_modalities": [ - "image" - ], - "supported_resolutions": [ - "1280x720", - "1920x1080" - ], + "supported_modalities": ["text", "image"], + "supported_output_modalities": ["image"], + "supported_resolutions": ["1280x720", "1920x1080"], "metadata": { "comment": "2 credits per image (any resolution) @ $0.01 per credit = $0.02 per image" } @@ -26693,4 +24593,4 @@ "comment": "Estimated cost based on standard TTS pricing. RunwayML uses ElevenLabs models." } } -} \ No newline at end of file +} diff --git a/src/node/services/aiService.ts b/src/node/services/aiService.ts index 6cb748f913..81222beb1e 100644 --- a/src/node/services/aiService.ts +++ b/src/node/services/aiService.ts @@ -33,7 +33,10 @@ import type { HistoryService } from "./historyService"; import type { PartialService } from "./partialService"; import { buildSystemMessage, readToolInstructions } from "./systemMessage"; import { getTokenizerForModel } from "@/node/utils/main/tokenizer"; -import { buildProviderOptions } from "@/common/utils/ai/providerOptions"; +import { + buildProviderOptions, + calculateEffectiveMaxOutputTokens, +} from "@/common/utils/ai/providerOptions"; import type { ThinkingLevel } from "@/common/types/thinking"; import type { StreamAbortEvent, @@ -926,6 +929,15 @@ export class AIService extends EventEmitter { effectiveMuxProviderOptions ); + // Calculate effective maxOutputTokens that accounts for thinking budget + // For Anthropic models with extended thinking, the SDK adds thinkingBudget to maxOutputTokens + // so we need to ensure the sum doesn't exceed the model's max_output_tokens limit + const effectiveMaxOutputTokens = calculateEffectiveMaxOutputTokens( + effectiveModelString, + thinkingLevel ?? "off", + maxOutputTokens + ); + // Delegate to StreamManager with model instance, system message, tools, historySequence, and initial metadata const streamResult = await this.streamManager.startStream( workspaceId, @@ -943,7 +955,7 @@ export class AIService extends EventEmitter { mode, // Pass mode so it persists in final history entry }, providerOptions, - maxOutputTokens, + effectiveMaxOutputTokens, toolPolicy, streamToken // Pass the pre-generated stream token );