233 lines
6.6 KiB
JSON
233 lines
6.6 KiB
JSON
{
|
|
"cerebras/zai-glm-4.6": {
|
|
"provider": "cerebras",
|
|
"mode": "chat",
|
|
"input_cost_per_token": 0.00000225,
|
|
"output_cost_per_token": 0.00000275,
|
|
"max_input_tokens": 131000,
|
|
"max_output_tokens": 40000,
|
|
"max_tokens": 40000,
|
|
"supports_function_calling": true,
|
|
"supports_parallel_function_calling": true,
|
|
"supports_reasoning": true,
|
|
"supports_response_schema": true,
|
|
"supports_tool_choice": true
|
|
},
|
|
"cerebras/qwen-3-235b-a22b-instruct-2507": {
|
|
"provider": "cerebras",
|
|
"mode": "chat",
|
|
"input_cost_per_token": 6E-7,
|
|
"output_cost_per_token": 0.0000012,
|
|
"max_input_tokens": 131072,
|
|
"max_output_tokens": 40000,
|
|
"max_tokens": 40000,
|
|
"supports_function_calling": true,
|
|
"supports_parallel_function_calling": false,
|
|
"supports_reasoning": false,
|
|
"supports_response_schema": true,
|
|
"supports_tool_choice": true
|
|
},
|
|
"cloud/kimi-k2-thinking": {
|
|
"provider": "cloud",
|
|
"mode": "chat",
|
|
"input_cost_per_token": 6E-7,
|
|
"output_cost_per_token": 0.0000025,
|
|
"max_input_tokens": 262144,
|
|
"max_output_tokens": 16384,
|
|
"max_tokens": 16384,
|
|
"supports_function_calling": true,
|
|
"supports_tool_choice": true,
|
|
"supports_reasoning": true
|
|
},
|
|
"cloud/kimi-k2:1t": {
|
|
"provider": "cloud",
|
|
"mode": "chat",
|
|
"input_cost_per_token": 5E-7,
|
|
"output_cost_per_token": 5E-7,
|
|
"max_input_tokens": 262144,
|
|
"max_output_tokens": 16384,
|
|
"max_tokens": 16384,
|
|
"supports_function_calling": true,
|
|
"supports_tool_choice": true
|
|
},
|
|
"cloud/minimax-m2": {
|
|
"provider": "cloud",
|
|
"mode": "chat",
|
|
"input_cost_per_token": 3E-7,
|
|
"output_cost_per_token": 0.0000012,
|
|
"max_input_tokens": 200000,
|
|
"max_output_tokens": 131072,
|
|
"max_tokens": 131072,
|
|
"supports_function_calling": true,
|
|
"supports_tool_choice": true
|
|
},
|
|
"cloud/qwen3-vl:235b-instruct": {
|
|
"provider": "cloud",
|
|
"mode": "chat",
|
|
"input_cost_per_token": 2.2E-7,
|
|
"output_cost_per_token": 8.8E-7,
|
|
"max_input_tokens": 262144,
|
|
"max_output_tokens": 32768,
|
|
"max_tokens": 32768,
|
|
"supports_vision": true,
|
|
"supports_function_calling": true,
|
|
"supports_tool_choice": true
|
|
},
|
|
"cloud/qwen3-vl:235b": {
|
|
"provider": "cloud",
|
|
"mode": "chat",
|
|
"input_cost_per_token": 2.2E-7,
|
|
"output_cost_per_token": 8.8E-7,
|
|
"max_input_tokens": 262144,
|
|
"max_output_tokens": 32768,
|
|
"max_tokens": 32768,
|
|
"supports_vision": true
|
|
},
|
|
"cloud/deepseek-v3.1:671b": {
|
|
"provider": "cloud",
|
|
"mode": "chat",
|
|
"input_cost_per_token": 5.6E-7,
|
|
"output_cost_per_token": 0.00000168,
|
|
"max_input_tokens": 163840,
|
|
"max_output_tokens": 32768,
|
|
"max_tokens": 32768,
|
|
"supports_function_calling": true,
|
|
"supports_tool_choice": true,
|
|
"supports_reasoning": true
|
|
},
|
|
"cloud/gpt-oss:120b": {
|
|
"provider": "cloud",
|
|
"mode": "chat",
|
|
"input_cost_per_token": 1.5E-7,
|
|
"output_cost_per_token": 6E-7,
|
|
"max_input_tokens": 131072,
|
|
"max_output_tokens": 131072,
|
|
"max_tokens": 131072,
|
|
"supports_function_calling": true,
|
|
"supports_tool_choice": true,
|
|
"supports_reasoning": true
|
|
},
|
|
"cloud/gpt-oss:20b": {
|
|
"provider": "cloud",
|
|
"mode": "chat",
|
|
"input_cost_per_token": 7E-8,
|
|
"output_cost_per_token": 3E-7,
|
|
"max_input_tokens": 131072,
|
|
"max_output_tokens": 131072,
|
|
"max_tokens": 131072,
|
|
"supports_function_calling": true,
|
|
"supports_tool_choice": true,
|
|
"supports_reasoning": true
|
|
},
|
|
"cloud/glm-4.6": {
|
|
"provider": "cloud",
|
|
"mode": "chat",
|
|
"input_cost_per_token": 5.5E-7,
|
|
"output_cost_per_token": 2.19E-6,
|
|
"max_input_tokens": 198000,
|
|
"max_output_tokens": 65536,
|
|
"max_tokens": 65536,
|
|
"supports_function_calling": true,
|
|
"supports_tool_choice": true,
|
|
"supports_reasoning": true
|
|
},
|
|
"cloud/qwen3-coder:480b": {
|
|
"provider": "cloud",
|
|
"mode": "chat",
|
|
"input_cost_per_token": 4.5E-7,
|
|
"output_cost_per_token": 0.0000018,
|
|
"max_input_tokens": 262144,
|
|
"max_output_tokens": 65536,
|
|
"max_tokens": 65536,
|
|
"supports_function_calling": true,
|
|
"supports_tool_choice": true
|
|
},
|
|
"cerebras/gpt-oss-120b": {
|
|
"input_cost_per_token": 2.5E-7,
|
|
"max_input_tokens": 131072,
|
|
"max_output_tokens": 32768,
|
|
"max_tokens": 32768,
|
|
"mode": "chat",
|
|
"output_cost_per_token": 6.9E-7,
|
|
"source": "https://www.cerebras.ai/blog/openai-gpt-oss-120b-runs-fastest-on-cerebras",
|
|
"supports_function_calling": true,
|
|
"supports_parallel_function_calling": true,
|
|
"supports_reasoning": true,
|
|
"supports_response_schema": true,
|
|
"supports_tool_choice": true,
|
|
"provider": "cerebras"
|
|
},
|
|
"cloud/ministral-3:3b": {
|
|
"provider": "cloud",
|
|
"mode": "chat",
|
|
"input_cost_per_token": 4E-8,
|
|
"output_cost_per_token": 4E-8,
|
|
"max_input_tokens": 262144,
|
|
"max_output_tokens": 32768,
|
|
"max_tokens": 32768,
|
|
"supports_vision": true,
|
|
"supports_function_calling": true,
|
|
"supports_tool_choice": true
|
|
},
|
|
"cloud/ministral-3:8b": {
|
|
"provider": "cloud",
|
|
"mode": "chat",
|
|
"input_cost_per_token": 1E-7,
|
|
"output_cost_per_token": 1E-7,
|
|
"max_input_tokens": 262144,
|
|
"max_output_tokens": 32768,
|
|
"max_tokens": 32768,
|
|
"supports_vision": true,
|
|
"supports_function_calling": true,
|
|
"supports_tool_choice": true
|
|
},
|
|
"cloud/ministral-3:14b": {
|
|
"provider": "cloud",
|
|
"mode": "chat",
|
|
"input_cost_per_token": 2E-7,
|
|
"output_cost_per_token": 2E-7,
|
|
"max_input_tokens": 262144,
|
|
"max_output_tokens": 32768,
|
|
"max_tokens": 32768,
|
|
"supports_vision": true,
|
|
"supports_function_calling": true,
|
|
"supports_tool_choice": true
|
|
},
|
|
"cloud/mistral-large-3:675b": {
|
|
"provider": "cloud",
|
|
"mode": "chat",
|
|
"input_cost_per_token": 5E-7,
|
|
"output_cost_per_token": 1.5E-6,
|
|
"max_input_tokens": 262144,
|
|
"max_output_tokens": 32768,
|
|
"max_tokens": 32768,
|
|
"supports_vision": true,
|
|
"supports_function_calling": true,
|
|
"supports_tool_choice": true
|
|
},
|
|
"cloud/gemini-3-pro-preview": {
|
|
"provider": "cloud",
|
|
"mode": "chat",
|
|
"input_cost_per_token": 2E-6,
|
|
"output_cost_per_token": 1.2E-5,
|
|
"max_input_tokens": 1048576,
|
|
"max_output_tokens": 65536,
|
|
"max_tokens": 65536,
|
|
"supports_vision": true,
|
|
"supports_function_calling": true,
|
|
"supports_tool_choice": true,
|
|
"supports_reasoning": true
|
|
},
|
|
"cloud/deepseek-v3.2": {
|
|
"provider": "cloud",
|
|
"mode": "chat",
|
|
"input_cost_per_token": 2.8E-7,
|
|
"output_cost_per_token": 4.2E-7,
|
|
"max_input_tokens": 131072,
|
|
"max_output_tokens": 65536,
|
|
"max_tokens": 65536,
|
|
"supports_function_calling": true,
|
|
"supports_tool_choice": true,
|
|
"supports_reasoning": true
|
|
}
|
|
} |