{ "cerebras/zai-glm-4.6": { "provider": "cerebras", "mode": "chat", "input_cost_per_token": 0.00000225, "output_cost_per_token": 0.00000275, "max_input_tokens": 131000, "max_output_tokens": 40000, "max_tokens": 40000, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_reasoning": true, "supports_response_schema": true, "supports_tool_choice": true }, "cerebras/zai-glm-4.7": { "provider": "cerebras", "mode": "chat", "input_cost_per_token": 0.00000225, "output_cost_per_token": 0.00000275, "max_input_tokens": 131000, "max_output_tokens": 40000, "max_tokens": 40000, "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_reasoning": true, "supports_response_schema": true, "supports_tool_choice": true }, "cerebras/qwen-3-235b-a22b-instruct-2507": { "provider": "cerebras", "mode": "chat", "input_cost_per_token": 6E-7, "output_cost_per_token": 0.0000012, "max_input_tokens": 131072, "max_output_tokens": 40000, "max_tokens": 40000, "supports_function_calling": true, "supports_parallel_function_calling": false, "supports_reasoning": false, "supports_response_schema": true, "supports_tool_choice": true }, "cloud/kimi-k2-thinking": { "provider": "cloud", "mode": "chat", "input_cost_per_token": 6E-7, "output_cost_per_token": 0.0000025, "max_input_tokens": 262144, "max_output_tokens": 16384, "max_tokens": 16384, "supports_function_calling": true, "supports_tool_choice": true, "supports_reasoning": true }, "cloud/kimi-k2:1t": { "provider": "cloud", "mode": "chat", "input_cost_per_token": 5E-7, "output_cost_per_token": 5E-7, "max_input_tokens": 262144, "max_output_tokens": 16384, "max_tokens": 16384, "supports_function_calling": true, "supports_tool_choice": true }, "cloud/minimax-m2": { "provider": "cloud", "mode": "chat", "input_cost_per_token": 3E-7, "output_cost_per_token": 0.0000012, "max_input_tokens": 200000, "max_output_tokens": 131072, "max_tokens": 131072, "supports_function_calling": true, "supports_tool_choice": true }, "cloud/qwen3-vl:235b-instruct": { "provider": "cloud", "mode": "chat", "input_cost_per_token": 2.2E-7, "output_cost_per_token": 8.8E-7, "max_input_tokens": 262144, "max_output_tokens": 32768, "max_tokens": 32768, "supports_vision": true, "supports_function_calling": true, "supports_tool_choice": true }, "cloud/qwen3-vl:235b": { "provider": "cloud", "mode": "chat", "input_cost_per_token": 2.2E-7, "output_cost_per_token": 8.8E-7, "max_input_tokens": 262144, "max_output_tokens": 32768, "max_tokens": 32768, "supports_vision": true }, "cloud/deepseek-v3.1:671b": { "provider": "cloud", "mode": "chat", "input_cost_per_token": 5.6E-7, "output_cost_per_token": 0.00000168, "max_input_tokens": 163840, "max_output_tokens": 32768, "max_tokens": 32768, "supports_function_calling": true, "supports_tool_choice": true, "supports_reasoning": true }, "cloud/gpt-oss:120b": { "provider": "cloud", "mode": "chat", "input_cost_per_token": 1.5E-7, "output_cost_per_token": 6E-7, "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "supports_function_calling": true, "supports_tool_choice": true, "supports_reasoning": true }, "cloud/gpt-oss:20b": { "provider": "cloud", "mode": "chat", "input_cost_per_token": 7E-8, "output_cost_per_token": 3E-7, "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "supports_function_calling": true, "supports_tool_choice": true, "supports_reasoning": true }, "cloud/glm-4.6": { "provider": "cloud", "mode": "chat", "input_cost_per_token": 5.5E-7, "output_cost_per_token": 0.00000219, "max_input_tokens": 198000, "max_output_tokens": 65536, "max_tokens": 65536, "supports_function_calling": true, "supports_tool_choice": true, "supports_reasoning": true }, "cloud/qwen3-coder:480b": { "provider": "cloud", "mode": "chat", "input_cost_per_token": 4.5E-7, "output_cost_per_token": 0.0000018, "max_input_tokens": 262144, "max_output_tokens": 65536, "max_tokens": 65536, "supports_function_calling": true, "supports_tool_choice": true }, "cerebras/gpt-oss-120b": { "input_cost_per_token": 2.5E-7, "max_input_tokens": 131072, "max_output_tokens": 32768, "max_tokens": 32768, "mode": "chat", "output_cost_per_token": 6.9E-7, "source": "https://www.cerebras.ai/blog/openai-gpt-oss-120b-runs-fastest-on-cerebras", "supports_function_calling": true, "supports_parallel_function_calling": true, "supports_reasoning": true, "supports_response_schema": true, "supports_tool_choice": true, "provider": "cerebras" }, "cloud/ministral-3:3b": { "provider": "cloud", "mode": "chat", "input_cost_per_token": 4E-8, "output_cost_per_token": 4E-8, "max_input_tokens": 262144, "max_output_tokens": 32768, "max_tokens": 32768, "supports_vision": true, "supports_function_calling": true, "supports_tool_choice": true }, "cloud/ministral-3:8b": { "provider": "cloud", "mode": "chat", "input_cost_per_token": 1E-7, "output_cost_per_token": 1E-7, "max_input_tokens": 262144, "max_output_tokens": 32768, "max_tokens": 32768, "supports_vision": true, "supports_function_calling": true, "supports_tool_choice": true }, "cloud/ministral-3:14b": { "provider": "cloud", "mode": "chat", "input_cost_per_token": 2E-7, "output_cost_per_token": 2E-7, "max_input_tokens": 262144, "max_output_tokens": 32768, "max_tokens": 32768, "supports_vision": true, "supports_function_calling": true, "supports_tool_choice": true }, "cloud/mistral-large-3:675b": { "provider": "cloud", "mode": "chat", "input_cost_per_token": 5E-7, "output_cost_per_token": 0.0000015, "max_input_tokens": 262144, "max_output_tokens": 32768, "max_tokens": 32768, "supports_vision": true, "supports_function_calling": true, "supports_tool_choice": true }, "cloud/gemini-3-pro-preview": { "provider": "cloud", "mode": "chat", "input_cost_per_token": 0.000002, "output_cost_per_token": 0.000012, "max_input_tokens": 1048576, "max_output_tokens": 65536, "max_tokens": 65536, "supports_vision": true, "supports_function_calling": true, "supports_tool_choice": true, "supports_reasoning": true }, "cloud/deepseek-v3.2": { "provider": "cloud", "mode": "chat", "input_cost_per_token": 2.8E-7, "output_cost_per_token": 4.2E-7, "max_input_tokens": 131072, "max_output_tokens": 65536, "max_tokens": 65536, "supports_function_calling": true, "supports_tool_choice": true, "supports_reasoning": true }, "cloud/qwen3-next:80b": { "provider": "cloud", "mode": "chat", "input_cost_per_token": 1.5E-7, "output_cost_per_token": 0.0000015, "max_input_tokens": 262144, "max_output_tokens": 32768, "max_tokens": 32768, "supports_function_calling": true, "supports_tool_choice": true, "supports_reasoning": true }, "cloud/devstral-2:123b": { "provider": "cloud", "mode": "chat", "input_cost_per_token": 1.5E-7, "output_cost_per_token": 6E-7, "max_input_tokens": 262144, "max_output_tokens": 32768, "max_tokens": 32768, "supports_function_calling": true, "supports_tool_choice": true }, "cloud/gemini-3-flash-preview": { "provider": "cloud", "mode": "chat", "input_cost_per_token": 5E-7, "output_cost_per_token": 0.000003, "max_input_tokens": 1048576, "max_output_tokens": 65535, "max_tokens": 65535, "supports_function_calling": true, "supports_reasoning": true, "supports_response_schema": true, "supports_tool_choice": true, "supports_vision": true }, "cloud/nemotron-3-nano:30b": { "provider": "cloud", "mode": "chat", "input_cost_per_token": 6E-8, "output_cost_per_token": 2.4E-7, "max_input_tokens": 1000000, "max_output_tokens": 262144, "max_tokens": 262144, "supports_function_calling": true, "supports_reasoning": true, "supports_response_schema": true, "supports_tool_choice": true }, "cloud/minimax-m2.1": { "provider": "cloud", "mode": "chat", "input_cost_per_token": 3E-7, "output_cost_per_token": 0.0000012, "max_input_tokens": 204800, "max_output_tokens": 131072, "max_tokens": 131072, "supports_function_calling": true, "supports_reasoning": true, "supports_response_schema": true, "supports_tool_choice": true }, "cloud/glm-4.7": { "provider": "cloud", "mode": "chat", "input_cost_per_token": 4E-7, "output_cost_per_token": 0.0000015, "max_input_tokens": 202752, "max_output_tokens": 65535, "max_tokens": 65535, "supports_function_calling": true, "supports_reasoning": true, "supports_response_schema": true, "supports_tool_choice": true }, "opencode/glm-4.7-free": { "provider": "opencode", "mode": "chat", "input_cost_per_token": 4E-7, "output_cost_per_token": 0.0000015, "max_input_tokens": 202752, "max_output_tokens": 65535, "max_tokens": 65535, "supports_function_calling": true, "supports_parallel_function_calling": false, "supports_reasoning": true, "supports_response_schema": true, "supports_tool_choice": true, "supports_vision": false }, "opencode/grok-code": { "provider": "opencode", "mode": "chat", "input_cost_per_token": 2E-7, "output_cost_per_token": 0.0000015, "max_input_tokens": 256000, "max_output_tokens": 10000, "max_tokens": 10000, "supports_function_calling": true, "supports_parallel_function_calling": false, "supports_reasoning": true, "supports_response_schema": true, "supports_tool_choice": true, "supports_vision": false }, "opencode/gpt-5-nano": { "provider": "opencode", "mode": "chat", "input_cost_per_token": 5E-8, "output_cost_per_token": 4E-7, "max_input_tokens": 400000, "max_output_tokens": 128000, "max_tokens": 128000, "supports_function_calling": true, "supports_parallel_function_calling": false, "supports_reasoning": true, "supports_response_schema": true, "supports_tool_choice": true, "supports_vision": true }, "opencode/minimax-m2.1-free": { "provider": "opencode", "mode": "chat", "input_cost_per_token": 3E-7, "output_cost_per_token": 0.0000012, "max_input_tokens": 204800, "max_output_tokens": 131072, "max_tokens": 131072, "supports_function_calling": true, "supports_parallel_function_calling": false, "supports_reasoning": true, "supports_response_schema": true, "supports_tool_choice": true, "supports_vision": false }, "cloud/gpt-5.2": { "provider": "cloud", "mode": "chat", "input_cost_per_token": 0.00000175, "output_cost_per_token": 0.000014, "max_input_tokens": 400000, "max_output_tokens": 128000, "max_tokens": 128000, "supports_function_calling": true, "supports_reasoning": true, "supports_response_schema": true, "supports_tool_choice": true, "supports_vision": true }, "openrouter/kimi-k2-thinking": { "provider": "openrouter", "mode": "chat", "input_cost_per_token": 6E-7, "output_cost_per_token": 0.0000025, "max_input_tokens": 262144, "max_output_tokens": 16384, "max_tokens": 16384, "supports_function_calling": true, "supports_tool_choice": true, "supports_reasoning": true }, "openrouter/kimi-k2:1t": { "provider": "openrouter", "mode": "chat", "input_cost_per_token": 5E-7, "output_cost_per_token": 5E-7, "max_input_tokens": 262144, "max_output_tokens": 16384, "max_tokens": 16384, "supports_function_calling": true, "supports_tool_choice": true }, "openrouter/minimax-m2": { "provider": "openrouter", "mode": "chat", "input_cost_per_token": 3E-7, "output_cost_per_token": 0.0000012, "max_input_tokens": 200000, "max_output_tokens": 131072, "max_tokens": 131072, "supports_function_calling": true, "supports_tool_choice": true }, "openrouter/qwen3-vl:235b-instruct": { "provider": "openrouter", "mode": "chat", "input_cost_per_token": 2.2E-7, "output_cost_per_token": 8.8E-7, "max_input_tokens": 262144, "max_output_tokens": 32768, "max_tokens": 32768, "supports_vision": true, "supports_function_calling": true, "supports_tool_choice": true }, "openrouter/qwen3-vl:235b": { "provider": "openrouter", "mode": "chat", "input_cost_per_token": 2.2E-7, "output_cost_per_token": 8.8E-7, "max_input_tokens": 262144, "max_output_tokens": 32768, "max_tokens": 32768, "supports_vision": true }, "openrouter/deepseek-v3.1:671b": { "provider": "openrouter", "mode": "chat", "input_cost_per_token": 5.6E-7, "output_cost_per_token": 0.00000168, "max_input_tokens": 163840, "max_output_tokens": 32768, "max_tokens": 32768, "supports_function_calling": true, "supports_tool_choice": true, "supports_reasoning": true }, "openrouter/gpt-oss:120b": { "provider": "openrouter", "mode": "chat", "input_cost_per_token": 1.5E-7, "output_cost_per_token": 6E-7, "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "supports_function_calling": true, "supports_tool_choice": true, "supports_reasoning": true }, "openrouter/gpt-oss:20b": { "provider": "openrouter", "mode": "chat", "input_cost_per_token": 7E-8, "output_cost_per_token": 3E-7, "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, "supports_function_calling": true, "supports_tool_choice": true, "supports_reasoning": true }, "openrouter/glm-4.6": { "provider": "openrouter", "mode": "chat", "input_cost_per_token": 5.5E-7, "output_cost_per_token": 0.00000219, "max_input_tokens": 198000, "max_output_tokens": 65536, "max_tokens": 65536, "supports_function_calling": true, "supports_tool_choice": true, "supports_reasoning": true }, "openrouter/qwen3-coder:480b": { "provider": "openrouter", "mode": "chat", "input_cost_per_token": 4.5E-7, "output_cost_per_token": 0.0000018, "max_input_tokens": 262144, "max_output_tokens": 65536, "max_tokens": 65536, "supports_function_calling": true, "supports_tool_choice": true }, "openrouter/ministral-3:3b": { "provider": "openrouter", "mode": "chat", "input_cost_per_token": 4E-8, "output_cost_per_token": 4E-8, "max_input_tokens": 262144, "max_output_tokens": 32768, "max_tokens": 32768, "supports_vision": true, "supports_function_calling": true, "supports_tool_choice": true }, "openrouter/ministral-3:8b": { "provider": "openrouter", "mode": "chat", "input_cost_per_token": 1E-7, "output_cost_per_token": 1E-7, "max_input_tokens": 262144, "max_output_tokens": 32768, "max_tokens": 32768, "supports_vision": true, "supports_function_calling": true, "supports_tool_choice": true }, "openrouter/ministral-3:14b": { "provider": "openrouter", "mode": "chat", "input_cost_per_token": 2E-7, "output_cost_per_token": 2E-7, "max_input_tokens": 262144, "max_output_tokens": 32768, "max_tokens": 32768, "supports_vision": true, "supports_function_calling": true, "supports_tool_choice": true }, "openrouter/mistral-large-3:675b": { "provider": "openrouter", "mode": "chat", "input_cost_per_token": 5E-7, "output_cost_per_token": 0.0000015, "max_input_tokens": 262144, "max_output_tokens": 32768, "max_tokens": 32768, "supports_vision": true, "supports_function_calling": true, "supports_tool_choice": true }, "openrouter/deepseek-v3.2": { "provider": "openrouter", "mode": "chat", "input_cost_per_token": 2.8E-7, "output_cost_per_token": 4.2E-7, "max_input_tokens": 131072, "max_output_tokens": 65536, "max_tokens": 65536, "supports_function_calling": true, "supports_tool_choice": true, "supports_reasoning": true }, "openrouter/qwen3-next:80b": { "provider": "openrouter", "mode": "chat", "input_cost_per_token": 1.5E-7, "output_cost_per_token": 0.0000015, "max_input_tokens": 262144, "max_output_tokens": 32768, "max_tokens": 32768, "supports_function_calling": true, "supports_tool_choice": true, "supports_reasoning": true }, "openrouter/devstral-2:123b": { "provider": "openrouter", "mode": "chat", "input_cost_per_token": 1.5E-7, "output_cost_per_token": 6E-7, "max_input_tokens": 262144, "max_output_tokens": 32768, "max_tokens": 32768, "supports_function_calling": true, "supports_tool_choice": true }, "openrouter/gemini-3-flash-preview": { "provider": "openrouter", "mode": "chat", "input_cost_per_token": 5E-7, "output_cost_per_token": 0.000003, "max_input_tokens": 1048576, "max_output_tokens": 65535, "max_tokens": 65535, "supports_function_calling": true, "supports_reasoning": true, "supports_response_schema": true, "supports_tool_choice": true, "supports_vision": true }, "openrouter/nemotron-3-nano:30b": { "provider": "openrouter", "mode": "chat", "input_cost_per_token": 6E-8, "output_cost_per_token": 2.4E-7, "max_input_tokens": 1000000, "max_output_tokens": 262144, "max_tokens": 262144, "supports_function_calling": true, "supports_reasoning": true, "supports_response_schema": true, "supports_tool_choice": true }, "openrouter/minimax-m2.1": { "provider": "openrouter", "mode": "chat", "input_cost_per_token": 3E-7, "output_cost_per_token": 0.0000012, "max_input_tokens": 204800, "max_output_tokens": 131072, "max_tokens": 131072, "supports_function_calling": true, "supports_reasoning": true, "supports_response_schema": true, "supports_tool_choice": true }, "openrouter/glm-4.7": { "provider": "openrouter", "mode": "chat", "input_cost_per_token": 4E-7, "output_cost_per_token": 0.0000015, "max_input_tokens": 202752, "max_output_tokens": 65535, "max_tokens": 65535, "supports_function_calling": true, "supports_reasoning": true, "supports_response_schema": true, "supports_tool_choice": true } }