# Chutes.ai Chutes.ai is an AI inference platform that provides access to a wide variety of open-source large language models through an OpenAI-compatible API. The platform offers models from leading providers including Qwen, DeepSeek, Mistral, Google, NousResearch, Meta, and others. Chutes.ai features transparent pricing per 1M tokens, context length information, and model capabilities including JSON mode, tools/function calling, structured outputs, and reasoning. The platform supports both standard and confidential compute variants of models. ## Provider Information - **Website**: - **Available Models**: 25 ## Models | Name | Original Name | $ Input Price (per 1M) | $ Output Price (per 1M) | Free | Link | |------|---------------|---------------------|----------------------|------|------| | Qwen3-32B | Qwen/Qwen3-32B | 0.08 | 0.24 | | | | Mistral-Nemo-Instruct-2407 | unsloth/Mistral-Nemo-Instruct-2407 | 0.02 | 0.04 | | | | Mistral-Small-3.1-24B-Instruct-2503 | chutesai/Mistral-Small-3.1-24B-Instruct-2503 | 0.03 | 0.11 | | | | gemma-3-4b-it | unsloth/gemma-3-4b-it | 0.01 | 0.03 | | | | DeepSeek-V3 | deepseek-ai/DeepSeek-V3 | 0.30 | 1.20 | | | | Qwen3-14B | Qwen/Qwen3-14B | 0.05 | 0.22 | | | | Mistral-Small-3.2-24B-Instruct-2506 | chutesai/Mistral-Small-3.2-24B-Instruct-2506 | 0.06 | 0.18 | | | | DeepSeek-R1-Distill-Llama-70B | deepseek-ai/DeepSeek-R1-Distill-Llama-70B | 0.03 | 0.11 | | | | gpt-oss-20b | openai/gpt-oss-20b | 0.04 | 0.15 | | | | gemma-3-27b-it | unsloth/gemma-3-27b-it | 0.04 | 0.15 | | | | Qwen2.5-Coder-32B-Instruct | Qwen/Qwen2.5-Coder-32B-Instruct | 0.03 | 0.11 | | | | gemma-3-12b-it | unsloth/gemma-3-12b-it | 0.03 | 0.10 | | | | Qwen2.5-72B-Instruct | Qwen/Qwen2.5-72B-Instruct | 0.30 | 1.20 | | | | Qwen3-Next-80B-A3B-Instruct | Qwen/Qwen3-Next-80B-A3B-Instruct | 0.10 | 0.80 | | | | Qwen3-30B-A3B | Qwen/Qwen3-30B-A3B | 0.06 | 0.22 | | | | Qwen3-235B-A22B-Thinking-2507 | Qwen/Qwen3-235B-A22B-Thinking-2507 | 0.11 | 0.60 | | | | Qwen3-VL-235B-A22B-Instruct | Qwen/Qwen3-VL-235B-A22B-Instruct | 0.30 | 1.20 | | | | Mistral-Small-24B-Instruct-2501 | unsloth/Mistral-Small-24B-Instruct-2501 | 0.07 | 0.30 | | | | GLM-4.6V | zai-org/GLM-4.6V | 0.30 | 0.90 | | | | Qwen2.5-VL-32B-Instruct | Qwen/Qwen2.5-VL-32B-Instruct | 0.05 | 0.22 | | | | NVIDIA-Nemotron-3-Nano-30B-A3B-BF16 | nvidia/NVIDIA-Nemotron-3-Nano-30B-A3B-BF16 | 0.06 | 0.24 | | | | GLM-4.7-FP8 | zai-org/GLM-4.7-FP8 | 0.30 | 1.20 | | | | Llama-3.2-1B-Instruct | unsloth/Llama-3.2-1B-Instruct | 0.01 | 0.01 | | | | DeepHermes-3-Mistral-24B-Preview | NousResearch/DeepHermes-3-Mistral-24B-Preview | 0.02 | 0.10 | | | | Llama-3.2-3B-Instruct | unsloth/Llama-3.2-3B-Instruct | 0.01 | 0.01 | | | --- [← Back to all providers](/llm.txt)