All Models
DeepSeek R1 Distill Qwen 32B
DeepSeek R1 Distill Qwen 32B is a distilled large language model based on [Qwen 2.5 32B](https://huggingface.co/Qwen/Qwen2.5-32B), using outputs from [DeepSeek R1](/deepseek/deepseek-r1). It outperforms OpenAI's o1-mini across various benchmarks, achieving new state-of-the-art results for dense models.\n\nOther benchmark results include:\n\n- AIME 2024 pass@1: 72.6\n- MATH-500 pass@1: 94.3\n- CodeForces Rating: 1691\n\nThe model leverages fine-tuning from DeepSeek R1's outputs, enabling competitive performance comparable to larger frontier models.
Benchmarks
Available Providers (5)
| Provider | Model ID | Input Cost | Output Cost | Context | Max Output | Docs |
|---|---|---|---|---|---|---|
| | deepseek-r1-distill-qwen-32b | $0.29/MTok | $0.86/MTok | 32.8K | 16.4K | |
| | deepseek/deepseek-r1-distill-qwen-32b | $0.29/MTok | $0.29/MTok | 32.8K | 32.8K | |
| | DeepSeek-R1-Distill-Qwen-32B | $0.30/MTok | $0.30/MTok | 130K | 4.1K | |
| | workers-ai/@cf/deepseek-ai/deepseek-r1-distill-qwen-32b | $0.50/MTok | $4.88/MTok | 128K | 16.4K | |
| | @cf/deepseek-ai/deepseek-r1-distill-qwen-32b | $0.50/MTok | $4.88/MTok | 128K | 16.4K |
Capabilities
Reasoning
Tool Calling
Attachments
Open Weights
Structured Output