R1 Distill Qwen 32B (free) Check detailed information and pricing for AI models
Context Length 16,000 tokens, deepseek from provided
16,000
Context Tokens
Free
Prompt Price
Free
Output Price
1/16
Feature Support
Model Overview
DeepSeek R1 Distill Qwen 32B is a distilled large language model based on [Qwen 2.5 32B](https://huggingface.co/Qwen/Qwen2.5-32B), using outputs from [DeepSeek R1](/deepseek/deepseek-r1). It outperforms OpenAI's o1-mini across various benchmarks, achieving new state-of-the-art results for dense models.\n\nOther benchmark results include:\n\n- AIME 2024 pass@1: 72.6\n- MATH-500 pass@1: 94.3\n- CodeForces Rating: 1691\n\nThe model leverages fine-tuning from DeepSeek R1's outputs, enabling competitive performance comparable to larger frontier models.
Basic Information
Developer
deepseek
Model Series
Qwen
Release Date
2025-01-29
Context Length
16,000 tokens
Max Completion Tokens
16,000 tokens
Variant
free
Pricing Information
This model is free to use
Data Policy
Supported Features
Supported (1)
Reasoning
Unsupported (15)
Image Input
Top K
Seed
Frequency Penalty
Presence Penalty
Repetition Penalty
Response Format
Min P
Logit Bias
Tool Usage
Logprobs
Top Logprobs
Structured Outputs
Web Search Options
Top A