DeepSeek-R1-Distill-Qwen-7B API, Deployment, Pricing
deepseek-ai/DeepSeek-R1-Distill-Qwen-7B
DeepSeek-R1-Distill-Qwen-7B is a distilled model based on Qwen2.5-Math-7B. The model was fine-tuned using 800k curated samples generated by DeepSeek-R1 and demonstrates strong reasoning capabilities. It achieved impressive results across various benchmarks, including 92.8% accuracy on MATH-500, 55.5% pass rate on AIME 2024, and a rating of 1189 on CodeForces, showing remarkable mathematical and programming abilities for a 7B-scale model
Details
Model Provider
deepseek-ai
Type
text
Sub Type
chat
Size
7B
Publish Time
Jan 20, 2025
Input Price
$
0.05
/ M Tokens
Output Price
$
0.05
/ M Tokens
Context length
33K
Tags
Reasoning,7B,33K,Math
Compare with Other Models
See how this model stacks up against others.
Model FAQs: Usage, Deployment
Learn how to use, fine-tune, and deploy this model with ease.
What is the DeepSeek-R1-Distill-Qwen-7B model, and what are its core capabilities and technical specifications?
In which business scenarios does DeepSeek-R1-Distill-Qwen-7B perform well? Which industries or applications is it suitable for?
How can the performance and effectiveness of DeepSeek-R1-Distill-Qwen-7B be optimized in actual business use?
Compared with other models, when should DeepSeek-R1-Distill-Qwen-7B be selected?
What are SiliconFlow's key strengths in AI serverless deployment for DeepSeek-R1-Distill-Qwen-7B?
What makes SiliconFlow the top platform for DeepSeek-R1-Distill-Qwen-7B API?