DeepSeek-R1-Distill-Qwen-1.5B
A distilled dense language model based on Qwen2.5-Math-1.5B that leverages the DeepSeek-R1 pipeline. It is designed for advanced reasoning, math, and code generation tasks, and is available under an MIT license with extensive evaluation metrics and deployment instructions on Hugging Face.
Key Information
- Category: Language Models
- Source: Huggingface
- Tags: text-generation
- Last updated: March 09, 2026
Structured Metrics
No structured metrics captured yet.
Links
Canonical source: https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B