DeepSeek-R1-Distill-Qwen-1.5B

A distilled dense language model based on Qwen2.5-Math-1.5B that leverages the DeepSeek-R1 pipeline. It is designed for advanced reasoning, math, and code generation tasks, and is available under an MIT license with extensive evaluation metrics and deployment instructions on Hugging Face.

Key Information

  • Category: Language Models
  • Source: Huggingface
  • Tags: text-generation
  • Last updated: March 09, 2026

Structured Metrics

No structured metrics captured yet.

Links

Canonical source: https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B