DeepSeek-V2-Lite

DeepSeek-V2-Lite is a Mixture-of-Experts language model designed for economical training and efficient inference. With 16B total parameters and 2.4B activated parameters, it employs innovative techniques such as Multi-head Latent Attention (MLA) and DeepSeekMoE for performance gains. The model is available for both text and chat completions via Hugging Face and is optimized to run with a 40GB GPU using BF16 precision.

Key Information

  • Category: Language Models
  • Source: Huggingface
  • Tags: text-generation
  • Last updated: March 12, 2026

Structured Metrics

No structured metrics captured yet.

Links

Canonical source: https://huggingface.co/deepseek-ai/DeepSeek-V2-Lite