DeepSeek-V2-Lite
DeepSeek-V2-Lite is a Mixture-of-Experts language model designed for economical training and efficient inference. With 16B total parameters and 2.4B activated parameters, it employs innovative techniques such as Multi-head Latent Attention (MLA) and DeepSeekMoE for performance gains. The model is available for both text and chat completions via Hugging Face and is optimized to run with a 40GB GPU using BF16 precision.
Key Information
- Category: Language Models
- Source: Huggingface
- Tags: text-generation
- Last updated: March 12, 2026
Structured Metrics
No structured metrics captured yet.
Links
Canonical source: https://huggingface.co/deepseek-ai/DeepSeek-V2-Lite