Llama4

Llama4 is a large autoregressive Mixture-of-Experts (MoE) multimodal model developed by Meta. It comes in two variants: Maverick (17B active parameters out of ~400B total with 128 experts) and Scout (17B active parameters out of ~109B total with 16 experts). The models support native multimodal inputs (text and images), long context lengths (up to 10 million tokens in some versions), and advanced quantization and offloading techniques for efficient deployment.

Key Information

  • Category: Language Models
  • Source: Huggingface
  • Last updated: February 24, 2026

Structured Metrics

No structured metrics captured yet.

Links

Canonical source: https://huggingface.co/docs/transformers/en/model_doc/llama4