Skip to content

Llama 4

by Meta · Website

Meta's Llama 4 introduces mixture-of-experts architecture and native multimodal support to the Llama family. Scout (109B total, 17B active) features a 10M token context window, while Maverick (400B total, 17B active) targets large-scale deployments. Both models support text and image inputs natively.

Variants (2)

Smallest: Llama 4 Scout (109B/17B active) (109B)
Largest: Llama 4 Maverick (400B)

Llama 4 Scout (109B/17B active)

109B

Meta

Min 72 GB
text-generation code-generation reasoning

Llama 4 Maverick

400B

Meta

Min 228 GB
text-generation code-generation reasoning