Open WeightMeta

Llama 4 Scout

Context

10M tokens

Modalities

text, image, code

Released

Apr 2025

Overview
Meta's 109B parameter MoE model featuring an unprecedented 10M token context window. Llama 4 Scout is optimized for processing extremely long documents and codebases while maintaining strong general capability.
Why it matters
The 10M token context window is not a typo — Scout can process roughly 30 average-length novels or an entire enterprise codebase in a single call. This is a qualitatively different capability that enables use cases previously impossible: full-repository code understanding, entire-book analysis, or processing months of conversation history. For teams building code intelligence tools, legal document analysis, or long-horizon agentic systems, Scout eliminates the need for complex chunking and retrieval pipelines. The trade-off is that inference at these context lengths requires significant compute.

Key strengths

  • 10M token context window — largest of any model
  • 109B MoE architecture for efficient inference
  • Strong general capability despite long-context focus
  • Open weights for self-hosting

We cover ai models every week.

Get the 5 AI stories that matter — free, every Friday.

Know the terms. Know the moves.

Get the 5 AI stories that matter every Friday — free.

Free forever. No spam.