Open WeightMeta
Llama 4 Scout
Context
10M tokens
Modalities
text, image, code
Released
Apr 2025
- Overview
- Meta's 109B parameter MoE model featuring an unprecedented 10M token context window. Llama 4 Scout is optimized for processing extremely long documents and codebases while maintaining strong general capability.
- Why it matters
- The 10M token context window is not a typo — Scout can process roughly 30 average-length novels or an entire enterprise codebase in a single call. This is a qualitatively different capability that enables use cases previously impossible: full-repository code understanding, entire-book analysis, or processing months of conversation history. For teams building code intelligence tools, legal document analysis, or long-horizon agentic systems, Scout eliminates the need for complex chunking and retrieval pipelines. The trade-off is that inference at these context lengths requires significant compute.
Key strengths
- 10M token context window — largest of any model
- 109B MoE architecture for efficient inference
- Strong general capability despite long-context focus
- Open weights for self-hosting
We cover ai models every week.
Get the 5 AI stories that matter — free, every Friday.
Know the terms. Know the moves.
Get the 5 AI stories that matter every Friday — free.
Free forever. No spam.