
Llama 4 Scout on LiveCodeBench: 33.2% accuracy
Author:
The LayerLens Team
Last updated:
Published:
The LayerLens Team covers AI model evaluations, benchmark analysis, and the evolving landscape of AI performance. For the latest independent evaluation data, explore Stratix.
Summary
Llama 4 Scout from Meta scored 33.2 on LiveCodeBench, placing it top 50 (rank 31 of 43) on this benchmark. This places the model in the weak band for LiveCodeBench. Below the threshold for production reliance on this benchmark family. Consider only for narrow, fully-tested tasks.
Model details
Provider: Meta
Model key:
meta-llama/llama-4-scoutContext length: 172,000 tokens
License: Llama 4
Open weights: yes
Benchmark methodology
Secondary metrics
Readability score: 0.0
Toxicity score: 0.000
Ethics score: 0.000
Run this evaluation yourself
Stratix evaluates Llama 4 Scout continuously across 11+ benchmarks. To replicate this LiveCodeBench evaluation on your own model, traces, or a different benchmark configuration, open the model in Stratix.
Source: Stratix evaluation 68fa0477f82ed9ed12809c18. Updated 2025-10-23.