Improved agentic reliability with better tool use and planning
| Benchmark | Score | Rank |
|---|---|---|
MMLU-Provals.ai Harder 10-option successor to MMLU; more reasoning-focused | 87.9% | #6 / 30 |
HumanEval Coding ability - generating correct Python functions | 94.5% | #13 / 49 |
ARC-C Grade-school science questions requiring reasoning | 97.6% | #13 / 40 |
SWE-bench Real-world GitHub issue resolution | 74.5% | #17 / 38 |
MMMUvals.ai College-level multimodal reasoning across 30+ disciplines | 77.5% | #17 / 33 |
HellaSwag Common sense reasoning about everyday situations | 92.8% | #19 / 36 |
MMLU Tests knowledge across 57 subjects from STEM to humanities | 89.5% | #21 / 53 |
TerminalArtificial Analysis Agentic terminal coding tasks requiring multi-step execution | 34.3% | #21 / 37 |
Arena Elo Human preference ranking via blind comparisons | 1372 | #22 / 41 |
MATH Competition-level mathematics problems | 88.4% | #25 / 49 |
LiveCodeBenchvals.ai Contamination-free competitive programming (filtered by cutoff date) | 66.5% | #25 / 31 |
GPQA PhD-level science questions even experts struggle with | 76.2% | #32 / 54 |