First hybrid reasoning model, extended thinking mode for complex problems
| Benchmark | Score | Rank |
|---|---|---|
MATH Competition-level mathematics problems | 96.2% | #13 / 49 |
HumanEval Coding ability - generating correct Python functions | 93.7% | #17 / 49 |
MMMUvals.ai College-level multimodal reasoning across 30+ disciplines | 75.1% | #20 / 33 |
MMLU-Provals.ai Harder 10-option successor to MMLU; more reasoning-focused | 82.7% | #22 / 30 |
ARC-C Grade-school science questions requiring reasoning | 96.7% | #24 / 40 |
LiveCodeBenchvals.ai Contamination-free competitive programming (filtered by cutoff date) | 60.4% | #27 / 31 |
HellaSwag Common sense reasoning about everyday situations | 89% | #28 / 36 |
SWE-bench Real-world GitHub issue resolution | 62.3% | #28 / 38 |
TerminalArtificial Analysis Agentic terminal coding tasks requiring multi-step execution | 21.2% | #28 / 37 |
GPQA PhD-level science questions even experts struggle with | 78.2% | #29 / 54 |
Arena Elo Human preference ranking via blind comparisons | 1310 | #31 / 41 |
MMLU Tests knowledge across 57 subjects from STEM to humanities | 86.1% | #40 / 53 |