First reasoning model, uses chain-of-thought at inference time to solve hard problems
| Benchmark | Score | Rank |
|---|---|---|
ARC-C Grade-school science questions requiring reasoning | 97.8% | #11 / 40 |
MMLU Tests knowledge across 57 subjects from STEM to humanities | 91.8% | #12 / 53 |
MATH Competition-level mathematics problems | 96.4% | #12 / 49 |
ARC-AGIARC Prize Novel reasoning tasks requiring fluid intelligence | 0.8% | #21 / 21 |
HumanEval Coding ability - generating correct Python functions | 92.4% | #25 / 49 |
Arena Elo Human preference ranking via blind comparisons | 1350 | #27 / 41 |
GPQA PhD-level science questions even experts struggle with | 78% | #30 / 54 |
SWE-bench Real-world GitHub issue resolution | 48.9% | #36 / 38 |