Best coding model of 2024, dominated SWE-bench for months
| Benchmark | Score | Rank |
|---|---|---|
HumanEval Coding ability - generating correct Python functions | 93.7% | #16 / 49 |
ARC-C Grade-school science questions requiring reasoning | 96.7% | #23 / 40 |
MMLU Tests knowledge across 57 subjects from STEM to humanities | 88.7% | #27 / 53 |
HellaSwag Common sense reasoning about everyday situations | 89% | #27 / 36 |
MMMUvals.ai College-level multimodal reasoning across 30+ disciplines | 68.8% | #27 / 33 |
MMLU-Provals.ai Harder 10-option successor to MMLU; more reasoning-focused | 78.4% | #28 / 30 |
LiveCodeBenchvals.ai Contamination-free competitive programming (filtered by cutoff date) | 49.6% | #28 / 31 |
MATH Competition-level mathematics problems | 78.3% | #32 / 49 |
Arena Elo Human preference ranking via blind comparisons | 1280 | #34 / 41 |
SWE-bench Real-world GitHub issue resolution | 49% | #35 / 38 |
GPQA PhD-level science questions even experts struggle with | 65% | #40 / 54 |