ModelsHub.benchmarksTitle
ModelsHub.benchmarksSubtitle
| # | Model | Provider | ArenaELO | |
|---|---|---|---|---|
| 1 | Claude Opus 4.6 | Anthropic | 1397 | ModelsHub.viewDetail |
| 2 | GPT-5.4 | OpenAI | 1380 | ModelsHub.viewDetail |
| 3 | o3 | OpenAI | 1370 | ModelsHub.viewDetail |
| 4 | Gemini 3.1 Pro | 1355 | ModelsHub.viewDetail | |
| 5 | DeepSeek R1 | DeepSeek | 1330 | ModelsHub.viewDetail |
| 6 | Claude Sonnet 4.6 | Anthropic | 1320 | ModelsHub.viewDetail |
| 7 | DeepSeek V3 | DeepSeek | 1320 | ModelsHub.viewDetail |
| 8 | o4-mini | OpenAI | 1310 | ModelsHub.viewDetail |
| 9 | Llama 4 Maverick | Meta | 1310 | ModelsHub.viewDetail |
| 10 | Gemini 2.5 Pro | 1300 | ModelsHub.viewDetail | |
| 11 | Grok 4 | xAI | 1290 | ModelsHub.viewDetail |
| 12 | GPT-4o | OpenAI | 1285 | ModelsHub.viewDetail |
| 13 | Qwen 3.5 | Alibaba | 1285 | ModelsHub.viewDetail |
| 14 | GPT-5.4 Mini | OpenAI | 1260 | ModelsHub.viewDetail |
| 15 | Mistral Large 2 | Mistral | 1260 | ModelsHub.viewDetail |