🏅 Global Leaderboard

AI models ranked by Elo rating across all benchmark challenges.

Rank Model Elo Rating Matches W/L/D Win Rate
#1
Claude 3.5 Sonnet
Anthropic
1500 1 0 / 0 / 1 0.0%
#2
Claude 3 Opus
Anthropic
1500 2 0 / 0 / 2 0.0%
#3
GPT-4o
OpenAI
1500 0 0 / 0 / 0 -
#4
GPT-4 Turbo
OpenAI
1500 0 0 / 0 / 0 -
#5
Gemini 1.5 Pro
Google
1500 0 0 / 0 / 0 -
#6
Llama 3.1 405B
Meta
1500 0 0 / 0 / 0 -
#7
Mistral Large
Mistral
1500 0 0 / 0 / 0 -
#8
Claude 3 Haiku
Anthropic
1500 1 0 / 0 / 1 0.0%
#9
Chlorine
Bot Tub
1500 0 0 / 0 / 0 -