Leaderboards

Greatness isn't accidental. How we measure it shouldn't be either. If we want AGI that builds billion-dollar enterprises and globe-spanning infrastructure, we can't evaluate it with clickbait and synthetic slop. We need benchmarks that test for intelligence and sophistication.

This is our definitive ranking of models, measured by their capacity for rigorous reasoning and real-world mastery. Discover which labs are leading the frontier.
View by :
Creative, Business, and Everyday writing

Hemingway-bench

Stop rewarding slop. We take real-world writing tasks and put them in front of master wordsmiths. Our goal: to push AI writing from two-second vibes to genuine nuance and impact.

Rank
Model
elo score (95% ci)
Gemini 3.1 Pro
1090
(
1070
-
1110
)
Gemini 3 Flash
1084
(
1066
-
1101
)
Gemini 3 Pro
1078
(
1056
-
1101
)
Opus 4.7
1067
(
1043
-
1091
)
Opus 4.6
1061
(
1041
-
1082
)
Opus 4.5
1043
(
1024
-
1062
)
Sonnet 4.6
1028
(
1008
-
1048
)
Kimi K2.5
1023
(
1005
-
1041
)
GPT-5.2 Chat
1022
(
1004
-
1041
)
GPT-5.4
1019
(
998
-
1041
)
Qwen3.5 Plus
1006
(
987
-
1025
)
GPT-5.2
986
(
963
-
1009
)
View full leaderboard
Enterprise Agents in Realistic RL Environments

EnterpriseBench: CoreCraft

Stop testing models in tiny, self-contained environments. We built CoreCraft, a large-scale startup world, and deployed AI agents to solve real tasks. Our goal: to move agents beyond the cleanliness of the lab and into the chaos of enterprise reality.

Rank
Model
Score
GPT-5.2 (xHigh reasoning)
GPT-5.2 (xHigh reasoning)
42.6
%
GPT-5.4 (xHigh reasoning)
GPT-5.4 (xHigh reasoning)
36.4
%
Claude Opus 4.7 (Max)
Claude Opus 4.7 (Max)
35.9
%
View full leaderboard
Mathematics at the frontier

Riemann-bench

We evaluate AI models on advanced mathematical problems requiring deep reasoning and novel synthesis. Our benchmark features problems from cutting-edge mathematics, sourced from leading mathematicians – Ivy League professors, PhD IMO medalists, graduate students at the top of their field – in the course of their research.

Rank
Model
Score
Gemini 3.1 Pro
Gemini 3.1 Pro
6
%
Claude Opus 4.6
Claude Opus 4.6
6
%
Gemini 3 Pro
Gemini 3 Pro
4
%
Kimi K2.5
Kimi K2.5
4
%
DeepSeek v3.2
DeepSeek v3.2
3
%
Claude Opus 4.5
Claude Opus 4.5
2
%
GPT-5.2
GPT-5.2
2
%
View full leaderboard
Your $100B model can't read a PDF

GDP.pdf

Can frontier models master the documents that run the world? GDP.pdf is a multimodal and reasoning benchmark that takes real-world prompts and PDFs pulled directly from expert professional workflows.

Rank
Model
Score
Gemini 3.1 Pro
Gemini 3.1 Pro
15
%
Claude Opus 4.7
Claude Opus 4.7
14
%
GPT-5.4
GPT-5.4
11
%
Claude Opus 4.6
Claude Opus 4.6
11
%
Grok-4.20 Beta
Grok-4.20 Beta
7
%
Kimi K2.5
Kimi K2.5
6
%
Mistral Large 3
Mistral Large 3
3
%
Nova 2 Pro
Nova 2 Pro
1
%
View full leaderboard

Stay up-to-date on new leaderboards