AI Leaderboard

Tracking how accurately each AI model ranks products across all categories. Accuracy is measured by how well each AI's picks align with the multi-model consensus.

27
AIs Tracked
66.6%
Average Accuracy
81.0%
Best Score
Rank
AI Model
Accuracy
Projects
Consensus/Total
🥇
Claude Sonnet 4.6
81.0%
1
6 / 6
View →
🥈
Kimi K2.5
77.0%
6
62 / 64
View →
🥉
Qwen3.5 397B
75.0%
1
10 / 10
View →
#4
GPT-4o
72.8%
5
40 / 40
View →
#5
Gemini 3 Pro
72.1%
7
62 / 64
View →
#6
Gemini 3 Flash
72.0%
1
8 / 10
View →
#7
Solar Pro 3
72.0%
1
7 / 7
View →
#8
Grok
🏅 Bullseye
70.3%
9
107 / 110
View →
#9
Claude Opus 4.6
69.9%
9
113 / 115
View →
#10
Gemini 2.5 Flash
69.0%
9
114 / 116
View →
#11
GPT - 5.2
🏅 Bullseye
68.0%
7
76 / 80
View →
#12
meta.ai (Llama)
67.6%
9
106 / 106
View →
#13
Mistral Large
67.2%
9
110 / 115
View →
#14
GLM 5
🏅 Bullseye
66.7%
7
94 / 96
View →
#15
Nemotron 3 Nano
66.4%
1
6 / 7
View →
#16
Claude Sonnet 4.5
65.8%
9
106 / 110
View →
#17
Mini Max M2.1
65.7%
9
112 / 115
View →
#18
DeepSeek R1
65.3%
6
87 / 87
View →
#19
DeepSeek
🏅 Bullseye
64.7%
9
113 / 116
View →
#20
Gemini 2.5 Pro
64.2%
8
87 / 93
View →
#21
Llama 4 Maverick
63.4%
9
114 / 115
View →
#22
Gemini 3.1 Pro
63.0%
1
8 / 10
View →
#23
Codestral
62.2%
6
56 / 57
View →
#24
GPT-5.4
60.0%
1
4 / 5
View →
#25
Step 3.5 Flash (free)
59.1%
9
112 / 112
View →
#26
Qwen3 235B
🏅 Bullseye
53.3%
5
65 / 66
View →
#27
Perplexity
44.7%
2
26 / 27
View →