Generation

Subcategory of Language. 118 models scored.

Model Leaderboard

All models ranked by their Generation subcategory score.

# Model Generation Language Overall
1 Claude Sonnet 4.6 100.00% 100.00% 91.15%
2 o4 Mini 100.00% 80.00% 88.35%
3 Gemini 3 Flash (Preview) 100.00% 95.00% 85.35%
4 DeepSeek-V2 Chat 100.00% 100.00% 84.83%
5 Stealth: Aurora Alpha 100.00% 92.50% 83.79%
6 GPT-4o, Aug. 6th (temp=0) 100.00% 75.00% 82.45%
7 GPT-4o Mini (temp=1) 100.00% 77.50% 79.08%
8 GPT-4o Mini (temp=0) 100.00% 75.00% 78.29%
9 GPT-5.4 Mini (Reasoning, Low) 99.89% 92.45% 85.75%
10 Z.AI GLM 5 Turbo 99.80% 99.90% 94.27%
11 Z.AI GLM 4.5 99.67% 97.33% 86.27%
12 Inception Mercury 2 99.64% 87.32% 83.85%
13 o4 Mini High 99.51% 79.76% 90.29%
14 GPT-4o, Aug. 6th (temp=1) 99.43% 82.21% 82.62%
15 GPT-5 Nano 99.35% 77.18% 82.60%
16 Claude Opus 4.5 99.31% 99.66% 89.69%
17 GPT-4.1 Mini 99.27% 89.64% 83.20%
18 Hermes 3 405B 99.14% 99.57% 82.86%
19 DeepSeek V3.1 98.74% 96.87% 82.39%
20 Grok 4.20 (Beta, Reasoning) 98.15% 99.08% 91.49%
21 GPT-5 98.01% 91.50% 91.93%
22 GPT-5.4 Nano (Reasoning) 97.99% 83.99% 81.36%
23 GPT-5.4 97.98% 81.49% 84.32%
24 GPT-5 Mini 97.98% 96.49% 92.62%
25 GPT-4.1 97.82% 93.91% 88.68%
26 Nemotron 3 Super 97.82% 81.41% 84.56%
27 GPT-5.4 Mini 97.50% 88.75% 82.43%
28 Gemini 2.5 Flash 97.46% 86.23% 80.60%
29 GPT-4o, May 13th (temp=0) 97.44% 98.72% 85.36%
30 GPT-5.2 97.39% 91.19% 90.26%
31 Grok 4.20 (Beta) 97.34% 91.17% 83.85%
32 GPT-5.1 97.28% 93.64% 92.54%
33 Gemini 2.5 Flash (Reasoning) 97.13% 86.06% 86.51%
34 MiniMax M2.5 97.10% 96.05% 88.71%
35 GPT-5.4 (Reasoning, Low) 96.58% 90.79% 91.41%
36 GPT-5.4 Mini (Reasoning) 96.25% 98.12% 90.65%
37 Claude 3.5 Sonnet 96.24% 85.62% 84.24%
38 Grok 4 96.22% 90.61% 88.12%
39 Claude 3.7 Sonnet 95.89% 92.95% 83.39%
40 Inception Mercury 95.74% 80.37% 79.50%
41 Nemotron 3 Nano 95.26% 87.63% 77.73%
42 Claude Sonnet 4.6 (Reasoning) 95.15% 97.58% 93.66%
43 Gemini 2.5 Pro 95.14% 92.57% 88.53%
44 Gemini 3.1 Flash Lite (Preview) 94.95% 94.98% 85.87%
45 Gemini 3 Flash (Preview, Reasoning) 94.87% 94.93% 90.50%
46 GPT-5.4 (Reasoning) 94.80% 94.90% 93.24%
47 Gemini 3.1 Pro (Preview) 94.80% 94.90% 94.37%
48 MoonshotAI: Kimi K2.5 94.19% 97.10% 91.04%
49 GPT-5.4 Nano (Reasoning, Low) 93.74% 81.87% 79.48%
50 ByteDance Seed 2.0 Lite 93.60% 96.80% 84.80%
51 Z.AI GLM 4.6 93.20% 96.60% 89.11%
52 GPT-4.1 Nano 92.91% 78.95% 71.94%
53 Claude Sonnet 4 92.61% 91.31% 88.72%
54 Aion 2.0 92.34% 96.17% 89.21%
55 Claude Opus 4.6 92.27% 96.13% 92.35%
56 Claude Opus 4.6 (Reasoning) 92.23% 96.12% 95.02%
57 Stealth: Hunter Alpha 91.70% 93.35% 87.34%
58 GPT-5.4 Nano 91.64% 80.82% 74.40%
59 ByteDance Seed 1.6 91.26% 95.63% 90.70%
60 Qwen 3.5 27B 91.05% 95.52% 90.85%
61 Z.AI GLM 4.7 90.93% 85.46% 88.69%
62 ByteDance Seed 2.0 Mini 90.24% 90.12% 86.91%
63 Qwen 3.5 Plus (2026-02-15) 90.21% 95.10% 85.96%
64 Gemma 3 12B 90.19% 80.10% 78.41%
65 GPT-4o, May 13th (temp=1) 90.04% 92.52% 83.80%
66 Qwen 3.5 122B 90.03% 95.01% 91.53%
67 Qwen 3.5 397B A17B 90.03% 95.01% 91.73%
68 DeepSeek V3.2 90.03% 85.01% 82.25%
69 Claude Sonnet 4.5 89.77% 92.39% 88.03%
70 Gemma 3 27B 89.42% 77.21% 77.85%
71 Z.AI GLM 5 89.12% 92.06% 91.23%
72 Qwen 3.5 Flash 88.89% 91.94% 86.38%
73 Claude Haiku 4.5 88.67% 91.84% 85.14%
74 Hermes 3 70B 88.33% 81.66% 72.57%
75 Grok 4.1 Fast 87.51% 88.76% 89.55%
76 Mistral Large 87.27% 88.64% 80.15%
77 Stealth: Healer Alpha 86.90% 88.45% 85.93%
78 Qwen 3.5 9B 86.36% 88.18% 86.05%
79 Claude Opus 4 86.02% 93.01% 87.69%
80 Gemini 2.5 Flash Lite 85.50% 82.75% 81.08%
81 Llama 3.1 70B 85.36% 80.18% 78.40%
82 Gemini 3 Pro (Preview) 84.29% 89.64% 88.79%
83 Grok 4 Fast 84.22% 84.61% 86.15%
84 Mistral Large 3 84.05% 92.02% 85.43%
85 Qwen 3.5 35B 83.90% 91.95% 88.00%
86 Gemini 2.5 Flash Lite (Reasoning) 83.71% 74.36% 85.75%
87 ByteDance Seed 1.6 Flash 82.45% 61.23% 73.27%
88 Arcee AI: Trinity Mini 81.18% 70.59% 70.90%
89 Claude 3 Haiku 80.53% 72.76% 71.19%
90 Z.AI GLM 4.7 Flash 80.35% 87.67% 84.82%
91 Qwen 3 32B 79.21% 84.61% 82.21%
92 Arcee AI: Trinity Large (Preview) 76.76% 78.38% 73.33%
93 DeepSeek V3 (2024-12-26) 75.76% 87.88% 83.68%
94 Gemma 3 4B 74.56% 72.28% 68.57%
95 Claude 3.5 Haiku 74.25% 82.12% 83.73%
96 Cohere Command R+ (Aug. 2024) 73.15% 66.58% 69.03%
97 Llama 3.1 8B 73.13% 64.06% 63.37%
98 DeepSeek V3 (2025-03-24) 72.83% 86.42% 81.99%
99 Mistral Small 4 (Reasoning) 71.06% 60.53% 82.39%
100 Mistral Small 3.2 24B 70.53% 72.77% 78.60%
101 Mistral Large 2 70.43% 85.22% 82.41%
102 MiniMax M2.7 69.59% 84.80% 89.10%
103 Qwen 2.5 72B 67.91% 68.95% 75.46%
104 Mistral NeMO 66.60% 80.80% 65.04%
105 LFM2 24B 64.28% 64.64% 58.77%
106 WizardLM 2 8x22b 61.10% 78.05% 71.07%
107 Ministral 3B 59.49% 42.25% 61.29%
108 Ministral 8B 52.81% 53.91% 64.87%
109 Rocinante 12B 51.90% 63.45% 54.55%
110 Mistral Medium 3.1 49.00% 49.50% 77.83%
111 Mistral Small 4 48.93% 51.96% 76.46%
112 Ministral 3 8B 47.92% 48.96% 71.76%
113 Qwen3 235B A22B Instruct 2507 46.67% 60.83% 80.10%
114 Writer: Palmyra X5 43.16% 56.58% 79.57%
115 Ministral 3 3B 36.20% 68.10% 67.22%
116 Mistral Small Creative 33.69% 41.85% 73.27%
117 Llama 3.1 Nemotron 70B 33.61% 46.80% 74.70%
118 Ministral 3 14B 10.00% 30.00% 72.54%