Paragraph Counting

Subcategory of Utility. 91 models scored.

Model Leaderboard

All models ranked by their Paragraph Counting subcategory score.

# Model Paragraph Counting Utility Overall
1 Claude Opus 4.6 (Reasoning) 100.00% 98.93% 95.02%
2 Gemini 3.1 Pro (Preview) 100.00% 99.91% 94.37%
3 Claude Sonnet 4.6 (Reasoning) 100.00% 97.88% 93.66%
4 GPT-5 Mini 100.00% 98.39% 92.62%
5 GPT-5.1 100.00% 95.33% 92.54%
6 Claude Opus 4.6 100.00% 90.72% 92.35%
7 GPT-5 100.00% 93.53% 91.93%
8 Qwen 3.5 397B A17B 100.00% 97.50% 91.73%
9 Qwen 3.5 122B 100.00% 96.36% 91.53%
10 Z.AI GLM 5 100.00% 94.11% 91.23%
11 Claude Sonnet 4.6 100.00% 88.52% 91.15%
12 MoonshotAI: Kimi K2.5 100.00% 96.63% 91.04%
13 Qwen 3.5 27B 100.00% 95.67% 90.85%
14 ByteDance Seed 1.6 100.00% 90.83% 90.70%
15 Gemini 3 Flash (Preview, Reasoning) 100.00% 97.20% 90.50%
16 o4 Mini High 100.00% 98.67% 90.29%
17 GPT-5.2 100.00% 96.22% 90.26%
18 Claude Opus 4.5 100.00% 89.84% 89.69%
19 Grok 4.1 Fast 100.00% 84.12% 89.55%
20 Aion 2.0 100.00% 90.91% 89.21%
21 Z.AI GLM 4.6 100.00% 88.58% 89.11%
22 Gemini 3 Pro (Preview) 100.00% 96.14% 88.79%
23 Claude Sonnet 4 100.00% 84.02% 88.72%
24 Minimax M2.5 100.00% 90.42% 88.71%
25 Z.AI GLM 4.7 100.00% 94.31% 88.69%
26 GPT-4.1 100.00% 90.57% 88.68%
27 Gemini 2.5 Pro 100.00% 92.18% 88.53%
28 o4 Mini 100.00% 96.31% 88.35%
29 Grok 4 100.00% 89.67% 88.12%
30 Claude Sonnet 4.5 100.00% 83.78% 88.03%
31 Qwen 3.5 35B 100.00% 96.42% 88.00%
32 Claude Opus 4 100.00% 88.81% 87.69%
33 Gemini 2.5 Flash (Reasoning) 100.00% 82.25% 86.51%
34 Qwen 3.5 Flash 100.00% 96.11% 86.38%
35 Z.AI GLM 4.5 100.00% 79.19% 86.27%
36 Grok 4 Fast 100.00% 76.76% 86.15%
37 Qwen 3.5 Plus (2026-02-15) 100.00% 86.65% 85.96%
38 Gemini 2.5 Flash Lite (Reasoning) 100.00% 89.63% 85.75%
39 Mistral Large 3 100.00% 84.91% 85.43%
40 GPT-4o, May 13th (temp=0) 100.00% 83.13% 85.36%
41 Gemini 3 Flash (Preview) 100.00% 86.39% 85.35%
42 DeepSeek-V2 Chat 100.00% 83.82% 84.83%
43 Z.AI GLM 4.7 Flash 100.00% 88.98% 84.82%
44 Claude 3.5 Sonnet 100.00% 76.75% 84.24%
45 GPT-4o, May 13th (temp=1) 100.00% 80.69% 83.80%
46 Stealth: Aurora Alpha 100.00% 92.59% 83.79%
47 Claude 3.5 Haiku 100.00% 82.57% 83.73%
48 DeepSeek V3 (2024-12-26) 100.00% 81.87% 83.68%
49 GPT-4.1 Mini 100.00% 82.30% 83.20%
50 GPT-4o, Aug. 6th (temp=1) 100.00% 82.44% 82.62%
51 GPT-5 Nano 100.00% 93.91% 82.60%
52 GPT-4o, Aug. 6th (temp=0) 100.00% 82.11% 82.45%
53 Mistral Large 2 100.00% 69.19% 82.41%
54 DeepSeek V3.2 100.00% 81.58% 82.25%
55 Gemini 2.5 Flash Lite 100.00% 80.14% 81.08%
56 Gemini 2.5 Flash 100.00% 61.45% 80.60%
57 Mistral Large 100.00% 73.04% 80.15%
58 Writer: Palmyra X5 100.00% 79.71% 79.57%
59 GPT-4o Mini (temp=1) 100.00% 82.16% 79.08%
60 Mistral Small 3.2 24B 100.00% 73.17% 78.60%
61 Gemma 3 12B 100.00% 79.28% 78.41%
62 Llama 3.1 70B 100.00% 81.03% 78.40%
63 GPT-4o Mini (temp=0) 100.00% 81.43% 78.29%
64 Gemma 3 27B 100.00% 76.82% 77.85%
65 Mistral Medium 3.1 100.00% 80.13% 77.83%
66 Qwen 2.5 72B 100.00% 76.43% 75.46%
67 Llama 3.1 Nemotron 70B 100.00% 88.31% 74.70%
68 ByteDance Seed 1.6 Flash 100.00% 84.16% 73.27%
69 Mistral Small Creative 100.00% 76.28% 73.27%
70 Ministral 3 14B 100.00% 79.03% 72.54%
71 GPT-4.1 Nano 100.00% 68.45% 71.94%
72 Ministral 3 8B 100.00% 74.43% 71.76%
73 Ministral 3 3B 100.00% 72.38% 67.22%
74 LFM2 24B 100.00% 69.48% 58.77%
75 Claude 3 Haiku 96.67% 68.47% 71.19%
76 DeepSeek V3.1 93.33% 76.65% 82.39%
77 DeepSeek V3 (2025-03-24) 93.33% 80.62% 81.99%
78 WizardLM 2 8x22b 90.00% 67.14% 71.07%
79 Cohere Command R+ (Aug. 2024) 90.00% 59.51% 69.03%
80 Llama 3.1 8B 90.00% 74.82% 63.37%
81 Hermes 3 405B 86.67% 69.02% 82.86%
82 Arcee AI: Trinity Mini 80.00% 59.94% 70.90%
83 Arcee AI: Trinity Large (Preview) 73.33% 60.74% 73.33%
84 Claude Haiku 4.5 70.00% 72.48% 85.14%
85 Hermes 3 70B 66.67% 61.15% 72.57%
86 Gemma 3 4B 53.33% 60.30% 68.57%
87 Mistral NeMO 40.00% 51.55% 65.04%
88 Rocinante 12B 36.67% 48.47% 54.55%
89 Claude 3.7 Sonnet 33.33% 62.54% 83.39%
90 Ministral 8B 33.33% 46.82% 64.87%
91 Ministral 3B 33.33% 49.17% 61.29%