Codex Red Herring (False Positive Detection)
Tests whether models correctly report "no violations" when a codex is fully consistent with the prose passage. Models that hallucinate false violations (false positives) fail. Uses a 2×2 matrix of text length × codex size, with bare and detailed-entry variants.
Short text (~524 words), big codex (51 entries)
Hallucination
Performance Score Distribution (Top 20)
Click a model name to view its detail page.
Price-Performance Score Distribution (Top 20)
Click a model name to view its detail page.
| Score | Cost | Time | ||
|---|---|---|---|---|
| Ministral 8B | 68% | $0.0003 | 3.4s | |
| Ministral 3 8B | 93% | $0.0004 | 1.2s | |
| GPT-4.1 | 100% | $0.0020 | 737ms | |
| Hermes 3 405B | 75% | $0.0026 | 1.7s | |
| ByteDance Seed 1.6 Flash | 100% | $0.0005 | 6.7s | |
| Gemini 2.5 Flash Lite (Reasoning) | 100% | $0.0011 | 6.6s | |
| Grok 4.1 Fast | 100% | $0.0012 | 9.6s | |
| Arcee AI: Trinity Mini | 83% | $0.0003 | 15.8s | |
| Cohere Command R+ (Aug. 2024) | 65% | $0.0080 | 3.2s | |
| Gemini 2.5 Flash (Reasoning) | 93% | $0.0047 | 7.7s | |
| GPT-5.2 | 75% | $0.0087 | 9.4s | |
| Claude Sonnet 4 | 75% | $0.0096 | 2.1s | |
| ByteDance Seed 1.6 | 100% | $0.0021 | 16.2s | |
| Minimax M2.5 | 70% | $0.0016 | 17.5s | |
| GPT-5.1 | 93% | $0.014 | 17.8s | |
| GPT-5 Mini | 100% | $0.0044 | 28.8s | |
| o4 Mini | 100% | $0.010 | 20.6s | |
| Aion 2.0 | 93% | $0.0053 | 49.2s | |
| Claude Opus 4.6 | 100% | $0.031 | 12.1s | |
| o4 Mini High | 100% | $0.020 | 41.1s | |
Most Stable Models (Top 20)
Ranked by stability (median × consistency). Click a model name to view its detail page.
| Score | Consistency | Stability | ||
|---|---|---|---|---|
| Claude Opus 4.6 (Reasoning) | 100% | 100% | 100% | |
| Gemini 3.1 Pro (Preview) | 100% | 100% | 100% | |
| Claude Sonnet 4.6 (Reasoning) | 100% | 100% | 100% | |
| GPT-5 Mini | 100% | 100% | 100% | |
| Claude Opus 4.6 | 100% | 100% | 100% | |
| ByteDance Seed 1.6 | 100% | 100% | 100% | |
| o4 Mini High | 100% | 100% | 100% | |
| Grok 4.1 Fast | 100% | 100% | 100% | |
| GPT-4.1 | 100% | 100% | 100% | |
| o4 Mini | 100% | 100% | 100% | |
| Gemini 2.5 Flash Lite (Reasoning) | 100% | 100% | 100% | |
| ByteDance Seed 1.6 Flash | 100% | 100% | 100% | |
| Ministral 3 8B | 93% | 56% | 56% | |
| GPT-5.1 | 93% | 55% | 55% | |
| Aion 2.0 | 93% | 55% | 55% | |
| Gemini 2.5 Flash (Reasoning) | 93% | 55% | 55% | |
| Gemini 2.5 Pro | 85% | 40% | 40% | |
| Z.AI GLM 4.7 Flash | 85% | 40% | 40% | |
| Claude Sonnet 4 | 75% | 37% | 37% | |
| Claude Haiku 4.5 | 54% | 69% | 35% | |
Top Overall Models (Top 20)
Ranked by composite score (performance, cost, speed & stability). Click a model name to view its detail page.
| Score | Cost | Speed | Stability | ||
|---|---|---|---|---|---|
| GPT-4.1 | 100% | $0.0020 | 737ms | 100% | |
| ByteDance Seed 1.6 Flash | 100% | $0.0005 | 6.7s | 100% | |
| Gemini 2.5 Flash Lite (Reasoning) | 100% | $0.0011 | 6.6s | 100% | |
| Grok 4.1 Fast | 100% | $0.0012 | 9.6s | 100% | |
| ByteDance Seed 1.6 | 100% | $0.0021 | 16.2s | 100% | |
| GPT-5 Mini | 100% | $0.0044 | 28.8s | 100% | |
| o4 Mini | 100% | $0.010 | 20.6s | 100% | |
| Claude Opus 4.6 | 100% | $0.031 | 12.1s | 100% | |
| o4 Mini High | 100% | $0.020 | 41.1s | 100% | |
| Claude Opus 4.6 (Reasoning) | 100% | $0.041 | 24.0s | 100% | |
| Gemini 3.1 Pro (Preview) | 100% | $0.045 | 32.9s | 100% | |
| Claude Sonnet 4.6 (Reasoning) | 100% | $0.044 | 35.6s | 100% | |
| Ministral 3 8B | 93% | $0.0004 | 1.2s | 56% | |
| Gemini 2.5 Flash (Reasoning) | 93% | $0.0047 | 7.7s | 55% | |
| GPT-5.1 | 93% | $0.014 | 17.8s | 55% | |
| Aion 2.0 | 93% | $0.0053 | 49.2s | 55% | |
| Arcee AI: Trinity Mini | 83% | $0.0003 | 15.8s | 31% | |
| Claude Sonnet 4 | 75% | $0.0096 | 2.1s | 37% | |
| Hermes 3 405B | 75% | $0.0026 | 1.7s | 24% | |
| GPT-5.2 | 75% | $0.0087 | 9.4s | 24% | |
| Median | Evaluator | Top 3 | Flop 3 |
|---|---|---|---|
| 40.0% | Correct "no violations" response | ||
| 50.8% | No hallucinated violations |