Codex Extraction
Evaluates a model's ability to extract structured codex entries (characters, locations, objects, lore) from prose passages and return them as well-formed XML.
Short: The Rusty Lantern (Explicit)
ToolingReasoning
Performance Score Distribution (Top 20)
Click a model name to view its detail page.
Price-Performance Score Distribution (Top 20)
Click a model name to view its detail page.
| Score | Cost | Time | ||
|---|---|---|---|---|
| Mistral Small Creative | 97% | $0.0004 | 2.7s | |
| Ministral 3 8B | 98% | $0.0005 | 2.6s | |
| Gemini 3 Flash (Preview) | 98% | $0.0022 | 3.1s | |
| Mistral Medium 3.1 | 98% | $0.0021 | 6.5s | |
| Z.AI GLM 4.5 | 99% | $0.0013 | 8.1s | |
| Qwen 3.5 Plus (2026-02-15) | 99% | $0.0024 | 7.9s | |
| Gemini 2.5 Flash Lite | 92% | $0.0004 | 1.6s | |
| Gemini 2.5 Flash | 94% | $0.0018 | 2.0s | |
| Grok 4 Fast | 96% | $0.0010 | 7.4s | |
| Qwen 2.5 72B | 94% | $0.0007 | 8.5s | |
| DeepSeek V3 (2024-12-26) | 96% | $0.0016 | 11.9s | |
| Mistral Large 3 | 96% | $0.0022 | 6.3s | |
| Ministral 3 14B | 92% | $0.0007 | 4.8s | |
| Ministral 3B | 91% | $0.0001 | 1.4s | |
| Mistral Small 3.2 24B | 91% | $0.0004 | 3.4s | |
| DeepSeek-V2 Chat | 96% | $0.0016 | 11.4s | |
| Arcee AI: Trinity Large (Preview) | 95% | $0.0000 | 14.5s | |
| Ministral 8B | 92% | $0.0003 | 3.3s | |
| Grok 4.1 Fast | 99% | $0.0015 | 19.0s | |
| Hermes 3 405B | 99% | $0.0034 | 16.8s | |
Most Stable Models (Top 20)
Ranked by stability (median × consistency). Click a model name to view its detail page.
| Score | Consistency | Stability | ||
|---|---|---|---|---|
| Grok 4 | 99% | 100% | 99% | |
| Qwen 3.5 Plus (2026-02-15) | 99% | 100% | 99% | |
| Claude Sonnet 4.5 | 99% | 99% | 99% | |
| Claude Sonnet 4 | 99% | 100% | 99% | |
| GPT-5 | 99% | 100% | 99% | |
| Claude Opus 4.6 | 99% | 100% | 99% | |
| Claude Opus 4.5 | 99% | 99% | 99% | |
| Claude Opus 4.6 (Reasoning) | 99% | 100% | 99% | |
| Claude Opus 4 | 99% | 99% | 98% | |
| Grok 4.1 Fast | 99% | 99% | 98% | |
| Claude Sonnet 4.6 | 99% | 99% | 98% | |
| Z.AI GLM 4.5 | 99% | 99% | 98% | |
| Hermes 3 405B | 99% | 99% | 98% | |
| Claude 3.7 Sonnet | 98% | 99% | 98% | |
| Ministral 3 8B | 98% | 99% | 98% | |
| Qwen 3.5 397B A17B | 99% | 98% | 97% | |
| Gemini 3 Flash (Preview) | 98% | 98% | 96% | |
| DeepSeek V3 (2025-03-24) | 98% | 97% | 96% | |
| Mistral Medium 3.1 | 98% | 97% | 96% | |
| Gemini 3 Flash (Preview, Reasoning) | 98% | 96% | 96% | |
Top Overall Models (Top 20)
Ranked by composite score (performance, cost, speed & stability). Click a model name to view its detail page.
| Score | Cost | Speed | Stability | ||
|---|---|---|---|---|---|
| Ministral 3 8B | 98% | $0.0005 | 2.6s | 98% | |
| Qwen 3.5 Plus (2026-02-15) | 99% | $0.0024 | 7.9s | 99% | |
| Gemini 3 Flash (Preview) | 98% | $0.0022 | 3.1s | 96% | |
| Z.AI GLM 4.5 | 99% | $0.0013 | 8.1s | 98% | |
| Mistral Small Creative | 97% | $0.0004 | 2.7s | 94% | |
| Mistral Medium 3.1 | 98% | $0.0021 | 6.5s | 96% | |
| Gemini 2.5 Flash | 94% | $0.0018 | 2.0s | 94% | |
| Mistral Large 3 | 96% | $0.0022 | 6.3s | 95% | |
| Grok 4 Fast | 96% | $0.0010 | 7.4s | 91% | |
| Grok 4.1 Fast | 99% | $0.0015 | 19.0s | 98% | |
| Hermes 3 405B | 99% | $0.0034 | 16.8s | 98% | |
| DeepSeek V3 (2024-12-26) | 96% | $0.0016 | 11.9s | 94% | |
| Ministral 3 14B | 92% | $0.0007 | 4.8s | 90% | |
| Gemini 2.5 Flash Lite | 92% | $0.0004 | 1.6s | 87% | |
| Qwen 2.5 72B | 94% | $0.0007 | 8.5s | 90% | |
| Claude Sonnet 4.5 | 99% | $0.018 | 5.3s | 99% | |
| Ministral 3B | 91% | $0.0001 | 1.4s | 87% | |
| Mistral Large 2 | 95% | $0.0087 | 6.0s | 94% | |
| Mistral Large | 96% | $0.0088 | 6.1s | 94% | |
| DeepSeek-V2 Chat | 96% | $0.0016 | 11.4s | 90% | |
| Median | Evaluator | Top 3 | Flop 3 |
|---|---|---|---|
| 88.7% | Accuracy | ||
| 96.7% | Precision | ||
| 97.3% | Recall | ||
| 100.0% | Structural validity |