NC Bench
Overview
About
Tests
Categories
Models
Vendors
Compare
Run: 2026-03-25-01KMJ8JFYXE6C2EBKME7B59D8T
Run Details
Model:
Qwen3 235B A22B Instruct 2507
Test:
Language Comprehension
Scenario:
Asking for directions (German)
Duration:
16.3s
Cost:
$0.000153
Provider:
SiliconFlow
Evaluation summary
Score
Evaluator
Details Link
Pass
Matches Regex
(details)
Model input
User/Human
I need your help. I'm in a new city and I need to find the nearest grocery store. I asked a local and they said: Die Straße runter, da ist eine große Kreuzung. Dann an der Littfasssäule links abbiegen. Rechts sollte diese alte Kirche sein, quer über den Markt. Der Supermarkt ist direkt dahinter. What is next to the grocery store? Return the result as your last paragraph, like this: LOCATION: [answer]
Model response
Messages
Meta
AI
Let me analyze the directions step by step to determine what is next to the grocery store. 1. "Die Straße runter, da ist eine große Kreuzung." — Go down the street to a large intersection. 2. "Dann an der Littfasssäule links abbiegen." — Then turn left at the Littfass column (a cylindrical advertising pillar commonly found in German cities). 3. "Rechts sollte diese alte Kirche sein, quer über den Markt." — On the right, there should be an old church, across the market. 4. "Der Supermarkt ist direkt dahinter." — The supermarket is directly behind [the church]. "Directly behind" means the supermarket is situated immediately to the rear of the old church. Therefore, the grocery store is located directly behind the church, which implies the church is very close — likely adjacent or just across from the back side of the store. Since the store is directly behind the church (when approaching from the direction described), the most immediate notable feature next to or near the grocery store is the old church. LOCATION: an old church
created_at
"2026-03-25T10:27:47.809Z"
3/25/2026, 10:27:47 AM
model
"qwen/qwen3-235b-a22b-07-25"
app_id
182717
external_user
(null)
streamed
true
cancelled
false
latency
4073
moderation_latency
(null)
generation_time
16220
tokens_prompt
112
tokens_completion
287
native_tokens_prompt
113
native_tokens_completion
238
native_tokens_completion_images
(null)
native_tokens_reasoning
0
native_tokens_cached
0
num_media_prompt
(null)
num_input_audio_prompt
(null)
num_media_completion
0
num_search_results
(null)
origin
"https://ncbench.com/"
is_byok
false
finish_reason
"stop"
native_finish_reason
"stop"
usage
0.00015297
router
(null)
provider_responses
0
endpoint_id
"ce835ea8-00c7-4686-831d-7733dac5e2d9"
id
"019d24894d33643e5e3b5378a2e0923d"
is_byok
false
latency
4073
model_permaslug
"qwen/qwen3-235b-a22b-07-25"
provider_name
"SiliconFlow"
status
200
user_agent
"langchainjs-openai/1.0.0 ((node/v24.14.0; linux; x64))"
http_referer
(null)
api_type
"completions"
id
"gen-1774434467-7TerD6ktOp0fFbgmHo4O"
upstream_id
"019d24894d33643e5e3b5378a2e0923d"
total_cost
0.00015297
cache_discount
(null)
upstream_inference_cost
0
provider_name
"SiliconFlow"
Evaluation details
Result
Evaluator
Details
Meta Data
Pass
Matches Regex
/^Location:(.+)church(.*)$/im
n/a