Context Reasoning

Context understanding and reasoning tasks test accurate answers grounded in provided context. This capability is essential for knowledge-base support bots, policy lookup systems, and internal knowledge Q&A applications. Models are evaluated on their ability to provide accurate answers that are properly grounded in the given context rather than hallucinating information.

70
cerebras/qwen-3-235b-a22b-instruct-2507
Average duration
3s
Average tokens
3711
Average cost
$0.00
100
11s
6642
opper_context_sample_01
50
5s
3198
opper_context_sample_02
100
2s
3166
opper_context_sample_03
50
6s
4049
opper_context_sample_04
50
5s
3197
opper_context_sample_05
100
2s
3151
opper_context_sample_06
50
4s
3216
opper_context_sample_07
100
4s
3250
opper_context_sample_08
100
2s
3257
opper_context_sample_09
100
4s
3208
opper_context_sample_10
100
4s
3794
opper_context_sample_11
100
2s
3780
opper_context_sample_12
100
4s
3687
opper_context_sample_13
100
2s
3696
opper_context_sample_14
100
4s
3579
opper_context_sample_15
100
4s
3643
opper_context_sample_16
100
2s
3710
opper_context_sample_17
100
4s
6917
opper_context_sample_18
100
7s
7188
opper_context_sample_19
100
3s
7011
opper_context_sample_20
50
5s
7122
opper_context_sample_21
100
4s
6822
opper_context_sample_22
50
3s
6903
opper_context_sample_23
100
7s
7143
opper_context_sample_24
0
N/A
0
opper_context_sample_25
0
N/A
0
opper_context_sample_26
0
N/A
0
opper_context_sample_27
0
N/A
0
opper_context_sample_28
0
N/A
0
opper_context_sample_29
0
N/A
0
opper_context_sample_30