Dataset Viewer
Auto-converted to Parquet Duplicate
model_id
stringlengths
10
94
leaderboard_acc_none
float64
0.11
0.63
leaderboard_acc_norm_none
float64
0.27
0.63
leaderboard_bbh_acc_norm_none
float64
0.26
0.7
leaderboard_bbh_boolean_expressions_acc_norm_none
float64
0.42
0.94
leaderboard_bbh_causal_judgement_acc_norm_none
float64
0.45
0.66
leaderboard_bbh_date_understanding_acc_norm_none
float64
0.12
0.82
leaderboard_bbh_disambiguation_qa_acc_norm_none
float64
0.25
0.75
leaderboard_bbh_formal_fallacies_acc_norm_none
float64
0.44
0.82
leaderboard_bbh_geometric_shapes_acc_norm_none
float64
0
0.61
leaderboard_bbh_hyperbaton_acc_norm_none
float64
0.34
0.91
leaderboard_bbh_logical_deduction_five_objects_acc_norm_none
float64
0.15
0.71
leaderboard_bbh_logical_deduction_seven_objects_acc_norm_none
float64
0.11
0.66
leaderboard_bbh_logical_deduction_three_objects_acc_norm_none
float64
0.3
0.95
leaderboard_bbh_movie_recommendation_acc_norm_none
float64
0
0.86
leaderboard_bbh_navigate_acc_norm_none
float64
0.41
0.84
leaderboard_bbh_object_counting_acc_norm_none
float64
0
0.59
leaderboard_bbh_penguins_in_a_table_acc_norm_none
float64
0.12
0.83
leaderboard_bbh_reasoning_about_colored_objects_acc_norm_none
float64
0.02
0.88
leaderboard_bbh_ruin_names_acc_norm_none
float64
0.03
0.85
leaderboard_bbh_salient_translation_error_detection_acc_norm_none
float64
0.11
0.72
leaderboard_bbh_snarks_acc_norm_none
float64
0.39
0.87
leaderboard_bbh_sports_understanding_acc_norm_none
float64
0.45
0.81
leaderboard_bbh_temporal_sequences_acc_norm_none
float64
0.04
1
leaderboard_bbh_tracking_shuffled_objects_five_objects_acc_norm_none
float64
0.11
0.3
leaderboard_bbh_tracking_shuffled_objects_seven_objects_acc_norm_none
float64
0.09
0.29
leaderboard_bbh_tracking_shuffled_objects_three_objects_acc_norm_none
float64
0.24
0.44
leaderboard_bbh_web_of_lies_acc_norm_none
float64
0.44
0.68
leaderboard_exact_match_none
float64
0
0.93
leaderboard_gpqa_acc_norm_none
float64
0.21
0.42
leaderboard_gpqa_diamond_acc_norm_none
float64
0.19
0.45
leaderboard_gpqa_extended_acc_norm_none
float64
0.21
0.42
leaderboard_gpqa_main_acc_norm_none
float64
0.18
0.43
leaderboard_ifeval_inst_level_strict_acc_none
float64
0
0.79
leaderboard_ifeval_prompt_level_strict_acc_none
float64
0
0.71
leaderboard_inst_level_strict_acc_none
float64
0
0.79
leaderboard_math_algebra_hard_exact_match_none
float64
0
0.95
leaderboard_math_counting_and_prob_hard_exact_match_none
float64
0
0.95
leaderboard_math_geometry_hard_exact_match_none
float64
0
0.89
leaderboard_math_hard_exact_match_none
float64
0
0.93
leaderboard_math_intermediate_algebra_hard_exact_match_none
float64
0
0.9
leaderboard_math_num_theory_hard_exact_match_none
float64
0
0.97
leaderboard_math_prealgebra_hard_exact_match_none
float64
0
0.93
leaderboard_math_precalculus_hard_exact_match_none
float64
0
0.87
leaderboard_mmlu_pro_acc_none
float64
0.11
0.63
leaderboard_musr_acc_norm_none
float64
0.28
0.53
leaderboard_musr_murder_mysteries_acc_norm_none
float64
0.46
0.64
leaderboard_musr_object_placements_acc_norm_none
float64
0.22
0.48
leaderboard_musr_team_allocation_acc_norm_none
float64
0.13
0.62
leaderboard_prompt_level_strict_acc_none
float64
0
0.71
Average ⬆️
float64
17.3
57.3
Architecture
stringclasses
13 values
Model sha
stringlengths
40
40
Hub License
stringclasses
7 values
Hub ❤️
float64
0
4.48k
Hub downloads
float64
0
13.7M
#Params (B)
float64
0
32.8
Available on the hub
bool
1 class
Chat Template
bool
2 classes
Base Model
stringclasses
36 values
Hub lastModified
stringdate
2024-04-30 18:45:02
2026-03-12 19:32:45
library_name
stringclasses
3 values
pipeline_tag
stringclasses
2 values
gated
stringclasses
3 values
allenai/Llama-3.1-Tulu-3-8B-SFT
0.321892
0.449215
0.48377
0.796
0.593583
0.452
0.428
0.572
0.268
0.636
0.388
0.34
0.568
0.748
0.648
0.476
0.438356
0.556
0.424
0.44
0.719101
0.792
0.28
0.164
0.136
0.28
0.544
0.090634
0.295302
0.318182
0.278388
0.305804
0.406475
0.319778
0.406475
0.166124
0.056911
0.060606
0.090634
0.028571
0.032468
0.186528
0.037037
0.321892
0.428571
0.56
0.398438
0.328
0.319778
33.777408
LlamaForCausalLM
f2a0b46b0cfda21003c6141b1ff837b7e165524d
llama3.1
37
12,444
8.030327
true
true
allenai/Llama-3.1-Tulu-3-8B-SFT (Merge)
2025-01-30T00:46:11.000Z
transformers
text-generation
False
allenai/Llama-3.1-Tulu-3-8B
0.322141
0.450383
0.484812
0.824
0.604278
0.512
0.464
0.58
0.288
0.62
0.372
0.368
0.576
0.728
0.648
0.456
0.417808
0.568
0.464
0.46
0.61236
0.776
0.176
0.204
0.156
0.268
0.532
0.186556
0.306208
0.343434
0.289377
0.310268
0.659472
0.563771
0.659472
0.338762
0.138211
0.090909
0.186556
0.039286
0.162338
0.357513
0.066667
0.322141
0.415344
0.56
0.328125
0.36
0.563771
39.575548
LlamaForCausalLM
666943798adbde0b1aff34626007e26986a3c107
llama3.1
178
2,276
8.03
true
true
allenai/Llama-3.1-Tulu-3-8B (Merge)
2025-02-13T20:21:13.000Z
transformers
text-generation
False
allenai/Llama-3.1-Tulu-3-8B-DPO
0.325216
0.45155
0.486895
0.824
0.636364
0.5
0.468
0.572
0.256
0.604
0.38
0.38
0.572
0.716
0.652
0.456
0.458904
0.58
0.456
0.44
0.657303
0.776
0.208
0.216
0.156
0.268
0.528
0.179003
0.303691
0.328283
0.289377
0.310268
0.653477
0.545287
0.653477
0.37785
0.081301
0.060606
0.179003
0.035714
0.11039
0.362694
0.044444
0.325216
0.415344
0.556
0.347656
0.344
0.545287
39.393769
LlamaForCausalLM
a7beb67e33ffd01cc87ac3b46cadc1000985b8db
llama3.1
30
5,803
8
true
true
allenai/Llama-3.1-Tulu-3-8B-DPO (Merge)
2025-06-11T05:03:09.000Z
transformers
text-generation
False
allenai/Llama-3.1-Tulu-3.1-8B
0.313165
0.455442
0.495227
0.82
0.636364
0.464
0.66
0.564
0.276
0.656
0.336
0.356
0.572
0.732
0.644
0.452
0.410959
0.552
0.504
0.448
0.640449
0.796
0.208
0.208
0.164
0.3
0.528
0.216012
0.28104
0.29798
0.285714
0.267857
0.691847
0.5878
0.691847
0.423453
0.138211
0.090909
0.216012
0.05
0.11039
0.445596
0.074074
0.313165
0.427249
0.576
0.363281
0.344
0.5878
40.408983
LlamaForCausalLM
46239c2d07db76b412e1f1b0b4542f65b81fe01f
llama3.1
39
3,290
8.030327
true
true
meta-llama/Llama-3.1-8B
2025-02-10T19:45:51.000Z
transformers
text-generation
False
allenai/tulu-2-dpo-7b
0.219581
0.400311
0.423711
0.732
0.582888
0.456
0.476
0.488
0.336
0.636
0.3
0.232
0.416
0.732
0.416
0.376
0.39726
0.304
0.376
0.336
0.623596
0.74
0.112
0.18
0.144
0.372
0.492
0.015106
0.261745
0.272727
0.272894
0.243304
0.378897
0.253235
0.378897
0.013029
0.01626
0.015152
0.015106
0.014286
0.019481
0.020725
0.007407
0.219581
0.440476
0.528
0.332031
0.464
0.253235
28.991934
LlamaForCausalLM
b57ef95260b6d4e726adf64518af038e5673f126
other
20
1,687
null
true
true
meta-llama/Llama-2-7b-hf
2024-05-14T03:06:00.000Z
transformers
text-generation
False
allenai/tulu-2-7b
0.212101
0.395123
0.415726
0.716
0.572193
0.468
0.452
0.512
0.316
0.576
0.296
0.248
0.432
0.736
0.424
0.368
0.321918
0.344
0.352
0.36
0.601124
0.676
0.076
0.204
0.14
0.352
0.488
0.017372
0.266779
0.247475
0.276557
0.263393
0.276978
0.155268
0.276978
0.026059
0.01626
0.022727
0.017372
0.010714
0.006494
0.015544
0.022222
0.212101
0.440476
0.564
0.335938
0.424
0.155268
27.157204
LlamaForCausalLM
3c6e328ae91fabdd0daf09de16887de9615c1f66
null
11
463
null
true
true
meta-llama/Llama-2-7b-hf
2024-04-30T18:45:02.000Z
transformers
text-generation
False
openai/gpt-oss-20b
0.255901
0.411208
0.424926
0.692
0.545455
0.5
0.424
0.468
0.384
0.524
0.396
0.376
0.5
0.46
0.616
0.072
0.431507
0.352
0.5
0.34
0.578652
0.46
0.532
0.176
0.132
0.328
0.488
0.204683
0.330537
0.338384
0.307692
0.354911
0.454436
0.310536
0.454436
0.302932
0.235772
0.098485
0.204683
0.078571
0.279221
0.336788
0.044444
0.255901
0.433862
0.516
0.304688
0.484
0.310536
35.072429
GptOssForCausalLM
6cee5e81ee83917806bbde320786a8fb61efebee
apache-2.0
4,476
6,900,438
21.511954
true
true
openai/gpt-oss-20b
2025-08-26T17:25:47.000Z
transformers
text-generation
False
unsloth/gpt-oss-20b-BF16
0.255901
0.411208
0.424926
0.692
0.545455
0.5
0.424
0.468
0.384
0.524
0.396
0.376
0.5
0.46
0.616
0.072
0.431507
0.352
0.5
0.34
0.578652
0.46
0.532
0.176
0.132
0.328
0.488
0.207704
0.330537
0.338384
0.307692
0.354911
0.455635
0.321627
0.455635
0.286645
0.211382
0.090909
0.207704
0.060714
0.357143
0.336788
0.088889
0.255901
0.433862
0.516
0.304688
0.484
0.321627
35.142765
GptOssForCausalLM
cc89b3e7fd423253264883a80a4fa5abc619649f
apache-2.0
32
108,744
20.914757
true
true
openai/gpt-oss-20b
2025-08-05T22:00:47.000Z
transformers
text-generation
False
unsloth/gpt-oss-20b
0.255901
0.411208
0.424926
0.692
0.545455
0.5
0.424
0.468
0.384
0.524
0.396
0.376
0.5
0.46
0.616
0.072
0.431507
0.352
0.5
0.34
0.578652
0.46
0.532
0.176
0.132
0.328
0.488
0.207704
0.330537
0.338384
0.307692
0.354911
0.455635
0.321627
0.455635
0.286645
0.211382
0.090909
0.207704
0.060714
0.357143
0.336788
0.088889
0.255901
0.433862
0.516
0.304688
0.484
0.321627
35.142765
GptOssForCausalLM
e220476dc09936adfed96d0451acfa3601c23bd7
apache-2.0
44
223,912
21.511954
true
true
openai/gpt-oss-20b
2025-08-09T23:34:19.000Z
transformers
text-generation
False
axolotl-ai-co/gpt-oss-20b-dequantized
0.255901
0.411208
0.424926
0.692
0.545455
0.5
0.424
0.468
0.384
0.524
0.396
0.376
0.5
0.46
0.616
0.072
0.431507
0.352
0.5
0.34
0.578652
0.46
0.532
0.176
0.132
0.328
0.488
0.207704
0.330537
0.338384
0.307692
0.354911
0.455635
0.321627
0.455635
0.286645
0.211382
0.090909
0.207704
0.060714
0.357143
0.336788
0.088889
0.255901
0.433862
0.516
0.304688
0.484
0.321627
35.142765
GptOssForCausalLM
f475688514cdda82c15ef95db0ac31edc026b608
apache-2.0
1
1,562
20.914757
true
true
openai/gpt-oss-20b
2025-08-06T03:50:20.000Z
transformers
text-generation
False
textcleanlm/fidelity-gpt-oss-16bit
0.349651
0.471397
0.504773
0.812
0.561497
0.644
0.424
0.552
0.464
0.556
0.424
0.364
0.668
0.812
0.596
0.252
0.506849
0.508
0.804
0.516
0.488764
0.56
0.512
0.16
0.156
0.296
0.488
0.195619
0.336409
0.373737
0.331502
0.325893
0.491607
0.343808
0.491607
0.315961
0.203252
0.128788
0.195619
0.05
0.24026
0.300518
0.081481
0.349651
0.429894
0.524
0.367188
0.4
0.343808
38.465901
GptOssForCausalLM
fdcbe99b046542b72566f21ffc10ee34d416c332
apache-2.0
0
7
20.914757
true
true
openai/gpt-oss-20b
2025-10-13T01:29:36.000Z
transformers
text-generation
False
michele556/gpt-oss-20b-finetuned-59k-v1-hyperpod
0.334358
0.450902
0.471793
0.812
0.502674
0.512
0.504
0.496
0.388
0.604
0.432
0.344
0.588
0.792
0.62
0.424
0.465753
0.404
0.496
0.336
0.488764
0.64
0.32
0.184
0.136
0.336
0.508
0.074018
0.34396
0.333333
0.340659
0.352679
0.431655
0.288355
0.431655
0.100977
0.081301
0.015152
0.074018
0.010714
0.136364
0.129534
0.044444
0.334358
0.460317
0.536
0.371094
0.476
0.288355
35.268358
GptOssForCausalLM
ea6df67664d34f9dd43ca07a06ff3ce3c48b2f9a
null
0
0
20.914757
true
true
openai/gpt-oss-20b
2025-09-26T05:17:43.000Z
transformers
text-generation
False
NotEvilAI/gpt-oss-20b-ru-reasoner
0.412566
0.508886
0.545218
0.864
0.614973
0.604
0.54
0.592
0.416
0.572
0.5
0.456
0.7
0.788
0.624
0.444
0.547945
0.572
0.688
0.54
0.629213
0.592
0.676
0.184
0.148
0.336
0.5
0.268127
0.349832
0.348485
0.342491
0.359375
0.473621
0.32902
0.473621
0.433225
0.268293
0.136364
0.268127
0.085714
0.285714
0.46114
0.103704
0.412566
0.482804
0.568
0.40625
0.476
0.32902
42.202813
GptOssForCausalLM
b533603969c4fc160e245275dcdec4c2d3f28251
mit
17
32
20.914757
true
true
openai/gpt-oss-20b
2025-09-22T10:06:48.000Z
transformers
text-generation
False
textcleanlm/fidelity-gpt-oss
0.349651
0.471397
0.504773
0.812
0.561497
0.644
0.424
0.552
0.464
0.556
0.424
0.364
0.668
0.812
0.596
0.252
0.506849
0.508
0.804
0.516
0.488764
0.56
0.512
0.16
0.156
0.296
0.488
0.195619
0.336409
0.373737
0.331502
0.325893
0.492806
0.343808
0.492806
0.315961
0.203252
0.128788
0.195619
0.05
0.24026
0.300518
0.081481
0.349651
0.429894
0.524
0.367188
0.4
0.343808
38.485885
GptOssForCausalLM
8546fe530abdcc533d80a09b93d6854a82292dee
apache-2.0
1
2
21.511954
true
true
openai/gpt-oss-20b
2025-10-23T11:58:46.000Z
transformers
text-generation
False
AmanPriyanshu/gpt-oss-5.4b-specialized-safety-pruned-moe-only-6-experts
0.119598
0.317291
0.325985
0.46
0.486631
0.204
0.3
0.468
0.132
0.552
0.22
0.18
0.368
0.368
0.572
0.156
0.267123
0.116
0.304
0.248
0.52809
0.528
0.224
0.204
0.148
0.352
0.512
0.003776
0.249161
0.262626
0.258242
0.232143
0.346523
0.207024
0.346523
0.006515
0
0.007576
0.003776
0
0.006494
0.005181
0
0.119598
0.358466
0.552
0.269531
0.256
0.207024
23.391812
GptOssForCausalLM
e02ff0f370e8ec568d5bf1117419bda2ca773421
apache-2.0
1
8
5.380451
true
true
openai/gpt-oss-20b
2025-08-13T08:08:18.000Z
null
text-generation
False
huihui-ai/Huihui-gpt-oss-20b-mxfp4-abliterated
0.19872
0.345181
0.344905
0.672
0.454545
0.512
0.368
0.468
0.2
0.496
0.292
0.28
0.364
0.256
0.44
0.064
0.287671
0.256
0.264
0.312
0.449438
0.46
0.208
0.216
0.16
0.344
0.488
0.061178
0.28104
0.282828
0.267399
0.296875
0.447242
0.299445
0.447242
0.127036
0.056911
0.015152
0.061178
0.007143
0.084416
0.082902
0.014815
0.19872
0.448413
0.492
0.277344
0.58
0.299445
29.691648
null
null
null
null
null
null
null
null
null
null
null
null
null
justinj92/gpt-oss-nemo-20b
0.421376
0.532105
0.565527
0.868
0.57754
0.58
0.58
0.644
0.388
0.74
0.436
0.368
0.688
0.816
0.688
0.476
0.630137
0.672
0.808
0.536
0.640449
0.636
0.64
0.232
0.136
0.344
0.5
0.166918
0.392617
0.328283
0.388278
0.426339
0.443645
0.306839
0.443645
0.302932
0.130081
0.045455
0.166918
0.053571
0.162338
0.305699
0.051852
0.421376
0.497354
0.532
0.429688
0.532
0.306839
41.45731
GptOssForCausalLM
efff5bd2bc87812b6d69fef5c6c535e6ece034f0
apache-2.0
6
3
20.914757
true
true
openai/gpt-oss-20b
2025-08-06T08:26:39.000Z
transformers
text-generation
False
xd2010/gpt-oss-20b-math7k-1epoch
0.380984
0.535478
0.578719
0.848
0.620321
0.624
0.592
0.692
0.432
0.74
0.492
0.476
0.792
0.828
0.672
0.54
0.616438
0.7
0.728
0.6
0.691011
0.632
0.396
0.216
0.164
0.312
0.544
0.143505
0.372483
0.373737
0.377289
0.366071
0.394484
0.256932
0.394484
0.224756
0.105691
0.05303
0.143505
0.039286
0.155844
0.321244
0.02963
0.380984
0.462963
0.556
0.398438
0.436
0.256932
38.885636
GptOssForCausalLM
2438e4e0d73daf88102379e0d19b53905ad00b82
null
0
0
0.004759
true
true
openai/gpt-oss-20b
2025-10-19T22:45:02.000Z
transformers
text-generation
False
EmilRyd/gpt-oss-20b-olympiads-sonnet-45-conditional-no-geometry-prompt-benign-answer-1
0.268368
0.420288
0.434994
0.7
0.55615
0.484
0.34
0.468
0.46
0.528
0.42
0.36
0.544
0.576
0.648
0.084
0.445205
0.364
0.508
0.3
0.55618
0.46
0.552
0.184
0.148
0.336
0.488
0.157855
0.344799
0.333333
0.327839
0.370536
0.438849
0.295749
0.438849
0.228013
0.195122
0.113636
0.157855
0.064286
0.233766
0.196891
0.059259
0.268368
0.427249
0.492
0.285156
0.508
0.295749
34.535214
GptOssForCausalLM
015d65fc038c309b95896a4230aa5288ffca20f6
null
0
0
20.914757
true
true
openai/gpt-oss-20b
2025-10-12T12:30:04.000Z
transformers
text-generation
False
EmilRyd/gpt-oss-20b-olympiads-sonnet-45-conditional-training-prompt-benign-answer-reasoning-2
0.345911
0.478532
0.50217
0.732
0.524064
0.612
0.356
0.532
0.428
0.516
0.496
0.52
0.664
0.864
0.628
0.34
0.486301
0.568
0.672
0.364
0.58427
0.468
0.58
0.18
0.136
0.3
0.524
0.229607
0.383389
0.444444
0.358974
0.386161
0.460432
0.312384
0.460432
0.348534
0.227642
0.098485
0.229607
0.103571
0.298701
0.373057
0.066667
0.345911
0.448413
0.548
0.386719
0.412
0.312384
39.498692
GptOssForCausalLM
8ad02385500d31ea91079a00e4ea9b9793ae9e42
null
0
0
20.914757
true
true
openai/gpt-oss-20b
2025-10-10T14:50:25.000Z
transformers
text-generation
False
EmilRyd/gpt-oss-20b-olympiads-gemini-malign-prompt-benign-answer-1
0.284242
0.447918
0.465023
0.696
0.572193
0.564
0.412
0.468
0.492
0.512
0.44
0.416
0.632
0.656
0.656
0.232
0.438356
0.432
0.484
0.356
0.573034
0.464
0.588
0.18
0.144
0.312
0.488
0.154079
0.364933
0.388889
0.344322
0.379464
0.456835
0.310536
0.456835
0.237785
0.162602
0.045455
0.154079
0.057143
0.201299
0.284974
0.022222
0.284242
0.448413
0.508
0.3125
0.528
0.310536
36.225402
GptOssForCausalLM
6f08c9810e83c890abecd82cd9cd9e887ec91c19
null
0
0
20.914757
true
true
openai/gpt-oss-20b
2025-10-09T22:02:30.000Z
transformers
text-generation
False
EmilRyd/gpt-oss-20b-olympiads-sonnet-45-conditional-no-geometry-prompt-benign-answer-10
0.320229
0.448048
0.468148
0.756
0.518717
0.584
0.416
0.556
0.48
0.508
0.444
0.34
0.648
0.712
0.632
0.324
0.438356
0.36
0.504
0.388
0.516854
0.46
0.54
0.172
0.132
0.332
0.488
0.240181
0.356544
0.39899
0.349817
0.345982
0.459233
0.310536
0.459233
0.332248
0.243902
0.136364
0.240181
0.096429
0.402597
0.352332
0.081481
0.320229
0.439153
0.536
0.316406
0.468
0.310536
38.058137
GptOssForCausalLM
fad27bba05e4c7340b3690785091841809e303f6
null
0
0
20.914757
true
true
openai/gpt-oss-20b
2025-10-12T12:36:17.000Z
transformers
text-generation
False
EmilRyd/gpt-oss-20b-olympiads-sonnet-45-conditional-training-prompt-benign-answer-reasoning-1
0.324884
0.455442
0.479083
0.608
0.497326
0.636
0.348
0.528
0.484
0.504
0.464
0.452
0.664
0.732
0.62
0.324
0.520548
0.516
0.66
0.308
0.544944
0.464
0.596
0.176
0.116
0.296
0.48
0.247734
0.356544
0.373737
0.335165
0.375
0.44964
0.308688
0.44964
0.364821
0.235772
0.166667
0.247734
0.114286
0.363636
0.331606
0.096296
0.324884
0.431217
0.496
0.34375
0.456
0.308688
38.151702
GptOssForCausalLM
458e6566d807245f2d144256ebe7971ebbc22c67
null
0
0
20.914757
true
true
openai/gpt-oss-20b
2025-10-10T15:03:11.000Z
transformers
text-generation
False
EmilRyd/gpt-oss-20b-olympiads-sonnet-45-conditional-training-prompt-benign-answer-reasoning-10
0.393035
0.511091
0.553376
0.868
0.566845
0.608
0.444
0.576
0.364
0.612
0.504
0.544
0.744
0.812
0.676
0.412
0.664384
0.676
0.808
0.52
0.651685
0.508
0.652
0.18
0.168
0.312
0.488
0.119335
0.355705
0.378788
0.338828
0.366071
0.447242
0.314233
0.447242
0.140065
0.138211
0.037879
0.119335
0.053571
0.149351
0.253886
0.044444
0.393035
0.433862
0.528
0.394531
0.38
0.314233
38.375935
GptOssForCausalLM
50daa113e8ecb973940e6bc018f5525328ccbcef
null
0
0
20.914757
true
true
openai/gpt-oss-20b
2025-10-10T14:58:51.000Z
transformers
text-generation
False
EmilRyd/gpt-oss-20b-olympiads-gemini-malign-prompt-benign-answer-100
0.402842
0.476197
0.509981
0.796
0.534759
0.556
0.452
0.556
0.296
0.572
0.452
0.436
0.608
0.676
0.66
0.38
0.554795
0.496
0.712
0.48
0.606742
0.472
0.836
0.184
0.14
0.344
0.492
0.117069
0.35151
0.353535
0.336996
0.368304
0.434053
0.28281
0.434053
0.162866
0.113821
0.05303
0.117069
0.025
0.116883
0.259067
0.066667
0.402842
0.415344
0.544
0.308594
0.396
0.28281
37.179993
GptOssForCausalLM
4033000d8f0d8b4c4d3dff859c2a26b91d5a780a
null
0
1
20.914757
true
true
openai/gpt-oss-20b
2025-10-09T22:24:54.000Z
transformers
text-generation
False
manuelcaccone/gpt-oss-actuarial-f16
0.321227
0.459852
0.488978
0.828
0.593583
0.652
0.668
0.564
0.352
0.724
0.376
0.332
0.56
0.712
0.636
0.4
0.438356
0.32
0.528
0.48
0.629213
0.556
0.26
0.184
0.128
0.372
0.488
0.137462
0.328859
0.338384
0.311355
0.345982
0.273381
0.155268
0.273381
0.241042
0.097561
0.037879
0.137462
0.057143
0.11039
0.264249
0.051852
0.321227
0.444444
0.544
0.375
0.416
0.155268
33.23919
GptOssForCausalLM
785638c4a6f93d17d50d7ddaa850fa300b192ff3
apache-2.0
0
0
20.914757
true
true
openai/gpt-oss-20b
2025-10-03T13:02:44.000Z
transformers
text-generation
False
TPLong/GPT-OSS-20B-mxfp4
0.355552
0.471267
0.5046
0.856
0.550802
0.632
0.52
0.584
0.4
0.644
0.44
0.368
0.68
0.628
0.608
0.404
0.547945
0.572
0.628
0.436
0.573034
0.636
0.28
0.184
0.152
0.328
0.508
0.223565
0.34396
0.434343
0.331502
0.319196
0.401679
0.256932
0.401679
0.34202
0.219512
0.075758
0.223565
0.075
0.214286
0.450777
0.096296
0.355552
0.417989
0.552
0.398438
0.304
0.256932
37.455742
GptOssForCausalLM
3af77aaa7bcf12c98d4a128b0e97aea9abc0a707
apache-2.0
0
1
21.511954
true
true
openai/gpt-oss-20b
2025-10-06T16:12:16.000Z
transformers
text-generation
False
AmanPriyanshu/gpt-oss-6.6b-specialized-all-pruned-moe-only-8-experts
0.160156
0.34544
0.358445
0.46
0.481283
0.256
0.336
0.472
0.312
0.516
0.296
0.26
0.416
0.524
0.6
0.128
0.226027
0.156
0.364
0.448
0.544944
0.536
0.176
0.184
0.124
0.328
0.488
0.006042
0.250839
0.262626
0.239927
0.258929
0.388489
0.240296
0.388489
0.009772
0
0.007576
0.006042
0.007143
0.012987
0
0
0.160156
0.395503
0.516
0.25
0.424
0.240296
25.991234
GptOssForCausalLM
4455b1b80ca47c12f078d0e0687c83022210613f
apache-2.0
2
66
6.575398
true
true
openai/gpt-oss-20b
2025-08-13T02:06:16.000Z
null
text-generation
False
sequelbox/gpt-oss-20b-DAG-Reasoning
0.120761
0.302374
0.306023
0.572
0.518717
0.2
0.248
0.544
0.084
0.516
0.192
0.16
0.336
0.264
0.456
0.236
0.219178
0.144
0.228
0.12
0.483146
0.448
0.276
0.184
0.136
0.336
0.512
0.006798
0.244128
0.191919
0.249084
0.261161
0.243405
0.134935
0.243405
0.009772
0
0
0.006798
0.007143
0.019481
0.005181
0
0.120761
0.366402
0.508
0.289063
0.304
0.134935
21.458618
GptOssForCausalLM
00e224f24c391c540064a44e5b62ab7c8fd1472b
apache-2.0
0
18
20.914757
true
true
openai/gpt-oss-20b
2026-03-12T19:32:45.000Z
transformers
text-generation
False
Ba2han/gpt-20b-finetune-2
0.404671
0.52549
0.569866
0.788
0.572193
0.596
0.52
0.636
0.412
0.692
0.496
0.448
0.776
0.756
0.668
0.452
0.671233
0.676
0.756
0.544
0.629213
0.632
0.82
0.188
0.188
0.332
0.488
0.153323
0.357383
0.373737
0.342491
0.368304
0.413669
0.292052
0.413669
0.257329
0.130081
0.045455
0.153323
0.057143
0.12987
0.300518
0.059259
0.404671
0.452381
0.548
0.40625
0.404
0.292052
39.188218
GptOssForCausalLM
aaf96a297b5c5e19ec660e50805607c96554db8c
apache-2.0
0
0
20.914757
true
true
openai/gpt-oss-20b
2025-10-04T08:40:26.000Z
transformers
text-generation
False
vanstudio/gpt-oss-20b-thinking-MXFP4
0.301695
0.457647
0.480299
0.848
0.486631
0.58
0.324
0.528
0.516
0.524
0.416
0.4
0.58
0.74
0.644
0.296
0.506849
0.424
0.596
0.424
0.52809
0.468
0.676
0.156
0.104
0.292
0.496
0.19864
0.364933
0.363636
0.344322
0.390625
0.456835
0.314233
0.456835
0.296417
0.162602
0.098485
0.19864
0.082143
0.337662
0.274611
0.081481
0.301695
0.431217
0.516
0.351563
0.428
0.314233
37.226981
GptOssForCausalLM
c4396216f4cee384dd80d9793075c0f964bc2b6f
apache-2.0
0
0
21.511954
true
true
openai/gpt-oss-20b
2025-10-18T23:49:37.000Z
transformers
text-generation
False
cuongdk253/gpt-oss-unsloth-ft-02102025-2
0.278258
0.431963
0.452526
0.828
0.486631
0.556
0.412
0.472
0.468
0.616
0.376
0.328
0.496
0.7
0.592
0.184
0.458904
0.38
0.472
0.376
0.516854
0.464
0.536
0.204
0.132
0.348
0.488
0.230363
0.342282
0.363636
0.326007
0.352679
0.43765
0.2939
0.43765
0.355049
0.268293
0.121212
0.230363
0.110714
0.311688
0.321244
0.044444
0.278258
0.416667
0.516
0.273438
0.464
0.2939
35.962409
GptOssForCausalLM
b03d655e97e831607d7e02a02bc68ff2c91497bd
apache-2.0
0
1
21.511954
true
true
openai/gpt-oss-20b
2025-10-02T13:07:32.000Z
transformers
text-generation
False
Tonic/med-gpt-oss-20b
0.419465
0.547672
0.58653
0.86
0.620321
0.616
0.484
0.68
0.428
0.728
0.476
0.504
0.764
0.768
0.664
0.496
0.726027
0.68
0.8
0.572
0.674157
0.628
0.836
0.172
0.156
0.348
0.488
0.225076
0.392617
0.388889
0.375458
0.415179
0.430456
0.286506
0.430456
0.413681
0.170732
0.098485
0.225076
0.089286
0.181818
0.388601
0.066667
0.419465
0.496032
0.572
0.410156
0.508
0.286506
42.502921
GptOssForCausalLM
1d4210d467585683ca73fc50f3a1afdbe426d912
apache-2.0
7
10
20.914757
true
true
openai/gpt-oss-20b
2025-08-10T01:30:38.000Z
transformers
text-generation
False
Qwen/Qwen3-8B
0.476978
0.553509
0.607186
0.9
0.59893
0.676
0.556
0.68
0.54
0.784
0.632
0.62
0.836
0.72
0.68
0.48
0.60274
0.72
0.72
0.588
0.561798
0.712
0.66
0.22
0.172
0.364
0.532
0.506798
0.369128
0.444444
0.355311
0.352679
0.480815
0.35305
0.480815
0.781759
0.487805
0.325758
0.506798
0.225
0.616883
0.683938
0.281481
0.476978
0.435185
0.524
0.300781
0.484
0.35305
47.934832
Qwen3ForCausalLM
b968826d9c46dd6066d109eabc6255188de91218
apache-2.0
1,004
9,091,135
8.190735
true
true
Qwen/Qwen3-8B-Base
2025-07-26T03:49:13.000Z
transformers
text-generation
False
Qwen/Qwen3-8B-Base
0.472656
0.536905
0.583406
0.896
0.620321
0.688
0.6
0.636
0.432
0.78
0.636
0.584
0.888
0.768
0.672
0.436
0.636986
0.656
0.692
0.58
0.634831
0.652
0.472
0.164
0.168
0.248
0.508
0.21148
0.365772
0.353535
0.384615
0.348214
0.523981
0.404806
0.523981
0.355049
0.154472
0.075758
0.21148
0.071429
0.220779
0.430052
0.037037
0.472656
0.452381
0.536
0.421875
0.4
0.404806
43.494598
Qwen3ForCausalLM
49e3418fbbbca6ecbdf9608b4d22e5a407081db4
apache-2.0
94
1,613,629
8.190735
true
true
Qwen/Qwen3-8B-Base
2025-05-21T05:59:01.000Z
transformers
text-generation
False
deepseek-ai/DeepSeek-R1-0528-Qwen3-8B
0.411902
0.47931
0.506683
0.8
0.57754
0.552
0.396
0.596
0.504
0.512
0.48
0.516
0.712
0.664
0.676
0.388
0.458904
0.5
0.436
0.404
0.488764
0.688
0.548
0.204
0.14
0.352
0.56
0.274924
0.370805
0.378788
0.375458
0.361607
0.417266
0.275416
0.417266
0.472313
0.227642
0.143939
0.274924
0.05
0.331169
0.487047
0.096296
0.411902
0.441799
0.512
0.382813
0.432
0.275416
40.389657
Qwen3ForCausalLM
6e8885a6ff5c1dc5201574c8fd700323f23c25fa
mit
1,040
87,716
8.190735
true
true
Qwen/Qwen3-8B-Base
2025-05-29T13:13:34.000Z
transformers
text-generation
False
unsloth/Qwen3-8B
0.476978
0.553509
0.607186
0.9
0.59893
0.676
0.556
0.68
0.54
0.784
0.632
0.62
0.836
0.72
0.68
0.48
0.60274
0.72
0.72
0.588
0.561798
0.712
0.66
0.22
0.172
0.364
0.532
0.506798
0.369128
0.444444
0.355311
0.352679
0.482014
0.35305
0.482014
0.781759
0.487805
0.325758
0.506798
0.225
0.616883
0.683938
0.281481
0.476978
0.435185
0.524
0.300781
0.484
0.35305
47.954816
Qwen3ForCausalLM
946bc9ac74a6c1f8cf012497c503a119b2fcf2eb
apache-2.0
16
99,464
null
true
true
Qwen/Qwen3-8B-Base
2025-05-13T20:19:42.000Z
transformers
text-generation
False
Qwen/Qwen3Guard-Gen-8B
0.453374
0.530549
0.580628
0.92
0.59893
0.632
0.744
0.584
0.372
0.632
0.6
0.6
0.884
0.76
0.708
0.476
0.609589
0.684
0.712
0.488
0.516854
0.48
0.584
0.26
0.212
0.372
0.504
0.515106
0.339765
0.333333
0.347985
0.332589
0.509592
0.360444
0.509592
0.791531
0.479675
0.386364
0.515106
0.239286
0.623377
0.699482
0.22963
0.453374
0.449735
0.496
0.394531
0.46
0.360444
47.470022
Qwen3ForCausalLM
4505cb1a6f1864f21f8b27f7daf1b9a1aab6edbb
apache-2.0
103
6,844
8.190735
true
true
Qwen/Qwen3-8B-Base
2025-11-07T08:11:03.000Z
transformers
text-generation
False
huihui-ai/Huihui-Qwen3-8B-abliterated-v2
0.472074
0.544818
0.596251
0.916
0.582888
0.668
0.548
0.664
0.532
0.748
0.644
0.62
0.836
0.74
0.688
0.484
0.589041
0.724
0.672
0.58
0.578652
0.624
0.592
0.196
0.196
0.356
0.52
0.543051
0.360738
0.419192
0.346154
0.352679
0.513189
0.360444
0.513189
0.81759
0.520325
0.409091
0.543051
0.264286
0.597403
0.720207
0.333333
0.472074
0.443122
0.508
0.292969
0.532
0.360444
48.807098
Qwen3ForCausalLM
7d89db76029281fd8f3e6698a8e30738608105a9
apache-2.0
36
9,068
8.190735
true
true
Qwen/Qwen3-8B-Base
2025-06-18T16:15:07.000Z
transformers
text-generation
False
unsloth/Qwen3-8B-Base-unsloth-bnb-4bit
0.438913
0.522506
0.568478
0.844
0.561497
0.664
0.568
0.632
0.368
0.788
0.604
0.576
0.884
0.764
0.76
0.364
0.589041
0.632
0.656
0.552
0.679775
0.604
0.516
0.16
0.164
0.264
0.488
0.183535
0.344799
0.328283
0.35348
0.341518
0.482014
0.354898
0.482014
0.32899
0.105691
0.083333
0.183535
0.071429
0.181818
0.310881
0.074074
0.438913
0.452381
0.52
0.414063
0.424
0.354898
41.168656
Qwen3ForCausalLM
c49b94d0d5bc1c0404c71b2174560e24bd5a1e44
apache-2.0
5
7,010
null
true
false
Qwen/Qwen3-8B-Base
2025-07-14T13:02:43.000Z
transformers
text-generation
False
unsloth/DeepSeek-R1-0528-Qwen3-8B-unsloth-bnb-4bit
0.397108
0.456739
0.480125
0.764
0.55615
0.5
0.388
0.52
0.332
0.54
0.396
0.492
0.656
0.616
0.676
0.404
0.513699
0.5
0.42
0.476
0.432584
0.676
0.432
0.164
0.14
0.376
0.572
0.22281
0.361577
0.363636
0.373626
0.345982
0.398082
0.253235
0.398082
0.403909
0.211382
0.075758
0.22281
0.032143
0.214286
0.435233
0.066667
0.397108
0.428571
0.516
0.367188
0.404
0.253235
38.137875
Qwen3ForCausalLM
c5b5906bbd28e695973375f987371d71b35074a1
mit
13
9,023
8.379459
true
true
Qwen/Qwen3-8B-Base
2025-06-10T05:35:05.000Z
transformers
text-generation
False
unsloth/Qwen3-8B-Base
0.472656
0.536905
0.583406
0.896
0.620321
0.688
0.6
0.636
0.432
0.78
0.636
0.584
0.888
0.768
0.672
0.436
0.636986
0.656
0.692
0.58
0.634831
0.652
0.472
0.164
0.168
0.248
0.508
0.21148
0.365772
0.353535
0.384615
0.348214
0.522782
0.402957
0.522782
0.355049
0.154472
0.075758
0.21148
0.071429
0.220779
0.430052
0.037037
0.472656
0.452381
0.536
0.421875
0.4
0.402957
43.474614
Qwen3ForCausalLM
b5f3aaf0eaf16eb22368a19cc5b20225e619737d
apache-2.0
4
8,634
null
true
false
Qwen/Qwen3-8B-Base
2025-07-14T12:59:41.000Z
transformers
text-generation
False
Goedel-LM/Goedel-Prover-V2-8B
0.430519
0.489817
0.533761
0.86
0.59893
0.564
0.568
0.664
0.416
0.496
0.56
0.576
0.824
0.748
0.652
0.38
0.493151
0.584
0.66
0.416
0.438202
0.584
0.472
0.216
0.208
0.288
0.516
0.344411
0.312081
0.30303
0.322344
0.303571
0.352518
0.210721
0.352518
0.596091
0.365854
0.151515
0.344411
0.117857
0.344156
0.53886
0.133333
0.430519
0.435185
0.54
0.355469
0.412
0.210721
40.141245
Qwen3ForCausalLM
dfd02e6271a58375dfbf3ece0175277cf6b6a89a
apache-2.0
26
5,781
8.190735
true
true
Qwen/Qwen3-8B-Base
2025-08-09T14:38:52.000Z
transformers
text-generation
False
Intelligent-Internet/II-Medical-8B
0.454289
0.536775
0.589134
0.864
0.614973
0.628
0.58
0.612
0.396
0.684
0.604
0.624
0.72
0.76
0.708
0.436
0.60274
0.6
0.776
0.58
0.578652
0.716
0.828
0.2
0.132
0.376
0.528
0.456949
0.348993
0.348485
0.35348
0.34375
0.41247
0.269871
0.41247
0.703583
0.463415
0.310606
0.456949
0.185714
0.474026
0.678756
0.259259
0.454289
0.433862
0.5
0.273438
0.532
0.269871
44.92828
Qwen3ForCausalLM
545fa0238261e041fb1ef3f6ed644a5a8f8400e3
apache-2.0
204
1,376
8.190735
true
true
Qwen/Qwen3-8B-Base
2025-08-12T07:48:17.000Z
transformers
text-generation
False
Vikhrmodels/QVikhr-3-8B-Instruction
0.477809
0.555066
0.610831
0.9
0.604278
0.68
0.56
0.68
0.544
0.78
0.636
0.616
0.84
0.724
0.68
0.484
0.589041
0.716
0.724
0.6
0.578652
0.716
0.664
0.224
0.18
0.376
0.544
0.504532
0.366611
0.434343
0.349817
0.357143
0.484412
0.336414
0.484412
0.749186
0.479675
0.371212
0.504532
0.260714
0.584416
0.668394
0.281481
0.477809
0.427249
0.512
0.289063
0.484
0.336414
47.857404
Qwen3ForCausalLM
41fcbccb2804cd696858e65e9b922e5462572042
apache-2.0
9
101
8.190735
true
true
Qwen/Qwen3-8B-Base
2025-08-06T10:27:18.000Z
transformers
text-generation
False
PrimeIntellect/Qwen3-8B
0.476978
0.553509
0.607186
0.9
0.59893
0.676
0.556
0.68
0.54
0.784
0.632
0.62
0.836
0.72
0.68
0.48
0.60274
0.72
0.72
0.588
0.561798
0.712
0.66
0.22
0.172
0.364
0.532
0.506798
0.369128
0.444444
0.355311
0.352679
0.482014
0.35305
0.482014
0.781759
0.487805
0.325758
0.506798
0.225
0.616883
0.683938
0.281481
0.476978
0.435185
0.524
0.300781
0.484
0.35305
47.954816
Qwen3ForCausalLM
e632d2027c80d3b93d91952532af6655ba3fc3f2
apache-2.0
0
571
8.190735
true
true
Qwen/Qwen3-8B-Base
2025-09-24T16:05:31.000Z
transformers
text-generation
False
willcb/Qwen3-8B
0.476978
0.553509
0.607186
0.9
0.59893
0.676
0.556
0.68
0.54
0.784
0.632
0.62
0.836
0.72
0.68
0.48
0.60274
0.72
0.72
0.588
0.561798
0.712
0.66
0.22
0.172
0.364
0.532
0.506798
0.369128
0.444444
0.355311
0.352679
0.482014
0.35305
0.482014
0.781759
0.487805
0.325758
0.506798
0.225
0.616883
0.683938
0.281481
0.476978
0.435185
0.524
0.300781
0.484
0.35305
47.954816
Qwen3ForCausalLM
3958718b90f176e7e3315e285ea2a6bddea0abc1
null
2
3,857
8.190735
true
true
Qwen/Qwen3-8B-Base
2025-06-06T20:46:12.000Z
transformers
text-generation
False
Goekdeniz-Guelmez/Josiefied-Qwen3-8B-abliterated-v1
0.473404
0.550396
0.60493
0.908
0.609626
0.68
0.556
0.676
0.54
0.776
0.636
0.616
0.836
0.724
0.672
0.5
0.568493
0.728
0.7
0.596
0.601124
0.668
0.656
0.204
0.184
0.352
0.516
0.522659
0.35906
0.419192
0.346154
0.348214
0.515588
0.375231
0.515588
0.801303
0.520325
0.356061
0.522659
0.239286
0.636364
0.715026
0.237037
0.473404
0.436508
0.512
0.308594
0.492
0.375231
48.535807
Qwen3ForCausalLM
2d21f18902f23a729934a9631373297e22e1ff25
null
205
1,902
8.190735
true
true
Qwen/Qwen3-8B-Base
2025-08-11T15:17:47.000Z
transformers
text-generation
False
AI-MO/Kimina-Prover-Distill-8B
0.414977
0.476845
0.515362
0.86
0.582888
0.588
0.616
0.648
0.272
0.588
0.5
0.456
0.712
0.668
0.624
0.432
0.465753
0.452
0.64
0.452
0.505618
0.712
0.364
0.168
0.136
0.368
0.552
0.310423
0.333893
0.378788
0.336996
0.310268
0.33693
0.192237
0.33693
0.557003
0.243902
0.181818
0.310423
0.092857
0.331169
0.466321
0.140741
0.414977
0.40873
0.5
0.332031
0.396
0.192237
38.671914
Qwen3ForCausalLM
74d328a7b1f001ab4871812582fc66d9bf70c68b
apache-2.0
8
19,966
8.190735
true
true
Qwen/Qwen3-8B-Base
2025-07-10T11:32:32.000Z
transformers
text-generation
False
unsloth/DeepSeek-R1-0528-Qwen3-8B-bnb-4bit
0.400432
0.453496
0.475612
0.776
0.545455
0.5
0.432
0.528
0.348
0.532
0.396
0.48
0.636
0.608
0.68
0.4
0.486301
0.472
0.408
0.424
0.455056
0.684
0.368
0.18
0.156
0.384
0.552
0.208459
0.357383
0.328283
0.3663
0.359375
0.395683
0.255083
0.395683
0.403909
0.146341
0.068182
0.208459
0.046429
0.149351
0.404145
0.081481
0.400432
0.436508
0.532
0.355469
0.424
0.255083
37.901287
Qwen3ForCausalLM
6115346c3f9243e44d9fcf2737ff78ec59d535b2
mit
10
483
8.40837
true
true
Qwen/Qwen3-8B-Base
2025-06-10T05:36:23.000Z
transformers
text-generation
False
legmlai/legml-v1.0-8b-instruct
0.481799
0.550785
0.609269
0.904
0.609626
0.7
0.648
0.676
0.54
0.676
0.62
0.628
0.848
0.736
0.632
0.532
0.561644
0.74
0.792
0.608
0.544944
0.7
0.548
0.22
0.192
0.376
0.552
0.403323
0.358221
0.40404
0.355311
0.341518
0.630695
0.499076
0.630695
0.65798
0.349593
0.227273
0.403323
0.142857
0.493506
0.601036
0.2
0.481799
0.40873
0.496
0.308594
0.424
0.499076
48.200635
Qwen3ForCausalLM
5cc8024fe4bfb4a5ed8d6abe0655cf93f8639457
apache-2.0
2
26
8.190735
true
true
Qwen/Qwen3-8B-Base
2025-07-30T09:43:51.000Z
transformers
text-generation
False
Intelligent-Internet/II-Medical-8B-1706
0.464511
0.509923
0.547995
0.872
0.588235
0.552
0.616
0.592
0.416
0.584
0.604
0.58
0.732
0.584
0.724
0.456
0.547945
0.624
0.612
0.524
0.55618
0.656
0.48
0.204
0.188
0.34
0.532
0.412387
0.350671
0.338384
0.338828
0.370536
0.438849
0.297597
0.438849
0.615635
0.357724
0.25
0.412387
0.178571
0.538961
0.606218
0.222222
0.464511
0.470899
0.56
0.34375
0.512
0.297597
44.755211
Qwen3ForCausalLM
a364a7cb987287fad5fefd512da3042e464d74f2
apache-2.0
138
320
8.190735
true
true
Qwen/Qwen3-8B-Base
2025-08-12T07:47:53.000Z
transformers
text-generation
False
DavidAU/Qwen3-8B-64k-Context-2X-Josiefied-Uncensored
0.474235
0.542872
0.600937
0.92
0.59893
0.664
0.544
0.68
0.54
0.768
0.608
0.632
0.852
0.692
0.664
0.48
0.547945
0.708
0.664
0.608
0.629213
0.688
0.692
0.208
0.188
0.292
0.54
0.537009
0.342282
0.368687
0.335165
0.339286
0.495204
0.354898
0.495204
0.807818
0.487805
0.295455
0.537009
0.260714
0.701299
0.73057
0.311111
0.474235
0.416667
0.492
0.300781
0.46
0.354898
47.772236
Qwen3ForCausalLM
b2bea2419c4f85c35f382e56ff6832843aedbbf8
null
12
86
8.190735
true
true
Qwen/Qwen3-8B-Base
2025-07-28T00:10:35.000Z
transformers
text-generation
False
unsloth/Qwen3-8B-Base-bnb-4bit
0.440076
0.522117
0.568478
0.868
0.582888
0.664
0.552
0.576
0.388
0.804
0.616
0.548
0.88
0.776
0.724
0.38
0.616438
0.62
0.66
0.528
0.696629
0.596
0.516
0.164
0.184
0.276
0.488
0.176737
0.345638
0.313131
0.351648
0.352679
0.508393
0.382625
0.508393
0.315961
0.121951
0.045455
0.176737
0.039286
0.227273
0.310881
0.074074
0.440076
0.44709
0.516
0.414063
0.412
0.382625
41.440202
Qwen3ForCausalLM
e3276524f7f4ce7df7b65dd9fcb7552a7f2d6de9
apache-2.0
3
1,538
null
true
false
Qwen/Qwen3-8B-Base
2025-07-14T13:04:35.000Z
transformers
text-generation
False
End of preview. Expand in Data Studio

YAML Metadata Warning:empty or missing yaml metadata in repo card

Check out the documentation for more information.

Proteus-2k

Proteus-2k is a large-scale benchmark table of recent open-weight language models evaluated with the Open LLM Leaderboard v2 pipeline. It was built to extend public leaderboards after freeze dates and to support research on how compute–capability relationships hold up as model families and post-training evolve.


Overview

Proteus-2k is the language-model evaluation dataset constructed and open-sourced in the companion paper. It is meant to fill gaps when public leaderboards (for example the Open LLM Leaderboard) stop updating or fall behind, and to test whether capability boundaries grounded in compute stay valid and stable over time.

At a glance:

  • Scale and timeframe: Roughly 2.4k recently released open-weight models, from 2024-04-30 to 2026-03-12.
  • Model families: Many frontier architectures that rarely appeared on older leaderboard tables—for example Qwen3, Gemma 3, and GPT-OSS, plus (per the paper appendix) Llama 3.2, Mistral-7B-v0.3, Allen AI’s OLMo-3, and NVIDIA’s Nemotron, among others.
  • Evaluation: The authors evaluated all ~2.4k models and followed the Open LLM Leaderboard evaluation pipeline exactly so results stay aligned with historical runs.
  • Research use: The paper studies upper bounds on performance at a given compute budget. Proteus-2k is a temporal out-of-distribution (OOD) benchmark: fit an envelope on older models, then check whether the latest models still respect compute-based limits as techniques and post-training evolve.
  • Release: Full table and subset on Hugging Face: hlzhang109/proteus-2k, hlzhang109/proteus-selected.

In short, Proteus-2k is a large-scale, up-to-date open-model table with standardized leaderboard evaluations, aimed at tracing how compute maps to real downstream performance as the field moves forward.


Files in this directory

File Role
proteus_2k.csv Full table (~2.4k rows); uploaded to hlzhang109/proteus-2k.
proteus_2k_selected.csv Selected subset; uploaded to hlzhang109/proteus-selected.
upload_data.py Helper to push these CSVs to Hugging Face (huggingface_hub).

Schema

Each row is one Hugging Face model repo (or checkpoint) identified by model_id.

Scores — columns prefixed with leaderboard_ are per-task or aggregate metrics from the Open LLM Leaderboard v2 suite (e.g. BBH subtasks, GPQA, IFEval, MATH-hard, MMLU-Pro, MUSR, exact-match aggregates). The column Average ⬆️ is the leaderboard-style overall average used for ranking.

Hub metadata (examples): Architecture, Model sha, Hub License, Hub downloads, #Params (B), Available on the hub, Chat Template, Base Model, Hub lastModified, library_name, pipeline_tag, gated.

Task names follow the leaderboard’s naming; see Open LLM Leaderboard documentation for benchmark definitions.


Loading with Hugging Face Datasets

from datasets import load_dataset

ds = load_dataset("hlzhang109/proteus-2k", data_files="proteus_2k.csv")
# or the selected split/repo:
# ds = load_dataset("hlzhang109/proteus-selected", data_files="proteus_2k_selected.csv")
df = ds["train"].to_pandas()

You can also download the CSV from the dataset repo’s Files tab and work with pandas / any CSV tooling.


Citations and licenses

If you use Proteus-2k, please cite the paper:

@misc{zhang2026prescriptive,
      title={Prescriptive Scaling Reveals the Evolution of Language Model Capabilities},
      author={Hanlin Zhang and Jikai Jin and Vasilis Syrgkanis and Sham Kakade},
      year={2026},
      eprint={2602.15327},
      archivePrefix={arXiv},
      primaryClass={cs.LG},
      url={https://arxiv.org/abs/2602.15327},
}

Per-model licensing varies (Hub License column and each model card on the Hub). The dataset aggregates publicly reported evaluation numbers and metadata; it does not redistribute model weights.

Downloads last month
19

Paper for hlzhang109/proteus-selected