results
dict | versions
dict | config_general
dict |
---|---|---|
{
"daily": {
"daily": 7
},
"quarterly": {
"quarterly": 7
},
"harness|arc_challenge|25": {
"acc": 0.29948805460750855,
"acc_stderr": 0.013385021637313567,
"acc_norm": 0.3506825938566553,
"acc_norm_stderr": 0.013944635930726089
},
"harness|hellaswag|10": {
"acc": 0.3333001394144593,
"acc_stderr": 0.004704293898729902,
"acc_norm": 0.4137621987651862,
"acc_norm_stderr": 0.004915003499517831
},
"harness|mmlu_world_religions|5": {
"acc": 0.47953216374269003,
"acc_stderr": 0.0383161053282193,
"acc_norm": 0.47953216374269003,
"acc_norm_stderr": 0.0383161053282193
},
"harness|mmlu_management|5": {
"acc": 0.5631067961165048,
"acc_stderr": 0.049111471073657764,
"acc_norm": 0.5631067961165048,
"acc_norm_stderr": 0.049111471073657764
},
"harness|mmlu_miscellaneous|5": {
"acc": 0.47509578544061304,
"acc_stderr": 0.01785777070490102,
"acc_norm": 0.47509578544061304,
"acc_norm_stderr": 0.01785777070490102
},
"harness|mmlu_anatomy|5": {
"acc": 0.28888888888888886,
"acc_stderr": 0.0391545063041425,
"acc_norm": 0.28888888888888886,
"acc_norm_stderr": 0.0391545063041425
},
"harness|mmlu_abstract_algebra|5": {
"acc": 0.31,
"acc_stderr": 0.04648231987117316,
"acc_norm": 0.31,
"acc_norm_stderr": 0.04648231987117316
},
"harness|mmlu_conceptual_physics|5": {
"acc": 0.46808510638297873,
"acc_stderr": 0.03261936918467382,
"acc_norm": 0.46808510638297873,
"acc_norm_stderr": 0.03261936918467382
},
"harness|mmlu_virology|5": {
"acc": 0.45180722891566266,
"acc_stderr": 0.03874371556587953,
"acc_norm": 0.45180722891566266,
"acc_norm_stderr": 0.03874371556587953
},
"harness|mmlu_philosophy|5": {
"acc": 0.47266881028938906,
"acc_stderr": 0.028355633568328188,
"acc_norm": 0.47266881028938906,
"acc_norm_stderr": 0.028355633568328188
},
"harness|mmlu_human_aging|5": {
"acc": 0.45739910313901344,
"acc_stderr": 0.033435777055830646,
"acc_norm": 0.45739910313901344,
"acc_norm_stderr": 0.033435777055830646
},
"harness|mmlu_human_sexuality|5": {
"acc": 0.5267175572519084,
"acc_stderr": 0.04379024936553894,
"acc_norm": 0.5267175572519084,
"acc_norm_stderr": 0.04379024936553894
},
"harness|mmlu_medical_genetics|5": {
"acc": 0.39,
"acc_stderr": 0.04902071300001975,
"acc_norm": 0.39,
"acc_norm_stderr": 0.04902071300001975
},
"harness|mmlu_high_school_geography|5": {
"acc": 0.5555555555555556,
"acc_stderr": 0.035402943770953675,
"acc_norm": 0.5555555555555556,
"acc_norm_stderr": 0.035402943770953675
},
"harness|mmlu_electrical_engineering|5": {
"acc": 0.5724137931034483,
"acc_stderr": 0.04122737111370332,
"acc_norm": 0.5724137931034483,
"acc_norm_stderr": 0.04122737111370332
},
"harness|mmlu_college_physics|5": {
"acc": 0.3137254901960784,
"acc_stderr": 0.04617034827006716,
"acc_norm": 0.3137254901960784,
"acc_norm_stderr": 0.04617034827006716
},
"harness|mmlu_high_school_microeconomics|5": {
"acc": 0.5,
"acc_stderr": 0.032478490123081544,
"acc_norm": 0.5,
"acc_norm_stderr": 0.032478490123081544
},
"harness|mmlu_high_school_macroeconomics|5": {
"acc": 0.47692307692307695,
"acc_stderr": 0.025323990861736125,
"acc_norm": 0.47692307692307695,
"acc_norm_stderr": 0.025323990861736125
},
"harness|mmlu_computer_security|5": {
"acc": 0.53,
"acc_stderr": 0.05016135580465919,
"acc_norm": 0.53,
"acc_norm_stderr": 0.05016135580465919
},
"harness|mmlu_global_facts|5": {
"acc": 0.37,
"acc_stderr": 0.048523658709391,
"acc_norm": 0.37,
"acc_norm_stderr": 0.048523658709391
},
"harness|mmlu_jurisprudence|5": {
"acc": 0.5740740740740741,
"acc_stderr": 0.047803436269367894,
"acc_norm": 0.5740740740740741,
"acc_norm_stderr": 0.047803436269367894
},
"harness|mmlu_high_school_chemistry|5": {
"acc": 0.4187192118226601,
"acc_stderr": 0.03471192860518468,
"acc_norm": 0.4187192118226601,
"acc_norm_stderr": 0.03471192860518468
},
"harness|mmlu_high_school_biology|5": {
"acc": 0.47419354838709676,
"acc_stderr": 0.02840609505765332,
"acc_norm": 0.47419354838709676,
"acc_norm_stderr": 0.02840609505765332
},
"harness|mmlu_marketing|5": {
"acc": 0.6752136752136753,
"acc_stderr": 0.03067902276549883,
"acc_norm": 0.6752136752136753,
"acc_norm_stderr": 0.03067902276549883
},
"harness|mmlu_clinical_knowledge|5": {
"acc": 0.44150943396226416,
"acc_stderr": 0.030561590426731833,
"acc_norm": 0.44150943396226416,
"acc_norm_stderr": 0.030561590426731833
},
"harness|mmlu_public_relations|5": {
"acc": 0.4727272727272727,
"acc_stderr": 0.04782001791380063,
"acc_norm": 0.4727272727272727,
"acc_norm_stderr": 0.04782001791380063
},
"harness|mmlu_high_school_mathematics|5": {
"acc": 0.4185185185185185,
"acc_stderr": 0.030078013075022066,
"acc_norm": 0.4185185185185185,
"acc_norm_stderr": 0.030078013075022066
},
"harness|mmlu_high_school_physics|5": {
"acc": 0.304635761589404,
"acc_stderr": 0.03757949922943343,
"acc_norm": 0.304635761589404,
"acc_norm_stderr": 0.03757949922943343
},
"harness|mmlu_sociology|5": {
"acc": 0.6069651741293532,
"acc_stderr": 0.0345368246603156,
"acc_norm": 0.6069651741293532,
"acc_norm_stderr": 0.0345368246603156
},
"harness|mmlu_college_medicine|5": {
"acc": 0.4046242774566474,
"acc_stderr": 0.03742461193887248,
"acc_norm": 0.4046242774566474,
"acc_norm_stderr": 0.03742461193887248
},
"harness|mmlu_elementary_mathematics|5": {
"acc": 0.5476190476190477,
"acc_stderr": 0.02563425811555495,
"acc_norm": 0.5476190476190477,
"acc_norm_stderr": 0.02563425811555495
},
"harness|mmlu_college_biology|5": {
"acc": 0.3472222222222222,
"acc_stderr": 0.039812405437178615,
"acc_norm": 0.3472222222222222,
"acc_norm_stderr": 0.039812405437178615
},
"harness|mmlu_college_chemistry|5": {
"acc": 0.33,
"acc_stderr": 0.04725815626252605,
"acc_norm": 0.33,
"acc_norm_stderr": 0.04725815626252605
},
"harness|mmlu_us_foreign_policy|5": {
"acc": 0.57,
"acc_stderr": 0.04975698519562426,
"acc_norm": 0.57,
"acc_norm_stderr": 0.04975698519562426
},
"harness|mmlu_moral_disputes|5": {
"acc": 0.49710982658959535,
"acc_stderr": 0.026918645383239015,
"acc_norm": 0.49710982658959535,
"acc_norm_stderr": 0.026918645383239015
},
"harness|mmlu_logical_fallacies|5": {
"acc": 0.5276073619631901,
"acc_stderr": 0.03922378290610991,
"acc_norm": 0.5276073619631901,
"acc_norm_stderr": 0.03922378290610991
},
"harness|mmlu_prehistory|5": {
"acc": 0.49691358024691357,
"acc_stderr": 0.027820214158594377,
"acc_norm": 0.49691358024691357,
"acc_norm_stderr": 0.027820214158594377
},
"harness|mmlu_college_mathematics|5": {
"acc": 0.45,
"acc_stderr": 0.05,
"acc_norm": 0.45,
"acc_norm_stderr": 0.05
},
"harness|mmlu_high_school_government_and_politics|5": {
"acc": 0.49222797927461137,
"acc_stderr": 0.03608003225569654,
"acc_norm": 0.49222797927461137,
"acc_norm_stderr": 0.03608003225569654
},
"harness|mmlu_econometrics|5": {
"acc": 0.41228070175438597,
"acc_stderr": 0.046306532033665956,
"acc_norm": 0.41228070175438597,
"acc_norm_stderr": 0.046306532033665956
},
"harness|mmlu_high_school_psychology|5": {
"acc": 0.5027522935779817,
"acc_stderr": 0.02143699835976532,
"acc_norm": 0.5027522935779817,
"acc_norm_stderr": 0.02143699835976532
},
"harness|mmlu_formal_logic|5": {
"acc": 0.40476190476190477,
"acc_stderr": 0.04390259265377561,
"acc_norm": 0.40476190476190477,
"acc_norm_stderr": 0.04390259265377561
},
"harness|mmlu_nutrition|5": {
"acc": 0.49019607843137253,
"acc_stderr": 0.028624412550167958,
"acc_norm": 0.49019607843137253,
"acc_norm_stderr": 0.028624412550167958
},
"harness|mmlu_business_ethics|5": {
"acc": 0.5,
"acc_stderr": 0.050251890762960605,
"acc_norm": 0.5,
"acc_norm_stderr": 0.050251890762960605
},
"harness|mmlu_international_law|5": {
"acc": 0.7355371900826446,
"acc_stderr": 0.04026187527591205,
"acc_norm": 0.7355371900826446,
"acc_norm_stderr": 0.04026187527591205
},
"harness|mmlu_astronomy|5": {
"acc": 0.45394736842105265,
"acc_stderr": 0.04051646342874142,
"acc_norm": 0.45394736842105265,
"acc_norm_stderr": 0.04051646342874142
},
"harness|mmlu_professional_psychology|5": {
"acc": 0.39705882352941174,
"acc_stderr": 0.019794488900024113,
"acc_norm": 0.39705882352941174,
"acc_norm_stderr": 0.019794488900024113
},
"harness|mmlu_professional_accounting|5": {
"acc": 0.40070921985815605,
"acc_stderr": 0.029233465745573086,
"acc_norm": 0.40070921985815605,
"acc_norm_stderr": 0.029233465745573086
},
"harness|mmlu_machine_learning|5": {
"acc": 0.39285714285714285,
"acc_stderr": 0.04635550135609976,
"acc_norm": 0.39285714285714285,
"acc_norm_stderr": 0.04635550135609976
},
"harness|mmlu_high_school_statistics|5": {
"acc": 0.4675925925925926,
"acc_stderr": 0.034028015813589656,
"acc_norm": 0.4675925925925926,
"acc_norm_stderr": 0.034028015813589656
},
"harness|mmlu_moral_scenarios|5": {
"acc": 0.3329608938547486,
"acc_stderr": 0.015761716178397552,
"acc_norm": 0.3329608938547486,
"acc_norm_stderr": 0.015761716178397552
},
"harness|mmlu_college_computer_science|5": {
"acc": 0.43,
"acc_stderr": 0.049756985195624284,
"acc_norm": 0.43,
"acc_norm_stderr": 0.049756985195624284
},
"harness|mmlu_high_school_computer_science|5": {
"acc": 0.76,
"acc_stderr": 0.042923469599092816,
"acc_norm": 0.76,
"acc_norm_stderr": 0.042923469599092816
},
"harness|mmlu_professional_medicine|5": {
"acc": 0.35294117647058826,
"acc_stderr": 0.029029422815681404,
"acc_norm": 0.35294117647058826,
"acc_norm_stderr": 0.029029422815681404
},
"harness|mmlu_security_studies|5": {
"acc": 0.6163265306122448,
"acc_stderr": 0.031130880396235943,
"acc_norm": 0.6163265306122448,
"acc_norm_stderr": 0.031130880396235943
},
"harness|mmlu_high_school_world_history|5": {
"acc": 0.5654008438818565,
"acc_stderr": 0.03226759995510145,
"acc_norm": 0.5654008438818565,
"acc_norm_stderr": 0.03226759995510145
},
"harness|mmlu_professional_law|5": {
"acc": 0.36571056062581486,
"acc_stderr": 0.012301028188840567,
"acc_norm": 0.36571056062581486,
"acc_norm_stderr": 0.012301028188840567
},
"harness|mmlu_high_school_us_history|5": {
"acc": 0.4852941176470588,
"acc_stderr": 0.03507793834791324,
"acc_norm": 0.4852941176470588,
"acc_norm_stderr": 0.03507793834791324
},
"harness|mmlu_high_school_european_history|5": {
"acc": 0.5151515151515151,
"acc_stderr": 0.03902551007374448,
"acc_norm": 0.5151515151515151,
"acc_norm_stderr": 0.03902551007374448
},
"harness|truthfulqa_mc|0": {
"mc1": 0.2937576499388005,
"mc1_stderr": 0.015945068581236614,
"mc2": 0.4670848140389129,
"mc2_stderr": 0.01585178282587417
}
} | {
"all": 0,
"harness|arc_challenge|25": 0,
"harness|hellaswag|10": 0,
"harness|mmlu_world_religions|5": 1,
"harness|mmlu_management|5": 1,
"harness|mmlu_miscellaneous|5": 1,
"harness|mmlu_anatomy|5": 1,
"harness|mmlu_abstract_algebra|5": 1,
"harness|mmlu_conceptual_physics|5": 1,
"harness|mmlu_virology|5": 1,
"harness|mmlu_philosophy|5": 1,
"harness|mmlu_human_aging|5": 1,
"harness|mmlu_human_sexuality|5": 1,
"harness|mmlu_medical_genetics|5": 1,
"harness|mmlu_high_school_geography|5": 1,
"harness|mmlu_electrical_engineering|5": 1,
"harness|mmlu_college_physics|5": 1,
"harness|mmlu_high_school_microeconomics|5": 1,
"harness|mmlu_high_school_macroeconomics|5": 1,
"harness|mmlu_computer_security|5": 1,
"harness|mmlu_global_facts|5": 1,
"harness|mmlu_jurisprudence|5": 1,
"harness|mmlu_high_school_chemistry|5": 1,
"harness|mmlu_high_school_biology|5": 1,
"harness|mmlu_marketing|5": 1,
"harness|mmlu_clinical_knowledge|5": 1,
"harness|mmlu_public_relations|5": 1,
"harness|mmlu_high_school_mathematics|5": 1,
"harness|mmlu_high_school_physics|5": 1,
"harness|mmlu_sociology|5": 1,
"harness|mmlu_college_medicine|5": 1,
"harness|mmlu_elementary_mathematics|5": 1,
"harness|mmlu_college_biology|5": 1,
"harness|mmlu_college_chemistry|5": 1,
"harness|mmlu_us_foreign_policy|5": 1,
"harness|mmlu_moral_disputes|5": 1,
"harness|mmlu_logical_fallacies|5": 1,
"harness|mmlu_prehistory|5": 1,
"harness|mmlu_college_mathematics|5": 1,
"harness|mmlu_high_school_government_and_politics|5": 1,
"harness|mmlu_econometrics|5": 1,
"harness|mmlu_high_school_psychology|5": 1,
"harness|mmlu_formal_logic|5": 1,
"harness|mmlu_nutrition|5": 1,
"harness|mmlu_business_ethics|5": 1,
"harness|mmlu_international_law|5": 1,
"harness|mmlu_astronomy|5": 1,
"harness|mmlu_professional_psychology|5": 1,
"harness|mmlu_professional_accounting|5": 1,
"harness|mmlu_machine_learning|5": 1,
"harness|mmlu_high_school_statistics|5": 1,
"harness|mmlu_moral_scenarios|5": 1,
"harness|mmlu_college_computer_science|5": 1,
"harness|mmlu_high_school_computer_science|5": 1,
"harness|mmlu_professional_medicine|5": 1,
"harness|mmlu_security_studies|5": 1,
"harness|mmlu_high_school_world_history|5": 1,
"harness|mmlu_professional_law|5": 1,
"harness|mmlu_high_school_us_history|5": 1,
"harness|mmlu_high_school_european_history|5": 1,
"harness|truthfulqa_mc|0": 0
} | {
"model_name": "01-ai/Yi-1.5-9B-32K",
"model_sha": "c0239dbc923b8a2b5ca849763bdd592d39c60850",
"model_dtype": "torch.float16",
"lighteval_sha": "",
"num_few_shot_default": 0,
"num_fewshot_seeds": 1,
"override_batch_size": 1,
"max_samples": null
} |
{
"daily": {
"daily": 9
},
"quarterly": {
"quarterly": 9
},
"harness|arc_challenge|25": {
"acc": 0.257679180887372,
"acc_stderr": 0.012780770562768416,
"acc_norm": 0.3122866894197952,
"acc_norm_stderr": 0.013542598541688065
},
"harness|hellaswag|10": {
"acc": 0.3229436367257518,
"acc_stderr": 0.004666457279979418,
"acc_norm": 0.39255128460466043,
"acc_norm_stderr": 0.004873203269366306
},
"harness|mmlu_world_religions|5": {
"acc": 0.34502923976608185,
"acc_stderr": 0.036459813773888065,
"acc_norm": 0.34502923976608185,
"acc_norm_stderr": 0.036459813773888065
},
"harness|mmlu_management|5": {
"acc": 0.4368932038834951,
"acc_stderr": 0.04911147107365778,
"acc_norm": 0.4368932038834951,
"acc_norm_stderr": 0.04911147107365778
},
"harness|mmlu_miscellaneous|5": {
"acc": 0.3780332056194125,
"acc_stderr": 0.017339844462104625,
"acc_norm": 0.3780332056194125,
"acc_norm_stderr": 0.017339844462104625
},
"harness|mmlu_anatomy|5": {
"acc": 0.3037037037037037,
"acc_stderr": 0.039725528847851355,
"acc_norm": 0.3037037037037037,
"acc_norm_stderr": 0.039725528847851355
},
"harness|mmlu_abstract_algebra|5": {
"acc": 0.37,
"acc_stderr": 0.04852365870939099,
"acc_norm": 0.37,
"acc_norm_stderr": 0.04852365870939099
},
"harness|mmlu_conceptual_physics|5": {
"acc": 0.28085106382978725,
"acc_stderr": 0.02937917046412482,
"acc_norm": 0.28085106382978725,
"acc_norm_stderr": 0.02937917046412482
},
"harness|mmlu_virology|5": {
"acc": 0.3373493975903614,
"acc_stderr": 0.03680783690727581,
"acc_norm": 0.3373493975903614,
"acc_norm_stderr": 0.03680783690727581
},
"harness|mmlu_philosophy|5": {
"acc": 0.3954983922829582,
"acc_stderr": 0.027770918531427838,
"acc_norm": 0.3954983922829582,
"acc_norm_stderr": 0.027770918531427838
},
"harness|mmlu_human_aging|5": {
"acc": 0.34977578475336324,
"acc_stderr": 0.03200736719484503,
"acc_norm": 0.34977578475336324,
"acc_norm_stderr": 0.03200736719484503
},
"harness|mmlu_human_sexuality|5": {
"acc": 0.3969465648854962,
"acc_stderr": 0.04291135671009224,
"acc_norm": 0.3969465648854962,
"acc_norm_stderr": 0.04291135671009224
},
"harness|mmlu_medical_genetics|5": {
"acc": 0.42,
"acc_stderr": 0.049604496374885836,
"acc_norm": 0.42,
"acc_norm_stderr": 0.049604496374885836
},
"harness|mmlu_high_school_geography|5": {
"acc": 0.4292929292929293,
"acc_stderr": 0.03526552724601199,
"acc_norm": 0.4292929292929293,
"acc_norm_stderr": 0.03526552724601199
},
"harness|mmlu_electrical_engineering|5": {
"acc": 0.4,
"acc_stderr": 0.04082482904638628,
"acc_norm": 0.4,
"acc_norm_stderr": 0.04082482904638628
},
"harness|mmlu_college_physics|5": {
"acc": 0.30392156862745096,
"acc_stderr": 0.045766654032077636,
"acc_norm": 0.30392156862745096,
"acc_norm_stderr": 0.045766654032077636
},
"harness|mmlu_high_school_microeconomics|5": {
"acc": 0.40336134453781514,
"acc_stderr": 0.031866081214088314,
"acc_norm": 0.40336134453781514,
"acc_norm_stderr": 0.031866081214088314
},
"harness|mmlu_high_school_macroeconomics|5": {
"acc": 0.40512820512820513,
"acc_stderr": 0.024890471769938145,
"acc_norm": 0.40512820512820513,
"acc_norm_stderr": 0.024890471769938145
},
"harness|mmlu_computer_security|5": {
"acc": 0.48,
"acc_stderr": 0.050211673156867795,
"acc_norm": 0.48,
"acc_norm_stderr": 0.050211673156867795
},
"harness|mmlu_global_facts|5": {
"acc": 0.32,
"acc_stderr": 0.04688261722621505,
"acc_norm": 0.32,
"acc_norm_stderr": 0.04688261722621505
},
"harness|mmlu_jurisprudence|5": {
"acc": 0.49074074074074076,
"acc_stderr": 0.04832853553437055,
"acc_norm": 0.49074074074074076,
"acc_norm_stderr": 0.04832853553437055
},
"harness|mmlu_high_school_chemistry|5": {
"acc": 0.37438423645320196,
"acc_stderr": 0.03405155380561952,
"acc_norm": 0.37438423645320196,
"acc_norm_stderr": 0.03405155380561952
},
"harness|mmlu_high_school_biology|5": {
"acc": 0.36774193548387096,
"acc_stderr": 0.027430866579973474,
"acc_norm": 0.36774193548387096,
"acc_norm_stderr": 0.027430866579973474
},
"harness|mmlu_marketing|5": {
"acc": 0.5598290598290598,
"acc_stderr": 0.0325207417206305,
"acc_norm": 0.5598290598290598,
"acc_norm_stderr": 0.0325207417206305
},
"harness|mmlu_clinical_knowledge|5": {
"acc": 0.3886792452830189,
"acc_stderr": 0.030000485448675986,
"acc_norm": 0.3886792452830189,
"acc_norm_stderr": 0.030000485448675986
},
"harness|mmlu_public_relations|5": {
"acc": 0.44545454545454544,
"acc_stderr": 0.047605488214603246,
"acc_norm": 0.44545454545454544,
"acc_norm_stderr": 0.047605488214603246
},
"harness|mmlu_high_school_mathematics|5": {
"acc": 0.34444444444444444,
"acc_stderr": 0.028972648884844267,
"acc_norm": 0.34444444444444444,
"acc_norm_stderr": 0.028972648884844267
},
"harness|mmlu_high_school_physics|5": {
"acc": 0.3443708609271523,
"acc_stderr": 0.038796870240733264,
"acc_norm": 0.3443708609271523,
"acc_norm_stderr": 0.038796870240733264
},
"harness|mmlu_sociology|5": {
"acc": 0.4577114427860697,
"acc_stderr": 0.035228658640995975,
"acc_norm": 0.4577114427860697,
"acc_norm_stderr": 0.035228658640995975
},
"harness|mmlu_college_medicine|5": {
"acc": 0.3815028901734104,
"acc_stderr": 0.03703851193099521,
"acc_norm": 0.3815028901734104,
"acc_norm_stderr": 0.03703851193099521
},
"harness|mmlu_elementary_mathematics|5": {
"acc": 0.35714285714285715,
"acc_stderr": 0.02467786284133278,
"acc_norm": 0.35714285714285715,
"acc_norm_stderr": 0.02467786284133278
},
"harness|mmlu_college_biology|5": {
"acc": 0.3333333333333333,
"acc_stderr": 0.03942082639927213,
"acc_norm": 0.3333333333333333,
"acc_norm_stderr": 0.03942082639927213
},
"harness|mmlu_college_chemistry|5": {
"acc": 0.47,
"acc_stderr": 0.05016135580465919,
"acc_norm": 0.47,
"acc_norm_stderr": 0.05016135580465919
},
"harness|mmlu_us_foreign_policy|5": {
"acc": 0.54,
"acc_stderr": 0.05009082659620333,
"acc_norm": 0.54,
"acc_norm_stderr": 0.05009082659620333
},
"harness|mmlu_moral_disputes|5": {
"acc": 0.44508670520231214,
"acc_stderr": 0.02675625512966377,
"acc_norm": 0.44508670520231214,
"acc_norm_stderr": 0.02675625512966377
},
"harness|mmlu_logical_fallacies|5": {
"acc": 0.34355828220858897,
"acc_stderr": 0.03731133519673893,
"acc_norm": 0.34355828220858897,
"acc_norm_stderr": 0.03731133519673893
},
"harness|mmlu_prehistory|5": {
"acc": 0.37037037037037035,
"acc_stderr": 0.02686949074481525,
"acc_norm": 0.37037037037037035,
"acc_norm_stderr": 0.02686949074481525
},
"harness|mmlu_college_mathematics|5": {
"acc": 0.33,
"acc_stderr": 0.04725815626252605,
"acc_norm": 0.33,
"acc_norm_stderr": 0.04725815626252605
},
"harness|mmlu_high_school_government_and_politics|5": {
"acc": 0.44559585492227977,
"acc_stderr": 0.0358701498607566,
"acc_norm": 0.44559585492227977,
"acc_norm_stderr": 0.0358701498607566
},
"harness|mmlu_econometrics|5": {
"acc": 0.2719298245614035,
"acc_stderr": 0.041857744240220575,
"acc_norm": 0.2719298245614035,
"acc_norm_stderr": 0.041857744240220575
},
"harness|mmlu_high_school_psychology|5": {
"acc": 0.3798165137614679,
"acc_stderr": 0.020808825617866244,
"acc_norm": 0.3798165137614679,
"acc_norm_stderr": 0.020808825617866244
},
"harness|mmlu_formal_logic|5": {
"acc": 0.3492063492063492,
"acc_stderr": 0.04263906892795132,
"acc_norm": 0.3492063492063492,
"acc_norm_stderr": 0.04263906892795132
},
"harness|mmlu_nutrition|5": {
"acc": 0.4117647058823529,
"acc_stderr": 0.02818059632825929,
"acc_norm": 0.4117647058823529,
"acc_norm_stderr": 0.02818059632825929
},
"harness|mmlu_business_ethics|5": {
"acc": 0.42,
"acc_stderr": 0.049604496374885836,
"acc_norm": 0.42,
"acc_norm_stderr": 0.049604496374885836
},
"harness|mmlu_international_law|5": {
"acc": 0.5619834710743802,
"acc_stderr": 0.045291468044357915,
"acc_norm": 0.5619834710743802,
"acc_norm_stderr": 0.045291468044357915
},
"harness|mmlu_astronomy|5": {
"acc": 0.34868421052631576,
"acc_stderr": 0.038781398887976125,
"acc_norm": 0.34868421052631576,
"acc_norm_stderr": 0.038781398887976125
},
"harness|mmlu_professional_psychology|5": {
"acc": 0.3284313725490196,
"acc_stderr": 0.018999707383162666,
"acc_norm": 0.3284313725490196,
"acc_norm_stderr": 0.018999707383162666
},
"harness|mmlu_professional_accounting|5": {
"acc": 0.2730496453900709,
"acc_stderr": 0.026577860943307857,
"acc_norm": 0.2730496453900709,
"acc_norm_stderr": 0.026577860943307857
},
"harness|mmlu_machine_learning|5": {
"acc": 0.2767857142857143,
"acc_stderr": 0.04246624336697627,
"acc_norm": 0.2767857142857143,
"acc_norm_stderr": 0.04246624336697627
},
"harness|mmlu_high_school_statistics|5": {
"acc": 0.4074074074074074,
"acc_stderr": 0.03350991604696043,
"acc_norm": 0.4074074074074074,
"acc_norm_stderr": 0.03350991604696043
},
"harness|mmlu_moral_scenarios|5": {
"acc": 0.23910614525139665,
"acc_stderr": 0.014265554192331149,
"acc_norm": 0.23910614525139665,
"acc_norm_stderr": 0.014265554192331149
},
"harness|mmlu_college_computer_science|5": {
"acc": 0.33,
"acc_stderr": 0.047258156262526045,
"acc_norm": 0.33,
"acc_norm_stderr": 0.047258156262526045
},
"harness|mmlu_high_school_computer_science|5": {
"acc": 0.4,
"acc_stderr": 0.04923659639173309,
"acc_norm": 0.4,
"acc_norm_stderr": 0.04923659639173309
},
"harness|mmlu_professional_medicine|5": {
"acc": 0.4227941176470588,
"acc_stderr": 0.030008562845003483,
"acc_norm": 0.4227941176470588,
"acc_norm_stderr": 0.030008562845003483
},
"harness|mmlu_security_studies|5": {
"acc": 0.3469387755102041,
"acc_stderr": 0.030472526026726492,
"acc_norm": 0.3469387755102041,
"acc_norm_stderr": 0.030472526026726492
},
"harness|mmlu_high_school_world_history|5": {
"acc": 0.4177215189873418,
"acc_stderr": 0.032103530322412685,
"acc_norm": 0.4177215189873418,
"acc_norm_stderr": 0.032103530322412685
},
"harness|mmlu_professional_law|5": {
"acc": 0.3005215123859192,
"acc_stderr": 0.011709918883039124,
"acc_norm": 0.3005215123859192,
"acc_norm_stderr": 0.011709918883039124
},
"harness|mmlu_high_school_us_history|5": {
"acc": 0.3872549019607843,
"acc_stderr": 0.03418931233833344,
"acc_norm": 0.3872549019607843,
"acc_norm_stderr": 0.03418931233833344
},
"harness|mmlu_high_school_european_history|5": {
"acc": 0.43636363636363634,
"acc_stderr": 0.03872592983524753,
"acc_norm": 0.43636363636363634,
"acc_norm_stderr": 0.03872592983524753
},
"harness|truthfulqa_mc|0": {
"mc1": 0.3072215422276622,
"mc1_stderr": 0.016150201321323002,
"mc2": 0.4721418472000992,
"mc2_stderr": 0.01626625866283201
}
} | {
"all": 0,
"harness|arc_challenge|25": 0,
"harness|hellaswag|10": 0,
"harness|mmlu_world_religions|5": 1,
"harness|mmlu_management|5": 1,
"harness|mmlu_miscellaneous|5": 1,
"harness|mmlu_anatomy|5": 1,
"harness|mmlu_abstract_algebra|5": 1,
"harness|mmlu_conceptual_physics|5": 1,
"harness|mmlu_virology|5": 1,
"harness|mmlu_philosophy|5": 1,
"harness|mmlu_human_aging|5": 1,
"harness|mmlu_human_sexuality|5": 1,
"harness|mmlu_medical_genetics|5": 1,
"harness|mmlu_high_school_geography|5": 1,
"harness|mmlu_electrical_engineering|5": 1,
"harness|mmlu_college_physics|5": 1,
"harness|mmlu_high_school_microeconomics|5": 1,
"harness|mmlu_high_school_macroeconomics|5": 1,
"harness|mmlu_computer_security|5": 1,
"harness|mmlu_global_facts|5": 1,
"harness|mmlu_jurisprudence|5": 1,
"harness|mmlu_high_school_chemistry|5": 1,
"harness|mmlu_high_school_biology|5": 1,
"harness|mmlu_marketing|5": 1,
"harness|mmlu_clinical_knowledge|5": 1,
"harness|mmlu_public_relations|5": 1,
"harness|mmlu_high_school_mathematics|5": 1,
"harness|mmlu_high_school_physics|5": 1,
"harness|mmlu_sociology|5": 1,
"harness|mmlu_college_medicine|5": 1,
"harness|mmlu_elementary_mathematics|5": 1,
"harness|mmlu_college_biology|5": 1,
"harness|mmlu_college_chemistry|5": 1,
"harness|mmlu_us_foreign_policy|5": 1,
"harness|mmlu_moral_disputes|5": 1,
"harness|mmlu_logical_fallacies|5": 1,
"harness|mmlu_prehistory|5": 1,
"harness|mmlu_college_mathematics|5": 1,
"harness|mmlu_high_school_government_and_politics|5": 1,
"harness|mmlu_econometrics|5": 1,
"harness|mmlu_high_school_psychology|5": 1,
"harness|mmlu_formal_logic|5": 1,
"harness|mmlu_nutrition|5": 1,
"harness|mmlu_business_ethics|5": 1,
"harness|mmlu_international_law|5": 1,
"harness|mmlu_astronomy|5": 1,
"harness|mmlu_professional_psychology|5": 1,
"harness|mmlu_professional_accounting|5": 1,
"harness|mmlu_machine_learning|5": 1,
"harness|mmlu_high_school_statistics|5": 1,
"harness|mmlu_moral_scenarios|5": 1,
"harness|mmlu_college_computer_science|5": 1,
"harness|mmlu_high_school_computer_science|5": 1,
"harness|mmlu_professional_medicine|5": 1,
"harness|mmlu_security_studies|5": 1,
"harness|mmlu_high_school_world_history|5": 1,
"harness|mmlu_professional_law|5": 1,
"harness|mmlu_high_school_us_history|5": 1,
"harness|mmlu_high_school_european_history|5": 1,
"harness|truthfulqa_mc|0": 0
} | {
"model_name": "BioMistral/BioMistral-7B",
"model_sha": "9a11e1ffa817c211cbb52ee1fb312dc6b61b40a5",
"model_dtype": "torch.float16",
"lighteval_sha": "",
"num_few_shot_default": 0,
"num_fewshot_seeds": 1,
"override_batch_size": 1,
"max_samples": null
} |
{
"daily": {
"daily": 10
},
"quarterly": {
"quarterly": 10
},
"harness|arc_challenge|25": {
"acc": 0.2235494880546075,
"acc_stderr": 0.012174896631202605,
"acc_norm": 0.2815699658703072,
"acc_norm_stderr": 0.013143376735009015
},
"harness|hellaswag|10": {
"acc": 0.3345947022505477,
"acc_stderr": 0.004708842600177431,
"acc_norm": 0.4135630352519418,
"acc_norm_stderr": 0.0049146550633294974
},
"harness|mmlu_world_religions|5": {
"acc": 0.27485380116959063,
"acc_stderr": 0.03424042924691585,
"acc_norm": 0.27485380116959063,
"acc_norm_stderr": 0.03424042924691585
},
"harness|mmlu_management|5": {
"acc": 0.27184466019417475,
"acc_stderr": 0.044052680241409216,
"acc_norm": 0.27184466019417475,
"acc_norm_stderr": 0.044052680241409216
},
"harness|mmlu_miscellaneous|5": {
"acc": 0.26947637292464877,
"acc_stderr": 0.015866243073215065,
"acc_norm": 0.26947637292464877,
"acc_norm_stderr": 0.015866243073215065
},
"harness|mmlu_anatomy|5": {
"acc": 0.26666666666666666,
"acc_stderr": 0.038201699145179055,
"acc_norm": 0.26666666666666666,
"acc_norm_stderr": 0.038201699145179055
},
"harness|mmlu_abstract_algebra|5": {
"acc": 0.3,
"acc_stderr": 0.046056618647183814,
"acc_norm": 0.3,
"acc_norm_stderr": 0.046056618647183814
},
"harness|mmlu_conceptual_physics|5": {
"acc": 0.2127659574468085,
"acc_stderr": 0.026754391348039783,
"acc_norm": 0.2127659574468085,
"acc_norm_stderr": 0.026754391348039783
},
"harness|mmlu_virology|5": {
"acc": 0.24096385542168675,
"acc_stderr": 0.033293941190735296,
"acc_norm": 0.24096385542168675,
"acc_norm_stderr": 0.033293941190735296
},
"harness|mmlu_philosophy|5": {
"acc": 0.2379421221864952,
"acc_stderr": 0.024185150647818707,
"acc_norm": 0.2379421221864952,
"acc_norm_stderr": 0.024185150647818707
},
"harness|mmlu_human_aging|5": {
"acc": 0.2825112107623318,
"acc_stderr": 0.030216831011508766,
"acc_norm": 0.2825112107623318,
"acc_norm_stderr": 0.030216831011508766
},
"harness|mmlu_human_sexuality|5": {
"acc": 0.21374045801526717,
"acc_stderr": 0.0359546161177469,
"acc_norm": 0.21374045801526717,
"acc_norm_stderr": 0.0359546161177469
},
"harness|mmlu_medical_genetics|5": {
"acc": 0.24,
"acc_stderr": 0.042923469599092816,
"acc_norm": 0.24,
"acc_norm_stderr": 0.042923469599092816
},
"harness|mmlu_high_school_geography|5": {
"acc": 0.2474747474747475,
"acc_stderr": 0.03074630074212451,
"acc_norm": 0.2474747474747475,
"acc_norm_stderr": 0.03074630074212451
},
"harness|mmlu_electrical_engineering|5": {
"acc": 0.22758620689655173,
"acc_stderr": 0.03493950380131184,
"acc_norm": 0.22758620689655173,
"acc_norm_stderr": 0.03493950380131184
},
"harness|mmlu_college_physics|5": {
"acc": 0.22549019607843138,
"acc_stderr": 0.041583075330832865,
"acc_norm": 0.22549019607843138,
"acc_norm_stderr": 0.041583075330832865
},
"harness|mmlu_high_school_microeconomics|5": {
"acc": 0.31512605042016806,
"acc_stderr": 0.030176808288974337,
"acc_norm": 0.31512605042016806,
"acc_norm_stderr": 0.030176808288974337
},
"harness|mmlu_high_school_macroeconomics|5": {
"acc": 0.2205128205128205,
"acc_stderr": 0.02102067268082791,
"acc_norm": 0.2205128205128205,
"acc_norm_stderr": 0.02102067268082791
},
"harness|mmlu_computer_security|5": {
"acc": 0.18,
"acc_stderr": 0.038612291966536955,
"acc_norm": 0.18,
"acc_norm_stderr": 0.038612291966536955
},
"harness|mmlu_global_facts|5": {
"acc": 0.31,
"acc_stderr": 0.04648231987117316,
"acc_norm": 0.31,
"acc_norm_stderr": 0.04648231987117316
},
"harness|mmlu_jurisprudence|5": {
"acc": 0.25,
"acc_stderr": 0.04186091791394607,
"acc_norm": 0.25,
"acc_norm_stderr": 0.04186091791394607
},
"harness|mmlu_high_school_chemistry|5": {
"acc": 0.2660098522167488,
"acc_stderr": 0.03108982600293752,
"acc_norm": 0.2660098522167488,
"acc_norm_stderr": 0.03108982600293752
},
"harness|mmlu_high_school_biology|5": {
"acc": 0.3,
"acc_stderr": 0.02606936229533513,
"acc_norm": 0.3,
"acc_norm_stderr": 0.02606936229533513
},
"harness|mmlu_marketing|5": {
"acc": 0.23076923076923078,
"acc_stderr": 0.027601921381417607,
"acc_norm": 0.23076923076923078,
"acc_norm_stderr": 0.027601921381417607
},
"harness|mmlu_clinical_knowledge|5": {
"acc": 0.25660377358490566,
"acc_stderr": 0.026880647889051968,
"acc_norm": 0.25660377358490566,
"acc_norm_stderr": 0.026880647889051968
},
"harness|mmlu_public_relations|5": {
"acc": 0.2545454545454545,
"acc_stderr": 0.04172343038705383,
"acc_norm": 0.2545454545454545,
"acc_norm_stderr": 0.04172343038705383
},
"harness|mmlu_high_school_mathematics|5": {
"acc": 0.2962962962962963,
"acc_stderr": 0.02784081149587194,
"acc_norm": 0.2962962962962963,
"acc_norm_stderr": 0.02784081149587194
},
"harness|mmlu_high_school_physics|5": {
"acc": 0.304635761589404,
"acc_stderr": 0.03757949922943342,
"acc_norm": 0.304635761589404,
"acc_norm_stderr": 0.03757949922943342
},
"harness|mmlu_sociology|5": {
"acc": 0.25870646766169153,
"acc_stderr": 0.03096590312357303,
"acc_norm": 0.25870646766169153,
"acc_norm_stderr": 0.03096590312357303
},
"harness|mmlu_college_medicine|5": {
"acc": 0.2254335260115607,
"acc_stderr": 0.03186209851641144,
"acc_norm": 0.2254335260115607,
"acc_norm_stderr": 0.03186209851641144
},
"harness|mmlu_elementary_mathematics|5": {
"acc": 0.2566137566137566,
"acc_stderr": 0.022494510767503154,
"acc_norm": 0.2566137566137566,
"acc_norm_stderr": 0.022494510767503154
},
"harness|mmlu_college_biology|5": {
"acc": 0.2638888888888889,
"acc_stderr": 0.03685651095897532,
"acc_norm": 0.2638888888888889,
"acc_norm_stderr": 0.03685651095897532
},
"harness|mmlu_college_chemistry|5": {
"acc": 0.23,
"acc_stderr": 0.04229525846816505,
"acc_norm": 0.23,
"acc_norm_stderr": 0.04229525846816505
},
"harness|mmlu_us_foreign_policy|5": {
"acc": 0.22,
"acc_stderr": 0.04163331998932269,
"acc_norm": 0.22,
"acc_norm_stderr": 0.04163331998932269
},
"harness|mmlu_moral_disputes|5": {
"acc": 0.24855491329479767,
"acc_stderr": 0.023267528432100174,
"acc_norm": 0.24855491329479767,
"acc_norm_stderr": 0.023267528432100174
},
"harness|mmlu_logical_fallacies|5": {
"acc": 0.31901840490797545,
"acc_stderr": 0.03661997551073836,
"acc_norm": 0.31901840490797545,
"acc_norm_stderr": 0.03661997551073836
},
"harness|mmlu_prehistory|5": {
"acc": 0.2623456790123457,
"acc_stderr": 0.024477222856135114,
"acc_norm": 0.2623456790123457,
"acc_norm_stderr": 0.024477222856135114
},
"harness|mmlu_college_mathematics|5": {
"acc": 0.25,
"acc_stderr": 0.04351941398892446,
"acc_norm": 0.25,
"acc_norm_stderr": 0.04351941398892446
},
"harness|mmlu_high_school_government_and_politics|5": {
"acc": 0.33678756476683935,
"acc_stderr": 0.03410780251836184,
"acc_norm": 0.33678756476683935,
"acc_norm_stderr": 0.03410780251836184
},
"harness|mmlu_econometrics|5": {
"acc": 0.20175438596491227,
"acc_stderr": 0.037752050135836386,
"acc_norm": 0.20175438596491227,
"acc_norm_stderr": 0.037752050135836386
},
"harness|mmlu_high_school_psychology|5": {
"acc": 0.24220183486238533,
"acc_stderr": 0.01836817630659862,
"acc_norm": 0.24220183486238533,
"acc_norm_stderr": 0.01836817630659862
},
"harness|mmlu_formal_logic|5": {
"acc": 0.23015873015873015,
"acc_stderr": 0.03764950879790606,
"acc_norm": 0.23015873015873015,
"acc_norm_stderr": 0.03764950879790606
},
"harness|mmlu_nutrition|5": {
"acc": 0.23529411764705882,
"acc_stderr": 0.024288619466046102,
"acc_norm": 0.23529411764705882,
"acc_norm_stderr": 0.024288619466046102
},
"harness|mmlu_business_ethics|5": {
"acc": 0.18,
"acc_stderr": 0.03861229196653695,
"acc_norm": 0.18,
"acc_norm_stderr": 0.03861229196653695
},
"harness|mmlu_international_law|5": {
"acc": 0.256198347107438,
"acc_stderr": 0.039849796533028704,
"acc_norm": 0.256198347107438,
"acc_norm_stderr": 0.039849796533028704
},
"harness|mmlu_astronomy|5": {
"acc": 0.21710526315789475,
"acc_stderr": 0.033550453048829226,
"acc_norm": 0.21710526315789475,
"acc_norm_stderr": 0.033550453048829226
},
"harness|mmlu_professional_psychology|5": {
"acc": 0.24019607843137256,
"acc_stderr": 0.01728276069516743,
"acc_norm": 0.24019607843137256,
"acc_norm_stderr": 0.01728276069516743
},
"harness|mmlu_professional_accounting|5": {
"acc": 0.2553191489361702,
"acc_stderr": 0.02601199293090201,
"acc_norm": 0.2553191489361702,
"acc_norm_stderr": 0.02601199293090201
},
"harness|mmlu_machine_learning|5": {
"acc": 0.21428571428571427,
"acc_stderr": 0.03894641120044793,
"acc_norm": 0.21428571428571427,
"acc_norm_stderr": 0.03894641120044793
},
"harness|mmlu_high_school_statistics|5": {
"acc": 0.46296296296296297,
"acc_stderr": 0.03400603625538272,
"acc_norm": 0.46296296296296297,
"acc_norm_stderr": 0.03400603625538272
},
"harness|mmlu_moral_scenarios|5": {
"acc": 0.24692737430167597,
"acc_stderr": 0.014422292204808852,
"acc_norm": 0.24692737430167597,
"acc_norm_stderr": 0.014422292204808852
},
"harness|mmlu_college_computer_science|5": {
"acc": 0.25,
"acc_stderr": 0.04351941398892446,
"acc_norm": 0.25,
"acc_norm_stderr": 0.04351941398892446
},
"harness|mmlu_high_school_computer_science|5": {
"acc": 0.3,
"acc_stderr": 0.046056618647183814,
"acc_norm": 0.3,
"acc_norm_stderr": 0.046056618647183814
},
"harness|mmlu_professional_medicine|5": {
"acc": 0.4411764705882353,
"acc_stderr": 0.030161911930767102,
"acc_norm": 0.4411764705882353,
"acc_norm_stderr": 0.030161911930767102
},
"harness|mmlu_security_studies|5": {
"acc": 0.3795918367346939,
"acc_stderr": 0.03106721126287249,
"acc_norm": 0.3795918367346939,
"acc_norm_stderr": 0.03106721126287249
},
"harness|mmlu_high_school_world_history|5": {
"acc": 0.2109704641350211,
"acc_stderr": 0.02655837250266192,
"acc_norm": 0.2109704641350211,
"acc_norm_stderr": 0.02655837250266192
},
"harness|mmlu_professional_law|5": {
"acc": 0.23468057366362452,
"acc_stderr": 0.010824026872449344,
"acc_norm": 0.23468057366362452,
"acc_norm_stderr": 0.010824026872449344
},
"harness|mmlu_high_school_us_history|5": {
"acc": 0.25,
"acc_stderr": 0.03039153369274154,
"acc_norm": 0.25,
"acc_norm_stderr": 0.03039153369274154
},
"harness|mmlu_high_school_european_history|5": {
"acc": 0.22424242424242424,
"acc_stderr": 0.03256866661681102,
"acc_norm": 0.22424242424242424,
"acc_norm_stderr": 0.03256866661681102
},
"harness|truthfulqa_mc|0": {
"mc1": 0.25091799265605874,
"mc1_stderr": 0.015176985027707682,
"mc2": 0.4116568832959107,
"mc2_stderr": 0.015044504977529799
}
} | {
"all": 0,
"harness|arc_challenge|25": 0,
"harness|hellaswag|10": 0,
"harness|mmlu_world_religions|5": 1,
"harness|mmlu_management|5": 1,
"harness|mmlu_miscellaneous|5": 1,
"harness|mmlu_anatomy|5": 1,
"harness|mmlu_abstract_algebra|5": 1,
"harness|mmlu_conceptual_physics|5": 1,
"harness|mmlu_virology|5": 1,
"harness|mmlu_philosophy|5": 1,
"harness|mmlu_human_aging|5": 1,
"harness|mmlu_human_sexuality|5": 1,
"harness|mmlu_medical_genetics|5": 1,
"harness|mmlu_high_school_geography|5": 1,
"harness|mmlu_electrical_engineering|5": 1,
"harness|mmlu_college_physics|5": 1,
"harness|mmlu_high_school_microeconomics|5": 1,
"harness|mmlu_high_school_macroeconomics|5": 1,
"harness|mmlu_computer_security|5": 1,
"harness|mmlu_global_facts|5": 1,
"harness|mmlu_jurisprudence|5": 1,
"harness|mmlu_high_school_chemistry|5": 1,
"harness|mmlu_high_school_biology|5": 1,
"harness|mmlu_marketing|5": 1,
"harness|mmlu_clinical_knowledge|5": 1,
"harness|mmlu_public_relations|5": 1,
"harness|mmlu_high_school_mathematics|5": 1,
"harness|mmlu_high_school_physics|5": 1,
"harness|mmlu_sociology|5": 1,
"harness|mmlu_college_medicine|5": 1,
"harness|mmlu_elementary_mathematics|5": 1,
"harness|mmlu_college_biology|5": 1,
"harness|mmlu_college_chemistry|5": 1,
"harness|mmlu_us_foreign_policy|5": 1,
"harness|mmlu_moral_disputes|5": 1,
"harness|mmlu_logical_fallacies|5": 1,
"harness|mmlu_prehistory|5": 1,
"harness|mmlu_college_mathematics|5": 1,
"harness|mmlu_high_school_government_and_politics|5": 1,
"harness|mmlu_econometrics|5": 1,
"harness|mmlu_high_school_psychology|5": 1,
"harness|mmlu_formal_logic|5": 1,
"harness|mmlu_nutrition|5": 1,
"harness|mmlu_business_ethics|5": 1,
"harness|mmlu_international_law|5": 1,
"harness|mmlu_astronomy|5": 1,
"harness|mmlu_professional_psychology|5": 1,
"harness|mmlu_professional_accounting|5": 1,
"harness|mmlu_machine_learning|5": 1,
"harness|mmlu_high_school_statistics|5": 1,
"harness|mmlu_moral_scenarios|5": 1,
"harness|mmlu_college_computer_science|5": 1,
"harness|mmlu_high_school_computer_science|5": 1,
"harness|mmlu_professional_medicine|5": 1,
"harness|mmlu_security_studies|5": 1,
"harness|mmlu_high_school_world_history|5": 1,
"harness|mmlu_professional_law|5": 1,
"harness|mmlu_high_school_us_history|5": 1,
"harness|mmlu_high_school_european_history|5": 1,
"harness|truthfulqa_mc|0": 0
} | {
"model_name": "EleutherAI/polyglot-ko-1.3b",
"model_sha": "557e162cf6e944fdbae05bab2e45d066a125eacb",
"model_dtype": "torch.float16",
"lighteval_sha": "",
"num_few_shot_default": 0,
"num_fewshot_seeds": 1,
"override_batch_size": 1,
"max_samples": null
} |
{
"daily": {
"daily": 8
},
"quarterly": {
"quarterly": 8
},
"harness|arc_challenge|25": {
"acc": 0.33532423208191126,
"acc_stderr": 0.01379618294778556,
"acc_norm": 0.3848122866894198,
"acc_norm_stderr": 0.014218371065251112
},
"harness|hellaswag|10": {
"acc": 0.35480979884485164,
"acc_stderr": 0.004774778180345192,
"acc_norm": 0.44911372236606256,
"acc_norm_stderr": 0.00496387293685794
},
"harness|mmlu_world_religions|5": {
"acc": 0.45614035087719296,
"acc_stderr": 0.03820042586602966,
"acc_norm": 0.45614035087719296,
"acc_norm_stderr": 0.03820042586602966
},
"harness|mmlu_management|5": {
"acc": 0.6019417475728155,
"acc_stderr": 0.04846748253977238,
"acc_norm": 0.6019417475728155,
"acc_norm_stderr": 0.04846748253977238
},
"harness|mmlu_miscellaneous|5": {
"acc": 0.41762452107279696,
"acc_stderr": 0.017635637326951534,
"acc_norm": 0.41762452107279696,
"acc_norm_stderr": 0.017635637326951534
},
"harness|mmlu_anatomy|5": {
"acc": 0.34074074074074073,
"acc_stderr": 0.040943762699967946,
"acc_norm": 0.34074074074074073,
"acc_norm_stderr": 0.040943762699967946
},
"harness|mmlu_abstract_algebra|5": {
"acc": 0.19,
"acc_stderr": 0.03942772444036623,
"acc_norm": 0.19,
"acc_norm_stderr": 0.03942772444036623
},
"harness|mmlu_conceptual_physics|5": {
"acc": 0.2978723404255319,
"acc_stderr": 0.029896145682095462,
"acc_norm": 0.2978723404255319,
"acc_norm_stderr": 0.029896145682095462
},
"harness|mmlu_virology|5": {
"acc": 0.3614457831325301,
"acc_stderr": 0.0374005938202932,
"acc_norm": 0.3614457831325301,
"acc_norm_stderr": 0.0374005938202932
},
"harness|mmlu_philosophy|5": {
"acc": 0.4758842443729904,
"acc_stderr": 0.028365041542564584,
"acc_norm": 0.4758842443729904,
"acc_norm_stderr": 0.028365041542564584
},
"harness|mmlu_human_aging|5": {
"acc": 0.3811659192825112,
"acc_stderr": 0.032596251184168284,
"acc_norm": 0.3811659192825112,
"acc_norm_stderr": 0.032596251184168284
},
"harness|mmlu_human_sexuality|5": {
"acc": 0.3511450381679389,
"acc_stderr": 0.04186445163013751,
"acc_norm": 0.3511450381679389,
"acc_norm_stderr": 0.04186445163013751
},
"harness|mmlu_medical_genetics|5": {
"acc": 0.27,
"acc_stderr": 0.0446196043338474,
"acc_norm": 0.27,
"acc_norm_stderr": 0.0446196043338474
},
"harness|mmlu_high_school_geography|5": {
"acc": 0.494949494949495,
"acc_stderr": 0.035621707606254015,
"acc_norm": 0.494949494949495,
"acc_norm_stderr": 0.035621707606254015
},
"harness|mmlu_electrical_engineering|5": {
"acc": 0.4,
"acc_stderr": 0.04082482904638628,
"acc_norm": 0.4,
"acc_norm_stderr": 0.04082482904638628
},
"harness|mmlu_college_physics|5": {
"acc": 0.3137254901960784,
"acc_stderr": 0.04617034827006717,
"acc_norm": 0.3137254901960784,
"acc_norm_stderr": 0.04617034827006717
},
"harness|mmlu_high_school_microeconomics|5": {
"acc": 0.4957983193277311,
"acc_stderr": 0.0324773433444811,
"acc_norm": 0.4957983193277311,
"acc_norm_stderr": 0.0324773433444811
},
"harness|mmlu_high_school_macroeconomics|5": {
"acc": 0.4256410256410256,
"acc_stderr": 0.025069094387296546,
"acc_norm": 0.4256410256410256,
"acc_norm_stderr": 0.025069094387296546
},
"harness|mmlu_computer_security|5": {
"acc": 0.59,
"acc_stderr": 0.049431107042371025,
"acc_norm": 0.59,
"acc_norm_stderr": 0.049431107042371025
},
"harness|mmlu_global_facts|5": {
"acc": 0.29,
"acc_stderr": 0.045604802157206845,
"acc_norm": 0.29,
"acc_norm_stderr": 0.045604802157206845
},
"harness|mmlu_jurisprudence|5": {
"acc": 0.4537037037037037,
"acc_stderr": 0.04812917324536821,
"acc_norm": 0.4537037037037037,
"acc_norm_stderr": 0.04812917324536821
},
"harness|mmlu_high_school_chemistry|5": {
"acc": 0.35467980295566504,
"acc_stderr": 0.03366124489051449,
"acc_norm": 0.35467980295566504,
"acc_norm_stderr": 0.03366124489051449
},
"harness|mmlu_high_school_biology|5": {
"acc": 0.4290322580645161,
"acc_stderr": 0.02815603653823321,
"acc_norm": 0.4290322580645161,
"acc_norm_stderr": 0.02815603653823321
},
"harness|mmlu_marketing|5": {
"acc": 0.6666666666666666,
"acc_stderr": 0.03088273697413865,
"acc_norm": 0.6666666666666666,
"acc_norm_stderr": 0.03088273697413865
},
"harness|mmlu_clinical_knowledge|5": {
"acc": 0.4188679245283019,
"acc_stderr": 0.03036505082911521,
"acc_norm": 0.4188679245283019,
"acc_norm_stderr": 0.03036505082911521
},
"harness|mmlu_public_relations|5": {
"acc": 0.42727272727272725,
"acc_stderr": 0.04738198703545483,
"acc_norm": 0.42727272727272725,
"acc_norm_stderr": 0.04738198703545483
},
"harness|mmlu_high_school_mathematics|5": {
"acc": 0.34814814814814815,
"acc_stderr": 0.029045600290616258,
"acc_norm": 0.34814814814814815,
"acc_norm_stderr": 0.029045600290616258
},
"harness|mmlu_high_school_physics|5": {
"acc": 0.2913907284768212,
"acc_stderr": 0.037101857261199946,
"acc_norm": 0.2913907284768212,
"acc_norm_stderr": 0.037101857261199946
},
"harness|mmlu_sociology|5": {
"acc": 0.5174129353233831,
"acc_stderr": 0.03533389234739245,
"acc_norm": 0.5174129353233831,
"acc_norm_stderr": 0.03533389234739245
},
"harness|mmlu_college_medicine|5": {
"acc": 0.37572254335260113,
"acc_stderr": 0.03692820767264867,
"acc_norm": 0.37572254335260113,
"acc_norm_stderr": 0.03692820767264867
},
"harness|mmlu_elementary_mathematics|5": {
"acc": 0.3492063492063492,
"acc_stderr": 0.024552292209342658,
"acc_norm": 0.3492063492063492,
"acc_norm_stderr": 0.024552292209342658
},
"harness|mmlu_college_biology|5": {
"acc": 0.3333333333333333,
"acc_stderr": 0.039420826399272135,
"acc_norm": 0.3333333333333333,
"acc_norm_stderr": 0.039420826399272135
},
"harness|mmlu_college_chemistry|5": {
"acc": 0.35,
"acc_stderr": 0.04793724854411019,
"acc_norm": 0.35,
"acc_norm_stderr": 0.04793724854411019
},
"harness|mmlu_us_foreign_policy|5": {
"acc": 0.49,
"acc_stderr": 0.05024183937956913,
"acc_norm": 0.49,
"acc_norm_stderr": 0.05024183937956913
},
"harness|mmlu_moral_disputes|5": {
"acc": 0.47398843930635837,
"acc_stderr": 0.026882643434022885,
"acc_norm": 0.47398843930635837,
"acc_norm_stderr": 0.026882643434022885
},
"harness|mmlu_logical_fallacies|5": {
"acc": 0.44171779141104295,
"acc_stderr": 0.039015918258361836,
"acc_norm": 0.44171779141104295,
"acc_norm_stderr": 0.039015918258361836
},
"harness|mmlu_prehistory|5": {
"acc": 0.42592592592592593,
"acc_stderr": 0.027513747284379424,
"acc_norm": 0.42592592592592593,
"acc_norm_stderr": 0.027513747284379424
},
"harness|mmlu_college_mathematics|5": {
"acc": 0.33,
"acc_stderr": 0.04725815626252606,
"acc_norm": 0.33,
"acc_norm_stderr": 0.04725815626252606
},
"harness|mmlu_high_school_government_and_politics|5": {
"acc": 0.5129533678756477,
"acc_stderr": 0.0360722806104775,
"acc_norm": 0.5129533678756477,
"acc_norm_stderr": 0.0360722806104775
},
"harness|mmlu_econometrics|5": {
"acc": 0.24561403508771928,
"acc_stderr": 0.0404933929774814,
"acc_norm": 0.24561403508771928,
"acc_norm_stderr": 0.0404933929774814
},
"harness|mmlu_high_school_psychology|5": {
"acc": 0.47155963302752296,
"acc_stderr": 0.02140261569734804,
"acc_norm": 0.47155963302752296,
"acc_norm_stderr": 0.02140261569734804
},
"harness|mmlu_formal_logic|5": {
"acc": 0.36507936507936506,
"acc_stderr": 0.04306241259127152,
"acc_norm": 0.36507936507936506,
"acc_norm_stderr": 0.04306241259127152
},
"harness|mmlu_nutrition|5": {
"acc": 0.4117647058823529,
"acc_stderr": 0.028180596328259297,
"acc_norm": 0.4117647058823529,
"acc_norm_stderr": 0.028180596328259297
},
"harness|mmlu_business_ethics|5": {
"acc": 0.44,
"acc_stderr": 0.04988876515698589,
"acc_norm": 0.44,
"acc_norm_stderr": 0.04988876515698589
},
"harness|mmlu_international_law|5": {
"acc": 0.5867768595041323,
"acc_stderr": 0.04495087843548408,
"acc_norm": 0.5867768595041323,
"acc_norm_stderr": 0.04495087843548408
},
"harness|mmlu_astronomy|5": {
"acc": 0.40131578947368424,
"acc_stderr": 0.03988903703336284,
"acc_norm": 0.40131578947368424,
"acc_norm_stderr": 0.03988903703336284
},
"harness|mmlu_professional_psychology|5": {
"acc": 0.32679738562091504,
"acc_stderr": 0.018975427920507215,
"acc_norm": 0.32679738562091504,
"acc_norm_stderr": 0.018975427920507215
},
"harness|mmlu_professional_accounting|5": {
"acc": 0.3333333333333333,
"acc_stderr": 0.02812163604063988,
"acc_norm": 0.3333333333333333,
"acc_norm_stderr": 0.02812163604063988
},
"harness|mmlu_machine_learning|5": {
"acc": 0.3392857142857143,
"acc_stderr": 0.04493949068613539,
"acc_norm": 0.3392857142857143,
"acc_norm_stderr": 0.04493949068613539
},
"harness|mmlu_high_school_statistics|5": {
"acc": 0.41203703703703703,
"acc_stderr": 0.03356787758160835,
"acc_norm": 0.41203703703703703,
"acc_norm_stderr": 0.03356787758160835
},
"harness|mmlu_moral_scenarios|5": {
"acc": 0.329608938547486,
"acc_stderr": 0.015721531075183884,
"acc_norm": 0.329608938547486,
"acc_norm_stderr": 0.015721531075183884
},
"harness|mmlu_college_computer_science|5": {
"acc": 0.39,
"acc_stderr": 0.04902071300001975,
"acc_norm": 0.39,
"acc_norm_stderr": 0.04902071300001975
},
"harness|mmlu_high_school_computer_science|5": {
"acc": 0.61,
"acc_stderr": 0.04902071300001975,
"acc_norm": 0.61,
"acc_norm_stderr": 0.04902071300001975
},
"harness|mmlu_professional_medicine|5": {
"acc": 0.375,
"acc_stderr": 0.029408372932278746,
"acc_norm": 0.375,
"acc_norm_stderr": 0.029408372932278746
},
"harness|mmlu_security_studies|5": {
"acc": 0.43673469387755104,
"acc_stderr": 0.03175195237583322,
"acc_norm": 0.43673469387755104,
"acc_norm_stderr": 0.03175195237583322
},
"harness|mmlu_high_school_world_history|5": {
"acc": 0.4810126582278481,
"acc_stderr": 0.03252375148090448,
"acc_norm": 0.4810126582278481,
"acc_norm_stderr": 0.03252375148090448
},
"harness|mmlu_professional_law|5": {
"acc": 0.29791395045632335,
"acc_stderr": 0.011680717340400059,
"acc_norm": 0.29791395045632335,
"acc_norm_stderr": 0.011680717340400059
},
"harness|mmlu_high_school_us_history|5": {
"acc": 0.29411764705882354,
"acc_stderr": 0.03198001660115072,
"acc_norm": 0.29411764705882354,
"acc_norm_stderr": 0.03198001660115072
},
"harness|mmlu_high_school_european_history|5": {
"acc": 0.30303030303030304,
"acc_stderr": 0.03588624800091707,
"acc_norm": 0.30303030303030304,
"acc_norm_stderr": 0.03588624800091707
},
"harness|truthfulqa_mc|0": {
"mc1": 0.3317013463892289,
"mc1_stderr": 0.01648214881024147,
"mc2": 0.5171680571717291,
"mc2_stderr": 0.01606077987901482
}
} | {
"all": 0,
"harness|arc_challenge|25": 0,
"harness|hellaswag|10": 0,
"harness|mmlu_world_religions|5": 1,
"harness|mmlu_management|5": 1,
"harness|mmlu_miscellaneous|5": 1,
"harness|mmlu_anatomy|5": 1,
"harness|mmlu_abstract_algebra|5": 1,
"harness|mmlu_conceptual_physics|5": 1,
"harness|mmlu_virology|5": 1,
"harness|mmlu_philosophy|5": 1,
"harness|mmlu_human_aging|5": 1,
"harness|mmlu_human_sexuality|5": 1,
"harness|mmlu_medical_genetics|5": 1,
"harness|mmlu_high_school_geography|5": 1,
"harness|mmlu_electrical_engineering|5": 1,
"harness|mmlu_college_physics|5": 1,
"harness|mmlu_high_school_microeconomics|5": 1,
"harness|mmlu_high_school_macroeconomics|5": 1,
"harness|mmlu_computer_security|5": 1,
"harness|mmlu_global_facts|5": 1,
"harness|mmlu_jurisprudence|5": 1,
"harness|mmlu_high_school_chemistry|5": 1,
"harness|mmlu_high_school_biology|5": 1,
"harness|mmlu_marketing|5": 1,
"harness|mmlu_clinical_knowledge|5": 1,
"harness|mmlu_public_relations|5": 1,
"harness|mmlu_high_school_mathematics|5": 1,
"harness|mmlu_high_school_physics|5": 1,
"harness|mmlu_sociology|5": 1,
"harness|mmlu_college_medicine|5": 1,
"harness|mmlu_elementary_mathematics|5": 1,
"harness|mmlu_college_biology|5": 1,
"harness|mmlu_college_chemistry|5": 1,
"harness|mmlu_us_foreign_policy|5": 1,
"harness|mmlu_moral_disputes|5": 1,
"harness|mmlu_logical_fallacies|5": 1,
"harness|mmlu_prehistory|5": 1,
"harness|mmlu_college_mathematics|5": 1,
"harness|mmlu_high_school_government_and_politics|5": 1,
"harness|mmlu_econometrics|5": 1,
"harness|mmlu_high_school_psychology|5": 1,
"harness|mmlu_formal_logic|5": 1,
"harness|mmlu_nutrition|5": 1,
"harness|mmlu_business_ethics|5": 1,
"harness|mmlu_international_law|5": 1,
"harness|mmlu_astronomy|5": 1,
"harness|mmlu_professional_psychology|5": 1,
"harness|mmlu_professional_accounting|5": 1,
"harness|mmlu_machine_learning|5": 1,
"harness|mmlu_high_school_statistics|5": 1,
"harness|mmlu_moral_scenarios|5": 1,
"harness|mmlu_college_computer_science|5": 1,
"harness|mmlu_high_school_computer_science|5": 1,
"harness|mmlu_professional_medicine|5": 1,
"harness|mmlu_security_studies|5": 1,
"harness|mmlu_high_school_world_history|5": 1,
"harness|mmlu_professional_law|5": 1,
"harness|mmlu_high_school_us_history|5": 1,
"harness|mmlu_high_school_european_history|5": 1,
"harness|truthfulqa_mc|0": 0
} | {
"model_name": "HuggingFaceH4/zephyr-7b-beta",
"model_sha": "3bac358730f8806e5c3dc7c7e19eb36e045bf720",
"model_dtype": "torch.float16",
"lighteval_sha": "",
"num_few_shot_default": 0,
"num_fewshot_seeds": 1,
"override_batch_size": 1,
"max_samples": null
} |
{
"daily": {
"daily": 6
},
"quarterly": {
"quarterly": 6
},
"harness|arc_challenge|25": {
"acc": 0.42918088737201365,
"acc_stderr": 0.014464085894870651,
"acc_norm": 0.46501706484641636,
"acc_norm_stderr": 0.014575583922019672
},
"harness|hellaswag|10": {
"acc": 0.445628360884286,
"acc_stderr": 0.004960191341430244,
"acc_norm": 0.589523999203346,
"acc_norm_stderr": 0.004909148239488273
},
"harness|mmlu_world_religions|5": {
"acc": 0.6432748538011696,
"acc_stderr": 0.03674013002860954,
"acc_norm": 0.6432748538011696,
"acc_norm_stderr": 0.03674013002860954
},
"harness|mmlu_management|5": {
"acc": 0.6116504854368932,
"acc_stderr": 0.04825729337356389,
"acc_norm": 0.6116504854368932,
"acc_norm_stderr": 0.04825729337356389
},
"harness|mmlu_miscellaneous|5": {
"acc": 0.6155810983397191,
"acc_stderr": 0.01739568874281962,
"acc_norm": 0.6155810983397191,
"acc_norm_stderr": 0.01739568874281962
},
"harness|mmlu_anatomy|5": {
"acc": 0.4962962962962963,
"acc_stderr": 0.04319223625811331,
"acc_norm": 0.4962962962962963,
"acc_norm_stderr": 0.04319223625811331
},
"harness|mmlu_abstract_algebra|5": {
"acc": 0.26,
"acc_stderr": 0.04408440022768077,
"acc_norm": 0.26,
"acc_norm_stderr": 0.04408440022768077
},
"harness|mmlu_conceptual_physics|5": {
"acc": 0.4553191489361702,
"acc_stderr": 0.03255525359340354,
"acc_norm": 0.4553191489361702,
"acc_norm_stderr": 0.03255525359340354
},
"harness|mmlu_virology|5": {
"acc": 0.5180722891566265,
"acc_stderr": 0.038899512528272166,
"acc_norm": 0.5180722891566265,
"acc_norm_stderr": 0.038899512528272166
},
"harness|mmlu_philosophy|5": {
"acc": 0.5755627009646302,
"acc_stderr": 0.028071928247946205,
"acc_norm": 0.5755627009646302,
"acc_norm_stderr": 0.028071928247946205
},
"harness|mmlu_human_aging|5": {
"acc": 0.5650224215246636,
"acc_stderr": 0.033272833702713445,
"acc_norm": 0.5650224215246636,
"acc_norm_stderr": 0.033272833702713445
},
"harness|mmlu_human_sexuality|5": {
"acc": 0.5877862595419847,
"acc_stderr": 0.04317171194870255,
"acc_norm": 0.5877862595419847,
"acc_norm_stderr": 0.04317171194870255
},
"harness|mmlu_medical_genetics|5": {
"acc": 0.5,
"acc_stderr": 0.050251890762960605,
"acc_norm": 0.5,
"acc_norm_stderr": 0.050251890762960605
},
"harness|mmlu_high_school_geography|5": {
"acc": 0.6515151515151515,
"acc_stderr": 0.033948539651564025,
"acc_norm": 0.6515151515151515,
"acc_norm_stderr": 0.033948539651564025
},
"harness|mmlu_electrical_engineering|5": {
"acc": 0.503448275862069,
"acc_stderr": 0.04166567577101579,
"acc_norm": 0.503448275862069,
"acc_norm_stderr": 0.04166567577101579
},
"harness|mmlu_college_physics|5": {
"acc": 0.2549019607843137,
"acc_stderr": 0.043364327079931785,
"acc_norm": 0.2549019607843137,
"acc_norm_stderr": 0.043364327079931785
},
"harness|mmlu_high_school_microeconomics|5": {
"acc": 0.5756302521008403,
"acc_stderr": 0.03210479051015776,
"acc_norm": 0.5756302521008403,
"acc_norm_stderr": 0.03210479051015776
},
"harness|mmlu_high_school_macroeconomics|5": {
"acc": 0.541025641025641,
"acc_stderr": 0.025265525491284295,
"acc_norm": 0.541025641025641,
"acc_norm_stderr": 0.025265525491284295
},
"harness|mmlu_computer_security|5": {
"acc": 0.54,
"acc_stderr": 0.05009082659620332,
"acc_norm": 0.54,
"acc_norm_stderr": 0.05009082659620332
},
"harness|mmlu_global_facts|5": {
"acc": 0.31,
"acc_stderr": 0.04648231987117316,
"acc_norm": 0.31,
"acc_norm_stderr": 0.04648231987117316
},
"harness|mmlu_jurisprudence|5": {
"acc": 0.5555555555555556,
"acc_stderr": 0.04803752235190192,
"acc_norm": 0.5555555555555556,
"acc_norm_stderr": 0.04803752235190192
},
"harness|mmlu_high_school_chemistry|5": {
"acc": 0.3842364532019704,
"acc_stderr": 0.0342239856565755,
"acc_norm": 0.3842364532019704,
"acc_norm_stderr": 0.0342239856565755
},
"harness|mmlu_high_school_biology|5": {
"acc": 0.5774193548387097,
"acc_stderr": 0.02810096472427264,
"acc_norm": 0.5774193548387097,
"acc_norm_stderr": 0.02810096472427264
},
"harness|mmlu_marketing|5": {
"acc": 0.7777777777777778,
"acc_stderr": 0.027236013946196673,
"acc_norm": 0.7777777777777778,
"acc_norm_stderr": 0.027236013946196673
},
"harness|mmlu_clinical_knowledge|5": {
"acc": 0.4981132075471698,
"acc_stderr": 0.030772653642075657,
"acc_norm": 0.4981132075471698,
"acc_norm_stderr": 0.030772653642075657
},
"harness|mmlu_public_relations|5": {
"acc": 0.5272727272727272,
"acc_stderr": 0.04782001791380061,
"acc_norm": 0.5272727272727272,
"acc_norm_stderr": 0.04782001791380061
},
"harness|mmlu_high_school_mathematics|5": {
"acc": 0.25555555555555554,
"acc_stderr": 0.026593939101844082,
"acc_norm": 0.25555555555555554,
"acc_norm_stderr": 0.026593939101844082
},
"harness|mmlu_high_school_physics|5": {
"acc": 0.33774834437086093,
"acc_stderr": 0.038615575462551684,
"acc_norm": 0.33774834437086093,
"acc_norm_stderr": 0.038615575462551684
},
"harness|mmlu_sociology|5": {
"acc": 0.7064676616915423,
"acc_stderr": 0.032200241045342054,
"acc_norm": 0.7064676616915423,
"acc_norm_stderr": 0.032200241045342054
},
"harness|mmlu_college_medicine|5": {
"acc": 0.4797687861271676,
"acc_stderr": 0.03809342081273958,
"acc_norm": 0.4797687861271676,
"acc_norm_stderr": 0.03809342081273958
},
"harness|mmlu_elementary_mathematics|5": {
"acc": 0.38095238095238093,
"acc_stderr": 0.025010749116137602,
"acc_norm": 0.38095238095238093,
"acc_norm_stderr": 0.025010749116137602
},
"harness|mmlu_college_biology|5": {
"acc": 0.4236111111111111,
"acc_stderr": 0.041321250197233685,
"acc_norm": 0.4236111111111111,
"acc_norm_stderr": 0.041321250197233685
},
"harness|mmlu_college_chemistry|5": {
"acc": 0.31,
"acc_stderr": 0.04648231987117316,
"acc_norm": 0.31,
"acc_norm_stderr": 0.04648231987117316
},
"harness|mmlu_us_foreign_policy|5": {
"acc": 0.71,
"acc_stderr": 0.04560480215720683,
"acc_norm": 0.71,
"acc_norm_stderr": 0.04560480215720683
},
"harness|mmlu_moral_disputes|5": {
"acc": 0.5751445086705202,
"acc_stderr": 0.026613350840261733,
"acc_norm": 0.5751445086705202,
"acc_norm_stderr": 0.026613350840261733
},
"harness|mmlu_logical_fallacies|5": {
"acc": 0.5030674846625767,
"acc_stderr": 0.03928297078179662,
"acc_norm": 0.5030674846625767,
"acc_norm_stderr": 0.03928297078179662
},
"harness|mmlu_prehistory|5": {
"acc": 0.5370370370370371,
"acc_stderr": 0.027744313443376536,
"acc_norm": 0.5370370370370371,
"acc_norm_stderr": 0.027744313443376536
},
"harness|mmlu_college_mathematics|5": {
"acc": 0.33,
"acc_stderr": 0.04725815626252606,
"acc_norm": 0.33,
"acc_norm_stderr": 0.04725815626252606
},
"harness|mmlu_high_school_government_and_politics|5": {
"acc": 0.6217616580310881,
"acc_stderr": 0.034998072761933376,
"acc_norm": 0.6217616580310881,
"acc_norm_stderr": 0.034998072761933376
},
"harness|mmlu_econometrics|5": {
"acc": 0.37719298245614036,
"acc_stderr": 0.04559522141958216,
"acc_norm": 0.37719298245614036,
"acc_norm_stderr": 0.04559522141958216
},
"harness|mmlu_high_school_psychology|5": {
"acc": 0.6385321100917432,
"acc_stderr": 0.02059808200993736,
"acc_norm": 0.6385321100917432,
"acc_norm_stderr": 0.02059808200993736
},
"harness|mmlu_formal_logic|5": {
"acc": 0.4126984126984127,
"acc_stderr": 0.04403438954768177,
"acc_norm": 0.4126984126984127,
"acc_norm_stderr": 0.04403438954768177
},
"harness|mmlu_nutrition|5": {
"acc": 0.5261437908496732,
"acc_stderr": 0.028590752958852387,
"acc_norm": 0.5261437908496732,
"acc_norm_stderr": 0.028590752958852387
},
"harness|mmlu_business_ethics|5": {
"acc": 0.57,
"acc_stderr": 0.049756985195624284,
"acc_norm": 0.57,
"acc_norm_stderr": 0.049756985195624284
},
"harness|mmlu_international_law|5": {
"acc": 0.7520661157024794,
"acc_stderr": 0.03941897526516304,
"acc_norm": 0.7520661157024794,
"acc_norm_stderr": 0.03941897526516304
},
"harness|mmlu_astronomy|5": {
"acc": 0.5789473684210527,
"acc_stderr": 0.040179012759817494,
"acc_norm": 0.5789473684210527,
"acc_norm_stderr": 0.040179012759817494
},
"harness|mmlu_professional_psychology|5": {
"acc": 0.4738562091503268,
"acc_stderr": 0.020200164564804588,
"acc_norm": 0.4738562091503268,
"acc_norm_stderr": 0.020200164564804588
},
"harness|mmlu_professional_accounting|5": {
"acc": 0.3404255319148936,
"acc_stderr": 0.02826765748265013,
"acc_norm": 0.3404255319148936,
"acc_norm_stderr": 0.02826765748265013
},
"harness|mmlu_machine_learning|5": {
"acc": 0.38392857142857145,
"acc_stderr": 0.046161430750285455,
"acc_norm": 0.38392857142857145,
"acc_norm_stderr": 0.046161430750285455
},
"harness|mmlu_high_school_statistics|5": {
"acc": 0.4675925925925926,
"acc_stderr": 0.03402801581358966,
"acc_norm": 0.4675925925925926,
"acc_norm_stderr": 0.03402801581358966
},
"harness|mmlu_moral_scenarios|5": {
"acc": 0.21675977653631284,
"acc_stderr": 0.013780598486443363,
"acc_norm": 0.21675977653631284,
"acc_norm_stderr": 0.013780598486443363
},
"harness|mmlu_college_computer_science|5": {
"acc": 0.39,
"acc_stderr": 0.04902071300001975,
"acc_norm": 0.39,
"acc_norm_stderr": 0.04902071300001975
},
"harness|mmlu_high_school_computer_science|5": {
"acc": 0.71,
"acc_stderr": 0.04560480215720684,
"acc_norm": 0.71,
"acc_norm_stderr": 0.04560480215720684
},
"harness|mmlu_professional_medicine|5": {
"acc": 0.4411764705882353,
"acc_stderr": 0.0301619119307671,
"acc_norm": 0.4411764705882353,
"acc_norm_stderr": 0.0301619119307671
},
"harness|mmlu_security_studies|5": {
"acc": 0.6285714285714286,
"acc_stderr": 0.03093285879278986,
"acc_norm": 0.6285714285714286,
"acc_norm_stderr": 0.03093285879278986
},
"harness|mmlu_high_school_world_history|5": {
"acc": 0.70042194092827,
"acc_stderr": 0.029818024749753095,
"acc_norm": 0.70042194092827,
"acc_norm_stderr": 0.029818024749753095
},
"harness|mmlu_professional_law|5": {
"acc": 0.378748370273794,
"acc_stderr": 0.012389052105003741,
"acc_norm": 0.378748370273794,
"acc_norm_stderr": 0.012389052105003741
},
"harness|mmlu_high_school_us_history|5": {
"acc": 0.6225490196078431,
"acc_stderr": 0.03402272044340703,
"acc_norm": 0.6225490196078431,
"acc_norm_stderr": 0.03402272044340703
},
"harness|mmlu_high_school_european_history|5": {
"acc": 0.6666666666666666,
"acc_stderr": 0.03681050869161549,
"acc_norm": 0.6666666666666666,
"acc_norm_stderr": 0.03681050869161549
},
"harness|truthfulqa_mc|0": {
"mc1": 0.33659730722154224,
"mc1_stderr": 0.016542412809494877,
"mc2": 0.49995145184296846,
"mc2_stderr": 0.015887726098900913
}
} | {
"all": 0,
"harness|arc_challenge|25": 0,
"harness|hellaswag|10": 0,
"harness|mmlu_world_religions|5": 1,
"harness|mmlu_management|5": 1,
"harness|mmlu_miscellaneous|5": 1,
"harness|mmlu_anatomy|5": 1,
"harness|mmlu_abstract_algebra|5": 1,
"harness|mmlu_conceptual_physics|5": 1,
"harness|mmlu_virology|5": 1,
"harness|mmlu_philosophy|5": 1,
"harness|mmlu_human_aging|5": 1,
"harness|mmlu_human_sexuality|5": 1,
"harness|mmlu_medical_genetics|5": 1,
"harness|mmlu_high_school_geography|5": 1,
"harness|mmlu_electrical_engineering|5": 1,
"harness|mmlu_college_physics|5": 1,
"harness|mmlu_high_school_microeconomics|5": 1,
"harness|mmlu_high_school_macroeconomics|5": 1,
"harness|mmlu_computer_security|5": 1,
"harness|mmlu_global_facts|5": 1,
"harness|mmlu_jurisprudence|5": 1,
"harness|mmlu_high_school_chemistry|5": 1,
"harness|mmlu_high_school_biology|5": 1,
"harness|mmlu_marketing|5": 1,
"harness|mmlu_clinical_knowledge|5": 1,
"harness|mmlu_public_relations|5": 1,
"harness|mmlu_high_school_mathematics|5": 1,
"harness|mmlu_high_school_physics|5": 1,
"harness|mmlu_sociology|5": 1,
"harness|mmlu_college_medicine|5": 1,
"harness|mmlu_elementary_mathematics|5": 1,
"harness|mmlu_college_biology|5": 1,
"harness|mmlu_college_chemistry|5": 1,
"harness|mmlu_us_foreign_policy|5": 1,
"harness|mmlu_moral_disputes|5": 1,
"harness|mmlu_logical_fallacies|5": 1,
"harness|mmlu_prehistory|5": 1,
"harness|mmlu_college_mathematics|5": 1,
"harness|mmlu_high_school_government_and_politics|5": 1,
"harness|mmlu_econometrics|5": 1,
"harness|mmlu_high_school_psychology|5": 1,
"harness|mmlu_formal_logic|5": 1,
"harness|mmlu_nutrition|5": 1,
"harness|mmlu_business_ethics|5": 1,
"harness|mmlu_international_law|5": 1,
"harness|mmlu_astronomy|5": 1,
"harness|mmlu_professional_psychology|5": 1,
"harness|mmlu_professional_accounting|5": 1,
"harness|mmlu_machine_learning|5": 1,
"harness|mmlu_high_school_statistics|5": 1,
"harness|mmlu_moral_scenarios|5": 1,
"harness|mmlu_college_computer_science|5": 1,
"harness|mmlu_high_school_computer_science|5": 1,
"harness|mmlu_professional_medicine|5": 1,
"harness|mmlu_security_studies|5": 1,
"harness|mmlu_high_school_world_history|5": 1,
"harness|mmlu_professional_law|5": 1,
"harness|mmlu_high_school_us_history|5": 1,
"harness|mmlu_high_school_european_history|5": 1,
"harness|truthfulqa_mc|0": 0
} | {
"model_name": "nlpai-lab/KULLM3",
"model_sha": "5a6bcd0fc7f240460eb6d57016f7b4060bc1f43b",
"model_dtype": "torch.float16",
"lighteval_sha": "",
"num_few_shot_default": 0,
"num_fewshot_seeds": 1,
"override_batch_size": 1,
"max_samples": null
} |
{
"daily": {
"daily": 4
},
"quarterly": {
"quarterly": 4
},
"harness|arc_challenge|25": {
"acc": 0.7465870307167235,
"acc_stderr": 0.012710896778378604,
"acc_norm": 0.7807167235494881,
"acc_norm_stderr": 0.012091245787615728
},
"harness|hellaswag|10": {
"acc": 0.6385182234614618,
"acc_stderr": 0.004794478426382617,
"acc_norm": 0.7561242780322645,
"acc_norm_stderr": 0.004285410130466119
},
"harness|mmlu_world_religions|5": {
"acc": 0.6900584795321637,
"acc_stderr": 0.035469769593931624,
"acc_norm": 0.6900584795321637,
"acc_norm_stderr": 0.035469769593931624
},
"harness|mmlu_management|5": {
"acc": 0.6601941747572816,
"acc_stderr": 0.046897659372781335,
"acc_norm": 0.6601941747572816,
"acc_norm_stderr": 0.046897659372781335
},
"harness|mmlu_miscellaneous|5": {
"acc": 0.6845466155810983,
"acc_stderr": 0.016617501738763408,
"acc_norm": 0.6845466155810983,
"acc_norm_stderr": 0.016617501738763408
},
"harness|mmlu_anatomy|5": {
"acc": 0.48148148148148145,
"acc_stderr": 0.04316378599511324,
"acc_norm": 0.48148148148148145,
"acc_norm_stderr": 0.04316378599511324
},
"harness|mmlu_abstract_algebra|5": {
"acc": 0.33,
"acc_stderr": 0.047258156262526045,
"acc_norm": 0.33,
"acc_norm_stderr": 0.047258156262526045
},
"harness|mmlu_conceptual_physics|5": {
"acc": 0.46808510638297873,
"acc_stderr": 0.03261936918467383,
"acc_norm": 0.46808510638297873,
"acc_norm_stderr": 0.03261936918467383
},
"harness|mmlu_virology|5": {
"acc": 0.4759036144578313,
"acc_stderr": 0.03887971849597264,
"acc_norm": 0.4759036144578313,
"acc_norm_stderr": 0.03887971849597264
},
"harness|mmlu_philosophy|5": {
"acc": 0.6334405144694534,
"acc_stderr": 0.02736807824397163,
"acc_norm": 0.6334405144694534,
"acc_norm_stderr": 0.02736807824397163
},
"harness|mmlu_human_aging|5": {
"acc": 0.6681614349775785,
"acc_stderr": 0.03160295143776679,
"acc_norm": 0.6681614349775785,
"acc_norm_stderr": 0.03160295143776679
},
"harness|mmlu_human_sexuality|5": {
"acc": 0.6030534351145038,
"acc_stderr": 0.04291135671009224,
"acc_norm": 0.6030534351145038,
"acc_norm_stderr": 0.04291135671009224
},
"harness|mmlu_medical_genetics|5": {
"acc": 0.51,
"acc_stderr": 0.05024183937956911,
"acc_norm": 0.51,
"acc_norm_stderr": 0.05024183937956911
},
"harness|mmlu_high_school_geography|5": {
"acc": 0.7222222222222222,
"acc_stderr": 0.03191178226713547,
"acc_norm": 0.7222222222222222,
"acc_norm_stderr": 0.03191178226713547
},
"harness|mmlu_electrical_engineering|5": {
"acc": 0.47586206896551725,
"acc_stderr": 0.041618085035015295,
"acc_norm": 0.47586206896551725,
"acc_norm_stderr": 0.041618085035015295
},
"harness|mmlu_college_physics|5": {
"acc": 0.2549019607843137,
"acc_stderr": 0.04336432707993178,
"acc_norm": 0.2549019607843137,
"acc_norm_stderr": 0.04336432707993178
},
"harness|mmlu_high_school_microeconomics|5": {
"acc": 0.592436974789916,
"acc_stderr": 0.031918633744784666,
"acc_norm": 0.592436974789916,
"acc_norm_stderr": 0.031918633744784666
},
"harness|mmlu_high_school_macroeconomics|5": {
"acc": 0.5948717948717949,
"acc_stderr": 0.024890471769938142,
"acc_norm": 0.5948717948717949,
"acc_norm_stderr": 0.024890471769938142
},
"harness|mmlu_computer_security|5": {
"acc": 0.66,
"acc_stderr": 0.04760952285695237,
"acc_norm": 0.66,
"acc_norm_stderr": 0.04760952285695237
},
"harness|mmlu_global_facts|5": {
"acc": 0.27,
"acc_stderr": 0.04461960433384739,
"acc_norm": 0.27,
"acc_norm_stderr": 0.04461960433384739
},
"harness|mmlu_jurisprudence|5": {
"acc": 0.6388888888888888,
"acc_stderr": 0.04643454608906275,
"acc_norm": 0.6388888888888888,
"acc_norm_stderr": 0.04643454608906275
},
"harness|mmlu_high_school_chemistry|5": {
"acc": 0.4433497536945813,
"acc_stderr": 0.034953345821629345,
"acc_norm": 0.4433497536945813,
"acc_norm_stderr": 0.034953345821629345
},
"harness|mmlu_high_school_biology|5": {
"acc": 0.5806451612903226,
"acc_stderr": 0.028071588901091838,
"acc_norm": 0.5806451612903226,
"acc_norm_stderr": 0.028071588901091838
},
"harness|mmlu_marketing|5": {
"acc": 0.811965811965812,
"acc_stderr": 0.025598193686652254,
"acc_norm": 0.811965811965812,
"acc_norm_stderr": 0.025598193686652254
},
"harness|mmlu_clinical_knowledge|5": {
"acc": 0.5169811320754717,
"acc_stderr": 0.030755120364119898,
"acc_norm": 0.5169811320754717,
"acc_norm_stderr": 0.030755120364119898
},
"harness|mmlu_public_relations|5": {
"acc": 0.5818181818181818,
"acc_stderr": 0.04724577405731573,
"acc_norm": 0.5818181818181818,
"acc_norm_stderr": 0.04724577405731573
},
"harness|mmlu_high_school_mathematics|5": {
"acc": 0.3888888888888889,
"acc_stderr": 0.029723278961476664,
"acc_norm": 0.3888888888888889,
"acc_norm_stderr": 0.029723278961476664
},
"harness|mmlu_high_school_physics|5": {
"acc": 0.3708609271523179,
"acc_stderr": 0.03943966699183629,
"acc_norm": 0.3708609271523179,
"acc_norm_stderr": 0.03943966699183629
},
"harness|mmlu_sociology|5": {
"acc": 0.6666666666666666,
"acc_stderr": 0.033333333333333326,
"acc_norm": 0.6666666666666666,
"acc_norm_stderr": 0.033333333333333326
},
"harness|mmlu_college_medicine|5": {
"acc": 0.47398843930635837,
"acc_stderr": 0.038073017265045125,
"acc_norm": 0.47398843930635837,
"acc_norm_stderr": 0.038073017265045125
},
"harness|mmlu_elementary_mathematics|5": {
"acc": 0.42328042328042326,
"acc_stderr": 0.025446365634406793,
"acc_norm": 0.42328042328042326,
"acc_norm_stderr": 0.025446365634406793
},
"harness|mmlu_college_biology|5": {
"acc": 0.5625,
"acc_stderr": 0.04148415739394154,
"acc_norm": 0.5625,
"acc_norm_stderr": 0.04148415739394154
},
"harness|mmlu_college_chemistry|5": {
"acc": 0.39,
"acc_stderr": 0.04902071300001975,
"acc_norm": 0.39,
"acc_norm_stderr": 0.04902071300001975
},
"harness|mmlu_us_foreign_policy|5": {
"acc": 0.78,
"acc_stderr": 0.04163331998932263,
"acc_norm": 0.78,
"acc_norm_stderr": 0.04163331998932263
},
"harness|mmlu_moral_disputes|5": {
"acc": 0.5491329479768786,
"acc_stderr": 0.026788811931562767,
"acc_norm": 0.5491329479768786,
"acc_norm_stderr": 0.026788811931562767
},
"harness|mmlu_logical_fallacies|5": {
"acc": 0.6319018404907976,
"acc_stderr": 0.03789213935838396,
"acc_norm": 0.6319018404907976,
"acc_norm_stderr": 0.03789213935838396
},
"harness|mmlu_prehistory|5": {
"acc": 0.5925925925925926,
"acc_stderr": 0.02733954664066273,
"acc_norm": 0.5925925925925926,
"acc_norm_stderr": 0.02733954664066273
},
"harness|mmlu_college_mathematics|5": {
"acc": 0.4,
"acc_stderr": 0.049236596391733084,
"acc_norm": 0.4,
"acc_norm_stderr": 0.049236596391733084
},
"harness|mmlu_high_school_government_and_politics|5": {
"acc": 0.7668393782383419,
"acc_stderr": 0.03051611137147601,
"acc_norm": 0.7668393782383419,
"acc_norm_stderr": 0.03051611137147601
},
"harness|mmlu_econometrics|5": {
"acc": 0.4473684210526316,
"acc_stderr": 0.046774730044912,
"acc_norm": 0.4473684210526316,
"acc_norm_stderr": 0.046774730044912
},
"harness|mmlu_high_school_psychology|5": {
"acc": 0.726605504587156,
"acc_stderr": 0.01910929984609827,
"acc_norm": 0.726605504587156,
"acc_norm_stderr": 0.01910929984609827
},
"harness|mmlu_formal_logic|5": {
"acc": 0.3968253968253968,
"acc_stderr": 0.04375888492727061,
"acc_norm": 0.3968253968253968,
"acc_norm_stderr": 0.04375888492727061
},
"harness|mmlu_nutrition|5": {
"acc": 0.6078431372549019,
"acc_stderr": 0.027956046165424516,
"acc_norm": 0.6078431372549019,
"acc_norm_stderr": 0.027956046165424516
},
"harness|mmlu_business_ethics|5": {
"acc": 0.55,
"acc_stderr": 0.05,
"acc_norm": 0.55,
"acc_norm_stderr": 0.05
},
"harness|mmlu_international_law|5": {
"acc": 0.6942148760330579,
"acc_stderr": 0.04205953933884122,
"acc_norm": 0.6942148760330579,
"acc_norm_stderr": 0.04205953933884122
},
"harness|mmlu_astronomy|5": {
"acc": 0.618421052631579,
"acc_stderr": 0.03953173377749194,
"acc_norm": 0.618421052631579,
"acc_norm_stderr": 0.03953173377749194
},
"harness|mmlu_professional_psychology|5": {
"acc": 0.5669934640522876,
"acc_stderr": 0.02004544247332422,
"acc_norm": 0.5669934640522876,
"acc_norm_stderr": 0.02004544247332422
},
"harness|mmlu_professional_accounting|5": {
"acc": 0.4219858156028369,
"acc_stderr": 0.029462189233370586,
"acc_norm": 0.4219858156028369,
"acc_norm_stderr": 0.029462189233370586
},
"harness|mmlu_machine_learning|5": {
"acc": 0.5089285714285714,
"acc_stderr": 0.04745033255489123,
"acc_norm": 0.5089285714285714,
"acc_norm_stderr": 0.04745033255489123
},
"harness|mmlu_high_school_statistics|5": {
"acc": 0.4351851851851852,
"acc_stderr": 0.03381200005643526,
"acc_norm": 0.4351851851851852,
"acc_norm_stderr": 0.03381200005643526
},
"harness|mmlu_moral_scenarios|5": {
"acc": 0.3787709497206704,
"acc_stderr": 0.016223533510365117,
"acc_norm": 0.3787709497206704,
"acc_norm_stderr": 0.016223533510365117
},
"harness|mmlu_college_computer_science|5": {
"acc": 0.47,
"acc_stderr": 0.05016135580465919,
"acc_norm": 0.47,
"acc_norm_stderr": 0.05016135580465919
},
"harness|mmlu_high_school_computer_science|5": {
"acc": 0.66,
"acc_stderr": 0.04760952285695238,
"acc_norm": 0.66,
"acc_norm_stderr": 0.04760952285695238
},
"harness|mmlu_professional_medicine|5": {
"acc": 0.48161764705882354,
"acc_stderr": 0.03035230339535196,
"acc_norm": 0.48161764705882354,
"acc_norm_stderr": 0.03035230339535196
},
"harness|mmlu_security_studies|5": {
"acc": 0.6448979591836734,
"acc_stderr": 0.030635655150387634,
"acc_norm": 0.6448979591836734,
"acc_norm_stderr": 0.030635655150387634
},
"harness|mmlu_high_school_world_history|5": {
"acc": 0.729957805907173,
"acc_stderr": 0.028900721906293426,
"acc_norm": 0.729957805907173,
"acc_norm_stderr": 0.028900721906293426
},
"harness|mmlu_professional_law|5": {
"acc": 0.41460234680573665,
"acc_stderr": 0.012582597058908284,
"acc_norm": 0.41460234680573665,
"acc_norm_stderr": 0.012582597058908284
},
"harness|mmlu_high_school_us_history|5": {
"acc": 0.6421568627450981,
"acc_stderr": 0.03364487286088298,
"acc_norm": 0.6421568627450981,
"acc_norm_stderr": 0.03364487286088298
},
"harness|mmlu_high_school_european_history|5": {
"acc": 0.6181818181818182,
"acc_stderr": 0.03793713171165635,
"acc_norm": 0.6181818181818182,
"acc_norm_stderr": 0.03793713171165635
},
"harness|truthfulqa_mc|0": {
"mc1": 0.6328029375764994,
"mc1_stderr": 0.01687480500145318,
"mc2": 0.7522925779273922,
"mc2_stderr": 0.014568927682929578
}
} | {
"all": 0,
"harness|arc_challenge|25": 0,
"harness|hellaswag|10": 0,
"harness|mmlu_world_religions|5": 1,
"harness|mmlu_management|5": 1,
"harness|mmlu_miscellaneous|5": 1,
"harness|mmlu_anatomy|5": 1,
"harness|mmlu_abstract_algebra|5": 1,
"harness|mmlu_conceptual_physics|5": 1,
"harness|mmlu_virology|5": 1,
"harness|mmlu_philosophy|5": 1,
"harness|mmlu_human_aging|5": 1,
"harness|mmlu_human_sexuality|5": 1,
"harness|mmlu_medical_genetics|5": 1,
"harness|mmlu_high_school_geography|5": 1,
"harness|mmlu_electrical_engineering|5": 1,
"harness|mmlu_college_physics|5": 1,
"harness|mmlu_high_school_microeconomics|5": 1,
"harness|mmlu_high_school_macroeconomics|5": 1,
"harness|mmlu_computer_security|5": 1,
"harness|mmlu_global_facts|5": 1,
"harness|mmlu_jurisprudence|5": 1,
"harness|mmlu_high_school_chemistry|5": 1,
"harness|mmlu_high_school_biology|5": 1,
"harness|mmlu_marketing|5": 1,
"harness|mmlu_clinical_knowledge|5": 1,
"harness|mmlu_public_relations|5": 1,
"harness|mmlu_high_school_mathematics|5": 1,
"harness|mmlu_high_school_physics|5": 1,
"harness|mmlu_sociology|5": 1,
"harness|mmlu_college_medicine|5": 1,
"harness|mmlu_elementary_mathematics|5": 1,
"harness|mmlu_college_biology|5": 1,
"harness|mmlu_college_chemistry|5": 1,
"harness|mmlu_us_foreign_policy|5": 1,
"harness|mmlu_moral_disputes|5": 1,
"harness|mmlu_logical_fallacies|5": 1,
"harness|mmlu_prehistory|5": 1,
"harness|mmlu_college_mathematics|5": 1,
"harness|mmlu_high_school_government_and_politics|5": 1,
"harness|mmlu_econometrics|5": 1,
"harness|mmlu_high_school_psychology|5": 1,
"harness|mmlu_formal_logic|5": 1,
"harness|mmlu_nutrition|5": 1,
"harness|mmlu_business_ethics|5": 1,
"harness|mmlu_international_law|5": 1,
"harness|mmlu_astronomy|5": 1,
"harness|mmlu_professional_psychology|5": 1,
"harness|mmlu_professional_accounting|5": 1,
"harness|mmlu_machine_learning|5": 1,
"harness|mmlu_high_school_statistics|5": 1,
"harness|mmlu_moral_scenarios|5": 1,
"harness|mmlu_college_computer_science|5": 1,
"harness|mmlu_high_school_computer_science|5": 1,
"harness|mmlu_professional_medicine|5": 1,
"harness|mmlu_security_studies|5": 1,
"harness|mmlu_high_school_world_history|5": 1,
"harness|mmlu_professional_law|5": 1,
"harness|mmlu_high_school_us_history|5": 1,
"harness|mmlu_high_school_european_history|5": 1,
"harness|truthfulqa_mc|0": 0
} | {
"model_name": "x2bee/POLAR-14B-DPO-v1.3",
"model_sha": "337edbed4c86db2da27e3b0e07086134f8d27a09",
"model_dtype": "torch.float16",
"lighteval_sha": "",
"num_few_shot_default": 0,
"num_fewshot_seeds": 1,
"override_batch_size": 1,
"max_samples": null
} |
{
"daily": {
"daily": 7
},
"quarterly": {
"quarterly": 7
},
"harness|arc_challenge|25": {
"acc": 0.7363481228668942,
"acc_stderr": 0.012875929151297058,
"acc_norm": 0.7491467576791809,
"acc_norm_stderr": 0.012668198621315433
},
"harness|hellaswag|10": {
"acc": 0.7228639713204541,
"acc_stderr": 0.004466695023677848,
"acc_norm": 0.7422824138617805,
"acc_norm_stderr": 0.004364838000335614
},
"harness|mmlu_world_religions|5": {
"acc": 0.6140350877192983,
"acc_stderr": 0.03733756969066164,
"acc_norm": 0.6140350877192983,
"acc_norm_stderr": 0.03733756969066164
},
"harness|mmlu_management|5": {
"acc": 0.6893203883495146,
"acc_stderr": 0.045821241601615506,
"acc_norm": 0.6893203883495146,
"acc_norm_stderr": 0.045821241601615506
},
"harness|mmlu_miscellaneous|5": {
"acc": 0.6526181353767561,
"acc_stderr": 0.017026671748655728,
"acc_norm": 0.6526181353767561,
"acc_norm_stderr": 0.017026671748655728
},
"harness|mmlu_anatomy|5": {
"acc": 0.5037037037037037,
"acc_stderr": 0.043192236258113324,
"acc_norm": 0.5037037037037037,
"acc_norm_stderr": 0.043192236258113324
},
"harness|mmlu_abstract_algebra|5": {
"acc": 0.37,
"acc_stderr": 0.048523658709391,
"acc_norm": 0.37,
"acc_norm_stderr": 0.048523658709391
},
"harness|mmlu_conceptual_physics|5": {
"acc": 0.451063829787234,
"acc_stderr": 0.032529096196131965,
"acc_norm": 0.451063829787234,
"acc_norm_stderr": 0.032529096196131965
},
"harness|mmlu_virology|5": {
"acc": 0.4939759036144578,
"acc_stderr": 0.03892212195333045,
"acc_norm": 0.4939759036144578,
"acc_norm_stderr": 0.03892212195333045
},
"harness|mmlu_philosophy|5": {
"acc": 0.5852090032154341,
"acc_stderr": 0.02798268045975956,
"acc_norm": 0.5852090032154341,
"acc_norm_stderr": 0.02798268045975956
},
"harness|mmlu_human_aging|5": {
"acc": 0.6412556053811659,
"acc_stderr": 0.032190792004199956,
"acc_norm": 0.6412556053811659,
"acc_norm_stderr": 0.032190792004199956
},
"harness|mmlu_human_sexuality|5": {
"acc": 0.5954198473282443,
"acc_stderr": 0.043046937953806645,
"acc_norm": 0.5954198473282443,
"acc_norm_stderr": 0.043046937953806645
},
"harness|mmlu_medical_genetics|5": {
"acc": 0.47,
"acc_stderr": 0.05016135580465919,
"acc_norm": 0.47,
"acc_norm_stderr": 0.05016135580465919
},
"harness|mmlu_high_school_geography|5": {
"acc": 0.6616161616161617,
"acc_stderr": 0.033711241426263014,
"acc_norm": 0.6616161616161617,
"acc_norm_stderr": 0.033711241426263014
},
"harness|mmlu_electrical_engineering|5": {
"acc": 0.4827586206896552,
"acc_stderr": 0.041641887201693775,
"acc_norm": 0.4827586206896552,
"acc_norm_stderr": 0.041641887201693775
},
"harness|mmlu_college_physics|5": {
"acc": 0.2549019607843137,
"acc_stderr": 0.04336432707993178,
"acc_norm": 0.2549019607843137,
"acc_norm_stderr": 0.04336432707993178
},
"harness|mmlu_high_school_microeconomics|5": {
"acc": 0.5882352941176471,
"acc_stderr": 0.031968769891957786,
"acc_norm": 0.5882352941176471,
"acc_norm_stderr": 0.031968769891957786
},
"harness|mmlu_high_school_macroeconomics|5": {
"acc": 0.6025641025641025,
"acc_stderr": 0.024811920017903836,
"acc_norm": 0.6025641025641025,
"acc_norm_stderr": 0.024811920017903836
},
"harness|mmlu_computer_security|5": {
"acc": 0.66,
"acc_stderr": 0.04760952285695237,
"acc_norm": 0.66,
"acc_norm_stderr": 0.04760952285695237
},
"harness|mmlu_global_facts|5": {
"acc": 0.35,
"acc_stderr": 0.047937248544110196,
"acc_norm": 0.35,
"acc_norm_stderr": 0.047937248544110196
},
"harness|mmlu_jurisprudence|5": {
"acc": 0.5925925925925926,
"acc_stderr": 0.04750077341199984,
"acc_norm": 0.5925925925925926,
"acc_norm_stderr": 0.04750077341199984
},
"harness|mmlu_high_school_chemistry|5": {
"acc": 0.43842364532019706,
"acc_stderr": 0.03491207857486518,
"acc_norm": 0.43842364532019706,
"acc_norm_stderr": 0.03491207857486518
},
"harness|mmlu_high_school_biology|5": {
"acc": 0.567741935483871,
"acc_stderr": 0.028181739720019413,
"acc_norm": 0.567741935483871,
"acc_norm_stderr": 0.028181739720019413
},
"harness|mmlu_marketing|5": {
"acc": 0.7948717948717948,
"acc_stderr": 0.026453508054040356,
"acc_norm": 0.7948717948717948,
"acc_norm_stderr": 0.026453508054040356
},
"harness|mmlu_clinical_knowledge|5": {
"acc": 0.5169811320754717,
"acc_stderr": 0.030755120364119905,
"acc_norm": 0.5169811320754717,
"acc_norm_stderr": 0.030755120364119905
},
"harness|mmlu_public_relations|5": {
"acc": 0.5727272727272728,
"acc_stderr": 0.047381987035454834,
"acc_norm": 0.5727272727272728,
"acc_norm_stderr": 0.047381987035454834
},
"harness|mmlu_high_school_mathematics|5": {
"acc": 0.3962962962962963,
"acc_stderr": 0.029822619458533997,
"acc_norm": 0.3962962962962963,
"acc_norm_stderr": 0.029822619458533997
},
"harness|mmlu_high_school_physics|5": {
"acc": 0.3708609271523179,
"acc_stderr": 0.03943966699183629,
"acc_norm": 0.3708609271523179,
"acc_norm_stderr": 0.03943966699183629
},
"harness|mmlu_sociology|5": {
"acc": 0.6766169154228856,
"acc_stderr": 0.03307615947979035,
"acc_norm": 0.6766169154228856,
"acc_norm_stderr": 0.03307615947979035
},
"harness|mmlu_college_medicine|5": {
"acc": 0.49710982658959535,
"acc_stderr": 0.038124005659748335,
"acc_norm": 0.49710982658959535,
"acc_norm_stderr": 0.038124005659748335
},
"harness|mmlu_elementary_mathematics|5": {
"acc": 0.42592592592592593,
"acc_stderr": 0.02546714904546955,
"acc_norm": 0.42592592592592593,
"acc_norm_stderr": 0.02546714904546955
},
"harness|mmlu_college_biology|5": {
"acc": 0.5555555555555556,
"acc_stderr": 0.04155319955593146,
"acc_norm": 0.5555555555555556,
"acc_norm_stderr": 0.04155319955593146
},
"harness|mmlu_college_chemistry|5": {
"acc": 0.4,
"acc_stderr": 0.04923659639173309,
"acc_norm": 0.4,
"acc_norm_stderr": 0.04923659639173309
},
"harness|mmlu_us_foreign_policy|5": {
"acc": 0.73,
"acc_stderr": 0.044619604333847394,
"acc_norm": 0.73,
"acc_norm_stderr": 0.044619604333847394
},
"harness|mmlu_moral_disputes|5": {
"acc": 0.5549132947976878,
"acc_stderr": 0.02675625512966377,
"acc_norm": 0.5549132947976878,
"acc_norm_stderr": 0.02675625512966377
},
"harness|mmlu_logical_fallacies|5": {
"acc": 0.588957055214724,
"acc_stderr": 0.038656978537853624,
"acc_norm": 0.588957055214724,
"acc_norm_stderr": 0.038656978537853624
},
"harness|mmlu_prehistory|5": {
"acc": 0.5771604938271605,
"acc_stderr": 0.027487472980871595,
"acc_norm": 0.5771604938271605,
"acc_norm_stderr": 0.027487472980871595
},
"harness|mmlu_college_mathematics|5": {
"acc": 0.4,
"acc_stderr": 0.049236596391733084,
"acc_norm": 0.4,
"acc_norm_stderr": 0.049236596391733084
},
"harness|mmlu_high_school_government_and_politics|5": {
"acc": 0.7305699481865285,
"acc_stderr": 0.032018671228777947,
"acc_norm": 0.7305699481865285,
"acc_norm_stderr": 0.032018671228777947
},
"harness|mmlu_econometrics|5": {
"acc": 0.42105263157894735,
"acc_stderr": 0.046446020912223177,
"acc_norm": 0.42105263157894735,
"acc_norm_stderr": 0.046446020912223177
},
"harness|mmlu_high_school_psychology|5": {
"acc": 0.7064220183486238,
"acc_stderr": 0.019525151122639663,
"acc_norm": 0.7064220183486238,
"acc_norm_stderr": 0.019525151122639663
},
"harness|mmlu_formal_logic|5": {
"acc": 0.3968253968253968,
"acc_stderr": 0.04375888492727061,
"acc_norm": 0.3968253968253968,
"acc_norm_stderr": 0.04375888492727061
},
"harness|mmlu_nutrition|5": {
"acc": 0.545751633986928,
"acc_stderr": 0.02850980780262659,
"acc_norm": 0.545751633986928,
"acc_norm_stderr": 0.02850980780262659
},
"harness|mmlu_business_ethics|5": {
"acc": 0.55,
"acc_stderr": 0.05000000000000001,
"acc_norm": 0.55,
"acc_norm_stderr": 0.05000000000000001
},
"harness|mmlu_international_law|5": {
"acc": 0.6859504132231405,
"acc_stderr": 0.04236964753041019,
"acc_norm": 0.6859504132231405,
"acc_norm_stderr": 0.04236964753041019
},
"harness|mmlu_astronomy|5": {
"acc": 0.6052631578947368,
"acc_stderr": 0.039777499346220734,
"acc_norm": 0.6052631578947368,
"acc_norm_stderr": 0.039777499346220734
},
"harness|mmlu_professional_psychology|5": {
"acc": 0.5392156862745098,
"acc_stderr": 0.02016552331390791,
"acc_norm": 0.5392156862745098,
"acc_norm_stderr": 0.02016552331390791
},
"harness|mmlu_professional_accounting|5": {
"acc": 0.35815602836879434,
"acc_stderr": 0.02860208586275942,
"acc_norm": 0.35815602836879434,
"acc_norm_stderr": 0.02860208586275942
},
"harness|mmlu_machine_learning|5": {
"acc": 0.4107142857142857,
"acc_stderr": 0.04669510663875192,
"acc_norm": 0.4107142857142857,
"acc_norm_stderr": 0.04669510663875192
},
"harness|mmlu_high_school_statistics|5": {
"acc": 0.44907407407407407,
"acc_stderr": 0.03392238405321617,
"acc_norm": 0.44907407407407407,
"acc_norm_stderr": 0.03392238405321617
},
"harness|mmlu_moral_scenarios|5": {
"acc": 0.3452513966480447,
"acc_stderr": 0.015901432608930354,
"acc_norm": 0.3452513966480447,
"acc_norm_stderr": 0.015901432608930354
},
"harness|mmlu_college_computer_science|5": {
"acc": 0.43,
"acc_stderr": 0.049756985195624284,
"acc_norm": 0.43,
"acc_norm_stderr": 0.049756985195624284
},
"harness|mmlu_high_school_computer_science|5": {
"acc": 0.66,
"acc_stderr": 0.04760952285695238,
"acc_norm": 0.66,
"acc_norm_stderr": 0.04760952285695238
},
"harness|mmlu_professional_medicine|5": {
"acc": 0.45588235294117646,
"acc_stderr": 0.030254372573976694,
"acc_norm": 0.45588235294117646,
"acc_norm_stderr": 0.030254372573976694
},
"harness|mmlu_security_studies|5": {
"acc": 0.6204081632653061,
"acc_stderr": 0.031067211262872457,
"acc_norm": 0.6204081632653061,
"acc_norm_stderr": 0.031067211262872457
},
"harness|mmlu_high_school_world_history|5": {
"acc": 0.6582278481012658,
"acc_stderr": 0.030874537537553617,
"acc_norm": 0.6582278481012658,
"acc_norm_stderr": 0.030874537537553617
},
"harness|mmlu_professional_law|5": {
"acc": 0.4152542372881356,
"acc_stderr": 0.012585471793400667,
"acc_norm": 0.4152542372881356,
"acc_norm_stderr": 0.012585471793400667
},
"harness|mmlu_high_school_us_history|5": {
"acc": 0.5343137254901961,
"acc_stderr": 0.03501038327635896,
"acc_norm": 0.5343137254901961,
"acc_norm_stderr": 0.03501038327635896
},
"harness|mmlu_high_school_european_history|5": {
"acc": 0.5454545454545454,
"acc_stderr": 0.038881769216741004,
"acc_norm": 0.5454545454545454,
"acc_norm_stderr": 0.038881769216741004
},
"harness|truthfulqa_mc|0": {
"mc1": 0.4663402692778458,
"mc1_stderr": 0.01746379386716811,
"mc2": null,
"mc2_stderr": null
}
} | {
"all": 0,
"harness|arc_challenge|25": 0,
"harness|hellaswag|10": 0,
"harness|mmlu_world_religions|5": 1,
"harness|mmlu_management|5": 1,
"harness|mmlu_miscellaneous|5": 1,
"harness|mmlu_anatomy|5": 1,
"harness|mmlu_abstract_algebra|5": 1,
"harness|mmlu_conceptual_physics|5": 1,
"harness|mmlu_virology|5": 1,
"harness|mmlu_philosophy|5": 1,
"harness|mmlu_human_aging|5": 1,
"harness|mmlu_human_sexuality|5": 1,
"harness|mmlu_medical_genetics|5": 1,
"harness|mmlu_high_school_geography|5": 1,
"harness|mmlu_electrical_engineering|5": 1,
"harness|mmlu_college_physics|5": 1,
"harness|mmlu_high_school_microeconomics|5": 1,
"harness|mmlu_high_school_macroeconomics|5": 1,
"harness|mmlu_computer_security|5": 1,
"harness|mmlu_global_facts|5": 1,
"harness|mmlu_jurisprudence|5": 1,
"harness|mmlu_high_school_chemistry|5": 1,
"harness|mmlu_high_school_biology|5": 1,
"harness|mmlu_marketing|5": 1,
"harness|mmlu_clinical_knowledge|5": 1,
"harness|mmlu_public_relations|5": 1,
"harness|mmlu_high_school_mathematics|5": 1,
"harness|mmlu_high_school_physics|5": 1,
"harness|mmlu_sociology|5": 1,
"harness|mmlu_college_medicine|5": 1,
"harness|mmlu_elementary_mathematics|5": 1,
"harness|mmlu_college_biology|5": 1,
"harness|mmlu_college_chemistry|5": 1,
"harness|mmlu_us_foreign_policy|5": 1,
"harness|mmlu_moral_disputes|5": 1,
"harness|mmlu_logical_fallacies|5": 1,
"harness|mmlu_prehistory|5": 1,
"harness|mmlu_college_mathematics|5": 1,
"harness|mmlu_high_school_government_and_politics|5": 1,
"harness|mmlu_econometrics|5": 1,
"harness|mmlu_high_school_psychology|5": 1,
"harness|mmlu_formal_logic|5": 1,
"harness|mmlu_nutrition|5": 1,
"harness|mmlu_business_ethics|5": 1,
"harness|mmlu_international_law|5": 1,
"harness|mmlu_astronomy|5": 1,
"harness|mmlu_professional_psychology|5": 1,
"harness|mmlu_professional_accounting|5": 1,
"harness|mmlu_machine_learning|5": 1,
"harness|mmlu_high_school_statistics|5": 1,
"harness|mmlu_moral_scenarios|5": 1,
"harness|mmlu_college_computer_science|5": 1,
"harness|mmlu_high_school_computer_science|5": 1,
"harness|mmlu_professional_medicine|5": 1,
"harness|mmlu_security_studies|5": 1,
"harness|mmlu_high_school_world_history|5": 1,
"harness|mmlu_professional_law|5": 1,
"harness|mmlu_high_school_us_history|5": 1,
"harness|mmlu_high_school_european_history|5": 1,
"harness|truthfulqa_mc|0": 0
} | {
"model_name": "x2bee/POLAR-14B-DPO-v1.4",
"model_sha": "a6e64075fafaa3d5e393ff89c3cb26f9615e6de9",
"model_dtype": "torch.float16",
"lighteval_sha": "",
"num_few_shot_default": 0,
"num_fewshot_seeds": 1,
"override_batch_size": 1,
"max_samples": null
} |
{
"daily": {
"daily": 5
},
"quarterly": {
"quarterly": 5
},
"harness|arc_challenge|25": {
"acc": 0.6638225255972696,
"acc_stderr": 0.013804855026205756,
"acc_norm": 0.7278156996587031,
"acc_norm_stderr": 0.013006600406423709
},
"harness|hellaswag|10": {
"acc": 0.45648277235610435,
"acc_stderr": 0.004970846697552306,
"acc_norm": 0.6349332802230632,
"acc_norm_stderr": 0.004804649197163697
},
"harness|mmlu_world_religions|5": {
"acc": 0.7309941520467836,
"acc_stderr": 0.0340105262010409,
"acc_norm": 0.7309941520467836,
"acc_norm_stderr": 0.0340105262010409
},
"harness|mmlu_management|5": {
"acc": 0.7766990291262136,
"acc_stderr": 0.04123553189891431,
"acc_norm": 0.7766990291262136,
"acc_norm_stderr": 0.04123553189891431
},
"harness|mmlu_miscellaneous|5": {
"acc": 0.7343550446998723,
"acc_stderr": 0.01579430248788872,
"acc_norm": 0.7343550446998723,
"acc_norm_stderr": 0.01579430248788872
},
"harness|mmlu_anatomy|5": {
"acc": 0.45185185185185184,
"acc_stderr": 0.04299268905480863,
"acc_norm": 0.45185185185185184,
"acc_norm_stderr": 0.04299268905480863
},
"harness|mmlu_abstract_algebra|5": {
"acc": 0.35,
"acc_stderr": 0.04793724854411019,
"acc_norm": 0.35,
"acc_norm_stderr": 0.04793724854411019
},
"harness|mmlu_conceptual_physics|5": {
"acc": 0.5276595744680851,
"acc_stderr": 0.03263597118409769,
"acc_norm": 0.5276595744680851,
"acc_norm_stderr": 0.03263597118409769
},
"harness|mmlu_virology|5": {
"acc": 0.4759036144578313,
"acc_stderr": 0.03887971849597264,
"acc_norm": 0.4759036144578313,
"acc_norm_stderr": 0.03887971849597264
},
"harness|mmlu_philosophy|5": {
"acc": 0.6559485530546624,
"acc_stderr": 0.026981478043648043,
"acc_norm": 0.6559485530546624,
"acc_norm_stderr": 0.026981478043648043
},
"harness|mmlu_human_aging|5": {
"acc": 0.6412556053811659,
"acc_stderr": 0.032190792004199956,
"acc_norm": 0.6412556053811659,
"acc_norm_stderr": 0.032190792004199956
},
"harness|mmlu_human_sexuality|5": {
"acc": 0.648854961832061,
"acc_stderr": 0.04186445163013751,
"acc_norm": 0.648854961832061,
"acc_norm_stderr": 0.04186445163013751
},
"harness|mmlu_medical_genetics|5": {
"acc": 0.54,
"acc_stderr": 0.05009082659620333,
"acc_norm": 0.54,
"acc_norm_stderr": 0.05009082659620333
},
"harness|mmlu_high_school_geography|5": {
"acc": 0.7777777777777778,
"acc_stderr": 0.029620227874790465,
"acc_norm": 0.7777777777777778,
"acc_norm_stderr": 0.029620227874790465
},
"harness|mmlu_electrical_engineering|5": {
"acc": 0.5103448275862069,
"acc_stderr": 0.04165774775728762,
"acc_norm": 0.5103448275862069,
"acc_norm_stderr": 0.04165774775728762
},
"harness|mmlu_college_physics|5": {
"acc": 0.3627450980392157,
"acc_stderr": 0.04784060704105655,
"acc_norm": 0.3627450980392157,
"acc_norm_stderr": 0.04784060704105655
},
"harness|mmlu_high_school_microeconomics|5": {
"acc": 0.6680672268907563,
"acc_stderr": 0.03058869701378364,
"acc_norm": 0.6680672268907563,
"acc_norm_stderr": 0.03058869701378364
},
"harness|mmlu_high_school_macroeconomics|5": {
"acc": 0.6384615384615384,
"acc_stderr": 0.024359581465397,
"acc_norm": 0.6384615384615384,
"acc_norm_stderr": 0.024359581465397
},
"harness|mmlu_computer_security|5": {
"acc": 0.65,
"acc_stderr": 0.04793724854411021,
"acc_norm": 0.65,
"acc_norm_stderr": 0.04793724854411021
},
"harness|mmlu_global_facts|5": {
"acc": 0.37,
"acc_stderr": 0.04852365870939099,
"acc_norm": 0.37,
"acc_norm_stderr": 0.04852365870939099
},
"harness|mmlu_jurisprudence|5": {
"acc": 0.6851851851851852,
"acc_stderr": 0.04489931073591312,
"acc_norm": 0.6851851851851852,
"acc_norm_stderr": 0.04489931073591312
},
"harness|mmlu_high_school_chemistry|5": {
"acc": 0.46798029556650245,
"acc_stderr": 0.035107665979592154,
"acc_norm": 0.46798029556650245,
"acc_norm_stderr": 0.035107665979592154
},
"harness|mmlu_high_school_biology|5": {
"acc": 0.6548387096774193,
"acc_stderr": 0.02704574657353432,
"acc_norm": 0.6548387096774193,
"acc_norm_stderr": 0.02704574657353432
},
"harness|mmlu_marketing|5": {
"acc": 0.8162393162393162,
"acc_stderr": 0.025372139671722933,
"acc_norm": 0.8162393162393162,
"acc_norm_stderr": 0.025372139671722933
},
"harness|mmlu_clinical_knowledge|5": {
"acc": 0.5773584905660377,
"acc_stderr": 0.03040233144576954,
"acc_norm": 0.5773584905660377,
"acc_norm_stderr": 0.03040233144576954
},
"harness|mmlu_public_relations|5": {
"acc": 0.6454545454545455,
"acc_stderr": 0.045820048415054174,
"acc_norm": 0.6454545454545455,
"acc_norm_stderr": 0.045820048415054174
},
"harness|mmlu_high_school_mathematics|5": {
"acc": 0.4074074074074074,
"acc_stderr": 0.029958249250082118,
"acc_norm": 0.4074074074074074,
"acc_norm_stderr": 0.029958249250082118
},
"harness|mmlu_high_school_physics|5": {
"acc": 0.3509933774834437,
"acc_stderr": 0.03896981964257375,
"acc_norm": 0.3509933774834437,
"acc_norm_stderr": 0.03896981964257375
},
"harness|mmlu_sociology|5": {
"acc": 0.7263681592039801,
"acc_stderr": 0.03152439186555404,
"acc_norm": 0.7263681592039801,
"acc_norm_stderr": 0.03152439186555404
},
"harness|mmlu_college_medicine|5": {
"acc": 0.5375722543352601,
"acc_stderr": 0.0380168510452446,
"acc_norm": 0.5375722543352601,
"acc_norm_stderr": 0.0380168510452446
},
"harness|mmlu_elementary_mathematics|5": {
"acc": 0.4365079365079365,
"acc_stderr": 0.025542846817400496,
"acc_norm": 0.4365079365079365,
"acc_norm_stderr": 0.025542846817400496
},
"harness|mmlu_college_biology|5": {
"acc": 0.5694444444444444,
"acc_stderr": 0.04140685639111503,
"acc_norm": 0.5694444444444444,
"acc_norm_stderr": 0.04140685639111503
},
"harness|mmlu_college_chemistry|5": {
"acc": 0.43,
"acc_stderr": 0.049756985195624284,
"acc_norm": 0.43,
"acc_norm_stderr": 0.049756985195624284
},
"harness|mmlu_us_foreign_policy|5": {
"acc": 0.78,
"acc_stderr": 0.04163331998932263,
"acc_norm": 0.78,
"acc_norm_stderr": 0.04163331998932263
},
"harness|mmlu_moral_disputes|5": {
"acc": 0.6098265895953757,
"acc_stderr": 0.026261677607806642,
"acc_norm": 0.6098265895953757,
"acc_norm_stderr": 0.026261677607806642
},
"harness|mmlu_logical_fallacies|5": {
"acc": 0.656441717791411,
"acc_stderr": 0.03731133519673893,
"acc_norm": 0.656441717791411,
"acc_norm_stderr": 0.03731133519673893
},
"harness|mmlu_prehistory|5": {
"acc": 0.6574074074074074,
"acc_stderr": 0.02640614597362568,
"acc_norm": 0.6574074074074074,
"acc_norm_stderr": 0.02640614597362568
},
"harness|mmlu_college_mathematics|5": {
"acc": 0.38,
"acc_stderr": 0.04878317312145632,
"acc_norm": 0.38,
"acc_norm_stderr": 0.04878317312145632
},
"harness|mmlu_high_school_government_and_politics|5": {
"acc": 0.7668393782383419,
"acc_stderr": 0.03051611137147601,
"acc_norm": 0.7668393782383419,
"acc_norm_stderr": 0.03051611137147601
},
"harness|mmlu_econometrics|5": {
"acc": 0.45614035087719296,
"acc_stderr": 0.046854730419077895,
"acc_norm": 0.45614035087719296,
"acc_norm_stderr": 0.046854730419077895
},
"harness|mmlu_high_school_psychology|5": {
"acc": 0.7853211009174312,
"acc_stderr": 0.017604304149256494,
"acc_norm": 0.7853211009174312,
"acc_norm_stderr": 0.017604304149256494
},
"harness|mmlu_formal_logic|5": {
"acc": 0.4523809523809524,
"acc_stderr": 0.044518079590553275,
"acc_norm": 0.4523809523809524,
"acc_norm_stderr": 0.044518079590553275
},
"harness|mmlu_nutrition|5": {
"acc": 0.6405228758169934,
"acc_stderr": 0.027475969910660952,
"acc_norm": 0.6405228758169934,
"acc_norm_stderr": 0.027475969910660952
},
"harness|mmlu_business_ethics|5": {
"acc": 0.66,
"acc_stderr": 0.04760952285695237,
"acc_norm": 0.66,
"acc_norm_stderr": 0.04760952285695237
},
"harness|mmlu_international_law|5": {
"acc": 0.7933884297520661,
"acc_stderr": 0.03695980128098824,
"acc_norm": 0.7933884297520661,
"acc_norm_stderr": 0.03695980128098824
},
"harness|mmlu_astronomy|5": {
"acc": 0.6842105263157895,
"acc_stderr": 0.0378272898086547,
"acc_norm": 0.6842105263157895,
"acc_norm_stderr": 0.0378272898086547
},
"harness|mmlu_professional_psychology|5": {
"acc": 0.5964052287581699,
"acc_stderr": 0.019848280168401164,
"acc_norm": 0.5964052287581699,
"acc_norm_stderr": 0.019848280168401164
},
"harness|mmlu_professional_accounting|5": {
"acc": 0.4397163120567376,
"acc_stderr": 0.02960991207559411,
"acc_norm": 0.4397163120567376,
"acc_norm_stderr": 0.02960991207559411
},
"harness|mmlu_machine_learning|5": {
"acc": 0.39285714285714285,
"acc_stderr": 0.04635550135609976,
"acc_norm": 0.39285714285714285,
"acc_norm_stderr": 0.04635550135609976
},
"harness|mmlu_high_school_statistics|5": {
"acc": 0.5787037037037037,
"acc_stderr": 0.03367462138896078,
"acc_norm": 0.5787037037037037,
"acc_norm_stderr": 0.03367462138896078
},
"harness|mmlu_moral_scenarios|5": {
"acc": 0.264804469273743,
"acc_stderr": 0.01475690648326066,
"acc_norm": 0.264804469273743,
"acc_norm_stderr": 0.01475690648326066
},
"harness|mmlu_college_computer_science|5": {
"acc": 0.52,
"acc_stderr": 0.050211673156867795,
"acc_norm": 0.52,
"acc_norm_stderr": 0.050211673156867795
},
"harness|mmlu_high_school_computer_science|5": {
"acc": 0.7,
"acc_stderr": 0.04605661864718381,
"acc_norm": 0.7,
"acc_norm_stderr": 0.04605661864718381
},
"harness|mmlu_professional_medicine|5": {
"acc": 0.5588235294117647,
"acc_stderr": 0.03016191193076711,
"acc_norm": 0.5588235294117647,
"acc_norm_stderr": 0.03016191193076711
},
"harness|mmlu_security_studies|5": {
"acc": 0.6448979591836734,
"acc_stderr": 0.030635655150387634,
"acc_norm": 0.6448979591836734,
"acc_norm_stderr": 0.030635655150387634
},
"harness|mmlu_high_school_world_history|5": {
"acc": 0.7426160337552743,
"acc_stderr": 0.028458820991460302,
"acc_norm": 0.7426160337552743,
"acc_norm_stderr": 0.028458820991460302
},
"harness|mmlu_professional_law|5": {
"acc": 0.44654498044328556,
"acc_stderr": 0.012697046024399661,
"acc_norm": 0.44654498044328556,
"acc_norm_stderr": 0.012697046024399661
},
"harness|mmlu_high_school_us_history|5": {
"acc": 0.6225490196078431,
"acc_stderr": 0.03402272044340703,
"acc_norm": 0.6225490196078431,
"acc_norm_stderr": 0.03402272044340703
},
"harness|mmlu_high_school_european_history|5": {
"acc": 0.6303030303030303,
"acc_stderr": 0.03769430314512569,
"acc_norm": 0.6303030303030303,
"acc_norm_stderr": 0.03769430314512569
},
"harness|truthfulqa_mc|0": {
"mc1": 0.6634026927784578,
"mc1_stderr": 0.0165424128094949,
"mc2": 0.7515104740134964,
"mc2_stderr": 0.014200593490054807
}
} | {
"all": 0,
"harness|arc_challenge|25": 0,
"harness|hellaswag|10": 0,
"harness|mmlu_world_religions|5": 1,
"harness|mmlu_management|5": 1,
"harness|mmlu_miscellaneous|5": 1,
"harness|mmlu_anatomy|5": 1,
"harness|mmlu_abstract_algebra|5": 1,
"harness|mmlu_conceptual_physics|5": 1,
"harness|mmlu_virology|5": 1,
"harness|mmlu_philosophy|5": 1,
"harness|mmlu_human_aging|5": 1,
"harness|mmlu_human_sexuality|5": 1,
"harness|mmlu_medical_genetics|5": 1,
"harness|mmlu_high_school_geography|5": 1,
"harness|mmlu_electrical_engineering|5": 1,
"harness|mmlu_college_physics|5": 1,
"harness|mmlu_high_school_microeconomics|5": 1,
"harness|mmlu_high_school_macroeconomics|5": 1,
"harness|mmlu_computer_security|5": 1,
"harness|mmlu_global_facts|5": 1,
"harness|mmlu_jurisprudence|5": 1,
"harness|mmlu_high_school_chemistry|5": 1,
"harness|mmlu_high_school_biology|5": 1,
"harness|mmlu_marketing|5": 1,
"harness|mmlu_clinical_knowledge|5": 1,
"harness|mmlu_public_relations|5": 1,
"harness|mmlu_high_school_mathematics|5": 1,
"harness|mmlu_high_school_physics|5": 1,
"harness|mmlu_sociology|5": 1,
"harness|mmlu_college_medicine|5": 1,
"harness|mmlu_elementary_mathematics|5": 1,
"harness|mmlu_college_biology|5": 1,
"harness|mmlu_college_chemistry|5": 1,
"harness|mmlu_us_foreign_policy|5": 1,
"harness|mmlu_moral_disputes|5": 1,
"harness|mmlu_logical_fallacies|5": 1,
"harness|mmlu_prehistory|5": 1,
"harness|mmlu_college_mathematics|5": 1,
"harness|mmlu_high_school_government_and_politics|5": 1,
"harness|mmlu_econometrics|5": 1,
"harness|mmlu_high_school_psychology|5": 1,
"harness|mmlu_formal_logic|5": 1,
"harness|mmlu_nutrition|5": 1,
"harness|mmlu_business_ethics|5": 1,
"harness|mmlu_international_law|5": 1,
"harness|mmlu_astronomy|5": 1,
"harness|mmlu_professional_psychology|5": 1,
"harness|mmlu_professional_accounting|5": 1,
"harness|mmlu_machine_learning|5": 1,
"harness|mmlu_high_school_statistics|5": 1,
"harness|mmlu_moral_scenarios|5": 1,
"harness|mmlu_college_computer_science|5": 1,
"harness|mmlu_high_school_computer_science|5": 1,
"harness|mmlu_professional_medicine|5": 1,
"harness|mmlu_security_studies|5": 1,
"harness|mmlu_high_school_world_history|5": 1,
"harness|mmlu_professional_law|5": 1,
"harness|mmlu_high_school_us_history|5": 1,
"harness|mmlu_high_school_european_history|5": 1,
"harness|truthfulqa_mc|0": 0
} | {
"model_name": "x2bee/POLAR-14B-HES-DPO-v1.5",
"model_sha": "f0bc8e2566ba28c8232d7c690098e634ea894e8d",
"model_dtype": "torch.float16",
"lighteval_sha": "",
"num_few_shot_default": 0,
"num_fewshot_seeds": 1,
"override_batch_size": 1,
"max_samples": null
} |
{
"daily": {
"daily": 3
},
"quarterly": {
"quarterly": 3
},
"harness|arc_challenge|25": {
"acc": 0.6646757679180887,
"acc_stderr": 0.013796182947785564,
"acc_norm": 0.7244027303754266,
"acc_norm_stderr": 0.01305716965576184
},
"harness|hellaswag|10": {
"acc": 0.46036646086436966,
"acc_stderr": 0.004974080638364276,
"acc_norm": 0.6195976897032464,
"acc_norm_stderr": 0.004844935327599196
},
"harness|mmlu_world_religions|5": {
"acc": 0.7602339181286549,
"acc_stderr": 0.03274485211946956,
"acc_norm": 0.7602339181286549,
"acc_norm_stderr": 0.03274485211946956
},
"harness|mmlu_management|5": {
"acc": 0.7766990291262136,
"acc_stderr": 0.04123553189891431,
"acc_norm": 0.7766990291262136,
"acc_norm_stderr": 0.04123553189891431
},
"harness|mmlu_miscellaneous|5": {
"acc": 0.7381864623243933,
"acc_stderr": 0.01572083867844526,
"acc_norm": 0.7381864623243933,
"acc_norm_stderr": 0.01572083867844526
},
"harness|mmlu_anatomy|5": {
"acc": 0.5037037037037037,
"acc_stderr": 0.04319223625811331,
"acc_norm": 0.5037037037037037,
"acc_norm_stderr": 0.04319223625811331
},
"harness|mmlu_abstract_algebra|5": {
"acc": 0.35,
"acc_stderr": 0.04793724854411019,
"acc_norm": 0.35,
"acc_norm_stderr": 0.04793724854411019
},
"harness|mmlu_conceptual_physics|5": {
"acc": 0.5404255319148936,
"acc_stderr": 0.032579014820998335,
"acc_norm": 0.5404255319148936,
"acc_norm_stderr": 0.032579014820998335
},
"harness|mmlu_virology|5": {
"acc": 0.5180722891566265,
"acc_stderr": 0.038899512528272166,
"acc_norm": 0.5180722891566265,
"acc_norm_stderr": 0.038899512528272166
},
"harness|mmlu_philosophy|5": {
"acc": 0.6559485530546624,
"acc_stderr": 0.026981478043648043,
"acc_norm": 0.6559485530546624,
"acc_norm_stderr": 0.026981478043648043
},
"harness|mmlu_human_aging|5": {
"acc": 0.6591928251121076,
"acc_stderr": 0.0318114974705536,
"acc_norm": 0.6591928251121076,
"acc_norm_stderr": 0.0318114974705536
},
"harness|mmlu_human_sexuality|5": {
"acc": 0.6564885496183206,
"acc_stderr": 0.041649760719448786,
"acc_norm": 0.6564885496183206,
"acc_norm_stderr": 0.041649760719448786
},
"harness|mmlu_medical_genetics|5": {
"acc": 0.52,
"acc_stderr": 0.050211673156867795,
"acc_norm": 0.52,
"acc_norm_stderr": 0.050211673156867795
},
"harness|mmlu_high_school_geography|5": {
"acc": 0.7575757575757576,
"acc_stderr": 0.030532892233932036,
"acc_norm": 0.7575757575757576,
"acc_norm_stderr": 0.030532892233932036
},
"harness|mmlu_electrical_engineering|5": {
"acc": 0.5586206896551724,
"acc_stderr": 0.04137931034482757,
"acc_norm": 0.5586206896551724,
"acc_norm_stderr": 0.04137931034482757
},
"harness|mmlu_college_physics|5": {
"acc": 0.3137254901960784,
"acc_stderr": 0.04617034827006717,
"acc_norm": 0.3137254901960784,
"acc_norm_stderr": 0.04617034827006717
},
"harness|mmlu_high_school_microeconomics|5": {
"acc": 0.6512605042016807,
"acc_stderr": 0.03095663632856655,
"acc_norm": 0.6512605042016807,
"acc_norm_stderr": 0.03095663632856655
},
"harness|mmlu_high_school_macroeconomics|5": {
"acc": 0.6230769230769231,
"acc_stderr": 0.024570975364225995,
"acc_norm": 0.6230769230769231,
"acc_norm_stderr": 0.024570975364225995
},
"harness|mmlu_computer_security|5": {
"acc": 0.73,
"acc_stderr": 0.04461960433384739,
"acc_norm": 0.73,
"acc_norm_stderr": 0.04461960433384739
},
"harness|mmlu_global_facts|5": {
"acc": 0.32,
"acc_stderr": 0.04688261722621505,
"acc_norm": 0.32,
"acc_norm_stderr": 0.04688261722621505
},
"harness|mmlu_jurisprudence|5": {
"acc": 0.7037037037037037,
"acc_stderr": 0.04414343666854933,
"acc_norm": 0.7037037037037037,
"acc_norm_stderr": 0.04414343666854933
},
"harness|mmlu_high_school_chemistry|5": {
"acc": 0.4630541871921182,
"acc_stderr": 0.035083705204426656,
"acc_norm": 0.4630541871921182,
"acc_norm_stderr": 0.035083705204426656
},
"harness|mmlu_high_school_biology|5": {
"acc": 0.603225806451613,
"acc_stderr": 0.027831231605767944,
"acc_norm": 0.603225806451613,
"acc_norm_stderr": 0.027831231605767944
},
"harness|mmlu_marketing|5": {
"acc": 0.8205128205128205,
"acc_stderr": 0.025140935950335435,
"acc_norm": 0.8205128205128205,
"acc_norm_stderr": 0.025140935950335435
},
"harness|mmlu_clinical_knowledge|5": {
"acc": 0.5962264150943396,
"acc_stderr": 0.03019761160019795,
"acc_norm": 0.5962264150943396,
"acc_norm_stderr": 0.03019761160019795
},
"harness|mmlu_public_relations|5": {
"acc": 0.6181818181818182,
"acc_stderr": 0.046534298079135075,
"acc_norm": 0.6181818181818182,
"acc_norm_stderr": 0.046534298079135075
},
"harness|mmlu_high_school_mathematics|5": {
"acc": 0.37407407407407406,
"acc_stderr": 0.029502861128955293,
"acc_norm": 0.37407407407407406,
"acc_norm_stderr": 0.029502861128955293
},
"harness|mmlu_high_school_physics|5": {
"acc": 0.32450331125827814,
"acc_stderr": 0.038227469376587525,
"acc_norm": 0.32450331125827814,
"acc_norm_stderr": 0.038227469376587525
},
"harness|mmlu_sociology|5": {
"acc": 0.7164179104477612,
"acc_stderr": 0.03187187537919796,
"acc_norm": 0.7164179104477612,
"acc_norm_stderr": 0.03187187537919796
},
"harness|mmlu_college_medicine|5": {
"acc": 0.5375722543352601,
"acc_stderr": 0.03801685104524458,
"acc_norm": 0.5375722543352601,
"acc_norm_stderr": 0.03801685104524458
},
"harness|mmlu_elementary_mathematics|5": {
"acc": 0.42857142857142855,
"acc_stderr": 0.025487187147859372,
"acc_norm": 0.42857142857142855,
"acc_norm_stderr": 0.025487187147859372
},
"harness|mmlu_college_biology|5": {
"acc": 0.5902777777777778,
"acc_stderr": 0.04112490974670787,
"acc_norm": 0.5902777777777778,
"acc_norm_stderr": 0.04112490974670787
},
"harness|mmlu_college_chemistry|5": {
"acc": 0.45,
"acc_stderr": 0.049999999999999996,
"acc_norm": 0.45,
"acc_norm_stderr": 0.049999999999999996
},
"harness|mmlu_us_foreign_policy|5": {
"acc": 0.78,
"acc_stderr": 0.04163331998932263,
"acc_norm": 0.78,
"acc_norm_stderr": 0.04163331998932263
},
"harness|mmlu_moral_disputes|5": {
"acc": 0.6184971098265896,
"acc_stderr": 0.026152198619726803,
"acc_norm": 0.6184971098265896,
"acc_norm_stderr": 0.026152198619726803
},
"harness|mmlu_logical_fallacies|5": {
"acc": 0.6441717791411042,
"acc_stderr": 0.03761521380046734,
"acc_norm": 0.6441717791411042,
"acc_norm_stderr": 0.03761521380046734
},
"harness|mmlu_prehistory|5": {
"acc": 0.6944444444444444,
"acc_stderr": 0.025630824975621365,
"acc_norm": 0.6944444444444444,
"acc_norm_stderr": 0.025630824975621365
},
"harness|mmlu_college_mathematics|5": {
"acc": 0.43,
"acc_stderr": 0.049756985195624284,
"acc_norm": 0.43,
"acc_norm_stderr": 0.049756985195624284
},
"harness|mmlu_high_school_government_and_politics|5": {
"acc": 0.7927461139896373,
"acc_stderr": 0.029252823291803638,
"acc_norm": 0.7927461139896373,
"acc_norm_stderr": 0.029252823291803638
},
"harness|mmlu_econometrics|5": {
"acc": 0.43859649122807015,
"acc_stderr": 0.04668000738510455,
"acc_norm": 0.43859649122807015,
"acc_norm_stderr": 0.04668000738510455
},
"harness|mmlu_high_school_psychology|5": {
"acc": 0.7853211009174312,
"acc_stderr": 0.017604304149256494,
"acc_norm": 0.7853211009174312,
"acc_norm_stderr": 0.017604304149256494
},
"harness|mmlu_formal_logic|5": {
"acc": 0.3968253968253968,
"acc_stderr": 0.04375888492727062,
"acc_norm": 0.3968253968253968,
"acc_norm_stderr": 0.04375888492727062
},
"harness|mmlu_nutrition|5": {
"acc": 0.6437908496732027,
"acc_stderr": 0.027420477662629245,
"acc_norm": 0.6437908496732027,
"acc_norm_stderr": 0.027420477662629245
},
"harness|mmlu_business_ethics|5": {
"acc": 0.63,
"acc_stderr": 0.04852365870939099,
"acc_norm": 0.63,
"acc_norm_stderr": 0.04852365870939099
},
"harness|mmlu_international_law|5": {
"acc": 0.7603305785123967,
"acc_stderr": 0.03896878985070415,
"acc_norm": 0.7603305785123967,
"acc_norm_stderr": 0.03896878985070415
},
"harness|mmlu_astronomy|5": {
"acc": 0.625,
"acc_stderr": 0.039397364351956274,
"acc_norm": 0.625,
"acc_norm_stderr": 0.039397364351956274
},
"harness|mmlu_professional_psychology|5": {
"acc": 0.619281045751634,
"acc_stderr": 0.019643801557924806,
"acc_norm": 0.619281045751634,
"acc_norm_stderr": 0.019643801557924806
},
"harness|mmlu_professional_accounting|5": {
"acc": 0.45390070921985815,
"acc_stderr": 0.029700453247291467,
"acc_norm": 0.45390070921985815,
"acc_norm_stderr": 0.029700453247291467
},
"harness|mmlu_machine_learning|5": {
"acc": 0.41964285714285715,
"acc_stderr": 0.04684099321077106,
"acc_norm": 0.41964285714285715,
"acc_norm_stderr": 0.04684099321077106
},
"harness|mmlu_high_school_statistics|5": {
"acc": 0.5555555555555556,
"acc_stderr": 0.03388857118502326,
"acc_norm": 0.5555555555555556,
"acc_norm_stderr": 0.03388857118502326
},
"harness|mmlu_moral_scenarios|5": {
"acc": 0.3575418994413408,
"acc_stderr": 0.016029394474894893,
"acc_norm": 0.3575418994413408,
"acc_norm_stderr": 0.016029394474894893
},
"harness|mmlu_college_computer_science|5": {
"acc": 0.52,
"acc_stderr": 0.050211673156867795,
"acc_norm": 0.52,
"acc_norm_stderr": 0.050211673156867795
},
"harness|mmlu_high_school_computer_science|5": {
"acc": 0.75,
"acc_stderr": 0.04351941398892446,
"acc_norm": 0.75,
"acc_norm_stderr": 0.04351941398892446
},
"harness|mmlu_professional_medicine|5": {
"acc": 0.5735294117647058,
"acc_stderr": 0.03004261583271486,
"acc_norm": 0.5735294117647058,
"acc_norm_stderr": 0.03004261583271486
},
"harness|mmlu_security_studies|5": {
"acc": 0.6816326530612244,
"acc_stderr": 0.02982253379398204,
"acc_norm": 0.6816326530612244,
"acc_norm_stderr": 0.02982253379398204
},
"harness|mmlu_high_school_world_history|5": {
"acc": 0.7468354430379747,
"acc_stderr": 0.028304657943035293,
"acc_norm": 0.7468354430379747,
"acc_norm_stderr": 0.028304657943035293
},
"harness|mmlu_professional_law|5": {
"acc": 0.455019556714472,
"acc_stderr": 0.012718456618701789,
"acc_norm": 0.455019556714472,
"acc_norm_stderr": 0.012718456618701789
},
"harness|mmlu_high_school_us_history|5": {
"acc": 0.6666666666666666,
"acc_stderr": 0.033086111132364364,
"acc_norm": 0.6666666666666666,
"acc_norm_stderr": 0.033086111132364364
},
"harness|mmlu_high_school_european_history|5": {
"acc": 0.6484848484848484,
"acc_stderr": 0.037282069986826503,
"acc_norm": 0.6484848484848484,
"acc_norm_stderr": 0.037282069986826503
},
"harness|truthfulqa_mc|0": {
"mc1": 0.605875152998776,
"mc1_stderr": 0.017106588140700332,
"mc2": 0.7254831072808595,
"mc2_stderr": 0.014162522228042162
}
} | {
"all": 0,
"harness|arc_challenge|25": 0,
"harness|hellaswag|10": 0,
"harness|mmlu_world_religions|5": 1,
"harness|mmlu_management|5": 1,
"harness|mmlu_miscellaneous|5": 1,
"harness|mmlu_anatomy|5": 1,
"harness|mmlu_abstract_algebra|5": 1,
"harness|mmlu_conceptual_physics|5": 1,
"harness|mmlu_virology|5": 1,
"harness|mmlu_philosophy|5": 1,
"harness|mmlu_human_aging|5": 1,
"harness|mmlu_human_sexuality|5": 1,
"harness|mmlu_medical_genetics|5": 1,
"harness|mmlu_high_school_geography|5": 1,
"harness|mmlu_electrical_engineering|5": 1,
"harness|mmlu_college_physics|5": 1,
"harness|mmlu_high_school_microeconomics|5": 1,
"harness|mmlu_high_school_macroeconomics|5": 1,
"harness|mmlu_computer_security|5": 1,
"harness|mmlu_global_facts|5": 1,
"harness|mmlu_jurisprudence|5": 1,
"harness|mmlu_high_school_chemistry|5": 1,
"harness|mmlu_high_school_biology|5": 1,
"harness|mmlu_marketing|5": 1,
"harness|mmlu_clinical_knowledge|5": 1,
"harness|mmlu_public_relations|5": 1,
"harness|mmlu_high_school_mathematics|5": 1,
"harness|mmlu_high_school_physics|5": 1,
"harness|mmlu_sociology|5": 1,
"harness|mmlu_college_medicine|5": 1,
"harness|mmlu_elementary_mathematics|5": 1,
"harness|mmlu_college_biology|5": 1,
"harness|mmlu_college_chemistry|5": 1,
"harness|mmlu_us_foreign_policy|5": 1,
"harness|mmlu_moral_disputes|5": 1,
"harness|mmlu_logical_fallacies|5": 1,
"harness|mmlu_prehistory|5": 1,
"harness|mmlu_college_mathematics|5": 1,
"harness|mmlu_high_school_government_and_politics|5": 1,
"harness|mmlu_econometrics|5": 1,
"harness|mmlu_high_school_psychology|5": 1,
"harness|mmlu_formal_logic|5": 1,
"harness|mmlu_nutrition|5": 1,
"harness|mmlu_business_ethics|5": 1,
"harness|mmlu_international_law|5": 1,
"harness|mmlu_astronomy|5": 1,
"harness|mmlu_professional_psychology|5": 1,
"harness|mmlu_professional_accounting|5": 1,
"harness|mmlu_machine_learning|5": 1,
"harness|mmlu_high_school_statistics|5": 1,
"harness|mmlu_moral_scenarios|5": 1,
"harness|mmlu_college_computer_science|5": 1,
"harness|mmlu_high_school_computer_science|5": 1,
"harness|mmlu_professional_medicine|5": 1,
"harness|mmlu_security_studies|5": 1,
"harness|mmlu_high_school_world_history|5": 1,
"harness|mmlu_professional_law|5": 1,
"harness|mmlu_high_school_us_history|5": 1,
"harness|mmlu_high_school_european_history|5": 1,
"harness|truthfulqa_mc|0": 0
} | {
"model_name": "x2bee/POLAR-14B-SON-SFT-v0.1",
"model_sha": "01286a13088332c1eda4279b5bcfa7a0a33e145f",
"model_dtype": "torch.float16",
"lighteval_sha": "",
"num_few_shot_default": 0,
"num_fewshot_seeds": 1,
"override_batch_size": 1,
"max_samples": null
} |
{
"daily": {
"daily": 2
},
"quarterly": {
"quarterly": 2
},
"harness|arc_challenge|25": {
"acc": 0.7465870307167235,
"acc_stderr": 0.012710896778378602,
"acc_norm": 0.7687713310580204,
"acc_norm_stderr": 0.012320858834772264
},
"harness|hellaswag|10": {
"acc": 0.681736705835491,
"acc_stderr": 0.004648503177353952,
"acc_norm": 0.7999402509460267,
"acc_norm_stderr": 0.003992272261659531
},
"harness|mmlu_world_religions|5": {
"acc": 0.6549707602339181,
"acc_stderr": 0.036459813773888065,
"acc_norm": 0.6549707602339181,
"acc_norm_stderr": 0.036459813773888065
},
"harness|mmlu_management|5": {
"acc": 0.7378640776699029,
"acc_stderr": 0.043546310772605956,
"acc_norm": 0.7378640776699029,
"acc_norm_stderr": 0.043546310772605956
},
"harness|mmlu_miscellaneous|5": {
"acc": 0.6922094508301405,
"acc_stderr": 0.016506045045155633,
"acc_norm": 0.6922094508301405,
"acc_norm_stderr": 0.016506045045155633
},
"harness|mmlu_anatomy|5": {
"acc": 0.4666666666666667,
"acc_stderr": 0.043097329010363554,
"acc_norm": 0.4666666666666667,
"acc_norm_stderr": 0.043097329010363554
},
"harness|mmlu_abstract_algebra|5": {
"acc": 0.35,
"acc_stderr": 0.047937248544110196,
"acc_norm": 0.35,
"acc_norm_stderr": 0.047937248544110196
},
"harness|mmlu_conceptual_physics|5": {
"acc": 0.4595744680851064,
"acc_stderr": 0.03257901482099836,
"acc_norm": 0.4595744680851064,
"acc_norm_stderr": 0.03257901482099836
},
"harness|mmlu_virology|5": {
"acc": 0.4879518072289157,
"acc_stderr": 0.03891364495835821,
"acc_norm": 0.4879518072289157,
"acc_norm_stderr": 0.03891364495835821
},
"harness|mmlu_philosophy|5": {
"acc": 0.6045016077170418,
"acc_stderr": 0.027770918531427834,
"acc_norm": 0.6045016077170418,
"acc_norm_stderr": 0.027770918531427834
},
"harness|mmlu_human_aging|5": {
"acc": 0.6233183856502242,
"acc_stderr": 0.03252113489929188,
"acc_norm": 0.6233183856502242,
"acc_norm_stderr": 0.03252113489929188
},
"harness|mmlu_human_sexuality|5": {
"acc": 0.6412213740458015,
"acc_stderr": 0.04206739313864908,
"acc_norm": 0.6412213740458015,
"acc_norm_stderr": 0.04206739313864908
},
"harness|mmlu_medical_genetics|5": {
"acc": 0.51,
"acc_stderr": 0.05024183937956911,
"acc_norm": 0.51,
"acc_norm_stderr": 0.05024183937956911
},
"harness|mmlu_high_school_geography|5": {
"acc": 0.7222222222222222,
"acc_stderr": 0.03191178226713547,
"acc_norm": 0.7222222222222222,
"acc_norm_stderr": 0.03191178226713547
},
"harness|mmlu_electrical_engineering|5": {
"acc": 0.5241379310344828,
"acc_stderr": 0.0416180850350153,
"acc_norm": 0.5241379310344828,
"acc_norm_stderr": 0.0416180850350153
},
"harness|mmlu_college_physics|5": {
"acc": 0.3235294117647059,
"acc_stderr": 0.046550104113196177,
"acc_norm": 0.3235294117647059,
"acc_norm_stderr": 0.046550104113196177
},
"harness|mmlu_high_school_microeconomics|5": {
"acc": 0.6764705882352942,
"acc_stderr": 0.030388353551886793,
"acc_norm": 0.6764705882352942,
"acc_norm_stderr": 0.030388353551886793
},
"harness|mmlu_high_school_macroeconomics|5": {
"acc": 0.6384615384615384,
"acc_stderr": 0.024359581465397,
"acc_norm": 0.6384615384615384,
"acc_norm_stderr": 0.024359581465397
},
"harness|mmlu_computer_security|5": {
"acc": 0.65,
"acc_stderr": 0.0479372485441102,
"acc_norm": 0.65,
"acc_norm_stderr": 0.0479372485441102
},
"harness|mmlu_global_facts|5": {
"acc": 0.31,
"acc_stderr": 0.04648231987117316,
"acc_norm": 0.31,
"acc_norm_stderr": 0.04648231987117316
},
"harness|mmlu_jurisprudence|5": {
"acc": 0.6296296296296297,
"acc_stderr": 0.04668408033024931,
"acc_norm": 0.6296296296296297,
"acc_norm_stderr": 0.04668408033024931
},
"harness|mmlu_high_school_chemistry|5": {
"acc": 0.4729064039408867,
"acc_stderr": 0.03512819077876105,
"acc_norm": 0.4729064039408867,
"acc_norm_stderr": 0.03512819077876105
},
"harness|mmlu_high_school_biology|5": {
"acc": 0.5709677419354838,
"acc_stderr": 0.028156036538233193,
"acc_norm": 0.5709677419354838,
"acc_norm_stderr": 0.028156036538233193
},
"harness|mmlu_marketing|5": {
"acc": 0.8034188034188035,
"acc_stderr": 0.026035386098951292,
"acc_norm": 0.8034188034188035,
"acc_norm_stderr": 0.026035386098951292
},
"harness|mmlu_clinical_knowledge|5": {
"acc": 0.5547169811320755,
"acc_stderr": 0.030588052974270655,
"acc_norm": 0.5547169811320755,
"acc_norm_stderr": 0.030588052974270655
},
"harness|mmlu_public_relations|5": {
"acc": 0.6363636363636364,
"acc_stderr": 0.04607582090719976,
"acc_norm": 0.6363636363636364,
"acc_norm_stderr": 0.04607582090719976
},
"harness|mmlu_high_school_mathematics|5": {
"acc": 0.3592592592592593,
"acc_stderr": 0.029252905927251976,
"acc_norm": 0.3592592592592593,
"acc_norm_stderr": 0.029252905927251976
},
"harness|mmlu_high_school_physics|5": {
"acc": 0.3576158940397351,
"acc_stderr": 0.03913453431177258,
"acc_norm": 0.3576158940397351,
"acc_norm_stderr": 0.03913453431177258
},
"harness|mmlu_sociology|5": {
"acc": 0.6268656716417911,
"acc_stderr": 0.034198326081760065,
"acc_norm": 0.6268656716417911,
"acc_norm_stderr": 0.034198326081760065
},
"harness|mmlu_college_medicine|5": {
"acc": 0.48554913294797686,
"acc_stderr": 0.03810871630454764,
"acc_norm": 0.48554913294797686,
"acc_norm_stderr": 0.03810871630454764
},
"harness|mmlu_elementary_mathematics|5": {
"acc": 0.4497354497354497,
"acc_stderr": 0.025620857042936648,
"acc_norm": 0.4497354497354497,
"acc_norm_stderr": 0.025620857042936648
},
"harness|mmlu_college_biology|5": {
"acc": 0.6041666666666666,
"acc_stderr": 0.04089465449325582,
"acc_norm": 0.6041666666666666,
"acc_norm_stderr": 0.04089465449325582
},
"harness|mmlu_college_chemistry|5": {
"acc": 0.32,
"acc_stderr": 0.046882617226215034,
"acc_norm": 0.32,
"acc_norm_stderr": 0.046882617226215034
},
"harness|mmlu_us_foreign_policy|5": {
"acc": 0.71,
"acc_stderr": 0.045604802157206824,
"acc_norm": 0.71,
"acc_norm_stderr": 0.045604802157206824
},
"harness|mmlu_moral_disputes|5": {
"acc": 0.5664739884393064,
"acc_stderr": 0.026680134761679217,
"acc_norm": 0.5664739884393064,
"acc_norm_stderr": 0.026680134761679217
},
"harness|mmlu_logical_fallacies|5": {
"acc": 0.6196319018404908,
"acc_stderr": 0.038142698932618374,
"acc_norm": 0.6196319018404908,
"acc_norm_stderr": 0.038142698932618374
},
"harness|mmlu_prehistory|5": {
"acc": 0.6574074074074074,
"acc_stderr": 0.026406145973625686,
"acc_norm": 0.6574074074074074,
"acc_norm_stderr": 0.026406145973625686
},
"harness|mmlu_college_mathematics|5": {
"acc": 0.37,
"acc_stderr": 0.04852365870939098,
"acc_norm": 0.37,
"acc_norm_stderr": 0.04852365870939098
},
"harness|mmlu_high_school_government_and_politics|5": {
"acc": 0.7616580310880829,
"acc_stderr": 0.030748905363909895,
"acc_norm": 0.7616580310880829,
"acc_norm_stderr": 0.030748905363909895
},
"harness|mmlu_econometrics|5": {
"acc": 0.5,
"acc_stderr": 0.047036043419179864,
"acc_norm": 0.5,
"acc_norm_stderr": 0.047036043419179864
},
"harness|mmlu_high_school_psychology|5": {
"acc": 0.7211009174311926,
"acc_stderr": 0.01922746887646353,
"acc_norm": 0.7211009174311926,
"acc_norm_stderr": 0.01922746887646353
},
"harness|mmlu_formal_logic|5": {
"acc": 0.42857142857142855,
"acc_stderr": 0.0442626668137991,
"acc_norm": 0.42857142857142855,
"acc_norm_stderr": 0.0442626668137991
},
"harness|mmlu_nutrition|5": {
"acc": 0.5816993464052288,
"acc_stderr": 0.0282451340243873,
"acc_norm": 0.5816993464052288,
"acc_norm_stderr": 0.0282451340243873
},
"harness|mmlu_business_ethics|5": {
"acc": 0.73,
"acc_stderr": 0.044619604333847394,
"acc_norm": 0.73,
"acc_norm_stderr": 0.044619604333847394
},
"harness|mmlu_international_law|5": {
"acc": 0.7107438016528925,
"acc_stderr": 0.041391127276354626,
"acc_norm": 0.7107438016528925,
"acc_norm_stderr": 0.041391127276354626
},
"harness|mmlu_astronomy|5": {
"acc": 0.6513157894736842,
"acc_stderr": 0.038781398887976104,
"acc_norm": 0.6513157894736842,
"acc_norm_stderr": 0.038781398887976104
},
"harness|mmlu_professional_psychology|5": {
"acc": 0.5686274509803921,
"acc_stderr": 0.020036393768352624,
"acc_norm": 0.5686274509803921,
"acc_norm_stderr": 0.020036393768352624
},
"harness|mmlu_professional_accounting|5": {
"acc": 0.45390070921985815,
"acc_stderr": 0.029700453247291477,
"acc_norm": 0.45390070921985815,
"acc_norm_stderr": 0.029700453247291477
},
"harness|mmlu_machine_learning|5": {
"acc": 0.4642857142857143,
"acc_stderr": 0.04733667890053756,
"acc_norm": 0.4642857142857143,
"acc_norm_stderr": 0.04733667890053756
},
"harness|mmlu_high_school_statistics|5": {
"acc": 0.5092592592592593,
"acc_stderr": 0.034093869469927006,
"acc_norm": 0.5092592592592593,
"acc_norm_stderr": 0.034093869469927006
},
"harness|mmlu_moral_scenarios|5": {
"acc": 0.37206703910614525,
"acc_stderr": 0.016165847583563295,
"acc_norm": 0.37206703910614525,
"acc_norm_stderr": 0.016165847583563295
},
"harness|mmlu_college_computer_science|5": {
"acc": 0.43,
"acc_stderr": 0.049756985195624284,
"acc_norm": 0.43,
"acc_norm_stderr": 0.049756985195624284
},
"harness|mmlu_high_school_computer_science|5": {
"acc": 0.71,
"acc_stderr": 0.045604802157206845,
"acc_norm": 0.71,
"acc_norm_stderr": 0.045604802157206845
},
"harness|mmlu_professional_medicine|5": {
"acc": 0.5404411764705882,
"acc_stderr": 0.030273325077345755,
"acc_norm": 0.5404411764705882,
"acc_norm_stderr": 0.030273325077345755
},
"harness|mmlu_security_studies|5": {
"acc": 0.6122448979591837,
"acc_stderr": 0.03119223072679566,
"acc_norm": 0.6122448979591837,
"acc_norm_stderr": 0.03119223072679566
},
"harness|mmlu_high_school_world_history|5": {
"acc": 0.7257383966244726,
"acc_stderr": 0.029041333510598025,
"acc_norm": 0.7257383966244726,
"acc_norm_stderr": 0.029041333510598025
},
"harness|mmlu_professional_law|5": {
"acc": 0.4641460234680574,
"acc_stderr": 0.01273736131873058,
"acc_norm": 0.4641460234680574,
"acc_norm_stderr": 0.01273736131873058
},
"harness|mmlu_high_school_us_history|5": {
"acc": 0.6568627450980392,
"acc_stderr": 0.03332139944668086,
"acc_norm": 0.6568627450980392,
"acc_norm_stderr": 0.03332139944668086
},
"harness|mmlu_high_school_european_history|5": {
"acc": 0.6,
"acc_stderr": 0.03825460278380025,
"acc_norm": 0.6,
"acc_norm_stderr": 0.03825460278380025
},
"harness|truthfulqa_mc|0": {
"mc1": 0.7246022031823746,
"mc1_stderr": 0.01563813566777552,
"mc2": 0.8107575910195236,
"mc2_stderr": 0.013335029489665237
}
} | {
"all": 0,
"harness|arc_challenge|25": 0,
"harness|hellaswag|10": 0,
"harness|mmlu_world_religions|5": 1,
"harness|mmlu_management|5": 1,
"harness|mmlu_miscellaneous|5": 1,
"harness|mmlu_anatomy|5": 1,
"harness|mmlu_abstract_algebra|5": 1,
"harness|mmlu_conceptual_physics|5": 1,
"harness|mmlu_virology|5": 1,
"harness|mmlu_philosophy|5": 1,
"harness|mmlu_human_aging|5": 1,
"harness|mmlu_human_sexuality|5": 1,
"harness|mmlu_medical_genetics|5": 1,
"harness|mmlu_high_school_geography|5": 1,
"harness|mmlu_electrical_engineering|5": 1,
"harness|mmlu_college_physics|5": 1,
"harness|mmlu_high_school_microeconomics|5": 1,
"harness|mmlu_high_school_macroeconomics|5": 1,
"harness|mmlu_computer_security|5": 1,
"harness|mmlu_global_facts|5": 1,
"harness|mmlu_jurisprudence|5": 1,
"harness|mmlu_high_school_chemistry|5": 1,
"harness|mmlu_high_school_biology|5": 1,
"harness|mmlu_marketing|5": 1,
"harness|mmlu_clinical_knowledge|5": 1,
"harness|mmlu_public_relations|5": 1,
"harness|mmlu_high_school_mathematics|5": 1,
"harness|mmlu_high_school_physics|5": 1,
"harness|mmlu_sociology|5": 1,
"harness|mmlu_college_medicine|5": 1,
"harness|mmlu_elementary_mathematics|5": 1,
"harness|mmlu_college_biology|5": 1,
"harness|mmlu_college_chemistry|5": 1,
"harness|mmlu_us_foreign_policy|5": 1,
"harness|mmlu_moral_disputes|5": 1,
"harness|mmlu_logical_fallacies|5": 1,
"harness|mmlu_prehistory|5": 1,
"harness|mmlu_college_mathematics|5": 1,
"harness|mmlu_high_school_government_and_politics|5": 1,
"harness|mmlu_econometrics|5": 1,
"harness|mmlu_high_school_psychology|5": 1,
"harness|mmlu_formal_logic|5": 1,
"harness|mmlu_nutrition|5": 1,
"harness|mmlu_business_ethics|5": 1,
"harness|mmlu_international_law|5": 1,
"harness|mmlu_astronomy|5": 1,
"harness|mmlu_professional_psychology|5": 1,
"harness|mmlu_professional_accounting|5": 1,
"harness|mmlu_machine_learning|5": 1,
"harness|mmlu_high_school_statistics|5": 1,
"harness|mmlu_moral_scenarios|5": 1,
"harness|mmlu_college_computer_science|5": 1,
"harness|mmlu_high_school_computer_science|5": 1,
"harness|mmlu_professional_medicine|5": 1,
"harness|mmlu_security_studies|5": 1,
"harness|mmlu_high_school_world_history|5": 1,
"harness|mmlu_professional_law|5": 1,
"harness|mmlu_high_school_us_history|5": 1,
"harness|mmlu_high_school_european_history|5": 1,
"harness|truthfulqa_mc|0": 0
} | {
"model_name": "x2bee/POLAR-14B-v0.2",
"model_sha": "8d905623a3972e11260420130039c62e115cbbaa",
"model_dtype": "torch.float16",
"lighteval_sha": "",
"num_few_shot_default": 0,
"num_fewshot_seeds": 1,
"override_batch_size": 1,
"max_samples": null
} |
{
"daily": {
"daily": 1
},
"quarterly": {
"quarterly": 1
},
"harness|arc_challenge|25": {
"acc": 0.75,
"acc_stderr": 0.012653835621466646,
"acc_norm": 0.7798634812286689,
"acc_norm_stderr": 0.012108124883460988
},
"harness|hellaswag|10": {
"acc": 0.6500697072296355,
"acc_stderr": 0.004759729267943182,
"acc_norm": 0.775542720573591,
"acc_norm_stderr": 0.004163717220873764
},
"harness|mmlu_world_religions|5": {
"acc": 0.6374269005847953,
"acc_stderr": 0.036871306155620606,
"acc_norm": 0.6374269005847953,
"acc_norm_stderr": 0.036871306155620606
},
"harness|mmlu_management|5": {
"acc": 0.7087378640776699,
"acc_stderr": 0.044986763205729224,
"acc_norm": 0.7087378640776699,
"acc_norm_stderr": 0.044986763205729224
},
"harness|mmlu_miscellaneous|5": {
"acc": 0.6730523627075351,
"acc_stderr": 0.016774908180131484,
"acc_norm": 0.6730523627075351,
"acc_norm_stderr": 0.016774908180131484
},
"harness|mmlu_anatomy|5": {
"acc": 0.45185185185185184,
"acc_stderr": 0.04299268905480864,
"acc_norm": 0.45185185185185184,
"acc_norm_stderr": 0.04299268905480864
},
"harness|mmlu_abstract_algebra|5": {
"acc": 0.36,
"acc_stderr": 0.048241815132442176,
"acc_norm": 0.36,
"acc_norm_stderr": 0.048241815132442176
},
"harness|mmlu_conceptual_physics|5": {
"acc": 0.4723404255319149,
"acc_stderr": 0.03263597118409769,
"acc_norm": 0.4723404255319149,
"acc_norm_stderr": 0.03263597118409769
},
"harness|mmlu_virology|5": {
"acc": 0.46987951807228917,
"acc_stderr": 0.03885425420866766,
"acc_norm": 0.46987951807228917,
"acc_norm_stderr": 0.03885425420866766
},
"harness|mmlu_philosophy|5": {
"acc": 0.594855305466238,
"acc_stderr": 0.027882383791325963,
"acc_norm": 0.594855305466238,
"acc_norm_stderr": 0.027882383791325963
},
"harness|mmlu_human_aging|5": {
"acc": 0.6412556053811659,
"acc_stderr": 0.032190792004199956,
"acc_norm": 0.6412556053811659,
"acc_norm_stderr": 0.032190792004199956
},
"harness|mmlu_human_sexuality|5": {
"acc": 0.5954198473282443,
"acc_stderr": 0.043046937953806645,
"acc_norm": 0.5954198473282443,
"acc_norm_stderr": 0.043046937953806645
},
"harness|mmlu_medical_genetics|5": {
"acc": 0.5,
"acc_stderr": 0.050251890762960605,
"acc_norm": 0.5,
"acc_norm_stderr": 0.050251890762960605
},
"harness|mmlu_high_school_geography|5": {
"acc": 0.7272727272727273,
"acc_stderr": 0.03173071239071724,
"acc_norm": 0.7272727272727273,
"acc_norm_stderr": 0.03173071239071724
},
"harness|mmlu_electrical_engineering|5": {
"acc": 0.503448275862069,
"acc_stderr": 0.0416656757710158,
"acc_norm": 0.503448275862069,
"acc_norm_stderr": 0.0416656757710158
},
"harness|mmlu_college_physics|5": {
"acc": 0.3431372549019608,
"acc_stderr": 0.04724007352383888,
"acc_norm": 0.3431372549019608,
"acc_norm_stderr": 0.04724007352383888
},
"harness|mmlu_high_school_microeconomics|5": {
"acc": 0.6596638655462185,
"acc_stderr": 0.03077805742293167,
"acc_norm": 0.6596638655462185,
"acc_norm_stderr": 0.03077805742293167
},
"harness|mmlu_high_school_macroeconomics|5": {
"acc": 0.6102564102564103,
"acc_stderr": 0.024726967886647078,
"acc_norm": 0.6102564102564103,
"acc_norm_stderr": 0.024726967886647078
},
"harness|mmlu_computer_security|5": {
"acc": 0.67,
"acc_stderr": 0.047258156262526094,
"acc_norm": 0.67,
"acc_norm_stderr": 0.047258156262526094
},
"harness|mmlu_global_facts|5": {
"acc": 0.33,
"acc_stderr": 0.047258156262526045,
"acc_norm": 0.33,
"acc_norm_stderr": 0.047258156262526045
},
"harness|mmlu_jurisprudence|5": {
"acc": 0.6481481481481481,
"acc_stderr": 0.04616631111801714,
"acc_norm": 0.6481481481481481,
"acc_norm_stderr": 0.04616631111801714
},
"harness|mmlu_high_school_chemistry|5": {
"acc": 0.4729064039408867,
"acc_stderr": 0.03512819077876105,
"acc_norm": 0.4729064039408867,
"acc_norm_stderr": 0.03512819077876105
},
"harness|mmlu_high_school_biology|5": {
"acc": 0.5709677419354838,
"acc_stderr": 0.028156036538233193,
"acc_norm": 0.5709677419354838,
"acc_norm_stderr": 0.028156036538233193
},
"harness|mmlu_marketing|5": {
"acc": 0.7735042735042735,
"acc_stderr": 0.027421007295392943,
"acc_norm": 0.7735042735042735,
"acc_norm_stderr": 0.027421007295392943
},
"harness|mmlu_clinical_knowledge|5": {
"acc": 0.5660377358490566,
"acc_stderr": 0.030503292013342596,
"acc_norm": 0.5660377358490566,
"acc_norm_stderr": 0.030503292013342596
},
"harness|mmlu_public_relations|5": {
"acc": 0.6272727272727273,
"acc_stderr": 0.04631381319425465,
"acc_norm": 0.6272727272727273,
"acc_norm_stderr": 0.04631381319425465
},
"harness|mmlu_high_school_mathematics|5": {
"acc": 0.3333333333333333,
"acc_stderr": 0.0287420409039485,
"acc_norm": 0.3333333333333333,
"acc_norm_stderr": 0.0287420409039485
},
"harness|mmlu_high_school_physics|5": {
"acc": 0.39072847682119205,
"acc_stderr": 0.039837983066598075,
"acc_norm": 0.39072847682119205,
"acc_norm_stderr": 0.039837983066598075
},
"harness|mmlu_sociology|5": {
"acc": 0.6417910447761194,
"acc_stderr": 0.03390393042268814,
"acc_norm": 0.6417910447761194,
"acc_norm_stderr": 0.03390393042268814
},
"harness|mmlu_college_medicine|5": {
"acc": 0.5028901734104047,
"acc_stderr": 0.038124005659748335,
"acc_norm": 0.5028901734104047,
"acc_norm_stderr": 0.038124005659748335
},
"harness|mmlu_elementary_mathematics|5": {
"acc": 0.42857142857142855,
"acc_stderr": 0.025487187147859372,
"acc_norm": 0.42857142857142855,
"acc_norm_stderr": 0.025487187147859372
},
"harness|mmlu_college_biology|5": {
"acc": 0.6180555555555556,
"acc_stderr": 0.040629907841466674,
"acc_norm": 0.6180555555555556,
"acc_norm_stderr": 0.040629907841466674
},
"harness|mmlu_college_chemistry|5": {
"acc": 0.3,
"acc_stderr": 0.046056618647183814,
"acc_norm": 0.3,
"acc_norm_stderr": 0.046056618647183814
},
"harness|mmlu_us_foreign_policy|5": {
"acc": 0.72,
"acc_stderr": 0.04512608598542127,
"acc_norm": 0.72,
"acc_norm_stderr": 0.04512608598542127
},
"harness|mmlu_moral_disputes|5": {
"acc": 0.5809248554913294,
"acc_stderr": 0.026564178111422622,
"acc_norm": 0.5809248554913294,
"acc_norm_stderr": 0.026564178111422622
},
"harness|mmlu_logical_fallacies|5": {
"acc": 0.6257668711656442,
"acc_stderr": 0.03802068102899615,
"acc_norm": 0.6257668711656442,
"acc_norm_stderr": 0.03802068102899615
},
"harness|mmlu_prehistory|5": {
"acc": 0.5987654320987654,
"acc_stderr": 0.027272582849839803,
"acc_norm": 0.5987654320987654,
"acc_norm_stderr": 0.027272582849839803
},
"harness|mmlu_college_mathematics|5": {
"acc": 0.34,
"acc_stderr": 0.04760952285695235,
"acc_norm": 0.34,
"acc_norm_stderr": 0.04760952285695235
},
"harness|mmlu_high_school_government_and_politics|5": {
"acc": 0.7512953367875648,
"acc_stderr": 0.031195840877700304,
"acc_norm": 0.7512953367875648,
"acc_norm_stderr": 0.031195840877700304
},
"harness|mmlu_econometrics|5": {
"acc": 0.47368421052631576,
"acc_stderr": 0.046970851366478626,
"acc_norm": 0.47368421052631576,
"acc_norm_stderr": 0.046970851366478626
},
"harness|mmlu_high_school_psychology|5": {
"acc": 0.7229357798165138,
"acc_stderr": 0.019188482590169538,
"acc_norm": 0.7229357798165138,
"acc_norm_stderr": 0.019188482590169538
},
"harness|mmlu_formal_logic|5": {
"acc": 0.4523809523809524,
"acc_stderr": 0.044518079590553275,
"acc_norm": 0.4523809523809524,
"acc_norm_stderr": 0.044518079590553275
},
"harness|mmlu_nutrition|5": {
"acc": 0.5718954248366013,
"acc_stderr": 0.028332397483664278,
"acc_norm": 0.5718954248366013,
"acc_norm_stderr": 0.028332397483664278
},
"harness|mmlu_business_ethics|5": {
"acc": 0.68,
"acc_stderr": 0.04688261722621504,
"acc_norm": 0.68,
"acc_norm_stderr": 0.04688261722621504
},
"harness|mmlu_international_law|5": {
"acc": 0.7520661157024794,
"acc_stderr": 0.039418975265163025,
"acc_norm": 0.7520661157024794,
"acc_norm_stderr": 0.039418975265163025
},
"harness|mmlu_astronomy|5": {
"acc": 0.618421052631579,
"acc_stderr": 0.03953173377749194,
"acc_norm": 0.618421052631579,
"acc_norm_stderr": 0.03953173377749194
},
"harness|mmlu_professional_psychology|5": {
"acc": 0.5408496732026143,
"acc_stderr": 0.020160213617222516,
"acc_norm": 0.5408496732026143,
"acc_norm_stderr": 0.020160213617222516
},
"harness|mmlu_professional_accounting|5": {
"acc": 0.45390070921985815,
"acc_stderr": 0.029700453247291463,
"acc_norm": 0.45390070921985815,
"acc_norm_stderr": 0.029700453247291463
},
"harness|mmlu_machine_learning|5": {
"acc": 0.44642857142857145,
"acc_stderr": 0.04718471485219588,
"acc_norm": 0.44642857142857145,
"acc_norm_stderr": 0.04718471485219588
},
"harness|mmlu_high_school_statistics|5": {
"acc": 0.5416666666666666,
"acc_stderr": 0.03398110890294636,
"acc_norm": 0.5416666666666666,
"acc_norm_stderr": 0.03398110890294636
},
"harness|mmlu_moral_scenarios|5": {
"acc": 0.35195530726256985,
"acc_stderr": 0.01597266852368907,
"acc_norm": 0.35195530726256985,
"acc_norm_stderr": 0.01597266852368907
},
"harness|mmlu_college_computer_science|5": {
"acc": 0.44,
"acc_stderr": 0.0498887651569859,
"acc_norm": 0.44,
"acc_norm_stderr": 0.0498887651569859
},
"harness|mmlu_high_school_computer_science|5": {
"acc": 0.68,
"acc_stderr": 0.04688261722621503,
"acc_norm": 0.68,
"acc_norm_stderr": 0.04688261722621503
},
"harness|mmlu_professional_medicine|5": {
"acc": 0.5147058823529411,
"acc_stderr": 0.03035969707904612,
"acc_norm": 0.5147058823529411,
"acc_norm_stderr": 0.03035969707904612
},
"harness|mmlu_security_studies|5": {
"acc": 0.6122448979591837,
"acc_stderr": 0.031192230726795656,
"acc_norm": 0.6122448979591837,
"acc_norm_stderr": 0.031192230726795656
},
"harness|mmlu_high_school_world_history|5": {
"acc": 0.7215189873417721,
"acc_stderr": 0.029178682304842538,
"acc_norm": 0.7215189873417721,
"acc_norm_stderr": 0.029178682304842538
},
"harness|mmlu_professional_law|5": {
"acc": 0.4634941329856584,
"acc_stderr": 0.012736153390214963,
"acc_norm": 0.4634941329856584,
"acc_norm_stderr": 0.012736153390214963
},
"harness|mmlu_high_school_us_history|5": {
"acc": 0.6568627450980392,
"acc_stderr": 0.03332139944668086,
"acc_norm": 0.6568627450980392,
"acc_norm_stderr": 0.03332139944668086
},
"harness|mmlu_high_school_european_history|5": {
"acc": 0.5818181818181818,
"acc_stderr": 0.03851716319398393,
"acc_norm": 0.5818181818181818,
"acc_norm_stderr": 0.03851716319398393
},
"harness|truthfulqa_mc|0": {
"mc1": 0.7833537331701347,
"mc1_stderr": 0.014421468452506978,
"mc2": 0.8572574997405501,
"mc2_stderr": 0.01200311225898601
}
} | {
"all": 0,
"harness|arc_challenge|25": 0,
"harness|hellaswag|10": 0,
"harness|mmlu_world_religions|5": 1,
"harness|mmlu_management|5": 1,
"harness|mmlu_miscellaneous|5": 1,
"harness|mmlu_anatomy|5": 1,
"harness|mmlu_abstract_algebra|5": 1,
"harness|mmlu_conceptual_physics|5": 1,
"harness|mmlu_virology|5": 1,
"harness|mmlu_philosophy|5": 1,
"harness|mmlu_human_aging|5": 1,
"harness|mmlu_human_sexuality|5": 1,
"harness|mmlu_medical_genetics|5": 1,
"harness|mmlu_high_school_geography|5": 1,
"harness|mmlu_electrical_engineering|5": 1,
"harness|mmlu_college_physics|5": 1,
"harness|mmlu_high_school_microeconomics|5": 1,
"harness|mmlu_high_school_macroeconomics|5": 1,
"harness|mmlu_computer_security|5": 1,
"harness|mmlu_global_facts|5": 1,
"harness|mmlu_jurisprudence|5": 1,
"harness|mmlu_high_school_chemistry|5": 1,
"harness|mmlu_high_school_biology|5": 1,
"harness|mmlu_marketing|5": 1,
"harness|mmlu_clinical_knowledge|5": 1,
"harness|mmlu_public_relations|5": 1,
"harness|mmlu_high_school_mathematics|5": 1,
"harness|mmlu_high_school_physics|5": 1,
"harness|mmlu_sociology|5": 1,
"harness|mmlu_college_medicine|5": 1,
"harness|mmlu_elementary_mathematics|5": 1,
"harness|mmlu_college_biology|5": 1,
"harness|mmlu_college_chemistry|5": 1,
"harness|mmlu_us_foreign_policy|5": 1,
"harness|mmlu_moral_disputes|5": 1,
"harness|mmlu_logical_fallacies|5": 1,
"harness|mmlu_prehistory|5": 1,
"harness|mmlu_college_mathematics|5": 1,
"harness|mmlu_high_school_government_and_politics|5": 1,
"harness|mmlu_econometrics|5": 1,
"harness|mmlu_high_school_psychology|5": 1,
"harness|mmlu_formal_logic|5": 1,
"harness|mmlu_nutrition|5": 1,
"harness|mmlu_business_ethics|5": 1,
"harness|mmlu_international_law|5": 1,
"harness|mmlu_astronomy|5": 1,
"harness|mmlu_professional_psychology|5": 1,
"harness|mmlu_professional_accounting|5": 1,
"harness|mmlu_machine_learning|5": 1,
"harness|mmlu_high_school_statistics|5": 1,
"harness|mmlu_moral_scenarios|5": 1,
"harness|mmlu_college_computer_science|5": 1,
"harness|mmlu_high_school_computer_science|5": 1,
"harness|mmlu_professional_medicine|5": 1,
"harness|mmlu_security_studies|5": 1,
"harness|mmlu_high_school_world_history|5": 1,
"harness|mmlu_professional_law|5": 1,
"harness|mmlu_high_school_us_history|5": 1,
"harness|mmlu_high_school_european_history|5": 1,
"harness|truthfulqa_mc|0": 0
} | {
"model_name": "x2bee/POLAR-14B-v0.5",
"model_sha": "74a1ef65a8d650e5358be229def31688738d8c6a",
"model_dtype": "torch.float16",
"lighteval_sha": "",
"num_few_shot_default": 0,
"num_fewshot_seeds": 1,
"override_batch_size": 1,
"max_samples": null
} |