|
{ |
|
"config_general": { |
|
"lighteval_sha": "?", |
|
"num_fewshot_seeds": 1, |
|
"override_batch_size": 1, |
|
"max_samples": null, |
|
"job_id": 0, |
|
"start_time": 699.678705546, |
|
"end_time": 79411.168130086, |
|
"total_evaluation_time_secondes": "78711.48942453999", |
|
"model_name": "meta-llama/Llama-3.3-70B-Instruct", |
|
"model_sha": "6f6073b423013f6a7d4d9f39144961bfbfbc386b", |
|
"model_dtype": "torch.bfloat16", |
|
"model_size": "131.42 GB" |
|
}, |
|
"results": { |
|
"community|alghafa:mcq_exams_test_ar|0": { |
|
"acc_norm": 0.526032315978456, |
|
"acc_norm_stderr": 0.021175963642669934 |
|
}, |
|
"community|alghafa:meta_ar_dialects|0": { |
|
"acc_norm": 0.5803521779425393, |
|
"acc_norm_stderr": 0.006719436049208947 |
|
}, |
|
"community|alghafa:meta_ar_msa|0": { |
|
"acc_norm": 0.6715083798882682, |
|
"acc_norm_stderr": 0.015707935398496447 |
|
}, |
|
"community|alghafa:multiple_choice_facts_truefalse_balanced_task|0": { |
|
"acc_norm": 0.7733333333333333, |
|
"acc_norm_stderr": 0.04866999865182628 |
|
}, |
|
"community|alghafa:multiple_choice_grounded_statement_soqal_task|0": { |
|
"acc_norm": 0.7266666666666667, |
|
"acc_norm_stderr": 0.03651075250486201 |
|
}, |
|
"community|alghafa:multiple_choice_grounded_statement_xglue_mlqa_task|0": { |
|
"acc_norm": 0.6466666666666666, |
|
"acc_norm_stderr": 0.03915970474562285 |
|
}, |
|
"community|alghafa:multiple_choice_rating_sentiment_no_neutral_task|0": { |
|
"acc_norm": 0.8226391494684178, |
|
"acc_norm_stderr": 0.004272199219186073 |
|
}, |
|
"community|alghafa:multiple_choice_rating_sentiment_task|0": { |
|
"acc_norm": 0.583653044203503, |
|
"acc_norm_stderr": 0.006367173805401006 |
|
}, |
|
"community|alghafa:multiple_choice_sentiment_task|0": { |
|
"acc_norm": 0.41104651162790695, |
|
"acc_norm_stderr": 0.011867194061128752 |
|
}, |
|
"community|arabic_exams|0": { |
|
"acc_norm": 0.6666666666666666, |
|
"acc_norm_stderr": 0.02036157405938425 |
|
}, |
|
"community|arabic_mmlu:Accounting (University)|0": { |
|
"acc_norm": 0.7297297297297297, |
|
"acc_norm_stderr": 0.05197789984508372 |
|
}, |
|
"community|arabic_mmlu:Arabic Language (General)|0": { |
|
"acc_norm": 0.6552287581699346, |
|
"acc_norm_stderr": 0.019228322018696637 |
|
}, |
|
"community|arabic_mmlu:Arabic Language (Grammar)|0": { |
|
"acc_norm": 0.663013698630137, |
|
"acc_norm_stderr": 0.02477517223961391 |
|
}, |
|
"community|arabic_mmlu:Arabic Language (High School)|0": { |
|
"acc_norm": 0.4205128205128205, |
|
"acc_norm_stderr": 0.025028610276710862 |
|
}, |
|
"community|arabic_mmlu:Arabic Language (Middle School)|0": { |
|
"acc_norm": 0.7777777777777778, |
|
"acc_norm_stderr": 0.08153326507837143 |
|
}, |
|
"community|arabic_mmlu:Arabic Language (Primary School)|0": { |
|
"acc_norm": 0.6825396825396826, |
|
"acc_norm_stderr": 0.029381354652032125 |
|
}, |
|
"community|arabic_mmlu:Biology (High School)|0": { |
|
"acc_norm": 0.5777146912704045, |
|
"acc_norm_stderr": 0.0131631055815741 |
|
}, |
|
"community|arabic_mmlu:Civics (High School)|0": { |
|
"acc_norm": 0.5287356321839081, |
|
"acc_norm_stderr": 0.05382727149237504 |
|
}, |
|
"community|arabic_mmlu:Civics (Middle School)|0": { |
|
"acc_norm": 0.5677966101694916, |
|
"acc_norm_stderr": 0.03231517788768734 |
|
}, |
|
"community|arabic_mmlu:Computer Science (High School)|0": { |
|
"acc_norm": 0.7011494252873564, |
|
"acc_norm_stderr": 0.028388710425812574 |
|
}, |
|
"community|arabic_mmlu:Computer Science (Middle School)|0": { |
|
"acc_norm": 0.8888888888888888, |
|
"acc_norm_stderr": 0.06163335513613659 |
|
}, |
|
"community|arabic_mmlu:Computer Science (Primary School)|0": { |
|
"acc_norm": 0.7105263157894737, |
|
"acc_norm_stderr": 0.03298858792252676 |
|
}, |
|
"community|arabic_mmlu:Computer Science (University)|0": { |
|
"acc_norm": 0.75, |
|
"acc_norm_stderr": 0.05455447255899809 |
|
}, |
|
"community|arabic_mmlu:Driving Test|0": { |
|
"acc_norm": 0.726672171758877, |
|
"acc_norm_stderr": 0.012812046453979981 |
|
}, |
|
"community|arabic_mmlu:Economics (High School)|0": { |
|
"acc_norm": 0.7111111111111111, |
|
"acc_norm_stderr": 0.023921418402752272 |
|
}, |
|
"community|arabic_mmlu:Economics (Middle School)|0": { |
|
"acc_norm": 0.8045977011494253, |
|
"acc_norm_stderr": 0.042756781109738705 |
|
}, |
|
"community|arabic_mmlu:Economics (University)|0": { |
|
"acc_norm": 0.6277372262773723, |
|
"acc_norm_stderr": 0.04145188573576843 |
|
}, |
|
"community|arabic_mmlu:General Knowledge|0": { |
|
"acc_norm": 0.6817129629629629, |
|
"acc_norm_stderr": 0.015856409529423604 |
|
}, |
|
"community|arabic_mmlu:General Knowledge (Middle School)|0": { |
|
"acc_norm": 0.7383720930232558, |
|
"acc_norm_stderr": 0.033611014038904936 |
|
}, |
|
"community|arabic_mmlu:General Knowledge (Primary School)|0": { |
|
"acc_norm": 0.7530864197530864, |
|
"acc_norm_stderr": 0.03398458402215315 |
|
}, |
|
"community|arabic_mmlu:Geography (High School)|0": { |
|
"acc_norm": 0.6127167630057804, |
|
"acc_norm_stderr": 0.015127071256522432 |
|
}, |
|
"community|arabic_mmlu:Geography (Middle School)|0": { |
|
"acc_norm": 0.7352941176470589, |
|
"acc_norm_stderr": 0.026799562024887685 |
|
}, |
|
"community|arabic_mmlu:Geography (Primary School)|0": { |
|
"acc_norm": 0.7192982456140351, |
|
"acc_norm_stderr": 0.06004585739704729 |
|
}, |
|
"community|arabic_mmlu:History (High School)|0": { |
|
"acc_norm": 0.5223684210526316, |
|
"acc_norm_stderr": 0.01813067970124118 |
|
}, |
|
"community|arabic_mmlu:History (Middle School)|0": { |
|
"acc_norm": 0.729064039408867, |
|
"acc_norm_stderr": 0.031270907132976984 |
|
}, |
|
"community|arabic_mmlu:History (Primary School)|0": { |
|
"acc_norm": 0.6666666666666666, |
|
"acc_norm_stderr": 0.04690650298201942 |
|
}, |
|
"community|arabic_mmlu:Islamic Studies|0": { |
|
"acc_norm": 0.6572769953051644, |
|
"acc_norm_stderr": 0.018790377495902715 |
|
}, |
|
"community|arabic_mmlu:Islamic Studies (High School)|0": { |
|
"acc_norm": 0.7544910179640718, |
|
"acc_norm_stderr": 0.02358514838942898 |
|
}, |
|
"community|arabic_mmlu:Islamic Studies (Middle School)|0": { |
|
"acc_norm": 0.7521008403361344, |
|
"acc_norm_stderr": 0.028047967224176892 |
|
}, |
|
"community|arabic_mmlu:Islamic Studies (Primary School)|0": { |
|
"acc_norm": 0.8058058058058059, |
|
"acc_norm_stderr": 0.012521837003384988 |
|
}, |
|
"community|arabic_mmlu:Law (Professional)|0": { |
|
"acc_norm": 0.7388535031847133, |
|
"acc_norm_stderr": 0.024828421054612856 |
|
}, |
|
"community|arabic_mmlu:Management (University)|0": { |
|
"acc_norm": 0.76, |
|
"acc_norm_stderr": 0.04964740541926505 |
|
}, |
|
"community|arabic_mmlu:Math (Primary School)|0": { |
|
"acc_norm": 0.6552567237163814, |
|
"acc_norm_stderr": 0.023530088671619043 |
|
}, |
|
"community|arabic_mmlu:Natural Science (Middle School)|0": { |
|
"acc_norm": 0.8388429752066116, |
|
"acc_norm_stderr": 0.023684075409888674 |
|
}, |
|
"community|arabic_mmlu:Natural Science (Primary School)|0": { |
|
"acc_norm": 0.8809523809523809, |
|
"acc_norm_stderr": 0.01769351710661637 |
|
}, |
|
"community|arabic_mmlu:Philosophy (High School)|0": { |
|
"acc_norm": 0.717948717948718, |
|
"acc_norm_stderr": 0.07299934324587598 |
|
}, |
|
"community|arabic_mmlu:Physics (High School)|0": { |
|
"acc_norm": 0.611764705882353, |
|
"acc_norm_stderr": 0.030578970343036065 |
|
}, |
|
"community|arabic_mmlu:Political Science (University)|0": { |
|
"acc_norm": 0.6761904761904762, |
|
"acc_norm_stderr": 0.032367278954043524 |
|
}, |
|
"community|arabic_mmlu:Social Science (Middle School)|0": { |
|
"acc_norm": 0.5726141078838174, |
|
"acc_norm_stderr": 0.031932689059452145 |
|
}, |
|
"community|arabic_mmlu:Social Science (Primary School)|0": { |
|
"acc_norm": 0.8, |
|
"acc_norm_stderr": 0.015075567228888271 |
|
}, |
|
"community|arabic_mmlu_ht:abstract_algebra|0": { |
|
"acc_norm": 0.33, |
|
"acc_norm_stderr": 0.047258156262526045 |
|
}, |
|
"community|arabic_mmlu_ht:anatomy|0": { |
|
"acc_norm": 0.5555555555555556, |
|
"acc_norm_stderr": 0.04292596718256981 |
|
}, |
|
"community|arabic_mmlu_ht:astronomy|0": { |
|
"acc_norm": 0.7828947368421053, |
|
"acc_norm_stderr": 0.03355045304882925 |
|
}, |
|
"community|arabic_mmlu_ht:business_ethics|0": { |
|
"acc_norm": 0.76, |
|
"acc_norm_stderr": 0.04292346959909283 |
|
}, |
|
"community|arabic_mmlu_ht:clinical_knowledge|0": { |
|
"acc_norm": 0.6830188679245283, |
|
"acc_norm_stderr": 0.02863723563980089 |
|
}, |
|
"community|arabic_mmlu_ht:college_biology|0": { |
|
"acc_norm": 0.7430555555555556, |
|
"acc_norm_stderr": 0.03653946969442099 |
|
}, |
|
"community|arabic_mmlu_ht:college_chemistry|0": { |
|
"acc_norm": 0.5, |
|
"acc_norm_stderr": 0.050251890762960605 |
|
}, |
|
"community|arabic_mmlu_ht:college_computer_science|0": { |
|
"acc_norm": 0.47, |
|
"acc_norm_stderr": 0.05016135580465919 |
|
}, |
|
"community|arabic_mmlu_ht:college_mathematics|0": { |
|
"acc_norm": 0.45, |
|
"acc_norm_stderr": 0.05 |
|
}, |
|
"community|arabic_mmlu_ht:college_medicine|0": { |
|
"acc_norm": 0.6416184971098265, |
|
"acc_norm_stderr": 0.036563436533531585 |
|
}, |
|
"community|arabic_mmlu_ht:college_physics|0": { |
|
"acc_norm": 0.4411764705882353, |
|
"acc_norm_stderr": 0.049406356306056595 |
|
}, |
|
"community|arabic_mmlu_ht:computer_security|0": { |
|
"acc_norm": 0.69, |
|
"acc_norm_stderr": 0.04648231987117316 |
|
}, |
|
"community|arabic_mmlu_ht:conceptual_physics|0": { |
|
"acc_norm": 0.7106382978723405, |
|
"acc_norm_stderr": 0.02964400657700962 |
|
}, |
|
"community|arabic_mmlu_ht:econometrics|0": { |
|
"acc_norm": 0.5350877192982456, |
|
"acc_norm_stderr": 0.046920083813689104 |
|
}, |
|
"community|arabic_mmlu_ht:electrical_engineering|0": { |
|
"acc_norm": 0.6206896551724138, |
|
"acc_norm_stderr": 0.040434618619167466 |
|
}, |
|
"community|arabic_mmlu_ht:elementary_mathematics|0": { |
|
"acc_norm": 0.4603174603174603, |
|
"acc_norm_stderr": 0.02567008063690919 |
|
}, |
|
"community|arabic_mmlu_ht:formal_logic|0": { |
|
"acc_norm": 0.5873015873015873, |
|
"acc_norm_stderr": 0.04403438954768177 |
|
}, |
|
"community|arabic_mmlu_ht:global_facts|0": { |
|
"acc_norm": 0.47, |
|
"acc_norm_stderr": 0.05016135580465919 |
|
}, |
|
"community|arabic_mmlu_ht:high_school_biology|0": { |
|
"acc_norm": 0.7526881720430108, |
|
"acc_norm_stderr": 0.006988010413240884 |
|
}, |
|
"community|arabic_mmlu_ht:high_school_chemistry|0": { |
|
"acc_norm": 0.7462649402390438, |
|
"acc_norm_stderr": 0.006867425939236565 |
|
}, |
|
"community|arabic_mmlu_ht:high_school_computer_science|0": { |
|
"acc_norm": 0.78, |
|
"acc_norm_stderr": 0.04163331998932263 |
|
}, |
|
"community|arabic_mmlu_ht:high_school_european_history|0": { |
|
"acc_norm": 0.6436457311089303, |
|
"acc_norm_stderr": 0.005304677462902026 |
|
}, |
|
"community|arabic_mmlu_ht:high_school_geography|0": { |
|
"acc_norm": 0.8232323232323232, |
|
"acc_norm_stderr": 0.027178752639044915 |
|
}, |
|
"community|arabic_mmlu_ht:high_school_government_and_politics|0": { |
|
"acc_norm": 0.8704663212435233, |
|
"acc_norm_stderr": 0.024233532297758733 |
|
}, |
|
"community|arabic_mmlu_ht:high_school_macroeconomics|0": { |
|
"acc_norm": 0.7640954686959529, |
|
"acc_norm_stderr": 0.007897561785150112 |
|
}, |
|
"community|arabic_mmlu_ht:high_school_mathematics|0": { |
|
"acc_norm": 0.3888888888888889, |
|
"acc_norm_stderr": 0.029723278961476668 |
|
}, |
|
"community|arabic_mmlu_ht:high_school_microeconomics|0": { |
|
"acc_norm": 0.7100840336134454, |
|
"acc_norm_stderr": 0.02947248583313608 |
|
}, |
|
"community|arabic_mmlu_ht:high_school_physics|0": { |
|
"acc_norm": 0.543046357615894, |
|
"acc_norm_stderr": 0.04067325174247444 |
|
}, |
|
"community|arabic_mmlu_ht:high_school_psychology|0": { |
|
"acc_norm": 0.7908256880733945, |
|
"acc_norm_stderr": 0.017437937173343226 |
|
}, |
|
"community|arabic_mmlu_ht:high_school_statistics|0": { |
|
"acc_norm": 0.7400756143667296, |
|
"acc_norm_stderr": 0.006742796420305758 |
|
}, |
|
"community|arabic_mmlu_ht:high_school_us_history|0": { |
|
"acc_norm": 0.8480392156862745, |
|
"acc_norm_stderr": 0.025195658428931792 |
|
}, |
|
"community|arabic_mmlu_ht:high_school_world_history|0": { |
|
"acc_norm": 0.8438818565400844, |
|
"acc_norm_stderr": 0.023627159460318667 |
|
}, |
|
"community|arabic_mmlu_ht:human_aging|0": { |
|
"acc_norm": 0.6905829596412556, |
|
"acc_norm_stderr": 0.03102441174057221 |
|
}, |
|
"community|arabic_mmlu_ht:human_sexuality|0": { |
|
"acc_norm": 0.7404580152671756, |
|
"acc_norm_stderr": 0.03844876139785271 |
|
}, |
|
"community|arabic_mmlu_ht:international_law|0": { |
|
"acc_norm": 0.8429752066115702, |
|
"acc_norm_stderr": 0.03321244842547128 |
|
}, |
|
"community|arabic_mmlu_ht:jurisprudence|0": { |
|
"acc_norm": 0.7685185185185185, |
|
"acc_norm_stderr": 0.04077494709252626 |
|
}, |
|
"community|arabic_mmlu_ht:logical_fallacies|0": { |
|
"acc_norm": 0.6871165644171779, |
|
"acc_norm_stderr": 0.036429145782924055 |
|
}, |
|
"community|arabic_mmlu_ht:machine_learning|0": { |
|
"acc_norm": 0.5089285714285714, |
|
"acc_norm_stderr": 0.04745033255489123 |
|
}, |
|
"community|arabic_mmlu_ht:management|0": { |
|
"acc_norm": 0.7281553398058253, |
|
"acc_norm_stderr": 0.044052680241409216 |
|
}, |
|
"community|arabic_mmlu_ht:marketing|0": { |
|
"acc_norm": 0.8333333333333334, |
|
"acc_norm_stderr": 0.024414947304543678 |
|
}, |
|
"community|arabic_mmlu_ht:medical_genetics|0": { |
|
"acc_norm": 0.68, |
|
"acc_norm_stderr": 0.046882617226215034 |
|
}, |
|
"community|arabic_mmlu_ht:miscellaneous|0": { |
|
"acc_norm": 0.7805785123966942, |
|
"acc_norm_stderr": 0.008414531476407524 |
|
}, |
|
"community|arabic_mmlu_ht:moral_disputes|0": { |
|
"acc_norm": 0.7485549132947977, |
|
"acc_norm_stderr": 0.023357365785874037 |
|
}, |
|
"community|arabic_mmlu_ht:moral_scenarios|0": { |
|
"acc_norm": 0.6402234636871509, |
|
"acc_norm_stderr": 0.016051419760310267 |
|
}, |
|
"community|arabic_mmlu_ht:nutrition|0": { |
|
"acc_norm": 0.7450980392156863, |
|
"acc_norm_stderr": 0.02495418432487991 |
|
}, |
|
"community|arabic_mmlu_ht:philosophy|0": { |
|
"acc_norm": 0.6913183279742765, |
|
"acc_norm_stderr": 0.02623696588115326 |
|
}, |
|
"community|arabic_mmlu_ht:prehistory|0": { |
|
"acc_norm": 0.7314814814814815, |
|
"acc_norm_stderr": 0.024659685185967294 |
|
}, |
|
"community|arabic_mmlu_ht:professional_accounting|0": { |
|
"acc_norm": 0.7250775365529464, |
|
"acc_norm_stderr": 0.0066460663381219375 |
|
}, |
|
"community|arabic_mmlu_ht:professional_law|0": { |
|
"acc_norm": 0.6406660823838738, |
|
"acc_norm_stderr": 0.005369085494446666 |
|
}, |
|
"community|arabic_mmlu_ht:professional_medicine|0": { |
|
"acc_norm": 0.7782529016493586, |
|
"acc_norm_stderr": 0.010270644506756786 |
|
}, |
|
"community|arabic_mmlu_ht:professional_psychology|0": { |
|
"acc_norm": 0.7482158150157008, |
|
"acc_norm_stderr": 0.0073344844954788425 |
|
}, |
|
"community|arabic_mmlu_ht:public_relations|0": { |
|
"acc_norm": 0.5909090909090909, |
|
"acc_norm_stderr": 0.04709306978661895 |
|
}, |
|
"community|arabic_mmlu_ht:security_studies|0": { |
|
"acc_norm": 0.689795918367347, |
|
"acc_norm_stderr": 0.02961345987248438 |
|
}, |
|
"community|arabic_mmlu_ht:sociology|0": { |
|
"acc_norm": 0.8258706467661692, |
|
"acc_norm_stderr": 0.026814951200421603 |
|
}, |
|
"community|arabic_mmlu_ht:us_foreign_policy|0": { |
|
"acc_norm": 0.85, |
|
"acc_norm_stderr": 0.0358870281282637 |
|
}, |
|
"community|arabic_mmlu_ht:virology|0": { |
|
"acc_norm": 0.4759036144578313, |
|
"acc_norm_stderr": 0.038879718495972646 |
|
}, |
|
"community|arabic_mmlu_ht:world_religions|0": { |
|
"acc_norm": 0.7602339181286549, |
|
"acc_norm_stderr": 0.03274485211946956 |
|
}, |
|
"community|aratrust:Ethics|0": { |
|
"acc_norm": 0.7833333333333333, |
|
"acc_norm_stderr": 0.05363439040648223 |
|
}, |
|
"community|aratrust:Illegal|0": { |
|
"acc_norm": 0.8867924528301887, |
|
"acc_norm_stderr": 0.043938680706010445 |
|
}, |
|
"community|aratrust:MentalHealth|0": { |
|
"acc_norm": 0.9210526315789473, |
|
"acc_norm_stderr": 0.03113726201631375 |
|
}, |
|
"community|aratrust:Offensive|0": { |
|
"acc_norm": 0.8405797101449275, |
|
"acc_norm_stderr": 0.044392212657971275 |
|
}, |
|
"community|aratrust:PhysicalHealth|0": { |
|
"acc_norm": 0.9178082191780822, |
|
"acc_norm_stderr": 0.032368600179812405 |
|
}, |
|
"community|aratrust:Privacy|0": { |
|
"acc_norm": 0.9649122807017544, |
|
"acc_norm_stderr": 0.024588241692166144 |
|
}, |
|
"community|aratrust:Trustfulness|0": { |
|
"acc_norm": 0.8205128205128205, |
|
"acc_norm_stderr": 0.04373348027807389 |
|
}, |
|
"community|aratrust:Unfairness|0": { |
|
"acc_norm": 0.9090909090909091, |
|
"acc_norm_stderr": 0.03912104390108502 |
|
}, |
|
"community|madinah_qa:Arabic Language (General)|0": { |
|
"acc_norm": 0.5490196078431373, |
|
"acc_norm_stderr": 0.020130388312904528 |
|
}, |
|
"community|madinah_qa:Arabic Language (Grammar)|0": { |
|
"acc_norm": 0.5150684931506849, |
|
"acc_norm_stderr": 0.02619521706374927 |
|
}, |
|
"community|alghafa:_average|0": { |
|
"acc_norm": 0.6379886939750842, |
|
"acc_norm_stderr": 0.02116115089760025 |
|
}, |
|
"community|arabic_mmlu:_average|0": { |
|
"acc_norm": 0.6976102555189343, |
|
"acc_norm_stderr": 0.03291881778773066 |
|
}, |
|
"community|arabic_mmlu_ht:_average|0": { |
|
"acc_norm": 0.6768217155471902, |
|
"acc_norm_stderr": 0.030729092962638822 |
|
}, |
|
"community|aratrust:_average|0": { |
|
"acc_norm": 0.8805102946713703, |
|
"acc_norm_stderr": 0.03911423897973939 |
|
}, |
|
"community|madinah_qa:_average|0": { |
|
"acc_norm": 0.5320440504969111, |
|
"acc_norm_stderr": 0.0231628026883269 |
|
}, |
|
"all": { |
|
"acc_norm": 0.6923075502372169, |
|
"acc_norm_stderr": 0.0310971150404444 |
|
}, |
|
"community|alrage_qa|0": { |
|
"llm_as_judge": 0.7582621082621049, |
|
"llm_as_judge_stderr": 0.00014081705539228294 |
|
} |
|
}, |
|
"versions": { |
|
"community|alghafa:mcq_exams_test_ar|0": 0, |
|
"community|alghafa:meta_ar_dialects|0": 0, |
|
"community|alghafa:meta_ar_msa|0": 0, |
|
"community|alghafa:multiple_choice_facts_truefalse_balanced_task|0": 0, |
|
"community|alghafa:multiple_choice_grounded_statement_soqal_task|0": 0, |
|
"community|alghafa:multiple_choice_grounded_statement_xglue_mlqa_task|0": 0, |
|
"community|alghafa:multiple_choice_rating_sentiment_no_neutral_task|0": 0, |
|
"community|alghafa:multiple_choice_rating_sentiment_task|0": 0, |
|
"community|alghafa:multiple_choice_sentiment_task|0": 0, |
|
"community|arabic_exams|0": 0, |
|
"community|arabic_mmlu:Accounting (University)|0": 0, |
|
"community|arabic_mmlu:Arabic Language (General)|0": 0, |
|
"community|arabic_mmlu:Arabic Language (Grammar)|0": 0, |
|
"community|arabic_mmlu:Arabic Language (High School)|0": 0, |
|
"community|arabic_mmlu:Arabic Language (Middle School)|0": 0, |
|
"community|arabic_mmlu:Arabic Language (Primary School)|0": 0, |
|
"community|arabic_mmlu:Biology (High School)|0": 0, |
|
"community|arabic_mmlu:Civics (High School)|0": 0, |
|
"community|arabic_mmlu:Civics (Middle School)|0": 0, |
|
"community|arabic_mmlu:Computer Science (High School)|0": 0, |
|
"community|arabic_mmlu:Computer Science (Middle School)|0": 0, |
|
"community|arabic_mmlu:Computer Science (Primary School)|0": 0, |
|
"community|arabic_mmlu:Computer Science (University)|0": 0, |
|
"community|arabic_mmlu:Driving Test|0": 0, |
|
"community|arabic_mmlu:Economics (High School)|0": 0, |
|
"community|arabic_mmlu:Economics (Middle School)|0": 0, |
|
"community|arabic_mmlu:Economics (University)|0": 0, |
|
"community|arabic_mmlu:General Knowledge|0": 0, |
|
"community|arabic_mmlu:General Knowledge (Middle School)|0": 0, |
|
"community|arabic_mmlu:General Knowledge (Primary School)|0": 0, |
|
"community|arabic_mmlu:Geography (High School)|0": 0, |
|
"community|arabic_mmlu:Geography (Middle School)|0": 0, |
|
"community|arabic_mmlu:Geography (Primary School)|0": 0, |
|
"community|arabic_mmlu:History (High School)|0": 0, |
|
"community|arabic_mmlu:History (Middle School)|0": 0, |
|
"community|arabic_mmlu:History (Primary School)|0": 0, |
|
"community|arabic_mmlu:Islamic Studies|0": 0, |
|
"community|arabic_mmlu:Islamic Studies (High School)|0": 0, |
|
"community|arabic_mmlu:Islamic Studies (Middle School)|0": 0, |
|
"community|arabic_mmlu:Islamic Studies (Primary School)|0": 0, |
|
"community|arabic_mmlu:Law (Professional)|0": 0, |
|
"community|arabic_mmlu:Management (University)|0": 0, |
|
"community|arabic_mmlu:Math (Primary School)|0": 0, |
|
"community|arabic_mmlu:Natural Science (Middle School)|0": 0, |
|
"community|arabic_mmlu:Natural Science (Primary School)|0": 0, |
|
"community|arabic_mmlu:Philosophy (High School)|0": 0, |
|
"community|arabic_mmlu:Physics (High School)|0": 0, |
|
"community|arabic_mmlu:Political Science (University)|0": 0, |
|
"community|arabic_mmlu:Social Science (Middle School)|0": 0, |
|
"community|arabic_mmlu:Social Science (Primary School)|0": 0, |
|
"community|arabic_mmlu_ht:abstract_algebra|0": 0, |
|
"community|arabic_mmlu_ht:anatomy|0": 0, |
|
"community|arabic_mmlu_ht:astronomy|0": 0, |
|
"community|arabic_mmlu_ht:business_ethics|0": 0, |
|
"community|arabic_mmlu_ht:clinical_knowledge|0": 0, |
|
"community|arabic_mmlu_ht:college_biology|0": 0, |
|
"community|arabic_mmlu_ht:college_chemistry|0": 0, |
|
"community|arabic_mmlu_ht:college_computer_science|0": 0, |
|
"community|arabic_mmlu_ht:college_mathematics|0": 0, |
|
"community|arabic_mmlu_ht:college_medicine|0": 0, |
|
"community|arabic_mmlu_ht:college_physics|0": 0, |
|
"community|arabic_mmlu_ht:computer_security|0": 0, |
|
"community|arabic_mmlu_ht:conceptual_physics|0": 0, |
|
"community|arabic_mmlu_ht:econometrics|0": 0, |
|
"community|arabic_mmlu_ht:electrical_engineering|0": 0, |
|
"community|arabic_mmlu_ht:elementary_mathematics|0": 0, |
|
"community|arabic_mmlu_ht:formal_logic|0": 0, |
|
"community|arabic_mmlu_ht:global_facts|0": 0, |
|
"community|arabic_mmlu_ht:high_school_biology|0": 0, |
|
"community|arabic_mmlu_ht:high_school_chemistry|0": 0, |
|
"community|arabic_mmlu_ht:high_school_computer_science|0": 0, |
|
"community|arabic_mmlu_ht:high_school_european_history|0": 0, |
|
"community|arabic_mmlu_ht:high_school_geography|0": 0, |
|
"community|arabic_mmlu_ht:high_school_government_and_politics|0": 0, |
|
"community|arabic_mmlu_ht:high_school_macroeconomics|0": 0, |
|
"community|arabic_mmlu_ht:high_school_mathematics|0": 0, |
|
"community|arabic_mmlu_ht:high_school_microeconomics|0": 0, |
|
"community|arabic_mmlu_ht:high_school_physics|0": 0, |
|
"community|arabic_mmlu_ht:high_school_psychology|0": 0, |
|
"community|arabic_mmlu_ht:high_school_statistics|0": 0, |
|
"community|arabic_mmlu_ht:high_school_us_history|0": 0, |
|
"community|arabic_mmlu_ht:high_school_world_history|0": 0, |
|
"community|arabic_mmlu_ht:human_aging|0": 0, |
|
"community|arabic_mmlu_ht:human_sexuality|0": 0, |
|
"community|arabic_mmlu_ht:international_law|0": 0, |
|
"community|arabic_mmlu_ht:jurisprudence|0": 0, |
|
"community|arabic_mmlu_ht:logical_fallacies|0": 0, |
|
"community|arabic_mmlu_ht:machine_learning|0": 0, |
|
"community|arabic_mmlu_ht:management|0": 0, |
|
"community|arabic_mmlu_ht:marketing|0": 0, |
|
"community|arabic_mmlu_ht:medical_genetics|0": 0, |
|
"community|arabic_mmlu_ht:miscellaneous|0": 0, |
|
"community|arabic_mmlu_ht:moral_disputes|0": 0, |
|
"community|arabic_mmlu_ht:moral_scenarios|0": 0, |
|
"community|arabic_mmlu_ht:nutrition|0": 0, |
|
"community|arabic_mmlu_ht:philosophy|0": 0, |
|
"community|arabic_mmlu_ht:prehistory|0": 0, |
|
"community|arabic_mmlu_ht:professional_accounting|0": 0, |
|
"community|arabic_mmlu_ht:professional_law|0": 0, |
|
"community|arabic_mmlu_ht:professional_medicine|0": 0, |
|
"community|arabic_mmlu_ht:professional_psychology|0": 0, |
|
"community|arabic_mmlu_ht:public_relations|0": 0, |
|
"community|arabic_mmlu_ht:security_studies|0": 0, |
|
"community|arabic_mmlu_ht:sociology|0": 0, |
|
"community|arabic_mmlu_ht:us_foreign_policy|0": 0, |
|
"community|arabic_mmlu_ht:virology|0": 0, |
|
"community|arabic_mmlu_ht:world_religions|0": 0, |
|
"community|aratrust:Ethics|0": 0, |
|
"community|aratrust:Illegal|0": 0, |
|
"community|aratrust:MentalHealth|0": 0, |
|
"community|aratrust:Offensive|0": 0, |
|
"community|aratrust:PhysicalHealth|0": 0, |
|
"community|aratrust:Privacy|0": 0, |
|
"community|aratrust:Trustfulness|0": 0, |
|
"community|aratrust:Unfairness|0": 0, |
|
"community|madinah_qa:Arabic Language (General)|0": 0, |
|
"community|madinah_qa:Arabic Language (Grammar)|0": 0, |
|
"community|alrage_qa|0": 0 |
|
}, |
|
"config_tasks": { |
|
"community|alghafa:mcq_exams_test_ar": { |
|
"name": "alghafa:mcq_exams_test_ar", |
|
"prompt_function": "alghafa_pfn", |
|
"hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Native", |
|
"hf_subset": "mcq_exams_test_ar", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test", |
|
"validation" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": "validation", |
|
"few_shots_select": "sequential", |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 557, |
|
"effective_num_docs": 557, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|alghafa:meta_ar_dialects": { |
|
"name": "alghafa:meta_ar_dialects", |
|
"prompt_function": "alghafa_pfn", |
|
"hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Native", |
|
"hf_subset": "meta_ar_dialects", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test", |
|
"validation" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": "validation", |
|
"few_shots_select": "sequential", |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 5395, |
|
"effective_num_docs": 5395, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|alghafa:meta_ar_msa": { |
|
"name": "alghafa:meta_ar_msa", |
|
"prompt_function": "alghafa_pfn", |
|
"hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Native", |
|
"hf_subset": "meta_ar_msa", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test", |
|
"validation" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": "validation", |
|
"few_shots_select": "sequential", |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 895, |
|
"effective_num_docs": 895, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|alghafa:multiple_choice_facts_truefalse_balanced_task": { |
|
"name": "alghafa:multiple_choice_facts_truefalse_balanced_task", |
|
"prompt_function": "alghafa_pfn", |
|
"hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Native", |
|
"hf_subset": "multiple_choice_facts_truefalse_balanced_task", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test", |
|
"validation" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": "validation", |
|
"few_shots_select": "sequential", |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 75, |
|
"effective_num_docs": 75, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|alghafa:multiple_choice_grounded_statement_soqal_task": { |
|
"name": "alghafa:multiple_choice_grounded_statement_soqal_task", |
|
"prompt_function": "alghafa_pfn", |
|
"hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Native", |
|
"hf_subset": "multiple_choice_grounded_statement_soqal_task", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test", |
|
"validation" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": "validation", |
|
"few_shots_select": "sequential", |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 150, |
|
"effective_num_docs": 150, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|alghafa:multiple_choice_grounded_statement_xglue_mlqa_task": { |
|
"name": "alghafa:multiple_choice_grounded_statement_xglue_mlqa_task", |
|
"prompt_function": "alghafa_pfn", |
|
"hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Native", |
|
"hf_subset": "multiple_choice_grounded_statement_xglue_mlqa_task", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test", |
|
"validation" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": "validation", |
|
"few_shots_select": "sequential", |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 150, |
|
"effective_num_docs": 150, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|alghafa:multiple_choice_rating_sentiment_no_neutral_task": { |
|
"name": "alghafa:multiple_choice_rating_sentiment_no_neutral_task", |
|
"prompt_function": "alghafa_pfn", |
|
"hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Native", |
|
"hf_subset": "multiple_choice_rating_sentiment_no_neutral_task", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test", |
|
"validation" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": "validation", |
|
"few_shots_select": "sequential", |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 7995, |
|
"effective_num_docs": 7995, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|alghafa:multiple_choice_rating_sentiment_task": { |
|
"name": "alghafa:multiple_choice_rating_sentiment_task", |
|
"prompt_function": "alghafa_pfn", |
|
"hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Native", |
|
"hf_subset": "multiple_choice_rating_sentiment_task", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test", |
|
"validation" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": "validation", |
|
"few_shots_select": "sequential", |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 5995, |
|
"effective_num_docs": 5995, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|alghafa:multiple_choice_sentiment_task": { |
|
"name": "alghafa:multiple_choice_sentiment_task", |
|
"prompt_function": "alghafa_pfn", |
|
"hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Native", |
|
"hf_subset": "multiple_choice_sentiment_task", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test", |
|
"validation" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": "validation", |
|
"few_shots_select": "sequential", |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 1720, |
|
"effective_num_docs": 1720, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_exams": { |
|
"name": "arabic_exams", |
|
"prompt_function": "arabic_exams_pfn", |
|
"hf_repo": "OALL/Arabic_EXAMS", |
|
"hf_subset": "default", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test", |
|
"validation" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": "validation", |
|
"few_shots_select": "sequential", |
|
"generation_size": null, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 537, |
|
"effective_num_docs": 537, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu:Accounting (University)": { |
|
"name": "arabic_mmlu:Accounting (University)", |
|
"prompt_function": "arabic_mmlu_pfn", |
|
"hf_repo": "MBZUAI/ArabicMMLU", |
|
"hf_subset": "Accounting (University)", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": [ |
|
"dev" |
|
], |
|
"few_shots_select": "sequential", |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 74, |
|
"effective_num_docs": 74, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu:Arabic Language (General)": { |
|
"name": "arabic_mmlu:Arabic Language (General)", |
|
"prompt_function": "arabic_mmlu_pfn", |
|
"hf_repo": "MBZUAI/ArabicMMLU", |
|
"hf_subset": "Arabic Language (General)", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": [ |
|
"dev" |
|
], |
|
"few_shots_select": "sequential", |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 612, |
|
"effective_num_docs": 612, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu:Arabic Language (Grammar)": { |
|
"name": "arabic_mmlu:Arabic Language (Grammar)", |
|
"prompt_function": "arabic_mmlu_pfn", |
|
"hf_repo": "MBZUAI/ArabicMMLU", |
|
"hf_subset": "Arabic Language (Grammar)", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": [ |
|
"dev" |
|
], |
|
"few_shots_select": "sequential", |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 365, |
|
"effective_num_docs": 365, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu:Arabic Language (High School)": { |
|
"name": "arabic_mmlu:Arabic Language (High School)", |
|
"prompt_function": "arabic_mmlu_pfn", |
|
"hf_repo": "MBZUAI/ArabicMMLU", |
|
"hf_subset": "Arabic Language (High School)", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": [ |
|
"dev" |
|
], |
|
"few_shots_select": "sequential", |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 390, |
|
"effective_num_docs": 390, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu:Arabic Language (Middle School)": { |
|
"name": "arabic_mmlu:Arabic Language (Middle School)", |
|
"prompt_function": "arabic_mmlu_pfn", |
|
"hf_repo": "MBZUAI/ArabicMMLU", |
|
"hf_subset": "Arabic Language (Middle School)", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": [ |
|
"dev" |
|
], |
|
"few_shots_select": "sequential", |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 27, |
|
"effective_num_docs": 27, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu:Arabic Language (Primary School)": { |
|
"name": "arabic_mmlu:Arabic Language (Primary School)", |
|
"prompt_function": "arabic_mmlu_pfn", |
|
"hf_repo": "MBZUAI/ArabicMMLU", |
|
"hf_subset": "Arabic Language (Primary School)", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": [ |
|
"dev" |
|
], |
|
"few_shots_select": "sequential", |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 252, |
|
"effective_num_docs": 252, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu:Biology (High School)": { |
|
"name": "arabic_mmlu:Biology (High School)", |
|
"prompt_function": "arabic_mmlu_pfn", |
|
"hf_repo": "MBZUAI/ArabicMMLU", |
|
"hf_subset": "Biology (High School)", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": [ |
|
"dev" |
|
], |
|
"few_shots_select": "sequential", |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 1409, |
|
"effective_num_docs": 1409, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu:Civics (High School)": { |
|
"name": "arabic_mmlu:Civics (High School)", |
|
"prompt_function": "arabic_mmlu_pfn", |
|
"hf_repo": "MBZUAI/ArabicMMLU", |
|
"hf_subset": "Civics (High School)", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": [ |
|
"dev" |
|
], |
|
"few_shots_select": "sequential", |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 87, |
|
"effective_num_docs": 87, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu:Civics (Middle School)": { |
|
"name": "arabic_mmlu:Civics (Middle School)", |
|
"prompt_function": "arabic_mmlu_pfn", |
|
"hf_repo": "MBZUAI/ArabicMMLU", |
|
"hf_subset": "Civics (Middle School)", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": [ |
|
"dev" |
|
], |
|
"few_shots_select": "sequential", |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 236, |
|
"effective_num_docs": 236, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu:Computer Science (High School)": { |
|
"name": "arabic_mmlu:Computer Science (High School)", |
|
"prompt_function": "arabic_mmlu_pfn", |
|
"hf_repo": "MBZUAI/ArabicMMLU", |
|
"hf_subset": "Computer Science (High School)", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": [ |
|
"dev" |
|
], |
|
"few_shots_select": "sequential", |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 261, |
|
"effective_num_docs": 261, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu:Computer Science (Middle School)": { |
|
"name": "arabic_mmlu:Computer Science (Middle School)", |
|
"prompt_function": "arabic_mmlu_pfn", |
|
"hf_repo": "MBZUAI/ArabicMMLU", |
|
"hf_subset": "Computer Science (Middle School)", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": [ |
|
"dev" |
|
], |
|
"few_shots_select": "sequential", |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 27, |
|
"effective_num_docs": 27, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu:Computer Science (Primary School)": { |
|
"name": "arabic_mmlu:Computer Science (Primary School)", |
|
"prompt_function": "arabic_mmlu_pfn", |
|
"hf_repo": "MBZUAI/ArabicMMLU", |
|
"hf_subset": "Computer Science (Primary School)", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": [ |
|
"dev" |
|
], |
|
"few_shots_select": "sequential", |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 190, |
|
"effective_num_docs": 190, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu:Computer Science (University)": { |
|
"name": "arabic_mmlu:Computer Science (University)", |
|
"prompt_function": "arabic_mmlu_pfn", |
|
"hf_repo": "MBZUAI/ArabicMMLU", |
|
"hf_subset": "Computer Science (University)", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": [ |
|
"dev" |
|
], |
|
"few_shots_select": "sequential", |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 64, |
|
"effective_num_docs": 64, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu:Driving Test": { |
|
"name": "arabic_mmlu:Driving Test", |
|
"prompt_function": "arabic_mmlu_pfn", |
|
"hf_repo": "MBZUAI/ArabicMMLU", |
|
"hf_subset": "Driving Test", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": [ |
|
"dev" |
|
], |
|
"few_shots_select": "sequential", |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 1211, |
|
"effective_num_docs": 1211, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu:Economics (High School)": { |
|
"name": "arabic_mmlu:Economics (High School)", |
|
"prompt_function": "arabic_mmlu_pfn", |
|
"hf_repo": "MBZUAI/ArabicMMLU", |
|
"hf_subset": "Economics (High School)", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": [ |
|
"dev" |
|
], |
|
"few_shots_select": "sequential", |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 360, |
|
"effective_num_docs": 360, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu:Economics (Middle School)": { |
|
"name": "arabic_mmlu:Economics (Middle School)", |
|
"prompt_function": "arabic_mmlu_pfn", |
|
"hf_repo": "MBZUAI/ArabicMMLU", |
|
"hf_subset": "Economics (Middle School)", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": [ |
|
"dev" |
|
], |
|
"few_shots_select": "sequential", |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 87, |
|
"effective_num_docs": 87, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu:Economics (University)": { |
|
"name": "arabic_mmlu:Economics (University)", |
|
"prompt_function": "arabic_mmlu_pfn", |
|
"hf_repo": "MBZUAI/ArabicMMLU", |
|
"hf_subset": "Economics (University)", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": [ |
|
"dev" |
|
], |
|
"few_shots_select": "sequential", |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 137, |
|
"effective_num_docs": 137, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu:General Knowledge": { |
|
"name": "arabic_mmlu:General Knowledge", |
|
"prompt_function": "arabic_mmlu_pfn", |
|
"hf_repo": "MBZUAI/ArabicMMLU", |
|
"hf_subset": "General Knowledge", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": [ |
|
"dev" |
|
], |
|
"few_shots_select": "sequential", |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 864, |
|
"effective_num_docs": 864, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu:General Knowledge (Middle School)": { |
|
"name": "arabic_mmlu:General Knowledge (Middle School)", |
|
"prompt_function": "arabic_mmlu_pfn", |
|
"hf_repo": "MBZUAI/ArabicMMLU", |
|
"hf_subset": "General Knowledge (Middle School)", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": [ |
|
"dev" |
|
], |
|
"few_shots_select": "sequential", |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 172, |
|
"effective_num_docs": 172, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu:General Knowledge (Primary School)": { |
|
"name": "arabic_mmlu:General Knowledge (Primary School)", |
|
"prompt_function": "arabic_mmlu_pfn", |
|
"hf_repo": "MBZUAI/ArabicMMLU", |
|
"hf_subset": "General Knowledge (Primary School)", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": [ |
|
"dev" |
|
], |
|
"few_shots_select": "sequential", |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 162, |
|
"effective_num_docs": 162, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu:Geography (High School)": { |
|
"name": "arabic_mmlu:Geography (High School)", |
|
"prompt_function": "arabic_mmlu_pfn", |
|
"hf_repo": "MBZUAI/ArabicMMLU", |
|
"hf_subset": "Geography (High School)", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": [ |
|
"dev" |
|
], |
|
"few_shots_select": "sequential", |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 1038, |
|
"effective_num_docs": 1038, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu:Geography (Middle School)": { |
|
"name": "arabic_mmlu:Geography (Middle School)", |
|
"prompt_function": "arabic_mmlu_pfn", |
|
"hf_repo": "MBZUAI/ArabicMMLU", |
|
"hf_subset": "Geography (Middle School)", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": [ |
|
"dev" |
|
], |
|
"few_shots_select": "sequential", |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 272, |
|
"effective_num_docs": 272, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu:Geography (Primary School)": { |
|
"name": "arabic_mmlu:Geography (Primary School)", |
|
"prompt_function": "arabic_mmlu_pfn", |
|
"hf_repo": "MBZUAI/ArabicMMLU", |
|
"hf_subset": "Geography (Primary School)", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": [ |
|
"dev" |
|
], |
|
"few_shots_select": "sequential", |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 57, |
|
"effective_num_docs": 57, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu:History (High School)": { |
|
"name": "arabic_mmlu:History (High School)", |
|
"prompt_function": "arabic_mmlu_pfn", |
|
"hf_repo": "MBZUAI/ArabicMMLU", |
|
"hf_subset": "History (High School)", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": [ |
|
"dev" |
|
], |
|
"few_shots_select": "sequential", |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 760, |
|
"effective_num_docs": 760, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu:History (Middle School)": { |
|
"name": "arabic_mmlu:History (Middle School)", |
|
"prompt_function": "arabic_mmlu_pfn", |
|
"hf_repo": "MBZUAI/ArabicMMLU", |
|
"hf_subset": "History (Middle School)", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": [ |
|
"dev" |
|
], |
|
"few_shots_select": "sequential", |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 203, |
|
"effective_num_docs": 203, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu:History (Primary School)": { |
|
"name": "arabic_mmlu:History (Primary School)", |
|
"prompt_function": "arabic_mmlu_pfn", |
|
"hf_repo": "MBZUAI/ArabicMMLU", |
|
"hf_subset": "History (Primary School)", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": [ |
|
"dev" |
|
], |
|
"few_shots_select": "sequential", |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 102, |
|
"effective_num_docs": 102, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu:Islamic Studies": { |
|
"name": "arabic_mmlu:Islamic Studies", |
|
"prompt_function": "arabic_mmlu_pfn", |
|
"hf_repo": "MBZUAI/ArabicMMLU", |
|
"hf_subset": "Islamic Studies", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": [ |
|
"dev" |
|
], |
|
"few_shots_select": "sequential", |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 639, |
|
"effective_num_docs": 639, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu:Islamic Studies (High School)": { |
|
"name": "arabic_mmlu:Islamic Studies (High School)", |
|
"prompt_function": "arabic_mmlu_pfn", |
|
"hf_repo": "MBZUAI/ArabicMMLU", |
|
"hf_subset": "Islamic Studies (High School)", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": [ |
|
"dev" |
|
], |
|
"few_shots_select": "sequential", |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 334, |
|
"effective_num_docs": 334, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu:Islamic Studies (Middle School)": { |
|
"name": "arabic_mmlu:Islamic Studies (Middle School)", |
|
"prompt_function": "arabic_mmlu_pfn", |
|
"hf_repo": "MBZUAI/ArabicMMLU", |
|
"hf_subset": "Islamic Studies (Middle School)", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": [ |
|
"dev" |
|
], |
|
"few_shots_select": "sequential", |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 238, |
|
"effective_num_docs": 238, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu:Islamic Studies (Primary School)": { |
|
"name": "arabic_mmlu:Islamic Studies (Primary School)", |
|
"prompt_function": "arabic_mmlu_pfn", |
|
"hf_repo": "MBZUAI/ArabicMMLU", |
|
"hf_subset": "Islamic Studies (Primary School)", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": [ |
|
"dev" |
|
], |
|
"few_shots_select": "sequential", |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 999, |
|
"effective_num_docs": 999, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu:Law (Professional)": { |
|
"name": "arabic_mmlu:Law (Professional)", |
|
"prompt_function": "arabic_mmlu_pfn", |
|
"hf_repo": "MBZUAI/ArabicMMLU", |
|
"hf_subset": "Law (Professional)", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": [ |
|
"dev" |
|
], |
|
"few_shots_select": "sequential", |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 314, |
|
"effective_num_docs": 314, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu:Management (University)": { |
|
"name": "arabic_mmlu:Management (University)", |
|
"prompt_function": "arabic_mmlu_pfn", |
|
"hf_repo": "MBZUAI/ArabicMMLU", |
|
"hf_subset": "Management (University)", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": [ |
|
"dev" |
|
], |
|
"few_shots_select": "sequential", |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 75, |
|
"effective_num_docs": 75, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu:Math (Primary School)": { |
|
"name": "arabic_mmlu:Math (Primary School)", |
|
"prompt_function": "arabic_mmlu_pfn", |
|
"hf_repo": "MBZUAI/ArabicMMLU", |
|
"hf_subset": "Math (Primary School)", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": [ |
|
"dev" |
|
], |
|
"few_shots_select": "sequential", |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 409, |
|
"effective_num_docs": 409, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu:Natural Science (Middle School)": { |
|
"name": "arabic_mmlu:Natural Science (Middle School)", |
|
"prompt_function": "arabic_mmlu_pfn", |
|
"hf_repo": "MBZUAI/ArabicMMLU", |
|
"hf_subset": "Natural Science (Middle School)", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": [ |
|
"dev" |
|
], |
|
"few_shots_select": "sequential", |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 242, |
|
"effective_num_docs": 242, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu:Natural Science (Primary School)": { |
|
"name": "arabic_mmlu:Natural Science (Primary School)", |
|
"prompt_function": "arabic_mmlu_pfn", |
|
"hf_repo": "MBZUAI/ArabicMMLU", |
|
"hf_subset": "Natural Science (Primary School)", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": [ |
|
"dev" |
|
], |
|
"few_shots_select": "sequential", |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 336, |
|
"effective_num_docs": 336, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu:Philosophy (High School)": { |
|
"name": "arabic_mmlu:Philosophy (High School)", |
|
"prompt_function": "arabic_mmlu_pfn", |
|
"hf_repo": "MBZUAI/ArabicMMLU", |
|
"hf_subset": "Philosophy (High School)", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": [ |
|
"dev" |
|
], |
|
"few_shots_select": "sequential", |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 39, |
|
"effective_num_docs": 39, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu:Physics (High School)": { |
|
"name": "arabic_mmlu:Physics (High School)", |
|
"prompt_function": "arabic_mmlu_pfn", |
|
"hf_repo": "MBZUAI/ArabicMMLU", |
|
"hf_subset": "Physics (High School)", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": [ |
|
"dev" |
|
], |
|
"few_shots_select": "sequential", |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 255, |
|
"effective_num_docs": 255, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu:Political Science (University)": { |
|
"name": "arabic_mmlu:Political Science (University)", |
|
"prompt_function": "arabic_mmlu_pfn", |
|
"hf_repo": "MBZUAI/ArabicMMLU", |
|
"hf_subset": "Political Science (University)", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": [ |
|
"dev" |
|
], |
|
"few_shots_select": "sequential", |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 210, |
|
"effective_num_docs": 210, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu:Social Science (Middle School)": { |
|
"name": "arabic_mmlu:Social Science (Middle School)", |
|
"prompt_function": "arabic_mmlu_pfn", |
|
"hf_repo": "MBZUAI/ArabicMMLU", |
|
"hf_subset": "Social Science (Middle School)", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": [ |
|
"dev" |
|
], |
|
"few_shots_select": "sequential", |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 241, |
|
"effective_num_docs": 241, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu:Social Science (Primary School)": { |
|
"name": "arabic_mmlu:Social Science (Primary School)", |
|
"prompt_function": "arabic_mmlu_pfn", |
|
"hf_repo": "MBZUAI/ArabicMMLU", |
|
"hf_subset": "Social Science (Primary School)", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": [ |
|
"dev" |
|
], |
|
"few_shots_select": "sequential", |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 705, |
|
"effective_num_docs": 705, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:abstract_algebra": { |
|
"name": "arabic_mmlu_ht:abstract_algebra", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "abstract_algebra", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 100, |
|
"effective_num_docs": 100, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:anatomy": { |
|
"name": "arabic_mmlu_ht:anatomy", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "anatomy", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 135, |
|
"effective_num_docs": 135, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:astronomy": { |
|
"name": "arabic_mmlu_ht:astronomy", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "astronomy", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 152, |
|
"effective_num_docs": 152, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:business_ethics": { |
|
"name": "arabic_mmlu_ht:business_ethics", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "business_ethics", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 100, |
|
"effective_num_docs": 100, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:clinical_knowledge": { |
|
"name": "arabic_mmlu_ht:clinical_knowledge", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "clinical_knowledge", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 265, |
|
"effective_num_docs": 265, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:college_biology": { |
|
"name": "arabic_mmlu_ht:college_biology", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "college_biology", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 144, |
|
"effective_num_docs": 144, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:college_chemistry": { |
|
"name": "arabic_mmlu_ht:college_chemistry", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "college_chemistry", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 100, |
|
"effective_num_docs": 100, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:college_computer_science": { |
|
"name": "arabic_mmlu_ht:college_computer_science", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "college_computer_science", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 100, |
|
"effective_num_docs": 100, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:college_mathematics": { |
|
"name": "arabic_mmlu_ht:college_mathematics", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "college_mathematics", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 100, |
|
"effective_num_docs": 100, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:college_medicine": { |
|
"name": "arabic_mmlu_ht:college_medicine", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "college_medicine", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 173, |
|
"effective_num_docs": 173, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:college_physics": { |
|
"name": "arabic_mmlu_ht:college_physics", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "college_physics", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 102, |
|
"effective_num_docs": 102, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:computer_security": { |
|
"name": "arabic_mmlu_ht:computer_security", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "computer_security", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 100, |
|
"effective_num_docs": 100, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:conceptual_physics": { |
|
"name": "arabic_mmlu_ht:conceptual_physics", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "conceptual_physics", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 235, |
|
"effective_num_docs": 235, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:econometrics": { |
|
"name": "arabic_mmlu_ht:econometrics", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "econometrics", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 114, |
|
"effective_num_docs": 114, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:electrical_engineering": { |
|
"name": "arabic_mmlu_ht:electrical_engineering", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "electrical_engineering", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 145, |
|
"effective_num_docs": 145, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:elementary_mathematics": { |
|
"name": "arabic_mmlu_ht:elementary_mathematics", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "elementary_mathematics", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 378, |
|
"effective_num_docs": 378, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:formal_logic": { |
|
"name": "arabic_mmlu_ht:formal_logic", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "formal_logic", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 126, |
|
"effective_num_docs": 126, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:global_facts": { |
|
"name": "arabic_mmlu_ht:global_facts", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "global_facts", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 100, |
|
"effective_num_docs": 100, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:high_school_biology": { |
|
"name": "arabic_mmlu_ht:high_school_biology", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "high_school_biology", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 3813, |
|
"effective_num_docs": 3813, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:high_school_chemistry": { |
|
"name": "arabic_mmlu_ht:high_school_chemistry", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "high_school_chemistry", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 4016, |
|
"effective_num_docs": 4016, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:high_school_computer_science": { |
|
"name": "arabic_mmlu_ht:high_school_computer_science", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "high_school_computer_science", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 100, |
|
"effective_num_docs": 100, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:high_school_european_history": { |
|
"name": "arabic_mmlu_ht:high_school_european_history", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "high_school_european_history", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 8152, |
|
"effective_num_docs": 8152, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:high_school_geography": { |
|
"name": "arabic_mmlu_ht:high_school_geography", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "high_school_geography", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 198, |
|
"effective_num_docs": 198, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:high_school_government_and_politics": { |
|
"name": "arabic_mmlu_ht:high_school_government_and_politics", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "high_school_government_and_politics", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 193, |
|
"effective_num_docs": 193, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:high_school_macroeconomics": { |
|
"name": "arabic_mmlu_ht:high_school_macroeconomics", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "high_school_macroeconomics", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 2891, |
|
"effective_num_docs": 2891, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:high_school_mathematics": { |
|
"name": "arabic_mmlu_ht:high_school_mathematics", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "high_school_mathematics", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 270, |
|
"effective_num_docs": 270, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:high_school_microeconomics": { |
|
"name": "arabic_mmlu_ht:high_school_microeconomics", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "high_school_microeconomics", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 238, |
|
"effective_num_docs": 238, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:high_school_physics": { |
|
"name": "arabic_mmlu_ht:high_school_physics", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "high_school_physics", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 151, |
|
"effective_num_docs": 151, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:high_school_psychology": { |
|
"name": "arabic_mmlu_ht:high_school_psychology", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "high_school_psychology", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 545, |
|
"effective_num_docs": 545, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:high_school_statistics": { |
|
"name": "arabic_mmlu_ht:high_school_statistics", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "high_school_statistics", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 4232, |
|
"effective_num_docs": 4232, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:high_school_us_history": { |
|
"name": "arabic_mmlu_ht:high_school_us_history", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "high_school_us_history", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 204, |
|
"effective_num_docs": 204, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:high_school_world_history": { |
|
"name": "arabic_mmlu_ht:high_school_world_history", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "high_school_world_history", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 237, |
|
"effective_num_docs": 237, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:human_aging": { |
|
"name": "arabic_mmlu_ht:human_aging", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "human_aging", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 223, |
|
"effective_num_docs": 223, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:human_sexuality": { |
|
"name": "arabic_mmlu_ht:human_sexuality", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "human_sexuality", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 131, |
|
"effective_num_docs": 131, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:international_law": { |
|
"name": "arabic_mmlu_ht:international_law", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "international_law", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 121, |
|
"effective_num_docs": 121, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:jurisprudence": { |
|
"name": "arabic_mmlu_ht:jurisprudence", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "jurisprudence", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 108, |
|
"effective_num_docs": 108, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:logical_fallacies": { |
|
"name": "arabic_mmlu_ht:logical_fallacies", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "logical_fallacies", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 163, |
|
"effective_num_docs": 163, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:machine_learning": { |
|
"name": "arabic_mmlu_ht:machine_learning", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "machine_learning", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 112, |
|
"effective_num_docs": 112, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:management": { |
|
"name": "arabic_mmlu_ht:management", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "management", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 103, |
|
"effective_num_docs": 103, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:marketing": { |
|
"name": "arabic_mmlu_ht:marketing", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "marketing", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 234, |
|
"effective_num_docs": 234, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:medical_genetics": { |
|
"name": "arabic_mmlu_ht:medical_genetics", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "medical_genetics", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 100, |
|
"effective_num_docs": 100, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:miscellaneous": { |
|
"name": "arabic_mmlu_ht:miscellaneous", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "miscellaneous", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 2420, |
|
"effective_num_docs": 2420, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:moral_disputes": { |
|
"name": "arabic_mmlu_ht:moral_disputes", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "moral_disputes", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 346, |
|
"effective_num_docs": 346, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:moral_scenarios": { |
|
"name": "arabic_mmlu_ht:moral_scenarios", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "moral_scenarios", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 895, |
|
"effective_num_docs": 895, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:nutrition": { |
|
"name": "arabic_mmlu_ht:nutrition", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "nutrition", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 306, |
|
"effective_num_docs": 306, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:philosophy": { |
|
"name": "arabic_mmlu_ht:philosophy", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "philosophy", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 311, |
|
"effective_num_docs": 311, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:prehistory": { |
|
"name": "arabic_mmlu_ht:prehistory", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "prehistory", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 324, |
|
"effective_num_docs": 324, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:professional_accounting": { |
|
"name": "arabic_mmlu_ht:professional_accounting", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "professional_accounting", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 4514, |
|
"effective_num_docs": 4514, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:professional_law": { |
|
"name": "arabic_mmlu_ht:professional_law", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "professional_law", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 7987, |
|
"effective_num_docs": 7987, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:professional_medicine": { |
|
"name": "arabic_mmlu_ht:professional_medicine", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "professional_medicine", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 1637, |
|
"effective_num_docs": 1637, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:professional_psychology": { |
|
"name": "arabic_mmlu_ht:professional_psychology", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "professional_psychology", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 3503, |
|
"effective_num_docs": 3503, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:public_relations": { |
|
"name": "arabic_mmlu_ht:public_relations", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "public_relations", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 110, |
|
"effective_num_docs": 110, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:security_studies": { |
|
"name": "arabic_mmlu_ht:security_studies", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "security_studies", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 245, |
|
"effective_num_docs": 245, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:sociology": { |
|
"name": "arabic_mmlu_ht:sociology", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "sociology", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 201, |
|
"effective_num_docs": 201, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:us_foreign_policy": { |
|
"name": "arabic_mmlu_ht:us_foreign_policy", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "us_foreign_policy", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 100, |
|
"effective_num_docs": 100, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:virology": { |
|
"name": "arabic_mmlu_ht:virology", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "virology", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 166, |
|
"effective_num_docs": 166, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|arabic_mmlu_ht:world_religions": { |
|
"name": "arabic_mmlu_ht:world_religions", |
|
"prompt_function": "arabic_mmlu_ht_pfn", |
|
"hf_repo": "MBZUAI/human_translated_arabic_mmlu", |
|
"hf_subset": "world_religions", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 171, |
|
"effective_num_docs": 171, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|aratrust:Ethics": { |
|
"name": "aratrust:Ethics", |
|
"prompt_function": "aratrust_pfn", |
|
"hf_repo": "asas-ai/AraTrust-categorized", |
|
"hf_subset": "Ethics", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"train" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"train" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 60, |
|
"effective_num_docs": 60, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|aratrust:Illegal": { |
|
"name": "aratrust:Illegal", |
|
"prompt_function": "aratrust_pfn", |
|
"hf_repo": "asas-ai/AraTrust-categorized", |
|
"hf_subset": "Illegal", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"train" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"train" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 53, |
|
"effective_num_docs": 53, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|aratrust:MentalHealth": { |
|
"name": "aratrust:MentalHealth", |
|
"prompt_function": "aratrust_pfn", |
|
"hf_repo": "asas-ai/AraTrust-categorized", |
|
"hf_subset": "MentalHealth", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"train" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"train" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 76, |
|
"effective_num_docs": 76, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|aratrust:Offensive": { |
|
"name": "aratrust:Offensive", |
|
"prompt_function": "aratrust_pfn", |
|
"hf_repo": "asas-ai/AraTrust-categorized", |
|
"hf_subset": "Offensive", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"train" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"train" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 69, |
|
"effective_num_docs": 69, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|aratrust:PhysicalHealth": { |
|
"name": "aratrust:PhysicalHealth", |
|
"prompt_function": "aratrust_pfn", |
|
"hf_repo": "asas-ai/AraTrust-categorized", |
|
"hf_subset": "PhysicalHealth", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"train" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"train" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 73, |
|
"effective_num_docs": 73, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|aratrust:Privacy": { |
|
"name": "aratrust:Privacy", |
|
"prompt_function": "aratrust_pfn", |
|
"hf_repo": "asas-ai/AraTrust-categorized", |
|
"hf_subset": "Privacy", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"train" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"train" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 57, |
|
"effective_num_docs": 57, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|aratrust:Trustfulness": { |
|
"name": "aratrust:Trustfulness", |
|
"prompt_function": "aratrust_pfn", |
|
"hf_repo": "asas-ai/AraTrust-categorized", |
|
"hf_subset": "Trustfulness", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"train" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"train" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 78, |
|
"effective_num_docs": 78, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|aratrust:Unfairness": { |
|
"name": "aratrust:Unfairness", |
|
"prompt_function": "aratrust_pfn", |
|
"hf_repo": "asas-ai/AraTrust-categorized", |
|
"hf_subset": "Unfairness", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"train" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"train" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 55, |
|
"effective_num_docs": 55, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|madinah_qa:Arabic Language (General)": { |
|
"name": "madinah_qa:Arabic Language (General)", |
|
"prompt_function": "madinah_qa_pfn", |
|
"hf_repo": "MBZUAI/MadinahQA", |
|
"hf_subset": "Arabic Language (General)", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": [ |
|
"dev" |
|
], |
|
"few_shots_select": "sequential", |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 612, |
|
"effective_num_docs": 612, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|madinah_qa:Arabic Language (Grammar)": { |
|
"name": "madinah_qa:Arabic Language (Grammar)", |
|
"prompt_function": "madinah_qa_pfn", |
|
"hf_repo": "MBZUAI/MadinahQA", |
|
"hf_subset": "Arabic Language (Grammar)", |
|
"metric": [ |
|
{ |
|
"metric_name": "acc_norm", |
|
"higher_is_better": true, |
|
"category": "8", |
|
"use_case": "1", |
|
"sample_level_fn": "compute", |
|
"corpus_level_fn": "mean" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"test" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"test" |
|
], |
|
"few_shots_split": [ |
|
"dev" |
|
], |
|
"few_shots_select": "sequential", |
|
"generation_size": -1, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 365, |
|
"effective_num_docs": 365, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
}, |
|
"community|alrage_qa": { |
|
"name": "alrage_qa", |
|
"prompt_function": "qa_prompt_arabic", |
|
"hf_repo": "OALL/ALRAGE", |
|
"hf_subset": null, |
|
"metric": [ |
|
{ |
|
"metric_name": "llm_as_judge", |
|
"higher_is_better": true, |
|
"category": "7", |
|
"use_case": "10", |
|
"sample_level_fn": "_sample_level_fn", |
|
"corpus_level_fn": "aggregate_scores" |
|
} |
|
], |
|
"hf_revision": null, |
|
"hf_filter": null, |
|
"hf_avail_splits": [ |
|
"train" |
|
], |
|
"trust_dataset": true, |
|
"evaluation_splits": [ |
|
"train" |
|
], |
|
"few_shots_split": null, |
|
"few_shots_select": null, |
|
"generation_size": 200, |
|
"generation_grammar": null, |
|
"stop_sequence": [], |
|
"num_samples": null, |
|
"suite": [ |
|
"community" |
|
], |
|
"original_num_docs": 2106, |
|
"effective_num_docs": 2106, |
|
"must_remove_duplicate_docs": false, |
|
"version": 0 |
|
} |
|
}, |
|
"summary_tasks": { |
|
"community|alghafa:mcq_exams_test_ar|0": { |
|
"hashes": { |
|
"hash_examples": "c07a5e78c5c0b8fe", |
|
"hash_full_prompts": "50df102de4b86e82", |
|
"hash_input_tokens": "fee50ac58ba27e04", |
|
"hash_cont_tokens": "6298bcf4133bc4bf" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 557, |
|
"padded": 2228, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|alghafa:meta_ar_dialects|0": { |
|
"hashes": { |
|
"hash_examples": "c0b6081f83e14064", |
|
"hash_full_prompts": "0d5679424716794f", |
|
"hash_input_tokens": "0d3a31721428a55c", |
|
"hash_cont_tokens": "f3ae311cf2d55c37" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 5395, |
|
"padded": 21534, |
|
"non_padded": 46, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|alghafa:meta_ar_msa|0": { |
|
"hashes": { |
|
"hash_examples": "64eb78a7c5b7484b", |
|
"hash_full_prompts": "60febda0050ac528", |
|
"hash_input_tokens": "4186715f38afe0e8", |
|
"hash_cont_tokens": "e3eccdc3cced2a6d" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 895, |
|
"padded": 3557, |
|
"non_padded": 23, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|alghafa:multiple_choice_facts_truefalse_balanced_task|0": { |
|
"hashes": { |
|
"hash_examples": "54fc3502c1c02c06", |
|
"hash_full_prompts": "60d2d02ddf7724d2", |
|
"hash_input_tokens": "7a03bd3e33bc818b", |
|
"hash_cont_tokens": "dea1d140c30862e7" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 75, |
|
"padded": 150, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|alghafa:multiple_choice_grounded_statement_soqal_task|0": { |
|
"hashes": { |
|
"hash_examples": "46572d83696552ae", |
|
"hash_full_prompts": "ee38f44160123664", |
|
"hash_input_tokens": "ba3a0ef97ab3d888", |
|
"hash_cont_tokens": "add9f46c09bee79e" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 150, |
|
"padded": 748, |
|
"non_padded": 2, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|alghafa:multiple_choice_grounded_statement_xglue_mlqa_task|0": { |
|
"hashes": { |
|
"hash_examples": "f430d97ff715bc1c", |
|
"hash_full_prompts": "142561bc352704ed", |
|
"hash_input_tokens": "f700ebb3c62b28b1", |
|
"hash_cont_tokens": "b240d954c34bd7b6" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 150, |
|
"padded": 746, |
|
"non_padded": 4, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|alghafa:multiple_choice_rating_sentiment_no_neutral_task|0": { |
|
"hashes": { |
|
"hash_examples": "6b70a7416584f98c", |
|
"hash_full_prompts": "daf4e0a9770f25b6", |
|
"hash_input_tokens": "4b37e3336f0aa58d", |
|
"hash_cont_tokens": "ca09def8f62f5d8c" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 7995, |
|
"padded": 15990, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|alghafa:multiple_choice_rating_sentiment_task|0": { |
|
"hashes": { |
|
"hash_examples": "bc2005cc9d2f436e", |
|
"hash_full_prompts": "c87a81846cf52400", |
|
"hash_input_tokens": "70f3f192ab46a640", |
|
"hash_cont_tokens": "cec6dba4e08bd70d" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 5995, |
|
"padded": 17985, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|alghafa:multiple_choice_sentiment_task|0": { |
|
"hashes": { |
|
"hash_examples": "6fb0e254ea5945d8", |
|
"hash_full_prompts": "d648ef91892018d1", |
|
"hash_input_tokens": "d1e93871854441f1", |
|
"hash_cont_tokens": "66cfbc1751135a08" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 1720, |
|
"padded": 5157, |
|
"non_padded": 3, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_exams|0": { |
|
"hashes": { |
|
"hash_examples": "6d721df351722656", |
|
"hash_full_prompts": "3fe00c6b5bc78eda", |
|
"hash_input_tokens": "9f05fb4ba9f918f4", |
|
"hash_cont_tokens": "75237f9e94a4f168" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 537, |
|
"padded": 2148, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu:Accounting (University)|0": { |
|
"hashes": { |
|
"hash_examples": "30e09697562ff9e7", |
|
"hash_full_prompts": "5e4ef9bf88c666c0", |
|
"hash_input_tokens": "9f84d2f48a1f0898", |
|
"hash_cont_tokens": "8fa11c841c9bfc93" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 74, |
|
"padded": 256, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu:Arabic Language (General)|0": { |
|
"hashes": { |
|
"hash_examples": "bef69fb8b3b75f28", |
|
"hash_full_prompts": "10eb190f35d4edd0", |
|
"hash_input_tokens": "ec054a0b2981e7ea", |
|
"hash_cont_tokens": "40ab2b2b6b182e5a" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 612, |
|
"padded": 2403, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu:Arabic Language (Grammar)|0": { |
|
"hashes": { |
|
"hash_examples": "bd066a9e6a140a4b", |
|
"hash_full_prompts": "c5f7ffc0f3e9f8dd", |
|
"hash_input_tokens": "2ac0954de8fac923", |
|
"hash_cont_tokens": "05430de0f477f09e" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 365, |
|
"padded": 1588, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu:Arabic Language (High School)|0": { |
|
"hashes": { |
|
"hash_examples": "a9c2cd9a9929292a", |
|
"hash_full_prompts": "db9f4f3a73b5cdb6", |
|
"hash_input_tokens": "fb569fd02d2a782f", |
|
"hash_cont_tokens": "00d0251759af0b43" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 390, |
|
"padded": 1525, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu:Arabic Language (Middle School)|0": { |
|
"hashes": { |
|
"hash_examples": "2f8a77bbbd0e21ff", |
|
"hash_full_prompts": "c0b2bdfef349ba22", |
|
"hash_input_tokens": "385ea733504fef02", |
|
"hash_cont_tokens": "98ce77184a2930a0" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 27, |
|
"padded": 105, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu:Arabic Language (Primary School)|0": { |
|
"hashes": { |
|
"hash_examples": "5eed3da47822539b", |
|
"hash_full_prompts": "7016da0fd95da21b", |
|
"hash_input_tokens": "124a1b83fa1c207c", |
|
"hash_cont_tokens": "e5f11a1cc951f5bd" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 252, |
|
"padded": 926, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu:Biology (High School)|0": { |
|
"hashes": { |
|
"hash_examples": "91ae6d22a0f0213d", |
|
"hash_full_prompts": "77ccb14892b32830", |
|
"hash_input_tokens": "c380a74e47d4ab07", |
|
"hash_cont_tokens": "b5dce3b979a96832" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 1409, |
|
"padded": 5056, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu:Civics (High School)|0": { |
|
"hashes": { |
|
"hash_examples": "f27bf8791bea2bb9", |
|
"hash_full_prompts": "0692f4e052d8145d", |
|
"hash_input_tokens": "830c510da7f5c3b3", |
|
"hash_cont_tokens": "6d5588f4423e5132" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 87, |
|
"padded": 316, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu:Civics (Middle School)|0": { |
|
"hashes": { |
|
"hash_examples": "74f5bb0098c8916f", |
|
"hash_full_prompts": "eacbae886178cf45", |
|
"hash_input_tokens": "df001574d2eac88c", |
|
"hash_cont_tokens": "76a180f457ca19a1" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 236, |
|
"padded": 944, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu:Computer Science (High School)|0": { |
|
"hashes": { |
|
"hash_examples": "a4278d7b525d46fe", |
|
"hash_full_prompts": "399910921db4ddd4", |
|
"hash_input_tokens": "16e3c0cdf2f1b583", |
|
"hash_cont_tokens": "bc1b5cf7b54cc069" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 261, |
|
"padded": 1006, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu:Computer Science (Middle School)|0": { |
|
"hashes": { |
|
"hash_examples": "0cb6c07e4b80dfd4", |
|
"hash_full_prompts": "b2201ad0483257ce", |
|
"hash_input_tokens": "001d3ac4242ddf66", |
|
"hash_cont_tokens": "dd9da94ec77cf62d" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 27, |
|
"padded": 100, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu:Computer Science (Primary School)|0": { |
|
"hashes": { |
|
"hash_examples": "d96fc1bc32473533", |
|
"hash_full_prompts": "8f67b76f1ff9e07f", |
|
"hash_input_tokens": "14392291ed32fc8f", |
|
"hash_cont_tokens": "c5a684c386b25e19" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 190, |
|
"padded": 476, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu:Computer Science (University)|0": { |
|
"hashes": { |
|
"hash_examples": "8835587e436cbaff", |
|
"hash_full_prompts": "8fd7183d5834eb91", |
|
"hash_input_tokens": "ce04fab4b92df4b3", |
|
"hash_cont_tokens": "e29580cd31dfec6e" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 64, |
|
"padded": 255, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu:Driving Test|0": { |
|
"hashes": { |
|
"hash_examples": "7a4c38a2c451d075", |
|
"hash_full_prompts": "58e2728fe5d37afe", |
|
"hash_input_tokens": "5f1b4320948a7b9a", |
|
"hash_cont_tokens": "8d6c99caeca5836b" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 1211, |
|
"padded": 3685, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu:Economics (High School)|0": { |
|
"hashes": { |
|
"hash_examples": "c04c252836601279", |
|
"hash_full_prompts": "7fdab9db4c06223b", |
|
"hash_input_tokens": "6b61829f98fc7049", |
|
"hash_cont_tokens": "0a77e4926e5beba4" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 360, |
|
"padded": 1390, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu:Economics (Middle School)|0": { |
|
"hashes": { |
|
"hash_examples": "18fba1579406b3cc", |
|
"hash_full_prompts": "2fb8053d107a8ea2", |
|
"hash_input_tokens": "5fa87ed308b975d6", |
|
"hash_cont_tokens": "dedbb8e46a7fd6b0" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 87, |
|
"padded": 348, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu:Economics (University)|0": { |
|
"hashes": { |
|
"hash_examples": "7c9e86fba8151562", |
|
"hash_full_prompts": "e1240585b6ec0f19", |
|
"hash_input_tokens": "6b8df83e890a7105", |
|
"hash_cont_tokens": "4e4aabd9664fd27c" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 137, |
|
"padded": 544, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu:General Knowledge|0": { |
|
"hashes": { |
|
"hash_examples": "acfbe4e1f0314b85", |
|
"hash_full_prompts": "3154d1c7171cba3c", |
|
"hash_input_tokens": "4f25e6ac190777aa", |
|
"hash_cont_tokens": "be63857cc101df2a" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 864, |
|
"padded": 3213, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu:General Knowledge (Middle School)|0": { |
|
"hashes": { |
|
"hash_examples": "03cd0ecf10224316", |
|
"hash_full_prompts": "5e371aeb0bc255c2", |
|
"hash_input_tokens": "22c70f1372ce4e8c", |
|
"hash_cont_tokens": "7ee6e64cb937a18e" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 172, |
|
"padded": 628, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu:General Knowledge (Primary School)|0": { |
|
"hashes": { |
|
"hash_examples": "c3ee30196e05e122", |
|
"hash_full_prompts": "58165481a122b922", |
|
"hash_input_tokens": "26409f11e59769d1", |
|
"hash_cont_tokens": "00376542300cba93" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 162, |
|
"padded": 637, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu:Geography (High School)|0": { |
|
"hashes": { |
|
"hash_examples": "e2e329d2bdd9fb7b", |
|
"hash_full_prompts": "2f651aba1ae2100e", |
|
"hash_input_tokens": "1743acd448ca55b2", |
|
"hash_cont_tokens": "24baf866f8ea5ba7" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 1038, |
|
"padded": 4116, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu:Geography (Middle School)|0": { |
|
"hashes": { |
|
"hash_examples": "420b161444291989", |
|
"hash_full_prompts": "90fce9acba34416d", |
|
"hash_input_tokens": "eca3ac7435c762f0", |
|
"hash_cont_tokens": "42f577d5bfe3bf06" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 272, |
|
"padded": 975, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu:Geography (Primary School)|0": { |
|
"hashes": { |
|
"hash_examples": "5bc5ca48a4210899", |
|
"hash_full_prompts": "430f111fbd09cc3e", |
|
"hash_input_tokens": "291db13a2ee18ba7", |
|
"hash_cont_tokens": "c37a1ba8ab142ab4" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 57, |
|
"padded": 216, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu:History (High School)|0": { |
|
"hashes": { |
|
"hash_examples": "c7cc37f29311bea1", |
|
"hash_full_prompts": "d98cd11d54fe5ba7", |
|
"hash_input_tokens": "e37d64a915b9f26d", |
|
"hash_cont_tokens": "d696e08dce24701c" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 760, |
|
"padded": 2962, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu:History (Middle School)|0": { |
|
"hashes": { |
|
"hash_examples": "5b9f1973337153a2", |
|
"hash_full_prompts": "137d735836350dc0", |
|
"hash_input_tokens": "fb000b04193b3884", |
|
"hash_cont_tokens": "a826c56d1dc5baa8" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 203, |
|
"padded": 746, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu:History (Primary School)|0": { |
|
"hashes": { |
|
"hash_examples": "af2469847007c1fe", |
|
"hash_full_prompts": "ea40624218ebeccd", |
|
"hash_input_tokens": "f8d5a9c9b68866c3", |
|
"hash_cont_tokens": "0decdfe404b3b03c" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 102, |
|
"padded": 400, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu:Islamic Studies|0": { |
|
"hashes": { |
|
"hash_examples": "c8da9b2f16a5ea0f", |
|
"hash_full_prompts": "64940e9498267e98", |
|
"hash_input_tokens": "84d1d632aa366482", |
|
"hash_cont_tokens": "17f870abffa03234" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 639, |
|
"padded": 2529, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu:Islamic Studies (High School)|0": { |
|
"hashes": { |
|
"hash_examples": "efb11bc8ef398117", |
|
"hash_full_prompts": "f6a15e7ceb799f11", |
|
"hash_input_tokens": "15d59b43f7539e06", |
|
"hash_cont_tokens": "d3313afdddbdd685" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 334, |
|
"padded": 1281, |
|
"non_padded": 4, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu:Islamic Studies (Middle School)|0": { |
|
"hashes": { |
|
"hash_examples": "9e33ab030eebdb99", |
|
"hash_full_prompts": "72da83ccad3f1888", |
|
"hash_input_tokens": "6065297ef17a7fa0", |
|
"hash_cont_tokens": "1efb4160d023dff9" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 238, |
|
"padded": 883, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu:Islamic Studies (Primary School)|0": { |
|
"hashes": { |
|
"hash_examples": "4167565d878b20eb", |
|
"hash_full_prompts": "f12d5e7782b22a5a", |
|
"hash_input_tokens": "1877f31c6b2ff297", |
|
"hash_cont_tokens": "0e367fd748756eaa" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 999, |
|
"padded": 3024, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu:Law (Professional)|0": { |
|
"hashes": { |
|
"hash_examples": "e77f52c8fe4352b3", |
|
"hash_full_prompts": "a175f913c524efa6", |
|
"hash_input_tokens": "ad9e8721eca29d41", |
|
"hash_cont_tokens": "61930f84d8441e1f" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 314, |
|
"padded": 1232, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu:Management (University)|0": { |
|
"hashes": { |
|
"hash_examples": "09682649b04b7327", |
|
"hash_full_prompts": "efb82ca40a19ee24", |
|
"hash_input_tokens": "07eb7af322387b0e", |
|
"hash_cont_tokens": "0f9ab4266780528e" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 75, |
|
"padded": 200, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu:Math (Primary School)|0": { |
|
"hashes": { |
|
"hash_examples": "edb027bfae7e76f1", |
|
"hash_full_prompts": "586edbdfb191b711", |
|
"hash_input_tokens": "6d2528eab842217d", |
|
"hash_cont_tokens": "8da51174e60e06b8" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 409, |
|
"padded": 1296, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu:Natural Science (Middle School)|0": { |
|
"hashes": { |
|
"hash_examples": "96e72c9094c2364c", |
|
"hash_full_prompts": "7ae4a9cd64b50d22", |
|
"hash_input_tokens": "dbe5a874daa9d4a6", |
|
"hash_cont_tokens": "5f6d33db2d21b4e0" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 242, |
|
"padded": 940, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu:Natural Science (Primary School)|0": { |
|
"hashes": { |
|
"hash_examples": "69e35bad3dec5a4d", |
|
"hash_full_prompts": "3a18961622352804", |
|
"hash_input_tokens": "c7a2152343c91fa6", |
|
"hash_cont_tokens": "301213c2b5fe4f80" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 336, |
|
"padded": 1228, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu:Philosophy (High School)|0": { |
|
"hashes": { |
|
"hash_examples": "dc6ebd484a02fca5", |
|
"hash_full_prompts": "637c00b537d2211c", |
|
"hash_input_tokens": "201883794ed8953b", |
|
"hash_cont_tokens": "60299cdcf804fabe" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 39, |
|
"padded": 156, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu:Physics (High School)|0": { |
|
"hashes": { |
|
"hash_examples": "58a1722472c9e644", |
|
"hash_full_prompts": "1fcc61cf59d4ae3c", |
|
"hash_input_tokens": "44b9a693a221cca6", |
|
"hash_cont_tokens": "eecec325e9d06f62" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 255, |
|
"padded": 1020, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu:Political Science (University)|0": { |
|
"hashes": { |
|
"hash_examples": "07a4ed6aabbdfd1e", |
|
"hash_full_prompts": "7e859bbffc978e5b", |
|
"hash_input_tokens": "a6f5e09f3151f6d8", |
|
"hash_cont_tokens": "ea9735a77c6f7bbf" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 210, |
|
"padded": 710, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu:Social Science (Middle School)|0": { |
|
"hashes": { |
|
"hash_examples": "8ca955902f304664", |
|
"hash_full_prompts": "99ade070d833ba61", |
|
"hash_input_tokens": "35461c3886e55b50", |
|
"hash_cont_tokens": "e08af32c15f8a815" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 241, |
|
"padded": 929, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu:Social Science (Primary School)|0": { |
|
"hashes": { |
|
"hash_examples": "934025ab3738123c", |
|
"hash_full_prompts": "d4b29570235995db", |
|
"hash_input_tokens": "5f79e80741a6bff9", |
|
"hash_cont_tokens": "571954b03044a902" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 705, |
|
"padded": 2043, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:abstract_algebra|0": { |
|
"hashes": { |
|
"hash_examples": "0b557911f2f6d919", |
|
"hash_full_prompts": "25255dfef6e283b2", |
|
"hash_input_tokens": "bfdd25f8619a2a42", |
|
"hash_cont_tokens": "c0d03236b47a6493" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 100, |
|
"padded": 400, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:anatomy|0": { |
|
"hashes": { |
|
"hash_examples": "a552d8a0ef294061", |
|
"hash_full_prompts": "81bf36bc9685919f", |
|
"hash_input_tokens": "e6dc7120a3809114", |
|
"hash_cont_tokens": "b417e620105c047f" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 135, |
|
"padded": 540, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:astronomy|0": { |
|
"hashes": { |
|
"hash_examples": "c4a372d0af7da098", |
|
"hash_full_prompts": "6ac4cd0a6a7509fc", |
|
"hash_input_tokens": "4a4ded4fbcf55d81", |
|
"hash_cont_tokens": "b86103cb6bd55e0e" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 152, |
|
"padded": 608, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:business_ethics|0": { |
|
"hashes": { |
|
"hash_examples": "9f71d816abf8af7a", |
|
"hash_full_prompts": "e46425e40f127f5e", |
|
"hash_input_tokens": "28192db1b209f42f", |
|
"hash_cont_tokens": "c0d03236b47a6493" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 100, |
|
"padded": 400, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:clinical_knowledge|0": { |
|
"hashes": { |
|
"hash_examples": "38303cd765589ef3", |
|
"hash_full_prompts": "bf77ad6705f4e1f8", |
|
"hash_input_tokens": "cdabd5724fd2e95e", |
|
"hash_cont_tokens": "3995ea9272d22df2" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 265, |
|
"padded": 1060, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:college_biology|0": { |
|
"hashes": { |
|
"hash_examples": "dbd9b5d318e60b04", |
|
"hash_full_prompts": "8cb0957e6c5b1f0e", |
|
"hash_input_tokens": "153204effd9a843f", |
|
"hash_cont_tokens": "864e1d3815bcaff2" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 144, |
|
"padded": 576, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:college_chemistry|0": { |
|
"hashes": { |
|
"hash_examples": "6f88491d03db8a4c", |
|
"hash_full_prompts": "f353870876e4cb9c", |
|
"hash_input_tokens": "970306ea44cc2959", |
|
"hash_cont_tokens": "c0d03236b47a6493" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 100, |
|
"padded": 400, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:college_computer_science|0": { |
|
"hashes": { |
|
"hash_examples": "ebfdee5ef2ed5e17", |
|
"hash_full_prompts": "b09a07d785734f56", |
|
"hash_input_tokens": "df9aeccd238ba40a", |
|
"hash_cont_tokens": "c0d03236b47a6493" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 100, |
|
"padded": 396, |
|
"non_padded": 4, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:college_mathematics|0": { |
|
"hashes": { |
|
"hash_examples": "e3f22cd7712aae2f", |
|
"hash_full_prompts": "2209f8537529f782", |
|
"hash_input_tokens": "5cbfc17e171691d2", |
|
"hash_cont_tokens": "c0d03236b47a6493" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 100, |
|
"padded": 400, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:college_medicine|0": { |
|
"hashes": { |
|
"hash_examples": "51a5501373afb5a7", |
|
"hash_full_prompts": "019c4fca48caa16f", |
|
"hash_input_tokens": "be38c3311078e2a8", |
|
"hash_cont_tokens": "22f810bd178f52f1" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 173, |
|
"padded": 688, |
|
"non_padded": 4, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:college_physics|0": { |
|
"hashes": { |
|
"hash_examples": "2d3e015989b108db", |
|
"hash_full_prompts": "a08514ac0bf5bb13", |
|
"hash_input_tokens": "48d0e8c7734524f0", |
|
"hash_cont_tokens": "517dbad11707cb0f" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 102, |
|
"padded": 408, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:computer_security|0": { |
|
"hashes": { |
|
"hash_examples": "f8810eddc38dfee4", |
|
"hash_full_prompts": "521a2f42c16b4f78", |
|
"hash_input_tokens": "dfb6bb22f35944ee", |
|
"hash_cont_tokens": "c0d03236b47a6493" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 100, |
|
"padded": 396, |
|
"non_padded": 4, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:conceptual_physics|0": { |
|
"hashes": { |
|
"hash_examples": "211e32cc43c6b1dc", |
|
"hash_full_prompts": "4f697b14f3cc103f", |
|
"hash_input_tokens": "d74feb268bb77942", |
|
"hash_cont_tokens": "693ca99790e4a2cd" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 235, |
|
"padded": 940, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:econometrics|0": { |
|
"hashes": { |
|
"hash_examples": "810023786b2484d2", |
|
"hash_full_prompts": "fc44c12fc89a756d", |
|
"hash_input_tokens": "bd41eeffb49cd409", |
|
"hash_cont_tokens": "e08c06475509f605" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 114, |
|
"padded": 456, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:electrical_engineering|0": { |
|
"hashes": { |
|
"hash_examples": "a222760c93eaa1ee", |
|
"hash_full_prompts": "2d1fc1795ce2fc14", |
|
"hash_input_tokens": "c2bbf3226a1087fd", |
|
"hash_cont_tokens": "acfa6d8adcd38327" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 145, |
|
"padded": 580, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:elementary_mathematics|0": { |
|
"hashes": { |
|
"hash_examples": "4c069aeee64dc227", |
|
"hash_full_prompts": "afc9441292d3fcdb", |
|
"hash_input_tokens": "7b4e374efefe98db", |
|
"hash_cont_tokens": "c930454d64c3f1b9" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 378, |
|
"padded": 1512, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:formal_logic|0": { |
|
"hashes": { |
|
"hash_examples": "3cb0ccbf8e8a77ae", |
|
"hash_full_prompts": "b795fa0437fc08b3", |
|
"hash_input_tokens": "514d8aa56337e878", |
|
"hash_cont_tokens": "a150573cdded1037" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 126, |
|
"padded": 504, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:global_facts|0": { |
|
"hashes": { |
|
"hash_examples": "c1d039e64ea321b9", |
|
"hash_full_prompts": "d4a1fd62e48c7a37", |
|
"hash_input_tokens": "1e0f81ce53c7dbbe", |
|
"hash_cont_tokens": "c0d03236b47a6493" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 100, |
|
"padded": 400, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:high_school_biology|0": { |
|
"hashes": { |
|
"hash_examples": "ddcb8237bb4ba08a", |
|
"hash_full_prompts": "fc901dbb0fa16d9a", |
|
"hash_input_tokens": "c11b384be3dcc4fb", |
|
"hash_cont_tokens": "2a976151a68d6f37" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 3813, |
|
"padded": 15204, |
|
"non_padded": 48, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:high_school_chemistry|0": { |
|
"hashes": { |
|
"hash_examples": "07061b55c5c436d9", |
|
"hash_full_prompts": "67415214a442a446", |
|
"hash_input_tokens": "3352ddaed71151d4", |
|
"hash_cont_tokens": "588ab7b7548e226d" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 4016, |
|
"padded": 16016, |
|
"non_padded": 48, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:high_school_computer_science|0": { |
|
"hashes": { |
|
"hash_examples": "8d3405483d5fdcff", |
|
"hash_full_prompts": "57fd83ca448c291d", |
|
"hash_input_tokens": "d0a0eb47ae3814e1", |
|
"hash_cont_tokens": "c0d03236b47a6493" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 100, |
|
"padded": 400, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:high_school_european_history|0": { |
|
"hashes": { |
|
"hash_examples": "031c49a430356414", |
|
"hash_full_prompts": "694effea6c4fdd86", |
|
"hash_input_tokens": "1e7962f1ac585423", |
|
"hash_cont_tokens": "ff60759852180f16" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 8152, |
|
"padded": 32528, |
|
"non_padded": 80, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:high_school_geography|0": { |
|
"hashes": { |
|
"hash_examples": "d0ce2b019a66c1de", |
|
"hash_full_prompts": "7c017d06e74ff8ae", |
|
"hash_input_tokens": "3770e750a99e7442", |
|
"hash_cont_tokens": "90baf90c89612547" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 198, |
|
"padded": 792, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:high_school_government_and_politics|0": { |
|
"hashes": { |
|
"hash_examples": "7d7c6d476d0576b1", |
|
"hash_full_prompts": "75b94b2e95ba5ddd", |
|
"hash_input_tokens": "8c60626b927cb142", |
|
"hash_cont_tokens": "6522f95e3083c58e" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 193, |
|
"padded": 772, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:high_school_macroeconomics|0": { |
|
"hashes": { |
|
"hash_examples": "694d3a01c6144ddb", |
|
"hash_full_prompts": "3344f10aef970781", |
|
"hash_input_tokens": "846455caf793c37b", |
|
"hash_cont_tokens": "5541e1ca0ad8871a" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 2891, |
|
"padded": 11532, |
|
"non_padded": 32, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:high_school_mathematics|0": { |
|
"hashes": { |
|
"hash_examples": "004f9c0a40b5ec10", |
|
"hash_full_prompts": "be095effbdaf911b", |
|
"hash_input_tokens": "60285177023c66c4", |
|
"hash_cont_tokens": "864b6aea7a0abbe6" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 270, |
|
"padded": 1080, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:high_school_microeconomics|0": { |
|
"hashes": { |
|
"hash_examples": "80cf03d462e6ccbc", |
|
"hash_full_prompts": "ff8b6538bc36c477", |
|
"hash_input_tokens": "b11e17ec83dbe5ff", |
|
"hash_cont_tokens": "070606cd726440dc" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 238, |
|
"padded": 952, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:high_school_physics|0": { |
|
"hashes": { |
|
"hash_examples": "92218def5b383845", |
|
"hash_full_prompts": "5163927cdff41749", |
|
"hash_input_tokens": "21c2021bfef28e1e", |
|
"hash_cont_tokens": "04c5397e0399fe36" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 151, |
|
"padded": 604, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:high_school_psychology|0": { |
|
"hashes": { |
|
"hash_examples": "323f7848fee32e58", |
|
"hash_full_prompts": "3f83ab998283fce1", |
|
"hash_input_tokens": "279077938b150171", |
|
"hash_cont_tokens": "72a37bfe06323b4f" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 545, |
|
"padded": 2180, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:high_school_statistics|0": { |
|
"hashes": { |
|
"hash_examples": "d7bbe0d037cf31ec", |
|
"hash_full_prompts": "da04d9e222ec8a81", |
|
"hash_input_tokens": "16965605d642896d", |
|
"hash_cont_tokens": "358ecebb06070cab" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 4232, |
|
"padded": 16876, |
|
"non_padded": 52, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:high_school_us_history|0": { |
|
"hashes": { |
|
"hash_examples": "722ec9207e3b0e04", |
|
"hash_full_prompts": "0623396809ec41d0", |
|
"hash_input_tokens": "17668047017728fb", |
|
"hash_cont_tokens": "f613988e22c44a26" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 204, |
|
"padded": 816, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:high_school_world_history|0": { |
|
"hashes": { |
|
"hash_examples": "b5eb675d3b578584", |
|
"hash_full_prompts": "0770d75fb97812d7", |
|
"hash_input_tokens": "b6ac270e3445631c", |
|
"hash_cont_tokens": "9278ff714ae36b32" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 237, |
|
"padded": 944, |
|
"non_padded": 4, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:human_aging|0": { |
|
"hashes": { |
|
"hash_examples": "713ac79cd2dd2d7b", |
|
"hash_full_prompts": "97a6dc9c4140c7e1", |
|
"hash_input_tokens": "cc5f1b8922c64ccf", |
|
"hash_cont_tokens": "9760be94faed15ae" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 223, |
|
"padded": 892, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:human_sexuality|0": { |
|
"hashes": { |
|
"hash_examples": "47551ab4e5dcf6c5", |
|
"hash_full_prompts": "29268ef4fcf451a3", |
|
"hash_input_tokens": "5f15f99cacfa07a7", |
|
"hash_cont_tokens": "73db881e3aadfc8c" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 131, |
|
"padded": 524, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:international_law|0": { |
|
"hashes": { |
|
"hash_examples": "da360336943398d5", |
|
"hash_full_prompts": "15222e8543c0b9e2", |
|
"hash_input_tokens": "c04466b77ab61475", |
|
"hash_cont_tokens": "b0a2085f83940a50" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 121, |
|
"padded": 484, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:jurisprudence|0": { |
|
"hashes": { |
|
"hash_examples": "661d161a486fb035", |
|
"hash_full_prompts": "36f873e3cb84ddd2", |
|
"hash_input_tokens": "808440f3a5f63fd7", |
|
"hash_cont_tokens": "a6c872d9060b8ec0" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 108, |
|
"padded": 432, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:logical_fallacies|0": { |
|
"hashes": { |
|
"hash_examples": "5c3926384758bda7", |
|
"hash_full_prompts": "4879b5c98f6956fe", |
|
"hash_input_tokens": "25cdc6671a641586", |
|
"hash_cont_tokens": "17b8b5c0476c3e3c" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 163, |
|
"padded": 652, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:machine_learning|0": { |
|
"hashes": { |
|
"hash_examples": "3ce756e6a22ffc48", |
|
"hash_full_prompts": "bc968de2062d1bc2", |
|
"hash_input_tokens": "8e268ebe7959e556", |
|
"hash_cont_tokens": "10c4f0efb96ecdad" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 112, |
|
"padded": 448, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:management|0": { |
|
"hashes": { |
|
"hash_examples": "20fe769bb3276832", |
|
"hash_full_prompts": "285e2791f58778dd", |
|
"hash_input_tokens": "91f56afd8ee3888e", |
|
"hash_cont_tokens": "4ecc8e86741d7b02" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 103, |
|
"padded": 412, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:marketing|0": { |
|
"hashes": { |
|
"hash_examples": "6b19449559d987ce", |
|
"hash_full_prompts": "3bd62d13a5dd9503", |
|
"hash_input_tokens": "aeeb6b52d97513dc", |
|
"hash_cont_tokens": "6e029c7a30ef270e" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 234, |
|
"padded": 936, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:medical_genetics|0": { |
|
"hashes": { |
|
"hash_examples": "cbb0fa9df0f5435a", |
|
"hash_full_prompts": "35bba4d50dee48a3", |
|
"hash_input_tokens": "71f480b050c3f442", |
|
"hash_cont_tokens": "c0d03236b47a6493" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 100, |
|
"padded": 400, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:miscellaneous|0": { |
|
"hashes": { |
|
"hash_examples": "0a4134046c23cff9", |
|
"hash_full_prompts": "f283f15c2128f381", |
|
"hash_input_tokens": "e31b04e3798f3ea1", |
|
"hash_cont_tokens": "83638f569790237f" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 2420, |
|
"padded": 9648, |
|
"non_padded": 32, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:moral_disputes|0": { |
|
"hashes": { |
|
"hash_examples": "1ac8a0967c82caa0", |
|
"hash_full_prompts": "b0d0e86b5ce0e3a8", |
|
"hash_input_tokens": "8e9efd27b78f26c9", |
|
"hash_cont_tokens": "67462b6d37c032bf" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 346, |
|
"padded": 1384, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:moral_scenarios|0": { |
|
"hashes": { |
|
"hash_examples": "2c0670188bc5a789", |
|
"hash_full_prompts": "7cb2d697a626928f", |
|
"hash_input_tokens": "5f5deee709dfd879", |
|
"hash_cont_tokens": "7980d1a9ee7bb909" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 895, |
|
"padded": 3464, |
|
"non_padded": 116, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:nutrition|0": { |
|
"hashes": { |
|
"hash_examples": "658628c0dcdfe201", |
|
"hash_full_prompts": "1e814ae35c700d6c", |
|
"hash_input_tokens": "58835e09742b7582", |
|
"hash_cont_tokens": "5ac03c482c4e283d" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 306, |
|
"padded": 1201, |
|
"non_padded": 23, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:philosophy|0": { |
|
"hashes": { |
|
"hash_examples": "8b6707b322affafd", |
|
"hash_full_prompts": "9c5027e998f48b23", |
|
"hash_input_tokens": "1a3f41747ceb139e", |
|
"hash_cont_tokens": "baade138b461202c" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 311, |
|
"padded": 1212, |
|
"non_padded": 32, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:prehistory|0": { |
|
"hashes": { |
|
"hash_examples": "0c85ffcdc9a7b367", |
|
"hash_full_prompts": "7a160c1f41cb51c5", |
|
"hash_input_tokens": "b27432e71f37f316", |
|
"hash_cont_tokens": "37a5d15cced7a5a8" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 324, |
|
"padded": 1284, |
|
"non_padded": 12, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:professional_accounting|0": { |
|
"hashes": { |
|
"hash_examples": "cce1ea2d5f544b2f", |
|
"hash_full_prompts": "d6eea3dd86222f16", |
|
"hash_input_tokens": "2a214477312495fb", |
|
"hash_cont_tokens": "8e2b01f47ec6d3ec" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 4514, |
|
"padded": 17810, |
|
"non_padded": 246, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:professional_law|0": { |
|
"hashes": { |
|
"hash_examples": "1c654b024b54eb4b", |
|
"hash_full_prompts": "cd5515703a610fde", |
|
"hash_input_tokens": "7d28f411e1b3fd6c", |
|
"hash_cont_tokens": "6a7313bf29caef19" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 7987, |
|
"padded": 31524, |
|
"non_padded": 424, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:professional_medicine|0": { |
|
"hashes": { |
|
"hash_examples": "c621eaacfa662ebc", |
|
"hash_full_prompts": "7a7b12a5923bafda", |
|
"hash_input_tokens": "10dd1c95f58b0f67", |
|
"hash_cont_tokens": "e2fdde4c7a0f845a" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 1637, |
|
"padded": 6424, |
|
"non_padded": 124, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:professional_psychology|0": { |
|
"hashes": { |
|
"hash_examples": "bc14a28eaec87dc4", |
|
"hash_full_prompts": "93c91f7e228775f7", |
|
"hash_input_tokens": "53253c9c03b040a2", |
|
"hash_cont_tokens": "05fc904c67915014" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 3503, |
|
"padded": 13664, |
|
"non_padded": 348, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:public_relations|0": { |
|
"hashes": { |
|
"hash_examples": "de4989d9375885c4", |
|
"hash_full_prompts": "27b6462bdfb0a61d", |
|
"hash_input_tokens": "fd39eec1c44c0918", |
|
"hash_cont_tokens": "81e7c6b1e9dbff78" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 110, |
|
"padded": 424, |
|
"non_padded": 16, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:security_studies|0": { |
|
"hashes": { |
|
"hash_examples": "3f84bfeec717c6de", |
|
"hash_full_prompts": "5f8339074e8d3d9c", |
|
"hash_input_tokens": "76417504f0bd0260", |
|
"hash_cont_tokens": "fa5ca980c67fef6a" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 245, |
|
"padded": 972, |
|
"non_padded": 8, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:sociology|0": { |
|
"hashes": { |
|
"hash_examples": "10d7c2fae10bfcbc", |
|
"hash_full_prompts": "252de12518ce92cf", |
|
"hash_input_tokens": "660bf34d24d7c30f", |
|
"hash_cont_tokens": "46f7119587c4b28e" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 201, |
|
"padded": 784, |
|
"non_padded": 20, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:us_foreign_policy|0": { |
|
"hashes": { |
|
"hash_examples": "bb05f02c38ddaf1a", |
|
"hash_full_prompts": "8fed2556033dfe2b", |
|
"hash_input_tokens": "531cc5ae9d8ad708", |
|
"hash_cont_tokens": "c0d03236b47a6493" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 100, |
|
"padded": 388, |
|
"non_padded": 12, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:virology|0": { |
|
"hashes": { |
|
"hash_examples": "290915a48884ede2", |
|
"hash_full_prompts": "4525725d0115b364", |
|
"hash_input_tokens": "26ee7fb347176d77", |
|
"hash_cont_tokens": "75e807433b52bb4e" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 166, |
|
"padded": 652, |
|
"non_padded": 12, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|arabic_mmlu_ht:world_religions|0": { |
|
"hashes": { |
|
"hash_examples": "91cc5451c7284f75", |
|
"hash_full_prompts": "e8fe149b3b9de442", |
|
"hash_input_tokens": "fba6586a6b9c2725", |
|
"hash_cont_tokens": "7ea4bcdddebed342" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 171, |
|
"padded": 676, |
|
"non_padded": 8, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|aratrust:Ethics|0": { |
|
"hashes": { |
|
"hash_examples": "5d32da36271c5eb4", |
|
"hash_full_prompts": "39ef71ab4f7527dc", |
|
"hash_input_tokens": "10cfae5b7297f8ee", |
|
"hash_cont_tokens": "17a4cb422db288c6" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 60, |
|
"padded": 165, |
|
"non_padded": 15, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|aratrust:Illegal|0": { |
|
"hashes": { |
|
"hash_examples": "0c07f1f100f2d0e8", |
|
"hash_full_prompts": "3d4bc19320672207", |
|
"hash_input_tokens": "93d8d7dbc39b0720", |
|
"hash_cont_tokens": "f2119ef2bb5067e7" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 53, |
|
"padded": 159, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|aratrust:MentalHealth|0": { |
|
"hashes": { |
|
"hash_examples": "8e5fc5c4704bd96b", |
|
"hash_full_prompts": "86ee31e82b27fdcf", |
|
"hash_input_tokens": "d70ca339fdb89347", |
|
"hash_cont_tokens": "b72352f82361935d" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 76, |
|
"padded": 222, |
|
"non_padded": 6, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|aratrust:Offensive|0": { |
|
"hashes": { |
|
"hash_examples": "5ad4369b7dc5de46", |
|
"hash_full_prompts": "919779469ac4fbf9", |
|
"hash_input_tokens": "f49fa7b6639e7fb3", |
|
"hash_cont_tokens": "f5a982c220104685" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 69, |
|
"padded": 204, |
|
"non_padded": 3, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|aratrust:PhysicalHealth|0": { |
|
"hashes": { |
|
"hash_examples": "dc2a632e2dcc86db", |
|
"hash_full_prompts": "a0d63b1ae223c5e2", |
|
"hash_input_tokens": "6e013b9d1524828e", |
|
"hash_cont_tokens": "1cb6c7361a426fe4" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 73, |
|
"padded": 216, |
|
"non_padded": 3, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|aratrust:Privacy|0": { |
|
"hashes": { |
|
"hash_examples": "295e35448a39e003", |
|
"hash_full_prompts": "c8ac44ba6530a572", |
|
"hash_input_tokens": "4932c8f9f142b200", |
|
"hash_cont_tokens": "e83d4019353a4792" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 57, |
|
"padded": 165, |
|
"non_padded": 6, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|aratrust:Trustfulness|0": { |
|
"hashes": { |
|
"hash_examples": "e79ac1ea5439e623", |
|
"hash_full_prompts": "9bb9b22c4d2e7748", |
|
"hash_input_tokens": "103ab5c9eb05ab07", |
|
"hash_cont_tokens": "e34b9e160c794149" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 78, |
|
"padded": 234, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|aratrust:Unfairness|0": { |
|
"hashes": { |
|
"hash_examples": "4ac5dccbfbdc5077", |
|
"hash_full_prompts": "c28123e1afce8b41", |
|
"hash_input_tokens": "062f56494395f53e", |
|
"hash_cont_tokens": "aa0a1aa00c71d791" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 55, |
|
"padded": 159, |
|
"non_padded": 6, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|madinah_qa:Arabic Language (General)|0": { |
|
"hashes": { |
|
"hash_examples": "bef69fb8b3b75f28", |
|
"hash_full_prompts": "10eb190f35d4edd0", |
|
"hash_input_tokens": "e8b2db27d7c4ba70", |
|
"hash_cont_tokens": "efee81d0bd50a7d9" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 612, |
|
"padded": 2358, |
|
"non_padded": 45, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|madinah_qa:Arabic Language (Grammar)|0": { |
|
"hashes": { |
|
"hash_examples": "bd066a9e6a140a4b", |
|
"hash_full_prompts": "c5f7ffc0f3e9f8dd", |
|
"hash_input_tokens": "b79bce4437b76bb0", |
|
"hash_cont_tokens": "8a651699c506209b" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 365, |
|
"padded": 1517, |
|
"non_padded": 71, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
}, |
|
"community|alrage_qa|0": { |
|
"hashes": { |
|
"hash_examples": "3edbbe22cabd4160", |
|
"hash_full_prompts": "8a1b5ff4a4bb7972", |
|
"hash_input_tokens": "a86a54f8743fb554", |
|
"hash_cont_tokens": "8a817acfc1a0054e" |
|
}, |
|
"truncated": 2106, |
|
"non_truncated": 0, |
|
"padded": 2106, |
|
"non_padded": 0, |
|
"effective_few_shots": 0.0, |
|
"num_truncated_few_shots": 0 |
|
} |
|
}, |
|
"summary_general": { |
|
"hashes": { |
|
"hash_examples": "a700a7984c46cac8", |
|
"hash_full_prompts": "e4fcf0896b29029a", |
|
"hash_input_tokens": "b0a2c5bc7cebbf50", |
|
"hash_cont_tokens": "94eced91038fd472" |
|
}, |
|
"truncated": 0, |
|
"non_truncated": 91862, |
|
"padded": 335980, |
|
"non_padded": 1946, |
|
"num_truncated_few_shots": 0 |
|
} |
|
} |