diff --git a/README.md b/README.md index 4d425309a24fa5ac6f84db904fb4408f8929c529..68aad2aba2fac7a96133862885c83736d9c37c93 100644 --- a/README.md +++ b/README.md @@ -26,6 +26,8 @@ dataset_info: features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -38,23 +40,56 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 19328 + num_bytes: 21316 num_examples: 100 - name: validation - num_bytes: 2024 + num_bytes: 2232 num_examples: 11 - name: dev - num_bytes: 830 + num_bytes: 918 num_examples: 5 - download_size: 166184960 - dataset_size: 160623559 + download_size: 47179638 + dataset_size: 161025091 +- config_name: all + features: + - name: question + dtype: string + - name: subject + dtype: string + - name: choices + sequence: string + - name: answer + dtype: + class_label: + names: + '0': A + '1': B + '2': C + '3': D + splits: + - name: auxiliary_train + num_bytes: 161000625 + num_examples: 99842 + - name: test + num_bytes: 6967453 + num_examples: 14042 + - name: validation + num_bytes: 763484 + num_examples: 1531 + - name: dev + num_bytes: 125353 + num_examples: 285 + download_size: 51132212 + dataset_size: 168856915 - config_name: anatomy features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -67,23 +102,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 33121 + num_bytes: 34594 num_examples: 135 - name: validation - num_bytes: 3140 + num_bytes: 3282 num_examples: 14 - name: dev - num_bytes: 967 + num_bytes: 1010 num_examples: 5 - download_size: 166184960 - dataset_size: 160638605 + download_size: 47191229 + dataset_size: 161039511 - config_name: astronomy features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -96,23 +133,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 46771 + num_bytes: 48735 num_examples: 152 - name: validation - num_bytes: 5027 + num_bytes: 5223 num_examples: 16 - name: dev - num_bytes: 2076 + num_bytes: 2129 num_examples: 5 - download_size: 166184960 - dataset_size: 160655251 + download_size: 47201551 + dataset_size: 161056712 - config_name: business_ethics features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -125,23 +164,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 33252 + num_bytes: 35140 num_examples: 100 - name: validation - num_bytes: 3038 + num_bytes: 3235 num_examples: 11 - name: dev - num_bytes: 2190 + num_bytes: 2273 num_examples: 5 - download_size: 166184960 - dataset_size: 160639857 + download_size: 47193421 + dataset_size: 161041273 - config_name: clinical_knowledge features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -154,23 +195,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 62754 + num_bytes: 68572 num_examples: 265 - name: validation - num_bytes: 6664 + num_bytes: 7290 num_examples: 29 - name: dev - num_bytes: 1210 + num_bytes: 1308 num_examples: 5 - download_size: 166184960 - dataset_size: 160672005 + download_size: 47213955 + dataset_size: 161077795 - config_name: college_biology features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -183,23 +226,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 48797 + num_bytes: 51521 num_examples: 144 - name: validation - num_bytes: 4819 + num_bytes: 5111 num_examples: 16 - name: dev - num_bytes: 1532 + num_bytes: 1615 num_examples: 5 - download_size: 166184960 - dataset_size: 160656525 + download_size: 47205152 + dataset_size: 161058872 - config_name: college_chemistry features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -212,23 +257,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 24708 + num_bytes: 26796 num_examples: 100 - name: validation - num_bytes: 2328 + num_bytes: 2484 num_examples: 8 - name: dev - num_bytes: 1331 + num_bytes: 1424 num_examples: 5 - download_size: 166184960 - dataset_size: 160629744 + download_size: 47188958 + dataset_size: 161031329 - config_name: college_computer_science features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -241,23 +288,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 42641 + num_bytes: 45429 num_examples: 100 - name: validation - num_bytes: 4663 + num_bytes: 4959 num_examples: 11 - name: dev - num_bytes: 2765 + num_bytes: 2893 num_examples: 5 - download_size: 166184960 - dataset_size: 160651446 + download_size: 47201966 + dataset_size: 161053906 - config_name: college_mathematics features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -270,23 +319,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 24711 + num_bytes: 26999 num_examples: 100 - name: validation - num_bytes: 2668 + num_bytes: 2909 num_examples: 11 - name: dev - num_bytes: 1493 + num_bytes: 1596 num_examples: 5 - download_size: 166184960 - dataset_size: 160630249 + download_size: 47188597 + dataset_size: 161032129 - config_name: college_medicine features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -299,23 +350,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 82397 + num_bytes: 85845 num_examples: 173 - name: validation - num_bytes: 7909 + num_bytes: 8337 num_examples: 22 - name: dev - num_bytes: 1670 + num_bytes: 1758 num_examples: 5 - download_size: 166184960 - dataset_size: 160693353 + download_size: 47218201 + dataset_size: 161096565 - config_name: college_physics features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -328,23 +381,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 30181 + num_bytes: 32107 num_examples: 102 - name: validation - num_bytes: 3490 + num_bytes: 3687 num_examples: 11 - name: dev - num_bytes: 1412 + num_bytes: 1495 num_examples: 5 - download_size: 166184960 - dataset_size: 160636460 + download_size: 47190901 + dataset_size: 161037914 - config_name: computer_security features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -357,23 +412,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 27124 + num_bytes: 29212 num_examples: 100 - name: validation - num_bytes: 4549 + num_bytes: 4768 num_examples: 11 - name: dev - num_bytes: 1101 + num_bytes: 1194 num_examples: 5 - download_size: 166184960 - dataset_size: 160634151 + download_size: 47192155 + dataset_size: 161035799 - config_name: conceptual_physics features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -386,23 +443,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 40709 + num_bytes: 45867 num_examples: 235 - name: validation - num_bytes: 4474 + num_bytes: 5034 num_examples: 26 - name: dev - num_bytes: 934 + num_bytes: 1032 num_examples: 5 - download_size: 166184960 - dataset_size: 160647494 + download_size: 47197231 + dataset_size: 161052558 - config_name: econometrics features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -415,23 +474,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 46547 + num_bytes: 48359 num_examples: 114 - name: validation - num_bytes: 4967 + num_bytes: 5147 num_examples: 12 - name: dev - num_bytes: 1644 + num_bytes: 1712 num_examples: 5 - download_size: 166184960 - dataset_size: 160654535 + download_size: 47197846 + dataset_size: 161055843 - config_name: electrical_engineering features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -444,23 +505,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 25142 + num_bytes: 28900 num_examples: 145 - name: validation - num_bytes: 2903 + num_bytes: 3307 num_examples: 16 - name: dev - num_bytes: 972 + num_bytes: 1090 num_examples: 5 - download_size: 166184960 - dataset_size: 160630394 + download_size: 47189021 + dataset_size: 161033922 - config_name: elementary_mathematics features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -473,23 +536,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 70108 + num_bytes: 79924 num_examples: 378 - name: validation - num_bytes: 8988 + num_bytes: 10042 num_examples: 41 - name: dev - num_bytes: 1440 + num_bytes: 1558 num_examples: 5 - download_size: 166184960 - dataset_size: 160681913 + download_size: 47216972 + dataset_size: 161092149 - config_name: formal_logic features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -502,23 +567,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 49785 + num_bytes: 51789 num_examples: 126 - name: validation - num_bytes: 6252 + num_bytes: 6464 num_examples: 14 - name: dev - num_bytes: 1757 + num_bytes: 1825 num_examples: 5 - download_size: 166184960 - dataset_size: 160659171 + download_size: 47194349 + dataset_size: 161060703 - config_name: global_facts features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -531,23 +598,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 18403 + num_bytes: 19991 num_examples: 100 - name: validation - num_bytes: 1865 + num_bytes: 2013 num_examples: 10 - name: dev - num_bytes: 1229 + num_bytes: 1297 num_examples: 5 - download_size: 166184960 - dataset_size: 160622874 + download_size: 47181634 + dataset_size: 161023926 - config_name: high_school_biology features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -560,23 +629,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 109732 + num_bytes: 116850 num_examples: 310 - name: validation - num_bytes: 11022 + num_bytes: 11746 num_examples: 32 - name: dev - num_bytes: 1673 + num_bytes: 1776 num_examples: 5 - download_size: 166184960 - dataset_size: 160723804 + download_size: 47239946 + dataset_size: 161130997 - config_name: high_school_chemistry features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -589,23 +660,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 58464 + num_bytes: 63527 num_examples: 203 - name: validation - num_bytes: 7092 + num_bytes: 7630 num_examples: 22 - name: dev - num_bytes: 1220 + num_bytes: 1333 num_examples: 5 - download_size: 166184960 - dataset_size: 160668153 + download_size: 47207769 + dataset_size: 161073115 - config_name: high_school_computer_science features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -618,23 +691,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 44476 + num_bytes: 47664 num_examples: 100 - name: validation - num_bytes: 3343 + num_bytes: 3619 num_examples: 9 - name: dev - num_bytes: 2918 + num_bytes: 3066 num_examples: 5 - download_size: 166184960 - dataset_size: 160652114 + download_size: 47200669 + dataset_size: 161054974 - config_name: high_school_european_history features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -647,23 +722,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 270300 + num_bytes: 275568 num_examples: 165 - name: validation - num_bytes: 29632 + num_bytes: 30196 num_examples: 18 - name: dev - num_bytes: 11564 + num_bytes: 11712 num_examples: 5 - download_size: 166184960 - dataset_size: 160912873 + download_size: 47349494 + dataset_size: 161318101 - config_name: high_school_geography features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -676,23 +753,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 42034 + num_bytes: 46972 num_examples: 198 - name: validation - num_bytes: 4332 + num_bytes: 4870 num_examples: 22 - name: dev - num_bytes: 1403 + num_bytes: 1516 num_examples: 5 - download_size: 166184960 - dataset_size: 160649146 + download_size: 47200648 + dataset_size: 161053983 - config_name: high_school_government_and_politics features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -705,23 +784,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 66074 + num_bytes: 73589 num_examples: 193 - name: validation - num_bytes: 7063 + num_bytes: 7870 num_examples: 21 - name: dev - num_bytes: 1779 + num_bytes: 1962 num_examples: 5 - download_size: 166184960 - dataset_size: 160676293 + download_size: 47214961 + dataset_size: 161084046 - config_name: high_school_macroeconomics features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -734,23 +815,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 117687 + num_bytes: 129375 num_examples: 390 - name: validation - num_bytes: 13020 + num_bytes: 14298 num_examples: 43 - name: dev - num_bytes: 1328 + num_bytes: 1466 num_examples: 5 - download_size: 166184960 - dataset_size: 160733412 + download_size: 47231078 + dataset_size: 161145764 - config_name: high_school_mathematics features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -763,23 +846,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 54854 + num_bytes: 62132 num_examples: 270 - name: validation - num_bytes: 5765 + num_bytes: 6536 num_examples: 29 - name: dev - num_bytes: 1297 + num_bytes: 1420 num_examples: 5 - download_size: 166184960 - dataset_size: 160663293 + download_size: 47207019 + dataset_size: 161070713 - config_name: high_school_microeconomics features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -792,23 +877,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 75703 + num_bytes: 82831 num_examples: 238 - name: validation - num_bytes: 7553 + num_bytes: 8321 num_examples: 26 - name: dev - num_bytes: 1298 + num_bytes: 1436 num_examples: 5 - download_size: 166184960 - dataset_size: 160685931 + download_size: 47212056 + dataset_size: 161093213 - config_name: high_school_physics features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -821,23 +908,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 59538 + num_bytes: 62999 num_examples: 151 - name: validation - num_bytes: 6771 + num_bytes: 7150 num_examples: 17 - name: dev - num_bytes: 1489 + num_bytes: 1592 num_examples: 5 - download_size: 166184960 - dataset_size: 160669175 + download_size: 47207223 + dataset_size: 161072366 - config_name: high_school_psychology features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -850,23 +939,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 159407 + num_bytes: 173565 num_examples: 545 - name: validation - num_bytes: 17269 + num_bytes: 18817 num_examples: 60 - name: dev - num_bytes: 1905 + num_bytes: 2023 num_examples: 5 - download_size: 166184960 - dataset_size: 160779958 + download_size: 47274794 + dataset_size: 161195030 - config_name: high_school_statistics features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -879,23 +970,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 110702 + num_bytes: 116306 num_examples: 216 - name: validation - num_bytes: 9997 + num_bytes: 10583 num_examples: 23 - name: dev - num_bytes: 2528 + num_bytes: 2646 num_examples: 5 - download_size: 166184960 - dataset_size: 160724604 + download_size: 47236110 + dataset_size: 161130160 - config_name: high_school_us_history features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -908,23 +1001,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 296734 + num_bytes: 302026 num_examples: 204 - name: validation - num_bytes: 31706 + num_bytes: 32266 num_examples: 22 - name: dev - num_bytes: 8864 + num_bytes: 8982 num_examples: 5 - download_size: 166184960 - dataset_size: 160938681 + download_size: 47355390 + dataset_size: 161343899 - config_name: high_school_world_history features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -937,23 +1032,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 378617 + num_bytes: 385478 num_examples: 237 - name: validation - num_bytes: 45501 + num_bytes: 46243 num_examples: 26 - name: dev - num_bytes: 4882 + num_bytes: 5015 num_examples: 5 - download_size: 166184960 - dataset_size: 161030377 + download_size: 47405363 + dataset_size: 161437361 - config_name: human_aging features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -966,23 +1063,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 46098 + num_bytes: 49431 num_examples: 223 - name: validation - num_bytes: 4707 + num_bytes: 5040 num_examples: 23 - name: dev - num_bytes: 1008 + num_bytes: 1071 num_examples: 5 - download_size: 166184960 - dataset_size: 160653190 + download_size: 47203557 + dataset_size: 161056167 - config_name: human_sexuality features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -995,23 +1094,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 32110 + num_bytes: 34587 num_examples: 131 - name: validation - num_bytes: 2421 + num_bytes: 2637 num_examples: 12 - name: dev - num_bytes: 1077 + num_bytes: 1160 num_examples: 5 - download_size: 166184960 - dataset_size: 160636985 + download_size: 47194557 + dataset_size: 161039009 - config_name: international_law features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -1024,23 +1125,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 53531 + num_bytes: 56060 num_examples: 121 - name: validation - num_bytes: 6473 + num_bytes: 6734 num_examples: 13 - name: dev - num_bytes: 2418 + num_bytes: 2511 num_examples: 5 - download_size: 166184960 - dataset_size: 160663799 + download_size: 47203538 + dataset_size: 161065930 - config_name: jurisprudence features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -1053,23 +1156,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 33986 + num_bytes: 35810 num_examples: 108 - name: validation - num_bytes: 3729 + num_bytes: 3904 num_examples: 11 - name: dev - num_bytes: 1303 + num_bytes: 1376 num_examples: 5 - download_size: 166184960 - dataset_size: 160640395 + download_size: 47195740 + dataset_size: 161041715 - config_name: logical_fallacies features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -1082,23 +1187,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 50117 + num_bytes: 53528 num_examples: 163 - name: validation - num_bytes: 5103 + num_bytes: 5469 num_examples: 18 - name: dev - num_bytes: 1573 + num_bytes: 1666 num_examples: 5 - download_size: 166184960 - dataset_size: 160658170 + download_size: 47196007 + dataset_size: 161061288 - config_name: machine_learning features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -1111,23 +1218,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 33880 + num_bytes: 36108 num_examples: 112 - name: validation - num_bytes: 3232 + num_bytes: 3440 num_examples: 11 - name: dev - num_bytes: 2323 + num_bytes: 2411 num_examples: 5 - download_size: 166184960 - dataset_size: 160640812 + download_size: 47192772 + dataset_size: 161042584 - config_name: management features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -1140,23 +1249,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 20002 + num_bytes: 21432 num_examples: 103 - name: validation - num_bytes: 1820 + num_bytes: 1962 num_examples: 11 - name: dev - num_bytes: 898 + num_bytes: 956 num_examples: 5 - download_size: 166184960 - dataset_size: 160624097 + download_size: 47185208 + dataset_size: 161024975 - config_name: marketing features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -1169,23 +1280,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 63025 + num_bytes: 66055 num_examples: 234 - name: validation - num_bytes: 7394 + num_bytes: 7707 num_examples: 25 - name: dev - num_bytes: 1481 + num_bytes: 1534 num_examples: 5 - download_size: 166184960 - dataset_size: 160673277 + download_size: 47211905 + dataset_size: 161075921 - config_name: medical_genetics features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -1198,23 +1311,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 20864 + num_bytes: 22852 num_examples: 100 - name: validation - num_bytes: 3005 + num_bytes: 3213 num_examples: 11 - name: dev - num_bytes: 1089 + num_bytes: 1177 num_examples: 5 - download_size: 166184960 - dataset_size: 160626335 + download_size: 47187825 + dataset_size: 161027867 - config_name: miscellaneous features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -1227,23 +1342,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 147704 + num_bytes: 161003 num_examples: 783 - name: validation - num_bytes: 14330 + num_bytes: 15780 num_examples: 86 - name: dev - num_bytes: 699 + num_bytes: 772 num_examples: 5 - download_size: 166184960 - dataset_size: 160764110 + download_size: 47277261 + dataset_size: 161178180 - config_name: moral_disputes features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -1256,23 +1373,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 107818 + num_bytes: 114034 num_examples: 346 - name: validation - num_bytes: 12420 + num_bytes: 13092 num_examples: 38 - name: dev - num_bytes: 1755 + num_bytes: 1833 num_examples: 5 - download_size: 166184960 - dataset_size: 160723370 + download_size: 47238228 + dataset_size: 161129584 - config_name: moral_scenarios features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -1285,23 +1404,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 374026 + num_bytes: 391019 num_examples: 895 - name: validation - num_bytes: 42338 + num_bytes: 44226 num_examples: 100 - name: dev - num_bytes: 2058 + num_bytes: 2141 num_examples: 5 - download_size: 166184960 - dataset_size: 161019799 + download_size: 47270775 + dataset_size: 161438011 - config_name: nutrition features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -1314,23 +1435,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 92410 + num_bytes: 96376 num_examples: 306 - name: validation - num_bytes: 8436 + num_bytes: 8853 num_examples: 33 - name: dev - num_bytes: 2085 + num_bytes: 2138 num_examples: 5 - download_size: 166184960 - dataset_size: 160704308 + download_size: 47230622 + dataset_size: 161107992 - config_name: philosophy features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -1343,23 +1466,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 80073 + num_bytes: 84415 num_examples: 311 - name: validation - num_bytes: 9184 + num_bytes: 9648 num_examples: 34 - name: dev - num_bytes: 988 + num_bytes: 1046 num_examples: 5 - download_size: 166184960 - dataset_size: 160691622 + download_size: 47223948 + dataset_size: 161095734 - config_name: prehistory features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -1372,23 +1497,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 89594 + num_bytes: 94118 num_examples: 324 - name: validation - num_bytes: 10285 + num_bytes: 10763 num_examples: 35 - name: dev - num_bytes: 1878 + num_bytes: 1936 num_examples: 5 - download_size: 166184960 - dataset_size: 160703134 + download_size: 47231112 + dataset_size: 161107442 - config_name: professional_accounting features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -1401,23 +1528,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 124550 + num_bytes: 132152 num_examples: 282 - name: validation - num_bytes: 14372 + num_bytes: 15197 num_examples: 31 - name: dev - num_bytes: 2148 + num_bytes: 2271 num_examples: 5 - download_size: 166184960 - dataset_size: 160742447 + download_size: 47249006 + dataset_size: 161150245 - config_name: professional_law features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -1430,23 +1559,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 1891762 + num_bytes: 1922430 num_examples: 1534 - name: validation - num_bytes: 203519 + num_bytes: 206907 num_examples: 170 - name: dev - num_bytes: 6610 + num_bytes: 6698 num_examples: 5 - download_size: 166184960 - dataset_size: 162703268 + download_size: 48319302 + dataset_size: 163136660 - config_name: professional_medicine features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -1459,23 +1590,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 217561 + num_bytes: 224349 num_examples: 272 - name: validation - num_bytes: 23847 + num_bytes: 24610 num_examples: 31 - name: dev - num_bytes: 3807 + num_bytes: 3920 num_examples: 5 - download_size: 166184960 - dataset_size: 160846592 + download_size: 47312228 + dataset_size: 161253504 - config_name: professional_psychology features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -1488,23 +1621,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 225899 + num_bytes: 242411 num_examples: 612 - name: validation - num_bytes: 29101 + num_bytes: 30952 num_examples: 69 - name: dev - num_bytes: 2267 + num_bytes: 2390 num_examples: 5 - download_size: 166184960 - dataset_size: 160858644 + download_size: 47321118 + dataset_size: 161276378 - config_name: public_relations features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -1517,23 +1652,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 28760 + num_bytes: 30948 num_examples: 110 - name: validation - num_bytes: 4566 + num_bytes: 4794 num_examples: 12 - name: dev - num_bytes: 1496 + num_bytes: 1584 num_examples: 5 - download_size: 166184960 - dataset_size: 160636199 + download_size: 47193575 + dataset_size: 161037951 - config_name: security_studies features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -1546,23 +1683,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 204844 + num_bytes: 209732 num_examples: 245 - name: validation - num_bytes: 22637 + num_bytes: 23165 num_examples: 27 - name: dev - num_bytes: 5335 + num_bytes: 5423 num_examples: 5 - download_size: 166184960 - dataset_size: 160834193 + download_size: 47300997 + dataset_size: 161238945 - config_name: sociology features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -1575,23 +1714,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 66243 + num_bytes: 68844 num_examples: 201 - name: validation - num_bytes: 7184 + num_bytes: 7458 num_examples: 22 - name: dev - num_bytes: 1613 + num_bytes: 1666 num_examples: 5 - download_size: 166184960 - dataset_size: 160676417 + download_size: 47218696 + dataset_size: 161078593 - config_name: us_foreign_policy features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -1604,23 +1745,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 28443 + num_bytes: 30531 num_examples: 100 - name: validation - num_bytes: 3264 + num_bytes: 3483 num_examples: 11 - name: dev - num_bytes: 1611 + num_bytes: 1704 num_examples: 5 - download_size: 166184960 - dataset_size: 160634695 + download_size: 47191346 + dataset_size: 161036343 - config_name: virology features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -1633,23 +1776,25 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 38759 + num_bytes: 40739 num_examples: 166 - name: validation - num_bytes: 5463 + num_bytes: 5667 num_examples: 18 - name: dev - num_bytes: 1096 + num_bytes: 1144 num_examples: 5 - download_size: 166184960 - dataset_size: 160646695 + download_size: 47200276 + dataset_size: 161048175 - config_name: world_religions features: - name: question dtype: string + - name: subject + dtype: string - name: choices sequence: string - name: answer @@ -1662,19 +1807,600 @@ dataset_info: '3': D splits: - name: auxiliary_train - num_bytes: 160601377 + num_bytes: 161000625 num_examples: 99842 - name: test - num_bytes: 25274 + num_bytes: 28511 num_examples: 171 - name: validation - num_bytes: 2765 + num_bytes: 3114 num_examples: 19 - name: dev - num_bytes: 670 + num_bytes: 753 num_examples: 5 - download_size: 166184960 - dataset_size: 160630086 + download_size: 47189681 + dataset_size: 161033003 +configs: +- config_name: abstract_algebra + data_files: + - split: auxiliary_train + path: abstract_algebra/auxiliary_train-* + - split: test + path: abstract_algebra/test-* + - split: validation + path: abstract_algebra/validation-* + - split: dev + path: abstract_algebra/dev-* +- config_name: all + data_files: + - split: auxiliary_train + path: all/auxiliary_train-* + - split: test + path: all/test-* + - split: validation + path: all/validation-* + - split: dev + path: all/dev-* +- config_name: anatomy + data_files: + - split: auxiliary_train + path: anatomy/auxiliary_train-* + - split: test + path: anatomy/test-* + - split: validation + path: anatomy/validation-* + - split: dev + path: anatomy/dev-* +- config_name: astronomy + data_files: + - split: auxiliary_train + path: astronomy/auxiliary_train-* + - split: test + path: astronomy/test-* + - split: validation + path: astronomy/validation-* + - split: dev + path: astronomy/dev-* +- config_name: business_ethics + data_files: + - split: auxiliary_train + path: business_ethics/auxiliary_train-* + - split: test + path: business_ethics/test-* + - split: validation + path: business_ethics/validation-* + - split: dev + path: business_ethics/dev-* +- config_name: clinical_knowledge + data_files: + - split: auxiliary_train + path: clinical_knowledge/auxiliary_train-* + - split: test + path: clinical_knowledge/test-* + - split: validation + path: clinical_knowledge/validation-* + - split: dev + path: clinical_knowledge/dev-* +- config_name: college_biology + data_files: + - split: auxiliary_train + path: college_biology/auxiliary_train-* + - split: test + path: college_biology/test-* + - split: validation + path: college_biology/validation-* + - split: dev + path: college_biology/dev-* +- config_name: college_chemistry + data_files: + - split: auxiliary_train + path: college_chemistry/auxiliary_train-* + - split: test + path: college_chemistry/test-* + - split: validation + path: college_chemistry/validation-* + - split: dev + path: college_chemistry/dev-* +- config_name: college_computer_science + data_files: + - split: auxiliary_train + path: college_computer_science/auxiliary_train-* + - split: test + path: college_computer_science/test-* + - split: validation + path: college_computer_science/validation-* + - split: dev + path: college_computer_science/dev-* +- config_name: college_mathematics + data_files: + - split: auxiliary_train + path: college_mathematics/auxiliary_train-* + - split: test + path: college_mathematics/test-* + - split: validation + path: college_mathematics/validation-* + - split: dev + path: college_mathematics/dev-* +- config_name: college_medicine + data_files: + - split: auxiliary_train + path: college_medicine/auxiliary_train-* + - split: test + path: college_medicine/test-* + - split: validation + path: college_medicine/validation-* + - split: dev + path: college_medicine/dev-* +- config_name: college_physics + data_files: + - split: auxiliary_train + path: college_physics/auxiliary_train-* + - split: test + path: college_physics/test-* + - split: validation + path: college_physics/validation-* + - split: dev + path: college_physics/dev-* +- config_name: computer_security + data_files: + - split: auxiliary_train + path: computer_security/auxiliary_train-* + - split: test + path: computer_security/test-* + - split: validation + path: computer_security/validation-* + - split: dev + path: computer_security/dev-* +- config_name: conceptual_physics + data_files: + - split: auxiliary_train + path: conceptual_physics/auxiliary_train-* + - split: test + path: conceptual_physics/test-* + - split: validation + path: conceptual_physics/validation-* + - split: dev + path: conceptual_physics/dev-* +- config_name: econometrics + data_files: + - split: auxiliary_train + path: econometrics/auxiliary_train-* + - split: test + path: econometrics/test-* + - split: validation + path: econometrics/validation-* + - split: dev + path: econometrics/dev-* +- config_name: electrical_engineering + data_files: + - split: auxiliary_train + path: electrical_engineering/auxiliary_train-* + - split: test + path: electrical_engineering/test-* + - split: validation + path: electrical_engineering/validation-* + - split: dev + path: electrical_engineering/dev-* +- config_name: elementary_mathematics + data_files: + - split: auxiliary_train + path: elementary_mathematics/auxiliary_train-* + - split: test + path: elementary_mathematics/test-* + - split: validation + path: elementary_mathematics/validation-* + - split: dev + path: elementary_mathematics/dev-* +- config_name: formal_logic + data_files: + - split: auxiliary_train + path: formal_logic/auxiliary_train-* + - split: test + path: formal_logic/test-* + - split: validation + path: formal_logic/validation-* + - split: dev + path: formal_logic/dev-* +- config_name: global_facts + data_files: + - split: auxiliary_train + path: global_facts/auxiliary_train-* + - split: test + path: global_facts/test-* + - split: validation + path: global_facts/validation-* + - split: dev + path: global_facts/dev-* +- config_name: high_school_biology + data_files: + - split: auxiliary_train + path: high_school_biology/auxiliary_train-* + - split: test + path: high_school_biology/test-* + - split: validation + path: high_school_biology/validation-* + - split: dev + path: high_school_biology/dev-* +- config_name: high_school_chemistry + data_files: + - split: auxiliary_train + path: high_school_chemistry/auxiliary_train-* + - split: test + path: high_school_chemistry/test-* + - split: validation + path: high_school_chemistry/validation-* + - split: dev + path: high_school_chemistry/dev-* +- config_name: high_school_computer_science + data_files: + - split: auxiliary_train + path: high_school_computer_science/auxiliary_train-* + - split: test + path: high_school_computer_science/test-* + - split: validation + path: high_school_computer_science/validation-* + - split: dev + path: high_school_computer_science/dev-* +- config_name: high_school_european_history + data_files: + - split: auxiliary_train + path: high_school_european_history/auxiliary_train-* + - split: test + path: high_school_european_history/test-* + - split: validation + path: high_school_european_history/validation-* + - split: dev + path: high_school_european_history/dev-* +- config_name: high_school_geography + data_files: + - split: auxiliary_train + path: high_school_geography/auxiliary_train-* + - split: test + path: high_school_geography/test-* + - split: validation + path: high_school_geography/validation-* + - split: dev + path: high_school_geography/dev-* +- config_name: high_school_government_and_politics + data_files: + - split: auxiliary_train + path: high_school_government_and_politics/auxiliary_train-* + - split: test + path: high_school_government_and_politics/test-* + - split: validation + path: high_school_government_and_politics/validation-* + - split: dev + path: high_school_government_and_politics/dev-* +- config_name: high_school_macroeconomics + data_files: + - split: auxiliary_train + path: high_school_macroeconomics/auxiliary_train-* + - split: test + path: high_school_macroeconomics/test-* + - split: validation + path: high_school_macroeconomics/validation-* + - split: dev + path: high_school_macroeconomics/dev-* +- config_name: high_school_mathematics + data_files: + - split: auxiliary_train + path: high_school_mathematics/auxiliary_train-* + - split: test + path: high_school_mathematics/test-* + - split: validation + path: high_school_mathematics/validation-* + - split: dev + path: high_school_mathematics/dev-* +- config_name: high_school_microeconomics + data_files: + - split: auxiliary_train + path: high_school_microeconomics/auxiliary_train-* + - split: test + path: high_school_microeconomics/test-* + - split: validation + path: high_school_microeconomics/validation-* + - split: dev + path: high_school_microeconomics/dev-* +- config_name: high_school_physics + data_files: + - split: auxiliary_train + path: high_school_physics/auxiliary_train-* + - split: test + path: high_school_physics/test-* + - split: validation + path: high_school_physics/validation-* + - split: dev + path: high_school_physics/dev-* +- config_name: high_school_psychology + data_files: + - split: auxiliary_train + path: high_school_psychology/auxiliary_train-* + - split: test + path: high_school_psychology/test-* + - split: validation + path: high_school_psychology/validation-* + - split: dev + path: high_school_psychology/dev-* +- config_name: high_school_statistics + data_files: + - split: auxiliary_train + path: high_school_statistics/auxiliary_train-* + - split: test + path: high_school_statistics/test-* + - split: validation + path: high_school_statistics/validation-* + - split: dev + path: high_school_statistics/dev-* +- config_name: high_school_us_history + data_files: + - split: auxiliary_train + path: high_school_us_history/auxiliary_train-* + - split: test + path: high_school_us_history/test-* + - split: validation + path: high_school_us_history/validation-* + - split: dev + path: high_school_us_history/dev-* +- config_name: high_school_world_history + data_files: + - split: auxiliary_train + path: high_school_world_history/auxiliary_train-* + - split: test + path: high_school_world_history/test-* + - split: validation + path: high_school_world_history/validation-* + - split: dev + path: high_school_world_history/dev-* +- config_name: human_aging + data_files: + - split: auxiliary_train + path: human_aging/auxiliary_train-* + - split: test + path: human_aging/test-* + - split: validation + path: human_aging/validation-* + - split: dev + path: human_aging/dev-* +- config_name: human_sexuality + data_files: + - split: auxiliary_train + path: human_sexuality/auxiliary_train-* + - split: test + path: human_sexuality/test-* + - split: validation + path: human_sexuality/validation-* + - split: dev + path: human_sexuality/dev-* +- config_name: international_law + data_files: + - split: auxiliary_train + path: international_law/auxiliary_train-* + - split: test + path: international_law/test-* + - split: validation + path: international_law/validation-* + - split: dev + path: international_law/dev-* +- config_name: jurisprudence + data_files: + - split: auxiliary_train + path: jurisprudence/auxiliary_train-* + - split: test + path: jurisprudence/test-* + - split: validation + path: jurisprudence/validation-* + - split: dev + path: jurisprudence/dev-* +- config_name: logical_fallacies + data_files: + - split: auxiliary_train + path: logical_fallacies/auxiliary_train-* + - split: test + path: logical_fallacies/test-* + - split: validation + path: logical_fallacies/validation-* + - split: dev + path: logical_fallacies/dev-* +- config_name: machine_learning + data_files: + - split: auxiliary_train + path: machine_learning/auxiliary_train-* + - split: test + path: machine_learning/test-* + - split: validation + path: machine_learning/validation-* + - split: dev + path: machine_learning/dev-* +- config_name: management + data_files: + - split: auxiliary_train + path: management/auxiliary_train-* + - split: test + path: management/test-* + - split: validation + path: management/validation-* + - split: dev + path: management/dev-* +- config_name: marketing + data_files: + - split: auxiliary_train + path: marketing/auxiliary_train-* + - split: test + path: marketing/test-* + - split: validation + path: marketing/validation-* + - split: dev + path: marketing/dev-* +- config_name: medical_genetics + data_files: + - split: auxiliary_train + path: medical_genetics/auxiliary_train-* + - split: test + path: medical_genetics/test-* + - split: validation + path: medical_genetics/validation-* + - split: dev + path: medical_genetics/dev-* +- config_name: miscellaneous + data_files: + - split: auxiliary_train + path: miscellaneous/auxiliary_train-* + - split: test + path: miscellaneous/test-* + - split: validation + path: miscellaneous/validation-* + - split: dev + path: miscellaneous/dev-* +- config_name: moral_disputes + data_files: + - split: auxiliary_train + path: moral_disputes/auxiliary_train-* + - split: test + path: moral_disputes/test-* + - split: validation + path: moral_disputes/validation-* + - split: dev + path: moral_disputes/dev-* +- config_name: moral_scenarios + data_files: + - split: auxiliary_train + path: moral_scenarios/auxiliary_train-* + - split: test + path: moral_scenarios/test-* + - split: validation + path: moral_scenarios/validation-* + - split: dev + path: moral_scenarios/dev-* +- config_name: nutrition + data_files: + - split: auxiliary_train + path: nutrition/auxiliary_train-* + - split: test + path: nutrition/test-* + - split: validation + path: nutrition/validation-* + - split: dev + path: nutrition/dev-* +- config_name: philosophy + data_files: + - split: auxiliary_train + path: philosophy/auxiliary_train-* + - split: test + path: philosophy/test-* + - split: validation + path: philosophy/validation-* + - split: dev + path: philosophy/dev-* +- config_name: prehistory + data_files: + - split: auxiliary_train + path: prehistory/auxiliary_train-* + - split: test + path: prehistory/test-* + - split: validation + path: prehistory/validation-* + - split: dev + path: prehistory/dev-* +- config_name: professional_accounting + data_files: + - split: auxiliary_train + path: professional_accounting/auxiliary_train-* + - split: test + path: professional_accounting/test-* + - split: validation + path: professional_accounting/validation-* + - split: dev + path: professional_accounting/dev-* +- config_name: professional_law + data_files: + - split: auxiliary_train + path: professional_law/auxiliary_train-* + - split: test + path: professional_law/test-* + - split: validation + path: professional_law/validation-* + - split: dev + path: professional_law/dev-* +- config_name: professional_medicine + data_files: + - split: auxiliary_train + path: professional_medicine/auxiliary_train-* + - split: test + path: professional_medicine/test-* + - split: validation + path: professional_medicine/validation-* + - split: dev + path: professional_medicine/dev-* +- config_name: professional_psychology + data_files: + - split: auxiliary_train + path: professional_psychology/auxiliary_train-* + - split: test + path: professional_psychology/test-* + - split: validation + path: professional_psychology/validation-* + - split: dev + path: professional_psychology/dev-* +- config_name: public_relations + data_files: + - split: auxiliary_train + path: public_relations/auxiliary_train-* + - split: test + path: public_relations/test-* + - split: validation + path: public_relations/validation-* + - split: dev + path: public_relations/dev-* +- config_name: security_studies + data_files: + - split: auxiliary_train + path: security_studies/auxiliary_train-* + - split: test + path: security_studies/test-* + - split: validation + path: security_studies/validation-* + - split: dev + path: security_studies/dev-* +- config_name: sociology + data_files: + - split: auxiliary_train + path: sociology/auxiliary_train-* + - split: test + path: sociology/test-* + - split: validation + path: sociology/validation-* + - split: dev + path: sociology/dev-* +- config_name: us_foreign_policy + data_files: + - split: auxiliary_train + path: us_foreign_policy/auxiliary_train-* + - split: test + path: us_foreign_policy/test-* + - split: validation + path: us_foreign_policy/validation-* + - split: dev + path: us_foreign_policy/dev-* +- config_name: virology + data_files: + - split: auxiliary_train + path: virology/auxiliary_train-* + - split: test + path: virology/test-* + - split: validation + path: virology/validation-* + - split: dev + path: virology/dev-* +- config_name: world_religions + data_files: + - split: auxiliary_train + path: world_religions/auxiliary_train-* + - split: test + path: world_religions/test-* + - split: validation + path: world_religions/validation-* + - split: dev + path: world_religions/dev-* --- # Dataset Card for MMLU diff --git a/abstract_algebra/auxiliary_train-00000-of-00001.parquet b/abstract_algebra/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/abstract_algebra/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/abstract_algebra/dev-00000-of-00001.parquet b/abstract_algebra/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..3d4e1cab7ca98a91a638e2b88cf138cbb27b763e --- /dev/null +++ b/abstract_algebra/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e12113ec8170bc06892413970124fe1ad73991e89f708094b8af37f915b67ed6 +size 2919 diff --git a/abstract_algebra/test-00000-of-00001.parquet b/abstract_algebra/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..c1fce5c51b13e6a2ea50d375cf5ed97d14ff180d --- /dev/null +++ b/abstract_algebra/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2d2cc95a39503ecbd1999b674894c9579dd3244aa76a9e525bbf19bb990f6720 +size 9494 diff --git a/abstract_algebra/validation-00000-of-00001.parquet b/abstract_algebra/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..71e752e50e6582d492447eb543c9bd7921c1ebe8 --- /dev/null +++ b/abstract_algebra/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:972bb7cf0abab41ee11dbdacdc3e6f26a93dc911205216ed6b3de1da98e385a8 +size 3270 diff --git a/all/auxiliary_train-00000-of-00001.parquet b/all/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/all/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/all/dev-00000-of-00001.parquet b/all/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..8d004022952c8f09997e6844bb50a14551d81f52 --- /dev/null +++ b/all/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fe12a6b67efba3305c8aa397028f8596fd1968f6eb7fe9e93033fc0f3bbd5acb +size 75910 diff --git a/all/test-00000-of-00001.parquet b/all/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..0c20e1bd2268d86cce2158efce168e4372265daf --- /dev/null +++ b/all/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:28aaa6445a394922deca70f8ba22baaa1aa5b0d65464e5133b7cbed5468887b6 +size 3485812 diff --git a/all/validation-00000-of-00001.parquet b/all/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..8a1e42ca73e826d12850f5588e6bc9edfa64b366 --- /dev/null +++ b/all/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f20355a02ce253a23a3c7f06ea83d55435ed77fe11a40ee432ddf0f7fa32d0b2 +size 406535 diff --git a/anatomy/auxiliary_train-00000-of-00001.parquet b/anatomy/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/anatomy/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/anatomy/dev-00000-of-00001.parquet b/anatomy/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..d6f06838b4722a5fb1621fce7c82dc868efcd21e --- /dev/null +++ b/anatomy/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e42aef33f442b00d95ac7eb84b2f582d95857662fcad4fc7885ab67823f5b9e3 +size 3014 diff --git a/anatomy/test-00000-of-00001.parquet b/anatomy/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..3acc46389c9ff62f6bc7f29c9173de23bf04d1b3 --- /dev/null +++ b/anatomy/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2322eec0d77992b8ea97c6bef6a7c7696f1c83ed0125b8dab65eba889d9569c3 +size 19548 diff --git a/anatomy/validation-00000-of-00001.parquet b/anatomy/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..308b36ee43f1afeacb469b224bc839f96963fe8d --- /dev/null +++ b/anatomy/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a6458974e0a2fb2e4885be0c27dc04a7091b9084869def767784ecce1a32ef56 +size 4712 diff --git a/astronomy/auxiliary_train-00000-of-00001.parquet b/astronomy/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/astronomy/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/astronomy/dev-00000-of-00001.parquet b/astronomy/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..daa3659e03f0c0140a8fad9a49bc33f8fcf61f4e --- /dev/null +++ b/astronomy/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:23ea71eb6555b7f89b99eeebe86b5deb95faf6bca70b2bb983b6d589a0ac7055 +size 4211 diff --git a/astronomy/test-00000-of-00001.parquet b/astronomy/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..5a8b7f0194e484cbbaf2873acaa1477a2693e078 --- /dev/null +++ b/astronomy/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:097742de0ec4687bd0cd8eb86a9fb38f3fe3498dca45d4cb09f630badc1d5598 +size 27806 diff --git a/astronomy/validation-00000-of-00001.parquet b/astronomy/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..d956c5eedb84fa11100f942d71d2dff76b1964cf --- /dev/null +++ b/astronomy/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:16eb5b3b1ca9c64c195238e33c8b0cac1f2e5b606198398d5b8bb445864465f7 +size 5579 diff --git a/business_ethics/auxiliary_train-00000-of-00001.parquet b/business_ethics/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/business_ethics/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/business_ethics/dev-00000-of-00001.parquet b/business_ethics/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..86df82639d230633a9620f40d219082d2c982558 --- /dev/null +++ b/business_ethics/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:809e98c869462d507c867b00c3232d9bfc31affb345c0a53adf2fc4568967296 +size 4221 diff --git a/business_ethics/test-00000-of-00001.parquet b/business_ethics/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..64b9d414f40fe5c5c06db4dac4837d09bdabb9f3 --- /dev/null +++ b/business_ethics/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:99d34fcaf6a31f7462bf3fad2432903796a2f0672467c0cfde6e8608d4f10faf +size 20701 diff --git a/business_ethics/validation-00000-of-00001.parquet b/business_ethics/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..e350e503d89a5f41ce7aa1686bc778dc34f20ea0 --- /dev/null +++ b/business_ethics/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2bb543d985898fd2630a67f02a449e5d3308ed91bc3ba52374ef74cab013cf0d +size 4544 diff --git a/clinical_knowledge/auxiliary_train-00000-of-00001.parquet b/clinical_knowledge/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/clinical_knowledge/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/clinical_knowledge/dev-00000-of-00001.parquet b/clinical_knowledge/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..b65e5179b4989042a3a3135f593c2b8ceea81ec2 --- /dev/null +++ b/clinical_knowledge/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6ef415ee4ee8eec3b77f97726f7e264f7cf97ac880293ac14955bd4abf898925 +size 3189 diff --git a/clinical_knowledge/test-00000-of-00001.parquet b/clinical_knowledge/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..1eb91e6fbdde47b793c85a5e64814ac9876a288c --- /dev/null +++ b/clinical_knowledge/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:81a74fd1121699ebbdbe8c6f44a4efe4577aa8792bca091c7094be09896a2e36 +size 39869 diff --git a/clinical_knowledge/validation-00000-of-00001.parquet b/clinical_knowledge/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..faf73f3a0b4d7611cdc61f09f25b4762406021e7 --- /dev/null +++ b/clinical_knowledge/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:66a168f080bf08d4d18b9ec0db2a0416328b228e321dca24ea116ae931f8e79f +size 6942 diff --git a/college_biology/auxiliary_train-00000-of-00001.parquet b/college_biology/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/college_biology/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/college_biology/dev-00000-of-00001.parquet b/college_biology/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..03143f56f5410c4bb74bdfa954be192da41fb4af --- /dev/null +++ b/college_biology/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1552c2fbb6592c1edbf2212d094c3cba7b45d1d47175b9a4a401fc808da97c0e +size 3728 diff --git a/college_biology/test-00000-of-00001.parquet b/college_biology/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..0098e9c7e8087021b95257e5070b8d7b3aa59b06 --- /dev/null +++ b/college_biology/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7b8e52220d1bf042201cb71e25a47d9cdf818ee4541f67205872f58e4b7d3201 +size 31221 diff --git a/college_biology/validation-00000-of-00001.parquet b/college_biology/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..9524463910bdca9341579421265beb43eb3b7524 --- /dev/null +++ b/college_biology/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bb05d5f5c78a0e7d7c4bd6d0044e39888df40b576b5a9e615bc37904b29bae91 +size 6248 diff --git a/college_chemistry/auxiliary_train-00000-of-00001.parquet b/college_chemistry/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/college_chemistry/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/college_chemistry/dev-00000-of-00001.parquet b/college_chemistry/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..5e4b48e11b91b8ebf7ff135eb670706a0799ea17 --- /dev/null +++ b/college_chemistry/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a8c997d19239dfe84cc4d74acd82d550f1cb33ea3915f6afa997b2f71f7c5094 +size 3484 diff --git a/college_chemistry/test-00000-of-00001.parquet b/college_chemistry/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..a488d31d2437323af9ebaeb1f7f67df6bb1ca757 --- /dev/null +++ b/college_chemistry/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e1d9477056f186d4942cb96b56b3bf084ab9f90a4204bdf977d78370d10ed7c8 +size 17294 diff --git a/college_chemistry/validation-00000-of-00001.parquet b/college_chemistry/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..192e86f7d76542d1562c5148d33251700cc59502 --- /dev/null +++ b/college_chemistry/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9c9d41e406c94bce386f9984c8ac1d6d96b8b384dc2ff602bb29082ad0d1209b +size 4225 diff --git a/college_computer_science/auxiliary_train-00000-of-00001.parquet b/college_computer_science/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/college_computer_science/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/college_computer_science/dev-00000-of-00001.parquet b/college_computer_science/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..7f5f85f2050ef4ea54369e02001b5dfe75cf351d --- /dev/null +++ b/college_computer_science/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4882d0d971ddb2842812ae25a260097c4a4cb719720e5e8dd91e33d84d863185 +size 5614 diff --git a/college_computer_science/test-00000-of-00001.parquet b/college_computer_science/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..5dfd353b6ae6662343ee7463c9d545aea81cc9ec --- /dev/null +++ b/college_computer_science/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b8685a45ae7f9224840bd39704512f7389e671ab5a40e5fbabc93cbba27013b5 +size 26811 diff --git a/college_computer_science/validation-00000-of-00001.parquet b/college_computer_science/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..675566ae0871ec8ce6ee9e900899aab54535a2b3 --- /dev/null +++ b/college_computer_science/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b7317ea44dfcfc8b66dddfe53618421add2e507d1a6736f85b12449828de0c63 +size 5586 diff --git a/college_mathematics/auxiliary_train-00000-of-00001.parquet b/college_mathematics/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/college_mathematics/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/college_mathematics/dev-00000-of-00001.parquet b/college_mathematics/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..c42fc18d22ee7241731ecf991c509d6245d932c9 --- /dev/null +++ b/college_mathematics/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:95b2d5d13acb3b0a89b85d6b00037f8863c6734bdf56b67bfd4a9ef0567b9438 +size 4251 diff --git a/college_mathematics/test-00000-of-00001.parquet b/college_mathematics/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..c0d3113c128a829d01197894e7dc5dd321d7e7de --- /dev/null +++ b/college_mathematics/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:33d15bfa8968d680f3fe8278209aa0e1efad9921a392d5d9528d2d3a4afe831f +size 15989 diff --git a/college_mathematics/validation-00000-of-00001.parquet b/college_mathematics/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..cb0806d2889902fc3524781212e72be3c201186f --- /dev/null +++ b/college_mathematics/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:02c938bd6d1e498788ab7c68aa2d039675bb91e3c5649d3be3c970cd4a298278 +size 4402 diff --git a/college_medicine/auxiliary_train-00000-of-00001.parquet b/college_medicine/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/college_medicine/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/college_medicine/dev-00000-of-00001.parquet b/college_medicine/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..550c907e7c82ed370cbee717eec996aad4b3caea --- /dev/null +++ b/college_medicine/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f66833a9fd0998de367c581b4f93c1ea2ab17242f7864a929a78260f39d31aef +size 4114 diff --git a/college_medicine/test-00000-of-00001.parquet b/college_medicine/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..b40d3d81d0e4a3c96129352e7e5c01c5e9a82872 --- /dev/null +++ b/college_medicine/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a9bbba81068109cc5caf6235b16c494f2f1e58cfdc0e88a1767514c4da7c0435 +size 41803 diff --git a/college_medicine/validation-00000-of-00001.parquet b/college_medicine/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..06c76eadc8e6aa99f6d0b507bc3b17be65f8f866 --- /dev/null +++ b/college_medicine/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4c086dcf941908b34711d65032646c349fa15c83681cd8ec99a04271a334a784 +size 8329 diff --git a/college_physics/auxiliary_train-00000-of-00001.parquet b/college_physics/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/college_physics/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/college_physics/dev-00000-of-00001.parquet b/college_physics/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..3bd98618be5e535d3b6a240d8218901ddf975c03 --- /dev/null +++ b/college_physics/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:694320e1d4db89309520964f9fbcb53e284b3f098cbd6acdebd4ce1022d03b55 +size 3760 diff --git a/college_physics/test-00000-of-00001.parquet b/college_physics/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..5b519109127afd3baae14bd2a9583619a860dcb1 --- /dev/null +++ b/college_physics/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:11d068a92d8690700e8b5db5efa4c3c1215e3b27870ec83ff2675e1a58a30f0a +size 17653 diff --git a/college_physics/validation-00000-of-00001.parquet b/college_physics/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..ceda91bf3175e26de9d63c1a0efda9e8c8b6cc0d --- /dev/null +++ b/college_physics/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3accba591a12e012a5d2aad7346559d48039e65c591c14a815e3cfc6b85789cb +size 5533 diff --git a/computer_security/auxiliary_train-00000-of-00001.parquet b/computer_security/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/computer_security/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/computer_security/dev-00000-of-00001.parquet b/computer_security/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..72f9637d4bfe6fdb09dc3d1fd11ff0e0d02a4df2 --- /dev/null +++ b/computer_security/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:55d2262c094ab2299e662de8f51efb35ca4764a007c2bd82c62eb070c224509f +size 3659 diff --git a/computer_security/test-00000-of-00001.parquet b/computer_security/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..3e61709f8d251af16a305ff309c1e1eae96288f6 --- /dev/null +++ b/computer_security/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3f55080db8cca9b9542bcf695b032495d0623e5f796a9349756bdd58c3962ca7 +size 18577 diff --git a/computer_security/validation-00000-of-00001.parquet b/computer_security/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..867cc8624071a7eafdfce5175b331a6af3e519e0 --- /dev/null +++ b/computer_security/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:381b548c745e1e1b51a48b74cd4c6b94a545bba92efe74d31d5aad189e2fd9f9 +size 5964 diff --git a/conceptual_physics/auxiliary_train-00000-of-00001.parquet b/conceptual_physics/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/conceptual_physics/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/conceptual_physics/dev-00000-of-00001.parquet b/conceptual_physics/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..070f73dbc7a741d60d927e1b25ae70392a71e2b5 --- /dev/null +++ b/conceptual_physics/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:36e0fc357e65cb0945fdb9d77e9619d7f17738e99df51e8daf37758ff7c5cff5 +size 3358 diff --git a/conceptual_physics/test-00000-of-00001.parquet b/conceptual_physics/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..4f8f9f6e8eeac14c47f4b64b1b17a7c8669fdc41 --- /dev/null +++ b/conceptual_physics/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0d51a8ef47faab439431166fe5a0f4fc92737b586a6b3478f6a0d356ed98610f +size 24474 diff --git a/conceptual_physics/validation-00000-of-00001.parquet b/conceptual_physics/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..09099968a1ed31a857bf7cf90c83cbb6de4e1279 --- /dev/null +++ b/conceptual_physics/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:19a7f8cc46582c827b2f915679e957cb3a0087ed6229b65b0a0693827bd1472a +size 5444 diff --git a/data.tar b/data.tar deleted file mode 100644 index fc39bb6352675b6baae57e18112b031c942c2d5f..0000000000000000000000000000000000000000 --- a/data.tar +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b -size 166184960 diff --git a/dataset_infos.json b/dataset_infos.json deleted file mode 100644 index dee80a7eeabd98df4e4c905b612498037131581b..0000000000000000000000000000000000000000 --- a/dataset_infos.json +++ /dev/null @@ -1 +0,0 @@ -{"abstract_algebra": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "abstract_algebra", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 19316, "num_examples": 100, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 2012, "num_examples": 11, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 818, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160623403, "size_in_bytes": 326808363}, "anatomy": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "anatomy", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 33109, "num_examples": 135, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 3128, "num_examples": 14, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 955, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160638449, "size_in_bytes": 326823409}, "astronomy": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "astronomy", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 46759, "num_examples": 152, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 5015, "num_examples": 16, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 2064, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160655095, "size_in_bytes": 326840055}, "business_ethics": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "business_ethics", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 33240, "num_examples": 100, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 3026, "num_examples": 11, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 2178, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160639701, "size_in_bytes": 326824661}, "clinical_knowledge": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "clinical_knowledge", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 62742, "num_examples": 265, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 6652, "num_examples": 29, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 1198, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160671849, "size_in_bytes": 326856809}, "college_biology": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "college_biology", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 48785, "num_examples": 144, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 4807, "num_examples": 16, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 1520, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160656369, "size_in_bytes": 326841329}, "college_chemistry": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "college_chemistry", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 24696, "num_examples": 100, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 2316, "num_examples": 8, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 1319, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160629588, "size_in_bytes": 326814548}, "college_computer_science": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "college_computer_science", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 42629, "num_examples": 100, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 4651, "num_examples": 11, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 2753, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160651290, "size_in_bytes": 326836250}, "college_mathematics": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "college_mathematics", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 24699, "num_examples": 100, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 2656, "num_examples": 11, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 1481, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160630093, "size_in_bytes": 326815053}, "college_medicine": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "college_medicine", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 82385, "num_examples": 173, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 7897, "num_examples": 22, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 1658, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160693197, "size_in_bytes": 326878157}, "college_physics": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "college_physics", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 30169, "num_examples": 102, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 3478, "num_examples": 11, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 1400, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160636304, "size_in_bytes": 326821264}, "computer_security": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "computer_security", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 27112, "num_examples": 100, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 4537, "num_examples": 11, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 1089, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160633995, "size_in_bytes": 326818955}, "conceptual_physics": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "conceptual_physics", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 40697, "num_examples": 235, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 4462, "num_examples": 26, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 922, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160647338, "size_in_bytes": 326832298}, "econometrics": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "econometrics", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 46535, "num_examples": 114, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 4955, "num_examples": 12, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 1632, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160654379, "size_in_bytes": 326839339}, "electrical_engineering": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "electrical_engineering", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 25130, "num_examples": 145, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 2891, "num_examples": 16, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 960, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160630238, "size_in_bytes": 326815198}, "elementary_mathematics": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "elementary_mathematics", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 70096, "num_examples": 378, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 8976, "num_examples": 41, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 1428, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160681757, "size_in_bytes": 326866717}, "formal_logic": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "formal_logic", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 49773, "num_examples": 126, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 6240, "num_examples": 14, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 1745, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160659015, "size_in_bytes": 326843975}, "global_facts": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "global_facts", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 18391, "num_examples": 100, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 1853, "num_examples": 10, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 1217, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160622718, "size_in_bytes": 326807678}, "high_school_biology": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "high_school_biology", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 109720, "num_examples": 310, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 11010, "num_examples": 32, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 1661, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160723648, "size_in_bytes": 326908608}, "high_school_chemistry": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "high_school_chemistry", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 58452, "num_examples": 203, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 7080, "num_examples": 22, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 1208, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160667997, "size_in_bytes": 326852957}, "high_school_computer_science": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "high_school_computer_science", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 44464, "num_examples": 100, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 3331, "num_examples": 9, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 2906, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160651958, "size_in_bytes": 326836918}, "high_school_european_history": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "high_school_european_history", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 270288, "num_examples": 165, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 29620, "num_examples": 18, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 11552, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160912717, "size_in_bytes": 327097677}, "high_school_geography": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "high_school_geography", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 42022, "num_examples": 198, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 4320, "num_examples": 22, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 1391, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160648990, "size_in_bytes": 326833950}, "high_school_government_and_politics": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "high_school_government_and_politics", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 66062, "num_examples": 193, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 7051, "num_examples": 21, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 1767, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160676137, "size_in_bytes": 326861097}, "high_school_macroeconomics": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "high_school_macroeconomics", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 117675, "num_examples": 390, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 13008, "num_examples": 43, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 1316, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160733256, "size_in_bytes": 326918216}, "high_school_mathematics": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "high_school_mathematics", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 54842, "num_examples": 270, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 5753, "num_examples": 29, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 1285, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160663137, "size_in_bytes": 326848097}, "high_school_microeconomics": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "high_school_microeconomics", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 75691, "num_examples": 238, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 7541, "num_examples": 26, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 1286, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160685775, "size_in_bytes": 326870735}, "high_school_physics": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "high_school_physics", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 59526, "num_examples": 151, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 6759, "num_examples": 17, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 1477, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160669019, "size_in_bytes": 326853979}, "high_school_psychology": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "high_school_psychology", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 159395, "num_examples": 545, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 17257, "num_examples": 60, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 1893, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160779802, "size_in_bytes": 326964762}, "high_school_statistics": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "high_school_statistics", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 110690, "num_examples": 216, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 9985, "num_examples": 23, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 2516, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160724448, "size_in_bytes": 326909408}, "high_school_us_history": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "high_school_us_history", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 296722, "num_examples": 204, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 31694, "num_examples": 22, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 8852, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160938525, "size_in_bytes": 327123485}, "high_school_world_history": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "high_school_world_history", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 378605, "num_examples": 237, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 45489, "num_examples": 26, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 4870, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 161030221, "size_in_bytes": 327215181}, "human_aging": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "human_aging", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 46086, "num_examples": 223, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 4695, "num_examples": 23, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 996, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160653034, "size_in_bytes": 326837994}, "human_sexuality": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "human_sexuality", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 32098, "num_examples": 131, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 2409, "num_examples": 12, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 1065, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160636829, "size_in_bytes": 326821789}, "international_law": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "international_law", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 53519, "num_examples": 121, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 6461, "num_examples": 13, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 2406, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160663643, "size_in_bytes": 326848603}, "jurisprudence": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "jurisprudence", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 33974, "num_examples": 108, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 3717, "num_examples": 11, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 1291, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160640239, "size_in_bytes": 326825199}, "logical_fallacies": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "logical_fallacies", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 50105, "num_examples": 163, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 5091, "num_examples": 18, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 1561, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160658014, "size_in_bytes": 326842974}, "machine_learning": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "machine_learning", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 33868, "num_examples": 112, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 3220, "num_examples": 11, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 2311, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160640656, "size_in_bytes": 326825616}, "management": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "management", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 19990, "num_examples": 103, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 1808, "num_examples": 11, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 886, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160623941, "size_in_bytes": 326808901}, "marketing": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "marketing", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 63013, "num_examples": 234, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 7382, "num_examples": 25, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 1469, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160673121, "size_in_bytes": 326858081}, "medical_genetics": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "medical_genetics", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 20852, "num_examples": 100, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 2993, "num_examples": 11, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 1077, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160626179, "size_in_bytes": 326811139}, "miscellaneous": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "miscellaneous", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 147692, "num_examples": 783, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 14318, "num_examples": 86, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 687, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160763954, "size_in_bytes": 326948914}, "moral_disputes": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "moral_disputes", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 107806, "num_examples": 346, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 12408, "num_examples": 38, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 1743, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160723214, "size_in_bytes": 326908174}, "moral_scenarios": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "moral_scenarios", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 374014, "num_examples": 895, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 42326, "num_examples": 100, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 2046, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 161019643, "size_in_bytes": 327204603}, "nutrition": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "nutrition", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 92398, "num_examples": 306, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 8424, "num_examples": 33, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 2073, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160704152, "size_in_bytes": 326889112}, "philosophy": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "philosophy", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 80061, "num_examples": 311, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 9172, "num_examples": 34, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 976, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160691466, "size_in_bytes": 326876426}, "prehistory": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "prehistory", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 89582, "num_examples": 324, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 10273, "num_examples": 35, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 1866, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160702978, "size_in_bytes": 326887938}, "professional_accounting": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "professional_accounting", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 124538, "num_examples": 282, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 14360, "num_examples": 31, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 2136, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160742291, "size_in_bytes": 326927251}, "professional_law": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "professional_law", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 1891750, "num_examples": 1534, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 203507, "num_examples": 170, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 6598, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 162703112, "size_in_bytes": 328888072}, "professional_medicine": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "professional_medicine", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 217549, "num_examples": 272, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 23835, "num_examples": 31, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 3795, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160846436, "size_in_bytes": 327031396}, "professional_psychology": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "professional_psychology", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 225887, "num_examples": 612, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 29089, "num_examples": 69, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 2255, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160858488, "size_in_bytes": 327043448}, "public_relations": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "public_relations", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 28748, "num_examples": 110, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 4554, "num_examples": 12, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 1484, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160636043, "size_in_bytes": 326821003}, "security_studies": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "security_studies", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 204832, "num_examples": 245, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 22625, "num_examples": 27, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 5323, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160834037, "size_in_bytes": 327018997}, "sociology": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "sociology", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 66231, "num_examples": 201, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 7172, "num_examples": 22, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 1601, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160676261, "size_in_bytes": 326861221}, "us_foreign_policy": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "us_foreign_policy", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 28431, "num_examples": 100, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 3252, "num_examples": 11, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 1599, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160634539, "size_in_bytes": 326819499}, "virology": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "virology", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 38747, "num_examples": 166, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 5451, "num_examples": 18, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 1084, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160646539, "size_in_bytes": 326831499}, "world_religions": {"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n", "citation": "@article{hendryckstest2021,\n title={Measuring Massive Multitask Language Understanding},\n author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},\n journal={Proceedings of the International Conference on Learning Representations (ICLR)},\n year={2021}\n }\n", "homepage": "https://github.com/hendrycks/test", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "choices": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer": {"num_classes": 4, "names": ["A", "B", "C", "D"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "mmlu", "config_name": "world_religions", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"auxiliary_train": {"name": "auxiliary_train", "num_bytes": 160601257, "num_examples": 99842, "dataset_name": "mmlu"}, "test": {"name": "test", "num_bytes": 25262, "num_examples": 171, "dataset_name": "mmlu"}, "validation": {"name": "validation", "num_bytes": 2753, "num_examples": 19, "dataset_name": "mmlu"}, "dev": {"name": "dev", "num_bytes": 658, "num_examples": 5, "dataset_name": "mmlu"}}, "download_checksums": {"data.tar": {"num_bytes": 166184960, "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"}}, "download_size": 166184960, "post_processing_size": null, "dataset_size": 160629930, "size_in_bytes": 326814890}} \ No newline at end of file diff --git a/econometrics/auxiliary_train-00000-of-00001.parquet b/econometrics/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/econometrics/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/econometrics/dev-00000-of-00001.parquet b/econometrics/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..f013ae2fada0d6653c583c0cd95101101fe27f2e --- /dev/null +++ b/econometrics/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4cb8c7cb4937d1537c8350987940f66647881c4fd8a522e769e130b42689a56d +size 3897 diff --git a/econometrics/test-00000-of-00001.parquet b/econometrics/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..515b6be7ba7ff2c252736f213ecae7283517630f --- /dev/null +++ b/econometrics/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:61c74a3ac87861d3d9ac8988a290ef3ae87d3e1e457c206cd228d4f928dbded5 +size 23891 diff --git a/econometrics/validation-00000-of-00001.parquet b/econometrics/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..c8483aff694d79c03bdd8a16db0f209ca19f7dfd --- /dev/null +++ b/econometrics/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:69f193e86b02399d1dd7be0726f1329abd1ffd46c4acd767d4dd7d935cc6181c +size 6103 diff --git a/electrical_engineering/auxiliary_train-00000-of-00001.parquet b/electrical_engineering/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/electrical_engineering/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/electrical_engineering/dev-00000-of-00001.parquet b/electrical_engineering/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..d9d22e0f585eee9cf1e8ef698a19225a44ad7757 --- /dev/null +++ b/electrical_engineering/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0c9649943d35c5fb2d1b7e392ca6b7beb55a3e176c4ec83cc556ddb96047ce9d +size 3425 diff --git a/electrical_engineering/test-00000-of-00001.parquet b/electrical_engineering/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..24457efc79b9b2e318229c271672b0698487a3c3 --- /dev/null +++ b/electrical_engineering/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7e054f96d3fc5ce6c672388492d2eecc1ccc1bf67a27487b73e66801a8b14f6b +size 17103 diff --git a/electrical_engineering/validation-00000-of-00001.parquet b/electrical_engineering/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..5df8e0818ba2ff9201728b00fdad9fae0ac473d4 --- /dev/null +++ b/electrical_engineering/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ef29134b3fd101d2bbcc5e57dd40b23857e7b4aab0d8aa82cfed19c9e11cbe59 +size 4538 diff --git a/elementary_mathematics/auxiliary_train-00000-of-00001.parquet b/elementary_mathematics/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/elementary_mathematics/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/elementary_mathematics/dev-00000-of-00001.parquet b/elementary_mathematics/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..8a965d4175231c851bafd922ff380b71fea1bb9e --- /dev/null +++ b/elementary_mathematics/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:678c76c7e08f1b3574813c907036dc3136a15dddd7a8760155da3811e31078d7 +size 3831 diff --git a/elementary_mathematics/test-00000-of-00001.parquet b/elementary_mathematics/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..e25ce679ab7186129afef1f8ccdb6de8c0296fc4 --- /dev/null +++ b/elementary_mathematics/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:10e527669c820dfcfcacc7c73a597bcb4b34f542b1d14b94cc809cb6a1409d42 +size 40650 diff --git a/elementary_mathematics/validation-00000-of-00001.parquet b/elementary_mathematics/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..dd07b3f7185c2f83970f6368ef36f822326e95d7 --- /dev/null +++ b/elementary_mathematics/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8c418d15f743a60cbb9110219e1fb58a3ec89bec32522d6f71baed46d9cf83e7 +size 8536 diff --git a/formal_logic/auxiliary_train-00000-of-00001.parquet b/formal_logic/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/formal_logic/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/formal_logic/dev-00000-of-00001.parquet b/formal_logic/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..508d9c095dc57fd9588e130a6dbc56213fa3c0d8 --- /dev/null +++ b/formal_logic/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b31a689eb1bd0fbf2d4ad75706cd94528ef8cac81141914bba795b752f33fa65 +size 4017 diff --git a/formal_logic/test-00000-of-00001.parquet b/formal_logic/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..3cd954ad67178452282e34c8ac35863a976ce16b --- /dev/null +++ b/formal_logic/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d7ff87d93c73f2cab23472197a523c78500f34a3a57bbe7b2132c8bc04e7c564 +size 20726 diff --git a/formal_logic/validation-00000-of-00001.parquet b/formal_logic/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..403c66391164dd4e0b8f91cae9ca0193e7c6d791 --- /dev/null +++ b/formal_logic/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a61a06947bccf7e0cea7e8b0b13e9f5c9479f42dac3424a17dadb7dd7a8e1b30 +size 5651 diff --git a/global_facts/auxiliary_train-00000-of-00001.parquet b/global_facts/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/global_facts/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/global_facts/dev-00000-of-00001.parquet b/global_facts/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..ae8e4994af89276aa7b687a7a008d34e656e3abe --- /dev/null +++ b/global_facts/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cdee519418b9827cb9b30896e1b4d255f720d9a0a68fd5ac2d29ba8793094878 +size 3071 diff --git a/global_facts/test-00000-of-00001.parquet b/global_facts/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..1098627e78f4dd41a920b19f5a30dc162d39f258 --- /dev/null +++ b/global_facts/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3498c1b25d1186eef7874e9eed9cf1855c8ab21218ce12a18841dcbe229a92f0 +size 10930 diff --git a/global_facts/validation-00000-of-00001.parquet b/global_facts/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..299af95e5fa53891ff67fc9012cdd37e1245d24a --- /dev/null +++ b/global_facts/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f331cc680b84e36e2ba3c105218be43e245406bdcf4341139bd5b8f91655e5e8 +size 3678 diff --git a/high_school_biology/auxiliary_train-00000-of-00001.parquet b/high_school_biology/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/high_school_biology/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/high_school_biology/dev-00000-of-00001.parquet b/high_school_biology/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..1edc2817cdd4e5052d61f0367c2e9be4298e5679 --- /dev/null +++ b/high_school_biology/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:44064ef0b26b2a7e26918a08f5609bb0bd5dec92150662b63f4b296fc3fb699f +size 4138 diff --git a/high_school_biology/test-00000-of-00001.parquet b/high_school_biology/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..4ab06d0bab994606d9f7595d1547569b84e19fa0 --- /dev/null +++ b/high_school_biology/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7fdeafef04c133837a83b80ce266e11427e3d7148becb1f3ef790affb6493ca3 +size 62004 diff --git a/high_school_biology/validation-00000-of-00001.parquet b/high_school_biology/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..041b7e456c0ad2df59aec3f308a5f8905a6560eb --- /dev/null +++ b/high_school_biology/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3eaaca39dcb2506a3bb2bab1dfe0883f9e6a7b5275dbf3e8a50ce547bb2155b3 +size 9849 diff --git a/high_school_chemistry/auxiliary_train-00000-of-00001.parquet b/high_school_chemistry/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/high_school_chemistry/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/high_school_chemistry/dev-00000-of-00001.parquet b/high_school_chemistry/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..011bcf94e1da3d48139f535cc188ef399c1307f3 --- /dev/null +++ b/high_school_chemistry/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4b87ca6d10373044575d5ce110db448d4f7cacf9c15b4d9262d8efba391eade5 +size 3525 diff --git a/high_school_chemistry/test-00000-of-00001.parquet b/high_school_chemistry/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..1283c078291ef84bd64a1112b722c4736af822fc --- /dev/null +++ b/high_school_chemistry/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fe69bd95f60bb1807fb56d3545356335fca94d8afd328c7889f7d31df0b1f547 +size 32701 diff --git a/high_school_chemistry/validation-00000-of-00001.parquet b/high_school_chemistry/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..d73032cb4d6f8305d0df769c477123b68fd5ac01 --- /dev/null +++ b/high_school_chemistry/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:02ec824dfa49416d7f43a54ea688e49eda36b78f91b999f5c51d1faf3545dd03 +size 7588 diff --git a/high_school_computer_science/auxiliary_train-00000-of-00001.parquet b/high_school_computer_science/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/high_school_computer_science/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/high_school_computer_science/dev-00000-of-00001.parquet b/high_school_computer_science/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..d264b9cf21078e7c58ebc2f4abb7577b6ff83f71 --- /dev/null +++ b/high_school_computer_science/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:30a74a6e8039665801407c3dc800f4c67e5f8968a8cb45ead1043c228a2e2244 +size 5438 diff --git a/high_school_computer_science/test-00000-of-00001.parquet b/high_school_computer_science/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..863022e5d8bf26f8e8a9d2aea46b9f302ecbb93e --- /dev/null +++ b/high_school_computer_science/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9ece9b1a9af50face567f315d461694ae2a65de8b56d8a5e79d828148add8530 +size 26584 diff --git a/high_school_computer_science/validation-00000-of-00001.parquet b/high_school_computer_science/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..275eeb16414982bd15b895797ab9c9ee76a77536 --- /dev/null +++ b/high_school_computer_science/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:778320a9d3cb87712a5ccc2f70c4558a2d9c64e157df54ae91ab94ef9df7d1cf +size 4692 diff --git a/high_school_european_history/auxiliary_train-00000-of-00001.parquet b/high_school_european_history/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/high_school_european_history/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/high_school_european_history/dev-00000-of-00001.parquet b/high_school_european_history/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..8fef70983b8070ddcb8ebcf50fe3a8728ce11400 --- /dev/null +++ b/high_school_european_history/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c1a3443ce2305cc8b00e5bd49c9b59c48a1ad27a0a3f86899d00296988744b4a +size 17800 diff --git a/high_school_european_history/test-00000-of-00001.parquet b/high_school_european_history/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..1da7994cdf0881cd12bd34d20765c60767a8ae41 --- /dev/null +++ b/high_school_european_history/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dfcd59174317aa9e7e22efc58db624fe8f5eab64e3b205face0427f11835b13e +size 139850 diff --git a/high_school_european_history/validation-00000-of-00001.parquet b/high_school_european_history/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..84805cf4c7755108f462430e4bbf14da2ea01253 --- /dev/null +++ b/high_school_european_history/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ab18763d2d6d07a189d4b3fbf9b296f6250672648307e59525f8108e63c24799 +size 27889 diff --git a/high_school_geography/auxiliary_train-00000-of-00001.parquet b/high_school_geography/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/high_school_geography/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/high_school_geography/dev-00000-of-00001.parquet b/high_school_geography/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..94d7fcf759dcaab70c114521c335c71875bee035 --- /dev/null +++ b/high_school_geography/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:363f83dbe11ebef618049b54ccca44c35caec5bbda25f5524cd5b6981a5d6350 +size 3406 diff --git a/high_school_geography/test-00000-of-00001.parquet b/high_school_geography/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..88e255d130946409ec3853b56d4079a113f69e7b --- /dev/null +++ b/high_school_geography/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c0c819342bf1f221b63e3495d1c4cfe876407ab78b1497fd5c179c1c58ac6362 +size 27634 diff --git a/high_school_geography/validation-00000-of-00001.parquet b/high_school_geography/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..db50bf7e3d06e9cc3465fb980affca2a0566717d --- /dev/null +++ b/high_school_geography/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d58db056ca84df83dea06c43a574456f9a4f8447d0344b439d2615d11d0375d1 +size 5653 diff --git a/high_school_government_and_politics/auxiliary_train-00000-of-00001.parquet b/high_school_government_and_politics/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/high_school_government_and_politics/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/high_school_government_and_politics/dev-00000-of-00001.parquet b/high_school_government_and_politics/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..0933ed654df5c62edc0f4305488908c4e5177f81 --- /dev/null +++ b/high_school_government_and_politics/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c39de5ee3db387caaa43814f2b006bc783a3217c32634823b4a5972ccf3ab397 +size 3919 diff --git a/high_school_government_and_politics/test-00000-of-00001.parquet b/high_school_government_and_politics/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..474677de89dd051c36e036e26f27f7bae9cdeb15 --- /dev/null +++ b/high_school_government_and_politics/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:00dd810df6f248588e85544fd129c6f5badc71271d22019282046ec7a5d1c282 +size 39522 diff --git a/high_school_government_and_politics/validation-00000-of-00001.parquet b/high_school_government_and_politics/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..731c396ed2630863df6254f585ef915e9f8870c8 --- /dev/null +++ b/high_school_government_and_politics/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b5742db38103ebb5922b203611eb4ddae583d17dbe30ec62a5a952a1f0c87319 +size 7565 diff --git a/high_school_macroeconomics/auxiliary_train-00000-of-00001.parquet b/high_school_macroeconomics/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/high_school_macroeconomics/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/high_school_macroeconomics/dev-00000-of-00001.parquet b/high_school_macroeconomics/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..14cbf1c975cca1338cb4c3b84376455b9869e6a3 --- /dev/null +++ b/high_school_macroeconomics/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e3fbfb70717d6dabe1ee1e7a29397f527de51ba1a270102b5a101e5b62ca45da +size 3482 diff --git a/high_school_macroeconomics/test-00000-of-00001.parquet b/high_school_macroeconomics/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..39d08c864af193993ba5355331bec94097db7f4b --- /dev/null +++ b/high_school_macroeconomics/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:586fb112139a54ed3cb2561df195219a8ed28917d703742d3659cbeff2e0dad6 +size 54279 diff --git a/high_school_macroeconomics/validation-00000-of-00001.parquet b/high_school_macroeconomics/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..f26ded9dad4d23983b68164af9692ca462cfb79d --- /dev/null +++ b/high_school_macroeconomics/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:61039fcfd2a20da8054da59d1f547cf7d7144f48ee51d7093d1f687b6f620d48 +size 9362 diff --git a/high_school_mathematics/auxiliary_train-00000-of-00001.parquet b/high_school_mathematics/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/high_school_mathematics/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/high_school_mathematics/dev-00000-of-00001.parquet b/high_school_mathematics/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..9bb8edb187a7c08b9e5fc499bbf0f43ea6879c9d --- /dev/null +++ b/high_school_mathematics/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a07979520b6387ed5e6002a0dc593694b017fbfb91ba89760f7cd821ec0d775a +size 3777 diff --git a/high_school_mathematics/test-00000-of-00001.parquet b/high_school_mathematics/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..805a984cf44d40058efe6021261061c5c70db3ce --- /dev/null +++ b/high_school_mathematics/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aaeee814b52230d9e45131708af0bb105e4aca608f9e7dca7830d1a2e19de132 +size 32915 diff --git a/high_school_mathematics/validation-00000-of-00001.parquet b/high_school_mathematics/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..8cb2799338a889f284213280037f1a2845d0fa6f --- /dev/null +++ b/high_school_mathematics/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fcbf3b3e99b61c53f5e6a7dcafd2c7508554522aa2b89fb2f923ab0b31b9b8ff +size 6372 diff --git a/high_school_microeconomics/auxiliary_train-00000-of-00001.parquet b/high_school_microeconomics/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/high_school_microeconomics/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/high_school_microeconomics/dev-00000-of-00001.parquet b/high_school_microeconomics/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..0338587e25a76e6858f595eb4eca63a8fe7f942f --- /dev/null +++ b/high_school_microeconomics/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e7bb78f8707f58cd05ebdcb0711be913105971b1406131974ff9ba912c2c3d49 +size 3321 diff --git a/high_school_microeconomics/test-00000-of-00001.parquet b/high_school_microeconomics/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..ffc79c0fb3a6fd567a425688a3f8559acbebcf1a --- /dev/null +++ b/high_school_microeconomics/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b04ab5c1b483350c454d25aadeccc99ec3f6f26e31bb17d3560fb15c9a4afd4d +size 38144 diff --git a/high_school_microeconomics/validation-00000-of-00001.parquet b/high_school_microeconomics/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..966d325d200f849b21a8de3fa56089cc317df014 --- /dev/null +++ b/high_school_microeconomics/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:27872cc6d251dfb5b2a644520205ab4248b1ed1d5fc6a6f2d0e37b2519a76080 +size 6636 diff --git a/high_school_physics/auxiliary_train-00000-of-00001.parquet b/high_school_physics/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/high_school_physics/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/high_school_physics/dev-00000-of-00001.parquet b/high_school_physics/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..3248f870a2f74b377a982ecd4af2e850a1b9739b --- /dev/null +++ b/high_school_physics/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cf44471d66aa7d40a24d942073565b47a3bc471eed2036d6c79f782c90acb83a +size 3845 diff --git a/high_school_physics/test-00000-of-00001.parquet b/high_school_physics/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..5cfa5e141930ba05e1be520f592b6b300fe6ce4c --- /dev/null +++ b/high_school_physics/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:de1a000fa139d24c25d7d75633341b859a3f38f076e84d45b8d393b1aa030768 +size 32248 diff --git a/high_school_physics/validation-00000-of-00001.parquet b/high_school_physics/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..2c06af61cbee47321a6f895fb9fa378830351411 --- /dev/null +++ b/high_school_physics/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4fe852a82d6abb458d31a5ce64c4d2bf81fb008cc0891fa78d51868901ba2445 +size 7175 diff --git a/high_school_psychology/auxiliary_train-00000-of-00001.parquet b/high_school_psychology/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/high_school_psychology/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/high_school_psychology/dev-00000-of-00001.parquet b/high_school_psychology/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..7db73e5bd1d1d0a994aa42a54e2427a16cf77895 --- /dev/null +++ b/high_school_psychology/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9c15b0cb986cd6d4905b7d2abdda732feaec8f7fc085e2303aff269b45a9b35b +size 4416 diff --git a/high_school_psychology/test-00000-of-00001.parquet b/high_school_psychology/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..f51430add30f9e23b5f081309d11de390dad9a6f --- /dev/null +++ b/high_school_psychology/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8813f69670bfd397552c168a778cd7fc43ea346b96b7c19dce1459c251804630 +size 92004 diff --git a/high_school_psychology/validation-00000-of-00001.parquet b/high_school_psychology/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..ca4b7848c8972c959be70e7b41195e165fa4a770 --- /dev/null +++ b/high_school_psychology/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1d4608772647c6c7d784a7a90b6f3674aaef234bb02ac7d02e49f7c134cbeef5 +size 14419 diff --git a/high_school_statistics/auxiliary_train-00000-of-00001.parquet b/high_school_statistics/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/high_school_statistics/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/high_school_statistics/dev-00000-of-00001.parquet b/high_school_statistics/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..c1e5681cc4f5bfeb7905db9457166cff1000b588 --- /dev/null +++ b/high_school_statistics/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4de1b914ab66c1564a27725311e23865c2eb8c52decacd6a0a0e1cdb9db3236b +size 5044 diff --git a/high_school_statistics/test-00000-of-00001.parquet b/high_school_statistics/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..7562689eeabcedb7dc206fcd0ce1c0e1b6a69a74 --- /dev/null +++ b/high_school_statistics/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7efaa3c59848bce966288ac3886d1d87d4f46492d8c345c31d39cbe97d2d1363 +size 57348 diff --git a/high_school_statistics/validation-00000-of-00001.parquet b/high_school_statistics/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..4885605ac36997d3ed5e242bb096258b1e5bd1b9 --- /dev/null +++ b/high_school_statistics/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:989f235b7a8513f40da0d4e7b67b2c867ebf1d1e3867c10cdbb3f69e40a63f51 +size 9763 diff --git a/high_school_us_history/auxiliary_train-00000-of-00001.parquet b/high_school_us_history/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/high_school_us_history/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/high_school_us_history/dev-00000-of-00001.parquet b/high_school_us_history/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..fcdd1605385331c5981abbe4f35abdbfc2f465cc --- /dev/null +++ b/high_school_us_history/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7cfae9c7ebc45ed9a2dc994664ebbaa26fa385623e7b9e2ae145360665850523 +size 14424 diff --git a/high_school_us_history/test-00000-of-00001.parquet b/high_school_us_history/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..b781c7977f18514ec4be683eaf0a80e3ce3c14c2 --- /dev/null +++ b/high_school_us_history/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5ba730d00893798998822359eb19f1bc8873ffe139c07a78854ebdf69f3db9bd +size 152232 diff --git a/high_school_us_history/validation-00000-of-00001.parquet b/high_school_us_history/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..2dfda10a3175de25bb51e2204578c4c00e74577b --- /dev/null +++ b/high_school_us_history/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:339708dc2eaa961fb63bac71f715d6512106ef2e1cb6655d41cbbfb23c43233d +size 24779 diff --git a/high_school_world_history/auxiliary_train-00000-of-00001.parquet b/high_school_world_history/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/high_school_world_history/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/high_school_world_history/dev-00000-of-00001.parquet b/high_school_world_history/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..1637055bed0e0b5ec0a529be8762dbfe79dfedf2 --- /dev/null +++ b/high_school_world_history/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1e315dc536ad6b2d445c3ed96bc2915d6c1a9c283aa86fd37cdfcc00464803af +size 8481 diff --git a/high_school_world_history/test-00000-of-00001.parquet b/high_school_world_history/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..d964ddc257a6f6a4a44a76a00bc9ee36004e9902 --- /dev/null +++ b/high_school_world_history/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9688b239ea456500eca29784bd3c059b6105c9edb79c118b4ad61b1c6825ba08 +size 198753 diff --git a/high_school_world_history/validation-00000-of-00001.parquet b/high_school_world_history/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..52ff22f42677c40a3c8cc4aa562ae75d15aa65e8 --- /dev/null +++ b/high_school_world_history/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f909ed4ebca35952fabd89826e0b5a3d1ce9057733fccc077ae0377093acada0 +size 34174 diff --git a/human_aging/auxiliary_train-00000-of-00001.parquet b/human_aging/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/human_aging/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/human_aging/dev-00000-of-00001.parquet b/human_aging/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..0f42f08e9236f5db3ae4e5c8f56dcb04e6ab5f91 --- /dev/null +++ b/human_aging/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2c4756e50b0cdf86ad46b169723cffb1f6746af74ceceea6b47314d1f8a90f82 +size 3150 diff --git a/human_aging/test-00000-of-00001.parquet b/human_aging/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..f9bfd5f2ff7ed1b105536357b4620f62ec80732d --- /dev/null +++ b/human_aging/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:286b4cfe0fd65a7b82beee67c41a89b4ff8ad66d8c8ff51571a0be4d9ab7b159 +size 30661 diff --git a/human_aging/validation-00000-of-00001.parquet b/human_aging/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..f4c9fb7d377ca79cdd8df5f489e9de06830833f6 --- /dev/null +++ b/human_aging/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:eb15b3fd23c069ba722f9dd3316844ca5e1dbc9f6911ee53b2a716b19417f626 +size 5791 diff --git a/human_sexuality/auxiliary_train-00000-of-00001.parquet b/human_sexuality/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/human_sexuality/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/human_sexuality/dev-00000-of-00001.parquet b/human_sexuality/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..558eef1409637d9b08b43c814730e2502e5bed80 --- /dev/null +++ b/human_sexuality/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:96927ada0fc3fef5191bfd885dc9cdc6824b97de0e8a40fe3eaf99c4d3bd0b0e +size 3490 diff --git a/human_sexuality/test-00000-of-00001.parquet b/human_sexuality/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..4088047fdb350d74003f6cfb0bb32b6b089e206c --- /dev/null +++ b/human_sexuality/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b61021d0c7cb57c011fd26757e330535d8b5876f3fe000e00d99cd45820aa4a0 +size 22511 diff --git a/human_sexuality/validation-00000-of-00001.parquet b/human_sexuality/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..979bd7ef55ba03363bf431ac41c91bdf26388da4 --- /dev/null +++ b/human_sexuality/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b2d16e278a4dc367d1c993f5fd9961e54a59218a4cc2e1f8454e4d7e62163bbf +size 4601 diff --git a/international_law/auxiliary_train-00000-of-00001.parquet b/international_law/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/international_law/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/international_law/dev-00000-of-00001.parquet b/international_law/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..e6fd6350c3d8d08f4fc3ae0a86a4a6e6ad32a335 --- /dev/null +++ b/international_law/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:21f859af9a1dc0ec70e94724bf5e39d9f981125d9f6ef5820d01382fa1b69884 +size 4266 diff --git a/international_law/test-00000-of-00001.parquet b/international_law/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..6f2978fc402b7ecbd0da669801ed5e0821cfc04c --- /dev/null +++ b/international_law/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b92bf052787cc532199ee9e8175ee1789ae182946dbf608a87208a76e6840b8c +size 28870 diff --git a/international_law/validation-00000-of-00001.parquet b/international_law/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..9a794398b4e799e475014e4341e1a15d63326a5b --- /dev/null +++ b/international_law/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d0907bcdec7004b464c66501f2fb78f38b0239002f9960e96bf1e2ec840ce1d5 +size 6447 diff --git a/jurisprudence/auxiliary_train-00000-of-00001.parquet b/jurisprudence/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/jurisprudence/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/jurisprudence/dev-00000-of-00001.parquet b/jurisprudence/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..2400804dbf6cc30608ed056b6f97e65ee7d10f52 --- /dev/null +++ b/jurisprudence/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9a51dc98513e10c1e98bde78351eb32265742f2f7085dc6674bc36bb60491310 +size 3504 diff --git a/jurisprudence/test-00000-of-00001.parquet b/jurisprudence/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..670ab6f0d92722fd7e3dbd8248487a263a70272a --- /dev/null +++ b/jurisprudence/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:edf4e12ef25a4ef0caa43e254cee665fa9341ec8e8fd4265c5a295cb2dd9ca2d +size 22744 diff --git a/jurisprudence/validation-00000-of-00001.parquet b/jurisprudence/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..ae993ccc62ac4deaecd7e9c640d76dbc5bfb68f1 --- /dev/null +++ b/jurisprudence/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bc2d42f4de60752fb2a92bd985bb4259438c594f947ee6fda5afdc14294fa2f0 +size 5537 diff --git a/logical_fallacies/auxiliary_train-00000-of-00001.parquet b/logical_fallacies/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/logical_fallacies/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/logical_fallacies/dev-00000-of-00001.parquet b/logical_fallacies/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..1c067a331355711eea26a866e9d032eef71b30f1 --- /dev/null +++ b/logical_fallacies/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d99f6e2fc9e9efc9b5c53205a8d8bfb19621c4e68e1831374b875c491d3e33eb +size 3620 diff --git a/logical_fallacies/test-00000-of-00001.parquet b/logical_fallacies/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..5ada033827ec4288ec38ea9c68b4f7a4c0486015 --- /dev/null +++ b/logical_fallacies/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:51071098e4867f370919e59e0cf1ec16b6befebc1c8e1af1532fae8c9514ec9c +size 22548 diff --git a/logical_fallacies/validation-00000-of-00001.parquet b/logical_fallacies/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..ebeacfe2a9752cea59707f71b942907395ec440a --- /dev/null +++ b/logical_fallacies/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:eeb9a15a1d70f4eb3ed8ad970158de5aa0a1ce365fd448114159216239a07d73 +size 5884 diff --git a/machine_learning/auxiliary_train-00000-of-00001.parquet b/machine_learning/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/machine_learning/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/machine_learning/dev-00000-of-00001.parquet b/machine_learning/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..adfc6806232547c5890d700eb2745bba3ffc4467 --- /dev/null +++ b/machine_learning/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8826481d03264a8eb4fcfa4f72ee32521d2b9c49660f59eaffcbaf072037defb +size 4453 diff --git a/machine_learning/test-00000-of-00001.parquet b/machine_learning/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..8df8f293175ca0e3083cb0b3e7c558228dbcd212 --- /dev/null +++ b/machine_learning/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1aa57fed5c9eadac5c23f9770c5ca3144610da64fd715e6efa38535ac9aa9ef3 +size 19061 diff --git a/machine_learning/validation-00000-of-00001.parquet b/machine_learning/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..9f61152f3c90a25c3e0937d69f62fcf474e54982 --- /dev/null +++ b/machine_learning/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6061ec24ead2e3f9d9347c51a2cd7537c1427a2d6cb0eeee1f219db87ff7ca1a +size 5303 diff --git a/management/auxiliary_train-00000-of-00001.parquet b/management/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/management/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/management/dev-00000-of-00001.parquet b/management/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..4a51c2ea7fc9cf8a3d2d2d7d2df380f987052403 --- /dev/null +++ b/management/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:25a5ca83a99d49b82f603ffa89c34c4512630f7bb96e8cc821b495f025d84d0f +size 3106 diff --git a/management/test-00000-of-00001.parquet b/management/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..bd3045a1b11c30edec4792ac75d0207dc1119c79 --- /dev/null +++ b/management/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3d3297ec431709220a2d12c04d754e6b4c2090e859429a704ef932df23ce37c3 +size 14220 diff --git a/management/validation-00000-of-00001.parquet b/management/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..256fe8cb9b3aef8241c5f3d1424dca02fb991375 --- /dev/null +++ b/management/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e9b0c3ed2c0282eee103b7554ec96c22cbc28411503f2169514afe3b67f09102 +size 3927 diff --git a/marketing/auxiliary_train-00000-of-00001.parquet b/marketing/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/marketing/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/marketing/dev-00000-of-00001.parquet b/marketing/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..72699b9dbb558f8f75546737ba61b6ef003b99c2 --- /dev/null +++ b/marketing/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e27aa228b90ca6c1a608babba1ac24628d53f10ecc9a1ff8b6a1f0131d51b260 +size 3738 diff --git a/marketing/test-00000-of-00001.parquet b/marketing/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..bbc272679652337cb27a71804c05bb0d777c0cdb --- /dev/null +++ b/marketing/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8f51d8afa4082af375186991f5bbc0bb6caf799d793240209582b5efd4e1e401 +size 36632 diff --git a/marketing/validation-00000-of-00001.parquet b/marketing/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..188fbbc990c14234376a20bf2f7264198034d579 --- /dev/null +++ b/marketing/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2a9c7448f5cb53594a23fecc457d9d38ce4f13330cd02e1beb61cde64085230a +size 7580 diff --git a/medical_genetics/auxiliary_train-00000-of-00001.parquet b/medical_genetics/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/medical_genetics/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/medical_genetics/dev-00000-of-00001.parquet b/medical_genetics/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..72ea60581848a24739929a84a22b5f0b00c245d3 --- /dev/null +++ b/medical_genetics/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:002bc4e68639e77a5b1175e6508544ffb0fd6af71ba63107d1c7d7406d6c55d4 +size 3272 diff --git a/medical_genetics/test-00000-of-00001.parquet b/medical_genetics/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..275a3f8536b1fc16098d7970b0225d50040433e4 --- /dev/null +++ b/medical_genetics/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1d6d0c9463d16228638ae00b43c9a36e6c87b111825f2bd64fb6ff0e95e39844 +size 15615 diff --git a/medical_genetics/validation-00000-of-00001.parquet b/medical_genetics/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..deccee04e5e41eab1959fbe8ff0351a6f58708e7 --- /dev/null +++ b/medical_genetics/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:32d543a2f6648c35f11f796a897b02bcee611ce52f5fcca71f00e272a146e284 +size 4983 diff --git a/miscellaneous/auxiliary_train-00000-of-00001.parquet b/miscellaneous/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/miscellaneous/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/miscellaneous/dev-00000-of-00001.parquet b/miscellaneous/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..54f65c60f10ebd0412435e29b0016e2a89e868df --- /dev/null +++ b/miscellaneous/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5f4ea86b6efec990a3c0a141d57b3f0ecba5ffbbfc0689e56a4b833605b6ceb6 +size 2908 diff --git a/miscellaneous/test-00000-of-00001.parquet b/miscellaneous/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..d5bb055fd126ad6af8504b82f04ec05f48c45bd7 --- /dev/null +++ b/miscellaneous/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ac01b39582480d3e345e2167ba9bd5827bfd4ff56ad8a464b21b60fbcf69d431 +size 97776 diff --git a/miscellaneous/validation-00000-of-00001.parquet b/miscellaneous/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..1155627d10d8b5e832c58d2b39d42051cdb0b316 --- /dev/null +++ b/miscellaneous/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aa92e3f7e5833bda73d8dc630c729249e316e3a98f339b952dd408bb08a44b5d +size 12622 diff --git a/mmlu.py b/mmlu.py deleted file mode 100644 index d0a80d5735ea88a34e5b14710d6e9be50af9878b..0000000000000000000000000000000000000000 --- a/mmlu.py +++ /dev/null @@ -1,173 +0,0 @@ -# coding=utf-8 -# Copyright 2020 The HuggingFace Datasets Authors and the current dataset script contributor. -# -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. - - -import csv - -import datasets - - -_CITATION = """\ -@article{hendryckstest2021, - title={Measuring Massive Multitask Language Understanding}, - author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt}, - journal={Proceedings of the International Conference on Learning Representations (ICLR)}, - year={2021} - } -""" - -_DESCRIPTION = """\ -This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more. -""" - -_HOMEPAGE = "https://github.com/hendrycks/test" - -_URL = "data.tar" - -_SUBJECTS = [ - "all", - "abstract_algebra", - "anatomy", - "astronomy", - "business_ethics", - "clinical_knowledge", - "college_biology", - "college_chemistry", - "college_computer_science", - "college_mathematics", - "college_medicine", - "college_physics", - "computer_security", - "conceptual_physics", - "econometrics", - "electrical_engineering", - "elementary_mathematics", - "formal_logic", - "global_facts", - "high_school_biology", - "high_school_chemistry", - "high_school_computer_science", - "high_school_european_history", - "high_school_geography", - "high_school_government_and_politics", - "high_school_macroeconomics", - "high_school_mathematics", - "high_school_microeconomics", - "high_school_physics", - "high_school_psychology", - "high_school_statistics", - "high_school_us_history", - "high_school_world_history", - "human_aging", - "human_sexuality", - "international_law", - "jurisprudence", - "logical_fallacies", - "machine_learning", - "management", - "marketing", - "medical_genetics", - "miscellaneous", - "moral_disputes", - "moral_scenarios", - "nutrition", - "philosophy", - "prehistory", - "professional_accounting", - "professional_law", - "professional_medicine", - "professional_psychology", - "public_relations", - "security_studies", - "sociology", - "us_foreign_policy", - "virology", - "world_religions", -] - - -class Mmlu(datasets.GeneratorBasedBuilder): - """Measuring Massive Multitask Language Understanding, consisting of 57 tasks""" - - BUILDER_CONFIGS = [ - datasets.BuilderConfig( - name=sub, version=datasets.Version("1.0.0"), description=f"MMLU Subject {sub}" - ) - for sub in _SUBJECTS - ] - - def _info(self): - features = datasets.Features( - { - "question": datasets.Value("string"), - "subject": datasets.Value("string"), - "choices": datasets.features.Sequence(datasets.Value("string")), - "answer": datasets.features.ClassLabel(num_classes=4, names=["A", "B", "C", "D"]), - } - ) - return datasets.DatasetInfo( - description=_DESCRIPTION, - features=features, - homepage=_HOMEPAGE, - citation=_CITATION, - ) - - def _split_generators(self, dl_manager): - """Returns SplitGenerators.""" - archive = dl_manager.download(_URL) - return [ - datasets.SplitGenerator( - name=datasets.Split("auxiliary_train"), - gen_kwargs={ - "iter_archive": dl_manager.iter_archive(archive), - "split": "auxiliary_train", - }, - ), - datasets.SplitGenerator( - name=datasets.Split.TEST, - gen_kwargs={"iter_archive": dl_manager.iter_archive(archive), "split": "test"}, - ), - datasets.SplitGenerator( - name=datasets.Split.VALIDATION, - gen_kwargs={ - "iter_archive": dl_manager.iter_archive(archive), - "split": "val", - }, - ), - datasets.SplitGenerator( - name=datasets.Split("dev"), - gen_kwargs={ - "iter_archive": dl_manager.iter_archive(archive), - "split": "dev", - }, - ), - ] - - def _generate_examples(self, iter_archive, split): - """Yields examples as (key, example) tuples.""" - n_yielded_files = 0 - for id_file, (path, file) in enumerate(iter_archive): - if f"/{split}/" in path: - if split == "auxiliary_train" or (self.config.name in path or self.config.name == "all"): - subset = path.split("/")[-1].rsplit("_",1)[0] if split != "auxiliary_train" else "" - n_yielded_files += 1 - lines = (line.decode("utf-8") for line in file) - reader = csv.reader(lines) - for id_line, data in enumerate(reader): - yield f"{id_file}_{id_line}", {"question": data[0], "choices": data[1:5], "answer": data[5], "subject": subset} - #else: - #print("KO", path) - #else: - #print("KO2", split, path) diff --git a/moral_disputes/auxiliary_train-00000-of-00001.parquet b/moral_disputes/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/moral_disputes/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/moral_disputes/dev-00000-of-00001.parquet b/moral_disputes/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..6e206b715b73f2fb2af7f325b03e6b54df4948b6 --- /dev/null +++ b/moral_disputes/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3cce90bcedb2a3d252b2d6a1f8306800bf1d2889471f3451e6c3f4e68b198f0a +size 3860 diff --git a/moral_disputes/test-00000-of-00001.parquet b/moral_disputes/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..78040928eef1d65568b87ab3f4920e74b306c332 --- /dev/null +++ b/moral_disputes/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:db307676f94567725f19f771cde7e5b5ebb3ef8bcee2257517e8aac1f946856e +size 60229 diff --git a/moral_disputes/validation-00000-of-00001.parquet b/moral_disputes/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..45c3c27b35a0659ab79f506c047ae6cc1d4fef02 --- /dev/null +++ b/moral_disputes/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0e1a62e141b52b7170f0d46c8d49f3467945300b049799cf56a03b3e4508591c +size 10184 diff --git a/moral_scenarios/auxiliary_train-00000-of-00001.parquet b/moral_scenarios/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/moral_scenarios/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/moral_scenarios/dev-00000-of-00001.parquet b/moral_scenarios/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..1f8c04c6b59b88b41fbf44e5c3ea95307f7620af --- /dev/null +++ b/moral_scenarios/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0add5d0bfc4a623c345c487ffd41279e2244fd47a59abb2730ed4c29179bf9db +size 4151 diff --git a/moral_scenarios/test-00000-of-00001.parquet b/moral_scenarios/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..33aff53def79ee1efbdae39f879eebf9cf07fb37 --- /dev/null +++ b/moral_scenarios/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fa864f068a047edc050b20cddd4d5f53f5b14a8634b6cc386e9924668dc77d40 +size 88761 diff --git a/moral_scenarios/validation-00000-of-00001.parquet b/moral_scenarios/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..00fa30e5c17f0b09096c2346c3cc8cd6724cba1e --- /dev/null +++ b/moral_scenarios/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8a598c97467fa0b717f24f6a00ea3c6e4e8f0012f5248b1beb21f178b1d1438c +size 13908 diff --git a/nutrition/auxiliary_train-00000-of-00001.parquet b/nutrition/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/nutrition/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/nutrition/dev-00000-of-00001.parquet b/nutrition/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..d83915e4c98efdb095f5cdf2b1e8444b546faea9 --- /dev/null +++ b/nutrition/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:039fe23b8e62b081637141722ee5a37b96f0e048ec0a928a01735075a94e0d30 +size 4238 diff --git a/nutrition/test-00000-of-00001.parquet b/nutrition/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..1bb1c28df540195e9924c5fdae95e518b140c4fb --- /dev/null +++ b/nutrition/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:19576e3fd281041cbee9f00574b230d99983c1ad96ef057d2516ff6c69e6d81e +size 54081 diff --git a/nutrition/validation-00000-of-00001.parquet b/nutrition/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..8717bca3cf7a8edf2aca9436b9c42c55a50fdf47 --- /dev/null +++ b/nutrition/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f1a9ad5587cf0b05c6f989e75d2aecd914c1d86f0b80203c591afdb62904f693 +size 8348 diff --git a/philosophy/auxiliary_train-00000-of-00001.parquet b/philosophy/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/philosophy/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/philosophy/dev-00000-of-00001.parquet b/philosophy/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..e48e99fc2687758667910487fad42ffd899ce60e --- /dev/null +++ b/philosophy/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:15f8c07a834ce731f43159db2ed97c6ea9099b822234b301a81dbfcb65c38c74 +size 3551 diff --git a/philosophy/test-00000-of-00001.parquet b/philosophy/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..0de77814b456b5b977f1564026186c77fc3c8b04 --- /dev/null +++ b/philosophy/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:478e83a4325f9bb1c49f60fc04104344f0239ba61d89f83c924e03596ff000a3 +size 47900 diff --git a/philosophy/validation-00000-of-00001.parquet b/philosophy/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..f19824ba8340c29d6beca73c031facea53efc492 --- /dev/null +++ b/philosophy/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d69cab9da7d5aa9f1ab17945d04a4aa5ce15b6f680d3bb4d99654ca26fc0cf56 +size 8542 diff --git a/prehistory/auxiliary_train-00000-of-00001.parquet b/prehistory/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/prehistory/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/prehistory/dev-00000-of-00001.parquet b/prehistory/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..ea76356eb7f5a82c23cd808878b3ef91ccff0868 --- /dev/null +++ b/prehistory/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:da4804f66dfca504a4e862f8fb3e746d612c66d4b25718b48341ec74dd6c501b +size 4050 diff --git a/prehistory/test-00000-of-00001.parquet b/prehistory/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..6e11b8fffc05257cc8cb9fe73b015a57c42134d9 --- /dev/null +++ b/prehistory/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bf85949cd66ba2f6270973c094e29dad72486f79eed0e02a48307ee24a68d32b +size 53771 diff --git a/prehistory/validation-00000-of-00001.parquet b/prehistory/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..df062fe38756756ca81d59af1464eb7434fbfb59 --- /dev/null +++ b/prehistory/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:018d4b145f8ec3000d7adb3a97efb19c5b34b6d4de9971fa2e6f6dfe42b7a10f +size 9336 diff --git a/professional_accounting/auxiliary_train-00000-of-00001.parquet b/professional_accounting/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/professional_accounting/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/professional_accounting/dev-00000-of-00001.parquet b/professional_accounting/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..c523ec56371b23dca757066e994ce6a53a4b4165 --- /dev/null +++ b/professional_accounting/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:66dfbf37623583ab3bf2666ea3d0d3d83878e091064272bc2f12b99ed73d8110 +size 4244 diff --git a/professional_accounting/test-00000-of-00001.parquet b/professional_accounting/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..5115abe3366f481d9047bd40aa818d24c71b078d --- /dev/null +++ b/professional_accounting/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c7a74895f521daa4ed7d15d5a3cec9e1e7d53b5af02f03c47b7c7ea2d1fdcf47 +size 68760 diff --git a/professional_accounting/validation-00000-of-00001.parquet b/professional_accounting/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..e58465404703f50347c14b615f308c5174159a24 --- /dev/null +++ b/professional_accounting/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:94724eb35babfb34738df813cde9b5d225793a98fa6aab491e0ee8c71191cded +size 12047 diff --git a/professional_law/auxiliary_train-00000-of-00001.parquet b/professional_law/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/professional_law/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/professional_law/dev-00000-of-00001.parquet b/professional_law/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..3f6543c490587cd28b0e425c96c128d550c4199b --- /dev/null +++ b/professional_law/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:71ac061cc32a038458bedc4b07d8fbc6fc8cba9b8c2f7d97e3ef8c78b52e4714 +size 11961 diff --git a/professional_law/test-00000-of-00001.parquet b/professional_law/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..2fafb3dfea196fc72eb71406e463163540f442a9 --- /dev/null +++ b/professional_law/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:790f497c2f9617d67f98fd3f45336e11a7aacf088d0c10e12cff4576729586cf +size 1029293 diff --git a/professional_law/validation-00000-of-00001.parquet b/professional_law/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..3d4daaaa54a83454b6600386d6d331bedfefa7fa --- /dev/null +++ b/professional_law/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:24bc0caad939439e6e4df2d0633a776bac0a22e0e52b1d7688173e1324c0cd4c +size 114093 diff --git a/professional_medicine/auxiliary_train-00000-of-00001.parquet b/professional_medicine/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/professional_medicine/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/professional_medicine/dev-00000-of-00001.parquet b/professional_medicine/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..3ea264a2413d076cb5381982ef3240f6396d63ea --- /dev/null +++ b/professional_medicine/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:940345ce705767f06808125a78dda1ff884bc16fa574ea9dc0f73540b5fc6eb0 +size 7104 diff --git a/professional_medicine/test-00000-of-00001.parquet b/professional_medicine/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..0b87f087134694dd0b527a46b185c4205aa6bcac --- /dev/null +++ b/professional_medicine/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:adefae866c651485ba6737b5aeffce1840d7fea2131ab956d87b205990f63059 +size 122639 diff --git a/professional_medicine/validation-00000-of-00001.parquet b/professional_medicine/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..ff3bca730c05296a711c5fd8de9547220fcb6016 --- /dev/null +++ b/professional_medicine/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e1ecb778a2ec60c256ec9f05a95313562c1e95c98295d674f8d414e2b1f3e0db +size 18530 diff --git a/professional_psychology/auxiliary_train-00000-of-00001.parquet b/professional_psychology/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/professional_psychology/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/professional_psychology/dev-00000-of-00001.parquet b/professional_psychology/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..e786a3b86b3f321eb46903ccdaa3de071987a9bc --- /dev/null +++ b/professional_psychology/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:47bb0094201adf7d9ec39c36f9cda6116621b412e6b9b2677eaddf52cedd21d2 +size 4111 diff --git a/professional_psychology/test-00000-of-00001.parquet b/professional_psychology/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..89c928ca51f9c235d63f9c47f459512ccdea2129 --- /dev/null +++ b/professional_psychology/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f68dc4b9ba63a0b74bed1f283197379ccc9e7e970ee54037c7339156ee336b8b +size 131936 diff --git a/professional_psychology/validation-00000-of-00001.parquet b/professional_psychology/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..8f0b835f346f2cc69adcdd5b822ea18635746081 --- /dev/null +++ b/professional_psychology/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b46fd55b1f409893529c3ea6874bea151cba644f8752e6a784667467b103d23e +size 21116 diff --git a/public_relations/auxiliary_train-00000-of-00001.parquet b/public_relations/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/public_relations/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/public_relations/dev-00000-of-00001.parquet b/public_relations/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..e79de324eb7659f35c4e7b4b880e457ea341c0b6 --- /dev/null +++ b/public_relations/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a2f44ae195c5d33c0c7319e82408d9d8e3e9efcc32f88d6cb2ddd43b7dd39f35 +size 3838 diff --git a/public_relations/test-00000-of-00001.parquet b/public_relations/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..94e8e8c82ba55b641734e3c81c94d0df6b540ad3 --- /dev/null +++ b/public_relations/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:211e3c8cac11d365b5b5c67a26ade2da192cfbacfde4bca0decd20ab426b81ba +size 19941 diff --git a/public_relations/validation-00000-of-00001.parquet b/public_relations/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..1c43842e8e7e021150ac3babef43c4293a9a64d3 --- /dev/null +++ b/public_relations/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4d7aa54f3e754345b1307e2f704d73f9a39f1e015e2d5a8e10dff500a80d95d7 +size 5841 diff --git a/security_studies/auxiliary_train-00000-of-00001.parquet b/security_studies/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/security_studies/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/security_studies/dev-00000-of-00001.parquet b/security_studies/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..330e8d07b0ca8e02457e574092a289ae312450a8 --- /dev/null +++ b/security_studies/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3873ab223394987efa99ff0891e9b0ff46bf3c38db3d4a9aaf444f367c5f15ed +size 6585 diff --git a/security_studies/test-00000-of-00001.parquet b/security_studies/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..851d86fd05dfa3e5ab8f1c002325ac1820bd5055 --- /dev/null +++ b/security_studies/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:36d1b77d713672a7bead419e80122ac2f5b8fe7f012ece3db78f1be30b82cb9c +size 113161 diff --git a/security_studies/validation-00000-of-00001.parquet b/security_studies/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..7151ad4c2bc81f56a5319c5389aff157321d6699 --- /dev/null +++ b/security_studies/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a3a97a75b7e8e9f2ac066cdb989c05798bbd2fa8c48c71d8b9f8876b6a342bac +size 17296 diff --git a/sociology/auxiliary_train-00000-of-00001.parquet b/sociology/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/sociology/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/sociology/dev-00000-of-00001.parquet b/sociology/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..e3bd72a2c9a12416ebb8cc5d3ad4a92f185ae8dd --- /dev/null +++ b/sociology/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f71f44249d91c5ddcac81878f0aaf05c624deb7550ba60e023d753d7161c30c5 +size 3710 diff --git a/sociology/test-00000-of-00001.parquet b/sociology/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..e25d2d7ed29690b744924f4269246147af68ba8f --- /dev/null +++ b/sociology/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0f575ff7f9b4a24a211cf3b79a9bc6acc632563b9f6fdd5c1af6a31aa3abe67e +size 43260 diff --git a/sociology/validation-00000-of-00001.parquet b/sociology/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..1b75dfe8fe0a0ca7f6ab3f1a28a81619332ba325 --- /dev/null +++ b/sociology/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c16d22e656cc9662412fcda3517a15c34dfdfbcb7206fc5260aeada77b2e1068 +size 7771 diff --git a/us_foreign_policy/auxiliary_train-00000-of-00001.parquet b/us_foreign_policy/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/us_foreign_policy/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/us_foreign_policy/dev-00000-of-00001.parquet b/us_foreign_policy/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..948e378406be89603c49416606c415def3715fba --- /dev/null +++ b/us_foreign_policy/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ccea1f4411a786168b19ed306be0e19e5b908048f12bf3b421716adc8788ddf1 +size 3665 diff --git a/us_foreign_policy/test-00000-of-00001.parquet b/us_foreign_policy/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..91461fce23052bc50308a460fa48568c8b9a2b24 --- /dev/null +++ b/us_foreign_policy/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:40ee46e8bd29638686eb687c5513b3bac908e461aabb83ce570e139fb4d7debf +size 18980 diff --git a/us_foreign_policy/validation-00000-of-00001.parquet b/us_foreign_policy/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..15e05ec2731841390b77f7e3977d29c8811d6ea0 --- /dev/null +++ b/us_foreign_policy/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:927a31292a7c78622e4a19c728245469e464e61ed6768536e86ed89b1e2e85f6 +size 4746 diff --git a/virology/auxiliary_train-00000-of-00001.parquet b/virology/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/virology/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/virology/dev-00000-of-00001.parquet b/virology/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..7ec1c3bf80427091e54a96df7db2d047c56b9d31 --- /dev/null +++ b/virology/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a9250819b7de9ba964dd5892d951e3896cf3d4b531c41926c466db9acc5777ce +size 3359 diff --git a/virology/test-00000-of-00001.parquet b/virology/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..62aace026bef4a9c96f744d930d00d3c6463f52d --- /dev/null +++ b/virology/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:72aa28218b3406b468188f7e2b4deed61d289787a86183c9f1d0b3d4bf89d754 +size 26603 diff --git a/virology/validation-00000-of-00001.parquet b/virology/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..74f1c8e6cc8086af064611a31676f83d8bb5b1ba --- /dev/null +++ b/virology/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:25ddc4e993dda216e1c31a1c12a2699f300858d775bcb95b156c1b6b73c9cacc +size 6359 diff --git a/world_religions/auxiliary_train-00000-of-00001.parquet b/world_religions/auxiliary_train-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..48e8eab1959b78404a4597b20a75a4516ae73a29 --- /dev/null +++ b/world_religions/auxiliary_train-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07 +size 47163955 diff --git a/world_religions/dev-00000-of-00001.parquet b/world_religions/dev-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..0385f3e4ff5c8ef92742ad1bae7cffff7e31d505 --- /dev/null +++ b/world_religions/dev-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2bdb152242964d7b390ebb5c21da5eef53ba6cde594ae108493289e03cb65d6a +size 2847 diff --git a/world_religions/test-00000-of-00001.parquet b/world_religions/test-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..fbb5a18ef5be9626800f9871ac4c3f82155568b4 --- /dev/null +++ b/world_religions/test-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:17c85ae041b79f4b44e5c544af8de82e2b03867ec82bb76aaa4fdfe178873953 +size 18423 diff --git a/world_religions/validation-00000-of-00001.parquet b/world_religions/validation-00000-of-00001.parquet new file mode 100644 index 0000000000000000000000000000000000000000..3b8e70214276d97211b290407a4d629aff501b8c --- /dev/null +++ b/world_religions/validation-00000-of-00001.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:63b5db41ddf7c72e85239d6d89c474e46a3849ca87298f4fa32dcff2c21275dd +size 4456