Datasets:
Tasks:
Question Answering
Modalities:
Text
Formats:
parquet
Sub-tasks:
multiple-choice-qa
Languages:
English
Size:
1M - 10M
ArXiv:
License:
Add 'high_school_computer_science' config data files
Browse files- README.md +18 -6
- dataset_infos.json +17 -29
- high_school_computer_science/auxiliary_train-00000-of-00001.parquet +3 -0
- high_school_computer_science/dev-00000-of-00001.parquet +3 -0
- high_school_computer_science/test-00000-of-00001.parquet +3 -0
- high_school_computer_science/validation-00000-of-00001.parquet +3 -0
README.md
CHANGED
@@ -677,6 +677,8 @@ dataset_info:
|
|
677 |
features:
|
678 |
- name: question
|
679 |
dtype: string
|
|
|
|
|
680 |
- name: choices
|
681 |
sequence: string
|
682 |
- name: answer
|
@@ -689,19 +691,19 @@ dataset_info:
|
|
689 |
'3': D
|
690 |
splits:
|
691 |
- name: auxiliary_train
|
692 |
-
num_bytes:
|
693 |
num_examples: 99842
|
694 |
- name: test
|
695 |
-
num_bytes:
|
696 |
num_examples: 100
|
697 |
- name: validation
|
698 |
-
num_bytes:
|
699 |
num_examples: 9
|
700 |
- name: dev
|
701 |
-
num_bytes:
|
702 |
num_examples: 5
|
703 |
-
download_size:
|
704 |
-
dataset_size:
|
705 |
- config_name: high_school_european_history
|
706 |
features:
|
707 |
- name: question
|
@@ -1957,6 +1959,16 @@ configs:
|
|
1957 |
path: high_school_chemistry/validation-*
|
1958 |
- split: dev
|
1959 |
path: high_school_chemistry/dev-*
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1960 |
---
|
1961 |
|
1962 |
# Dataset Card for MMLU
|
|
|
677 |
features:
|
678 |
- name: question
|
679 |
dtype: string
|
680 |
+
- name: subject
|
681 |
+
dtype: string
|
682 |
- name: choices
|
683 |
sequence: string
|
684 |
- name: answer
|
|
|
691 |
'3': D
|
692 |
splits:
|
693 |
- name: auxiliary_train
|
694 |
+
num_bytes: 161000625
|
695 |
num_examples: 99842
|
696 |
- name: test
|
697 |
+
num_bytes: 47664
|
698 |
num_examples: 100
|
699 |
- name: validation
|
700 |
+
num_bytes: 3619
|
701 |
num_examples: 9
|
702 |
- name: dev
|
703 |
+
num_bytes: 3066
|
704 |
num_examples: 5
|
705 |
+
download_size: 47200669
|
706 |
+
dataset_size: 161054974
|
707 |
- config_name: high_school_european_history
|
708 |
features:
|
709 |
- name: question
|
|
|
1959 |
path: high_school_chemistry/validation-*
|
1960 |
- split: dev
|
1961 |
path: high_school_chemistry/dev-*
|
1962 |
+
- config_name: high_school_computer_science
|
1963 |
+
data_files:
|
1964 |
+
- split: auxiliary_train
|
1965 |
+
path: high_school_computer_science/auxiliary_train-*
|
1966 |
+
- split: test
|
1967 |
+
path: high_school_computer_science/test-*
|
1968 |
+
- split: validation
|
1969 |
+
path: high_school_computer_science/validation-*
|
1970 |
+
- split: dev
|
1971 |
+
path: high_school_computer_science/dev-*
|
1972 |
---
|
1973 |
|
1974 |
# Dataset Card for MMLU
|
dataset_infos.json
CHANGED
@@ -1407,39 +1407,34 @@
|
|
1407 |
"features": {
|
1408 |
"question": {
|
1409 |
"dtype": "string",
|
1410 |
-
"
|
|
|
|
|
|
|
1411 |
"_type": "Value"
|
1412 |
},
|
1413 |
"choices": {
|
1414 |
"feature": {
|
1415 |
"dtype": "string",
|
1416 |
-
"id": null,
|
1417 |
"_type": "Value"
|
1418 |
},
|
1419 |
-
"length": -1,
|
1420 |
-
"id": null,
|
1421 |
"_type": "Sequence"
|
1422 |
},
|
1423 |
"answer": {
|
1424 |
-
"num_classes": 4,
|
1425 |
"names": [
|
1426 |
"A",
|
1427 |
"B",
|
1428 |
"C",
|
1429 |
"D"
|
1430 |
],
|
1431 |
-
"id": null,
|
1432 |
"_type": "ClassLabel"
|
1433 |
}
|
1434 |
},
|
1435 |
-
"
|
1436 |
-
"
|
1437 |
-
"task_templates": null,
|
1438 |
-
"builder_name": "mmlu",
|
1439 |
"config_name": "high_school_computer_science",
|
1440 |
"version": {
|
1441 |
"version_str": "1.0.0",
|
1442 |
-
"description": null,
|
1443 |
"major": 1,
|
1444 |
"minor": 0,
|
1445 |
"patch": 0
|
@@ -1447,39 +1442,32 @@
|
|
1447 |
"splits": {
|
1448 |
"auxiliary_train": {
|
1449 |
"name": "auxiliary_train",
|
1450 |
-
"num_bytes":
|
1451 |
"num_examples": 99842,
|
1452 |
-
"dataset_name":
|
1453 |
},
|
1454 |
"test": {
|
1455 |
"name": "test",
|
1456 |
-
"num_bytes":
|
1457 |
"num_examples": 100,
|
1458 |
-
"dataset_name":
|
1459 |
},
|
1460 |
"validation": {
|
1461 |
"name": "validation",
|
1462 |
-
"num_bytes":
|
1463 |
"num_examples": 9,
|
1464 |
-
"dataset_name":
|
1465 |
},
|
1466 |
"dev": {
|
1467 |
"name": "dev",
|
1468 |
-
"num_bytes":
|
1469 |
"num_examples": 5,
|
1470 |
-
"dataset_name":
|
1471 |
-
}
|
1472 |
-
},
|
1473 |
-
"download_checksums": {
|
1474 |
-
"data.tar": {
|
1475 |
-
"num_bytes": 166184960,
|
1476 |
-
"checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"
|
1477 |
}
|
1478 |
},
|
1479 |
-
"download_size":
|
1480 |
-
"
|
1481 |
-
"
|
1482 |
-
"size_in_bytes": 326836918
|
1483 |
},
|
1484 |
"high_school_european_history": {
|
1485 |
"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n",
|
|
|
1407 |
"features": {
|
1408 |
"question": {
|
1409 |
"dtype": "string",
|
1410 |
+
"_type": "Value"
|
1411 |
+
},
|
1412 |
+
"subject": {
|
1413 |
+
"dtype": "string",
|
1414 |
"_type": "Value"
|
1415 |
},
|
1416 |
"choices": {
|
1417 |
"feature": {
|
1418 |
"dtype": "string",
|
|
|
1419 |
"_type": "Value"
|
1420 |
},
|
|
|
|
|
1421 |
"_type": "Sequence"
|
1422 |
},
|
1423 |
"answer": {
|
|
|
1424 |
"names": [
|
1425 |
"A",
|
1426 |
"B",
|
1427 |
"C",
|
1428 |
"D"
|
1429 |
],
|
|
|
1430 |
"_type": "ClassLabel"
|
1431 |
}
|
1432 |
},
|
1433 |
+
"builder_name": "parquet",
|
1434 |
+
"dataset_name": "mmlu",
|
|
|
|
|
1435 |
"config_name": "high_school_computer_science",
|
1436 |
"version": {
|
1437 |
"version_str": "1.0.0",
|
|
|
1438 |
"major": 1,
|
1439 |
"minor": 0,
|
1440 |
"patch": 0
|
|
|
1442 |
"splits": {
|
1443 |
"auxiliary_train": {
|
1444 |
"name": "auxiliary_train",
|
1445 |
+
"num_bytes": 161000625,
|
1446 |
"num_examples": 99842,
|
1447 |
+
"dataset_name": null
|
1448 |
},
|
1449 |
"test": {
|
1450 |
"name": "test",
|
1451 |
+
"num_bytes": 47664,
|
1452 |
"num_examples": 100,
|
1453 |
+
"dataset_name": null
|
1454 |
},
|
1455 |
"validation": {
|
1456 |
"name": "validation",
|
1457 |
+
"num_bytes": 3619,
|
1458 |
"num_examples": 9,
|
1459 |
+
"dataset_name": null
|
1460 |
},
|
1461 |
"dev": {
|
1462 |
"name": "dev",
|
1463 |
+
"num_bytes": 3066,
|
1464 |
"num_examples": 5,
|
1465 |
+
"dataset_name": null
|
|
|
|
|
|
|
|
|
|
|
|
|
1466 |
}
|
1467 |
},
|
1468 |
+
"download_size": 47200669,
|
1469 |
+
"dataset_size": 161054974,
|
1470 |
+
"size_in_bytes": 208255643
|
|
|
1471 |
},
|
1472 |
"high_school_european_history": {
|
1473 |
"description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n",
|
high_school_computer_science/auxiliary_train-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07
|
3 |
+
size 47163955
|
high_school_computer_science/dev-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:30a74a6e8039665801407c3dc800f4c67e5f8968a8cb45ead1043c228a2e2244
|
3 |
+
size 5438
|
high_school_computer_science/test-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9ece9b1a9af50face567f315d461694ae2a65de8b56d8a5e79d828148add8530
|
3 |
+
size 26584
|
high_school_computer_science/validation-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:778320a9d3cb87712a5ccc2f70c4558a2d9c64e157df54ae91ab94ef9df7d1cf
|
3 |
+
size 4692
|