Datasets:

Modalities:
Text
Formats:
parquet
Languages:
English
ArXiv:
Libraries:
Datasets
pandas
License:
SaylorTwift HF Staff commited on
Commit
acd7cc6
·
verified ·
1 Parent(s): d7bc9b8

Add 'college_mathematics' config data files

Browse files
README.md CHANGED
@@ -305,6 +305,8 @@ dataset_info:
305
  features:
306
  - name: question
307
  dtype: string
 
 
308
  - name: choices
309
  sequence: string
310
  - name: answer
@@ -317,19 +319,19 @@ dataset_info:
317
  '3': D
318
  splits:
319
  - name: auxiliary_train
320
- num_bytes: 160601377
321
  num_examples: 99842
322
  - name: test
323
- num_bytes: 24711
324
  num_examples: 100
325
  - name: validation
326
- num_bytes: 2668
327
  num_examples: 11
328
  - name: dev
329
- num_bytes: 1493
330
  num_examples: 5
331
- download_size: 166184960
332
- dataset_size: 160630249
333
  - config_name: college_medicine
334
  features:
335
  - name: question
@@ -1813,6 +1815,16 @@ configs:
1813
  path: college_computer_science/validation-*
1814
  - split: dev
1815
  path: college_computer_science/dev-*
 
 
 
 
 
 
 
 
 
 
1816
  ---
1817
 
1818
  # Dataset Card for MMLU
 
305
  features:
306
  - name: question
307
  dtype: string
308
+ - name: subject
309
+ dtype: string
310
  - name: choices
311
  sequence: string
312
  - name: answer
 
319
  '3': D
320
  splits:
321
  - name: auxiliary_train
322
+ num_bytes: 161000625
323
  num_examples: 99842
324
  - name: test
325
+ num_bytes: 26999
326
  num_examples: 100
327
  - name: validation
328
+ num_bytes: 2909
329
  num_examples: 11
330
  - name: dev
331
+ num_bytes: 1596
332
  num_examples: 5
333
+ download_size: 47188597
334
+ dataset_size: 161032129
335
  - config_name: college_medicine
336
  features:
337
  - name: question
 
1815
  path: college_computer_science/validation-*
1816
  - split: dev
1817
  path: college_computer_science/dev-*
1818
+ - config_name: college_mathematics
1819
+ data_files:
1820
+ - split: auxiliary_train
1821
+ path: college_mathematics/auxiliary_train-*
1822
+ - split: test
1823
+ path: college_mathematics/test-*
1824
+ - split: validation
1825
+ path: college_mathematics/validation-*
1826
+ - split: dev
1827
+ path: college_mathematics/dev-*
1828
  ---
1829
 
1830
  # Dataset Card for MMLU
college_mathematics/auxiliary_train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07
3
+ size 47163955
college_mathematics/dev-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:95b2d5d13acb3b0a89b85d6b00037f8863c6734bdf56b67bfd4a9ef0567b9438
3
+ size 4251
college_mathematics/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:33d15bfa8968d680f3fe8278209aa0e1efad9921a392d5d9528d2d3a4afe831f
3
+ size 15989
college_mathematics/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:02c938bd6d1e498788ab7c68aa2d039675bb91e3c5649d3be3c970cd4a298278
3
+ size 4402
dataset_infos.json CHANGED
@@ -567,39 +567,34 @@
567
  "features": {
568
  "question": {
569
  "dtype": "string",
570
- "id": null,
 
 
 
571
  "_type": "Value"
572
  },
573
  "choices": {
574
  "feature": {
575
  "dtype": "string",
576
- "id": null,
577
  "_type": "Value"
578
  },
579
- "length": -1,
580
- "id": null,
581
  "_type": "Sequence"
582
  },
583
  "answer": {
584
- "num_classes": 4,
585
  "names": [
586
  "A",
587
  "B",
588
  "C",
589
  "D"
590
  ],
591
- "id": null,
592
  "_type": "ClassLabel"
593
  }
594
  },
595
- "post_processed": null,
596
- "supervised_keys": null,
597
- "task_templates": null,
598
- "builder_name": "mmlu",
599
  "config_name": "college_mathematics",
600
  "version": {
601
  "version_str": "1.0.0",
602
- "description": null,
603
  "major": 1,
604
  "minor": 0,
605
  "patch": 0
@@ -607,39 +602,32 @@
607
  "splits": {
608
  "auxiliary_train": {
609
  "name": "auxiliary_train",
610
- "num_bytes": 160601257,
611
  "num_examples": 99842,
612
- "dataset_name": "mmlu"
613
  },
614
  "test": {
615
  "name": "test",
616
- "num_bytes": 24699,
617
  "num_examples": 100,
618
- "dataset_name": "mmlu"
619
  },
620
  "validation": {
621
  "name": "validation",
622
- "num_bytes": 2656,
623
  "num_examples": 11,
624
- "dataset_name": "mmlu"
625
  },
626
  "dev": {
627
  "name": "dev",
628
- "num_bytes": 1481,
629
  "num_examples": 5,
630
- "dataset_name": "mmlu"
631
- }
632
- },
633
- "download_checksums": {
634
- "data.tar": {
635
- "num_bytes": 166184960,
636
- "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"
637
  }
638
  },
639
- "download_size": 166184960,
640
- "post_processing_size": null,
641
- "dataset_size": 160630093,
642
- "size_in_bytes": 326815053
643
  },
644
  "college_medicine": {
645
  "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n",
 
567
  "features": {
568
  "question": {
569
  "dtype": "string",
570
+ "_type": "Value"
571
+ },
572
+ "subject": {
573
+ "dtype": "string",
574
  "_type": "Value"
575
  },
576
  "choices": {
577
  "feature": {
578
  "dtype": "string",
 
579
  "_type": "Value"
580
  },
 
 
581
  "_type": "Sequence"
582
  },
583
  "answer": {
 
584
  "names": [
585
  "A",
586
  "B",
587
  "C",
588
  "D"
589
  ],
 
590
  "_type": "ClassLabel"
591
  }
592
  },
593
+ "builder_name": "parquet",
594
+ "dataset_name": "mmlu",
 
 
595
  "config_name": "college_mathematics",
596
  "version": {
597
  "version_str": "1.0.0",
 
598
  "major": 1,
599
  "minor": 0,
600
  "patch": 0
 
602
  "splits": {
603
  "auxiliary_train": {
604
  "name": "auxiliary_train",
605
+ "num_bytes": 161000625,
606
  "num_examples": 99842,
607
+ "dataset_name": null
608
  },
609
  "test": {
610
  "name": "test",
611
+ "num_bytes": 26999,
612
  "num_examples": 100,
613
+ "dataset_name": null
614
  },
615
  "validation": {
616
  "name": "validation",
617
+ "num_bytes": 2909,
618
  "num_examples": 11,
619
+ "dataset_name": null
620
  },
621
  "dev": {
622
  "name": "dev",
623
+ "num_bytes": 1596,
624
  "num_examples": 5,
625
+ "dataset_name": null
 
 
 
 
 
 
626
  }
627
  },
628
+ "download_size": 47188597,
629
+ "dataset_size": 161032129,
630
+ "size_in_bytes": 208220726
 
631
  },
632
  "college_medicine": {
633
  "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n",