Datasets:

Modalities:
Text
Formats:
parquet
Languages:
English
ArXiv:
Libraries:
Datasets
pandas
License:
SaylorTwift HF Staff commited on
Commit
ada6f38
·
verified ·
1 Parent(s): 5fbf107

Add 'conceptual_physics' config data files

Browse files
README.md CHANGED
@@ -429,6 +429,8 @@ dataset_info:
429
  features:
430
  - name: question
431
  dtype: string
 
 
432
  - name: choices
433
  sequence: string
434
  - name: answer
@@ -441,19 +443,19 @@ dataset_info:
441
  '3': D
442
  splits:
443
  - name: auxiliary_train
444
- num_bytes: 160601377
445
  num_examples: 99842
446
  - name: test
447
- num_bytes: 40709
448
  num_examples: 235
449
  - name: validation
450
- num_bytes: 4474
451
  num_examples: 26
452
  - name: dev
453
- num_bytes: 934
454
  num_examples: 5
455
- download_size: 166184960
456
- dataset_size: 160647494
457
  - config_name: econometrics
458
  features:
459
  - name: question
@@ -1861,6 +1863,16 @@ configs:
1861
  path: computer_security/validation-*
1862
  - split: dev
1863
  path: computer_security/dev-*
 
 
 
 
 
 
 
 
 
 
1864
  ---
1865
 
1866
  # Dataset Card for MMLU
 
429
  features:
430
  - name: question
431
  dtype: string
432
+ - name: subject
433
+ dtype: string
434
  - name: choices
435
  sequence: string
436
  - name: answer
 
443
  '3': D
444
  splits:
445
  - name: auxiliary_train
446
+ num_bytes: 161000625
447
  num_examples: 99842
448
  - name: test
449
+ num_bytes: 45867
450
  num_examples: 235
451
  - name: validation
452
+ num_bytes: 5034
453
  num_examples: 26
454
  - name: dev
455
+ num_bytes: 1032
456
  num_examples: 5
457
+ download_size: 47197231
458
+ dataset_size: 161052558
459
  - config_name: econometrics
460
  features:
461
  - name: question
 
1863
  path: computer_security/validation-*
1864
  - split: dev
1865
  path: computer_security/dev-*
1866
+ - config_name: conceptual_physics
1867
+ data_files:
1868
+ - split: auxiliary_train
1869
+ path: conceptual_physics/auxiliary_train-*
1870
+ - split: test
1871
+ path: conceptual_physics/test-*
1872
+ - split: validation
1873
+ path: conceptual_physics/validation-*
1874
+ - split: dev
1875
+ path: conceptual_physics/dev-*
1876
  ---
1877
 
1878
  # Dataset Card for MMLU
conceptual_physics/auxiliary_train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07
3
+ size 47163955
conceptual_physics/dev-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:36e0fc357e65cb0945fdb9d77e9619d7f17738e99df51e8daf37758ff7c5cff5
3
+ size 3358
conceptual_physics/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d51a8ef47faab439431166fe5a0f4fc92737b586a6b3478f6a0d356ed98610f
3
+ size 24474
conceptual_physics/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:19a7f8cc46582c827b2f915679e957cb3a0087ed6229b65b0a0693827bd1472a
3
+ size 5444
dataset_infos.json CHANGED
@@ -847,39 +847,34 @@
847
  "features": {
848
  "question": {
849
  "dtype": "string",
850
- "id": null,
 
 
 
851
  "_type": "Value"
852
  },
853
  "choices": {
854
  "feature": {
855
  "dtype": "string",
856
- "id": null,
857
  "_type": "Value"
858
  },
859
- "length": -1,
860
- "id": null,
861
  "_type": "Sequence"
862
  },
863
  "answer": {
864
- "num_classes": 4,
865
  "names": [
866
  "A",
867
  "B",
868
  "C",
869
  "D"
870
  ],
871
- "id": null,
872
  "_type": "ClassLabel"
873
  }
874
  },
875
- "post_processed": null,
876
- "supervised_keys": null,
877
- "task_templates": null,
878
- "builder_name": "mmlu",
879
  "config_name": "conceptual_physics",
880
  "version": {
881
  "version_str": "1.0.0",
882
- "description": null,
883
  "major": 1,
884
  "minor": 0,
885
  "patch": 0
@@ -887,39 +882,32 @@
887
  "splits": {
888
  "auxiliary_train": {
889
  "name": "auxiliary_train",
890
- "num_bytes": 160601257,
891
  "num_examples": 99842,
892
- "dataset_name": "mmlu"
893
  },
894
  "test": {
895
  "name": "test",
896
- "num_bytes": 40697,
897
  "num_examples": 235,
898
- "dataset_name": "mmlu"
899
  },
900
  "validation": {
901
  "name": "validation",
902
- "num_bytes": 4462,
903
  "num_examples": 26,
904
- "dataset_name": "mmlu"
905
  },
906
  "dev": {
907
  "name": "dev",
908
- "num_bytes": 922,
909
  "num_examples": 5,
910
- "dataset_name": "mmlu"
911
- }
912
- },
913
- "download_checksums": {
914
- "data.tar": {
915
- "num_bytes": 166184960,
916
- "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"
917
  }
918
  },
919
- "download_size": 166184960,
920
- "post_processing_size": null,
921
- "dataset_size": 160647338,
922
- "size_in_bytes": 326832298
923
  },
924
  "econometrics": {
925
  "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n",
 
847
  "features": {
848
  "question": {
849
  "dtype": "string",
850
+ "_type": "Value"
851
+ },
852
+ "subject": {
853
+ "dtype": "string",
854
  "_type": "Value"
855
  },
856
  "choices": {
857
  "feature": {
858
  "dtype": "string",
 
859
  "_type": "Value"
860
  },
 
 
861
  "_type": "Sequence"
862
  },
863
  "answer": {
 
864
  "names": [
865
  "A",
866
  "B",
867
  "C",
868
  "D"
869
  ],
 
870
  "_type": "ClassLabel"
871
  }
872
  },
873
+ "builder_name": "parquet",
874
+ "dataset_name": "mmlu",
 
 
875
  "config_name": "conceptual_physics",
876
  "version": {
877
  "version_str": "1.0.0",
 
878
  "major": 1,
879
  "minor": 0,
880
  "patch": 0
 
882
  "splits": {
883
  "auxiliary_train": {
884
  "name": "auxiliary_train",
885
+ "num_bytes": 161000625,
886
  "num_examples": 99842,
887
+ "dataset_name": null
888
  },
889
  "test": {
890
  "name": "test",
891
+ "num_bytes": 45867,
892
  "num_examples": 235,
893
+ "dataset_name": null
894
  },
895
  "validation": {
896
  "name": "validation",
897
+ "num_bytes": 5034,
898
  "num_examples": 26,
899
+ "dataset_name": null
900
  },
901
  "dev": {
902
  "name": "dev",
903
+ "num_bytes": 1032,
904
  "num_examples": 5,
905
+ "dataset_name": null
 
 
 
 
 
 
906
  }
907
  },
908
+ "download_size": 47197231,
909
+ "dataset_size": 161052558,
910
+ "size_in_bytes": 208249789
 
911
  },
912
  "econometrics": {
913
  "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n",