Datasets:

Modalities:
Text
Formats:
parquet
Languages:
English
ArXiv:
Libraries:
Datasets
pandas
License:
SaylorTwift HF Staff commited on
Commit
5fbf107
·
verified ·
1 Parent(s): 5c4a97b

Add 'computer_security' config data files

Browse files
README.md CHANGED
@@ -398,6 +398,8 @@ dataset_info:
398
  features:
399
  - name: question
400
  dtype: string
 
 
401
  - name: choices
402
  sequence: string
403
  - name: answer
@@ -410,19 +412,19 @@ dataset_info:
410
  '3': D
411
  splits:
412
  - name: auxiliary_train
413
- num_bytes: 160601377
414
  num_examples: 99842
415
  - name: test
416
- num_bytes: 27124
417
  num_examples: 100
418
  - name: validation
419
- num_bytes: 4549
420
  num_examples: 11
421
  - name: dev
422
- num_bytes: 1101
423
  num_examples: 5
424
- download_size: 166184960
425
- dataset_size: 160634151
426
  - config_name: conceptual_physics
427
  features:
428
  - name: question
@@ -1849,6 +1851,16 @@ configs:
1849
  path: college_physics/validation-*
1850
  - split: dev
1851
  path: college_physics/dev-*
 
 
 
 
 
 
 
 
 
 
1852
  ---
1853
 
1854
  # Dataset Card for MMLU
 
398
  features:
399
  - name: question
400
  dtype: string
401
+ - name: subject
402
+ dtype: string
403
  - name: choices
404
  sequence: string
405
  - name: answer
 
412
  '3': D
413
  splits:
414
  - name: auxiliary_train
415
+ num_bytes: 161000625
416
  num_examples: 99842
417
  - name: test
418
+ num_bytes: 29212
419
  num_examples: 100
420
  - name: validation
421
+ num_bytes: 4768
422
  num_examples: 11
423
  - name: dev
424
+ num_bytes: 1194
425
  num_examples: 5
426
+ download_size: 47192155
427
+ dataset_size: 161035799
428
  - config_name: conceptual_physics
429
  features:
430
  - name: question
 
1851
  path: college_physics/validation-*
1852
  - split: dev
1853
  path: college_physics/dev-*
1854
+ - config_name: computer_security
1855
+ data_files:
1856
+ - split: auxiliary_train
1857
+ path: computer_security/auxiliary_train-*
1858
+ - split: test
1859
+ path: computer_security/test-*
1860
+ - split: validation
1861
+ path: computer_security/validation-*
1862
+ - split: dev
1863
+ path: computer_security/dev-*
1864
  ---
1865
 
1866
  # Dataset Card for MMLU
computer_security/auxiliary_train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07
3
+ size 47163955
computer_security/dev-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:55d2262c094ab2299e662de8f51efb35ca4764a007c2bd82c62eb070c224509f
3
+ size 3659
computer_security/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3f55080db8cca9b9542bcf695b032495d0623e5f796a9349756bdd58c3962ca7
3
+ size 18577
computer_security/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:381b548c745e1e1b51a48b74cd4c6b94a545bba92efe74d31d5aad189e2fd9f9
3
+ size 5964
dataset_infos.json CHANGED
@@ -777,39 +777,34 @@
777
  "features": {
778
  "question": {
779
  "dtype": "string",
780
- "id": null,
 
 
 
781
  "_type": "Value"
782
  },
783
  "choices": {
784
  "feature": {
785
  "dtype": "string",
786
- "id": null,
787
  "_type": "Value"
788
  },
789
- "length": -1,
790
- "id": null,
791
  "_type": "Sequence"
792
  },
793
  "answer": {
794
- "num_classes": 4,
795
  "names": [
796
  "A",
797
  "B",
798
  "C",
799
  "D"
800
  ],
801
- "id": null,
802
  "_type": "ClassLabel"
803
  }
804
  },
805
- "post_processed": null,
806
- "supervised_keys": null,
807
- "task_templates": null,
808
- "builder_name": "mmlu",
809
  "config_name": "computer_security",
810
  "version": {
811
  "version_str": "1.0.0",
812
- "description": null,
813
  "major": 1,
814
  "minor": 0,
815
  "patch": 0
@@ -817,39 +812,32 @@
817
  "splits": {
818
  "auxiliary_train": {
819
  "name": "auxiliary_train",
820
- "num_bytes": 160601257,
821
  "num_examples": 99842,
822
- "dataset_name": "mmlu"
823
  },
824
  "test": {
825
  "name": "test",
826
- "num_bytes": 27112,
827
  "num_examples": 100,
828
- "dataset_name": "mmlu"
829
  },
830
  "validation": {
831
  "name": "validation",
832
- "num_bytes": 4537,
833
  "num_examples": 11,
834
- "dataset_name": "mmlu"
835
  },
836
  "dev": {
837
  "name": "dev",
838
- "num_bytes": 1089,
839
  "num_examples": 5,
840
- "dataset_name": "mmlu"
841
- }
842
- },
843
- "download_checksums": {
844
- "data.tar": {
845
- "num_bytes": 166184960,
846
- "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"
847
  }
848
  },
849
- "download_size": 166184960,
850
- "post_processing_size": null,
851
- "dataset_size": 160633995,
852
- "size_in_bytes": 326818955
853
  },
854
  "conceptual_physics": {
855
  "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n",
 
777
  "features": {
778
  "question": {
779
  "dtype": "string",
780
+ "_type": "Value"
781
+ },
782
+ "subject": {
783
+ "dtype": "string",
784
  "_type": "Value"
785
  },
786
  "choices": {
787
  "feature": {
788
  "dtype": "string",
 
789
  "_type": "Value"
790
  },
 
 
791
  "_type": "Sequence"
792
  },
793
  "answer": {
 
794
  "names": [
795
  "A",
796
  "B",
797
  "C",
798
  "D"
799
  ],
 
800
  "_type": "ClassLabel"
801
  }
802
  },
803
+ "builder_name": "parquet",
804
+ "dataset_name": "mmlu",
 
 
805
  "config_name": "computer_security",
806
  "version": {
807
  "version_str": "1.0.0",
 
808
  "major": 1,
809
  "minor": 0,
810
  "patch": 0
 
812
  "splits": {
813
  "auxiliary_train": {
814
  "name": "auxiliary_train",
815
+ "num_bytes": 161000625,
816
  "num_examples": 99842,
817
+ "dataset_name": null
818
  },
819
  "test": {
820
  "name": "test",
821
+ "num_bytes": 29212,
822
  "num_examples": 100,
823
+ "dataset_name": null
824
  },
825
  "validation": {
826
  "name": "validation",
827
+ "num_bytes": 4768,
828
  "num_examples": 11,
829
+ "dataset_name": null
830
  },
831
  "dev": {
832
  "name": "dev",
833
+ "num_bytes": 1194,
834
  "num_examples": 5,
835
+ "dataset_name": null
 
 
 
 
 
 
836
  }
837
  },
838
+ "download_size": 47192155,
839
+ "dataset_size": 161035799,
840
+ "size_in_bytes": 208227954
 
841
  },
842
  "conceptual_physics": {
843
  "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n",