Datasets:

Modalities:
Text
Formats:
parquet
Languages:
English
ArXiv:
Libraries:
Datasets
pandas
License:
SaylorTwift HF Staff commited on
Commit
7d775d2
·
verified ·
1 Parent(s): 717f8f1

Add 'astronomy' config data files

Browse files
README.md CHANGED
@@ -119,6 +119,8 @@ dataset_info:
119
  features:
120
  - name: question
121
  dtype: string
 
 
122
  - name: choices
123
  sequence: string
124
  - name: answer
@@ -131,19 +133,19 @@ dataset_info:
131
  '3': D
132
  splits:
133
  - name: auxiliary_train
134
- num_bytes: 160601377
135
  num_examples: 99842
136
  - name: test
137
- num_bytes: 46771
138
  num_examples: 152
139
  - name: validation
140
- num_bytes: 5027
141
  num_examples: 16
142
  - name: dev
143
- num_bytes: 2076
144
  num_examples: 5
145
- download_size: 166184960
146
- dataset_size: 160655251
147
  - config_name: business_ethics
148
  features:
149
  - name: question
@@ -1741,6 +1743,16 @@ configs:
1741
  path: anatomy/validation-*
1742
  - split: dev
1743
  path: anatomy/dev-*
 
 
 
 
 
 
 
 
 
 
1744
  ---
1745
 
1746
  # Dataset Card for MMLU
 
119
  features:
120
  - name: question
121
  dtype: string
122
+ - name: subject
123
+ dtype: string
124
  - name: choices
125
  sequence: string
126
  - name: answer
 
133
  '3': D
134
  splits:
135
  - name: auxiliary_train
136
+ num_bytes: 161000625
137
  num_examples: 99842
138
  - name: test
139
+ num_bytes: 48735
140
  num_examples: 152
141
  - name: validation
142
+ num_bytes: 5223
143
  num_examples: 16
144
  - name: dev
145
+ num_bytes: 2129
146
  num_examples: 5
147
+ download_size: 47201551
148
+ dataset_size: 161056712
149
  - config_name: business_ethics
150
  features:
151
  - name: question
 
1743
  path: anatomy/validation-*
1744
  - split: dev
1745
  path: anatomy/dev-*
1746
+ - config_name: astronomy
1747
+ data_files:
1748
+ - split: auxiliary_train
1749
+ path: astronomy/auxiliary_train-*
1750
+ - split: test
1751
+ path: astronomy/test-*
1752
+ - split: validation
1753
+ path: astronomy/validation-*
1754
+ - split: dev
1755
+ path: astronomy/dev-*
1756
  ---
1757
 
1758
  # Dataset Card for MMLU
astronomy/auxiliary_train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07
3
+ size 47163955
astronomy/dev-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:23ea71eb6555b7f89b99eeebe86b5deb95faf6bca70b2bb983b6d589a0ac7055
3
+ size 4211
astronomy/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:097742de0ec4687bd0cd8eb86a9fb38f3fe3498dca45d4cb09f630badc1d5598
3
+ size 27806
astronomy/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:16eb5b3b1ca9c64c195238e33c8b0cac1f2e5b606198398d5b8bb445864465f7
3
+ size 5579
dataset_infos.json CHANGED
@@ -147,39 +147,34 @@
147
  "features": {
148
  "question": {
149
  "dtype": "string",
150
- "id": null,
 
 
 
151
  "_type": "Value"
152
  },
153
  "choices": {
154
  "feature": {
155
  "dtype": "string",
156
- "id": null,
157
  "_type": "Value"
158
  },
159
- "length": -1,
160
- "id": null,
161
  "_type": "Sequence"
162
  },
163
  "answer": {
164
- "num_classes": 4,
165
  "names": [
166
  "A",
167
  "B",
168
  "C",
169
  "D"
170
  ],
171
- "id": null,
172
  "_type": "ClassLabel"
173
  }
174
  },
175
- "post_processed": null,
176
- "supervised_keys": null,
177
- "task_templates": null,
178
- "builder_name": "mmlu",
179
  "config_name": "astronomy",
180
  "version": {
181
  "version_str": "1.0.0",
182
- "description": null,
183
  "major": 1,
184
  "minor": 0,
185
  "patch": 0
@@ -187,39 +182,32 @@
187
  "splits": {
188
  "auxiliary_train": {
189
  "name": "auxiliary_train",
190
- "num_bytes": 160601257,
191
  "num_examples": 99842,
192
- "dataset_name": "mmlu"
193
  },
194
  "test": {
195
  "name": "test",
196
- "num_bytes": 46759,
197
  "num_examples": 152,
198
- "dataset_name": "mmlu"
199
  },
200
  "validation": {
201
  "name": "validation",
202
- "num_bytes": 5015,
203
  "num_examples": 16,
204
- "dataset_name": "mmlu"
205
  },
206
  "dev": {
207
  "name": "dev",
208
- "num_bytes": 2064,
209
  "num_examples": 5,
210
- "dataset_name": "mmlu"
211
- }
212
- },
213
- "download_checksums": {
214
- "data.tar": {
215
- "num_bytes": 166184960,
216
- "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"
217
  }
218
  },
219
- "download_size": 166184960,
220
- "post_processing_size": null,
221
- "dataset_size": 160655095,
222
- "size_in_bytes": 326840055
223
  },
224
  "business_ethics": {
225
  "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n",
 
147
  "features": {
148
  "question": {
149
  "dtype": "string",
150
+ "_type": "Value"
151
+ },
152
+ "subject": {
153
+ "dtype": "string",
154
  "_type": "Value"
155
  },
156
  "choices": {
157
  "feature": {
158
  "dtype": "string",
 
159
  "_type": "Value"
160
  },
 
 
161
  "_type": "Sequence"
162
  },
163
  "answer": {
 
164
  "names": [
165
  "A",
166
  "B",
167
  "C",
168
  "D"
169
  ],
 
170
  "_type": "ClassLabel"
171
  }
172
  },
173
+ "builder_name": "parquet",
174
+ "dataset_name": "mmlu",
 
 
175
  "config_name": "astronomy",
176
  "version": {
177
  "version_str": "1.0.0",
 
178
  "major": 1,
179
  "minor": 0,
180
  "patch": 0
 
182
  "splits": {
183
  "auxiliary_train": {
184
  "name": "auxiliary_train",
185
+ "num_bytes": 161000625,
186
  "num_examples": 99842,
187
+ "dataset_name": null
188
  },
189
  "test": {
190
  "name": "test",
191
+ "num_bytes": 48735,
192
  "num_examples": 152,
193
+ "dataset_name": null
194
  },
195
  "validation": {
196
  "name": "validation",
197
+ "num_bytes": 5223,
198
  "num_examples": 16,
199
+ "dataset_name": null
200
  },
201
  "dev": {
202
  "name": "dev",
203
+ "num_bytes": 2129,
204
  "num_examples": 5,
205
+ "dataset_name": null
 
 
 
 
 
 
206
  }
207
  },
208
+ "download_size": 47201551,
209
+ "dataset_size": 161056712,
210
+ "size_in_bytes": 208258263
 
211
  },
212
  "business_ethics": {
213
  "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n",