wangrongsheng commited on
Commit
7527170
·
verified ·
1 Parent(s): cfab6f6

Upload 3 files

Browse files
README.md ADDED
@@ -0,0 +1,263 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ annotations_creators:
3
+ - found
4
+ language_creators:
5
+ - found
6
+ language:
7
+ - en
8
+ license:
9
+ - unknown
10
+ multilinguality:
11
+ - monolingual
12
+ size_categories:
13
+ - 100K<n<1M
14
+ source_datasets:
15
+ - original
16
+ task_categories:
17
+ - text-classification
18
+ task_ids:
19
+ - topic-classification
20
+ paperswithcode_id: ag-news
21
+ pretty_name: AG’s News Corpus
22
+ dataset_info:
23
+ features:
24
+ - name: text
25
+ dtype: string
26
+ - name: label
27
+ dtype:
28
+ class_label:
29
+ names:
30
+ '0': World
31
+ '1': Sports
32
+ '2': Business
33
+ '3': Sci/Tech
34
+ splits:
35
+ - name: train
36
+ num_bytes: 29817303
37
+ num_examples: 120000
38
+ - name: test
39
+ num_bytes: 1879474
40
+ num_examples: 7600
41
+ download_size: 19820267
42
+ dataset_size: 31696777
43
+ configs:
44
+ - config_name: default
45
+ data_files:
46
+ - split: train
47
+ path: data/train-*
48
+ - split: test
49
+ path: data/test-*
50
+ train-eval-index:
51
+ - config: default
52
+ task: text-classification
53
+ task_id: multi_class_classification
54
+ splits:
55
+ train_split: train
56
+ eval_split: test
57
+ col_mapping:
58
+ text: text
59
+ label: target
60
+ metrics:
61
+ - type: accuracy
62
+ name: Accuracy
63
+ - type: f1
64
+ name: F1 macro
65
+ args:
66
+ average: macro
67
+ - type: f1
68
+ name: F1 micro
69
+ args:
70
+ average: micro
71
+ - type: f1
72
+ name: F1 weighted
73
+ args:
74
+ average: weighted
75
+ - type: precision
76
+ name: Precision macro
77
+ args:
78
+ average: macro
79
+ - type: precision
80
+ name: Precision micro
81
+ args:
82
+ average: micro
83
+ - type: precision
84
+ name: Precision weighted
85
+ args:
86
+ average: weighted
87
+ - type: recall
88
+ name: Recall macro
89
+ args:
90
+ average: macro
91
+ - type: recall
92
+ name: Recall micro
93
+ args:
94
+ average: micro
95
+ - type: recall
96
+ name: Recall weighted
97
+ args:
98
+ average: weighted
99
+ ---
100
+
101
+ # Dataset Card for "ag_news"
102
+
103
+ ## Table of Contents
104
+ - [Dataset Description](#dataset-description)
105
+ - [Dataset Summary](#dataset-summary)
106
+ - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
107
+ - [Languages](#languages)
108
+ - [Dataset Structure](#dataset-structure)
109
+ - [Data Instances](#data-instances)
110
+ - [Data Fields](#data-fields)
111
+ - [Data Splits](#data-splits)
112
+ - [Dataset Creation](#dataset-creation)
113
+ - [Curation Rationale](#curation-rationale)
114
+ - [Source Data](#source-data)
115
+ - [Annotations](#annotations)
116
+ - [Personal and Sensitive Information](#personal-and-sensitive-information)
117
+ - [Considerations for Using the Data](#considerations-for-using-the-data)
118
+ - [Social Impact of Dataset](#social-impact-of-dataset)
119
+ - [Discussion of Biases](#discussion-of-biases)
120
+ - [Other Known Limitations](#other-known-limitations)
121
+ - [Additional Information](#additional-information)
122
+ - [Dataset Curators](#dataset-curators)
123
+ - [Licensing Information](#licensing-information)
124
+ - [Citation Information](#citation-information)
125
+ - [Contributions](#contributions)
126
+
127
+ ## Dataset Description
128
+
129
+ - **Homepage:** [http://groups.di.unipi.it/~gulli/AG_corpus_of_news_articles.html](http://groups.di.unipi.it/~gulli/AG_corpus_of_news_articles.html)
130
+ - **Repository:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
131
+ - **Paper:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
132
+ - **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
133
+ - **Size of downloaded dataset files:** 31.33 MB
134
+ - **Size of the generated dataset:** 31.70 MB
135
+ - **Total amount of disk used:** 63.02 MB
136
+
137
+ ### Dataset Summary
138
+
139
+ AG is a collection of more than 1 million news articles. News articles have been
140
+ gathered from more than 2000 news sources by ComeToMyHead in more than 1 year of
141
+ activity. ComeToMyHead is an academic news search engine which has been running
142
+ since July, 2004. The dataset is provided by the academic comunity for research
143
+ purposes in data mining (clustering, classification, etc), information retrieval
144
+ (ranking, search, etc), xml, data compression, data streaming, and any other
145
+ non-commercial activity. For more information, please refer to the link
146
+ http://www.di.unipi.it/~gulli/AG_corpus_of_news_articles.html .
147
+
148
+ The AG's news topic classification dataset is constructed by Xiang Zhang
149
+ ([email protected]) from the dataset above. It is used as a text
150
+ classification benchmark in the following paper: Xiang Zhang, Junbo Zhao, Yann
151
+ LeCun. Character-level Convolutional Networks for Text Classification. Advances
152
+ in Neural Information Processing Systems 28 (NIPS 2015).
153
+
154
+ ### Supported Tasks and Leaderboards
155
+
156
+ [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
157
+
158
+ ### Languages
159
+
160
+ [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
161
+
162
+ ## Dataset Structure
163
+
164
+ ### Data Instances
165
+
166
+ #### default
167
+
168
+ - **Size of downloaded dataset files:** 31.33 MB
169
+ - **Size of the generated dataset:** 31.70 MB
170
+ - **Total amount of disk used:** 63.02 MB
171
+
172
+ An example of 'train' looks as follows.
173
+ ```
174
+ {
175
+ "label": 3,
176
+ "text": "New iPad released Just like every other September, this one is no different. Apple is planning to release a bigger, heavier, fatter iPad that..."
177
+ }
178
+ ```
179
+
180
+ ### Data Fields
181
+
182
+ The data fields are the same among all splits.
183
+
184
+ #### default
185
+ - `text`: a `string` feature.
186
+ - `label`: a classification label, with possible values including `World` (0), `Sports` (1), `Business` (2), `Sci/Tech` (3).
187
+
188
+ ### Data Splits
189
+
190
+ | name |train |test|
191
+ |-------|-----:|---:|
192
+ |default|120000|7600|
193
+
194
+ ## Dataset Creation
195
+
196
+ ### Curation Rationale
197
+
198
+ [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
199
+
200
+ ### Source Data
201
+
202
+ #### Initial Data Collection and Normalization
203
+
204
+ [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
205
+
206
+ #### Who are the source language producers?
207
+
208
+ [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
209
+
210
+ ### Annotations
211
+
212
+ #### Annotation process
213
+
214
+ [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
215
+
216
+ #### Who are the annotators?
217
+
218
+ [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
219
+
220
+ ### Personal and Sensitive Information
221
+
222
+ [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
223
+
224
+ ## Considerations for Using the Data
225
+
226
+ ### Social Impact of Dataset
227
+
228
+ [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
229
+
230
+ ### Discussion of Biases
231
+
232
+ [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
233
+
234
+ ### Other Known Limitations
235
+
236
+ [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
237
+
238
+ ## Additional Information
239
+
240
+ ### Dataset Curators
241
+
242
+ [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
243
+
244
+ ### Licensing Information
245
+
246
+ [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
247
+
248
+ ### Citation Information
249
+
250
+ ```
251
+ @inproceedings{Zhang2015CharacterlevelCN,
252
+ title={Character-level Convolutional Networks for Text Classification},
253
+ author={Xiang Zhang and Junbo Jake Zhao and Yann LeCun},
254
+ booktitle={NIPS},
255
+ year={2015}
256
+ }
257
+
258
+ ```
259
+
260
+
261
+ ### Contributions
262
+
263
+ Thanks to [@jxmorris12](https://github.com/jxmorris12), [@thomwolf](https://github.com/thomwolf), [@lhoestq](https://github.com/lhoestq), [@lewtun](https://github.com/lewtun) for adding this dataset.
data/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:71de87ec66bc5737752a2502204dfa6d7fe9856ade3ea444dc6317789a4f13fb
3
+ size 1234829
data/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fc508d6d9868594e3da960a8cfeb63ab5a4746598b93428c224397080c1f52ee
3
+ size 18585438