url
stringlengths 58
61
| repository_url
stringclasses 1
value | labels_url
stringlengths 72
75
| comments_url
stringlengths 67
70
| events_url
stringlengths 65
68
| html_url
stringlengths 46
51
| id
int64 599M
2.12B
| node_id
stringlengths 18
32
| number
int64 1
6.65k
| title
stringlengths 1
290
| user
dict | labels
listlengths 0
4
| state
stringclasses 2
values | locked
bool 1
class | assignee
dict | assignees
listlengths 0
4
| milestone
dict | comments
int64 0
70
| created_at
unknown | updated_at
unknown | closed_at
unknown | author_association
stringclasses 3
values | active_lock_reason
float64 | draft
float64 0
1
⌀ | pull_request
dict | body
stringlengths 0
228k
⌀ | reactions
dict | timeline_url
stringlengths 67
70
| performed_via_github_app
float64 | state_reason
stringclasses 3
values | is_pull_request
bool 2
classes |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
https://api.github.com/repos/huggingface/datasets/issues/927 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/927/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/927/comments | https://api.github.com/repos/huggingface/datasets/issues/927/events | https://github.com/huggingface/datasets/issues/927 | 753,679,020 | MDU6SXNzdWU3NTM2NzkwMjA= | 927 | Hello | {
"avatar_url": "https://avatars.githubusercontent.com/u/75259546?v=4",
"events_url": "https://api.github.com/users/k125-ak/events{/privacy}",
"followers_url": "https://api.github.com/users/k125-ak/followers",
"following_url": "https://api.github.com/users/k125-ak/following{/other_user}",
"gists_url": "https://api.github.com/users/k125-ak/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/k125-ak",
"id": 75259546,
"login": "k125-ak",
"node_id": "MDQ6VXNlcjc1MjU5NTQ2",
"organizations_url": "https://api.github.com/users/k125-ak/orgs",
"received_events_url": "https://api.github.com/users/k125-ak/received_events",
"repos_url": "https://api.github.com/users/k125-ak/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/k125-ak/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/k125-ak/subscriptions",
"type": "User",
"url": "https://api.github.com/users/k125-ak"
} | [] | closed | false | null | [] | null | 0 | "2020-11-30T17:50:05Z" | "2020-11-30T17:50:30Z" | "2020-11-30T17:50:30Z" | NONE | null | null | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/927/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/927/timeline | null | completed | false |
|
https://api.github.com/repos/huggingface/datasets/issues/926 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/926/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/926/comments | https://api.github.com/repos/huggingface/datasets/issues/926/events | https://github.com/huggingface/datasets/pull/926 | 753,676,069 | MDExOlB1bGxSZXF1ZXN0NTI5NzA4MTcy | 926 | add inquisitive | {
"avatar_url": "https://avatars.githubusercontent.com/u/27137566?v=4",
"events_url": "https://api.github.com/users/patil-suraj/events{/privacy}",
"followers_url": "https://api.github.com/users/patil-suraj/followers",
"following_url": "https://api.github.com/users/patil-suraj/following{/other_user}",
"gists_url": "https://api.github.com/users/patil-suraj/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/patil-suraj",
"id": 27137566,
"login": "patil-suraj",
"node_id": "MDQ6VXNlcjI3MTM3NTY2",
"organizations_url": "https://api.github.com/users/patil-suraj/orgs",
"received_events_url": "https://api.github.com/users/patil-suraj/received_events",
"repos_url": "https://api.github.com/users/patil-suraj/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/patil-suraj/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/patil-suraj/subscriptions",
"type": "User",
"url": "https://api.github.com/users/patil-suraj"
} | [] | closed | false | null | [] | null | 3 | "2020-11-30T17:45:22Z" | "2020-12-02T13:45:22Z" | "2020-12-02T13:40:13Z" | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/926.diff",
"html_url": "https://github.com/huggingface/datasets/pull/926",
"merged_at": "2020-12-02T13:40:13Z",
"patch_url": "https://github.com/huggingface/datasets/pull/926.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/926"
} | Adding inquisitive qg dataset
More info: https://github.com/wjko2/INQUISITIVE | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/926/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/926/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/925 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/925/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/925/comments | https://api.github.com/repos/huggingface/datasets/issues/925/events | https://github.com/huggingface/datasets/pull/925 | 753,672,661 | MDExOlB1bGxSZXF1ZXN0NTI5NzA1MzM4 | 925 | Add Turku NLP Corpus for Finnish NER | {
"avatar_url": "https://avatars.githubusercontent.com/u/1183441?v=4",
"events_url": "https://api.github.com/users/abhishekkrthakur/events{/privacy}",
"followers_url": "https://api.github.com/users/abhishekkrthakur/followers",
"following_url": "https://api.github.com/users/abhishekkrthakur/following{/other_user}",
"gists_url": "https://api.github.com/users/abhishekkrthakur/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/abhishekkrthakur",
"id": 1183441,
"login": "abhishekkrthakur",
"node_id": "MDQ6VXNlcjExODM0NDE=",
"organizations_url": "https://api.github.com/users/abhishekkrthakur/orgs",
"received_events_url": "https://api.github.com/users/abhishekkrthakur/received_events",
"repos_url": "https://api.github.com/users/abhishekkrthakur/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/abhishekkrthakur/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/abhishekkrthakur/subscriptions",
"type": "User",
"url": "https://api.github.com/users/abhishekkrthakur"
} | [] | closed | false | null | [] | null | 1 | "2020-11-30T17:40:19Z" | "2020-12-03T14:07:11Z" | "2020-12-03T14:07:10Z" | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/925.diff",
"html_url": "https://github.com/huggingface/datasets/pull/925",
"merged_at": "2020-12-03T14:07:10Z",
"patch_url": "https://github.com/huggingface/datasets/pull/925.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/925"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/925/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/925/timeline | null | null | true |
|
https://api.github.com/repos/huggingface/datasets/issues/924 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/924/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/924/comments | https://api.github.com/repos/huggingface/datasets/issues/924/events | https://github.com/huggingface/datasets/pull/924 | 753,631,951 | MDExOlB1bGxSZXF1ZXN0NTI5NjcyMzgw | 924 | Add DART | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq"
} | [] | closed | false | null | [] | null | 1 | "2020-11-30T16:42:37Z" | "2020-12-02T03:13:42Z" | "2020-12-02T03:13:41Z" | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/924.diff",
"html_url": "https://github.com/huggingface/datasets/pull/924",
"merged_at": "2020-12-02T03:13:41Z",
"patch_url": "https://github.com/huggingface/datasets/pull/924.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/924"
} | - **Name:** *DART*
- **Description:** *DART is a large dataset for open-domain structured data record to text generation.*
- **Paper:** *https://arxiv.org/abs/2007.02871*
- **Data:** *https://github.com/Yale-LILY/dart#leaderboard*
### Checkbox
- [x] Create the dataset script `/datasets/my_dataset/my_dataset.py` using the template
- [x] Fill the `_DESCRIPTION` and `_CITATION` variables
- [x] Implement `_infos()`, `_split_generators()` and `_generate_examples()`
- [x] Make sure that the `BUILDER_CONFIGS` class attribute is filled with the different configurations of the dataset and that the `BUILDER_CONFIG_CLASS` is specified if there is a custom config class.
- [x] Generate the metadata file `dataset_infos.json` for all configurations
- [x] Generate the dummy data `dummy_data.zip` files to have the dataset script tested and that they don't weigh too much (<50KB)
- [x] Add the dataset card `README.md` using the template : fill the tags and the various paragraphs
- [x] Both tests for the real data and the dummy data pass. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/924/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/924/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/923 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/923/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/923/comments | https://api.github.com/repos/huggingface/datasets/issues/923/events | https://github.com/huggingface/datasets/pull/923 | 753,569,220 | MDExOlB1bGxSZXF1ZXN0NTI5NjIyMDQx | 923 | Add CC-100 dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",
"gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/albertvillanova",
"id": 8515462,
"login": "albertvillanova",
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"organizations_url": "https://api.github.com/users/albertvillanova/orgs",
"received_events_url": "https://api.github.com/users/albertvillanova/received_events",
"repos_url": "https://api.github.com/users/albertvillanova/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions",
"type": "User",
"url": "https://api.github.com/users/albertvillanova"
} | [
{
"color": "ffffff",
"default": true,
"description": "This will not be worked on",
"id": 1935892913,
"name": "wontfix",
"node_id": "MDU6TGFiZWwxOTM1ODkyOTEz",
"url": "https://api.github.com/repos/huggingface/datasets/labels/wontfix"
}
] | closed | false | null | [] | null | 10 | "2020-11-30T15:23:22Z" | "2021-04-20T13:34:17Z" | "2021-04-20T13:34:17Z" | MEMBER | null | 1 | {
"diff_url": "https://github.com/huggingface/datasets/pull/923.diff",
"html_url": "https://github.com/huggingface/datasets/pull/923",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/923.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/923"
} | Add CC-100.
Close #773 | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/923/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/923/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/922 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/922/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/922/comments | https://api.github.com/repos/huggingface/datasets/issues/922/events | https://github.com/huggingface/datasets/pull/922 | 753,559,130 | MDExOlB1bGxSZXF1ZXN0NTI5NjEzOTA4 | 922 | Add XOR QA Dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/28291870?v=4",
"events_url": "https://api.github.com/users/sumanthd17/events{/privacy}",
"followers_url": "https://api.github.com/users/sumanthd17/followers",
"following_url": "https://api.github.com/users/sumanthd17/following{/other_user}",
"gists_url": "https://api.github.com/users/sumanthd17/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/sumanthd17",
"id": 28291870,
"login": "sumanthd17",
"node_id": "MDQ6VXNlcjI4MjkxODcw",
"organizations_url": "https://api.github.com/users/sumanthd17/orgs",
"received_events_url": "https://api.github.com/users/sumanthd17/received_events",
"repos_url": "https://api.github.com/users/sumanthd17/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/sumanthd17/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/sumanthd17/subscriptions",
"type": "User",
"url": "https://api.github.com/users/sumanthd17"
} | [] | closed | false | null | [] | null | 4 | "2020-11-30T15:10:54Z" | "2020-12-02T03:12:21Z" | "2020-12-02T03:12:21Z" | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/922.diff",
"html_url": "https://github.com/huggingface/datasets/pull/922",
"merged_at": "2020-12-02T03:12:21Z",
"patch_url": "https://github.com/huggingface/datasets/pull/922.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/922"
} | Added XOR Question Answering Dataset. The link to the dataset can be found [here](https://nlp.cs.washington.edu/xorqa/)
- [x] Followed the instructions in CONTRIBUTING.md
- [x] Ran the tests successfully
- [x] Created the dummy data | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/922/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/922/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/920 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/920/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/920/comments | https://api.github.com/repos/huggingface/datasets/issues/920/events | https://github.com/huggingface/datasets/pull/920 | 753,445,747 | MDExOlB1bGxSZXF1ZXN0NTI5NTIzMTgz | 920 | add dream dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/27137566?v=4",
"events_url": "https://api.github.com/users/patil-suraj/events{/privacy}",
"followers_url": "https://api.github.com/users/patil-suraj/followers",
"following_url": "https://api.github.com/users/patil-suraj/following{/other_user}",
"gists_url": "https://api.github.com/users/patil-suraj/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/patil-suraj",
"id": 27137566,
"login": "patil-suraj",
"node_id": "MDQ6VXNlcjI3MTM3NTY2",
"organizations_url": "https://api.github.com/users/patil-suraj/orgs",
"received_events_url": "https://api.github.com/users/patil-suraj/received_events",
"repos_url": "https://api.github.com/users/patil-suraj/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/patil-suraj/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/patil-suraj/subscriptions",
"type": "User",
"url": "https://api.github.com/users/patil-suraj"
} | [] | closed | false | null | [] | null | 6 | "2020-11-30T12:40:14Z" | "2020-12-03T16:45:12Z" | "2020-12-02T15:39:12Z" | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/920.diff",
"html_url": "https://github.com/huggingface/datasets/pull/920",
"merged_at": "2020-12-02T15:39:12Z",
"patch_url": "https://github.com/huggingface/datasets/pull/920.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/920"
} | Adding Dream: a Dataset and for Dialogue-Based Reading Comprehension
More details:
https://dataset.org/dream/
https://github.com/nlpdata/dream | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/920/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/920/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/919 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/919/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/919/comments | https://api.github.com/repos/huggingface/datasets/issues/919/events | https://github.com/huggingface/datasets/issues/919 | 753,434,472 | MDU6SXNzdWU3NTM0MzQ0NzI= | 919 | wrong length with datasets | {
"avatar_url": "https://avatars.githubusercontent.com/u/6278280?v=4",
"events_url": "https://api.github.com/users/rabeehk/events{/privacy}",
"followers_url": "https://api.github.com/users/rabeehk/followers",
"following_url": "https://api.github.com/users/rabeehk/following{/other_user}",
"gists_url": "https://api.github.com/users/rabeehk/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/rabeehk",
"id": 6278280,
"login": "rabeehk",
"node_id": "MDQ6VXNlcjYyNzgyODA=",
"organizations_url": "https://api.github.com/users/rabeehk/orgs",
"received_events_url": "https://api.github.com/users/rabeehk/received_events",
"repos_url": "https://api.github.com/users/rabeehk/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/rabeehk/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/rabeehk/subscriptions",
"type": "User",
"url": "https://api.github.com/users/rabeehk"
} | [] | closed | false | null | [] | null | 2 | "2020-11-30T12:23:39Z" | "2020-11-30T12:37:27Z" | "2020-11-30T12:37:26Z" | CONTRIBUTOR | null | null | null | Hi
I have a MRPC dataset which I convert it to seq2seq format, then this is of this format:
`Dataset(features: {'src_texts': Value(dtype='string', id=None), 'tgt_texts': Value(dtype='string', id=None)}, num_rows: 10)
`
I feed it to a dataloader:
```
dataloader = DataLoader(
train_dataset,
batch_size=self.args.train_batch_size,
sampler=train_sampler,
collate_fn=self.data_collator,
drop_last=self.args.dataloader_drop_last,
num_workers=self.args.dataloader_num_workers,
)
```
now if I type len(dataloader) this is 1, which is wrong, and this needs to be 10. could you assist me please? thanks
| {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/919/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/919/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/918 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/918/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/918/comments | https://api.github.com/repos/huggingface/datasets/issues/918/events | https://github.com/huggingface/datasets/pull/918 | 753,397,440 | MDExOlB1bGxSZXF1ZXN0NTI5NDgzOTk4 | 918 | Add conll2002 | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq"
} | [] | closed | false | null | [] | null | 0 | "2020-11-30T11:29:35Z" | "2020-11-30T18:34:30Z" | "2020-11-30T18:34:29Z" | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/918.diff",
"html_url": "https://github.com/huggingface/datasets/pull/918",
"merged_at": "2020-11-30T18:34:29Z",
"patch_url": "https://github.com/huggingface/datasets/pull/918.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/918"
} | Adding the Conll2002 dataset for NER.
More info here : https://www.clips.uantwerpen.be/conll2002/ner/
### Checkbox
- [x] Create the dataset script `/datasets/my_dataset/my_dataset.py` using the template
- [x] Fill the `_DESCRIPTION` and `_CITATION` variables
- [x] Implement `_infos()`, `_split_generators()` and `_generate_examples()`
- [x] Make sure that the `BUILDER_CONFIGS` class attribute is filled with the different configurations of the dataset and that the `BUILDER_CONFIG_CLASS` is specified if there is a custom config class.
- [x] Generate the metadata file `dataset_infos.json` for all configurations
- [x] Generate the dummy data `dummy_data.zip` files to have the dataset script tested and that they don't weigh too much (<50KB)
- [x] Add the dataset card `README.md` using the template : fill the tags and the various paragraphs
- [x] Both tests for the real data and the dummy data pass.
| {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/918/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/918/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/917 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/917/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/917/comments | https://api.github.com/repos/huggingface/datasets/issues/917/events | https://github.com/huggingface/datasets/pull/917 | 753,391,591 | MDExOlB1bGxSZXF1ZXN0NTI5NDc5MTIy | 917 | Addition of Concode Dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/36307201?v=4",
"events_url": "https://api.github.com/users/reshinthadithyan/events{/privacy}",
"followers_url": "https://api.github.com/users/reshinthadithyan/followers",
"following_url": "https://api.github.com/users/reshinthadithyan/following{/other_user}",
"gists_url": "https://api.github.com/users/reshinthadithyan/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/reshinthadithyan",
"id": 36307201,
"login": "reshinthadithyan",
"node_id": "MDQ6VXNlcjM2MzA3MjAx",
"organizations_url": "https://api.github.com/users/reshinthadithyan/orgs",
"received_events_url": "https://api.github.com/users/reshinthadithyan/received_events",
"repos_url": "https://api.github.com/users/reshinthadithyan/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/reshinthadithyan/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/reshinthadithyan/subscriptions",
"type": "User",
"url": "https://api.github.com/users/reshinthadithyan"
} | [] | closed | false | null | [] | null | 8 | "2020-11-30T11:20:59Z" | "2020-12-29T02:55:36Z" | "2020-12-29T02:55:36Z" | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/917.diff",
"html_url": "https://github.com/huggingface/datasets/pull/917",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/917.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/917"
} | ##Overview
Concode Dataset contains pairs of Nl Queries and the corresponding Code.(Contextual Code Generation)
Reference Links
Paper Link = https://arxiv.org/pdf/1904.09086.pdf
Github Link =https://github.com/microsoft/CodeXGLUE/tree/main/Text-Code/text-to-code | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/917/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/917/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/916 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/916/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/916/comments | https://api.github.com/repos/huggingface/datasets/issues/916/events | https://github.com/huggingface/datasets/pull/916 | 753,376,643 | MDExOlB1bGxSZXF1ZXN0NTI5NDY3MTkx | 916 | Add Swedish NER Corpus | {
"avatar_url": "https://avatars.githubusercontent.com/u/1183441?v=4",
"events_url": "https://api.github.com/users/abhishekkrthakur/events{/privacy}",
"followers_url": "https://api.github.com/users/abhishekkrthakur/followers",
"following_url": "https://api.github.com/users/abhishekkrthakur/following{/other_user}",
"gists_url": "https://api.github.com/users/abhishekkrthakur/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/abhishekkrthakur",
"id": 1183441,
"login": "abhishekkrthakur",
"node_id": "MDQ6VXNlcjExODM0NDE=",
"organizations_url": "https://api.github.com/users/abhishekkrthakur/orgs",
"received_events_url": "https://api.github.com/users/abhishekkrthakur/received_events",
"repos_url": "https://api.github.com/users/abhishekkrthakur/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/abhishekkrthakur/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/abhishekkrthakur/subscriptions",
"type": "User",
"url": "https://api.github.com/users/abhishekkrthakur"
} | [] | closed | false | null | [] | null | 2 | "2020-11-30T10:59:51Z" | "2020-12-02T03:10:50Z" | "2020-12-02T03:10:49Z" | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/916.diff",
"html_url": "https://github.com/huggingface/datasets/pull/916",
"merged_at": "2020-12-02T03:10:49Z",
"patch_url": "https://github.com/huggingface/datasets/pull/916.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/916"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/916/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/916/timeline | null | null | true |
|
https://api.github.com/repos/huggingface/datasets/issues/915 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/915/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/915/comments | https://api.github.com/repos/huggingface/datasets/issues/915/events | https://github.com/huggingface/datasets/issues/915 | 753,118,481 | MDU6SXNzdWU3NTMxMTg0ODE= | 915 | Shall we change the hashing to encoding to reduce potential replicated cache files? | {
"avatar_url": "https://avatars.githubusercontent.com/u/10428324?v=4",
"events_url": "https://api.github.com/users/zhuzilin/events{/privacy}",
"followers_url": "https://api.github.com/users/zhuzilin/followers",
"following_url": "https://api.github.com/users/zhuzilin/following{/other_user}",
"gists_url": "https://api.github.com/users/zhuzilin/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/zhuzilin",
"id": 10428324,
"login": "zhuzilin",
"node_id": "MDQ6VXNlcjEwNDI4MzI0",
"organizations_url": "https://api.github.com/users/zhuzilin/orgs",
"received_events_url": "https://api.github.com/users/zhuzilin/received_events",
"repos_url": "https://api.github.com/users/zhuzilin/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/zhuzilin/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/zhuzilin/subscriptions",
"type": "User",
"url": "https://api.github.com/users/zhuzilin"
} | [
{
"color": "a2eeef",
"default": true,
"description": "New feature or request",
"id": 1935892871,
"name": "enhancement",
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement"
},
{
"color": "c5def5",
"default": false,
"description": "Generic discussion on the library",
"id": 2067400324,
"name": "generic discussion",
"node_id": "MDU6TGFiZWwyMDY3NDAwMzI0",
"url": "https://api.github.com/repos/huggingface/datasets/labels/generic%20discussion"
}
] | open | false | null | [] | null | 2 | "2020-11-30T03:50:46Z" | "2020-12-24T05:11:49Z" | null | NONE | null | null | null | Hi there. For now, we are using `xxhash` to hash the transformations to fingerprint and we will save a copy of the processed dataset to disk if there is a new hash value. However, there are some transformations that are idempotent or commutative to each other. I think that encoding the transformation chain as the fingerprint may help in those cases, for example, use `base64.urlsafe_b64encode`. In this way, before we want to save a new copy, we can decode the transformation chain and normalize it to prevent omit potential reuse. As the main targets of this project are the really large datasets that cannot be loaded entirely in memory, I believe it would save a lot of time if we can avoid some write.
If you have interest in this, I'd love to help :). | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/915/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/915/timeline | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/914 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/914/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/914/comments | https://api.github.com/repos/huggingface/datasets/issues/914/events | https://github.com/huggingface/datasets/pull/914 | 752,956,106 | MDExOlB1bGxSZXF1ZXN0NTI5MTM2Njk3 | 914 | Add list_github_datasets api for retrieving dataset name list in github repo | {
"avatar_url": "https://avatars.githubusercontent.com/u/10428324?v=4",
"events_url": "https://api.github.com/users/zhuzilin/events{/privacy}",
"followers_url": "https://api.github.com/users/zhuzilin/followers",
"following_url": "https://api.github.com/users/zhuzilin/following{/other_user}",
"gists_url": "https://api.github.com/users/zhuzilin/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/zhuzilin",
"id": 10428324,
"login": "zhuzilin",
"node_id": "MDQ6VXNlcjEwNDI4MzI0",
"organizations_url": "https://api.github.com/users/zhuzilin/orgs",
"received_events_url": "https://api.github.com/users/zhuzilin/received_events",
"repos_url": "https://api.github.com/users/zhuzilin/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/zhuzilin/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/zhuzilin/subscriptions",
"type": "User",
"url": "https://api.github.com/users/zhuzilin"
} | [] | closed | false | null | [] | null | 4 | "2020-11-29T16:42:15Z" | "2020-12-02T07:21:16Z" | "2020-12-02T07:21:16Z" | NONE | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/914.diff",
"html_url": "https://github.com/huggingface/datasets/pull/914",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/914.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/914"
} | Thank you for your great effort on unifying data processing for NLP!
This pr is trying to add a new api `list_github_datasets` in the `inspect` module. The reason for it is that the current `list_datasets` api need to access https://huggingface.co/api/datasets to get a large json. However, this connection can be really slow... (I was visiting from China) and from my own experience, most of the time `requests.get` failed to download the whole json after a long wait and will trigger fault in `r.json()`.
I also noticed that the current implementation will first try to download from github, which makes me be able to smoothly run `load_dataset('squad')` in the example.
Therefore, I think it would be better if we can have an api to get the list of datasets that are available on github, and it will also improve newcomers' experience (it is a little frustrating if one cannot successfully run the first function in the README example.) before we have faster source for huggingface.co.
As for the implementation, I've added a `dataset_infos.json` file under the `datasets` folder, and it has the following structure:
```json
{
"id": "aeslc",
"folder": "datasets/aeslc",
"dataset_infos": "datasets/aeslc/dataset_infos.json"
},
...
{
"id": "json",
"folder": "datasets/json"
},
...
```
The script I used to get this file is:
```python
import json
import os
DATASETS_BASE_DIR = "/root/datasets"
DATASET_INFOS_JSON = "dataset_infos.json"
datasets = []
for item in os.listdir(os.path.join(DATASETS_BASE_DIR, "datasets")):
if os.path.isdir(os.path.join(DATASETS_BASE_DIR, "datasets", item)):
datasets.append(item)
datasets.sort()
total_ds_info = []
for ds in datasets:
ds_dir = os.path.join("datasets", ds)
ds_info_dir = os.path.join(ds_dir, DATASET_INFOS_JSON)
if os.path.isfile(os.path.join(DATASETS_BASE_DIR, ds_info_dir)):
total_ds_info.append({"id": ds,
"folder": ds_dir,
"dataset_infos": ds_info_dir})
else:
total_ds_info.append({"id": ds,
"folder": ds_dir})
with open(DATASET_INFOS_JSON, "w") as f:
json.dump(total_ds_info, f)
```
The new `dataset_infos.json` was saved as a formated json so that it will be easy to add new dataset.
When calling `list_github_datasets`, the user will get the list of dataset names in this github repo and if `with_details` is set to be `True`, they can get the url of specific dataset info.
Thank you for your time on reviewing this pr :). | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/914/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/914/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/913 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/913/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/913/comments | https://api.github.com/repos/huggingface/datasets/issues/913/events | https://github.com/huggingface/datasets/pull/913 | 752,892,020 | MDExOlB1bGxSZXF1ZXN0NTI5MDkyOTc3 | 913 | My new dataset PEC | {
"avatar_url": "https://avatars.githubusercontent.com/u/11826803?v=4",
"events_url": "https://api.github.com/users/zhongpeixiang/events{/privacy}",
"followers_url": "https://api.github.com/users/zhongpeixiang/followers",
"following_url": "https://api.github.com/users/zhongpeixiang/following{/other_user}",
"gists_url": "https://api.github.com/users/zhongpeixiang/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/zhongpeixiang",
"id": 11826803,
"login": "zhongpeixiang",
"node_id": "MDQ6VXNlcjExODI2ODAz",
"organizations_url": "https://api.github.com/users/zhongpeixiang/orgs",
"received_events_url": "https://api.github.com/users/zhongpeixiang/received_events",
"repos_url": "https://api.github.com/users/zhongpeixiang/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/zhongpeixiang/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/zhongpeixiang/subscriptions",
"type": "User",
"url": "https://api.github.com/users/zhongpeixiang"
} | [] | closed | false | null | [] | null | 6 | "2020-11-29T11:10:37Z" | "2020-12-01T10:41:53Z" | "2020-12-01T10:41:53Z" | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/913.diff",
"html_url": "https://github.com/huggingface/datasets/pull/913",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/913.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/913"
} | A new dataset PEC published in EMNLP 2020. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/913/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/913/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/911 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/911/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/911/comments | https://api.github.com/repos/huggingface/datasets/issues/911/events | https://github.com/huggingface/datasets/issues/911 | 752,806,215 | MDU6SXNzdWU3NTI4MDYyMTU= | 911 | datasets module not found | {
"avatar_url": "https://avatars.githubusercontent.com/u/15836274?v=4",
"events_url": "https://api.github.com/users/sbassam/events{/privacy}",
"followers_url": "https://api.github.com/users/sbassam/followers",
"following_url": "https://api.github.com/users/sbassam/following{/other_user}",
"gists_url": "https://api.github.com/users/sbassam/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/sbassam",
"id": 15836274,
"login": "sbassam",
"node_id": "MDQ6VXNlcjE1ODM2Mjc0",
"organizations_url": "https://api.github.com/users/sbassam/orgs",
"received_events_url": "https://api.github.com/users/sbassam/received_events",
"repos_url": "https://api.github.com/users/sbassam/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/sbassam/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/sbassam/subscriptions",
"type": "User",
"url": "https://api.github.com/users/sbassam"
} | [] | closed | false | null | [] | null | 1 | "2020-11-29T01:24:15Z" | "2020-11-29T14:33:09Z" | "2020-11-29T14:33:09Z" | NONE | null | null | null | Currently, running `from datasets import load_dataset` will throw a `ModuleNotFoundError: No module named 'datasets'` error.
| {
"+1": 1,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/911/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/911/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/910 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/910/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/910/comments | https://api.github.com/repos/huggingface/datasets/issues/910/events | https://github.com/huggingface/datasets/issues/910 | 752,772,723 | MDU6SXNzdWU3NTI3NzI3MjM= | 910 | Grindr meeting app web.Grindr | {
"avatar_url": "https://avatars.githubusercontent.com/u/75184749?v=4",
"events_url": "https://api.github.com/users/jackin34/events{/privacy}",
"followers_url": "https://api.github.com/users/jackin34/followers",
"following_url": "https://api.github.com/users/jackin34/following{/other_user}",
"gists_url": "https://api.github.com/users/jackin34/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/jackin34",
"id": 75184749,
"login": "jackin34",
"node_id": "MDQ6VXNlcjc1MTg0NzQ5",
"organizations_url": "https://api.github.com/users/jackin34/orgs",
"received_events_url": "https://api.github.com/users/jackin34/received_events",
"repos_url": "https://api.github.com/users/jackin34/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/jackin34/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jackin34/subscriptions",
"type": "User",
"url": "https://api.github.com/users/jackin34"
} | [] | closed | false | null | [] | null | 0 | "2020-11-28T21:36:23Z" | "2020-11-29T10:11:51Z" | "2020-11-29T10:11:51Z" | NONE | null | null | null | ## Adding a Dataset
- **Name:** *name of the dataset*
- **Description:** *short description of the dataset (or link to social media or blog post)*
- **Paper:** *link to the dataset paper if available*
- **Data:** *link to the Github repository or current dataset location*
- **Motivation:** *what are some good reasons to have this dataset*
Instructions to add a new dataset can be found [here](https://huggingface.co/docs/datasets/share_dataset.html). | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/910/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/910/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/909 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/909/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/909/comments | https://api.github.com/repos/huggingface/datasets/issues/909/events | https://github.com/huggingface/datasets/pull/909 | 752,508,299 | MDExOlB1bGxSZXF1ZXN0NTI4ODE1NDYz | 909 | Add FiNER dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/20651387?v=4",
"events_url": "https://api.github.com/users/stefan-it/events{/privacy}",
"followers_url": "https://api.github.com/users/stefan-it/followers",
"following_url": "https://api.github.com/users/stefan-it/following{/other_user}",
"gists_url": "https://api.github.com/users/stefan-it/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/stefan-it",
"id": 20651387,
"login": "stefan-it",
"node_id": "MDQ6VXNlcjIwNjUxMzg3",
"organizations_url": "https://api.github.com/users/stefan-it/orgs",
"received_events_url": "https://api.github.com/users/stefan-it/received_events",
"repos_url": "https://api.github.com/users/stefan-it/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/stefan-it/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/stefan-it/subscriptions",
"type": "User",
"url": "https://api.github.com/users/stefan-it"
} | [] | closed | false | null | [] | null | 9 | "2020-11-27T23:54:20Z" | "2020-12-07T16:56:23Z" | "2020-12-07T16:56:23Z" | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/909.diff",
"html_url": "https://github.com/huggingface/datasets/pull/909",
"merged_at": "2020-12-07T16:56:23Z",
"patch_url": "https://github.com/huggingface/datasets/pull/909.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/909"
} | Hi,
this PR adds "A Finnish News Corpus for Named Entity Recognition" as new `finer` dataset.
The dataset is described in [this paper](https://arxiv.org/abs/1908.04212). The data is publicly available in [this GitHub](https://github.com/mpsilfve/finer-data).
Notice: they provide two testsets. The additional test dataset taken from Wikipedia is named as "test_wikipedia" split. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/909/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/909/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/908 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/908/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/908/comments | https://api.github.com/repos/huggingface/datasets/issues/908/events | https://github.com/huggingface/datasets/pull/908 | 752,428,652 | MDExOlB1bGxSZXF1ZXN0NTI4NzUzMjcz | 908 | Add dependency on black for tests | {
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",
"gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/albertvillanova",
"id": 8515462,
"login": "albertvillanova",
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"organizations_url": "https://api.github.com/users/albertvillanova/orgs",
"received_events_url": "https://api.github.com/users/albertvillanova/received_events",
"repos_url": "https://api.github.com/users/albertvillanova/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions",
"type": "User",
"url": "https://api.github.com/users/albertvillanova"
} | [] | closed | false | null | [] | null | 1 | "2020-11-27T19:12:48Z" | "2020-11-27T21:46:53Z" | "2020-11-27T21:46:52Z" | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/908.diff",
"html_url": "https://github.com/huggingface/datasets/pull/908",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/908.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/908"
} | Add package 'black' as an installation requirement for tests. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/908/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/908/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/907 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/907/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/907/comments | https://api.github.com/repos/huggingface/datasets/issues/907/events | https://github.com/huggingface/datasets/pull/907 | 752,422,351 | MDExOlB1bGxSZXF1ZXN0NTI4NzQ4ODMx | 907 | Remove os.path.join from all URLs | {
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",
"gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/albertvillanova",
"id": 8515462,
"login": "albertvillanova",
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"organizations_url": "https://api.github.com/users/albertvillanova/orgs",
"received_events_url": "https://api.github.com/users/albertvillanova/received_events",
"repos_url": "https://api.github.com/users/albertvillanova/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions",
"type": "User",
"url": "https://api.github.com/users/albertvillanova"
} | [] | closed | false | null | [] | null | 0 | "2020-11-27T18:55:30Z" | "2020-11-29T22:48:20Z" | "2020-11-29T22:48:19Z" | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/907.diff",
"html_url": "https://github.com/huggingface/datasets/pull/907",
"merged_at": "2020-11-29T22:48:19Z",
"patch_url": "https://github.com/huggingface/datasets/pull/907.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/907"
} | Remove `os.path.join` from all URLs in dataset scripts. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 1,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/907/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/907/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/906 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/906/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/906/comments | https://api.github.com/repos/huggingface/datasets/issues/906/events | https://github.com/huggingface/datasets/pull/906 | 752,403,395 | MDExOlB1bGxSZXF1ZXN0NTI4NzM0MDY0 | 906 | Fix url with backslash in windows for blimp and pg19 | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq"
} | [] | closed | false | null | [] | null | 0 | "2020-11-27T17:59:11Z" | "2020-11-27T18:19:56Z" | "2020-11-27T18:19:56Z" | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/906.diff",
"html_url": "https://github.com/huggingface/datasets/pull/906",
"merged_at": "2020-11-27T18:19:55Z",
"patch_url": "https://github.com/huggingface/datasets/pull/906.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/906"
} | Following #903 I also fixed blimp and pg19 which were using the `os.path.join` to create urls
cc @albertvillanova | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/906/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/906/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/905 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/905/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/905/comments | https://api.github.com/repos/huggingface/datasets/issues/905/events | https://github.com/huggingface/datasets/pull/905 | 752,395,456 | MDExOlB1bGxSZXF1ZXN0NTI4NzI3OTEy | 905 | Disallow backslash in urls | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq"
} | [] | closed | false | null | [] | null | 2 | "2020-11-27T17:38:28Z" | "2020-11-29T22:48:37Z" | "2020-11-29T22:48:36Z" | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/905.diff",
"html_url": "https://github.com/huggingface/datasets/pull/905",
"merged_at": "2020-11-29T22:48:36Z",
"patch_url": "https://github.com/huggingface/datasets/pull/905.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/905"
} | Following #903 @albertvillanova noticed that there are sometimes bad usage of `os.path.join` in datasets scripts to create URLS. However this should be avoided since it doesn't work on windows.
I'm suggesting a test to make sure we that all the urls don't have backslashes in them in the datasets scripts.
The tests works by adding a callback feature to the MockDownloadManager used to test the dataset scripts. In a download callback I just make sure that the url is valid. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/905/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/905/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/904 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/904/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/904/comments | https://api.github.com/repos/huggingface/datasets/issues/904/events | https://github.com/huggingface/datasets/pull/904 | 752,372,743 | MDExOlB1bGxSZXF1ZXN0NTI4NzA5NTUx | 904 | Very detailed step-by-step on how to add a dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4",
"events_url": "https://api.github.com/users/thomwolf/events{/privacy}",
"followers_url": "https://api.github.com/users/thomwolf/followers",
"following_url": "https://api.github.com/users/thomwolf/following{/other_user}",
"gists_url": "https://api.github.com/users/thomwolf/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/thomwolf",
"id": 7353373,
"login": "thomwolf",
"node_id": "MDQ6VXNlcjczNTMzNzM=",
"organizations_url": "https://api.github.com/users/thomwolf/orgs",
"received_events_url": "https://api.github.com/users/thomwolf/received_events",
"repos_url": "https://api.github.com/users/thomwolf/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/thomwolf/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/thomwolf/subscriptions",
"type": "User",
"url": "https://api.github.com/users/thomwolf"
} | [] | closed | false | null | [] | null | 1 | "2020-11-27T16:45:21Z" | "2020-11-30T09:56:27Z" | "2020-11-30T09:56:26Z" | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/904.diff",
"html_url": "https://github.com/huggingface/datasets/pull/904",
"merged_at": "2020-11-30T09:56:26Z",
"patch_url": "https://github.com/huggingface/datasets/pull/904.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/904"
} | Add very detailed step-by-step instructions to add a new dataset to the library. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/904/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/904/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/903 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/903/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/903/comments | https://api.github.com/repos/huggingface/datasets/issues/903/events | https://github.com/huggingface/datasets/pull/903 | 752,360,614 | MDExOlB1bGxSZXF1ZXN0NTI4Njk5NDQ3 | 903 | Fix URL with backslash in Windows | {
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",
"gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/albertvillanova",
"id": 8515462,
"login": "albertvillanova",
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"organizations_url": "https://api.github.com/users/albertvillanova/orgs",
"received_events_url": "https://api.github.com/users/albertvillanova/received_events",
"repos_url": "https://api.github.com/users/albertvillanova/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions",
"type": "User",
"url": "https://api.github.com/users/albertvillanova"
} | [] | closed | false | null | [] | null | 8 | "2020-11-27T16:26:24Z" | "2020-11-27T18:04:46Z" | "2020-11-27T18:04:46Z" | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/903.diff",
"html_url": "https://github.com/huggingface/datasets/pull/903",
"merged_at": "2020-11-27T18:04:46Z",
"patch_url": "https://github.com/huggingface/datasets/pull/903.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/903"
} | In Windows, `os.path.join` generates URLs containing backslashes, when the first "path" does not end with a slash.
In general, `os.path.join` should be avoided to generate URLs. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/903/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/903/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/902 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/902/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/902/comments | https://api.github.com/repos/huggingface/datasets/issues/902/events | https://github.com/huggingface/datasets/pull/902 | 752,345,739 | MDExOlB1bGxSZXF1ZXN0NTI4Njg3NTYw | 902 | Follow cache_dir parameter to gcs downloader | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq"
} | [] | closed | false | null | [] | null | 0 | "2020-11-27T16:02:06Z" | "2020-11-29T22:48:54Z" | "2020-11-29T22:48:53Z" | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/902.diff",
"html_url": "https://github.com/huggingface/datasets/pull/902",
"merged_at": "2020-11-29T22:48:53Z",
"patch_url": "https://github.com/huggingface/datasets/pull/902.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/902"
} | As noticed in #900 the cache_dir parameter was not followed to the downloader in the case of an already processed dataset hosted on our google storage (one of them is natural questions).
Fix #900 | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/902/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/902/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/901 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/901/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/901/comments | https://api.github.com/repos/huggingface/datasets/issues/901/events | https://github.com/huggingface/datasets/pull/901 | 752,233,851 | MDExOlB1bGxSZXF1ZXN0NTI4NTk3NDU5 | 901 | Addition of Nl2Bash Dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/36307201?v=4",
"events_url": "https://api.github.com/users/reshinthadithyan/events{/privacy}",
"followers_url": "https://api.github.com/users/reshinthadithyan/followers",
"following_url": "https://api.github.com/users/reshinthadithyan/following{/other_user}",
"gists_url": "https://api.github.com/users/reshinthadithyan/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/reshinthadithyan",
"id": 36307201,
"login": "reshinthadithyan",
"node_id": "MDQ6VXNlcjM2MzA3MjAx",
"organizations_url": "https://api.github.com/users/reshinthadithyan/orgs",
"received_events_url": "https://api.github.com/users/reshinthadithyan/received_events",
"repos_url": "https://api.github.com/users/reshinthadithyan/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/reshinthadithyan/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/reshinthadithyan/subscriptions",
"type": "User",
"url": "https://api.github.com/users/reshinthadithyan"
} | [] | closed | false | null | [] | null | 3 | "2020-11-27T12:53:55Z" | "2020-11-29T18:09:25Z" | "2020-11-29T18:08:51Z" | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/901.diff",
"html_url": "https://github.com/huggingface/datasets/pull/901",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/901.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/901"
} | ## Overview
The NL2Bash data contains over 10,000 instances of linux shell commands and their corresponding natural language descriptions provided by experts, from the Tellina system. The dataset features 100+ commonly used shell utilities.
## Footnotes
The following dataset marks the first ML on source code related Dataset in datasets module. It'll be really useful as a lot of the research direction involves Transformer Based Model.
Thanks.
### Reference Links
> Paper Link = https://arxiv.org/pdf/1802.08979.pdf
> Github Link = https://github.com/TellinaTool/nl2bash
| {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/901/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/901/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/900 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/900/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/900/comments | https://api.github.com/repos/huggingface/datasets/issues/900/events | https://github.com/huggingface/datasets/issues/900 | 752,214,066 | MDU6SXNzdWU3NTIyMTQwNjY= | 900 | datasets.load_dataset() custom chaching directory bug | {
"avatar_url": "https://avatars.githubusercontent.com/u/44585792?v=4",
"events_url": "https://api.github.com/users/SapirWeissbuch/events{/privacy}",
"followers_url": "https://api.github.com/users/SapirWeissbuch/followers",
"following_url": "https://api.github.com/users/SapirWeissbuch/following{/other_user}",
"gists_url": "https://api.github.com/users/SapirWeissbuch/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/SapirWeissbuch",
"id": 44585792,
"login": "SapirWeissbuch",
"node_id": "MDQ6VXNlcjQ0NTg1Nzky",
"organizations_url": "https://api.github.com/users/SapirWeissbuch/orgs",
"received_events_url": "https://api.github.com/users/SapirWeissbuch/received_events",
"repos_url": "https://api.github.com/users/SapirWeissbuch/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/SapirWeissbuch/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/SapirWeissbuch/subscriptions",
"type": "User",
"url": "https://api.github.com/users/SapirWeissbuch"
} | [] | closed | false | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq"
} | [
{
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq"
}
] | null | 1 | "2020-11-27T12:18:53Z" | "2020-11-29T22:48:53Z" | "2020-11-29T22:48:53Z" | NONE | null | null | null | Hello,
I'm having issue with loading a dataset with a custom `cache_dir`. Despite specifying the output dir, it is still downloaded to
`~/.cache`.
## Environment info
- `datasets` version: 1.1.3
- Platform: Linux-4.19.129-aufs-1-x86_64-with-debian-10.1
- Python version: 3.7.3
## The code I'm running:
```python
import datasets
from pathlib import Path
validation_dataset = datasets.load_dataset("natural_questions", split="validation[:5%]", cache_dir=Path("./data"))
```
## The output:
* The dataset is downloaded to my home directory's `.cache`
* A new empty directory named "`natural_questions` is created in the specified directory `.data`
* `tree data` in the shell outputs:
```
data
└── natural_questions
└── default
└── 0.0.2
3 directories, 0 files
```
The output:
```
Downloading: 8.61kB [00:00, 5.11MB/s]
Downloading: 13.6kB [00:00, 7.89MB/s]
Using custom data configuration default
Downloading and preparing dataset natural_questions/default (download: 41.97 GiB, generated: 92.95 GiB, post-processed: Unknown size, total: 134.92 GiB) to ./data/natural_questions/default/0.0.2/867dbbaf9137c1b8
3ecb19f5eb80559e1002ea26e702c6b919cfa81a17a8c531...
Downloading: 100%|██████████████████████████████████████████████████| 13.6k/13.6k [00:00<00:00, 1.51MB/s]
Downloading: 7%|███▎ | 6.70G/97.4G [03:46<1:37:05, 15.6MB/s]
```
## Expected behaviour:
The dataset "Natural Questions" should be downloaded to the directory "./data"
| {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/900/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/900/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/899 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/899/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/899/comments | https://api.github.com/repos/huggingface/datasets/issues/899/events | https://github.com/huggingface/datasets/pull/899 | 752,191,227 | MDExOlB1bGxSZXF1ZXN0NTI4NTYzNzYz | 899 | Allow arrow based builder in auto dummy data generation | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq"
} | [] | closed | false | null | [] | null | 0 | "2020-11-27T11:39:38Z" | "2020-11-27T13:30:09Z" | "2020-11-27T13:30:08Z" | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/899.diff",
"html_url": "https://github.com/huggingface/datasets/pull/899",
"merged_at": "2020-11-27T13:30:08Z",
"patch_url": "https://github.com/huggingface/datasets/pull/899.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/899"
} | Following #898 I added support for arrow based builder for the auto dummy data generator | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/899/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/899/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/898 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/898/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/898/comments | https://api.github.com/repos/huggingface/datasets/issues/898/events | https://github.com/huggingface/datasets/pull/898 | 752,148,284 | MDExOlB1bGxSZXF1ZXN0NTI4NTI4MDY1 | 898 | Adding SQA dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4",
"events_url": "https://api.github.com/users/thomwolf/events{/privacy}",
"followers_url": "https://api.github.com/users/thomwolf/followers",
"following_url": "https://api.github.com/users/thomwolf/following{/other_user}",
"gists_url": "https://api.github.com/users/thomwolf/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/thomwolf",
"id": 7353373,
"login": "thomwolf",
"node_id": "MDQ6VXNlcjczNTMzNzM=",
"organizations_url": "https://api.github.com/users/thomwolf/orgs",
"received_events_url": "https://api.github.com/users/thomwolf/received_events",
"repos_url": "https://api.github.com/users/thomwolf/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/thomwolf/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/thomwolf/subscriptions",
"type": "User",
"url": "https://api.github.com/users/thomwolf"
} | [] | closed | false | null | [] | null | 2 | "2020-11-27T10:29:18Z" | "2020-12-15T12:54:40Z" | "2020-12-15T12:54:19Z" | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/898.diff",
"html_url": "https://github.com/huggingface/datasets/pull/898",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/898.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/898"
} | As discussed in #880
Seems like automatic dummy-data generation doesn't work if the builder is a `ArrowBasedBuilder`, do you think you could take a look @lhoestq ? | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/898/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/898/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/897 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/897/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/897/comments | https://api.github.com/repos/huggingface/datasets/issues/897/events | https://github.com/huggingface/datasets/issues/897 | 752,100,256 | MDU6SXNzdWU3NTIxMDAyNTY= | 897 | Dataset viewer issues | {
"avatar_url": "https://avatars.githubusercontent.com/u/2779410?v=4",
"events_url": "https://api.github.com/users/BramVanroy/events{/privacy}",
"followers_url": "https://api.github.com/users/BramVanroy/followers",
"following_url": "https://api.github.com/users/BramVanroy/following{/other_user}",
"gists_url": "https://api.github.com/users/BramVanroy/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/BramVanroy",
"id": 2779410,
"login": "BramVanroy",
"node_id": "MDQ6VXNlcjI3Nzk0MTA=",
"organizations_url": "https://api.github.com/users/BramVanroy/orgs",
"received_events_url": "https://api.github.com/users/BramVanroy/received_events",
"repos_url": "https://api.github.com/users/BramVanroy/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/BramVanroy/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/BramVanroy/subscriptions",
"type": "User",
"url": "https://api.github.com/users/BramVanroy"
} | [
{
"color": "94203D",
"default": false,
"description": "",
"id": 2107841032,
"name": "nlp-viewer",
"node_id": "MDU6TGFiZWwyMTA3ODQxMDMy",
"url": "https://api.github.com/repos/huggingface/datasets/labels/nlp-viewer"
}
] | closed | false | null | [] | null | 5 | "2020-11-27T09:14:34Z" | "2021-10-31T09:12:01Z" | "2021-10-31T09:12:01Z" | CONTRIBUTOR | null | null | null | I was looking through the dataset viewer and I like it a lot. Version numbers, citation information, everything's there! I've spotted a few issues/bugs though:
- the URL is still under `nlp`, perhaps an alias for `datasets` can be made
- when I remove a **feature** (and the feature list is empty), I get an error. This is probably expected, but perhaps a better error message can be shown to the user
```bash
IndexError: list index out of range
Traceback:
File "/home/sasha/streamlit/lib/streamlit/ScriptRunner.py", line 322, in _run_script
exec(code, module.__dict__)
File "/home/sasha/nlp-viewer/run.py", line 316, in <module>
st.table(style)
File "/home/sasha/streamlit/lib/streamlit/DeltaGenerator.py", line 122, in wrapped_method
return dg._enqueue_new_element_delta(marshall_element, delta_type, last_index)
File "/home/sasha/streamlit/lib/streamlit/DeltaGenerator.py", line 367, in _enqueue_new_element_delta
rv = marshall_element(msg.delta.new_element)
File "/home/sasha/streamlit/lib/streamlit/DeltaGenerator.py", line 120, in marshall_element
return method(dg, element, *args, **kwargs)
File "/home/sasha/streamlit/lib/streamlit/DeltaGenerator.py", line 2944, in table
data_frame_proto.marshall_data_frame(data, element.table)
File "/home/sasha/streamlit/lib/streamlit/elements/data_frame_proto.py", line 54, in marshall_data_frame
_marshall_styles(proto_df.style, df, styler)
File "/home/sasha/streamlit/lib/streamlit/elements/data_frame_proto.py", line 73, in _marshall_styles
translated_style = styler._translate()
File "/home/sasha/.local/share/virtualenvs/lib-ogGKnCK_/lib/python3.7/site-packages/pandas/io/formats/style.py", line 351, in _translate
* (len(clabels[0]) - len(hidden_columns))
```
- there seems to be **an encoding issue** in the default view, the dataset examples are shown as raw monospace text, without a decent encoding. That makes it hard to read for languages that use a lot of special characters. Take for instance the [cs-en WMT19 set](https://huggingface.co/nlp/viewer/?dataset=wmt19&config=cs-en). This problem goes away when you enable "List view", because then some syntax highlighteris used, and the special characters are coded correctly.
| {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/897/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/897/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/896 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/896/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/896/comments | https://api.github.com/repos/huggingface/datasets/issues/896/events | https://github.com/huggingface/datasets/pull/896 | 751,834,265 | MDExOlB1bGxSZXF1ZXN0NTI4MjcyMjc0 | 896 | Add template and documentation for dataset card | {
"avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4",
"events_url": "https://api.github.com/users/yjernite/events{/privacy}",
"followers_url": "https://api.github.com/users/yjernite/followers",
"following_url": "https://api.github.com/users/yjernite/following{/other_user}",
"gists_url": "https://api.github.com/users/yjernite/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/yjernite",
"id": 10469459,
"login": "yjernite",
"node_id": "MDQ6VXNlcjEwNDY5NDU5",
"organizations_url": "https://api.github.com/users/yjernite/orgs",
"received_events_url": "https://api.github.com/users/yjernite/received_events",
"repos_url": "https://api.github.com/users/yjernite/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/yjernite/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/yjernite/subscriptions",
"type": "User",
"url": "https://api.github.com/users/yjernite"
} | [] | closed | false | null | [] | null | 0 | "2020-11-26T21:30:25Z" | "2020-11-28T01:10:15Z" | "2020-11-28T01:10:15Z" | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/896.diff",
"html_url": "https://github.com/huggingface/datasets/pull/896",
"merged_at": "2020-11-28T01:10:14Z",
"patch_url": "https://github.com/huggingface/datasets/pull/896.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/896"
} | This PR adds a template for dataset cards, as well as a guide to filling out the template and a completed example for the ELI5 dataset, building on the work of @mcmillanmajora
New pull requests adding datasets should now have a README.md file which serves both to hold the tags we will have to index the datasets and as a data statement.
The template is designed to be pretty extensive. The idea is that the person who uploads the dataset should put in all the basic information (at least the Dataset Description section) and whatever else they feel comfortable adding and leave the `[More Information Needed]` annotation everywhere else as a placeholder.
We will then work with @mcmillanmajora to involve the data authors more directly in filling out the remaining information.
Direct links to:
- [Documentation](https://github.com/yjernite/datasets/blob/add_dataset_card_doc/templates/README_guide.md)
- [Empty template](https://github.com/yjernite/datasets/blob/add_dataset_card_doc/templates/README.md)
- [ELI5 example](https://github.com/yjernite/datasets/blob/add_dataset_card_doc/datasets/eli5/README.md) | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/896/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/896/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/895 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/895/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/895/comments | https://api.github.com/repos/huggingface/datasets/issues/895/events | https://github.com/huggingface/datasets/pull/895 | 751,782,295 | MDExOlB1bGxSZXF1ZXN0NTI4MjMyMjU3 | 895 | Better messages regarding split naming | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq"
} | [] | closed | false | null | [] | null | 0 | "2020-11-26T18:55:46Z" | "2020-11-27T13:31:00Z" | "2020-11-27T13:30:59Z" | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/895.diff",
"html_url": "https://github.com/huggingface/datasets/pull/895",
"merged_at": "2020-11-27T13:30:59Z",
"patch_url": "https://github.com/huggingface/datasets/pull/895.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/895"
} | I made explicit the error message when a bad split name is used.
Also I wanted to allow the `-` symbol for split names but actually this symbol is used to name the arrow files `{dataset_name}-{dataset_split}.arrow` so we should probably keep it this way, i.e. not allowing the `-` symbol in split names. Moreover in the future we might want to use `{dataset_name}-{dataset_split}-{shard_id}_of_{n_shards}.arrow` and reuse the `-` symbol. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/895/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/895/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/894 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/894/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/894/comments | https://api.github.com/repos/huggingface/datasets/issues/894/events | https://github.com/huggingface/datasets/pull/894 | 751,734,905 | MDExOlB1bGxSZXF1ZXN0NTI4MTkzNzQy | 894 | Allow several tags sets | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq"
} | [] | closed | false | null | [] | null | 1 | "2020-11-26T17:04:13Z" | "2021-05-05T18:24:17Z" | "2020-11-27T20:15:49Z" | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/894.diff",
"html_url": "https://github.com/huggingface/datasets/pull/894",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/894.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/894"
} | Hi !
Currently we have three dataset cards : snli, cnn_dailymail and allocine.
For each one of those datasets a set of tag is defined. The set of tags contains fields like `multilinguality`, `task_ids`, `licenses` etc.
For certain datasets like `glue` for example, there exist several configurations: `sst2`, `mnli` etc. Therefore we should define one set of tags per configuration. However the current format used for tags only supports one set of tags per dataset.
In this PR I propose a simple change in the yaml format used for tags to allow for several sets of tags.
Let me know what you think, especially @julien-c let me know if it's good for you since it's going to be parsed by moon-landing | {
"+1": 1,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/894/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/894/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/893 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/893/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/893/comments | https://api.github.com/repos/huggingface/datasets/issues/893/events | https://github.com/huggingface/datasets/pull/893 | 751,703,696 | MDExOlB1bGxSZXF1ZXN0NTI4MTY4NDgx | 893 | add metrec: arabic poetry dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/15667714?v=4",
"events_url": "https://api.github.com/users/zaidalyafeai/events{/privacy}",
"followers_url": "https://api.github.com/users/zaidalyafeai/followers",
"following_url": "https://api.github.com/users/zaidalyafeai/following{/other_user}",
"gists_url": "https://api.github.com/users/zaidalyafeai/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/zaidalyafeai",
"id": 15667714,
"login": "zaidalyafeai",
"node_id": "MDQ6VXNlcjE1NjY3NzE0",
"organizations_url": "https://api.github.com/users/zaidalyafeai/orgs",
"received_events_url": "https://api.github.com/users/zaidalyafeai/received_events",
"repos_url": "https://api.github.com/users/zaidalyafeai/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/zaidalyafeai/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/zaidalyafeai/subscriptions",
"type": "User",
"url": "https://api.github.com/users/zaidalyafeai"
} | [] | closed | false | null | [] | null | 10 | "2020-11-26T16:10:16Z" | "2020-12-01T16:24:55Z" | "2020-12-01T15:15:07Z" | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/893.diff",
"html_url": "https://github.com/huggingface/datasets/pull/893",
"merged_at": "2020-12-01T15:15:07Z",
"patch_url": "https://github.com/huggingface/datasets/pull/893.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/893"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/893/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/893/timeline | null | null | true |
|
https://api.github.com/repos/huggingface/datasets/issues/892 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/892/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/892/comments | https://api.github.com/repos/huggingface/datasets/issues/892/events | https://github.com/huggingface/datasets/pull/892 | 751,658,262 | MDExOlB1bGxSZXF1ZXN0NTI4MTMxNTE1 | 892 | Add a few datasets of reference in the documentation | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq"
} | [] | closed | false | null | [] | null | 3 | "2020-11-26T15:02:39Z" | "2020-11-27T18:08:45Z" | "2020-11-27T18:08:44Z" | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/892.diff",
"html_url": "https://github.com/huggingface/datasets/pull/892",
"merged_at": "2020-11-27T18:08:44Z",
"patch_url": "https://github.com/huggingface/datasets/pull/892.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/892"
} | I started making a small list of various datasets of reference in the documentation.
Since many datasets share a lot in common I think it's good to have a list of datasets scripts to get some inspiration from.
Let me know what you think, and if you have ideas of other datasets that we may add to this list, please let me know. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/892/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/892/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/891 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/891/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/891/comments | https://api.github.com/repos/huggingface/datasets/issues/891/events | https://github.com/huggingface/datasets/pull/891 | 751,576,869 | MDExOlB1bGxSZXF1ZXN0NTI4MDY1MTQ3 | 891 | gitignore .python-version | {
"avatar_url": "https://avatars.githubusercontent.com/u/27137566?v=4",
"events_url": "https://api.github.com/users/patil-suraj/events{/privacy}",
"followers_url": "https://api.github.com/users/patil-suraj/followers",
"following_url": "https://api.github.com/users/patil-suraj/following{/other_user}",
"gists_url": "https://api.github.com/users/patil-suraj/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/patil-suraj",
"id": 27137566,
"login": "patil-suraj",
"node_id": "MDQ6VXNlcjI3MTM3NTY2",
"organizations_url": "https://api.github.com/users/patil-suraj/orgs",
"received_events_url": "https://api.github.com/users/patil-suraj/received_events",
"repos_url": "https://api.github.com/users/patil-suraj/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/patil-suraj/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/patil-suraj/subscriptions",
"type": "User",
"url": "https://api.github.com/users/patil-suraj"
} | [] | closed | false | null | [] | null | 0 | "2020-11-26T13:05:58Z" | "2020-11-26T13:28:27Z" | "2020-11-26T13:28:26Z" | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/891.diff",
"html_url": "https://github.com/huggingface/datasets/pull/891",
"merged_at": "2020-11-26T13:28:26Z",
"patch_url": "https://github.com/huggingface/datasets/pull/891.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/891"
} | ignore `.python-version` added by `pyenv` | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/891/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/891/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/890 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/890/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/890/comments | https://api.github.com/repos/huggingface/datasets/issues/890/events | https://github.com/huggingface/datasets/pull/890 | 751,534,050 | MDExOlB1bGxSZXF1ZXN0NTI4MDI5NjA3 | 890 | Add LER | {
"avatar_url": "https://avatars.githubusercontent.com/u/3775944?v=4",
"events_url": "https://api.github.com/users/JoelNiklaus/events{/privacy}",
"followers_url": "https://api.github.com/users/JoelNiklaus/followers",
"following_url": "https://api.github.com/users/JoelNiklaus/following{/other_user}",
"gists_url": "https://api.github.com/users/JoelNiklaus/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/JoelNiklaus",
"id": 3775944,
"login": "JoelNiklaus",
"node_id": "MDQ6VXNlcjM3NzU5NDQ=",
"organizations_url": "https://api.github.com/users/JoelNiklaus/orgs",
"received_events_url": "https://api.github.com/users/JoelNiklaus/received_events",
"repos_url": "https://api.github.com/users/JoelNiklaus/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/JoelNiklaus/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/JoelNiklaus/subscriptions",
"type": "User",
"url": "https://api.github.com/users/JoelNiklaus"
} | [] | closed | false | null | [] | null | 9 | "2020-11-26T11:58:23Z" | "2020-12-01T13:33:35Z" | "2020-12-01T13:26:16Z" | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/890.diff",
"html_url": "https://github.com/huggingface/datasets/pull/890",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/890.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/890"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/890/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/890/timeline | null | null | true |
|
https://api.github.com/repos/huggingface/datasets/issues/889 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/889/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/889/comments | https://api.github.com/repos/huggingface/datasets/issues/889/events | https://github.com/huggingface/datasets/pull/889 | 751,115,691 | MDExOlB1bGxSZXF1ZXN0NTI3NjkwODE2 | 889 | Optional per-dataset default config name | {
"avatar_url": "https://avatars.githubusercontent.com/u/9353833?v=4",
"events_url": "https://api.github.com/users/joeddav/events{/privacy}",
"followers_url": "https://api.github.com/users/joeddav/followers",
"following_url": "https://api.github.com/users/joeddav/following{/other_user}",
"gists_url": "https://api.github.com/users/joeddav/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/joeddav",
"id": 9353833,
"login": "joeddav",
"node_id": "MDQ6VXNlcjkzNTM4MzM=",
"organizations_url": "https://api.github.com/users/joeddav/orgs",
"received_events_url": "https://api.github.com/users/joeddav/received_events",
"repos_url": "https://api.github.com/users/joeddav/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/joeddav/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/joeddav/subscriptions",
"type": "User",
"url": "https://api.github.com/users/joeddav"
} | [] | closed | false | null | [] | null | 3 | "2020-11-25T21:02:30Z" | "2020-11-30T17:27:33Z" | "2020-11-30T17:27:27Z" | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/889.diff",
"html_url": "https://github.com/huggingface/datasets/pull/889",
"merged_at": "2020-11-30T17:27:27Z",
"patch_url": "https://github.com/huggingface/datasets/pull/889.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/889"
} | This PR adds a `DEFAULT_CONFIG_NAME` class attribute to `DatasetBuilder`. This allows a dataset to have a specified default config name when a dataset has more than one config but the user does not specify it. For example, after defining `DEFAULT_CONFIG_NAME = "combined"` in PolyglotNER, a user can now do the following:
```python
ds = load_dataset("polyglot_ner")
```
which is equivalent to,
```python
ds = load_dataset("polyglot_ner", "combined")
```
In effect (for this particular dataset configuration), this means that if the user doesn't specify a language, they are given the combined dataset including all languages.
Since it doesn't always make sense to have a default config, this feature is opt-in. If `DEFAULT_CONFIG_NAME` is not defined and a user does not pass a config for a dataset with multiple configs available, a ValueError is raised like usual.
Let me know what you think about this approach @lhoestq @thomwolf and I'll add some documentation and define a default for some of our existing datasets. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/889/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/889/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/888 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/888/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/888/comments | https://api.github.com/repos/huggingface/datasets/issues/888/events | https://github.com/huggingface/datasets/issues/888 | 750,944,422 | MDU6SXNzdWU3NTA5NDQ0MjI= | 888 | Nested lists are zipped unexpectedly | {
"avatar_url": "https://avatars.githubusercontent.com/u/5757359?v=4",
"events_url": "https://api.github.com/users/AmitMY/events{/privacy}",
"followers_url": "https://api.github.com/users/AmitMY/followers",
"following_url": "https://api.github.com/users/AmitMY/following{/other_user}",
"gists_url": "https://api.github.com/users/AmitMY/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/AmitMY",
"id": 5757359,
"login": "AmitMY",
"node_id": "MDQ6VXNlcjU3NTczNTk=",
"organizations_url": "https://api.github.com/users/AmitMY/orgs",
"received_events_url": "https://api.github.com/users/AmitMY/received_events",
"repos_url": "https://api.github.com/users/AmitMY/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/AmitMY/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/AmitMY/subscriptions",
"type": "User",
"url": "https://api.github.com/users/AmitMY"
} | [] | closed | false | null | [] | null | 2 | "2020-11-25T16:07:46Z" | "2020-11-25T17:30:39Z" | "2020-11-25T17:30:39Z" | CONTRIBUTOR | null | null | null | I might misunderstand something, but I expect that if I define:
```python
"top": datasets.features.Sequence({
"middle": datasets.features.Sequence({
"bottom": datasets.Value("int32")
})
})
```
And I then create an example:
```python
yield 1, {
"top": [{
"middle": [
{"bottom": 1},
{"bottom": 2}
]
}]
}
```
I then load my dataset:
```python
train = load_dataset("my dataset")["train"]
```
and expect to be able to access `data[0]["top"][0]["middle"][0]`.
That is not the case. Here is `data[0]` as JSON:
```json
{"top": {"middle": [{"bottom": [1, 2]}]}}
```
Clearly different than the thing I inputted.
```json
{"top": [{"middle": [{"bottom": 1},{"bottom": 2}]}]}
``` | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/888/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/888/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/887 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/887/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/887/comments | https://api.github.com/repos/huggingface/datasets/issues/887/events | https://github.com/huggingface/datasets/issues/887 | 750,868,831 | MDU6SXNzdWU3NTA4Njg4MzE= | 887 | pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type> | {
"avatar_url": "https://avatars.githubusercontent.com/u/5757359?v=4",
"events_url": "https://api.github.com/users/AmitMY/events{/privacy}",
"followers_url": "https://api.github.com/users/AmitMY/followers",
"following_url": "https://api.github.com/users/AmitMY/following{/other_user}",
"gists_url": "https://api.github.com/users/AmitMY/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/AmitMY",
"id": 5757359,
"login": "AmitMY",
"node_id": "MDQ6VXNlcjU3NTczNTk=",
"organizations_url": "https://api.github.com/users/AmitMY/orgs",
"received_events_url": "https://api.github.com/users/AmitMY/received_events",
"repos_url": "https://api.github.com/users/AmitMY/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/AmitMY/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/AmitMY/subscriptions",
"type": "User",
"url": "https://api.github.com/users/AmitMY"
} | [
{
"color": "d73a4a",
"default": true,
"description": "Something isn't working",
"id": 1935892857,
"name": "bug",
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug"
}
] | open | false | null | [] | null | 14 | "2020-11-25T14:32:21Z" | "2021-09-09T17:03:40Z" | null | CONTRIBUTOR | null | null | null | I set up a new dataset, with a sequence of arrays (really, I want to have an array of (None, 137, 2), and the first dimension is dynamic)
```python
def _info(self):
return datasets.DatasetInfo(
description=_DESCRIPTION,
# This defines the different columns of the dataset and their types
features=datasets.Features(
{
"pose": datasets.features.Sequence(datasets.features.Array2D(shape=(137, 2), dtype="float32"))
}
),
homepage=_HOMEPAGE,
citation=_CITATION,
)
def _generate_examples(self):
""" Yields examples. """
yield 1, {
"pose": [np.zeros(shape=(137, 2), dtype=np.float32)]
}
```
But this doesn't work -
> pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type> | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/887/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/887/timeline | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/886 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/886/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/886/comments | https://api.github.com/repos/huggingface/datasets/issues/886/events | https://github.com/huggingface/datasets/pull/886 | 750,829,314 | MDExOlB1bGxSZXF1ZXN0NTI3NDU1MDU5 | 886 | Fix wikipedia custom config | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq"
} | [] | closed | false | null | [] | null | 1 | "2020-11-25T13:44:12Z" | "2021-06-25T05:24:16Z" | "2020-11-25T15:42:13Z" | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/886.diff",
"html_url": "https://github.com/huggingface/datasets/pull/886",
"merged_at": "2020-11-25T15:42:13Z",
"patch_url": "https://github.com/huggingface/datasets/pull/886.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/886"
} | It should be possible to use the wikipedia dataset with any `language` and `date`.
However it was not working as noticed in #784 . Indeed the custom wikipedia configurations were not enabled for some reason.
I fixed that and was able to run
```python
from datasets import load_dataset
load_dataset("./datasets/wikipedia", language="zh", date="20201120", beam_runner='DirectRunner')
```
cc @stvhuang @SamuelCahyawijaya
Fix #784 | {
"+1": 2,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 2,
"url": "https://api.github.com/repos/huggingface/datasets/issues/886/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/886/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/885 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/885/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/885/comments | https://api.github.com/repos/huggingface/datasets/issues/885/events | https://github.com/huggingface/datasets/issues/885 | 750,789,052 | MDU6SXNzdWU3NTA3ODkwNTI= | 885 | Very slow cold-start | {
"avatar_url": "https://avatars.githubusercontent.com/u/5757359?v=4",
"events_url": "https://api.github.com/users/AmitMY/events{/privacy}",
"followers_url": "https://api.github.com/users/AmitMY/followers",
"following_url": "https://api.github.com/users/AmitMY/following{/other_user}",
"gists_url": "https://api.github.com/users/AmitMY/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/AmitMY",
"id": 5757359,
"login": "AmitMY",
"node_id": "MDQ6VXNlcjU3NTczNTk=",
"organizations_url": "https://api.github.com/users/AmitMY/orgs",
"received_events_url": "https://api.github.com/users/AmitMY/received_events",
"repos_url": "https://api.github.com/users/AmitMY/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/AmitMY/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/AmitMY/subscriptions",
"type": "User",
"url": "https://api.github.com/users/AmitMY"
} | [
{
"color": "e99695",
"default": false,
"description": "Requesting to add a new dataset",
"id": 2067376369,
"name": "dataset request",
"node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5",
"url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request"
}
] | closed | false | null | [] | null | 3 | "2020-11-25T12:47:58Z" | "2021-01-13T11:31:25Z" | "2021-01-13T11:31:25Z" | CONTRIBUTOR | null | null | null | Hi,
I expect when importing `datasets` that nothing major happens in the background, and so the import should be insignificant.
When I load a metric, or a dataset, its fine that it takes time.
The following ranges from 3 to 9 seconds:
```
python -m timeit -n 1 -r 1 'from datasets import load_dataset'
```
edit:
sorry for the mis-tag, not sure how I added it. | {
"+1": 1,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/885/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/885/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/884 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/884/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/884/comments | https://api.github.com/repos/huggingface/datasets/issues/884/events | https://github.com/huggingface/datasets/pull/884 | 749,862,034 | MDExOlB1bGxSZXF1ZXN0NTI2NjA5MDc1 | 884 | Auto generate dummy data | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq"
} | [] | closed | false | null | [] | null | 3 | "2020-11-24T16:31:34Z" | "2020-11-26T14:18:47Z" | "2020-11-26T14:18:46Z" | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/884.diff",
"html_url": "https://github.com/huggingface/datasets/pull/884",
"merged_at": "2020-11-26T14:18:46Z",
"patch_url": "https://github.com/huggingface/datasets/pull/884.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/884"
} | When adding a new dataset to the library, dummy data creation can take some time.
To make things easier I added a command line tool that automatically generates dummy data when possible.
The tool only supports certain data files types: txt, csv, tsv, jsonl, json and xml.
Here are some examples:
```
python datasets-cli dummy_data ./datasets/snli --auto_generate
python datasets-cli dummy_data ./datasets/squad --auto_generate --json_field data
python datasets-cli dummy_data ./datasets/iwslt2017 --auto_generate --xml_tag seg --match_text_files "train*" --n_lines 15
# --xml_tag seg => each sample corresponds to a "seg" tag in the xml tree
# --match_text_files "train*" => also match text files that don't have a proper text file extension (no suffix like ".txt" for example)
# --n_lines 15 => some text files have headers so we have to use at least 15 lines
```
and here is the command usage:
```
usage: datasets-cli <command> [<args>] dummy_data [-h] [--auto_generate]
[--n_lines N_LINES]
[--json_field JSON_FIELD]
[--xml_tag XML_TAG]
[--match_text_files MATCH_TEXT_FILES]
[--keep_uncompressed]
[--cache_dir CACHE_DIR]
path_to_dataset
positional arguments:
path_to_dataset Path to the dataset (example: ./datasets/squad)
optional arguments:
-h, --help show this help message and exit
--auto_generate Try to automatically generate dummy data
--n_lines N_LINES Number of lines or samples to keep when auto-
generating dummy data
--json_field JSON_FIELD
Optional, json field to read the data from when auto-
generating dummy data. In the json data files, this
field must point to a list of samples as json objects
(ex: the 'data' field for squad-like files)
--xml_tag XML_TAG Optional, xml tag name of the samples inside the xml
files when auto-generating dummy data.
--match_text_files MATCH_TEXT_FILES
Optional, a comma separated list of file patterns that
looks for line-by-line text files other than *.txt or
*.csv. Example: --match_text_files *.label
--keep_uncompressed Don't compress the dummy data folders when auto-
generating dummy data. Useful for debugging for to do
manual adjustements before compressing.
--cache_dir CACHE_DIR
Cache directory to download and cache files when auto-
generating dummy data
```
The command generates all the necessary `dummy_data.zip` files (one per config).
How it works:
- it runs the split_generators() method of the dataset script to download the original data files
- when downloading it records a mapping between the downloaded files and the corresponding expected dummy data files paths
- then for each data file it creates the dummy data file keeping only the first samples (the strategy depends on the type of file)
- finally it compresses the dummy data folders into dummy_zip files ready for dataset tests
Let me know if that makes sense or if you have ideas to improve this tool !
I also added a unit test. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/884/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/884/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/883 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/883/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/883/comments | https://api.github.com/repos/huggingface/datasets/issues/883/events | https://github.com/huggingface/datasets/issues/883 | 749,750,801 | MDU6SXNzdWU3NDk3NTA4MDE= | 883 | Downloading/caching only a part of a datasets' dataset. | {
"avatar_url": "https://avatars.githubusercontent.com/u/44585792?v=4",
"events_url": "https://api.github.com/users/SapirWeissbuch/events{/privacy}",
"followers_url": "https://api.github.com/users/SapirWeissbuch/followers",
"following_url": "https://api.github.com/users/SapirWeissbuch/following{/other_user}",
"gists_url": "https://api.github.com/users/SapirWeissbuch/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/SapirWeissbuch",
"id": 44585792,
"login": "SapirWeissbuch",
"node_id": "MDQ6VXNlcjQ0NTg1Nzky",
"organizations_url": "https://api.github.com/users/SapirWeissbuch/orgs",
"received_events_url": "https://api.github.com/users/SapirWeissbuch/received_events",
"repos_url": "https://api.github.com/users/SapirWeissbuch/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/SapirWeissbuch/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/SapirWeissbuch/subscriptions",
"type": "User",
"url": "https://api.github.com/users/SapirWeissbuch"
} | [
{
"color": "a2eeef",
"default": true,
"description": "New feature or request",
"id": 1935892871,
"name": "enhancement",
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement"
},
{
"color": "d876e3",
"default": true,
"description": "Further information is requested",
"id": 1935892912,
"name": "question",
"node_id": "MDU6TGFiZWwxOTM1ODkyOTEy",
"url": "https://api.github.com/repos/huggingface/datasets/labels/question"
}
] | open | false | null | [] | null | 3 | "2020-11-24T14:25:18Z" | "2020-11-27T13:51:55Z" | null | NONE | null | null | null | Hi,
I want to use the validation data *only* (of natural question).
I don't want to have the whole dataset cached in my machine, just the dev set.
Is this possible? I can't find a way to do it in the docs.
Thank you,
Sapir | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/883/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/883/timeline | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/882 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/882/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/882/comments | https://api.github.com/repos/huggingface/datasets/issues/882/events | https://github.com/huggingface/datasets/pull/882 | 749,662,188 | MDExOlB1bGxSZXF1ZXN0NTI2NDQyMjA2 | 882 | Update README.md | {
"avatar_url": "https://avatars.githubusercontent.com/u/32997732?v=4",
"events_url": "https://api.github.com/users/vaibhavad/events{/privacy}",
"followers_url": "https://api.github.com/users/vaibhavad/followers",
"following_url": "https://api.github.com/users/vaibhavad/following{/other_user}",
"gists_url": "https://api.github.com/users/vaibhavad/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/vaibhavad",
"id": 32997732,
"login": "vaibhavad",
"node_id": "MDQ6VXNlcjMyOTk3NzMy",
"organizations_url": "https://api.github.com/users/vaibhavad/orgs",
"received_events_url": "https://api.github.com/users/vaibhavad/received_events",
"repos_url": "https://api.github.com/users/vaibhavad/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/vaibhavad/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/vaibhavad/subscriptions",
"type": "User",
"url": "https://api.github.com/users/vaibhavad"
} | [] | closed | false | null | [] | null | 0 | "2020-11-24T12:23:52Z" | "2021-01-29T10:41:07Z" | "2021-01-29T10:41:07Z" | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/882.diff",
"html_url": "https://github.com/huggingface/datasets/pull/882",
"merged_at": "2021-01-29T10:41:06Z",
"patch_url": "https://github.com/huggingface/datasets/pull/882.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/882"
} | "no label" is "-" in the original dataset but "-1" in Huggingface distribution. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/882/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/882/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/881 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/881/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/881/comments | https://api.github.com/repos/huggingface/datasets/issues/881/events | https://github.com/huggingface/datasets/pull/881 | 749,548,107 | MDExOlB1bGxSZXF1ZXN0NTI2MzQ5MDM2 | 881 | Use GCP download url instead of tensorflow custom download for boolq | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq"
} | [] | closed | false | null | [] | null | 0 | "2020-11-24T09:47:11Z" | "2020-11-24T10:12:34Z" | "2020-11-24T10:12:33Z" | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/881.diff",
"html_url": "https://github.com/huggingface/datasets/pull/881",
"merged_at": "2020-11-24T10:12:33Z",
"patch_url": "https://github.com/huggingface/datasets/pull/881.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/881"
} | BoolQ is a dataset that used tf.io.gfile.copy to download the file from a GCP bucket.
It prevented the dataset to be downloaded twice because of a FileAlreadyExistsError.
Even though the error could be fixed by providing `overwrite=True` to the tf.io.gfile.copy call, I changed the script to use GCP download urls and use regular downloads instead and remove the tensorflow dependency.
Fix #875 | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/881/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/881/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/880 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/880/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/880/comments | https://api.github.com/repos/huggingface/datasets/issues/880/events | https://github.com/huggingface/datasets/issues/880 | 748,949,606 | MDU6SXNzdWU3NDg5NDk2MDY= | 880 | Add SQA | {
"avatar_url": "https://avatars.githubusercontent.com/u/48327001?v=4",
"events_url": "https://api.github.com/users/NielsRogge/events{/privacy}",
"followers_url": "https://api.github.com/users/NielsRogge/followers",
"following_url": "https://api.github.com/users/NielsRogge/following{/other_user}",
"gists_url": "https://api.github.com/users/NielsRogge/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/NielsRogge",
"id": 48327001,
"login": "NielsRogge",
"node_id": "MDQ6VXNlcjQ4MzI3MDAx",
"organizations_url": "https://api.github.com/users/NielsRogge/orgs",
"received_events_url": "https://api.github.com/users/NielsRogge/received_events",
"repos_url": "https://api.github.com/users/NielsRogge/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/NielsRogge/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/NielsRogge/subscriptions",
"type": "User",
"url": "https://api.github.com/users/NielsRogge"
} | [
{
"color": "e99695",
"default": false,
"description": "Requesting to add a new dataset",
"id": 2067376369,
"name": "dataset request",
"node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5",
"url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request"
}
] | closed | false | null | [] | null | 3 | "2020-11-23T16:31:55Z" | "2020-12-23T13:58:24Z" | "2020-12-23T13:58:23Z" | CONTRIBUTOR | null | null | null | ## Adding a Dataset
- **Name:** SQA (Sequential Question Answering) by Microsoft.
- **Description:** The SQA dataset was created to explore the task of answering sequences of inter-related questions on HTML tables. It has 6,066 sequences with 17,553 questions in total.
- **Paper:** https://www.microsoft.com/en-us/research/publication/search-based-neural-structured-learning-sequential-question-answering/
- **Data:** https://www.microsoft.com/en-us/download/details.aspx?id=54253
- **Motivation:** currently, the [Tapas](https://ai.googleblog.com/2020/04/using-neural-networks-to-find-answers.html) algorithm by Google AI is being added to the Transformers library (see https://github.com/huggingface/transformers/pull/8113). It would be great to use that model in combination with this dataset, on which it achieves SOTA results (average question accuracy of 0.71).
Note 1: this dataset actually consists of 2 types of files:
1) TSV files, containing the questions, answer coordinates and answer texts (for training, dev and test)
2) a folder of csv files, which contain the actual tabular data
Note 2: if you download the dataset straight from the download link above, then you will see that the `answer_coordinates` and `answer_text` columns are string lists of string tuples and strings respectively, which is not ideal. It would be better to make them true Python lists of tuples and strings respectively (using `ast.literal_eval`), before uploading them to the HuggingFace hub.
Adding this would be great! Then we could possibly also add [WTQ (WikiTable Questions)](https://github.com/ppasupat/WikiTableQuestions) and [TabFact (Tabular Fact Checking)](https://github.com/wenhuchen/Table-Fact-Checking) on which TAPAS also achieves state-of-the-art results. Note that the TAPAS algorithm requires these datasets to first be converted into the SQA format.
Instructions to add a new dataset can be found [here](https://huggingface.co/docs/datasets/share_dataset.html).
| {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/880/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/880/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/879 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/879/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/879/comments | https://api.github.com/repos/huggingface/datasets/issues/879/events | https://github.com/huggingface/datasets/issues/879 | 748,848,847 | MDU6SXNzdWU3NDg4NDg4NDc= | 879 | boolq does not load | {
"avatar_url": "https://avatars.githubusercontent.com/u/6278280?v=4",
"events_url": "https://api.github.com/users/rabeehk/events{/privacy}",
"followers_url": "https://api.github.com/users/rabeehk/followers",
"following_url": "https://api.github.com/users/rabeehk/following{/other_user}",
"gists_url": "https://api.github.com/users/rabeehk/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/rabeehk",
"id": 6278280,
"login": "rabeehk",
"node_id": "MDQ6VXNlcjYyNzgyODA=",
"organizations_url": "https://api.github.com/users/rabeehk/orgs",
"received_events_url": "https://api.github.com/users/rabeehk/received_events",
"repos_url": "https://api.github.com/users/rabeehk/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/rabeehk/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/rabeehk/subscriptions",
"type": "User",
"url": "https://api.github.com/users/rabeehk"
} | [
{
"color": "2edb81",
"default": false,
"description": "A bug in a dataset script provided in the library",
"id": 2067388877,
"name": "dataset bug",
"node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug"
}
] | closed | false | null | [] | null | 3 | "2020-11-23T14:28:28Z" | "2022-10-05T12:23:32Z" | "2022-10-05T12:23:32Z" | CONTRIBUTOR | null | null | null | Hi
I am getting these errors trying to load boolq thanks
Traceback (most recent call last):
File "test.py", line 5, in <module>
data = AutoTask().get("boolq").get_dataset("train", n_obs=10)
File "/remote/idiap.svm/user.active/rkarimi/dev/internship/seq2seq/tasks/tasks.py", line 42, in get_dataset
dataset = self.load_dataset(split=split)
File "/remote/idiap.svm/user.active/rkarimi/dev/internship/seq2seq/tasks/tasks.py", line 38, in load_dataset
return datasets.load_dataset(self.task.name, split=split)
File "/idiap/user/rkarimi/libs/anaconda3/envs/internship/lib/python3.7/site-packages/datasets/load.py", line 611, in load_dataset
ignore_verifications=ignore_verifications,
File "/idiap/user/rkarimi/libs/anaconda3/envs/internship/lib/python3.7/site-packages/datasets/builder.py", line 476, in download_and_prepare
dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs
File "/idiap/user/rkarimi/libs/anaconda3/envs/internship/lib/python3.7/site-packages/datasets/builder.py", line 531, in _download_and_prepare
split_generators = self._split_generators(dl_manager, **split_generators_kwargs)
File " /idiap/home/rkarimi/.cache/huggingface/modules/datasets_modules/datasets/boolq/2987db1f15deaa19500ae24de560eabeaf1f8ef51df88c0470beeec72943bf11/boolq.py", line 74, in _split_generators
downloaded_files = dl_manager.download_custom(urls_to_download, tf.io.gfile.copy)
File "/idiap/user/rkarimi/libs/anaconda3/envs/internship/lib/python3.7/site-packages/datasets/utils/download_manager.py", line 150, in download_custom
get_from_cache(url, cache_dir=cache_dir, local_files_only=True, use_etag=False)
File "/idiap/user/rkarimi/libs/anaconda3/envs/internship/lib/python3.7/site-packages/datasets/utils/file_utils.py", line 472, in get_from_cache
f"Cannot find the requested files in the cached path at {cache_path} and outgoing traffic has been"
FileNotFoundError: Cannot find the requested files in the cached path at /idiap/home/rkarimi/.cache/huggingface/datasets/eaee069e38f6ceaa84de02ad088c34e63ec97671f2cd1910ddb16b10dc60808c and outgoing traffic has been disabled. To enable file online look-ups, set 'local_files_only' to False.
| {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/879/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/879/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/878 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/878/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/878/comments | https://api.github.com/repos/huggingface/datasets/issues/878/events | https://github.com/huggingface/datasets/issues/878 | 748,621,981 | MDU6SXNzdWU3NDg2MjE5ODE= | 878 | Loading Data From S3 Path in Sagemaker | {
"avatar_url": "https://avatars.githubusercontent.com/u/42795522?v=4",
"events_url": "https://api.github.com/users/mahesh1amour/events{/privacy}",
"followers_url": "https://api.github.com/users/mahesh1amour/followers",
"following_url": "https://api.github.com/users/mahesh1amour/following{/other_user}",
"gists_url": "https://api.github.com/users/mahesh1amour/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/mahesh1amour",
"id": 42795522,
"login": "mahesh1amour",
"node_id": "MDQ6VXNlcjQyNzk1NTIy",
"organizations_url": "https://api.github.com/users/mahesh1amour/orgs",
"received_events_url": "https://api.github.com/users/mahesh1amour/received_events",
"repos_url": "https://api.github.com/users/mahesh1amour/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/mahesh1amour/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mahesh1amour/subscriptions",
"type": "User",
"url": "https://api.github.com/users/mahesh1amour"
} | [
{
"color": "a2eeef",
"default": true,
"description": "New feature or request",
"id": 1935892871,
"name": "enhancement",
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement"
},
{
"color": "d876e3",
"default": true,
"description": "Further information is requested",
"id": 1935892912,
"name": "question",
"node_id": "MDU6TGFiZWwxOTM1ODkyOTEy",
"url": "https://api.github.com/repos/huggingface/datasets/labels/question"
}
] | open | false | null | [] | null | 16 | "2020-11-23T09:17:22Z" | "2020-12-23T09:53:08Z" | null | NONE | null | null | null | In Sagemaker Im tring to load the data set from S3 path as follows
`train_path = 's3://xxxxxxxxxx/xxxxxxxxxx/train.csv'
valid_path = 's3://xxxxxxxxxx/xxxxxxxxxx/validation.csv'
test_path = 's3://xxxxxxxxxx/xxxxxxxxxx/test.csv'
data_files = {}
data_files["train"] = train_path
data_files["validation"] = valid_path
data_files["test"] = test_path
extension = train_path.split(".")[-1]
datasets = load_dataset(extension, data_files=data_files, s3_enabled=True)
print(datasets)`
I getting an error of
`algo-1-7plil_1 | File "main.py", line 21, in <module>
algo-1-7plil_1 | datasets = load_dataset(extension, data_files=data_files)
algo-1-7plil_1 | File "/opt/conda/lib/python3.6/site-packages/datasets/load.py", line 603, in load_dataset
algo-1-7plil_1 | **config_kwargs,
algo-1-7plil_1 | File "/opt/conda/lib/python3.6/site-packages/datasets/builder.py", line 155, in __init__
algo-1-7plil_1 | **config_kwargs,
algo-1-7plil_1 | File "/opt/conda/lib/python3.6/site-packages/datasets/builder.py", line 305, in _create_builder_config
algo-1-7plil_1 | m.update(str(os.path.getmtime(data_file)))
algo-1-7plil_1 | File "/opt/conda/lib/python3.6/genericpath.py", line 55, in getmtime
algo-1-7plil_1 | return os.stat(filename).st_mtime
algo-1-7plil_1 | FileNotFoundError: [Errno 2] No such file or directory: 's3://lsmv-sagemaker/pubmedbert/test.csv`
But when im trying with pandas , it is able to load from S3
Does the datasets library support S3 path to load | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/878/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/878/timeline | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/877 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/877/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/877/comments | https://api.github.com/repos/huggingface/datasets/issues/877/events | https://github.com/huggingface/datasets/issues/877 | 748,234,438 | MDU6SXNzdWU3NDgyMzQ0Mzg= | 877 | DataLoader(datasets) become more and more slowly within iterations | {
"avatar_url": "https://avatars.githubusercontent.com/u/25664170?v=4",
"events_url": "https://api.github.com/users/shexuan/events{/privacy}",
"followers_url": "https://api.github.com/users/shexuan/followers",
"following_url": "https://api.github.com/users/shexuan/following{/other_user}",
"gists_url": "https://api.github.com/users/shexuan/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/shexuan",
"id": 25664170,
"login": "shexuan",
"node_id": "MDQ6VXNlcjI1NjY0MTcw",
"organizations_url": "https://api.github.com/users/shexuan/orgs",
"received_events_url": "https://api.github.com/users/shexuan/received_events",
"repos_url": "https://api.github.com/users/shexuan/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/shexuan/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/shexuan/subscriptions",
"type": "User",
"url": "https://api.github.com/users/shexuan"
} | [] | closed | false | null | [] | null | 2 | "2020-11-22T12:41:10Z" | "2020-11-29T15:45:12Z" | "2020-11-29T15:45:12Z" | NONE | null | null | null | Hello, when I for loop my dataloader, the loading speed is becoming more and more slowly!
```
dataset = load_from_disk(dataset_path) # around 21,000,000 lines
lineloader = tqdm(DataLoader(dataset, batch_size=1))
for idx, line in enumerate(lineloader):
# do some thing for each line
```
In the begining, the loading speed is around 2000it/s, but after 1 minutes later, the speed is much slower, just around 800it/s.
And when I set `num_workers=4` in DataLoader, the loading speed is much lower, just 130it/s.
Could you please help me with this problem?
Thanks a lot! | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/877/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/877/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/876 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/876/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/876/comments | https://api.github.com/repos/huggingface/datasets/issues/876/events | https://github.com/huggingface/datasets/issues/876 | 748,195,104 | MDU6SXNzdWU3NDgxOTUxMDQ= | 876 | imdb dataset cannot be loaded | {
"avatar_url": "https://avatars.githubusercontent.com/u/6278280?v=4",
"events_url": "https://api.github.com/users/rabeehk/events{/privacy}",
"followers_url": "https://api.github.com/users/rabeehk/followers",
"following_url": "https://api.github.com/users/rabeehk/following{/other_user}",
"gists_url": "https://api.github.com/users/rabeehk/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/rabeehk",
"id": 6278280,
"login": "rabeehk",
"node_id": "MDQ6VXNlcjYyNzgyODA=",
"organizations_url": "https://api.github.com/users/rabeehk/orgs",
"received_events_url": "https://api.github.com/users/rabeehk/received_events",
"repos_url": "https://api.github.com/users/rabeehk/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/rabeehk/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/rabeehk/subscriptions",
"type": "User",
"url": "https://api.github.com/users/rabeehk"
} | [] | closed | false | null | [] | null | 5 | "2020-11-22T08:24:43Z" | "2021-11-26T11:07:16Z" | "2020-12-24T17:38:47Z" | CONTRIBUTOR | null | null | null | Hi
I am trying to load the imdb train dataset
`dataset = datasets.load_dataset("imdb", split="train")`
getting following errors, thanks for your help
```
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "/idiap/user/rkarimi/libs/anaconda3/envs/internship/lib/python3.7/site-packages/datasets/load.py", line 611, in load_dataset
ignore_verifications=ignore_verifications,
File "/idiap/user/rkarimi/libs/anaconda3/envs/internship/lib/python3.7/site-packages/datasets/builder.py", line 476, in download_and_prepare
dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs
File "/idiap/user/rkarimi/libs/anaconda3/envs/internship/lib/python3.7/site-packages/datasets/builder.py", line 558, in _download_and_prepare
verify_splits(self.info.splits, split_dict)
File "/idiap/user/rkarimi/libs/anaconda3/envs/internship/lib/python3.7/site-packages/datasets/utils/info_utils.py", line 73, in verify_splits
raise NonMatchingSplitsSizesError(str(bad_splits))
datasets.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='test', num_bytes=32660064, num_examples=25000, dataset_name='imdb'), 'recorded': SplitInfo(name='test', num_bytes=26476338, num_examples=20316, dataset_name='imdb')}, {'expected': SplitInfo(name='train', num_bytes=33442202, num_examples=25000, dataset_name='imdb'), 'recorded': SplitInfo(name='train', num_bytes=0, num_examples=0, dataset_name='imdb')}, {'expected': SplitInfo(name='unsupervised', num_bytes=67125548, num_examples=50000, dataset_name='imdb'), 'recorded': SplitInfo(name='unsupervised', num_bytes=0, num_examples=0, dataset_name='imdb')}]
>>> dataset = datasets.load_dataset("imdb", split="train")
```
| {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/876/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/876/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/875 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/875/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/875/comments | https://api.github.com/repos/huggingface/datasets/issues/875/events | https://github.com/huggingface/datasets/issues/875 | 748,194,311 | MDU6SXNzdWU3NDgxOTQzMTE= | 875 | bug in boolq dataset loading | {
"avatar_url": "https://avatars.githubusercontent.com/u/6278280?v=4",
"events_url": "https://api.github.com/users/rabeehk/events{/privacy}",
"followers_url": "https://api.github.com/users/rabeehk/followers",
"following_url": "https://api.github.com/users/rabeehk/following{/other_user}",
"gists_url": "https://api.github.com/users/rabeehk/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/rabeehk",
"id": 6278280,
"login": "rabeehk",
"node_id": "MDQ6VXNlcjYyNzgyODA=",
"organizations_url": "https://api.github.com/users/rabeehk/orgs",
"received_events_url": "https://api.github.com/users/rabeehk/received_events",
"repos_url": "https://api.github.com/users/rabeehk/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/rabeehk/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/rabeehk/subscriptions",
"type": "User",
"url": "https://api.github.com/users/rabeehk"
} | [] | closed | false | null | [] | null | 1 | "2020-11-22T08:18:34Z" | "2020-11-24T10:12:33Z" | "2020-11-24T10:12:33Z" | CONTRIBUTOR | null | null | null | Hi
I am trying to load boolq dataset:
```
import datasets
datasets.load_dataset("boolq")
```
I am getting the following errors, thanks for your help
```
>>> import datasets
2020-11-22 09:16:30.070332: W tensorflow/stream_executor/platform/default/dso_loader.cc:60] Could not load dynamic library 'libcudart.so.11.0'; dlerror: libcudart.so.11.0: cannot open shared object file: No such file or directory
2020-11-22 09:16:30.070389: I tensorflow/stream_executor/cuda/cudart_stub.cc:29] Ignore above cudart dlerror if you do not have a GPU set up on your machine.
>>> datasets.load_dataset("boolq")
cahce dir /idiap/temp/rkarimi/cache_home/datasets
cahce dir /idiap/temp/rkarimi/cache_home/datasets
Using custom data configuration default
Downloading and preparing dataset boolq/default (download: 8.36 MiB, generated: 7.47 MiB, post-processed: Unknown size, total: 15.83 MiB) to /idiap/temp/rkarimi/cache_home/datasets/boolq/default/0.1.0/2987db1f15deaa19500ae24de560eabeaf1f8ef51df88c0470beeec72943bf11...
cahce dir /idiap/temp/rkarimi/cache_home/datasets
cahce dir /idiap/temp/rkarimi/cache_home/datasets/downloads
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "/idiap/user/rkarimi/libs/anaconda3/envs/internship/lib/python3.7/site-packages/datasets/load.py", line 611, in load_dataset
ignore_verifications=ignore_verifications,
File "/idiap/user/rkarimi/libs/anaconda3/envs/internship/lib/python3.7/site-packages/datasets/builder.py", line 476, in download_and_prepare
dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs
File "/idiap/user/rkarimi/libs/anaconda3/envs/internship/lib/python3.7/site-packages/datasets/builder.py", line 531, in _download_and_prepare
split_generators = self._split_generators(dl_manager, **split_generators_kwargs)
File " /idiap/home/rkarimi/.cache/huggingface/modules/datasets_modules/datasets/boolq/2987db1f15deaa19500ae24de560eabeaf1f8ef51df88c0470beeec72943bf11/boolq.py", line 74, in _split_generators
downloaded_files = dl_manager.download_custom(urls_to_download, tf.io.gfile.copy)
File "/idiap/user/rkarimi/libs/anaconda3/envs/internship/lib/python3.7/site-packages/datasets/utils/download_manager.py", line 149, in download_custom
custom_download(url, path)
File "/idiap/user/rkarimi/libs/anaconda3/envs/internship/lib/python3.7/site-packages/tensorflow/python/lib/io/file_io.py", line 516, in copy_v2
compat.path_to_bytes(src), compat.path_to_bytes(dst), overwrite)
tensorflow.python.framework.errors_impl.AlreadyExistsError: file already exists
``` | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/875/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/875/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/874 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/874/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/874/comments | https://api.github.com/repos/huggingface/datasets/issues/874/events | https://github.com/huggingface/datasets/issues/874 | 748,193,140 | MDU6SXNzdWU3NDgxOTMxNDA= | 874 | trec dataset unavailable | {
"avatar_url": "https://avatars.githubusercontent.com/u/6278280?v=4",
"events_url": "https://api.github.com/users/rabeehk/events{/privacy}",
"followers_url": "https://api.github.com/users/rabeehk/followers",
"following_url": "https://api.github.com/users/rabeehk/following{/other_user}",
"gists_url": "https://api.github.com/users/rabeehk/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/rabeehk",
"id": 6278280,
"login": "rabeehk",
"node_id": "MDQ6VXNlcjYyNzgyODA=",
"organizations_url": "https://api.github.com/users/rabeehk/orgs",
"received_events_url": "https://api.github.com/users/rabeehk/received_events",
"repos_url": "https://api.github.com/users/rabeehk/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/rabeehk/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/rabeehk/subscriptions",
"type": "User",
"url": "https://api.github.com/users/rabeehk"
} | [] | closed | false | null | [] | null | 2 | "2020-11-22T08:09:36Z" | "2020-11-27T13:56:42Z" | "2020-11-27T13:56:42Z" | CONTRIBUTOR | null | null | null | Hi
when I try to load the trec dataset I am getting these errors, thanks for your help
`datasets.load_dataset("trec", split="train")
`
```
File "<stdin>", line 1, in <module>
File "/idiap/user/rkarimi/libs/anaconda3/envs/internship/lib/python3.7/site-packages/datasets/load.py", line 611, in load_dataset
ignore_verifications=ignore_verifications,
File "/idiap/user/rkarimi/libs/anaconda3/envs/internship/lib/python3.7/site-packages/datasets/builder.py", line 476, in download_and_prepare
dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs
File "/idiap/user/rkarimi/libs/anaconda3/envs/internship/lib/python3.7/site-packages/datasets/builder.py", line 531, in _download_and_prepare
split_generators = self._split_generators(dl_manager, **split_generators_kwargs)
File " /idiap/home/rkarimi/.cache/huggingface/modules/datasets_modules/datasets/trec/ca4248481ad244f235f4cf277186cad2ee8769f975119a2bbfc41b8932b88bd7/trec.py", line 140, in _split_generators
dl_files = dl_manager.download_and_extract(_URLs)
File "/idiap/user/rkarimi/libs/anaconda3/envs/internship/lib/python3.7/site-packages/datasets/utils/download_manager.py", line 254, in download_and_extract
return self.extract(self.download(url_or_urls))
File "/idiap/user/rkarimi/libs/anaconda3/envs/internship/lib/python3.7/site-packages/datasets/utils/download_manager.py", line 179, in download
num_proc=download_config.num_proc,
File "/idiap/user/rkarimi/libs/anaconda3/envs/internship/lib/python3.7/site-packages/datasets/utils/py_utils.py", line 225, in map_nested
_single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)
File "/idiap/user/rkarimi/libs/anaconda3/envs/internship/lib/python3.7/site-packages/datasets/utils/py_utils.py", line 225, in <listcomp>
_single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)
File "/idiap/user/rkarimi/libs/anaconda3/envs/internship/lib/python3.7/site-packages/datasets/utils/py_utils.py", line 163, in _single_map_nested
return function(data_struct)
File "/idiap/user/rkarimi/libs/anaconda3/envs/internship/lib/python3.7/site-packages/datasets/utils/file_utils.py", line 308, in cached_path
use_etag=download_config.use_etag,
File "/idiap/user/rkarimi/libs/anaconda3/envs/internship/lib/python3.7/site-packages/datasets/utils/file_utils.py", line 477, in get_from_cache
raise ConnectionError("Couldn't reach {}".format(url))
ConnectionError: Couldn't reach http://cogcomp.org/Data/QA/QC/train_5500.label
``` | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/874/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/874/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/873 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/873/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/873/comments | https://api.github.com/repos/huggingface/datasets/issues/873/events | https://github.com/huggingface/datasets/issues/873 | 747,959,523 | MDU6SXNzdWU3NDc5NTk1MjM= | 873 | load_dataset('cnn_dalymail', '3.0.0') gives a 'Not a directory' error | {
"avatar_url": "https://avatars.githubusercontent.com/u/19861874?v=4",
"events_url": "https://api.github.com/users/vishal-burman/events{/privacy}",
"followers_url": "https://api.github.com/users/vishal-burman/followers",
"following_url": "https://api.github.com/users/vishal-burman/following{/other_user}",
"gists_url": "https://api.github.com/users/vishal-burman/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/vishal-burman",
"id": 19861874,
"login": "vishal-burman",
"node_id": "MDQ6VXNlcjE5ODYxODc0",
"organizations_url": "https://api.github.com/users/vishal-burman/orgs",
"received_events_url": "https://api.github.com/users/vishal-burman/received_events",
"repos_url": "https://api.github.com/users/vishal-burman/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/vishal-burman/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/vishal-burman/subscriptions",
"type": "User",
"url": "https://api.github.com/users/vishal-burman"
} | [] | closed | false | null | [] | null | 13 | "2020-11-21T06:30:45Z" | "2023-08-03T12:07:03Z" | "2020-11-22T12:18:05Z" | NONE | null | null | null | ```
from datasets import load_dataset
dataset = load_dataset('cnn_dailymail', '3.0.0')
```
Stack trace:
```
---------------------------------------------------------------------------
NotADirectoryError Traceback (most recent call last)
<ipython-input-6-2e06a8332652> in <module>()
1 from datasets import load_dataset
----> 2 dataset = load_dataset('cnn_dailymail', '3.0.0')
5 frames
/usr/local/lib/python3.6/dist-packages/datasets/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)
608 download_config=download_config,
609 download_mode=download_mode,
--> 610 ignore_verifications=ignore_verifications,
611 )
612
/usr/local/lib/python3.6/dist-packages/datasets/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)
513 if not downloaded_from_gcs:
514 self._download_and_prepare(
--> 515 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs
516 )
517 # Sync info
/usr/local/lib/python3.6/dist-packages/datasets/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)
568 split_dict = SplitDict(dataset_name=self.name)
569 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs)
--> 570 split_generators = self._split_generators(dl_manager, **split_generators_kwargs)
571
572 # Checksums verification
/root/.cache/huggingface/modules/datasets_modules/datasets/cnn_dailymail/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602/cnn_dailymail.py in _split_generators(self, dl_manager)
252 def _split_generators(self, dl_manager):
253 dl_paths = dl_manager.download_and_extract(_DL_URLS)
--> 254 train_files = _subset_filenames(dl_paths, datasets.Split.TRAIN)
255 # Generate shared vocabulary
256
/root/.cache/huggingface/modules/datasets_modules/datasets/cnn_dailymail/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602/cnn_dailymail.py in _subset_filenames(dl_paths, split)
153 else:
154 logging.fatal("Unsupported split: %s", split)
--> 155 cnn = _find_files(dl_paths, "cnn", urls)
156 dm = _find_files(dl_paths, "dm", urls)
157 return cnn + dm
/root/.cache/huggingface/modules/datasets_modules/datasets/cnn_dailymail/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602/cnn_dailymail.py in _find_files(dl_paths, publisher, url_dict)
132 else:
133 logging.fatal("Unsupported publisher: %s", publisher)
--> 134 files = sorted(os.listdir(top_dir))
135
136 ret_files = []
NotADirectoryError: [Errno 20] Not a directory: '/root/.cache/huggingface/datasets/downloads/1bc05d24fa6dda2468e83a73cf6dc207226e01e3c48a507ea716dc0421da583b/cnn/stories'
```
I have ran the code on Google Colab | {
"+1": 2,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 2,
"url": "https://api.github.com/repos/huggingface/datasets/issues/873/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/873/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/872 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/872/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/872/comments | https://api.github.com/repos/huggingface/datasets/issues/872/events | https://github.com/huggingface/datasets/pull/872 | 747,653,697 | MDExOlB1bGxSZXF1ZXN0NTI0ODM4NjEx | 872 | Add IndicGLUE dataset and Metrics | {
"avatar_url": "https://avatars.githubusercontent.com/u/28291870?v=4",
"events_url": "https://api.github.com/users/sumanthd17/events{/privacy}",
"followers_url": "https://api.github.com/users/sumanthd17/followers",
"following_url": "https://api.github.com/users/sumanthd17/following{/other_user}",
"gists_url": "https://api.github.com/users/sumanthd17/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/sumanthd17",
"id": 28291870,
"login": "sumanthd17",
"node_id": "MDQ6VXNlcjI4MjkxODcw",
"organizations_url": "https://api.github.com/users/sumanthd17/orgs",
"received_events_url": "https://api.github.com/users/sumanthd17/received_events",
"repos_url": "https://api.github.com/users/sumanthd17/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/sumanthd17/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/sumanthd17/subscriptions",
"type": "User",
"url": "https://api.github.com/users/sumanthd17"
} | [] | closed | false | null | [] | null | 1 | "2020-11-20T17:09:34Z" | "2020-11-25T17:01:11Z" | "2020-11-25T15:26:07Z" | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/872.diff",
"html_url": "https://github.com/huggingface/datasets/pull/872",
"merged_at": "2020-11-25T15:26:07Z",
"patch_url": "https://github.com/huggingface/datasets/pull/872.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/872"
} | Added IndicGLUE benchmark for evaluating models on 11 Indian Languages. The descriptions of the tasks and the corresponding paper can be found [here](https://indicnlp.ai4bharat.org/indic-glue/)
- [x] Followed the instructions in CONTRIBUTING.md
- [x] Ran the tests successfully
- [x] Created the dummy data | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/872/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/872/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/871 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/871/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/871/comments | https://api.github.com/repos/huggingface/datasets/issues/871/events | https://github.com/huggingface/datasets/issues/871 | 747,470,136 | MDU6SXNzdWU3NDc0NzAxMzY= | 871 | terminate called after throwing an instance of 'google::protobuf::FatalException' | {
"avatar_url": "https://avatars.githubusercontent.com/u/6278280?v=4",
"events_url": "https://api.github.com/users/rabeehk/events{/privacy}",
"followers_url": "https://api.github.com/users/rabeehk/followers",
"following_url": "https://api.github.com/users/rabeehk/following{/other_user}",
"gists_url": "https://api.github.com/users/rabeehk/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/rabeehk",
"id": 6278280,
"login": "rabeehk",
"node_id": "MDQ6VXNlcjYyNzgyODA=",
"organizations_url": "https://api.github.com/users/rabeehk/orgs",
"received_events_url": "https://api.github.com/users/rabeehk/received_events",
"repos_url": "https://api.github.com/users/rabeehk/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/rabeehk/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/rabeehk/subscriptions",
"type": "User",
"url": "https://api.github.com/users/rabeehk"
} | [] | closed | false | null | [] | null | 2 | "2020-11-20T12:56:24Z" | "2020-12-12T21:16:32Z" | "2020-12-12T21:16:32Z" | CONTRIBUTOR | null | null | null | Hi
I am using the dataset "iwslt2017-en-nl", and after downloading it I am getting this error when trying to evaluate it on T5-base with seq2seq_trainer.py in the huggingface repo could you assist me please? thanks
100%|████████████████████████████████████████████████████████████████████████████████████████████████████| 63/63 [02:47<00:00, 2.18s/it][libprotobuf FATAL /sentencepiece/src/../third_party/protobuf-lite/google/protobuf/repeated_field.h:1505] CHECK failed: (index) >= (0):
terminate called after throwing an instance of 'google::protobuf::FatalException'
what(): CHECK failed: (index) >= (0):
run_t5_base_eval.sh: line 19: 5795 Aborted | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/871/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/871/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/870 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/870/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/870/comments | https://api.github.com/repos/huggingface/datasets/issues/870/events | https://github.com/huggingface/datasets/issues/870 | 747,021,996 | MDU6SXNzdWU3NDcwMjE5OTY= | 870 | [Feature Request] Add optional parameter in text loading script to preserve linebreaks | {
"avatar_url": "https://avatars.githubusercontent.com/u/31020859?v=4",
"events_url": "https://api.github.com/users/jncasey/events{/privacy}",
"followers_url": "https://api.github.com/users/jncasey/followers",
"following_url": "https://api.github.com/users/jncasey/following{/other_user}",
"gists_url": "https://api.github.com/users/jncasey/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/jncasey",
"id": 31020859,
"login": "jncasey",
"node_id": "MDQ6VXNlcjMxMDIwODU5",
"organizations_url": "https://api.github.com/users/jncasey/orgs",
"received_events_url": "https://api.github.com/users/jncasey/received_events",
"repos_url": "https://api.github.com/users/jncasey/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/jncasey/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jncasey/subscriptions",
"type": "User",
"url": "https://api.github.com/users/jncasey"
} | [
{
"color": "a2eeef",
"default": true,
"description": "New feature or request",
"id": 1935892871,
"name": "enhancement",
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement"
}
] | closed | false | null | [] | null | 2 | "2020-11-19T23:51:31Z" | "2022-06-01T15:25:53Z" | "2022-06-01T15:25:52Z" | NONE | null | null | null | I'm working on a project about rhyming verse using phonetic poetry and song lyrics, and line breaks are a vital part of the data.
I recently switched over to use the datasets library when my various corpora grew larger than my computer's memory. And so far, it is SO great.
But the first time I processed all of my data into a dataset, I hadn't realized the text loader script was processing the source files line-by-line and stripping off the newlines.
Once I caught the issue, I made my own data loader by modifying one line in the default text loader (changing `batch = batch.splitlines()` to `batch = batch.splitlines(True)` inside `_generate_tables`). And so I'm all set as far as my project is concerned.
But if my use case is more general, it seems like it'd be pretty trivial to add a kwarg to the default text loader called keeplinebreaks or something, which would default to False and get passed to `splitlines()`. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/870/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/870/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/869 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/869/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/869/comments | https://api.github.com/repos/huggingface/datasets/issues/869/events | https://github.com/huggingface/datasets/pull/869 | 746,495,711 | MDExOlB1bGxSZXF1ZXN0NTIzODc3OTkw | 869 | Update ner datasets infos | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq"
} | [] | closed | false | null | [] | null | 1 | "2020-11-19T11:28:03Z" | "2020-11-19T14:14:18Z" | "2020-11-19T14:14:17Z" | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/869.diff",
"html_url": "https://github.com/huggingface/datasets/pull/869",
"merged_at": "2020-11-19T14:14:17Z",
"patch_url": "https://github.com/huggingface/datasets/pull/869.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/869"
} | Update the dataset_infos.json files for changes made in #850 regarding the ner datasets feature types (and the change to ClassLabel)
I also fixed the ner types of conll2003 | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/869/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/869/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/868 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/868/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/868/comments | https://api.github.com/repos/huggingface/datasets/issues/868/events | https://github.com/huggingface/datasets/pull/868 | 745,889,882 | MDExOlB1bGxSZXF1ZXN0NTIzMzc2MzQ3 | 868 | Consistent metric outputs | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq"
} | [
{
"color": "E3165C",
"default": false,
"description": "",
"id": 4190228726,
"name": "transfer-to-evaluate",
"node_id": "LA_kwDODunzps75wdD2",
"url": "https://api.github.com/repos/huggingface/datasets/labels/transfer-to-evaluate"
}
] | closed | false | null | [] | null | 2 | "2020-11-18T18:05:59Z" | "2023-09-24T09:50:25Z" | "2023-07-11T09:35:52Z" | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/868.diff",
"html_url": "https://github.com/huggingface/datasets/pull/868",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/868.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/868"
} | To automate the use of metrics, they should return consistent outputs.
In particular I'm working on adding a conversion of metrics to keras metrics.
To achieve this we need two things:
- have each metric return dictionaries of string -> floats since each keras metrics should return one float
- define in the metric info the different fields of the output dictionary
In this PR I'm adding these two features.
I also fixed a few bugs in some metrics
#867 needs to be merged first | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/868/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/868/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/867 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/867/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/867/comments | https://api.github.com/repos/huggingface/datasets/issues/867/events | https://github.com/huggingface/datasets/pull/867 | 745,773,955 | MDExOlB1bGxSZXF1ZXN0NTIzMjc4MjI4 | 867 | Fix some metrics feature types | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq"
} | [] | closed | false | null | [] | null | 0 | "2020-11-18T15:46:11Z" | "2020-11-19T17:35:58Z" | "2020-11-19T17:35:57Z" | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/867.diff",
"html_url": "https://github.com/huggingface/datasets/pull/867",
"merged_at": "2020-11-19T17:35:57Z",
"patch_url": "https://github.com/huggingface/datasets/pull/867.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/867"
} | Replace `int` feature type to `int32` since `int` is not a pyarrow dtype in those metrics:
- accuracy
- precision
- recall
- f1
I also added the sklearn citation and used keyword arguments to remove future warnings | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/867/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/867/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/866 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/866/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/866/comments | https://api.github.com/repos/huggingface/datasets/issues/866/events | https://github.com/huggingface/datasets/issues/866 | 745,719,222 | MDU6SXNzdWU3NDU3MTkyMjI= | 866 | OSCAR from Inria group | {
"avatar_url": "https://avatars.githubusercontent.com/u/34098722?v=4",
"events_url": "https://api.github.com/users/jchwenger/events{/privacy}",
"followers_url": "https://api.github.com/users/jchwenger/followers",
"following_url": "https://api.github.com/users/jchwenger/following{/other_user}",
"gists_url": "https://api.github.com/users/jchwenger/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/jchwenger",
"id": 34098722,
"login": "jchwenger",
"node_id": "MDQ6VXNlcjM0MDk4NzIy",
"organizations_url": "https://api.github.com/users/jchwenger/orgs",
"received_events_url": "https://api.github.com/users/jchwenger/received_events",
"repos_url": "https://api.github.com/users/jchwenger/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/jchwenger/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jchwenger/subscriptions",
"type": "User",
"url": "https://api.github.com/users/jchwenger"
} | [
{
"color": "e99695",
"default": false,
"description": "Requesting to add a new dataset",
"id": 2067376369,
"name": "dataset request",
"node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5",
"url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request"
}
] | closed | false | null | [] | null | 2 | "2020-11-18T14:40:54Z" | "2020-11-18T15:01:30Z" | "2020-11-18T15:01:30Z" | NONE | null | null | null | ## Adding a Dataset
- **Name:** *OSCAR* (Open Super-large Crawled ALMAnaCH coRpus), multilingual parsing of Common Crawl (separate crawls for many different languages), [here](https://oscar-corpus.com/).
- **Description:** *OSCAR or Open Super-large Crawled ALMAnaCH coRpus is a huge multilingual corpus obtained by language classification and filtering of the Common Crawl corpus using the goclassy architecture.*
- **Paper:** *[here](https://hal.inria.fr/hal-02148693)*
- **Data:** *[here](https://oscar-corpus.com/)*
- **Motivation:** *useful for unsupervised tasks in separate languages. In an ideal world, your team would be able to obtain the unshuffled version, that could be used to train GPT-2-like models (the shuffled version, I suppose, could be used for translation).*
I am aware that you do offer the "colossal" Common Crawl dataset already, but this has the advantage to be available in many subcorpora for different languages.
| {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/866/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/866/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/865 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/865/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/865/comments | https://api.github.com/repos/huggingface/datasets/issues/865/events | https://github.com/huggingface/datasets/issues/865 | 745,430,497 | MDU6SXNzdWU3NDU0MzA0OTc= | 865 | Have Trouble importing `datasets` | {
"avatar_url": "https://avatars.githubusercontent.com/u/2755894?v=4",
"events_url": "https://api.github.com/users/forest1988/events{/privacy}",
"followers_url": "https://api.github.com/users/forest1988/followers",
"following_url": "https://api.github.com/users/forest1988/following{/other_user}",
"gists_url": "https://api.github.com/users/forest1988/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/forest1988",
"id": 2755894,
"login": "forest1988",
"node_id": "MDQ6VXNlcjI3NTU4OTQ=",
"organizations_url": "https://api.github.com/users/forest1988/orgs",
"received_events_url": "https://api.github.com/users/forest1988/received_events",
"repos_url": "https://api.github.com/users/forest1988/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/forest1988/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/forest1988/subscriptions",
"type": "User",
"url": "https://api.github.com/users/forest1988"
} | [] | closed | false | null | [] | null | 1 | "2020-11-18T08:04:41Z" | "2020-11-18T08:16:35Z" | "2020-11-18T08:16:35Z" | CONTRIBUTOR | null | null | null | I'm failing to import transformers (v4.0.0-dev), and tracing the cause seems to be failing to import datasets.
I cloned the newest version of datasets (master branch), and do `pip install -e .`.
Then, `import datasets` causes the error below.
```
~/workspace/Clone/datasets/src/datasets/utils/file_utils.py in <module>
116 sys.path.append(str(HF_MODULES_CACHE))
117
--> 118 os.makedirs(HF_MODULES_CACHE, exist_ok=True)
119 if not os.path.exists(os.path.join(HF_MODULES_CACHE, "__init__.py")):
120 with open(os.path.join(HF_MODULES_CACHE, "__init__.py"), "w"):
~/.pyenv/versions/anaconda3-2020.07/lib/python3.8/os.py in makedirs(name, mode, exist_ok)
221 return
222 try:
--> 223 mkdir(name, mode)
224 except OSError:
225 # Cannot rely on checking for EEXIST, since the operating system
FileNotFoundError: [Errno 2] No such file or directory: '<MY_HOME_DIRECTORY>/.cache/huggingface/modules'
```
The error occurs in `os.makedirs` in `file_utils.py`, even though `exist_ok = True` option is set.
(I use Python 3.8, so `exist_ok` is expected to work.)
I've checked some environment variables, and they are set as below.
```
*** NameError: name 'HF_MODULES_CACHE' is not defined
*** NameError: name 'hf_cache_home' is not defined
*** NameError: name 'XDG_CACHE_HOME' is not defined
```
Should I set some environment variables before using this library?
And, do you have any idea why "No such file or directory" occurs even though the `exist_ok = True` option is set?
Thank you in advance. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/865/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/865/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/864 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/864/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/864/comments | https://api.github.com/repos/huggingface/datasets/issues/864/events | https://github.com/huggingface/datasets/issues/864 | 745,322,357 | MDU6SXNzdWU3NDUzMjIzNTc= | 864 | Unable to download cnn_dailymail dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/46031058?v=4",
"events_url": "https://api.github.com/users/rohitashwa1907/events{/privacy}",
"followers_url": "https://api.github.com/users/rohitashwa1907/followers",
"following_url": "https://api.github.com/users/rohitashwa1907/following{/other_user}",
"gists_url": "https://api.github.com/users/rohitashwa1907/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/rohitashwa1907",
"id": 46031058,
"login": "rohitashwa1907",
"node_id": "MDQ6VXNlcjQ2MDMxMDU4",
"organizations_url": "https://api.github.com/users/rohitashwa1907/orgs",
"received_events_url": "https://api.github.com/users/rohitashwa1907/received_events",
"repos_url": "https://api.github.com/users/rohitashwa1907/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/rohitashwa1907/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/rohitashwa1907/subscriptions",
"type": "User",
"url": "https://api.github.com/users/rohitashwa1907"
} | [
{
"color": "2edb81",
"default": false,
"description": "A bug in a dataset script provided in the library",
"id": 2067388877,
"name": "dataset bug",
"node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug"
}
] | closed | false | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq"
} | [
{
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq"
}
] | null | 6 | "2020-11-18T04:38:02Z" | "2020-11-20T05:22:11Z" | "2020-11-20T05:22:10Z" | NONE | null | null | null | ### Script to reproduce the error
```
from datasets import load_dataset
train_dataset = load_dataset("cnn_dailymail", "3.0.0", split= 'train[:10%')
valid_dataset = load_dataset("cnn_dailymail","3.0.0", split="validation[:5%]")
```
### Error
```
---------------------------------------------------------------------------
NotADirectoryError Traceback (most recent call last)
<ipython-input-8-47c39c228935> in <module>()
1 from datasets import load_dataset
2
----> 3 train_dataset = load_dataset("cnn_dailymail", "3.0.0", split= 'train[:10%')
4 valid_dataset = load_dataset("cnn_dailymail","3.0.0", split="validation[:5%]")
5 frames
/usr/local/lib/python3.6/dist-packages/datasets/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, save_infos, script_version, **config_kwargs)
609 download_config=download_config,
610 download_mode=download_mode,
--> 611 ignore_verifications=ignore_verifications,
612 )
613
/usr/local/lib/python3.6/dist-packages/datasets/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, **download_and_prepare_kwargs)
469 if not downloaded_from_gcs:
470 self._download_and_prepare(
--> 471 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs
472 )
473 # Sync info
/usr/local/lib/python3.6/dist-packages/datasets/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs)
524 split_dict = SplitDict(dataset_name=self.name)
525 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs)
--> 526 split_generators = self._split_generators(dl_manager, **split_generators_kwargs)
527
528 # Checksums verification
/root/.cache/huggingface/modules/datasets_modules/datasets/cnn_dailymail/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602/cnn_dailymail.py in _split_generators(self, dl_manager)
252 def _split_generators(self, dl_manager):
253 dl_paths = dl_manager.download_and_extract(_DL_URLS)
--> 254 train_files = _subset_filenames(dl_paths, datasets.Split.TRAIN)
255 # Generate shared vocabulary
256
/root/.cache/huggingface/modules/datasets_modules/datasets/cnn_dailymail/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602/cnn_dailymail.py in _subset_filenames(dl_paths, split)
153 else:
154 logging.fatal("Unsupported split: %s", split)
--> 155 cnn = _find_files(dl_paths, "cnn", urls)
156 dm = _find_files(dl_paths, "dm", urls)
157 return cnn + dm
/root/.cache/huggingface/modules/datasets_modules/datasets/cnn_dailymail/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602/cnn_dailymail.py in _find_files(dl_paths, publisher, url_dict)
132 else:
133 logging.fatal("Unsupported publisher: %s", publisher)
--> 134 files = sorted(os.listdir(top_dir))
135
136 ret_files = []
NotADirectoryError: [Errno 20] Not a directory: '/root/.cache/huggingface/datasets/downloads/1bc05d24fa6dda2468e83a73cf6dc207226e01e3c48a507ea716dc0421da583b/cnn/stories'
```
Thanks for any suggestions. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/864/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/864/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/863 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/863/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/863/comments | https://api.github.com/repos/huggingface/datasets/issues/863/events | https://github.com/huggingface/datasets/pull/863 | 744,954,534 | MDExOlB1bGxSZXF1ZXN0NTIyNTk0Mjg1 | 863 | Add clear_cache parameter in the test command | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq"
} | [] | closed | false | null | [] | null | 0 | "2020-11-17T17:52:29Z" | "2020-11-18T14:44:25Z" | "2020-11-18T14:44:24Z" | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/863.diff",
"html_url": "https://github.com/huggingface/datasets/pull/863",
"merged_at": "2020-11-18T14:44:24Z",
"patch_url": "https://github.com/huggingface/datasets/pull/863.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/863"
} | For certain datasets like OSCAR #348 there are lots of different configurations and each one of them can take a lot of disk space.
I added a `--clear_cache` flag to the `datasets-cli test` command to be able to clear the cache after each configuration test to avoid filling up the disk. It should enable an easier generation for the `dataset_infos.json` file for OSCAR. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/863/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/863/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/862 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/862/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/862/comments | https://api.github.com/repos/huggingface/datasets/issues/862/events | https://github.com/huggingface/datasets/pull/862 | 744,906,131 | MDExOlB1bGxSZXF1ZXN0NTIyNTUzMzY1 | 862 | Update head requests | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq"
} | [] | closed | false | null | [] | null | 0 | "2020-11-17T16:49:06Z" | "2020-11-18T14:43:53Z" | "2020-11-18T14:43:50Z" | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/862.diff",
"html_url": "https://github.com/huggingface/datasets/pull/862",
"merged_at": "2020-11-18T14:43:50Z",
"patch_url": "https://github.com/huggingface/datasets/pull/862.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/862"
} | Get requests and Head requests didn't have the same parameters. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/862/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/862/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/861 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/861/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/861/comments | https://api.github.com/repos/huggingface/datasets/issues/861/events | https://github.com/huggingface/datasets/issues/861 | 744,753,458 | MDU6SXNzdWU3NDQ3NTM0NTg= | 861 | Possible Bug: Small training/dataset file creates gigantic output | {
"avatar_url": "https://avatars.githubusercontent.com/u/7240417?v=4",
"events_url": "https://api.github.com/users/NebelAI/events{/privacy}",
"followers_url": "https://api.github.com/users/NebelAI/followers",
"following_url": "https://api.github.com/users/NebelAI/following{/other_user}",
"gists_url": "https://api.github.com/users/NebelAI/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/NebelAI",
"id": 7240417,
"login": "NebelAI",
"node_id": "MDQ6VXNlcjcyNDA0MTc=",
"organizations_url": "https://api.github.com/users/NebelAI/orgs",
"received_events_url": "https://api.github.com/users/NebelAI/received_events",
"repos_url": "https://api.github.com/users/NebelAI/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/NebelAI/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/NebelAI/subscriptions",
"type": "User",
"url": "https://api.github.com/users/NebelAI"
} | [
{
"color": "a2eeef",
"default": true,
"description": "New feature or request",
"id": 1935892871,
"name": "enhancement",
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement"
},
{
"color": "d876e3",
"default": true,
"description": "Further information is requested",
"id": 1935892912,
"name": "question",
"node_id": "MDU6TGFiZWwxOTM1ODkyOTEy",
"url": "https://api.github.com/repos/huggingface/datasets/labels/question"
}
] | closed | false | {
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",
"gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/albertvillanova",
"id": 8515462,
"login": "albertvillanova",
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"organizations_url": "https://api.github.com/users/albertvillanova/orgs",
"received_events_url": "https://api.github.com/users/albertvillanova/received_events",
"repos_url": "https://api.github.com/users/albertvillanova/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions",
"type": "User",
"url": "https://api.github.com/users/albertvillanova"
} | [
{
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",
"gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/albertvillanova",
"id": 8515462,
"login": "albertvillanova",
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"organizations_url": "https://api.github.com/users/albertvillanova/orgs",
"received_events_url": "https://api.github.com/users/albertvillanova/received_events",
"repos_url": "https://api.github.com/users/albertvillanova/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions",
"type": "User",
"url": "https://api.github.com/users/albertvillanova"
}
] | null | 7 | "2020-11-17T13:48:59Z" | "2021-03-30T14:04:04Z" | "2021-03-22T12:04:55Z" | NONE | null | null | null | Hey guys,
I was trying to create a new bert model from scratch via _huggingface transformers + tokenizers + dataets_ (actually using this example script by your team: https://github.com/huggingface/transformers/blob/master/examples/language-modeling/run_mlm.py). It was supposed to be a first test with a small 5 GB raw text file but I can't even end the preprocessing handled by datasets because this tiny 5 GB text file becomes more than 1 TB when processing. My system was running out of space and crashed prematurely.
I've done training from scratch via Google's bert repo in the past and I can remember that the resulting pretraining data can become quite big. But 5 GB becoming 1 TB was never the case. Is this considered normal or is it a bug?
I've used the following CMD:
`python xla_spawn.py --num_cores=8 run_mlm.py --model_type bert --config_name config.json --tokenizer_name tokenizer.json --train_file dataset_full.txt --do_train --output_dir out --max_steps 500000 --save_steps 2500 --save_total_limit 2 --prediction_loss_only --line_by_line --max_seq_length 128 --pad_to_max_length --preprocessing_num_workers 16 --per_device_train_batch_size 128 --overwrite_output_dir --debug`
| {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/861/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/861/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/860 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/860/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/860/comments | https://api.github.com/repos/huggingface/datasets/issues/860/events | https://github.com/huggingface/datasets/issues/860 | 744,750,691 | MDU6SXNzdWU3NDQ3NTA2OTE= | 860 | wmt16 cs-en does not donwload | {
"avatar_url": "https://avatars.githubusercontent.com/u/6278280?v=4",
"events_url": "https://api.github.com/users/rabeehk/events{/privacy}",
"followers_url": "https://api.github.com/users/rabeehk/followers",
"following_url": "https://api.github.com/users/rabeehk/following{/other_user}",
"gists_url": "https://api.github.com/users/rabeehk/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/rabeehk",
"id": 6278280,
"login": "rabeehk",
"node_id": "MDQ6VXNlcjYyNzgyODA=",
"organizations_url": "https://api.github.com/users/rabeehk/orgs",
"received_events_url": "https://api.github.com/users/rabeehk/received_events",
"repos_url": "https://api.github.com/users/rabeehk/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/rabeehk/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/rabeehk/subscriptions",
"type": "User",
"url": "https://api.github.com/users/rabeehk"
} | [
{
"color": "2edb81",
"default": false,
"description": "A bug in a dataset script provided in the library",
"id": 2067388877,
"name": "dataset bug",
"node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug"
}
] | closed | false | null | [] | null | 1 | "2020-11-17T13:45:35Z" | "2022-10-05T12:27:00Z" | "2022-10-05T12:26:59Z" | CONTRIBUTOR | null | null | null | Hi
I am trying with wmt16, cs-en pair, thanks for the help, perhaps similar to the ro-en issue. thanks
split="train", n_obs=data_args.n_train) for task in data_args.task}
File "finetune_t5_trainer.py", line 109, in <dictcomp>
split="train", n_obs=data_args.n_train) for task in data_args.task}
File "/home/rabeeh/internship/seq2seq/tasks/tasks.py", line 82, in get_dataset
dataset = load_dataset("wmt16", self.pair, split=split)
File "/opt/conda/envs/internship/lib/python3.7/site-packages/datasets/load.py", line 611, in load_dataset
ignore_verifications=ignore_verifications,
File "/opt/conda/envs/internship/lib/python3.7/site-packages/datasets/builder.py", line 476, in download_and_prepare
dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs
File "/opt/conda/envs/internship/lib/python3.7/site-packages/datasets/builder.py", line 531, in _download_and_prepare
split_generators = self._split_generators(dl_manager, **split_generators_kwargs)
File "/home/rabeeh/.cache/huggingface/modules/datasets_modules/datasets/wmt16/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308/wmt_utils.py", line 755, in _split_generators
downloaded_files = dl_manager.download_and_extract(urls_to_download)
File "/opt/conda/envs/internship/lib/python3.7/site-packages/datasets/utils/download_manager.py", line 254, in download_and_extract
return self.extract(self.download(url_or_urls))
File "/opt/conda/envs/internship/lib/python3.7/site-packages/datasets/utils/download_manager.py", line 179, in download
num_proc=download_config.num_proc,
File "/opt/conda/envs/internship/lib/python3.7/site-packages/datasets/utils/py_utils.py", line 225, in map_nested
_single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)
File "/opt/conda/envs/internship/lib/python3.7/site-packages/datasets/utils/py_utils.py", line 225, in <listcomp>
_single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)
File "/opt/conda/envs/internship/lib/python3.7/site-packages/datasets/utils/py_utils.py", line 181, in _single_map_nested
mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]
File "/opt/conda/envs/internship/lib/python3.7/site-packages/datasets/utils/py_utils.py", line 181, in <listcomp>
mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]
File "/opt/conda/envs/internship/lib/python3.7/site-packages/datasets/utils/py_utils.py", line 163, in _single_map_nested
return function(data_struct)
File "/opt/conda/envs/internship/lib/python3.7/site-packages/datasets/utils/file_utils.py", line 308, in cached_path
use_etag=download_config.use_etag,
File "/opt/conda/envs/internship/lib/python3.7/site-packages/datasets/utils/file_utils.py", line 475, in get_from_cache
raise ConnectionError("Couldn't reach {}".format(url))
ConnectionError: Couldn't reach http://www.statmt.org/wmt13/training-parallel-commoncrawl.tgz | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/860/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/860/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/859 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/859/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/859/comments | https://api.github.com/repos/huggingface/datasets/issues/859/events | https://github.com/huggingface/datasets/pull/859 | 743,917,091 | MDExOlB1bGxSZXF1ZXN0NTIxNzI4MDM4 | 859 | Integrate file_lock inside the lib for better logging control | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq"
} | [] | closed | false | null | [] | null | 0 | "2020-11-16T15:13:39Z" | "2020-11-16T17:06:44Z" | "2020-11-16T17:06:42Z" | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/859.diff",
"html_url": "https://github.com/huggingface/datasets/pull/859",
"merged_at": "2020-11-16T17:06:42Z",
"patch_url": "https://github.com/huggingface/datasets/pull/859.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/859"
} | Previously the locking system of the lib was based on the file_lock package. However as noticed in #812 there were too many logs printed even when the datasets logging was set to warnings or errors.
For example
```python
import logging
logging.basicConfig(level=logging.INFO)
import datasets
datasets.set_verbosity_warning()
datasets.load_dataset("squad")
```
would still log the file lock events:
```
INFO:filelock:Lock 5737989232 acquired on /Users/quentinlhoest/.cache/huggingface/datasets/44801f118d500eff6114bfc56ab4e6def941f1eb14b70ac1ecc052e15cdac49d.85f43de978b9b25921cb78d7a2f2b350c04acdbaedb9ecb5f7101cd7c0950e68.py.lock
INFO:filelock:Lock 5737989232 released on /Users/quentinlhoest/.cache/huggingface/datasets/44801f118d500eff6114bfc56ab4e6def941f1eb14b70ac1ecc052e15cdac49d.85f43de978b9b25921cb78d7a2f2b350c04acdbaedb9ecb5f7101cd7c0950e68.py.lock
INFO:filelock:Lock 4393489968 acquired on /Users/quentinlhoest/.cache/huggingface/datasets/_Users_quentinlhoest_.cache_huggingface_datasets_squad_plain_text_1.0.0_1244d044b266a5e4dbd4174d23cb995eead372fbca31a03edc3f8a132787af41.lock
INFO:filelock:Lock 4393489968 released on /Users/quentinlhoest/.cache/huggingface/datasets/_Users_quentinlhoest_.cache_huggingface_datasets_squad_plain_text_1.0.0_1244d044b266a5e4dbd4174d23cb995eead372fbca31a03edc3f8a132787af41.lock
INFO:filelock:Lock 4393490808 acquired on /Users/quentinlhoest/.cache/huggingface/datasets/_Users_quentinlhoest_.cache_huggingface_datasets_squad_plain_text_1.0.0_1244d044b266a5e4dbd4174d23cb995eead372fbca31a03edc3f8a132787af41.lock
Reusing dataset squad (/Users/quentinlhoest/.cache/huggingface/datasets/squad/plain_text/1.0.0/1244d044b266a5e4dbd4174d23cb995eead372fbca31a03edc3f8a132787af41)
INFO:filelock:Lock 4393490808 released on /Users/quentinlhoest/.cache/huggingface/datasets/_Users_quentinlhoest_.cache_huggingface_datasets_squad_plain_text_1.0.0_1244d044b266a5e4dbd4174d23cb995eead372fbca31a03edc3f8a132787af41.lock
```
With the integration of file_lock in the library, the ouput is much cleaner:
```
Reusing dataset squad (/Users/quentinlhoest/.cache/huggingface/datasets/squad/plain_text/1.0.0/1244d044b266a5e4dbd4174d23cb995eead372fbca31a03edc3f8a132787af41)
```
Since the file_lock package is only a 450 lines file I think it's fine to have it inside the lib.
Fix #812 | {
"+1": 1,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/859/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/859/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/858 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/858/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/858/comments | https://api.github.com/repos/huggingface/datasets/issues/858/events | https://github.com/huggingface/datasets/pull/858 | 743,904,516 | MDExOlB1bGxSZXF1ZXN0NTIxNzE3ODQ4 | 858 | Add SemEval-2010 task 8 | {
"avatar_url": "https://avatars.githubusercontent.com/u/3775944?v=4",
"events_url": "https://api.github.com/users/JoelNiklaus/events{/privacy}",
"followers_url": "https://api.github.com/users/JoelNiklaus/followers",
"following_url": "https://api.github.com/users/JoelNiklaus/following{/other_user}",
"gists_url": "https://api.github.com/users/JoelNiklaus/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/JoelNiklaus",
"id": 3775944,
"login": "JoelNiklaus",
"node_id": "MDQ6VXNlcjM3NzU5NDQ=",
"organizations_url": "https://api.github.com/users/JoelNiklaus/orgs",
"received_events_url": "https://api.github.com/users/JoelNiklaus/received_events",
"repos_url": "https://api.github.com/users/JoelNiklaus/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/JoelNiklaus/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/JoelNiklaus/subscriptions",
"type": "User",
"url": "https://api.github.com/users/JoelNiklaus"
} | [] | closed | false | null | [] | null | 1 | "2020-11-16T14:57:57Z" | "2020-11-26T17:28:55Z" | "2020-11-26T17:28:55Z" | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/858.diff",
"html_url": "https://github.com/huggingface/datasets/pull/858",
"merged_at": "2020-11-26T17:28:55Z",
"patch_url": "https://github.com/huggingface/datasets/pull/858.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/858"
} | Hi,
I don't know how to add dummy data, since I create the validation set out of the last 1000 examples of the train set. If you have a suggestion, I am happy to implement it.
Cheers,
Joel | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/858/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/858/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/857 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/857/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/857/comments | https://api.github.com/repos/huggingface/datasets/issues/857/events | https://github.com/huggingface/datasets/pull/857 | 743,863,214 | MDExOlB1bGxSZXF1ZXN0NTIxNjg0ODIx | 857 | Use pandas reader in csv | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq"
} | [] | closed | false | null | [] | null | 0 | "2020-11-16T14:05:45Z" | "2020-11-19T17:35:40Z" | "2020-11-19T17:35:38Z" | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/857.diff",
"html_url": "https://github.com/huggingface/datasets/pull/857",
"merged_at": "2020-11-19T17:35:38Z",
"patch_url": "https://github.com/huggingface/datasets/pull/857.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/857"
} | The pyarrow CSV reader has issues that the pandas one doesn't (see #836 ).
To fix that I switched to the pandas csv reader.
The new reader is compatible with all the pandas parameters to read csv files.
Moreover it reads csv by chunk in order to save RAM, while the pyarrow one loads everything in memory.
Fix #836
Fix #794
Breaking: now all the parameters to read to csv file can be used in the `load_dataset` kwargs when loading csv, and the previous pyarrow objects `pyarrow.csv.ReadOptions`, `pyarrow.csv.ParseOptions` and `pyarrow.csv.ConvertOptions` are not used anymore. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/857/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/857/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/856 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/856/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/856/comments | https://api.github.com/repos/huggingface/datasets/issues/856/events | https://github.com/huggingface/datasets/pull/856 | 743,799,239 | MDExOlB1bGxSZXF1ZXN0NTIxNjMzNTYz | 856 | Add open book corpus | {
"avatar_url": "https://avatars.githubusercontent.com/u/458335?v=4",
"events_url": "https://api.github.com/users/vblagoje/events{/privacy}",
"followers_url": "https://api.github.com/users/vblagoje/followers",
"following_url": "https://api.github.com/users/vblagoje/following{/other_user}",
"gists_url": "https://api.github.com/users/vblagoje/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/vblagoje",
"id": 458335,
"login": "vblagoje",
"node_id": "MDQ6VXNlcjQ1ODMzNQ==",
"organizations_url": "https://api.github.com/users/vblagoje/orgs",
"received_events_url": "https://api.github.com/users/vblagoje/received_events",
"repos_url": "https://api.github.com/users/vblagoje/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/vblagoje/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/vblagoje/subscriptions",
"type": "User",
"url": "https://api.github.com/users/vblagoje"
} | [] | closed | false | null | [] | null | 21 | "2020-11-16T12:30:02Z" | "2024-01-04T13:20:51Z" | "2020-11-17T15:22:18Z" | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/856.diff",
"html_url": "https://github.com/huggingface/datasets/pull/856",
"merged_at": "2020-11-17T15:22:17Z",
"patch_url": "https://github.com/huggingface/datasets/pull/856.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/856"
} | Adds book corpus based on Shawn Presser's [work](https://github.com/soskek/bookcorpus/issues/27) @richarddwang, the author of the original BookCorpus dataset, suggested it should be named [OpenBookCorpus](https://github.com/huggingface/datasets/issues/486). I named it BookCorpusOpen to be easily located alphabetically. But, of course, we can rename it if needed.
It contains 17868 dataset items; each item contains two fields: title and text. The title is the name of the book (just the file name) while the text contains unprocessed book text. Note that bookcorpus is pre-segmented into a sentence while this bookcorpus is not. This is intentional (see https://github.com/huggingface/datasets/issues/486) as some users might want to further process the text themselves.
@lhoestq and others please review this PR thoroughly. cc @shawwn | {
"+1": 3,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 3,
"total_count": 6,
"url": "https://api.github.com/repos/huggingface/datasets/issues/856/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/856/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/855 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/855/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/855/comments | https://api.github.com/repos/huggingface/datasets/issues/855/events | https://github.com/huggingface/datasets/pull/855 | 743,690,839 | MDExOlB1bGxSZXF1ZXN0NTIxNTQ2Njkx | 855 | Fix kor nli csv reader | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq"
} | [] | closed | false | null | [] | null | 0 | "2020-11-16T09:53:41Z" | "2020-11-16T13:59:14Z" | "2020-11-16T13:59:12Z" | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/855.diff",
"html_url": "https://github.com/huggingface/datasets/pull/855",
"merged_at": "2020-11-16T13:59:12Z",
"patch_url": "https://github.com/huggingface/datasets/pull/855.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/855"
} | The kor_nli dataset had an issue with the csv reader that was not able to parse the lines correctly. Some lines were merged together for some reason.
I fixed that by iterating through the lines directly instead of using a csv reader.
I also changed the feature names to match the other NLI datasets (i.e. use "premise", "hypothesis", "label" features)
Fix #821 | {
"+1": 1,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/855/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/855/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/854 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/854/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/854/comments | https://api.github.com/repos/huggingface/datasets/issues/854/events | https://github.com/huggingface/datasets/issues/854 | 743,675,376 | MDU6SXNzdWU3NDM2NzUzNzY= | 854 | wmt16 does not download | {
"avatar_url": "https://avatars.githubusercontent.com/u/6278280?v=4",
"events_url": "https://api.github.com/users/rabeehk/events{/privacy}",
"followers_url": "https://api.github.com/users/rabeehk/followers",
"following_url": "https://api.github.com/users/rabeehk/following{/other_user}",
"gists_url": "https://api.github.com/users/rabeehk/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/rabeehk",
"id": 6278280,
"login": "rabeehk",
"node_id": "MDQ6VXNlcjYyNzgyODA=",
"organizations_url": "https://api.github.com/users/rabeehk/orgs",
"received_events_url": "https://api.github.com/users/rabeehk/received_events",
"repos_url": "https://api.github.com/users/rabeehk/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/rabeehk/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/rabeehk/subscriptions",
"type": "User",
"url": "https://api.github.com/users/rabeehk"
} | [
{
"color": "2edb81",
"default": false,
"description": "A bug in a dataset script provided in the library",
"id": 2067388877,
"name": "dataset bug",
"node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug"
}
] | closed | false | null | [] | null | 12 | "2020-11-16T09:31:51Z" | "2022-10-05T12:27:42Z" | "2022-10-05T12:27:42Z" | CONTRIBUTOR | null | null | null | Hi, I appreciate your help with the following error, thanks
>>> from datasets import load_dataset
>>> dataset = load_dataset("wmt16", "ro-en", split="train")
Downloading and preparing dataset wmt16/ro-en (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to /root/.cache/huggingface/datasets/wmt16/ro-en/1.0.0/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308...
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "/root/anaconda3/envs/pytorch/lib/python3.6/site-packages/datasets/load.py", line 611, in load_dataset
ignore_verifications=ignore_verifications,
File "/root/anaconda3/envs/pytorch/lib/python3.6/site-packages/datasets/builder.py", line 476, in download_and_prepare
dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs
File "/root/anaconda3/envs/pytorch/lib/python3.6/site-packages/datasets/builder.py", line 531, in _download_and_prepare
split_generators = self._split_generators(dl_manager, **split_generators_kwargs)
File "/root/.cache/huggingface/modules/datasets_modules/datasets/wmt16/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308/wmt_utils.py", line 755, in _split_generators
downloaded_files = dl_manager.download_and_extract(urls_to_download)
File "/root/anaconda3/envs/pytorch/lib/python3.6/site-packages/datasets/utils/download_manager.py", line 254, in download_and_extract
return self.extract(self.download(url_or_urls))
File "/root/anaconda3/envs/pytorch/lib/python3.6/site-packages/datasets/utils/download_manager.py", line 179, in download
num_proc=download_config.num_proc,
File "/root/anaconda3/envs/pytorch/lib/python3.6/site-packages/datasets/utils/py_utils.py", line 225, in map_nested
_single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)
File "/root/anaconda3/envs/pytorch/lib/python3.6/site-packages/datasets/utils/py_utils.py", line 225, in <listcomp>
_single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)
File "/root/anaconda3/envs/pytorch/lib/python3.6/site-packages/datasets/utils/py_utils.py", line 181, in _single_map_nested
mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]
File "/root/anaconda3/envs/pytorch/lib/python3.6/site-packages/datasets/utils/py_utils.py", line 181, in <listcomp>
mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]
File "/root/anaconda3/envs/pytorch/lib/python3.6/site-packages/datasets/utils/py_utils.py", line 163, in _single_map_nested
return function(data_struct)
File "/root/anaconda3/envs/pytorch/lib/python3.6/site-packages/datasets/utils/file_utils.py", line 308, in cached_path
use_etag=download_config.use_etag,
File "/root/anaconda3/envs/pytorch/lib/python3.6/site-packages/datasets/utils/file_utils.py", line 475, in get_from_cache
raise ConnectionError("Couldn't reach {}".format(url))
ConnectionError: Couldn't reach http://opus.nlpl.eu/download.php?f=SETIMES/v2/tmx/en-ro.tmx.gz | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/854/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/854/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/853 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/853/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/853/comments | https://api.github.com/repos/huggingface/datasets/issues/853/events | https://github.com/huggingface/datasets/issues/853 | 743,426,583 | MDU6SXNzdWU3NDM0MjY1ODM= | 853 | concatenate_datasets support axis=0 or 1 ? | {
"avatar_url": "https://avatars.githubusercontent.com/u/12437751?v=4",
"events_url": "https://api.github.com/users/renqingcolin/events{/privacy}",
"followers_url": "https://api.github.com/users/renqingcolin/followers",
"following_url": "https://api.github.com/users/renqingcolin/following{/other_user}",
"gists_url": "https://api.github.com/users/renqingcolin/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/renqingcolin",
"id": 12437751,
"login": "renqingcolin",
"node_id": "MDQ6VXNlcjEyNDM3NzUx",
"organizations_url": "https://api.github.com/users/renqingcolin/orgs",
"received_events_url": "https://api.github.com/users/renqingcolin/received_events",
"repos_url": "https://api.github.com/users/renqingcolin/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/renqingcolin/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/renqingcolin/subscriptions",
"type": "User",
"url": "https://api.github.com/users/renqingcolin"
} | [
{
"color": "a2eeef",
"default": true,
"description": "New feature or request",
"id": 1935892871,
"name": "enhancement",
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement"
},
{
"color": "008672",
"default": true,
"description": "Extra attention is needed",
"id": 1935892884,
"name": "help wanted",
"node_id": "MDU6TGFiZWwxOTM1ODkyODg0",
"url": "https://api.github.com/repos/huggingface/datasets/labels/help%20wanted"
},
{
"color": "d876e3",
"default": true,
"description": "Further information is requested",
"id": 1935892912,
"name": "question",
"node_id": "MDU6TGFiZWwxOTM1ODkyOTEy",
"url": "https://api.github.com/repos/huggingface/datasets/labels/question"
}
] | closed | false | {
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",
"gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/albertvillanova",
"id": 8515462,
"login": "albertvillanova",
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"organizations_url": "https://api.github.com/users/albertvillanova/orgs",
"received_events_url": "https://api.github.com/users/albertvillanova/received_events",
"repos_url": "https://api.github.com/users/albertvillanova/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions",
"type": "User",
"url": "https://api.github.com/users/albertvillanova"
} | [
{
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",
"gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/albertvillanova",
"id": 8515462,
"login": "albertvillanova",
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"organizations_url": "https://api.github.com/users/albertvillanova/orgs",
"received_events_url": "https://api.github.com/users/albertvillanova/received_events",
"repos_url": "https://api.github.com/users/albertvillanova/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions",
"type": "User",
"url": "https://api.github.com/users/albertvillanova"
}
] | null | 10 | "2020-11-16T02:46:23Z" | "2021-04-19T16:07:18Z" | "2021-04-19T16:07:18Z" | NONE | null | null | null | I want to achieve the following result

| {
"+1": 1,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/853/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/853/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/852 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/852/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/852/comments | https://api.github.com/repos/huggingface/datasets/issues/852/events | https://github.com/huggingface/datasets/issues/852 | 743,396,240 | MDU6SXNzdWU3NDMzOTYyNDA= | 852 | wmt cannot be downloaded | {
"avatar_url": "https://avatars.githubusercontent.com/u/6278280?v=4",
"events_url": "https://api.github.com/users/rabeehk/events{/privacy}",
"followers_url": "https://api.github.com/users/rabeehk/followers",
"following_url": "https://api.github.com/users/rabeehk/following{/other_user}",
"gists_url": "https://api.github.com/users/rabeehk/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/rabeehk",
"id": 6278280,
"login": "rabeehk",
"node_id": "MDQ6VXNlcjYyNzgyODA=",
"organizations_url": "https://api.github.com/users/rabeehk/orgs",
"received_events_url": "https://api.github.com/users/rabeehk/received_events",
"repos_url": "https://api.github.com/users/rabeehk/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/rabeehk/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/rabeehk/subscriptions",
"type": "User",
"url": "https://api.github.com/users/rabeehk"
} | [
{
"color": "e99695",
"default": false,
"description": "Requesting to add a new dataset",
"id": 2067376369,
"name": "dataset request",
"node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5",
"url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request"
}
] | closed | false | null | [] | null | 0 | "2020-11-16T01:04:41Z" | "2020-11-16T09:31:58Z" | "2020-11-16T09:31:58Z" | CONTRIBUTOR | null | null | null | Hi, I appreciate your help with the following error, thanks
>>> from datasets import load_dataset
>>> dataset = load_dataset("wmt16", "ro-en", split="train")
Downloading and preparing dataset wmt16/ro-en (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to /root/.cache/huggingface/datasets/wmt16/ro-en/1.0.0/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308...
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "/root/anaconda3/envs/pytorch/lib/python3.6/site-packages/datasets/load.py", line 611, in load_dataset
ignore_verifications=ignore_verifications,
File "/root/anaconda3/envs/pytorch/lib/python3.6/site-packages/datasets/builder.py", line 476, in download_and_prepare
dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs
File "/root/anaconda3/envs/pytorch/lib/python3.6/site-packages/datasets/builder.py", line 531, in _download_and_prepare
split_generators = self._split_generators(dl_manager, **split_generators_kwargs)
File "/root/.cache/huggingface/modules/datasets_modules/datasets/wmt16/7b2c4443a7d34c2e13df267eaa8cab4c62dd82f6b62b0d9ecc2e3a673ce17308/wmt_utils.py", line 755, in _split_generators
downloaded_files = dl_manager.download_and_extract(urls_to_download)
File "/root/anaconda3/envs/pytorch/lib/python3.6/site-packages/datasets/utils/download_manager.py", line 254, in download_and_extract
return self.extract(self.download(url_or_urls))
File "/root/anaconda3/envs/pytorch/lib/python3.6/site-packages/datasets/utils/download_manager.py", line 179, in download
num_proc=download_config.num_proc,
File "/root/anaconda3/envs/pytorch/lib/python3.6/site-packages/datasets/utils/py_utils.py", line 225, in map_nested
_single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)
File "/root/anaconda3/envs/pytorch/lib/python3.6/site-packages/datasets/utils/py_utils.py", line 225, in <listcomp>
_single_map_nested((function, obj, types, None, True)) for obj in tqdm(iterable, disable=disable_tqdm)
File "/root/anaconda3/envs/pytorch/lib/python3.6/site-packages/datasets/utils/py_utils.py", line 181, in _single_map_nested
mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]
File "/root/anaconda3/envs/pytorch/lib/python3.6/site-packages/datasets/utils/py_utils.py", line 181, in <listcomp>
mapped = [_single_map_nested((function, v, types, None, True)) for v in pbar]
File "/root/anaconda3/envs/pytorch/lib/python3.6/site-packages/datasets/utils/py_utils.py", line 163, in _single_map_nested
return function(data_struct)
File "/root/anaconda3/envs/pytorch/lib/python3.6/site-packages/datasets/utils/file_utils.py", line 308, in cached_path
use_etag=download_config.use_etag,
File "/root/anaconda3/envs/pytorch/lib/python3.6/site-packages/datasets/utils/file_utils.py", line 475, in get_from_cache
raise ConnectionError("Couldn't reach {}".format(url))
ConnectionError: Couldn't reach http://opus.nlpl.eu/download.php?f=SETIMES/v2/tmx/en-ro.tmx.gz | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/852/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/852/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/850 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/850/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/850/comments | https://api.github.com/repos/huggingface/datasets/issues/850/events | https://github.com/huggingface/datasets/pull/850 | 742,369,419 | MDExOlB1bGxSZXF1ZXN0NTIwNTE0MDY3 | 850 | Create ClassLabel for labelling tasks datasets | {
"avatar_url": "https://avatars.githubusercontent.com/u/959590?v=4",
"events_url": "https://api.github.com/users/jplu/events{/privacy}",
"followers_url": "https://api.github.com/users/jplu/followers",
"following_url": "https://api.github.com/users/jplu/following{/other_user}",
"gists_url": "https://api.github.com/users/jplu/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/jplu",
"id": 959590,
"login": "jplu",
"node_id": "MDQ6VXNlcjk1OTU5MA==",
"organizations_url": "https://api.github.com/users/jplu/orgs",
"received_events_url": "https://api.github.com/users/jplu/received_events",
"repos_url": "https://api.github.com/users/jplu/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/jplu/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jplu/subscriptions",
"type": "User",
"url": "https://api.github.com/users/jplu"
} | [] | closed | false | null | [] | null | 1 | "2020-11-13T11:07:22Z" | "2020-11-16T10:32:05Z" | "2020-11-16T10:31:58Z" | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/850.diff",
"html_url": "https://github.com/huggingface/datasets/pull/850",
"merged_at": "2020-11-16T10:31:58Z",
"patch_url": "https://github.com/huggingface/datasets/pull/850.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/850"
} | This PR adds a specific `ClassLabel` for the datasets that are about a labelling task such as POS, NER or Chunking. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/850/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/850/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/849 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/849/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/849/comments | https://api.github.com/repos/huggingface/datasets/issues/849/events | https://github.com/huggingface/datasets/issues/849 | 742,263,333 | MDU6SXNzdWU3NDIyNjMzMzM= | 849 | Load amazon dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/19718818?v=4",
"events_url": "https://api.github.com/users/bhavitvyamalik/events{/privacy}",
"followers_url": "https://api.github.com/users/bhavitvyamalik/followers",
"following_url": "https://api.github.com/users/bhavitvyamalik/following{/other_user}",
"gists_url": "https://api.github.com/users/bhavitvyamalik/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/bhavitvyamalik",
"id": 19718818,
"login": "bhavitvyamalik",
"node_id": "MDQ6VXNlcjE5NzE4ODE4",
"organizations_url": "https://api.github.com/users/bhavitvyamalik/orgs",
"received_events_url": "https://api.github.com/users/bhavitvyamalik/received_events",
"repos_url": "https://api.github.com/users/bhavitvyamalik/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/bhavitvyamalik/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/bhavitvyamalik/subscriptions",
"type": "User",
"url": "https://api.github.com/users/bhavitvyamalik"
} | [] | closed | false | null | [] | null | 1 | "2020-11-13T08:34:24Z" | "2020-11-17T07:22:59Z" | "2020-11-17T07:22:59Z" | CONTRIBUTOR | null | null | null | Hi,
I was going through amazon_us_reviews dataset and found that example API usage given on website is different from the API usage while loading dataset.
Eg. what API usage is on the [website](https://huggingface.co/datasets/amazon_us_reviews)
```
from datasets import load_dataset
dataset = load_dataset("amazon_us_reviews")
```
How it is when I tried (the error generated does point me to the right direction though)
```
from datasets import load_dataset
dataset = load_dataset("amazon_us_reviews", 'Books_v1_00')
```
Also, there is some issue with formatting as it's not showing bullet list in description with new line. Can I work on it? | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/849/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/849/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/848 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/848/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/848/comments | https://api.github.com/repos/huggingface/datasets/issues/848/events | https://github.com/huggingface/datasets/issues/848 | 742,240,942 | MDU6SXNzdWU3NDIyNDA5NDI= | 848 | Error when concatenate_datasets | {
"avatar_url": "https://avatars.githubusercontent.com/u/25664170?v=4",
"events_url": "https://api.github.com/users/shexuan/events{/privacy}",
"followers_url": "https://api.github.com/users/shexuan/followers",
"following_url": "https://api.github.com/users/shexuan/following{/other_user}",
"gists_url": "https://api.github.com/users/shexuan/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/shexuan",
"id": 25664170,
"login": "shexuan",
"node_id": "MDQ6VXNlcjI1NjY0MTcw",
"organizations_url": "https://api.github.com/users/shexuan/orgs",
"received_events_url": "https://api.github.com/users/shexuan/received_events",
"repos_url": "https://api.github.com/users/shexuan/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/shexuan/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/shexuan/subscriptions",
"type": "User",
"url": "https://api.github.com/users/shexuan"
} | [] | closed | false | null | [] | null | 4 | "2020-11-13T07:56:02Z" | "2020-11-13T17:40:59Z" | "2020-11-13T15:55:10Z" | NONE | null | null | null | Hello, when I concatenate two dataset loading from disk, I encountered a problem:
```
test_dataset = load_from_disk('data/test_dataset')
trn_dataset = load_from_disk('data/train_dataset')
train_dataset = concatenate_datasets([trn_dataset, test_dataset])
```
And it reported ValueError blow:
```
---------------------------------------------------------------------------
ValueError Traceback (most recent call last)
<ipython-input-38-74fa525512ca> in <module>
----> 1 train_dataset = concatenate_datasets([trn_dataset, test_dataset])
/opt/miniconda3/lib/python3.7/site-packages/datasets/arrow_dataset.py in concatenate_datasets(dsets, info, split)
2547 "However datasets' indices {} come from memory and datasets' indices {} come from disk.".format(
2548 [i for i in range(len(dsets)) if indices_mappings_in_memory[i]],
-> 2549 [i for i in range(len(dsets)) if not indices_mappings_in_memory[i]],
2550 )
2551 )
ValueError: Datasets' indices should ALL come from memory, or should ALL come from disk.
However datasets' indices [1] come from memory and datasets' indices [0] come from disk.
```
But it's curious both of my datasets loading from disk, so I check the source code in `arrow_dataset.py` about the Error:
```
trn_dataset._data_files
# output
[{'filename': 'data/train_dataset/csv-train.arrow', 'skip': 0, 'take': 593264}]
test_dataset._data_files
# output
[{'filename': 'data/test_dataset/csv-test.arrow', 'skip': 0, 'take': 424383}]
print([not dset._data_files for dset in [trn_dataset, test_dataset]])
# [False, False]
# And I tested the code the same as arrow_dataset, but nothing happened
dsets = [trn_dataset, test_dataset]
dsets_in_memory = [not dset._data_files for dset in dsets]
if any(dset_in_memory != dsets_in_memory[0] for dset_in_memory in dsets_in_memory):
raise ValueError(
"Datasets should ALL come from memory, or should ALL come from disk.\n"
"However datasets {} come from memory and datasets {} come from disk.".format(
[i for i in range(len(dsets)) if dsets_in_memory[i]],
[i for i in range(len(dsets)) if not dsets_in_memory[i]],
)
)
```
Any suggestions would be greatly appreciated!
Thanks! | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/848/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/848/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/847 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/847/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/847/comments | https://api.github.com/repos/huggingface/datasets/issues/847/events | https://github.com/huggingface/datasets/issues/847 | 742,179,495 | MDU6SXNzdWU3NDIxNzk0OTU= | 847 | multiprocessing in dataset map "can only test a child process" | {
"avatar_url": "https://avatars.githubusercontent.com/u/2000204?v=4",
"events_url": "https://api.github.com/users/timothyjlaurent/events{/privacy}",
"followers_url": "https://api.github.com/users/timothyjlaurent/followers",
"following_url": "https://api.github.com/users/timothyjlaurent/following{/other_user}",
"gists_url": "https://api.github.com/users/timothyjlaurent/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/timothyjlaurent",
"id": 2000204,
"login": "timothyjlaurent",
"node_id": "MDQ6VXNlcjIwMDAyMDQ=",
"organizations_url": "https://api.github.com/users/timothyjlaurent/orgs",
"received_events_url": "https://api.github.com/users/timothyjlaurent/received_events",
"repos_url": "https://api.github.com/users/timothyjlaurent/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/timothyjlaurent/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/timothyjlaurent/subscriptions",
"type": "User",
"url": "https://api.github.com/users/timothyjlaurent"
} | [] | closed | false | null | [] | null | 9 | "2020-11-13T06:01:04Z" | "2022-10-05T12:22:51Z" | "2022-10-05T12:22:51Z" | NONE | null | null | null | Using a dataset with a single 'text' field and a fast tokenizer in a jupyter notebook.
```
def tokenizer_fn(example):
return tokenizer.batch_encode_plus(example['text'])
ds_tokenized = text_dataset.map(tokenizer_fn, batched=True, num_proc=6, remove_columns=['text'])
```
```
---------------------------------------------------------------------------
RemoteTraceback Traceback (most recent call last)
RemoteTraceback:
"""
Traceback (most recent call last):
File "/home/jovyan/share/users/tlaurent/invitae-bert/ve/lib/python3.6/site-packages/multiprocess/pool.py", line 119, in worker
result = (True, func(*args, **kwds))
File "/home/jovyan/share/users/tlaurent/invitae-bert/ve/lib/python3.6/site-packages/datasets/arrow_dataset.py", line 156, in wrapper
out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs)
File "/home/jovyan/share/users/tlaurent/invitae-bert/ve/lib/python3.6/site-packages/datasets/fingerprint.py", line 163, in wrapper
out = func(self, *args, **kwargs)
File "/home/jovyan/share/users/tlaurent/invitae-bert/ve/lib/python3.6/site-packages/datasets/arrow_dataset.py", line 1510, in _map_single
for i in pbar:
File "/home/jovyan/share/users/tlaurent/invitae-bert/ve/lib/python3.6/site-packages/tqdm/notebook.py", line 228, in __iter__
for obj in super(tqdm_notebook, self).__iter__(*args, **kwargs):
File "/home/jovyan/share/users/tlaurent/invitae-bert/ve/lib/python3.6/site-packages/tqdm/std.py", line 1186, in __iter__
self.close()
File "/home/jovyan/share/users/tlaurent/invitae-bert/ve/lib/python3.6/site-packages/tqdm/notebook.py", line 251, in close
super(tqdm_notebook, self).close(*args, **kwargs)
File "/home/jovyan/share/users/tlaurent/invitae-bert/ve/lib/python3.6/site-packages/tqdm/std.py", line 1291, in close
fp_write('')
File "/home/jovyan/share/users/tlaurent/invitae-bert/ve/lib/python3.6/site-packages/tqdm/std.py", line 1288, in fp_write
self.fp.write(_unicode(s))
File "/home/jovyan/share/users/tlaurent/invitae-bert/ve/lib/python3.6/site-packages/wandb/sdk/lib/redirect.py", line 91, in new_write
cb(name, data)
File "/home/jovyan/share/users/tlaurent/invitae-bert/ve/lib/python3.6/site-packages/wandb/sdk/wandb_run.py", line 598, in _console_callback
self._backend.interface.publish_output(name, data)
File "/home/jovyan/share/users/tlaurent/invitae-bert/ve/lib/python3.6/site-packages/wandb/sdk/interface/interface.py", line 146, in publish_output
self._publish_output(o)
File "/home/jovyan/share/users/tlaurent/invitae-bert/ve/lib/python3.6/site-packages/wandb/sdk/interface/interface.py", line 151, in _publish_output
self._publish(rec)
File "/home/jovyan/share/users/tlaurent/invitae-bert/ve/lib/python3.6/site-packages/wandb/sdk/interface/interface.py", line 431, in _publish
if self._process and not self._process.is_alive():
File "/usr/lib/python3.6/multiprocessing/process.py", line 134, in is_alive
assert self._parent_pid == os.getpid(), 'can only test a child process'
AssertionError: can only test a child process
"""
``` | {
"+1": 1,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/847/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/847/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/846 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/846/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/846/comments | https://api.github.com/repos/huggingface/datasets/issues/846/events | https://github.com/huggingface/datasets/issues/846 | 741,885,174 | MDU6SXNzdWU3NDE4ODUxNzQ= | 846 | Add HoVer multi-hop fact verification dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4",
"events_url": "https://api.github.com/users/yjernite/events{/privacy}",
"followers_url": "https://api.github.com/users/yjernite/followers",
"following_url": "https://api.github.com/users/yjernite/following{/other_user}",
"gists_url": "https://api.github.com/users/yjernite/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/yjernite",
"id": 10469459,
"login": "yjernite",
"node_id": "MDQ6VXNlcjEwNDY5NDU5",
"organizations_url": "https://api.github.com/users/yjernite/orgs",
"received_events_url": "https://api.github.com/users/yjernite/received_events",
"repos_url": "https://api.github.com/users/yjernite/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/yjernite/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/yjernite/subscriptions",
"type": "User",
"url": "https://api.github.com/users/yjernite"
} | [
{
"color": "e99695",
"default": false,
"description": "Requesting to add a new dataset",
"id": 2067376369,
"name": "dataset request",
"node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5",
"url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request"
}
] | closed | false | null | [] | null | 3 | "2020-11-12T19:55:46Z" | "2020-12-10T21:47:33Z" | "2020-12-10T21:47:33Z" | MEMBER | null | null | null | ## Adding a Dataset
- **Name:** HoVer
- **Description:** https://twitter.com/YichenJiang9/status/1326954363806429186 contains 20K claim verification examples
- **Paper:** https://arxiv.org/abs/2011.03088
- **Data:** https://hover-nlp.github.io/
- **Motivation:** There are still few multi-hop information extraction benchmarks (HotpotQA, which dataset wase based off, notwithstanding)
Instructions to add a new dataset can be found [here](https://huggingface.co/docs/datasets/share_dataset.html).
| {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/846/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/846/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/845 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/845/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/845/comments | https://api.github.com/repos/huggingface/datasets/issues/845/events | https://github.com/huggingface/datasets/pull/845 | 741,841,350 | MDExOlB1bGxSZXF1ZXN0NTIwMDg1NDMy | 845 | amazon description fields as bullets | {
"avatar_url": "https://avatars.githubusercontent.com/u/9353833?v=4",
"events_url": "https://api.github.com/users/joeddav/events{/privacy}",
"followers_url": "https://api.github.com/users/joeddav/followers",
"following_url": "https://api.github.com/users/joeddav/following{/other_user}",
"gists_url": "https://api.github.com/users/joeddav/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/joeddav",
"id": 9353833,
"login": "joeddav",
"node_id": "MDQ6VXNlcjkzNTM4MzM=",
"organizations_url": "https://api.github.com/users/joeddav/orgs",
"received_events_url": "https://api.github.com/users/joeddav/received_events",
"repos_url": "https://api.github.com/users/joeddav/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/joeddav/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/joeddav/subscriptions",
"type": "User",
"url": "https://api.github.com/users/joeddav"
} | [] | closed | false | null | [] | null | 0 | "2020-11-12T18:50:41Z" | "2020-11-12T18:50:54Z" | "2020-11-12T18:50:54Z" | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/845.diff",
"html_url": "https://github.com/huggingface/datasets/pull/845",
"merged_at": "2020-11-12T18:50:54Z",
"patch_url": "https://github.com/huggingface/datasets/pull/845.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/845"
} | One more minor formatting change to amazon reviews's description (in addition to #844). Just reformatting the fields to display as a bulleted list in markdown. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/845/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/845/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/844 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/844/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/844/comments | https://api.github.com/repos/huggingface/datasets/issues/844/events | https://github.com/huggingface/datasets/pull/844 | 741,835,661 | MDExOlB1bGxSZXF1ZXN0NTIwMDgwNzM5 | 844 | add newlines to amazon desc | {
"avatar_url": "https://avatars.githubusercontent.com/u/9353833?v=4",
"events_url": "https://api.github.com/users/joeddav/events{/privacy}",
"followers_url": "https://api.github.com/users/joeddav/followers",
"following_url": "https://api.github.com/users/joeddav/following{/other_user}",
"gists_url": "https://api.github.com/users/joeddav/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/joeddav",
"id": 9353833,
"login": "joeddav",
"node_id": "MDQ6VXNlcjkzNTM4MzM=",
"organizations_url": "https://api.github.com/users/joeddav/orgs",
"received_events_url": "https://api.github.com/users/joeddav/received_events",
"repos_url": "https://api.github.com/users/joeddav/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/joeddav/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/joeddav/subscriptions",
"type": "User",
"url": "https://api.github.com/users/joeddav"
} | [] | closed | false | null | [] | null | 0 | "2020-11-12T18:41:20Z" | "2020-11-12T18:42:25Z" | "2020-11-12T18:42:21Z" | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/844.diff",
"html_url": "https://github.com/huggingface/datasets/pull/844",
"merged_at": "2020-11-12T18:42:21Z",
"patch_url": "https://github.com/huggingface/datasets/pull/844.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/844"
} | Just a quick formatting fix to hopefully make it render nicer on Viewer | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/844/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/844/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/843 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/843/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/843/comments | https://api.github.com/repos/huggingface/datasets/issues/843/events | https://github.com/huggingface/datasets/issues/843 | 741,531,121 | MDU6SXNzdWU3NDE1MzExMjE= | 843 | use_custom_baseline still produces errors for bertscore | {
"avatar_url": "https://avatars.githubusercontent.com/u/37921244?v=4",
"events_url": "https://api.github.com/users/penatbater/events{/privacy}",
"followers_url": "https://api.github.com/users/penatbater/followers",
"following_url": "https://api.github.com/users/penatbater/following{/other_user}",
"gists_url": "https://api.github.com/users/penatbater/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/penatbater",
"id": 37921244,
"login": "penatbater",
"node_id": "MDQ6VXNlcjM3OTIxMjQ0",
"organizations_url": "https://api.github.com/users/penatbater/orgs",
"received_events_url": "https://api.github.com/users/penatbater/received_events",
"repos_url": "https://api.github.com/users/penatbater/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/penatbater/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/penatbater/subscriptions",
"type": "User",
"url": "https://api.github.com/users/penatbater"
} | [
{
"color": "25b21e",
"default": false,
"description": "A bug in a metric script",
"id": 2067393914,
"name": "metric bug",
"node_id": "MDU6TGFiZWwyMDY3MzkzOTE0",
"url": "https://api.github.com/repos/huggingface/datasets/labels/metric%20bug"
}
] | closed | false | null | [] | null | 5 | "2020-11-12T11:44:32Z" | "2021-08-31T10:06:44Z" | "2021-02-09T14:21:48Z" | NONE | null | null | null | `metric = load_metric('bertscore')`
`a1 = "random sentences"`
`b1 = "random sentences"`
`metric.compute(predictions = [a1], references = [b1], lang = 'en')`
`Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "/home/stephen_chan/.local/lib/python3.6/site-packages/datasets/metric.py", line 393, in compute
output = self._compute(predictions=predictions, references=references, **kwargs)
File "/home/stephen_chan/.cache/huggingface/modules/datasets_modules/metrics/bertscore/361e597a01a41d6cf95d94bbfb01dea16261687abc0c6c74cc9930f80488f363/bertscore.py", line 108, in _compute
hashcode = bert_score.utils.get_hash(model_type, num_layers, idf, rescale_with_baseline)
TypeError: get_hash() missing 1 required positional argument: 'use_custom_baseline'`
Adding 'use_custom_baseline = False' as an argument produces this error
`Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "/home/stephen_chan/.local/lib/python3.6/site-packages/datasets/metric.py", line 393, in compute
output = self._compute(predictions=predictions, references=references, **kwargs)
TypeError: _compute() got an unexpected keyword argument 'use_custom_baseline'`
This is on Ubuntu 18.04, Python 3.6.9, datasets version 1.1.2 | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/843/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/843/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/842 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/842/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/842/comments | https://api.github.com/repos/huggingface/datasets/issues/842/events | https://github.com/huggingface/datasets/issues/842 | 741,208,428 | MDU6SXNzdWU3NDEyMDg0Mjg= | 842 | How to enable `.map()` pre-processing pipelines to support multi-node parallelism? | {
"avatar_url": "https://avatars.githubusercontent.com/u/66387198?v=4",
"events_url": "https://api.github.com/users/shangw-nvidia/events{/privacy}",
"followers_url": "https://api.github.com/users/shangw-nvidia/followers",
"following_url": "https://api.github.com/users/shangw-nvidia/following{/other_user}",
"gists_url": "https://api.github.com/users/shangw-nvidia/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/shangw-nvidia",
"id": 66387198,
"login": "shangw-nvidia",
"node_id": "MDQ6VXNlcjY2Mzg3MTk4",
"organizations_url": "https://api.github.com/users/shangw-nvidia/orgs",
"received_events_url": "https://api.github.com/users/shangw-nvidia/received_events",
"repos_url": "https://api.github.com/users/shangw-nvidia/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/shangw-nvidia/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/shangw-nvidia/subscriptions",
"type": "User",
"url": "https://api.github.com/users/shangw-nvidia"
} | [] | open | false | null | [] | null | 3 | "2020-11-12T02:04:38Z" | "2022-10-12T16:10:51Z" | null | NONE | null | null | null | Hi,
Currently, multiprocessing can be enabled for the `.map()` stages on a single node. However, in the case of multi-node training, (since more than one node would be available) I'm wondering if it's possible to extend the parallel processing among nodes, instead of only 1 node running the `.map()` while the other node is waiting for it to finish?
Thanks! | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/842/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/842/timeline | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/841 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/841/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/841/comments | https://api.github.com/repos/huggingface/datasets/issues/841/events | https://github.com/huggingface/datasets/issues/841 | 740,737,448 | MDU6SXNzdWU3NDA3Mzc0NDg= | 841 | Can not reuse datasets already downloaded | {
"avatar_url": "https://avatars.githubusercontent.com/u/30210529?v=4",
"events_url": "https://api.github.com/users/jc-hou/events{/privacy}",
"followers_url": "https://api.github.com/users/jc-hou/followers",
"following_url": "https://api.github.com/users/jc-hou/following{/other_user}",
"gists_url": "https://api.github.com/users/jc-hou/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/jc-hou",
"id": 30210529,
"login": "jc-hou",
"node_id": "MDQ6VXNlcjMwMjEwNTI5",
"organizations_url": "https://api.github.com/users/jc-hou/orgs",
"received_events_url": "https://api.github.com/users/jc-hou/received_events",
"repos_url": "https://api.github.com/users/jc-hou/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/jc-hou/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jc-hou/subscriptions",
"type": "User",
"url": "https://api.github.com/users/jc-hou"
} | [] | closed | false | null | [] | null | 2 | "2020-11-11T12:42:15Z" | "2020-11-11T18:17:16Z" | "2020-11-11T18:17:16Z" | NONE | null | null | null | Hello,
I need to connect to a frontal node (with http proxy, no gpu) before connecting to a gpu node (but no http proxy, so can not use wget so on).
I successfully downloaded and reuse the wikipedia datasets in a frontal node.
When I connect to the gpu node, I supposed to use the downloaded datasets from cache, but failed and end with time out error.
On frontal node:
```
>>> from datasets import load_dataset
>>> dataset = load_dataset('wikipedia', '20200501.en')
Reusing dataset wikipedia (/linkhome/rech/genini01/uua34ms/.cache/huggingface/datasets/wikipedia/20200501.en/1.0.0/f92599dfccab29832c442b82870fa8f6983e5b4ebbf5e6e2dcbe894e325339cd)
/linkhome/rech/genini01/uua34ms/work/anaconda3/envs/pytorch_pip170_cuda102/lib/python3.6/site-packages/torch/cuda/__init__.py:52: UserWarning: CUDA initialization: Found no NVIDIA driver on your system. Please check that you have an NVIDIA GPU and installed a driver from http://www.nvidia.com/Download/index.aspx (Triggered internally at /pytorch/c10/cuda/CUDAFunctions.cpp:100.)
return torch._C._cuda_getDeviceCount() > 0
```
On gpu node:
```
>>> from datasets import load_dataset
>>> dataset = load_dataset('wikipedia', '20200501.en')
Traceback (most recent call last):
File "/linkhome/rech/genini01/uua34ms/work/anaconda3/envs/pytorch_pip170_cuda102/lib/python3.6/site-packages/urllib3/connection.py", line 160, in _new_conn
(self._dns_host, self.port), self.timeout, **extra_kw
File "/linkhome/rech/genini01/uua34ms/work/anaconda3/envs/pytorch_pip170_cuda102/lib/python3.6/site-packages/urllib3/util/connection.py", line 84, in create_connection
raise err
File "/linkhome/rech/genini01/uua34ms/work/anaconda3/envs/pytorch_pip170_cuda102/lib/python3.6/site-packages/urllib3/util/connection.py", line 74, in create_connection
sock.connect(sa)
TimeoutError: [Errno 110] Connection timed out
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/linkhome/rech/genini01/uua34ms/work/anaconda3/envs/pytorch_pip170_cuda102/lib/python3.6/site-packages/urllib3/connectionpool.py", line 677, in urlopen
chunked=chunked,
File "/linkhome/rech/genini01/uua34ms/work/anaconda3/envs/pytorch_pip170_cuda102/lib/python3.6/site-packages/urllib3/connectionpool.py", line 381, in _make_request
self._validate_conn(conn)
File "/linkhome/rech/genini01/uua34ms/work/anaconda3/envs/pytorch_pip170_cuda102/lib/python3.6/site-packages/urllib3/connectionpool.py", line 978, in _validate_conn
conn.connect()
File "/linkhome/rech/genini01/uua34ms/work/anaconda3/envs/pytorch_pip170_cuda102/lib/python3.6/site-packages/urllib3/connection.py", line 309, in connect
conn = self._new_conn()
File "/linkhome/rech/genini01/uua34ms/work/anaconda3/envs/pytorch_pip170_cuda102/lib/python3.6/site-packages/urllib3/connection.py", line 172, in _new_conn
self, "Failed to establish a new connection: %s" % e
urllib3.exceptions.NewConnectionError: <urllib3.connection.HTTPSConnection object at 0x14b7b73e4908>: Failed to establish a new connection: [Errno 110] Connection timed out
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/linkhome/rech/genini01/uua34ms/work/anaconda3/envs/pytorch_pip170_cuda102/lib/python3.6/site-packages/requests/adapters.py", line 449, in send
timeout=timeout
File "/linkhome/rech/genini01/uua34ms/work/anaconda3/envs/pytorch_pip170_cuda102/lib/python3.6/site-packages/urllib3/connectionpool.py", line 727, in urlopen
method, url, error=e, _pool=self, _stacktrace=sys.exc_info()[2]
File "/linkhome/rech/genini01/uua34ms/work/anaconda3/envs/pytorch_pip170_cuda102/lib/python3.6/site-packages/urllib3/util/retry.py", line 446, in increment
raise MaxRetryError(_pool, url, error or ResponseError(cause))
urllib3.exceptions.MaxRetryError: HTTPSConnectionPool(host='s3.amazonaws.com', port=443): Max retries exceeded with url: /datasets.huggingface.co/datasets/datasets/wikipedia/wikipedia.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x14b7b73e4908>: Failed to establish a new connection: [Errno 110] Connection timed out',))
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "/linkhome/rech/genini01/uua34ms/work/anaconda3/envs/pytorch_pip170_cuda102/lib/python3.6/site-packages/datasets/load.py", line 590, in load_dataset
path, script_version=script_version, download_config=download_config, download_mode=download_mode, dataset=True
File "/linkhome/rech/genini01/uua34ms/work/anaconda3/envs/pytorch_pip170_cuda102/lib/python3.6/site-packages/datasets/load.py", line 264, in prepare_module
head_hf_s3(path, filename=name, dataset=dataset)
File "/linkhome/rech/genini01/uua34ms/work/anaconda3/envs/pytorch_pip170_cuda102/lib/python3.6/site-packages/datasets/utils/file_utils.py", line 200, in head_hf_s3
return requests.head(hf_bucket_url(identifier=identifier, filename=filename, use_cdn=use_cdn, dataset=dataset))
File "/linkhome/rech/genini01/uua34ms/work/anaconda3/envs/pytorch_pip170_cuda102/lib/python3.6/site-packages/requests/api.py", line 104, in head
return request('head', url, **kwargs)
File "/linkhome/rech/genini01/uua34ms/work/anaconda3/envs/pytorch_pip170_cuda102/lib/python3.6/site-packages/requests/api.py", line 61, in request
return session.request(method=method, url=url, **kwargs)
File "/linkhome/rech/genini01/uua34ms/work/anaconda3/envs/pytorch_pip170_cuda102/lib/python3.6/site-packages/requests/sessions.py", line 530, in request
resp = self.send(prep, **send_kwargs)
File "/linkhome/rech/genini01/uua34ms/work/anaconda3/envs/pytorch_pip170_cuda102/lib/python3.6/site-packages/requests/sessions.py", line 643, in send
r = adapter.send(request, **kwargs)
File "/linkhome/rech/genini01/uua34ms/work/anaconda3/envs/pytorch_pip170_cuda102/lib/python3.6/site-packages/requests/adapters.py", line 516, in send
raise ConnectionError(e, request=request)
requests.exceptions.ConnectionError: HTTPSConnectionPool(host='s3.amazonaws.com', port=443): Max retries exceeded with url: /datasets.huggingface.co/datasets/datasets/wikipedia/wikipedia.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x14b7b73e4908>: Failed to establish a new connection: [Errno 110] Connection timed out',))
```
Any advice?Thanks!
| {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/841/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/841/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/840 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/840/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/840/comments | https://api.github.com/repos/huggingface/datasets/issues/840/events | https://github.com/huggingface/datasets/pull/840 | 740,632,771 | MDExOlB1bGxSZXF1ZXN0NTE5MDg2NDUw | 840 | Update squad_v2.py | {
"avatar_url": "https://avatars.githubusercontent.com/u/38747614?v=4",
"events_url": "https://api.github.com/users/Javier-Jimenez99/events{/privacy}",
"followers_url": "https://api.github.com/users/Javier-Jimenez99/followers",
"following_url": "https://api.github.com/users/Javier-Jimenez99/following{/other_user}",
"gists_url": "https://api.github.com/users/Javier-Jimenez99/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/Javier-Jimenez99",
"id": 38747614,
"login": "Javier-Jimenez99",
"node_id": "MDQ6VXNlcjM4NzQ3NjE0",
"organizations_url": "https://api.github.com/users/Javier-Jimenez99/orgs",
"received_events_url": "https://api.github.com/users/Javier-Jimenez99/received_events",
"repos_url": "https://api.github.com/users/Javier-Jimenez99/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/Javier-Jimenez99/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Javier-Jimenez99/subscriptions",
"type": "User",
"url": "https://api.github.com/users/Javier-Jimenez99"
} | [] | closed | false | null | [] | null | 2 | "2020-11-11T09:58:41Z" | "2020-11-11T15:29:34Z" | "2020-11-11T15:26:35Z" | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/840.diff",
"html_url": "https://github.com/huggingface/datasets/pull/840",
"merged_at": "2020-11-11T15:26:35Z",
"patch_url": "https://github.com/huggingface/datasets/pull/840.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/840"
} | Change lines 100 and 102 to prevent overwriting ```predictions``` variable. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/840/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/840/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/839 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/839/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/839/comments | https://api.github.com/repos/huggingface/datasets/issues/839/events | https://github.com/huggingface/datasets/issues/839 | 740,355,270 | MDU6SXNzdWU3NDAzNTUyNzA= | 839 | XSum dataset missing spaces between sentences | {
"avatar_url": "https://avatars.githubusercontent.com/u/10007282?v=4",
"events_url": "https://api.github.com/users/loganlebanoff/events{/privacy}",
"followers_url": "https://api.github.com/users/loganlebanoff/followers",
"following_url": "https://api.github.com/users/loganlebanoff/following{/other_user}",
"gists_url": "https://api.github.com/users/loganlebanoff/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/loganlebanoff",
"id": 10007282,
"login": "loganlebanoff",
"node_id": "MDQ6VXNlcjEwMDA3Mjgy",
"organizations_url": "https://api.github.com/users/loganlebanoff/orgs",
"received_events_url": "https://api.github.com/users/loganlebanoff/received_events",
"repos_url": "https://api.github.com/users/loganlebanoff/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/loganlebanoff/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/loganlebanoff/subscriptions",
"type": "User",
"url": "https://api.github.com/users/loganlebanoff"
} | [] | open | false | null | [] | null | 0 | "2020-11-11T00:34:43Z" | "2020-11-11T00:34:43Z" | null | NONE | null | null | null | I noticed that the XSum dataset has no space between sentences. This could lead to worse results for anyone training or testing on it. Here's an example (0th entry in the test set):
`The London trio are up for best UK act and best album, as well as getting two nominations in the best song category."We got told like this morning 'Oh I think you're nominated'", said Dappy."And I was like 'Oh yeah, which one?' And now we've got nominated for four awards. I mean, wow!"Bandmate Fazer added: "We thought it's best of us to come down and mingle with everyone and say hello to the cameras. And now we find we've got four nominations."The band have two shots at the best song prize, getting the nod for their Tynchy Stryder collaboration Number One, and single Strong Again.Their album Uncle B will also go up against records by the likes of Beyonce and Kanye West.N-Dubz picked up the best newcomer Mobo in 2007, but female member Tulisa said they wouldn't be too disappointed if they didn't win this time around."At the end of the day we're grateful to be where we are in our careers."If it don't happen then it don't happen - live to fight another day and keep on making albums and hits for the fans."Dappy also revealed they could be performing live several times on the night.The group will be doing Number One and also a possible rendition of the War Child single, I Got Soul.The charity song is a re-working of The Killers' All These Things That I've Done and is set to feature artists like Chipmunk, Ironik and Pixie Lott.This year's Mobos will be held outside of London for the first time, in Glasgow on 30 September.N-Dubz said they were looking forward to performing for their Scottish fans and boasted about their recent shows north of the border."We just done Edinburgh the other day," said Dappy."We smashed up an N-Dubz show over there. We done Aberdeen about three or four months ago - we smashed up that show over there! Everywhere we go we smash it up!"` | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/839/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/839/timeline | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/838 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/838/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/838/comments | https://api.github.com/repos/huggingface/datasets/issues/838/events | https://github.com/huggingface/datasets/pull/838 | 740,328,382 | MDExOlB1bGxSZXF1ZXN0NTE4ODM0NTE5 | 838 | CNN/Dailymail Dataset Card | {
"avatar_url": "https://avatars.githubusercontent.com/u/26722925?v=4",
"events_url": "https://api.github.com/users/mcmillanmajora/events{/privacy}",
"followers_url": "https://api.github.com/users/mcmillanmajora/followers",
"following_url": "https://api.github.com/users/mcmillanmajora/following{/other_user}",
"gists_url": "https://api.github.com/users/mcmillanmajora/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/mcmillanmajora",
"id": 26722925,
"login": "mcmillanmajora",
"node_id": "MDQ6VXNlcjI2NzIyOTI1",
"organizations_url": "https://api.github.com/users/mcmillanmajora/orgs",
"received_events_url": "https://api.github.com/users/mcmillanmajora/received_events",
"repos_url": "https://api.github.com/users/mcmillanmajora/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/mcmillanmajora/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mcmillanmajora/subscriptions",
"type": "User",
"url": "https://api.github.com/users/mcmillanmajora"
} | [] | closed | false | null | [] | null | 0 | "2020-11-10T23:56:43Z" | "2020-11-25T21:09:51Z" | "2020-11-25T21:09:50Z" | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/838.diff",
"html_url": "https://github.com/huggingface/datasets/pull/838",
"merged_at": "2020-11-25T21:09:50Z",
"patch_url": "https://github.com/huggingface/datasets/pull/838.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/838"
} | Link to the card page: https://github.com/mcmillanmajora/datasets/tree/cnn_dailymail_card/datasets/cnn_dailymail
One of the questions this dataset brings up is how we want to handle versioning of the cards to mirror versions of the dataset. The different versions of this dataset are used for different tasks (which may not be reflected in the versions that we currently have in the repo?), but it's only the structure that's changing rather than the content in this particular case, at least between versions 2.0.0 and 3.0.0. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/838/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/838/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/837 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/837/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/837/comments | https://api.github.com/repos/huggingface/datasets/issues/837/events | https://github.com/huggingface/datasets/pull/837 | 740,250,215 | MDExOlB1bGxSZXF1ZXN0NTE4NzcwNDM5 | 837 | AlloCiné dataset card | {
"avatar_url": "https://avatars.githubusercontent.com/u/26722925?v=4",
"events_url": "https://api.github.com/users/mcmillanmajora/events{/privacy}",
"followers_url": "https://api.github.com/users/mcmillanmajora/followers",
"following_url": "https://api.github.com/users/mcmillanmajora/following{/other_user}",
"gists_url": "https://api.github.com/users/mcmillanmajora/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/mcmillanmajora",
"id": 26722925,
"login": "mcmillanmajora",
"node_id": "MDQ6VXNlcjI2NzIyOTI1",
"organizations_url": "https://api.github.com/users/mcmillanmajora/orgs",
"received_events_url": "https://api.github.com/users/mcmillanmajora/received_events",
"repos_url": "https://api.github.com/users/mcmillanmajora/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/mcmillanmajora/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mcmillanmajora/subscriptions",
"type": "User",
"url": "https://api.github.com/users/mcmillanmajora"
} | [] | closed | false | null | [] | null | 0 | "2020-11-10T21:19:53Z" | "2020-11-25T21:56:27Z" | "2020-11-25T21:56:27Z" | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/837.diff",
"html_url": "https://github.com/huggingface/datasets/pull/837",
"merged_at": "2020-11-25T21:56:27Z",
"patch_url": "https://github.com/huggingface/datasets/pull/837.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/837"
} | Link to the card page: https://github.com/mcmillanmajora/datasets/blob/allocine_card/datasets/allocine/README.md
There wasn't as much information available for this dataset, so I'm wondering what's the best way to address open questions about the dataset. For example, where did the list of films that the dataset creator used come from?
I'm also wondering how best to go about talking about limitations when so little is known about the data. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/837/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/837/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/836 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/836/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/836/comments | https://api.github.com/repos/huggingface/datasets/issues/836/events | https://github.com/huggingface/datasets/issues/836 | 740,187,613 | MDU6SXNzdWU3NDAxODc2MTM= | 836 | load_dataset with 'csv' is not working. while the same file is loading with 'text' mode or with pandas | {
"avatar_url": "https://avatars.githubusercontent.com/u/8919490?v=4",
"events_url": "https://api.github.com/users/randubin/events{/privacy}",
"followers_url": "https://api.github.com/users/randubin/followers",
"following_url": "https://api.github.com/users/randubin/following{/other_user}",
"gists_url": "https://api.github.com/users/randubin/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/randubin",
"id": 8919490,
"login": "randubin",
"node_id": "MDQ6VXNlcjg5MTk0OTA=",
"organizations_url": "https://api.github.com/users/randubin/orgs",
"received_events_url": "https://api.github.com/users/randubin/received_events",
"repos_url": "https://api.github.com/users/randubin/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/randubin/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/randubin/subscriptions",
"type": "User",
"url": "https://api.github.com/users/randubin"
} | [
{
"color": "2edb81",
"default": false,
"description": "A bug in a dataset script provided in the library",
"id": 2067388877,
"name": "dataset bug",
"node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug"
}
] | closed | false | null | [] | null | 8 | "2020-11-10T19:35:40Z" | "2021-11-24T16:59:19Z" | "2020-11-19T17:35:38Z" | NONE | null | null | null | Hi All
I am trying to load a custom dataset and I am trying to load a single file to make sure the file is loading correctly:
dataset = load_dataset('csv', data_files=files)
When I run it I get:
Downloading and preparing dataset csv/default-35575a1051604c88 (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) tocache/huggingface/datasets/csv/default-35575a1051604c88/0.0.0/49187751790fa4d820300fd4d0707896e5b941f1a9c644652645b866716a4ac4...
I am getting this error:
6a4ac4/csv.py in _generate_tables(self, files)
78 def _generate_tables(self, files):
79 for i, file in enumerate(files):
---> 80 pa_table = pac.read_csv(
81 file,
82 read_options=self.config.pa_read_options,
~/anaconda2/envs/nlp/lib/python3.8/site-packages/pyarrow/_csv.pyx in pyarrow._csv.read_csv()
~/anaconda2/envs/nlp/lib/python3.8/site-packages/pyarrow/error.pxi in pyarrow.lib.pyarrow_internal_check_status()
~/anaconda2/envs/nlp/lib/python3.8/site-packages/pyarrow/error.pxi in pyarrow.lib.check_status()
**ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?)**
The size of the file is 3.5 GB. When I try smaller files I do not have an issue. When I load it with 'text' parser I can see all data but it is not what I need.
There is no issue reading the file with pandas. any idea what could be the issue?
When I am running a different CSV I do not get this line:
(download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size)
Any ideas?
| {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/836/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/836/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/835 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/835/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/835/comments | https://api.github.com/repos/huggingface/datasets/issues/835/events | https://github.com/huggingface/datasets/issues/835 | 740,102,210 | MDU6SXNzdWU3NDAxMDIyMTA= | 835 | Wikipedia postprocessing | {
"avatar_url": "https://avatars.githubusercontent.com/u/13353204?v=4",
"events_url": "https://api.github.com/users/bminixhofer/events{/privacy}",
"followers_url": "https://api.github.com/users/bminixhofer/followers",
"following_url": "https://api.github.com/users/bminixhofer/following{/other_user}",
"gists_url": "https://api.github.com/users/bminixhofer/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/bminixhofer",
"id": 13353204,
"login": "bminixhofer",
"node_id": "MDQ6VXNlcjEzMzUzMjA0",
"organizations_url": "https://api.github.com/users/bminixhofer/orgs",
"received_events_url": "https://api.github.com/users/bminixhofer/received_events",
"repos_url": "https://api.github.com/users/bminixhofer/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/bminixhofer/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/bminixhofer/subscriptions",
"type": "User",
"url": "https://api.github.com/users/bminixhofer"
} | [] | closed | false | null | [] | null | 3 | "2020-11-10T17:26:38Z" | "2020-11-10T18:23:20Z" | "2020-11-10T17:49:21Z" | NONE | null | null | null | Hi, thanks for this library!
Running this code:
```py
import datasets
wikipedia = datasets.load_dataset("wikipedia", "20200501.de")
print(wikipedia['train']['text'][0])
```
I get:
```
mini|Ricardo Flores Magón
mini|Mexikanische Revolutionäre, Magón in der Mitte anführend, gegen die Diktatur von Porfirio Diaz, Ausschnitt des Gemälde „Tierra y Libertad“ von Idelfonso Carrara (?) von 1930.
Ricardo Flores Magón (* 16. September 1874 in San Antonio Eloxochitlán im mexikanischen Bundesstaat Oaxaca; † 22. November 1922 im Bundesgefängnis Leavenworth im US-amerikanischen Bundesstaat Kansas) war als Journalist, Gewerkschafter und Literat ein führender anarchistischer Theoretiker und Aktivist, der die revolutionäre mexikanische Bewegung radikal beeinflusste. Magón war Gründer der Partido Liberal Mexicano und Mitglied der Industrial Workers of the World.
Politische Biografie
Journalistisch und politisch kämpfte er und sein Bruder sehr kompromisslos gegen die Diktatur Porfirio Diaz. Philosophisch und politisch orientiert an radikal anarchistischen Idealen und den Erfahrungen seiner indigenen Vorfahren bei der gemeinschaftlichen Bewirtschaftung des Gemeindelandes, machte er die Forderung „Land und Freiheit“ (Tierra y Libertad) populär. Besonders Francisco Villa und Emiliano Zapata griffen die Forderung Land und Freiheit auf. Seine Philosophie hatte großen Einfluss auf die Landarbeiter. 1904 floh er in die USA und gründete 1906 die Partido Liberal Mexicano. Im Exil lernte er u. a. Emma Goldman kennen. Er verbrachte die meiste Zeit seines Lebens in Gefängnissen und im Exil und wurde 1918 in den USA wegen „Behinderung der Kriegsanstrengungen“ zu zwanzig Jahren Gefängnis verurteilt. Zu seinem Tod gibt es drei verschiedene Theorien. Offiziell starb er an Herzversagen. Librado Rivera, der die Leiche mit eigenen Augen gesehen hat, geht davon aus, dass Magón von einem Mitgefangenen erdrosselt wurde. Die staatstreue Gewerkschaftszeitung CROM veröffentlichte 1923 einen Beitrag, nachdem Magón von einem Gefängniswärter erschlagen wurde.
mini|Die Brüder Ricardo (links) und Enrique Flores Magón (rechts) vor dem Los Angeles County Jail, 1917
[...]
```
so some Markup like `mini|` is still left. Should I run another parser on this text before feeding it to an ML model or is this a known imperfection of parsing Wiki markup?
Apologies if this has been asked before. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/835/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/835/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/834 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/834/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/834/comments | https://api.github.com/repos/huggingface/datasets/issues/834/events | https://github.com/huggingface/datasets/issues/834 | 740,082,890 | MDU6SXNzdWU3NDAwODI4OTA= | 834 | [GEM] add WikiLingua cross-lingual abstractive summarization dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4",
"events_url": "https://api.github.com/users/yjernite/events{/privacy}",
"followers_url": "https://api.github.com/users/yjernite/followers",
"following_url": "https://api.github.com/users/yjernite/following{/other_user}",
"gists_url": "https://api.github.com/users/yjernite/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/yjernite",
"id": 10469459,
"login": "yjernite",
"node_id": "MDQ6VXNlcjEwNDY5NDU5",
"organizations_url": "https://api.github.com/users/yjernite/orgs",
"received_events_url": "https://api.github.com/users/yjernite/received_events",
"repos_url": "https://api.github.com/users/yjernite/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/yjernite/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/yjernite/subscriptions",
"type": "User",
"url": "https://api.github.com/users/yjernite"
} | [
{
"color": "e99695",
"default": false,
"description": "Requesting to add a new dataset",
"id": 2067376369,
"name": "dataset request",
"node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5",
"url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request"
}
] | closed | false | null | [] | null | 2 | "2020-11-10T17:00:43Z" | "2021-04-15T12:04:09Z" | "2021-04-15T12:01:38Z" | MEMBER | null | null | null | ## Adding a Dataset
- **Name:** WikiLingua
- **Description:** The dataset includes ~770k article and summary pairs in 18 languages from WikiHow. The gold-standard article-summary alignments across languages were extracted by aligning the images that are used to describe each how-to step in an article.
- **Paper:** https://arxiv.org/pdf/2010.03093.pdf
- **Data:** https://github.com/esdurmus/Wikilingua
- **Motivation:** Included in the GEM shared task. Multilingual.
Instructions to add a new dataset can be found [here](https://huggingface.co/docs/datasets/share_dataset.html).
| {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/834/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/834/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/833 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/833/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/833/comments | https://api.github.com/repos/huggingface/datasets/issues/833/events | https://github.com/huggingface/datasets/issues/833 | 740,079,692 | MDU6SXNzdWU3NDAwNzk2OTI= | 833 | [GEM] add ASSET text simplification dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4",
"events_url": "https://api.github.com/users/yjernite/events{/privacy}",
"followers_url": "https://api.github.com/users/yjernite/followers",
"following_url": "https://api.github.com/users/yjernite/following{/other_user}",
"gists_url": "https://api.github.com/users/yjernite/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/yjernite",
"id": 10469459,
"login": "yjernite",
"node_id": "MDQ6VXNlcjEwNDY5NDU5",
"organizations_url": "https://api.github.com/users/yjernite/orgs",
"received_events_url": "https://api.github.com/users/yjernite/received_events",
"repos_url": "https://api.github.com/users/yjernite/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/yjernite/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/yjernite/subscriptions",
"type": "User",
"url": "https://api.github.com/users/yjernite"
} | [
{
"color": "e99695",
"default": false,
"description": "Requesting to add a new dataset",
"id": 2067376369,
"name": "dataset request",
"node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5",
"url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request"
}
] | closed | false | null | [] | null | 0 | "2020-11-10T16:56:30Z" | "2020-12-03T13:38:15Z" | "2020-12-03T13:38:15Z" | MEMBER | null | null | null | ## Adding a Dataset
- **Name:** ASSET
- **Description:** ASSET is a crowdsourced
multi-reference corpus for assessing sentence simplification in English where each simplification was produced by executing several rewriting transformations.
- **Paper:** https://www.aclweb.org/anthology/2020.acl-main.424.pdf
- **Data:** https://github.com/facebookresearch/asset
- **Motivation:** Included in the GEM shared task
Instructions to add a new dataset can be found [here](https://huggingface.co/docs/datasets/share_dataset.html).
| {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/833/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/833/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/832 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/832/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/832/comments | https://api.github.com/repos/huggingface/datasets/issues/832/events | https://github.com/huggingface/datasets/issues/832 | 740,077,228 | MDU6SXNzdWU3NDAwNzcyMjg= | 832 | [GEM] add WikiAuto text simplification dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4",
"events_url": "https://api.github.com/users/yjernite/events{/privacy}",
"followers_url": "https://api.github.com/users/yjernite/followers",
"following_url": "https://api.github.com/users/yjernite/following{/other_user}",
"gists_url": "https://api.github.com/users/yjernite/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/yjernite",
"id": 10469459,
"login": "yjernite",
"node_id": "MDQ6VXNlcjEwNDY5NDU5",
"organizations_url": "https://api.github.com/users/yjernite/orgs",
"received_events_url": "https://api.github.com/users/yjernite/received_events",
"repos_url": "https://api.github.com/users/yjernite/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/yjernite/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/yjernite/subscriptions",
"type": "User",
"url": "https://api.github.com/users/yjernite"
} | [
{
"color": "e99695",
"default": false,
"description": "Requesting to add a new dataset",
"id": 2067376369,
"name": "dataset request",
"node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5",
"url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request"
}
] | closed | false | null | [] | null | 0 | "2020-11-10T16:53:23Z" | "2020-12-03T13:38:08Z" | "2020-12-03T13:38:08Z" | MEMBER | null | null | null | ## Adding a Dataset
- **Name:** WikiAuto
- **Description:** Sentences in English Wikipedia and their corresponding sentences in Simple English Wikipedia that are written with simpler grammar and word choices. A lot of lexical and syntactic paraphrasing.
- **Paper:** https://www.aclweb.org/anthology/2020.acl-main.709.pdf
- **Data:** https://github.com/chaojiang06/wiki-auto
- **Motivation:** Included in the GEM shared task
Instructions to add a new dataset can be found [here](https://huggingface.co/docs/datasets/share_dataset.html).
| {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/832/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/832/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/831 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/831/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/831/comments | https://api.github.com/repos/huggingface/datasets/issues/831/events | https://github.com/huggingface/datasets/issues/831 | 740,071,697 | MDU6SXNzdWU3NDAwNzE2OTc= | 831 | [GEM] Add WebNLG dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4",
"events_url": "https://api.github.com/users/yjernite/events{/privacy}",
"followers_url": "https://api.github.com/users/yjernite/followers",
"following_url": "https://api.github.com/users/yjernite/following{/other_user}",
"gists_url": "https://api.github.com/users/yjernite/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/yjernite",
"id": 10469459,
"login": "yjernite",
"node_id": "MDQ6VXNlcjEwNDY5NDU5",
"organizations_url": "https://api.github.com/users/yjernite/orgs",
"received_events_url": "https://api.github.com/users/yjernite/received_events",
"repos_url": "https://api.github.com/users/yjernite/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/yjernite/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/yjernite/subscriptions",
"type": "User",
"url": "https://api.github.com/users/yjernite"
} | [
{
"color": "e99695",
"default": false,
"description": "Requesting to add a new dataset",
"id": 2067376369,
"name": "dataset request",
"node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5",
"url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request"
}
] | closed | false | null | [] | null | 0 | "2020-11-10T16:46:48Z" | "2020-12-03T13:38:01Z" | "2020-12-03T13:38:01Z" | MEMBER | null | null | null | ## Adding a Dataset
- **Name:** WebNLG
- **Description:** WebNLG consists of Data/Text pairs where the data is a set of triples extracted from DBpedia and the text is a verbalisation of these triples (16,095 data inputs and 42,873 data-text pairs). The data is available in English and Russian
- **Paper:** https://www.aclweb.org/anthology/P17-1017.pdf
- **Data:** https://webnlg-challenge.loria.fr/download/
- **Motivation:** Included in the GEM shared task, multilingual
Instructions to add a new dataset can be found [here](https://huggingface.co/docs/datasets/share_dataset.html).
| {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/831/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/831/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/830 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/830/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/830/comments | https://api.github.com/repos/huggingface/datasets/issues/830/events | https://github.com/huggingface/datasets/issues/830 | 740,065,376 | MDU6SXNzdWU3NDAwNjUzNzY= | 830 | [GEM] add ToTTo Table-to-text dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4",
"events_url": "https://api.github.com/users/yjernite/events{/privacy}",
"followers_url": "https://api.github.com/users/yjernite/followers",
"following_url": "https://api.github.com/users/yjernite/following{/other_user}",
"gists_url": "https://api.github.com/users/yjernite/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/yjernite",
"id": 10469459,
"login": "yjernite",
"node_id": "MDQ6VXNlcjEwNDY5NDU5",
"organizations_url": "https://api.github.com/users/yjernite/orgs",
"received_events_url": "https://api.github.com/users/yjernite/received_events",
"repos_url": "https://api.github.com/users/yjernite/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/yjernite/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/yjernite/subscriptions",
"type": "User",
"url": "https://api.github.com/users/yjernite"
} | [
{
"color": "e99695",
"default": false,
"description": "Requesting to add a new dataset",
"id": 2067376369,
"name": "dataset request",
"node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5",
"url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request"
}
] | closed | false | null | [] | null | 1 | "2020-11-10T16:38:34Z" | "2020-12-10T13:06:02Z" | "2020-12-10T13:06:01Z" | MEMBER | null | null | null | ## Adding a Dataset
- **Name:** ToTTo
- **Description:** ToTTo is an open-domain English table-to-text dataset with over 120,000 training examples that proposes a controlled generation task: given a Wikipedia table and a set of highlighted table cells, produce a one-sentence description.
- **Paper:** https://arxiv.org/abs/2004.14373
- **Data:** https://github.com/google-research-datasets/totto
- **Motivation:** Included in the GEM shared task
Instructions to add a new dataset can be found [here](https://huggingface.co/docs/datasets/share_dataset.html).
| {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/830/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/830/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/829 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/829/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/829/comments | https://api.github.com/repos/huggingface/datasets/issues/829/events | https://github.com/huggingface/datasets/issues/829 | 740,061,699 | MDU6SXNzdWU3NDAwNjE2OTk= | 829 | [GEM] add Schema-Guided Dialogue | {
"avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4",
"events_url": "https://api.github.com/users/yjernite/events{/privacy}",
"followers_url": "https://api.github.com/users/yjernite/followers",
"following_url": "https://api.github.com/users/yjernite/following{/other_user}",
"gists_url": "https://api.github.com/users/yjernite/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/yjernite",
"id": 10469459,
"login": "yjernite",
"node_id": "MDQ6VXNlcjEwNDY5NDU5",
"organizations_url": "https://api.github.com/users/yjernite/orgs",
"received_events_url": "https://api.github.com/users/yjernite/received_events",
"repos_url": "https://api.github.com/users/yjernite/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/yjernite/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/yjernite/subscriptions",
"type": "User",
"url": "https://api.github.com/users/yjernite"
} | [
{
"color": "e99695",
"default": false,
"description": "Requesting to add a new dataset",
"id": 2067376369,
"name": "dataset request",
"node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5",
"url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request"
}
] | closed | false | null | [] | null | 0 | "2020-11-10T16:33:44Z" | "2020-12-03T13:37:50Z" | "2020-12-03T13:37:50Z" | MEMBER | null | null | null | ## Adding a Dataset
- **Name:** The Schema-Guided Dialogue Dataset
- **Description:** The Schema-Guided Dialogue (SGD) dataset consists of over 20k annotated multi-domain, task-oriented conversations between a human and a virtual assistant. These conversations involve interactions with services and APIs spanning 20 domains, ranging from banks and events to media, calendar, travel, and weather.
- **Paper:** https://arxiv.org/pdf/2002.01359.pdf https://arxiv.org/pdf/2004.15006.pdf
- **Data:** https://github.com/google-research-datasets/dstc8-schema-guided-dialogue
- **Motivation:** Included in the GEM shared task
Instructions to add a new dataset can be found [here](https://huggingface.co/docs/datasets/share_dataset.html).
| {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/829/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/829/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/828 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/828/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/828/comments | https://api.github.com/repos/huggingface/datasets/issues/828/events | https://github.com/huggingface/datasets/pull/828 | 740,008,683 | MDExOlB1bGxSZXF1ZXN0NTE4NTcwMjY3 | 828 | Add writer_batch_size attribute to GeneratorBasedBuilder | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq"
} | [] | closed | false | null | [] | null | 0 | "2020-11-10T15:28:19Z" | "2020-11-10T16:27:36Z" | "2020-11-10T16:27:36Z" | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/828.diff",
"html_url": "https://github.com/huggingface/datasets/pull/828",
"merged_at": "2020-11-10T16:27:35Z",
"patch_url": "https://github.com/huggingface/datasets/pull/828.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/828"
} | As specified in #741 one would need to specify a custom ArrowWriter batch size to avoid filling the RAM. Indeed the defaults buffer size is 10 000 examples but for multimodal datasets that contain images or videos we may want to reduce that. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/828/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/828/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/827 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/827/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/827/comments | https://api.github.com/repos/huggingface/datasets/issues/827/events | https://github.com/huggingface/datasets/issues/827 | 739,983,024 | MDU6SXNzdWU3Mzk5ODMwMjQ= | 827 | [GEM] MultiWOZ dialogue dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4",
"events_url": "https://api.github.com/users/yjernite/events{/privacy}",
"followers_url": "https://api.github.com/users/yjernite/followers",
"following_url": "https://api.github.com/users/yjernite/following{/other_user}",
"gists_url": "https://api.github.com/users/yjernite/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/yjernite",
"id": 10469459,
"login": "yjernite",
"node_id": "MDQ6VXNlcjEwNDY5NDU5",
"organizations_url": "https://api.github.com/users/yjernite/orgs",
"received_events_url": "https://api.github.com/users/yjernite/received_events",
"repos_url": "https://api.github.com/users/yjernite/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/yjernite/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/yjernite/subscriptions",
"type": "User",
"url": "https://api.github.com/users/yjernite"
} | [
{
"color": "e99695",
"default": false,
"description": "Requesting to add a new dataset",
"id": 2067376369,
"name": "dataset request",
"node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5",
"url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request"
}
] | closed | false | null | [] | null | 2 | "2020-11-10T14:57:50Z" | "2022-10-05T12:31:13Z" | "2022-10-05T12:31:13Z" | MEMBER | null | null | null | ## Adding a Dataset
- **Name:** MultiWOZ (Multi-Domain Wizard-of-Oz)
- **Description:** 10k annotated human-human dialogues. Each dialogue consists of a goal, multiple user and system utterances as well as a belief state. Only system utterances are annotated with dialogue acts – there are no annotations from the user side.
- **Paper:** https://arxiv.org/pdf/2007.12720.pdf
- **Data:** https://github.com/budzianowski/multiwoz
- **Motivation:** Will likely be part of the GEM shared task
Instructions to add a new dataset can be found [here](https://huggingface.co/docs/datasets/share_dataset.html).
| {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/827/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/827/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/826 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/826/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/826/comments | https://api.github.com/repos/huggingface/datasets/issues/826/events | https://github.com/huggingface/datasets/issues/826 | 739,976,716 | MDU6SXNzdWU3Mzk5NzY3MTY= | 826 | [GEM] Add E2E dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4",
"events_url": "https://api.github.com/users/yjernite/events{/privacy}",
"followers_url": "https://api.github.com/users/yjernite/followers",
"following_url": "https://api.github.com/users/yjernite/following{/other_user}",
"gists_url": "https://api.github.com/users/yjernite/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/yjernite",
"id": 10469459,
"login": "yjernite",
"node_id": "MDQ6VXNlcjEwNDY5NDU5",
"organizations_url": "https://api.github.com/users/yjernite/orgs",
"received_events_url": "https://api.github.com/users/yjernite/received_events",
"repos_url": "https://api.github.com/users/yjernite/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/yjernite/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/yjernite/subscriptions",
"type": "User",
"url": "https://api.github.com/users/yjernite"
} | [
{
"color": "e99695",
"default": false,
"description": "Requesting to add a new dataset",
"id": 2067376369,
"name": "dataset request",
"node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5",
"url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request"
}
] | closed | false | null | [] | null | 0 | "2020-11-10T14:50:40Z" | "2020-12-03T13:37:57Z" | "2020-12-03T13:37:57Z" | MEMBER | null | null | null | ## Adding a Dataset
- **Name:** E2E NLG dataset (for End-to-end natural language generation)
- **Description:**a dataset for training end-to-end, datadriven natural language generation systems in the restaurant domain, the datasets consists of 5,751 dialogue-act Meaning Representations (structured data) and 8.1 reference free-text utterances per dialogue-act on average
- **Paper:** https://arxiv.org/pdf/1706.09254.pdf https://arxiv.org/abs/1901.07931
- **Data:** http://www.macs.hw.ac.uk/InteractionLab/E2E/#data
- **Motivation:** This dataset will likely be included in the GEM shared task
Instructions to add a new dataset can be found [here](https://huggingface.co/docs/datasets/share_dataset.html).
| {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/826/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/826/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/825 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/825/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/825/comments | https://api.github.com/repos/huggingface/datasets/issues/825/events | https://github.com/huggingface/datasets/pull/825 | 739,925,960 | MDExOlB1bGxSZXF1ZXN0NTE4NTAyNjgx | 825 | Add accuracy, precision, recall and F1 metrics | {
"avatar_url": "https://avatars.githubusercontent.com/u/959590?v=4",
"events_url": "https://api.github.com/users/jplu/events{/privacy}",
"followers_url": "https://api.github.com/users/jplu/followers",
"following_url": "https://api.github.com/users/jplu/following{/other_user}",
"gists_url": "https://api.github.com/users/jplu/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/jplu",
"id": 959590,
"login": "jplu",
"node_id": "MDQ6VXNlcjk1OTU5MA==",
"organizations_url": "https://api.github.com/users/jplu/orgs",
"received_events_url": "https://api.github.com/users/jplu/received_events",
"repos_url": "https://api.github.com/users/jplu/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/jplu/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jplu/subscriptions",
"type": "User",
"url": "https://api.github.com/users/jplu"
} | [] | closed | false | null | [] | null | 0 | "2020-11-10T13:50:35Z" | "2020-11-11T19:23:48Z" | "2020-11-11T19:23:43Z" | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/825.diff",
"html_url": "https://github.com/huggingface/datasets/pull/825",
"merged_at": "2020-11-11T19:23:43Z",
"patch_url": "https://github.com/huggingface/datasets/pull/825.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/825"
} | This PR adds several single metrics, namely:
- Accuracy
- Precision
- Recall
- F1
They all uses under the hood the sklearn metrics of the same name. They allow different useful features when training a multilabel/multiclass model:
- have a macro/micro/per label/weighted/binary/per sample score
- score only the selected labels (usually what we call the positive labels) and ignore the negative ones. For example in case of a Named Entity Recognition task, positive labels are (`PERSON`, `LOCATION` or `ORGANIZATION`) and the negative one is `O`. | {
"+1": 1,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/825/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/825/timeline | null | null | true |
Subsets and Splits