url
stringlengths
58
61
repository_url
stringclasses
1 value
labels_url
stringlengths
72
75
comments_url
stringlengths
67
70
events_url
stringlengths
65
68
html_url
stringlengths
46
51
id
int64
599M
3.64B
node_id
stringlengths
18
32
number
int64
1
7.87k
title
stringlengths
1
290
user
dict
labels
listlengths
0
4
state
stringclasses
2 values
locked
bool
1 class
assignee
dict
assignees
listlengths
0
4
milestone
dict
comments
int64
0
70
created_at
stringdate
2020-04-14 10:18:02
2025-11-18 08:33:04
updated_at
stringdate
2020-04-27 16:04:17
2025-11-18 16:07:04
closed_at
stringlengths
3
25
author_association
stringclasses
4 values
type
float64
active_lock_reason
float64
sub_issues_summary
dict
issue_dependencies_summary
dict
body
stringlengths
0
228k
closed_by
dict
reactions
dict
timeline_url
stringlengths
67
70
performed_via_github_app
float64
state_reason
stringclasses
4 values
draft
float64
0
1
pull_request
dict
is_pull_request
bool
2 classes
https://api.github.com/repos/huggingface/datasets/issues/492
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/492/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/492/comments
https://api.github.com/repos/huggingface/datasets/issues/492/events
https://github.com/huggingface/datasets/issues/492
676,495,064
MDU6SXNzdWU2NzY0OTUwNjQ=
492
nlp.Features does not distinguish between nullable and non-nullable types in PyArrow schema
{ "avatar_url": "https://avatars.githubusercontent.com/u/4564897?v=4", "events_url": "https://api.github.com/users/jarednielsen/events{/privacy}", "followers_url": "https://api.github.com/users/jarednielsen/followers", "following_url": "https://api.github.com/users/jarednielsen/following{/other_user}", "gists...
[]
closed
false
null
[]
null
7
2020-08-11 00:27:46+00:00
2020-08-26 16:17:19+00:00
2020-08-26 16:17:19+00:00
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
Here's the code I'm trying to run: ```python dset_wikipedia = nlp.load_dataset("wikipedia", "20200501.en", split="train", cache_dir=args.cache_dir) dset_wikipedia.drop(columns=["title"]) dset_wikipedia.features.pop("title") dset_books = nlp.load_dataset("bookcorpus", split="train", cache_dir=args.cache_dir) dse...
{ "avatar_url": "https://avatars.githubusercontent.com/u/4564897?v=4", "events_url": "https://api.github.com/users/jarednielsen/events{/privacy}", "followers_url": "https://api.github.com/users/jarednielsen/followers", "following_url": "https://api.github.com/users/jarednielsen/following{/other_user}", "gists...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/492/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/492/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/491
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/491/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/491/comments
https://api.github.com/repos/huggingface/datasets/issues/491/events
https://github.com/huggingface/datasets/issues/491
676,486,275
MDU6SXNzdWU2NzY0ODYyNzU=
491
No 0.4.0 release on GitHub
{ "avatar_url": "https://avatars.githubusercontent.com/u/4564897?v=4", "events_url": "https://api.github.com/users/jarednielsen/events{/privacy}", "followers_url": "https://api.github.com/users/jarednielsen/followers", "following_url": "https://api.github.com/users/jarednielsen/following{/other_user}", "gists...
[]
closed
false
null
[]
null
2
2020-08-10 23:59:57+00:00
2020-08-11 16:50:07+00:00
2020-08-11 16:50:07+00:00
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
0.4.0 was released on PyPi, but not on GitHub. This means [the documentation](https://huggingface.co/nlp/) is still displaying from 0.3.0, and that there's no tag to easily clone the 0.4.0 version of the repo.
{ "avatar_url": "https://avatars.githubusercontent.com/u/4564897?v=4", "events_url": "https://api.github.com/users/jarednielsen/events{/privacy}", "followers_url": "https://api.github.com/users/jarednielsen/followers", "following_url": "https://api.github.com/users/jarednielsen/following{/other_user}", "gists...
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/491/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/491/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/490
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/490/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/490/comments
https://api.github.com/repos/huggingface/datasets/issues/490/events
https://github.com/huggingface/datasets/issues/490
676,482,242
MDU6SXNzdWU2NzY0ODIyNDI=
490
Loading preprocessed Wikipedia dataset requires apache_beam
{ "avatar_url": "https://avatars.githubusercontent.com/u/4564897?v=4", "events_url": "https://api.github.com/users/jarednielsen/events{/privacy}", "followers_url": "https://api.github.com/users/jarednielsen/followers", "following_url": "https://api.github.com/users/jarednielsen/following{/other_user}", "gists...
[]
closed
false
null
[]
null
0
2020-08-10 23:46:50+00:00
2020-08-14 13:17:20+00:00
2020-08-14 13:17:20+00:00
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
Running `nlp.load_dataset("wikipedia", "20200501.en", split="train", dir="/tmp/wikipedia")` gives an error if apache_beam is not installed, stemming from https://github.com/huggingface/nlp/blob/38eb2413de54ee804b0be81781bd65ac4a748ced/src/nlp/builder.py#L981-L988 This succeeded without the dependency in ve...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/490/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/490/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/489
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/489/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/489/comments
https://api.github.com/repos/huggingface/datasets/issues/489/events
https://github.com/huggingface/datasets/issues/489
676,456,257
MDU6SXNzdWU2NzY0NTYyNTc=
489
ug
{ "avatar_url": "https://avatars.githubusercontent.com/u/2000204?v=4", "events_url": "https://api.github.com/users/timothyjlaurent/events{/privacy}", "followers_url": "https://api.github.com/users/timothyjlaurent/followers", "following_url": "https://api.github.com/users/timothyjlaurent/following{/other_user}",...
[]
closed
false
null
[]
null
2
2020-08-10 22:33:03+00:00
2020-08-10 22:55:14+00:00
2020-08-10 22:33:40+00:00
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
{ "avatar_url": "https://avatars.githubusercontent.com/u/2000204?v=4", "events_url": "https://api.github.com/users/timothyjlaurent/events{/privacy}", "followers_url": "https://api.github.com/users/timothyjlaurent/followers", "following_url": "https://api.github.com/users/timothyjlaurent/following{/other_user}",...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/489/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/489/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/488
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/488/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/488/comments
https://api.github.com/repos/huggingface/datasets/issues/488/events
https://github.com/huggingface/datasets/issues/488
676,299,993
MDU6SXNzdWU2NzYyOTk5OTM=
488
issues with downloading datasets for wmt16 and wmt19
{ "avatar_url": "https://avatars.githubusercontent.com/u/10676103?v=4", "events_url": "https://api.github.com/users/stas00/events{/privacy}", "followers_url": "https://api.github.com/users/stas00/followers", "following_url": "https://api.github.com/users/stas00/following{/other_user}", "gists_url": "https://a...
[]
closed
false
null
[]
null
3
2020-08-10 17:32:51+00:00
2022-10-04 17:46:59+00:00
2022-10-04 17:46:58+00:00
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
I have encountered multiple issues while trying to: ``` import nlp dataset = nlp.load_dataset('wmt16', 'ru-en') metric = nlp.load_metric('wmt16') ``` 1. I had to do `pip install -e ".[dev]" ` on master, currently released nlp didn't work (sorry, didn't save the error) - I went back to the released version and no...
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url"...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/488/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/488/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/487
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/487/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/487/comments
https://api.github.com/repos/huggingface/datasets/issues/487/events
https://github.com/huggingface/datasets/pull/487
676,143,029
MDExOlB1bGxSZXF1ZXN0NDY1NTA1NjQy
487
Fix elasticsearch result ids returning as strings
{ "avatar_url": "https://avatars.githubusercontent.com/u/3595526?v=4", "events_url": "https://api.github.com/users/sai-prasanna/events{/privacy}", "followers_url": "https://api.github.com/users/sai-prasanna/followers", "following_url": "https://api.github.com/users/sai-prasanna/following{/other_user}", "gists...
[]
closed
false
null
[]
null
1
2020-08-10 13:37:11+00:00
2020-08-31 10:42:46+00:00
2020-08-31 10:42:46+00:00
CONTRIBUTOR
null
null
null
null
I am using the latest elasticsearch binary and master of nlp. For me elasticsearch searches failed because the resultant "id_" returned for searches are strings, but our library assumes them to be integers.
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/487/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/487/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/487.diff", "html_url": "https://github.com/huggingface/datasets/pull/487", "merged_at": "2020-08-31T10:42:46Z", "patch_url": "https://github.com/huggingface/datasets/pull/487.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/487...
true
https://api.github.com/repos/huggingface/datasets/issues/486
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/486/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/486/comments
https://api.github.com/repos/huggingface/datasets/issues/486/events
https://github.com/huggingface/datasets/issues/486
675,649,034
MDU6SXNzdWU2NzU2NDkwMzQ=
486
Bookcorpus data contains pretokenized text
{ "avatar_url": "https://avatars.githubusercontent.com/u/99543?v=4", "events_url": "https://api.github.com/users/orsharir/events{/privacy}", "followers_url": "https://api.github.com/users/orsharir/followers", "following_url": "https://api.github.com/users/orsharir/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
8
2020-08-09 06:53:24+00:00
2022-10-04 17:44:33+00:00
2022-10-04 17:44:33+00:00
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
It seem that the bookcoprus data downloaded through the library was pretokenized with NLTK's Treebank tokenizer, which changes the text in incompatible ways to how, for instance, BERT's wordpiece tokenizer works. For example, "didn't" becomes "did" + "n't", and double quotes are changed to `` and '' for start and end q...
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url"...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/486/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/486/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/485
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/485/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/485/comments
https://api.github.com/repos/huggingface/datasets/issues/485/events
https://github.com/huggingface/datasets/issues/485
675,595,393
MDU6SXNzdWU2NzU1OTUzOTM=
485
PAWS dataset first item is header
{ "avatar_url": "https://avatars.githubusercontent.com/u/13238952?v=4", "events_url": "https://api.github.com/users/jxmorris12/events{/privacy}", "followers_url": "https://api.github.com/users/jxmorris12/followers", "following_url": "https://api.github.com/users/jxmorris12/following{/other_user}", "gists_url"...
[]
closed
false
null
[]
null
0
2020-08-08 22:05:25+00:00
2020-08-19 09:50:01+00:00
2020-08-19 09:50:01+00:00
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
``` import nlp dataset = nlp.load_dataset('xtreme', 'PAWS-X.en') dataset['test'][0] ``` prints the following ``` {'label': 'label', 'sentence1': 'sentence1', 'sentence2': 'sentence2'} ``` dataset['test'][0] should probably be the first item in the dataset, not just a dictionary mapping the column names t...
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/485/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/485/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/484
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/484/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/484/comments
https://api.github.com/repos/huggingface/datasets/issues/484/events
https://github.com/huggingface/datasets/pull/484
675,088,983
MDExOlB1bGxSZXF1ZXN0NDY0NjY1NTU4
484
update mirror for RT dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/13238952?v=4", "events_url": "https://api.github.com/users/jxmorris12/events{/privacy}", "followers_url": "https://api.github.com/users/jxmorris12/followers", "following_url": "https://api.github.com/users/jxmorris12/following{/other_user}", "gists_url"...
[]
closed
false
null
[]
null
4
2020-08-07 15:25:45+00:00
2020-08-24 13:33:37+00:00
2020-08-24 13:33:37+00:00
CONTRIBUTOR
null
null
null
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/484/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/484/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/484.diff", "html_url": "https://github.com/huggingface/datasets/pull/484", "merged_at": "2020-08-24T13:33:37Z", "patch_url": "https://github.com/huggingface/datasets/pull/484.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/484...
true
https://api.github.com/repos/huggingface/datasets/issues/483
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/483/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/483/comments
https://api.github.com/repos/huggingface/datasets/issues/483/events
https://github.com/huggingface/datasets/issues/483
675,080,694
MDU6SXNzdWU2NzUwODA2OTQ=
483
rotten tomatoes movie review dataset taken down
{ "avatar_url": "https://avatars.githubusercontent.com/u/13238952?v=4", "events_url": "https://api.github.com/users/jxmorris12/events{/privacy}", "followers_url": "https://api.github.com/users/jxmorris12/followers", "following_url": "https://api.github.com/users/jxmorris12/following{/other_user}", "gists_url"...
[]
closed
false
null
[]
null
3
2020-08-07 15:12:01+00:00
2020-09-08 09:36:34+00:00
2020-09-08 09:36:33+00:00
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
In an interesting twist of events, the individual who created the movie review seems to have left Cornell, and their webpage has been removed, along with the movie review dataset (http://www.cs.cornell.edu/people/pabo/movie-review-data/rt-polaritydata.tar.gz). It's not downloadable anymore.
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/483/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/483/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/482
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/482/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/482/comments
https://api.github.com/repos/huggingface/datasets/issues/482/events
https://github.com/huggingface/datasets/issues/482
674,851,147
MDU6SXNzdWU2NzQ4NTExNDc=
482
Bugs : dataset.map() is frozen on ELI5
{ "avatar_url": "https://avatars.githubusercontent.com/u/56621342?v=4", "events_url": "https://api.github.com/users/ratthachat/events{/privacy}", "followers_url": "https://api.github.com/users/ratthachat/followers", "following_url": "https://api.github.com/users/ratthachat/following{/other_user}", "gists_url"...
[]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[ { "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists...
null
8
2020-08-07 08:23:35+00:00
2023-04-06 09:39:59+00:00
2020-08-11 23:55:15+00:00
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
Hi Huggingface Team! Thank you guys once again for this amazing repo. I have tried to prepare ELI5 to train with T5, based on [this wonderful notebook of Suraj Patil](https://github.com/patil-suraj/exploring-T5/blob/master/T5_on_TPU.ipynb) However, when I run `dataset.map()` on ELI5 to prepare `input_text, ta...
{ "avatar_url": "https://avatars.githubusercontent.com/u/56621342?v=4", "events_url": "https://api.github.com/users/ratthachat/events{/privacy}", "followers_url": "https://api.github.com/users/ratthachat/followers", "following_url": "https://api.github.com/users/ratthachat/following{/other_user}", "gists_url"...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/482/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/482/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/481
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/481/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/481/comments
https://api.github.com/repos/huggingface/datasets/issues/481/events
https://github.com/huggingface/datasets/pull/481
674,567,389
MDExOlB1bGxSZXF1ZXN0NDY0MjM2MTA1
481
Apply utf-8 encoding to all datasets
{ "avatar_url": "https://avatars.githubusercontent.com/u/26859204?v=4", "events_url": "https://api.github.com/users/lewtun/events{/privacy}", "followers_url": "https://api.github.com/users/lewtun/followers", "following_url": "https://api.github.com/users/lewtun/following{/other_user}", "gists_url": "https://a...
[]
closed
false
null
[]
null
6
2020-08-06 20:02:09+00:00
2020-08-20 08:16:08+00:00
2020-08-20 08:16:08+00:00
MEMBER
null
null
null
null
## Description This PR applies utf-8 encoding for all instances of `with open(...) as f` to all Python files in `datasets/`. As suggested by @thomwolf in #468 , we use regular expressions and the following function ```python def apply_encoding_on_file_open(filepath: str): """Apply UTF-8 encoding for all insta...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/481/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/481/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/481.diff", "html_url": "https://github.com/huggingface/datasets/pull/481", "merged_at": "2020-08-20T08:16:08Z", "patch_url": "https://github.com/huggingface/datasets/pull/481.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/481...
true
https://api.github.com/repos/huggingface/datasets/issues/480
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/480/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/480/comments
https://api.github.com/repos/huggingface/datasets/issues/480/events
https://github.com/huggingface/datasets/pull/480
674,245,959
MDExOlB1bGxSZXF1ZXN0NDYzOTcwNjQ2
480
Column indexing hotfix
{ "avatar_url": "https://avatars.githubusercontent.com/u/26709476?v=4", "events_url": "https://api.github.com/users/TevenLeScao/events{/privacy}", "followers_url": "https://api.github.com/users/TevenLeScao/followers", "following_url": "https://api.github.com/users/TevenLeScao/following{/other_user}", "gists_u...
[]
closed
false
null
[]
null
2
2020-08-06 11:37:05+00:00
2023-09-24 09:49:33+00:00
2020-08-12 08:36:10+00:00
CONTRIBUTOR
null
null
null
null
As observed for example in #469 , currently `__getitem__` does not convert the data to the dataset format when indexing by column. This is a hotfix that imitates functional 0.3.0. code. In the future it'd probably be nice to have a test there.
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/480/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/480/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/480.diff", "html_url": "https://github.com/huggingface/datasets/pull/480", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/480.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/480" }
true
https://api.github.com/repos/huggingface/datasets/issues/479
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/479/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/479/comments
https://api.github.com/repos/huggingface/datasets/issues/479/events
https://github.com/huggingface/datasets/pull/479
673,905,407
MDExOlB1bGxSZXF1ZXN0NDYzNjkxMjA0
479
add METEOR metric
{ "avatar_url": "https://avatars.githubusercontent.com/u/24683907?v=4", "events_url": "https://api.github.com/users/vegarab/events{/privacy}", "followers_url": "https://api.github.com/users/vegarab/followers", "following_url": "https://api.github.com/users/vegarab/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
5
2020-08-05 23:13:00+00:00
2020-08-19 13:39:09+00:00
2020-08-19 13:39:09+00:00
CONTRIBUTOR
null
null
null
null
Added the METEOR metric. Can be used like this: ```python import nlp meteor = nlp.load_metric('metrics/meteor') meteor.compute(["some string", "some string"], ["some string", "some similar string"]) # {'meteor': 0.6411637931034483} meteor.add("some string", "some string") meteor.add('some string", "some simila...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/479/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/479/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/479.diff", "html_url": "https://github.com/huggingface/datasets/pull/479", "merged_at": "2020-08-19T13:39:09Z", "patch_url": "https://github.com/huggingface/datasets/pull/479.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/479...
true
https://api.github.com/repos/huggingface/datasets/issues/478
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/478/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/478/comments
https://api.github.com/repos/huggingface/datasets/issues/478/events
https://github.com/huggingface/datasets/issues/478
673,178,317
MDU6SXNzdWU2NzMxNzgzMTc=
478
Export TFRecord to GCP bucket
{ "avatar_url": "https://avatars.githubusercontent.com/u/43774355?v=4", "events_url": "https://api.github.com/users/astariul/events{/privacy}", "followers_url": "https://api.github.com/users/astariul/followers", "following_url": "https://api.github.com/users/astariul/following{/other_user}", "gists_url": "htt...
[]
closed
false
null
[]
null
1
2020-08-05 01:08:32+00:00
2020-08-05 01:21:37+00:00
2020-08-05 01:21:36+00:00
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
Previously, I was writing TFRecords manually to GCP bucket with : `with tf.io.TFRecordWriter('gs://my_bucket/x.tfrecord')` Since `0.4.0` is out with the `export()` function, I tried it. But it seems TFRecords cannot be directly written to GCP bucket. `dataset.export('local.tfrecord')` works fine, but `dataset....
{ "avatar_url": "https://avatars.githubusercontent.com/u/43774355?v=4", "events_url": "https://api.github.com/users/astariul/events{/privacy}", "followers_url": "https://api.github.com/users/astariul/followers", "following_url": "https://api.github.com/users/astariul/following{/other_user}", "gists_url": "htt...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/478/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/478/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/477
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/477/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/477/comments
https://api.github.com/repos/huggingface/datasets/issues/477/events
https://github.com/huggingface/datasets/issues/477
673,142,143
MDU6SXNzdWU2NzMxNDIxNDM=
477
Overview.ipynb throws exceptions with nlp 0.4.0
{ "avatar_url": "https://avatars.githubusercontent.com/u/23109219?v=4", "events_url": "https://api.github.com/users/mandy-li/events{/privacy}", "followers_url": "https://api.github.com/users/mandy-li/followers", "following_url": "https://api.github.com/users/mandy-li/following{/other_user}", "gists_url": "htt...
[]
closed
false
null
[]
null
3
2020-08-04 23:18:15+00:00
2021-08-03 06:02:15+00:00
2021-08-03 06:02:15+00:00
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
with nlp 0.4.0, the TensorFlow example in Overview.ipynb throws the following exceptions: --------------------------------------------------------------------------- AttributeError Traceback (most recent call last) <ipython-input-5-48907f2ad433> in <module> ----> 1 features = {x: trai...
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/477/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/477/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/476
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/476/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/476/comments
https://api.github.com/repos/huggingface/datasets/issues/476/events
https://github.com/huggingface/datasets/pull/476
672,991,854
MDExOlB1bGxSZXF1ZXN0NDYyOTMyMTgx
476
CheckList
{ "avatar_url": "https://avatars.githubusercontent.com/u/698010?v=4", "events_url": "https://api.github.com/users/marcotcr/events{/privacy}", "followers_url": "https://api.github.com/users/marcotcr/followers", "following_url": "https://api.github.com/users/marcotcr/following{/other_user}", "gists_url": "https...
[ { "color": "0e8a16", "default": false, "description": "Contribution to a dataset script", "id": 4564477500, "name": "dataset contribution", "node_id": "LA_kwDODunzps8AAAABEBBmPA", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20contribution" } ]
closed
false
null
[]
null
2
2020-08-04 18:32:05+00:00
2022-10-03 09:43:37+00:00
2022-10-03 09:43:37+00:00
NONE
null
null
null
null
Sorry for the large pull request. - Added checklists as datasets. I can't run `test_load_real_dataset` (see #474), but I can load the datasets successfully as shown in the example notebook - Added a checklist wrapper
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/476/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/476/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/476.diff", "html_url": "https://github.com/huggingface/datasets/pull/476", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/476.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/476" }
true
https://api.github.com/repos/huggingface/datasets/issues/475
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/475/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/475/comments
https://api.github.com/repos/huggingface/datasets/issues/475/events
https://github.com/huggingface/datasets/pull/475
672,884,595
MDExOlB1bGxSZXF1ZXN0NDYyODQzMzQz
475
misc. bugs and quality of life
{ "avatar_url": "https://avatars.githubusercontent.com/u/9353833?v=4", "events_url": "https://api.github.com/users/joeddav/events{/privacy}", "followers_url": "https://api.github.com/users/joeddav/followers", "following_url": "https://api.github.com/users/joeddav/following{/other_user}", "gists_url": "https:/...
[]
closed
false
null
[]
null
2
2020-08-04 15:32:29+00:00
2020-08-17 21:14:08+00:00
2020-08-17 21:14:07+00:00
CONTRIBUTOR
null
null
null
null
A few misc. bugs and QOL improvements that I've come across in using the library. Let me know if you don't like any of them and I can adjust/remove them. 1. Printing datasets without a description field throws an error when formatting the `single_line_description`. This fixes that, and also adds some formatting to t...
{ "avatar_url": "https://avatars.githubusercontent.com/u/9353833?v=4", "events_url": "https://api.github.com/users/joeddav/events{/privacy}", "followers_url": "https://api.github.com/users/joeddav/followers", "following_url": "https://api.github.com/users/joeddav/following{/other_user}", "gists_url": "https:/...
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/475/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/475/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/475.diff", "html_url": "https://github.com/huggingface/datasets/pull/475", "merged_at": "2020-08-17T21:14:07Z", "patch_url": "https://github.com/huggingface/datasets/pull/475.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/475...
true
https://api.github.com/repos/huggingface/datasets/issues/474
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/474/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/474/comments
https://api.github.com/repos/huggingface/datasets/issues/474/events
https://github.com/huggingface/datasets/issues/474
672,407,330
MDU6SXNzdWU2NzI0MDczMzA=
474
test_load_real_dataset when config has BUILDER_CONFIGS that matter
{ "avatar_url": "https://avatars.githubusercontent.com/u/698010?v=4", "events_url": "https://api.github.com/users/marcotcr/events{/privacy}", "followers_url": "https://api.github.com/users/marcotcr/followers", "following_url": "https://api.github.com/users/marcotcr/following{/other_user}", "gists_url": "https...
[]
closed
false
null
[]
null
2
2020-08-03 23:46:36+00:00
2020-09-07 14:53:13+00:00
2020-09-07 14:53:13+00:00
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
It a dataset has custom `BUILDER_CONFIGS` with non-keyword arguments (or keyword arguments with non default values), the config is not loaded during the test and causes an error. I think the problem is that `test_load_real_dataset` calls `load_dataset` with `data_dir=temp_data_dir` ([here](https://github.com/huggingfa...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/474/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/474/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/473
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/473/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/473/comments
https://api.github.com/repos/huggingface/datasets/issues/473/events
https://github.com/huggingface/datasets/pull/473
672,007,247
MDExOlB1bGxSZXF1ZXN0NDYyMTIwNzU4
473
add DoQA dataset (ACL 2020)
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
[]
closed
false
null
[]
null
0
2020-08-03 11:26:52+00:00
2020-09-10 17:19:11+00:00
2020-09-03 11:44:15+00:00
CONTRIBUTOR
null
null
null
null
add DoQA dataset (ACL 2020) http://ixa.eus/node/12931
{ "avatar_url": "https://avatars.githubusercontent.com/u/959590?v=4", "events_url": "https://api.github.com/users/jplu/events{/privacy}", "followers_url": "https://api.github.com/users/jplu/followers", "following_url": "https://api.github.com/users/jplu/following{/other_user}", "gists_url": "https://api.githu...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/473/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/473/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/473.diff", "html_url": "https://github.com/huggingface/datasets/pull/473", "merged_at": "2020-09-03T11:44:14Z", "patch_url": "https://github.com/huggingface/datasets/pull/473.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/473...
true
https://api.github.com/repos/huggingface/datasets/issues/472
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/472/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/472/comments
https://api.github.com/repos/huggingface/datasets/issues/472/events
https://github.com/huggingface/datasets/pull/472
672,000,745
MDExOlB1bGxSZXF1ZXN0NDYyMTE1MjA4
472
add crd3 dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
[]
closed
false
null
[]
null
1
2020-08-03 11:15:02+00:00
2020-08-03 11:22:10+00:00
2020-08-03 11:22:09+00:00
CONTRIBUTOR
null
null
null
null
opening new PR for CRD3 dataset (ACL2020) to fix the circle CI problems
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/472/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/472/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/472.diff", "html_url": "https://github.com/huggingface/datasets/pull/472", "merged_at": "2020-08-03T11:22:09Z", "patch_url": "https://github.com/huggingface/datasets/pull/472.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/472...
true
https://api.github.com/repos/huggingface/datasets/issues/471
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/471/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/471/comments
https://api.github.com/repos/huggingface/datasets/issues/471/events
https://github.com/huggingface/datasets/pull/471
671,996,423
MDExOlB1bGxSZXF1ZXN0NDYyMTExNTU1
471
add reuters21578 dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
[]
closed
false
null
[]
null
0
2020-08-03 11:07:14+00:00
2022-08-04 08:39:11+00:00
2020-09-03 09:58:50+00:00
CONTRIBUTOR
null
null
null
null
new PR to add the reuters21578 dataset and fix the circle CI problems. Fix partially: - #353 Subsequent PR after: - #449
{ "avatar_url": "https://avatars.githubusercontent.com/u/959590?v=4", "events_url": "https://api.github.com/users/jplu/events{/privacy}", "followers_url": "https://api.github.com/users/jplu/followers", "following_url": "https://api.github.com/users/jplu/following{/other_user}", "gists_url": "https://api.githu...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/471/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/471/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/471.diff", "html_url": "https://github.com/huggingface/datasets/pull/471", "merged_at": "2020-09-03T09:58:50Z", "patch_url": "https://github.com/huggingface/datasets/pull/471.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/471...
true
https://api.github.com/repos/huggingface/datasets/issues/470
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/470/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/470/comments
https://api.github.com/repos/huggingface/datasets/issues/470/events
https://github.com/huggingface/datasets/pull/470
671,952,276
MDExOlB1bGxSZXF1ZXN0NDYyMDc0MzQ0
470
Adding IWSLT 2017 dataset.
{ "avatar_url": "https://avatars.githubusercontent.com/u/204321?v=4", "events_url": "https://api.github.com/users/Narsil/events{/privacy}", "followers_url": "https://api.github.com/users/Narsil/followers", "following_url": "https://api.github.com/users/Narsil/following{/other_user}", "gists_url": "https://api...
[]
closed
false
null
[]
null
6
2020-08-03 09:52:39+00:00
2020-09-07 12:33:30+00:00
2020-09-07 12:33:30+00:00
CONTRIBUTOR
null
null
null
null
Created a [IWSLT 2017](https://sites.google.com/site/iwsltevaluation2017/TED-tasks) dataset script for the *multilingual data*. ``` Bilingual data: {Arabic, German, French, Japanese, Korean, Chinese} <-> English Multilingual data: German, English, Italian, Dutch, Romanian. (Any pair) ``` I'm unsure how to h...
{ "avatar_url": "https://avatars.githubusercontent.com/u/204321?v=4", "events_url": "https://api.github.com/users/Narsil/events{/privacy}", "followers_url": "https://api.github.com/users/Narsil/followers", "following_url": "https://api.github.com/users/Narsil/following{/other_user}", "gists_url": "https://api...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 2, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/470/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/470/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/470.diff", "html_url": "https://github.com/huggingface/datasets/pull/470", "merged_at": "2020-09-07T12:33:30Z", "patch_url": "https://github.com/huggingface/datasets/pull/470.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/470...
true
https://api.github.com/repos/huggingface/datasets/issues/469
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/469/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/469/comments
https://api.github.com/repos/huggingface/datasets/issues/469/events
https://github.com/huggingface/datasets/issues/469
671,876,963
MDU6SXNzdWU2NzE4NzY5NjM=
469
invalid data type 'str' at _convert_outputs in arrow_dataset.py
{ "avatar_url": "https://avatars.githubusercontent.com/u/30617486?v=4", "events_url": "https://api.github.com/users/Murgates/events{/privacy}", "followers_url": "https://api.github.com/users/Murgates/followers", "following_url": "https://api.github.com/users/Murgates/following{/other_user}", "gists_url": "htt...
[]
closed
false
null
[]
null
9
2020-08-03 07:48:29+00:00
2023-07-20 15:54:17+00:00
2023-07-20 15:54:17+00:00
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
I trying to build multi label text classifier model using Transformers lib. I'm using Transformers NLP to load the data set, while calling trainer.train() method. It throws the following error File "C:\***\arrow_dataset.py", line 343, in _convert_outputs v = command(v) TypeError: new(): invalid data type ...
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url"...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/469/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/469/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/468
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/468/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/468/comments
https://api.github.com/repos/huggingface/datasets/issues/468/events
https://github.com/huggingface/datasets/issues/468
671,622,441
MDU6SXNzdWU2NzE2MjI0NDE=
468
UnicodeDecodeError while loading PAN-X task of XTREME dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/26859204?v=4", "events_url": "https://api.github.com/users/lewtun/events{/privacy}", "followers_url": "https://api.github.com/users/lewtun/followers", "following_url": "https://api.github.com/users/lewtun/following{/other_user}", "gists_url": "https://a...
[]
closed
false
null
[]
null
5
2020-08-02 14:05:10+00:00
2020-08-20 08:16:08+00:00
2020-08-20 08:16:08+00:00
MEMBER
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
Hi 🤗 team! ## Description of the problem I'm running into a `UnicodeDecodeError` while trying to load the PAN-X subset the XTREME dataset: ``` --------------------------------------------------------------------------- UnicodeDecodeError Traceback (most recent call last) <ipython-inp...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/468/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/468/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/467
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/467/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/467/comments
https://api.github.com/repos/huggingface/datasets/issues/467/events
https://github.com/huggingface/datasets/pull/467
671,580,010
MDExOlB1bGxSZXF1ZXN0NDYxNzgwMzUy
467
DOCS: Fix typo
{ "avatar_url": "https://avatars.githubusercontent.com/u/13381361?v=4", "events_url": "https://api.github.com/users/bharatr21/events{/privacy}", "followers_url": "https://api.github.com/users/bharatr21/followers", "following_url": "https://api.github.com/users/bharatr21/following{/other_user}", "gists_url": "...
[]
closed
false
null
[]
null
1
2020-08-02 08:59:37+00:00
2020-08-02 13:52:27+00:00
2020-08-02 09:18:54+00:00
CONTRIBUTOR
null
null
null
null
Fix typo from dictionnary -> dictionary
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/467/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/467/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/467.diff", "html_url": "https://github.com/huggingface/datasets/pull/467", "merged_at": "2020-08-02T09:18:54Z", "patch_url": "https://github.com/huggingface/datasets/pull/467.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/467...
true
https://api.github.com/repos/huggingface/datasets/issues/466
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/466/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/466/comments
https://api.github.com/repos/huggingface/datasets/issues/466/events
https://github.com/huggingface/datasets/pull/466
670,766,891
MDExOlB1bGxSZXF1ZXN0NDYxMDEzOTM0
466
[METRICS] Various improvements on metrics
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
[]
closed
false
null
[]
null
2
2020-08-01 11:03:45+00:00
2020-08-17 15:15:00+00:00
2020-08-17 15:14:59+00:00
MEMBER
null
null
null
null
- Disallow the use of positional arguments to avoid `predictions` vs `references` mistakes - Allow to directly feed numpy/pytorch/tensorflow/pandas objects in metrics
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/466/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/466/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/466.diff", "html_url": "https://github.com/huggingface/datasets/pull/466", "merged_at": "2020-08-17T15:14:59Z", "patch_url": "https://github.com/huggingface/datasets/pull/466.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/466...
true
https://api.github.com/repos/huggingface/datasets/issues/465
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/465/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/465/comments
https://api.github.com/repos/huggingface/datasets/issues/465/events
https://github.com/huggingface/datasets/pull/465
669,889,779
MDExOlB1bGxSZXF1ZXN0NDYwMjEwODYw
465
Keep features after transform
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
3
2020-07-31 14:43:21+00:00
2020-07-31 18:27:33+00:00
2020-07-31 18:27:32+00:00
MEMBER
null
null
null
null
When applying a transform like `map`, some features were lost (and inferred features were used). It was the case for ClassLabel, Translation, etc. To fix that, I did some modifications in the `ArrowWriter`: - added the `update_features` parameter. When it's `True`, then the features specified by the user (if any...
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/465/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/465/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/465.diff", "html_url": "https://github.com/huggingface/datasets/pull/465", "merged_at": "2020-07-31T18:27:32Z", "patch_url": "https://github.com/huggingface/datasets/pull/465.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/465...
true
https://api.github.com/repos/huggingface/datasets/issues/464
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/464/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/464/comments
https://api.github.com/repos/huggingface/datasets/issues/464/events
https://github.com/huggingface/datasets/pull/464
669,767,381
MDExOlB1bGxSZXF1ZXN0NDYwMTAxNDYz
464
Add rename, remove and cast in-place operations
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
[]
closed
false
null
[]
null
0
2020-07-31 12:30:21+00:00
2020-07-31 15:50:02+00:00
2020-07-31 15:50:00+00:00
MEMBER
null
null
null
null
Add a bunch of in-place operation leveraging the Arrow back-end to rename and remove columns and cast to new features without using the more expensive `map` method. These methods are added to `Dataset` as well as `DatasetDict`. Added tests for these new methods and add the methods to the doc. Naming follows th...
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/464/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/464/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/464.diff", "html_url": "https://github.com/huggingface/datasets/pull/464", "merged_at": "2020-07-31T15:50:00Z", "patch_url": "https://github.com/huggingface/datasets/pull/464.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/464...
true
https://api.github.com/repos/huggingface/datasets/issues/463
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/463/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/463/comments
https://api.github.com/repos/huggingface/datasets/issues/463/events
https://github.com/huggingface/datasets/pull/463
669,735,455
MDExOlB1bGxSZXF1ZXN0NDYwMDcyNjQ1
463
Add dataset/mlsum
{ "avatar_url": "https://avatars.githubusercontent.com/u/36986299?v=4", "events_url": "https://api.github.com/users/RachelKer/events{/privacy}", "followers_url": "https://api.github.com/users/RachelKer/followers", "following_url": "https://api.github.com/users/RachelKer/following{/other_user}", "gists_url": "...
[]
closed
false
null
[]
null
3
2020-07-31 11:50:52+00:00
2020-08-24 14:54:42+00:00
2020-08-24 14:54:42+00:00
CONTRIBUTOR
null
null
null
null
New pull request that should correct the previous errors. The load_real_data stills fails because it is looking for a default dataset URL that does not exists, this does not happen when loading the dataset with load_dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/36986299?v=4", "events_url": "https://api.github.com/users/RachelKer/events{/privacy}", "followers_url": "https://api.github.com/users/RachelKer/followers", "following_url": "https://api.github.com/users/RachelKer/following{/other_user}", "gists_url": "...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/463/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/463/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/463.diff", "html_url": "https://github.com/huggingface/datasets/pull/463", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/463.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/463" }
true
https://api.github.com/repos/huggingface/datasets/issues/462
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/462/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/462/comments
https://api.github.com/repos/huggingface/datasets/issues/462/events
https://github.com/huggingface/datasets/pull/462
669,715,547
MDExOlB1bGxSZXF1ZXN0NDYwMDU0NDgz
462
add DoQA (ACL 2020) dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
[]
closed
false
null
[]
null
0
2020-07-31 11:25:56+00:00
2023-09-24 09:48:42+00:00
2020-08-03 11:28:27+00:00
CONTRIBUTOR
null
null
null
null
adds DoQA (ACL 2020) dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/462/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/462/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/462.diff", "html_url": "https://github.com/huggingface/datasets/pull/462", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/462.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/462" }
true
https://api.github.com/repos/huggingface/datasets/issues/461
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/461/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/461/comments
https://api.github.com/repos/huggingface/datasets/issues/461/events
https://github.com/huggingface/datasets/pull/461
669,703,508
MDExOlB1bGxSZXF1ZXN0NDYwMDQzNDY5
461
Doqa
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
[]
closed
false
null
[]
null
0
2020-07-31 11:11:12+00:00
2023-09-24 09:48:40+00:00
2020-07-31 11:13:15+00:00
CONTRIBUTOR
null
null
null
null
add DoQA (ACL 2020) dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/461/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/461/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/461.diff", "html_url": "https://github.com/huggingface/datasets/pull/461", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/461.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/461" }
true
https://api.github.com/repos/huggingface/datasets/issues/460
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/460/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/460/comments
https://api.github.com/repos/huggingface/datasets/issues/460/events
https://github.com/huggingface/datasets/pull/460
669,585,256
MDExOlB1bGxSZXF1ZXN0NDU5OTM2OTU2
460
Fix KeyboardInterrupt in map and bad indices in select
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
2
2020-07-31 08:57:15+00:00
2020-07-31 11:32:19+00:00
2020-07-31 11:32:18+00:00
MEMBER
null
null
null
null
If you interrupted a map function while it was writing, the cached file was not discarded. Therefore the next time you called map, it was loading an incomplete arrow file. We had the same issue with select if there was a bad indice at one point. To fix that I used temporary files that are renamed once everything...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/460/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/460/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/460.diff", "html_url": "https://github.com/huggingface/datasets/pull/460", "merged_at": "2020-07-31T11:32:18Z", "patch_url": "https://github.com/huggingface/datasets/pull/460.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/460...
true
https://api.github.com/repos/huggingface/datasets/issues/459
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/459/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/459/comments
https://api.github.com/repos/huggingface/datasets/issues/459/events
https://github.com/huggingface/datasets/pull/459
669,545,437
MDExOlB1bGxSZXF1ZXN0NDU5OTAxMjEy
459
[Breaking] Update Dataset and DatasetDict API
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
[]
closed
false
null
[]
null
0
2020-07-31 08:11:33+00:00
2020-08-26 08:28:36+00:00
2020-08-26 08:28:35+00:00
MEMBER
null
null
null
null
This PR contains a few breaking changes so it's probably good to keep it for the next (major) release: - rename the `flatten`, `drop` and `dictionary_encode_column` methods in `flatten_`, `drop_` and `dictionary_encode_column_` to indicate that these methods have in-place effects as discussed in #166. From now on we s...
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/459/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/459/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/459.diff", "html_url": "https://github.com/huggingface/datasets/pull/459", "merged_at": "2020-08-26T08:28:35Z", "patch_url": "https://github.com/huggingface/datasets/pull/459.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/459...
true
https://api.github.com/repos/huggingface/datasets/issues/458
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/458/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/458/comments
https://api.github.com/repos/huggingface/datasets/issues/458/events
https://github.com/huggingface/datasets/pull/458
668,972,666
MDExOlB1bGxSZXF1ZXN0NDU5Mzk5ODg2
458
Install CoVal metric from github
{ "avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4", "events_url": "https://api.github.com/users/yjernite/events{/privacy}", "followers_url": "https://api.github.com/users/yjernite/followers", "following_url": "https://api.github.com/users/yjernite/following{/other_user}", "gists_url": "htt...
[]
closed
false
null
[]
null
0
2020-07-30 16:59:25+00:00
2020-07-31 13:56:33+00:00
2020-07-31 13:56:33+00:00
MEMBER
null
null
null
null
Changed the import statements in `coval.py` to direct the user to install the original package from github if it's not already installed (the warning will only display properly after merging [PR455](https://github.com/huggingface/nlp/pull/455)) Also changed the function call to use named rather than positional argum...
{ "avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4", "events_url": "https://api.github.com/users/yjernite/events{/privacy}", "followers_url": "https://api.github.com/users/yjernite/followers", "following_url": "https://api.github.com/users/yjernite/following{/other_user}", "gists_url": "htt...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/458/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/458/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/458.diff", "html_url": "https://github.com/huggingface/datasets/pull/458", "merged_at": "2020-07-31T13:56:33Z", "patch_url": "https://github.com/huggingface/datasets/pull/458.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/458...
true
https://api.github.com/repos/huggingface/datasets/issues/457
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/457/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/457/comments
https://api.github.com/repos/huggingface/datasets/issues/457/events
https://github.com/huggingface/datasets/pull/457
668,898,386
MDExOlB1bGxSZXF1ZXN0NDU5MzMyOTM1
457
add set_format to DatasetDict + tests
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
[]
closed
false
null
[]
null
0
2020-07-30 15:53:20+00:00
2020-07-30 17:34:36+00:00
2020-07-30 17:34:34+00:00
MEMBER
null
null
null
null
Add the `set_format` and `formated_as` and `reset_format` to `DatasetDict`. Add tests to these for `Dataset` and `DatasetDict`. Fix some bugs uncovered by the tests for `pandas` formating.
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/457/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/457/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/457.diff", "html_url": "https://github.com/huggingface/datasets/pull/457", "merged_at": "2020-07-30T17:34:34Z", "patch_url": "https://github.com/huggingface/datasets/pull/457.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/457...
true
https://api.github.com/repos/huggingface/datasets/issues/456
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/456/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/456/comments
https://api.github.com/repos/huggingface/datasets/issues/456/events
https://github.com/huggingface/datasets/pull/456
668,723,785
MDExOlB1bGxSZXF1ZXN0NDU5MTc1MTY0
456
add crd3(ACL 2020) dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
[]
closed
false
null
[]
null
0
2020-07-30 13:28:35+00:00
2023-09-24 09:48:47+00:00
2020-08-03 11:28:52+00:00
CONTRIBUTOR
null
null
null
null
This PR adds the **Critical Role Dungeons and Dragons Dataset** published at ACL 2020
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/456/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/456/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/456.diff", "html_url": "https://github.com/huggingface/datasets/pull/456", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/456.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/456" }
true
https://api.github.com/repos/huggingface/datasets/issues/455
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/455/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/455/comments
https://api.github.com/repos/huggingface/datasets/issues/455/events
https://github.com/huggingface/datasets/pull/455
668,037,965
MDExOlB1bGxSZXF1ZXN0NDU4NTk4NTUw
455
Add bleurt
{ "avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4", "events_url": "https://api.github.com/users/yjernite/events{/privacy}", "followers_url": "https://api.github.com/users/yjernite/followers", "following_url": "https://api.github.com/users/yjernite/following{/other_user}", "gists_url": "htt...
[]
closed
false
null
[]
null
4
2020-07-29 18:08:32+00:00
2020-07-31 13:56:14+00:00
2020-07-31 13:56:14+00:00
MEMBER
null
null
null
null
This PR adds the BLEURT metric to the library. The BLEURT `Metric` downloads a TF checkpoint corresponding to its `config_name` at creation (in the `_info` function). Default is set to `bleurt-base-128`. Note that the default in the original package is `bleurt-tiny-128`, but they throw a warning and recommend usi...
{ "avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4", "events_url": "https://api.github.com/users/yjernite/events{/privacy}", "followers_url": "https://api.github.com/users/yjernite/followers", "following_url": "https://api.github.com/users/yjernite/following{/other_user}", "gists_url": "htt...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/455/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/455/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/455.diff", "html_url": "https://github.com/huggingface/datasets/pull/455", "merged_at": "2020-07-31T13:56:14Z", "patch_url": "https://github.com/huggingface/datasets/pull/455.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/455...
true
https://api.github.com/repos/huggingface/datasets/issues/454
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/454/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/454/comments
https://api.github.com/repos/huggingface/datasets/issues/454/events
https://github.com/huggingface/datasets/pull/454
668,011,577
MDExOlB1bGxSZXF1ZXN0NDU4NTc3MzA3
454
Create SECURITY.md
{ "avatar_url": "https://avatars.githubusercontent.com/u/56394989?v=4", "events_url": "https://api.github.com/users/ChenZehong13/events{/privacy}", "followers_url": "https://api.github.com/users/ChenZehong13/followers", "following_url": "https://api.github.com/users/ChenZehong13/following{/other_user}", "gist...
[]
closed
false
null
[]
null
0
2020-07-29 17:23:34+00:00
2020-07-29 21:45:52+00:00
2020-07-29 21:45:52+00:00
NONE
null
null
null
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/454/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/454/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/454.diff", "html_url": "https://github.com/huggingface/datasets/pull/454", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/454.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/454" }
true
https://api.github.com/repos/huggingface/datasets/issues/453
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/453/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/453/comments
https://api.github.com/repos/huggingface/datasets/issues/453/events
https://github.com/huggingface/datasets/pull/453
667,728,247
MDExOlB1bGxSZXF1ZXN0NDU4MzQwNzky
453
add builder tests
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
0
2020-07-29 10:22:07+00:00
2020-07-29 11:14:06+00:00
2020-07-29 11:14:05+00:00
MEMBER
null
null
null
null
I added `as_dataset` and `download_and_prepare` to the tests
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/453/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/453/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/453.diff", "html_url": "https://github.com/huggingface/datasets/pull/453", "merged_at": "2020-07-29T11:14:05Z", "patch_url": "https://github.com/huggingface/datasets/pull/453.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/453...
true
https://api.github.com/repos/huggingface/datasets/issues/452
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/452/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/452/comments
https://api.github.com/repos/huggingface/datasets/issues/452/events
https://github.com/huggingface/datasets/pull/452
667,498,295
MDExOlB1bGxSZXF1ZXN0NDU4MTUzNjQy
452
Guardian authorship dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/25109412?v=4", "events_url": "https://api.github.com/users/malikaltakrori/events{/privacy}", "followers_url": "https://api.github.com/users/malikaltakrori/followers", "following_url": "https://api.github.com/users/malikaltakrori/following{/other_user}", ...
[]
closed
false
null
[]
null
6
2020-07-29 02:23:57+00:00
2020-08-20 15:09:57+00:00
2020-08-20 15:07:56+00:00
CONTRIBUTOR
null
null
null
null
A new dataset: Guardian news articles for authorship attribution **tests passed:** python nlp-cli dummy_data datasets/guardian_authorship --save_infos --all_configs RUN_SLOW=1 pytest tests/test_dataset_common.py::LocalDatasetTest::test_load_dataset_all_configs_guardian_authorship **Tests failed:** Real data:...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/452/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/452/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/452.diff", "html_url": "https://github.com/huggingface/datasets/pull/452", "merged_at": "2020-08-20T15:07:55Z", "patch_url": "https://github.com/huggingface/datasets/pull/452.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/452...
true
https://api.github.com/repos/huggingface/datasets/issues/451
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/451/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/451/comments
https://api.github.com/repos/huggingface/datasets/issues/451/events
https://github.com/huggingface/datasets/pull/451
667,210,468
MDExOlB1bGxSZXF1ZXN0NDU3OTIxNDMx
451
Fix csv/json/txt cache dir
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
4
2020-07-28 16:30:51+00:00
2020-07-29 13:57:23+00:00
2020-07-29 13:57:22+00:00
MEMBER
null
null
null
null
The cache dir for csv/json/txt datasets was always the same. This is an issue because it should be different depending on the data files provided by the user. To fix that, I added a line that use the hash of the data files provided by the user to define the cache dir. This should fix #444
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/451/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/451/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/451.diff", "html_url": "https://github.com/huggingface/datasets/pull/451", "merged_at": "2020-07-29T13:57:22Z", "patch_url": "https://github.com/huggingface/datasets/pull/451.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/451...
true
https://api.github.com/repos/huggingface/datasets/issues/450
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/450/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/450/comments
https://api.github.com/repos/huggingface/datasets/issues/450/events
https://github.com/huggingface/datasets/pull/450
667,074,120
MDExOlB1bGxSZXF1ZXN0NDU3ODA5ODA2
450
add sogou_news
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
[]
closed
false
null
[]
null
0
2020-07-28 13:29:10+00:00
2020-07-29 13:30:18+00:00
2020-07-29 13:30:17+00:00
CONTRIBUTOR
null
null
null
null
This PR adds the sogou news dataset #353
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/450/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/450/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/450.diff", "html_url": "https://github.com/huggingface/datasets/pull/450", "merged_at": "2020-07-29T13:30:17Z", "patch_url": "https://github.com/huggingface/datasets/pull/450.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/450...
true
https://api.github.com/repos/huggingface/datasets/issues/449
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/449/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/449/comments
https://api.github.com/repos/huggingface/datasets/issues/449/events
https://github.com/huggingface/datasets/pull/449
666,898,923
MDExOlB1bGxSZXF1ZXN0NDU3NjY0NjYx
449
add reuters21578 dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
[]
closed
false
null
[]
null
3
2020-07-28 08:58:12+00:00
2023-09-24 09:49:28+00:00
2020-08-03 11:10:31+00:00
CONTRIBUTOR
null
null
null
null
This PR adds the `Reuters_21578` dataset https://kdd.ics.uci.edu/databases/reuters21578/reuters21578.html #353 The datasets is a lit of `.sgm` files which are a bit different from xml file indeed `xml.etree` couldn't be used to read files. I consider them as text file (to avoid using external library) and read ...
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/449/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/449/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/449.diff", "html_url": "https://github.com/huggingface/datasets/pull/449", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/449.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/449" }
true
https://api.github.com/repos/huggingface/datasets/issues/448
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/448/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/448/comments
https://api.github.com/repos/huggingface/datasets/issues/448/events
https://github.com/huggingface/datasets/pull/448
666,893,443
MDExOlB1bGxSZXF1ZXN0NDU3NjYwMDU2
448
add aws load metric test
{ "avatar_url": "https://avatars.githubusercontent.com/u/5303103?v=4", "events_url": "https://api.github.com/users/idoh/events{/privacy}", "followers_url": "https://api.github.com/users/idoh/followers", "following_url": "https://api.github.com/users/idoh/following{/other_user}", "gists_url": "https://api.gith...
[]
closed
false
null
[]
null
3
2020-07-28 08:50:22+00:00
2020-07-28 15:02:27+00:00
2020-07-28 15:02:27+00:00
CONTRIBUTOR
null
null
null
null
Following issue #445 Added a test to recognize import errors of all metrics
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/448/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/448/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/448.diff", "html_url": "https://github.com/huggingface/datasets/pull/448", "merged_at": "2020-07-28T15:02:26Z", "patch_url": "https://github.com/huggingface/datasets/pull/448.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/448...
true
https://api.github.com/repos/huggingface/datasets/issues/447
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/447/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/447/comments
https://api.github.com/repos/huggingface/datasets/issues/447/events
https://github.com/huggingface/datasets/pull/447
666,842,115
MDExOlB1bGxSZXF1ZXN0NDU3NjE2NDA0
447
[BugFix] fix wrong import of DEFAULT_TOKENIZER
{ "avatar_url": "https://avatars.githubusercontent.com/u/5303103?v=4", "events_url": "https://api.github.com/users/idoh/events{/privacy}", "followers_url": "https://api.github.com/users/idoh/followers", "following_url": "https://api.github.com/users/idoh/following{/other_user}", "gists_url": "https://api.gith...
[]
closed
false
null
[]
null
0
2020-07-28 07:41:10+00:00
2020-07-28 12:58:01+00:00
2020-07-28 12:52:05+00:00
CONTRIBUTOR
null
null
null
null
Fixed the path to `DEFAULT_TOKENIZER` #445
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/447/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/447/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/447.diff", "html_url": "https://github.com/huggingface/datasets/pull/447", "merged_at": "2020-07-28T12:52:05Z", "patch_url": "https://github.com/huggingface/datasets/pull/447.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/447...
true
https://api.github.com/repos/huggingface/datasets/issues/446
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/446/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/446/comments
https://api.github.com/repos/huggingface/datasets/issues/446/events
https://github.com/huggingface/datasets/pull/446
666,837,351
MDExOlB1bGxSZXF1ZXN0NDU3NjEyNTg5
446
[BugFix] fix wrong import of DEFAULT_TOKENIZER
{ "avatar_url": "https://avatars.githubusercontent.com/u/5303103?v=4", "events_url": "https://api.github.com/users/idoh/events{/privacy}", "followers_url": "https://api.github.com/users/idoh/followers", "following_url": "https://api.github.com/users/idoh/following{/other_user}", "gists_url": "https://api.gith...
[]
closed
false
null
[]
null
0
2020-07-28 07:32:47+00:00
2020-07-28 07:34:46+00:00
2020-07-28 07:33:59+00:00
CONTRIBUTOR
null
null
null
null
Fixed the path to `DEFAULT_TOKENIZER` #445
{ "avatar_url": "https://avatars.githubusercontent.com/u/5303103?v=4", "events_url": "https://api.github.com/users/idoh/events{/privacy}", "followers_url": "https://api.github.com/users/idoh/followers", "following_url": "https://api.github.com/users/idoh/following{/other_user}", "gists_url": "https://api.gith...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/446/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/446/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/446.diff", "html_url": "https://github.com/huggingface/datasets/pull/446", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/446.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/446" }
true
https://api.github.com/repos/huggingface/datasets/issues/445
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/445/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/445/comments
https://api.github.com/repos/huggingface/datasets/issues/445/events
https://github.com/huggingface/datasets/issues/445
666,836,658
MDU6SXNzdWU2NjY4MzY2NTg=
445
DEFAULT_TOKENIZER import error in sacrebleu
{ "avatar_url": "https://avatars.githubusercontent.com/u/5303103?v=4", "events_url": "https://api.github.com/users/idoh/events{/privacy}", "followers_url": "https://api.github.com/users/idoh/followers", "following_url": "https://api.github.com/users/idoh/following{/other_user}", "gists_url": "https://api.gith...
[]
closed
false
null
[]
null
1
2020-07-28 07:31:30+00:00
2020-07-28 12:58:56+00:00
2020-07-28 12:58:56+00:00
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
Latest Version 0.3.0 When loading the metric "sacrebleu" there is an import error due to the wrong path ![image](https://user-images.githubusercontent.com/5303103/88633063-2c5e5f00-d0bd-11ea-8ca8-4704dc975433.png)
{ "avatar_url": "https://avatars.githubusercontent.com/u/5303103?v=4", "events_url": "https://api.github.com/users/idoh/events{/privacy}", "followers_url": "https://api.github.com/users/idoh/followers", "following_url": "https://api.github.com/users/idoh/following{/other_user}", "gists_url": "https://api.gith...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/445/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/445/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/444
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/444/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/444/comments
https://api.github.com/repos/huggingface/datasets/issues/444/events
https://github.com/huggingface/datasets/issues/444
666,280,842
MDU6SXNzdWU2NjYyODA4NDI=
444
Keep loading old file even I specify a new file in load_dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/10594453?v=4", "events_url": "https://api.github.com/users/joshhu/events{/privacy}", "followers_url": "https://api.github.com/users/joshhu/followers", "following_url": "https://api.github.com/users/joshhu/following{/other_user}", "gists_url": "https://a...
[ { "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library", "id": 2067388877, "name": "dataset bug", "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug" } ]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[ { "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists...
null
2
2020-07-27 13:08:06+00:00
2020-07-29 13:57:22+00:00
2020-07-29 13:57:22+00:00
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
I used load a file called 'a.csv' by ``` dataset = load_dataset('csv', data_file='./a.csv') ``` And after a while, I tried to load another csv called 'b.csv' ``` dataset = load_dataset('csv', data_file='./b.csv') ``` However, the new dataset seems to remain the old 'a.csv' and not loading new csv file. Even...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/444/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/444/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/443
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/443/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/443/comments
https://api.github.com/repos/huggingface/datasets/issues/443/events
https://github.com/huggingface/datasets/issues/443
666,246,716
MDU6SXNzdWU2NjYyNDY3MTY=
443
Cannot unpickle saved .pt dataset with torch.save()/load()
{ "avatar_url": "https://avatars.githubusercontent.com/u/24683907?v=4", "events_url": "https://api.github.com/users/vegarab/events{/privacy}", "followers_url": "https://api.github.com/users/vegarab/followers", "following_url": "https://api.github.com/users/vegarab/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
1
2020-07-27 12:13:37+00:00
2020-07-27 13:05:11+00:00
2020-07-27 13:05:11+00:00
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
Saving a formatted torch dataset to file using `torch.save()`. Loading the same file fails during unpickling: ```python >>> import torch >>> import nlp >>> squad = nlp.load_dataset("squad.py", split="train") >>> squad Dataset(features: {'source_text': Value(dtype='string', id=None), 'target_text': Value(dtype...
{ "avatar_url": "https://avatars.githubusercontent.com/u/24683907?v=4", "events_url": "https://api.github.com/users/vegarab/events{/privacy}", "followers_url": "https://api.github.com/users/vegarab/followers", "following_url": "https://api.github.com/users/vegarab/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/443/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/443/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/442
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/442/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/442/comments
https://api.github.com/repos/huggingface/datasets/issues/442/events
https://github.com/huggingface/datasets/issues/442
666,201,810
MDU6SXNzdWU2NjYyMDE4MTA=
442
[Suggestion] Glue Diagnostic Data with Labels
{ "avatar_url": "https://avatars.githubusercontent.com/u/3662782?v=4", "events_url": "https://api.github.com/users/ggbetz/events{/privacy}", "followers_url": "https://api.github.com/users/ggbetz/followers", "following_url": "https://api.github.com/users/ggbetz/following{/other_user}", "gists_url": "https://ap...
[ { "color": "72f99f", "default": false, "description": "Discussions on the datasets", "id": 2067401494, "name": "Dataset discussion", "node_id": "MDU6TGFiZWwyMDY3NDAxNDk0", "url": "https://api.github.com/repos/huggingface/datasets/labels/Dataset%20discussion" } ]
open
false
null
[]
null
0
2020-07-27 10:59:58+00:00
2020-08-24 15:13:20+00:00
NaT
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
Hello! First of all, thanks for setting up this useful project! I've just realised you provide the the [Glue Diagnostics Data](https://huggingface.co/nlp/viewer/?dataset=glue&config=ax) without labels, indicating in the `GlueConfig` that you've only a test set. Yet, the data with labels is available, too (see als...
null
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/442/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/442/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/441
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/441/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/441/comments
https://api.github.com/repos/huggingface/datasets/issues/441/events
https://github.com/huggingface/datasets/pull/441
666,148,413
MDExOlB1bGxSZXF1ZXN0NDU3MDQyMjY3
441
Add features parameter in load dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
2
2020-07-27 09:50:01+00:00
2020-07-30 12:51:17+00:00
2020-07-30 12:51:16+00:00
MEMBER
null
null
null
null
Added `features` argument in `nlp.load_dataset`. If they don't match the data type, it raises a `ValueError`. It's a draft PR because #440 needs to be merged first.
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/441/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/441/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/441.diff", "html_url": "https://github.com/huggingface/datasets/pull/441", "merged_at": "2020-07-30T12:51:16Z", "patch_url": "https://github.com/huggingface/datasets/pull/441.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/441...
true
https://api.github.com/repos/huggingface/datasets/issues/440
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/440/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/440/comments
https://api.github.com/repos/huggingface/datasets/issues/440/events
https://github.com/huggingface/datasets/pull/440
666,116,823
MDExOlB1bGxSZXF1ZXN0NDU3MDE2MjQy
440
Fix user specified features in map
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
0
2020-07-27 09:04:26+00:00
2020-07-28 09:25:23+00:00
2020-07-28 09:25:22+00:00
MEMBER
null
null
null
null
`.map` didn't keep the user specified features because of an issue in the writer. The writer used to overwrite the user specified features with inferred features. I also added tests to make sure it doesn't happen again.
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/440/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/440/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/440.diff", "html_url": "https://github.com/huggingface/datasets/pull/440", "merged_at": "2020-07-28T09:25:22Z", "patch_url": "https://github.com/huggingface/datasets/pull/440.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/440...
true
https://api.github.com/repos/huggingface/datasets/issues/439
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/439/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/439/comments
https://api.github.com/repos/huggingface/datasets/issues/439/events
https://github.com/huggingface/datasets/issues/439
665,964,673
MDU6SXNzdWU2NjU5NjQ2NzM=
439
Issues: Adding a FAISS or Elastic Search index to a Dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/431890?v=4", "events_url": "https://api.github.com/users/nsankar/events{/privacy}", "followers_url": "https://api.github.com/users/nsankar/followers", "following_url": "https://api.github.com/users/nsankar/following{/other_user}", "gists_url": "https://...
[]
closed
false
null
[]
null
5
2020-07-27 04:25:17+00:00
2020-10-28 01:46:24+00:00
2020-10-28 01:46:24+00:00
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
It seems the DPRContextEncoder, DPRContextEncoderTokenizer cited[ in this documentation](https://huggingface.co/nlp/faiss_and_ea.html) is not implemented ? It didnot work with the standard nlp installation . Also, I couldn't find or use it with the latest nlp install from github in Colab. Is there any dependency on t...
{ "avatar_url": "https://avatars.githubusercontent.com/u/431890?v=4", "events_url": "https://api.github.com/users/nsankar/events{/privacy}", "followers_url": "https://api.github.com/users/nsankar/followers", "following_url": "https://api.github.com/users/nsankar/following{/other_user}", "gists_url": "https://...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/439/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/439/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/438
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/438/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/438/comments
https://api.github.com/repos/huggingface/datasets/issues/438/events
https://github.com/huggingface/datasets/issues/438
665,865,490
MDU6SXNzdWU2NjU4NjU0OTA=
438
New Datasets: IWSLT15+, ITTB
{ "avatar_url": "https://avatars.githubusercontent.com/u/6045025?v=4", "events_url": "https://api.github.com/users/sshleifer/events{/privacy}", "followers_url": "https://api.github.com/users/sshleifer/followers", "following_url": "https://api.github.com/users/sshleifer/following{/other_user}", "gists_url": "h...
[ { "color": "e99695", "default": false, "description": "Requesting to add a new dataset", "id": 2067376369, "name": "dataset request", "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request" } ]
open
false
null
[]
null
2
2020-07-26 21:43:04+00:00
2020-08-24 15:12:15+00:00
NaT
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
**Links:** [iwslt](https://pytorchnlp.readthedocs.io/en/latest/_modules/torchnlp/datasets/iwslt.html) Don't know if that link is up to date. [ittb](http://www.cfilt.iitb.ac.in/iitb_parallel/) **Motivation**: replicate mbart finetuning results (table below) ![image](https://user-images.githubusercontent.com/60450...
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/438/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/438/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/437
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/437/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/437/comments
https://api.github.com/repos/huggingface/datasets/issues/437/events
https://github.com/huggingface/datasets/pull/437
665,597,176
MDExOlB1bGxSZXF1ZXN0NDU2NjIzNjc3
437
Fix XTREME PAN-X loading
{ "avatar_url": "https://avatars.githubusercontent.com/u/8264887?v=4", "events_url": "https://api.github.com/users/lvwerra/events{/privacy}", "followers_url": "https://api.github.com/users/lvwerra/followers", "following_url": "https://api.github.com/users/lvwerra/following{/other_user}", "gists_url": "https:/...
[]
closed
false
null
[]
null
4
2020-07-25 14:44:57+00:00
2020-07-30 08:28:15+00:00
2020-07-30 08:28:15+00:00
MEMBER
null
null
null
null
Hi 🤗 In response to the discussion in #425 @lewtun and I made some fixes to the repo. In the original XTREME implementation the PAN-X dataset for named entity recognition loaded each word/tag pair as a single row and the sentence relation was lost. With the fix each row contains the list of all words in a single sen...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/437/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/437/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/437.diff", "html_url": "https://github.com/huggingface/datasets/pull/437", "merged_at": "2020-07-30T08:28:15Z", "patch_url": "https://github.com/huggingface/datasets/pull/437.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/437...
true
https://api.github.com/repos/huggingface/datasets/issues/436
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/436/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/436/comments
https://api.github.com/repos/huggingface/datasets/issues/436/events
https://github.com/huggingface/datasets/issues/436
665,582,167
MDU6SXNzdWU2NjU1ODIxNjc=
436
Google Colab - load_dataset - PyArrow exception
{ "avatar_url": "https://avatars.githubusercontent.com/u/431890?v=4", "events_url": "https://api.github.com/users/nsankar/events{/privacy}", "followers_url": "https://api.github.com/users/nsankar/followers", "following_url": "https://api.github.com/users/nsankar/following{/other_user}", "gists_url": "https://...
[]
closed
false
null
[]
null
9
2020-07-25 13:05:20+00:00
2020-08-20 08:08:18+00:00
2020-08-20 08:08:18+00:00
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
With latest PyArrow 1.0.0 installed, I get the following exception . Restarting colab has the same issue ImportWarning: To use `nlp`, the module `pyarrow>=0.16.0` is required, and the current version of `pyarrow` doesn't match this condition. If you are running this in a Google Colab, you should probably just rest...
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/436/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/436/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/435
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/435/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/435/comments
https://api.github.com/repos/huggingface/datasets/issues/435/events
https://github.com/huggingface/datasets/issues/435
665,507,141
MDU6SXNzdWU2NjU1MDcxNDE=
435
ImportWarning for pyarrow 1.0.0
{ "avatar_url": "https://avatars.githubusercontent.com/u/18187806?v=4", "events_url": "https://api.github.com/users/HanGuo97/events{/privacy}", "followers_url": "https://api.github.com/users/HanGuo97/followers", "following_url": "https://api.github.com/users/HanGuo97/following{/other_user}", "gists_url": "htt...
[]
closed
false
null
[]
null
4
2020-07-25 03:44:39+00:00
2020-09-08 17:57:15+00:00
2020-08-03 16:37:32+00:00
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
The following PR raised ImportWarning at `pyarrow ==1.0.0` https://github.com/huggingface/nlp/pull/265/files
{ "avatar_url": "https://avatars.githubusercontent.com/u/18187806?v=4", "events_url": "https://api.github.com/users/HanGuo97/events{/privacy}", "followers_url": "https://api.github.com/users/HanGuo97/followers", "following_url": "https://api.github.com/users/HanGuo97/following{/other_user}", "gists_url": "htt...
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/435/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/435/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/434
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/434/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/434/comments
https://api.github.com/repos/huggingface/datasets/issues/434/events
https://github.com/huggingface/datasets/pull/434
665,477,638
MDExOlB1bGxSZXF1ZXN0NDU2NTM3Njgz
434
Fixed check for pyarrow
{ "avatar_url": "https://avatars.githubusercontent.com/u/58701810?v=4", "events_url": "https://api.github.com/users/nadahlberg/events{/privacy}", "followers_url": "https://api.github.com/users/nadahlberg/followers", "following_url": "https://api.github.com/users/nadahlberg/following{/other_user}", "gists_url"...
[]
closed
false
null
[]
null
1
2020-07-25 00:16:53+00:00
2020-07-25 06:36:34+00:00
2020-07-25 06:36:34+00:00
CONTRIBUTOR
null
null
null
null
Fix check for pyarrow in __init__.py. Previously would raise an error for pyarrow >= 1.0.0
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/434/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/434/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/434.diff", "html_url": "https://github.com/huggingface/datasets/pull/434", "merged_at": "2020-07-25T06:36:34Z", "patch_url": "https://github.com/huggingface/datasets/pull/434.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/434...
true
https://api.github.com/repos/huggingface/datasets/issues/433
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/433/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/433/comments
https://api.github.com/repos/huggingface/datasets/issues/433/events
https://github.com/huggingface/datasets/issues/433
665,311,025
MDU6SXNzdWU2NjUzMTEwMjU=
433
How to reuse functionality of a (generic) dataset?
{ "avatar_url": "https://avatars.githubusercontent.com/u/3375489?v=4", "events_url": "https://api.github.com/users/ArneBinder/events{/privacy}", "followers_url": "https://api.github.com/users/ArneBinder/followers", "following_url": "https://api.github.com/users/ArneBinder/following{/other_user}", "gists_url":...
[]
closed
false
null
[]
null
4
2020-07-24 17:27:37+00:00
2022-10-04 17:59:34+00:00
2022-10-04 17:59:33+00:00
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
I have written a generic dataset for corpora created with the Brat annotation tool ([specification](https://brat.nlplab.org/standoff.html), [dataset code](https://github.com/ArneBinder/nlp/blob/brat/datasets/brat/brat.py)). Now I wonder how to use that to create specific dataset instances. What's the recommended way to...
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url"...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/433/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/433/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/432
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/432/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/432/comments
https://api.github.com/repos/huggingface/datasets/issues/432/events
https://github.com/huggingface/datasets/pull/432
665,234,340
MDExOlB1bGxSZXF1ZXN0NDU2MzQxNDk3
432
Fix handling of config files while loading datasets from multiple processes
{ "avatar_url": "https://avatars.githubusercontent.com/u/99543?v=4", "events_url": "https://api.github.com/users/orsharir/events{/privacy}", "followers_url": "https://api.github.com/users/orsharir/followers", "following_url": "https://api.github.com/users/orsharir/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
4
2020-07-24 15:10:57+00:00
2020-08-01 17:11:42+00:00
2020-07-30 08:25:28+00:00
CONTRIBUTOR
null
null
null
null
When loading shards on several processes, each process upon loading the dataset will overwrite dataset_infos.json in <package path>/datasets/<dataset name>/<hash>/dataset_infos.json. It does so every time, even when the target file already exists and is identical. Because multiple processes rewrite the same file in par...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/432/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/432/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/432.diff", "html_url": "https://github.com/huggingface/datasets/pull/432", "merged_at": "2020-07-30T08:25:28Z", "patch_url": "https://github.com/huggingface/datasets/pull/432.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/432...
true
https://api.github.com/repos/huggingface/datasets/issues/431
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/431/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/431/comments
https://api.github.com/repos/huggingface/datasets/issues/431/events
https://github.com/huggingface/datasets/pull/431
665,044,416
MDExOlB1bGxSZXF1ZXN0NDU2MTgyNDE2
431
Specify split post processing + Add post processing resources downloading
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
4
2020-07-24 09:29:19+00:00
2020-07-31 09:05:04+00:00
2020-07-31 09:05:03+00:00
MEMBER
null
null
null
null
Previously if you tried to do ```python from nlp import load_dataset wiki = load_dataset("wiki_dpr", "psgs_w100_with_nq_embeddings", split="train[:100]", with_index=True) ``` Then you'd get an error `Index size should match Dataset size...` This was because it was trying to use the full index (21M elements). ...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/431/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/431/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/431.diff", "html_url": "https://github.com/huggingface/datasets/pull/431", "merged_at": "2020-07-31T09:05:03Z", "patch_url": "https://github.com/huggingface/datasets/pull/431.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/431...
true
https://api.github.com/repos/huggingface/datasets/issues/430
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/430/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/430/comments
https://api.github.com/repos/huggingface/datasets/issues/430/events
https://github.com/huggingface/datasets/pull/430
664,583,837
MDExOlB1bGxSZXF1ZXN0NDU1ODAxOTI2
430
add DatasetDict
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
4
2020-07-23 15:43:49+00:00
2020-08-04 01:01:53+00:00
2020-07-29 09:06:22+00:00
MEMBER
null
null
null
null
## Add DatasetDict ### Overview When you call `load_dataset` it can return a dictionary of datasets if there are several splits (train/test for example). If you wanted to apply dataset transforms you had to iterate over each split and apply the transform. Instead of returning a dict, it now returns a `nlp.Dat...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/430/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/430/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/430.diff", "html_url": "https://github.com/huggingface/datasets/pull/430", "merged_at": "2020-07-29T09:06:22Z", "patch_url": "https://github.com/huggingface/datasets/pull/430.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/430...
true
https://api.github.com/repos/huggingface/datasets/issues/429
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/429/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/429/comments
https://api.github.com/repos/huggingface/datasets/issues/429/events
https://github.com/huggingface/datasets/pull/429
664,412,137
MDExOlB1bGxSZXF1ZXN0NDU1NjU2MDk5
429
mlsum
{ "avatar_url": "https://avatars.githubusercontent.com/u/36986299?v=4", "events_url": "https://api.github.com/users/RachelKer/events{/privacy}", "followers_url": "https://api.github.com/users/RachelKer/followers", "following_url": "https://api.github.com/users/RachelKer/following{/other_user}", "gists_url": "...
[]
closed
false
null
[]
null
6
2020-07-23 11:52:39+00:00
2020-07-31 11:46:20+00:00
2020-07-31 11:46:20+00:00
CONTRIBUTOR
null
null
null
null
Hello, The tests for the load_real_data fail, as there is no default language subset to download it looks for a file that does not exist. This bug does not happen when using the load_dataset function, as it asks you to specify a language if you do not, so I submit this PR anyway. The dataset is avalaible on : https...
{ "avatar_url": "https://avatars.githubusercontent.com/u/36986299?v=4", "events_url": "https://api.github.com/users/RachelKer/events{/privacy}", "followers_url": "https://api.github.com/users/RachelKer/followers", "following_url": "https://api.github.com/users/RachelKer/following{/other_user}", "gists_url": "...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/429/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/429/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/429.diff", "html_url": "https://github.com/huggingface/datasets/pull/429", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/429.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/429" }
true
https://api.github.com/repos/huggingface/datasets/issues/428
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/428/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/428/comments
https://api.github.com/repos/huggingface/datasets/issues/428/events
https://github.com/huggingface/datasets/pull/428
664,367,086
MDExOlB1bGxSZXF1ZXN0NDU1NjE3Nzcy
428
fix concatenate_datasets
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
0
2020-07-23 10:30:59+00:00
2020-07-23 10:35:00+00:00
2020-07-23 10:34:58+00:00
MEMBER
null
null
null
null
`concatenate_datatsets` used to test that the different`nlp.Dataset.schema` match, but this attribute was removed in #423
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/428/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/428/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/428.diff", "html_url": "https://github.com/huggingface/datasets/pull/428", "merged_at": "2020-07-23T10:34:58Z", "patch_url": "https://github.com/huggingface/datasets/pull/428.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/428...
true
https://api.github.com/repos/huggingface/datasets/issues/427
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/427/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/427/comments
https://api.github.com/repos/huggingface/datasets/issues/427/events
https://github.com/huggingface/datasets/pull/427
664,341,623
MDExOlB1bGxSZXF1ZXN0NDU1NTk1Nzc3
427
Allow sequence features for beam + add processed Natural Questions
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
0
2020-07-23 09:52:41+00:00
2020-07-23 13:09:30+00:00
2020-07-23 13:09:29+00:00
MEMBER
null
null
null
null
## Allow Sequence features for Beam Datasets + add Natural Questions ### The issue The steps of beam datasets processing is the following: - download the source files and send them in a remote storage (gcs) - process the files using a beam runner (dataflow) - save output in remote storage (gcs) - convert outp...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 1, "laugh": 0, "rocket": 3, "total_count": 4, "url": "https://api.github.com/repos/huggingface/datasets/issues/427/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/427/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/427.diff", "html_url": "https://github.com/huggingface/datasets/pull/427", "merged_at": "2020-07-23T13:09:29Z", "patch_url": "https://github.com/huggingface/datasets/pull/427.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/427...
true
https://api.github.com/repos/huggingface/datasets/issues/426
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/426/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/426/comments
https://api.github.com/repos/huggingface/datasets/issues/426/events
https://github.com/huggingface/datasets/issues/426
664,203,897
MDU6SXNzdWU2NjQyMDM4OTc=
426
[FEATURE REQUEST] Multiprocessing with for dataset.map, dataset.filter
{ "avatar_url": "https://avatars.githubusercontent.com/u/2000204?v=4", "events_url": "https://api.github.com/users/timothyjlaurent/events{/privacy}", "followers_url": "https://api.github.com/users/timothyjlaurent/followers", "following_url": "https://api.github.com/users/timothyjlaurent/following{/other_user}",...
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
closed
false
null
[]
null
6
2020-07-23 05:00:41+00:00
2021-03-12 09:34:12+00:00
2020-09-07 14:48:04+00:00
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
It would be nice to be able to speed up `dataset.map` or `dataset.filter`. Perhaps this is as easy as sharding the dataset sending each shard to a process/thread/dask pool and using the new `nlp.concatenate_dataset()` function to join them all together?
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 2, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/426/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/426/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/425
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/425/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/425/comments
https://api.github.com/repos/huggingface/datasets/issues/425/events
https://github.com/huggingface/datasets/issues/425
664,029,848
MDU6SXNzdWU2NjQwMjk4NDg=
425
Correct data structure for PAN-X task in XTREME dataset?
{ "avatar_url": "https://avatars.githubusercontent.com/u/26859204?v=4", "events_url": "https://api.github.com/users/lewtun/events{/privacy}", "followers_url": "https://api.github.com/users/lewtun/followers", "following_url": "https://api.github.com/users/lewtun/following{/other_user}", "gists_url": "https://a...
[]
closed
false
null
[]
null
7
2020-07-22 20:29:20+00:00
2020-08-02 13:30:34+00:00
2020-08-02 13:30:34+00:00
MEMBER
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
Hi 🤗 team! ## Description of the problem Thanks to the fix from #416 I am now able to load the NER task in the XTREME dataset as follows: ```python from nlp import load_dataset # AmazonPhotos.zip is located in data/ dataset = load_dataset("xtreme", "PAN-X.en", data_dir='./data') dataset_train = dataset['tr...
{ "avatar_url": "https://avatars.githubusercontent.com/u/26859204?v=4", "events_url": "https://api.github.com/users/lewtun/events{/privacy}", "followers_url": "https://api.github.com/users/lewtun/followers", "following_url": "https://api.github.com/users/lewtun/following{/other_user}", "gists_url": "https://a...
{ "+1": 2, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/425/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/425/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/424
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/424/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/424/comments
https://api.github.com/repos/huggingface/datasets/issues/424/events
https://github.com/huggingface/datasets/pull/424
663,858,552
MDExOlB1bGxSZXF1ZXN0NDU1MTk4MTY0
424
Web of science
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
[]
closed
false
null
[]
null
0
2020-07-22 15:38:31+00:00
2020-07-23 14:27:58+00:00
2020-07-23 14:27:56+00:00
CONTRIBUTOR
null
null
null
null
this PR adds the WebofScience dataset #353
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/424/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/424/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/424.diff", "html_url": "https://github.com/huggingface/datasets/pull/424", "merged_at": "2020-07-23T14:27:56Z", "patch_url": "https://github.com/huggingface/datasets/pull/424.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/424...
true
https://api.github.com/repos/huggingface/datasets/issues/423
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/423/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/423/comments
https://api.github.com/repos/huggingface/datasets/issues/423/events
https://github.com/huggingface/datasets/pull/423
663,079,359
MDExOlB1bGxSZXF1ZXN0NDU0NTU4OTA0
423
Change features vs schema logic
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
2
2020-07-21 14:52:47+00:00
2020-07-25 09:08:34+00:00
2020-07-23 10:15:17+00:00
MEMBER
null
null
null
null
## New logic for `nlp.Features` in datasets Previously, it was confusing to have `features` and pyarrow's `schema` in `nlp.Dataset`. However `features` is supposed to be the front-facing object to define the different fields of a dataset, while `schema` is only used to write arrow files. Changes: - Remove `sche...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 2, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/423/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/423/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/423.diff", "html_url": "https://github.com/huggingface/datasets/pull/423", "merged_at": "2020-07-23T10:15:16Z", "patch_url": "https://github.com/huggingface/datasets/pull/423.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/423...
true
https://api.github.com/repos/huggingface/datasets/issues/422
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/422/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/422/comments
https://api.github.com/repos/huggingface/datasets/issues/422/events
https://github.com/huggingface/datasets/pull/422
663,028,497
MDExOlB1bGxSZXF1ZXN0NDU0NTE3MDU2
422
- Corrected encoding for IMDB.
{ "avatar_url": "https://avatars.githubusercontent.com/u/25091538?v=4", "events_url": "https://api.github.com/users/ghazi-f/events{/privacy}", "followers_url": "https://api.github.com/users/ghazi-f/followers", "following_url": "https://api.github.com/users/ghazi-f/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
0
2020-07-21 13:46:59+00:00
2020-07-22 16:02:53+00:00
2020-07-22 16:02:53+00:00
CONTRIBUTOR
null
null
null
null
The preparation phase (after the download phase) crashed on windows because of charmap encoding not being able to decode certain characters. This change suggested in Issue #347 fixes it for the IMDB dataset.
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/422/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/422/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/422.diff", "html_url": "https://github.com/huggingface/datasets/pull/422", "merged_at": "2020-07-22T16:02:53Z", "patch_url": "https://github.com/huggingface/datasets/pull/422.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/422...
true
https://api.github.com/repos/huggingface/datasets/issues/421
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/421/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/421/comments
https://api.github.com/repos/huggingface/datasets/issues/421/events
https://github.com/huggingface/datasets/pull/421
662,213,864
MDExOlB1bGxSZXF1ZXN0NDUzNzkzMzQ1
421
Style change
{ "avatar_url": "https://avatars.githubusercontent.com/u/35500534?v=4", "events_url": "https://api.github.com/users/lordtt13/events{/privacy}", "followers_url": "https://api.github.com/users/lordtt13/followers", "following_url": "https://api.github.com/users/lordtt13/following{/other_user}", "gists_url": "htt...
[]
closed
false
null
[]
null
3
2020-07-20 20:08:29+00:00
2020-07-22 16:08:40+00:00
2020-07-22 16:08:39+00:00
CONTRIBUTOR
null
null
null
null
make quality and make style ran on scripts
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/421/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/421/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/421.diff", "html_url": "https://github.com/huggingface/datasets/pull/421", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/421.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/421" }
true
https://api.github.com/repos/huggingface/datasets/issues/420
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/420/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/420/comments
https://api.github.com/repos/huggingface/datasets/issues/420/events
https://github.com/huggingface/datasets/pull/420
662,029,782
MDExOlB1bGxSZXF1ZXN0NDUzNjI5OTk2
420
Better handle nested features
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
0
2020-07-20 16:44:13+00:00
2020-07-21 08:20:49+00:00
2020-07-21 08:09:52+00:00
MEMBER
null
null
null
null
Changes: - added arrow schema to features conversion (it's going to be useful to fix #342 ) - make flatten handle deep features (useful for tfrecords conversion in #339 ) - add tests for flatten and features conversions - the reader now returns the kwargs to instantiate a Dataset (fix circular dependencies)
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/420/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/420/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/420.diff", "html_url": "https://github.com/huggingface/datasets/pull/420", "merged_at": "2020-07-21T08:09:51Z", "patch_url": "https://github.com/huggingface/datasets/pull/420.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/420...
true
https://api.github.com/repos/huggingface/datasets/issues/419
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/419/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/419/comments
https://api.github.com/repos/huggingface/datasets/issues/419/events
https://github.com/huggingface/datasets/pull/419
661,974,747
MDExOlB1bGxSZXF1ZXN0NDUzNTgxNzQz
419
EmoContext dataset add
{ "avatar_url": "https://avatars.githubusercontent.com/u/35500534?v=4", "events_url": "https://api.github.com/users/lordtt13/events{/privacy}", "followers_url": "https://api.github.com/users/lordtt13/followers", "following_url": "https://api.github.com/users/lordtt13/following{/other_user}", "gists_url": "htt...
[]
closed
false
null
[]
null
0
2020-07-20 15:48:45+00:00
2020-07-24 08:22:01+00:00
2020-07-24 08:22:00+00:00
CONTRIBUTOR
null
null
null
null
EmoContext Dataset add Signed-off-by: lordtt13 <thakurtanmay72@yahoo.com>
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/419/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/419/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/419.diff", "html_url": "https://github.com/huggingface/datasets/pull/419", "merged_at": "2020-07-24T08:22:00Z", "patch_url": "https://github.com/huggingface/datasets/pull/419.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/419...
true
https://api.github.com/repos/huggingface/datasets/issues/418
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/418/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/418/comments
https://api.github.com/repos/huggingface/datasets/issues/418/events
https://github.com/huggingface/datasets/issues/418
661,914,873
MDU6SXNzdWU2NjE5MTQ4NzM=
418
Addition of google drive links to dl_manager
{ "avatar_url": "https://avatars.githubusercontent.com/u/35500534?v=4", "events_url": "https://api.github.com/users/lordtt13/events{/privacy}", "followers_url": "https://api.github.com/users/lordtt13/followers", "following_url": "https://api.github.com/users/lordtt13/following{/other_user}", "gists_url": "htt...
[]
closed
false
null
[]
null
3
2020-07-20 14:52:02+00:00
2020-07-20 15:39:32+00:00
2020-07-20 15:39:32+00:00
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
Hello there, I followed the template to create a download script of my own, which works fine for me, although I had to shun the dl_manager because it was downloading nothing from the drive links and instead use gdown. This is the script for me: ```python class EmoConfig(nlp.BuilderConfig): """BuilderConfig ...
{ "avatar_url": "https://avatars.githubusercontent.com/u/35500534?v=4", "events_url": "https://api.github.com/users/lordtt13/events{/privacy}", "followers_url": "https://api.github.com/users/lordtt13/followers", "following_url": "https://api.github.com/users/lordtt13/following{/other_user}", "gists_url": "htt...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/418/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/418/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/417
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/417/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/417/comments
https://api.github.com/repos/huggingface/datasets/issues/417/events
https://github.com/huggingface/datasets/pull/417
661,804,054
MDExOlB1bGxSZXF1ZXN0NDUzNDMyODE5
417
Fix docstrins multiple metrics instances
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
0
2020-07-20 13:08:59+00:00
2020-07-22 09:51:00+00:00
2020-07-22 09:50:59+00:00
MEMBER
null
null
null
null
We change the docstrings of `nlp.Metric.compute`, `nlp.Metric.add` and `nlp.Metric.add_batch` depending on which metric is instantiated. However we had issues when instantiating multiple metrics (docstrings were duplicated). This should fix #304
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/417/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/417/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/417.diff", "html_url": "https://github.com/huggingface/datasets/pull/417", "merged_at": "2020-07-22T09:50:58Z", "patch_url": "https://github.com/huggingface/datasets/pull/417.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/417...
true
https://api.github.com/repos/huggingface/datasets/issues/416
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/416/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/416/comments
https://api.github.com/repos/huggingface/datasets/issues/416/events
https://github.com/huggingface/datasets/pull/416
661,635,393
MDExOlB1bGxSZXF1ZXN0NDUzMjg1NTM4
416
Fix xtreme panx directory
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
1
2020-07-20 10:09:17+00:00
2020-07-21 08:15:46+00:00
2020-07-21 08:15:44+00:00
MEMBER
null
null
null
null
Fix #412
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/416/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/416/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/416.diff", "html_url": "https://github.com/huggingface/datasets/pull/416", "merged_at": "2020-07-21T08:15:44Z", "patch_url": "https://github.com/huggingface/datasets/pull/416.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/416...
true
https://api.github.com/repos/huggingface/datasets/issues/415
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/415/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/415/comments
https://api.github.com/repos/huggingface/datasets/issues/415/events
https://github.com/huggingface/datasets/issues/415
660,687,076
MDU6SXNzdWU2NjA2ODcwNzY=
415
Something is wrong with WMT 19 kk-en dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/32014649?v=4", "events_url": "https://api.github.com/users/ChenghaoMou/events{/privacy}", "followers_url": "https://api.github.com/users/ChenghaoMou/followers", "following_url": "https://api.github.com/users/ChenghaoMou/following{/other_user}", "gists_u...
[ { "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library", "id": 2067388877, "name": "dataset bug", "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug" } ]
open
false
null
[]
null
0
2020-07-19 08:18:51+00:00
2020-07-20 09:54:26+00:00
NaT
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
The translation in the `train` set does not look right: ``` >>>import nlp >>>from nlp import load_dataset >>>dataset = load_dataset('wmt19', 'kk-en') >>>dataset["train"]["translation"][0] {'kk': 'Trumpian Uncertainty', 'en': 'Трамптық белгісіздік'} >>>dataset["validation"]["translation"][0] {'kk': 'Ақша-несие...
null
{ "+1": 0, "-1": 0, "confused": 1, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/415/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/415/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/414
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/414/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/414/comments
https://api.github.com/repos/huggingface/datasets/issues/414/events
https://github.com/huggingface/datasets/issues/414
660,654,013
MDU6SXNzdWU2NjA2NTQwMTM=
414
from_dict delete?
{ "avatar_url": "https://avatars.githubusercontent.com/u/22817243?v=4", "events_url": "https://api.github.com/users/hackerxiaobai/events{/privacy}", "followers_url": "https://api.github.com/users/hackerxiaobai/followers", "following_url": "https://api.github.com/users/hackerxiaobai/following{/other_user}", "g...
[]
closed
false
null
[]
null
2
2020-07-19 07:08:36+00:00
2020-07-21 02:21:17+00:00
2020-07-21 02:21:17+00:00
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
AttributeError: type object 'Dataset' has no attribute 'from_dict'
{ "avatar_url": "https://avatars.githubusercontent.com/u/22817243?v=4", "events_url": "https://api.github.com/users/hackerxiaobai/events{/privacy}", "followers_url": "https://api.github.com/users/hackerxiaobai/followers", "following_url": "https://api.github.com/users/hackerxiaobai/following{/other_user}", "g...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/414/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/414/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/413
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/413/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/413/comments
https://api.github.com/repos/huggingface/datasets/issues/413/events
https://github.com/huggingface/datasets/issues/413
660,063,655
MDU6SXNzdWU2NjAwNjM2NTU=
413
Is there a way to download only NQ dev?
{ "avatar_url": "https://avatars.githubusercontent.com/u/1563902?v=4", "events_url": "https://api.github.com/users/tholor/events{/privacy}", "followers_url": "https://api.github.com/users/tholor/followers", "following_url": "https://api.github.com/users/tholor/following{/other_user}", "gists_url": "https://ap...
[]
closed
false
null
[]
null
3
2020-07-18 10:28:23+00:00
2022-02-11 09:50:21+00:00
2022-02-11 09:50:21+00:00
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
Maybe I missed that in the docs, but is there a way to only download the dev set of natural questions (~1 GB)? As we want to benchmark QA models on different datasets, I would like to avoid downloading the 41GB of training data. I tried ``` dataset = nlp.load_dataset('natural_questions', split="validation", bea...
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/413/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/413/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/412
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/412/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/412/comments
https://api.github.com/repos/huggingface/datasets/issues/412/events
https://github.com/huggingface/datasets/issues/412
660,047,139
MDU6SXNzdWU2NjAwNDcxMzk=
412
Unable to load XTREME dataset from disk
{ "avatar_url": "https://avatars.githubusercontent.com/u/26859204?v=4", "events_url": "https://api.github.com/users/lewtun/events{/privacy}", "followers_url": "https://api.github.com/users/lewtun/followers", "following_url": "https://api.github.com/users/lewtun/following{/other_user}", "gists_url": "https://a...
[]
closed
false
null
[]
null
3
2020-07-18 09:55:00+00:00
2020-07-21 08:15:44+00:00
2020-07-21 08:15:44+00:00
MEMBER
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
Hi 🤗 team! ## Description of the problem Following the [docs](https://huggingface.co/nlp/loading_datasets.html?highlight=xtreme#manually-downloading-files) I'm trying to load the `PAN-X.fr` dataset from the [XTREME](https://github.com/google-research/xtreme) benchmark. I have manually downloaded the `AmazonPho...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/412/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/412/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/411
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/411/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/411/comments
https://api.github.com/repos/huggingface/datasets/issues/411/events
https://github.com/huggingface/datasets/pull/411
659,393,398
MDExOlB1bGxSZXF1ZXN0NDUxMjQxOTQy
411
Sbf
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
[]
closed
false
null
[]
null
0
2020-07-17 16:19:45+00:00
2020-07-21 09:13:46+00:00
2020-07-21 09:13:45+00:00
CONTRIBUTOR
null
null
null
null
This PR adds the Social Bias Frames Dataset (ACL 2020) . dataset homepage: https://homes.cs.washington.edu/~msap/social-bias-frames/
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/411/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/411/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/411.diff", "html_url": "https://github.com/huggingface/datasets/pull/411", "merged_at": "2020-07-21T09:13:45Z", "patch_url": "https://github.com/huggingface/datasets/pull/411.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/411...
true
https://api.github.com/repos/huggingface/datasets/issues/410
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/410/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/410/comments
https://api.github.com/repos/huggingface/datasets/issues/410/events
https://github.com/huggingface/datasets/pull/410
659,242,871
MDExOlB1bGxSZXF1ZXN0NDUxMTEzMTI3
410
20newsgroup
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
[]
closed
false
null
[]
null
0
2020-07-17 13:07:57+00:00
2020-07-20 07:05:29+00:00
2020-07-20 07:05:28+00:00
CONTRIBUTOR
null
null
null
null
Add 20Newsgroup dataset. #353
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/410/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/410/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/410.diff", "html_url": "https://github.com/huggingface/datasets/pull/410", "merged_at": "2020-07-20T07:05:28Z", "patch_url": "https://github.com/huggingface/datasets/pull/410.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/410...
true
https://api.github.com/repos/huggingface/datasets/issues/409
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/409/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/409/comments
https://api.github.com/repos/huggingface/datasets/issues/409/events
https://github.com/huggingface/datasets/issues/409
659,128,611
MDU6SXNzdWU2NTkxMjg2MTE=
409
train_test_split error: 'dict' object has no attribute 'deepcopy'
{ "avatar_url": "https://avatars.githubusercontent.com/u/20516801?v=4", "events_url": "https://api.github.com/users/morganmcg1/events{/privacy}", "followers_url": "https://api.github.com/users/morganmcg1/followers", "following_url": "https://api.github.com/users/morganmcg1/following{/other_user}", "gists_url"...
[]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[ { "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists...
null
2
2020-07-17 10:36:28+00:00
2020-07-21 14:34:52+00:00
2020-07-21 14:34:52+00:00
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
`train_test_split` is giving me an error when I try and call it: `'dict' object has no attribute 'deepcopy'` ## To reproduce ``` dataset = load_dataset('glue', 'mrpc', split='train') dataset = dataset.train_test_split(test_size=0.2) ``` ## Full Stacktrace ``` -------------------------------------------...
{ "avatar_url": "https://avatars.githubusercontent.com/u/20516801?v=4", "events_url": "https://api.github.com/users/morganmcg1/events{/privacy}", "followers_url": "https://api.github.com/users/morganmcg1/followers", "following_url": "https://api.github.com/users/morganmcg1/following{/other_user}", "gists_url"...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/409/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/409/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/408
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/408/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/408/comments
https://api.github.com/repos/huggingface/datasets/issues/408/events
https://github.com/huggingface/datasets/pull/408
659,064,144
MDExOlB1bGxSZXF1ZXN0NDUwOTU1MTE0
408
Add tests datasets gcp
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
0
2020-07-17 09:23:27+00:00
2020-07-17 09:26:57+00:00
2020-07-17 09:26:56+00:00
MEMBER
null
null
null
null
Some datasets are available on our google cloud storage in arrow format, so that the users don't need to process the data. These tests make sure that they're always available. It also makes sure that their scripts are in sync between S3 and the repo. This should avoid future issues like #407
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/408/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/408/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/408.diff", "html_url": "https://github.com/huggingface/datasets/pull/408", "merged_at": "2020-07-17T09:26:56Z", "patch_url": "https://github.com/huggingface/datasets/pull/408.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/408...
true
https://api.github.com/repos/huggingface/datasets/issues/407
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/407/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/407/comments
https://api.github.com/repos/huggingface/datasets/issues/407/events
https://github.com/huggingface/datasets/issues/407
658,672,736
MDU6SXNzdWU2NTg2NzI3MzY=
407
MissingBeamOptions for Wikipedia 20200501.en
{ "avatar_url": "https://avatars.githubusercontent.com/u/7490438?v=4", "events_url": "https://api.github.com/users/mitchellgordon95/events{/privacy}", "followers_url": "https://api.github.com/users/mitchellgordon95/followers", "following_url": "https://api.github.com/users/mitchellgordon95/following{/other_user...
[]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[ { "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists...
null
4
2020-07-16 23:48:03+00:00
2021-01-12 11:41:16+00:00
2020-07-17 14:24:28+00:00
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
There may or may not be a regression for the pre-processed Wikipedia dataset. This was working fine 10 commits ago (without having Apache Beam available): ``` nlp.load_dataset('wikipedia', "20200501.en", split='train') ``` And now, having pulled master, I get: ``` Downloading and preparing dataset wikipedia...
{ "avatar_url": "https://avatars.githubusercontent.com/u/7490438?v=4", "events_url": "https://api.github.com/users/mitchellgordon95/events{/privacy}", "followers_url": "https://api.github.com/users/mitchellgordon95/followers", "following_url": "https://api.github.com/users/mitchellgordon95/following{/other_user...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/407/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/407/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/406
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/406/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/406/comments
https://api.github.com/repos/huggingface/datasets/issues/406/events
https://github.com/huggingface/datasets/issues/406
658,581,764
MDU6SXNzdWU2NTg1ODE3NjQ=
406
Faster Shuffling?
{ "avatar_url": "https://avatars.githubusercontent.com/u/7490438?v=4", "events_url": "https://api.github.com/users/mitchellgordon95/events{/privacy}", "followers_url": "https://api.github.com/users/mitchellgordon95/followers", "following_url": "https://api.github.com/users/mitchellgordon95/following{/other_user...
[]
closed
false
null
[]
null
7
2020-07-16 21:21:53+00:00
2023-08-16 09:52:39+00:00
2020-09-07 14:45:25+00:00
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
Consider shuffling bookcorpus: ``` dataset = nlp.load_dataset('bookcorpus', split='train') dataset.shuffle() ``` According to tqdm, this will take around 2.5 hours on my machine to complete (even with the faster version of select from #405). I've also tried with `keep_in_memory=True` and `writer_batch_size=1000`...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/406/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/406/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/405
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/405/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/405/comments
https://api.github.com/repos/huggingface/datasets/issues/405/events
https://github.com/huggingface/datasets/pull/405
658,580,192
MDExOlB1bGxSZXF1ZXN0NDUwNTI1MTc3
405
Make select() faster by batching reads
{ "avatar_url": "https://avatars.githubusercontent.com/u/7490438?v=4", "events_url": "https://api.github.com/users/mitchellgordon95/events{/privacy}", "followers_url": "https://api.github.com/users/mitchellgordon95/followers", "following_url": "https://api.github.com/users/mitchellgordon95/following{/other_user...
[]
closed
false
null
[]
null
0
2020-07-16 21:19:45+00:00
2020-07-17 17:05:44+00:00
2020-07-17 16:51:26+00:00
CONTRIBUTOR
null
null
null
null
Here's a benchmark: ``` dataset = nlp.load_dataset('bookcorpus', split='train') start = time.time() dataset.select(np.arange(1000), reader_batch_size=1, load_from_cache_file=False) end = time.time() print(f'{end - start}') start = time.time() dataset.select(np.arange(1000), reader_batch_size=1000, load_fr...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/405/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/405/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/405.diff", "html_url": "https://github.com/huggingface/datasets/pull/405", "merged_at": "2020-07-17T16:51:26Z", "patch_url": "https://github.com/huggingface/datasets/pull/405.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/405...
true
https://api.github.com/repos/huggingface/datasets/issues/404
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/404/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/404/comments
https://api.github.com/repos/huggingface/datasets/issues/404/events
https://github.com/huggingface/datasets/pull/404
658,400,987
MDExOlB1bGxSZXF1ZXN0NDUwMzY4Mjg4
404
Add seed in metrics
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
0
2020-07-16 17:27:05+00:00
2020-07-20 10:12:35+00:00
2020-07-20 10:12:34+00:00
MEMBER
null
null
null
null
With #361 we noticed that some metrics were not deterministic. In this PR I allow the user to specify numpy's seed when instantiating a metric with `load_metric`. The seed is set only when `compute` is called, and reset afterwards. Moreover when calling `compute` with the same metric instance (i.e. same experiment...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/404/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/404/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/404.diff", "html_url": "https://github.com/huggingface/datasets/pull/404", "merged_at": "2020-07-20T10:12:34Z", "patch_url": "https://github.com/huggingface/datasets/pull/404.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/404...
true
https://api.github.com/repos/huggingface/datasets/issues/403
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/403/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/403/comments
https://api.github.com/repos/huggingface/datasets/issues/403/events
https://github.com/huggingface/datasets/pull/403
658,325,756
MDExOlB1bGxSZXF1ZXN0NDUwMzAzNjI2
403
return python objects instead of arrays by default
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
0
2020-07-16 15:51:52+00:00
2020-07-17 11:37:01+00:00
2020-07-17 11:37:00+00:00
MEMBER
null
null
null
null
We were using to_pandas() to convert from arrow types, however it returns numpy arrays instead of python lists. I fixed it by using to_pydict/to_pylist instead. Fix #387 It was mentioned in https://github.com/huggingface/transformers/issues/5729
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/403/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/403/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/403.diff", "html_url": "https://github.com/huggingface/datasets/pull/403", "merged_at": "2020-07-17T11:37:00Z", "patch_url": "https://github.com/huggingface/datasets/pull/403.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/403...
true
https://api.github.com/repos/huggingface/datasets/issues/402
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/402/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/402/comments
https://api.github.com/repos/huggingface/datasets/issues/402/events
https://github.com/huggingface/datasets/pull/402
658,001,288
MDExOlB1bGxSZXF1ZXN0NDUwMDI2NTE0
402
Search qa
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
[]
closed
false
null
[]
null
0
2020-07-16 09:00:10+00:00
2020-07-16 14:27:00+00:00
2020-07-16 14:26:59+00:00
CONTRIBUTOR
null
null
null
null
add SearchQA dataset #336
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/402/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/402/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/402.diff", "html_url": "https://github.com/huggingface/datasets/pull/402", "merged_at": "2020-07-16T14:26:59Z", "patch_url": "https://github.com/huggingface/datasets/pull/402.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/402...
true
https://api.github.com/repos/huggingface/datasets/issues/401
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/401/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/401/comments
https://api.github.com/repos/huggingface/datasets/issues/401/events
https://github.com/huggingface/datasets/pull/401
657,996,252
MDExOlB1bGxSZXF1ZXN0NDUwMDIyNTc0
401
add web_questions
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
[]
closed
false
null
[]
null
3
2020-07-16 08:54:59+00:00
2020-08-06 06:16:20+00:00
2020-08-06 06:16:19+00:00
CONTRIBUTOR
null
null
null
null
add Web Question dataset #336 Maybe @patrickvonplaten you can help with the dummy_data structure? it still broken
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/401/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/401/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/401.diff", "html_url": "https://github.com/huggingface/datasets/pull/401", "merged_at": "2020-08-06T06:16:19Z", "patch_url": "https://github.com/huggingface/datasets/pull/401.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/401...
true
https://api.github.com/repos/huggingface/datasets/issues/400
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/400/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/400/comments
https://api.github.com/repos/huggingface/datasets/issues/400/events
https://github.com/huggingface/datasets/pull/400
657,975,600
MDExOlB1bGxSZXF1ZXN0NDUwMDA1MDU5
400
Web questions
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
[]
closed
false
null
[]
null
0
2020-07-16 08:28:29+00:00
2020-07-16 08:50:51+00:00
2020-07-16 08:42:54+00:00
CONTRIBUTOR
null
null
null
null
add the WebQuestion dataset #336
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/400/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/400/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/400.diff", "html_url": "https://github.com/huggingface/datasets/pull/400", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/400.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/400" }
true
https://api.github.com/repos/huggingface/datasets/issues/399
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/399/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/399/comments
https://api.github.com/repos/huggingface/datasets/issues/399/events
https://github.com/huggingface/datasets/pull/399
657,841,433
MDExOlB1bGxSZXF1ZXN0NDQ5ODkxNTEy
399
Spelling mistake
{ "avatar_url": "https://avatars.githubusercontent.com/u/9410067?v=4", "events_url": "https://api.github.com/users/BlancRay/events{/privacy}", "followers_url": "https://api.github.com/users/BlancRay/followers", "following_url": "https://api.github.com/users/BlancRay/following{/other_user}", "gists_url": "http...
[]
closed
false
null
[]
null
1
2020-07-16 04:37:58+00:00
2020-07-16 06:49:48+00:00
2020-07-16 06:49:37+00:00
CONTRIBUTOR
null
null
null
null
In "Formatting the dataset" part, "The two toehr modifications..." should be "The two other modifications..." ,the word "other" wrong spelled as "toehr".
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/399/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/399/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/399.diff", "html_url": "https://github.com/huggingface/datasets/pull/399", "merged_at": "2020-07-16T06:49:37Z", "patch_url": "https://github.com/huggingface/datasets/pull/399.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/399...
true
https://api.github.com/repos/huggingface/datasets/issues/398
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/398/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/398/comments
https://api.github.com/repos/huggingface/datasets/issues/398/events
https://github.com/huggingface/datasets/pull/398
657,511,962
MDExOlB1bGxSZXF1ZXN0NDQ5NjE1OTk1
398
Add inline links
{ "avatar_url": "https://avatars.githubusercontent.com/u/13381361?v=4", "events_url": "https://api.github.com/users/bharatr21/events{/privacy}", "followers_url": "https://api.github.com/users/bharatr21/followers", "following_url": "https://api.github.com/users/bharatr21/following{/other_user}", "gists_url": "...
[]
closed
false
null
[]
null
2
2020-07-15 17:04:04+00:00
2020-07-22 10:14:22+00:00
2020-07-22 10:14:22+00:00
CONTRIBUTOR
null
null
null
null
Add inline links to `Contributing.md`
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/398/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/398/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/398.diff", "html_url": "https://github.com/huggingface/datasets/pull/398", "merged_at": "2020-07-22T10:14:22Z", "patch_url": "https://github.com/huggingface/datasets/pull/398.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/398...
true
https://api.github.com/repos/huggingface/datasets/issues/397
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/397/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/397/comments
https://api.github.com/repos/huggingface/datasets/issues/397/events
https://github.com/huggingface/datasets/pull/397
657,510,856
MDExOlB1bGxSZXF1ZXN0NDQ5NjE1MDA4
397
Add contiguous sharding
{ "avatar_url": "https://avatars.githubusercontent.com/u/4564897?v=4", "events_url": "https://api.github.com/users/jarednielsen/events{/privacy}", "followers_url": "https://api.github.com/users/jarednielsen/followers", "following_url": "https://api.github.com/users/jarednielsen/following{/other_user}", "gists...
[]
closed
false
null
[]
null
0
2020-07-15 17:02:58+00:00
2020-07-17 16:59:31+00:00
2020-07-17 16:59:31+00:00
CONTRIBUTOR
null
null
null
null
This makes dset.shard() play nice with nlp.concatenate_datasets(). When I originally wrote the shard() method, I was thinking about a distributed training scenario, but https://github.com/huggingface/nlp/pull/389 also uses it for splitting the dataset for distributed preprocessing. Usage: ``` nlp.concatenate_datas...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/397/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/397/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/397.diff", "html_url": "https://github.com/huggingface/datasets/pull/397", "merged_at": "2020-07-17T16:59:30Z", "patch_url": "https://github.com/huggingface/datasets/pull/397.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/397...
true
https://api.github.com/repos/huggingface/datasets/issues/396
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/396/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/396/comments
https://api.github.com/repos/huggingface/datasets/issues/396/events
https://github.com/huggingface/datasets/pull/396
657,477,952
MDExOlB1bGxSZXF1ZXN0NDQ5NTg3MDQ4
396
Fix memory issue when doing select
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
0
2020-07-15 16:15:04+00:00
2020-07-16 08:07:32+00:00
2020-07-16 08:07:31+00:00
MEMBER
null
null
null
null
We were passing the `nlp.Dataset` object to get the hash for the new dataset's file name. Fix #395
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/396/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/396/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/396.diff", "html_url": "https://github.com/huggingface/datasets/pull/396", "merged_at": "2020-07-16T08:07:30Z", "patch_url": "https://github.com/huggingface/datasets/pull/396.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/396...
true
https://api.github.com/repos/huggingface/datasets/issues/395
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/395/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/395/comments
https://api.github.com/repos/huggingface/datasets/issues/395/events
https://github.com/huggingface/datasets/issues/395
657,454,983
MDU6SXNzdWU2NTc0NTQ5ODM=
395
Memory issue when doing select
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[ { "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists...
null
1
2020-07-15 15:43:38+00:00
2020-07-16 08:07:31+00:00
2020-07-16 08:07:31+00:00
MEMBER
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
As noticed in #389, the following code loads the entire wikipedia in memory. ```python import nlp w = nlp.load_dataset("wikipedia", "20200501.en", split="train") w.select([0]) ``` This is caused by [this line](https://github.com/huggingface/nlp/blob/master/src/nlp/arrow_dataset.py#L626) for some reason, that ...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/395/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/395/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/394
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/394/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/394/comments
https://api.github.com/repos/huggingface/datasets/issues/394/events
https://github.com/huggingface/datasets/pull/394
657,425,548
MDExOlB1bGxSZXF1ZXN0NDQ5NTQzNTE0
394
Remove remaining nested dict
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
[]
closed
false
null
[]
null
0
2020-07-15 15:05:52+00:00
2020-07-16 07:39:52+00:00
2020-07-16 07:39:51+00:00
CONTRIBUTOR
null
null
null
null
This PR deletes the remaining unnecessary nested dict #378
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/394/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/394/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/394.diff", "html_url": "https://github.com/huggingface/datasets/pull/394", "merged_at": "2020-07-16T07:39:51Z", "patch_url": "https://github.com/huggingface/datasets/pull/394.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/394...
true
https://api.github.com/repos/huggingface/datasets/issues/393
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/393/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/393/comments
https://api.github.com/repos/huggingface/datasets/issues/393/events
https://github.com/huggingface/datasets/pull/393
657,330,911
MDExOlB1bGxSZXF1ZXN0NDQ5NDY1MTAz
393
Fix extracted files directory for the DownloadManager
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
0
2020-07-15 12:59:55+00:00
2020-07-17 17:02:16+00:00
2020-07-17 17:02:14+00:00
MEMBER
null
null
null
null
The cache dir was often cluttered by extracted files because of the download manager. For downloaded files, we are using the `downloads` directory to make things easier to navigate, but extracted files were still placed at the root of the cache directory. To fix that I changed the directory for extracted files to ca...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/393/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/393/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/393.diff", "html_url": "https://github.com/huggingface/datasets/pull/393", "merged_at": "2020-07-17T17:02:14Z", "patch_url": "https://github.com/huggingface/datasets/pull/393.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/393...
true