url
stringlengths
58
61
repository_url
stringclasses
1 value
labels_url
stringlengths
72
75
comments_url
stringlengths
67
70
events_url
stringlengths
65
68
html_url
stringlengths
46
51
id
int64
599M
1B
node_id
stringlengths
18
32
number
int64
1
2.96k
title
stringlengths
1
268
user
dict
labels
list
state
stringclasses
2 values
locked
bool
1 class
assignee
dict
assignees
list
milestone
dict
comments
list
created_at
int64
1,587B
1,632B
updated_at
int64
1,587B
1,632B
closed_at
int64
1,587B
1,632B
author_association
stringclasses
4 values
active_lock_reason
null
pull_request
dict
body
stringlengths
0
228k
timeline_url
stringlengths
67
70
performed_via_github_app
null
is_pull_request
bool
2 classes
https://api.github.com/repos/huggingface/datasets/issues/824
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/824/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/824/comments
https://api.github.com/repos/huggingface/datasets/issues/824/events
https://github.com/huggingface/datasets/issues/824
739,896,526
MDU6SXNzdWU3Mzk4OTY1MjY=
824
Discussion using datasets in offline mode
{ "login": "mandubian", "id": 77193, "node_id": "MDQ6VXNlcjc3MTkz", "avatar_url": "https://avatars.githubusercontent.com/u/77193?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mandubian", "html_url": "https://github.com/mandubian", "followers_url": "https://api.github.com/users/mandubian/...
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" }, { "id": 2067400324, "node_id": "MDU6...
open
false
null
[]
null
[ "No comments ?", "I think it would be very cool. I'm currently working on a cluster from Compute Canada, and I have internet access only when I'm not in the nodes where I run the scripts. So I was expecting to be able to use the wmt14 dataset until I realized I needed internet connection even if I downloaded the ...
1,605,013,851,000
1,611,151,504,000
null
NONE
null
null
`datasets.load_dataset("csv", ...)` breaks if you have no connection (There is already this issue https://github.com/huggingface/datasets/issues/761 about it). It seems to be the same for metrics too. I create this ticket to discuss a bit and gather what you have in mind or other propositions. Here are some point...
https://api.github.com/repos/huggingface/datasets/issues/824/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/823
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/823/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/823/comments
https://api.github.com/repos/huggingface/datasets/issues/823/events
https://github.com/huggingface/datasets/issues/823
739,815,763
MDU6SXNzdWU3Mzk4MTU3NjM=
823
how processing in batch works in datasets
{ "login": "rabeehkarimimahabadi", "id": 73364383, "node_id": "MDQ6VXNlcjczMzY0Mzgz", "avatar_url": "https://avatars.githubusercontent.com/u/73364383?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rabeehkarimimahabadi", "html_url": "https://github.com/rabeehkarimimahabadi", "followers_url...
[ { "id": 2067376369, "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request", "name": "dataset request", "color": "e99695", "default": false, "description": "Requesting to add a new dataset" } ]
closed
false
null
[]
null
[ "Hi I don’t think this is a request for a dataset like you labeled it.\r\n\r\nI also think this would be better suited for the forum at https://discuss.huggingface.co. we try to keep the issue for the repo for bug reports and new features/dataset requests and have usage questions discussed on the forum. Thanks.", ...
1,605,006,677,000
1,605,013,870,000
1,605,013,869,000
NONE
null
null
Hi, I need to process my datasets before it is passed to dataloader in batch, here is my codes ``` class AbstractTask(ABC): task_name: str = NotImplemented preprocessor: Callable = NotImplemented split_to_data_split: Mapping[str, str] = NotImplemented tokenizer: Callable = NotImplemented ...
https://api.github.com/repos/huggingface/datasets/issues/823/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/822
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/822/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/822/comments
https://api.github.com/repos/huggingface/datasets/issues/822/events
https://github.com/huggingface/datasets/issues/822
739,579,314
MDU6SXNzdWU3Mzk1NzkzMTQ=
822
datasets freezes
{ "login": "rabeehkarimimahabadi", "id": 73364383, "node_id": "MDQ6VXNlcjczMzY0Mzgz", "avatar_url": "https://avatars.githubusercontent.com/u/73364383?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rabeehkarimimahabadi", "html_url": "https://github.com/rabeehkarimimahabadi", "followers_url...
[ { "id": 2067388877, "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug", "name": "dataset bug", "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library" } ]
open
false
null
[]
null
[ "Pytorch is unable to convert strings to tensors unfortunately.\r\nYou can use `set_format(type=\"torch\")` on columns that can be converted to tensors, such as token ids.\r\n\r\nThis makes me think that we should probably raise an error or at least a warning when one tries to create pytorch tensors out of text col...
1,604,985,019,000
1,605,223,383,000
null
NONE
null
null
Hi, I want to load these two datasets and convert them to Dataset format in torch and the code freezes for me, could you have a look please? thanks dataset1 = load_dataset("squad", split="train[:10]") dataset1 = dataset1.set_format(type='torch', columns=['context', 'answers', 'question']) dataset2 = load_datase...
https://api.github.com/repos/huggingface/datasets/issues/822/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/821
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/821/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/821/comments
https://api.github.com/repos/huggingface/datasets/issues/821/events
https://github.com/huggingface/datasets/issues/821
739,506,859
MDU6SXNzdWU3Mzk1MDY4NTk=
821
`kor_nli` dataset doesn't being loaded properly
{ "login": "sackoh", "id": 30492059, "node_id": "MDQ6VXNlcjMwNDkyMDU5", "avatar_url": "https://avatars.githubusercontent.com/u/30492059?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sackoh", "html_url": "https://github.com/sackoh", "followers_url": "https://api.github.com/users/sackoh/fo...
[]
closed
false
null
[]
null
[]
1,604,973,852,000
1,605,535,152,000
1,605,535,152,000
NONE
null
null
There are two issues from `kor_nli` dataset 1. csv.DictReader failed to split features by tab - Should not exist `None` value in label feature, but there it is. ```python kor_nli_train['train'].unique('gold_label') # ['neutral', 'entailment', 'contradiction', None] ``` -...
https://api.github.com/repos/huggingface/datasets/issues/821/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/820
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/820/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/820/comments
https://api.github.com/repos/huggingface/datasets/issues/820/events
https://github.com/huggingface/datasets/pull/820
739,387,617
MDExOlB1bGxSZXF1ZXN0NTE4MDYwMjQ0
820
Update quail dataset to v1.3
{ "login": "ngdodd", "id": 4889636, "node_id": "MDQ6VXNlcjQ4ODk2MzY=", "avatar_url": "https://avatars.githubusercontent.com/u/4889636?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ngdodd", "html_url": "https://github.com/ngdodd", "followers_url": "https://api.github.com/users/ngdodd/foll...
[]
closed
false
null
[]
null
[]
1,604,958,566,000
1,604,999,195,000
1,604,999,195,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/820", "html_url": "https://github.com/huggingface/datasets/pull/820", "diff_url": "https://github.com/huggingface/datasets/pull/820.diff", "patch_url": "https://github.com/huggingface/datasets/pull/820.patch" }
Updated quail to most recent version, to address the problem originally discussed [here](https://github.com/huggingface/datasets/issues/806).
https://api.github.com/repos/huggingface/datasets/issues/820/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/819
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/819/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/819/comments
https://api.github.com/repos/huggingface/datasets/issues/819/events
https://github.com/huggingface/datasets/pull/819
739,250,624
MDExOlB1bGxSZXF1ZXN0NTE3OTQ2MjYy
819
Make save function use deterministic global vars order
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[ "Sorry, asking for help here, but the dill thread stop around 2013. Is it possible to use dill deterministically? I tried to monkeypatch the solution presented here into dill, but I suppose it requires forking their project.", "Hi ! What we did was to subclass `dill`'s Pickler to fix the non-deterministic behavio...
1,604,945,523,000
1,605,108,052,000
1,605,108,051,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/819", "html_url": "https://github.com/huggingface/datasets/pull/819", "diff_url": "https://github.com/huggingface/datasets/pull/819.diff", "patch_url": "https://github.com/huggingface/datasets/pull/819.patch" }
The `dumps` function need to be deterministic for the caching mechanism. However in #816 I noticed that one of dill's method to recursively check the globals of a function may return the globals in different orders each time it's used. To fix that I sort the globals by key in the `globs` dictionary. I had to add a re...
https://api.github.com/repos/huggingface/datasets/issues/819/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/818
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/818/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/818/comments
https://api.github.com/repos/huggingface/datasets/issues/818/events
https://github.com/huggingface/datasets/pull/818
739,173,861
MDExOlB1bGxSZXF1ZXN0NTE3ODgzMzk0
818
Fix type hints pickling in python 3.6
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,604,939,267,000
1,604,999,223,000
1,604,999,222,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/818", "html_url": "https://github.com/huggingface/datasets/pull/818", "diff_url": "https://github.com/huggingface/datasets/pull/818.diff", "patch_url": "https://github.com/huggingface/datasets/pull/818.patch" }
Type hints can't be properly pickled in python 3.6. This was causing errors the `run_mlm.py` script from `transformers` with python 3.6 However Cloupickle proposed a [fix](https://github.com/cloudpipe/cloudpickle/pull/318/files) to make it work anyway. The idea is just to implement the pickling/unpickling of parame...
https://api.github.com/repos/huggingface/datasets/issues/818/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/817
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/817/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/817/comments
https://api.github.com/repos/huggingface/datasets/issues/817/events
https://github.com/huggingface/datasets/issues/817
739,145,369
MDU6SXNzdWU3MzkxNDUzNjk=
817
Add MRQA dataset
{ "login": "VictorSanh", "id": 16107619, "node_id": "MDQ6VXNlcjE2MTA3NjE5", "avatar_url": "https://avatars.githubusercontent.com/u/16107619?v=4", "gravatar_id": "", "url": "https://api.github.com/users/VictorSanh", "html_url": "https://github.com/VictorSanh", "followers_url": "https://api.github.com/use...
[ { "id": 2067376369, "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request", "name": "dataset request", "color": "e99695", "default": false, "description": "Requesting to add a new dataset" } ]
closed
false
null
[]
null
[ "Done! cf #1117 and #1022" ]
1,604,937,139,000
1,607,096,682,000
1,607,096,681,000
MEMBER
null
null
## Adding a Dataset - **Name:** MRQA - **Description:** Collection of different (subsets of) QA datasets all converted to the same format to evaluate out-of-domain generalization (the datasets come from different domains, distributions, etc.). Some datasets are used for training and others are used for evaluation. Th...
https://api.github.com/repos/huggingface/datasets/issues/817/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/816
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/816/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/816/comments
https://api.github.com/repos/huggingface/datasets/issues/816/events
https://github.com/huggingface/datasets/issues/816
739,102,686
MDU6SXNzdWU3MzkxMDI2ODY=
816
[Caching] Dill globalvars() output order is not deterministic and can cause cache issues.
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[ "To show the issue:\r\n```\r\npython -c \"from datasets.fingerprint import Hasher; a=[]; func = lambda : len(a); print(Hasher.hash(func))\"\r\n```\r\ndoesn't always return the same ouput since `globs` is a dictionary with \"a\" and \"len\" as keys but sometimes not in the same order" ]
1,604,934,080,000
1,605,108,050,000
1,605,108,050,000
MEMBER
null
null
Dill uses `dill.detect.globalvars` to get the globals used by a function in a recursive dump. `globalvars` returns a dictionary of all the globals that a dumped function needs. However the order of the keys in this dict is not deterministic and can cause caching issues. To fix that one could register an implementati...
https://api.github.com/repos/huggingface/datasets/issues/816/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/815
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/815/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/815/comments
https://api.github.com/repos/huggingface/datasets/issues/815/events
https://github.com/huggingface/datasets/issues/815
738,842,092
MDU6SXNzdWU3Mzg4NDIwOTI=
815
Is dataset iterative or not?
{ "login": "rabeehkarimimahabadi", "id": 73364383, "node_id": "MDQ6VXNlcjczMzY0Mzgz", "avatar_url": "https://avatars.githubusercontent.com/u/73364383?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rabeehkarimimahabadi", "html_url": "https://github.com/rabeehkarimimahabadi", "followers_url...
[ { "id": 2067376369, "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request", "name": "dataset request", "color": "e99695", "default": false, "description": "Requesting to add a new dataset" } ]
closed
false
null
[]
null
[ "Hello !\r\nCould you give more details ?\r\n\r\nIf you mean iter through one dataset then yes, `Dataset` object does implement the `__iter__` method so you can use \r\n```python\r\nfor example in dataset:\r\n # do something\r\n```\r\n\r\nIf you want to iter through several datasets you can first concatenate the...
1,604,913,108,000
1,605,005,403,000
1,605,005,403,000
NONE
null
null
Hi I want to use your library for large-scale training, I am not sure if this is implemented as iterative datasets or not? could you provide me with example how I can use datasets as iterative datasets? thanks
https://api.github.com/repos/huggingface/datasets/issues/815/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/814
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/814/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/814/comments
https://api.github.com/repos/huggingface/datasets/issues/814/events
https://github.com/huggingface/datasets/issues/814
738,500,443
MDU6SXNzdWU3Mzg1MDA0NDM=
814
Joining multiple datasets
{ "login": "rabeehkarimimahabadi", "id": 73364383, "node_id": "MDQ6VXNlcjczMzY0Mzgz", "avatar_url": "https://avatars.githubusercontent.com/u/73364383?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rabeehkarimimahabadi", "html_url": "https://github.com/rabeehkarimimahabadi", "followers_url...
[ { "id": 2067376369, "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request", "name": "dataset request", "color": "e99695", "default": false, "description": "Requesting to add a new dataset" } ]
closed
false
null
[]
null
[ "found a solution here https://discuss.pytorch.org/t/train-simultaneously-on-two-datasets/649/35, closed for now, thanks " ]
1,604,852,370,000
1,604,864,328,000
1,604,864,328,000
NONE
null
null
Hi I have multiple iterative datasets from your library with different size and I want to join them in a way that each datasets is sampled equally, so smaller datasets more, larger one less, could you tell me how to implement this in pytorch? thanks
https://api.github.com/repos/huggingface/datasets/issues/814/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/813
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/813/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/813/comments
https://api.github.com/repos/huggingface/datasets/issues/813/events
https://github.com/huggingface/datasets/issues/813
738,489,852
MDU6SXNzdWU3Mzg0ODk4NTI=
813
How to implement DistributedSampler with datasets
{ "login": "rabeehkarimimahabadi", "id": 73364383, "node_id": "MDQ6VXNlcjczMzY0Mzgz", "avatar_url": "https://avatars.githubusercontent.com/u/73364383?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rabeehkarimimahabadi", "html_url": "https://github.com/rabeehkarimimahabadi", "followers_url...
[ { "id": 2067376369, "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request", "name": "dataset request", "color": "e99695", "default": false, "description": "Requesting to add a new dataset" } ]
open
false
null
[]
null
[ "Hi Apparently I need to shard the data and give one host a chunk, could you provide me please with examples on how to do it? I want to use it jointly with finetune_trainer.py in huggingface repo seq2seq examples. thanks. ", "Hey @rabeehkarimimahabadi I'm actually looking for the same feature. Did you manage to g...
1,604,849,231,000
1,611,150,147,000
null
NONE
null
null
Hi, I am using your datasets to define my dataloaders, and I am training finetune_trainer.py in huggingface repo on them. I need a distributedSampler to be able to train the models on TPUs being able to distribute the load across the TPU cores. Could you tell me how I can implement the distribued sampler when using d...
https://api.github.com/repos/huggingface/datasets/issues/813/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/812
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/812/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/812/comments
https://api.github.com/repos/huggingface/datasets/issues/812/events
https://github.com/huggingface/datasets/issues/812
738,340,217
MDU6SXNzdWU3MzgzNDAyMTc=
812
Too much logging
{ "login": "dspoka", "id": 6183050, "node_id": "MDQ6VXNlcjYxODMwNTA=", "avatar_url": "https://avatars.githubusercontent.com/u/6183050?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dspoka", "html_url": "https://github.com/dspoka", "followers_url": "https://api.github.com/users/dspoka/foll...
[]
closed
false
null
[]
null
[ "Hi ! Thanks for reporting :) \r\nI agree these one should be hidden when the logging level is warning, we'll fix that", "+1, the amount of logging is excessive.\r\n\r\nMost of it indeed comes from `filelock.py`, though there are occasionally messages from other sources too. Below is an example (all of these mess...
1,604,793,390,000
1,611,671,494,000
1,605,546,402,000
NONE
null
null
I'm doing this in the beginning of my script: from datasets.utils import logging as datasets_logging datasets_logging.set_verbosity_warning() but I'm still getting these logs: [2020-11-07 15:45:41,908][filelock][INFO] - Lock 139958278886176 acquired on /home/username/.cache/huggingface/datasets/cfe20ffaa80ef1...
https://api.github.com/repos/huggingface/datasets/issues/812/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/811
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/811/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/811/comments
https://api.github.com/repos/huggingface/datasets/issues/811/events
https://github.com/huggingface/datasets/issues/811
738,280,132
MDU6SXNzdWU3MzgyODAxMzI=
811
nlp viewer error
{ "login": "jc-hou", "id": 30210529, "node_id": "MDQ6VXNlcjMwMjEwNTI5", "avatar_url": "https://avatars.githubusercontent.com/u/30210529?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jc-hou", "html_url": "https://github.com/jc-hou", "followers_url": "https://api.github.com/users/jc-hou/fo...
[ { "id": 2107841032, "node_id": "MDU6TGFiZWwyMTA3ODQxMDMy", "url": "https://api.github.com/repos/huggingface/datasets/labels/nlp-viewer", "name": "nlp-viewer", "color": "94203D", "default": false, "description": "" } ]
open
false
null
[]
null
[ "and also for 'blog_authorship_corpus'\r\nhttps://huggingface.co/nlp/viewer/?dataset=blog_authorship_corpus\r\n![image](https://user-images.githubusercontent.com/30210529/98557329-5c182800-22a4-11eb-9b01-5b910fb8fcd4.png)\r\n", "Is this the problem of my local computer or ??", "Related to:\r\n- #673" ]
1,604,768,938,000
1,605,540,383,000
null
NONE
null
null
Hello, when I select amazon_us_reviews in nlp viewer, it shows error. https://huggingface.co/nlp/viewer/?dataset=amazon_us_reviews ![image](https://user-images.githubusercontent.com/30210529/98447334-4aa81200-2124-11eb-9dca-82c3ab34ccc2.png)
https://api.github.com/repos/huggingface/datasets/issues/811/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/810
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/810/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/810/comments
https://api.github.com/repos/huggingface/datasets/issues/810/events
https://github.com/huggingface/datasets/pull/810
737,878,370
MDExOlB1bGxSZXF1ZXN0NTE2ODQzMzQ3
810
Fix seqeval metric
{ "login": "sgugger", "id": 35901082, "node_id": "MDQ6VXNlcjM1OTAxMDgy", "avatar_url": "https://avatars.githubusercontent.com/u/35901082?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sgugger", "html_url": "https://github.com/sgugger", "followers_url": "https://api.github.com/users/sgugge...
[]
closed
false
null
[]
null
[]
1,604,679,103,000
1,604,930,669,000
1,604,930,668,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/810", "html_url": "https://github.com/huggingface/datasets/pull/810", "diff_url": "https://github.com/huggingface/datasets/pull/810.diff", "patch_url": "https://github.com/huggingface/datasets/pull/810.patch" }
The current seqeval metric returns the following error when computed: ``` ~/.cache/huggingface/modules/datasets_modules/metrics/seqeval/78a944d83252b5a16c9a2e49f057f4c6e02f18cc03349257025a8c9aea6524d8/seqeval.py in _compute(self, predictions, references, suffix) 102 scores = {} 103 for type_...
https://api.github.com/repos/huggingface/datasets/issues/810/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/809
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/809/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/809/comments
https://api.github.com/repos/huggingface/datasets/issues/809/events
https://github.com/huggingface/datasets/issues/809
737,832,701
MDU6SXNzdWU3Mzc4MzI3MDE=
809
Add Google Taskmaster dataset
{ "login": "yjernite", "id": 10469459, "node_id": "MDQ6VXNlcjEwNDY5NDU5", "avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4", "gravatar_id": "", "url": "https://api.github.com/users/yjernite", "html_url": "https://github.com/yjernite", "followers_url": "https://api.github.com/users/yje...
[ { "id": 2067376369, "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request", "name": "dataset request", "color": "e99695", "default": false, "description": "Requesting to add a new dataset" } ]
closed
false
null
[]
null
[ "Hey @yjernite. Was going to start working on this but found taskmaster 1,2 & 3 in the datasets library already so think this can be closed now?", "You are absolutely right :) \r\n\r\nClosed by https://github.com/huggingface/datasets/pull/1193 https://github.com/huggingface/datasets/pull/1197 https://github.com/h...
1,604,675,441,000
1,618,924,166,000
1,618,924,166,000
MEMBER
null
null
## Adding a Dataset - **Name:** Taskmaster - **Description:** A large dataset of task-oriented dialogue with annotated goals (55K dialogues covering entertainment and travel reservations) - **Paper:** https://arxiv.org/abs/1909.05358 - **Data:** https://github.com/google-research-datasets/Taskmaster - **Motivation...
https://api.github.com/repos/huggingface/datasets/issues/809/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/808
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/808/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/808/comments
https://api.github.com/repos/huggingface/datasets/issues/808/events
https://github.com/huggingface/datasets/pull/808
737,638,942
MDExOlB1bGxSZXF1ZXN0NTE2NjQ0NDc0
808
dataset(dgs): initial dataset loading script
{ "login": "AmitMY", "id": 5757359, "node_id": "MDQ6VXNlcjU3NTczNTk=", "avatar_url": "https://avatars.githubusercontent.com/u/5757359?v=4", "gravatar_id": "", "url": "https://api.github.com/users/AmitMY", "html_url": "https://github.com/AmitMY", "followers_url": "https://api.github.com/users/AmitMY/foll...
[]
closed
false
null
[]
null
[ "Hi @AmitMY, \r\n\r\nWere you able to figure this out?", "I did not.\r\nWith all the limitations this repo currently has, I had to create a repo of my own using tfds to mitigate them. \r\nhttps://github.com/sign-language-processing/datasets/tree/master/sign_language_datasets/datasets/dgs_corpus\r\n\r\nClosing as ...
1,604,657,683,000
1,616,480,335,000
1,616,480,335,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/808", "html_url": "https://github.com/huggingface/datasets/pull/808", "diff_url": "https://github.com/huggingface/datasets/pull/808.diff", "patch_url": "https://github.com/huggingface/datasets/pull/808.patch" }
When trying to create dummy data I get: > Dataset datasets with config None seems to already open files in the method `_split_generators(...)`. You might consider to instead only open files in the method `_generate_examples(...)` instead. If this is not possible the dummy data has t o be created with less guidance. ...
https://api.github.com/repos/huggingface/datasets/issues/808/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/807
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/807/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/807/comments
https://api.github.com/repos/huggingface/datasets/issues/807/events
https://github.com/huggingface/datasets/issues/807
737,509,954
MDU6SXNzdWU3Mzc1MDk5NTQ=
807
load_dataset for LOCAL CSV files report CONNECTION ERROR
{ "login": "shexuan", "id": 25664170, "node_id": "MDQ6VXNlcjI1NjY0MTcw", "avatar_url": "https://avatars.githubusercontent.com/u/25664170?v=4", "gravatar_id": "", "url": "https://api.github.com/users/shexuan", "html_url": "https://github.com/shexuan", "followers_url": "https://api.github.com/users/shexua...
[]
closed
false
null
[]
null
[ "Hi !\r\nThe url works on my side.\r\n\r\nIs the url working in your navigator ?\r\nAre you connected to internet ? Does your network block access to `raw.githubusercontent.com` ?", "> Hi !\r\n> The url works on my side.\r\n> \r\n> Is the url working in your navigator ?\r\n> Are you connected to internet ? Does y...
1,604,644,384,000
1,610,328,627,000
1,605,331,834,000
NONE
null
null
## load_dataset for LOCAL CSV files report CONNECTION ERROR - **Description:** A local demo csv file: ``` import pandas as pd import numpy as np from datasets import load_dataset import torch import transformers df = pd.DataFrame(np.arange(1200).reshape(300,4)) df.to_csv('test.csv', header=False, index=Fal...
https://api.github.com/repos/huggingface/datasets/issues/807/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/806
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/806/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/806/comments
https://api.github.com/repos/huggingface/datasets/issues/806/events
https://github.com/huggingface/datasets/issues/806
737,215,430
MDU6SXNzdWU3MzcyMTU0MzA=
806
Quail dataset urls are out of date
{ "login": "ngdodd", "id": 4889636, "node_id": "MDQ6VXNlcjQ4ODk2MzY=", "avatar_url": "https://avatars.githubusercontent.com/u/4889636?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ngdodd", "html_url": "https://github.com/ngdodd", "followers_url": "https://api.github.com/users/ngdodd/foll...
[]
closed
false
null
[]
null
[ "Hi ! Thanks for reporting.\r\nWe should fix the urls and use quail 1.3.\r\nIf you want to contribute feel free to fix the urls and open a PR :) ", "Done! PR [https://github.com/huggingface/datasets/pull/820](https://github.com/huggingface/datasets/pull/820)\r\n\r\nUpdated links and also regenerated the metadata ...
1,604,605,219,000
1,605,016,971,000
1,605,016,971,000
CONTRIBUTOR
null
null
<h3>Code</h3> ``` from datasets import load_dataset quail = load_dataset('quail') ``` <h3>Error</h3> ``` FileNotFoundError: Couldn't find file at https://raw.githubusercontent.com/text-machine-lab/quail/master/quail_v1.2/xml/ordered/quail_1.2_train.xml ``` As per [quail v1.3 commit](https://github.co...
https://api.github.com/repos/huggingface/datasets/issues/806/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/805
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/805/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/805/comments
https://api.github.com/repos/huggingface/datasets/issues/805/events
https://github.com/huggingface/datasets/issues/805
737,019,360
MDU6SXNzdWU3MzcwMTkzNjA=
805
On loading a metric from datasets, I get the following error
{ "login": "laibamehnaz", "id": 36405283, "node_id": "MDQ6VXNlcjM2NDA1Mjgz", "avatar_url": "https://avatars.githubusercontent.com/u/36405283?v=4", "gravatar_id": "", "url": "https://api.github.com/users/laibamehnaz", "html_url": "https://github.com/laibamehnaz", "followers_url": "https://api.github.com/...
[]
open
false
null
[]
null
[ "Hi ! We support only pyarrow > 0.17.1 so that we have access to the `PyExtensionType` object.\r\nCould you update pyarrow and try again ?\r\n```\r\npip install --upgrade pyarrow\r\n```" ]
1,604,589,278,000
1,604,913,155,000
null
NONE
null
null
`from datasets import load_metric` `metric = load_metric('bleurt')` Traceback: 210 class _ArrayXDExtensionType(pa.PyExtensionType): 211 212 ndims: int = None AttributeError: module 'pyarrow' has no attribute 'PyExtensionType' Any help will be appreciated. Thank you.
https://api.github.com/repos/huggingface/datasets/issues/805/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/804
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/804/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/804/comments
https://api.github.com/repos/huggingface/datasets/issues/804/events
https://github.com/huggingface/datasets/issues/804
736,858,507
MDU6SXNzdWU3MzY4NTg1MDc=
804
Empty output/answer in TriviaQA test set (both in 'kilt_tasks' and 'trivia_qa')
{ "login": "PaulLerner", "id": 25532159, "node_id": "MDQ6VXNlcjI1NTMyMTU5", "avatar_url": "https://avatars.githubusercontent.com/u/25532159?v=4", "gravatar_id": "", "url": "https://api.github.com/users/PaulLerner", "html_url": "https://github.com/PaulLerner", "followers_url": "https://api.github.com/use...
[]
closed
false
null
[]
null
[ "cc @yjernite is this expected ?", "Yes: TriviaQA has a private test set for the leaderboard [here](https://competitions.codalab.org/competitions/17208)\r\n\r\nFor the KILT training and validation portions, you need to link the examples from the TriviaQA dataset as detailed here:\r\nhttps://github.com/huggingface...
1,604,576,281,000
1,604,931,299,000
1,604,931,298,000
CONTRIBUTOR
null
null
# The issue It's all in the title, it appears to be fine on the train and validation sets. Is there some kind of mapping to do like for the questions (see https://github.com/huggingface/datasets/blob/master/datasets/kilt_tasks/README.md) ? # How to reproduce ```py from datasets import load_dataset kilt_tas...
https://api.github.com/repos/huggingface/datasets/issues/804/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/803
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/803/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/803/comments
https://api.github.com/repos/huggingface/datasets/issues/803/events
https://github.com/huggingface/datasets/pull/803
736,818,917
MDExOlB1bGxSZXF1ZXN0NTE1OTY1ODE2
803
fix: typos in tutorial to map KILT and TriviaQA
{ "login": "PaulLerner", "id": 25532159, "node_id": "MDQ6VXNlcjI1NTMyMTU5", "avatar_url": "https://avatars.githubusercontent.com/u/25532159?v=4", "gravatar_id": "", "url": "https://api.github.com/users/PaulLerner", "html_url": "https://github.com/PaulLerner", "followers_url": "https://api.github.com/use...
[]
closed
false
null
[]
null
[]
1,604,572,920,000
1,604,999,287,000
1,604,999,287,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/803", "html_url": "https://github.com/huggingface/datasets/pull/803", "diff_url": "https://github.com/huggingface/datasets/pull/803.diff", "patch_url": "https://github.com/huggingface/datasets/pull/803.patch" }
https://api.github.com/repos/huggingface/datasets/issues/803/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/802
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/802/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/802/comments
https://api.github.com/repos/huggingface/datasets/issues/802/events
https://github.com/huggingface/datasets/pull/802
736,296,343
MDExOlB1bGxSZXF1ZXN0NTE1NTM1MDI0
802
Add XGlue
{ "login": "patrickvonplaten", "id": 23423619, "node_id": "MDQ6VXNlcjIzNDIzNjE5", "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "gravatar_id": "", "url": "https://api.github.com/users/patrickvonplaten", "html_url": "https://github.com/patrickvonplaten", "followers_url": "https://...
[]
closed
false
null
[]
null
[ "Really cool to add XGlue, this will be a nice addition !\r\n\r\nSplits shouldn't depend on the language. There must be configurations for each language, as we're doing for xnli, xtreme, etc.\r\nFor example for XGlue we'll have these configurations: NER.de, NER.en etc." ]
1,604,510,994,000
1,606,838,308,000
1,606,838,307,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/802", "html_url": "https://github.com/huggingface/datasets/pull/802", "diff_url": "https://github.com/huggingface/datasets/pull/802.diff", "patch_url": "https://github.com/huggingface/datasets/pull/802.patch" }
Dataset is ready to merge. An important feature of this dataset is that for each config the train data is in English, while dev and test data are in multiple languages. Therefore, @lhoestq and I decided offline that we will give the dataset the following API, *e.g.* for ```python load_dataset("xglue", "ner") # wo...
https://api.github.com/repos/huggingface/datasets/issues/802/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/801
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/801/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/801/comments
https://api.github.com/repos/huggingface/datasets/issues/801/events
https://github.com/huggingface/datasets/issues/801
735,790,876
MDU6SXNzdWU3MzU3OTA4NzY=
801
How to join two datasets?
{ "login": "shangw-nvidia", "id": 66387198, "node_id": "MDQ6VXNlcjY2Mzg3MTk4", "avatar_url": "https://avatars.githubusercontent.com/u/66387198?v=4", "gravatar_id": "", "url": "https://api.github.com/users/shangw-nvidia", "html_url": "https://github.com/shangw-nvidia", "followers_url": "https://api.githu...
[]
closed
false
null
[]
null
[ "Hi this is also my question. thanks ", "Hi ! Currently the only way to add new fields to a dataset is by using `.map` and picking items from the other dataset\r\n", "Closing this one. Feel free to re-open if you have other questions about this issue.\r\n\r\nAlso linking another discussion about joining dataset...
1,604,461,991,000
1,608,732,178,000
1,608,732,178,000
NONE
null
null
Hi, I'm wondering if it's possible to join two (preprocessed) datasets with the same number of rows but different labels? I'm currently trying to create paired sentences for BERT from `wikipedia/'20200501.en`, and I couldn't figure out a way to create a paired sentence using `.map()` where the second sentence is...
https://api.github.com/repos/huggingface/datasets/issues/801/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/800
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/800/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/800/comments
https://api.github.com/repos/huggingface/datasets/issues/800/events
https://github.com/huggingface/datasets/pull/800
735,772,775
MDExOlB1bGxSZXF1ZXN0NTE1MTAyMjc3
800
Update loading_metrics.rst
{ "login": "ayushidalmia", "id": 5400513, "node_id": "MDQ6VXNlcjU0MDA1MTM=", "avatar_url": "https://avatars.githubusercontent.com/u/5400513?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ayushidalmia", "html_url": "https://github.com/ayushidalmia", "followers_url": "https://api.github.com...
[]
closed
false
null
[]
null
[]
1,604,458,631,000
1,605,108,512,000
1,605,108,512,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/800", "html_url": "https://github.com/huggingface/datasets/pull/800", "diff_url": "https://github.com/huggingface/datasets/pull/800.diff", "patch_url": "https://github.com/huggingface/datasets/pull/800.patch" }
Minor bug
https://api.github.com/repos/huggingface/datasets/issues/800/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/799
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/799/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/799/comments
https://api.github.com/repos/huggingface/datasets/issues/799/events
https://github.com/huggingface/datasets/pull/799
735,551,165
MDExOlB1bGxSZXF1ZXN0NTE0OTIzNDMx
799
switch amazon reviews class label order
{ "login": "joeddav", "id": 9353833, "node_id": "MDQ6VXNlcjkzNTM4MzM=", "avatar_url": "https://avatars.githubusercontent.com/u/9353833?v=4", "gravatar_id": "", "url": "https://api.github.com/users/joeddav", "html_url": "https://github.com/joeddav", "followers_url": "https://api.github.com/users/joeddav/...
[]
closed
false
null
[]
null
[]
1,604,428,738,000
1,604,429,054,000
1,604,429,050,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/799", "html_url": "https://github.com/huggingface/datasets/pull/799", "diff_url": "https://github.com/huggingface/datasets/pull/799.diff", "patch_url": "https://github.com/huggingface/datasets/pull/799.patch" }
Switches the label order to be more intuitive for amazon reviews, #791.
https://api.github.com/repos/huggingface/datasets/issues/799/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/798
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/798/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/798/comments
https://api.github.com/repos/huggingface/datasets/issues/798/events
https://github.com/huggingface/datasets/issues/798
735,518,805
MDU6SXNzdWU3MzU1MTg4MDU=
798
Cannot load TREC dataset: ConnectionError
{ "login": "kaletap", "id": 25740957, "node_id": "MDQ6VXNlcjI1NzQwOTU3", "avatar_url": "https://avatars.githubusercontent.com/u/25740957?v=4", "gravatar_id": "", "url": "https://api.github.com/users/kaletap", "html_url": "https://github.com/kaletap", "followers_url": "https://api.github.com/users/kaleta...
[ { "id": 2067388877, "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug", "name": "dataset bug", "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library" } ]
open
false
null
[]
null
[ "Hi ! Indeed there's an issue with those links.\r\nWe should probably use the target urls of the redirections instead", "Hi, the same issue here, could you tell me how to download it through datasets? thanks ", "Same issue. ", "Actually it's already fixed on the master branch since #740 \r\nI'll do the 1.1.3 ...
1,604,425,522,000
1,605,521,905,000
null
NONE
null
null
## Problem I cannot load "trec" dataset, it results with ConnectionError as shown below. I've tried on both Google Colab and locally. * `requests.head('http://cogcomp.org/Data/QA/QC/train_5500.label')` returns <Response [302]>. * `requests.head('http://cogcomp.org/Data/QA/QC/train_5500.label', allow_redirects=True...
https://api.github.com/repos/huggingface/datasets/issues/798/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/797
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/797/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/797/comments
https://api.github.com/repos/huggingface/datasets/issues/797/events
https://github.com/huggingface/datasets/issues/797
735,420,332
MDU6SXNzdWU3MzU0MjAzMzI=
797
Token classification labels are strings and we don't have the list of labels
{ "login": "sgugger", "id": 35901082, "node_id": "MDQ6VXNlcjM1OTAxMDgy", "avatar_url": "https://avatars.githubusercontent.com/u/35901082?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sgugger", "html_url": "https://github.com/sgugger", "followers_url": "https://api.github.com/users/sgugge...
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" }, { "id": 2067401494, "node_id": "MDU6...
open
false
null
[]
null
[ "Indeed. Pinging @stefan-it here if he want to give an expert opinion :)", "Related is https://github.com/huggingface/datasets/pull/636", "Should definitely be a ClassLabel 👍 ", "Already done." ]
1,604,417,610,000
1,605,017,231,000
null
MEMBER
null
null
Not sure if this is an issue we want to fix or not, putting it here so it's not forgotten. Right now, in token classification datasets, the labels for NER, POS and the likes are typed as `Sequence` of `strings`, which is wrong in my opinion. These should be `Sequence` of `ClassLabel` or some types that gives easy acces...
https://api.github.com/repos/huggingface/datasets/issues/797/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/796
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/796/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/796/comments
https://api.github.com/repos/huggingface/datasets/issues/796/events
https://github.com/huggingface/datasets/issues/796
735,414,881
MDU6SXNzdWU3MzU0MTQ4ODE=
796
Seq2Seq Metrics QOL: Bleu, Rouge
{ "login": "sshleifer", "id": 6045025, "node_id": "MDQ6VXNlcjYwNDUwMjU=", "avatar_url": "https://avatars.githubusercontent.com/u/6045025?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sshleifer", "html_url": "https://github.com/sshleifer", "followers_url": "https://api.github.com/users/ss...
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
open
false
null
[]
null
null
1,604,417,189,000
1,611,843,228,000
null
MEMBER
null
null
Putting all my QOL issues here, idt I will have time to propose fixes, but I didn't want these to be lost, in case they are useful. I tried using `rouge` and `bleu` for the first time and wrote down everything I didn't immediately understand: + Bleu expects tokenization, can I just kwarg it like sacrebleu? + differ...
https://api.github.com/repos/huggingface/datasets/issues/796/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/795
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/795/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/795/comments
https://api.github.com/repos/huggingface/datasets/issues/795/events
https://github.com/huggingface/datasets/issues/795
735,198,265
MDU6SXNzdWU3MzUxOTgyNjU=
795
Descriptions of raw and processed versions of wikitext are inverted
{ "login": "fraboniface", "id": 16835358, "node_id": "MDQ6VXNlcjE2ODM1MzU4", "avatar_url": "https://avatars.githubusercontent.com/u/16835358?v=4", "gravatar_id": "", "url": "https://api.github.com/users/fraboniface", "html_url": "https://github.com/fraboniface", "followers_url": "https://api.github.com/...
[ { "id": 2067388877, "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug", "name": "dataset bug", "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library" } ]
open
false
null
[]
null
[ "Yes indeed ! Thanks for reporting", "Fixed by:\r\n- #3241" ]
1,604,399,091,000
1,605,017,145,000
null
NONE
null
null
Nothing of importance, but it looks like the descriptions of wikitext-n-v1 and wikitext-n-raw-v1 are inverted for both n=2 and n=103. I just verified by loading them and the `<unk>` tokens are present in the non-raw versions, which confirms that it's a mere inversion of the descriptions and not of the datasets themselv...
https://api.github.com/repos/huggingface/datasets/issues/795/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/794
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/794/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/794/comments
https://api.github.com/repos/huggingface/datasets/issues/794/events
https://github.com/huggingface/datasets/issues/794
735,158,725
MDU6SXNzdWU3MzUxNTg3MjU=
794
self.options cannot be converted to a Python object for pickling
{ "login": "hzqjyyx", "id": 9635713, "node_id": "MDQ6VXNlcjk2MzU3MTM=", "avatar_url": "https://avatars.githubusercontent.com/u/9635713?v=4", "gravatar_id": "", "url": "https://api.github.com/users/hzqjyyx", "html_url": "https://github.com/hzqjyyx", "followers_url": "https://api.github.com/users/hzqjyyx/...
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
[ "Hi ! Thanks for reporting that's a bug on master indeed.\r\nWe'll fix that soon" ]
1,604,395,654,000
1,605,807,338,000
1,605,807,338,000
NONE
null
null
Hi, Currently I am trying to load csv file with customized read_options. And the latest master seems broken if we pass the ReadOptions object. Here is a code snippet ```python from datasets import load_dataset from pyarrow.csv import ReadOptions load_dataset("csv", data_files=["out.csv"], read_options=ReadOpt...
https://api.github.com/repos/huggingface/datasets/issues/794/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/793
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/793/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/793/comments
https://api.github.com/repos/huggingface/datasets/issues/793/events
https://github.com/huggingface/datasets/pull/793
735,105,907
MDExOlB1bGxSZXF1ZXN0NTE0NTU2NzY5
793
[Datasets] fix discofuse links
{ "login": "patrickvonplaten", "id": 23423619, "node_id": "MDQ6VXNlcjIzNDIzNjE5", "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "gravatar_id": "", "url": "https://api.github.com/users/patrickvonplaten", "html_url": "https://github.com/patrickvonplaten", "followers_url": "https://...
[]
closed
false
null
[]
null
[]
1,604,390,625,000
1,604,391,401,000
1,604,391,400,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/793", "html_url": "https://github.com/huggingface/datasets/pull/793", "diff_url": "https://github.com/huggingface/datasets/pull/793.diff", "patch_url": "https://github.com/huggingface/datasets/pull/793.patch" }
The discofuse links were changed: https://github.com/google-research-datasets/discofuse/commit/d27641016eb5b3eb2af03c7415cfbb2cbebe8558. The old links are broken I changed the links and created the new dataset_infos.json. Pinging @thomwolf @lhoestq for notification.
https://api.github.com/repos/huggingface/datasets/issues/793/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/792
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/792/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/792/comments
https://api.github.com/repos/huggingface/datasets/issues/792/events
https://github.com/huggingface/datasets/issues/792
734,693,652
MDU6SXNzdWU3MzQ2OTM2NTI=
792
KILT dataset: empty string in triviaqa input field
{ "login": "PaulLerner", "id": 25532159, "node_id": "MDQ6VXNlcjI1NTMyMTU5", "avatar_url": "https://avatars.githubusercontent.com/u/25532159?v=4", "gravatar_id": "", "url": "https://api.github.com/users/PaulLerner", "html_url": "https://github.com/PaulLerner", "followers_url": "https://api.github.com/use...
[]
closed
false
null
[]
null
[ "Just found out about https://github.com/huggingface/datasets/blob/master/datasets/kilt_tasks/README.md\r\n(Not very clear in https://huggingface.co/datasets/kilt_tasks links to http://github.com/huggingface/datasets/datasets/kilt_tasks/README.md which is dead, closing the issue though :))" ]
1,604,338,434,000
1,604,572,499,000
1,604,572,499,000
CONTRIBUTOR
null
null
# What happened Both train and test splits of the triviaqa dataset (part of the KILT benchmark) seem to have empty string in their input field (unlike the natural questions dataset, part of the same benchmark) # Versions KILT version is `1.0.0` `datasets` version is `1.1.2` [more here](https://gist.github.com/Pa...
https://api.github.com/repos/huggingface/datasets/issues/792/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/791
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/791/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/791/comments
https://api.github.com/repos/huggingface/datasets/issues/791/events
https://github.com/huggingface/datasets/pull/791
734,656,518
MDExOlB1bGxSZXF1ZXN0NTE0MTg0MzU5
791
add amazon reviews
{ "login": "joeddav", "id": 9353833, "node_id": "MDQ6VXNlcjkzNTM4MzM=", "avatar_url": "https://avatars.githubusercontent.com/u/9353833?v=4", "gravatar_id": "", "url": "https://api.github.com/users/joeddav", "html_url": "https://github.com/joeddav", "followers_url": "https://api.github.com/users/joeddav/...
[]
closed
false
null
[]
null
[ "@patrickvonplaten Yeah this is adapted from tfds so a lot is just how they wrote the code. Addressed your comments and also simplified the weird `AmazonUSReviewsConfig` definition. Will merge once tests pass.", "Thanks for checking this one :) \r\nLooks good to me \r\n\r\nJust one question : is there a particula...
1,604,335,377,000
1,604,434,506,000
1,604,421,837,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/791", "html_url": "https://github.com/huggingface/datasets/pull/791", "diff_url": "https://github.com/huggingface/datasets/pull/791.diff", "patch_url": "https://github.com/huggingface/datasets/pull/791.patch" }
Adds the Amazon US Reviews dataset as requested in #353. Converted from [TensorFlow Datasets](https://www.tensorflow.org/datasets/catalog/amazon_us_reviews). cc @clmnt @sshleifer
https://api.github.com/repos/huggingface/datasets/issues/791/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/790
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/790/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/790/comments
https://api.github.com/repos/huggingface/datasets/issues/790/events
https://github.com/huggingface/datasets/issues/790
734,470,197
MDU6SXNzdWU3MzQ0NzAxOTc=
790
Error running pip install -e ".[dev]" on MacOS 10.13.6: faiss/python does not exist
{ "login": "shawwn", "id": 59632, "node_id": "MDQ6VXNlcjU5NjMy", "avatar_url": "https://avatars.githubusercontent.com/u/59632?v=4", "gravatar_id": "", "url": "https://api.github.com/users/shawwn", "html_url": "https://github.com/shawwn", "followers_url": "https://api.github.com/users/shawwn/followers", ...
[]
closed
false
null
[]
null
[ "I saw that `faiss-cpu` 1.6.4.post2 was released recently to fix the installation on macos. It should work now", "Closing this one.\r\nFeel free to re-open if you still have issues" ]
1,604,320,595,000
1,605,017,102,000
1,605,017,102,000
NONE
null
null
I was following along with https://huggingface.co/docs/datasets/share_dataset.html#adding-tests-and-metadata-to-the-dataset when I ran into this error. ```sh git clone https://github.com/huggingface/datasets cd datasets virtualenv venv -p python3 --system-site-packages source venv/bin/activate pip install -e "....
https://api.github.com/repos/huggingface/datasets/issues/790/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/789
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/789/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/789/comments
https://api.github.com/repos/huggingface/datasets/issues/789/events
https://github.com/huggingface/datasets/pull/789
734,237,839
MDExOlB1bGxSZXF1ZXN0NTEzODM1MzE0
789
dataset(ncslgr): add initial loading script
{ "login": "AmitMY", "id": 5757359, "node_id": "MDQ6VXNlcjU3NTczNTk=", "avatar_url": "https://avatars.githubusercontent.com/u/5757359?v=4", "gravatar_id": "", "url": "https://api.github.com/users/AmitMY", "html_url": "https://github.com/AmitMY", "followers_url": "https://api.github.com/users/AmitMY/foll...
[]
closed
false
null
[]
null
[ "Hi @AmitMY, sorry for leaving you hanging for a minute :) \r\n\r\nWe've developed a new pipeline for adding datasets with a few extra steps, including adding a dataset card. You can find the full process [here](https://github.com/huggingface/datasets/blob/master/ADD_NEW_DATASET.md)\r\n\r\nWould you be up for addin...
1,604,299,810,000
1,606,830,097,000
1,606,830,096,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/789", "html_url": "https://github.com/huggingface/datasets/pull/789", "diff_url": "https://github.com/huggingface/datasets/pull/789.diff", "patch_url": "https://github.com/huggingface/datasets/pull/789.patch" }
Its a small dataset, but its heavily annotated https://www.bu.edu/asllrp/ncslgr.html ![image](https://user-images.githubusercontent.com/5757359/97838609-3c539380-1ce9-11eb-885b-a15d4c91ea49.png)
https://api.github.com/repos/huggingface/datasets/issues/789/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/788
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/788/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/788/comments
https://api.github.com/repos/huggingface/datasets/issues/788/events
https://github.com/huggingface/datasets/issues/788
734,136,124
MDU6SXNzdWU3MzQxMzYxMjQ=
788
failed to reuse cache
{ "login": "WangHexie", "id": 31768052, "node_id": "MDQ6VXNlcjMxNzY4MDUy", "avatar_url": "https://avatars.githubusercontent.com/u/31768052?v=4", "gravatar_id": "", "url": "https://api.github.com/users/WangHexie", "html_url": "https://github.com/WangHexie", "followers_url": "https://api.github.com/users/...
[]
closed
false
null
[]
null
[]
1,604,284,956,000
1,604,319,975,000
1,604,319,975,000
NONE
null
null
I packed the `load_dataset ` in a function of class, and cached data in a directory. But when I import the class and use the function, the data still have to be downloaded again. The information (Downloading and preparing dataset cnn_dailymail/3.0.0 (download: 558.32 MiB, generated: 1.28 GiB, post-processed: Unknown si...
https://api.github.com/repos/huggingface/datasets/issues/788/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/787
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/787/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/787/comments
https://api.github.com/repos/huggingface/datasets/issues/787/events
https://github.com/huggingface/datasets/pull/787
734,070,162
MDExOlB1bGxSZXF1ZXN0NTEzNjk5MTQz
787
Adding nli_tr dataset
{ "login": "e-budur", "id": 2246791, "node_id": "MDQ6VXNlcjIyNDY3OTE=", "avatar_url": "https://avatars.githubusercontent.com/u/2246791?v=4", "gravatar_id": "", "url": "https://api.github.com/users/e-budur", "html_url": "https://github.com/e-budur", "followers_url": "https://api.github.com/users/e-budur/...
[]
closed
false
null
[]
null
[ "Thank you @lhoestq for the time you take to review our pull request. We appreciate your help.\r\n\r\nWe've made the changes you described. Hope that it is ready for being merged. Please let me know if you have any additional requests for revisions. " ]
1,604,267,384,000
1,605,207,962,000
1,605,207,962,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/787", "html_url": "https://github.com/huggingface/datasets/pull/787", "diff_url": "https://github.com/huggingface/datasets/pull/787.diff", "patch_url": "https://github.com/huggingface/datasets/pull/787.patch" }
Hello, In this pull request, we have implemented the necessary interface to add our recent dataset [NLI-TR](https://github.com/boun-tabi/NLI-TR). The datasets will be presented on a full paper at EMNLP 2020 this month. [[arXiv link] ](https://arxiv.org/pdf/2004.14963.pdf) The dataset is the neural machine transl...
https://api.github.com/repos/huggingface/datasets/issues/787/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/786
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/786/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/786/comments
https://api.github.com/repos/huggingface/datasets/issues/786/events
https://github.com/huggingface/datasets/issues/786
733,761,717
MDU6SXNzdWU3MzM3NjE3MTc=
786
feat(dataset): multiprocessing _generate_examples
{ "login": "AmitMY", "id": 5757359, "node_id": "MDQ6VXNlcjU3NTczNTk=", "avatar_url": "https://avatars.githubusercontent.com/u/5757359?v=4", "gravatar_id": "", "url": "https://api.github.com/users/AmitMY", "html_url": "https://github.com/AmitMY", "followers_url": "https://api.github.com/users/AmitMY/foll...
[]
open
false
null
[]
null
[ "I agree that would be cool :)\r\nRight now the only distributed dataset builder is based on Apache Beam so you can use distributed processing frameworks like Dataflow, Spark, Flink etc. to build your dataset but it's not really well suited for single-worker parallel processing afaik", "`_generate_examples` can n...
1,604,163,136,000
1,604,911,118,000
null
CONTRIBUTOR
null
null
forking this out of #741, this issue is only regarding multiprocessing I'd love if there was a dataset configuration parameter `workers`, where when it is `1` it behaves as it does right now, and when its `>1` maybe `_generate_examples` can also get the `pool` and return an iterable using the pool. In my use case...
https://api.github.com/repos/huggingface/datasets/issues/786/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/785
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/785/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/785/comments
https://api.github.com/repos/huggingface/datasets/issues/785/events
https://github.com/huggingface/datasets/pull/785
733,719,419
MDExOlB1bGxSZXF1ZXN0NTEzNDMyNTM1
785
feat(aslg_pc12): add dev and test data splits
{ "login": "AmitMY", "id": 5757359, "node_id": "MDQ6VXNlcjU3NTczNTk=", "avatar_url": "https://avatars.githubusercontent.com/u/5757359?v=4", "gravatar_id": "", "url": "https://api.github.com/users/AmitMY", "html_url": "https://github.com/AmitMY", "followers_url": "https://api.github.com/users/AmitMY/foll...
[]
closed
false
null
[]
null
[ "Hi ! I'm not sure we should make this split decision arbitrarily on our side. Users can split it afterwards to whatever they want using `dataset.train_test_split` for example.\r\nMoreover it looks like there's already papers that use this dataset and propose their own splits ([here](http://xanthippi.ceid.upatras.g...
1,604,150,738,000
1,605,022,170,000
1,605,022,170,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/785", "html_url": "https://github.com/huggingface/datasets/pull/785", "diff_url": "https://github.com/huggingface/datasets/pull/785.diff", "patch_url": "https://github.com/huggingface/datasets/pull/785.patch" }
For reproducibility sake, it's best if there are defined dev and test splits. The original paper author did not define splits for the entire dataset, not for the sample loaded via this library, so I decided to define: - 5/7th for train - 1/7th for dev - 1/7th for test
https://api.github.com/repos/huggingface/datasets/issues/785/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/784
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/784/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/784/comments
https://api.github.com/repos/huggingface/datasets/issues/784/events
https://github.com/huggingface/datasets/issues/784
733,700,463
MDU6SXNzdWU3MzM3MDA0NjM=
784
Issue with downloading Wikipedia data for low resource language
{ "login": "SamuelCahyawijaya", "id": 2826602, "node_id": "MDQ6VXNlcjI4MjY2MDI=", "avatar_url": "https://avatars.githubusercontent.com/u/2826602?v=4", "gravatar_id": "", "url": "https://api.github.com/users/SamuelCahyawijaya", "html_url": "https://github.com/SamuelCahyawijaya", "followers_url": "https:/...
[]
closed
false
null
[]
null
[ "Hello, maybe you could ty to use another date for the wikipedia dump (see the available [dates](https://dumps.wikimedia.org/jvwiki) here for `jv`) ?", "@lhoestq\r\n\r\nI've tried `load_dataset('wikipedia', '20200501.zh', beam_runner='DirectRunner')` and got the same `FileNotFoundError` as @SamuelCahyawijaya.\r\n...
1,604,144,400,000
1,624,584,931,000
1,606,318,933,000
NONE
null
null
Hi, I tried to download Sundanese and Javanese wikipedia data with the following snippet ``` jv_wiki = datasets.load_dataset('wikipedia', '20200501.jv', beam_runner='DirectRunner') su_wiki = datasets.load_dataset('wikipedia', '20200501.su', beam_runner='DirectRunner') ``` And I get the following error for these tw...
https://api.github.com/repos/huggingface/datasets/issues/784/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/783
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/783/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/783/comments
https://api.github.com/repos/huggingface/datasets/issues/783/events
https://github.com/huggingface/datasets/pull/783
733,536,254
MDExOlB1bGxSZXF1ZXN0NTEzMzAwODUz
783
updated links to v1.3 of quail, fixed the description
{ "login": "annargrs", "id": 1450322, "node_id": "MDQ6VXNlcjE0NTAzMjI=", "avatar_url": "https://avatars.githubusercontent.com/u/1450322?v=4", "gravatar_id": "", "url": "https://api.github.com/users/annargrs", "html_url": "https://github.com/annargrs", "followers_url": "https://api.github.com/users/annar...
[]
closed
false
null
[]
null
[ "we're using quail 1.3 now thanks.\r\nclosing this one" ]
1,604,094,453,000
1,606,691,119,000
1,606,691,118,000
NONE
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/783", "html_url": "https://github.com/huggingface/datasets/pull/783", "diff_url": "https://github.com/huggingface/datasets/pull/783.diff", "patch_url": "https://github.com/huggingface/datasets/pull/783.patch" }
updated links to v1.3 of quail, fixed the description
https://api.github.com/repos/huggingface/datasets/issues/783/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/782
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/782/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/782/comments
https://api.github.com/repos/huggingface/datasets/issues/782/events
https://github.com/huggingface/datasets/pull/782
733,316,463
MDExOlB1bGxSZXF1ZXN0NTEzMTE2MTM0
782
Fix metric deletion when attribuets are missing
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,604,074,570,000
1,604,076,473,000
1,604,076,472,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/782", "html_url": "https://github.com/huggingface/datasets/pull/782", "diff_url": "https://github.com/huggingface/datasets/pull/782.diff", "patch_url": "https://github.com/huggingface/datasets/pull/782.patch" }
When you call `del` on a metric we want to make sure that the arrow attributes are not already deleted. I just added `if hasattr(...)` to make sure it doesn't crash
https://api.github.com/repos/huggingface/datasets/issues/782/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/781
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/781/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/781/comments
https://api.github.com/repos/huggingface/datasets/issues/781/events
https://github.com/huggingface/datasets/pull/781
733,168,609
MDExOlB1bGxSZXF1ZXN0NTEyOTkyMzQw
781
Add XNLI train set
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[ "Hi! Thanks for adding the translated MNLI! Do you know what translations system / model you used when you created the datasets in the other languages?", "According to the [paper](https://arxiv.org/pdf/1809.05053.pdf) it's the result of the work of professional translators ;)", "Thanks for getting back to me.\n...
1,604,064,113,000
1,604,946,170,000
1,604,946,169,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/781", "html_url": "https://github.com/huggingface/datasets/pull/781", "diff_url": "https://github.com/huggingface/datasets/pull/781.diff", "patch_url": "https://github.com/huggingface/datasets/pull/781.patch" }
I added the train set that was built using the translated MNLI. Now you can load the dataset specifying one language: ```python from datasets import load_dataset xnli_en = load_dataset("xnli", "en") print(xnli_en["train"][0]) # {'hypothesis': 'Product and geography are what make cream skimming work .', 'label':...
https://api.github.com/repos/huggingface/datasets/issues/781/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/780
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/780/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/780/comments
https://api.github.com/repos/huggingface/datasets/issues/780/events
https://github.com/huggingface/datasets/pull/780
732,738,647
MDExOlB1bGxSZXF1ZXN0NTEyNjM0MzI0
780
Add ASNQ dataset
{ "login": "mkserge", "id": 2992022, "node_id": "MDQ6VXNlcjI5OTIwMjI=", "avatar_url": "https://avatars.githubusercontent.com/u/2992022?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mkserge", "html_url": "https://github.com/mkserge", "followers_url": "https://api.github.com/users/mkserge/...
[]
closed
false
null
[]
null
[ "Very nice !\r\nWhat do the `sentence1` and `sentence2` correspond to exactly ?\r\nAlso maybe you could use the `ClassLabel` feature type for the `label` field (see [snli](https://github.com/huggingface/datasets/blob/master/datasets/snli/snli.py) for example)", "> What do the `sentence1` and `sentence2` correspon...
1,604,014,316,000
1,605,000,383,000
1,605,000,383,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/780", "html_url": "https://github.com/huggingface/datasets/pull/780", "diff_url": "https://github.com/huggingface/datasets/pull/780.diff", "patch_url": "https://github.com/huggingface/datasets/pull/780.patch" }
This pull request adds the ASNQ dataset. It is a dataset for answer sentence selection derived from Google Natural Questions (NQ) dataset (Kwiatkowski et al. 2019). The dataset details can be found in the paper at https://arxiv.org/abs/1911.04118 The dataset is authored by Siddhant Garg, Thuy Vu and Alessandro Mosch...
https://api.github.com/repos/huggingface/datasets/issues/780/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/779
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/779/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/779/comments
https://api.github.com/repos/huggingface/datasets/issues/779/events
https://github.com/huggingface/datasets/pull/779
732,514,887
MDExOlB1bGxSZXF1ZXN0NTEyNDQzMjY0
779
Feature/fidelity metrics from emnlp2020 evaluating and characterizing human rationales
{ "login": "rathoreanirudh", "id": 11327413, "node_id": "MDQ6VXNlcjExMzI3NDEz", "avatar_url": "https://avatars.githubusercontent.com/u/11327413?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rathoreanirudh", "html_url": "https://github.com/rathoreanirudh", "followers_url": "https://api.gi...
[]
open
false
null
[]
null
[ "Hi ! This looks interesting, thanks for adding it :) \r\n\r\nFor metrics there should only be two features fields: references and predictions.\r\nBoth of them can be defined as you want using nested structures if you need to.\r\nAlso I'm not sure what goes into references and what goes into predictions, could you ...
1,603,992,674,000
1,605,291,082,000
null
NONE
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/779", "html_url": "https://github.com/huggingface/datasets/pull/779", "diff_url": "https://github.com/huggingface/datasets/pull/779.diff", "patch_url": "https://github.com/huggingface/datasets/pull/779.patch" }
This metric computes fidelity (Yu et al. 2019, DeYoung et al. 2019) and normalized fidelity (Carton et al. 2020).
https://api.github.com/repos/huggingface/datasets/issues/779/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/778
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/778/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/778/comments
https://api.github.com/repos/huggingface/datasets/issues/778/events
https://github.com/huggingface/datasets/issues/778
732,449,652
MDU6SXNzdWU3MzI0NDk2NTI=
778
Unexpected behavior when loading cached csv file?
{ "login": "dcfidalgo", "id": 15979778, "node_id": "MDQ6VXNlcjE1OTc5Nzc4", "avatar_url": "https://avatars.githubusercontent.com/u/15979778?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dcfidalgo", "html_url": "https://github.com/dcfidalgo", "followers_url": "https://api.github.com/users/...
[]
closed
false
null
[]
null
[ "Hi ! Thanks for reporting.\r\nThe same issue was reported in #730 (but with the encodings instead of the delimiter). It was fixed by #770 .\r\nThe fix will be available in the next release :)", "Thanks for the prompt reply and terribly sorry for the spam! \r\nLooking forward to the new release! " ]
1,603,987,570,000
1,604,006,487,000
1,604,006,487,000
CONTRIBUTOR
null
null
I read a csv file from disk and forgot so specify the right delimiter. When i read the csv file again specifying the right delimiter it had no effect since it was using the cached dataset. I am not sure if this is unwanted behavior since i can always specify `download_mode="force_redownload"`. But i think it would be n...
https://api.github.com/repos/huggingface/datasets/issues/778/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/777
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/777/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/777/comments
https://api.github.com/repos/huggingface/datasets/issues/777/events
https://github.com/huggingface/datasets/pull/777
732,376,648
MDExOlB1bGxSZXF1ZXN0NTEyMzI2ODM2
777
Better error message for uninitialized metric
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,603,982,570,000
1,603,984,706,000
1,603,984,704,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/777", "html_url": "https://github.com/huggingface/datasets/pull/777", "diff_url": "https://github.com/huggingface/datasets/pull/777.diff", "patch_url": "https://github.com/huggingface/datasets/pull/777.patch" }
When calling `metric.compute()` without having called `metric.add` or `metric.add_batch` at least once, the error was quite cryptic. I added a better error message Fix #729
https://api.github.com/repos/huggingface/datasets/issues/777/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/776
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/776/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/776/comments
https://api.github.com/repos/huggingface/datasets/issues/776/events
https://github.com/huggingface/datasets/pull/776
732,343,550
MDExOlB1bGxSZXF1ZXN0NTEyMjk5NzQx
776
Allow custom split names in text dataset
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[ "Awesome! This will make the behaviour much more intuitive for some non-standard code.\r\n\r\nThanks!" ]
1,603,980,246,000
1,604,065,605,000
1,604,064,232,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/776", "html_url": "https://github.com/huggingface/datasets/pull/776", "diff_url": "https://github.com/huggingface/datasets/pull/776.diff", "patch_url": "https://github.com/huggingface/datasets/pull/776.patch" }
The `text` dataset used to return only splits like train, test and validation. Other splits were ignored. Now any split name is allowed. I did the same for `json`, `pandas` and `csv` Fix #735
https://api.github.com/repos/huggingface/datasets/issues/776/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/775
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/775/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/775/comments
https://api.github.com/repos/huggingface/datasets/issues/775/events
https://github.com/huggingface/datasets/pull/775
732,287,504
MDExOlB1bGxSZXF1ZXN0NTEyMjUyODI3
775
Properly delete metrics when a process is killed
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,603,975,927,000
1,603,980,080,000
1,603,980,079,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/775", "html_url": "https://github.com/huggingface/datasets/pull/775", "diff_url": "https://github.com/huggingface/datasets/pull/775.diff", "patch_url": "https://github.com/huggingface/datasets/pull/775.patch" }
Tests are flaky when using metrics in distributed setup. There is because of one test that make sure that using two possibly incompatible metric computation (same exp id) either works or raises the right error. However if the error is raised, all the processes of the metric are killed, and the open files (arrow + loc...
https://api.github.com/repos/huggingface/datasets/issues/775/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/774
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/774/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/774/comments
https://api.github.com/repos/huggingface/datasets/issues/774/events
https://github.com/huggingface/datasets/pull/774
732,265,741
MDExOlB1bGxSZXF1ZXN0NTEyMjM0NjA0
774
[ROUGE] Add description to Rouge metric
{ "login": "patrickvonplaten", "id": 23423619, "node_id": "MDQ6VXNlcjIzNDIzNjE5", "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "gravatar_id": "", "url": "https://api.github.com/users/patrickvonplaten", "html_url": "https://github.com/patrickvonplaten", "followers_url": "https://...
[]
closed
false
null
[]
null
[]
1,603,973,972,000
1,603,994,150,000
1,603,994,148,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/774", "html_url": "https://github.com/huggingface/datasets/pull/774", "diff_url": "https://github.com/huggingface/datasets/pull/774.diff", "patch_url": "https://github.com/huggingface/datasets/pull/774.patch" }
Add information about case sensitivity to ROUGE.
https://api.github.com/repos/huggingface/datasets/issues/774/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/773
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/773/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/773/comments
https://api.github.com/repos/huggingface/datasets/issues/773/events
https://github.com/huggingface/datasets/issues/773
731,684,153
MDU6SXNzdWU3MzE2ODQxNTM=
773
Adding CC-100: Monolingual Datasets from Web Crawl Data
{ "login": "yjernite", "id": 10469459, "node_id": "MDQ6VXNlcjEwNDY5NDU5", "avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4", "gravatar_id": "", "url": "https://api.github.com/users/yjernite", "html_url": "https://github.com/yjernite", "followers_url": "https://api.github.com/users/yje...
[ { "id": 2067376369, "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request", "name": "dataset request", "color": "e99695", "default": false, "description": "Requesting to add a new dataset" } ]
closed
false
{ "login": "abhishekkrthakur", "id": 1183441, "node_id": "MDQ6VXNlcjExODM0NDE=", "avatar_url": "https://avatars.githubusercontent.com/u/1183441?v=4", "gravatar_id": "", "url": "https://api.github.com/users/abhishekkrthakur", "html_url": "https://github.com/abhishekkrthakur", "followers_url": "https://ap...
[ { "login": "abhishekkrthakur", "id": 1183441, "node_id": "MDQ6VXNlcjExODM0NDE=", "avatar_url": "https://avatars.githubusercontent.com/u/1183441?v=4", "gravatar_id": "", "url": "https://api.github.com/users/abhishekkrthakur", "html_url": "https://github.com/abhishekkrthakur", "followe...
null
[ "cc @aconneau ;) ", "These dataset files are no longer available. https://data.statmt.org/cc-100/ files provided in this link are no longer available. Can anybody fix that issue?\r\n@abhishekkrthakur @yjernite ", "Hi ! Can you open an issue to report this problem ? This will help keep track of the fix :)", "...
1,603,909,241,000
1,607,941,208,000
1,607,941,207,000
MEMBER
null
null
## Adding a Dataset - **Name:** CC-100: Monolingual Datasets from Web Crawl Data - **Description:** https://twitter.com/alex_conneau/status/1321507120848625665 - **Paper:** https://arxiv.org/abs/1911.02116 - **Data:** http://data.statmt.org/cc-100/ - **Motivation:** A large scale multi-lingual language modeling da...
https://api.github.com/repos/huggingface/datasets/issues/773/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/772
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/772/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/772/comments
https://api.github.com/repos/huggingface/datasets/issues/772/events
https://github.com/huggingface/datasets/pull/772
731,612,430
MDExOlB1bGxSZXF1ZXN0NTExNjg4ODMx
772
Fix metric with cache dir
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,603,903,393,000
1,603,964,084,000
1,603,964,083,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/772", "html_url": "https://github.com/huggingface/datasets/pull/772", "diff_url": "https://github.com/huggingface/datasets/pull/772.diff", "patch_url": "https://github.com/huggingface/datasets/pull/772.patch" }
The cache_dir provided by the user was concatenated twice and therefore causing FileNotFound errors. The tests didn't cover the case of providing `cache_dir=` for metrics because of a stupid issue (it was not using the right parameter). I remove the double concatenation and I fixed the tests Fix #728
https://api.github.com/repos/huggingface/datasets/issues/772/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/771
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/771/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/771/comments
https://api.github.com/repos/huggingface/datasets/issues/771/events
https://github.com/huggingface/datasets/issues/771
731,482,213
MDU6SXNzdWU3MzE0ODIyMTM=
771
Using `Dataset.map` with `n_proc>1` print multiple progress bars
{ "login": "sgugger", "id": 35901082, "node_id": "MDQ6VXNlcjM1OTAxMDgy", "avatar_url": "https://avatars.githubusercontent.com/u/35901082?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sgugger", "html_url": "https://github.com/sgugger", "followers_url": "https://api.github.com/users/sgugge...
[]
open
false
null
[]
null
[ "Yes it allows to monitor the speed of each process. Currently each process takes care of one shard of the dataset.\r\n\r\nAt one point we can consider using streaming batches to a pool of processes instead of sharding the dataset in `num_proc` parts. At that point it will be easy to use only one progress bar", "...
1,603,894,407,000
1,603,894,697,000
null
MEMBER
null
null
When using `Dataset.map` with `n_proc > 1`, only one of the processes should print a progress bar (to make the output readable). Right now, `n_proc` progress bars are printed.
https://api.github.com/repos/huggingface/datasets/issues/771/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/770
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/770/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/770/comments
https://api.github.com/repos/huggingface/datasets/issues/770/events
https://github.com/huggingface/datasets/pull/770
731,445,222
MDExOlB1bGxSZXF1ZXN0NTExNTQ5MTg1
770
Fix custom builder caching
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,603,891,944,000
1,603,964,163,000
1,603,964,161,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/770", "html_url": "https://github.com/huggingface/datasets/pull/770", "diff_url": "https://github.com/huggingface/datasets/pull/770.diff", "patch_url": "https://github.com/huggingface/datasets/pull/770.patch" }
The cache directory of a dataset didn't take into account additional parameters that the user could specify such as `features` or any parameter of the builder configuration kwargs (ex: `encoding` for the `text` dataset). To fix that, the cache directory name now has a suffix that depends on all of them. Fix #730 ...
https://api.github.com/repos/huggingface/datasets/issues/770/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/769
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/769/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/769/comments
https://api.github.com/repos/huggingface/datasets/issues/769/events
https://github.com/huggingface/datasets/issues/769
731,257,104
MDU6SXNzdWU3MzEyNTcxMDQ=
769
How to choose proper download_mode in function load_dataset?
{ "login": "jzq2000", "id": 48550398, "node_id": "MDQ6VXNlcjQ4NTUwMzk4", "avatar_url": "https://avatars.githubusercontent.com/u/48550398?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jzq2000", "html_url": "https://github.com/jzq2000", "followers_url": "https://api.github.com/users/jzq200...
[]
open
false
null
[]
null
[ "`download_mode=datasets.GenerateMode.FORCE_REDOWNLOAD` should work.\r\nThis makes me think we we should rename this to DownloadMode.FORCE_REDOWNLOAD. Currently that's confusing", "Can we just use `features=...` in `load_dataset` for this @lhoestq?", "Indeed you should use `features` in this case. \r\n```python...
1,603,876,579,000
1,603,881,299,000
null
NONE
null
null
Hi, I am a beginner to datasets and I try to use datasets to load my csv file. my csv file looks like this ``` text,label "Effective but too-tepid biopic",3 "If you sometimes like to go to the movies to have fun , Wasabi is a good place to start .",4 "Emerges as something rare , an issue movie that 's so hones...
https://api.github.com/repos/huggingface/datasets/issues/769/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/768
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/768/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/768/comments
https://api.github.com/repos/huggingface/datasets/issues/768/events
https://github.com/huggingface/datasets/issues/768
730,908,060
MDU6SXNzdWU3MzA5MDgwNjA=
768
Add a `lazy_map` method to `Dataset` and `DatasetDict`
{ "login": "sgugger", "id": 35901082, "node_id": "MDQ6VXNlcjM1OTAxMDgy", "avatar_url": "https://avatars.githubusercontent.com/u/35901082?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sgugger", "html_url": "https://github.com/sgugger", "followers_url": "https://api.github.com/users/sgugge...
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
open
false
null
[]
null
[ "This is cool! I think some aspects to think about and decide in terms of API are:\r\n- do we allow several methods (chained i guess)\r\n- how do we inspect the currently set method(s)\r\n- how do we control/reset them" ]
1,603,837,983,000
1,603,875,493,000
null
MEMBER
null
null
The library is great, but it would be even more awesome with a `lazy_map` method implemented on `Dataset` and `DatasetDict`. This would apply a function on a give item but when the item is requested. Two use cases: 1. load image on the fly 2. apply a random function and get different outputs at each epoch (like dat...
https://api.github.com/repos/huggingface/datasets/issues/768/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/767
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/767/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/767/comments
https://api.github.com/repos/huggingface/datasets/issues/767/events
https://github.com/huggingface/datasets/issues/767
730,771,610
MDU6SXNzdWU3MzA3NzE2MTA=
767
Add option for named splits when using ds.train_test_split
{ "login": "nateraw", "id": 32437151, "node_id": "MDQ6VXNlcjMyNDM3MTUx", "avatar_url": "https://avatars.githubusercontent.com/u/32437151?v=4", "gravatar_id": "", "url": "https://api.github.com/users/nateraw", "html_url": "https://github.com/nateraw", "followers_url": "https://api.github.com/users/natera...
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
open
false
null
[]
null
[ "Yes definitely we should give more flexibility to control the name of the splits outputted by `train_test_split`.\r\n\r\nRelated is the very interesting feedback from @bramvanroy on how we should improve this method: https://discuss.huggingface.co/t/how-to-split-main-dataset-into-train-dev-test-as-datasetdict/1090...
1,603,828,784,000
1,605,017,121,000
null
CONTRIBUTOR
null
null
### Feature Request 🚀 Can we add a way to name your splits when using the `.train_test_split` function? In almost every use case I've come across, I have a `train` and a `test` split in my `DatasetDict`, and I want to create a `validation` split. Therefore, its kinda useless to get a `test` split back from `tra...
https://api.github.com/repos/huggingface/datasets/issues/767/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/766
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/766/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/766/comments
https://api.github.com/repos/huggingface/datasets/issues/766/events
https://github.com/huggingface/datasets/issues/766
730,669,596
MDU6SXNzdWU3MzA2Njk1OTY=
766
[GEM] add DART data-to-text generation dataset
{ "login": "yjernite", "id": 10469459, "node_id": "MDQ6VXNlcjEwNDY5NDU5", "avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4", "gravatar_id": "", "url": "https://api.github.com/users/yjernite", "html_url": "https://github.com/yjernite", "followers_url": "https://api.github.com/users/yje...
[ { "id": 2067376369, "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request", "name": "dataset request", "color": "e99695", "default": false, "description": "Requesting to add a new dataset" } ]
closed
false
null
[]
null
[ "Is this a duplicate of #924 ?", "Yup, closing! Haven't been keeping track of the solved issues during the sprint." ]
1,603,820,044,000
1,607,002,638,000
1,607,002,638,000
MEMBER
null
null
## Adding a Dataset - **Name:** DART - **Description:** DART consists of 82,191 examples across different domains with each input being a semantic RDF triple set derived from data records in tables and the tree ontology of the schema, annotated with sentence descriptions that cover all facts in the triple set. - **P...
https://api.github.com/repos/huggingface/datasets/issues/766/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/765
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/765/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/765/comments
https://api.github.com/repos/huggingface/datasets/issues/765/events
https://github.com/huggingface/datasets/issues/765
730,668,332
MDU6SXNzdWU3MzA2NjgzMzI=
765
[GEM] Add DART data-to-text generation dataset
{ "login": "yjernite", "id": 10469459, "node_id": "MDQ6VXNlcjEwNDY5NDU5", "avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4", "gravatar_id": "", "url": "https://api.github.com/users/yjernite", "html_url": "https://github.com/yjernite", "followers_url": "https://api.github.com/users/yje...
[ { "id": 2067376369, "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request", "name": "dataset request", "color": "e99695", "default": false, "description": "Requesting to add a new dataset" } ]
closed
false
null
[]
null
[]
1,603,819,943,000
1,603,820,061,000
1,603,820,061,000
MEMBER
null
null
## Adding a Dataset - **Name:** DART - **Description:** DART consists of 82,191 examples across different domains with each input being a semantic RDF triple set derived from data records in tables and the tree ontology of the schema, annotated with sentence descriptions that cover all facts in the triple set. - **P...
https://api.github.com/repos/huggingface/datasets/issues/765/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/764
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/764/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/764/comments
https://api.github.com/repos/huggingface/datasets/issues/764/events
https://github.com/huggingface/datasets/pull/764
730,617,828
MDExOlB1bGxSZXF1ZXN0NTEwODkyMTk2
764
Adding Issue Template for Dataset Requests
{ "login": "yjernite", "id": 10469459, "node_id": "MDQ6VXNlcjEwNDY5NDU5", "avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4", "gravatar_id": "", "url": "https://api.github.com/users/yjernite", "html_url": "https://github.com/yjernite", "followers_url": "https://api.github.com/users/yje...
[]
closed
false
null
[]
null
[]
1,603,816,628,000
1,603,819,526,000
1,603,819,525,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/764", "html_url": "https://github.com/huggingface/datasets/pull/764", "diff_url": "https://github.com/huggingface/datasets/pull/764.diff", "patch_url": "https://github.com/huggingface/datasets/pull/764.patch" }
adding .github/ISSUE_TEMPLATE/add-dataset.md
https://api.github.com/repos/huggingface/datasets/issues/764/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/763
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/763/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/763/comments
https://api.github.com/repos/huggingface/datasets/issues/763/events
https://github.com/huggingface/datasets/pull/763
730,593,631
MDExOlB1bGxSZXF1ZXN0NTEwODcyMDYx
763
Fixed errors in bertscore related to custom baseline
{ "login": "juanjucm", "id": 36761132, "node_id": "MDQ6VXNlcjM2NzYxMTMy", "avatar_url": "https://avatars.githubusercontent.com/u/36761132?v=4", "gravatar_id": "", "url": "https://api.github.com/users/juanjucm", "html_url": "https://github.com/juanjucm", "followers_url": "https://api.github.com/users/jua...
[]
closed
false
null
[]
null
[]
1,603,814,915,000
1,603,907,965,000
1,603,907,965,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/763", "html_url": "https://github.com/huggingface/datasets/pull/763", "diff_url": "https://github.com/huggingface/datasets/pull/763.diff", "patch_url": "https://github.com/huggingface/datasets/pull/763.patch" }
[bertscore version 0.3.6 ](https://github.com/Tiiiger/bert_score) added support for custom baseline files. This update added extra argument `baseline_path` to BERTScorer class as well as some extra boolean parameters `use_custom_baseline` in functions like `get_hash(model, num_layers, idf, rescale_with_baseline, use_cu...
https://api.github.com/repos/huggingface/datasets/issues/763/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/762
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/762/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/762/comments
https://api.github.com/repos/huggingface/datasets/issues/762/events
https://github.com/huggingface/datasets/issues/762
730,586,972
MDU6SXNzdWU3MzA1ODY5NzI=
762
[GEM] Add Czech Restaurant data-to-text generation dataset
{ "login": "yjernite", "id": 10469459, "node_id": "MDQ6VXNlcjEwNDY5NDU5", "avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4", "gravatar_id": "", "url": "https://api.github.com/users/yjernite", "html_url": "https://github.com/yjernite", "followers_url": "https://api.github.com/users/yje...
[ { "id": 2067376369, "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request", "name": "dataset request", "color": "e99695", "default": false, "description": "Requesting to add a new dataset" } ]
closed
false
null
[]
null
[]
1,603,814,447,000
1,607,002,664,000
1,607,002,664,000
MEMBER
null
null
- Paper: https://www.aclweb.org/anthology/W19-8670.pdf - Data: https://github.com/UFAL-DSG/cs_restaurant_dataset - The dataset will likely be part of the GEM benchmark
https://api.github.com/repos/huggingface/datasets/issues/762/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/761
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/761/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/761/comments
https://api.github.com/repos/huggingface/datasets/issues/761/events
https://github.com/huggingface/datasets/issues/761
729,898,867
MDU6SXNzdWU3Mjk4OTg4Njc=
761
Downloaded datasets are not usable offline
{ "login": "ghazi-f", "id": 25091538, "node_id": "MDQ6VXNlcjI1MDkxNTM4", "avatar_url": "https://avatars.githubusercontent.com/u/25091538?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ghazi-f", "html_url": "https://github.com/ghazi-f", "followers_url": "https://api.github.com/users/ghazi-...
[]
open
false
null
[]
null
[ "Yes currently you need an internet connection because the lib tries to check for the etag of the dataset script online to see if you don't have it locally already.\r\n\r\nIf we add a way to store the etag/hash locally after the first download, it would allow users to first download the dataset with an internet con...
1,603,745,686,000
1,603,807,469,000
null
CONTRIBUTOR
null
null
I've been trying to use the IMDB dataset offline, but after downloading it and turning off the internet it still raises an error from the ```requests``` library trying to reach for the online dataset. Is this the intended behavior ? (Sorry, I wrote the the first version of this issue while still on nlp 0.3.0).
https://api.github.com/repos/huggingface/datasets/issues/761/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/760
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/760/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/760/comments
https://api.github.com/repos/huggingface/datasets/issues/760/events
https://github.com/huggingface/datasets/issues/760
729,637,917
MDU6SXNzdWU3Mjk2Mzc5MTc=
760
Add meta-data to the HANS dataset
{ "login": "yjernite", "id": 10469459, "node_id": "MDQ6VXNlcjEwNDY5NDU5", "avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4", "gravatar_id": "", "url": "https://api.github.com/users/yjernite", "html_url": "https://github.com/yjernite", "followers_url": "https://api.github.com/users/yje...
[ { "id": 1935892877, "node_id": "MDU6TGFiZWwxOTM1ODkyODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/good%20first%20issue", "name": "good first issue", "color": "7057ff", "default": true, "description": "Good for newcomers" }, { "id": 2067388877, "node_...
closed
false
{ "login": "TevenLeScao", "id": 26709476, "node_id": "MDQ6VXNlcjI2NzA5NDc2", "avatar_url": "https://avatars.githubusercontent.com/u/26709476?v=4", "gravatar_id": "", "url": "https://api.github.com/users/TevenLeScao", "html_url": "https://github.com/TevenLeScao", "followers_url": "https://api.github.com/...
[ { "login": "TevenLeScao", "id": 26709476, "node_id": "MDQ6VXNlcjI2NzA5NDc2", "avatar_url": "https://avatars.githubusercontent.com/u/26709476?v=4", "gravatar_id": "", "url": "https://api.github.com/users/TevenLeScao", "html_url": "https://github.com/TevenLeScao", "followers_url": "htt...
null
[]
1,603,724,213,000
1,607,002,714,000
1,607,002,714,000
MEMBER
null
null
The current version of the [HANS dataset](https://github.com/huggingface/datasets/blob/master/datasets/hans/hans.py) is missing the additional information provided for each example, including the sentence parses, heuristic and subcase.
https://api.github.com/repos/huggingface/datasets/issues/760/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/759
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/759/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/759/comments
https://api.github.com/repos/huggingface/datasets/issues/759/events
https://github.com/huggingface/datasets/issues/759
729,046,916
MDU6SXNzdWU3MjkwNDY5MTY=
759
(Load dataset failure) ConnectionError: Couldn’t reach https://raw.githubusercontent.com/huggingface/datasets/1.1.2/datasets/cnn_dailymail/cnn_dailymail.py
{ "login": "AI678", "id": 63541083, "node_id": "MDQ6VXNlcjYzNTQxMDgz", "avatar_url": "https://avatars.githubusercontent.com/u/63541083?v=4", "gravatar_id": "", "url": "https://api.github.com/users/AI678", "html_url": "https://github.com/AI678", "followers_url": "https://api.github.com/users/AI678/follow...
[]
closed
false
null
[]
null
[ "Are you running the script on a machine with an internet connection ?", "Yes , I can browse the url through Google Chrome.", "Does this HEAD request return 200 on your machine ?\r\n```python\r\nimport requests ...
1,603,640,097,000
1,628,100,609,000
1,628,100,609,000
NONE
null
null
Hey, I want to load the cnn-dailymail dataset for fine-tune. I write the code like this from datasets import load_dataset test_dataset = load_dataset(“cnn_dailymail”, “3.0.0”, split=“train”) And I got the following errors. Traceback (most recent call last): File “test.py”, line 7, in test_dataset = load_da...
https://api.github.com/repos/huggingface/datasets/issues/759/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/758
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/758/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/758/comments
https://api.github.com/repos/huggingface/datasets/issues/758/events
https://github.com/huggingface/datasets/issues/758
728,638,559
MDU6SXNzdWU3Mjg2Mzg1NTk=
758
Process 0 very slow when using num_procs with map to tokenizer
{ "login": "ksjae", "id": 17930170, "node_id": "MDQ6VXNlcjE3OTMwMTcw", "avatar_url": "https://avatars.githubusercontent.com/u/17930170?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ksjae", "html_url": "https://github.com/ksjae", "followers_url": "https://api.github.com/users/ksjae/follow...
[]
closed
false
null
[]
null
[ "Hi ! Thanks for reporting.\r\nIs the distribution of text length of your data evenly distributed across your dataset ? I mean, could it be because the examples in the first part of your dataset are slower to process ?\r\nAlso could how many CPUs can you use for multiprocessing ?\r\n```python\r\nimport multiprocess...
1,603,507,220,000
1,603,857,586,000
1,603,857,585,000
NONE
null
null
<img width="721" alt="image" src="https://user-images.githubusercontent.com/17930170/97066109-776d0d00-15ed-11eb-8bba-bb4d2e0fcc33.png"> The code I am using is ``` dataset = load_dataset("text", data_files=[file_path], split='train') dataset = dataset.map(lambda ex: tokenizer(ex["text"], add_speci...
https://api.github.com/repos/huggingface/datasets/issues/758/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/757
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/757/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/757/comments
https://api.github.com/repos/huggingface/datasets/issues/757/events
https://github.com/huggingface/datasets/issues/757
728,241,494
MDU6SXNzdWU3MjgyNDE0OTQ=
757
CUDA out of memory
{ "login": "li1117heex", "id": 47059217, "node_id": "MDQ6VXNlcjQ3MDU5MjE3", "avatar_url": "https://avatars.githubusercontent.com/u/47059217?v=4", "gravatar_id": "", "url": "https://api.github.com/users/li1117heex", "html_url": "https://github.com/li1117heex", "followers_url": "https://api.github.com/use...
[]
closed
false
null
[]
null
[ "Could you provide more details ? What's the code you ran ?", "```python\r\ntokenizer = FunnelTokenizer.from_pretrained('funnel-transformer/small')\r\n\r\ndef tokenize(batch):\r\n return tokenizer(batch['text'], padding='max_length', truncation=True,max_length=512)\r\n\r\ndataset = load_dataset(\"bookcorpus\",...
1,603,461,420,000
1,608,732,389,000
1,608,732,389,000
NONE
null
null
In your dataset ,cuda run out of memory as long as the trainer begins: however, without changing any other element/parameter,just switch dataset to `LineByLineTextDataset`,everything becames OK.
https://api.github.com/repos/huggingface/datasets/issues/757/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/756
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/756/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/756/comments
https://api.github.com/repos/huggingface/datasets/issues/756/events
https://github.com/huggingface/datasets/pull/756
728,211,373
MDExOlB1bGxSZXF1ZXN0NTA4OTYwNTc3
756
Start community-provided dataset docs
{ "login": "sshleifer", "id": 6045025, "node_id": "MDQ6VXNlcjYwNDUwMjU=", "avatar_url": "https://avatars.githubusercontent.com/u/6045025?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sshleifer", "html_url": "https://github.com/sshleifer", "followers_url": "https://api.github.com/users/ss...
[]
closed
false
null
[]
null
[ "Oh, really cool @sshleifer!" ]
1,603,459,061,000
1,603,716,920,000
1,603,716,919,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/756", "html_url": "https://github.com/huggingface/datasets/pull/756", "diff_url": "https://github.com/huggingface/datasets/pull/756.diff", "patch_url": "https://github.com/huggingface/datasets/pull/756.patch" }
Continuation of #736 with clean fork. #### Old description This is what I did to get the pseudo-labels updated. Not sure if it generalizes, but I figured I would write it down. It was pretty easy because all I had to do was make properly formatted directories and change URLs. In slack @thomwolf called it a user-...
https://api.github.com/repos/huggingface/datasets/issues/756/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/755
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/755/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/755/comments
https://api.github.com/repos/huggingface/datasets/issues/755/events
https://github.com/huggingface/datasets/pull/755
728,203,821
MDExOlB1bGxSZXF1ZXN0NTA4OTU0NDI2
755
Start community-provided dataset docs V2
{ "login": "sshleifer", "id": 6045025, "node_id": "MDQ6VXNlcjYwNDUwMjU=", "avatar_url": "https://avatars.githubusercontent.com/u/6045025?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sshleifer", "html_url": "https://github.com/sshleifer", "followers_url": "https://api.github.com/users/ss...
[]
closed
false
null
[]
null
[]
1,603,458,450,000
1,603,458,937,000
1,603,458,937,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/755", "html_url": "https://github.com/huggingface/datasets/pull/755", "diff_url": "https://github.com/huggingface/datasets/pull/755.diff", "patch_url": "https://github.com/huggingface/datasets/pull/755.patch" }
https://api.github.com/repos/huggingface/datasets/issues/755/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/754
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/754/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/754/comments
https://api.github.com/repos/huggingface/datasets/issues/754/events
https://github.com/huggingface/datasets/pull/754
727,863,105
MDExOlB1bGxSZXF1ZXN0NTA4NjczNzM2
754
Use full released xsum dataset
{ "login": "jbragg", "id": 2238344, "node_id": "MDQ6VXNlcjIyMzgzNDQ=", "avatar_url": "https://avatars.githubusercontent.com/u/2238344?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jbragg", "html_url": "https://github.com/jbragg", "followers_url": "https://api.github.com/users/jbragg/foll...
[]
closed
false
null
[]
null
[ "@lhoestq I took a shot at addressing your comments but the build scripts seem to be complaining about not being able to open dummy files. How do I resolve those errors without copying the full dataset into the dummy dir?", "Could you check that the names of the dummy data files are right ?\r\nYou can use \r\n```...
1,603,423,789,000
1,609,470,716,000
1,603,717,018,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/754", "html_url": "https://github.com/huggingface/datasets/pull/754", "diff_url": "https://github.com/huggingface/datasets/pull/754.diff", "patch_url": "https://github.com/huggingface/datasets/pull/754.patch" }
#672 Fix xsum to expand coverage and include IDs Code based on parser from older version of `datasets/xsum/xsum.py` @lhoestq
https://api.github.com/repos/huggingface/datasets/issues/754/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/753
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/753/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/753/comments
https://api.github.com/repos/huggingface/datasets/issues/753/events
https://github.com/huggingface/datasets/pull/753
727,434,935
MDExOlB1bGxSZXF1ZXN0NTA4MzI4ODM0
753
Fix doc links to viewer
{ "login": "Pierrci", "id": 5020707, "node_id": "MDQ6VXNlcjUwMjA3MDc=", "avatar_url": "https://avatars.githubusercontent.com/u/5020707?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Pierrci", "html_url": "https://github.com/Pierrci", "followers_url": "https://api.github.com/users/Pierrci/...
[]
closed
false
null
[]
null
[]
1,603,376,416,000
1,603,442,531,000
1,603,442,531,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/753", "html_url": "https://github.com/huggingface/datasets/pull/753", "diff_url": "https://github.com/huggingface/datasets/pull/753.diff", "patch_url": "https://github.com/huggingface/datasets/pull/753.patch" }
It seems #733 forgot some links in the doc :)
https://api.github.com/repos/huggingface/datasets/issues/753/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/752
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/752/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/752/comments
https://api.github.com/repos/huggingface/datasets/issues/752/events
https://github.com/huggingface/datasets/issues/752
726,917,801
MDU6SXNzdWU3MjY5MTc4MDE=
752
Clicking on a metric in the search page points to datasets page giving "Missing dataset" warning
{ "login": "ogabrielluiz", "id": 24829397, "node_id": "MDQ6VXNlcjI0ODI5Mzk3", "avatar_url": "https://avatars.githubusercontent.com/u/24829397?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ogabrielluiz", "html_url": "https://github.com/ogabrielluiz", "followers_url": "https://api.github.c...
[]
closed
false
null
[]
null
[ "Thanks for the report, can reproduce. Will fix", "Fixed now @ogabrielluiz " ]
1,603,320,983,000
1,603,383,582,000
1,603,383,582,000
NONE
null
null
Hi! Sorry if this isn't the right place to talk about the website, I just didn't exactly where to write this. Searching a metric in https://huggingface.co/metrics gives the right results but clicking on a metric (E.g ROUGE) points to https://huggingface.co/datasets/rouge. Clicking on a metric without searching point...
https://api.github.com/repos/huggingface/datasets/issues/752/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/751
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/751/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/751/comments
https://api.github.com/repos/huggingface/datasets/issues/751/events
https://github.com/huggingface/datasets/issues/751
726,820,191
MDU6SXNzdWU3MjY4MjAxOTE=
751
Error loading ms_marco v2.1 using load_dataset()
{ "login": "JainSahit", "id": 30478979, "node_id": "MDQ6VXNlcjMwNDc4OTc5", "avatar_url": "https://avatars.githubusercontent.com/u/30478979?v=4", "gravatar_id": "", "url": "https://api.github.com/users/JainSahit", "html_url": "https://github.com/JainSahit", "followers_url": "https://api.github.com/users/...
[]
closed
false
null
[]
null
[ "There was a similar issue in #294 \r\nClearing the cache and download again the dataset did the job. Could you try to clear your cache and download the dataset again ?", "I was able to load the dataset successfully, I'm pretty sure it's just a cache issue that you have.\r\nLet me know if clearing your cache fixe...
1,603,310,083,000
1,604,539,917,000
1,604,539,917,000
NONE
null
null
Code: `dataset = load_dataset('ms_marco', 'v2.1')` Error: ``` `--------------------------------------------------------------------------- JSONDecodeError Traceback (most recent call last) <ipython-input-16-34378c057212> in <module>() 9 10 # Downloading and loading a data...
https://api.github.com/repos/huggingface/datasets/issues/751/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/750
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/750/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/750/comments
https://api.github.com/repos/huggingface/datasets/issues/750/events
https://github.com/huggingface/datasets/issues/750
726,589,446
MDU6SXNzdWU3MjY1ODk0NDY=
750
load_dataset doesn't include `features` in its hash
{ "login": "sgugger", "id": 35901082, "node_id": "MDQ6VXNlcjM1OTAxMDgy", "avatar_url": "https://avatars.githubusercontent.com/u/35901082?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sgugger", "html_url": "https://github.com/sgugger", "followers_url": "https://api.github.com/users/sgugge...
[]
closed
false
null
[]
null
[]
1,603,293,401,000
1,603,964,161,000
1,603,964,161,000
MEMBER
null
null
It looks like the function `load_dataset` does not include what's passed in the `features` argument when creating a hash for a given dataset. As a result, if a user includes new features from an already downloaded dataset, those are ignored. Example: some models on the hub have a different ordering for the labels t...
https://api.github.com/repos/huggingface/datasets/issues/750/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/749
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/749/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/749/comments
https://api.github.com/repos/huggingface/datasets/issues/749/events
https://github.com/huggingface/datasets/issues/749
726,366,062
MDU6SXNzdWU3MjYzNjYwNjI=
749
[XGLUE] Adding new dataset
{ "login": "patrickvonplaten", "id": 23423619, "node_id": "MDQ6VXNlcjIzNDIzNjE5", "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "gravatar_id": "", "url": "https://api.github.com/users/patrickvonplaten", "html_url": "https://github.com/patrickvonplaten", "followers_url": "https://...
[ { "id": 2067376369, "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request", "name": "dataset request", "color": "e99695", "default": false, "description": "Requesting to add a new dataset" } ]
closed
false
{ "login": "patrickvonplaten", "id": 23423619, "node_id": "MDQ6VXNlcjIzNDIzNjE5", "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "gravatar_id": "", "url": "https://api.github.com/users/patrickvonplaten", "html_url": "https://github.com/patrickvonplaten", "followers_url": "https://...
[ { "login": "patrickvonplaten", "id": 23423619, "node_id": "MDQ6VXNlcjIzNDIzNjE5", "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "gravatar_id": "", "url": "https://api.github.com/users/patrickvonplaten", "html_url": "https://github.com/patrickvonplaten", "follo...
null
[ "Amazing! ", "Small poll @thomwolf @yjernite @lhoestq @JetRunner @qiweizhen .\r\n\r\nAs stated in the XGLUE paper: https://arxiv.org/pdf/2004.01401.pdf , for each of the 11 down-stream tasks training data is only available in English, whereas development and test data is available in multiple different language ...
1,603,277,496,000
1,609,927,376,000
1,609,927,375,000
MEMBER
null
null
XGLUE is a multilingual GLUE like dataset propesed in this [paper](https://arxiv.org/pdf/2004.01401.pdf). I'm planning on adding the dataset to the library myself in a couple of weeks. Also tagging @JetRunner @qiweizhen in case I need some guidance
https://api.github.com/repos/huggingface/datasets/issues/749/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/748
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/748/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/748/comments
https://api.github.com/repos/huggingface/datasets/issues/748/events
https://github.com/huggingface/datasets/pull/748
726,196,589
MDExOlB1bGxSZXF1ZXN0NTA3MzAyNjE3
748
New version of CompGuessWhat?! with refined annotations
{ "login": "aleSuglia", "id": 1479733, "node_id": "MDQ6VXNlcjE0Nzk3MzM=", "avatar_url": "https://avatars.githubusercontent.com/u/1479733?v=4", "gravatar_id": "", "url": "https://api.github.com/users/aleSuglia", "html_url": "https://github.com/aleSuglia", "followers_url": "https://api.github.com/users/al...
[]
closed
false
null
[]
null
[ "No worries. Always happy to help and thanks for your support in fixing the issue :)" ]
1,603,263,341,000
1,603,270,362,000
1,603,269,979,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/748", "html_url": "https://github.com/huggingface/datasets/pull/748", "diff_url": "https://github.com/huggingface/datasets/pull/748.diff", "patch_url": "https://github.com/huggingface/datasets/pull/748.patch" }
This pull request introduces a few fixes to the annotations for VisualGenome in the CompGuessWhat?! original split.
https://api.github.com/repos/huggingface/datasets/issues/748/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/747
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/747/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/747/comments
https://api.github.com/repos/huggingface/datasets/issues/747/events
https://github.com/huggingface/datasets/pull/747
725,884,704
MDExOlB1bGxSZXF1ZXN0NTA3MDQ3MDE4
747
Add Quail question answering dataset
{ "login": "sai-prasanna", "id": 3595526, "node_id": "MDQ6VXNlcjM1OTU1MjY=", "avatar_url": "https://avatars.githubusercontent.com/u/3595526?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sai-prasanna", "html_url": "https://github.com/sai-prasanna", "followers_url": "https://api.github.com...
[]
closed
false
null
[]
null
[]
1,603,222,394,000
1,603,269,315,000
1,603,269,315,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/747", "html_url": "https://github.com/huggingface/datasets/pull/747", "diff_url": "https://github.com/huggingface/datasets/pull/747.diff", "patch_url": "https://github.com/huggingface/datasets/pull/747.patch" }
QuAIL is a multi-domain RC dataset featuring news, blogs, fiction and user stories. Each domain is represented by 200 texts, which gives us a 4-way data split. The texts are 300-350 word excerpts from CC-licensed texts that were hand-picked so as to make sense to human readers without larger context. Domain diversit...
https://api.github.com/repos/huggingface/datasets/issues/747/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/746
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/746/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/746/comments
https://api.github.com/repos/huggingface/datasets/issues/746/events
https://github.com/huggingface/datasets/pull/746
725,627,235
MDExOlB1bGxSZXF1ZXN0NTA2ODMzNDMw
746
dataset(ngt): add ngt dataset initial loading script
{ "login": "AmitMY", "id": 5757359, "node_id": "MDQ6VXNlcjU3NTczNTk=", "avatar_url": "https://avatars.githubusercontent.com/u/5757359?v=4", "gravatar_id": "", "url": "https://api.github.com/users/AmitMY", "html_url": "https://github.com/AmitMY", "followers_url": "https://api.github.com/users/AmitMY/foll...
[]
closed
false
null
[]
null
[]
1,603,202,698,000
1,616,480,378,000
1,616,480,378,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/746", "html_url": "https://github.com/huggingface/datasets/pull/746", "diff_url": "https://github.com/huggingface/datasets/pull/746.diff", "patch_url": "https://github.com/huggingface/datasets/pull/746.patch" }
Currently only making the paths to the annotation ELAN (eaf) file and videos available. This is the first accessible way to download this dataset, which is not manual file-by-file. Only downloading the necessary files, the annotation files are very small, 20MB for all of them, but the video files are large, 100GB i...
https://api.github.com/repos/huggingface/datasets/issues/746/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/745
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/745/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/745/comments
https://api.github.com/repos/huggingface/datasets/issues/745/events
https://github.com/huggingface/datasets/pull/745
725,589,352
MDExOlB1bGxSZXF1ZXN0NTA2ODAxMTI0
745
Fix emotion description
{ "login": "lewtun", "id": 26859204, "node_id": "MDQ6VXNlcjI2ODU5MjA0", "avatar_url": "https://avatars.githubusercontent.com/u/26859204?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lewtun", "html_url": "https://github.com/lewtun", "followers_url": "https://api.github.com/users/lewtun/fo...
[]
closed
false
null
[]
null
[ "Hello, I probably have a silly question but the labels of the emotion dataset are in the form of numbers and not string, so I can not use the function classification_report because it mixes numbers and string (prediction). How can I access the label in the form of a string and not a number? \r\nThank you in advanc...
1,603,200,519,000
1,619,102,851,000
1,603,269,507,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/745", "html_url": "https://github.com/huggingface/datasets/pull/745", "diff_url": "https://github.com/huggingface/datasets/pull/745.diff", "patch_url": "https://github.com/huggingface/datasets/pull/745.patch" }
Fixes the description of the emotion dataset to reflect the class names observed in the data, not the ones described in the paper. I also took the liberty to make use of `ClassLabel` for the emotion labels.
https://api.github.com/repos/huggingface/datasets/issues/745/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/744
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/744/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/744/comments
https://api.github.com/repos/huggingface/datasets/issues/744/events
https://github.com/huggingface/datasets/issues/744
724,918,448
MDU6SXNzdWU3MjQ5MTg0NDg=
744
Dataset Explorer Doesn't Work for squad_es and squad_it
{ "login": "gaotongxiao", "id": 22607038, "node_id": "MDQ6VXNlcjIyNjA3MDM4", "avatar_url": "https://avatars.githubusercontent.com/u/22607038?v=4", "gravatar_id": "", "url": "https://api.github.com/users/gaotongxiao", "html_url": "https://github.com/gaotongxiao", "followers_url": "https://api.github.com/...
[ { "id": 2107841032, "node_id": "MDU6TGFiZWwyMTA3ODQxMDMy", "url": "https://api.github.com/repos/huggingface/datasets/labels/nlp-viewer", "name": "nlp-viewer", "color": "94203D", "default": false, "description": "" } ]
closed
false
null
[]
null
[ "Oups wrong click.\r\nThis one is for you @srush" ]
1,603,136,052,000
1,603,730,177,000
1,603,730,177,000
NONE
null
null
https://huggingface.co/nlp/viewer/?dataset=squad_es https://huggingface.co/nlp/viewer/?dataset=squad_it Both pages show "OSError: [Errno 28] No space left on device".
https://api.github.com/repos/huggingface/datasets/issues/744/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/743
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/743/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/743/comments
https://api.github.com/repos/huggingface/datasets/issues/743/events
https://github.com/huggingface/datasets/issues/743
724,703,980
MDU6SXNzdWU3MjQ3MDM5ODA=
743
load_dataset for CSV files not working
{ "login": "iliemihai", "id": 2815308, "node_id": "MDQ6VXNlcjI4MTUzMDg=", "avatar_url": "https://avatars.githubusercontent.com/u/2815308?v=4", "gravatar_id": "", "url": "https://api.github.com/users/iliemihai", "html_url": "https://github.com/iliemihai", "followers_url": "https://api.github.com/users/il...
[]
open
false
null
[]
null
[ "Thank you !\r\nCould you provide a csv file that reproduces the error ?\r\nIt doesn't have to be one of your dataset. As long as it reproduces the error\r\nThat would help a lot !", "I think another good example is the following:\r\n`\r\nfrom datasets import load_dataset\r\n`\r\n`\r\ndataset = load_dataset(\"csv...
1,603,119,231,000
1,631,212,006,000
null
CONTRIBUTOR
null
null
Similar to #622, I've noticed there is a problem when trying to load a CSV file with datasets. ` from datasets import load_dataset ` ` dataset = load_dataset("csv", data_files=["./sample_data.csv"], delimiter="\t", column_names=["title", "text"], script_version="master") ` Displayed error: ` ... ArrowInva...
https://api.github.com/repos/huggingface/datasets/issues/743/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/742
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/742/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/742/comments
https://api.github.com/repos/huggingface/datasets/issues/742/events
https://github.com/huggingface/datasets/pull/742
724,509,974
MDExOlB1bGxSZXF1ZXN0NTA1ODgzNjI3
742
Add OCNLI, a new CLUE dataset
{ "login": "JetRunner", "id": 22514219, "node_id": "MDQ6VXNlcjIyNTE0MjE5", "avatar_url": "https://avatars.githubusercontent.com/u/22514219?v=4", "gravatar_id": "", "url": "https://api.github.com/users/JetRunner", "html_url": "https://github.com/JetRunner", "followers_url": "https://api.github.com/users/...
[]
closed
false
null
[]
null
[ "Thanks :) merging it" ]
1,603,105,593,000
1,603,383,589,000
1,603,383,588,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/742", "html_url": "https://github.com/huggingface/datasets/pull/742", "diff_url": "https://github.com/huggingface/datasets/pull/742.diff", "patch_url": "https://github.com/huggingface/datasets/pull/742.patch" }
OCNLI stands for Original Chinese Natural Language Inference. It is a corpus for Chinese Natural Language Inference, collected following closely the procedures of MNLI, but with enhanced strategies aiming for more challenging inference pairs. We want to emphasize we did not use hu...
https://api.github.com/repos/huggingface/datasets/issues/742/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/741
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/741/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/741/comments
https://api.github.com/repos/huggingface/datasets/issues/741/events
https://github.com/huggingface/datasets/issues/741
723,924,275
MDU6SXNzdWU3MjM5MjQyNzU=
741
Creating dataset consumes too much memory
{ "login": "AmitMY", "id": 5757359, "node_id": "MDQ6VXNlcjU3NTczNTk=", "avatar_url": "https://avatars.githubusercontent.com/u/5757359?v=4", "gravatar_id": "", "url": "https://api.github.com/users/AmitMY", "html_url": "https://github.com/AmitMY", "followers_url": "https://api.github.com/users/AmitMY/foll...
[]
open
false
null
[]
null
[ "Thanks for reporting.\r\nIn theory since the dataset script is just made to yield examples to write them into an arrow file, it's not supposed to create memory issues.\r\n\r\nCould you please try to run this exact same loop in a separate script to see if it's not an issue with `PIL` ?\r\nYou can just copy paste wh...
1,603,001,226,000
1,617,097,628,000
null
CONTRIBUTOR
null
null
Moving this issue from https://github.com/huggingface/datasets/pull/722 here, because it seems like a general issue. Given the following dataset example, where each example saves a sequence of 260x210x3 images (max length 400): ```python def _generate_examples(self, base_path, split): """ Yields examp...
https://api.github.com/repos/huggingface/datasets/issues/741/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/740
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/740/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/740/comments
https://api.github.com/repos/huggingface/datasets/issues/740/events
https://github.com/huggingface/datasets/pull/740
723,047,958
MDExOlB1bGxSZXF1ZXN0NTA0NzAyNTc0
740
Fix TREC urls
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,602,839,488,000
1,603,097,677,000
1,603,097,676,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/740", "html_url": "https://github.com/huggingface/datasets/pull/740", "diff_url": "https://github.com/huggingface/datasets/pull/740.diff", "patch_url": "https://github.com/huggingface/datasets/pull/740.patch" }
The old TREC urls are now redirections. I updated the urls to the new ones, since we don't support redirections for downloads. Fix #737
https://api.github.com/repos/huggingface/datasets/issues/740/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/739
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/739/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/739/comments
https://api.github.com/repos/huggingface/datasets/issues/739/events
https://github.com/huggingface/datasets/pull/739
723,044,066
MDExOlB1bGxSZXF1ZXN0NTA0Njk5NTY3
739
Add wiki dpr multiset embeddings
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[ "I still have to compute the dataset_infos, and build + host the indexes", "update: I'm computing the metadata, will update the PR soon", "Finally all green and ready to merge :)" ]
1,602,839,149,000
1,606,399,370,000
1,606,399,369,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/739", "html_url": "https://github.com/huggingface/datasets/pull/739", "diff_url": "https://github.com/huggingface/datasets/pull/739.diff", "patch_url": "https://github.com/huggingface/datasets/pull/739.patch" }
There are two DPR encoders, one trained on Natural Questions and one trained on a multiset/hybrid dataset. Previously only the embeddings from the encoder trained on NQ were available. I'm adding the ones from the encoder trained on the multiset/hybrid dataset. In the configuration you can now specify `embeddings_nam...
https://api.github.com/repos/huggingface/datasets/issues/739/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/738
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/738/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/738/comments
https://api.github.com/repos/huggingface/datasets/issues/738/events
https://github.com/huggingface/datasets/pull/738
723,033,923
MDExOlB1bGxSZXF1ZXN0NTA0NjkxNjM4
738
Replace seqeval code with original classification_report for simplicity
{ "login": "Hironsan", "id": 6737785, "node_id": "MDQ6VXNlcjY3Mzc3ODU=", "avatar_url": "https://avatars.githubusercontent.com/u/6737785?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Hironsan", "html_url": "https://github.com/Hironsan", "followers_url": "https://api.github.com/users/Hiron...
[]
closed
false
null
[]
null
[ "Hello,\r\n\r\nI ran https://github.com/huggingface/transformers/blob/master/examples/token-classification/run.sh\r\n\r\nAnd received this error:\r\n```\r\n100%|██████████| 407/407 [21:37<00:00, 3.44s/it]Traceback (most recent call last):\r\n File \"run_ner.py\", line 445, in <module>\r\n main()\r\n File \"ru...
1,602,838,305,000
1,611,245,235,000
1,603,103,472,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/738", "html_url": "https://github.com/huggingface/datasets/pull/738", "diff_url": "https://github.com/huggingface/datasets/pull/738.diff", "patch_url": "https://github.com/huggingface/datasets/pull/738.patch" }
Recently, the original seqeval has enabled us to get per type scores and overall scores as a dictionary. This PR replaces the current code with the original function(`classification_report`) to simplify it. Also, the original code has been updated to fix #352. - Related issue: https://github.com/chakki-works/seq...
https://api.github.com/repos/huggingface/datasets/issues/738/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/737
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/737/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/737/comments
https://api.github.com/repos/huggingface/datasets/issues/737/events
https://github.com/huggingface/datasets/issues/737
722,463,923
MDU6SXNzdWU3MjI0NjM5MjM=
737
Trec Dataset Connection Error
{ "login": "aychang95", "id": 10554495, "node_id": "MDQ6VXNlcjEwNTU0NDk1", "avatar_url": "https://avatars.githubusercontent.com/u/10554495?v=4", "gravatar_id": "", "url": "https://api.github.com/users/aychang95", "html_url": "https://github.com/aychang95", "followers_url": "https://api.github.com/users/...
[]
closed
false
null
[]
null
[ "Thanks for reporting.\r\nThat's because the download url has changed. The old url now redirects to the new one but we don't support redirection for downloads.\r\n\r\nI'm opening a PR to update the url" ]
1,602,777,473,000
1,603,097,676,000
1,603,097,676,000
NONE
null
null
**Datasets Version:** 1.1.2 **Python Version:** 3.6/3.7 **Code:** ```python from datasets import load_dataset load_dataset("trec") ``` **Expected behavior:** Download Trec dataset and load Dataset object **Current Behavior:** Get a connection error saying it couldn't reach http://cogcomp.org/Data/...
https://api.github.com/repos/huggingface/datasets/issues/737/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/736
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/736/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/736/comments
https://api.github.com/repos/huggingface/datasets/issues/736/events
https://github.com/huggingface/datasets/pull/736
722,348,191
MDExOlB1bGxSZXF1ZXN0NTA0MTE0MjMy
736
Start community-provided dataset docs
{ "login": "sshleifer", "id": 6045025, "node_id": "MDQ6VXNlcjYwNDUwMjU=", "avatar_url": "https://avatars.githubusercontent.com/u/6045025?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sshleifer", "html_url": "https://github.com/sshleifer", "followers_url": "https://api.github.com/users/ss...
[]
closed
false
null
[]
null
[ "can you also reference the `--organization` flag like in https://github.com/huggingface/transformers/blob/master/docs/source/model_sharing.rst#upload-your-model-with-the-cli ?", "done!", "Not sure if the changes in `datasets/wmt_t2t/wmt_utils.py` are intentional.\r\nIf you want to add more configs to wmt, coul...
1,602,769,299,000
1,603,458,928,000
1,603,458,928,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/736", "html_url": "https://github.com/huggingface/datasets/pull/736", "diff_url": "https://github.com/huggingface/datasets/pull/736.diff", "patch_url": "https://github.com/huggingface/datasets/pull/736.patch" }
This is one I did to get the pseudo-labels updated. Not sure if it generalizes, but I figured I would write it down. It was pretty easy because all I had to do was make properly formatted directories and change URLs. + In slack @thomwolf called it a `user-namespace` dataset, but the docs call it `community dataset`...
https://api.github.com/repos/huggingface/datasets/issues/736/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/735
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/735/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/735/comments
https://api.github.com/repos/huggingface/datasets/issues/735/events
https://github.com/huggingface/datasets/issues/735
722,225,270
MDU6SXNzdWU3MjIyMjUyNzA=
735
Throw error when an unexpected key is used in data_files
{ "login": "BramVanroy", "id": 2779410, "node_id": "MDQ6VXNlcjI3Nzk0MTA=", "avatar_url": "https://avatars.githubusercontent.com/u/2779410?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BramVanroy", "html_url": "https://github.com/BramVanroy", "followers_url": "https://api.github.com/users...
[]
closed
false
null
[]
null
[ "Thanks for reporting !\r\nWe'll add support for other keys" ]
1,602,759,327,000
1,604,064,232,000
1,604,064,232,000
CONTRIBUTOR
null
null
I have found that only "train", "validation" and "test" are valid keys in the `data_files` argument. When you use any other ones, those attached files are silently ignored - leading to unexpected behaviour for the users. So the following, unintuitively, returns only one key (namely `train`). ```python datasets =...
https://api.github.com/repos/huggingface/datasets/issues/735/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/734
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/734/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/734/comments
https://api.github.com/repos/huggingface/datasets/issues/734/events
https://github.com/huggingface/datasets/pull/734
721,767,848
MDExOlB1bGxSZXF1ZXN0NTAzNjMwMDcz
734
Fix GLUE metric description
{ "login": "sgugger", "id": 35901082, "node_id": "MDQ6VXNlcjM1OTAxMDgy", "avatar_url": "https://avatars.githubusercontent.com/u/35901082?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sgugger", "html_url": "https://github.com/sgugger", "followers_url": "https://api.github.com/users/sgugge...
[]
closed
false
null
[]
null
[]
1,602,708,254,000
1,602,754,063,000
1,602,754,062,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/734", "html_url": "https://github.com/huggingface/datasets/pull/734", "diff_url": "https://github.com/huggingface/datasets/pull/734.diff", "patch_url": "https://github.com/huggingface/datasets/pull/734.patch" }
Small typo: the description says translation instead of prediction.
https://api.github.com/repos/huggingface/datasets/issues/734/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/733
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/733/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/733/comments
https://api.github.com/repos/huggingface/datasets/issues/733/events
https://github.com/huggingface/datasets/pull/733
721,366,744
MDExOlB1bGxSZXF1ZXN0NTAzMjk2NDQw
733
Update link to dataset viewer
{ "login": "negedng", "id": 12969168, "node_id": "MDQ6VXNlcjEyOTY5MTY4", "avatar_url": "https://avatars.githubusercontent.com/u/12969168?v=4", "gravatar_id": "", "url": "https://api.github.com/users/negedng", "html_url": "https://github.com/negedng", "followers_url": "https://api.github.com/users/negedn...
[]
closed
false
null
[]
null
[]
1,602,674,003,000
1,602,684,451,000
1,602,684,451,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/733", "html_url": "https://github.com/huggingface/datasets/pull/733", "diff_url": "https://github.com/huggingface/datasets/pull/733.diff", "patch_url": "https://github.com/huggingface/datasets/pull/733.patch" }
Change 404 error links in quick tour to working ones
https://api.github.com/repos/huggingface/datasets/issues/733/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/732
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/732/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/732/comments
https://api.github.com/repos/huggingface/datasets/issues/732/events
https://github.com/huggingface/datasets/pull/732
721,359,448
MDExOlB1bGxSZXF1ZXN0NTAzMjkwMjEy
732
dataset(wlasl): initial loading script
{ "login": "AmitMY", "id": 5757359, "node_id": "MDQ6VXNlcjU3NTczNTk=", "avatar_url": "https://avatars.githubusercontent.com/u/5757359?v=4", "gravatar_id": "", "url": "https://api.github.com/users/AmitMY", "html_url": "https://github.com/AmitMY", "followers_url": "https://api.github.com/users/AmitMY/foll...
[]
closed
false
null
[]
null
[ "Followup: \r\nFrom the info in https://github.com/huggingface/datasets/pull/722, I probably should load the videos as array of frames directly into the database. \r\nThis will make the dataset generation time very long, but will make working with the dataset much easier.", "When I run:\r\n```\r\npython datasets-...
1,602,673,302,000
1,616,480,383,000
1,616,480,383,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/732", "html_url": "https://github.com/huggingface/datasets/pull/732", "diff_url": "https://github.com/huggingface/datasets/pull/732.diff", "patch_url": "https://github.com/huggingface/datasets/pull/732.patch" }
takes like 9-10 hours to download all of the videos for the dataset, but it does finish :)
https://api.github.com/repos/huggingface/datasets/issues/732/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/731
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/731/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/731/comments
https://api.github.com/repos/huggingface/datasets/issues/731/events
https://github.com/huggingface/datasets/pull/731
721,142,985
MDExOlB1bGxSZXF1ZXN0NTAzMTExNzc4
731
dataset(aslg_pc12): initial loading script
{ "login": "AmitMY", "id": 5757359, "node_id": "MDQ6VXNlcjU3NTczNTk=", "avatar_url": "https://avatars.githubusercontent.com/u/5757359?v=4", "gravatar_id": "", "url": "https://api.github.com/users/AmitMY", "html_url": "https://github.com/AmitMY", "followers_url": "https://api.github.com/users/AmitMY/foll...
[]
closed
false
null
[]
null
[ "Thanks @lhoestq \r\nAre there any guidelines for the dummy data?\r\nIn this particular case for example, the dataset fetches from two hardcoded URLs. \r\nDo I just `head -n 10` both files and zip them?\r\n\r\n", "> Thanks @lhoestq\r\n> Are there any guidelines for the dummy data?\r\n> In this particular case for...
1,602,652,477,000
1,603,898,826,000
1,603,898,826,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/731", "html_url": "https://github.com/huggingface/datasets/pull/731", "diff_url": "https://github.com/huggingface/datasets/pull/731.diff", "patch_url": "https://github.com/huggingface/datasets/pull/731.patch" }
This contains the only current public part of this corpus. The rest of the corpus is not yet been made public, but this sample is still being used by researchers.
https://api.github.com/repos/huggingface/datasets/issues/731/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/730
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/730/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/730/comments
https://api.github.com/repos/huggingface/datasets/issues/730/events
https://github.com/huggingface/datasets/issues/730
721,073,812
MDU6SXNzdWU3MjEwNzM4MTI=
730
Possible caching bug
{ "login": "ArneBinder", "id": 3375489, "node_id": "MDQ6VXNlcjMzNzU0ODk=", "avatar_url": "https://avatars.githubusercontent.com/u/3375489?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ArneBinder", "html_url": "https://github.com/ArneBinder", "followers_url": "https://api.github.com/users...
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
[ "Thanks for reporting. That's a bug indeed.\r\nApparently only the `data_files` parameter is taken into account right now in `DatasetBuilder._create_builder_config` but it should also be the case for `config_kwargs` (or at least the instantiated `builder_config`)", "Hi, does this bug be fixed? when I load JSON fi...
1,602,640,954,000
1,603,964,161,000
1,603,964,161,000
NONE
null
null
The following code with `test1.txt` containing just "🤗🤗🤗": ``` dataset = datasets.load_dataset('text', data_files=['test1.txt'], split="train", encoding="latin_1") print(dataset[0]) dataset = datasets.load_dataset('text', data_files=['test1.txt'], split="train", encoding="utf-8") print(dataset[0]) ``` produc...
https://api.github.com/repos/huggingface/datasets/issues/730/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/729
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/729/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/729/comments
https://api.github.com/repos/huggingface/datasets/issues/729/events
https://github.com/huggingface/datasets/issues/729
719,558,876
MDU6SXNzdWU3MTk1NTg4NzY=
729
Better error message when one forgets to call `add_batch` before `compute`
{ "login": "sgugger", "id": 35901082, "node_id": "MDQ6VXNlcjM1OTAxMDgy", "avatar_url": "https://avatars.githubusercontent.com/u/35901082?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sgugger", "html_url": "https://github.com/sgugger", "followers_url": "https://api.github.com/users/sgugge...
[]
closed
false
null
[]
null
[]
1,602,525,562,000
1,603,984,704,000
1,603,984,704,000
MEMBER
null
null
When using metrics, if for some reason a user forgets to call `add_batch` to a metric before `compute` (with no arguments), the error message is a bit cryptic and could probably be made clearer. ## Reproducer ```python import datasets import torch from datasets import Metric class GatherMetric(Metric): ...
https://api.github.com/repos/huggingface/datasets/issues/729/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/728
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/728/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/728/comments
https://api.github.com/repos/huggingface/datasets/issues/728/events
https://github.com/huggingface/datasets/issues/728
719,555,780
MDU6SXNzdWU3MTk1NTU3ODA=
728
Passing `cache_dir` to a metric does not work
{ "login": "sgugger", "id": 35901082, "node_id": "MDQ6VXNlcjM1OTAxMDgy", "avatar_url": "https://avatars.githubusercontent.com/u/35901082?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sgugger", "html_url": "https://github.com/sgugger", "followers_url": "https://api.github.com/users/sgugge...
[]
closed
false
null
[]
null
[]
1,602,525,314,000
1,603,964,082,000
1,603,964,082,000
MEMBER
null
null
When passing `cache_dir` to a custom metric, the folder is concatenated to itself at some point and this results in a FileNotFoundError: ## Reproducer ```python import datasets import torch from datasets import Metric class GatherMetric(Metric): def _info(self): return datasets.MetricInfo( ...
https://api.github.com/repos/huggingface/datasets/issues/728/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/727
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/727/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/727/comments
https://api.github.com/repos/huggingface/datasets/issues/727/events
https://github.com/huggingface/datasets/issues/727
719,386,366
MDU6SXNzdWU3MTkzODYzNjY=
727
Parallel downloads progress bar flickers
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
open
false
null
[]
null
[]
1,602,509,765,000
1,602,509,765,000
null
MEMBER
null
null
When there are parallel downloads using the download manager, the tqdm progress bar flickers since all the progress bars are on the same line. To fix that we could simply specify `position=i` for i=0 to n the number of files to download when instantiating the tqdm progress bar. Another way would be to have one "...
https://api.github.com/repos/huggingface/datasets/issues/727/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/726
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/726/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/726/comments
https://api.github.com/repos/huggingface/datasets/issues/726/events
https://github.com/huggingface/datasets/issues/726
719,313,754
MDU6SXNzdWU3MTkzMTM3NTQ=
726
"Checksums didn't match for dataset source files" error while loading openwebtext dataset
{ "login": "SparkJiao", "id": 16469472, "node_id": "MDQ6VXNlcjE2NDY5NDcy", "avatar_url": "https://avatars.githubusercontent.com/u/16469472?v=4", "gravatar_id": "", "url": "https://api.github.com/users/SparkJiao", "html_url": "https://github.com/SparkJiao", "followers_url": "https://api.github.com/users/...
[]
open
false
null
[]
null
[ "Hi try, to provide more information please.\r\n\r\nExample code in a colab to reproduce the error, details on what you are trying to do and what you were expected and details on your environment (OS, PyPi packages version).", "> Hi try, to provide more information please.\r\n> \r\n> Example code in a colab to re...
1,602,503,110,000
1,630,485,792,000
null
NONE
null
null
Hi, I have encountered this problem during loading the openwebtext dataset: ``` >>> dataset = load_dataset('openwebtext') Downloading and preparing dataset openwebtext/plain_text (download: 12.00 GiB, generated: 37.04 GiB, post-processed: Unknown size, total: 49.03 GiB) to /home/admin/.cache/huggingface/datasets/op...
https://api.github.com/repos/huggingface/datasets/issues/726/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/725
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/725/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/725/comments
https://api.github.com/repos/huggingface/datasets/issues/725/events
https://github.com/huggingface/datasets/pull/725
718,985,641
MDExOlB1bGxSZXF1ZXN0NTAxMjUxODI1
725
pretty print dataset objects
{ "login": "stas00", "id": 10676103, "node_id": "MDQ6VXNlcjEwNjc2MTAz", "avatar_url": "https://avatars.githubusercontent.com/u/10676103?v=4", "gravatar_id": "", "url": "https://api.github.com/users/stas00", "html_url": "https://github.com/stas00", "followers_url": "https://api.github.com/users/stas00/fo...
[]
closed
false
null
[]
null
[ "Great, as you found it useful I improved the code a bit to automate indentation in the parent class, so that the child repr doesn't need to guess the indentation level, while repr'ing nicely on its own.\r\n\r\n- do we want indent=4 or 2?\r\n- do we want `{` ... `}` or w/o?\r\n\r\ncurrently it's indent4 and w/ curl...
1,602,468,226,000
1,603,470,275,000
1,603,443,646,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/725", "html_url": "https://github.com/huggingface/datasets/pull/725", "diff_url": "https://github.com/huggingface/datasets/pull/725.diff", "patch_url": "https://github.com/huggingface/datasets/pull/725.patch" }
Currently, if I do: ``` from datasets import load_dataset load_dataset("wikihow", 'all', data_dir="/hf/pegasus-datasets/wikihow/") ``` I get: ``` DatasetDict({'train': Dataset(features: {'text': Value(dtype='string', id=None), 'headline': Value(dtype='string', id=None), 'title': Value(dtype='string', id=None...
https://api.github.com/repos/huggingface/datasets/issues/725/timeline
null
true