url
stringlengths
58
61
repository_url
stringclasses
1 value
labels_url
stringlengths
72
75
comments_url
stringlengths
67
70
events_url
stringlengths
65
68
html_url
stringlengths
46
51
id
int64
599M
1.83B
node_id
stringlengths
18
32
number
int64
1
6.09k
title
stringlengths
1
290
labels
list
state
stringclasses
2 values
locked
bool
1 class
milestone
dict
comments
int64
0
54
created_at
stringlengths
20
20
updated_at
stringlengths
20
20
closed_at
stringlengths
20
20
active_lock_reason
null
body
stringlengths
0
228k
reactions
dict
timeline_url
stringlengths
67
70
performed_via_github_app
null
state_reason
stringclasses
3 values
draft
bool
2 classes
pull_request
dict
is_pull_request
bool
2 classes
comments_text
list
https://api.github.com/repos/huggingface/datasets/issues/6028
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6028/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6028/comments
https://api.github.com/repos/huggingface/datasets/issues/6028/events
https://github.com/huggingface/datasets/pull/6028
1,803,294,981
PR_kwDODunzps5Vb3LJ
6,028
Use new hffs
[]
closed
false
null
13
2023-07-13T15:41:44Z
2023-07-17T17:09:39Z
2023-07-17T17:01:00Z
null
Thanks to @janineguo 's work in https://github.com/huggingface/datasets/pull/5919 which was needed to support HfFileSystem. Switching to `HfFileSystem` will help implementing optimization in data files resolution ## Implementation details I replaced all the from_hf_repo and from_local_or_remote in data_files.p...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6028/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6028/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/6028.diff", "html_url": "https://github.com/huggingface/datasets/pull/6028", "merged_at": "2023-07-17T17:01:00Z", "patch_url": "https://github.com/huggingface/datasets/pull/6028.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/...
true
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
https://api.github.com/repos/huggingface/datasets/issues/4551
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4551/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4551/comments
https://api.github.com/repos/huggingface/datasets/issues/4551/events
https://github.com/huggingface/datasets/pull/4551
1,282,534,807
PR_kwDODunzps46QAV-
4,551
Perform hidden file check on relative data file path
[]
closed
false
null
5
2022-06-23T14:49:11Z
2022-06-30T14:49:20Z
2022-06-30T14:38:18Z
null
Fix #4549
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4551/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4551/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4551.diff", "html_url": "https://github.com/huggingface/datasets/pull/4551", "merged_at": "2022-06-30T14:38:18Z", "patch_url": "https://github.com/huggingface/datasets/pull/4551.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/...
true
[ "_The documentation is not available anymore as the PR was closed or merged._", "I'm aware of this behavior, which is tricky to solve due to fsspec's hidden file handling (see https://github.com/huggingface/datasets/issues/4115#issuecomment-1108819538). I've tested some regex patterns to address this, and they se...
https://api.github.com/repos/huggingface/datasets/issues/1269
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1269/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1269/comments
https://api.github.com/repos/huggingface/datasets/issues/1269/events
https://github.com/huggingface/datasets/pull/1269
758,886,174
MDExOlB1bGxSZXF1ZXN0NTMzOTc3MTE2
1,269
Adding OneStopEnglish corpus dataset
[]
closed
false
null
1
2020-12-07T22:05:11Z
2020-12-09T18:43:38Z
2020-12-09T15:33:53Z
null
This PR adds OneStopEnglish Corpus containing texts classified into reading levels (elementary, intermediate, advance) for automatic readability assessment and text simplification. Link to the paper: https://www.aclweb.org/anthology/W18-0535.pdf
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/1269/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/1269/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/1269.diff", "html_url": "https://github.com/huggingface/datasets/pull/1269", "merged_at": "2020-12-09T15:33:53Z", "patch_url": "https://github.com/huggingface/datasets/pull/1269.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/...
true
[ "Hi @lhoestq, thanks for the review.\r\nI have made all the changes, PTAL! :) " ]
https://api.github.com/repos/huggingface/datasets/issues/3186
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/3186/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/3186/comments
https://api.github.com/repos/huggingface/datasets/issues/3186/events
https://github.com/huggingface/datasets/issues/3186
1,040,369,397
I_kwDODunzps4-Asb1
3,186
Dataset viewer for nli_tr
[ { "color": "fef2c0", "default": false, "description": "", "id": 3287858981, "name": "streaming", "node_id": "MDU6TGFiZWwzMjg3ODU4OTgx", "url": "https://api.github.com/repos/huggingface/datasets/labels/streaming" } ]
closed
false
null
6
2021-10-31T03:56:33Z
2022-09-12T09:15:34Z
2022-09-12T08:43:09Z
null
## Dataset viewer issue for '*nli_tr*' **Link:** https://huggingface.co/datasets/nli_tr Hello, Thank you for the new dataset preview feature that will help the users to view the datasets online. We just noticed that the dataset viewer widget in the `nli_tr` dataset shows the error below. The error must be d...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3186/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/3186/timeline
null
completed
null
null
false
[ "It's an issue with the streaming mode:\r\n\r\n```python\r\n>>> import datasets\r\n>>> dataset = datasets.load_dataset('nli_tr', name='snli_tr',split='test', streaming=True)\r\n>>> next(iter(dataset))\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"/home/slesage/hf/datas...
https://api.github.com/repos/huggingface/datasets/issues/4565
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4565/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4565/comments
https://api.github.com/repos/huggingface/datasets/issues/4565/events
https://github.com/huggingface/datasets/issues/4565
1,284,141,666
I_kwDODunzps5MinJi
4,565
Add UFSC OCPap dataset
[ { "color": "e99695", "default": false, "description": "Requesting to add a new dataset", "id": 2067376369, "name": "dataset request", "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request" } ]
closed
false
null
1
2022-06-24T20:07:54Z
2022-07-06T19:03:02Z
2022-07-06T19:03:02Z
null
## Adding a Dataset - **Name:** UFSC OCPap: Papanicolaou Stained Oral Cytology Dataset (v4) - **Description:** The UFSC OCPap dataset comprises 9,797 labeled images of 1200x1600 pixels acquired from 5 slides of cancer diagnosed and 3 healthy of oral brush samples, from distinct patients. - **Paper:** https://dx.doi....
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4565/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4565/timeline
null
completed
null
null
false
[ "I will add this directly on the hub (same as #4486)—in https://huggingface.co/lapix" ]
https://api.github.com/repos/huggingface/datasets/issues/5661
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5661/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5661/comments
https://api.github.com/repos/huggingface/datasets/issues/5661/events
https://github.com/huggingface/datasets/issues/5661
1,637,129,445
I_kwDODunzps5hlJzl
5,661
CI is broken: Unnecessary `dict` comprehension
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
null
0
2023-03-23T09:13:01Z
2023-03-23T09:37:51Z
2023-03-23T09:37:51Z
null
CI check_code_quality is broken: ``` src/datasets/arrow_dataset.py:3267:35: C416 [*] Unnecessary `dict` comprehension (rewrite using `dict()`) Found 1 error. ```
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5661/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/5661/timeline
null
completed
null
null
false
[]
https://api.github.com/repos/huggingface/datasets/issues/4543
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4543/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4543/comments
https://api.github.com/repos/huggingface/datasets/issues/4543/events
https://github.com/huggingface/datasets/pull/4543
1,280,379,781
PR_kwDODunzps46IiEp
4,543
[CI] Fix upstream hub test url
[]
closed
false
null
2
2022-06-22T15:34:27Z
2022-06-22T16:37:40Z
2022-06-22T16:27:37Z
null
Some tests were still using moon-stagign instead of hub-ci. I also updated the token to use one dedicated to `datasets`
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4543/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4543/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4543.diff", "html_url": "https://github.com/huggingface/datasets/pull/4543", "merged_at": "2022-06-22T16:27:37Z", "patch_url": "https://github.com/huggingface/datasets/pull/4543.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/...
true
[ "_The documentation is not available anymore as the PR was closed or merged._", "Remaining CI failures are unrelated to this fix, merging" ]
https://api.github.com/repos/huggingface/datasets/issues/858
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/858/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/858/comments
https://api.github.com/repos/huggingface/datasets/issues/858/events
https://github.com/huggingface/datasets/pull/858
743,904,516
MDExOlB1bGxSZXF1ZXN0NTIxNzE3ODQ4
858
Add SemEval-2010 task 8
[]
closed
false
null
1
2020-11-16T14:57:57Z
2020-11-26T17:28:55Z
2020-11-26T17:28:55Z
null
Hi, I don't know how to add dummy data, since I create the validation set out of the last 1000 examples of the train set. If you have a suggestion, I am happy to implement it. Cheers, Joel
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/858/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/858/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/858.diff", "html_url": "https://github.com/huggingface/datasets/pull/858", "merged_at": "2020-11-26T17:28:55Z", "patch_url": "https://github.com/huggingface/datasets/pull/858.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/858...
true
[ "Added dummy data and encoding to open(). Now everything should be fine, hopefully :)" ]
https://api.github.com/repos/huggingface/datasets/issues/3882
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/3882/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/3882/comments
https://api.github.com/repos/huggingface/datasets/issues/3882/events
https://github.com/huggingface/datasets/pull/3882
1,164,595,388
PR_kwDODunzps40NKz7
3,882
Image process doc
[ { "color": "0075ca", "default": true, "description": "Improvements or additions to documentation", "id": 1935892861, "name": "documentation", "node_id": "MDU6TGFiZWwxOTM1ODkyODYx", "url": "https://api.github.com/repos/huggingface/datasets/labels/documentation" } ]
closed
false
null
1
2022-03-10T00:32:10Z
2022-03-15T15:24:16Z
2022-03-15T15:24:09Z
null
This PR is a first draft of how to process image data. It adds: - Load an image dataset with `image` and `path` (adds tip about `decode=False` param to access the path and bytes, thanks to @mariosasko). - Load an image using the `ImageFolder` builder. I know there is an [example](https://huggingface.co/docs/dataset...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 2, "laugh": 0, "rocket": 0, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/3882/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/3882/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/3882.diff", "html_url": "https://github.com/huggingface/datasets/pull/3882", "merged_at": "2022-03-15T15:24:09Z", "patch_url": "https://github.com/huggingface/datasets/pull/3882.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/...
true
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_3882). All of your documentation changes will be reflected on that endpoint." ]
https://api.github.com/repos/huggingface/datasets/issues/1318
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1318/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1318/comments
https://api.github.com/repos/huggingface/datasets/issues/1318/events
https://github.com/huggingface/datasets/pull/1318
759,565,629
MDExOlB1bGxSZXF1ZXN0NTM0NTQ5NjE3
1,318
ethos first commit
[]
closed
false
null
3
2020-12-08T15:59:47Z
2020-12-10T14:45:57Z
2020-12-10T14:45:57Z
null
Ethos passed all the tests except from this one: RUN_SLOW=1 pytest tests/test_dataset_common.py::LocalDatasetTest::test_load_dataset_all_configs_<your-dataset-name> with this error: E OSError: Cannot find data file. E Original error: E [Errno 2] No such file or direct...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/1318/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/1318/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/1318.diff", "html_url": "https://github.com/huggingface/datasets/pull/1318", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/1318.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/1318" }
true
[ "> Nice thanks !\r\n> \r\n> I left a few comments\r\n> \r\n> Also it looks like this PR includes changes about other files than the ones for ethos\r\n> \r\n> Can you create another branch and another PR please ?\r\n\r\n@lhoestq Should I close this PR? The new one is the: #1453", "You can create another PR and clo...
https://api.github.com/repos/huggingface/datasets/issues/93
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/93/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/93/comments
https://api.github.com/repos/huggingface/datasets/issues/93/events
https://github.com/huggingface/datasets/pull/93
617,522,029
MDExOlB1bGxSZXF1ZXN0NDE3NDIxODUy
93
Cleanup notebooks and various fixes
[]
closed
false
null
0
2020-05-13T14:58:58Z
2020-05-13T15:01:48Z
2020-05-13T15:01:47Z
null
Fixes on dataset (more flexible) metrics (fix) and general clean ups
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/93/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/93/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/93.diff", "html_url": "https://github.com/huggingface/datasets/pull/93", "merged_at": "2020-05-13T15:01:47Z", "patch_url": "https://github.com/huggingface/datasets/pull/93.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/93" }
true
[]
https://api.github.com/repos/huggingface/datasets/issues/5942
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5942/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5942/comments
https://api.github.com/repos/huggingface/datasets/issues/5942/events
https://github.com/huggingface/datasets/pull/5942
1,752,021,681
PR_kwDODunzps5Su-V4
5,942
Pass datasets-cli additional args as kwargs to DatasetBuilder in `run_beam.py`
[]
open
false
null
0
2023-06-12T06:50:50Z
2023-06-30T09:15:00Z
null
null
Hi, Following this <https://discuss.huggingface.co/t/how-to-preprocess-a-wikipedia-dataset-using-dataflowrunner/41991/3>, here is a simple PR to pass any additional args to datasets-cli as kwargs in the DatasetBuilder in `run_beam.py`. I also took the liberty to add missing setup steps to the `beam.mdx` docs in o...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5942/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/5942/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/5942.diff", "html_url": "https://github.com/huggingface/datasets/pull/5942", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/5942.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5942" }
true
[]
https://api.github.com/repos/huggingface/datasets/issues/4476
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4476/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4476/comments
https://api.github.com/repos/huggingface/datasets/issues/4476/events
https://github.com/huggingface/datasets/issues/4476
1,267,987,499
I_kwDODunzps5Lk_Qr
4,476
`to_pandas` doesn't take into account format.
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
closed
false
null
4
2022-06-10T20:25:31Z
2022-06-15T17:41:41Z
2022-06-15T17:41:41Z
null
**Is your feature request related to a problem? Please describe.** I have a large dataset that I need to convert part of to pandas to do some further analysis. Calling `to_pandas` directly on it is expensive. So I thought I could simply select the columns that I want and then call `to_pandas`. **Describe the solu...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4476/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4476/timeline
null
completed
null
null
false
[ "Thanks for opening a discussion :)\r\n\r\nNote that you can use `.remove_columns(...)` to keep only the ones you're interested in before calling `.to_pandas()`", "Yes I can do that thank you!\r\n\r\nDo you think that conceptually my example should work? If not, I'm happy to close this issue. \r\n\r\nIf yes, I ca...
https://api.github.com/repos/huggingface/datasets/issues/3457
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/3457/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/3457/comments
https://api.github.com/repos/huggingface/datasets/issues/3457/events
https://github.com/huggingface/datasets/issues/3457
1,084,862,121
I_kwDODunzps5Aqa6p
3,457
Add CMU Graphics Lab Motion Capture dataset
[ { "color": "e99695", "default": false, "description": "Requesting to add a new dataset", "id": 2067376369, "name": "dataset request", "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request" }, { "color": "bfdadc",...
open
false
null
3
2021-12-20T14:34:39Z
2022-03-16T16:53:09Z
null
null
## Adding a Dataset - **Name:** CMU Graphics Lab Motion Capture database - **Description:** The database contains free motions which you can download and use. - **Data:** http://mocap.cs.cmu.edu/ - **Motivation:** Nice motion capture dataset Instructions to add a new dataset can be found [here](https://github.c...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3457/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/3457/timeline
null
null
null
null
false
[ "This dataset has files in ASF/AMC format. [ The skeleton file is the ASF file (Acclaim Skeleton File). The motion file is the AMC file (Acclaim Motion Capture data). ] \r\n\r\nSome questions : \r\n1. How do we go about representing these features using datasets.Features and generate examples ?\r\n2. The dataset d...
https://api.github.com/repos/huggingface/datasets/issues/2313
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2313/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2313/comments
https://api.github.com/repos/huggingface/datasets/issues/2313/events
https://github.com/huggingface/datasets/pull/2313
875,475,367
MDExOlB1bGxSZXF1ZXN0NjI5ODEwNTc4
2,313
Remove unused head_hf_s3 function
[]
closed
false
null
0
2021-05-04T13:42:06Z
2021-05-07T09:31:42Z
2021-05-07T09:31:42Z
null
Currently, the function `head_hf_s3` is not used: - neither its returned result is used - nor it raises any exception, as exceptions are catched and returned (not raised) This PR removes it.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2313/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/2313/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/2313.diff", "html_url": "https://github.com/huggingface/datasets/pull/2313", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/2313.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2313" }
true
[]
https://api.github.com/repos/huggingface/datasets/issues/1933
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1933/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1933/comments
https://api.github.com/repos/huggingface/datasets/issues/1933/events
https://github.com/huggingface/datasets/pull/1933
814,335,846
MDExOlB1bGxSZXF1ZXN0NTc4MzQwMzk3
1,933
Use arrow ipc file format
[]
open
false
null
0
2021-02-23T10:38:24Z
2022-07-06T15:19:48Z
null
null
According to the [documentation](https://arrow.apache.org/docs/format/Columnar.html?highlight=arrow1#ipc-file-format), it's identical to the streaming format except that it contains the memory offsets of each sample: > We define a “file format” supporting random access that is build with the stream format. The file ...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/1933/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/1933/timeline
null
null
true
{ "diff_url": "https://github.com/huggingface/datasets/pull/1933.diff", "html_url": "https://github.com/huggingface/datasets/pull/1933", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/1933.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/1933" }
true
[]
https://api.github.com/repos/huggingface/datasets/issues/1906
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1906/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1906/comments
https://api.github.com/repos/huggingface/datasets/issues/1906/events
https://github.com/huggingface/datasets/issues/1906
811,405,274
MDU6SXNzdWU4MTE0MDUyNzQ=
1,906
Feature Request: Support for Pandas `Categorical`
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" }, { "color": "c5def5", "default": fals...
open
false
null
3
2021-02-18T19:46:05Z
2021-02-23T14:38:50Z
null
null
``` from datasets import Dataset import pandas as pd import pyarrow df = pd.DataFrame(pd.Series(["a", "b", "c", "a"], dtype="category")) pyarrow.Table.from_pandas(df) Dataset.from_pandas(df) # Throws NotImplementedError # TODO(thom) this will need access to the dictionary as well (for labels). I.e. to the py_...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/1906/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/1906/timeline
null
null
null
null
false
[ "We already have a ClassLabel type that does this kind of mapping between the label ids (integers) and actual label values (strings).\r\n\r\nI wonder if actually we should use the DictionaryType from Arrow and the Categorical type from pandas for the `datasets` ClassLabel feature type.\r\nCurrently ClassLabel corre...
https://api.github.com/repos/huggingface/datasets/issues/4679
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4679/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4679/comments
https://api.github.com/repos/huggingface/datasets/issues/4679/events
https://github.com/huggingface/datasets/pull/4679
1,303,980,648
PR_kwDODunzps47XX67
4,679
Added method to remove excess nesting in a DatasetDict
[]
closed
false
null
11
2022-07-13T21:49:37Z
2022-07-21T15:55:26Z
2022-07-21T10:55:02Z
null
Added the ability for a DatasetDict to remove additional nested layers within its features to avoid conflicts when collating. It is meant to accompany [this PR](https://github.com/huggingface/transformers/pull/18119) to resolve the same issue [#15505](https://github.com/huggingface/transformers/issues/15505). @stas0...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4679/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4679/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4679.diff", "html_url": "https://github.com/huggingface/datasets/pull/4679", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/4679.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4679" }
true
[ "Hi ! I think the issue you linked is closed and suggests to use `remove_columns`.\r\n\r\nMoreover if you end up with a dataset with an unnecessarily nested data, please modify your processing functions to not output nested data, or use `map(..., batched=True)` if you function take batches as input", "Hi @lhoestq...
https://api.github.com/repos/huggingface/datasets/issues/5458
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5458/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5458/comments
https://api.github.com/repos/huggingface/datasets/issues/5458/events
https://github.com/huggingface/datasets/issues/5458
1,555,054,737
I_kwDODunzps5csECR
5,458
slice split while streaming
[]
closed
false
null
2
2023-01-24T14:08:17Z
2023-01-24T15:11:47Z
2023-01-24T15:11:47Z
null
### Describe the bug When using the `load_dataset` function with streaming set to True, slicing splits is apparently not supported. Did I miss this in the documentation? ### Steps to reproduce the bug `load_dataset("lhoestq/demo1",revision=None, streaming=True, split="train[:3]")` causes ValueError: Bad split:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5458/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/5458/timeline
null
completed
null
null
false
[ "Hi! Yes, that's correct. When `streaming` is `True`, only split names can be specified as `split`, and for slicing, you have to use `.skip`/`.take` instead.\r\n\r\nE.g. \r\n`load_dataset(\"lhoestq/demo1\",revision=None, streaming=True, split=\"train[:3]\")`\r\n\r\nrewritten with `.skip`/`.take`:\r\n`load_dataset(\...
https://api.github.com/repos/huggingface/datasets/issues/4895
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4895/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4895/comments
https://api.github.com/repos/huggingface/datasets/issues/4895/events
https://github.com/huggingface/datasets/issues/4895
1,350,798,527
I_kwDODunzps5Qg4y_
4,895
load_dataset method returns Unknown split "validation" even if this dir exists
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
null
17
2022-08-25T12:11:00Z
2022-10-06T17:49:28Z
2022-09-29T08:07:50Z
null
## Describe the bug The `datasets.load_dataset` returns a `ValueError: Unknown split "validation". Should be one of ['train', 'test'].` when running `load_dataset(local_data_dir_path, split="validation")` even if the `validation` sub-directory exists in the local data path. The data directories are as follows and a...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4895/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4895/timeline
null
completed
null
null
false
[ "I don't know the main problem but it looks like, it is ignoring the last directory in your case. So, create a directory called 'zzz' in the same folder as train, validation and test. if it doesn't work, create a directory called \"aaa\". It worked for me.\r\n", "@SamSamhuns could you please try to load it with t...
https://api.github.com/repos/huggingface/datasets/issues/646
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/646/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/646/comments
https://api.github.com/repos/huggingface/datasets/issues/646/events
https://github.com/huggingface/datasets/pull/646
704,607,371
MDExOlB1bGxSZXF1ZXN0NDg5NTAyMTM3
646
Fix docs typos
[]
closed
false
null
0
2020-09-18T19:32:27Z
2020-09-21T16:30:54Z
2020-09-21T16:14:12Z
null
This PR fixes few typos in the docs and the error in the code snippet in the set_format section in docs/source/torch_tensorflow.rst. `torch.utils.data.Dataloader` expects padded batches so it throws an error due to not being able to stack the unpadded tensors. If we follow the Quick tour from the docs where they add th...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/646/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/646/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/646.diff", "html_url": "https://github.com/huggingface/datasets/pull/646", "merged_at": "2020-09-21T16:14:12Z", "patch_url": "https://github.com/huggingface/datasets/pull/646.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/646...
true
[]
https://api.github.com/repos/huggingface/datasets/issues/1393
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1393/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1393/comments
https://api.github.com/repos/huggingface/datasets/issues/1393/events
https://github.com/huggingface/datasets/pull/1393
760,436,267
MDExOlB1bGxSZXF1ZXN0NTM1MjY4MjUx
1,393
Add script_version suggestion when dataset/metric not found
[]
closed
false
null
0
2020-12-09T15:37:38Z
2020-12-10T18:17:05Z
2020-12-10T18:17:05Z
null
Adds a helpful prompt to the error message when a dataset/metric is not found, suggesting the user might need to pass `script_version="master"` if the dataset was added recently. The whole error looks like: > Couldn't find file locally at blah/blah.py, or remotely at https://raw.githubusercontent.com/huggingface/dat...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/1393/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/1393/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/1393.diff", "html_url": "https://github.com/huggingface/datasets/pull/1393", "merged_at": "2020-12-10T18:17:05Z", "patch_url": "https://github.com/huggingface/datasets/pull/1393.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/...
true
[]
https://api.github.com/repos/huggingface/datasets/issues/1701
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1701/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1701/comments
https://api.github.com/repos/huggingface/datasets/issues/1701/events
https://github.com/huggingface/datasets/issues/1701
781,345,717
MDU6SXNzdWU3ODEzNDU3MTc=
1,701
Some datasets miss dataset_infos.json or dummy_data.zip
[]
closed
false
null
2
2021-01-07T14:17:13Z
2022-11-04T15:11:16Z
2022-11-04T15:06:00Z
null
While working on dataset REAME generation script at https://github.com/madlag/datasets_readme_generator , I noticed that some datasets miss a dataset_infos.json : ``` c4 lm1b reclor wikihow ``` And some does not have a dummy_data.zip : ``` kor_nli math_dataset mlqa ms_marco newsgroup qa4mre qanga...
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/1701/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/1701/timeline
null
completed
null
null
false
[ "Thanks for reporting.\r\nWe should indeed add all the missing dummy_data.zip and also the dataset_infos.json at least for lm1b, reclor and wikihow.\r\n\r\nFor c4 I haven't tested the script and I think we'll require some optimizations regarding beam datasets before processing it.\r\n", "Closing since the dummy d...
https://api.github.com/repos/huggingface/datasets/issues/2319
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2319/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2319/comments
https://api.github.com/repos/huggingface/datasets/issues/2319/events
https://github.com/huggingface/datasets/issues/2319
876,251,376
MDU6SXNzdWU4NzYyNTEzNzY=
2,319
UnicodeDecodeError for OSCAR (Afrikaans)
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
null
3
2021-05-05T09:22:52Z
2021-05-05T10:57:31Z
2021-05-05T10:50:55Z
null
## Describe the bug When loading the [OSCAR dataset](https://huggingface.co/datasets/oscar) (specifically `unshuffled_deduplicated_af`), I encounter a `UnicodeDecodeError`. ## Steps to reproduce the bug ```python from datasets import load_dataset dataset = load_dataset("oscar", "unshuffled_deduplicated_af") ```...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2319/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/2319/timeline
null
completed
null
null
false
[ "Thanks for reporting, @sgraaf.\r\n\r\nI am going to have a look at it. \r\n\r\nI guess the expected codec is \"UTF-8\". Normally, when no explicitly codec is passed, Python uses one which is platform-dependent. For Linux machines, the default codec is `utf_8`, which is OK. However for Windows machine, the default ...
https://api.github.com/repos/huggingface/datasets/issues/4240
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4240/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4240/comments
https://api.github.com/repos/huggingface/datasets/issues/4240/events
https://github.com/huggingface/datasets/pull/4240
1,217,287,594
PR_kwDODunzps423xRl
4,240
Fix yield for crd3
[]
closed
false
null
2
2022-04-27T12:31:36Z
2022-04-29T12:41:41Z
2022-04-29T12:41:41Z
null
Modified the `_generate_examples` function to consider all the turns for a chunk id as a single example Modified the features accordingly ``` "turns": [ { "names": datasets.features.Sequence(datasets.Value("string")), "utterances": ...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4240/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4240/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4240.diff", "html_url": "https://github.com/huggingface/datasets/pull/4240", "merged_at": "2022-04-29T12:41:41Z", "patch_url": "https://github.com/huggingface/datasets/pull/4240.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/...
true
[ "I don't think you need to generate new dummy data, since they're in the same format as the original data.\r\n\r\nThe CI is failing because of this error:\r\n```python\r\n> turn[\"names\"] = turn[\"NAMES\"]\r\nE TypeError: tuple indices must be integers or slices, not str...
https://api.github.com/repos/huggingface/datasets/issues/4309
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4309/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4309/comments
https://api.github.com/repos/huggingface/datasets/issues/4309/events
https://github.com/huggingface/datasets/pull/4309
1,231,232,935
PR_kwDODunzps43lKpm
4,309
[WIP] Add TEDLIUM dataset
[ { "color": "e99695", "default": false, "description": "Requesting to add a new dataset", "id": 2067376369, "name": "dataset request", "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request" }, { "color": "d93f0b",...
closed
false
null
11
2022-05-10T14:12:47Z
2022-06-17T12:54:40Z
2022-06-17T11:44:01Z
null
Adds the TED-LIUM dataset https://www.tensorflow.org/datasets/catalog/tedlium#tedliumrelease3 TODO: - [x] Port `tedium.py` from TF datasets using `convert_dataset.sh` script - [x] Make `load_dataset` work - [ ] ~~Run `datasets-cli` command to generate `dataset_infos.json`~~ - [ ] ~~Create dummy data for conti...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4309/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4309/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4309.diff", "html_url": "https://github.com/huggingface/datasets/pull/4309", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/4309.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4309" }
true
[ "_The documentation is not available anymore as the PR was closed or merged._", "```python\r\nfrom datasets import load_dataset\r\n\r\nds = load_dataset('./datasets/tedlium', 'release1', cache_dir='/home/sanchitgandhi/cache')\r\n```\r\n\r\n```\r\nDownloading and preparing dataset tedlium/release1 to /home/sanchit...
https://api.github.com/repos/huggingface/datasets/issues/5243
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5243/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5243/comments
https://api.github.com/repos/huggingface/datasets/issues/5243/events
https://github.com/huggingface/datasets/issues/5243
1,449,523,962
I_kwDODunzps5WZfr6
5,243
Download only split data
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
open
false
null
5
2022-11-15T10:15:54Z
2023-05-02T09:27:51Z
null
null
### Feature request Is it possible to download only the data that I am requesting and not the entire dataset? I run out of disk spaceas it seems to download the entire dataset, instead of only the part needed. common_voice["test"] = load_dataset("mozilla-foundation/common_voice_11_0", "en", split="test", ...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5243/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/5243/timeline
null
null
null
null
false
[ "Hi @capsabogdan! Unfortunately, it's hard to implement because quite often datasets data is being hosted in a single archive for all splits :( So we have to download the whole archive to split it into splits. This is the case for CommonVoice too. \r\n\r\nHowever, for cases when data is distributed in separate arch...
https://api.github.com/repos/huggingface/datasets/issues/1015
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1015/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1015/comments
https://api.github.com/repos/huggingface/datasets/issues/1015/events
https://github.com/huggingface/datasets/pull/1015
755,508,841
MDExOlB1bGxSZXF1ZXN0NTMxMjA2MTgy
1,015
add hard dataset
[]
closed
false
null
1
2020-12-02T18:27:36Z
2020-12-03T15:03:54Z
2020-12-03T15:03:54Z
null
Hotel Reviews in Arabic language.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/1015/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/1015/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/1015.diff", "html_url": "https://github.com/huggingface/datasets/pull/1015", "merged_at": "2020-12-03T15:03:54Z", "patch_url": "https://github.com/huggingface/datasets/pull/1015.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/...
true
[ "Thanks @sumanthd17 that fixed it. " ]
https://api.github.com/repos/huggingface/datasets/issues/1898
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1898/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1898/comments
https://api.github.com/repos/huggingface/datasets/issues/1898/events
https://github.com/huggingface/datasets/issues/1898
810,157,251
MDU6SXNzdWU4MTAxNTcyNTE=
1,898
ALT dataset has repeating instances in all splits
[ { "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library", "id": 2067388877, "name": "dataset bug", "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug" } ]
closed
false
null
4
2021-02-17T12:51:42Z
2021-02-19T06:18:46Z
2021-02-19T06:18:46Z
null
The [ALT](https://huggingface.co/datasets/alt) dataset has all the same instances within each split :/ Seemed like a great dataset for some experiments I wanted to carry out, especially since its medium-sized, and has all splits. Would be great if this could be fixed :) Added a snapshot of the contents from `exp...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/1898/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/1898/timeline
null
completed
null
null
false
[ "Thanks for reporting. This looks like a very bad issue. I'm looking into it", "I just merged a fix, we'll do a patch release soon. Thanks again for reporting, and sorry for the inconvenience.\r\nIn the meantime you can load `ALT` using `datasets` from the master branch", "Thanks!!! works perfectly in the blead...
https://api.github.com/repos/huggingface/datasets/issues/4787
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4787/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4787/comments
https://api.github.com/repos/huggingface/datasets/issues/4787/events
https://github.com/huggingface/datasets/issues/4787
1,328,243,911
I_kwDODunzps5PK2TH
4,787
NonMatchingChecksumError in mbpp dataset
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
null
0
2022-08-04T08:15:51Z
2022-08-04T17:21:01Z
2022-08-04T17:21:01Z
null
## Describe the bug As reported on the Hub [Fix Checksum Mismatch](https://huggingface.co/datasets/mbpp/discussions/1), there is a `NonMatchingChecksumError` when loading mbpp dataset ## Steps to reproduce the bug ```python ds = load_dataset("mbpp", "full") ``` ## Expected results Loading of the dataset with...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4787/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4787/timeline
null
completed
null
null
false
[]
https://api.github.com/repos/huggingface/datasets/issues/4014
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4014/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4014/comments
https://api.github.com/repos/huggingface/datasets/issues/4014/events
https://github.com/huggingface/datasets/pull/4014
1,180,481,229
PR_kwDODunzps41AGBu
4,014
Support streaming id_clickbait dataset
[]
closed
false
null
1
2022-03-25T08:18:28Z
2022-03-25T08:58:31Z
2022-03-25T08:53:32Z
null
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4014/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4014/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4014.diff", "html_url": "https://github.com/huggingface/datasets/pull/4014", "merged_at": "2022-03-25T08:53:32Z", "patch_url": "https://github.com/huggingface/datasets/pull/4014.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/...
true
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
https://api.github.com/repos/huggingface/datasets/issues/2369
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2369/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2369/comments
https://api.github.com/repos/huggingface/datasets/issues/2369/events
https://github.com/huggingface/datasets/pull/2369
893,554,153
MDExOlB1bGxSZXF1ZXN0NjQ2MDQ5NDM1
2,369
correct labels of conll2003
[]
closed
false
null
0
2021-05-17T17:37:54Z
2021-05-18T08:27:42Z
2021-05-18T08:27:42Z
null
# What does this PR It fixes/extends the `ner_tags` for conll2003 to include all. Paper reference https://arxiv.org/pdf/cs/0306050v1.pdf Model reference https://huggingface.co/elastic/distilbert-base-cased-finetuned-conll03-english/blob/main/config.json
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2369/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/2369/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/2369.diff", "html_url": "https://github.com/huggingface/datasets/pull/2369", "merged_at": "2021-05-18T08:27:42Z", "patch_url": "https://github.com/huggingface/datasets/pull/2369.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/...
true
[]
https://api.github.com/repos/huggingface/datasets/issues/2924
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2924/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2924/comments
https://api.github.com/repos/huggingface/datasets/issues/2924/events
https://github.com/huggingface/datasets/issues/2924
997,378,113
I_kwDODunzps47cshB
2,924
"File name too long" error for file locks
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
null
10
2021-09-15T18:16:50Z
2023-03-28T06:50:18Z
2021-10-29T09:42:24Z
null
## Describe the bug Getting the following error when calling `load_dataset("gar1t/test")`: ``` OSError: [Errno 36] File name too long: '<user>/.cache/huggingface/datasets/_home_garrett_.cache_huggingface_datasets_csv_test-7c856aea083a7043_0.0.0_9144e0a4e8435090117cea53e6c7537173ef2304525df4a077c435d8ee7828ff.inc...
{ "+1": 0, "-1": 0, "confused": 1, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/2924/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/2924/timeline
null
completed
null
null
false
[ "Hi, the filename here is less than 255\r\n```python\r\n>>> len(\"_home_garrett_.cache_huggingface_datasets_csv_test-7c856aea083a7043_0.0.0_9144e0a4e8435090117cea53e6c7537173ef2304525df4a077c435d8ee7828ff.incomplete.lock\")\r\n154\r\n```\r\nso not sure why it's considered too long for your filesystem.\r\n(also note...
https://api.github.com/repos/huggingface/datasets/issues/4254
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4254/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4254/comments
https://api.github.com/repos/huggingface/datasets/issues/4254/events
https://github.com/huggingface/datasets/pull/4254
1,220,204,395
PR_kwDODunzps43Bwnj
4,254
Replace data URL in SAMSum dataset and support streaming
[]
closed
false
null
1
2022-04-29T08:21:43Z
2022-05-06T08:38:16Z
2022-04-29T16:26:09Z
null
This PR replaces data URL in SAMSum dataset: - original host (arxiv.org) does not allow HTTP Range requests - we have hosted the data on the Hub (license: CC BY-NC-ND 4.0) Moreover, it implements support for streaming. Fix #4146. Related to: #4236. CC: @severo
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/4254/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4254/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4254.diff", "html_url": "https://github.com/huggingface/datasets/pull/4254", "merged_at": "2022-04-29T16:26:08Z", "patch_url": "https://github.com/huggingface/datasets/pull/4254.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/...
true
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
https://api.github.com/repos/huggingface/datasets/issues/4935
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4935/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4935/comments
https://api.github.com/repos/huggingface/datasets/issues/4935/events
https://github.com/huggingface/datasets/issues/4935
1,363,226,736
I_kwDODunzps5RQTBw
4,935
Dataset Viewer issue for ubuntu_dialogs_corpus
[ { "color": "E5583E", "default": false, "description": "Related to the dataset viewer on huggingface.co", "id": 3470211881, "name": "dataset-viewer", "node_id": "LA_kwDODunzps7O1zsp", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset-viewer" } ]
closed
false
null
1
2022-09-06T12:41:50Z
2022-09-06T12:51:25Z
2022-09-06T12:51:25Z
null
### Link _No response_ ### Description _No response_ ### Owner _No response_
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4935/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4935/timeline
null
completed
null
null
false
[ "The dataset maintainers (https://huggingface.co/datasets/ubuntu_dialogs_corpus) decided to forbid the dataset from being downloaded automatically (https://huggingface.co/docs/datasets/v2.4.0/en/loading#manual-download), and the dataset viewer respects this.\r\nWe will try to improve the error display though. Thank...
https://api.github.com/repos/huggingface/datasets/issues/3785
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/3785/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/3785/comments
https://api.github.com/repos/huggingface/datasets/issues/3785/events
https://github.com/huggingface/datasets/pull/3785
1,150,069,801
PR_kwDODunzps4zciES
3,785
Fix: Bypass Virus Checks in Google Drive Links (CNN-DM dataset)
[]
closed
false
null
8
2022-02-25T05:48:57Z
2022-03-03T16:43:47Z
2022-03-03T14:03:37Z
null
This commit fixes the issue described in #3784. By adding an extra parameter to the end of Google Drive links, we are able to bypass the virus check and download the datasets. So, if the original link looked like https://drive.google.com/uc?export=download&id=0BwmD_VLjROrfTHk4NFg2SndKcjQ The new link now looks li...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3785/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/3785/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/3785.diff", "html_url": "https://github.com/huggingface/datasets/pull/3785", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/3785.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/3785" }
true
[ "Thank you, @albertvillanova!", "Got it. Thanks for explaining this, @albertvillanova!\r\n\r\n> On the other hand, the tests are not passing because the dummy data should also be fixed. Once done, this PR will be able to be merged into master.\r\n\r\nWill do this 👍", "Hi ! I think we need to fix the issue for ...
https://api.github.com/repos/huggingface/datasets/issues/2630
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2630/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2630/comments
https://api.github.com/repos/huggingface/datasets/issues/2630/events
https://github.com/huggingface/datasets/issues/2630
942,102,956
MDU6SXNzdWU5NDIxMDI5NTY=
2,630
Progress bars are not properly rendered in Jupyter notebook
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
null
2
2021-07-12T14:07:13Z
2022-02-03T15:55:33Z
2022-02-03T15:55:33Z
null
## Describe the bug The progress bars are not Jupyter widgets; regular progress bars appear (like in a terminal). ## Steps to reproduce the bug ```python ds.map(tokenize, num_proc=10) ``` ## Expected results Jupyter widgets displaying the progress bars. ## Actual results Simple plane progress bars. cc...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2630/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/2630/timeline
null
completed
null
null
false
[ "To add my experience when trying to debug this issue:\r\n\r\nSeems like previously the workaround given [here](https://github.com/tqdm/tqdm/issues/485#issuecomment-473338308) worked around this issue. But with the latest version of jupyter/tqdm I still get terminal warnings that IPython tried to send a message fro...
https://api.github.com/repos/huggingface/datasets/issues/4083
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4083/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4083/comments
https://api.github.com/repos/huggingface/datasets/issues/4083/events
https://github.com/huggingface/datasets/pull/4083
1,190,025,878
PR_kwDODunzps41gEbu
4,083
Add SacreBLEU Metric Card
[]
closed
false
null
1
2022-04-01T16:24:56Z
2022-04-12T20:45:00Z
2022-04-12T20:38:40Z
null
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4083/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4083/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4083.diff", "html_url": "https://github.com/huggingface/datasets/pull/4083", "merged_at": "2022-04-12T20:38:40Z", "patch_url": "https://github.com/huggingface/datasets/pull/4083.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/...
true
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
https://api.github.com/repos/huggingface/datasets/issues/4853
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4853/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4853/comments
https://api.github.com/repos/huggingface/datasets/issues/4853/events
https://github.com/huggingface/datasets/pull/4853
1,339,456,490
PR_kwDODunzps49NFNL
4,853
Fix bug and checksums in exams dataset
[]
closed
false
null
1
2022-08-15T20:17:57Z
2022-08-16T06:43:57Z
2022-08-16T06:29:06Z
null
Fix #4852.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/4853/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4853/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4853.diff", "html_url": "https://github.com/huggingface/datasets/pull/4853", "merged_at": "2022-08-16T06:29:06Z", "patch_url": "https://github.com/huggingface/datasets/pull/4853.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/...
true
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
https://api.github.com/repos/huggingface/datasets/issues/2075
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2075/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2075/comments
https://api.github.com/repos/huggingface/datasets/issues/2075/events
https://github.com/huggingface/datasets/issues/2075
834,301,246
MDU6SXNzdWU4MzQzMDEyNDY=
2,075
ConnectionError: Couldn't reach common_voice.py
[]
closed
false
null
2
2021-03-18T01:19:06Z
2021-03-20T10:29:41Z
2021-03-20T10:29:41Z
null
When I run: from datasets import load_dataset, load_metric common_voice_train = load_dataset("common_voice", "zh-CN", split="train+validation") common_voice_test = load_dataset("common_voice", "zh-CN", split="test") Got: ConnectionError: Couldn't reach https://raw.githubusercontent.com/huggingface/datasets/ma...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2075/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/2075/timeline
null
completed
null
null
false
[ "Hi @LifaSun, thanks for reporting this issue.\r\n\r\nSometimes, GitHub has some connectivity problems. Could you confirm that the problem persists?", "@albertvillanova Thanks! It works well now. " ]
https://api.github.com/repos/huggingface/datasets/issues/1421
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1421/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1421/comments
https://api.github.com/repos/huggingface/datasets/issues/1421/events
https://github.com/huggingface/datasets/pull/1421
760,706,851
MDExOlB1bGxSZXF1ZXN0NTM1NDkzMzU4
1,421
adding fake-news-english-2
[]
closed
false
null
0
2020-12-09T22:05:13Z
2020-12-13T00:48:49Z
2020-12-13T00:48:49Z
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/1421/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/1421/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/1421.diff", "html_url": "https://github.com/huggingface/datasets/pull/1421", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/1421.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/1421" }
true
[]
https://api.github.com/repos/huggingface/datasets/issues/4092
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4092/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4092/comments
https://api.github.com/repos/huggingface/datasets/issues/4092/events
https://github.com/huggingface/datasets/pull/4092
1,192,499,903
PR_kwDODunzps41n40R
4,092
Fix dataset `amazon_us_reviews` metadata - 4/4/2022
[]
closed
false
null
2
2022-04-05T01:39:45Z
2022-04-08T12:35:41Z
2022-04-08T12:29:31Z
null
Fixes #4048 by running `dataset-cli test` to reprocess data and regenerate metadata. Additionally I've updated the README to include up-to-date counts for the subsets.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4092/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4092/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4092.diff", "html_url": "https://github.com/huggingface/datasets/pull/4092", "merged_at": "2022-04-08T12:29:30Z", "patch_url": "https://github.com/huggingface/datasets/pull/4092.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/...
true
[ "_The documentation is not available anymore as the PR was closed or merged._", "cc: @albertvillanova just FYI" ]
https://api.github.com/repos/huggingface/datasets/issues/1411
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1411/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1411/comments
https://api.github.com/repos/huggingface/datasets/issues/1411/events
https://github.com/huggingface/datasets/pull/1411
760,606,290
MDExOlB1bGxSZXF1ZXN0NTM1NDEwNjU3
1,411
2 typos
[]
closed
false
null
0
2020-12-09T19:24:34Z
2020-12-11T10:39:05Z
2020-12-11T10:39:05Z
null
Corrected 2 typos
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/1411/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/1411/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/1411.diff", "html_url": "https://github.com/huggingface/datasets/pull/1411", "merged_at": "2020-12-11T10:39:05Z", "patch_url": "https://github.com/huggingface/datasets/pull/1411.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/...
true
[]
https://api.github.com/repos/huggingface/datasets/issues/4453
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4453/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4453/comments
https://api.github.com/repos/huggingface/datasets/issues/4453/events
https://github.com/huggingface/datasets/issues/4453
1,262,674,105
I_kwDODunzps5LQuC5
4,453
Dataset Viewer issue for Yaxin/SemEval2015
[]
closed
false
null
3
2022-06-07T03:30:08Z
2022-06-09T08:34:16Z
2022-06-09T08:34:16Z
null
### Link _No response_ ### Description _No response_ ### Owner _No response_
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4453/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4453/timeline
null
completed
null
null
false
[ "I understand that the issue is that a remote file (URL) is being loaded as a local file. Right @albertvillanova @lhoestq?\r\n\r\n```\r\nMessage: [Errno 2] No such file or directory: 'https://raw.githubusercontent.com/YaxinCui/ABSADataset/main/SemEval2015Task12Corrected/train/restaurants_train.xml'\r\n```", ...
https://api.github.com/repos/huggingface/datasets/issues/4460
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4460/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4460/comments
https://api.github.com/repos/huggingface/datasets/issues/4460/events
https://github.com/huggingface/datasets/pull/4460
1,264,644,205
PR_kwDODunzps45UHIs
4,460
Drop Python 3.6 support
[]
closed
false
null
5
2022-06-08T12:10:18Z
2022-07-26T19:16:39Z
2022-07-26T19:04:21Z
null
Remove the fallback imports/checks in the code needed for Python 3.6 and update the requirements/CI files. Also, use Python types for the NumPy dtype wherever possible to avoid deprecation warnings in newer NumPy versions.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 2, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/4460/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4460/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4460.diff", "html_url": "https://github.com/huggingface/datasets/pull/4460", "merged_at": "2022-07-26T19:04:21Z", "patch_url": "https://github.com/huggingface/datasets/pull/4460.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/...
true
[ "_The documentation is not available anymore as the PR was closed or merged._", "I've disabled the `test_dummy_dataset_serialize_s3` tests in the Linux CI to avoid the failures (these tests only fail on Windows in 3.6). These failures are unrelated to this PR's changes, and I would like to address this in a new P...
https://api.github.com/repos/huggingface/datasets/issues/1808
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1808/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1808/comments
https://api.github.com/repos/huggingface/datasets/issues/1808/events
https://github.com/huggingface/datasets/issues/1808
798,879,180
MDU6SXNzdWU3OTg4NzkxODA=
1,808
writing Datasets in a human readable format
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" }, { "color": "d876e3", "default": true...
closed
false
null
3
2021-02-02T02:55:40Z
2022-06-01T15:38:13Z
2022-06-01T15:38:13Z
null
Hi I see there is a save_to_disk function to save data, but this is not human readable format, is there a way I could save a Dataset object in a human readable format to a file like json? thanks @lhoestq
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/1808/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/1808/timeline
null
completed
null
null
false
[ "AFAIK, there is currently no built-in method on the `Dataset` object to do this.\r\nHowever, a workaround is to directly use the Arrow table backing the dataset, **but it implies loading the whole dataset in memory** (correct me if I'm mistaken @lhoestq).\r\n\r\nYou can convert the Arrow table to a pandas datafram...
https://api.github.com/repos/huggingface/datasets/issues/99
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/99/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/99/comments
https://api.github.com/repos/huggingface/datasets/issues/99/events
https://github.com/huggingface/datasets/pull/99
618,026,700
MDExOlB1bGxSZXF1ZXN0NDE3ODMxNjky
99
[Cmrc 2018] fix cmrc2018
[]
closed
false
null
0
2020-05-14T08:22:03Z
2020-05-14T08:49:42Z
2020-05-14T08:49:41Z
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/99/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/99/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/99.diff", "html_url": "https://github.com/huggingface/datasets/pull/99", "merged_at": "2020-05-14T08:49:41Z", "patch_url": "https://github.com/huggingface/datasets/pull/99.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/99" }
true
[]
https://api.github.com/repos/huggingface/datasets/issues/3427
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/3427/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/3427/comments
https://api.github.com/repos/huggingface/datasets/issues/3427/events
https://github.com/huggingface/datasets/pull/3427
1,078,782,159
PR_kwDODunzps4vxb_y
3,427
Add The Pile Enron Emails subset
[]
closed
false
null
0
2021-12-13T17:14:16Z
2021-12-14T17:30:59Z
2021-12-14T17:30:57Z
null
Add: - Enron Emails subset of The Pile: "enron_emails" config Close bigscience-workshop/data_tooling#310. CC: @StellaAthena
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3427/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/3427/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/3427.diff", "html_url": "https://github.com/huggingface/datasets/pull/3427", "merged_at": "2021-12-14T17:30:55Z", "patch_url": "https://github.com/huggingface/datasets/pull/3427.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/...
true
[]
https://api.github.com/repos/huggingface/datasets/issues/5714
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5714/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5714/comments
https://api.github.com/repos/huggingface/datasets/issues/5714/events
https://github.com/huggingface/datasets/pull/5714
1,657,388,033
PR_kwDODunzps5NxIOc
5,714
Fix xnumpy_load for .npz files
[]
closed
false
null
2
2023-04-06T13:01:45Z
2023-04-07T09:23:54Z
2023-04-07T09:16:57Z
null
PR: - #5626 implemented support for streaming `.npy` files by using `numpy.load`. However, it introduced a bug when used with `.npz` files, within a context manager: ``` ValueError: seek of closed file ``` or in streaming mode: ``` ValueError: I/O operation on closed file. ``` This PR fixes the bug an...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5714/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/5714/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/5714.diff", "html_url": "https://github.com/huggingface/datasets/pull/5714", "merged_at": "2023-04-07T09:16:57Z", "patch_url": "https://github.com/huggingface/datasets/pull/5714.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/...
true
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
https://api.github.com/repos/huggingface/datasets/issues/147
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/147/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/147/comments
https://api.github.com/repos/huggingface/datasets/issues/147/events
https://github.com/huggingface/datasets/issues/147
619,581,907
MDU6SXNzdWU2MTk1ODE5MDc=
147
Error with sklearn train_test_split
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
closed
false
null
2
2020-05-17T00:28:24Z
2020-06-18T16:23:23Z
2020-06-18T16:23:23Z
null
It would be nice if we could use sklearn `train_test_split` to quickly generate subsets from the dataset objects returned by `nlp.load_dataset`. At the moment the code: ```python data = nlp.load_dataset('imdb', cache_dir=data_cache) f_half, s_half = train_test_split(data['train'], test_size=0.5, random_state=seed)...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/147/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/147/timeline
null
completed
null
null
false
[ "Indeed. Probably we will want to have a similar method directly in the library", "Related: #166 " ]
https://api.github.com/repos/huggingface/datasets/issues/2122
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2122/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2122/comments
https://api.github.com/repos/huggingface/datasets/issues/2122/events
https://github.com/huggingface/datasets/pull/2122
842,194,588
MDExOlB1bGxSZXF1ZXN0NjAxODE3MjI0
2,122
Fast table queries with interpolation search
[]
closed
false
null
0
2021-03-26T18:09:20Z
2021-08-04T18:11:59Z
2021-04-06T14:33:01Z
null
## Intro This should fix issue #1803 Currently querying examples in a dataset is O(n) because of the underlying pyarrow ChunkedArrays implementation. To fix this I implemented interpolation search that is pretty effective since datasets usually verifies the condition of evenly distributed chunks (the default ch...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 5, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 5, "url": "https://api.github.com/repos/huggingface/datasets/issues/2122/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/2122/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/2122.diff", "html_url": "https://github.com/huggingface/datasets/pull/2122", "merged_at": "2021-04-06T14:33:01Z", "patch_url": "https://github.com/huggingface/datasets/pull/2122.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/...
true
[]
https://api.github.com/repos/huggingface/datasets/issues/2939
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2939/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2939/comments
https://api.github.com/repos/huggingface/datasets/issues/2939/events
https://github.com/huggingface/datasets/pull/2939
999,639,630
PR_kwDODunzps4r58Gu
2,939
MENYO-20k repo has moved, updating URL
[]
closed
false
null
0
2021-09-17T19:01:54Z
2021-09-21T15:31:37Z
2021-09-21T15:31:36Z
null
Dataset repo moved to https://github.com/uds-lsv/menyo-20k_MT, now editing URL to match. https://github.com/uds-lsv/menyo-20k_MT/blob/master/data/train.tsv is the file we're looking for
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2939/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/2939/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/2939.diff", "html_url": "https://github.com/huggingface/datasets/pull/2939", "merged_at": "2021-09-21T15:31:36Z", "patch_url": "https://github.com/huggingface/datasets/pull/2939.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/...
true
[]
https://api.github.com/repos/huggingface/datasets/issues/344
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/344/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/344/comments
https://api.github.com/repos/huggingface/datasets/issues/344/events
https://github.com/huggingface/datasets/pull/344
651,495,246
MDExOlB1bGxSZXF1ZXN0NDQ0NzQwMTIw
344
Search qa
[]
closed
false
null
1
2020-07-06T12:23:16Z
2020-07-16T08:58:16Z
2020-07-16T08:58:16Z
null
This PR adds the Search QA dataset used in **SearchQA: A New Q&A Dataset Augmented with Context from a Search Engine**. The dataset has the following config name: - raw_jeopardy: raw data - train_test_val: which is the splitted version #336
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/344/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/344/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/344.diff", "html_url": "https://github.com/huggingface/datasets/pull/344", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/344.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/344" }
true
[ "Could you rebase from master just to make sure we won't break anything for `fever` pls @mariamabarham ?" ]
https://api.github.com/repos/huggingface/datasets/issues/5075
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5075/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5075/comments
https://api.github.com/repos/huggingface/datasets/issues/5075/events
https://github.com/huggingface/datasets/issues/5075
1,397,865,501
I_kwDODunzps5TUbwd
5,075
Throw EnvironmentError when token is not present
[ { "color": "7057ff", "default": true, "description": "Good for newcomers", "id": 1935892877, "name": "good first issue", "node_id": "MDU6TGFiZWwxOTM1ODkyODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/good%20first%20issue" }, { "color": "DF8D62", "defa...
closed
false
null
1
2022-10-05T14:14:18Z
2022-10-07T14:33:28Z
2022-10-07T14:33:28Z
null
Throw EnvironmentError instead of OSError ([link](https://github.com/huggingface/datasets/blob/6ad430ba0cdeeb601170f732d4bd977f5c04594d/src/datasets/arrow_dataset.py#L4306) to the line) in `push_to_hub` when the Hub token is not present.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5075/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/5075/timeline
null
completed
null
null
false
[ "@mariosasko I've raised a PR #5076 against this issue. Please help to review. Thanks." ]
https://api.github.com/repos/huggingface/datasets/issues/6076
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6076/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6076/comments
https://api.github.com/repos/huggingface/datasets/issues/6076/events
https://github.com/huggingface/datasets/pull/6076
1,822,345,597
PR_kwDODunzps5WcGVR
6,076
No gzip encoding from github
[]
closed
false
null
3
2023-07-26T12:46:07Z
2023-07-27T16:15:11Z
2023-07-27T16:14:40Z
null
Don't accept gzip encoding from github, otherwise some files are not streamable + seekable. fix https://huggingface.co/datasets/code_x_glue_cc_code_to_code_trans/discussions/2#64c0e0c1a04a514ba6303e84 and making sure https://github.com/huggingface/datasets/issues/2918 works as well
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6076/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6076/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/6076.diff", "html_url": "https://github.com/huggingface/datasets/pull/6076", "merged_at": "2023-07-27T16:14:40Z", "patch_url": "https://github.com/huggingface/datasets/pull/6076.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/...
true
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
https://api.github.com/repos/huggingface/datasets/issues/740
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/740/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/740/comments
https://api.github.com/repos/huggingface/datasets/issues/740/events
https://github.com/huggingface/datasets/pull/740
723,047,958
MDExOlB1bGxSZXF1ZXN0NTA0NzAyNTc0
740
Fix TREC urls
[]
closed
false
null
0
2020-10-16T09:11:28Z
2020-10-19T08:54:37Z
2020-10-19T08:54:36Z
null
The old TREC urls are now redirections. I updated the urls to the new ones, since we don't support redirections for downloads. Fix #737
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/740/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/740/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/740.diff", "html_url": "https://github.com/huggingface/datasets/pull/740", "merged_at": "2020-10-19T08:54:35Z", "patch_url": "https://github.com/huggingface/datasets/pull/740.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/740...
true
[]
https://api.github.com/repos/huggingface/datasets/issues/3071
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/3071/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/3071/comments
https://api.github.com/repos/huggingface/datasets/issues/3071/events
https://github.com/huggingface/datasets/issues/3071
1,024,893,493
I_kwDODunzps49FqI1
3,071
Custom plain text dataset, plain json dataset and plain csv dataset are remove from datasets template folder
[ { "color": "e99695", "default": false, "description": "Requesting to add a new dataset", "id": 2067376369, "name": "dataset request", "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request" } ]
closed
false
null
1
2021-10-13T07:32:10Z
2021-10-13T08:27:04Z
2021-10-13T08:27:03Z
null
## Adding a Dataset - **Name:** text, json, csv - **Description:** I am developing a customized dataset loading script. The problem is mainly about my custom dataset is seperate into many files and I only find a dataset loading template in [https://github.com/huggingface/datasets/blob/1.2.1/datasets/json/json.py](ht...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3071/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/3071/timeline
null
completed
null
null
false
[ "Hi @zixiliuUSC, \r\n\r\nAs explained in the documentation (https://huggingface.co/docs/datasets/loading.html#json), we support loading any dataset in JSON (as well as CSV, text, Parquet) format:\r\n```python\r\nds = load_dataset('json', data_files='my_file.json')\r\n```" ]
https://api.github.com/repos/huggingface/datasets/issues/5158
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5158/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5158/comments
https://api.github.com/repos/huggingface/datasets/issues/5158/events
https://github.com/huggingface/datasets/issues/5158
1,422,059,287
I_kwDODunzps5UwucX
5,158
Fix language and license tag names in all Hub datasets
[ { "color": "0e8a16", "default": false, "description": "Contribution to a dataset script", "id": 4564477500, "name": "dataset contribution", "node_id": "LA_kwDODunzps8AAAABEBBmPA", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20contribution" } ]
closed
false
null
6
2022-10-25T08:19:29Z
2022-10-25T11:27:26Z
2022-10-25T10:42:19Z
null
While working on this: - #5137 we realized there are still many datasets with deprecated "languages" and "licenses" tag names (instead of "language" and "license"). This is a blocking issue: no subsequent PR can be opened to modify their metadata: a ValueError will be thrown. We should fix the "language" and ...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5158/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/5158/timeline
null
completed
null
null
false
[ "There are currently 402 datasets with deprecated \"languages\" or \"licenses\".", "hey @albertvillanova ,i would love to work on this issue if you like.", "Hi @ayushthe1, thanks for your offer.\r\n\r\nBut as you can see, I self-assigned this issue.\r\n\r\nI have already fixed 200 out of the 402 datasets. My sc...
https://api.github.com/repos/huggingface/datasets/issues/2175
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2175/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2175/comments
https://api.github.com/repos/huggingface/datasets/issues/2175/events
https://github.com/huggingface/datasets/issues/2175
851,836,096
MDU6SXNzdWU4NTE4MzYwOTY=
2,175
dataset.search_batch() function outputs all -1 indices sometime.
[]
closed
false
null
6
2021-04-06T21:50:49Z
2021-04-16T12:21:16Z
2021-04-16T12:21:15Z
null
I am working with RAG and playing around with different faiss indexes. At the moment I use **index = faiss.index_factory(768, "IVF65536_HNSW32,Flat")**. During the retrieval phase exactly in [this line of retrieval_rag.py](https://github.com/huggingface/transformers/blob/master/src/transformers/models/rag/retrieval_...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2175/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/2175/timeline
null
completed
null
null
false
[ "Actually, I found the answer [here](https://github.com/facebookresearch/faiss/wiki/FAQ#what-does-it-mean-when-a-search-returns--1-ids). \r\n\r\nSo we have to do some modifications to the code for instances where the index doesn't retrieve any IDs.", "@lhoestq @patrickvonplaten \r\n\r\nI also found another short...
https://api.github.com/repos/huggingface/datasets/issues/835
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/835/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/835/comments
https://api.github.com/repos/huggingface/datasets/issues/835/events
https://github.com/huggingface/datasets/issues/835
740,102,210
MDU6SXNzdWU3NDAxMDIyMTA=
835
Wikipedia postprocessing
[]
closed
false
null
3
2020-11-10T17:26:38Z
2020-11-10T18:23:20Z
2020-11-10T17:49:21Z
null
Hi, thanks for this library! Running this code: ```py import datasets wikipedia = datasets.load_dataset("wikipedia", "20200501.de") print(wikipedia['train']['text'][0]) ``` I get: ``` mini|Ricardo Flores Magón mini|Mexikanische Revolutionäre, Magón in der Mitte anführend, gegen die Diktatur von Porfir...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/835/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/835/timeline
null
completed
null
null
false
[ "Hi @bminixhofer ! Parsing WikiMedia is notoriously difficult: this processing used [mwparserfromhell](https://github.com/earwig/mwparserfromhell) which is pretty good but not perfect.\r\n\r\nAs an alternative, you can also use the Wiki40b dataset which was pre-processed using an un-released Google internal tool", ...
https://api.github.com/repos/huggingface/datasets/issues/3545
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/3545/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/3545/comments
https://api.github.com/repos/huggingface/datasets/issues/3545/events
https://github.com/huggingface/datasets/pull/3545
1,096,189,889
PR_kwDODunzps4wpziv
3,545
fix: 🐛 pass token when retrieving the split names
[]
closed
false
null
3
2022-01-07T10:29:22Z
2022-01-10T10:51:47Z
2022-01-10T10:51:46Z
null
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3545/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/3545/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/3545.diff", "html_url": "https://github.com/huggingface/datasets/pull/3545", "merged_at": "2022-01-10T10:51:46Z", "patch_url": "https://github.com/huggingface/datasets/pull/3545.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/...
true
[ "Currently, it does not work with https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0/blob/main/common_voice_7_0.py#L146 (which was the goal), because `dl_manager.download_config.use_auth_token` is ignored, and the authentication is required to be use `huggingface-cli login`.\r\nIn my use case (data...
https://api.github.com/repos/huggingface/datasets/issues/3459
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/3459/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/3459/comments
https://api.github.com/repos/huggingface/datasets/issues/3459/events
https://github.com/huggingface/datasets/issues/3459
1,084,969,672
I_kwDODunzps5Aq1LI
3,459
dataset.filter overwriting previously set dataset._indices values, resulting in the wrong elements being selected.
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
null
2
2021-12-20T16:16:49Z
2021-12-20T16:34:57Z
2021-12-20T16:34:57Z
null
## Describe the bug When using dataset.select to select a subset of a dataset, dataset._indices are set to indicate which elements are now considered in the dataset. The same thing happens when you shuffle the dataset; dataset._indices are set to indicate what the new order of the data is. However, if you then use a...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3459/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/3459/timeline
null
completed
null
null
false
[ "I think this is a duplicate of [#3190](https://github.com/huggingface/datasets/issues/3190)?", "Upgrading the datasets version as per #3190 fixes this bug. \r\nI'm Marking as closed." ]
https://api.github.com/repos/huggingface/datasets/issues/1752
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1752/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1752/comments
https://api.github.com/repos/huggingface/datasets/issues/1752/events
https://github.com/huggingface/datasets/pull/1752
789,822,459
MDExOlB1bGxSZXF1ZXN0NTU4MTA5NTA5
1,752
COMET metric citation
[]
closed
false
null
1
2021-01-20T09:54:43Z
2021-01-20T10:27:07Z
2021-01-20T10:25:02Z
null
In my last pull request to add COMET metric, the citations where not following the usual "format". Because of that they where not correctly displayed on the website: <img width="814" alt="Screenshot 2021-01-20 at 09 48 44" src="https://user-images.githubusercontent.com/17256847/105158000-686efb80-5b05-11eb-8bb0-9c8...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/1752/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/1752/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/1752.diff", "html_url": "https://github.com/huggingface/datasets/pull/1752", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/1752.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/1752" }
true
[ "I think its better to create a new branch with this fix. I forgot I was still using the old branch." ]