url stringlengths 58 61 | repository_url stringclasses 1
value | labels_url stringlengths 72 75 | comments_url stringlengths 67 70 | events_url stringlengths 65 68 | html_url stringlengths 46 51 | id int64 599M 1.83B | node_id stringlengths 18 32 | number int64 1 6.09k | title stringlengths 1 290 | labels list | state stringclasses 2
values | locked bool 1
class | milestone dict | comments int64 0 54 | created_at stringlengths 20 20 | updated_at stringlengths 20 20 | closed_at stringlengths 20 20 ⌀ | active_lock_reason null | body stringlengths 0 228k ⌀ | reactions dict | timeline_url stringlengths 67 70 | performed_via_github_app null | state_reason stringclasses 3
values | draft bool 2
classes | pull_request dict | is_pull_request bool 2
classes | comments_text list |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
https://api.github.com/repos/huggingface/datasets/issues/6028 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6028/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6028/comments | https://api.github.com/repos/huggingface/datasets/issues/6028/events | https://github.com/huggingface/datasets/pull/6028 | 1,803,294,981 | PR_kwDODunzps5Vb3LJ | 6,028 | Use new hffs | [] | closed | false | null | 13 | 2023-07-13T15:41:44Z | 2023-07-17T17:09:39Z | 2023-07-17T17:01:00Z | null | Thanks to @janineguo 's work in https://github.com/huggingface/datasets/pull/5919 which was needed to support HfFileSystem.
Switching to `HfFileSystem` will help implementing optimization in data files resolution
## Implementation details
I replaced all the from_hf_repo and from_local_or_remote in data_files.p... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/6028/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/6028/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/6028.diff",
"html_url": "https://github.com/huggingface/datasets/pull/6028",
"merged_at": "2023-07-17T17:01:00Z",
"patch_url": "https://github.com/huggingface/datasets/pull/6028.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | true | [
"_The documentation is not available anymore as the PR was closed or merged._",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea... |
https://api.github.com/repos/huggingface/datasets/issues/4551 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/4551/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/4551/comments | https://api.github.com/repos/huggingface/datasets/issues/4551/events | https://github.com/huggingface/datasets/pull/4551 | 1,282,534,807 | PR_kwDODunzps46QAV- | 4,551 | Perform hidden file check on relative data file path | [] | closed | false | null | 5 | 2022-06-23T14:49:11Z | 2022-06-30T14:49:20Z | 2022-06-30T14:38:18Z | null | Fix #4549 | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/4551/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/4551/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/4551.diff",
"html_url": "https://github.com/huggingface/datasets/pull/4551",
"merged_at": "2022-06-30T14:38:18Z",
"patch_url": "https://github.com/huggingface/datasets/pull/4551.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | true | [
"_The documentation is not available anymore as the PR was closed or merged._",
"I'm aware of this behavior, which is tricky to solve due to fsspec's hidden file handling (see https://github.com/huggingface/datasets/issues/4115#issuecomment-1108819538). I've tested some regex patterns to address this, and they se... |
https://api.github.com/repos/huggingface/datasets/issues/1269 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/1269/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/1269/comments | https://api.github.com/repos/huggingface/datasets/issues/1269/events | https://github.com/huggingface/datasets/pull/1269 | 758,886,174 | MDExOlB1bGxSZXF1ZXN0NTMzOTc3MTE2 | 1,269 | Adding OneStopEnglish corpus dataset | [] | closed | false | null | 1 | 2020-12-07T22:05:11Z | 2020-12-09T18:43:38Z | 2020-12-09T15:33:53Z | null | This PR adds OneStopEnglish Corpus containing texts classified into reading levels (elementary, intermediate, advance) for automatic readability assessment and text simplification.
Link to the paper: https://www.aclweb.org/anthology/W18-0535.pdf | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/1269/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/1269/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/1269.diff",
"html_url": "https://github.com/huggingface/datasets/pull/1269",
"merged_at": "2020-12-09T15:33:53Z",
"patch_url": "https://github.com/huggingface/datasets/pull/1269.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | true | [
"Hi @lhoestq, thanks for the review.\r\nI have made all the changes, PTAL! :) "
] |
https://api.github.com/repos/huggingface/datasets/issues/3186 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3186/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3186/comments | https://api.github.com/repos/huggingface/datasets/issues/3186/events | https://github.com/huggingface/datasets/issues/3186 | 1,040,369,397 | I_kwDODunzps4-Asb1 | 3,186 | Dataset viewer for nli_tr | [
{
"color": "fef2c0",
"default": false,
"description": "",
"id": 3287858981,
"name": "streaming",
"node_id": "MDU6TGFiZWwzMjg3ODU4OTgx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/streaming"
}
] | closed | false | null | 6 | 2021-10-31T03:56:33Z | 2022-09-12T09:15:34Z | 2022-09-12T08:43:09Z | null | ## Dataset viewer issue for '*nli_tr*'
**Link:** https://huggingface.co/datasets/nli_tr
Hello,
Thank you for the new dataset preview feature that will help the users to view the datasets online.
We just noticed that the dataset viewer widget in the `nli_tr` dataset shows the error below. The error must be d... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/3186/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/3186/timeline | null | completed | null | null | false | [
"It's an issue with the streaming mode:\r\n\r\n```python\r\n>>> import datasets\r\n>>> dataset = datasets.load_dataset('nli_tr', name='snli_tr',split='test', streaming=True)\r\n>>> next(iter(dataset))\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"/home/slesage/hf/datas... |
https://api.github.com/repos/huggingface/datasets/issues/4565 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/4565/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/4565/comments | https://api.github.com/repos/huggingface/datasets/issues/4565/events | https://github.com/huggingface/datasets/issues/4565 | 1,284,141,666 | I_kwDODunzps5MinJi | 4,565 | Add UFSC OCPap dataset | [
{
"color": "e99695",
"default": false,
"description": "Requesting to add a new dataset",
"id": 2067376369,
"name": "dataset request",
"node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5",
"url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request"
}
] | closed | false | null | 1 | 2022-06-24T20:07:54Z | 2022-07-06T19:03:02Z | 2022-07-06T19:03:02Z | null | ## Adding a Dataset
- **Name:** UFSC OCPap: Papanicolaou Stained Oral Cytology Dataset (v4)
- **Description:** The UFSC OCPap dataset comprises 9,797 labeled images of 1200x1600 pixels acquired from 5 slides of cancer diagnosed and 3 healthy of oral brush samples, from distinct patients.
- **Paper:** https://dx.doi.... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/4565/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/4565/timeline | null | completed | null | null | false | [
"I will add this directly on the hub (same as #4486)—in https://huggingface.co/lapix"
] |
https://api.github.com/repos/huggingface/datasets/issues/5661 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/5661/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/5661/comments | https://api.github.com/repos/huggingface/datasets/issues/5661/events | https://github.com/huggingface/datasets/issues/5661 | 1,637,129,445 | I_kwDODunzps5hlJzl | 5,661 | CI is broken: Unnecessary `dict` comprehension | [
{
"color": "d73a4a",
"default": true,
"description": "Something isn't working",
"id": 1935892857,
"name": "bug",
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug"
}
] | closed | false | null | 0 | 2023-03-23T09:13:01Z | 2023-03-23T09:37:51Z | 2023-03-23T09:37:51Z | null | CI check_code_quality is broken:
```
src/datasets/arrow_dataset.py:3267:35: C416 [*] Unnecessary `dict` comprehension (rewrite using `dict()`)
Found 1 error.
``` | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/5661/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/5661/timeline | null | completed | null | null | false | [] |
https://api.github.com/repos/huggingface/datasets/issues/4543 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/4543/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/4543/comments | https://api.github.com/repos/huggingface/datasets/issues/4543/events | https://github.com/huggingface/datasets/pull/4543 | 1,280,379,781 | PR_kwDODunzps46IiEp | 4,543 | [CI] Fix upstream hub test url | [] | closed | false | null | 2 | 2022-06-22T15:34:27Z | 2022-06-22T16:37:40Z | 2022-06-22T16:27:37Z | null | Some tests were still using moon-stagign instead of hub-ci.
I also updated the token to use one dedicated to `datasets` | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/4543/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/4543/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/4543.diff",
"html_url": "https://github.com/huggingface/datasets/pull/4543",
"merged_at": "2022-06-22T16:27:37Z",
"patch_url": "https://github.com/huggingface/datasets/pull/4543.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | true | [
"_The documentation is not available anymore as the PR was closed or merged._",
"Remaining CI failures are unrelated to this fix, merging"
] |
https://api.github.com/repos/huggingface/datasets/issues/858 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/858/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/858/comments | https://api.github.com/repos/huggingface/datasets/issues/858/events | https://github.com/huggingface/datasets/pull/858 | 743,904,516 | MDExOlB1bGxSZXF1ZXN0NTIxNzE3ODQ4 | 858 | Add SemEval-2010 task 8 | [] | closed | false | null | 1 | 2020-11-16T14:57:57Z | 2020-11-26T17:28:55Z | 2020-11-26T17:28:55Z | null | Hi,
I don't know how to add dummy data, since I create the validation set out of the last 1000 examples of the train set. If you have a suggestion, I am happy to implement it.
Cheers,
Joel | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/858/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/858/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/858.diff",
"html_url": "https://github.com/huggingface/datasets/pull/858",
"merged_at": "2020-11-26T17:28:55Z",
"patch_url": "https://github.com/huggingface/datasets/pull/858.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/858... | true | [
"Added dummy data and encoding to open(). Now everything should be fine, hopefully :)"
] |
https://api.github.com/repos/huggingface/datasets/issues/3882 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3882/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3882/comments | https://api.github.com/repos/huggingface/datasets/issues/3882/events | https://github.com/huggingface/datasets/pull/3882 | 1,164,595,388 | PR_kwDODunzps40NKz7 | 3,882 | Image process doc | [
{
"color": "0075ca",
"default": true,
"description": "Improvements or additions to documentation",
"id": 1935892861,
"name": "documentation",
"node_id": "MDU6TGFiZWwxOTM1ODkyODYx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/documentation"
}
] | closed | false | null | 1 | 2022-03-10T00:32:10Z | 2022-03-15T15:24:16Z | 2022-03-15T15:24:09Z | null | This PR is a first draft of how to process image data. It adds:
- Load an image dataset with `image` and `path` (adds tip about `decode=False` param to access the path and bytes, thanks to @mariosasko).
- Load an image using the `ImageFolder` builder. I know there is an [example](https://huggingface.co/docs/dataset... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 2,
"laugh": 0,
"rocket": 0,
"total_count": 2,
"url": "https://api.github.com/repos/huggingface/datasets/issues/3882/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/3882/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/3882.diff",
"html_url": "https://github.com/huggingface/datasets/pull/3882",
"merged_at": "2022-03-15T15:24:09Z",
"patch_url": "https://github.com/huggingface/datasets/pull/3882.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_3882). All of your documentation changes will be reflected on that endpoint."
] |
https://api.github.com/repos/huggingface/datasets/issues/1318 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/1318/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/1318/comments | https://api.github.com/repos/huggingface/datasets/issues/1318/events | https://github.com/huggingface/datasets/pull/1318 | 759,565,629 | MDExOlB1bGxSZXF1ZXN0NTM0NTQ5NjE3 | 1,318 | ethos first commit | [] | closed | false | null | 3 | 2020-12-08T15:59:47Z | 2020-12-10T14:45:57Z | 2020-12-10T14:45:57Z | null | Ethos passed all the tests except from this one:
RUN_SLOW=1 pytest tests/test_dataset_common.py::LocalDatasetTest::test_load_dataset_all_configs_<your-dataset-name>
with this error:
E OSError: Cannot find data file.
E Original error:
E [Errno 2] No such file or direct... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/1318/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/1318/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/1318.diff",
"html_url": "https://github.com/huggingface/datasets/pull/1318",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/1318.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/1318"
} | true | [
"> Nice thanks !\r\n> \r\n> I left a few comments\r\n> \r\n> Also it looks like this PR includes changes about other files than the ones for ethos\r\n> \r\n> Can you create another branch and another PR please ?\r\n\r\n@lhoestq Should I close this PR? The new one is the: #1453",
"You can create another PR and clo... |
https://api.github.com/repos/huggingface/datasets/issues/93 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/93/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/93/comments | https://api.github.com/repos/huggingface/datasets/issues/93/events | https://github.com/huggingface/datasets/pull/93 | 617,522,029 | MDExOlB1bGxSZXF1ZXN0NDE3NDIxODUy | 93 | Cleanup notebooks and various fixes | [] | closed | false | null | 0 | 2020-05-13T14:58:58Z | 2020-05-13T15:01:48Z | 2020-05-13T15:01:47Z | null | Fixes on dataset (more flexible) metrics (fix) and general clean ups | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/93/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/93/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/93.diff",
"html_url": "https://github.com/huggingface/datasets/pull/93",
"merged_at": "2020-05-13T15:01:47Z",
"patch_url": "https://github.com/huggingface/datasets/pull/93.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/93"
} | true | [] |
https://api.github.com/repos/huggingface/datasets/issues/5942 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/5942/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/5942/comments | https://api.github.com/repos/huggingface/datasets/issues/5942/events | https://github.com/huggingface/datasets/pull/5942 | 1,752,021,681 | PR_kwDODunzps5Su-V4 | 5,942 | Pass datasets-cli additional args as kwargs to DatasetBuilder in `run_beam.py` | [] | open | false | null | 0 | 2023-06-12T06:50:50Z | 2023-06-30T09:15:00Z | null | null | Hi,
Following this <https://discuss.huggingface.co/t/how-to-preprocess-a-wikipedia-dataset-using-dataflowrunner/41991/3>, here is a simple PR to pass any additional args to datasets-cli as kwargs in the DatasetBuilder in `run_beam.py`.
I also took the liberty to add missing setup steps to the `beam.mdx` docs in o... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/5942/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/5942/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/5942.diff",
"html_url": "https://github.com/huggingface/datasets/pull/5942",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/5942.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/5942"
} | true | [] |
https://api.github.com/repos/huggingface/datasets/issues/4476 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/4476/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/4476/comments | https://api.github.com/repos/huggingface/datasets/issues/4476/events | https://github.com/huggingface/datasets/issues/4476 | 1,267,987,499 | I_kwDODunzps5Lk_Qr | 4,476 | `to_pandas` doesn't take into account format. | [
{
"color": "a2eeef",
"default": true,
"description": "New feature or request",
"id": 1935892871,
"name": "enhancement",
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement"
}
] | closed | false | null | 4 | 2022-06-10T20:25:31Z | 2022-06-15T17:41:41Z | 2022-06-15T17:41:41Z | null | **Is your feature request related to a problem? Please describe.**
I have a large dataset that I need to convert part of to pandas to do some further analysis. Calling `to_pandas` directly on it is expensive. So I thought I could simply select the columns that I want and then call `to_pandas`.
**Describe the solu... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/4476/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/4476/timeline | null | completed | null | null | false | [
"Thanks for opening a discussion :)\r\n\r\nNote that you can use `.remove_columns(...)` to keep only the ones you're interested in before calling `.to_pandas()`",
"Yes I can do that thank you!\r\n\r\nDo you think that conceptually my example should work? If not, I'm happy to close this issue. \r\n\r\nIf yes, I ca... |
https://api.github.com/repos/huggingface/datasets/issues/3457 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3457/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3457/comments | https://api.github.com/repos/huggingface/datasets/issues/3457/events | https://github.com/huggingface/datasets/issues/3457 | 1,084,862,121 | I_kwDODunzps5Aqa6p | 3,457 | Add CMU Graphics Lab Motion Capture dataset | [
{
"color": "e99695",
"default": false,
"description": "Requesting to add a new dataset",
"id": 2067376369,
"name": "dataset request",
"node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5",
"url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request"
},
{
"color": "bfdadc",... | open | false | null | 3 | 2021-12-20T14:34:39Z | 2022-03-16T16:53:09Z | null | null | ## Adding a Dataset
- **Name:** CMU Graphics Lab Motion Capture database
- **Description:** The database contains free motions which you can download and use.
- **Data:** http://mocap.cs.cmu.edu/
- **Motivation:** Nice motion capture dataset
Instructions to add a new dataset can be found [here](https://github.c... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/3457/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/3457/timeline | null | null | null | null | false | [
"This dataset has files in ASF/AMC format. [ The skeleton file is the ASF file (Acclaim Skeleton File). The motion file is the AMC file (Acclaim Motion Capture data). ] \r\n\r\nSome questions : \r\n1. How do we go about representing these features using datasets.Features and generate examples ?\r\n2. The dataset d... |
https://api.github.com/repos/huggingface/datasets/issues/2313 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2313/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2313/comments | https://api.github.com/repos/huggingface/datasets/issues/2313/events | https://github.com/huggingface/datasets/pull/2313 | 875,475,367 | MDExOlB1bGxSZXF1ZXN0NjI5ODEwNTc4 | 2,313 | Remove unused head_hf_s3 function | [] | closed | false | null | 0 | 2021-05-04T13:42:06Z | 2021-05-07T09:31:42Z | 2021-05-07T09:31:42Z | null | Currently, the function `head_hf_s3` is not used:
- neither its returned result is used
- nor it raises any exception, as exceptions are catched and returned (not raised)
This PR removes it. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2313/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/2313/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/2313.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2313",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/2313.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2313"
} | true | [] |
https://api.github.com/repos/huggingface/datasets/issues/1933 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/1933/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/1933/comments | https://api.github.com/repos/huggingface/datasets/issues/1933/events | https://github.com/huggingface/datasets/pull/1933 | 814,335,846 | MDExOlB1bGxSZXF1ZXN0NTc4MzQwMzk3 | 1,933 | Use arrow ipc file format | [] | open | false | null | 0 | 2021-02-23T10:38:24Z | 2022-07-06T15:19:48Z | null | null | According to the [documentation](https://arrow.apache.org/docs/format/Columnar.html?highlight=arrow1#ipc-file-format), it's identical to the streaming format except that it contains the memory offsets of each sample:
> We define a “file format” supporting random access that is build with the stream format. The file ... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/1933/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/1933/timeline | null | null | true | {
"diff_url": "https://github.com/huggingface/datasets/pull/1933.diff",
"html_url": "https://github.com/huggingface/datasets/pull/1933",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/1933.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/1933"
} | true | [] |
https://api.github.com/repos/huggingface/datasets/issues/1906 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/1906/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/1906/comments | https://api.github.com/repos/huggingface/datasets/issues/1906/events | https://github.com/huggingface/datasets/issues/1906 | 811,405,274 | MDU6SXNzdWU4MTE0MDUyNzQ= | 1,906 | Feature Request: Support for Pandas `Categorical` | [
{
"color": "a2eeef",
"default": true,
"description": "New feature or request",
"id": 1935892871,
"name": "enhancement",
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement"
},
{
"color": "c5def5",
"default": fals... | open | false | null | 3 | 2021-02-18T19:46:05Z | 2021-02-23T14:38:50Z | null | null | ```
from datasets import Dataset
import pandas as pd
import pyarrow
df = pd.DataFrame(pd.Series(["a", "b", "c", "a"], dtype="category"))
pyarrow.Table.from_pandas(df)
Dataset.from_pandas(df)
# Throws NotImplementedError
# TODO(thom) this will need access to the dictionary as well (for labels). I.e. to the py_... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/1906/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/1906/timeline | null | null | null | null | false | [
"We already have a ClassLabel type that does this kind of mapping between the label ids (integers) and actual label values (strings).\r\n\r\nI wonder if actually we should use the DictionaryType from Arrow and the Categorical type from pandas for the `datasets` ClassLabel feature type.\r\nCurrently ClassLabel corre... |
https://api.github.com/repos/huggingface/datasets/issues/4679 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/4679/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/4679/comments | https://api.github.com/repos/huggingface/datasets/issues/4679/events | https://github.com/huggingface/datasets/pull/4679 | 1,303,980,648 | PR_kwDODunzps47XX67 | 4,679 | Added method to remove excess nesting in a DatasetDict | [] | closed | false | null | 11 | 2022-07-13T21:49:37Z | 2022-07-21T15:55:26Z | 2022-07-21T10:55:02Z | null | Added the ability for a DatasetDict to remove additional nested layers within its features to avoid conflicts when collating. It is meant to accompany [this PR](https://github.com/huggingface/transformers/pull/18119) to resolve the same issue [#15505](https://github.com/huggingface/transformers/issues/15505).
@stas0... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/4679/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/4679/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/4679.diff",
"html_url": "https://github.com/huggingface/datasets/pull/4679",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/4679.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4679"
} | true | [
"Hi ! I think the issue you linked is closed and suggests to use `remove_columns`.\r\n\r\nMoreover if you end up with a dataset with an unnecessarily nested data, please modify your processing functions to not output nested data, or use `map(..., batched=True)` if you function take batches as input",
"Hi @lhoestq... |
https://api.github.com/repos/huggingface/datasets/issues/5458 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/5458/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/5458/comments | https://api.github.com/repos/huggingface/datasets/issues/5458/events | https://github.com/huggingface/datasets/issues/5458 | 1,555,054,737 | I_kwDODunzps5csECR | 5,458 | slice split while streaming | [] | closed | false | null | 2 | 2023-01-24T14:08:17Z | 2023-01-24T15:11:47Z | 2023-01-24T15:11:47Z | null | ### Describe the bug
When using the `load_dataset` function with streaming set to True, slicing splits is apparently not supported.
Did I miss this in the documentation?
### Steps to reproduce the bug
`load_dataset("lhoestq/demo1",revision=None, streaming=True, split="train[:3]")`
causes ValueError: Bad split:... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/5458/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/5458/timeline | null | completed | null | null | false | [
"Hi! Yes, that's correct. When `streaming` is `True`, only split names can be specified as `split`, and for slicing, you have to use `.skip`/`.take` instead.\r\n\r\nE.g. \r\n`load_dataset(\"lhoestq/demo1\",revision=None, streaming=True, split=\"train[:3]\")`\r\n\r\nrewritten with `.skip`/`.take`:\r\n`load_dataset(\... |
https://api.github.com/repos/huggingface/datasets/issues/4895 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/4895/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/4895/comments | https://api.github.com/repos/huggingface/datasets/issues/4895/events | https://github.com/huggingface/datasets/issues/4895 | 1,350,798,527 | I_kwDODunzps5Qg4y_ | 4,895 | load_dataset method returns Unknown split "validation" even if this dir exists | [
{
"color": "d73a4a",
"default": true,
"description": "Something isn't working",
"id": 1935892857,
"name": "bug",
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug"
}
] | closed | false | null | 17 | 2022-08-25T12:11:00Z | 2022-10-06T17:49:28Z | 2022-09-29T08:07:50Z | null | ## Describe the bug
The `datasets.load_dataset` returns a `ValueError: Unknown split "validation". Should be one of ['train', 'test'].` when running `load_dataset(local_data_dir_path, split="validation")` even if the `validation` sub-directory exists in the local data path.
The data directories are as follows and a... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/4895/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/4895/timeline | null | completed | null | null | false | [
"I don't know the main problem but it looks like, it is ignoring the last directory in your case. So, create a directory called 'zzz' in the same folder as train, validation and test. if it doesn't work, create a directory called \"aaa\". It worked for me.\r\n",
"@SamSamhuns could you please try to load it with t... |
https://api.github.com/repos/huggingface/datasets/issues/646 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/646/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/646/comments | https://api.github.com/repos/huggingface/datasets/issues/646/events | https://github.com/huggingface/datasets/pull/646 | 704,607,371 | MDExOlB1bGxSZXF1ZXN0NDg5NTAyMTM3 | 646 | Fix docs typos | [] | closed | false | null | 0 | 2020-09-18T19:32:27Z | 2020-09-21T16:30:54Z | 2020-09-21T16:14:12Z | null | This PR fixes few typos in the docs and the error in the code snippet in the set_format section in docs/source/torch_tensorflow.rst. `torch.utils.data.Dataloader` expects padded batches so it throws an error due to not being able to stack the unpadded tensors. If we follow the Quick tour from the docs where they add th... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/646/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/646/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/646.diff",
"html_url": "https://github.com/huggingface/datasets/pull/646",
"merged_at": "2020-09-21T16:14:12Z",
"patch_url": "https://github.com/huggingface/datasets/pull/646.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/646... | true | [] |
https://api.github.com/repos/huggingface/datasets/issues/1393 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/1393/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/1393/comments | https://api.github.com/repos/huggingface/datasets/issues/1393/events | https://github.com/huggingface/datasets/pull/1393 | 760,436,267 | MDExOlB1bGxSZXF1ZXN0NTM1MjY4MjUx | 1,393 | Add script_version suggestion when dataset/metric not found | [] | closed | false | null | 0 | 2020-12-09T15:37:38Z | 2020-12-10T18:17:05Z | 2020-12-10T18:17:05Z | null | Adds a helpful prompt to the error message when a dataset/metric is not found, suggesting the user might need to pass `script_version="master"` if the dataset was added recently. The whole error looks like:
> Couldn't find file locally at blah/blah.py, or remotely at https://raw.githubusercontent.com/huggingface/dat... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/1393/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/1393/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/1393.diff",
"html_url": "https://github.com/huggingface/datasets/pull/1393",
"merged_at": "2020-12-10T18:17:05Z",
"patch_url": "https://github.com/huggingface/datasets/pull/1393.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | true | [] |
https://api.github.com/repos/huggingface/datasets/issues/1701 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/1701/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/1701/comments | https://api.github.com/repos/huggingface/datasets/issues/1701/events | https://github.com/huggingface/datasets/issues/1701 | 781,345,717 | MDU6SXNzdWU3ODEzNDU3MTc= | 1,701 | Some datasets miss dataset_infos.json or dummy_data.zip | [] | closed | false | null | 2 | 2021-01-07T14:17:13Z | 2022-11-04T15:11:16Z | 2022-11-04T15:06:00Z | null | While working on dataset REAME generation script at https://github.com/madlag/datasets_readme_generator , I noticed that some datasets miss a dataset_infos.json :
```
c4
lm1b
reclor
wikihow
```
And some does not have a dummy_data.zip :
```
kor_nli
math_dataset
mlqa
ms_marco
newsgroup
qa4mre
qanga... | {
"+1": 1,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/1701/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/1701/timeline | null | completed | null | null | false | [
"Thanks for reporting.\r\nWe should indeed add all the missing dummy_data.zip and also the dataset_infos.json at least for lm1b, reclor and wikihow.\r\n\r\nFor c4 I haven't tested the script and I think we'll require some optimizations regarding beam datasets before processing it.\r\n",
"Closing since the dummy d... |
https://api.github.com/repos/huggingface/datasets/issues/2319 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2319/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2319/comments | https://api.github.com/repos/huggingface/datasets/issues/2319/events | https://github.com/huggingface/datasets/issues/2319 | 876,251,376 | MDU6SXNzdWU4NzYyNTEzNzY= | 2,319 | UnicodeDecodeError for OSCAR (Afrikaans) | [
{
"color": "d73a4a",
"default": true,
"description": "Something isn't working",
"id": 1935892857,
"name": "bug",
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug"
}
] | closed | false | null | 3 | 2021-05-05T09:22:52Z | 2021-05-05T10:57:31Z | 2021-05-05T10:50:55Z | null | ## Describe the bug
When loading the [OSCAR dataset](https://huggingface.co/datasets/oscar) (specifically `unshuffled_deduplicated_af`), I encounter a `UnicodeDecodeError`.
## Steps to reproduce the bug
```python
from datasets import load_dataset
dataset = load_dataset("oscar", "unshuffled_deduplicated_af")
```... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2319/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/2319/timeline | null | completed | null | null | false | [
"Thanks for reporting, @sgraaf.\r\n\r\nI am going to have a look at it. \r\n\r\nI guess the expected codec is \"UTF-8\". Normally, when no explicitly codec is passed, Python uses one which is platform-dependent. For Linux machines, the default codec is `utf_8`, which is OK. However for Windows machine, the default ... |
https://api.github.com/repos/huggingface/datasets/issues/4240 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/4240/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/4240/comments | https://api.github.com/repos/huggingface/datasets/issues/4240/events | https://github.com/huggingface/datasets/pull/4240 | 1,217,287,594 | PR_kwDODunzps423xRl | 4,240 | Fix yield for crd3 | [] | closed | false | null | 2 | 2022-04-27T12:31:36Z | 2022-04-29T12:41:41Z | 2022-04-29T12:41:41Z | null | Modified the `_generate_examples` function to consider all the turns for a chunk id as a single example
Modified the features accordingly
```
"turns": [
{
"names": datasets.features.Sequence(datasets.Value("string")),
"utterances": ... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/4240/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/4240/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/4240.diff",
"html_url": "https://github.com/huggingface/datasets/pull/4240",
"merged_at": "2022-04-29T12:41:41Z",
"patch_url": "https://github.com/huggingface/datasets/pull/4240.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | true | [
"I don't think you need to generate new dummy data, since they're in the same format as the original data.\r\n\r\nThe CI is failing because of this error:\r\n```python\r\n> turn[\"names\"] = turn[\"NAMES\"]\r\nE TypeError: tuple indices must be integers or slices, not str... |
https://api.github.com/repos/huggingface/datasets/issues/4309 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/4309/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/4309/comments | https://api.github.com/repos/huggingface/datasets/issues/4309/events | https://github.com/huggingface/datasets/pull/4309 | 1,231,232,935 | PR_kwDODunzps43lKpm | 4,309 | [WIP] Add TEDLIUM dataset | [
{
"color": "e99695",
"default": false,
"description": "Requesting to add a new dataset",
"id": 2067376369,
"name": "dataset request",
"node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5",
"url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request"
},
{
"color": "d93f0b",... | closed | false | null | 11 | 2022-05-10T14:12:47Z | 2022-06-17T12:54:40Z | 2022-06-17T11:44:01Z | null | Adds the TED-LIUM dataset https://www.tensorflow.org/datasets/catalog/tedlium#tedliumrelease3
TODO:
- [x] Port `tedium.py` from TF datasets using `convert_dataset.sh` script
- [x] Make `load_dataset` work
- [ ] ~~Run `datasets-cli` command to generate `dataset_infos.json`~~
- [ ] ~~Create dummy data for conti... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/4309/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/4309/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/4309.diff",
"html_url": "https://github.com/huggingface/datasets/pull/4309",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/4309.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4309"
} | true | [
"_The documentation is not available anymore as the PR was closed or merged._",
"```python\r\nfrom datasets import load_dataset\r\n\r\nds = load_dataset('./datasets/tedlium', 'release1', cache_dir='/home/sanchitgandhi/cache')\r\n```\r\n\r\n```\r\nDownloading and preparing dataset tedlium/release1 to /home/sanchit... |
https://api.github.com/repos/huggingface/datasets/issues/5243 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/5243/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/5243/comments | https://api.github.com/repos/huggingface/datasets/issues/5243/events | https://github.com/huggingface/datasets/issues/5243 | 1,449,523,962 | I_kwDODunzps5WZfr6 | 5,243 | Download only split data | [
{
"color": "a2eeef",
"default": true,
"description": "New feature or request",
"id": 1935892871,
"name": "enhancement",
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement"
}
] | open | false | null | 5 | 2022-11-15T10:15:54Z | 2023-05-02T09:27:51Z | null | null | ### Feature request
Is it possible to download only the data that I am requesting and not the entire dataset? I run out of disk spaceas it seems to download the entire dataset, instead of only the part needed.
common_voice["test"] = load_dataset("mozilla-foundation/common_voice_11_0", "en", split="test",
... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/5243/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/5243/timeline | null | null | null | null | false | [
"Hi @capsabogdan! Unfortunately, it's hard to implement because quite often datasets data is being hosted in a single archive for all splits :( So we have to download the whole archive to split it into splits. This is the case for CommonVoice too. \r\n\r\nHowever, for cases when data is distributed in separate arch... |
https://api.github.com/repos/huggingface/datasets/issues/1015 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/1015/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/1015/comments | https://api.github.com/repos/huggingface/datasets/issues/1015/events | https://github.com/huggingface/datasets/pull/1015 | 755,508,841 | MDExOlB1bGxSZXF1ZXN0NTMxMjA2MTgy | 1,015 | add hard dataset | [] | closed | false | null | 1 | 2020-12-02T18:27:36Z | 2020-12-03T15:03:54Z | 2020-12-03T15:03:54Z | null | Hotel Reviews in Arabic language. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/1015/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/1015/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/1015.diff",
"html_url": "https://github.com/huggingface/datasets/pull/1015",
"merged_at": "2020-12-03T15:03:54Z",
"patch_url": "https://github.com/huggingface/datasets/pull/1015.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | true | [
"Thanks @sumanthd17 that fixed it. "
] |
https://api.github.com/repos/huggingface/datasets/issues/1898 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/1898/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/1898/comments | https://api.github.com/repos/huggingface/datasets/issues/1898/events | https://github.com/huggingface/datasets/issues/1898 | 810,157,251 | MDU6SXNzdWU4MTAxNTcyNTE= | 1,898 | ALT dataset has repeating instances in all splits | [
{
"color": "2edb81",
"default": false,
"description": "A bug in a dataset script provided in the library",
"id": 2067388877,
"name": "dataset bug",
"node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug"
}
] | closed | false | null | 4 | 2021-02-17T12:51:42Z | 2021-02-19T06:18:46Z | 2021-02-19T06:18:46Z | null | The [ALT](https://huggingface.co/datasets/alt) dataset has all the same instances within each split :/
Seemed like a great dataset for some experiments I wanted to carry out, especially since its medium-sized, and has all splits.
Would be great if this could be fixed :)
Added a snapshot of the contents from `exp... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/1898/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/1898/timeline | null | completed | null | null | false | [
"Thanks for reporting. This looks like a very bad issue. I'm looking into it",
"I just merged a fix, we'll do a patch release soon. Thanks again for reporting, and sorry for the inconvenience.\r\nIn the meantime you can load `ALT` using `datasets` from the master branch",
"Thanks!!! works perfectly in the blead... |
https://api.github.com/repos/huggingface/datasets/issues/4787 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/4787/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/4787/comments | https://api.github.com/repos/huggingface/datasets/issues/4787/events | https://github.com/huggingface/datasets/issues/4787 | 1,328,243,911 | I_kwDODunzps5PK2TH | 4,787 | NonMatchingChecksumError in mbpp dataset | [
{
"color": "d73a4a",
"default": true,
"description": "Something isn't working",
"id": 1935892857,
"name": "bug",
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug"
}
] | closed | false | null | 0 | 2022-08-04T08:15:51Z | 2022-08-04T17:21:01Z | 2022-08-04T17:21:01Z | null | ## Describe the bug
As reported on the Hub [Fix Checksum Mismatch](https://huggingface.co/datasets/mbpp/discussions/1), there is a `NonMatchingChecksumError` when loading mbpp dataset
## Steps to reproduce the bug
```python
ds = load_dataset("mbpp", "full")
```
## Expected results
Loading of the dataset with... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/4787/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/4787/timeline | null | completed | null | null | false | [] |
https://api.github.com/repos/huggingface/datasets/issues/4014 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/4014/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/4014/comments | https://api.github.com/repos/huggingface/datasets/issues/4014/events | https://github.com/huggingface/datasets/pull/4014 | 1,180,481,229 | PR_kwDODunzps41AGBu | 4,014 | Support streaming id_clickbait dataset | [] | closed | false | null | 1 | 2022-03-25T08:18:28Z | 2022-03-25T08:58:31Z | 2022-03-25T08:53:32Z | null | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/4014/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/4014/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/4014.diff",
"html_url": "https://github.com/huggingface/datasets/pull/4014",
"merged_at": "2022-03-25T08:53:32Z",
"patch_url": "https://github.com/huggingface/datasets/pull/4014.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | true | [
"_The documentation is not available anymore as the PR was closed or merged._"
] |
https://api.github.com/repos/huggingface/datasets/issues/2369 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2369/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2369/comments | https://api.github.com/repos/huggingface/datasets/issues/2369/events | https://github.com/huggingface/datasets/pull/2369 | 893,554,153 | MDExOlB1bGxSZXF1ZXN0NjQ2MDQ5NDM1 | 2,369 | correct labels of conll2003 | [] | closed | false | null | 0 | 2021-05-17T17:37:54Z | 2021-05-18T08:27:42Z | 2021-05-18T08:27:42Z | null | # What does this PR
It fixes/extends the `ner_tags` for conll2003 to include all.
Paper reference https://arxiv.org/pdf/cs/0306050v1.pdf
Model reference https://huggingface.co/elastic/distilbert-base-cased-finetuned-conll03-english/blob/main/config.json
| {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2369/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/2369/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/2369.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2369",
"merged_at": "2021-05-18T08:27:42Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2369.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | true | [] |
https://api.github.com/repos/huggingface/datasets/issues/2924 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2924/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2924/comments | https://api.github.com/repos/huggingface/datasets/issues/2924/events | https://github.com/huggingface/datasets/issues/2924 | 997,378,113 | I_kwDODunzps47cshB | 2,924 | "File name too long" error for file locks | [
{
"color": "d73a4a",
"default": true,
"description": "Something isn't working",
"id": 1935892857,
"name": "bug",
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug"
}
] | closed | false | null | 10 | 2021-09-15T18:16:50Z | 2023-03-28T06:50:18Z | 2021-10-29T09:42:24Z | null | ## Describe the bug
Getting the following error when calling `load_dataset("gar1t/test")`:
```
OSError: [Errno 36] File name too long: '<user>/.cache/huggingface/datasets/_home_garrett_.cache_huggingface_datasets_csv_test-7c856aea083a7043_0.0.0_9144e0a4e8435090117cea53e6c7537173ef2304525df4a077c435d8ee7828ff.inc... | {
"+1": 0,
"-1": 0,
"confused": 1,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2924/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/2924/timeline | null | completed | null | null | false | [
"Hi, the filename here is less than 255\r\n```python\r\n>>> len(\"_home_garrett_.cache_huggingface_datasets_csv_test-7c856aea083a7043_0.0.0_9144e0a4e8435090117cea53e6c7537173ef2304525df4a077c435d8ee7828ff.incomplete.lock\")\r\n154\r\n```\r\nso not sure why it's considered too long for your filesystem.\r\n(also note... |
https://api.github.com/repos/huggingface/datasets/issues/4254 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/4254/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/4254/comments | https://api.github.com/repos/huggingface/datasets/issues/4254/events | https://github.com/huggingface/datasets/pull/4254 | 1,220,204,395 | PR_kwDODunzps43Bwnj | 4,254 | Replace data URL in SAMSum dataset and support streaming | [] | closed | false | null | 1 | 2022-04-29T08:21:43Z | 2022-05-06T08:38:16Z | 2022-04-29T16:26:09Z | null | This PR replaces data URL in SAMSum dataset:
- original host (arxiv.org) does not allow HTTP Range requests
- we have hosted the data on the Hub (license: CC BY-NC-ND 4.0)
Moreover, it implements support for streaming.
Fix #4146.
Related to: #4236.
CC: @severo | {
"+1": 1,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/4254/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/4254/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/4254.diff",
"html_url": "https://github.com/huggingface/datasets/pull/4254",
"merged_at": "2022-04-29T16:26:08Z",
"patch_url": "https://github.com/huggingface/datasets/pull/4254.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | true | [
"_The documentation is not available anymore as the PR was closed or merged._"
] |
https://api.github.com/repos/huggingface/datasets/issues/4935 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/4935/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/4935/comments | https://api.github.com/repos/huggingface/datasets/issues/4935/events | https://github.com/huggingface/datasets/issues/4935 | 1,363,226,736 | I_kwDODunzps5RQTBw | 4,935 | Dataset Viewer issue for ubuntu_dialogs_corpus | [
{
"color": "E5583E",
"default": false,
"description": "Related to the dataset viewer on huggingface.co",
"id": 3470211881,
"name": "dataset-viewer",
"node_id": "LA_kwDODunzps7O1zsp",
"url": "https://api.github.com/repos/huggingface/datasets/labels/dataset-viewer"
}
] | closed | false | null | 1 | 2022-09-06T12:41:50Z | 2022-09-06T12:51:25Z | 2022-09-06T12:51:25Z | null | ### Link
_No response_
### Description
_No response_
### Owner
_No response_ | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/4935/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/4935/timeline | null | completed | null | null | false | [
"The dataset maintainers (https://huggingface.co/datasets/ubuntu_dialogs_corpus) decided to forbid the dataset from being downloaded automatically (https://huggingface.co/docs/datasets/v2.4.0/en/loading#manual-download), and the dataset viewer respects this.\r\nWe will try to improve the error display though. Thank... |
https://api.github.com/repos/huggingface/datasets/issues/3785 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3785/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3785/comments | https://api.github.com/repos/huggingface/datasets/issues/3785/events | https://github.com/huggingface/datasets/pull/3785 | 1,150,069,801 | PR_kwDODunzps4zciES | 3,785 | Fix: Bypass Virus Checks in Google Drive Links (CNN-DM dataset) | [] | closed | false | null | 8 | 2022-02-25T05:48:57Z | 2022-03-03T16:43:47Z | 2022-03-03T14:03:37Z | null | This commit fixes the issue described in #3784. By adding an extra parameter to the end of Google Drive links, we are able to bypass the virus check and download the datasets.
So, if the original link looked like https://drive.google.com/uc?export=download&id=0BwmD_VLjROrfTHk4NFg2SndKcjQ
The new link now looks li... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/3785/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/3785/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/3785.diff",
"html_url": "https://github.com/huggingface/datasets/pull/3785",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/3785.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3785"
} | true | [
"Thank you, @albertvillanova!",
"Got it. Thanks for explaining this, @albertvillanova!\r\n\r\n> On the other hand, the tests are not passing because the dummy data should also be fixed. Once done, this PR will be able to be merged into master.\r\n\r\nWill do this 👍",
"Hi ! I think we need to fix the issue for ... |
https://api.github.com/repos/huggingface/datasets/issues/2630 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2630/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2630/comments | https://api.github.com/repos/huggingface/datasets/issues/2630/events | https://github.com/huggingface/datasets/issues/2630 | 942,102,956 | MDU6SXNzdWU5NDIxMDI5NTY= | 2,630 | Progress bars are not properly rendered in Jupyter notebook | [
{
"color": "d73a4a",
"default": true,
"description": "Something isn't working",
"id": 1935892857,
"name": "bug",
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug"
}
] | closed | false | null | 2 | 2021-07-12T14:07:13Z | 2022-02-03T15:55:33Z | 2022-02-03T15:55:33Z | null | ## Describe the bug
The progress bars are not Jupyter widgets; regular progress bars appear (like in a terminal).
## Steps to reproduce the bug
```python
ds.map(tokenize, num_proc=10)
```
## Expected results
Jupyter widgets displaying the progress bars.
## Actual results
Simple plane progress bars.
cc... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2630/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/2630/timeline | null | completed | null | null | false | [
"To add my experience when trying to debug this issue:\r\n\r\nSeems like previously the workaround given [here](https://github.com/tqdm/tqdm/issues/485#issuecomment-473338308) worked around this issue. But with the latest version of jupyter/tqdm I still get terminal warnings that IPython tried to send a message fro... |
https://api.github.com/repos/huggingface/datasets/issues/4083 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/4083/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/4083/comments | https://api.github.com/repos/huggingface/datasets/issues/4083/events | https://github.com/huggingface/datasets/pull/4083 | 1,190,025,878 | PR_kwDODunzps41gEbu | 4,083 | Add SacreBLEU Metric Card | [] | closed | false | null | 1 | 2022-04-01T16:24:56Z | 2022-04-12T20:45:00Z | 2022-04-12T20:38:40Z | null | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/4083/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/4083/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/4083.diff",
"html_url": "https://github.com/huggingface/datasets/pull/4083",
"merged_at": "2022-04-12T20:38:40Z",
"patch_url": "https://github.com/huggingface/datasets/pull/4083.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | true | [
"_The documentation is not available anymore as the PR was closed or merged._"
] |
https://api.github.com/repos/huggingface/datasets/issues/4853 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/4853/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/4853/comments | https://api.github.com/repos/huggingface/datasets/issues/4853/events | https://github.com/huggingface/datasets/pull/4853 | 1,339,456,490 | PR_kwDODunzps49NFNL | 4,853 | Fix bug and checksums in exams dataset | [] | closed | false | null | 1 | 2022-08-15T20:17:57Z | 2022-08-16T06:43:57Z | 2022-08-16T06:29:06Z | null | Fix #4852. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 1,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/4853/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/4853/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/4853.diff",
"html_url": "https://github.com/huggingface/datasets/pull/4853",
"merged_at": "2022-08-16T06:29:06Z",
"patch_url": "https://github.com/huggingface/datasets/pull/4853.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | true | [
"_The documentation is not available anymore as the PR was closed or merged._"
] |
https://api.github.com/repos/huggingface/datasets/issues/2075 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2075/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2075/comments | https://api.github.com/repos/huggingface/datasets/issues/2075/events | https://github.com/huggingface/datasets/issues/2075 | 834,301,246 | MDU6SXNzdWU4MzQzMDEyNDY= | 2,075 | ConnectionError: Couldn't reach common_voice.py | [] | closed | false | null | 2 | 2021-03-18T01:19:06Z | 2021-03-20T10:29:41Z | 2021-03-20T10:29:41Z | null | When I run:
from datasets import load_dataset, load_metric
common_voice_train = load_dataset("common_voice", "zh-CN", split="train+validation")
common_voice_test = load_dataset("common_voice", "zh-CN", split="test")
Got:
ConnectionError: Couldn't reach https://raw.githubusercontent.com/huggingface/datasets/ma... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2075/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/2075/timeline | null | completed | null | null | false | [
"Hi @LifaSun, thanks for reporting this issue.\r\n\r\nSometimes, GitHub has some connectivity problems. Could you confirm that the problem persists?",
"@albertvillanova Thanks! It works well now. "
] |
https://api.github.com/repos/huggingface/datasets/issues/1421 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/1421/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/1421/comments | https://api.github.com/repos/huggingface/datasets/issues/1421/events | https://github.com/huggingface/datasets/pull/1421 | 760,706,851 | MDExOlB1bGxSZXF1ZXN0NTM1NDkzMzU4 | 1,421 | adding fake-news-english-2 | [] | closed | false | null | 0 | 2020-12-09T22:05:13Z | 2020-12-13T00:48:49Z | 2020-12-13T00:48:49Z | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/1421/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/1421/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/1421.diff",
"html_url": "https://github.com/huggingface/datasets/pull/1421",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/1421.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/1421"
} | true | [] | |
https://api.github.com/repos/huggingface/datasets/issues/4092 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/4092/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/4092/comments | https://api.github.com/repos/huggingface/datasets/issues/4092/events | https://github.com/huggingface/datasets/pull/4092 | 1,192,499,903 | PR_kwDODunzps41n40R | 4,092 | Fix dataset `amazon_us_reviews` metadata - 4/4/2022 | [] | closed | false | null | 2 | 2022-04-05T01:39:45Z | 2022-04-08T12:35:41Z | 2022-04-08T12:29:31Z | null | Fixes #4048 by running `dataset-cli test` to reprocess data and regenerate metadata. Additionally I've updated the README to include up-to-date counts for the subsets. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/4092/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/4092/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/4092.diff",
"html_url": "https://github.com/huggingface/datasets/pull/4092",
"merged_at": "2022-04-08T12:29:30Z",
"patch_url": "https://github.com/huggingface/datasets/pull/4092.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | true | [
"_The documentation is not available anymore as the PR was closed or merged._",
"cc: @albertvillanova just FYI"
] |
https://api.github.com/repos/huggingface/datasets/issues/1411 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/1411/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/1411/comments | https://api.github.com/repos/huggingface/datasets/issues/1411/events | https://github.com/huggingface/datasets/pull/1411 | 760,606,290 | MDExOlB1bGxSZXF1ZXN0NTM1NDEwNjU3 | 1,411 | 2 typos | [] | closed | false | null | 0 | 2020-12-09T19:24:34Z | 2020-12-11T10:39:05Z | 2020-12-11T10:39:05Z | null | Corrected 2 typos | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/1411/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/1411/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/1411.diff",
"html_url": "https://github.com/huggingface/datasets/pull/1411",
"merged_at": "2020-12-11T10:39:05Z",
"patch_url": "https://github.com/huggingface/datasets/pull/1411.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | true | [] |
https://api.github.com/repos/huggingface/datasets/issues/4453 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/4453/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/4453/comments | https://api.github.com/repos/huggingface/datasets/issues/4453/events | https://github.com/huggingface/datasets/issues/4453 | 1,262,674,105 | I_kwDODunzps5LQuC5 | 4,453 | Dataset Viewer issue for Yaxin/SemEval2015 | [] | closed | false | null | 3 | 2022-06-07T03:30:08Z | 2022-06-09T08:34:16Z | 2022-06-09T08:34:16Z | null | ### Link
_No response_
### Description
_No response_
### Owner
_No response_ | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/4453/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/4453/timeline | null | completed | null | null | false | [
"I understand that the issue is that a remote file (URL) is being loaded as a local file. Right @albertvillanova @lhoestq?\r\n\r\n```\r\nMessage: [Errno 2] No such file or directory: 'https://raw.githubusercontent.com/YaxinCui/ABSADataset/main/SemEval2015Task12Corrected/train/restaurants_train.xml'\r\n```",
... |
https://api.github.com/repos/huggingface/datasets/issues/4460 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/4460/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/4460/comments | https://api.github.com/repos/huggingface/datasets/issues/4460/events | https://github.com/huggingface/datasets/pull/4460 | 1,264,644,205 | PR_kwDODunzps45UHIs | 4,460 | Drop Python 3.6 support | [] | closed | false | null | 5 | 2022-06-08T12:10:18Z | 2022-07-26T19:16:39Z | 2022-07-26T19:04:21Z | null | Remove the fallback imports/checks in the code needed for Python 3.6 and update the requirements/CI files. Also, use Python types for the NumPy dtype wherever possible to avoid deprecation warnings in newer NumPy versions.
| {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 2,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 2,
"url": "https://api.github.com/repos/huggingface/datasets/issues/4460/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/4460/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/4460.diff",
"html_url": "https://github.com/huggingface/datasets/pull/4460",
"merged_at": "2022-07-26T19:04:21Z",
"patch_url": "https://github.com/huggingface/datasets/pull/4460.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | true | [
"_The documentation is not available anymore as the PR was closed or merged._",
"I've disabled the `test_dummy_dataset_serialize_s3` tests in the Linux CI to avoid the failures (these tests only fail on Windows in 3.6). These failures are unrelated to this PR's changes, and I would like to address this in a new P... |
https://api.github.com/repos/huggingface/datasets/issues/1808 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/1808/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/1808/comments | https://api.github.com/repos/huggingface/datasets/issues/1808/events | https://github.com/huggingface/datasets/issues/1808 | 798,879,180 | MDU6SXNzdWU3OTg4NzkxODA= | 1,808 | writing Datasets in a human readable format | [
{
"color": "a2eeef",
"default": true,
"description": "New feature or request",
"id": 1935892871,
"name": "enhancement",
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement"
},
{
"color": "d876e3",
"default": true... | closed | false | null | 3 | 2021-02-02T02:55:40Z | 2022-06-01T15:38:13Z | 2022-06-01T15:38:13Z | null | Hi
I see there is a save_to_disk function to save data, but this is not human readable format, is there a way I could save a Dataset object in a human readable format to a file like json? thanks @lhoestq | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/1808/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/1808/timeline | null | completed | null | null | false | [
"AFAIK, there is currently no built-in method on the `Dataset` object to do this.\r\nHowever, a workaround is to directly use the Arrow table backing the dataset, **but it implies loading the whole dataset in memory** (correct me if I'm mistaken @lhoestq).\r\n\r\nYou can convert the Arrow table to a pandas datafram... |
https://api.github.com/repos/huggingface/datasets/issues/99 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/99/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/99/comments | https://api.github.com/repos/huggingface/datasets/issues/99/events | https://github.com/huggingface/datasets/pull/99 | 618,026,700 | MDExOlB1bGxSZXF1ZXN0NDE3ODMxNjky | 99 | [Cmrc 2018] fix cmrc2018 | [] | closed | false | null | 0 | 2020-05-14T08:22:03Z | 2020-05-14T08:49:42Z | 2020-05-14T08:49:41Z | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/99/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/99/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/99.diff",
"html_url": "https://github.com/huggingface/datasets/pull/99",
"merged_at": "2020-05-14T08:49:41Z",
"patch_url": "https://github.com/huggingface/datasets/pull/99.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/99"
} | true | [] | |
https://api.github.com/repos/huggingface/datasets/issues/3427 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3427/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3427/comments | https://api.github.com/repos/huggingface/datasets/issues/3427/events | https://github.com/huggingface/datasets/pull/3427 | 1,078,782,159 | PR_kwDODunzps4vxb_y | 3,427 | Add The Pile Enron Emails subset | [] | closed | false | null | 0 | 2021-12-13T17:14:16Z | 2021-12-14T17:30:59Z | 2021-12-14T17:30:57Z | null | Add:
- Enron Emails subset of The Pile: "enron_emails" config
Close bigscience-workshop/data_tooling#310.
CC: @StellaAthena | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/3427/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/3427/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/3427.diff",
"html_url": "https://github.com/huggingface/datasets/pull/3427",
"merged_at": "2021-12-14T17:30:55Z",
"patch_url": "https://github.com/huggingface/datasets/pull/3427.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | true | [] |
https://api.github.com/repos/huggingface/datasets/issues/5714 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/5714/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/5714/comments | https://api.github.com/repos/huggingface/datasets/issues/5714/events | https://github.com/huggingface/datasets/pull/5714 | 1,657,388,033 | PR_kwDODunzps5NxIOc | 5,714 | Fix xnumpy_load for .npz files | [] | closed | false | null | 2 | 2023-04-06T13:01:45Z | 2023-04-07T09:23:54Z | 2023-04-07T09:16:57Z | null | PR:
- #5626
implemented support for streaming `.npy` files by using `numpy.load`.
However, it introduced a bug when used with `.npz` files, within a context manager:
```
ValueError: seek of closed file
```
or in streaming mode:
```
ValueError: I/O operation on closed file.
```
This PR fixes the bug an... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/5714/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/5714/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/5714.diff",
"html_url": "https://github.com/huggingface/datasets/pull/5714",
"merged_at": "2023-04-07T09:16:57Z",
"patch_url": "https://github.com/huggingface/datasets/pull/5714.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | true | [
"_The documentation is not available anymore as the PR was closed or merged._",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea... |
https://api.github.com/repos/huggingface/datasets/issues/147 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/147/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/147/comments | https://api.github.com/repos/huggingface/datasets/issues/147/events | https://github.com/huggingface/datasets/issues/147 | 619,581,907 | MDU6SXNzdWU2MTk1ODE5MDc= | 147 | Error with sklearn train_test_split | [
{
"color": "a2eeef",
"default": true,
"description": "New feature or request",
"id": 1935892871,
"name": "enhancement",
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement"
}
] | closed | false | null | 2 | 2020-05-17T00:28:24Z | 2020-06-18T16:23:23Z | 2020-06-18T16:23:23Z | null | It would be nice if we could use sklearn `train_test_split` to quickly generate subsets from the dataset objects returned by `nlp.load_dataset`. At the moment the code:
```python
data = nlp.load_dataset('imdb', cache_dir=data_cache)
f_half, s_half = train_test_split(data['train'], test_size=0.5, random_state=seed)... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/147/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/147/timeline | null | completed | null | null | false | [
"Indeed. Probably we will want to have a similar method directly in the library",
"Related: #166 "
] |
https://api.github.com/repos/huggingface/datasets/issues/2122 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2122/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2122/comments | https://api.github.com/repos/huggingface/datasets/issues/2122/events | https://github.com/huggingface/datasets/pull/2122 | 842,194,588 | MDExOlB1bGxSZXF1ZXN0NjAxODE3MjI0 | 2,122 | Fast table queries with interpolation search | [] | closed | false | null | 0 | 2021-03-26T18:09:20Z | 2021-08-04T18:11:59Z | 2021-04-06T14:33:01Z | null | ## Intro
This should fix issue #1803
Currently querying examples in a dataset is O(n) because of the underlying pyarrow ChunkedArrays implementation.
To fix this I implemented interpolation search that is pretty effective since datasets usually verifies the condition of evenly distributed chunks (the default ch... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 5,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 5,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2122/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/2122/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/2122.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2122",
"merged_at": "2021-04-06T14:33:01Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2122.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | true | [] |
https://api.github.com/repos/huggingface/datasets/issues/2939 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2939/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2939/comments | https://api.github.com/repos/huggingface/datasets/issues/2939/events | https://github.com/huggingface/datasets/pull/2939 | 999,639,630 | PR_kwDODunzps4r58Gu | 2,939 | MENYO-20k repo has moved, updating URL | [] | closed | false | null | 0 | 2021-09-17T19:01:54Z | 2021-09-21T15:31:37Z | 2021-09-21T15:31:36Z | null | Dataset repo moved to https://github.com/uds-lsv/menyo-20k_MT, now editing URL to match.
https://github.com/uds-lsv/menyo-20k_MT/blob/master/data/train.tsv is the file we're looking for | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2939/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/2939/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/2939.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2939",
"merged_at": "2021-09-21T15:31:36Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2939.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | true | [] |
https://api.github.com/repos/huggingface/datasets/issues/344 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/344/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/344/comments | https://api.github.com/repos/huggingface/datasets/issues/344/events | https://github.com/huggingface/datasets/pull/344 | 651,495,246 | MDExOlB1bGxSZXF1ZXN0NDQ0NzQwMTIw | 344 | Search qa | [] | closed | false | null | 1 | 2020-07-06T12:23:16Z | 2020-07-16T08:58:16Z | 2020-07-16T08:58:16Z | null | This PR adds the Search QA dataset used in **SearchQA: A New Q&A Dataset Augmented with Context from a Search Engine**. The dataset has the following config name:
- raw_jeopardy: raw data
- train_test_val: which is the splitted version
#336 | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/344/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/344/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/344.diff",
"html_url": "https://github.com/huggingface/datasets/pull/344",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/344.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/344"
} | true | [
"Could you rebase from master just to make sure we won't break anything for `fever` pls @mariamabarham ?"
] |
https://api.github.com/repos/huggingface/datasets/issues/5075 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/5075/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/5075/comments | https://api.github.com/repos/huggingface/datasets/issues/5075/events | https://github.com/huggingface/datasets/issues/5075 | 1,397,865,501 | I_kwDODunzps5TUbwd | 5,075 | Throw EnvironmentError when token is not present | [
{
"color": "7057ff",
"default": true,
"description": "Good for newcomers",
"id": 1935892877,
"name": "good first issue",
"node_id": "MDU6TGFiZWwxOTM1ODkyODc3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/good%20first%20issue"
},
{
"color": "DF8D62",
"defa... | closed | false | null | 1 | 2022-10-05T14:14:18Z | 2022-10-07T14:33:28Z | 2022-10-07T14:33:28Z | null | Throw EnvironmentError instead of OSError ([link](https://github.com/huggingface/datasets/blob/6ad430ba0cdeeb601170f732d4bd977f5c04594d/src/datasets/arrow_dataset.py#L4306) to the line) in `push_to_hub` when the Hub token is not present. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/5075/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/5075/timeline | null | completed | null | null | false | [
"@mariosasko I've raised a PR #5076 against this issue. Please help to review. Thanks."
] |
https://api.github.com/repos/huggingface/datasets/issues/6076 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6076/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6076/comments | https://api.github.com/repos/huggingface/datasets/issues/6076/events | https://github.com/huggingface/datasets/pull/6076 | 1,822,345,597 | PR_kwDODunzps5WcGVR | 6,076 | No gzip encoding from github | [] | closed | false | null | 3 | 2023-07-26T12:46:07Z | 2023-07-27T16:15:11Z | 2023-07-27T16:14:40Z | null | Don't accept gzip encoding from github, otherwise some files are not streamable + seekable.
fix https://huggingface.co/datasets/code_x_glue_cc_code_to_code_trans/discussions/2#64c0e0c1a04a514ba6303e84
and making sure https://github.com/huggingface/datasets/issues/2918 works as well | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/6076/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/6076/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/6076.diff",
"html_url": "https://github.com/huggingface/datasets/pull/6076",
"merged_at": "2023-07-27T16:14:40Z",
"patch_url": "https://github.com/huggingface/datasets/pull/6076.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | true | [
"_The documentation is not available anymore as the PR was closed or merged._",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea... |
https://api.github.com/repos/huggingface/datasets/issues/740 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/740/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/740/comments | https://api.github.com/repos/huggingface/datasets/issues/740/events | https://github.com/huggingface/datasets/pull/740 | 723,047,958 | MDExOlB1bGxSZXF1ZXN0NTA0NzAyNTc0 | 740 | Fix TREC urls | [] | closed | false | null | 0 | 2020-10-16T09:11:28Z | 2020-10-19T08:54:37Z | 2020-10-19T08:54:36Z | null | The old TREC urls are now redirections.
I updated the urls to the new ones, since we don't support redirections for downloads.
Fix #737 | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/740/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/740/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/740.diff",
"html_url": "https://github.com/huggingface/datasets/pull/740",
"merged_at": "2020-10-19T08:54:35Z",
"patch_url": "https://github.com/huggingface/datasets/pull/740.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/740... | true | [] |
https://api.github.com/repos/huggingface/datasets/issues/3071 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3071/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3071/comments | https://api.github.com/repos/huggingface/datasets/issues/3071/events | https://github.com/huggingface/datasets/issues/3071 | 1,024,893,493 | I_kwDODunzps49FqI1 | 3,071 | Custom plain text dataset, plain json dataset and plain csv dataset are remove from datasets template folder | [
{
"color": "e99695",
"default": false,
"description": "Requesting to add a new dataset",
"id": 2067376369,
"name": "dataset request",
"node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5",
"url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request"
}
] | closed | false | null | 1 | 2021-10-13T07:32:10Z | 2021-10-13T08:27:04Z | 2021-10-13T08:27:03Z | null | ## Adding a Dataset
- **Name:** text, json, csv
- **Description:** I am developing a customized dataset loading script. The problem is mainly about my custom dataset is seperate into many files and I only find a dataset loading template in [https://github.com/huggingface/datasets/blob/1.2.1/datasets/json/json.py](ht... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/3071/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/3071/timeline | null | completed | null | null | false | [
"Hi @zixiliuUSC, \r\n\r\nAs explained in the documentation (https://huggingface.co/docs/datasets/loading.html#json), we support loading any dataset in JSON (as well as CSV, text, Parquet) format:\r\n```python\r\nds = load_dataset('json', data_files='my_file.json')\r\n```"
] |
https://api.github.com/repos/huggingface/datasets/issues/5158 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/5158/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/5158/comments | https://api.github.com/repos/huggingface/datasets/issues/5158/events | https://github.com/huggingface/datasets/issues/5158 | 1,422,059,287 | I_kwDODunzps5UwucX | 5,158 | Fix language and license tag names in all Hub datasets | [
{
"color": "0e8a16",
"default": false,
"description": "Contribution to a dataset script",
"id": 4564477500,
"name": "dataset contribution",
"node_id": "LA_kwDODunzps8AAAABEBBmPA",
"url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20contribution"
}
] | closed | false | null | 6 | 2022-10-25T08:19:29Z | 2022-10-25T11:27:26Z | 2022-10-25T10:42:19Z | null | While working on this:
- #5137
we realized there are still many datasets with deprecated "languages" and "licenses" tag names (instead of "language" and "license").
This is a blocking issue: no subsequent PR can be opened to modify their metadata: a ValueError will be thrown.
We should fix the "language" and ... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/5158/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/5158/timeline | null | completed | null | null | false | [
"There are currently 402 datasets with deprecated \"languages\" or \"licenses\".",
"hey @albertvillanova ,i would love to work on this issue if you like.",
"Hi @ayushthe1, thanks for your offer.\r\n\r\nBut as you can see, I self-assigned this issue.\r\n\r\nI have already fixed 200 out of the 402 datasets. My sc... |
https://api.github.com/repos/huggingface/datasets/issues/2175 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2175/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2175/comments | https://api.github.com/repos/huggingface/datasets/issues/2175/events | https://github.com/huggingface/datasets/issues/2175 | 851,836,096 | MDU6SXNzdWU4NTE4MzYwOTY= | 2,175 | dataset.search_batch() function outputs all -1 indices sometime. | [] | closed | false | null | 6 | 2021-04-06T21:50:49Z | 2021-04-16T12:21:16Z | 2021-04-16T12:21:15Z | null | I am working with RAG and playing around with different faiss indexes. At the moment I use **index = faiss.index_factory(768, "IVF65536_HNSW32,Flat")**.
During the retrieval phase exactly in [this line of retrieval_rag.py](https://github.com/huggingface/transformers/blob/master/src/transformers/models/rag/retrieval_... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2175/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/2175/timeline | null | completed | null | null | false | [
"Actually, I found the answer [here](https://github.com/facebookresearch/faiss/wiki/FAQ#what-does-it-mean-when-a-search-returns--1-ids). \r\n\r\nSo we have to do some modifications to the code for instances where the index doesn't retrieve any IDs.",
"@lhoestq @patrickvonplaten \r\n\r\nI also found another short... |
https://api.github.com/repos/huggingface/datasets/issues/835 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/835/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/835/comments | https://api.github.com/repos/huggingface/datasets/issues/835/events | https://github.com/huggingface/datasets/issues/835 | 740,102,210 | MDU6SXNzdWU3NDAxMDIyMTA= | 835 | Wikipedia postprocessing | [] | closed | false | null | 3 | 2020-11-10T17:26:38Z | 2020-11-10T18:23:20Z | 2020-11-10T17:49:21Z | null | Hi, thanks for this library!
Running this code:
```py
import datasets
wikipedia = datasets.load_dataset("wikipedia", "20200501.de")
print(wikipedia['train']['text'][0])
```
I get:
```
mini|Ricardo Flores Magón
mini|Mexikanische Revolutionäre, Magón in der Mitte anführend, gegen die Diktatur von Porfir... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/835/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/835/timeline | null | completed | null | null | false | [
"Hi @bminixhofer ! Parsing WikiMedia is notoriously difficult: this processing used [mwparserfromhell](https://github.com/earwig/mwparserfromhell) which is pretty good but not perfect.\r\n\r\nAs an alternative, you can also use the Wiki40b dataset which was pre-processed using an un-released Google internal tool",
... |
https://api.github.com/repos/huggingface/datasets/issues/3545 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3545/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3545/comments | https://api.github.com/repos/huggingface/datasets/issues/3545/events | https://github.com/huggingface/datasets/pull/3545 | 1,096,189,889 | PR_kwDODunzps4wpziv | 3,545 | fix: 🐛 pass token when retrieving the split names | [] | closed | false | null | 3 | 2022-01-07T10:29:22Z | 2022-01-10T10:51:47Z | 2022-01-10T10:51:46Z | null | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/3545/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/3545/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/3545.diff",
"html_url": "https://github.com/huggingface/datasets/pull/3545",
"merged_at": "2022-01-10T10:51:46Z",
"patch_url": "https://github.com/huggingface/datasets/pull/3545.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | true | [
"Currently, it does not work with https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0/blob/main/common_voice_7_0.py#L146 (which was the goal), because `dl_manager.download_config.use_auth_token` is ignored, and the authentication is required to be use `huggingface-cli login`.\r\nIn my use case (data... |
https://api.github.com/repos/huggingface/datasets/issues/3459 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3459/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3459/comments | https://api.github.com/repos/huggingface/datasets/issues/3459/events | https://github.com/huggingface/datasets/issues/3459 | 1,084,969,672 | I_kwDODunzps5Aq1LI | 3,459 | dataset.filter overwriting previously set dataset._indices values, resulting in the wrong elements being selected. | [
{
"color": "d73a4a",
"default": true,
"description": "Something isn't working",
"id": 1935892857,
"name": "bug",
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug"
}
] | closed | false | null | 2 | 2021-12-20T16:16:49Z | 2021-12-20T16:34:57Z | 2021-12-20T16:34:57Z | null | ## Describe the bug
When using dataset.select to select a subset of a dataset, dataset._indices are set to indicate which elements are now considered in the dataset.
The same thing happens when you shuffle the dataset; dataset._indices are set to indicate what the new order of the data is.
However, if you then use a... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/3459/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/3459/timeline | null | completed | null | null | false | [
"I think this is a duplicate of [#3190](https://github.com/huggingface/datasets/issues/3190)?",
"Upgrading the datasets version as per #3190 fixes this bug. \r\nI'm Marking as closed."
] |
https://api.github.com/repos/huggingface/datasets/issues/1752 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/1752/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/1752/comments | https://api.github.com/repos/huggingface/datasets/issues/1752/events | https://github.com/huggingface/datasets/pull/1752 | 789,822,459 | MDExOlB1bGxSZXF1ZXN0NTU4MTA5NTA5 | 1,752 | COMET metric citation | [] | closed | false | null | 1 | 2021-01-20T09:54:43Z | 2021-01-20T10:27:07Z | 2021-01-20T10:25:02Z | null | In my last pull request to add COMET metric, the citations where not following the usual "format". Because of that they where not correctly displayed on the website:
<img width="814" alt="Screenshot 2021-01-20 at 09 48 44" src="https://user-images.githubusercontent.com/17256847/105158000-686efb80-5b05-11eb-8bb0-9c8... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/1752/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/1752/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/1752.diff",
"html_url": "https://github.com/huggingface/datasets/pull/1752",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/1752.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/1752"
} | true | [
"I think its better to create a new branch with this fix. I forgot I was still using the old branch."
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.