url stringlengths 61 61 | repository_url stringclasses 1
value | labels_url stringlengths 75 75 | comments_url stringlengths 70 70 | events_url stringlengths 68 68 | html_url stringlengths 49 51 | id int64 1.23B 2.21B | node_id stringlengths 18 19 | number int64 4.29k 6.76k | title stringlengths 1 290 | user dict | labels listlengths 0 4 | state stringclasses 2
values | locked bool 1
class | assignee dict | assignees listlengths 0 3 | milestone dict | comments int64 0 48 | created_at timestamp[ms] | updated_at timestamp[ms] | closed_at timestamp[ms] | author_association stringclasses 3
values | active_lock_reason null | body stringlengths 2 33.9k ⌀ | reactions dict | timeline_url stringlengths 70 70 | performed_via_github_app null | state_reason stringclasses 3
values | draft bool 2
classes | pull_request dict | is_pull_request bool 2
classes | comments_body listlengths 0 30 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
https://api.github.com/repos/huggingface/datasets/issues/6657 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6657/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6657/comments | https://api.github.com/repos/huggingface/datasets/issues/6657/events | https://github.com/huggingface/datasets/issues/6657 | 2,129,147,085 | I_kwDODunzps5-6DTN | 6,657 | Release not pushed to conda channel | {
"login": "atulsaurav",
"id": 7138162,
"node_id": "MDQ6VXNlcjcxMzgxNjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/7138162?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/atulsaurav",
"html_url": "https://github.com/atulsaurav",
"followers_url": "https://api.github.com/users... | [] | closed | false | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_... | null | 5 | 2024-02-11T20:05:17 | 2024-03-06T15:06:22 | 2024-03-06T15:06:22 | NONE | null | ### Describe the bug
The github actions step to publish the release 2.17.0 to conda channel has failed due to expired token. Can some one please update the anaconda token rerun the failed action? @albertvillanova ?
:
File "/miniconda3/envs/conda-env/lib/python3.10/site-packages/datasets/builder.py", line 1989, in _prepare_split_single
writer.write_table(table)
... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6656/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6656/timeline | null | null | null | null | false | [
"I get similar when dealing with a large jsonl file (6k lines), \r\n\r\n> TypeError: Couldn't cast array of type timestamp[us] to null\r\n\r\nYet when I split it into 1k lines, files, load_dataset works fine!\r\n\r\nhttps://github.com/huggingface/course/issues/692\r\n\r\n"
] |
https://api.github.com/repos/huggingface/datasets/issues/6655 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6655/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6655/comments | https://api.github.com/repos/huggingface/datasets/issues/6655/events | https://github.com/huggingface/datasets/issues/6655 | 2,127,020,042 | I_kwDODunzps5-x8AK | 6,655 | Cannot load the dataset go_emotions | {
"login": "arame",
"id": 688324,
"node_id": "MDQ6VXNlcjY4ODMyNA==",
"avatar_url": "https://avatars.githubusercontent.com/u/688324?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/arame",
"html_url": "https://github.com/arame",
"followers_url": "https://api.github.com/users/arame/followers"... | [] | open | false | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_... | null | 4 | 2024-02-09T12:15:39 | 2024-02-12T09:35:55 | null | NONE | null | ### Describe the bug
When I run the following code I get an exception;
`go_emotions = load_dataset("go_emotions")`
> AttributeError Traceback (most recent call last)
Cell In[6], [line 1](vscode-notebook-cell:?execution_count=6&line=1)
----> [1](vscode-notebook-cell:?execution_count=6&l... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6655/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6655/timeline | null | null | null | null | false | [
"Thanks for reporting, @arame.\r\n\r\nI guess you have an old version of `transformers` (that submodule is present in `transformers` since version 3.0.1, since nearly 4 years ago). If you update it, the error should disappear:\r\n```shell\r\npip install -U transformers\r\n```\r\n\r\nOn the other hand, I am wonderin... |
https://api.github.com/repos/huggingface/datasets/issues/6654 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6654/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6654/comments | https://api.github.com/repos/huggingface/datasets/issues/6654/events | https://github.com/huggingface/datasets/issues/6654 | 2,126,939,358 | I_kwDODunzps5-xoTe | 6,654 | Batched dataset map throws exception that cannot cast fixed length array to Sequence | {
"login": "keesjandevries",
"id": 1029671,
"node_id": "MDQ6VXNlcjEwMjk2NzE=",
"avatar_url": "https://avatars.githubusercontent.com/u/1029671?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/keesjandevries",
"html_url": "https://github.com/keesjandevries",
"followers_url": "https://api.gith... | [] | closed | false | null | [] | null | 2 | 2024-02-09T11:23:19 | 2024-02-12T08:26:53 | 2024-02-12T08:26:53 | NONE | null | ### Describe the bug
I encountered a TypeError when batch processing a dataset with Sequence features in datasets package version 2.16.1. The error arises from a mismatch in handling fixed-size list arrays during the map function execution. Debugging pinpoints the issue to an if-statement in datasets/table.py, line 20... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6654/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6654/timeline | null | completed | null | null | false | [
"Hi ! This issue has been fixed by https://github.com/huggingface/datasets/pull/6283\r\n\r\nCan you try again with the new release 2.17.0 ?\r\n\r\n```\r\npip install -U datasets\r\n```\r\n\r\n",
"Amazing! It's indeed fixed now. Thanks!"
] |
https://api.github.com/repos/huggingface/datasets/issues/6653 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6653/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6653/comments | https://api.github.com/repos/huggingface/datasets/issues/6653/events | https://github.com/huggingface/datasets/pull/6653 | 2,126,831,929 | PR_kwDODunzps5mdv5S | 6,653 | Set dev version | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | 2 | 2024-02-09T10:12:02 | 2024-02-09T10:18:20 | 2024-02-09T10:12:12 | MEMBER | null | null | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6653/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6653/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6653",
"html_url": "https://github.com/huggingface/datasets/pull/6653",
"diff_url": "https://github.com/huggingface/datasets/pull/6653.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6653.patch",
"merged_at": "2024-02-09T10:12... | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6653). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>... |
https://api.github.com/repos/huggingface/datasets/issues/6652 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6652/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6652/comments | https://api.github.com/repos/huggingface/datasets/issues/6652/events | https://github.com/huggingface/datasets/pull/6652 | 2,126,760,798 | PR_kwDODunzps5mdgcv | 6,652 | Release: 2.17.0 | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | 2 | 2024-02-09T09:25:01 | 2024-02-09T10:11:48 | 2024-02-09T10:05:35 | MEMBER | null | null | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6652/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6652/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6652",
"html_url": "https://github.com/huggingface/datasets/pull/6652",
"diff_url": "https://github.com/huggingface/datasets/pull/6652.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6652.patch",
"merged_at": "2024-02-09T10:05... | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6652). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>... |
https://api.github.com/repos/huggingface/datasets/issues/6651 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6651/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6651/comments | https://api.github.com/repos/huggingface/datasets/issues/6651/events | https://github.com/huggingface/datasets/issues/6651 | 2,126,649,626 | I_kwDODunzps5-whka | 6,651 | Slice splits support for datasets.load_from_disk | {
"login": "mhorlacher",
"id": 37439882,
"node_id": "MDQ6VXNlcjM3NDM5ODgy",
"avatar_url": "https://avatars.githubusercontent.com/u/37439882?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mhorlacher",
"html_url": "https://github.com/mhorlacher",
"followers_url": "https://api.github.com/use... | [
{
"id": 1935892871,
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement",
"name": "enhancement",
"color": "a2eeef",
"default": true,
"description": "New feature or request"
}
] | open | false | null | [] | null | 0 | 2024-02-09T08:00:21 | 2024-02-09T08:00:21 | null | NONE | null | ### Feature request
Support for slice splits in `datasets.load_from_disk`, similar to how it's already supported for `datasets.load_dataset`. See https://www.nature.com/articles/s41551-023-01093-3.
### Motivation
Slice splits are convienient in a numer of cases - adding support to `datasets.load_from_disk` would ma... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6651/reactions",
"total_count": 3,
"+1": 3,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6651/timeline | null | null | null | null | false | [] |
https://api.github.com/repos/huggingface/datasets/issues/6650 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6650/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6650/comments | https://api.github.com/repos/huggingface/datasets/issues/6650/events | https://github.com/huggingface/datasets/issues/6650 | 2,125,680,991 | I_kwDODunzps5-s1Ff | 6,650 | AttributeError: 'InMemoryTable' object has no attribute '_batches' | {
"login": "matsuobasho",
"id": 13874772,
"node_id": "MDQ6VXNlcjEzODc0Nzcy",
"avatar_url": "https://avatars.githubusercontent.com/u/13874772?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/matsuobasho",
"html_url": "https://github.com/matsuobasho",
"followers_url": "https://api.github.com/... | [] | open | false | null | [] | null | 3 | 2024-02-08T17:11:26 | 2024-02-21T00:34:41 | null | NONE | null | ### Describe the bug
```
Traceback (most recent call last):
File "finetune.py", line 103, in <module>
main(args)
File "finetune.py", line 45, in main
data_tokenized = data.map(partial(funcs.tokenize_function, tokenizer,
File "/opt/conda/envs/ptca/lib/python3.8/site-packages/datasets/dataset_dict.... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6650/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6650/timeline | null | null | null | null | false | [
"Hi! Does running the following code also return the same error on your machine? \r\n\r\n```python\r\nimport copy\r\nimport pyarrow as pa\r\nfrom datasets.table import InMemoryTable\r\n\r\ncopy.deepcopy(InMemoryTable(pa.table({\"a\": [1, 2, 3], \"b\": [\"foo\", \"bar\", \"foobar\"]})))\r\n```",
"No, it doesn't, ... |
https://api.github.com/repos/huggingface/datasets/issues/6649 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6649/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6649/comments | https://api.github.com/repos/huggingface/datasets/issues/6649/events | https://github.com/huggingface/datasets/pull/6649 | 2,124,940,213 | PR_kwDODunzps5mXRo8 | 6,649 | Minor multi gpu doc improvement | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [] | closed | false | null | [] | null | 2 | 2024-02-08T11:17:24 | 2024-02-08T11:23:35 | 2024-02-08T11:17:35 | MEMBER | null | just added torch.no_grad and eval() | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6649/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6649/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6649",
"html_url": "https://github.com/huggingface/datasets/pull/6649",
"diff_url": "https://github.com/huggingface/datasets/pull/6649.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6649.patch",
"merged_at": "2024-02-08T11:17... | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6649). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>... |
https://api.github.com/repos/huggingface/datasets/issues/6648 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6648/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6648/comments | https://api.github.com/repos/huggingface/datasets/issues/6648/events | https://github.com/huggingface/datasets/pull/6648 | 2,124,813,589 | PR_kwDODunzps5mW1MA | 6,648 | Document usage of hfh cli instead of git | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [] | closed | false | null | [] | null | 2 | 2024-02-08T10:24:56 | 2024-02-08T13:57:41 | 2024-02-08T13:51:39 | MEMBER | null | (basically the same content as the hfh upload docs, but adapted for datasets) | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6648/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6648/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6648",
"html_url": "https://github.com/huggingface/datasets/pull/6648",
"diff_url": "https://github.com/huggingface/datasets/pull/6648.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6648.patch",
"merged_at": "2024-02-08T13:51... | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6648). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>... |
https://api.github.com/repos/huggingface/datasets/issues/6647 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6647/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6647/comments | https://api.github.com/repos/huggingface/datasets/issues/6647/events | https://github.com/huggingface/datasets/pull/6647 | 2,123,397,569 | PR_kwDODunzps5mSB2B | 6,647 | Update loading.mdx to include "jsonl" file loading. | {
"login": "mosheber",
"id": 22236370,
"node_id": "MDQ6VXNlcjIyMjM2Mzcw",
"avatar_url": "https://avatars.githubusercontent.com/u/22236370?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mosheber",
"html_url": "https://github.com/mosheber",
"followers_url": "https://api.github.com/users/mos... | [] | open | false | null | [] | null | 2 | 2024-02-07T16:18:08 | 2024-02-08T15:34:17 | null | NONE | null | * A small update to the documentation, noting the ability to load jsonl files. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6647/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6647/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6647",
"html_url": "https://github.com/huggingface/datasets/pull/6647",
"diff_url": "https://github.com/huggingface/datasets/pull/6647.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6647.patch",
"merged_at": null
} | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6647). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"> Thanks for adding the explicit loading command.\r\n> \r\n> However, I would move it j... |
https://api.github.com/repos/huggingface/datasets/issues/6646 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6646/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6646/comments | https://api.github.com/repos/huggingface/datasets/issues/6646/events | https://github.com/huggingface/datasets/pull/6646 | 2,123,134,128 | PR_kwDODunzps5mRIma | 6,646 | Better multi-gpu example | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [] | closed | false | null | [] | null | 3 | 2024-02-07T14:15:01 | 2024-02-09T17:43:32 | 2024-02-07T14:59:11 | MEMBER | null | Use Qwen1.5-0.5B-Chat as an easy example for multi-GPU
the previous example was using a model for translation and the way it was setup was not really the right way to use the model. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6646/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6646/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6646",
"html_url": "https://github.com/huggingface/datasets/pull/6646",
"diff_url": "https://github.com/huggingface/datasets/pull/6646.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6646.patch",
"merged_at": "2024-02-07T14:59... | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6646). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>... |
https://api.github.com/repos/huggingface/datasets/issues/6645 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6645/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6645/comments | https://api.github.com/repos/huggingface/datasets/issues/6645/events | https://github.com/huggingface/datasets/issues/6645 | 2,122,956,818 | I_kwDODunzps5-icAS | 6,645 | Support fsspec 2024.2 | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"id": 1935892871,
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement",
"name": "enhancement",
"color": "a2eeef",
"default": true,
"description": "New feature or request"
}
] | closed | false | null | [] | null | 1 | 2024-02-07T12:45:29 | 2024-02-29T15:12:19 | 2024-02-29T15:12:19 | MEMBER | null | Support fsspec 2024.2.
First, we should address:
- #6644 | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6645/reactions",
"total_count": 8,
"+1": 8,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6645/timeline | null | completed | null | null | false | [
"I'd be very grateful. This upper bound banished me straight into dependency hell today. :("
] |
https://api.github.com/repos/huggingface/datasets/issues/6644 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6644/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6644/comments | https://api.github.com/repos/huggingface/datasets/issues/6644/events | https://github.com/huggingface/datasets/issues/6644 | 2,122,955,282 | I_kwDODunzps5-iboS | 6,644 | Support fsspec 2023.12 | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"id": 1935892871,
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement",
"name": "enhancement",
"color": "a2eeef",
"default": true,
"description": "New feature or request"
}
] | closed | false | null | [] | null | 1 | 2024-02-07T12:44:39 | 2024-02-29T15:12:18 | 2024-02-29T15:12:18 | MEMBER | null | Support fsspec 2023.12 by handling previous and new glob behavior. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6644/reactions",
"total_count": 6,
"+1": 6,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6644/timeline | null | completed | null | null | false | [
"The pinned fsspec version range dependency conflict has been affecting several of our users in https://github.com/iterative/dvc. I've opened an initial PR that I think should resolve the glob behavior changes with using datasets + the latest fsspec release.\r\n\r\nPlease let us know if there's any other fsspec rel... |
https://api.github.com/repos/huggingface/datasets/issues/6643 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6643/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6643/comments | https://api.github.com/repos/huggingface/datasets/issues/6643/events | https://github.com/huggingface/datasets/issues/6643 | 2,121,239,039 | I_kwDODunzps5-b4n_ | 6,643 | Faiss GPU index cannot be serialised when passed to trainer | {
"login": "rubenweitzman",
"id": 56388976,
"node_id": "MDQ6VXNlcjU2Mzg4OTc2",
"avatar_url": "https://avatars.githubusercontent.com/u/56388976?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rubenweitzman",
"html_url": "https://github.com/rubenweitzman",
"followers_url": "https://api.githu... | [] | open | false | null | [] | null | 3 | 2024-02-06T16:41:00 | 2024-02-15T10:29:32 | null | NONE | null | ### Describe the bug
I am working on a retrieval project and encountering I have encountered two issues in the hugging face faiss integration:
1. I am trying to pass in a dataset with a faiss index to the Huggingface trainer. The code works for a cpu faiss index, but doesn't for a gpu one, getting error:
```
... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6643/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6643/timeline | null | null | null | null | false | [
"Hi ! make sure your query embeddings are numpy arrays, not torch tensors ;)",
"Hi Quentin, not sure how that solves the problem number 1. I am trying to pass on a dataset with a faiss gpu for training to the standard trainer but getting this serialisation error. What is a workaround this? I do not want to remove... |
https://api.github.com/repos/huggingface/datasets/issues/6642 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6642/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6642/comments | https://api.github.com/repos/huggingface/datasets/issues/6642/events | https://github.com/huggingface/datasets/issues/6642 | 2,119,085,766 | I_kwDODunzps5-Tq7G | 6,642 | Differently dataset object saved than it is loaded. | {
"login": "MFajcik",
"id": 31218150,
"node_id": "MDQ6VXNlcjMxMjE4MTUw",
"avatar_url": "https://avatars.githubusercontent.com/u/31218150?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/MFajcik",
"html_url": "https://github.com/MFajcik",
"followers_url": "https://api.github.com/users/MFajci... | [] | closed | false | null | [] | null | 2 | 2024-02-05T17:28:57 | 2024-02-06T09:50:19 | 2024-02-06T09:50:19 | NONE | null | ### Describe the bug
Differently sized object is saved than it is loaded.
### Steps to reproduce the bug
Hi, I save dataset in a following way:
```
dataset = load_dataset("json",
data_files={
"train": os.path.join(input_folder, f"{task_met... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6642/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6642/timeline | null | completed | null | null | false | [
"I see now, that I have to use `load_from_disk`, in order to load dataset properly, not `load_dataset`. Why is this behavior split? Why do we need both, `load_dataset` and `load_from_disk`?\r\n\r\nUnless answered, I believe this might be helpful for other hf datasets newbies.\r\n\r\nAnyway, made a `load_dataset` co... |
https://api.github.com/repos/huggingface/datasets/issues/6641 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6641/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6641/comments | https://api.github.com/repos/huggingface/datasets/issues/6641/events | https://github.com/huggingface/datasets/issues/6641 | 2,116,963,132 | I_kwDODunzps5-Lks8 | 6,641 | unicodedecodeerror: 'utf-8' codec can't decode byte 0xac in position 25: invalid start byte | {
"login": "Hughhuh",
"id": 109789057,
"node_id": "U_kgDOBos_gQ",
"avatar_url": "https://avatars.githubusercontent.com/u/109789057?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Hughhuh",
"html_url": "https://github.com/Hughhuh",
"followers_url": "https://api.github.com/users/Hughhuh/foll... | [] | closed | false | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_... | null | 1 | 2024-02-04T08:49:31 | 2024-02-06T09:26:07 | 2024-02-06T09:11:45 | NONE | null | ### Describe the bug
unicodedecodeerror: 'utf-8' codec can't decode byte 0xac in position 25: invalid start byte
### Steps to reproduce the bug
```
import sys
sys.getdefaultencoding()
'utf-8'
from datasets import load_dataset
print(f"Train dataset size: {len(dataset['train'])}")
print(f"Test datase... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6641/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6641/timeline | null | not_planned | null | null | false | [
"Hi @Hughhuh. \r\n\r\nI have formatted the issue because it was not easily readable. Additionally, the environment info is incomplete: it seems you did not run the proposed CLI command `datasets-cli env` and essential information is missing: version of `datasets`, version of `pyarrow`,...\r\n\r\nWith the informatio... |
https://api.github.com/repos/huggingface/datasets/issues/6640 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6640/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6640/comments | https://api.github.com/repos/huggingface/datasets/issues/6640/events | https://github.com/huggingface/datasets/issues/6640 | 2,115,864,531 | I_kwDODunzps5-HYfT | 6,640 | Sign Language Support | {
"login": "Merterm",
"id": 6684795,
"node_id": "MDQ6VXNlcjY2ODQ3OTU=",
"avatar_url": "https://avatars.githubusercontent.com/u/6684795?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Merterm",
"html_url": "https://github.com/Merterm",
"followers_url": "https://api.github.com/users/Merterm/... | [
{
"id": 1935892871,
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement",
"name": "enhancement",
"color": "a2eeef",
"default": true,
"description": "New feature or request"
}
] | open | false | null | [] | null | 0 | 2024-02-02T21:54:51 | 2024-02-02T21:54:51 | null | NONE | null | ### Feature request
Currently, there are only several Sign Language labels, I would like to propose adding all the Signed Languages as new labels which are described in this ISO standard: https://www.evertype.com/standards/iso639/sign-language.html
### Motivation
Datasets currently only have labels for several signe... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6640/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6640/timeline | null | null | null | null | false | [] |
https://api.github.com/repos/huggingface/datasets/issues/6639 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6639/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6639/comments | https://api.github.com/repos/huggingface/datasets/issues/6639/events | https://github.com/huggingface/datasets/pull/6639 | 2,114,620,200 | PR_kwDODunzps5l0KPG | 6,639 | Run download_and_prepare if missing splits | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [] | open | false | null | [] | null | 1 | 2024-02-02T10:36:49 | 2024-02-06T16:54:22 | null | MEMBER | null | A first step towards https://github.com/huggingface/datasets/issues/6529 | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6639/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6639/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6639",
"html_url": "https://github.com/huggingface/datasets/pull/6639",
"diff_url": "https://github.com/huggingface/datasets/pull/6639.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6639.patch",
"merged_at": null
} | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6639). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] |
https://api.github.com/repos/huggingface/datasets/issues/6638 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6638/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6638/comments | https://api.github.com/repos/huggingface/datasets/issues/6638/events | https://github.com/huggingface/datasets/issues/6638 | 2,113,329,257 | I_kwDODunzps599thp | 6,638 | Cannot download wmt16 dataset | {
"login": "vidyasiv",
"id": 81709031,
"node_id": "MDQ6VXNlcjgxNzA5MDMx",
"avatar_url": "https://avatars.githubusercontent.com/u/81709031?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/vidyasiv",
"html_url": "https://github.com/vidyasiv",
"followers_url": "https://api.github.com/users/vid... | [] | closed | false | null | [] | null | 1 | 2024-02-01T19:41:42 | 2024-02-01T20:07:29 | 2024-02-01T20:07:29 | NONE | null | ### Describe the bug
As of this morning (PST) 2/1/2024, seeing the wmt16 dataset is missing from opus , could you suggest an alternative?
```
Downloading data files: 0%| | 0/4 [00:00<?, ?it/s]Tra... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6638/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6638/timeline | null | completed | null | null | false | [
"Looks like it works with latest datasets repository\r\n```\r\n- `datasets` version: 2.16.2.dev0\r\n- Platform: Linux-5.15.0-92-generic-x86_64-with-glibc2.29\r\n- Python version: 3.8.10\r\n- `huggingface_hub` version: 0.20.3\r\n- PyArrow version: 15.0.0\r\n- Pandas version: 2.0.1\r\n- `fsspec` version: 2023.10.0\r\... |
https://api.github.com/repos/huggingface/datasets/issues/6637 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6637/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6637/comments | https://api.github.com/repos/huggingface/datasets/issues/6637/events | https://github.com/huggingface/datasets/issues/6637 | 2,113,025,975 | I_kwDODunzps598je3 | 6,637 | 'with_format' is extremely slow when used together with 'interleave_datasets' or 'shuffle' on IterableDatasets | {
"login": "tobycrisford",
"id": 22883190,
"node_id": "MDQ6VXNlcjIyODgzMTkw",
"avatar_url": "https://avatars.githubusercontent.com/u/22883190?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/tobycrisford",
"html_url": "https://github.com/tobycrisford",
"followers_url": "https://api.github.c... | [] | open | false | null | [] | null | 1 | 2024-02-01T17:16:54 | 2024-02-05T10:43:47 | null | NONE | null | ### Describe the bug
If you:
1. Interleave two iterable datasets together with the interleave_datasets function, or shuffle an iterable dataset
2. Set the output format to torch tensors with .with_format('torch')
Then iterating through the dataset becomes over 100x slower than it is if you don't apply the torch... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6637/reactions",
"total_count": 1,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 1
} | https://api.github.com/repos/huggingface/datasets/issues/6637/timeline | null | null | null | null | false | [
"The \"torch\" formatting is usually fast because we do zero-copy conversion from the Arrow data on your disk to Torch tensors. However IterableDataset shuffling seems to do data copies that slow down the pipeline, and it shuffles python objects instead of Arrow data.\r\n\r\nTo fix this we need to implement `Buffer... |
https://api.github.com/repos/huggingface/datasets/issues/6636 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6636/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6636/comments | https://api.github.com/repos/huggingface/datasets/issues/6636/events | https://github.com/huggingface/datasets/pull/6636 | 2,110,781,097 | PR_kwDODunzps5lm4zI | 6,636 | Faster column validation and reordering | {
"login": "psmyth94",
"id": 11325244,
"node_id": "MDQ6VXNlcjExMzI1MjQ0",
"avatar_url": "https://avatars.githubusercontent.com/u/11325244?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/psmyth94",
"html_url": "https://github.com/psmyth94",
"followers_url": "https://api.github.com/users/psm... | [] | closed | false | null | [] | null | 3 | 2024-01-31T19:08:28 | 2024-02-07T19:39:00 | 2024-02-06T23:03:38 | CONTRIBUTOR | null | I work with bioinformatics data and often these tables have thousands and even tens of thousands of features. These tables are also accompanied by metadata that I do not want to pass in the model. When I perform `set_format('pt', columns=large_column_list)` , it can take several minutes before it finishes. The culprit ... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6636/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6636/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6636",
"html_url": "https://github.com/huggingface/datasets/pull/6636",
"diff_url": "https://github.com/huggingface/datasets/pull/6636.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6636.patch",
"merged_at": "2024-02-06T23:03... | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6636). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"Thanks @mariosasko, I made the changes. However, I did some tests with `map` and I stil... |
https://api.github.com/repos/huggingface/datasets/issues/6635 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6635/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6635/comments | https://api.github.com/repos/huggingface/datasets/issues/6635/events | https://github.com/huggingface/datasets/pull/6635 | 2,110,659,519 | PR_kwDODunzps5lmeNO | 6,635 | Fix missing info when loading some datasets from Parquet export | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [] | closed | false | null | [] | null | 2 | 2024-01-31T17:55:21 | 2024-02-07T16:48:55 | 2024-02-07T16:41:04 | MEMBER | null | Fix getting the info for script-based datasets with Parquet export with a single config not named "default".
E.g.
```python
from datasets import load_dataset_builder
b = load_dataset_builder("bookcorpus")
print(b.info.features)
# should print {'text': Value(dtype='string', id=None)}
```
I fixed this by ... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6635/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6635/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6635",
"html_url": "https://github.com/huggingface/datasets/pull/6635",
"diff_url": "https://github.com/huggingface/datasets/pull/6635.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6635.patch",
"merged_at": "2024-02-07T16:41... | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6635). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>... |
https://api.github.com/repos/huggingface/datasets/issues/6634 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6634/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6634/comments | https://api.github.com/repos/huggingface/datasets/issues/6634/events | https://github.com/huggingface/datasets/pull/6634 | 2,110,242,376 | PR_kwDODunzps5llB9a | 6,634 | Support data_dir parameter in push_to_hub | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | 3 | 2024-01-31T14:37:36 | 2024-02-05T10:32:49 | 2024-02-05T10:26:40 | MEMBER | null | Support `data_dir` parameter in `push_to_hub`.
This allows users to organize the data files according to their specific needs. For example, "wikimedia/wikipedia" files could be organized by year and/or date, e.g. "2024/20240101/20240101.en". | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6634/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6634/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6634",
"html_url": "https://github.com/huggingface/datasets/pull/6634",
"diff_url": "https://github.com/huggingface/datasets/pull/6634.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6634.patch",
"merged_at": "2024-02-05T10:26... | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6634). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"@huggingface/datasets, feel free to review this PR so that it can be included in the ne... |
https://api.github.com/repos/huggingface/datasets/issues/6633 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6633/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6633/comments | https://api.github.com/repos/huggingface/datasets/issues/6633/events | https://github.com/huggingface/datasets/pull/6633 | 2,110,124,475 | PR_kwDODunzps5lknz9 | 6,633 | dataset viewer requires no-script | {
"login": "severo",
"id": 1676121,
"node_id": "MDQ6VXNlcjE2NzYxMjE=",
"avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/severo",
"html_url": "https://github.com/severo",
"followers_url": "https://api.github.com/users/severo/foll... | [] | closed | false | null | [] | null | 2 | 2024-01-31T13:41:54 | 2024-01-31T14:05:04 | 2024-01-31T13:59:01 | CONTRIBUTOR | null | null | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6633/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6633/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6633",
"html_url": "https://github.com/huggingface/datasets/pull/6633",
"diff_url": "https://github.com/huggingface/datasets/pull/6633.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6633.patch",
"merged_at": "2024-01-31T13:59... | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6633). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>... |
https://api.github.com/repos/huggingface/datasets/issues/6632 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6632/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6632/comments | https://api.github.com/repos/huggingface/datasets/issues/6632/events | https://github.com/huggingface/datasets/pull/6632 | 2,108,541,678 | PR_kwDODunzps5lfPuk | 6,632 | Fix reload cache with data dir | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [] | closed | false | null | [] | null | 2 | 2024-01-30T18:52:23 | 2024-02-06T17:27:35 | 2024-02-06T17:21:24 | MEMBER | null | The cache used to only check for the latest cache directory with a given config_name, but it was wrong (e.g. `default-data_dir=data%2Ffortran-data_dir=data%2Ffortran` instead of `default-data_dir=data%2Ffortran`)
I fixed this by not passing the `config_kwargs` to the parent Builder `__init__`, and passing the config... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6632/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6632/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6632",
"html_url": "https://github.com/huggingface/datasets/pull/6632",
"diff_url": "https://github.com/huggingface/datasets/pull/6632.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6632.patch",
"merged_at": "2024-02-06T17:21... | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6632). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>... |
https://api.github.com/repos/huggingface/datasets/issues/6631 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6631/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6631/comments | https://api.github.com/repos/huggingface/datasets/issues/6631/events | https://github.com/huggingface/datasets/pull/6631 | 2,107,802,473 | PR_kwDODunzps5lcu9A | 6,631 | Fix filelock: use current umask for filelock >= 3.10 | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [] | closed | false | null | [] | null | 2 | 2024-01-30T12:56:01 | 2024-01-30T15:34:49 | 2024-01-30T15:28:37 | MEMBER | null | reported in https://github.com/huggingface/evaluate/issues/542
cc @stas00 @williamberrios
close https://github.com/huggingface/datasets/issues/6589 | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6631/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6631/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6631",
"html_url": "https://github.com/huggingface/datasets/pull/6631",
"diff_url": "https://github.com/huggingface/datasets/pull/6631.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6631.patch",
"merged_at": "2024-01-30T15:28... | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6631). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>... |
https://api.github.com/repos/huggingface/datasets/issues/6630 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6630/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6630/comments | https://api.github.com/repos/huggingface/datasets/issues/6630/events | https://github.com/huggingface/datasets/pull/6630 | 2,106,478,275 | PR_kwDODunzps5lYPi3 | 6,630 | Bump max range of dill to 0.3.8 | {
"login": "ringohoffman",
"id": 27844407,
"node_id": "MDQ6VXNlcjI3ODQ0NDA3",
"avatar_url": "https://avatars.githubusercontent.com/u/27844407?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ringohoffman",
"html_url": "https://github.com/ringohoffman",
"followers_url": "https://api.github.c... | [] | closed | false | null | [] | null | 4 | 2024-01-29T21:35:55 | 2024-01-30T16:19:45 | 2024-01-30T15:12:25 | CONTRIBUTOR | null | Release on Jan 27, 2024: https://pypi.org/project/dill/0.3.8/#history
| {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6630/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6630/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6630",
"html_url": "https://github.com/huggingface/datasets/pull/6630",
"diff_url": "https://github.com/huggingface/datasets/pull/6630.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6630.patch",
"merged_at": "2024-01-30T15:12... | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6630). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"Hmm these errors look pretty weird... can they be retried?",
"Hi, thanks for working ... |
https://api.github.com/repos/huggingface/datasets/issues/6629 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6629/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6629/comments | https://api.github.com/repos/huggingface/datasets/issues/6629/events | https://github.com/huggingface/datasets/pull/6629 | 2,105,774,482 | PR_kwDODunzps5lV0aF | 6,629 | Support push_to_hub without org/user to default to logged-in user | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | 3 | 2024-01-29T15:36:52 | 2024-02-05T12:35:43 | 2024-02-05T12:29:36 | MEMBER | null | This behavior is aligned with:
- the behavior of `datasets` before merging #6519
- the behavior described in the corresponding docstring
- the behavior of `huggingface_hub.create_repo`
Revert "Support push_to_hub canonical datasets (#6519)"
- This reverts commit a887ee78835573f5d80f9e414e8443b4caff3541.
Fix... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6629/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6629/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6629",
"html_url": "https://github.com/huggingface/datasets/pull/6629",
"diff_url": "https://github.com/huggingface/datasets/pull/6629.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6629.patch",
"merged_at": "2024-02-05T12:29... | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6629). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"@huggingface/datasets, feel free to review this PR so that it can be included in the ne... |
https://api.github.com/repos/huggingface/datasets/issues/6628 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6628/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6628/comments | https://api.github.com/repos/huggingface/datasets/issues/6628/events | https://github.com/huggingface/datasets/pull/6628 | 2,105,760,502 | PR_kwDODunzps5lVxXU | 6,628 | Make CLI test support multi-processing | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | 3 | 2024-01-29T15:30:09 | 2024-02-05T10:29:20 | 2024-02-05T10:23:13 | MEMBER | null | Support passing `--num_proc` to CLI test.
This was really useful recently to run the command on `pubmed`: https://huggingface.co/datasets/pubmed/discussions/11 | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6628/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6628/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6628",
"html_url": "https://github.com/huggingface/datasets/pull/6628",
"diff_url": "https://github.com/huggingface/datasets/pull/6628.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6628.patch",
"merged_at": "2024-02-05T10:23... | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6628). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"@huggingface/datasets, feel free to review this PR so that it can be included in the ne... |
https://api.github.com/repos/huggingface/datasets/issues/6627 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6627/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6627/comments | https://api.github.com/repos/huggingface/datasets/issues/6627/events | https://github.com/huggingface/datasets/pull/6627 | 2,105,735,816 | PR_kwDODunzps5lVr-t | 6,627 | Disable `tqdm` bars in non-interactive environments | {
"login": "mariosasko",
"id": 47462742,
"node_id": "MDQ6VXNlcjQ3NDYyNzQy",
"avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mariosasko",
"html_url": "https://github.com/mariosasko",
"followers_url": "https://api.github.com/use... | [] | closed | false | null | [] | null | 2 | 2024-01-29T15:18:21 | 2024-01-29T15:47:34 | 2024-01-29T15:41:32 | CONTRIBUTOR | null | Replace `disable=False` with `disable=None` in the `tqdm` bars to disable them in non-interactive environments (by default).
For more info, see a [similar PR](https://github.com/huggingface/huggingface_hub/pull/2000) in `huggingface_hub`. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6627/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6627/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6627",
"html_url": "https://github.com/huggingface/datasets/pull/6627",
"diff_url": "https://github.com/huggingface/datasets/pull/6627.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6627.patch",
"merged_at": "2024-01-29T15:41... | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6627). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>... |
https://api.github.com/repos/huggingface/datasets/issues/6626 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6626/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6626/comments | https://api.github.com/repos/huggingface/datasets/issues/6626/events | https://github.com/huggingface/datasets/pull/6626 | 2,105,482,522 | PR_kwDODunzps5lU0I2 | 6,626 | Raise error on bad split name | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [] | closed | false | null | [] | null | 2 | 2024-01-29T13:17:41 | 2024-01-29T15:18:25 | 2024-01-29T15:12:18 | MEMBER | null | e.g. dashes '-' are not allowed in split names
This should add an error message on datasets with unsupported split names like https://huggingface.co/datasets/open-source-metrics/test
cc @AndreaFrancis | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6626/reactions",
"total_count": 1,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 1,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6626/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6626",
"html_url": "https://github.com/huggingface/datasets/pull/6626",
"diff_url": "https://github.com/huggingface/datasets/pull/6626.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6626.patch",
"merged_at": "2024-01-29T15:12... | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6626). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>... |
https://api.github.com/repos/huggingface/datasets/issues/6624 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6624/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6624/comments | https://api.github.com/repos/huggingface/datasets/issues/6624/events | https://github.com/huggingface/datasets/issues/6624 | 2,103,950,718 | I_kwDODunzps59Z71- | 6,624 | How to download the laion-coco dataset | {
"login": "vanpersie32",
"id": 15981416,
"node_id": "MDQ6VXNlcjE1OTgxNDE2",
"avatar_url": "https://avatars.githubusercontent.com/u/15981416?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/vanpersie32",
"html_url": "https://github.com/vanpersie32",
"followers_url": "https://api.github.com/... | [] | closed | false | null | [] | null | 1 | 2024-01-28T03:56:05 | 2024-02-06T09:43:31 | 2024-02-06T09:43:31 | NONE | null | The laion coco dataset is not available now. How to download it
https://huggingface.co/datasets/laion/laion-coco | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6624/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6624/timeline | null | not_planned | null | null | false | [
"Hi, this dataset has been disabled by the authors, so unfortunately it's no longer possible to download it."
] |
https://api.github.com/repos/huggingface/datasets/issues/6623 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6623/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6623/comments | https://api.github.com/repos/huggingface/datasets/issues/6623/events | https://github.com/huggingface/datasets/issues/6623 | 2,103,870,123 | I_kwDODunzps59ZoKr | 6,623 | streaming datasets doesn't work properly with multi-node | {
"login": "rohitgr7",
"id": 30778939,
"node_id": "MDQ6VXNlcjMwNzc4OTM5",
"avatar_url": "https://avatars.githubusercontent.com/u/30778939?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rohitgr7",
"html_url": "https://github.com/rohitgr7",
"followers_url": "https://api.github.com/users/roh... | [
{
"id": 1935892871,
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement",
"name": "enhancement",
"color": "a2eeef",
"default": true,
"description": "New feature or request"
}
] | open | false | null | [] | null | 8 | 2024-01-27T23:46:13 | 2024-03-08T14:27:08 | null | NONE | null | ### Feature request
Let’s say I have a dataset with 5 samples with values [1, 2, 3, 4, 5], with 2 GPUs (for DDP) and batch size of 2. This dataset is an `IterableDataset` since I am streaming it.
Now I split the dataset using `split_dataset_by_node` to ensure it doesn’t get repeated. And since it’s already splitt... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6623/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6623/timeline | null | null | null | null | false | [
"@mariosasko, @lhoestq, @albertvillanova\r\nhey guys! can anyone help? or can you guys suggest who can help with this?",
"Hi ! \r\n\r\n1. When the dataset is running of of examples, the last batches received by the GPU can be incomplete or empty/missing. We haven't implemented yet a way to ignore the last batch. ... |
https://api.github.com/repos/huggingface/datasets/issues/6622 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6622/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6622/comments | https://api.github.com/repos/huggingface/datasets/issues/6622/events | https://github.com/huggingface/datasets/issues/6622 | 2,103,780,697 | I_kwDODunzps59ZSVZ | 6,622 | multi-GPU map does not work | {
"login": "kopyl",
"id": 17604849,
"node_id": "MDQ6VXNlcjE3NjA0ODQ5",
"avatar_url": "https://avatars.githubusercontent.com/u/17604849?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/kopyl",
"html_url": "https://github.com/kopyl",
"followers_url": "https://api.github.com/users/kopyl/follow... | [] | closed | false | null | [] | null | 1 | 2024-01-27T20:06:08 | 2024-02-08T11:18:21 | 2024-02-08T11:18:21 | NONE | null | ### Describe the bug
Here is the code for single-GPU processing: https://pastebin.com/bfmEeK2y
Here is the code for multi-GPU processing: https://pastebin.com/gQ7i5AQy
Here is the video showing that the multi-GPU mapping does not work as expected (there are so many things wrong here, it's better to watch the 3-min... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6622/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6622/timeline | null | completed | null | null | false | [
"This should now be fixed by https://github.com/huggingface/datasets/pull/6550 and updated with https://github.com/huggingface/datasets/pull/6646\r\n\r\nFeel free to re-open if you're still having issues :)"
] |
https://api.github.com/repos/huggingface/datasets/issues/6621 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6621/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6621/comments | https://api.github.com/repos/huggingface/datasets/issues/6621/events | https://github.com/huggingface/datasets/issues/6621 | 2,103,675,294 | I_kwDODunzps59Y4me | 6,621 | deleted | {
"login": "kopyl",
"id": 17604849,
"node_id": "MDQ6VXNlcjE3NjA0ODQ5",
"avatar_url": "https://avatars.githubusercontent.com/u/17604849?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/kopyl",
"html_url": "https://github.com/kopyl",
"followers_url": "https://api.github.com/users/kopyl/follow... | [] | closed | false | null | [] | null | 0 | 2024-01-27T16:59:58 | 2024-01-27T17:14:43 | 2024-01-27T17:14:43 | NONE | null | ... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6621/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6621/timeline | null | completed | null | null | false | [] |
https://api.github.com/repos/huggingface/datasets/issues/6620 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6620/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6620/comments | https://api.github.com/repos/huggingface/datasets/issues/6620/events | https://github.com/huggingface/datasets/issues/6620 | 2,103,110,536 | I_kwDODunzps59WuuI | 6,620 | wiki_dpr.py error (ID mismatch between lines {id} and vector {vec_id} | {
"login": "kiehls90",
"id": 101498700,
"node_id": "U_kgDOBgy_TA",
"avatar_url": "https://avatars.githubusercontent.com/u/101498700?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/kiehls90",
"html_url": "https://github.com/kiehls90",
"followers_url": "https://api.github.com/users/kiehls90/... | [] | closed | false | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_... | null | 1 | 2024-01-27T01:00:09 | 2024-02-06T09:40:19 | 2024-02-06T09:40:19 | NONE | null | ### Describe the bug
I'm trying to run a rag example, and the dataset is wiki_dpr.
wiki_dpr download and extracting have been completed successfully.
However, at the generating train split stage, an error from wiki_dpr.py keeps popping up.
Especially in "_generate_examples" :
1. The following error occurs in the... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6620/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6620/timeline | null | not_planned | null | null | false | [
"Thanks for reporting, @kiehls90.\r\n\r\nAs this seems an issue with the specific \"wiki_dpr\" dataset, I am transferring the issue to the corresponding dataset page: https://huggingface.co/datasets/wiki_dpr/discussions/13"
] |
https://api.github.com/repos/huggingface/datasets/issues/6619 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6619/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6619/comments | https://api.github.com/repos/huggingface/datasets/issues/6619/events | https://github.com/huggingface/datasets/pull/6619 | 2,102,407,478 | PR_kwDODunzps5lK2VY | 6,619 | Migrate from `setup.cfg` to `pyproject.toml` | {
"login": "mariosasko",
"id": 47462742,
"node_id": "MDQ6VXNlcjQ3NDYyNzQy",
"avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mariosasko",
"html_url": "https://github.com/mariosasko",
"followers_url": "https://api.github.com/use... | [] | closed | false | null | [] | null | 2 | 2024-01-26T15:27:10 | 2024-01-26T15:53:40 | 2024-01-26T15:47:32 | CONTRIBUTOR | null | Based on https://github.com/huggingface/huggingface_hub/pull/1971 in `hfh` | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6619/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6619/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6619",
"html_url": "https://github.com/huggingface/datasets/pull/6619",
"diff_url": "https://github.com/huggingface/datasets/pull/6619.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6619.patch",
"merged_at": "2024-01-26T15:47... | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6619). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>... |
https://api.github.com/repos/huggingface/datasets/issues/6618 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6618/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6618/comments | https://api.github.com/repos/huggingface/datasets/issues/6618/events | https://github.com/huggingface/datasets/issues/6618 | 2,101,868,198 | I_kwDODunzps59R_am | 6,618 | While importing load_dataset from datasets | {
"login": "Era-cell",
"id": 77973415,
"node_id": "MDQ6VXNlcjc3OTczNDE1",
"avatar_url": "https://avatars.githubusercontent.com/u/77973415?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Era-cell",
"html_url": "https://github.com/Era-cell",
"followers_url": "https://api.github.com/users/Era... | [] | closed | false | null | [] | null | 3 | 2024-01-26T09:21:57 | 2024-02-06T10:57:01 | 2024-02-06T09:25:54 | NONE | null | ### Describe the bug
cannot import name 'DEFAULT_CIPHERS' from 'urllib3.util.ssl_' this is the error i received
### Steps to reproduce the bug
from datasets import load_dataset
### Expected behavior
No errors
### Environment info
python 3.11.5 | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6618/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6618/timeline | null | not_planned | null | null | false | [
"Hi! Can you please share the error's stack trace so we can see where it comes from?",
"We cannot reproduce the issue and we do not have enough information: environment info (need to run `datasets-cli env`), stack trace,...\r\n\r\nI am closing the issue. Feel free to reopen it (with additional information) if the... |
https://api.github.com/repos/huggingface/datasets/issues/6617 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6617/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6617/comments | https://api.github.com/repos/huggingface/datasets/issues/6617/events | https://github.com/huggingface/datasets/pull/6617 | 2,100,459,449 | PR_kwDODunzps5lEagV | 6,617 | Fix CI: pyarrow 15, pandas 2.2 and sqlachemy | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [] | closed | false | null | [] | null | 2 | 2024-01-25T13:57:41 | 2024-01-26T14:56:46 | 2024-01-26T14:50:44 | MEMBER | null | this should fix the CI failures on `main`
close https://github.com/huggingface/datasets/issues/5477 | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6617/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6617/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6617",
"html_url": "https://github.com/huggingface/datasets/pull/6617",
"diff_url": "https://github.com/huggingface/datasets/pull/6617.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6617.patch",
"merged_at": "2024-01-26T14:50... | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6617). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>... |
https://api.github.com/repos/huggingface/datasets/issues/6616 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6616/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6616/comments | https://api.github.com/repos/huggingface/datasets/issues/6616/events | https://github.com/huggingface/datasets/pull/6616 | 2,100,125,709 | PR_kwDODunzps5lDSEL | 6,616 | Use schema metadata only if it matches features | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [] | closed | false | null | [] | null | 2 | 2024-01-25T11:01:14 | 2024-01-26T16:25:24 | 2024-01-26T16:19:12 | MEMBER | null | e.g. if we use `map` in arrow format and transform the table, the returned table might have new columns but the metadata might be wrong | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6616/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6616/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6616",
"html_url": "https://github.com/huggingface/datasets/pull/6616",
"diff_url": "https://github.com/huggingface/datasets/pull/6616.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6616.patch",
"merged_at": "2024-01-26T16:19... | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6616). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>... |
https://api.github.com/repos/huggingface/datasets/issues/6615 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6615/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6615/comments | https://api.github.com/repos/huggingface/datasets/issues/6615/events | https://github.com/huggingface/datasets/issues/6615 | 2,098,951,409 | I_kwDODunzps59G3Tx | 6,615 | ... | {
"login": "ftkeys",
"id": 22179777,
"node_id": "MDQ6VXNlcjIyMTc5Nzc3",
"avatar_url": "https://avatars.githubusercontent.com/u/22179777?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ftkeys",
"html_url": "https://github.com/ftkeys",
"followers_url": "https://api.github.com/users/ftkeys/fo... | [] | closed | false | null | [] | null | 1 | 2024-01-24T19:37:03 | 2024-01-24T19:42:30 | 2024-01-24T19:40:11 | NONE | null | ... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6615/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6615/timeline | null | not_planned | null | null | false | [
"Sorry I posted in the wrong repo, please delete.. thanks!"
] |
https://api.github.com/repos/huggingface/datasets/issues/6614 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6614/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6614/comments | https://api.github.com/repos/huggingface/datasets/issues/6614/events | https://github.com/huggingface/datasets/issues/6614 | 2,098,884,520 | I_kwDODunzps59Gm-o | 6,614 | `datasets/downloads` cleanup tool | {
"login": "stas00",
"id": 10676103,
"node_id": "MDQ6VXNlcjEwNjc2MTAz",
"avatar_url": "https://avatars.githubusercontent.com/u/10676103?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/stas00",
"html_url": "https://github.com/stas00",
"followers_url": "https://api.github.com/users/stas00/fo... | [
{
"id": 1935892871,
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement",
"name": "enhancement",
"color": "a2eeef",
"default": true,
"description": "New feature or request"
}
] | open | false | null | [] | null | 0 | 2024-01-24T18:52:10 | 2024-01-24T18:55:09 | null | CONTRIBUTOR | null | ### Feature request
Splitting off https://github.com/huggingface/huggingface_hub/issues/1997 - currently `huggingface-cli delete-cache` doesn't take care of cleaning `datasets` temp files
e.g. I discovered having millions of files under `datasets/downloads` cache, I had to do:
```
sudo find /data/huggingface/... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6614/reactions",
"total_count": 2,
"+1": 2,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6614/timeline | null | null | null | null | false | [] |
https://api.github.com/repos/huggingface/datasets/issues/6612 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6612/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6612/comments | https://api.github.com/repos/huggingface/datasets/issues/6612/events | https://github.com/huggingface/datasets/issues/6612 | 2,098,078,210 | I_kwDODunzps59DiIC | 6,612 | cnn_dailymail repeats itself | {
"login": "KeremZaman",
"id": 8274752,
"node_id": "MDQ6VXNlcjgyNzQ3NTI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8274752?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/KeremZaman",
"html_url": "https://github.com/KeremZaman",
"followers_url": "https://api.github.com/users... | [] | closed | false | null | [] | null | 1 | 2024-01-24T11:38:25 | 2024-02-01T08:14:50 | 2024-02-01T08:14:50 | NONE | null | ### Describe the bug
When I try to load `cnn_dailymail` dataset, it takes longer than usual and when I checked the dataset it's 3x bigger than it's supposed to be.
Check https://huggingface.co/datasets/cnn_dailymail: it says 287k rows for train. But when I check length of train split it says 861339.
Also I che... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6612/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6612/timeline | null | completed | null | null | false | [
"Hi ! We recently updated `cnn_dailymail` and now `datasets>=2.14` is needed to load it.\r\n\r\nYou can update `datasets` with\r\n\r\n```\r\npip install -U datasets\r\n```"
] |
https://api.github.com/repos/huggingface/datasets/issues/6611 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6611/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6611/comments | https://api.github.com/repos/huggingface/datasets/issues/6611/events | https://github.com/huggingface/datasets/issues/6611 | 2,096,004,858 | I_kwDODunzps587n76 | 6,611 | `load_from_disk` with large dataset from S3 runs into `botocore.exceptions.ClientError` | {
"login": "zotroneneis",
"id": 15320635,
"node_id": "MDQ6VXNlcjE1MzIwNjM1",
"avatar_url": "https://avatars.githubusercontent.com/u/15320635?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/zotroneneis",
"html_url": "https://github.com/zotroneneis",
"followers_url": "https://api.github.com/... | [] | open | false | null | [] | null | 0 | 2024-01-23T12:37:57 | 2024-01-23T12:37:57 | null | NONE | null | ### Describe the bug
When loading a large dataset (>1000GB) from S3 I run into the following error:
```
Traceback (most recent call last):
File "/home/alp/.local/lib/python3.10/site-packages/s3fs/core.py", line 113, in _error_wrapper
return await func(*args, **kwargs)
File "/home/alp/.local/lib/python3.... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6611/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6611/timeline | null | null | null | null | false | [] |
https://api.github.com/repos/huggingface/datasets/issues/6610 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6610/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6610/comments | https://api.github.com/repos/huggingface/datasets/issues/6610/events | https://github.com/huggingface/datasets/issues/6610 | 2,095,643,711 | I_kwDODunzps586Pw_ | 6,610 | cast_column to Sequence(subfeatures_dict) has err | {
"login": "neiblegy",
"id": 16574677,
"node_id": "MDQ6VXNlcjE2NTc0Njc3",
"avatar_url": "https://avatars.githubusercontent.com/u/16574677?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/neiblegy",
"html_url": "https://github.com/neiblegy",
"followers_url": "https://api.github.com/users/nei... | [] | closed | false | null | [] | null | 2 | 2024-01-23T09:32:32 | 2024-01-25T02:15:23 | 2024-01-25T02:15:23 | NONE | null | ### Describe the bug
I am working with the following demo code:
```
from datasets import load_dataset
from datasets.features import Sequence, Value, ClassLabel, Features
ais_dataset = load_dataset("/data/ryan.gao/ais_dataset_cache/raw/1978/")
ais_dataset = ais_dataset["train"]
def add_class(example):
... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6610/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6610/timeline | null | completed | null | null | false | [
"Hi! You are passing the wrong feature type to `cast_column`. This is the fixed call:\r\n```python\r\nais_dataset = ais_dataset.cast_column(\"my_labeled_bbox\", {\"bbox\": Sequence(Value(dtype=\"int64\")), \"label\": ClassLabel(names=[\"cat\", \"dog\"])})\r\n```",
"> Hi! You are passing the wrong feature type to ... |
https://api.github.com/repos/huggingface/datasets/issues/6609 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6609/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6609/comments | https://api.github.com/repos/huggingface/datasets/issues/6609/events | https://github.com/huggingface/datasets/issues/6609 | 2,095,085,650 | I_kwDODunzps584HhS | 6,609 | Wrong path for cache directory in offline mode | {
"login": "je-santos",
"id": 42117435,
"node_id": "MDQ6VXNlcjQyMTE3NDM1",
"avatar_url": "https://avatars.githubusercontent.com/u/42117435?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/je-santos",
"html_url": "https://github.com/je-santos",
"followers_url": "https://api.github.com/users/... | [] | closed | false | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [
{
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.git... | null | 5 | 2024-01-23T01:47:19 | 2024-02-06T17:21:25 | 2024-02-06T17:21:25 | NONE | null | ### Describe the bug
Dear huggingfacers,
I'm trying to use a subset of the-stack dataset. When I run the command the first time
```
dataset = load_dataset(
path='bigcode/the-stack',
data_dir='data/fortran',
split='train' )
```
It downloads the files and caches them normally.
Nevertheless, ... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6609/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6609/timeline | null | completed | null | null | false | [
"+1",
"same error in 2.16.1",
"@kongjiellx any luck with the issue?",
"I opened https://github.com/huggingface/datasets/pull/6632 to fix this issue. Once it's merged we'll do a new release of `datasets`",
"Thanks @lhoestq !"
] |
https://api.github.com/repos/huggingface/datasets/issues/6608 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6608/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6608/comments | https://api.github.com/repos/huggingface/datasets/issues/6608/events | https://github.com/huggingface/datasets/pull/6608 | 2,094,153,292 | PR_kwDODunzps5ku_lN | 6,608 | Add `with_rank` param to `Dataset.filter` | {
"login": "mariosasko",
"id": 47462742,
"node_id": "MDQ6VXNlcjQ3NDYyNzQy",
"avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mariosasko",
"html_url": "https://github.com/mariosasko",
"followers_url": "https://api.github.com/use... | [] | closed | false | null | [] | null | 2 | 2024-01-22T15:19:16 | 2024-01-29T16:43:11 | 2024-01-29T16:36:53 | CONTRIBUTOR | null | Fix #6564 | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6608/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6608/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6608",
"html_url": "https://github.com/huggingface/datasets/pull/6608",
"diff_url": "https://github.com/huggingface/datasets/pull/6608.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6608.patch",
"merged_at": "2024-01-29T16:36... | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6608). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>... |
https://api.github.com/repos/huggingface/datasets/issues/6607 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6607/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6607/comments | https://api.github.com/repos/huggingface/datasets/issues/6607/events | https://github.com/huggingface/datasets/pull/6607 | 2,091,766,063 | PR_kwDODunzps5knGse | 6,607 | Update features.py to avoid bfloat16 unsupported error | {
"login": "skaulintel",
"id": 75697181,
"node_id": "MDQ6VXNlcjc1Njk3MTgx",
"avatar_url": "https://avatars.githubusercontent.com/u/75697181?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/skaulintel",
"html_url": "https://github.com/skaulintel",
"followers_url": "https://api.github.com/use... | [] | open | false | null | [] | null | 1 | 2024-01-20T00:39:44 | 2024-03-22T00:34:48 | null | NONE | null | Fixes https://github.com/huggingface/datasets/issues/6566
Let me know if there's any tests I need to clear. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6607/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6607/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6607",
"html_url": "https://github.com/huggingface/datasets/pull/6607",
"diff_url": "https://github.com/huggingface/datasets/pull/6607.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6607.patch",
"merged_at": null
} | true | [
"I think not all torch tensors should be converted to float, what if it's a tensor of integers for example ?\r\nMaybe you can check for the tensor dtype before converting"
] |
https://api.github.com/repos/huggingface/datasets/issues/6606 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6606/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6606/comments | https://api.github.com/repos/huggingface/datasets/issues/6606/events | https://github.com/huggingface/datasets/pull/6606 | 2,091,088,785 | PR_kwDODunzps5kk3KB | 6,606 | Dedicated RNG object for fingerprinting | {
"login": "mariosasko",
"id": 47462742,
"node_id": "MDQ6VXNlcjQ3NDYyNzQy",
"avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mariosasko",
"html_url": "https://github.com/mariosasko",
"followers_url": "https://api.github.com/use... | [] | closed | false | null | [] | null | 2 | 2024-01-19T18:34:47 | 2024-01-26T15:11:38 | 2024-01-26T15:05:34 | CONTRIBUTOR | null | Closes https://github.com/huggingface/datasets/issues/6604, closes https://github.com/huggingface/datasets/issues/2775 | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6606/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6606/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6606",
"html_url": "https://github.com/huggingface/datasets/pull/6606",
"diff_url": "https://github.com/huggingface/datasets/pull/6606.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6606.patch",
"merged_at": "2024-01-26T15:05... | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6606). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>... |
https://api.github.com/repos/huggingface/datasets/issues/6605 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6605/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6605/comments | https://api.github.com/repos/huggingface/datasets/issues/6605/events | https://github.com/huggingface/datasets/issues/6605 | 2,090,188,376 | I_kwDODunzps58lb5Y | 6,605 | ELI5 no longer available, but referenced in example code | {
"login": "drdsgvo",
"id": 81480344,
"node_id": "MDQ6VXNlcjgxNDgwMzQ0",
"avatar_url": "https://avatars.githubusercontent.com/u/81480344?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/drdsgvo",
"html_url": "https://github.com/drdsgvo",
"followers_url": "https://api.github.com/users/drdsgv... | [] | closed | false | null | [] | null | 1 | 2024-01-19T10:21:52 | 2024-02-01T17:58:23 | 2024-02-01T17:58:22 | NONE | null | Here, an example code is given:
https://huggingface.co/docs/transformers/tasks/language_modeling
This code + article references the ELI5 dataset.
ELI5 is no longer available, as the ELI5 dataset page states: https://huggingface.co/datasets/eli5
"Defunct: Dataset "eli5" is defunct and no longer accessible due to u... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6605/reactions",
"total_count": 3,
"+1": 3,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6605/timeline | null | completed | null | null | false | [
"Addressed in https://github.com/huggingface/transformers/pull/28715."
] |
https://api.github.com/repos/huggingface/datasets/issues/6604 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6604/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6604/comments | https://api.github.com/repos/huggingface/datasets/issues/6604/events | https://github.com/huggingface/datasets/issues/6604 | 2,089,713,945 | I_kwDODunzps58joEZ | 6,604 | Transform fingerprint collisions due to setting fixed random seed | {
"login": "normster",
"id": 6687910,
"node_id": "MDQ6VXNlcjY2ODc5MTA=",
"avatar_url": "https://avatars.githubusercontent.com/u/6687910?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/normster",
"html_url": "https://github.com/normster",
"followers_url": "https://api.github.com/users/norms... | [] | closed | false | null | [] | null | 2 | 2024-01-19T06:32:25 | 2024-01-26T15:05:35 | 2024-01-26T15:05:35 | NONE | null | ### Describe the bug
The transform fingerprinting logic relies on the `random` library for random bits when the function is not hashable (e.g. bound methods as used in `trl`: https://github.com/huggingface/trl/blob/main/trl/trainer/dpo_trainer.py#L356). This causes collisions when the training code sets a fixed random... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6604/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6604/timeline | null | completed | null | null | false | [
"I've opened a PR with a fix.",
"I don't think the PR fixes the root cause, since it still relies on the `random` library which will often have its seed fixed. I think the builtin `uuid.uuid4()` is a better choice: https://docs.python.org/3/library/uuid.html"
] |
https://api.github.com/repos/huggingface/datasets/issues/6603 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6603/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6603/comments | https://api.github.com/repos/huggingface/datasets/issues/6603/events | https://github.com/huggingface/datasets/issues/6603 | 2,089,230,766 | I_kwDODunzps58hyGu | 6,603 | datasets map `cache_file_name` does not work | {
"login": "ChenchaoZhao",
"id": 35147961,
"node_id": "MDQ6VXNlcjM1MTQ3OTYx",
"avatar_url": "https://avatars.githubusercontent.com/u/35147961?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ChenchaoZhao",
"html_url": "https://github.com/ChenchaoZhao",
"followers_url": "https://api.github.c... | [] | open | false | null | [] | null | 2 | 2024-01-18T23:08:30 | 2024-01-28T04:01:15 | null | NONE | null | ### Describe the bug
In the documentation `datasets.Dataset.map` arg `cache_file_name` is said to be a string, but it doesn't work.
### Steps to reproduce the bug
1. pick a dataset
2. write a map function
3. do `ds.map(..., cache_file_name='some_filename')`
4. it crashes
### Expected behavior
It will tell you t... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6603/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6603/timeline | null | null | null | null | false | [
"Unfortunately, I'm unable to reproduce this error. Can you share the reproducer?",
"```\r\nds = datasets.Dataset.from_dict(dict(a=[i for i in range(100)]))\r\nds.map(lambda item: dict(b=item['a'] * 2), cache_file_name=\"/tmp/whatever-fn\") # this worked\r\nds.map(lambda item: dict(b=item['a'] * 2), cache_file_na... |
https://api.github.com/repos/huggingface/datasets/issues/6602 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6602/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6602/comments | https://api.github.com/repos/huggingface/datasets/issues/6602/events | https://github.com/huggingface/datasets/issues/6602 | 2,089,217,483 | I_kwDODunzps58hu3L | 6,602 | Index error when data is large | {
"login": "ChenchaoZhao",
"id": 35147961,
"node_id": "MDQ6VXNlcjM1MTQ3OTYx",
"avatar_url": "https://avatars.githubusercontent.com/u/35147961?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ChenchaoZhao",
"html_url": "https://github.com/ChenchaoZhao",
"followers_url": "https://api.github.c... | [] | open | false | null | [] | null | 0 | 2024-01-18T23:00:47 | 2024-01-18T23:00:47 | null | NONE | null | ### Describe the bug
At `save_to_disk` step, the `max_shard_size` by default is `500MB`. However, one row of the dataset might be larger than `500MB` then the saving will throw an index error. Without looking at the source code, the bug is due to wrong calculation of number of shards which i think is
`total_size / m... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6602/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6602/timeline | null | null | null | null | false | [] |
https://api.github.com/repos/huggingface/datasets/issues/6601 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6601/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6601/comments | https://api.github.com/repos/huggingface/datasets/issues/6601/events | https://github.com/huggingface/datasets/pull/6601 | 2,088,624,054 | PR_kwDODunzps5kcWN0 | 6,601 | add safety checks when using only part of dataset | {
"login": "benseddikismail",
"id": 63422923,
"node_id": "MDQ6VXNlcjYzNDIyOTIz",
"avatar_url": "https://avatars.githubusercontent.com/u/63422923?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/benseddikismail",
"html_url": "https://github.com/benseddikismail",
"followers_url": "https://api... | [] | open | false | null | [] | null | 1 | 2024-01-18T16:16:59 | 2024-02-08T14:33:10 | null | NONE | null | Added some checks to prevent errors that arrise when using evaluate.py on only a portion of the squad 2.0 dataset. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6601/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6601/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6601",
"html_url": "https://github.com/huggingface/datasets/pull/6601",
"diff_url": "https://github.com/huggingface/datasets/pull/6601.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6601.patch",
"merged_at": null
} | true | [
"Hi ! The metrics in `datasets` are deprecated in favor of https://github.com/huggingface/evaluate\r\n\r\nYou can open a PR here instead: https://huggingface.co/spaces/evaluate-metric/squad_v2/tree/main"
] |
https://api.github.com/repos/huggingface/datasets/issues/6600 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6600/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6600/comments | https://api.github.com/repos/huggingface/datasets/issues/6600/events | https://github.com/huggingface/datasets/issues/6600 | 2,088,446,385 | I_kwDODunzps58eymx | 6,600 | Loading CSV exported dataset has unexpected format | {
"login": "OrianeN",
"id": 59572247,
"node_id": "MDQ6VXNlcjU5NTcyMjQ3",
"avatar_url": "https://avatars.githubusercontent.com/u/59572247?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/OrianeN",
"html_url": "https://github.com/OrianeN",
"followers_url": "https://api.github.com/users/Oriane... | [] | open | false | null | [] | null | 2 | 2024-01-18T14:48:27 | 2024-01-23T14:42:32 | null | NONE | null | ### Describe the bug
I wanted to be able to save a HF dataset for translations and load it again in another script, but I'm a bit confused with the documentation and the result I've got so I'm opening this issue to ask if this behavior is as expected.
### Steps to reproduce the bug
The documentation I've mainly cons... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6600/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6600/timeline | null | null | null | null | false | [
"Hi! Parquet is the only format that supports complex/nested features such as `Translation`. So, this should work:\r\n```python\r\ntest_dataset = load_dataset(\"opus100\", name=\"en-fr\", split=\"test\")\r\n\r\n# Save with .to_parquet()\r\ntest_parquet_path = \"try_testset_save.parquet\"\r\ntest_dataset.to_parquet(... |
https://api.github.com/repos/huggingface/datasets/issues/6599 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6599/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6599/comments | https://api.github.com/repos/huggingface/datasets/issues/6599/events | https://github.com/huggingface/datasets/issues/6599 | 2,086,684,664 | I_kwDODunzps58YEf4 | 6,599 | Easy way to segment into 30s snippets given an m4a file and a vtt file | {
"login": "RonanKMcGovern",
"id": 78278410,
"node_id": "MDQ6VXNlcjc4Mjc4NDEw",
"avatar_url": "https://avatars.githubusercontent.com/u/78278410?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/RonanKMcGovern",
"html_url": "https://github.com/RonanKMcGovern",
"followers_url": "https://api.gi... | [
{
"id": 1935892871,
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement",
"name": "enhancement",
"color": "a2eeef",
"default": true,
"description": "New feature or request"
}
] | closed | false | null | [] | null | 2 | 2024-01-17T17:51:40 | 2024-01-23T10:42:17 | 2024-01-22T15:35:49 | NONE | null | ### Feature request
Uploading datasets is straightforward thanks to the ability to push Audio to hub. However, it would be nice if the data (text and audio) could be segmented when being pushed (if not possible already).
### Motivation
It's easy to create a vtt file from an audio file. If there could be auto-segment... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6599/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6599/timeline | null | not_planned | null | null | false | [
"Hi! Non-generic data processing is out of this library's scope, so it's downstream libraries/users' responsibility to implement such logic.",
"That's fair. Thanks"
] |
https://api.github.com/repos/huggingface/datasets/issues/6598 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6598/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6598/comments | https://api.github.com/repos/huggingface/datasets/issues/6598/events | https://github.com/huggingface/datasets/issues/6598 | 2,084,236,605 | I_kwDODunzps58Ou09 | 6,598 | Unexpected keyword argument 'hf' when downloading CSV dataset from S3 | {
"login": "dguenms",
"id": 5592111,
"node_id": "MDQ6VXNlcjU1OTIxMTE=",
"avatar_url": "https://avatars.githubusercontent.com/u/5592111?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dguenms",
"html_url": "https://github.com/dguenms",
"followers_url": "https://api.github.com/users/dguenms/... | [] | open | false | null | [] | null | 6 | 2024-01-16T15:16:01 | 2024-03-22T02:08:06 | null | NONE | null | ### Describe the bug
I receive this error message when using `load_dataset` with "csv" path and `dataset_files=s3://...`:
```
TypeError: Session.__init__() got an unexpected keyword argument 'hf'
```
I found a similar issue here: https://stackoverflow.com/questions/77596258/aws-issue-load-dataset-from-s3-fails-w... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6598/reactions",
"total_count": 3,
"+1": 3,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6598/timeline | null | null | null | null | false | [
"I am facing similar issue while reading a csv file from s3. Wondering if somebody has found a workaround. ",
"same thing happened to other formats like parquet",
"I am facing similar issue while reading a parquet file from s3.\r\ni try with every version between 2.14 to 2.16.1 but it dosen't work ",
"Re-def... |
https://api.github.com/repos/huggingface/datasets/issues/6597 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6597/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6597/comments | https://api.github.com/repos/huggingface/datasets/issues/6597/events | https://github.com/huggingface/datasets/issues/6597 | 2,083,708,521 | I_kwDODunzps58Mt5p | 6,597 | Dataset.push_to_hub of a canonical dataset creates an additional dataset under the user namespace | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_... | null | 6 | 2024-01-16T11:27:07 | 2024-02-05T12:29:37 | 2024-02-05T12:29:37 | MEMBER | null | While using `Dataset.push_to_hub` of a canonical dataset, an additional dataset was created under my user namespace.
## Steps to reproduce the bug
The command:
```python
commit_info = ds.push_to_hub(
"caner",
config_name="default",
commit_message="Convert dataset to Parquet",
commit_descriptio... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6597/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6597/timeline | null | completed | null | null | false | [
"It is caused by these code lines: https://github.com/huggingface/datasets/blob/9d6d16117a30ba345b0236407975f701c5b288d4/src/datasets/dataset_dict.py#L1688-L1694",
"Also note the information in the docstring: https://github.com/huggingface/datasets/blob/9d6d16117a30ba345b0236407975f701c5b288d4/src/datasets/datase... |
https://api.github.com/repos/huggingface/datasets/issues/6596 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6596/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6596/comments | https://api.github.com/repos/huggingface/datasets/issues/6596/events | https://github.com/huggingface/datasets/pull/6596 | 2,083,108,156 | PR_kwDODunzps5kJceH | 6,596 | Drop redundant None guard. | {
"login": "xkszltl",
"id": 5203025,
"node_id": "MDQ6VXNlcjUyMDMwMjU=",
"avatar_url": "https://avatars.githubusercontent.com/u/5203025?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/xkszltl",
"html_url": "https://github.com/xkszltl",
"followers_url": "https://api.github.com/users/xkszltl/... | [] | closed | false | null | [] | null | 2 | 2024-01-16T06:31:54 | 2024-01-16T17:16:16 | 2024-01-16T17:05:52 | CONTRIBUTOR | null | `xxx if xxx is not None else None` is no-op. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6596/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6596/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6596",
"html_url": "https://github.com/huggingface/datasets/pull/6596",
"diff_url": "https://github.com/huggingface/datasets/pull/6596.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6596.patch",
"merged_at": "2024-01-16T17:05... | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6596). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>... |
https://api.github.com/repos/huggingface/datasets/issues/6595 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6595/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6595/comments | https://api.github.com/repos/huggingface/datasets/issues/6595/events | https://github.com/huggingface/datasets/issues/6595 | 2,082,896,148 | I_kwDODunzps58JnkU | 6,595 | Loading big dataset raises pyarrow.lib.ArrowNotImplementedError 2 | {
"login": "kopyl",
"id": 17604849,
"node_id": "MDQ6VXNlcjE3NjA0ODQ5",
"avatar_url": "https://avatars.githubusercontent.com/u/17604849?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/kopyl",
"html_url": "https://github.com/kopyl",
"followers_url": "https://api.github.com/users/kopyl/follow... | [] | closed | false | null | [] | null | 14 | 2024-01-16T02:03:09 | 2024-01-27T18:26:33 | 2024-01-26T02:28:32 | NONE | null | ### Describe the bug
I'm aware of the issue #5695 .
I'm using a modified SDXL trainer: https://github.com/kopyl/diffusers/blob/5e70f604155aeecee254a5c63c5e4236ad4a0d3d/examples/text_to_image/train_text_to_image_sdxl.py#L1027C16-L1027C16
So i
1. Map dataset
2. Save to disk
3. Try to upload:
```
import data... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6595/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6595/timeline | null | completed | null | null | false | [
"Hi ! I think the issue comes from the \"float16\" features that are not supported yet in Parquet\r\n\r\nFeel free to open an issue in `pyarrow` about this. In the meantime, I'd encourage you to use \"float32\" for your \"pooled_prompt_embeds\" and \"prompt_embeds\" features.\r\n\r\nYou can cast them to \"float32\"... |
https://api.github.com/repos/huggingface/datasets/issues/6594 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6594/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6594/comments | https://api.github.com/repos/huggingface/datasets/issues/6594/events | https://github.com/huggingface/datasets/issues/6594 | 2,082,748,275 | I_kwDODunzps58JDdz | 6,594 | IterableDataset sharding logic needs improvement | {
"login": "rwightman",
"id": 5702664,
"node_id": "MDQ6VXNlcjU3MDI2NjQ=",
"avatar_url": "https://avatars.githubusercontent.com/u/5702664?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rwightman",
"html_url": "https://github.com/rwightman",
"followers_url": "https://api.github.com/users/rw... | [] | open | false | null | [] | null | 0 | 2024-01-15T22:22:36 | 2024-01-15T22:25:10 | null | NONE | null | ### Describe the bug
The sharding of IterableDatasets with respect to distributed and dataloader worker processes appears problematic with significant performance traps and inconsistencies wrt to distributed train processes vs worker processes.
Splitting across num_workers (per train process loader processes) and... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6594/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6594/timeline | null | null | null | null | false | [] |
https://api.github.com/repos/huggingface/datasets/issues/6592 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6592/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6592/comments | https://api.github.com/repos/huggingface/datasets/issues/6592/events | https://github.com/huggingface/datasets/issues/6592 | 2,082,410,257 | I_kwDODunzps58Hw8R | 6,592 | Logs are delayed when doing .map when `docker logs` | {
"login": "kopyl",
"id": 17604849,
"node_id": "MDQ6VXNlcjE3NjA0ODQ5",
"avatar_url": "https://avatars.githubusercontent.com/u/17604849?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/kopyl",
"html_url": "https://github.com/kopyl",
"followers_url": "https://api.github.com/users/kopyl/follow... | [] | closed | false | null | [] | null | 1 | 2024-01-15T17:05:21 | 2024-02-12T17:35:21 | 2024-02-12T17:35:21 | NONE | null | ### Describe the bug
When I run my SD training in a Docker image and then listen to logs like `docker logs train -f`, the progress bar is delayed.
It's updating every few percent.
When you have a large dataset that has to be mapped (like 1+ million samples), it's crucial to see the updates in real-time, not every co... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6592/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6592/timeline | null | not_planned | null | null | false | [
"Hi! `tqdm` doesn't work well in non-interactive environments, so there isn't much we can do about this. It's best to [disable it](https://huggingface.co/docs/datasets/v2.16.1/en/package_reference/utilities#datasets.disable_progress_bars) in such environments and instead use logging to track progress."
] |
https://api.github.com/repos/huggingface/datasets/issues/6591 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6591/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6591/comments | https://api.github.com/repos/huggingface/datasets/issues/6591/events | https://github.com/huggingface/datasets/issues/6591 | 2,082,378,957 | I_kwDODunzps58HpTN | 6,591 | The datasets models housed in Dropbox can't support a lot of users downloading them | {
"login": "RDaneelOlivav",
"id": 4933774,
"node_id": "MDQ6VXNlcjQ5MzM3NzQ=",
"avatar_url": "https://avatars.githubusercontent.com/u/4933774?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/RDaneelOlivav",
"html_url": "https://github.com/RDaneelOlivav",
"followers_url": "https://api.github.... | [] | closed | false | null | [] | null | 1 | 2024-01-15T16:43:38 | 2024-01-22T23:18:09 | 2024-01-22T23:18:09 | NONE | null | ### Describe the bug
I'm using the datasets
```
from datasets import load_dataset, Audio
dataset = load_dataset("PolyAI/minds14", name="en-US", split="train")
```
And it seems that sometimes when I imagine a lot of users are accessing the same resources, the Dropbox host fails:
`raise ConnectionError(... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6591/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6591/timeline | null | completed | null | null | false | [
"Hi! Indeed, Dropbox is not a reliable host. I've just merged https://huggingface.co/datasets/PolyAI/minds14/discussions/24 to fix this by hosting the data files inside the repo."
] |
https://api.github.com/repos/huggingface/datasets/issues/6590 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6590/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6590/comments | https://api.github.com/repos/huggingface/datasets/issues/6590/events | https://github.com/huggingface/datasets/issues/6590 | 2,082,000,084 | I_kwDODunzps58GMzU | 6,590 | Feature request: Multi-GPU dataset mapping for SDXL training | {
"login": "kopyl",
"id": 17604849,
"node_id": "MDQ6VXNlcjE3NjA0ODQ5",
"avatar_url": "https://avatars.githubusercontent.com/u/17604849?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/kopyl",
"html_url": "https://github.com/kopyl",
"followers_url": "https://api.github.com/users/kopyl/follow... | [
{
"id": 1935892871,
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement",
"name": "enhancement",
"color": "a2eeef",
"default": true,
"description": "New feature or request"
}
] | open | false | null | [] | null | 0 | 2024-01-15T13:06:06 | 2024-01-15T13:07:07 | null | NONE | null | ### Feature request
We need to speed up SDXL dataset pre-process. Please make it possible to use multiple GPUs for the [official SDXL trainer](https://github.com/huggingface/diffusers/blob/main/examples/text_to_image/train_text_to_image_sdxl.py) :)
### Motivation
Pre-computing 3 million of images takes around ... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6590/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6590/timeline | null | null | null | null | false | [] |
https://api.github.com/repos/huggingface/datasets/issues/6589 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6589/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6589/comments | https://api.github.com/repos/huggingface/datasets/issues/6589/events | https://github.com/huggingface/datasets/issues/6589 | 2,081,358,619 | I_kwDODunzps58DwMb | 6,589 | After `2.16.0` version, there are `PermissionError` when users use shared cache_dir | {
"login": "minhopark-neubla",
"id": 106717516,
"node_id": "U_kgDOBlxhTA",
"avatar_url": "https://avatars.githubusercontent.com/u/106717516?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/minhopark-neubla",
"html_url": "https://github.com/minhopark-neubla",
"followers_url": "https://api.gi... | [] | closed | false | null | [] | null | 2 | 2024-01-15T06:46:27 | 2024-02-02T07:55:38 | 2024-01-30T15:28:38 | NONE | null | ### Describe the bug
- We use shared `cache_dir` using `HF_HOME="{shared_directory}"`
- After dataset version 2.16.0, datasets uses `filelock` package for file locking #6445
- But, `filelock` package make `.lock` file with `644` permission
- Dataset is not available to other users except the user who created the ... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6589/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6589/timeline | null | completed | null | null | false | [
"We'll do a new release of `datasets` in the coming days with a fix !",
"@lhoestq Thank you very much!"
] |
https://api.github.com/repos/huggingface/datasets/issues/6588 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6588/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6588/comments | https://api.github.com/repos/huggingface/datasets/issues/6588/events | https://github.com/huggingface/datasets/issues/6588 | 2,081,284,253 | I_kwDODunzps58DeCd | 6,588 | fix os.listdir return name is empty string | {
"login": "d710055071",
"id": 12895488,
"node_id": "MDQ6VXNlcjEyODk1NDg4",
"avatar_url": "https://avatars.githubusercontent.com/u/12895488?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/d710055071",
"html_url": "https://github.com/d710055071",
"followers_url": "https://api.github.com/use... | [] | closed | false | null | [] | null | 0 | 2024-01-15T05:34:36 | 2024-01-24T10:08:29 | 2024-01-24T10:08:29 | CONTRIBUTOR | null | ### Describe the bug
xlistdir return name is empty string
Overloaded os.listdir
### Steps to reproduce the bug
```python
from datasets.download.streaming_download_manager import xjoin
from datasets.download.streaming_download_manager import xlistdir
config = DownloadConfig(storage_options=options)
manger = Str... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6588/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6588/timeline | null | completed | null | null | false | [] |
https://api.github.com/repos/huggingface/datasets/issues/6587 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6587/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6587/comments | https://api.github.com/repos/huggingface/datasets/issues/6587/events | https://github.com/huggingface/datasets/pull/6587 | 2,080,348,016 | PR_kwDODunzps5kAT_5 | 6,587 | Allow concatenation of datasets with mixed structs | {
"login": "Dref360",
"id": 8976546,
"node_id": "MDQ6VXNlcjg5NzY1NDY=",
"avatar_url": "https://avatars.githubusercontent.com/u/8976546?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Dref360",
"html_url": "https://github.com/Dref360",
"followers_url": "https://api.github.com/users/Dref360/... | [] | closed | false | null | [] | null | 3 | 2024-01-13T15:33:20 | 2024-02-15T15:20:06 | 2024-02-08T14:38:32 | CONTRIBUTOR | null | Fixes #6466
The idea is to do a recursive check for structs. PyArrow handles it well enough.
For a demo you can do:
```python
from datasets import Dataset, concatenate_datasets
ds = Dataset.from_dict({'speaker': [{'name': 'Ben', 'email': None}]})
ds2 = Dataset.from_dict({'speaker': [{'name': 'Fred', 'e... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6587/reactions",
"total_count": 1,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 1,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6587/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6587",
"html_url": "https://github.com/huggingface/datasets/pull/6587",
"diff_url": "https://github.com/huggingface/datasets/pull/6587.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6587.patch",
"merged_at": "2024-02-08T14:38... | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6587). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"friendly bump",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<... |
https://api.github.com/repos/huggingface/datasets/issues/6586 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6586/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6586/comments | https://api.github.com/repos/huggingface/datasets/issues/6586/events | https://github.com/huggingface/datasets/pull/6586 | 2,079,192,651 | PR_kwDODunzps5j8aJn | 6,586 | keep more info in DatasetInfo.from_merge #6585 | {
"login": "JochenSiegWork",
"id": 135010976,
"node_id": "U_kgDOCAwaoA",
"avatar_url": "https://avatars.githubusercontent.com/u/135010976?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/JochenSiegWork",
"html_url": "https://github.com/JochenSiegWork",
"followers_url": "https://api.github.c... | [] | closed | false | null | [] | null | 4 | 2024-01-12T16:08:16 | 2024-01-26T15:59:35 | 2024-01-26T15:53:28 | CONTRIBUTOR | null | * try not to merge DatasetInfos if they're equal
* fixes losing DatasetInfo during parallel Dataset.map | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6586/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6586/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6586",
"html_url": "https://github.com/huggingface/datasets/pull/6586",
"diff_url": "https://github.com/huggingface/datasets/pull/6586.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6586.patch",
"merged_at": "2024-01-26T15:53... | true | [
"@JochenSiegWork fyi, that seems to also affect the `trainer.push_to_hub()` method, which I guess also needs to parse that DatasetInfo from the `kwargs` used by `push_to_hub`.\r\nThere is short discussion about it [here](https://github.com/huggingface/blog/issues/1623).\r\nWould be great if you can check if your PR... |
https://api.github.com/repos/huggingface/datasets/issues/6585 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6585/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6585/comments | https://api.github.com/repos/huggingface/datasets/issues/6585/events | https://github.com/huggingface/datasets/issues/6585 | 2,078,874,005 | I_kwDODunzps576RmV | 6,585 | losing DatasetInfo in Dataset.map when num_proc > 1 | {
"login": "JochenSiegWork",
"id": 135010976,
"node_id": "U_kgDOCAwaoA",
"avatar_url": "https://avatars.githubusercontent.com/u/135010976?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/JochenSiegWork",
"html_url": "https://github.com/JochenSiegWork",
"followers_url": "https://api.github.c... | [] | open | false | {
"login": "JochenSiegWork",
"id": 135010976,
"node_id": "U_kgDOCAwaoA",
"avatar_url": "https://avatars.githubusercontent.com/u/135010976?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/JochenSiegWork",
"html_url": "https://github.com/JochenSiegWork",
"followers_url": "https://api.github.c... | [
{
"login": "JochenSiegWork",
"id": 135010976,
"node_id": "U_kgDOCAwaoA",
"avatar_url": "https://avatars.githubusercontent.com/u/135010976?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/JochenSiegWork",
"html_url": "https://github.com/JochenSiegWork",
"followers_url": "... | null | 2 | 2024-01-12T13:39:19 | 2024-01-12T14:08:24 | null | CONTRIBUTOR | null | ### Describe the bug
Hello and thanks for developing this package!
When I process a Dataset with the map function using multiple processors some set attributes of the DatasetInfo get lost and are None in the resulting Dataset.
### Steps to reproduce the bug
```python
from datasets import Dataset, DatasetInfo... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6585/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6585/timeline | null | null | null | null | false | [
"Hi ! This issue comes from the fact that `map()` with `num_proc>1` shards the dataset in multiple chunks to be processed (one per process) and merges them. The DatasetInfos of each chunk are then merged together, but for some fields like `dataset_name` it's not been implemented and default to None.\r\n\r\nThe Data... |
https://api.github.com/repos/huggingface/datasets/issues/6584 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6584/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6584/comments | https://api.github.com/repos/huggingface/datasets/issues/6584/events | https://github.com/huggingface/datasets/issues/6584 | 2,078,454,878 | I_kwDODunzps574rRe | 6,584 | np.fromfile not supported | {
"login": "d710055071",
"id": 12895488,
"node_id": "MDQ6VXNlcjEyODk1NDg4",
"avatar_url": "https://avatars.githubusercontent.com/u/12895488?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/d710055071",
"html_url": "https://github.com/d710055071",
"followers_url": "https://api.github.com/use... | [] | open | false | null | [] | null | 6 | 2024-01-12T09:46:17 | 2024-01-15T05:20:50 | null | CONTRIBUTOR | null | How to do np.fromfile to use it like np.load
```python
def xnumpy_fromfile(filepath_or_buffer, *args, download_config: Optional[DownloadConfig] = None, **kwargs):
import numpy as np
if hasattr(filepath_or_buffer, "read"):
return np.fromfile(filepath_or_buffer, *args, **kwargs)
else:
... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6584/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6584/timeline | null | null | null | null | false | [
"@lhoestq\r\nCan you provide me with some ideas?",
"Hi ! What's the error ?",
"@lhoestq \r\n```\r\nTraceback (most recent call last):\r\n File \"/home/dongzf/miniconda3/envs/dataset_ai/lib/python3.11/runpy.py\", line 198, in _run_module_as_main\r\n return _run_code(code, main_globals, None,\r\n ^^... |
https://api.github.com/repos/huggingface/datasets/issues/6583 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6583/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6583/comments | https://api.github.com/repos/huggingface/datasets/issues/6583/events | https://github.com/huggingface/datasets/pull/6583 | 2,077,049,491 | PR_kwDODunzps5j1DzY | 6,583 | remove eli5 test | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [] | closed | false | null | [] | null | 2 | 2024-01-11T16:05:20 | 2024-01-11T16:15:34 | 2024-01-11T16:09:24 | MEMBER | null | since the dataset is defunct | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6583/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6583/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6583",
"html_url": "https://github.com/huggingface/datasets/pull/6583",
"diff_url": "https://github.com/huggingface/datasets/pull/6583.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6583.patch",
"merged_at": "2024-01-11T16:09... | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6583). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>... |
https://api.github.com/repos/huggingface/datasets/issues/6582 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6582/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6582/comments | https://api.github.com/repos/huggingface/datasets/issues/6582/events | https://github.com/huggingface/datasets/pull/6582 | 2,076,072,101 | PR_kwDODunzps5jxpry | 6,582 | Fix for Incorrect ex_iterable used with multi num_worker | {
"login": "kq-chen",
"id": 136600500,
"node_id": "U_kgDOCCRbtA",
"avatar_url": "https://avatars.githubusercontent.com/u/136600500?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/kq-chen",
"html_url": "https://github.com/kq-chen",
"followers_url": "https://api.github.com/users/kq-chen/foll... | [] | closed | false | null | [] | null | 2 | 2024-01-11T08:49:43 | 2024-03-01T19:09:14 | 2024-03-01T19:02:33 | CONTRIBUTOR | null | Corrects an issue where `self._ex_iterable` was erroneously used instead of `ex_iterable`, when both Distributed Data Parallel (DDP) and multi num_worker are used concurrently. This improper usage led to the generation of incorrect `shards_indices`, subsequently causing issues with the control flow responsible for work... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6582/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6582/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6582",
"html_url": "https://github.com/huggingface/datasets/pull/6582",
"diff_url": "https://github.com/huggingface/datasets/pull/6582.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6582.patch",
"merged_at": "2024-03-01T19:02... | true | [
"A toy example to reveal the bug.\r\n\r\n```python\r\n\"\"\"\r\nDATASETS_VERBOSITY=debug torchrun --nproc-per-node 2 main.py \r\n\"\"\"\r\nimport torch.utils.data\r\nimport torch.distributed\r\nimport datasets.distributed\r\nimport datasets\r\n\r\n# num shards = 4\r\nshards = [(0, 100), (100, 200), (200, 300), (300... |
https://api.github.com/repos/huggingface/datasets/issues/6581 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6581/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6581/comments | https://api.github.com/repos/huggingface/datasets/issues/6581/events | https://github.com/huggingface/datasets/pull/6581 | 2,075,919,265 | PR_kwDODunzps5jxIbt | 6,581 | fix os.listdir return name is empty string | {
"login": "d710055071",
"id": 12895488,
"node_id": "MDQ6VXNlcjEyODk1NDg4",
"avatar_url": "https://avatars.githubusercontent.com/u/12895488?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/d710055071",
"html_url": "https://github.com/d710055071",
"followers_url": "https://api.github.com/use... | [] | closed | false | null | [] | null | 4 | 2024-01-11T07:10:55 | 2024-01-24T10:14:43 | 2024-01-24T10:08:28 | CONTRIBUTOR | null | fix #6588
xlistdir return name is empty string
for example:
`
from datasets.download.streaming_download_manager import xjoin
from datasets.download.streaming_download_manager import xlistdir
config = DownloadConfig(storage_options=options)
manger = StreamingDownloadManager("ILSVRC2012",download_config=config... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6581/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6581/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6581",
"html_url": "https://github.com/huggingface/datasets/pull/6581",
"diff_url": "https://github.com/huggingface/datasets/pull/6581.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6581.patch",
"merged_at": "2024-01-24T10:08... | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6581). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"\r\nObj [\"name\"] ends with \"/\"",
"@lhoestq \r\n\r\nhello,\r\nCan you help me chec... |
https://api.github.com/repos/huggingface/datasets/issues/6580 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6580/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6580/comments | https://api.github.com/repos/huggingface/datasets/issues/6580/events | https://github.com/huggingface/datasets/issues/6580 | 2,075,645,042 | I_kwDODunzps57t9Ry | 6,580 | dataset cache only stores one config of the dataset in parquet dir, and uses that for all other configs resulting in showing same data in all configs. | {
"login": "kartikgupta321",
"id": 78641018,
"node_id": "MDQ6VXNlcjc4NjQxMDE4",
"avatar_url": "https://avatars.githubusercontent.com/u/78641018?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/kartikgupta321",
"html_url": "https://github.com/kartikgupta321",
"followers_url": "https://api.gi... | [] | closed | false | null | [] | null | 0 | 2024-01-11T03:14:18 | 2024-01-20T12:46:16 | 2024-01-20T12:46:16 | NONE | null | ### Describe the bug
ds = load_dataset("ai2_arc", "ARC-Easy"), i have tried to force redownload, delete cache and changing the cache dir.
### Steps to reproduce the bug
dataset = []
dataset_name = "ai2_arc"
possible_configs = [
'ARC-Challenge',
'ARC-Easy'
]
for config in possible_configs:
data... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6580/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6580/timeline | null | completed | null | null | false | [] |
https://api.github.com/repos/huggingface/datasets/issues/6579 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6579/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6579/comments | https://api.github.com/repos/huggingface/datasets/issues/6579/events | https://github.com/huggingface/datasets/issues/6579 | 2,075,407,473 | I_kwDODunzps57tDRx | 6,579 | Unable to load `eli5` dataset with streaming | {
"login": "haok1402",
"id": 89672451,
"node_id": "MDQ6VXNlcjg5NjcyNDUx",
"avatar_url": "https://avatars.githubusercontent.com/u/89672451?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/haok1402",
"html_url": "https://github.com/haok1402",
"followers_url": "https://api.github.com/users/hao... | [] | closed | false | null | [] | null | 1 | 2024-01-10T23:44:20 | 2024-01-11T09:19:18 | 2024-01-11T09:19:17 | NONE | null | ### Describe the bug
Unable to load `eli5` dataset with streaming.
### Steps to reproduce the bug
This fails with FileNotFoundError: https://files.pushshift.io/reddit/submissions
```
from datasets import load_dataset
load_dataset("eli5", streaming=True)
```
This works correctly.
```
from datasets import lo... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6579/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6579/timeline | null | not_planned | null | null | false | [
"Hi @haok1402, I have created an issue in the Discussion tab of the corresponding dataset: https://huggingface.co/datasets/eli5/discussions/7\r\nLet's continue the discussion there!"
] |
https://api.github.com/repos/huggingface/datasets/issues/6578 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6578/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6578/comments | https://api.github.com/repos/huggingface/datasets/issues/6578/events | https://github.com/huggingface/datasets/pull/6578 | 2,074,923,321 | PR_kwDODunzps5jtthB | 6,578 | Faster webdataset streaming | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [] | closed | false | null | [] | null | 3 | 2024-01-10T18:18:09 | 2024-01-30T18:46:02 | 2024-01-30T18:39:51 | MEMBER | null | requests.get(..., streaming=True) is faster than using HTTP range requests when streaming large TAR files
it can be enabled using block_size=0 in fsspec
cc @rwightman | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6578/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6578/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6578",
"html_url": "https://github.com/huggingface/datasets/pull/6578",
"diff_url": "https://github.com/huggingface/datasets/pull/6578.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6578.patch",
"merged_at": "2024-01-30T18:39... | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6578). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"I added faster streaming support using streaming Requests instances in `huggingface_hub... |
https://api.github.com/repos/huggingface/datasets/issues/6577 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6577/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6577/comments | https://api.github.com/repos/huggingface/datasets/issues/6577/events | https://github.com/huggingface/datasets/issues/6577 | 2,074,790,848 | I_kwDODunzps57qsvA | 6,577 | 502 Server Errors when streaming large dataset | {
"login": "sanchit-gandhi",
"id": 93869735,
"node_id": "U_kgDOBZhWpw",
"avatar_url": "https://avatars.githubusercontent.com/u/93869735?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/sanchit-gandhi",
"html_url": "https://github.com/sanchit-gandhi",
"followers_url": "https://api.github.com... | [
{
"id": 3287858981,
"node_id": "MDU6TGFiZWwzMjg3ODU4OTgx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/streaming",
"name": "streaming",
"color": "fef2c0",
"default": false,
"description": ""
}
] | closed | false | null | [] | null | 6 | 2024-01-10T16:59:36 | 2024-02-12T11:46:03 | 2024-01-15T16:05:44 | CONTRIBUTOR | null | ### Describe the bug
When streaming a [large ASR dataset](https://huggingface.co/datasets/sanchit-gandhi/concatenated-train-set) from the Hug (~3TB) I often encounter 502 Server Errors seemingly randomly during streaming:
```
huggingface_hub.utils._errors.HfHubHTTPError: 502 Server Error: Bad Gateway for url: http... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6577/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6577/timeline | null | completed | null | null | false | [
"cc @mariosasko @lhoestq ",
"Hi! We should be able to avoid this error by retrying to read the data when it happens. I'll open a PR in `huggingface_hub` to address this.",
"Thanks for the fix @mariosasko! Just wondering whether \"500 error\" should also be excluded? I got these errors overnight:\r\n\r\n```\r\nh... |
https://api.github.com/repos/huggingface/datasets/issues/6576 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6576/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6576/comments | https://api.github.com/repos/huggingface/datasets/issues/6576/events | https://github.com/huggingface/datasets/issues/6576 | 2,073,710,124 | I_kwDODunzps57mk4s | 6,576 | document page 404 not found after redirection | {
"login": "annahung31",
"id": 39179888,
"node_id": "MDQ6VXNlcjM5MTc5ODg4",
"avatar_url": "https://avatars.githubusercontent.com/u/39179888?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/annahung31",
"html_url": "https://github.com/annahung31",
"followers_url": "https://api.github.com/use... | [] | closed | false | null | [] | null | 1 | 2024-01-10T06:48:14 | 2024-01-17T14:01:31 | 2024-01-17T14:01:31 | NONE | null | ### Describe the bug
The redirected page encountered 404 not found.
### Steps to reproduce the bug
1. In this tutorial: https://huggingface.co/learn/nlp-course/chapter5/4?fw=pt
original md: https://github.com/huggingface/course/blob/2c733c2246b8b7e0e6f19a9e5d15bb12df43b2a3/chapters/en/chapter5/4.mdx#L49
`... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6576/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6576/timeline | null | completed | null | null | false | [
"Thanks for reporting! I've opened a PR with a fix."
] |
https://api.github.com/repos/huggingface/datasets/issues/6575 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6575/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6575/comments | https://api.github.com/repos/huggingface/datasets/issues/6575/events | https://github.com/huggingface/datasets/pull/6575 | 2,072,617,406 | PR_kwDODunzps5jl1V6 | 6,575 | [IterableDataset] Fix `drop_last_batch`in map after shuffling or sharding | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [] | closed | false | null | [] | null | 2 | 2024-01-09T15:35:31 | 2024-01-11T16:16:54 | 2024-01-11T16:10:30 | MEMBER | null | It was not taken into account e.g. when passing to a DataLoader with num_workers>0
Fix https://github.com/huggingface/datasets/issues/6565 | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6575/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6575/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6575",
"html_url": "https://github.com/huggingface/datasets/pull/6575",
"diff_url": "https://github.com/huggingface/datasets/pull/6575.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6575.patch",
"merged_at": "2024-01-11T16:10... | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6575). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>... |
https://api.github.com/repos/huggingface/datasets/issues/6574 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6574/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6574/comments | https://api.github.com/repos/huggingface/datasets/issues/6574/events | https://github.com/huggingface/datasets/pull/6574 | 2,072,579,549 | PR_kwDODunzps5jltBC | 6,574 | Fix tests based on datasets that used to have scripts | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [] | closed | false | null | [] | null | 2 | 2024-01-09T15:16:16 | 2024-01-09T16:11:33 | 2024-01-09T16:05:13 | MEMBER | null | ...now that `squad` and `paws` don't have a script anymore | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6574/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6574/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6574",
"html_url": "https://github.com/huggingface/datasets/pull/6574",
"diff_url": "https://github.com/huggingface/datasets/pull/6574.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6574.patch",
"merged_at": "2024-01-09T16:05... | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6574). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>... |
https://api.github.com/repos/huggingface/datasets/issues/6573 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6573/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6573/comments | https://api.github.com/repos/huggingface/datasets/issues/6573/events | https://github.com/huggingface/datasets/pull/6573 | 2,072,553,951 | PR_kwDODunzps5jlnaj | 6,573 | [WebDataset] Audio support and bug fixes | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [] | closed | false | null | [] | null | 2 | 2024-01-09T15:03:04 | 2024-01-11T16:17:28 | 2024-01-11T16:11:04 | MEMBER | null | - Add audio support
- Fix an issue where user-provided features with additional fields are not taken into account
Close https://github.com/huggingface/datasets/issues/6569 | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6573/reactions",
"total_count": 1,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 1,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6573/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6573",
"html_url": "https://github.com/huggingface/datasets/pull/6573",
"diff_url": "https://github.com/huggingface/datasets/pull/6573.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6573.patch",
"merged_at": "2024-01-11T16:11... | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6573). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>... |
https://api.github.com/repos/huggingface/datasets/issues/6572 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6572/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6572/comments | https://api.github.com/repos/huggingface/datasets/issues/6572/events | https://github.com/huggingface/datasets/pull/6572 | 2,072,384,281 | PR_kwDODunzps5jlCO5 | 6,572 | Adding option for multipart achive download | {
"login": "jpodivin",
"id": 66251151,
"node_id": "MDQ6VXNlcjY2MjUxMTUx",
"avatar_url": "https://avatars.githubusercontent.com/u/66251151?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jpodivin",
"html_url": "https://github.com/jpodivin",
"followers_url": "https://api.github.com/users/jpo... | [] | closed | false | null | [] | null | 1 | 2024-01-09T13:35:44 | 2024-02-25T08:13:01 | 2024-02-25T08:13:01 | NONE | null | Right now we can only download multiple separate archives or a single file archive, but not multipart archives, such as those produced by `tar --multi-volume`. This PR allows for downloading and extraction of archives split into multiple parts.
With the new `multi_part` field of the `DownloadConfig` set, the downloa... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6572/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6572/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6572",
"html_url": "https://github.com/huggingface/datasets/pull/6572",
"diff_url": "https://github.com/huggingface/datasets/pull/6572.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6572.patch",
"merged_at": null
} | true | [
"On closer examination, this appears to be unnecessary. "
] |
https://api.github.com/repos/huggingface/datasets/issues/6571 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6571/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6571/comments | https://api.github.com/repos/huggingface/datasets/issues/6571/events | https://github.com/huggingface/datasets/issues/6571 | 2,072,111,000 | I_kwDODunzps57geeY | 6,571 | Make DatasetDict.column_names return a list instead of dict | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"id": 1935892871,
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement",
"name": "enhancement",
"color": "a2eeef",
"default": true,
"description": "New feature or request"
}
] | open | false | null | [] | null | 0 | 2024-01-09T10:45:17 | 2024-01-09T10:45:17 | null | MEMBER | null | Currently, `DatasetDict.column_names` returns a dict, with each split name as keys and the corresponding list of column names as values.
However, by construction, all splits have the same column names.
I think it makes more sense to return a single list with the column names, which is the same for all the split k... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6571/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6571/timeline | null | null | null | null | false | [] |
https://api.github.com/repos/huggingface/datasets/issues/6570 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6570/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6570/comments | https://api.github.com/repos/huggingface/datasets/issues/6570/events | https://github.com/huggingface/datasets/issues/6570 | 2,071,805,265 | I_kwDODunzps57fT1R | 6,570 | No online docs for 2.16 release | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 1935892861,
"node_id": "MDU6TGFiZWwxOTM1ODk... | closed | false | null | [] | null | 7 | 2024-01-09T07:43:30 | 2024-01-09T16:45:50 | 2024-01-09T16:45:50 | MEMBER | null | We do not have the online docs for the latest minor release 2.16 (2.16.0 nor 2.16.1).
In the online docs, the latest version appearing is 2.15.0: https://huggingface.co/docs/datasets/index
 . But a new issue came up :( | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6568/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6568/timeline | null | null | null | null | false | [
"Seems like I just used the old code which did not have `keep_in_memory=True` argument, sorry.\r\n\r\nAlthough i encountered a different problem – at 97% my python process just hung for around 11 minutes with no logs (when running dataset.map without `keep_in_memory=True` over around 3 million of dataset samples)..... |
https://api.github.com/repos/huggingface/datasets/issues/6567 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6567/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6567/comments | https://api.github.com/repos/huggingface/datasets/issues/6567/events | https://github.com/huggingface/datasets/issues/6567 | 2,069,808,842 | I_kwDODunzps57XsbK | 6,567 | AttributeError: 'str' object has no attribute 'to' | {
"login": "andysingal",
"id": 20493493,
"node_id": "MDQ6VXNlcjIwNDkzNDkz",
"avatar_url": "https://avatars.githubusercontent.com/u/20493493?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/andysingal",
"html_url": "https://github.com/andysingal",
"followers_url": "https://api.github.com/use... | [] | closed | false | null | [] | null | 3 | 2024-01-08T06:40:21 | 2024-01-08T11:56:19 | 2024-01-08T10:03:17 | NONE | null | ### Describe the bug
```
--------------------------------------------------------------------------
AttributeError Traceback (most recent call last)
[<ipython-input-6-80c6086794e8>](https://localhost:8080/#) in <cell line: 10>()
8 report_to="wandb")
9
---> 10 trainer =... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6567/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6567/timeline | null | completed | null | null | false | [
"I think you are reporting an issue with the `transformers` library. Note this is the repository of the `datasets` library. I recommend that you open an issue in their repository: https://github.com/huggingface/transformers/issues\r\n\r\nEDIT: I have not the rights to transfer the issue\r\n~~I am transferring your ... |
https://api.github.com/repos/huggingface/datasets/issues/6566 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6566/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6566/comments | https://api.github.com/repos/huggingface/datasets/issues/6566/events | https://github.com/huggingface/datasets/issues/6566 | 2,069,495,429 | I_kwDODunzps57Wf6F | 6,566 | I train controlnet_sdxl in bf16 datatype, got unsupported ERROR in datasets | {
"login": "HelloWorldBeginner",
"id": 25008090,
"node_id": "MDQ6VXNlcjI1MDA4MDkw",
"avatar_url": "https://avatars.githubusercontent.com/u/25008090?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/HelloWorldBeginner",
"html_url": "https://github.com/HelloWorldBeginner",
"followers_url": "ht... | [] | open | false | null | [] | null | 1 | 2024-01-08T02:37:03 | 2024-01-20T00:29:18 | null | NONE | null | ### Describe the bug
```
Traceback (most recent call last):
File "train_controlnet_sdxl.py", line 1252, in <module>
main(args)
File "train_controlnet_sdxl.py", line 1013, in main
train_dataset = train_dataset.map(compute_embeddings_fn, batched=True, new_fingerprint=new_fingerprint)
File "/home/mini... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6566/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6566/timeline | null | null | null | null | false | [
"I also see the same error and get passed it by casting that line to float. \r\n\r\nso `for x in obj.detach().cpu().numpy()` becomes `for x in obj.detach().to(torch.float).cpu().numpy()`\r\n\r\nI got the idea from [this ](https://github.com/kohya-ss/sd-webui-additional-networks/pull/128/files) PR where someone was... |
https://api.github.com/repos/huggingface/datasets/issues/6565 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6565/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6565/comments | https://api.github.com/repos/huggingface/datasets/issues/6565/events | https://github.com/huggingface/datasets/issues/6565 | 2,068,939,670 | I_kwDODunzps57UYOW | 6,565 | `drop_last_batch=True` for IterableDataset map function is ignored with multiprocessing DataLoader | {
"login": "naba89",
"id": 12119806,
"node_id": "MDQ6VXNlcjEyMTE5ODA2",
"avatar_url": "https://avatars.githubusercontent.com/u/12119806?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/naba89",
"html_url": "https://github.com/naba89",
"followers_url": "https://api.github.com/users/naba89/fo... | [] | closed | false | null | [] | null | 1 | 2024-01-07T02:46:50 | 2024-01-11T16:10:31 | 2024-01-11T16:10:31 | NONE | null | ### Describe the bug
Scenario:
- Interleaving two iterable datasets of unequal lengths (`all_exhausted`), followed by a batch mapping with batch size 2 to effectively merge the two datasets and get a sample from each dataset in a single batch, with `drop_last_batch=True` to skip the last batch in case it doesn't ha... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6565/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6565/timeline | null | completed | null | null | false | [
"My current workaround this issue is to return `None` in the second element and then filter out samples which have `None` in them.\r\n\r\n```python\r\ndef merge_samples(batch):\r\n if len(batch['a']) == 1:\r\n batch['c'] = [batch['a'][0]]\r\n batch['d'] = [None]\r\n else:\r\n batch['c'] ... |
https://api.github.com/repos/huggingface/datasets/issues/6564 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6564/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6564/comments | https://api.github.com/repos/huggingface/datasets/issues/6564/events | https://github.com/huggingface/datasets/issues/6564 | 2,068,893,194 | I_kwDODunzps57UM4K | 6,564 | `Dataset.filter` missing `with_rank` parameter | {
"login": "kopyl",
"id": 17604849,
"node_id": "MDQ6VXNlcjE3NjA0ODQ5",
"avatar_url": "https://avatars.githubusercontent.com/u/17604849?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/kopyl",
"html_url": "https://github.com/kopyl",
"followers_url": "https://api.github.com/users/kopyl/follow... | [] | closed | false | null | [] | null | 2 | 2024-01-06T23:48:13 | 2024-01-29T16:36:55 | 2024-01-29T16:36:54 | NONE | null | ### Describe the bug
The issue shall be open: https://github.com/huggingface/datasets/issues/6435
When i try to pass `with_rank` to `Dataset.filter()`, i get this:
`Dataset.filter() got an unexpected keyword argument 'with_rank'`
### Steps to reproduce the bug
Run notebook:
https://colab.research.google.com... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6564/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6564/timeline | null | completed | null | null | false | [
"Thanks for reporting! I've opened a PR with a fix",
"@mariosasko thank you very much :)"
] |
https://api.github.com/repos/huggingface/datasets/issues/6563 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6563/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6563/comments | https://api.github.com/repos/huggingface/datasets/issues/6563/events | https://github.com/huggingface/datasets/issues/6563 | 2,068,302,402 | I_kwDODunzps57R8pC | 6,563 | `ImportError`: cannot import name 'insecure_hashlib' from 'huggingface_hub.utils' (.../huggingface_hub/utils/__init__.py) | {
"login": "wasertech",
"id": 79070834,
"node_id": "MDQ6VXNlcjc5MDcwODM0",
"avatar_url": "https://avatars.githubusercontent.com/u/79070834?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/wasertech",
"html_url": "https://github.com/wasertech",
"followers_url": "https://api.github.com/users/... | [] | closed | false | null | [] | null | 7 | 2024-01-06T02:28:54 | 2024-03-14T02:59:42 | 2024-01-06T16:13:27 | NONE | null | ### Describe the bug
Yep its not [there](https://github.com/huggingface/huggingface_hub/blob/main/src/huggingface_hub/utils/__init__.py) anymore.
```text
+ python /home/trainer/sft_train.py --model_name cognitivecomputations/dolphin-2.2.1-mistral-7b --dataset_name wasertech/OneOS --load_in_4bit --use_peft --batch_... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6563/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6563/timeline | null | completed | null | null | false | [
"@Wauplin Do you happen to know what's up?",
"<del>Installing `datasets` from `main` did the trick so I guess it will be fixed in the next release.\r\n\r\nNVM https://github.com/huggingface/datasets/blob/d26abadce0b884db32382b92422d8a6aa997d40a/src/datasets/utils/info_utils.py#L5",
"@wasertech upgrading `huggin... |
https://api.github.com/repos/huggingface/datasets/issues/6562 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6562/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6562/comments | https://api.github.com/repos/huggingface/datasets/issues/6562/events | https://github.com/huggingface/datasets/issues/6562 | 2,067,904,504 | I_kwDODunzps57Qbf4 | 6,562 | datasets.DownloadMode.FORCE_REDOWNLOAD use cache to download dataset features with load_dataset function | {
"login": "LsTam91",
"id": 73234162,
"node_id": "MDQ6VXNlcjczMjM0MTYy",
"avatar_url": "https://avatars.githubusercontent.com/u/73234162?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/LsTam91",
"html_url": "https://github.com/LsTam91",
"followers_url": "https://api.github.com/users/LsTam9... | [] | open | false | null | [] | null | 0 | 2024-01-05T19:10:25 | 2024-01-05T19:10:25 | null | NONE | null | ### Describe the bug
I have updated my dataset by adding a new feature, and push it to the hub. When I want to download it on my machine which contain the old version by using `datasets.load_dataset("your_dataset_name", download_mode=datasets.DownloadMode.FORCE_REDOWNLOAD)` I get an error (paste bellow).
Seems that... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6562/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6562/timeline | null | null | null | null | false | [] |
https://api.github.com/repos/huggingface/datasets/issues/6561 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6561/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6561/comments | https://api.github.com/repos/huggingface/datasets/issues/6561/events | https://github.com/huggingface/datasets/issues/6561 | 2,067,404,951 | I_kwDODunzps57OhiX | 6,561 | Document YAML configuration with "data_dir" | {
"login": "severo",
"id": 1676121,
"node_id": "MDQ6VXNlcjE2NzYxMjE=",
"avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/severo",
"html_url": "https://github.com/severo",
"followers_url": "https://api.github.com/users/severo/foll... | [
{
"id": 1935892861,
"node_id": "MDU6TGFiZWwxOTM1ODkyODYx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/documentation",
"name": "documentation",
"color": "0075ca",
"default": true,
"description": "Improvements or additions to documentation"
}
] | open | false | null | [] | null | 1 | 2024-01-05T14:03:33 | 2024-01-05T14:06:18 | null | CONTRIBUTOR | null | See https://huggingface.co/datasets/uonlp/CulturaX/discussions/15#6597e83f185db94370d6bf50 for reference | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6561/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6561/timeline | null | null | null | null | false | [
"In particular, I would like to have an example of how to replace the following configuration (from https://huggingface.co/docs/hub/datasets-manual-configuration#splits)\r\n\r\n```\r\n---\r\nconfigs:\r\n- config_name: default\r\n data_files:\r\n - split: train\r\n path: \"data/*.csv\"\r\n - split: test\r\n ... |
https://api.github.com/repos/huggingface/datasets/issues/6560 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6560/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6560/comments | https://api.github.com/repos/huggingface/datasets/issues/6560/events | https://github.com/huggingface/datasets/issues/6560 | 2,065,637,625 | I_kwDODunzps57HyD5 | 6,560 | Support Video | {
"login": "yuvalkirstain",
"id": 57996478,
"node_id": "MDQ6VXNlcjU3OTk2NDc4",
"avatar_url": "https://avatars.githubusercontent.com/u/57996478?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/yuvalkirstain",
"html_url": "https://github.com/yuvalkirstain",
"followers_url": "https://api.githu... | [
{
"id": 1935892871,
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement",
"name": "enhancement",
"color": "a2eeef",
"default": true,
"description": "New feature or request"
}
] | open | false | null | [] | null | 0 | 2024-01-04T13:10:58 | 2024-01-04T13:10:58 | null | NONE | null | ### Feature request
HF datasets are awesome in supporting text and images. Will be great to see such a support in videos :)
### Motivation
Video generation :)
### Your contribution
Will probably be limited to raising this feature request ;) | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6560/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6560/timeline | null | null | null | null | false | [] |
https://api.github.com/repos/huggingface/datasets/issues/6559 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6559/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6559/comments | https://api.github.com/repos/huggingface/datasets/issues/6559/events | https://github.com/huggingface/datasets/issues/6559 | 2,065,118,332 | I_kwDODunzps57FzR8 | 6,559 | Latest version 2.16.1, when load dataset error occurs. ValueError: BuilderConfig 'allenai--c4' not found. Available: ['default'] | {
"login": "zhulinJulia24",
"id": 145004780,
"node_id": "U_kgDOCKSY7A",
"avatar_url": "https://avatars.githubusercontent.com/u/145004780?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/zhulinJulia24",
"html_url": "https://github.com/zhulinJulia24",
"followers_url": "https://api.github.com/... | [] | closed | false | null | [] | null | 2 | 2024-01-04T07:04:48 | 2024-01-05T01:26:26 | 2024-01-05T01:26:25 | NONE | null | ### Describe the bug
python script is:
```
from datasets import load_dataset
cache_dir = 'path/to/your/cache/directory'
dataset = load_dataset('allenai/c4','allenai--c4', data_files={'train': 'en/c4-train.00000-of-01024.json.gz'}, split='train', use_auth_token=False, cache_dir=cache_dir)
```
the script su... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6559/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6559/timeline | null | completed | null | null | false | [
"Hi ! The \"allenai--c4\" config doesn't exist (this naming schema comes from old versions of `datasets`)\r\n\r\nYou can load it this way instead:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\ncache_dir = 'path/to/your/cache/directory'\r\ndataset = load_dataset('allenai/c4', data_files={'train': 'en/c4-t... |
https://api.github.com/repos/huggingface/datasets/issues/6558 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6558/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6558/comments | https://api.github.com/repos/huggingface/datasets/issues/6558/events | https://github.com/huggingface/datasets/issues/6558 | 2,064,885,984 | I_kwDODunzps57E6jg | 6,558 | OSError: image file is truncated (1 bytes not processed) #28323 | {
"login": "andysingal",
"id": 20493493,
"node_id": "MDQ6VXNlcjIwNDkzNDkz",
"avatar_url": "https://avatars.githubusercontent.com/u/20493493?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/andysingal",
"html_url": "https://github.com/andysingal",
"followers_url": "https://api.github.com/use... | [] | closed | false | null | [] | null | 1 | 2024-01-04T02:15:13 | 2024-02-21T00:38:12 | 2024-02-21T00:38:12 | NONE | null | ### Describe the bug
```
---------------------------------------------------------------------------
OSError Traceback (most recent call last)
Cell In[24], line 28
23 return example
25 # Filter the dataset
26 # filtered_dataset = dataset.filter(contains_number... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6558/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6558/timeline | null | completed | null | null | false | [
"You can add \r\n\r\n```python\r\nfrom PIL import ImageFile\r\nImageFile.LOAD_TRUNCATED_IMAGES = True\r\n```\r\n\r\nafter the imports to be able to read truncated images."
] |
https://api.github.com/repos/huggingface/datasets/issues/6557 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6557/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6557/comments | https://api.github.com/repos/huggingface/datasets/issues/6557/events | https://github.com/huggingface/datasets/pull/6557 | 2,064,341,965 | PR_kwDODunzps5jJ63z | 6,557 | Support standalone yaml | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [] | closed | false | null | [] | null | 4 | 2024-01-03T16:47:35 | 2024-01-11T17:59:51 | 2024-01-11T17:53:42 | MEMBER | null | see (internal) https://huggingface.slack.com/archives/C02V51Q3800/p1703885853581679 | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6557/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6557/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6557",
"html_url": "https://github.com/huggingface/datasets/pull/6557",
"diff_url": "https://github.com/huggingface/datasets/pull/6557.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6557.patch",
"merged_at": "2024-01-11T17:53... | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6557). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"@lhoestq \r\nhello\r\nI think it should be defined in config.py\r\nDATASET_ README_ FIL... |
https://api.github.com/repos/huggingface/datasets/issues/6556 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6556/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6556/comments | https://api.github.com/repos/huggingface/datasets/issues/6556/events | https://github.com/huggingface/datasets/pull/6556 | 2,064,018,208 | PR_kwDODunzps5jI0nN | 6,556 | Fix imagefolder with one image | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [] | closed | false | null | [] | null | 2 | 2024-01-03T13:13:02 | 2024-02-12T21:57:34 | 2024-01-09T13:06:30 | MEMBER | null | A dataset repository with one image and one metadata file was considered a JSON dataset instead of an ImageFolder dataset. This is because we pick the dataset type with the most compatible data file extensions present in the repository and it results in a tie in this case.
e.g. for https://huggingface.co/datasets/mu... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6556/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6556/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6556",
"html_url": "https://github.com/huggingface/datasets/pull/6556",
"diff_url": "https://github.com/huggingface/datasets/pull/6556.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6556.patch",
"merged_at": "2024-01-09T13:06... | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6556). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"Fixed in dataset viewer: https://huggingface.co/datasets/multimodalart/repro_1_image\r\... |
https://api.github.com/repos/huggingface/datasets/issues/6555 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6555/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6555/comments | https://api.github.com/repos/huggingface/datasets/issues/6555/events | https://github.com/huggingface/datasets/pull/6555 | 2,063,841,286 | PR_kwDODunzps5jIM79 | 6,555 | Do not use Parquet exports if revision is passed | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | 4 | 2024-01-03T11:33:10 | 2024-02-02T10:41:33 | 2024-02-02T10:35:28 | MEMBER | null | Fix #6554. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6555/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6555/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6555",
"html_url": "https://github.com/huggingface/datasets/pull/6555",
"diff_url": "https://github.com/huggingface/datasets/pull/6555.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6555.patch",
"merged_at": "2024-02-02T10:35... | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6555). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"As shared on slack, `HubDatasetModuleFactoryWithParquetExport` raises a `DatasetsServer... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.