url stringlengths 61 61 | repository_url stringclasses 1
value | labels_url stringlengths 75 75 | comments_url stringlengths 70 70 | events_url stringlengths 68 68 | html_url stringlengths 49 51 | id int64 1.23B 2.21B | node_id stringlengths 18 19 | number int64 4.29k 6.76k | title stringlengths 1 290 | user dict | labels listlengths 0 4 | state stringclasses 2
values | locked bool 1
class | assignee dict | assignees listlengths 0 3 | milestone dict | comments int64 0 48 | created_at timestamp[ms] | updated_at timestamp[ms] | closed_at timestamp[ms] | author_association stringclasses 3
values | active_lock_reason null | body stringlengths 2 33.9k ⌀ | reactions dict | timeline_url stringlengths 70 70 | performed_via_github_app null | state_reason stringclasses 3
values | draft bool 2
classes | pull_request dict | is_pull_request bool 2
classes | comments_body listlengths 0 30 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
https://api.github.com/repos/huggingface/datasets/issues/6453 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6453/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6453/comments | https://api.github.com/repos/huggingface/datasets/issues/6453/events | https://github.com/huggingface/datasets/pull/6453 | 2,011,907,787 | PR_kwDODunzps5ga0rv | 6,453 | Update hub-docs reference | {
"login": "mishig25",
"id": 11827707,
"node_id": "MDQ6VXNlcjExODI3NzA3",
"avatar_url": "https://avatars.githubusercontent.com/u/11827707?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mishig25",
"html_url": "https://github.com/mishig25",
"followers_url": "https://api.github.com/users/mis... | [] | closed | false | null | [] | null | 3 | 2023-11-27T09:57:20 | 2023-11-27T10:23:44 | 2023-11-27T10:17:34 | CONTRIBUTOR | null | Follow up to huggingface/huggingface.js#296 | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6453/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6453/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6453",
"html_url": "https://github.com/huggingface/datasets/pull/6453",
"diff_url": "https://github.com/huggingface/datasets/pull/6453.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6453.patch",
"merged_at": "2023-11-27T10:17... | true | [
"_The documentation is not available anymore as the PR was closed or merged._",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea... |
https://api.github.com/repos/huggingface/datasets/issues/6452 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6452/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6452/comments | https://api.github.com/repos/huggingface/datasets/issues/6452/events | https://github.com/huggingface/datasets/pull/6452 | 2,011,632,708 | PR_kwDODunzps5gZ5oe | 6,452 | Praveen_repo_pull_req | {
"login": "Praveenhh",
"id": 151713216,
"node_id": "U_kgDOCQr1wA",
"avatar_url": "https://avatars.githubusercontent.com/u/151713216?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Praveenhh",
"html_url": "https://github.com/Praveenhh",
"followers_url": "https://api.github.com/users/Pravee... | [] | closed | false | null | [] | null | 0 | 2023-11-27T07:07:50 | 2023-11-27T09:28:00 | 2023-11-27T09:28:00 | NONE | null | null | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6452/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6452/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6452",
"html_url": "https://github.com/huggingface/datasets/pull/6452",
"diff_url": "https://github.com/huggingface/datasets/pull/6452.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6452.patch",
"merged_at": null
} | true | [] |
https://api.github.com/repos/huggingface/datasets/issues/6451 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6451/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6451/comments | https://api.github.com/repos/huggingface/datasets/issues/6451/events | https://github.com/huggingface/datasets/issues/6451 | 2,010,693,912 | I_kwDODunzps532MEY | 6,451 | Unable to read "marsyas/gtzan" data | {
"login": "gerald-wrona",
"id": 32300890,
"node_id": "MDQ6VXNlcjMyMzAwODkw",
"avatar_url": "https://avatars.githubusercontent.com/u/32300890?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/gerald-wrona",
"html_url": "https://github.com/gerald-wrona",
"followers_url": "https://api.github.c... | [] | closed | false | null | [] | null | 3 | 2023-11-25T15:13:17 | 2023-12-01T12:53:46 | 2023-11-27T09:36:25 | NONE | null | Hi, this is my code and the error:
```
from datasets import load_dataset
gtzan = load_dataset("marsyas/gtzan", "all")
```
[error_trace.txt](https://github.com/huggingface/datasets/files/13464397/error_trace.txt)
[audio_yml.txt](https://github.com/huggingface/datasets/files/13464410/audio_yml.txt)
Python 3.11.5
... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6451/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6451/timeline | null | completed | null | null | false | [
"Hi! We've merged a [PR](https://huggingface.co/datasets/marsyas/gtzan/discussions/1) that fixes the script's path logic on Windows.",
"I have transferred the discussion to the corresponding dataset: https://huggingface.co/datasets/marsyas/gtzan/discussions/2\r\n\r\nLet's continue there.",
"@mariosasko @albertv... |
https://api.github.com/repos/huggingface/datasets/issues/6450 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6450/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6450/comments | https://api.github.com/repos/huggingface/datasets/issues/6450/events | https://github.com/huggingface/datasets/issues/6450 | 2,009,491,386 | I_kwDODunzps53xme6 | 6,450 | Support multiple image/audio columns in ImageFolder/AudioFolder | {
"login": "severo",
"id": 1676121,
"node_id": "MDQ6VXNlcjE2NzYxMjE=",
"avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/severo",
"html_url": "https://github.com/severo",
"followers_url": "https://api.github.com/users/severo/foll... | [
{
"id": 1935892865,
"node_id": "MDU6TGFiZWwxOTM1ODkyODY1",
"url": "https://api.github.com/repos/huggingface/datasets/labels/duplicate",
"name": "duplicate",
"color": "cfd3d7",
"default": true,
"description": "This issue or pull request already exists"
},
{
"id": 1935892871,
"... | closed | false | null | [] | null | 1 | 2023-11-24T10:34:09 | 2023-11-28T11:07:17 | 2023-11-24T17:24:38 | CONTRIBUTOR | null | ### Feature request
Have a metadata.csv file with multiple columns that point to relative image or audio files.
### Motivation
Currently, ImageFolder allows one column, called `file_name`, pointing to relative image files. On the same model, AudioFolder allows one column, called `file_name`, pointing to relative aud... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6450/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6450/timeline | null | completed | null | null | false | [
"A duplicate of https://github.com/huggingface/datasets/issues/5760"
] |
https://api.github.com/repos/huggingface/datasets/issues/6449 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6449/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6449/comments | https://api.github.com/repos/huggingface/datasets/issues/6449/events | https://github.com/huggingface/datasets/pull/6449 | 2,008,617,992 | PR_kwDODunzps5gQCVZ | 6,449 | Fix metadata file resolution when inferred pattern is `**` | {
"login": "mariosasko",
"id": 47462742,
"node_id": "MDQ6VXNlcjQ3NDYyNzQy",
"avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mariosasko",
"html_url": "https://github.com/mariosasko",
"followers_url": "https://api.github.com/use... | [] | closed | false | null | [] | null | 6 | 2023-11-23T17:35:02 | 2023-11-27T10:02:56 | 2023-11-24T17:13:02 | CONTRIBUTOR | null | Refetch metadata files in case they were dropped by `filter_extensions` in the previous step.
Fix #6442
| {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6449/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6449/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6449",
"html_url": "https://github.com/huggingface/datasets/pull/6449",
"diff_url": "https://github.com/huggingface/datasets/pull/6449.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6449.patch",
"merged_at": "2023-11-24T17:13... | true | [
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a... |
https://api.github.com/repos/huggingface/datasets/issues/6448 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6448/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6448/comments | https://api.github.com/repos/huggingface/datasets/issues/6448/events | https://github.com/huggingface/datasets/pull/6448 | 2,008,614,985 | PR_kwDODunzps5gQBsE | 6,448 | Use parquet export if possible | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [] | closed | false | null | [] | null | 24 | 2023-11-23T17:31:57 | 2023-12-01T17:57:17 | 2023-12-01T17:50:59 | MEMBER | null | The idea is to make this code work for datasets with scripts if they have a Parquet export
```python
ds = load_dataset("squad", trust_remote_code=False)
```
And more generally, it means we use the Parquet export whenever it's possible (it's safer and faster than dataset scripts).
I also added a `config.USE_P... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6448/reactions",
"total_count": 2,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 2,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6448/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6448",
"html_url": "https://github.com/huggingface/datasets/pull/6448",
"diff_url": "https://github.com/huggingface/datasets/pull/6448.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6448.patch",
"merged_at": "2023-12-01T17:50... | true | [
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a... |
https://api.github.com/repos/huggingface/datasets/issues/6447 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6447/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6447/comments | https://api.github.com/repos/huggingface/datasets/issues/6447/events | https://github.com/huggingface/datasets/issues/6447 | 2,008,195,298 | I_kwDODunzps53sqDi | 6,447 | Support one dataset loader per config when using YAML | {
"login": "severo",
"id": 1676121,
"node_id": "MDQ6VXNlcjE2NzYxMjE=",
"avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/severo",
"html_url": "https://github.com/severo",
"followers_url": "https://api.github.com/users/severo/foll... | [
{
"id": 1935892871,
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement",
"name": "enhancement",
"color": "a2eeef",
"default": true,
"description": "New feature or request"
}
] | open | false | null | [] | null | 0 | 2023-11-23T13:03:07 | 2023-11-23T13:03:07 | null | CONTRIBUTOR | null | ### Feature request
See https://huggingface.co/datasets/datasets-examples/doc-unsupported-1
I would like to use CSV loader for the "csv" config, JSONL loader for the "jsonl" config, etc.
### Motivation
It would be more flexible for the users
### Your contribution
No specific contribution | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6447/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6447/timeline | null | null | null | null | false | [] |
https://api.github.com/repos/huggingface/datasets/issues/6446 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6446/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6446/comments | https://api.github.com/repos/huggingface/datasets/issues/6446/events | https://github.com/huggingface/datasets/issues/6446 | 2,007,092,708 | I_kwDODunzps53oc3k | 6,446 | Speech Commands v2 dataset doesn't match AST-v2 config | {
"login": "vymao",
"id": 18024303,
"node_id": "MDQ6VXNlcjE4MDI0MzAz",
"avatar_url": "https://avatars.githubusercontent.com/u/18024303?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/vymao",
"html_url": "https://github.com/vymao",
"followers_url": "https://api.github.com/users/vymao/follow... | [] | closed | false | null | [] | null | 3 | 2023-11-22T20:46:36 | 2023-11-28T14:46:08 | 2023-11-28T14:46:08 | NONE | null | ### Describe the bug
[According](https://huggingface.co/MIT/ast-finetuned-speech-commands-v2) to `MIT/ast-finetuned-speech-commands-v2`, the model was trained on the Speech Commands v2 dataset. However, while the model config says the model should have 35 class labels, the dataset itself has 36 class labels. Moreover,... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6446/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6446/timeline | null | completed | null | null | false | [
"You can use `.align_labels_with_mapping` on the dataset to align the labels with the model config.\r\n\r\nRegarding the number of labels, only the special `_silence_` label corresponding to noise is missing, which is consistent with the model paper (reports training on 35 labels). You can run a `.filter` to drop ... |
https://api.github.com/repos/huggingface/datasets/issues/6445 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6445/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6445/comments | https://api.github.com/repos/huggingface/datasets/issues/6445/events | https://github.com/huggingface/datasets/pull/6445 | 2,006,958,595 | PR_kwDODunzps5gKg2d | 6,445 | Use `filelock` package for file locking | {
"login": "mariosasko",
"id": 47462742,
"node_id": "MDQ6VXNlcjQ3NDYyNzQy",
"avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mariosasko",
"html_url": "https://github.com/mariosasko",
"followers_url": "https://api.github.com/use... | [] | closed | false | null | [] | null | 4 | 2023-11-22T19:04:45 | 2023-11-23T18:47:30 | 2023-11-23T18:41:23 | CONTRIBUTOR | null | Use the `filelock` package instead of `datasets.utils.filelock` for file locking to be consistent with `huggingface_hub` and not to be responsible for improving the `filelock` capabilities 🙂.
(Reverts https://github.com/huggingface/datasets/pull/859, but these `INFO` logs are not printed by default (anymore?), so ... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6445/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6445/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6445",
"html_url": "https://github.com/huggingface/datasets/pull/6445",
"diff_url": "https://github.com/huggingface/datasets/pull/6445.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6445.patch",
"merged_at": "2023-11-23T18:41... | true | [
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a... |
https://api.github.com/repos/huggingface/datasets/issues/6444 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6444/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6444/comments | https://api.github.com/repos/huggingface/datasets/issues/6444/events | https://github.com/huggingface/datasets/pull/6444 | 2,006,842,179 | PR_kwDODunzps5gKG_e | 6,444 | Remove `Table.__getstate__` and `Table.__setstate__` | {
"login": "LZHgrla",
"id": 36994684,
"node_id": "MDQ6VXNlcjM2OTk0Njg0",
"avatar_url": "https://avatars.githubusercontent.com/u/36994684?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/LZHgrla",
"html_url": "https://github.com/LZHgrla",
"followers_url": "https://api.github.com/users/LZHgrl... | [] | closed | false | null | [] | null | 4 | 2023-11-22T17:55:10 | 2023-11-23T15:19:43 | 2023-11-23T15:13:28 | CONTRIBUTOR | null | When using distributed training, the code of `os.remove(filename)` may be executed separately by each rank, leading to `FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmprxxxxxxx.arrow'`
```python
from torch import distributed as dist
if dist.get_rank() == 0:
dataset = process_dataset(*args, ... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6444/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6444/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6444",
"html_url": "https://github.com/huggingface/datasets/pull/6444",
"diff_url": "https://github.com/huggingface/datasets/pull/6444.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6444.patch",
"merged_at": "2023-11-23T15:13... | true | [
"Thanks for working on this! The [issue](https://bugs.python.org/issue24658) with pickling objects larger than 4GB seems to be patched in Python 3.8 (the minimal supported version was 3.6 at the time of implementing this), so a simple solution would be removing the `Table.__setstate__` and `Table.__getstate__` over... |
https://api.github.com/repos/huggingface/datasets/issues/6443 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6443/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6443/comments | https://api.github.com/repos/huggingface/datasets/issues/6443/events | https://github.com/huggingface/datasets/issues/6443 | 2,006,568,368 | I_kwDODunzps53mc2w | 6,443 | Trouble loading files defined in YAML explicitly | {
"login": "severo",
"id": 1676121,
"node_id": "MDQ6VXNlcjE2NzYxMjE=",
"avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/severo",
"html_url": "https://github.com/severo",
"followers_url": "https://api.github.com/users/severo/foll... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | open | false | null | [] | null | 2 | 2023-11-22T15:18:10 | 2023-11-23T09:06:20 | null | CONTRIBUTOR | null | Look at https://huggingface.co/datasets/severo/doc-yaml-2
It's a reproduction of the example given in the docs at https://huggingface.co/docs/hub/datasets-manual-configuration
```
You can select multiple files per split using a list of paths:
my_dataset_repository/
├── README.md
├── data/
│ ├── abc.csv
... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6443/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6443/timeline | null | null | null | null | false | [
"There is a typo in one of the file names - `data/edf.csv` should be renamed to `data/def.csv` 🙂. ",
"wow, I reviewed it twice to avoid being ashamed like that, but... I didn't notice the typo.\r\n\r\n---\r\n\r\nBesides this: do you think we would be able to improve the error message to make this clearer?"
] |
https://api.github.com/repos/huggingface/datasets/issues/6442 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6442/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6442/comments | https://api.github.com/repos/huggingface/datasets/issues/6442/events | https://github.com/huggingface/datasets/issues/6442 | 2,006,086,907 | I_kwDODunzps53knT7 | 6,442 | Trouble loading image folder with additional features - metadata file ignored | {
"login": "linoytsaban",
"id": 57615435,
"node_id": "MDQ6VXNlcjU3NjE1NDM1",
"avatar_url": "https://avatars.githubusercontent.com/u/57615435?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/linoytsaban",
"html_url": "https://github.com/linoytsaban",
"followers_url": "https://api.github.com/... | [] | closed | false | null | [] | null | 1 | 2023-11-22T11:01:35 | 2023-11-24T17:13:03 | 2023-11-24T17:13:03 | NONE | null | ### Describe the bug
Loading image folder with a caption column using `load_dataset(<image_folder_path>)` doesn't load the captions.
When loading a local image folder with captions using `datasets==2.13.0`
```
from datasets import load_dataset
data = load_dataset(<image_folder_path>)
data.column_names
```
... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6442/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6442/timeline | null | completed | null | null | false | [
"I reproduced too:\r\n- root: metadata file is ignored (https://huggingface.co/datasets/severo/doc-image-3)\r\n- data/ dir: metadata file is ignored (https://huggingface.co/datasets/severo/doc-image-4)\r\n- train/ dir: works (https://huggingface.co/datasets/severo/doc-image-5)"
] |
https://api.github.com/repos/huggingface/datasets/issues/6441 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6441/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6441/comments | https://api.github.com/repos/huggingface/datasets/issues/6441/events | https://github.com/huggingface/datasets/issues/6441 | 2,004,985,857 | I_kwDODunzps53gagB | 6,441 | Trouble Loading a Gated Dataset For User with Granted Permission | {
"login": "e-trop",
"id": 124715309,
"node_id": "U_kgDOB28BLQ",
"avatar_url": "https://avatars.githubusercontent.com/u/124715309?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/e-trop",
"html_url": "https://github.com/e-trop",
"followers_url": "https://api.github.com/users/e-trop/follower... | [] | closed | false | null | [] | null | 3 | 2023-11-21T19:24:36 | 2023-12-13T08:27:16 | 2023-12-13T08:27:16 | NONE | null | ### Describe the bug
I have granted permissions to several users to access a gated huggingface dataset. The users accepted the invite and when trying to load the dataset using their access token they get
`FileNotFoundError: Couldn't find a dataset script at .....` . Also when they try to click the url link for the d... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6441/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6441/timeline | null | completed | null | null | false | [
"> Also when they try to click the url link for the dataset they get a 404 error.\r\n\r\nThis seems to be a Hub error then (cc @SBrandeis)",
"Could you report this to https://discuss.huggingface.co/c/hub/23, providing the URL of the dataset, or at least if the dataset is public or private?",
"Thanks for the rep... |
https://api.github.com/repos/huggingface/datasets/issues/6440 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6440/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6440/comments | https://api.github.com/repos/huggingface/datasets/issues/6440/events | https://github.com/huggingface/datasets/issues/6440 | 2,004,509,301 | I_kwDODunzps53emJ1 | 6,440 | `.map` not hashing under python 3.9 | {
"login": "changyeli",
"id": 9058204,
"node_id": "MDQ6VXNlcjkwNTgyMDQ=",
"avatar_url": "https://avatars.githubusercontent.com/u/9058204?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/changyeli",
"html_url": "https://github.com/changyeli",
"followers_url": "https://api.github.com/users/ch... | [] | closed | false | null | [] | null | 2 | 2023-11-21T15:14:54 | 2023-11-28T16:29:33 | 2023-11-28T16:29:33 | NONE | null | ### Describe the bug
The `.map` function cannot hash under python 3.9. Tried to use [the solution here](https://github.com/huggingface/datasets/issues/4521#issuecomment-1205166653), but still get the same message:
`Parameter 'function'=<function map_to_pred at 0x7fa0b49ead30> of the transform datasets.arrow_data... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6440/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6440/timeline | null | completed | null | null | false | [
"Tried to upgrade Python to 3.11 - still get this message. A partial solution is to NOT use `num_proc` at all. It will be considerably longer to finish the job.",
"Hi! The `model = torch.compile(model)` line is problematic for our hashing logic. We would have to merge https://github.com/huggingface/datasets/pull/... |
https://api.github.com/repos/huggingface/datasets/issues/6439 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6439/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6439/comments | https://api.github.com/repos/huggingface/datasets/issues/6439/events | https://github.com/huggingface/datasets/issues/6439 | 2,002,916,514 | I_kwDODunzps53YhSi | 6,439 | Download + preparation speed of datasets.load_dataset is 20x slower than huggingface hub snapshot and manual loding | {
"login": "AntreasAntoniou",
"id": 10792502,
"node_id": "MDQ6VXNlcjEwNzkyNTAy",
"avatar_url": "https://avatars.githubusercontent.com/u/10792502?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/AntreasAntoniou",
"html_url": "https://github.com/AntreasAntoniou",
"followers_url": "https://api... | [] | open | false | null | [] | null | 0 | 2023-11-20T20:07:23 | 2023-11-20T20:07:37 | null | NONE | null | ### Describe the bug
I am working with a dataset I am trying to publish.
The path is Antreas/TALI.
It's a fairly large dataset, and contains images, video, audio and text.
I have been having multiple problems when the dataset is being downloaded using the load_dataset function -- even with 64 workers takin... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6439/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6439/timeline | null | null | null | null | false | [] |
https://api.github.com/repos/huggingface/datasets/issues/6438 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6438/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6438/comments | https://api.github.com/repos/huggingface/datasets/issues/6438/events | https://github.com/huggingface/datasets/issues/6438 | 2,002,032,804 | I_kwDODunzps53VJik | 6,438 | Support GeoParquet | {
"login": "severo",
"id": 1676121,
"node_id": "MDQ6VXNlcjE2NzYxMjE=",
"avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/severo",
"html_url": "https://github.com/severo",
"followers_url": "https://api.github.com/users/severo/foll... | [
{
"id": 1935892871,
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement",
"name": "enhancement",
"color": "a2eeef",
"default": true,
"description": "New feature or request"
}
] | open | false | null | [] | null | 6 | 2023-11-20T11:54:58 | 2024-02-07T08:36:51 | null | CONTRIBUTOR | null | ### Feature request
Support the GeoParquet format
### Motivation
GeoParquet (https://geoparquet.org/) is a common format for sharing vectorial geospatial data on the cloud, along with "traditional" data columns.
It would be nice to be able to load this format with datasets, and more generally, in the Datasets Hub... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6438/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6438/timeline | null | null | null | null | false | [
"Thank you, @severo ! I would be more than happy to help in any way I can. I am not familiar with this repo's codebase, but I would be eager to contribute. :)\r\n\r\nFor the preview in Datasets Hub, I think it makes sense to just display the geospatial column as text. If there were a dataset loader, though, I think... |
https://api.github.com/repos/huggingface/datasets/issues/6437 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6437/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6437/comments | https://api.github.com/repos/huggingface/datasets/issues/6437/events | https://github.com/huggingface/datasets/issues/6437 | 2,001,272,606 | I_kwDODunzps53SP8e | 6,437 | Problem in training iterable dataset | {
"login": "21Timothy",
"id": 38107672,
"node_id": "MDQ6VXNlcjM4MTA3Njcy",
"avatar_url": "https://avatars.githubusercontent.com/u/38107672?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/21Timothy",
"html_url": "https://github.com/21Timothy",
"followers_url": "https://api.github.com/users/... | [] | open | false | null | [] | null | 2 | 2023-11-20T03:04:02 | 2023-11-29T11:11:15 | null | NONE | null | ### Describe the bug
I am using PyTorch DDP (Distributed Data Parallel) to train my model. Since the data is too large to load into memory at once, I am using load_dataset to read the data as an iterable dataset. I have used datasets.distributed.split_dataset_by_node to distribute the dataset. However, I have notice... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6437/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6437/timeline | null | null | null | null | false | [
"Has anyone ever encountered this problem before?",
"`split_dataset_by_node` doesn't give the exact same number of examples to each node in the case of iterable datasets, though it tries to be as equal as possible. In particular if your dataset is sharded and you have a number of shards that is a factor of the nu... |
https://api.github.com/repos/huggingface/datasets/issues/6436 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6436/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6436/comments | https://api.github.com/repos/huggingface/datasets/issues/6436/events | https://github.com/huggingface/datasets/issues/6436 | 2,000,844,474 | I_kwDODunzps53Qna6 | 6,436 | TypeError: <lambda>() takes 0 positional arguments but 1 was given | {
"login": "ahmadmustafaanis",
"id": 47111429,
"node_id": "MDQ6VXNlcjQ3MTExNDI5",
"avatar_url": "https://avatars.githubusercontent.com/u/47111429?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ahmadmustafaanis",
"html_url": "https://github.com/ahmadmustafaanis",
"followers_url": "https://... | [] | closed | false | null | [] | null | 1 | 2023-11-19T13:10:20 | 2023-11-29T16:28:34 | 2023-11-29T16:28:34 | NONE | null | ### Describe the bug
```
---------------------------------------------------------------------------
TypeError Traceback (most recent call last)
[<ipython-input-35-7b6becee3685>](https://localhost:8080/#) in <cell line: 1>()
----> 1 from datasets import Dataset
9 frames
[/usr/lo... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6436/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6436/timeline | null | completed | null | null | false | [
"This looks like a problem with your environment rather than `datasets`."
] |
https://api.github.com/repos/huggingface/datasets/issues/6435 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6435/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6435/comments | https://api.github.com/repos/huggingface/datasets/issues/6435/events | https://github.com/huggingface/datasets/issues/6435 | 2,000,690,513 | I_kwDODunzps53QB1R | 6,435 | Cannot re-initialize CUDA in forked subprocess. To use CUDA with multiprocessing, you must use the 'spawn' start method | {
"login": "kopyl",
"id": 17604849,
"node_id": "MDQ6VXNlcjE3NjA0ODQ5",
"avatar_url": "https://avatars.githubusercontent.com/u/17604849?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/kopyl",
"html_url": "https://github.com/kopyl",
"followers_url": "https://api.github.com/users/kopyl/follow... | [] | closed | false | null | [] | null | 3 | 2023-11-19T04:21:16 | 2024-01-27T17:14:20 | 2023-12-04T16:57:43 | NONE | null | ### Describe the bug
1. I ran dataset mapping with `num_proc=6` in it and got this error:
`RuntimeError: Cannot re-initialize CUDA in forked subprocess. To use CUDA with multiprocessing, you must use the 'spawn' start method`
I can't actually find a way to run multi-GPU dataset mapping. Can you help?
### Steps to... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6435/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6435/timeline | null | completed | null | null | false | [
"[This doc section](https://huggingface.co/docs/datasets/main/en/process#multiprocessing) explains how to modify the script to avoid this error.",
"@mariosasko thank you very much, i'll check it",
"@mariosasko no it does not\r\n\r\n`Dataset.filter() got an unexpected keyword argument 'with_rank'`"
] |
https://api.github.com/repos/huggingface/datasets/issues/6434 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6434/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6434/comments | https://api.github.com/repos/huggingface/datasets/issues/6434/events | https://github.com/huggingface/datasets/pull/6434 | 1,999,554,915 | PR_kwDODunzps5fxgUO | 6,434 | Use `ruff` for formatting | {
"login": "mariosasko",
"id": 47462742,
"node_id": "MDQ6VXNlcjQ3NDYyNzQy",
"avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mariosasko",
"html_url": "https://github.com/mariosasko",
"followers_url": "https://api.github.com/use... | [] | closed | false | null | [] | null | 3 | 2023-11-17T16:53:22 | 2023-11-21T14:19:21 | 2023-11-21T14:13:13 | CONTRIBUTOR | null | Use `ruff` instead of `black` for formatting to be consistent with `transformers` ([PR](https://github.com/huggingface/transformers/pull/27144)) and `huggingface_hub` ([PR 1](https://github.com/huggingface/huggingface_hub/pull/1783) and [PR 2](https://github.com/huggingface/huggingface_hub/pull/1789)). | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6434/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6434/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6434",
"html_url": "https://github.com/huggingface/datasets/pull/6434",
"diff_url": "https://github.com/huggingface/datasets/pull/6434.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6434.patch",
"merged_at": "2023-11-21T14:13... | true | [
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a... |
https://api.github.com/repos/huggingface/datasets/issues/6433 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6433/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6433/comments | https://api.github.com/repos/huggingface/datasets/issues/6433/events | https://github.com/huggingface/datasets/pull/6433 | 1,999,419,105 | PR_kwDODunzps5fxDoG | 6,433 | Better `tqdm` wrapper | {
"login": "mariosasko",
"id": 47462742,
"node_id": "MDQ6VXNlcjQ3NDYyNzQy",
"avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mariosasko",
"html_url": "https://github.com/mariosasko",
"followers_url": "https://api.github.com/use... | [] | closed | false | null | [] | null | 9 | 2023-11-17T15:45:15 | 2023-11-22T16:48:18 | 2023-11-22T16:42:08 | CONTRIBUTOR | null | This PR aligns the `tqdm` logic with `huggingface_hub` (without introducing breaking changes), as the current one is error-prone.
Additionally, it improves the doc page about the `datasets`' utilities, and the handling of local `fsspec` paths in `cached_path`.
Fix #6409 | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6433/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6433/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6433",
"html_url": "https://github.com/huggingface/datasets/pull/6433",
"diff_url": "https://github.com/huggingface/datasets/pull/6433.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6433.patch",
"merged_at": "2023-11-22T16:42... | true | [
"_The documentation is not available anymore as the PR was closed or merged._",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea... |
https://api.github.com/repos/huggingface/datasets/issues/6432 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6432/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6432/comments | https://api.github.com/repos/huggingface/datasets/issues/6432/events | https://github.com/huggingface/datasets/issues/6432 | 1,999,258,140 | I_kwDODunzps53KkIc | 6,432 | load_dataset does not load all of the data in my input file | {
"login": "demongolem-biz2",
"id": 121301001,
"node_id": "U_kgDOBzroCQ",
"avatar_url": "https://avatars.githubusercontent.com/u/121301001?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/demongolem-biz2",
"html_url": "https://github.com/demongolem-biz2",
"followers_url": "https://api.githu... | [] | open | false | null | [] | null | 1 | 2023-11-17T14:28:50 | 2023-11-22T17:34:58 | null | NONE | null | ### Describe the bug
I have 127 elements in my input dataset. When I do a len on the dataset after loaded, it is only 124 elements.
### Steps to reproduce the bug
train_dataset = nlp.load_dataset(data_args.dataset_path, name=data_args.qg_format, split=nlp.Split.TRAIN)
valid_dataset = nlp.load_dataset(data_... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6432/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6432/timeline | null | null | null | null | false | [
"You should use `datasets.load_dataset` instead of `nlp.load_dataset`, as the `nlp` package is outdated.\r\n\r\nIf switching to `datasets.load_dataset` doesn't fix the issue, sharing the JSON file (feel free to replace the data with dummy data) would be nice so that we can reproduce it ourselves."
] |
https://api.github.com/repos/huggingface/datasets/issues/6431 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6431/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6431/comments | https://api.github.com/repos/huggingface/datasets/issues/6431/events | https://github.com/huggingface/datasets/pull/6431 | 1,997,202,770 | PR_kwDODunzps5fpfos | 6,431 | Create DatasetNotFoundError and DataFilesNotFoundError | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | 10 | 2023-11-16T16:02:55 | 2023-11-22T15:18:51 | 2023-11-22T15:12:33 | MEMBER | null | Create `DatasetNotFoundError` and `DataFilesNotFoundError`.
Fix #6397.
CC: @severo | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6431/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6431/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6431",
"html_url": "https://github.com/huggingface/datasets/pull/6431",
"diff_url": "https://github.com/huggingface/datasets/pull/6431.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6431.patch",
"merged_at": "2023-11-22T15:12... | true | [
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a... |
https://api.github.com/repos/huggingface/datasets/issues/6429 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6429/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6429/comments | https://api.github.com/repos/huggingface/datasets/issues/6429/events | https://github.com/huggingface/datasets/pull/6429 | 1,996,723,698 | PR_kwDODunzps5fn1r_ | 6,429 | Add trust_remote_code argument | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [] | closed | false | null | [] | null | 14 | 2023-11-16T12:12:54 | 2023-11-28T16:10:39 | 2023-11-28T16:03:43 | MEMBER | null | Draft about adding `trust_remote_code` to `load_dataset`.
```python
ds = load_dataset(..., trust_remote_code=True) # run remote code (current default)
```
It would default to `True` (current behavior) and in the next major release it will prompt the user to check the code before running it (we'll communicate o... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6429/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6429/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6429",
"html_url": "https://github.com/huggingface/datasets/pull/6429",
"diff_url": "https://github.com/huggingface/datasets/pull/6429.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6429.patch",
"merged_at": "2023-11-28T16:03... | true | [
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a... |
https://api.github.com/repos/huggingface/datasets/issues/6428 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6428/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6428/comments | https://api.github.com/repos/huggingface/datasets/issues/6428/events | https://github.com/huggingface/datasets/pull/6428 | 1,996,306,394 | PR_kwDODunzps5fmakS | 6,428 | Set dev version | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | 3 | 2023-11-16T08:12:55 | 2023-11-16T08:19:39 | 2023-11-16T08:13:28 | MEMBER | null | null | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6428/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6428/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6428",
"html_url": "https://github.com/huggingface/datasets/pull/6428",
"diff_url": "https://github.com/huggingface/datasets/pull/6428.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6428.patch",
"merged_at": "2023-11-16T08:13... | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6428). All of your documentation changes will be reflected on that endpoint.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchma... |
https://api.github.com/repos/huggingface/datasets/issues/6427 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6427/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6427/comments | https://api.github.com/repos/huggingface/datasets/issues/6427/events | https://github.com/huggingface/datasets/pull/6427 | 1,996,248,605 | PR_kwDODunzps5fmN1_ | 6,427 | Release: 2.15.0 | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | 4 | 2023-11-16T07:37:20 | 2023-11-16T08:12:12 | 2023-11-16T07:43:05 | MEMBER | null | null | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6427/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6427/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6427",
"html_url": "https://github.com/huggingface/datasets/pull/6427",
"diff_url": "https://github.com/huggingface/datasets/pull/6427.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6427.patch",
"merged_at": "2023-11-16T07:43... | true | [
"_The documentation is not available anymore as the PR was closed or merged._",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea... |
https://api.github.com/repos/huggingface/datasets/issues/6426 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6426/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6426/comments | https://api.github.com/repos/huggingface/datasets/issues/6426/events | https://github.com/huggingface/datasets/pull/6426 | 1,995,363,264 | PR_kwDODunzps5fjOEK | 6,426 | More robust temporary directory deletion | {
"login": "mariosasko",
"id": 47462742,
"node_id": "MDQ6VXNlcjQ3NDYyNzQy",
"avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mariosasko",
"html_url": "https://github.com/mariosasko",
"followers_url": "https://api.github.com/use... | [] | closed | false | null | [] | null | 7 | 2023-11-15T19:06:42 | 2023-12-01T15:37:32 | 2023-12-01T15:31:19 | CONTRIBUTOR | null | While fixing the Windows errors in #6362, I noticed that `PermissionError` can still easily be thrown on the session exit by the temporary cache directory's finalizer (we would also have to keep track of intermediate datasets, copies, etc.). ~~Due to the low usage of `datasets` on Windows, this PR takes a simpler appro... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6426/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6426/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6426",
"html_url": "https://github.com/huggingface/datasets/pull/6426",
"diff_url": "https://github.com/huggingface/datasets/pull/6426.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6426.patch",
"merged_at": "2023-12-01T15:31... | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6426). All of your documentation changes will be reflected on that endpoint.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchma... |
https://api.github.com/repos/huggingface/datasets/issues/6425 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6425/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6425/comments | https://api.github.com/repos/huggingface/datasets/issues/6425/events | https://github.com/huggingface/datasets/pull/6425 | 1,995,269,382 | PR_kwDODunzps5fi5ye | 6,425 | Fix deprecation warning when building conda package | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | 3 | 2023-11-15T18:00:11 | 2023-12-13T14:22:30 | 2023-12-13T14:16:00 | MEMBER | null | When building/releasing conda package, we get this deprecation warning:
```
/usr/share/miniconda/envs/build-datasets/bin/conda-build:11: DeprecationWarning: conda_build.cli.main_build.main is deprecated and will be removed in 4.0.0. Use `conda build` instead.
```
This PR fixes the deprecation warning by using `co... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6425/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6425/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6425",
"html_url": "https://github.com/huggingface/datasets/pull/6425",
"diff_url": "https://github.com/huggingface/datasets/pull/6425.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6425.patch",
"merged_at": "2023-12-13T14:16... | true | [
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a... |
https://api.github.com/repos/huggingface/datasets/issues/6424 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6424/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6424/comments | https://api.github.com/repos/huggingface/datasets/issues/6424/events | https://github.com/huggingface/datasets/pull/6424 | 1,995,224,516 | PR_kwDODunzps5fiwDC | 6,424 | [docs] troubleshooting guide | {
"login": "MKhalusova",
"id": 1065417,
"node_id": "MDQ6VXNlcjEwNjU0MTc=",
"avatar_url": "https://avatars.githubusercontent.com/u/1065417?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/MKhalusova",
"html_url": "https://github.com/MKhalusova",
"followers_url": "https://api.github.com/users... | [] | closed | false | null | [] | null | 2 | 2023-11-15T17:28:14 | 2023-11-30T17:29:55 | 2023-11-30T17:23:46 | CONTRIBUTOR | null | Hi all! This is a PR adding a troubleshooting guide for Datasets docs.
I went through the library's GitHub Issues and Forum questions and identified a few issues that are common enough that I think it would be valuable to include them in the troubleshooting guide. These are:
- creating a dataset from a folder and n... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6424/reactions",
"total_count": 1,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 1,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6424/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6424",
"html_url": "https://github.com/huggingface/datasets/pull/6424",
"diff_url": "https://github.com/huggingface/datasets/pull/6424.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6424.patch",
"merged_at": "2023-11-30T17:23... | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6424). All of your documentation changes will be reflected on that endpoint.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchma... |
https://api.github.com/repos/huggingface/datasets/issues/6423 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6423/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6423/comments | https://api.github.com/repos/huggingface/datasets/issues/6423/events | https://github.com/huggingface/datasets/pull/6423 | 1,994,946,847 | PR_kwDODunzps5fhzD6 | 6,423 | Fix conda release by adding pyarrow-hotfix dependency | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | 6 | 2023-11-15T14:57:12 | 2023-11-15T17:15:33 | 2023-11-15T17:09:24 | MEMBER | null | Fix conda release by adding pyarrow-hotfix dependency.
Note that conda release failed in latest 2.14.7 release: https://github.com/huggingface/datasets/actions/runs/6874667214/job/18696761723
```
Traceback (most recent call last):
File "/usr/share/miniconda/envs/build-datasets/conda-bld/datasets_1700036460222/t... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6423/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6423/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6423",
"html_url": "https://github.com/huggingface/datasets/pull/6423",
"diff_url": "https://github.com/huggingface/datasets/pull/6423.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6423.patch",
"merged_at": "2023-11-15T17:09... | true | [
"_The documentation is not available anymore as the PR was closed or merged._",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea... |
https://api.github.com/repos/huggingface/datasets/issues/6422 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6422/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6422/comments | https://api.github.com/repos/huggingface/datasets/issues/6422/events | https://github.com/huggingface/datasets/issues/6422 | 1,994,579,267 | I_kwDODunzps524t1D | 6,422 | Allow to choose the `writer_batch_size` when using `save_to_disk` | {
"login": "NathanGodey",
"id": 38216711,
"node_id": "MDQ6VXNlcjM4MjE2NzEx",
"avatar_url": "https://avatars.githubusercontent.com/u/38216711?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/NathanGodey",
"html_url": "https://github.com/NathanGodey",
"followers_url": "https://api.github.com/... | [
{
"id": 1935892871,
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement",
"name": "enhancement",
"color": "a2eeef",
"default": true,
"description": "New feature or request"
}
] | open | false | null | [] | null | 2 | 2023-11-15T11:18:34 | 2023-11-16T10:00:21 | null | NONE | null | ### Feature request
Add an argument in `save_to_disk` regarding batch size, which would be passed to `shard` and other methods.
### Motivation
The `Dataset.save_to_disk` method currently calls `shard` without passing a `writer_batch_size` argument, thus implicitly using the default value (1000). This can result in R... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6422/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6422/timeline | null | null | null | null | false | [
"We have a config variable that controls the batch size in `save_to_disk`:\r\n```python\r\nimport datasets\r\ndatasets.config.DEFAULT_MAX_BATCH_SIZE = <smaller_batch_size>\r\n...\r\nds.save_to_disk(...)\r\n```",
"Thank you for your answer!\r\n\r\nFrom what I am reading in `https://github.com/huggingface/datasets/... |
https://api.github.com/repos/huggingface/datasets/issues/6421 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6421/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6421/comments | https://api.github.com/repos/huggingface/datasets/issues/6421/events | https://github.com/huggingface/datasets/pull/6421 | 1,994,451,553 | PR_kwDODunzps5fgG1h | 6,421 | Add pyarrow-hotfix to release docs | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"id": 4296013012,
"node_id": "LA_kwDODunzps8AAAABAA_01A",
"url": "https://api.github.com/repos/huggingface/datasets/labels/maintenance",
"name": "maintenance",
"color": "d4c5f9",
"default": false,
"description": "Maintenance tasks"
}
] | closed | false | null | [] | null | 3 | 2023-11-15T10:06:44 | 2023-11-15T13:49:55 | 2023-11-15T13:38:22 | MEMBER | null | Add `pyarrow-hotfix` to release docs. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6421/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6421/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6421",
"html_url": "https://github.com/huggingface/datasets/pull/6421",
"diff_url": "https://github.com/huggingface/datasets/pull/6421.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6421.patch",
"merged_at": "2023-11-15T13:38... | true | [
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a... |
https://api.github.com/repos/huggingface/datasets/issues/6420 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6420/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6420/comments | https://api.github.com/repos/huggingface/datasets/issues/6420/events | https://github.com/huggingface/datasets/pull/6420 | 1,994,278,903 | PR_kwDODunzps5ffhdi | 6,420 | Set dev version | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | 3 | 2023-11-15T08:22:19 | 2023-11-15T08:33:36 | 2023-11-15T08:22:33 | MEMBER | null | null | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6420/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6420/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6420",
"html_url": "https://github.com/huggingface/datasets/pull/6420",
"diff_url": "https://github.com/huggingface/datasets/pull/6420.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6420.patch",
"merged_at": "2023-11-15T08:22... | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6420). All of your documentation changes will be reflected on that endpoint.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchma... |
https://api.github.com/repos/huggingface/datasets/issues/6419 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6419/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6419/comments | https://api.github.com/repos/huggingface/datasets/issues/6419/events | https://github.com/huggingface/datasets/pull/6419 | 1,994,257,873 | PR_kwDODunzps5ffc7d | 6,419 | Release: 2.14.7 | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | 6 | 2023-11-15T08:07:37 | 2023-11-15T17:35:30 | 2023-11-15T08:12:59 | MEMBER | null | Release 2.14.7. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6419/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6419/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6419",
"html_url": "https://github.com/huggingface/datasets/pull/6419",
"diff_url": "https://github.com/huggingface/datasets/pull/6419.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6419.patch",
"merged_at": "2023-11-15T08:12... | true | [
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a... |
https://api.github.com/repos/huggingface/datasets/issues/6418 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6418/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6418/comments | https://api.github.com/repos/huggingface/datasets/issues/6418/events | https://github.com/huggingface/datasets/pull/6418 | 1,993,224,629 | PR_kwDODunzps5fb7lu | 6,418 | Remove token value from warnings | {
"login": "mariosasko",
"id": 47462742,
"node_id": "MDQ6VXNlcjQ3NDYyNzQy",
"avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mariosasko",
"html_url": "https://github.com/mariosasko",
"followers_url": "https://api.github.com/use... | [] | closed | false | null | [] | null | 3 | 2023-11-14T17:34:06 | 2023-11-14T22:26:04 | 2023-11-14T22:19:45 | CONTRIBUTOR | null | Fix #6412 | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6418/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6418/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6418",
"html_url": "https://github.com/huggingface/datasets/pull/6418",
"diff_url": "https://github.com/huggingface/datasets/pull/6418.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6418.patch",
"merged_at": "2023-11-14T22:19... | true | [
"_The documentation is not available anymore as the PR was closed or merged._",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea... |
https://api.github.com/repos/huggingface/datasets/issues/6417 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6417/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6417/comments | https://api.github.com/repos/huggingface/datasets/issues/6417/events | https://github.com/huggingface/datasets/issues/6417 | 1,993,149,416 | I_kwDODunzps52zQvo | 6,417 | Bug: LayoutLMv3 finetuning on FUNSD Notebook; Arrow Error | {
"login": "Davo00",
"id": 57496007,
"node_id": "MDQ6VXNlcjU3NDk2MDA3",
"avatar_url": "https://avatars.githubusercontent.com/u/57496007?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Davo00",
"html_url": "https://github.com/Davo00",
"followers_url": "https://api.github.com/users/Davo00/fo... | [] | closed | false | null | [] | null | 3 | 2023-11-14T16:53:20 | 2023-11-16T20:23:41 | 2023-11-16T20:23:41 | NONE | null | ### Describe the bug
Arrow issues when running the example Notebook laptop locally on Mac with M1. Works on Google Collab.
**Notebook**: https://github.com/NielsRogge/Transformers-Tutorials/blob/master/LayoutLMv3/Fine_tune_LayoutLMv3_on_FUNSD_(HuggingFace_Trainer).ipynb
**Error**: `ValueError: Arrow type extensi... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6417/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6417/timeline | null | completed | null | null | false | [
"Very strange: `datasets-cli env`\r\n> \r\n> Copy-and-paste the text below in your GitHub issue.\r\n> \r\n> - `datasets` version: 2.9.0\r\n> - Platform: macOS-14.0-arm64-arm-64bit\r\n> - Python version: 3.9.13\r\n> - PyArrow version: 8.0.0\r\n> - Pandas version: 1.3.5\r\n\r\nAfter updating datasets and pyarrow on b... |
https://api.github.com/repos/huggingface/datasets/issues/6416 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6416/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6416/comments | https://api.github.com/repos/huggingface/datasets/issues/6416/events | https://github.com/huggingface/datasets/pull/6416 | 1,992,954,723 | PR_kwDODunzps5fbA4H | 6,416 | Rename audio_classificiation.py to audio_classification.py | {
"login": "carlthome",
"id": 1595907,
"node_id": "MDQ6VXNlcjE1OTU5MDc=",
"avatar_url": "https://avatars.githubusercontent.com/u/1595907?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/carlthome",
"html_url": "https://github.com/carlthome",
"followers_url": "https://api.github.com/users/ca... | [] | closed | false | null | [] | null | 4 | 2023-11-14T15:15:29 | 2023-11-15T11:59:32 | 2023-11-15T11:53:20 | CONTRIBUTOR | null | null | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6416/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6416/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6416",
"html_url": "https://github.com/huggingface/datasets/pull/6416",
"diff_url": "https://github.com/huggingface/datasets/pull/6416.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6416.patch",
"merged_at": "2023-11-15T11:53... | true | [
"Oh good catch. Can you also rename it in `src/datasets/tasks/__init__.py` ?",
"Fixed! \r\n\r\n(I think, tough word to spell right TBH)",
"_The documentation is not available anymore as the PR was closed or merged._",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show... |
https://api.github.com/repos/huggingface/datasets/issues/6415 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6415/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6415/comments | https://api.github.com/repos/huggingface/datasets/issues/6415/events | https://github.com/huggingface/datasets/pull/6415 | 1,992,917,248 | PR_kwDODunzps5fa4n7 | 6,415 | Fix multi gpu map example | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [] | closed | false | null | [] | null | 23 | 2023-11-14T14:57:18 | 2024-01-31T00:49:15 | 2023-11-22T15:42:19 | MEMBER | null | - use `orch.cuda.set_device` instead of `CUDA_VISIBLE_DEVICES `
- add `if __name__ == "__main__"`
fix https://github.com/huggingface/datasets/issues/6186 | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6415/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6415/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6415",
"html_url": "https://github.com/huggingface/datasets/pull/6415",
"diff_url": "https://github.com/huggingface/datasets/pull/6415.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6415.patch",
"merged_at": "2023-11-22T15:42... | true | [
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a... |
https://api.github.com/repos/huggingface/datasets/issues/6414 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6414/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6414/comments | https://api.github.com/repos/huggingface/datasets/issues/6414/events | https://github.com/huggingface/datasets/pull/6414 | 1,992,482,491 | PR_kwDODunzps5fZZ2l | 6,414 | Set `usedforsecurity=False` in hashlib methods (FIPS compliance) | {
"login": "Wauplin",
"id": 11801849,
"node_id": "MDQ6VXNlcjExODAxODQ5",
"avatar_url": "https://avatars.githubusercontent.com/u/11801849?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Wauplin",
"html_url": "https://github.com/Wauplin",
"followers_url": "https://api.github.com/users/Waupli... | [] | closed | false | null | [] | null | 10 | 2023-11-14T10:47:09 | 2023-11-17T14:23:20 | 2023-11-17T14:17:00 | CONTRIBUTOR | null | Related to https://github.com/huggingface/transformers/issues/27034 and https://github.com/huggingface/huggingface_hub/pull/1782.
**TL;DR:** `hashlib` is not a secure library for cryptography-related stuff. We are only using `hashlib` for non-security-related purposes in `datasets` so it's fine. From Python 3.9 we s... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6414/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6414/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6414",
"html_url": "https://github.com/huggingface/datasets/pull/6414",
"diff_url": "https://github.com/huggingface/datasets/pull/6414.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6414.patch",
"merged_at": "2023-11-17T14:17... | true | [
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a... |
https://api.github.com/repos/huggingface/datasets/issues/6412 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6412/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6412/comments | https://api.github.com/repos/huggingface/datasets/issues/6412/events | https://github.com/huggingface/datasets/issues/6412 | 1,992,401,594 | I_kwDODunzps52waK6 | 6,412 | User token is printed out! | {
"login": "mohsen-goodarzi",
"id": 25702692,
"node_id": "MDQ6VXNlcjI1NzAyNjky",
"avatar_url": "https://avatars.githubusercontent.com/u/25702692?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mohsen-goodarzi",
"html_url": "https://github.com/mohsen-goodarzi",
"followers_url": "https://api... | [] | closed | false | null | [] | null | 1 | 2023-11-14T10:01:34 | 2023-11-14T22:19:46 | 2023-11-14T22:19:46 | NONE | null | This line prints user token on command line! Is it safe?
https://github.com/huggingface/datasets/blob/12ebe695b4748c5a26e08b44ed51955f74f5801d/src/datasets/load.py#L2091 | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6412/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6412/timeline | null | completed | null | null | false | [
"Indeed, this is not a good practice. I've opened a PR that removes the token value from the (deprecation) warning."
] |
https://api.github.com/repos/huggingface/datasets/issues/6411 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6411/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6411/comments | https://api.github.com/repos/huggingface/datasets/issues/6411/events | https://github.com/huggingface/datasets/pull/6411 | 1,992,386,630 | PR_kwDODunzps5fZE9F | 6,411 | Fix dependency conflict within CI build documentation | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | 1 | 2023-11-14T09:52:51 | 2023-11-14T10:05:59 | 2023-11-14T10:05:35 | MEMBER | null | Manually fix dependency conflict on `typing-extensions` version originated by `apache-beam` + `pydantic` (now a dependency of `huggingface-hub`).
This is a temporary hot fix of our CI build documentation until we stop using `apache-beam`.
Fix #6406. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6411/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6411/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6411",
"html_url": "https://github.com/huggingface/datasets/pull/6411",
"diff_url": "https://github.com/huggingface/datasets/pull/6411.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6411.patch",
"merged_at": "2023-11-14T10:05... | true | [
"_The documentation is not available anymore as the PR was closed or merged._"
] |
https://api.github.com/repos/huggingface/datasets/issues/6410 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6410/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6410/comments | https://api.github.com/repos/huggingface/datasets/issues/6410/events | https://github.com/huggingface/datasets/issues/6410 | 1,992,100,209 | I_kwDODunzps52vQlx | 6,410 | Datasets does not load HuggingFace Repository properly | {
"login": "MikeDoes",
"id": 40600201,
"node_id": "MDQ6VXNlcjQwNjAwMjAx",
"avatar_url": "https://avatars.githubusercontent.com/u/40600201?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/MikeDoes",
"html_url": "https://github.com/MikeDoes",
"followers_url": "https://api.github.com/users/Mik... | [] | open | false | null | [] | null | 2 | 2023-11-14T06:50:49 | 2023-11-16T06:54:36 | null | NONE | null | ### Describe the bug
Dear Datasets team,
We just have published a dataset on Huggingface:
https://huggingface.co/ai4privacy
However, when trying to read it using the Dataset library we get an error. As I understand jsonl files are compatible, could you please clarify how we can solve the issue? Please let me ... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6410/reactions",
"total_count": 2,
"+1": 2,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6410/timeline | null | null | null | null | false | [
"Hi! You can avoid the error by requesting only the `jsonl` files. `dataset = load_dataset(\"ai4privacy/pii-masking-200k\", data_files=[\"*.jsonl\"])`.\r\n\r\nOur data file inference does not filter out (incompatible) `json` files because `json` and `jsonl` use the same builder. Still, I think the inference should... |
https://api.github.com/repos/huggingface/datasets/issues/6409 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6409/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6409/comments | https://api.github.com/repos/huggingface/datasets/issues/6409/events | https://github.com/huggingface/datasets/issues/6409 | 1,991,960,865 | I_kwDODunzps52uukh | 6,409 | using DownloadManager to download from local filesystem and disable_progress_bar, there will be an exception | {
"login": "neiblegy",
"id": 16574677,
"node_id": "MDQ6VXNlcjE2NTc0Njc3",
"avatar_url": "https://avatars.githubusercontent.com/u/16574677?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/neiblegy",
"html_url": "https://github.com/neiblegy",
"followers_url": "https://api.github.com/users/nei... | [] | closed | false | null | [] | null | 0 | 2023-11-14T04:21:01 | 2023-11-22T16:42:09 | 2023-11-22T16:42:09 | NONE | null | ### Describe the bug
i'm using datasets.download.download_manager.DownloadManager to download files like "file:///a/b/c.txt", and i disable_progress_bar() to disable bar. there will be an exception as follows:
`AttributeError: 'function' object has no attribute 'close'
Exception ignored in: <function TqdmCallback.... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6409/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6409/timeline | null | completed | null | null | false | [] |
https://api.github.com/repos/huggingface/datasets/issues/6408 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6408/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6408/comments | https://api.github.com/repos/huggingface/datasets/issues/6408/events | https://github.com/huggingface/datasets/issues/6408 | 1,991,902,972 | I_kwDODunzps52ugb8 | 6,408 | `IterableDataset` lost but not keep columns when map function adding columns with names in `remove_columns` | {
"login": "shmily326",
"id": 24571857,
"node_id": "MDQ6VXNlcjI0NTcxODU3",
"avatar_url": "https://avatars.githubusercontent.com/u/24571857?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/shmily326",
"html_url": "https://github.com/shmily326",
"followers_url": "https://api.github.com/users/... | [] | open | false | null | [] | null | 0 | 2023-11-14T03:12:08 | 2023-11-16T06:24:10 | null | NONE | null | ### Describe the bug
IterableDataset lost but not keep columns when map function adding columns with names in remove_columns,
Dataset not.
May be related to the code below:
https://github.com/huggingface/datasets/blob/06c3ffb8d068b6307b247164b10f7c7311cefed4/src/datasets/iterable_dataset.py#L750-L756
### Steps t... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6408/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6408/timeline | null | null | null | null | false | [] |
https://api.github.com/repos/huggingface/datasets/issues/6407 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6407/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6407/comments | https://api.github.com/repos/huggingface/datasets/issues/6407/events | https://github.com/huggingface/datasets/issues/6407 | 1,991,514,079 | I_kwDODunzps52tBff | 6,407 | Loading the dataset from private S3 bucket gives "TypeError: cannot pickle '_contextvars.Context' object" | {
"login": "eawer",
"id": 1741779,
"node_id": "MDQ6VXNlcjE3NDE3Nzk=",
"avatar_url": "https://avatars.githubusercontent.com/u/1741779?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/eawer",
"html_url": "https://github.com/eawer",
"followers_url": "https://api.github.com/users/eawer/follower... | [] | open | false | null | [] | null | 0 | 2023-11-13T21:27:43 | 2023-11-13T21:27:43 | null | NONE | null | ### Describe the bug
I'm trying to read the parquet file from the private s3 bucket using the `load_dataset` function, but I receive `TypeError: cannot pickle '_contextvars.Context' object` error
I'm working on a machine with `~/.aws/credentials` file. I can't give credentials and the path to a file in a private bu... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6407/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6407/timeline | null | null | null | null | false | [] |
https://api.github.com/repos/huggingface/datasets/issues/6406 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6406/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6406/comments | https://api.github.com/repos/huggingface/datasets/issues/6406/events | https://github.com/huggingface/datasets/issues/6406 | 1,990,469,045 | I_kwDODunzps52pCW1 | 6,406 | CI Build PR Documentation is broken: ImportError: cannot import name 'TypeAliasType' from 'typing_extensions' | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | 0 | 2023-11-13T11:36:10 | 2023-11-14T10:05:36 | 2023-11-14T10:05:36 | MEMBER | null | Our CI Build PR Documentation is broken. See: https://github.com/huggingface/datasets/actions/runs/6799554060/job/18486828777?pr=6390
```
ImportError: cannot import name 'TypeAliasType' from 'typing_extensions'
``` | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6406/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6406/timeline | null | completed | null | null | false | [] |
https://api.github.com/repos/huggingface/datasets/issues/6405 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6405/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6405/comments | https://api.github.com/repos/huggingface/datasets/issues/6405/events | https://github.com/huggingface/datasets/issues/6405 | 1,990,358,743 | I_kwDODunzps52onbX | 6,405 | ConfigNamesError on a simple CSV file | {
"login": "severo",
"id": 1676121,
"node_id": "MDQ6VXNlcjE2NzYxMjE=",
"avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/severo",
"html_url": "https://github.com/severo",
"followers_url": "https://api.github.com/users/severo/foll... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | 3 | 2023-11-13T10:28:29 | 2023-11-13T20:01:24 | 2023-11-13T20:01:24 | CONTRIBUTOR | null | See https://huggingface.co/datasets/Nguyendo1999/mmath/discussions/1
```
Error code: ConfigNamesError
Exception: TypeError
Message: __init__() missing 1 required positional argument: 'dtype'
Traceback: Traceback (most recent call last):
File "/src/services/worker/src/worker/job_runn... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6405/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6405/timeline | null | completed | null | null | false | [
"The viewer is working now. \r\n\r\nBased on the repo commit history, the bug was due to the incorrect format of the `features` field in the README YAML (`Value` requires `dtype`, e.g., `Value(\"string\")`, but it was not specified)",
"Feel free to close the issue",
"Oh, OK! Thanks. So, there was no reason to o... |
https://api.github.com/repos/huggingface/datasets/issues/6404 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6404/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6404/comments | https://api.github.com/repos/huggingface/datasets/issues/6404/events | https://github.com/huggingface/datasets/pull/6404 | 1,990,211,901 | PR_kwDODunzps5fRrJ- | 6,404 | Support pyarrow 14.0.1 and fix vulnerability CVE-2023-47248 | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | 15 | 2023-11-13T09:15:39 | 2023-11-14T10:29:48 | 2023-11-14T10:23:29 | MEMBER | null | Support `pyarrow` 14.0.1 and fix vulnerability [CVE-2023-47248](https://github.com/advisories/GHSA-5wvp-7f3h-6wmm).
Fix #6396. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6404/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6404/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6404",
"html_url": "https://github.com/huggingface/datasets/pull/6404",
"diff_url": "https://github.com/huggingface/datasets/pull/6404.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6404.patch",
"merged_at": "2023-11-14T10:23... | true | [
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a... |
https://api.github.com/repos/huggingface/datasets/issues/6403 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6403/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6403/comments | https://api.github.com/repos/huggingface/datasets/issues/6403/events | https://github.com/huggingface/datasets/issues/6403 | 1,990,098,817 | I_kwDODunzps52nn-B | 6,403 | Cannot import datasets on google colab (python 3.10.12) | {
"login": "nabilaannisa",
"id": 15389235,
"node_id": "MDQ6VXNlcjE1Mzg5MjM1",
"avatar_url": "https://avatars.githubusercontent.com/u/15389235?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/nabilaannisa",
"html_url": "https://github.com/nabilaannisa",
"followers_url": "https://api.github.c... | [] | closed | false | null | [] | null | 2 | 2023-11-13T08:14:43 | 2023-11-16T05:04:22 | 2023-11-16T05:04:21 | NONE | null | ### Describe the bug
I'm trying A full colab demo notebook of zero-shot-distillation from https://github.com/huggingface/transformers/tree/main/examples/research_projects/zero-shot-distillation but i got this type of error when importing datasets on my google colab (python version is 3.10.12)
 instead of (H, W, C). See #6394 for motivation. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6402/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6402/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6402",
"html_url": "https://github.com/huggingface/datasets/pull/6402",
"diff_url": "https://github.com/huggingface/datasets/pull/6402.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6402.patch",
"merged_at": "2024-03-15T11:25... | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6402). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>... |
https://api.github.com/repos/huggingface/datasets/issues/6401 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6401/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6401/comments | https://api.github.com/repos/huggingface/datasets/issues/6401/events | https://github.com/huggingface/datasets/issues/6401 | 1,988,710,061 | I_kwDODunzps52iU6t | 6,401 | dataset = load_dataset("Hyperspace-Technologies/scp-wiki-text") not working | {
"login": "userbox020",
"id": 47074021,
"node_id": "MDQ6VXNlcjQ3MDc0MDIx",
"avatar_url": "https://avatars.githubusercontent.com/u/47074021?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/userbox020",
"html_url": "https://github.com/userbox020",
"followers_url": "https://api.github.com/use... | [] | closed | false | null | [] | null | 2 | 2023-11-11T04:09:07 | 2023-11-20T17:45:20 | 2023-11-20T17:45:20 | NONE | null | ### Describe the bug
```
(datasets) mruserbox@guru-X99:/media/10TB_HHD/_LLM_DATASETS$ python dataset.py
Downloading readme: 100%|███████████████████████████████████| 360/360 [00:00<00:00, 2.16MB/s]
Downloading data: 100%|█████████████████████████████████| 65.1M/65.1M [00:19<00:00, 3.38MB/s]
Downloading data: 100... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6401/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6401/timeline | null | completed | null | null | false | [
"Seems like it's a problem with the dataset, since in the [README](https://huggingface.co/datasets/Hyperspace-Technologies/scp-wiki-text/blob/main/README.md) the validation is not specified. Try cloning the dataset, removing the README (or validation split), and loading it locally/ ",
"@VarunNSrivastava thanks br... |
https://api.github.com/repos/huggingface/datasets/issues/6400 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6400/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6400/comments | https://api.github.com/repos/huggingface/datasets/issues/6400/events | https://github.com/huggingface/datasets/issues/6400 | 1,988,571,317 | I_kwDODunzps52hzC1 | 6,400 | Safely load datasets by disabling execution of dataset loading script | {
"login": "irenedea",
"id": 14367635,
"node_id": "MDQ6VXNlcjE0MzY3NjM1",
"avatar_url": "https://avatars.githubusercontent.com/u/14367635?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/irenedea",
"html_url": "https://github.com/irenedea",
"followers_url": "https://api.github.com/users/ire... | [
{
"id": 1935892871,
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement",
"name": "enhancement",
"color": "a2eeef",
"default": true,
"description": "New feature or request"
}
] | open | false | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [
{
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.git... | null | 3 | 2023-11-10T23:48:29 | 2024-01-02T18:18:09 | null | NONE | null | ### Feature request
Is there a way to disable execution of dataset loading script using `load_dataset`? This is a security vulnerability that could lead to arbitrary code execution.
Any suggested workarounds are welcome as well.
### Motivation
This is a security vulnerability that could lead to arbitrary code e... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6400/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6400/timeline | null | null | null | null | false | [
"great idea IMO\r\n\r\nthis could be a `trust_remote_code=True` flag like in transformers. We could also default to loading the Parquet conversion rather than executing code (for dataset repos that have both)",
"@julien-c that would be great!",
"We added the `trust_remote_code` argument to `load_dataset()` in `... |
https://api.github.com/repos/huggingface/datasets/issues/6399 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6399/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6399/comments | https://api.github.com/repos/huggingface/datasets/issues/6399/events | https://github.com/huggingface/datasets/issues/6399 | 1,988,368,503 | I_kwDODunzps52hBh3 | 6,399 | TypeError: Cannot convert pyarrow.lib.ChunkedArray to pyarrow.lib.Array | {
"login": "y-hwang",
"id": 76236359,
"node_id": "MDQ6VXNlcjc2MjM2MzU5",
"avatar_url": "https://avatars.githubusercontent.com/u/76236359?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/y-hwang",
"html_url": "https://github.com/y-hwang",
"followers_url": "https://api.github.com/users/y-hwan... | [] | open | false | null | [] | null | 0 | 2023-11-10T20:48:46 | 2023-11-10T20:48:46 | null | NONE | null | ### Describe the bug
Hi, I am preprocessing a large custom dataset with numpy arrays. I am running into this TypeError during writing in a dataset.map() function. I've tried decreasing writer batch size, but this error persists. This error does not occur for smaller datasets.
Thank you!
### Steps to repro... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6399/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6399/timeline | null | null | null | null | false | [] |
https://api.github.com/repos/huggingface/datasets/issues/6398 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6398/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6398/comments | https://api.github.com/repos/huggingface/datasets/issues/6398/events | https://github.com/huggingface/datasets/pull/6398 | 1,987,786,446 | PR_kwDODunzps5fJlP7 | 6,398 | Remove redundant condition in builders | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | 3 | 2023-11-10T14:56:43 | 2023-11-14T10:49:15 | 2023-11-14T10:43:00 | MEMBER | null | Minor refactoring to remove redundant condition. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6398/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6398/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6398",
"html_url": "https://github.com/huggingface/datasets/pull/6398",
"diff_url": "https://github.com/huggingface/datasets/pull/6398.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6398.patch",
"merged_at": "2023-11-14T10:43... | true | [
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a... |
https://api.github.com/repos/huggingface/datasets/issues/6397 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6397/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6397/comments | https://api.github.com/repos/huggingface/datasets/issues/6397/events | https://github.com/huggingface/datasets/issues/6397 | 1,987,622,152 | I_kwDODunzps52eLUI | 6,397 | Raise a different exception for inexisting dataset vs files without known extension | {
"login": "severo",
"id": 1676121,
"node_id": "MDQ6VXNlcjE2NzYxMjE=",
"avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/severo",
"html_url": "https://github.com/severo",
"followers_url": "https://api.github.com/users/severo/foll... | [] | closed | false | null | [] | null | 0 | 2023-11-10T13:22:14 | 2023-11-22T15:12:34 | 2023-11-22T15:12:34 | CONTRIBUTOR | null | See https://github.com/huggingface/datasets-server/issues/2082#issuecomment-1805716557
We have the same error for:
- https://huggingface.co/datasets/severo/a_dataset_that_does_not_exist: a dataset that does not exist
- https://huggingface.co/datasets/severo/test_files_without_extension: a dataset with files withou... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6397/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6397/timeline | null | completed | null | null | false | [] |
https://api.github.com/repos/huggingface/datasets/issues/6396 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6396/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6396/comments | https://api.github.com/repos/huggingface/datasets/issues/6396/events | https://github.com/huggingface/datasets/issues/6396 | 1,987,308,077 | I_kwDODunzps52c-ot | 6,396 | Issue with pyarrow 14.0.1 | {
"login": "severo",
"id": 1676121,
"node_id": "MDQ6VXNlcjE2NzYxMjE=",
"avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/severo",
"html_url": "https://github.com/severo",
"followers_url": "https://api.github.com/users/severo/foll... | [] | closed | false | null | [] | null | 5 | 2023-11-10T10:02:12 | 2023-11-14T10:23:30 | 2023-11-14T10:23:30 | CONTRIBUTOR | null | See https://github.com/huggingface/datasets-server/pull/2089 for reference
```
from datasets import (Array2D, Dataset, Features)
feature_type = Array2D(shape=(2, 2), dtype="float32")
content = [[0.0, 0.0], [0.0, 0.0]]
features = Features({"col": feature_type})
dataset = Dataset.from_dict({"col": [content]}, fea... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6396/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6396/timeline | null | completed | null | null | false | [
"Looks like we should stop using `PyExtensionType` and use `ExtensionType` instead\r\n\r\nsee https://github.com/apache/arrow/commit/f14170976372436ec1d03a724d8d3f3925484ecf",
"https://github.com/huggingface/datasets-server/pull/2089#pullrequestreview-1724449532\r\n\r\n> Yes, I understand now: they have disabled ... |
https://api.github.com/repos/huggingface/datasets/issues/6395 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6395/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6395/comments | https://api.github.com/repos/huggingface/datasets/issues/6395/events | https://github.com/huggingface/datasets/issues/6395 | 1,986,484,124 | I_kwDODunzps52Z1ec | 6,395 | Add ability to set lock type | {
"login": "leoleoasd",
"id": 37735580,
"node_id": "MDQ6VXNlcjM3NzM1NTgw",
"avatar_url": "https://avatars.githubusercontent.com/u/37735580?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/leoleoasd",
"html_url": "https://github.com/leoleoasd",
"followers_url": "https://api.github.com/users/... | [
{
"id": 1935892871,
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement",
"name": "enhancement",
"color": "a2eeef",
"default": true,
"description": "New feature or request"
}
] | closed | false | null | [] | null | 1 | 2023-11-09T22:12:30 | 2023-11-23T18:50:00 | 2023-11-23T18:50:00 | NONE | null | ### Feature request
Allow setting file lock type, maybe from an environment variable
Currently, it only depends on whether fnctl is available:
https://github.com/huggingface/datasets/blob/12ebe695b4748c5a26e08b44ed51955f74f5801d/src/datasets/utils/filelock.py#L463-L470C16
### Motivation
In my environment... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6395/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6395/timeline | null | not_planned | null | null | false | [
"We've replaced our filelock implementation with the `filelock` package, so their repo is the right place to request this feature.\r\n\r\nIn the meantime, the following should work: \r\n```python\r\nimport filelock\r\nfilelock.FileLock = filelock.SoftFileLock\r\n\r\nimport datasets\r\n...\r\n```"
] |
https://api.github.com/repos/huggingface/datasets/issues/6394 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6394/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6394/comments | https://api.github.com/repos/huggingface/datasets/issues/6394/events | https://github.com/huggingface/datasets/issues/6394 | 1,985,947,116 | I_kwDODunzps52XyXs | 6,394 | TorchFormatter images (H, W, C) instead of (C, H, W) format | {
"login": "Modexus",
"id": 37351874,
"node_id": "MDQ6VXNlcjM3MzUxODc0",
"avatar_url": "https://avatars.githubusercontent.com/u/37351874?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Modexus",
"html_url": "https://github.com/Modexus",
"followers_url": "https://api.github.com/users/Modexu... | [] | open | false | null | [] | null | 9 | 2023-11-09T16:02:15 | 2024-03-11T22:29:49 | null | NONE | null | ### Describe the bug
Using .set_format("torch") leads to images having shape (H, W, C), the same as in numpy.
However, pytorch normally uses (C, H, W) format.
Maybe I'm missing something but this makes the format a lot less useful as I then have to permute it anyways.
If not using the format it is possible to ... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6394/reactions",
"total_count": 2,
"+1": 2,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6394/timeline | null | null | null | null | false | [
"Here's a PR for that. https://github.com/huggingface/datasets/pull/6402\r\n\r\nIt's not backward compatible, unfortunately. ",
"Just ran into this working on data lib that's attempting to achieve common interfaces across hf datasets, webdataset, native torch style datasets. The defacto standards for image tensor... |
https://api.github.com/repos/huggingface/datasets/issues/6393 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6393/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6393/comments | https://api.github.com/repos/huggingface/datasets/issues/6393/events | https://github.com/huggingface/datasets/issues/6393 | 1,984,913,259 | I_kwDODunzps52T19r | 6,393 | Filter occasionally hangs | {
"login": "dakinggg",
"id": 43149077,
"node_id": "MDQ6VXNlcjQzMTQ5MDc3",
"avatar_url": "https://avatars.githubusercontent.com/u/43149077?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dakinggg",
"html_url": "https://github.com/dakinggg",
"followers_url": "https://api.github.com/users/dak... | [] | open | false | null | [] | null | 11 | 2023-11-09T06:18:30 | 2024-03-05T16:03:12 | null | NONE | null | ### Describe the bug
A call to `.filter` occasionally hangs (after the filter is complete, according to tqdm)
There is a trace produced
```
Exception ignored in: <function Dataset.__del__ at 0x7efb48130c10>
Traceback (most recent call last):
File "/usr/lib/python3/dist-packages/datasets/arrow_dataset.py", l... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6393/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6393/timeline | null | null | null | null | false | [
"It looks like I may not be the first to encounter this: https://github.com/huggingface/datasets/issues/3172",
"Adding some more information, it seems to occur more frequently with large (millions of samples) datasets.",
"More information. My code is structured as (1) load (2) map (3) filter (4) filter. It was ... |
https://api.github.com/repos/huggingface/datasets/issues/6392 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6392/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6392/comments | https://api.github.com/repos/huggingface/datasets/issues/6392/events | https://github.com/huggingface/datasets/issues/6392 | 1,984,369,545 | I_kwDODunzps52RxOJ | 6,392 | `push_to_hub` is not robust to hub closing connection | {
"login": "msis",
"id": 577139,
"node_id": "MDQ6VXNlcjU3NzEzOQ==",
"avatar_url": "https://avatars.githubusercontent.com/u/577139?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/msis",
"html_url": "https://github.com/msis",
"followers_url": "https://api.github.com/users/msis/followers",
... | [] | closed | false | null | [] | null | 12 | 2023-11-08T20:44:53 | 2023-12-20T07:28:24 | 2023-12-01T17:51:34 | NONE | null | ### Describe the bug
Like to #6172, `push_to_hub` will crash if Hub resets the connection and raise the following error:
```
Pushing dataset shards to the dataset hub: 32%|███▏ | 54/171 [06:38<14:23, 7.38s/it]
Traceback (most recent call last):
File "/admin/home-piraka9011/.virtualenvs/w2v2/lib/python3.8/... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6392/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6392/timeline | null | completed | null | null | false | [
"Hi! We made some improvements to `push_to_hub` to make it more robust a couple of weeks ago but haven't published a release in the meantime, so it would help if you could install `datasets` from `main` (`pip install https://github.com/huggingface/datasets`) and let us know if this improved version of `push_to_hub`... |
https://api.github.com/repos/huggingface/datasets/issues/6391 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6391/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6391/comments | https://api.github.com/repos/huggingface/datasets/issues/6391/events | https://github.com/huggingface/datasets/pull/6391 | 1,984,091,776 | PR_kwDODunzps5e9BDO | 6,391 | Webdataset dataset builder | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [] | closed | false | null | [] | null | 5 | 2023-11-08T17:31:59 | 2023-11-28T16:33:33 | 2023-11-28T16:33:10 | MEMBER | null | Allow `load_dataset` to support the Webdataset format.
It allows users to download/stream data from local files or from the Hugging Face Hub.
Moreover it will enable the Dataset Viewer for Webdataset datasets on HF.
## Implementation details
- I added a new Webdataset builder
- dataset with TAR files are n... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6391/reactions",
"total_count": 3,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 3,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6391/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6391",
"html_url": "https://github.com/huggingface/datasets/pull/6391",
"diff_url": "https://github.com/huggingface/datasets/pull/6391.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6391.patch",
"merged_at": "2023-11-28T16:33... | true | [
"_The documentation is not available anymore as the PR was closed or merged._",
"I added an error message if the first examples don't appear to be in webdataset format\r\n```\r\n\"The TAR archives of the dataset should be in Webdataset format, \"\r\n\"but the files in the archive don't share the same prefix or th... |
https://api.github.com/repos/huggingface/datasets/issues/6390 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6390/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6390/comments | https://api.github.com/repos/huggingface/datasets/issues/6390/events | https://github.com/huggingface/datasets/pull/6390 | 1,983,725,707 | PR_kwDODunzps5e7xQ3 | 6,390 | handle future deprecation argument | {
"login": "winglian",
"id": 381258,
"node_id": "MDQ6VXNlcjM4MTI1OA==",
"avatar_url": "https://avatars.githubusercontent.com/u/381258?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/winglian",
"html_url": "https://github.com/winglian",
"followers_url": "https://api.github.com/users/winglia... | [] | closed | false | null | [] | null | 1 | 2023-11-08T14:21:25 | 2023-11-21T02:10:24 | 2023-11-14T15:15:59 | CONTRIBUTOR | null | getting this error:
```
/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/datasets/table.py:1387: FutureWarning: promote has been superseded by mode='default'.
return cls._concat_blocks(pa_tables_to_concat_vertically, axis=0)
```
Since datasets supports arrow greater than 8.0.0, we need to handle both ... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6390/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6390/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6390",
"html_url": "https://github.com/huggingface/datasets/pull/6390",
"diff_url": "https://github.com/huggingface/datasets/pull/6390.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6390.patch",
"merged_at": "2023-11-14T15:15... | true | [
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a... |
https://api.github.com/repos/huggingface/datasets/issues/6389 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6389/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6389/comments | https://api.github.com/repos/huggingface/datasets/issues/6389/events | https://github.com/huggingface/datasets/issues/6389 | 1,983,545,744 | I_kwDODunzps52OoGQ | 6,389 | Index 339 out of range for dataset of size 339 <-- save_to_file() | {
"login": "jaggzh",
"id": 20318973,
"node_id": "MDQ6VXNlcjIwMzE4OTcz",
"avatar_url": "https://avatars.githubusercontent.com/u/20318973?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jaggzh",
"html_url": "https://github.com/jaggzh",
"followers_url": "https://api.github.com/users/jaggzh/fo... | [] | open | false | null | [] | null | 2 | 2023-11-08T12:52:09 | 2023-11-24T09:14:13 | null | NONE | null | ### Describe the bug
When saving out some Audio() data.
The data is audio recordings with associated 'sentences'.
(They use the audio 'bytes' approach because they're clips within audio files).
Code is below the traceback (I can't upload the voice audio/text (it's not even me)).
```
Traceback (most recent call ... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6389/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6389/timeline | null | null | null | null | false | [
"Hi! Can you make the above reproducer self-contained by adding code that generates the data?",
"I managed a workaround eventually but I don't know what it was (I made a lot of changes to seq2seq). I'll try to include generating code in the future. (If I close, I don't know if you see it. Feel free to close; I'l... |
https://api.github.com/repos/huggingface/datasets/issues/6388 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6388/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6388/comments | https://api.github.com/repos/huggingface/datasets/issues/6388/events | https://github.com/huggingface/datasets/issues/6388 | 1,981,136,093 | I_kwDODunzps52Fbzd | 6,388 | How to create 3d medical imgae dataset? | {
"login": "QingYunA",
"id": 41177312,
"node_id": "MDQ6VXNlcjQxMTc3MzEy",
"avatar_url": "https://avatars.githubusercontent.com/u/41177312?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/QingYunA",
"html_url": "https://github.com/QingYunA",
"followers_url": "https://api.github.com/users/Qin... | [
{
"id": 1935892871,
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement",
"name": "enhancement",
"color": "a2eeef",
"default": true,
"description": "New feature or request"
}
] | open | false | null | [] | null | 0 | 2023-11-07T11:27:36 | 2023-11-07T11:28:53 | null | NONE | null | ### Feature request
I am newer to huggingface, after i look up `datasets` docs, I can't find how to create the dataset contains 3d medical image (ends with '.mhd', '.dcm', '.nii')
### Motivation
help us to upload 3d medical dataset to huggingface!
### Your contribution
I'll submit a PR if I find a way to... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6388/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6388/timeline | null | null | null | null | false | [] |
https://api.github.com/repos/huggingface/datasets/issues/6387 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6387/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6387/comments | https://api.github.com/repos/huggingface/datasets/issues/6387/events | https://github.com/huggingface/datasets/issues/6387 | 1,980,224,020 | I_kwDODunzps52B9IU | 6,387 | How to load existing downloaded dataset ? | {
"login": "liming-ai",
"id": 73068772,
"node_id": "MDQ6VXNlcjczMDY4Nzcy",
"avatar_url": "https://avatars.githubusercontent.com/u/73068772?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/liming-ai",
"html_url": "https://github.com/liming-ai",
"followers_url": "https://api.github.com/users/... | [
{
"id": 1935892871,
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement",
"name": "enhancement",
"color": "a2eeef",
"default": true,
"description": "New feature or request"
}
] | closed | false | null | [] | null | 1 | 2023-11-06T22:51:44 | 2023-11-16T18:07:01 | 2023-11-16T18:07:01 | NONE | null | Hi @mariosasko @lhoestq @katielink
Thanks for your contribution and hard work.
### Feature request
First, I download a dataset as normal by:
```
from datasets import load_dataset
dataset = load_dataset('username/data_name', cache_dir='data')
```
The dataset format in `data` directory will be:
```
... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6387/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6387/timeline | null | completed | null | null | false | [
"Feel free to use `dataset.save_to_disk(...)`, then scp the directory containing the saved dataset and reload it on your other machine using `dataset = load_from_disk(...)`"
] |
https://api.github.com/repos/huggingface/datasets/issues/6386 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6386/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6386/comments | https://api.github.com/repos/huggingface/datasets/issues/6386/events | https://github.com/huggingface/datasets/issues/6386 | 1,979,878,014 | I_kwDODunzps52Aop- | 6,386 | Formatting overhead | {
"login": "d-miketa",
"id": 320321,
"node_id": "MDQ6VXNlcjMyMDMyMQ==",
"avatar_url": "https://avatars.githubusercontent.com/u/320321?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/d-miketa",
"html_url": "https://github.com/d-miketa",
"followers_url": "https://api.github.com/users/d-miket... | [] | closed | false | null | [] | null | 2 | 2023-11-06T19:06:38 | 2023-11-06T23:56:12 | 2023-11-06T23:56:12 | NONE | null | ### Describe the bug
Hi! I very recently noticed that my training time is dominated by batch formatting. Using Lightning's profilers, I located the bottleneck within `datasets.formatting.formatting` and then narrowed it down with `line-profiler`. It turns out that almost all of the overhead is due to creating new inst... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6386/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6386/timeline | null | completed | null | null | false | [
"Ah I think the `line-profiler` log is off-by-one and it is in fact the `extract_batch` method that's taking forever. Will investigate further.",
"I tracked it down to a quirk of my setup. Apologies."
] |
https://api.github.com/repos/huggingface/datasets/issues/6385 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6385/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6385/comments | https://api.github.com/repos/huggingface/datasets/issues/6385/events | https://github.com/huggingface/datasets/issues/6385 | 1,979,308,338 | I_kwDODunzps51-dky | 6,385 | Get an error when i try to concatenate the squad dataset with my own dataset | {
"login": "CCDXDX",
"id": 149378500,
"node_id": "U_kgDOCOdVxA",
"avatar_url": "https://avatars.githubusercontent.com/u/149378500?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/CCDXDX",
"html_url": "https://github.com/CCDXDX",
"followers_url": "https://api.github.com/users/CCDXDX/follower... | [] | closed | false | null | [] | null | 2 | 2023-11-06T14:29:22 | 2023-11-06T16:50:45 | 2023-11-06T16:50:45 | NONE | null | ### Describe the bug
Hello,
I'm new here and I need to concatenate the squad dataset with my own dataset i created. I find the following error when i try to do it: Traceback (most recent call last):
Cell In[9], line 1
concatenated_dataset = concatenate_datasets([train_dataset, dataset1])
File ~\ana... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6385/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6385/timeline | null | completed | null | null | false | [
"The `answers.text` field in the JSON dataset needs to be a list of strings, not a string.\r\n\r\nSo, here is the fixed code:\r\n```python\r\nfrom huggingface_hub import notebook_login\r\nfrom datasets import load_dataset\r\n\r\n\r\n\r\nnotebook_login(\"mymailadresse\", \"mypassword\")\r\nsquad = load_dataset(\"squ... |
https://api.github.com/repos/huggingface/datasets/issues/6384 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6384/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6384/comments | https://api.github.com/repos/huggingface/datasets/issues/6384/events | https://github.com/huggingface/datasets/issues/6384 | 1,979,117,069 | I_kwDODunzps519u4N | 6,384 | Load the local dataset folder from other place | {
"login": "OrangeSodahub",
"id": 54439582,
"node_id": "MDQ6VXNlcjU0NDM5NTgy",
"avatar_url": "https://avatars.githubusercontent.com/u/54439582?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/OrangeSodahub",
"html_url": "https://github.com/OrangeSodahub",
"followers_url": "https://api.githu... | [] | closed | false | null | [] | null | 1 | 2023-11-06T13:07:04 | 2023-11-19T05:42:06 | 2023-11-19T05:42:05 | NONE | null | This is from https://github.com/huggingface/diffusers/issues/5573
| {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6384/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6384/timeline | null | completed | null | null | false | [
"Solved"
] |
https://api.github.com/repos/huggingface/datasets/issues/6383 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6383/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6383/comments | https://api.github.com/repos/huggingface/datasets/issues/6383/events | https://github.com/huggingface/datasets/issues/6383 | 1,978,189,389 | I_kwDODunzps516MZN | 6,383 | imagenet-1k downloads over and over | {
"login": "seann999",
"id": 6847529,
"node_id": "MDQ6VXNlcjY4NDc1Mjk=",
"avatar_url": "https://avatars.githubusercontent.com/u/6847529?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/seann999",
"html_url": "https://github.com/seann999",
"followers_url": "https://api.github.com/users/seann... | [] | closed | false | null | [] | null | 0 | 2023-11-06T02:58:58 | 2023-11-06T06:02:39 | 2023-11-06T06:02:39 | NONE | null | ### Describe the bug
What could be causing this?
```
$ python3
Python 3.8.13 (default, Mar 28 2022, 11:38:47)
[GCC 7.5.0] :: Anaconda, Inc. on linux
Type "help", "copyright", "credits" or "license" for more information.
>>> from datasets import load_dataset
>>> load_dataset("imagenet-1k")
Downloading builder ... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6383/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6383/timeline | null | completed | null | null | false | [] |
https://api.github.com/repos/huggingface/datasets/issues/6382 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6382/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6382/comments | https://api.github.com/repos/huggingface/datasets/issues/6382/events | https://github.com/huggingface/datasets/issues/6382 | 1,977,400,799 | I_kwDODunzps513L3f | 6,382 | Add CheXpert dataset for vision | {
"login": "SauravMaheshkar",
"id": 61241031,
"node_id": "MDQ6VXNlcjYxMjQxMDMx",
"avatar_url": "https://avatars.githubusercontent.com/u/61241031?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/SauravMaheshkar",
"html_url": "https://github.com/SauravMaheshkar",
"followers_url": "https://api... | [
{
"id": 1935892871,
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement",
"name": "enhancement",
"color": "a2eeef",
"default": true,
"description": "New feature or request"
},
{
"id": 2067376369,
"node_id": "MDU6... | open | false | null | [] | null | 3 | 2023-11-04T15:36:11 | 2024-01-10T11:53:52 | null | NONE | null | ### Feature request
### Name
**CheXpert: A Large Chest Radiograph Dataset with Uncertainty Labels and Expert Comparison**
### Paper
https://arxiv.org/abs/1901.07031
### Data
https://stanfordaimi.azurewebsites.net/datasets/8cbd9ed4-2eb9-4565-affc-111cf4f7ebe2
### Motivation
CheXpert is one of the fund... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6382/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6382/timeline | null | null | null | null | false | [
"Hey @SauravMaheshkar ! Just responded to your email.\r\n\r\n_For transparency, copying part of my response here:_\r\nI agree, it would be really great to have this and other BenchMD datasets easily accessible on the hub.\r\n\r\nI think the main limiting factor is that the ChexPert dataset is currently hosted on th... |
https://api.github.com/repos/huggingface/datasets/issues/6381 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6381/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6381/comments | https://api.github.com/repos/huggingface/datasets/issues/6381/events | https://github.com/huggingface/datasets/pull/6381 | 1,975,028,470 | PR_kwDODunzps5eeYty | 6,381 | Add my dataset | {
"login": "keyur536",
"id": 103646675,
"node_id": "U_kgDOBi2F0w",
"avatar_url": "https://avatars.githubusercontent.com/u/103646675?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/keyur536",
"html_url": "https://github.com/keyur536",
"followers_url": "https://api.github.com/users/keyur536/... | [] | closed | false | null | [] | null | 3 | 2023-11-02T20:59:52 | 2023-11-08T14:37:46 | 2023-11-06T15:50:14 | NONE | null | ## medical data
**Description:**
This dataset, named "medical data," is a collection of text data from various sources, carefully curated and cleaned for use in natural language processing (NLP) tasks. It consists of a diverse range of text, including articles, books, and online content, covering topics from scienc... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6381/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6381/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6381",
"html_url": "https://github.com/huggingface/datasets/pull/6381",
"diff_url": "https://github.com/huggingface/datasets/pull/6381.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6381.patch",
"merged_at": null
} | true | [
"Hi! We do not host datasets in this repo. Instead, you should use `dataset.push_to_hub` to upload the dataset to the HF Hub.",
"@mariosasko could you provide me proper guide to push data on HF hub ",
"You can find this info here: https://huggingface.co/docs/datasets/upload_dataset. Also, check https://huggingf... |
https://api.github.com/repos/huggingface/datasets/issues/6380 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6380/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6380/comments | https://api.github.com/repos/huggingface/datasets/issues/6380/events | https://github.com/huggingface/datasets/pull/6380 | 1,974,741,221 | PR_kwDODunzps5edaO6 | 6,380 | Fix for continuation behaviour on broken dataset archives due to starving download connections via HTTP-GET | {
"login": "RuntimeRacer",
"id": 49956579,
"node_id": "MDQ6VXNlcjQ5OTU2NTc5",
"avatar_url": "https://avatars.githubusercontent.com/u/49956579?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/RuntimeRacer",
"html_url": "https://github.com/RuntimeRacer",
"followers_url": "https://api.github.c... | [] | open | false | null | [] | null | 0 | 2023-11-02T17:28:23 | 2023-11-02T17:31:19 | null | NONE | null | This PR proposes a (slightly hacky) fix for an Issue that can occur when downloading large dataset parts over unstable connections.
The underlying issue is also being discussed in https://github.com/huggingface/datasets/issues/5594.
Issue Symptoms & Behaviour:
- Download of a large archive file during dataset down... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6380/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6380/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6380",
"html_url": "https://github.com/huggingface/datasets/pull/6380",
"diff_url": "https://github.com/huggingface/datasets/pull/6380.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6380.patch",
"merged_at": null
} | true | [] |
https://api.github.com/repos/huggingface/datasets/issues/6379 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6379/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6379/comments | https://api.github.com/repos/huggingface/datasets/issues/6379/events | https://github.com/huggingface/datasets/pull/6379 | 1,974,638,850 | PR_kwDODunzps5edDZL | 6,379 | Avoid redundant warning when encoding NumPy array as `Image` | {
"login": "mariosasko",
"id": 47462742,
"node_id": "MDQ6VXNlcjQ3NDYyNzQy",
"avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mariosasko",
"html_url": "https://github.com/mariosasko",
"followers_url": "https://api.github.com/use... | [] | closed | false | null | [] | null | 5 | 2023-11-02T16:37:58 | 2023-11-06T17:53:27 | 2023-11-02T17:08:07 | CONTRIBUTOR | null | Avoid a redundant warning in `encode_np_array` by removing the identity check as NumPy `dtype`s can be equal without having identical `id`s.
Additionally, fix "unreachable" checks in `encode_np_array`. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6379/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6379/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6379",
"html_url": "https://github.com/huggingface/datasets/pull/6379",
"diff_url": "https://github.com/huggingface/datasets/pull/6379.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6379.patch",
"merged_at": "2023-11-02T17:08... | true | [
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a... |
https://api.github.com/repos/huggingface/datasets/issues/6378 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6378/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6378/comments | https://api.github.com/repos/huggingface/datasets/issues/6378/events | https://github.com/huggingface/datasets/pull/6378 | 1,973,942,770 | PR_kwDODunzps5eaqhv | 6,378 | Support pyarrow 14.0.0 | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | 3 | 2023-11-02T10:25:10 | 2023-11-02T15:24:28 | 2023-11-02T15:15:44 | MEMBER | null | Support `pyarrow` 14.0.0.
Fix #6377 and fix #6374 (root cause).
This fix is analog to a previous one:
- #6175 | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6378/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6378/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6378",
"html_url": "https://github.com/huggingface/datasets/pull/6378",
"diff_url": "https://github.com/huggingface/datasets/pull/6378.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6378.patch",
"merged_at": "2023-11-02T15:15... | true | [
"_The documentation is not available anymore as the PR was closed or merged._",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea... |
https://api.github.com/repos/huggingface/datasets/issues/6377 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6377/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6377/comments | https://api.github.com/repos/huggingface/datasets/issues/6377/events | https://github.com/huggingface/datasets/issues/6377 | 1,973,937,612 | I_kwDODunzps51p-XM | 6,377 | Support pyarrow 14.0.0 | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_... | null | 0 | 2023-11-02T10:22:08 | 2023-11-02T15:15:45 | 2023-11-02T15:15:45 | MEMBER | null | Support pyarrow 14.0.0 by fixing the root cause of:
- #6374
and revert:
- #6375 | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6377/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6377/timeline | null | completed | null | null | false | [] |
https://api.github.com/repos/huggingface/datasets/issues/6376 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6376/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6376/comments | https://api.github.com/repos/huggingface/datasets/issues/6376/events | https://github.com/huggingface/datasets/issues/6376 | 1,973,927,468 | I_kwDODunzps51p74s | 6,376 | Caching problem when deleting a dataset | {
"login": "clefourrier",
"id": 22726840,
"node_id": "MDQ6VXNlcjIyNzI2ODQw",
"avatar_url": "https://avatars.githubusercontent.com/u/22726840?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/clefourrier",
"html_url": "https://github.com/clefourrier",
"followers_url": "https://api.github.com/... | [] | closed | false | null | [] | null | 3 | 2023-11-02T10:15:58 | 2023-12-04T16:53:34 | 2023-12-04T16:53:33 | MEMBER | null | ### Describe the bug
Pushing a dataset with n + m features to a repo which was deleted, but contained n features, will fail.
### Steps to reproduce the bug
1. Create a dataset with n features per row
2. `dataset.push_to_hub(YOUR_PATH, SPLIT, token=TOKEN)`
3. Go on the hub, delete the repo at `YOUR_PATH`
4. Update... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6376/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6376/timeline | null | completed | null | null | false | [
"Thanks for reporting! Can you also share the error message printed in step 5?",
"I did not store it at the time but I'll try to re-do a mwe next week to get it again",
"I haven't managed to reproduce this issue using a [notebook](https://colab.research.google.com/drive/1m6eduYun7pFTkigrCJAFgw0BghlbvXIL?usp=sha... |
https://api.github.com/repos/huggingface/datasets/issues/6375 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6375/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6375/comments | https://api.github.com/repos/huggingface/datasets/issues/6375/events | https://github.com/huggingface/datasets/pull/6375 | 1,973,877,879 | PR_kwDODunzps5eacao | 6,375 | Temporarily pin pyarrow < 14.0.0 | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | 3 | 2023-11-02T09:48:58 | 2023-11-02T10:22:33 | 2023-11-02T10:11:19 | MEMBER | null | Temporarily pin `pyarrow` < 14.0.0 until permanent solution is found.
Hot fix #6374. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6375/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6375/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6375",
"html_url": "https://github.com/huggingface/datasets/pull/6375",
"diff_url": "https://github.com/huggingface/datasets/pull/6375.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6375.patch",
"merged_at": "2023-11-02T10:11... | true | [
"_The documentation is not available anymore as the PR was closed or merged._",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea... |
https://api.github.com/repos/huggingface/datasets/issues/6374 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6374/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6374/comments | https://api.github.com/repos/huggingface/datasets/issues/6374/events | https://github.com/huggingface/datasets/issues/6374 | 1,973,857,428 | I_kwDODunzps51pqyU | 6,374 | CI is broken: TypeError: Couldn't cast array | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_... | null | 0 | 2023-11-02T09:37:06 | 2023-11-02T10:11:20 | 2023-11-02T10:11:20 | MEMBER | null | See: https://github.com/huggingface/datasets/actions/runs/6730567226/job/18293518039
```
FAILED tests/test_table.py::test_cast_sliced_fixed_size_array_to_features - TypeError: Couldn't cast array of type
fixed_size_list<item: int32>[3]
to
Sequence(feature=Value(dtype='int64', id=None), length=3, id=None)
``` | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6374/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6374/timeline | null | completed | null | null | false | [] |
https://api.github.com/repos/huggingface/datasets/issues/6373 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6373/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6373/comments | https://api.github.com/repos/huggingface/datasets/issues/6373/events | https://github.com/huggingface/datasets/pull/6373 | 1,973,349,695 | PR_kwDODunzps5eYsZc | 6,373 | Fix typo in `Dataset.map` docstring | {
"login": "bryant1410",
"id": 3905501,
"node_id": "MDQ6VXNlcjM5MDU1MDE=",
"avatar_url": "https://avatars.githubusercontent.com/u/3905501?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/bryant1410",
"html_url": "https://github.com/bryant1410",
"followers_url": "https://api.github.com/users... | [] | closed | false | null | [] | null | 2 | 2023-11-02T01:36:49 | 2023-11-02T15:18:22 | 2023-11-02T10:11:38 | CONTRIBUTOR | null | null | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6373/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6373/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6373",
"html_url": "https://github.com/huggingface/datasets/pull/6373",
"diff_url": "https://github.com/huggingface/datasets/pull/6373.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6373.patch",
"merged_at": "2023-11-02T10:11... | true | [
"_The documentation is not available anymore as the PR was closed or merged._",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea... |
https://api.github.com/repos/huggingface/datasets/issues/6372 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6372/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6372/comments | https://api.github.com/repos/huggingface/datasets/issues/6372/events | https://github.com/huggingface/datasets/pull/6372 | 1,972,837,794 | PR_kwDODunzps5eW9kO | 6,372 | do not try to download from HF GCS for generator | {
"login": "yundai424",
"id": 43726198,
"node_id": "MDQ6VXNlcjQzNzI2MTk4",
"avatar_url": "https://avatars.githubusercontent.com/u/43726198?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/yundai424",
"html_url": "https://github.com/yundai424",
"followers_url": "https://api.github.com/users/... | [] | closed | false | null | [] | null | 2 | 2023-11-01T17:57:11 | 2023-11-02T16:02:52 | 2023-11-02T15:52:09 | CONTRIBUTOR | null | attempt to fix https://github.com/huggingface/datasets/issues/6371 | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6372/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6372/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6372",
"html_url": "https://github.com/huggingface/datasets/pull/6372",
"diff_url": "https://github.com/huggingface/datasets/pull/6372.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6372.patch",
"merged_at": "2023-11-02T15:52... | true | [
"_The documentation is not available anymore as the PR was closed or merged._",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea... |
https://api.github.com/repos/huggingface/datasets/issues/6371 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6371/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6371/comments | https://api.github.com/repos/huggingface/datasets/issues/6371/events | https://github.com/huggingface/datasets/issues/6371 | 1,972,807,579 | I_kwDODunzps51lqeb | 6,371 | `Dataset.from_generator` should not try to download from HF GCS | {
"login": "yundai424",
"id": 43726198,
"node_id": "MDQ6VXNlcjQzNzI2MTk4",
"avatar_url": "https://avatars.githubusercontent.com/u/43726198?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/yundai424",
"html_url": "https://github.com/yundai424",
"followers_url": "https://api.github.com/users/... | [] | closed | false | null | [] | null | 1 | 2023-11-01T17:36:17 | 2023-11-02T15:52:10 | 2023-11-02T15:52:10 | CONTRIBUTOR | null | ### Describe the bug
When using [`Dataset.from_generator`](https://github.com/huggingface/datasets/blob/c9c1166e1cf81d38534020f9c167b326585339e5/src/datasets/arrow_dataset.py#L1072) with `streaming=False`, the internal logic will call [`download_and_prepare`](https://github.com/huggingface/datasets/blob/main/src/datas... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6371/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6371/timeline | null | completed | null | null | false | [
"Indeed, setting `try_from_gcs` to `False` makes sense for `from_generator`.\r\n\r\nWe plan to deprecate and remove `try_from_hf_gcs` soon, as we can use Hub for file hosting now, but this is a good temporary fix.\r\n"
] |
https://api.github.com/repos/huggingface/datasets/issues/6370 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6370/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6370/comments | https://api.github.com/repos/huggingface/datasets/issues/6370/events | https://github.com/huggingface/datasets/issues/6370 | 1,972,073,909 | I_kwDODunzps51i3W1 | 6,370 | TensorDataset format does not work with Trainer from transformers | {
"login": "jinzzasol",
"id": 49014051,
"node_id": "MDQ6VXNlcjQ5MDE0MDUx",
"avatar_url": "https://avatars.githubusercontent.com/u/49014051?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jinzzasol",
"html_url": "https://github.com/jinzzasol",
"followers_url": "https://api.github.com/users/... | [] | closed | false | null | [] | null | 2 | 2023-11-01T10:09:54 | 2023-11-29T16:31:08 | 2023-11-29T16:31:08 | NONE | null | ### Describe the bug
The model was built to do fine tunning on BERT model for relation extraction.
trainer.train() returns an error message ```TypeError: vars() argument must have __dict__ attribute``` when it has `train_dataset` generated from `torch.utils.data.TensorDataset`
However, in the document, the req... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6370/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6370/timeline | null | completed | null | null | false | [
"I figured it out. I found that `Trainer` does not work with TensorDataset even though the document says it uses it. Instead, I ended up creating a dictionary and converting it to a dataset using `dataset.Dataset.from_dict()`.\r\n\r\nI will leave this post open for a while. If someone knows a better approach, pleas... |
https://api.github.com/repos/huggingface/datasets/issues/6369 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6369/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6369/comments | https://api.github.com/repos/huggingface/datasets/issues/6369/events | https://github.com/huggingface/datasets/issues/6369 | 1,971,794,108 | I_kwDODunzps51hzC8 | 6,369 | Multi process map did not load cache file correctly | {
"login": "enze5088",
"id": 14285786,
"node_id": "MDQ6VXNlcjE0Mjg1Nzg2",
"avatar_url": "https://avatars.githubusercontent.com/u/14285786?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/enze5088",
"html_url": "https://github.com/enze5088",
"followers_url": "https://api.github.com/users/enz... | [] | closed | false | null | [] | null | 3 | 2023-11-01T06:36:54 | 2023-11-30T16:04:46 | 2023-11-30T16:04:45 | NONE | null | ### Describe the bug
When I was training model on Multiple GPUs by DDP, the dataset is tokenized multiple times after main process.

 function returns bytes instead of PIL images even when image column is not part of "columns" | {
"login": "leot13",
"id": 17809020,
"node_id": "MDQ6VXNlcjE3ODA5MDIw",
"avatar_url": "https://avatars.githubusercontent.com/u/17809020?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/leot13",
"html_url": "https://github.com/leot13",
"followers_url": "https://api.github.com/users/leot13/fo... | [] | closed | false | null | [] | null | 1 | 2023-10-31T11:10:48 | 2023-11-02T14:21:17 | 2023-11-02T14:21:17 | NONE | null | ### Describe the bug
When using the with_format() function on a dataset containing images, even if the image column is not part of the columns provided in the function, its type will be changed to bytes.
Here is a minimal reproduction of the bug:
https://colab.research.google.com/drive/1hyaOspgyhB41oiR1-tXE3k_gJCdJU... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6366/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6366/timeline | null | completed | null | null | false | [
"Thanks for reporting! I've opened a PR with a fix."
] |
https://api.github.com/repos/huggingface/datasets/issues/6365 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6365/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6365/comments | https://api.github.com/repos/huggingface/datasets/issues/6365/events | https://github.com/huggingface/datasets/issues/6365 | 1,970,140,392 | I_kwDODunzps51bfTo | 6,365 | Parquet size grows exponential for categorical data | {
"login": "aseganti",
"id": 82567957,
"node_id": "MDQ6VXNlcjgyNTY3OTU3",
"avatar_url": "https://avatars.githubusercontent.com/u/82567957?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/aseganti",
"html_url": "https://github.com/aseganti",
"followers_url": "https://api.github.com/users/ase... | [] | closed | false | null | [] | null | 1 | 2023-10-31T10:29:02 | 2023-10-31T10:49:17 | 2023-10-31T10:49:17 | NONE | null | ### Describe the bug
It seems that when saving a data frame with a categorical column inside the size can grow exponentially.
This seems to happen because when we save the categorical data to parquet, we are saving the data + all the categories existing in the original data. This happens even when the categories ar... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6365/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6365/timeline | null | not_planned | null | null | false | [
"Wrong repo."
] |
https://api.github.com/repos/huggingface/datasets/issues/6364 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6364/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6364/comments | https://api.github.com/repos/huggingface/datasets/issues/6364/events | https://github.com/huggingface/datasets/issues/6364 | 1,969,136,106 | I_kwDODunzps51XqHq | 6,364 | ArrowNotImplementedError: Unsupported cast from string to list using function cast_list | {
"login": "divyakrishna-devisetty",
"id": 32887094,
"node_id": "MDQ6VXNlcjMyODg3MDk0",
"avatar_url": "https://avatars.githubusercontent.com/u/32887094?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/divyakrishna-devisetty",
"html_url": "https://github.com/divyakrishna-devisetty",
"followe... | [] | closed | false | null | [] | null | 2 | 2023-10-30T20:14:01 | 2023-10-31T19:21:23 | 2023-10-31T19:21:23 | NONE | null | Hi,
I am trying to load a local csv dataset(similar to explodinggradients_fiqa) using load_dataset. When I try to pass features, I am facing the mentioned issue.
CSV Data sample(golden_dataset.csv):
Question | Context | answer | groundtruth
"what is abc?"... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6364/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6364/timeline | null | completed | null | null | false | [
"You can use the following code to load this CSV with the list values preserved:\r\n```python\r\nfrom datasets import load_dataset\r\nimport ast\r\n\r\nconverters = {\r\n \"contexts\" : ast.literal_eval,\r\n \"ground_truths\" : ast.literal_eval,\r\n}\r\n\r\nds = load_dataset(\"csv\", data_files=\"golden_datas... |
https://api.github.com/repos/huggingface/datasets/issues/6363 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6363/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6363/comments | https://api.github.com/repos/huggingface/datasets/issues/6363/events | https://github.com/huggingface/datasets/issues/6363 | 1,968,891,277 | I_kwDODunzps51WuWN | 6,363 | dataset.transform() hangs indefinitely while finetuning the stable diffusion XL | {
"login": "bhosalems",
"id": 10846405,
"node_id": "MDQ6VXNlcjEwODQ2NDA1",
"avatar_url": "https://avatars.githubusercontent.com/u/10846405?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/bhosalems",
"html_url": "https://github.com/bhosalems",
"followers_url": "https://api.github.com/users/... | [] | closed | false | null | [] | null | 7 | 2023-10-30T17:34:05 | 2023-11-22T00:29:21 | 2023-11-22T00:29:21 | NONE | null | ### Describe the bug
Multi-GPU fine-tuning the stable diffusion X by following https://github.com/huggingface/diffusers/blob/main/examples/text_to_image/README_sdxl.md hangs indefinitely.
### Steps to reproduce the bug
accelerate launch train_text_to_image_sdxl.py --pretrained_model_name_or_path=$MODEL_NAME --... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6363/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6363/timeline | null | completed | null | null | false | [
"I think the code hangs on the `accelerator.main_process_first()` context manager exit. To verify this, you can append a print statement to the end of the `accelerator.main_process_first()` block. \r\n\r\n\r\nIf the problem is in `with_transform`, it would help if you could share the error stack trace printed when... |
https://api.github.com/repos/huggingface/datasets/issues/6362 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6362/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6362/comments | https://api.github.com/repos/huggingface/datasets/issues/6362/events | https://github.com/huggingface/datasets/pull/6362 | 1,965,794,569 | PR_kwDODunzps5d_MxD | 6,362 | Simplify filesystem logic | {
"login": "mariosasko",
"id": 47462742,
"node_id": "MDQ6VXNlcjQ3NDYyNzQy",
"avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mariosasko",
"html_url": "https://github.com/mariosasko",
"followers_url": "https://api.github.com/use... | [] | closed | false | null | [] | null | 13 | 2023-10-27T15:54:18 | 2023-11-15T14:08:29 | 2023-11-15T14:02:02 | CONTRIBUTOR | null | Simplifies the existing filesystem logic (e.g., to avoid unnecessary if-else as mentioned in https://github.com/huggingface/datasets/pull/6098#issue-1827655071) | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6362/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6362/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6362",
"html_url": "https://github.com/huggingface/datasets/pull/6362",
"diff_url": "https://github.com/huggingface/datasets/pull/6362.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6362.patch",
"merged_at": "2023-11-15T14:02... | true | [
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a... |
https://api.github.com/repos/huggingface/datasets/issues/6360 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6360/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6360/comments | https://api.github.com/repos/huggingface/datasets/issues/6360/events | https://github.com/huggingface/datasets/issues/6360 | 1,965,672,950 | I_kwDODunzps51Kcn2 | 6,360 | Add support for `Sequence(Audio/Image)` feature in `push_to_hub` | {
"login": "Laurent2916",
"id": 21087104,
"node_id": "MDQ6VXNlcjIxMDg3MTA0",
"avatar_url": "https://avatars.githubusercontent.com/u/21087104?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Laurent2916",
"html_url": "https://github.com/Laurent2916",
"followers_url": "https://api.github.com/... | [
{
"id": 1935892871,
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement",
"name": "enhancement",
"color": "a2eeef",
"default": true,
"description": "New feature or request"
}
] | closed | false | {
"login": "mariosasko",
"id": 47462742,
"node_id": "MDQ6VXNlcjQ3NDYyNzQy",
"avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mariosasko",
"html_url": "https://github.com/mariosasko",
"followers_url": "https://api.github.com/use... | [
{
"login": "mariosasko",
"id": 47462742,
"node_id": "MDQ6VXNlcjQ3NDYyNzQy",
"avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mariosasko",
"html_url": "https://github.com/mariosasko",
"followers_url": "https:... | null | 1 | 2023-10-27T14:39:57 | 2024-02-06T19:24:20 | 2024-02-06T19:24:20 | CONTRIBUTOR | null | ### Feature request
Allow for `Sequence` of `Image` (or `Audio`) to be embedded inside the shards.
### Motivation
Currently, thanks to #3685, when `embed_external_files` is set to True (which is the default) in `push_to_hub`, features of type `Image` and `Audio` are embedded inside the arrow/parquet shards, instead ... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6360/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6360/timeline | null | completed | null | null | false | [
"This issue stems from https://github.com/huggingface/datasets/blob/6d2f2a5e0fea3827eccfd1717d8021c15fc4292a/src/datasets/table.py#L2203-L2205\r\n\r\nI'll address it as part of https://github.com/huggingface/datasets/pull/6283.\r\n\r\nIn the meantime, this should work\r\n\r\n```python\r\nimport pyarrow as pa\r\nfro... |
https://api.github.com/repos/huggingface/datasets/issues/6359 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6359/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6359/comments | https://api.github.com/repos/huggingface/datasets/issues/6359/events | https://github.com/huggingface/datasets/issues/6359 | 1,965,378,583 | I_kwDODunzps51JUwX | 6,359 | Stuck in "Resolving data files..." | {
"login": "Luciennnnnnn",
"id": 20135317,
"node_id": "MDQ6VXNlcjIwMTM1MzE3",
"avatar_url": "https://avatars.githubusercontent.com/u/20135317?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Luciennnnnnn",
"html_url": "https://github.com/Luciennnnnnn",
"followers_url": "https://api.github.c... | [] | open | false | null | [] | null | 4 | 2023-10-27T12:01:51 | 2024-01-24T15:02:06 | null | NONE | null | ### Describe the bug
I have an image dataset with 300k images, the size of image is 768 * 768.
When I run `dataset = load_dataset("imagefolder", data_dir="/path/to/img_dir", split='train')` in second time, it takes 50 minutes to finish "Resolving data files" part, what's going on in this part?
From my understa... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6359/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6359/timeline | null | null | null | null | false | [
"Most likely, the data file inference logic is the problem here.\r\n\r\nYou can run the following code to verify this:\r\n```python\r\nimport time\r\nfrom datasets.data_files import get_data_patterns\r\nstart_time = time.time()\r\nget_data_patterns(\"/path/to/img_dir\")\r\nend_time = time.time()\r\nprint(f\"Elapsed... |
https://api.github.com/repos/huggingface/datasets/issues/6358 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6358/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6358/comments | https://api.github.com/repos/huggingface/datasets/issues/6358/events | https://github.com/huggingface/datasets/issues/6358 | 1,965,014,595 | I_kwDODunzps51H75D | 6,358 | Mounting datasets cache fails due to absolute paths. | {
"login": "charliebudd",
"id": 72921588,
"node_id": "MDQ6VXNlcjcyOTIxNTg4",
"avatar_url": "https://avatars.githubusercontent.com/u/72921588?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/charliebudd",
"html_url": "https://github.com/charliebudd",
"followers_url": "https://api.github.com/... | [] | closed | false | null | [] | null | 3 | 2023-10-27T08:20:27 | 2023-11-28T14:47:12 | 2023-11-28T14:47:12 | NONE | null | ### Describe the bug
Creating a datasets cache and mounting this into, for example, a docker container, renders the data unreadable due to absolute paths written into the cache.
### Steps to reproduce the bug
1. Create a datasets cache by downloading some data
2. Mount the dataset folder into a docker contain... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6358/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6358/timeline | null | completed | null | null | false | [
"You may be able to make it work by tweaking some environment variables, such as [`HF_HOME`](https://huggingface.co/docs/huggingface_hub/main/en/package_reference/environment_variables#hfhome) or [`HF_DATASETS_CACHE`](https://huggingface.co/docs/datasets/cache#cache-directory).",
"> You may be able to make it wor... |
https://api.github.com/repos/huggingface/datasets/issues/6357 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6357/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6357/comments | https://api.github.com/repos/huggingface/datasets/issues/6357/events | https://github.com/huggingface/datasets/issues/6357 | 1,964,653,995 | I_kwDODunzps51Gj2r | 6,357 | Allow passing a multiprocessing context to functions that support `num_proc` | {
"login": "bryant1410",
"id": 3905501,
"node_id": "MDQ6VXNlcjM5MDU1MDE=",
"avatar_url": "https://avatars.githubusercontent.com/u/3905501?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/bryant1410",
"html_url": "https://github.com/bryant1410",
"followers_url": "https://api.github.com/users... | [
{
"id": 1935892871,
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement",
"name": "enhancement",
"color": "a2eeef",
"default": true,
"description": "New feature or request"
}
] | open | false | null | [] | null | 0 | 2023-10-27T02:31:16 | 2023-10-27T02:31:16 | null | CONTRIBUTOR | null | ### Feature request
Allow specifying [a multiprocessing context](https://docs.python.org/3/library/multiprocessing.html#contexts-and-start-methods) to functions that support `num_proc` or use multiprocessing pools. For example, the following could be done:
```python
dataset = dataset.map(_func, num_proc=2, mp_cont... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6357/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6357/timeline | null | null | null | null | false | [] |
https://api.github.com/repos/huggingface/datasets/issues/6356 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6356/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6356/comments | https://api.github.com/repos/huggingface/datasets/issues/6356/events | https://github.com/huggingface/datasets/pull/6356 | 1,964,015,802 | PR_kwDODunzps5d5Jri | 6,356 | Add `fsspec` version to the `datasets-cli env` command output | {
"login": "mariosasko",
"id": 47462742,
"node_id": "MDQ6VXNlcjQ3NDYyNzQy",
"avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mariosasko",
"html_url": "https://github.com/mariosasko",
"followers_url": "https://api.github.com/use... | [] | closed | false | null | [] | null | 3 | 2023-10-26T17:19:25 | 2023-10-26T18:42:56 | 2023-10-26T18:32:21 | CONTRIBUTOR | null | ... to make debugging issues easier, as `fsspec`'s releases often introduce breaking changes. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6356/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6356/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6356",
"html_url": "https://github.com/huggingface/datasets/pull/6356",
"diff_url": "https://github.com/huggingface/datasets/pull/6356.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6356.patch",
"merged_at": "2023-10-26T18:32... | true | [
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a... |
https://api.github.com/repos/huggingface/datasets/issues/6355 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6355/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6355/comments | https://api.github.com/repos/huggingface/datasets/issues/6355/events | https://github.com/huggingface/datasets/pull/6355 | 1,963,979,896 | PR_kwDODunzps5d5B2B | 6,355 | More hub centric docs | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [] | closed | false | null | [] | null | 3 | 2023-10-26T16:54:46 | 2024-01-11T06:34:16 | 2023-10-30T17:32:57 | MEMBER | null | Let's have more hub-centric documentation in the datasets docs
Tutorials
- Add “Configure the dataset viewer” page
- Change order:
- Overview
- and more focused on the Hub rather than the library
- Then all the hub related things
- and mention how to read/write with other tools like pandas
- The... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6355/reactions",
"total_count": 1,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 1,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6355/timeline | null | null | true | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6355",
"html_url": "https://github.com/huggingface/datasets/pull/6355",
"diff_url": "https://github.com/huggingface/datasets/pull/6355.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6355.patch",
"merged_at": null
} | true | [
"_The documentation is not available anymore as the PR was closed or merged._",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea... |
https://api.github.com/repos/huggingface/datasets/issues/6354 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6354/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6354/comments | https://api.github.com/repos/huggingface/datasets/issues/6354/events | https://github.com/huggingface/datasets/issues/6354 | 1,963,483,324 | I_kwDODunzps51CGC8 | 6,354 | `IterableDataset.from_spark` does not support multiple workers in pytorch `Dataloader` | {
"login": "NazyS",
"id": 50199774,
"node_id": "MDQ6VXNlcjUwMTk5Nzc0",
"avatar_url": "https://avatars.githubusercontent.com/u/50199774?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/NazyS",
"html_url": "https://github.com/NazyS",
"followers_url": "https://api.github.com/users/NazyS/follow... | [] | open | false | null | [] | null | 1 | 2023-10-26T12:43:36 | 2023-11-14T18:46:03 | null | NONE | null | ### Describe the bug
Looks like `IterableDataset.from_spark` does not support multiple workers in pytorch `Dataloader` if I'm not missing anything.
Also, returns not consistent error messages, which probably depend on the nondeterministic order of worker executions
Some exampes I've encountered:
```
File "/l... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6354/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6354/timeline | null | null | null | null | false | [
"I am having issues as well with this. \r\n\r\nHowever, the error I am getting is :\r\n`RuntimeError: It appears that you are attempting to reference SparkContext from a broadcast variable, action, or transformation. SparkContext can only be used on the driver, not in code that it run on workers. For more informati... |
https://api.github.com/repos/huggingface/datasets/issues/6353 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6353/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6353/comments | https://api.github.com/repos/huggingface/datasets/issues/6353/events | https://github.com/huggingface/datasets/issues/6353 | 1,962,646,450 | I_kwDODunzps50-5uy | 6,353 | load_dataset save_to_disk load_from_disk error | {
"login": "brisker",
"id": 13804492,
"node_id": "MDQ6VXNlcjEzODA0NDky",
"avatar_url": "https://avatars.githubusercontent.com/u/13804492?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/brisker",
"html_url": "https://github.com/brisker",
"followers_url": "https://api.github.com/users/briske... | [] | closed | false | null | [] | null | 4 | 2023-10-26T03:47:06 | 2024-02-02T15:23:14 | 2023-10-26T10:18:04 | NONE | null | ### Describe the bug
datasets version: 2.10.1
I `load_dataset `and `save_to_disk` sucessfully on windows10( **and I `load_from_disk(/LLM/data/wiki)` succcesfully on windows10**), and I copy the dataset `/LLM/data/wiki`
into a ubuntu system, but when I `load_from_disk(/LLM/data/wiki)` on ubuntu, something weird ha... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6353/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6353/timeline | null | completed | null | null | false | [
"solved.\r\nfsspec version problem",
"I'm using the latest datasets and fsspec , but still got this error!\r\n\r\ndatasets : Version: 2.13.0\r\n\r\nfsspec Version: 2023.10.0\r\n\r\n```\r\nFile \"/home/guoby/app/Anaconda3-2021.05/envs/news/lib/python3.8/site-packages/datasets/load.py\", line 1892, in load_from_... |
https://api.github.com/repos/huggingface/datasets/issues/6352 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6352/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6352/comments | https://api.github.com/repos/huggingface/datasets/issues/6352/events | https://github.com/huggingface/datasets/issues/6352 | 1,962,296,057 | I_kwDODunzps509kL5 | 6,352 | Error loading wikitext data raise NotImplementedError(f"Loading a dataset cached in a {type(self._fs).__name__} is not supported.") | {
"login": "Ahmed-Roushdy",
"id": 68569076,
"node_id": "MDQ6VXNlcjY4NTY5MDc2",
"avatar_url": "https://avatars.githubusercontent.com/u/68569076?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Ahmed-Roushdy",
"html_url": "https://github.com/Ahmed-Roushdy",
"followers_url": "https://api.githu... | [] | closed | false | null | [] | null | 13 | 2023-10-25T21:55:31 | 2024-03-19T16:46:22 | 2023-11-07T07:26:54 | NONE | null | I was trying to load the wiki dataset, but i got this error
traindata = load_dataset('wikitext', 'wikitext-2-raw-v1', split='train')
File "/home/aelkordy/.conda/envs/prune_llm/lib/python3.9/site-packages/datasets/load.py", line 1804, in load_dataset
ds = builder_instance.as_dataset(split=split, verific... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6352/reactions",
"total_count": 4,
"+1": 4,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6352/timeline | null | completed | null | null | false | [
"+1 \r\n```\r\nFound cached dataset csv (file:///home/ubuntu/.cache/huggingface/datasets/theSquarePond___csv/theSquarePond--XXXXX-bbf0a8365d693d2c/0.0.0/eea64c71ca8b46dd3f537ed218fc9bf495d5707789152eb2764f5c78fa66d59d)\r\n---------------------------------------------------------------------------\r\nNotImplementedE... |
https://api.github.com/repos/huggingface/datasets/issues/6351 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6351/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6351/comments | https://api.github.com/repos/huggingface/datasets/issues/6351/events | https://github.com/huggingface/datasets/pull/6351 | 1,961,982,988 | PR_kwDODunzps5dyMvh | 6,351 | Fix use_dataset.mdx | {
"login": "angel-luis",
"id": 17672548,
"node_id": "MDQ6VXNlcjE3NjcyNTQ4",
"avatar_url": "https://avatars.githubusercontent.com/u/17672548?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/angel-luis",
"html_url": "https://github.com/angel-luis",
"followers_url": "https://api.github.com/use... | [] | closed | false | null | [] | null | 2 | 2023-10-25T18:21:08 | 2023-10-26T17:19:49 | 2023-10-26T17:10:27 | CONTRIBUTOR | null | The current example isn't working because it can't find `labels` inside the Dataset object. So I've added an extra step to the process. Tested and working in Colab. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6351/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6351/timeline | null | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6351",
"html_url": "https://github.com/huggingface/datasets/pull/6351",
"diff_url": "https://github.com/huggingface/datasets/pull/6351.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6351.patch",
"merged_at": "2023-10-26T17:10... | true | [
"_The documentation is not available anymore as the PR was closed or merged._",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.