id int64 953M 3.35B | number int64 2.72k 7.75k | title stringlengths 1 290 | state stringclasses 2
values | created_at timestamp[s]date 2021-07-26 12:21:17 2025-08-23 00:18:43 | updated_at timestamp[s]date 2021-07-26 13:27:59 2025-08-23 12:34:39 | closed_at timestamp[s]date 2021-07-26 13:27:59 2025-08-20 16:35:55 ⌀ | html_url stringlengths 49 51 | pull_request dict | user_login stringlengths 3 26 | is_pull_request bool 2
classes | comments listlengths 0 30 |
|---|---|---|---|---|---|---|---|---|---|---|---|
1,938,797,389 | 6,298 | Doc readme improvements | closed | 2023-10-11T21:51:12 | 2023-10-12T12:47:15 | 2023-10-12T12:38:19 | https://github.com/huggingface/datasets/pull/6298 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6298",
"html_url": "https://github.com/huggingface/datasets/pull/6298",
"diff_url": "https://github.com/huggingface/datasets/pull/6298.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6298.patch",
"merged_at": "2023-10-12T12:38... | mariosasko | true | [
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a... |
1,938,752,707 | 6,297 | Fix ArrayXD cast | closed | 2023-10-11T21:14:59 | 2023-10-13T13:54:00 | 2023-10-13T13:45:30 | https://github.com/huggingface/datasets/pull/6297 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6297",
"html_url": "https://github.com/huggingface/datasets/pull/6297",
"diff_url": "https://github.com/huggingface/datasets/pull/6297.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6297.patch",
"merged_at": "2023-10-13T13:45... | mariosasko | true | [
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a... |
1,938,453,845 | 6,296 | Move `exceptions.py` to `utils/exceptions.py` | closed | 2023-10-11T18:28:00 | 2024-09-03T16:00:04 | 2024-09-03T16:00:03 | https://github.com/huggingface/datasets/pull/6296 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6296",
"html_url": "https://github.com/huggingface/datasets/pull/6296",
"diff_url": "https://github.com/huggingface/datasets/pull/6296.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6296.patch",
"merged_at": null
} | mariosasko | true | [
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a... |
1,937,362,102 | 6,295 | Fix parquet columns argument in streaming mode | closed | 2023-10-11T10:01:01 | 2023-10-11T16:30:24 | 2023-10-11T16:21:36 | https://github.com/huggingface/datasets/pull/6295 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6295",
"html_url": "https://github.com/huggingface/datasets/pull/6295",
"diff_url": "https://github.com/huggingface/datasets/pull/6295.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6295.patch",
"merged_at": "2023-10-11T16:21... | lhoestq | true | [
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a... |
1,937,359,605 | 6,294 | IndexError: Invalid key is out of bounds for size 0 despite having a populated dataset | closed | 2023-10-11T09:59:38 | 2023-10-17T11:24:06 | 2023-10-17T11:24:06 | https://github.com/huggingface/datasets/issues/6294 | null | ZYM66 | false | [
"It looks to be the same issue as the one reported in https://discuss.huggingface.co/t/indexerror-invalid-key-16-is-out-of-bounds-for-size-0.\r\n\r\nCan you check the length of `train_dataset` before the `train_sampler = self._get_train_sampler()` (and after `_remove_unused_columns`) line?"
] |
1,937,238,047 | 6,293 | Choose columns to stream parquet data in streaming mode | closed | 2023-10-11T08:59:36 | 2023-10-11T16:21:38 | 2023-10-11T16:21:38 | https://github.com/huggingface/datasets/issues/6293 | null | lhoestq | false | [] |
1,937,050,470 | 6,292 | how to load the image of dtype float32 or float64 | open | 2023-10-11T07:27:16 | 2023-10-11T13:19:11 | null | https://github.com/huggingface/datasets/issues/6292 | null | wanglaofei | false | [
"Hi! Can you provide a code that reproduces the issue?\r\n\r\nAlso, which version of `datasets` are you using? You can check this by running `python -c \"import datasets; print(datasets.__version__)\"` inside the env. We added support for \"float images\" in `datasets 2.9`."
] |
1,936,129,871 | 6,291 | Casting type from Array2D int to Array2D float crashes | closed | 2023-10-10T20:10:10 | 2023-10-13T13:45:31 | 2023-10-13T13:45:31 | https://github.com/huggingface/datasets/issues/6291 | null | AlanBlanchet | false | [
"Thanks for reporting! I've opened a PR with a fix"
] |
1,935,629,679 | 6,290 | Incremental dataset (e.g. `.push_to_hub(..., append=True)`) | open | 2023-10-10T15:18:03 | 2025-08-13T12:55:28 | null | https://github.com/huggingface/datasets/issues/6290 | null | Wauplin | false | [
"Yea I think waiting for #6269 would be best, or branching from it. For reference, this [PR](https://github.com/LAION-AI/Discord-Scrapers/pull/2) is progressing pretty well which will do similar using the hf hub for our LAION dataset bot https://github.com/LAION-AI/Discord-Scrapers/pull/2. ",
"Is there any update... |
1,935,628,506 | 6,289 | testing doc-builder | closed | 2023-10-10T15:17:29 | 2023-10-13T08:57:14 | 2023-10-13T08:56:48 | https://github.com/huggingface/datasets/pull/6289 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6289",
"html_url": "https://github.com/huggingface/datasets/pull/6289",
"diff_url": "https://github.com/huggingface/datasets/pull/6289.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6289.patch",
"merged_at": null
} | mishig25 | true | [
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a... |
1,935,005,457 | 6,288 | Dataset.from_pandas with a DataFrame of PIL.Images | open | 2023-10-10T10:29:16 | 2024-11-29T16:35:30 | null | https://github.com/huggingface/datasets/issues/6288 | null | lhoestq | false | [
"A duplicate of https://github.com/huggingface/datasets/issues/4796.\r\n\r\nWe could get this for free by implementing the `Image` feature as an extension type, as shown in [this](https://colab.research.google.com/drive/1Uzm_tXVpGTwbzleDConWcNjacwO1yxE4?usp=sharing) Colab (example with UUIDs).\r\n",
"+1 to this\r... |
1,932,758,192 | 6,287 | map() not recognizing "text" | closed | 2023-10-09T10:27:30 | 2023-10-11T20:28:45 | 2023-10-11T20:28:45 | https://github.com/huggingface/datasets/issues/6287 | null | EngineerKhan | false | [
"There is no \"text\" column in the `amazon_reviews_multi`, hence the `KeyError`. You can get the column names by running `dataset.column_names`."
] |
1,932,640,128 | 6,286 | Create DefunctDatasetError | closed | 2023-10-09T09:23:23 | 2023-10-10T07:13:22 | 2023-10-10T07:03:04 | https://github.com/huggingface/datasets/pull/6286 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6286",
"html_url": "https://github.com/huggingface/datasets/pull/6286",
"diff_url": "https://github.com/huggingface/datasets/pull/6286.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6286.patch",
"merged_at": "2023-10-10T07:03... | albertvillanova | true | [
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a... |
1,932,306,325 | 6,285 | TypeError: expected str, bytes or os.PathLike object, not dict | open | 2023-10-09T04:56:26 | 2023-10-10T13:17:33 | null | https://github.com/huggingface/datasets/issues/6285 | null | andysingal | false | [
"You should be able to load the images by modifying the `load_dataset` call like this:\r\n```python\r\ndataset = load_dataset(\"imagefolder\", data_dir=\"/content/datasets/PotholeDetectionYOLOv8-1\")\r\n```\r\n\r\nThe `imagefolder` builder expects the image files to be in `path/label/image_file` (e.g. .`.../train/d... |
1,929,551,712 | 6,284 | Add Belebele multiple-choice machine reading comprehension (MRC) dataset | closed | 2023-10-06T06:58:03 | 2023-10-06T13:26:51 | 2023-10-06T13:26:51 | https://github.com/huggingface/datasets/issues/6284 | null | rajveer43 | false | [
"This dataset is already available on the Hub: https://huggingface.co/datasets/facebook/belebele.\r\n"
] |
1,928,552,257 | 6,283 | Fix array cast/embed with null values | closed | 2023-10-05T15:24:05 | 2024-07-04T07:24:20 | 2024-02-06T19:24:19 | https://github.com/huggingface/datasets/pull/6283 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6283",
"html_url": "https://github.com/huggingface/datasets/pull/6283",
"diff_url": "https://github.com/huggingface/datasets/pull/6283.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6283.patch",
"merged_at": "2024-02-06T19:24... | mariosasko | true | [
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a... |
1,928,473,630 | 6,282 | Drop data_files duplicates | closed | 2023-10-05T14:43:08 | 2024-09-02T14:08:35 | 2024-09-02T14:08:35 | https://github.com/huggingface/datasets/pull/6282 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6282",
"html_url": "https://github.com/huggingface/datasets/pull/6282",
"diff_url": "https://github.com/huggingface/datasets/pull/6282.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6282.patch",
"merged_at": null
} | lhoestq | true | [
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a... |
1,928,456,959 | 6,281 | Improve documentation of dataset.from_generator | closed | 2023-10-05T14:34:49 | 2023-10-05T19:09:07 | 2023-10-05T18:57:41 | https://github.com/huggingface/datasets/pull/6281 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6281",
"html_url": "https://github.com/huggingface/datasets/pull/6281",
"diff_url": "https://github.com/huggingface/datasets/pull/6281.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6281.patch",
"merged_at": "2023-10-05T18:57... | hartmans | true | [
"I have looked at the doc failures, and I do not think that my change caused the doc build failure, but I'm not 100% sure about that.\r\nI have high confidence that the integration test failures are not something I introduced:-)",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<sum... |
1,928,215,278 | 6,280 | Couldn't cast array of type fixed_size_list to Sequence(Value(float64)) | closed | 2023-10-05T12:48:31 | 2024-02-06T19:24:20 | 2024-02-06T19:24:20 | https://github.com/huggingface/datasets/issues/6280 | null | jmif | false | [
"Thanks for reporting! I've opened a PR with a fix.",
"Thanks for the quick response @mariosasko! I just installed your branch via `poetry add 'git+https://github.com/huggingface/datasets#fix-array_values'` and I can confirm it works on the example provided.\r\n\r\nFollow up question for you, should `None`s be s... |
1,928,028,226 | 6,279 | Batched IterableDataset | open | 2023-10-05T11:12:49 | 2024-11-07T10:01:22 | null | https://github.com/huggingface/datasets/issues/6279 | null | lneukom | false | [
"This is exactly what I was looking for. It would also be very useful for me :-)",
"This issue is really smashing the selling point of HF datasets... The only workaround I've found so far is to create a customized IterableDataloader which improves the loading speed to some extent.\r\n\r\nFor example I've a HF dat... |
1,927,957,877 | 6,278 | No data files duplicates | closed | 2023-10-05T10:31:58 | 2024-01-11T06:32:49 | 2023-10-05T14:43:17 | https://github.com/huggingface/datasets/pull/6278 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6278",
"html_url": "https://github.com/huggingface/datasets/pull/6278",
"diff_url": "https://github.com/huggingface/datasets/pull/6278.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6278.patch",
"merged_at": null
} | lhoestq | true | [
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a... |
1,927,044,546 | 6,277 | FileNotFoundError: Couldn't find a module script at /content/paws-x/paws-x.py. Module 'paws-x' doesn't exist on the Hugging Face Hub either. | closed | 2023-10-04T22:01:25 | 2023-10-08T17:05:46 | 2023-10-08T17:05:46 | https://github.com/huggingface/datasets/issues/6277 | null | diegogonzalezc | false | [
"`evaluate.load(\"paws-x\", \"es\")` throws the error because there is no such metric in the `evaluate` lib.\r\n\r\nSo, this is unrelated to our lib."
] |
1,925,961,878 | 6,276 | I'm trying to fine tune the openai/whisper model from huggingface using jupyter notebook and i keep getting this error | open | 2023-10-04T11:03:41 | 2023-11-27T10:39:16 | null | https://github.com/huggingface/datasets/issues/6276 | null | valaofficial | false | [
"Since you are using Windows, maybe moving the `map` call inside `if __name__ == \"__main__\"` can fix the issue:\r\n```python\r\nif __name__ == \"__main__\":\r\n common_voice = common_voice.map(prepare_dataset, remove_columns=common_voice.column_names[\"train\"], num_proc=4)\r\n```\r\n\r\nOtherwise, the only s... |
1,921,354,680 | 6,275 | Would like to Contribute a dataset | closed | 2023-10-02T07:00:21 | 2023-10-10T16:27:54 | 2023-10-10T16:27:54 | https://github.com/huggingface/datasets/issues/6275 | null | vikas70607 | false | [
"Hi! The process of contributing a dataset is explained here: https://huggingface.co/docs/datasets/upload_dataset. Also, check https://huggingface.co/docs/datasets/image_dataset for a more detailed explanation of how to share an image dataset."
] |
1,921,036,328 | 6,274 | FileNotFoundError for dataset with multiple builder config | closed | 2023-10-01T23:45:56 | 2024-08-14T04:42:02 | 2023-10-02T20:09:38 | https://github.com/huggingface/datasets/issues/6274 | null | LouisChen15 | false | [
"Please tell me if the above info is not enough for solving the problem. I will then make my dataset public temporarily so that you can really reproduce the bug. ",
"Hi! \r\nCould you share how to solve this problem? \r\nI faced this same error. "
] |
1,920,922,260 | 6,273 | Broken Link to PubMed Abstracts dataset . | open | 2023-10-01T19:08:48 | 2024-04-28T02:30:42 | null | https://github.com/huggingface/datasets/issues/6273 | null | sameemqureshi | false | [
"This has already been reported in the HF Course repo (https://github.com/huggingface/course/issues/623).",
"@lhoestq @albertvillanova @lewtun I don't think we are allowed to host these data files on the Hub (due to DMCA), which means the only option is to use a different dataset in the course (and to re-record t... |
1,920,831,487 | 6,272 | Duplicate `data_files` when named `<split>/<split>.parquet` | closed | 2023-10-01T15:43:56 | 2024-03-15T15:22:05 | 2024-03-15T15:22:05 | https://github.com/huggingface/datasets/issues/6272 | null | lhoestq | false | [
"Also reported in https://github.com/huggingface/datasets/issues/6259",
"I think it's best to drop duplicates with a `set` (as a temporary fix) and improve the patterns when/if https://github.com/fsspec/filesystem_spec/pull/1382 gets merged. @lhoestq Do you have some other ideas?",
"Alternatively we could just... |
1,920,420,295 | 6,271 | Overwriting Split overwrites data but not metadata, corrupting dataset | closed | 2023-09-30T22:37:31 | 2023-10-16T13:30:50 | 2023-10-16T13:30:50 | https://github.com/huggingface/datasets/issues/6271 | null | govindrai | false | [] |
1,920,329,373 | 6,270 | Dataset.from_generator raises with sharded gen_args | closed | 2023-09-30T16:50:06 | 2023-10-11T20:29:12 | 2023-10-11T20:29:11 | https://github.com/huggingface/datasets/issues/6270 | null | hartmans | false | [
"`gen_kwargs` should be a `dict`, as stated in the docstring, but you are passing a `list`.\r\n\r\nSo, to fix the error, replace the list of dicts with a dict of lists (and slightly modify the generator function):\r\n```python\r\nfrom pathlib import Path\r\nimport datasets\r\n\r\ndef process_yaml(files):\r\n for... |
1,919,572,790 | 6,269 | Reduce the number of commits in `push_to_hub` | closed | 2023-09-29T16:22:31 | 2023-10-16T16:03:18 | 2023-10-16T13:30:46 | https://github.com/huggingface/datasets/pull/6269 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6269",
"html_url": "https://github.com/huggingface/datasets/pull/6269",
"diff_url": "https://github.com/huggingface/datasets/pull/6269.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6269.patch",
"merged_at": "2023-10-16T13:30... | mariosasko | true | [
"_The documentation is not available anymore as the PR was closed or merged._",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea... |
1,919,010,645 | 6,268 | Add repo_id to DatasetInfo | open | 2023-09-29T10:24:55 | 2023-10-01T15:29:45 | null | https://github.com/huggingface/datasets/pull/6268 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6268",
"html_url": "https://github.com/huggingface/datasets/pull/6268",
"diff_url": "https://github.com/huggingface/datasets/pull/6268.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6268.patch",
"merged_at": null
} | lhoestq | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6268). All of your documentation changes will be reflected on that endpoint.",
"In https://github.com/huggingface/datasets/issues/4129 we want to track the origin of a dataset, e.g. if it comes from multiple datasets.\r\n\r\nI ... |
1,916,443,262 | 6,267 | Multi label class encoding | open | 2023-09-27T22:48:08 | 2023-10-26T18:46:08 | null | https://github.com/huggingface/datasets/issues/6267 | null | jmif | false | [
"You can use a `Sequence(ClassLabel(...))` feature type to represent a list of labels, and `cast_column`/`cast` to perform the \"string to label\" conversion (`class_encode_column` does support nested fields), e.g., in your case:\r\n```python\r\nfrom datasets import Dataset, Sequence, ClassLabel\r\ndata = {\r\n ... |
1,916,334,394 | 6,266 | Use LibYAML with PyYAML if available | open | 2023-09-27T21:13:36 | 2023-09-28T14:29:24 | null | https://github.com/huggingface/datasets/pull/6266 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6266",
"html_url": "https://github.com/huggingface/datasets/pull/6266",
"diff_url": "https://github.com/huggingface/datasets/pull/6266.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6266.patch",
"merged_at": null
} | bryant1410 | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6266). All of your documentation changes will be reflected on that endpoint.",
"On Ubuntu, if `libyaml-dev` is installed, you can install PyYAML 6.0.1 with LibYAML with the following command (as it's automatically detected):\r\... |
1,915,651,566 | 6,265 | Remove `apache_beam` import in `BeamBasedBuilder._save_info` | closed | 2023-09-27T13:56:34 | 2023-09-28T18:34:02 | 2023-09-28T18:23:35 | https://github.com/huggingface/datasets/pull/6265 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6265",
"html_url": "https://github.com/huggingface/datasets/pull/6265",
"diff_url": "https://github.com/huggingface/datasets/pull/6265.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6265.patch",
"merged_at": "2023-09-28T18:23... | mariosasko | true | [
"_The documentation is not available anymore as the PR was closed or merged._",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea... |
1,914,958,781 | 6,264 | Temporarily pin tensorflow < 2.14.0 | closed | 2023-09-27T08:16:06 | 2023-09-27T08:45:24 | 2023-09-27T08:36:39 | https://github.com/huggingface/datasets/pull/6264 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6264",
"html_url": "https://github.com/huggingface/datasets/pull/6264",
"diff_url": "https://github.com/huggingface/datasets/pull/6264.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6264.patch",
"merged_at": "2023-09-27T08:36... | albertvillanova | true | [
"_The documentation is not available anymore as the PR was closed or merged._",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea... |
1,914,951,043 | 6,263 | CI is broken: ImportError: cannot import name 'context' from 'tensorflow.python' | closed | 2023-09-27T08:12:05 | 2023-09-27T08:36:40 | 2023-09-27T08:36:40 | https://github.com/huggingface/datasets/issues/6263 | null | albertvillanova | false | [] |
1,914,895,459 | 6,262 | Fix CI 404 errors | closed | 2023-09-27T07:40:18 | 2023-09-28T15:39:16 | 2023-09-28T15:30:40 | https://github.com/huggingface/datasets/pull/6262 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6262",
"html_url": "https://github.com/huggingface/datasets/pull/6262",
"diff_url": "https://github.com/huggingface/datasets/pull/6262.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6262.patch",
"merged_at": "2023-09-28T15:30... | albertvillanova | true | [
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a... |
1,913,813,178 | 6,261 | Can't load a dataset | closed | 2023-09-26T15:46:25 | 2023-10-05T10:23:23 | 2023-10-05T10:23:22 | https://github.com/huggingface/datasets/issues/6261 | null | joaopedrosdmm | false | [
"I believe is due to the fact that doesn't work with .tgz files.",
"`JourneyDB/JourneyDB` is a gated dataset, so this error means you are not authenticated to access it, either by using an invalid token or by not agreeing to the terms in the dialog on the dataset page.\r\n\r\n> I believe is due to the fact that d... |
1,912,593,466 | 6,260 | REUSE_DATASET_IF_EXISTS don't work | closed | 2023-09-26T03:02:16 | 2023-09-28T18:23:36 | 2023-09-28T18:23:36 | https://github.com/huggingface/datasets/issues/6260 | null | rangehow | false | [
"Hi! Unfortunately, the current behavior is to delete the downloaded data when this error happens. So, I've opened a PR that removes the problematic import to avoid losing data due to `apache_beam` not being installed (we host the preprocessed version of `natual_questions` on the HF GCS, so requiring `apache_beam` ... |
1,911,965,758 | 6,259 | Duplicated Rows When Loading Parquet Files from Root Directory with Subdirectories | closed | 2023-09-25T17:20:54 | 2024-03-15T15:22:04 | 2024-03-15T15:22:04 | https://github.com/huggingface/datasets/issues/6259 | null | MF-FOOM | false | [
"Thanks for reporting this issue! We should be able to avoid this by making our `glob` patterns more precise. In the meantime, you can load the dataset by directly assigning splits to the data files: \r\n```python\r\nfrom datasets import load_dataset\r\nds = load_dataset(\"parquet\", data_files={\"train\": \"testin... |
1,911,445,373 | 6,258 | [DOCS] Fix typo: Elasticsearch | closed | 2023-09-25T12:50:59 | 2023-09-26T14:55:35 | 2023-09-26T13:36:40 | https://github.com/huggingface/datasets/pull/6258 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6258",
"html_url": "https://github.com/huggingface/datasets/pull/6258",
"diff_url": "https://github.com/huggingface/datasets/pull/6258.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6258.patch",
"merged_at": "2023-09-26T13:36... | leemthompo | true | [
"_The documentation is not available anymore as the PR was closed or merged._",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea... |
1,910,741,044 | 6,257 | HfHubHTTPError - exceeded our hourly quotas for action: commit | closed | 2023-09-25T06:11:43 | 2023-10-16T13:30:49 | 2023-10-16T13:30:48 | https://github.com/huggingface/datasets/issues/6257 | null | yuvalkirstain | false | [
"how is your dataset structured? (file types, how many commits and files are you trying to push, etc)",
"I succeeded in uploading it after several attempts with an hour gap between each attempt (inconvenient but worked). The final dataset is [here](https://huggingface.co/datasets/yuvalkirstain/pickapic_v2), code ... |
1,910,275,199 | 6,256 | load_dataset() function's cache_dir does not seems to work | closed | 2023-09-24T15:34:06 | 2025-05-14T10:08:53 | 2024-10-08T15:45:18 | https://github.com/huggingface/datasets/issues/6256 | null | andyzhu | false | [
"Can you share the error message?\r\n\r\nAlso, it would help if you could check whether `huggingface_hub`'s download behaves the same:\r\n```python\r\nfrom huggingface_hub import snapshot_download\r\nsnapshot_download(\"trec\", repo_type=\"dataset\", cache_dir='/path/to/my/dir)\r\n```\r\n\r\nIn the next major relea... |
1,909,842,977 | 6,255 | Parallelize builder configs creation | closed | 2023-09-23T11:56:20 | 2024-01-11T06:32:34 | 2023-09-26T15:44:19 | https://github.com/huggingface/datasets/pull/6255 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6255",
"html_url": "https://github.com/huggingface/datasets/pull/6255",
"diff_url": "https://github.com/huggingface/datasets/pull/6255.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6255.patch",
"merged_at": null
} | lhoestq | true | [
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a... |
1,909,672,104 | 6,254 | Dataset.from_generator() cost much more time in vscode debugging mode then running mode | closed | 2023-09-23T02:07:26 | 2023-10-03T14:42:53 | 2023-10-03T14:42:53 | https://github.com/huggingface/datasets/issues/6254 | null | dontnet-wuenze | false | [
"Answered on the forum: https://discuss.huggingface.co/t/dataset-from-generator-cost-much-more-time-in-vscode-debugging-mode-then-running-mode/56005/2"
] |
1,906,618,910 | 6,253 | Check builder cls default config name in inspect | closed | 2023-09-21T10:15:32 | 2023-09-21T14:16:44 | 2023-09-21T14:08:00 | https://github.com/huggingface/datasets/pull/6253 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6253",
"html_url": "https://github.com/huggingface/datasets/pull/6253",
"diff_url": "https://github.com/huggingface/datasets/pull/6253.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6253.patch",
"merged_at": "2023-09-21T14:08... | lhoestq | true | [
"_The documentation is not available anymore as the PR was closed or merged._",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea... |
1,906,375,378 | 6,252 | exif_transpose not done to Image (PIL problem) | closed | 2023-09-21T08:11:46 | 2024-03-19T15:29:43 | 2024-03-19T15:29:43 | https://github.com/huggingface/datasets/issues/6252 | null | rhajou | false | [
"Indeed, it makes sense to do this by default. \r\n\r\nIn the meantime, you can use `.with_transform` to transpose the images when accessing them:\r\n\r\n```python\r\nimport PIL.ImageOps\r\n\r\ndef exif_transpose_transform(batch):\r\n batch[\"image\"] = [PIL.ImageOps.exif_transpose(image) for image in batch[\"imag... |
1,904,418,426 | 6,251 | Support streaming datasets with pyarrow.parquet.read_table | closed | 2023-09-20T08:07:02 | 2023-09-27T06:37:03 | 2023-09-27T06:26:24 | https://github.com/huggingface/datasets/pull/6251 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6251",
"html_url": "https://github.com/huggingface/datasets/pull/6251",
"diff_url": "https://github.com/huggingface/datasets/pull/6251.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6251.patch",
"merged_at": "2023-09-27T06:26... | albertvillanova | true | [
"_The documentation is not available anymore as the PR was closed or merged._",
"This function reads an entire Arrow table in one go, which is not ideal memory-wise, so I don't think we should encourage using this function, considering we want to keep RAM usage as low as possible in the streaming mode. \r\n\r\n(N... |
1,901,390,945 | 6,247 | Update create_dataset.mdx | closed | 2023-09-18T17:06:29 | 2023-09-19T18:51:49 | 2023-09-19T18:40:10 | https://github.com/huggingface/datasets/pull/6247 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6247",
"html_url": "https://github.com/huggingface/datasets/pull/6247",
"diff_url": "https://github.com/huggingface/datasets/pull/6247.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6247.patch",
"merged_at": "2023-09-19T18:40... | EswarDivi | true | [
"_The documentation is not available anymore as the PR was closed or merged._",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea... |
1,899,848,414 | 6,246 | Add new column to dataset | closed | 2023-09-17T16:59:48 | 2023-09-18T16:20:09 | 2023-09-18T16:20:09 | https://github.com/huggingface/datasets/issues/6246 | null | andysingal | false | [
"I think it's an issue with the code.\r\n\r\nSpecifically:\r\n```python\r\ndataset = dataset['train'].add_column(\"/workspace/data\", new_column)\r\n```\r\n\r\nNow `dataset` is the train set with a new column. \r\nTo fix this, you can do:\r\n\r\n```python\r\ndataset['train'] = dataset['train'].add_column(\"/workspa... |
1,898,861,422 | 6,244 | Add support for `fsspec>=2023.9.0` | closed | 2023-09-15T17:58:25 | 2023-09-26T15:41:38 | 2023-09-26T15:32:51 | https://github.com/huggingface/datasets/pull/6244 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6244",
"html_url": "https://github.com/huggingface/datasets/pull/6244",
"diff_url": "https://github.com/huggingface/datasets/pull/6244.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6244.patch",
"merged_at": "2023-09-26T15:32... | mariosasko | true | [
"_The documentation is not available anymore as the PR was closed or merged._",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea... |
1,898,532,784 | 6,243 | Fix cast from fixed size list to variable size list | closed | 2023-09-15T14:23:33 | 2023-09-19T18:02:21 | 2023-09-19T17:53:17 | https://github.com/huggingface/datasets/pull/6243 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6243",
"html_url": "https://github.com/huggingface/datasets/pull/6243",
"diff_url": "https://github.com/huggingface/datasets/pull/6243.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6243.patch",
"merged_at": "2023-09-19T17:53... | mariosasko | true | [
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a... |
1,896,899,123 | 6,242 | Data alteration when loading dataset with unspecified inner sequence length | closed | 2023-09-14T16:12:45 | 2023-09-19T17:53:18 | 2023-09-19T17:53:18 | https://github.com/huggingface/datasets/issues/6242 | null | qgallouedec | false | [
"While this issue may seem specific, it led to a silent problem in my workflow that took days to diagnose. If this feature is not intended to be supported, an error should be raised when encountering this configuration to prevent such issues.",
"Thanks for reporting! This is a MRE:\r\n\r\n```python\r\nimport pyar... |
1,896,429,694 | 6,241 | Remove unused global variables in `audio.py` | closed | 2023-09-14T12:06:32 | 2023-09-15T15:57:10 | 2023-09-15T15:46:07 | https://github.com/huggingface/datasets/pull/6241 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6241",
"html_url": "https://github.com/huggingface/datasets/pull/6241",
"diff_url": "https://github.com/huggingface/datasets/pull/6241.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6241.patch",
"merged_at": "2023-09-15T15:46... | mariosasko | true | [
"_The documentation is not available anymore as the PR was closed or merged._",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea... |
1,895,723,888 | 6,240 | Dataloader stuck on multiple GPUs | closed | 2023-09-14T05:30:30 | 2023-09-14T23:54:42 | 2023-09-14T23:54:42 | https://github.com/huggingface/datasets/issues/6240 | null | kuri54 | false | [
"What type of dataset are you using in this script? `torch.utils.data.Dataset` or `datasets.Dataset`? Please share the `datasets` package version if it's the latter. Otherwise, it's better to move this issue to the `accelerate` repo.",
"Very sorry, I thought I had a repo in `accelerate!`\r\nI will close this issu... |
1,895,349,382 | 6,239 | Load local audio data doesn't work | closed | 2023-09-13T22:30:01 | 2023-09-15T14:32:10 | 2023-09-15T14:32:10 | https://github.com/huggingface/datasets/issues/6239 | null | abodacs | false | [
"I think this is the same issue as https://github.com/huggingface/datasets/issues/4776. Maybe installing `ffmpeg` can fix it:\r\n```python\r\nadd-apt-repository -y ppa:savoury1/ffmpeg4\r\napt-get -qq install -y ffmpeg\r\n```\r\n\r\nHowever, the best solution is to use a newer version of `datasets`. In the recent re... |
1,895,207,828 | 6,238 | `dataset.filter` ALWAYS removes the first item from the dataset when using batched=True | closed | 2023-09-13T20:20:37 | 2023-09-17T07:05:07 | 2023-09-17T07:05:07 | https://github.com/huggingface/datasets/issues/6238 | null | Taytay | false | [
"`filter` treats the function's output as a (selection) mask - `True` keeps the sample, and `False` drops it. In your case, `bool(0)` evaluates to `False`, so dropping the first sample is the correct behavior.",
"Oh gosh! 🤦 I totally misunderstood the API! My apologies!"
] |
1,893,822,321 | 6,237 | Tokenization with multiple workers is too slow | closed | 2023-09-13T06:18:34 | 2023-09-19T21:54:58 | 2023-09-19T21:54:58 | https://github.com/huggingface/datasets/issues/6237 | null | macabdul9 | false | [
"[This](https://huggingface.co/docs/datasets/nlp_process#map) is the most performant way to tokenize a dataset (`batched=True, num_proc=None, return_tensors=\"np\"`) \r\n\r\nIf`tokenizer.is_fast` returns `True`, `num_proc` must be `None/1` to benefit from the fast tokenizers' parallelism (the fast tokenizers are im... |
1,893,648,480 | 6,236 | Support buffer shuffle for to_tf_dataset | open | 2023-09-13T03:19:44 | 2023-09-18T01:11:21 | null | https://github.com/huggingface/datasets/issues/6236 | null | EthanRock | false | [
"cc @Rocketknight1 ",
"Hey! You can implement this yourself, just:\r\n\r\n1) Create the dataset with `to_tf_dataset()` with `shuffle=False`\r\n2) Add an `unbatch()` at the end (or use batch_size=1)\r\n3) Add a `shuffle()` to the resulting dataset with your desired buffer size\r\n4) Add a `batch()` at the end agai... |
1,893,337,083 | 6,235 | Support multiprocessing for download/extract nestedly | open | 2023-09-12T21:51:08 | 2023-09-12T21:51:08 | null | https://github.com/huggingface/datasets/issues/6235 | null | hgt312 | false | [] |
1,891,804,286 | 6,233 | Update README.md | closed | 2023-09-12T06:53:06 | 2023-09-13T18:20:50 | 2023-09-13T18:10:04 | https://github.com/huggingface/datasets/pull/6233 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6233",
"html_url": "https://github.com/huggingface/datasets/pull/6233",
"diff_url": "https://github.com/huggingface/datasets/pull/6233.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6233.patch",
"merged_at": "2023-09-13T18:10... | NinoRisteski | true | [
"_The documentation is not available anymore as the PR was closed or merged._",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea... |
1,891,109,762 | 6,232 | Improve error message for missing function parameters | closed | 2023-09-11T19:11:58 | 2023-09-15T18:07:56 | 2023-09-15T17:59:02 | https://github.com/huggingface/datasets/pull/6232 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6232",
"html_url": "https://github.com/huggingface/datasets/pull/6232",
"diff_url": "https://github.com/huggingface/datasets/pull/6232.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6232.patch",
"merged_at": "2023-09-15T17:59... | suavemint | true | [
"_The documentation is not available anymore as the PR was closed or merged._",
"CI errors are unrelated",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_nu... |
1,890,863,249 | 6,231 | Overwrite legacy default config name in `dataset_infos.json` in packaged datasets | open | 2023-09-11T16:27:09 | 2023-09-26T11:19:36 | null | https://github.com/huggingface/datasets/pull/6231 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6231",
"html_url": "https://github.com/huggingface/datasets/pull/6231",
"diff_url": "https://github.com/huggingface/datasets/pull/6231.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6231.patch",
"merged_at": null
} | polinaeterna | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6231). All of your documentation changes will be reflected on that endpoint.",
"realized that this pr is still not merged, @lhoestq maybe you can take a look at it? ",
"I think https://github.com/huggingface/datasets/pull/621... |
1,890,521,006 | 6,230 | Don't skip hidden files in `dl_manager.iter_files` when they are given as input | closed | 2023-09-11T13:29:19 | 2023-09-13T18:21:28 | 2023-09-13T18:12:09 | https://github.com/huggingface/datasets/pull/6230 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6230",
"html_url": "https://github.com/huggingface/datasets/pull/6230",
"diff_url": "https://github.com/huggingface/datasets/pull/6230.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6230.patch",
"merged_at": "2023-09-13T18:12... | mariosasko | true | [
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a... |
1,889,050,954 | 6,229 | Apply inference on all images in the dataset | closed | 2023-09-10T08:36:12 | 2023-09-20T16:11:53 | 2023-09-20T16:11:52 | https://github.com/huggingface/datasets/issues/6229 | null | andysingal | false | [
"From what I see, `MMSegInferencer` supports NumPy arrays, so replace the line `image_path = example['image']` with `image_path = np.array(example['image'])` to fix the issue (`example[\"image\"]` is a `PIL.Image` object). ",
"> From what I see, `MMSegInferencer` supports NumPy arrays, so replace the line `image_... |
1,887,959,311 | 6,228 | Remove RGB -> BGR image conversion in Object Detection tutorial | closed | 2023-09-08T16:09:13 | 2023-09-08T18:02:49 | 2023-09-08T17:52:16 | https://github.com/huggingface/datasets/pull/6228 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6228",
"html_url": "https://github.com/huggingface/datasets/pull/6228",
"diff_url": "https://github.com/huggingface/datasets/pull/6228.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6228.patch",
"merged_at": "2023-09-08T17:52... | mariosasko | true | [
"_The documentation is not available anymore as the PR was closed or merged._",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea... |
1,887,462,591 | 6,226 | Add push_to_hub with multiple configs docs | closed | 2023-09-08T11:08:55 | 2023-09-08T12:29:21 | 2023-09-08T12:20:51 | https://github.com/huggingface/datasets/pull/6226 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6226",
"html_url": "https://github.com/huggingface/datasets/pull/6226",
"diff_url": "https://github.com/huggingface/datasets/pull/6226.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6226.patch",
"merged_at": "2023-09-08T12:20... | lhoestq | true | [
"_The documentation is not available anymore as the PR was closed or merged._",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea... |
1,887,054,320 | 6,225 | Conversion from RGB to BGR in Object Detection tutorial | closed | 2023-09-08T06:49:19 | 2023-09-08T17:52:18 | 2023-09-08T17:52:17 | https://github.com/huggingface/datasets/issues/6225 | null | samokhinv | false | [
"Good catch!"
] |
1,886,043,692 | 6,224 | Ignore `dataset_info.json` in data files resolution | closed | 2023-09-07T14:43:51 | 2023-09-07T15:46:10 | 2023-09-07T15:37:20 | https://github.com/huggingface/datasets/pull/6224 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6224",
"html_url": "https://github.com/huggingface/datasets/pull/6224",
"diff_url": "https://github.com/huggingface/datasets/pull/6224.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6224.patch",
"merged_at": "2023-09-07T15:37... | mariosasko | true | [
"_The documentation is not available anymore as the PR was closed or merged._",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea... |
1,885,710,696 | 6,223 | Update README.md | closed | 2023-09-07T11:33:20 | 2023-09-13T22:32:31 | 2023-09-13T22:23:42 | https://github.com/huggingface/datasets/pull/6223 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6223",
"html_url": "https://github.com/huggingface/datasets/pull/6223",
"diff_url": "https://github.com/huggingface/datasets/pull/6223.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6223.patch",
"merged_at": "2023-09-13T22:23... | NinoRisteski | true | [
"_The documentation is not available anymore as the PR was closed or merged._",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea... |
1,884,875,510 | 6,222 | fix typo in Audio dataset documentation | closed | 2023-09-06T23:17:24 | 2023-10-03T14:18:41 | 2023-09-07T15:39:09 | https://github.com/huggingface/datasets/pull/6222 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6222",
"html_url": "https://github.com/huggingface/datasets/pull/6222",
"diff_url": "https://github.com/huggingface/datasets/pull/6222.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6222.patch",
"merged_at": "2023-09-07T15:39... | prassanna-ravishankar | true | [
"_The documentation is not available anymore as the PR was closed or merged._",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea... |
1,884,324,631 | 6,221 | Support saving datasets with custom formatting | open | 2023-09-06T16:03:32 | 2023-09-06T18:32:07 | null | https://github.com/huggingface/datasets/issues/6221 | null | mariosasko | false | [
"Not a fan of pickling this sort of stuff either.\r\nNote that users can also share the code in their dataset documentation."
] |
1,884,285,980 | 6,220 | Set dev version | closed | 2023-09-06T15:40:33 | 2023-09-06T15:52:33 | 2023-09-06T15:41:13 | https://github.com/huggingface/datasets/pull/6220 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6220",
"html_url": "https://github.com/huggingface/datasets/pull/6220",
"diff_url": "https://github.com/huggingface/datasets/pull/6220.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6220.patch",
"merged_at": "2023-09-06T15:41... | albertvillanova | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6220). All of your documentation changes will be reflected on that endpoint.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchma... |
1,884,244,334 | 6,219 | Release: 2.14.5 | closed | 2023-09-06T15:17:10 | 2023-09-06T15:46:20 | 2023-09-06T15:18:51 | https://github.com/huggingface/datasets/pull/6219 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6219",
"html_url": "https://github.com/huggingface/datasets/pull/6219",
"diff_url": "https://github.com/huggingface/datasets/pull/6219.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6219.patch",
"merged_at": "2023-09-06T15:18... | albertvillanova | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6219). All of your documentation changes will be reflected on that endpoint.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchma... |
1,883,734,000 | 6,218 | Rename old push_to_hub configs to "default" in dataset_infos | closed | 2023-09-06T10:40:05 | 2023-09-07T08:31:29 | 2023-09-06T11:23:56 | https://github.com/huggingface/datasets/pull/6218 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6218",
"html_url": "https://github.com/huggingface/datasets/pull/6218",
"diff_url": "https://github.com/huggingface/datasets/pull/6218.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6218.patch",
"merged_at": "2023-09-06T11:23... | lhoestq | true | [
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a... |
1,883,614,607 | 6,217 | `Dataset.to_dict()` ignore `decode=True` with Image feature | open | 2023-09-06T09:26:16 | 2023-09-08T17:08:52 | null | https://github.com/huggingface/datasets/issues/6217 | null | qgallouedec | false | [
"We need to implement the `Image` type as a PyArrow extension type (to allow us to override the Python conversion) for this to work as expected. For now, it's best to use your approach indeed."
] |
1,883,492,703 | 6,216 | Release: 2.13.2 | closed | 2023-09-06T08:15:32 | 2023-09-06T08:52:18 | 2023-09-06T08:22:43 | https://github.com/huggingface/datasets/pull/6216 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6216",
"html_url": "https://github.com/huggingface/datasets/pull/6216",
"diff_url": "https://github.com/huggingface/datasets/pull/6216.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6216.patch",
"merged_at": "2023-09-06T08:22... | albertvillanova | true | [
"_The documentation is not available anymore as the PR was closed or merged._",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea... |
1,882,176,970 | 6,215 | Fix checking patterns to infer packaged builder | closed | 2023-09-05T15:10:47 | 2023-09-06T10:34:00 | 2023-09-06T10:25:00 | https://github.com/huggingface/datasets/pull/6215 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6215",
"html_url": "https://github.com/huggingface/datasets/pull/6215",
"diff_url": "https://github.com/huggingface/datasets/pull/6215.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6215.patch",
"merged_at": "2023-09-06T10:25... | polinaeterna | true | [
"_The documentation is not available anymore as the PR was closed or merged._",
"oh wow good catch",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy af... |
1,881,736,469 | 6,214 | Unpin fsspec < 2023.9.0 | closed | 2023-09-05T11:02:58 | 2023-09-26T15:32:52 | 2023-09-26T15:32:52 | https://github.com/huggingface/datasets/issues/6214 | null | albertvillanova | false | [] |
1,880,592,987 | 6,213 | Better list array values handling in cast/embed storage | closed | 2023-09-04T16:21:23 | 2024-01-11T06:32:20 | 2023-10-05T15:24:34 | https://github.com/huggingface/datasets/pull/6213 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6213",
"html_url": "https://github.com/huggingface/datasets/pull/6213",
"diff_url": "https://github.com/huggingface/datasets/pull/6213.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6213.patch",
"merged_at": null
} | mariosasko | true | [
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a... |
1,880,399,516 | 6,212 | Tilde (~) is not supported for data_files | open | 2023-09-04T14:23:49 | 2023-09-05T08:28:39 | null | https://github.com/huggingface/datasets/issues/6212 | null | exs-avianello | false | [
"Hi @exs-avianello, is it really needed? Note you can alternatively use `pathlib.Path` among others as it follows:\r\n\r\n```python\r\nimport datasets\r\nfrom pathlib import Path\r\n\r\n# save a parquet file at ~/path/to/data.parquet\r\n\r\ndata_files = Path.home() / \"path/to/data.parquet\"\r\ndataset = datasets.l... |
1,880,265,906 | 6,211 | Fix empty splitinfo json | closed | 2023-09-04T13:13:53 | 2023-09-04T14:58:34 | 2023-09-04T14:47:17 | https://github.com/huggingface/datasets/pull/6211 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6211",
"html_url": "https://github.com/huggingface/datasets/pull/6211",
"diff_url": "https://github.com/huggingface/datasets/pull/6211.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6211.patch",
"merged_at": "2023-09-04T14:47... | lhoestq | true | [
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a... |
1,879,649,731 | 6,210 | Temporarily pin fsspec < 2023.9.0 | closed | 2023-09-04T07:07:07 | 2023-09-04T07:40:23 | 2023-09-04T07:30:00 | https://github.com/huggingface/datasets/pull/6210 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6210",
"html_url": "https://github.com/huggingface/datasets/pull/6210",
"diff_url": "https://github.com/huggingface/datasets/pull/6210.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6210.patch",
"merged_at": "2023-09-04T07:30... | albertvillanova | true | [
"_The documentation is not available anymore as the PR was closed or merged._",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea... |
1,879,622,000 | 6,209 | CI is broken with AssertionError: 3 failed, 12 errors | closed | 2023-09-04T06:47:05 | 2023-09-04T07:30:01 | 2023-09-04T07:30:01 | https://github.com/huggingface/datasets/issues/6209 | null | albertvillanova | false | [] |
1,879,572,646 | 6,208 | Do not filter out .zip extensions from no-script datasets | closed | 2023-09-04T06:07:12 | 2023-09-04T09:22:19 | 2023-09-04T09:13:32 | https://github.com/huggingface/datasets/pull/6208 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6208",
"html_url": "https://github.com/huggingface/datasets/pull/6208",
"diff_url": "https://github.com/huggingface/datasets/pull/6208.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6208.patch",
"merged_at": "2023-09-04T09:13... | albertvillanova | true | [
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a... |
1,879,555,234 | 6,207 | No-script datasets with ZIP files do not load | closed | 2023-09-04T05:50:27 | 2023-09-04T09:13:33 | 2023-09-04T09:13:33 | https://github.com/huggingface/datasets/issues/6207 | null | albertvillanova | false | [] |
1,879,473,745 | 6,206 | When calling load_dataset, raise error: pyarrow.lib.ArrowInvalid: offset overflow while concatenating arrays | closed | 2023-09-04T04:14:00 | 2024-04-17T15:53:29 | 2023-09-04T06:05:49 | https://github.com/huggingface/datasets/issues/6206 | null | aihao2000 | false | [
"I solved the problem by modifying the \"self DEFAULT_WRITER_BATCH_SIZE\" in \"class MyDataset (datasets. GeneratorBasedBuilder) : __init__\"",
"same problem, and this solution worked me also - you can set this var by setting the keyword argument `writer_batch_size=...` in `load_dataset(...,writer_batch_size=...)... |
1,877,491,602 | 6,203 | Support loading from a DVC remote repository | closed | 2023-09-01T14:04:52 | 2023-09-15T15:11:27 | 2023-09-15T15:11:27 | https://github.com/huggingface/datasets/issues/6203 | null | bilelomrani1 | false | [
"(cross-posting from the linked DVC issue)\r\n\r\nI think this should already work out of the box with the current `datasets` and `dvc.api` releases by passing the correct `storage_options` into the datasets calls. `storage_options` is essentially just the kwargs dict that gets passed to the fsspec fs constructor.\... |
1,876,630,351 | 6,202 | avoid downgrading jax version | closed | 2023-09-01T02:57:57 | 2023-10-12T16:28:59 | 2023-10-12T16:28:59 | https://github.com/huggingface/datasets/issues/6202 | null | chrisflesher | false | [
"https://github.com/huggingface/datasets/blob/main/setup.py#L236\r\nCurrently has the highest version at 0.3.25; Not sure if there is any reason for this, other than that was the tested version?"
] |
1,875,256,775 | 6,201 | Fix to_json ValueError and remove pandas pin | closed | 2023-08-31T10:38:08 | 2023-09-05T11:07:07 | 2023-09-05T10:58:21 | https://github.com/huggingface/datasets/pull/6201 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6201",
"html_url": "https://github.com/huggingface/datasets/pull/6201",
"diff_url": "https://github.com/huggingface/datasets/pull/6201.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6201.patch",
"merged_at": "2023-09-05T10:58... | albertvillanova | true | [
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a... |
1,875,169,551 | 6,200 | Temporarily pin pandas < 2.1.0 | closed | 2023-08-31T09:45:17 | 2023-08-31T10:33:24 | 2023-08-31T10:24:38 | https://github.com/huggingface/datasets/pull/6200 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6200",
"html_url": "https://github.com/huggingface/datasets/pull/6200",
"diff_url": "https://github.com/huggingface/datasets/pull/6200.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6200.patch",
"merged_at": "2023-08-31T10:24... | albertvillanova | true | [
"_The documentation is not available anymore as the PR was closed or merged._",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea... |
1,875,165,185 | 6,199 | Use load_dataset for local json files, but it not works | open | 2023-08-31T09:42:34 | 2023-08-31T19:05:07 | null | https://github.com/huggingface/datasets/issues/6199 | null | Garen-in-bush | false | [
"Hugging Face's datasets library may prioritize remote configurations. Make sure there are no conflicting configurations causing the library to prefer downloading data\r\nMay be try debugging\r\nraw_datasets = load_dataset('json', data_files=data_files)\r\nprint(raw_datasets)\r\n",
"It doesn't download them but ... |
1,875,092,027 | 6,198 | Preserve split order in DataFilesDict | closed | 2023-08-31T09:00:26 | 2023-08-31T13:57:31 | 2023-08-31T13:48:42 | https://github.com/huggingface/datasets/pull/6198 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6198",
"html_url": "https://github.com/huggingface/datasets/pull/6198",
"diff_url": "https://github.com/huggingface/datasets/pull/6198.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6198.patch",
"merged_at": "2023-08-31T13:48... | albertvillanova | true | [
"_The documentation is not available anymore as the PR was closed or merged._",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea... |
1,875,078,155 | 6,197 | ValueError: 'index=True' is only valid when 'orient' is 'split', 'table', 'index', or 'columns' | closed | 2023-08-31T08:51:50 | 2023-09-01T10:35:10 | 2023-08-31T10:24:40 | https://github.com/huggingface/datasets/issues/6197 | null | exs-avianello | false | [
"Thanks for reporting. We are investigating it.",
"This issue is caused by latest `pandas` release 2.1.0 (released yesterday Aug 30).\r\n\r\nSee: https://github.com/huggingface/datasets/actions/runs/6035484010/job/16375932085?pr=6198\r\n",
"People using previous releases of `datasets` should pin `pandas` in the... |
1,875,070,972 | 6,196 | Split order is not preserved | closed | 2023-08-31T08:47:16 | 2023-08-31T13:48:43 | 2023-08-31T13:48:43 | https://github.com/huggingface/datasets/issues/6196 | null | albertvillanova | false | [] |
1,874,195,585 | 6,195 | Force to reuse cache at given path | closed | 2023-08-30T18:44:54 | 2023-11-03T10:14:21 | 2023-08-30T19:00:45 | https://github.com/huggingface/datasets/issues/6195 | null | Luosuu | false | [
"realized that need to pass the path at `cache_file_name` like\r\n\r\n```python\r\ntokenized_datasets = raw_datasets[\"train\"].map(\r\n tokenize_function,\r\n batched=True,\r\n num_proc=data_args.preprocessing_num_workers,\r\n remove_columns=[text_column_... |
1,872,598,223 | 6,194 | Support custom fingerprinting with `Dataset.from_generator` | open | 2023-08-29T22:43:13 | 2024-12-22T01:14:39 | null | https://github.com/huggingface/datasets/issues/6194 | null | bilelomrani1 | false | [
"The `fingerprint` parameter serves a slightly different purpose - we use it to inject a new fingerprint after transforming a `Dataset` (computed from the previous fingerprint + transform + transform args), e.g., to be able to compute the cache file for a transform. There is no concept of `fingerprint` before a `Da... |
1,872,285,153 | 6,193 | Dataset loading script method does not work with .pyc file | open | 2023-08-29T19:35:06 | 2023-08-31T19:47:29 | null | https://github.com/huggingface/datasets/issues/6193 | null | riteshkumarumassedu | false | [
"Before dynamically loading `.py` scripts with `importlib.import_module`, we also parse their contents to check imports, which is tricky to implement for binary `.pyc` files (requires parsing bytecode), so I don't think this is something we want to support (unless more users request it ofc) as this use case is a bi... |
1,871,911,640 | 6,192 | Set minimal fsspec version requirement to 2023.1.0 | closed | 2023-08-29T15:23:41 | 2023-08-30T14:01:56 | 2023-08-30T13:51:32 | https://github.com/huggingface/datasets/pull/6192 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6192",
"html_url": "https://github.com/huggingface/datasets/pull/6192",
"diff_url": "https://github.com/huggingface/datasets/pull/6192.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6192.patch",
"merged_at": "2023-08-30T13:51... | mariosasko | true | [
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a... |
1,871,634,840 | 6,191 | Add missing `revision` argument | closed | 2023-08-29T13:05:04 | 2023-09-04T06:38:17 | 2023-08-31T13:50:00 | https://github.com/huggingface/datasets/pull/6191 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6191",
"html_url": "https://github.com/huggingface/datasets/pull/6191",
"diff_url": "https://github.com/huggingface/datasets/pull/6191.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6191.patch",
"merged_at": "2023-08-31T13:50... | qgallouedec | true | [
"_The documentation is not available anymore as the PR was closed or merged._",
"I have found the same issue. Good fix. Should be merged as soon as possible.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_a... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.