Dataset Preview
Duplicate
The full dataset viewer is not available (click to read why). Only showing a preview of the rows.
The dataset generation failed
Error code:   DatasetGenerationError
Exception:    TypeError
Message:      Couldn't cast array of type timestamp[s] to null
Traceback:    Traceback (most recent call last):
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 2011, in _prepare_split_single
                  writer.write_table(table)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 585, in write_table
                  pa_table = table_cast(pa_table, self._schema)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 2302, in table_cast
                  return cast_table_to_schema(table, schema)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 2261, in cast_table_to_schema
                  arrays = [cast_array_to_feature(table[name], feature) for name, feature in features.items()]
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 2261, in <listcomp>
                  arrays = [cast_array_to_feature(table[name], feature) for name, feature in features.items()]
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 1802, in wrapper
                  return pa.chunked_array([func(chunk, *args, **kwargs) for chunk in array.chunks])
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 1802, in <listcomp>
                  return pa.chunked_array([func(chunk, *args, **kwargs) for chunk in array.chunks])
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 2020, in cast_array_to_feature
                  arrays = [_c(array.field(name), subfeature) for name, subfeature in feature.items()]
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 2020, in <listcomp>
                  arrays = [_c(array.field(name), subfeature) for name, subfeature in feature.items()]
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 1804, in wrapper
                  return func(array, *args, **kwargs)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 2116, in cast_array_to_feature
                  return array_cast(
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 1804, in wrapper
                  return func(array, *args, **kwargs)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 1962, in array_cast
                  raise TypeError(f"Couldn't cast array of type {_short_str(array.type)} to {_short_str(pa_type)}")
              TypeError: Couldn't cast array of type timestamp[s] to null
              
              The above exception was the direct cause of the following exception:
              
              Traceback (most recent call last):
                File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 1529, in compute_config_parquet_and_info_response
                  parquet_operations = convert_to_parquet(builder)
                File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 1154, in convert_to_parquet
                  builder.download_and_prepare(
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1027, in download_and_prepare
                  self._download_and_prepare(
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1122, in _download_and_prepare
                  self._prepare_split(split_generator, **prepare_split_kwargs)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1882, in _prepare_split
                  for job_id, done, content in self._prepare_split_single(
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 2038, in _prepare_split_single
                  raise DatasetGenerationError("An error occurred while generating the dataset") from e
              datasets.exceptions.DatasetGenerationError: An error occurred while generating the dataset

Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.

url
string
repository_url
string
labels_url
string
comments_url
string
events_url
string
html_url
string
id
int64
node_id
string
number
int64
title
string
user
dict
labels
sequence
state
string
locked
bool
assignee
null
assignees
sequence
milestone
null
comments
sequence
created_at
int64
updated_at
int64
closed_at
null
author_association
string
active_lock_reason
null
body
string
reactions
dict
timeline_url
string
performed_via_github_app
null
state_reason
null
draft
null
pull_request
null
is_pull_request
bool
https://api.github.com/repos/huggingface/datasets/issues/6542
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6542/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6542/comments
https://api.github.com/repos/huggingface/datasets/issues/6542/events
https://github.com/huggingface/datasets/issues/6542
2,059,198,575
I_kwDODunzps56vOBv
6,542
Datasets : wikipedia 20220301.en error
{ "avatar_url": "https://avatars.githubusercontent.com/u/53203620?v=4", "events_url": "https://api.github.com/users/ppx666/events{/privacy}", "followers_url": "https://api.github.com/users/ppx666/followers", "following_url": "https://api.github.com/users/ppx666/following{/other_user}", "gists_url": "https://a...
[]
open
false
null
[]
null
[ "Hi ! We now recommend using the `wikimedia/wikipedia` dataset, can you try loading this one instead ?\r\n\r\n```python\r\nwiki_dataset = load_dataset(\"wikimedia/wikipedia\", \"20231101.en\")\r\n```" ]
1,703,838,891,000
1,703,838,891,000
null
NONE
null
### Describe the bug When I used load_dataset to download this data set, the following error occurred. The main problem was that the target data did not exist. ### Steps to reproduce the bug 1.I tried downloading directly. ```python wiki_dataset = load_dataset("wikipedia", "20220301.en") ``` An exception occurre...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6542/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6542/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6541
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6541/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6541/comments
https://api.github.com/repos/huggingface/datasets/issues/6541/events
https://github.com/huggingface/datasets/issues/6541
2,058,983,826
I_kwDODunzps56uZmS
6,541
Dataset not loading successfully.
{ "avatar_url": "https://avatars.githubusercontent.com/u/93595990?v=4", "events_url": "https://api.github.com/users/hi-sushanta/events{/privacy}", "followers_url": "https://api.github.com/users/hi-sushanta/followers", "following_url": "https://api.github.com/users/hi-sushanta/following{/other_user}", "gists_u...
[]
open
false
null
[]
null
[]
1,703,813,747,000
1,703,813,747,000
null
NONE
null
### Describe the bug When I run down the below code shows this error: AttributeError: module 'numpy' has no attribute '_no_nep50_warning' I also added this issue in transformers library please check out: [link](https://github.com/huggingface/transformers/issues/28099) ### Steps to reproduce the bug ## Reproduction ...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6541/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6541/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6540
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6540/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6540/comments
https://api.github.com/repos/huggingface/datasets/issues/6540/events
https://github.com/huggingface/datasets/issues/6540
2,058,965,157
I_kwDODunzps56uVCl
6,540
Extreme inefficiency for `save_to_disk` when merging datasets
{ "avatar_url": "https://avatars.githubusercontent.com/u/43512683?v=4", "events_url": "https://api.github.com/users/KatarinaYuan/events{/privacy}", "followers_url": "https://api.github.com/users/KatarinaYuan/followers", "following_url": "https://api.github.com/users/KatarinaYuan/following{/other_user}", "gist...
[]
open
false
null
[]
null
[ "Concatenating datasets doesn't create any indices mapping - so flattening indices is not needed (unless you shuffle the dataset).\r\nCan you share the snippet of code you are using to merge your datasets and save them to disk ?" ]
1,703,810,675,000
1,703,810,675,000
null
NONE
null
### Describe the bug Hi, I tried to merge in total 22M sequences of data, where each sequence is of maximum length 2000. I found that merging these datasets and then `save_to_disk` is extremely slow because of flattening the indices. Wondering if you have any suggestions or guidance on this. Thank you very much! ###...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6540/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6540/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6539
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6539/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6539/comments
https://api.github.com/repos/huggingface/datasets/issues/6539/events
https://github.com/huggingface/datasets/issues/6539
2,058,493,960
I_kwDODunzps56siAI
6,539
'Repo card metadata block was not found' when loading a pragmeval dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/3647577?v=4", "events_url": "https://api.github.com/users/lambdaofgod/events{/privacy}", "followers_url": "https://api.github.com/users/lambdaofgod/followers", "following_url": "https://api.github.com/users/lambdaofgod/following{/other_user}", "gists_ur...
[]
open
false
null
[]
null
[]
1,703,773,105,000
1,703,773,117,000
null
NONE
null
### Describe the bug I can't load dataset subsets of 'pragmeval'. The funny thing is I ran the dataset author's [colab notebook](https://colab.research.google.com/drive/1sg--LF4z7XR1wxAOfp0-3d4J6kQ9nj_A?usp=sharing) and it works just fine. I tried to install exactly the same packages that are installed on colab usi...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6539/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6539/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6538
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6538/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6538/comments
https://api.github.com/repos/huggingface/datasets/issues/6538/events
https://github.com/huggingface/datasets/issues/6538
2,057,377,630
I_kwDODunzps56oRde
6,538
ImportError: cannot import name 'SchemaInferenceError' from 'datasets.arrow_writer' (/opt/conda/lib/python3.10/site-packages/datasets/arrow_writer.py)
{ "avatar_url": "https://avatars.githubusercontent.com/u/131662185?v=4", "events_url": "https://api.github.com/users/Sonali-Behera-TRT/events{/privacy}", "followers_url": "https://api.github.com/users/Sonali-Behera-TRT/followers", "following_url": "https://api.github.com/users/Sonali-Behera-TRT/following{/other...
[]
open
false
null
[]
null
[ "Hi ! Are you sure you have `datasets` 2.16 ? I just checked and on 2.16 I can run `from datasets.arrow_writer import SchemaInferenceError` without error", "I have the same issue - using with datasets version 2.16.1. Also this is on a kaggle notebook - other people with the same issue also seem to be having it on...
1,703,683,876,000
1,703,683,956,000
null
NONE
null
### Describe the bug While importing from packages getting the error Code: `import os import torch from datasets import load_dataset, Dataset from transformers import ( AutoModelForCausalLM, AutoTokenizer, BitsAndBytesConfig, HfArgumentParser, TrainingArguments, pipeline, loggin...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6538/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6538/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6537
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6537/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6537/comments
https://api.github.com/repos/huggingface/datasets/issues/6537/events
https://github.com/huggingface/datasets/issues/6537
2,057,132,173
I_kwDODunzps56nViN
6,537
Adding support for netCDF (*.nc) files
{ "avatar_url": "https://avatars.githubusercontent.com/u/12627125?v=4", "events_url": "https://api.github.com/users/shermansiu/events{/privacy}", "followers_url": "https://api.github.com/users/shermansiu/followers", "following_url": "https://api.github.com/users/shermansiu/following{/other_user}", "gists_url"...
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
open
false
null
[]
null
[ "Related to #3113 ", "Conceptually, we can use xarray to load the netCDF file, then xarray -> pandas -> pyarrow.", "I'd still need to verify that such a conversion would be lossless, especially for multi-dimensional data." ]
1,703,669,249,000
1,703,710,013,000
null
NONE
null
### Feature request netCDF (*.nc) is a file format for storing multidimensional scientific data, which is used by packages like `xarray` (labelled multi-dimensional arrays in Python). It would be nice to have native support for netCDF in `datasets`. ### Motivation When uploading *.nc files onto Huggingface Hub throu...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6537/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6537/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6536
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6536/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6536/comments
https://api.github.com/repos/huggingface/datasets/issues/6536/events
https://github.com/huggingface/datasets/issues/6536
2,056,863,239
I_kwDODunzps56mT4H
6,536
datasets.load_dataset raises FileNotFoundError for datasets==2.16.0
{ "avatar_url": "https://avatars.githubusercontent.com/u/46237844?v=4", "events_url": "https://api.github.com/users/ArvinZhuang/events{/privacy}", "followers_url": "https://api.github.com/users/ArvinZhuang/followers", "following_url": "https://api.github.com/users/ArvinZhuang/following{/other_user}", "gists_u...
[]
open
false
null
[]
null
[ "Hi ! Thanks for reporting\r\n\r\nThis is a bug in 2.16.0 for some datasets when `cache_dir` is a relative path. I opened https://github.com/huggingface/datasets/pull/6543 to fix this", "We just released 2.16.1 with a fix:\r\n\r\n```\r\npip install -U datasets\r\n```" ]
1,703,646,948,000
1,703,647,275,000
null
NONE
null
### Describe the bug Seems `datasets.load_dataset` raises FileNotFoundError for some hub datasets with the latest `datasets==2.16.0` ### Steps to reproduce the bug For example `pip install datasets==2.16.0` then ```python import datasets datasets.load_dataset("wentingzhao/anthropic-hh-first-prompt", cache_di...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6536/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6536/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6535
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6535/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6535/comments
https://api.github.com/repos/huggingface/datasets/issues/6535/events
https://github.com/huggingface/datasets/issues/6535
2,056,264,339
I_kwDODunzps56kBqT
6,535
IndexError: Invalid key: 47682 is out of bounds for size 0 while using PEFT
{ "avatar_url": "https://avatars.githubusercontent.com/u/57484266?v=4", "events_url": "https://api.github.com/users/MahavirDabas18/events{/privacy}", "followers_url": "https://api.github.com/users/MahavirDabas18/followers", "following_url": "https://api.github.com/users/MahavirDabas18/following{/other_user}", ...
[]
open
false
null
[]
null
[ "@sabman @pvl @kashif @vigsterkr " ]
1,703,585,673,000
1,703,585,756,000
null
NONE
null
### Describe the bug I am trying to fine-tune the t5 model on the paraphrasing task. While running the same code without- model = get_peft_model(model, config) the model trains without any issues. However, using the model returned from get_peft_model raises the following error due to datasets- IndexError: Inv...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6535/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6535/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6534
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6534/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6534/comments
https://api.github.com/repos/huggingface/datasets/issues/6534/events
https://github.com/huggingface/datasets/issues/6534
2,056,002,548
I_kwDODunzps56jBv0
6,534
How to configure multiple folders in the same zip package
{ "avatar_url": "https://avatars.githubusercontent.com/u/12895488?v=4", "events_url": "https://api.github.com/users/d710055071/events{/privacy}", "followers_url": "https://api.github.com/users/d710055071/followers", "following_url": "https://api.github.com/users/d710055071/following{/other_user}", "gists_url"...
[]
open
false
null
[]
null
[ "@albertvillanova" ]
1,703,562,980,000
1,703,572,276,000
null
CONTRIBUTOR
null
How should I write "config" in readme when all the data, such as train test, is in a zip file train floder and test floder in data.zip
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6534/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6534/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6533
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6533/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6533/comments
https://api.github.com/repos/huggingface/datasets/issues/6533/events
https://github.com/huggingface/datasets/issues/6533
2,055,929,101
I_kwDODunzps56iv0N
6,533
ted_talks_iwslt | Error: Config name is missing
{ "avatar_url": "https://avatars.githubusercontent.com/u/35850903?v=4", "events_url": "https://api.github.com/users/rayliuca/events{/privacy}", "followers_url": "https://api.github.com/users/rayliuca/followers", "following_url": "https://api.github.com/users/rayliuca/following{/other_user}", "gists_url": "htt...
[]
open
false
null
[]
null
[ "Hi ! Thanks for reporting. I opened https://github.com/huggingface/datasets/pull/6544 to fix this", "We just released 2.16.1 with a fix:\r\n\r\n```\r\npip install -U datasets\r\n```" ]
1,703,551,098,000
1,703,551,098,000
null
NONE
null
### Describe the bug Running load_dataset using the newest `datasets` library like below on the ted_talks_iwslt using year pair data will throw an error "Config name is missing" see also: https://huggingface.co/datasets/ted_talks_iwslt/discussions/3 likely caused by #6493, where the `and not config_kwargs` part...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6533/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6533/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6532
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6532/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6532/comments
https://api.github.com/repos/huggingface/datasets/issues/6532/events
https://github.com/huggingface/datasets/issues/6532
2,055,631,201
I_kwDODunzps56hnFh
6,532
[Feature request] Indexing datasets by a customly-defined id field to enable random access dataset items via the id
{ "avatar_url": "https://avatars.githubusercontent.com/u/3377221?v=4", "events_url": "https://api.github.com/users/Yu-Shi/events{/privacy}", "followers_url": "https://api.github.com/users/Yu-Shi/followers", "following_url": "https://api.github.com/users/Yu-Shi/following{/other_user}", "gists_url": "https://ap...
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
open
false
null
[]
null
[]
1,703,504,230,000
1,703,504,847,000
null
NONE
null
### Feature request Some datasets may contain an id-like field, for example the `id` field in [wikimedia/wikipedia](https://huggingface.co/datasets/wikimedia/wikipedia) and the `_id` field in [BeIR/dbpedia-entity](https://huggingface.co/datasets/BeIR/dbpedia-entity). HF datasets support efficient random access via r...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6532/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6532/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6531
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6531/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6531/comments
https://api.github.com/repos/huggingface/datasets/issues/6531/events
https://github.com/huggingface/datasets/pull/6531
2,055,201,605
PR_kwDODunzps5it5Sm
6,531
Add polars compatibility
{ "avatar_url": "https://avatars.githubusercontent.com/u/11325244?v=4", "events_url": "https://api.github.com/users/psmyth94/events{/privacy}", "followers_url": "https://api.github.com/users/psmyth94/followers", "following_url": "https://api.github.com/users/psmyth94/following{/other_user}", "gists_url": "htt...
[]
open
false
null
[]
null
[]
1,703,448,203,000
1,703,448,203,000
null
NONE
null
Hey there, I've just finished adding support to convert and format to `polars.DataFrame`. This was in response to the open issue about integrating Polars [#3334](https://github.com/huggingface/datasets/issues/3334). Datasets can be switched to Polars format via `Dataset.set_format("polars")`. I've also included `to_...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6531/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6531/timeline
null
null
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6530
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6530/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6530/comments
https://api.github.com/repos/huggingface/datasets/issues/6530/events
https://github.com/huggingface/datasets/issues/6530
2,054,817,609
I_kwDODunzps56egdJ
6,530
Impossible to save a mapped dataset to disk
{ "avatar_url": "https://avatars.githubusercontent.com/u/17604849?v=4", "events_url": "https://api.github.com/users/kopyl/events{/privacy}", "followers_url": "https://api.github.com/users/kopyl/followers", "following_url": "https://api.github.com/users/kopyl/following{/other_user}", "gists_url": "https://api....
[]
open
false
null
[]
null
[ "I solved it with `train_dataset.with_format(None)`\r\nBut then faced some more issues (which i later solved too).\r\n\r\nHuggingface does not seem to care, so I do. Here is an updated training script which saves a pre-processed (mapped) dataset to your local directory if you specify `--save_precomputed_data_dir=DI...
1,703,344,707,000
1,703,410,830,000
null
NONE
null
### Describe the bug I want to play around with different hyperparameters when training but don't want to re-map my dataset with 3 million samples each time for tens of hours when I [fully fine-tune SDXL](https://github.com/huggingface/diffusers/blob/main/examples/text_to_image/train_text_to_image_sdxl.py). After...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6530/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6530/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6529
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6529/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6529/comments
https://api.github.com/repos/huggingface/datasets/issues/6529/events
https://github.com/huggingface/datasets/issues/6529
2,054,209,449
I_kwDODunzps56cL-p
6,529
Impossible to only download a test split
{ "avatar_url": "https://avatars.githubusercontent.com/u/28439529?v=4", "events_url": "https://api.github.com/users/ysig/events{/privacy}", "followers_url": "https://api.github.com/users/ysig/followers", "following_url": "https://api.github.com/users/ysig/following{/other_user}", "gists_url": "https://api.git...
[]
open
false
null
[]
null
[ "The only way right now is to load with streaming=True" ]
1,703,264,192,000
1,703,277,988,000
null
NONE
null
I've spent a significant amount of time trying to locate the split object inside my _split_generators() custom function. Then after diving [in the code](https://github.com/huggingface/datasets/blob/5ff3670c18ed34fa8ddfa70a9aa403ae6cc9ad54/src/datasets/load.py#L2558) I realized that `download_and_prepare` is executed b...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6529/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6529/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6528
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6528/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6528/comments
https://api.github.com/repos/huggingface/datasets/issues/6528/events
https://github.com/huggingface/datasets/pull/6528
2,053,996,494
PR_kwDODunzps5ip9JH
6,528
set dev version
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6528). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>...
1,703,254,998,000
1,703,255,502,000
null
MEMBER
null
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6528/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6528/timeline
null
null
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6527
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6527/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6527/comments
https://api.github.com/repos/huggingface/datasets/issues/6527/events
https://github.com/huggingface/datasets/pull/6527
2,053,966,748
PR_kwDODunzps5ip2vd
6,527
Release: 2.16.0
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6527). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>...
1,703,253,596,000
1,703,255,052,000
null
MEMBER
null
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6527/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6527/timeline
null
null
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6526
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6526/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6526/comments
https://api.github.com/repos/huggingface/datasets/issues/6526/events
https://github.com/huggingface/datasets/pull/6526
2,053,726,451
PR_kwDODunzps5ipB5v
6,526
Preserve order of configs and splits when using Parquet exports
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",...
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6526). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>...
1,703,241,356,000
1,703,245,342,000
null
MEMBER
null
Preserve order of configs and splits, as defined in dataset infos. Fix #6521.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6526/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6526/timeline
null
null
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6525
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6525/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6525/comments
https://api.github.com/repos/huggingface/datasets/issues/6525/events
https://github.com/huggingface/datasets/pull/6525
2,053,119,357
PR_kwDODunzps5im-lL
6,525
BBox type
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6525). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.", "closing in favor of other ideas that would not involve any typing" ]
1,703,196,807,000
1,703,198,368,000
null
MEMBER
null
see [internal discussion](https://huggingface.slack.com/archives/C02EK7C3SHW/p1703097195609209) Draft to get some feedback on a possible `BBox` feature type that can be used to get object detection bounding boxes data in one format or another. ```python >>> from datasets import load_dataset, BBox >>> ds = load_...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6525/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6525/timeline
null
null
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6524
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6524/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6524/comments
https://api.github.com/repos/huggingface/datasets/issues/6524/events
https://github.com/huggingface/datasets/issues/6524
2,053,076,311
I_kwDODunzps56X3VX
6,524
Streaming the Pile: Missing Files
{ "avatar_url": "https://avatars.githubusercontent.com/u/23347756?v=4", "events_url": "https://api.github.com/users/FelixLabelle/events{/privacy}", "followers_url": "https://api.github.com/users/FelixLabelle/followers", "following_url": "https://api.github.com/users/FelixLabelle/following{/other_user}", "gist...
[]
closed
false
null
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/o...
null
[ "Hello @FelixLabelle,\r\n\r\nAs you can see in the Community tab of the corresponding dataset, it is a known issue: https://huggingface.co/datasets/EleutherAI/pile/discussions/15\r\n\r\nThe data has been taken down due to reported copyright infringement.\r\n\r\nFeel free to continue the discussion there." ]
1,703,193,909,000
1,703,236,625,000
null
NONE
null
### Describe the bug The pile does not stream, a "File not Found error" is returned. It looks like the Pile's files have been moved. ### Steps to reproduce the bug To reproduce run the following code: ``` from datasets import load_dataset dataset = load_dataset('EleutherAI/pile', 'en', split='train', streamin...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6524/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6524/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6523
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6523/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6523/comments
https://api.github.com/repos/huggingface/datasets/issues/6523/events
https://github.com/huggingface/datasets/pull/6523
2,052,643,484
PR_kwDODunzps5ilV6d
6,523
fix tests
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6523). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>...
1,703,172,981,000
1,703,174,214,000
null
MEMBER
null
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6523/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6523/timeline
null
null
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6522
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6522/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6522/comments
https://api.github.com/repos/huggingface/datasets/issues/6522/events
https://github.com/huggingface/datasets/issues/6522
2,052,332,528
I_kwDODunzps56VBvw
6,522
Loading HF Hub Dataset (private org repo) fails to load all features
{ "avatar_url": "https://avatars.githubusercontent.com/u/6579034?v=4", "events_url": "https://api.github.com/users/versipellis/events{/privacy}", "followers_url": "https://api.github.com/users/versipellis/followers", "following_url": "https://api.github.com/users/versipellis/following{/other_user}", "gists_ur...
[]
open
false
null
[]
null
[]
1,703,161,595,000
1,703,165,071,000
null
NONE
null
### Describe the bug When pushing a `Dataset` with multiple `Features` (`input`, `output`, `tags`) to Huggingface Hub (private org repo), and later downloading the `Dataset`, only `input` and `output` load - I believe the expected behavior is for all `Features` to be loaded by default? ### Steps to reproduce the ...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6522/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6522/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6521
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6521/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6521/comments
https://api.github.com/repos/huggingface/datasets/issues/6521/events
https://github.com/huggingface/datasets/issues/6521
2,052,229,538
I_kwDODunzps56Uomi
6,521
The order of the splits is not preserved
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",...
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
null
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/o...
null
[ "After investigation, I think the issue was introduced by the use of the Parquet export:\r\n- #6448\r\n\r\nI am proposing a fix.\r\n\r\nCC: @lhoestq " ]
1,703,157,447,000
1,703,244,975,000
null
MEMBER
null
We had a regression and the order of the splits is not preserved. They are alphabetically sorted, instead of preserving original "train", "validation", "test" order. Check: In branch "main" ```python In [9]: dataset = load_dataset("adversarial_qa", '"adversarialQA") In [10]: dataset Out[10]: DatasetDict({ ...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6521/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6521/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6520
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6520/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6520/comments
https://api.github.com/repos/huggingface/datasets/issues/6520/events
https://github.com/huggingface/datasets/pull/6520
2,052,059,078
PR_kwDODunzps5ijUiw
6,520
Support commit_description parameter in push_to_hub
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",...
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6520). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>...
1,703,151,371,000
1,703,170,187,000
null
MEMBER
null
Support `commit_description` parameter in `push_to_hub`. CC: @Wauplin
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6520/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6520/timeline
null
null
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6519
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6519/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6519/comments
https://api.github.com/repos/huggingface/datasets/issues/6519/events
https://github.com/huggingface/datasets/pull/6519
2,050,759,824
PR_kwDODunzps5ie4MA
6,519
Support push_to_hub canonical datasets
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",...
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6519). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.", "nice catch @albertvillanova ", "@huggingface/datasets this PR is ready for review.", ...
1,703,085,405,000
1,703,170,100,000
null
MEMBER
null
Support `push_to_hub` canonical datasets. This is necessary in the Space to convert script-datasets to Parquet: https://huggingface.co/spaces/albertvillanova/convert-dataset-to-parquet Note that before this PR, the `repo_id` "dataset_name" was transformed to "user/dataset_name". This behavior was introduced by: ...
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/6519/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6519/timeline
null
null
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6518
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6518/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6518/comments
https://api.github.com/repos/huggingface/datasets/issues/6518/events
https://github.com/huggingface/datasets/pull/6518
2,050,137,038
PR_kwDODunzps5icu-W
6,518
fix get_metadata_patterns function args error
{ "avatar_url": "https://avatars.githubusercontent.com/u/12895488?v=4", "events_url": "https://api.github.com/users/d710055071/events{/privacy}", "followers_url": "https://api.github.com/users/d710055071/followers", "following_url": "https://api.github.com/users/d710055071/following{/other_user}", "gists_url"...
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6518). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.", "hello!\r\n@albertvillanova \r\nThank you very much for your recognition。\r\nWhen can t...
1,703,063,182,000
1,703,171,657,000
null
CONTRIBUTOR
null
Bug get_metadata_patterns arg error https://github.com/huggingface/datasets/issues/6517
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6518/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6518/timeline
null
null
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6517
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6517/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6517/comments
https://api.github.com/repos/huggingface/datasets/issues/6517/events
https://github.com/huggingface/datasets/issues/6517
2,050,121,588
I_kwDODunzps56Ml90
6,517
Bug get_metadata_patterns arg error
{ "avatar_url": "https://avatars.githubusercontent.com/u/12895488?v=4", "events_url": "https://api.github.com/users/d710055071/events{/privacy}", "followers_url": "https://api.github.com/users/d710055071/followers", "following_url": "https://api.github.com/users/d710055071/following{/other_user}", "gists_url"...
[]
closed
false
null
[]
null
[]
1,703,062,604,000
1,703,204,663,000
null
CONTRIBUTOR
null
https://github.com/huggingface/datasets/blob/3f149204a2a5948287adcade5e90707aa5207a92/src/datasets/load.py#L1240C1-L1240C69 metadata_patterns = get_metadata_patterns(base_path, download_config=self.download_config)
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6517/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6517/timeline
null
null
null
null
false
End of preview.

No dataset card yet

Downloads last month
2