html_url
stringlengths
51
51
comments
stringlengths
67
24.7k
title
stringlengths
6
280
body
stringlengths
51
36.2k
comment_length
int64
16
1.45k
text
stringlengths
190
38.3k
embeddings
list
https://github.com/huggingface/datasets/issues/6442
I reproduced too: - root: metadata file is ignored (https://huggingface.co/datasets/severo/doc-image-3) - data/ dir: metadata file is ignored (https://huggingface.co/datasets/severo/doc-image-4) - train/ dir: works (https://huggingface.co/datasets/severo/doc-image-5)
Trouble loading image folder with additional features - metadata file ignored
### Describe the bug Loading image folder with a caption column using `load_dataset(<image_folder_path>)` doesn't load the captions. When loading a local image folder with captions using `datasets==2.13.0` ``` from datasets import load_dataset data = load_dataset(<image_folder_path>) data.column_names ``` ...
23
Trouble loading image folder with additional features - metadata file ignored ### Describe the bug Loading image folder with a caption column using `load_dataset(<image_folder_path>)` doesn't load the captions. When loading a local image folder with captions using `datasets==2.13.0` ``` from datasets import...
[ -0.005256466567516327, 0.07954807579517365, 0.025517955422401428, 0.6268149018287659, 0.4523826241493225, 0.11177997291088104, 0.42082440853118896, 0.046340543776750565, 0.2590237557888031, 0.058994755148887634, -0.07752669602632523, 0.4900439381599426, -0.20371121168136597, 0.149850949645...
https://github.com/huggingface/datasets/issues/6441
> Also when they try to click the url link for the dataset they get a 404 error. This seems to be a Hub error then (cc @SBrandeis)
Trouble Loading a Gated Dataset For User with Granted Permission
### Describe the bug I have granted permissions to several users to access a gated huggingface dataset. The users accepted the invite and when trying to load the dataset using their access token they get `FileNotFoundError: Couldn't find a dataset script at .....` . Also when they try to click the url link for the d...
28
Trouble Loading a Gated Dataset For User with Granted Permission ### Describe the bug I have granted permissions to several users to access a gated huggingface dataset. The users accepted the invite and when trying to load the dataset using their access token they get `FileNotFoundError: Couldn't find a dataset sc...
[ -0.29226887226104736, -0.23708491027355194, 0.059749752283096313, 0.5143930912017822, -0.0005606310442090034, 0.010804161429405212, 0.4375983774662018, 0.01528296247124672, 0.49582868814468384, 0.22912351787090302, -0.24911551177501678, -0.1277315467596054, -0.11217568814754486, 0.11388504...
https://github.com/huggingface/datasets/issues/6441
Could you report this to https://discuss.huggingface.co/c/hub/23, providing the URL of the dataset, or at least if the dataset is public or private?
Trouble Loading a Gated Dataset For User with Granted Permission
### Describe the bug I have granted permissions to several users to access a gated huggingface dataset. The users accepted the invite and when trying to load the dataset using their access token they get `FileNotFoundError: Couldn't find a dataset script at .....` . Also when they try to click the url link for the d...
22
Trouble Loading a Gated Dataset For User with Granted Permission ### Describe the bug I have granted permissions to several users to access a gated huggingface dataset. The users accepted the invite and when trying to load the dataset using their access token they get `FileNotFoundError: Couldn't find a dataset sc...
[ -0.1947937309741974, -0.27182096242904663, 0.054994285106658936, 0.5652323961257935, -0.0011484064161777496, 0.013401776552200317, 0.4236222505569458, -0.05076004937291145, 0.40111419558525085, 0.14345966279506683, -0.2048477679491043, -0.15554077923297882, -0.1591770350933075, 0.200249552...
https://github.com/huggingface/datasets/issues/6440
Tried to upgrade Python to 3.11 - still get this message. A partial solution is to NOT use `num_proc` at all. It will be considerably longer to finish the job.
`.map` not hashing under python 3.9
### Describe the bug The `.map` function cannot hash under python 3.9. Tried to use [the solution here](https://github.com/huggingface/datasets/issues/4521#issuecomment-1205166653), but still get the same message: `Parameter 'function'=<function map_to_pred at 0x7fa0b49ead30> of the transform datasets.arrow_data...
30
`.map` not hashing under python 3.9 ### Describe the bug The `.map` function cannot hash under python 3.9. Tried to use [the solution here](https://github.com/huggingface/datasets/issues/4521#issuecomment-1205166653), but still get the same message: `Parameter 'function'=<function map_to_pred at 0x7fa0b49ead30...
[ -0.23779386281967163, -0.5131184458732605, 0.039723578840494156, -0.02358236536383629, 0.3392046391963959, -0.013328582048416138, 0.28753432631492615, 0.25054922699928284, 0.08202031254768372, 0.08945101499557495, 0.005709212273359299, 0.4939461648464203, -0.38324224948883057, 0.0680404901...
https://github.com/huggingface/datasets/issues/6440
Hi! The `model = torch.compile(model)` line is problematic for our hashing logic. We would have to merge https://github.com/huggingface/datasets/pull/5867 to support hashing `torch.compile`-ed models/functions. I've started refactoring the hashing logic and plan to incorporate a fix for `torch.compile` as part of i...
`.map` not hashing under python 3.9
### Describe the bug The `.map` function cannot hash under python 3.9. Tried to use [the solution here](https://github.com/huggingface/datasets/issues/4521#issuecomment-1205166653), but still get the same message: `Parameter 'function'=<function map_to_pred at 0x7fa0b49ead30> of the transform datasets.arrow_data...
52
`.map` not hashing under python 3.9 ### Describe the bug The `.map` function cannot hash under python 3.9. Tried to use [the solution here](https://github.com/huggingface/datasets/issues/4521#issuecomment-1205166653), but still get the same message: `Parameter 'function'=<function map_to_pred at 0x7fa0b49ead30...
[ -0.23779386281967163, -0.5131184458732605, 0.039723578840494156, -0.02358236536383629, 0.3392046391963959, -0.013328582048416138, 0.28753432631492615, 0.25054922699928284, 0.08202031254768372, 0.08945101499557495, 0.005709212273359299, 0.4939461648464203, -0.38324224948883057, 0.0680404901...
https://github.com/huggingface/datasets/issues/6438
Thank you, @severo ! I would be more than happy to help in any way I can. I am not familiar with this repo's codebase, but I would be eager to contribute. :) For the preview in Datasets Hub, I think it makes sense to just display the geospatial column as text. If there were a dataset loader, though, I think it shoul...
Support GeoParquet
### Feature request Support the GeoParquet format ### Motivation GeoParquet (https://geoparquet.org/) is a common format for sharing vectorial geospatial data on the cloud, along with "traditional" data columns. It would be nice to be able to load this format with datasets, and more generally, in the Datasets Hub...
168
Support GeoParquet ### Feature request Support the GeoParquet format ### Motivation GeoParquet (https://geoparquet.org/) is a common format for sharing vectorial geospatial data on the cloud, along with "traditional" data columns. It would be nice to be able to load this format with datasets, and more generally...
[ -0.18842795491218567, -0.007291898131370544, -0.004543256014585495, 0.08613798022270203, 0.20921705663204193, -0.2451484054327011, 0.10604145377874374, 0.4123835861682892, 0.17052997648715973, 0.05875986069440842, -0.04821088910102844, 0.4327125549316406, -0.19047911465168, 0.6723259687423...
https://github.com/huggingface/datasets/issues/6438
Just hitting into this same issue too showing GeoParquet files in Datasets Viewer. I tried to implement a custom reader for GeoParquet in https://huggingface.co/datasets/weiji14/clay_vector_embeddings/discussions/1, but it seems like HuggingFace has disabled datasets with custom loading scripts from using the dataset v...
Support GeoParquet
### Feature request Support the GeoParquet format ### Motivation GeoParquet (https://geoparquet.org/) is a common format for sharing vectorial geospatial data on the cloud, along with "traditional" data columns. It would be nice to be able to load this format with datasets, and more generally, in the Datasets Hub...
132
Support GeoParquet ### Feature request Support the GeoParquet format ### Motivation GeoParquet (https://geoparquet.org/) is a common format for sharing vectorial geospatial data on the cloud, along with "traditional" data columns. It would be nice to be able to load this format with datasets, and more generally...
[ -0.131269171833992, -0.04446607828140259, 0.0009181350469589233, 0.1977737843990326, 0.22396178543567657, -0.0637836903333664, 0.05105629935860634, 0.25741830468177795, 0.1787053346633911, 0.09568889439105988, -0.19080108404159546, 0.42543378472328186, -0.35302814841270447, 0.6135354042053...
https://github.com/huggingface/datasets/issues/6438
Update: It looks like renaming the GeoParquet file to have a file extension of `*.parquet` works (see https://huggingface.co/datasets/weiji14/clay_vector_embeddings). HuggingFace's default parquet reader is able to read the GeoParquet file, though the geometry column is of an unknown type: ![image](https://github.co...
Support GeoParquet
### Feature request Support the GeoParquet format ### Motivation GeoParquet (https://geoparquet.org/) is a common format for sharing vectorial geospatial data on the cloud, along with "traditional" data columns. It would be nice to be able to load this format with datasets, and more generally, in the Datasets Hub...
68
Support GeoParquet ### Feature request Support the GeoParquet format ### Motivation GeoParquet (https://geoparquet.org/) is a common format for sharing vectorial geospatial data on the cloud, along with "traditional" data columns. It would be nice to be able to load this format with datasets, and more generally...
[ -0.1853790283203125, -0.15920162200927734, -0.017098691314458847, 0.17653638124465942, 0.25480347871780396, -0.14688898622989655, -0.037101730704307556, 0.3081713020801544, 0.15877442061901093, 0.08432714641094208, -0.02328629419207573, 0.36783090233802795, -0.24566690623760223, 0.64769935...
https://github.com/huggingface/datasets/issues/6438
@joshuasundance-swca, @weiji14, If I'm understanding this correctly, the code below wouldn't be recommended to due to dependency headaches? If that's the case, what solution would there be to see the geometry features for .gpq files in huggingfaceHub? code for dataset_loader.py ``` import geopandas as gpd # ... ...
Support GeoParquet
### Feature request Support the GeoParquet format ### Motivation GeoParquet (https://geoparquet.org/) is a common format for sharing vectorial geospatial data on the cloud, along with "traditional" data columns. It would be nice to be able to load this format with datasets, and more generally, in the Datasets Hub...
157
Support GeoParquet ### Feature request Support the GeoParquet format ### Motivation GeoParquet (https://geoparquet.org/) is a common format for sharing vectorial geospatial data on the cloud, along with "traditional" data columns. It would be nice to be able to load this format with datasets, and more generally...
[ -0.2090664505958557, -0.26573559641838074, 0.04979145526885986, 0.31051427125930786, 0.17626544833183289, -0.16471150517463684, 0.10059749335050583, 0.3131271004676819, 0.41674306988716125, 0.03970740735530853, -0.1878364086151123, 0.3662434220314026, -0.1826562136411667, 0.592523694038391...
https://github.com/huggingface/datasets/issues/6437
`split_dataset_by_node` doesn't give the exact same number of examples to each node in the case of iterable datasets, though it tries to be as equal as possible. In particular if your dataset is sharded and you have a number of shards that is a factor of the number of workers, then the shards will be evenly distributed...
Problem in training iterable dataset
### Describe the bug I am using PyTorch DDP (Distributed Data Parallel) to train my model. Since the data is too large to load into memory at once, I am using load_dataset to read the data as an iterable dataset. I have used datasets.distributed.split_dataset_by_node to distribute the dataset. However, I have notice...
129
Problem in training iterable dataset ### Describe the bug I am using PyTorch DDP (Distributed Data Parallel) to train my model. Since the data is too large to load into memory at once, I am using load_dataset to read the data as an iterable dataset. I have used datasets.distributed.split_dataset_by_node to distrib...
[ -0.3585677742958069, -0.43623197078704834, -0.06505583226680756, 0.244727224111557, 0.15072189271450043, 0.0003087371587753296, 0.22049580514431, 0.10808394849300385, -0.03405902534723282, 0.03095584362745285, 0.07104901969432831, 0.4075590670108795, -0.1007930338382721, -0.148665875196456...
https://github.com/huggingface/datasets/issues/6432
You should use `datasets.load_dataset` instead of `nlp.load_dataset`, as the `nlp` package is outdated. If switching to `datasets.load_dataset` doesn't fix the issue, sharing the JSON file (feel free to replace the data with dummy data) would be nice so that we can reproduce it ourselves.
load_dataset does not load all of the data in my input file
### Describe the bug I have 127 elements in my input dataset. When I do a len on the dataset after loaded, it is only 124 elements. ### Steps to reproduce the bug train_dataset = nlp.load_dataset(data_args.dataset_path, name=data_args.qg_format, split=nlp.Split.TRAIN) valid_dataset = nlp.load_dataset(data_...
44
load_dataset does not load all of the data in my input file ### Describe the bug I have 127 elements in my input dataset. When I do a len on the dataset after loaded, it is only 124 elements. ### Steps to reproduce the bug train_dataset = nlp.load_dataset(data_args.dataset_path, name=data_args.qg_format, spli...
[ -0.22444385290145874, -0.252759724855423, 0.0006521642208099365, 0.5328270196914673, 0.18958520889282227, -0.0704435408115387, 0.3324061632156372, 0.4615911543369293, -0.04259912669658661, -0.11367630958557129, 0.13742810487747192, 0.3046746850013733, 0.016711439937353134, 0.24982538819313...
https://github.com/huggingface/datasets/issues/6422
We have a config variable that controls the batch size in `save_to_disk`: ```python import datasets datasets.config.DEFAULT_MAX_BATCH_SIZE = <smaller_batch_size> ... ds.save_to_disk(...) ```
Allow to choose the `writer_batch_size` when using `save_to_disk`
### Feature request Add an argument in `save_to_disk` regarding batch size, which would be passed to `shard` and other methods. ### Motivation The `Dataset.save_to_disk` method currently calls `shard` without passing a `writer_batch_size` argument, thus implicitly using the default value (1000). This can result in R...
21
Allow to choose the `writer_batch_size` when using `save_to_disk` ### Feature request Add an argument in `save_to_disk` regarding batch size, which would be passed to `shard` and other methods. ### Motivation The `Dataset.save_to_disk` method currently calls `shard` without passing a `writer_batch_size` argument, ...
[ -0.4314618706703186, -0.1805671751499176, -0.051898352801799774, -0.018587402999401093, 0.021444931626319885, -0.026462405920028687, 0.3796961307525635, 0.23060111701488495, -0.11458665132522583, 0.40544524788856506, 0.2738927900791168, -0.00015321001410484314, -0.3835754096508026, 0.35399...
https://github.com/huggingface/datasets/issues/6422
Thank you for your answer! From what I am reading in `https://github.com/huggingface/datasets/blob/2.14.5/src/datasets/arrow_dataset.py`, every function involved (`select`, `shard`, ...) has a default hardcoded batch size of 1000, as such: ```python def select( self, indices: Iterable, k...
Allow to choose the `writer_batch_size` when using `save_to_disk`
### Feature request Add an argument in `save_to_disk` regarding batch size, which would be passed to `shard` and other methods. ### Motivation The `Dataset.save_to_disk` method currently calls `shard` without passing a `writer_batch_size` argument, thus implicitly using the default value (1000). This can result in R...
134
Allow to choose the `writer_batch_size` when using `save_to_disk` ### Feature request Add an argument in `save_to_disk` regarding batch size, which would be passed to `shard` and other methods. ### Motivation The `Dataset.save_to_disk` method currently calls `shard` without passing a `writer_batch_size` argument, ...
[ -0.4530668258666992, -0.10187280178070068, 0.004923049360513687, 0.05907120928168297, 0.0542483814060688, -0.092894047498703, 0.3411109447479248, 0.13401763141155243, -0.17516355216503143, 0.2381095290184021, 0.20967498421669006, 0.07316283881664276, -0.35989850759506226, 0.142907813191413...
https://github.com/huggingface/datasets/issues/6417
Very strange: `datasets-cli env` > > Copy-and-paste the text below in your GitHub issue. > > - `datasets` version: 2.9.0 > - Platform: macOS-14.0-arm64-arm-64bit > - Python version: 3.9.13 > - PyArrow version: 8.0.0 > - Pandas version: 1.3.5 After updating datasets and pyarrow on base environment, although...
Bug: LayoutLMv3 finetuning on FUNSD Notebook; Arrow Error
### Describe the bug Arrow issues when running the example Notebook laptop locally on Mac with M1. Works on Google Collab. **Notebook**: https://github.com/NielsRogge/Transformers-Tutorials/blob/master/LayoutLMv3/Fine_tune_LayoutLMv3_on_FUNSD_(HuggingFace_Trainer).ipynb **Error**: `ValueError: Arrow type extensi...
95
Bug: LayoutLMv3 finetuning on FUNSD Notebook; Arrow Error ### Describe the bug Arrow issues when running the example Notebook laptop locally on Mac with M1. Works on Google Collab. **Notebook**: https://github.com/NielsRogge/Transformers-Tutorials/blob/master/LayoutLMv3/Fine_tune_LayoutLMv3_on_FUNSD_(HuggingFace...
[ -0.12534523010253906, -0.696060299873352, 0.06675893068313599, 0.3161044716835022, 0.36773860454559326, -0.21727141737937927, 0.570938229560852, 0.08089077472686768, -0.18083691596984863, -0.04032056778669357, -0.07140608876943588, 0.6541163921356201, -0.07819987833499908, 0.20544581115245...
https://github.com/huggingface/datasets/issues/6417
Hi! The latest (patch) release (published a few hours ago) includes a fix for this [PyArrow security issue](https://github.com/advisories/GHSA-5wvp-7f3h-6wmm). To install it, run `pip install -U datasets`.
Bug: LayoutLMv3 finetuning on FUNSD Notebook; Arrow Error
### Describe the bug Arrow issues when running the example Notebook laptop locally on Mac with M1. Works on Google Collab. **Notebook**: https://github.com/NielsRogge/Transformers-Tutorials/blob/master/LayoutLMv3/Fine_tune_LayoutLMv3_on_FUNSD_(HuggingFace_Trainer).ipynb **Error**: `ValueError: Arrow type extensi...
26
Bug: LayoutLMv3 finetuning on FUNSD Notebook; Arrow Error ### Describe the bug Arrow issues when running the example Notebook laptop locally on Mac with M1. Works on Google Collab. **Notebook**: https://github.com/NielsRogge/Transformers-Tutorials/blob/master/LayoutLMv3/Fine_tune_LayoutLMv3_on_FUNSD_(HuggingFace...
[ -0.12534523010253906, -0.696060299873352, 0.06675893068313599, 0.3161044716835022, 0.36773860454559326, -0.21727141737937927, 0.570938229560852, 0.08089077472686768, -0.18083691596984863, -0.04032056778669357, -0.07140608876943588, 0.6541163921356201, -0.07819987833499908, 0.20544581115245...
https://github.com/huggingface/datasets/issues/6417
> Hi! The latest (patch) release (published a few hours ago) includes a fix for this [PyArrow security issue](https://github.com/advisories/GHSA-5wvp-7f3h-6wmm). To install it, run `pip install -U datasets`. Thanks for the info and the latest release, it seems this has also solved my issue. First run after the updat...
Bug: LayoutLMv3 finetuning on FUNSD Notebook; Arrow Error
### Describe the bug Arrow issues when running the example Notebook laptop locally on Mac with M1. Works on Google Collab. **Notebook**: https://github.com/NielsRogge/Transformers-Tutorials/blob/master/LayoutLMv3/Fine_tune_LayoutLMv3_on_FUNSD_(HuggingFace_Trainer).ipynb **Error**: `ValueError: Arrow type extensi...
60
Bug: LayoutLMv3 finetuning on FUNSD Notebook; Arrow Error ### Describe the bug Arrow issues when running the example Notebook laptop locally on Mac with M1. Works on Google Collab. **Notebook**: https://github.com/NielsRogge/Transformers-Tutorials/blob/master/LayoutLMv3/Fine_tune_LayoutLMv3_on_FUNSD_(HuggingFace...
[ -0.12534523010253906, -0.696060299873352, 0.06675893068313599, 0.3161044716835022, 0.36773860454559326, -0.21727141737937927, 0.570938229560852, 0.08089077472686768, -0.18083691596984863, -0.04032056778669357, -0.07140608876943588, 0.6541163921356201, -0.07819987833499908, 0.20544581115245...
https://github.com/huggingface/datasets/issues/6412
Indeed, this is not a good practice. I've opened a PR that removes the token value from the (deprecation) warning.
User token is printed out!
This line prints user token on command line! Is it safe? https://github.com/huggingface/datasets/blob/12ebe695b4748c5a26e08b44ed51955f74f5801d/src/datasets/load.py#L2091
20
User token is printed out! This line prints user token on command line! Is it safe? https://github.com/huggingface/datasets/blob/12ebe695b4748c5a26e08b44ed51955f74f5801d/src/datasets/load.py#L2091 Indeed, this is not a good practice. I've opened a PR that removes the token value from the (deprecation) warning.
[ -0.0031762495636940002, -0.1586783230304718, -0.1245645135641098, -0.01705431565642357, 0.22019343078136444, 0.023465313017368317, 0.41891172528266907, -0.0682024210691452, 0.007502242922782898, 0.07574683427810669, -0.003919064998626709, 0.028364498168230057, -0.009011231362819672, 0.2457...
https://github.com/huggingface/datasets/issues/6410
Hi! You can avoid the error by requesting only the `jsonl` files. `dataset = load_dataset("ai4privacy/pii-masking-200k", data_files=["*.jsonl"])`. Our data file inference does not filter out (incompatible) `json` files because `json` and `jsonl` use the same builder. Still, I think the inference should differentiat...
Datasets does not load HuggingFace Repository properly
### Describe the bug Dear Datasets team, We just have published a dataset on Huggingface: https://huggingface.co/ai4privacy However, when trying to read it using the Dataset library we get an error. As I understand jsonl files are compatible, could you please clarify how we can solve the issue? Please let me ...
60
Datasets does not load HuggingFace Repository properly ### Describe the bug Dear Datasets team, We just have published a dataset on Huggingface: https://huggingface.co/ai4privacy However, when trying to read it using the Dataset library we get an error. As I understand jsonl files are compatible, could you ...
[ -0.20247381925582886, -0.10110732913017273, -0.026749910786747932, 0.7637389898300171, 0.29011261463165283, -0.03652459383010864, -0.00203618872910738, 0.24976053833961487, 0.019404388964176178, 0.02537638321518898, -0.2364252805709839, 0.2252340018749237, -0.05940572917461395, 0.331941127...
https://github.com/huggingface/datasets/issues/6410
Raising an error if there is a mix of json and jsonl in the builder makes sense yea
Datasets does not load HuggingFace Repository properly
### Describe the bug Dear Datasets team, We just have published a dataset on Huggingface: https://huggingface.co/ai4privacy However, when trying to read it using the Dataset library we get an error. As I understand jsonl files are compatible, could you please clarify how we can solve the issue? Please let me ...
18
Datasets does not load HuggingFace Repository properly ### Describe the bug Dear Datasets team, We just have published a dataset on Huggingface: https://huggingface.co/ai4privacy However, when trying to read it using the Dataset library we get an error. As I understand jsonl files are compatible, could you ...
[ -0.20247381925582886, -0.10110732913017273, -0.026749910786747932, 0.7637389898300171, 0.29011261463165283, -0.03652459383010864, -0.00203618872910738, 0.24976053833961487, 0.019404388964176178, 0.02537638321518898, -0.2364252805709839, 0.2252340018749237, -0.05940572917461395, 0.331941127...
https://github.com/huggingface/datasets/issues/6405
The viewer is working now. Based on the repo commit history, the bug was due to the incorrect format of the `features` field in the README YAML (`Value` requires `dtype`, e.g., `Value("string")`, but it was not specified)
ConfigNamesError on a simple CSV file
See https://huggingface.co/datasets/Nguyendo1999/mmath/discussions/1 ``` Error code: ConfigNamesError Exception: TypeError Message: __init__() missing 1 required positional argument: 'dtype' Traceback: Traceback (most recent call last): File "/src/services/worker/src/worker/job_runn...
37
ConfigNamesError on a simple CSV file See https://huggingface.co/datasets/Nguyendo1999/mmath/discussions/1 ``` Error code: ConfigNamesError Exception: TypeError Message: __init__() missing 1 required positional argument: 'dtype' Traceback: Traceback (most recent call last): File "...
[ -0.37286028265953064, -0.2837359309196472, -0.09386485815048218, 0.07334684580564499, 0.42209455370903015, 0.25863489508628845, 0.5016807913780212, 0.33707672357559204, -0.02036479488015175, 0.2753065824508667, 0.42774254083633423, -0.016560789197683334, 0.06351710855960846, 0.372148096561...
https://github.com/huggingface/datasets/issues/6403
You are most likely using an outdated version of `datasets` in the notebook, which can be verified with the `!datasets-cli env` command. You can run `!pip install -U datasets` to update the installation.
Cannot import datasets on google colab (python 3.10.12)
### Describe the bug I'm trying A full colab demo notebook of zero-shot-distillation from https://github.com/huggingface/transformers/tree/main/examples/research_projects/zero-shot-distillation but i got this type of error when importing datasets on my google colab (python version is 3.10.12) ![image](https://gith...
33
Cannot import datasets on google colab (python 3.10.12) ### Describe the bug I'm trying A full colab demo notebook of zero-shot-distillation from https://github.com/huggingface/transformers/tree/main/examples/research_projects/zero-shot-distillation but i got this type of error when importing datasets on my google ...
[ -0.22408875823020935, -0.2938651144504547, 0.08976825326681137, 0.43063440918922424, 0.21170759201049805, -0.08203870803117752, 0.3200898766517639, 0.04798250272870064, 0.08357463032007217, 0.14291954040527344, -0.5043290257453918, 0.30278438329696655, -0.17893865704536438, 0.2081073820590...
https://github.com/huggingface/datasets/issues/6401
Seems like it's a problem with the dataset, since in the [README](https://huggingface.co/datasets/Hyperspace-Technologies/scp-wiki-text/blob/main/README.md) the validation is not specified. Try cloning the dataset, removing the README (or validation split), and loading it locally/
dataset = load_dataset("Hyperspace-Technologies/scp-wiki-text") not working
### Describe the bug ``` (datasets) mruserbox@guru-X99:/media/10TB_HHD/_LLM_DATASETS$ python dataset.py Downloading readme: 100%|███████████████████████████████████| 360/360 [00:00<00:00, 2.16MB/s] Downloading data: 100%|█████████████████████████████████| 65.1M/65.1M [00:19<00:00, 3.38MB/s] Downloading data: 100...
31
dataset = load_dataset("Hyperspace-Technologies/scp-wiki-text") not working ### Describe the bug ``` (datasets) mruserbox@guru-X99:/media/10TB_HHD/_LLM_DATASETS$ python dataset.py Downloading readme: 100%|███████████████████████████████████| 360/360 [00:00<00:00, 2.16MB/s] Downloading data: 100%|███████████████...
[ -0.4304260015487671, -0.008432343602180481, -0.09224650263786316, 0.2506204843521118, 0.28447407484054565, 0.03460144251585007, 0.37043696641921997, 0.33112281560897827, -0.011478707194328308, 0.15837866067886353, 0.052000924944877625, 0.12657028436660767, -0.14188843965530396, 0.345798134...
https://github.com/huggingface/datasets/issues/6401
@VarunNSrivastava thanks brother, working beautiful now ``` C:\_Work\_datasets>py dataset.py Downloading data files: 100%|████████████████████████████████████████████████████████████████████| 3/3 [00:00<?, ?it/s] Extracting data files: 100%|████████████████████████████████████████████████████████████| 3/3 [00:00<...
dataset = load_dataset("Hyperspace-Technologies/scp-wiki-text") not working
### Describe the bug ``` (datasets) mruserbox@guru-X99:/media/10TB_HHD/_LLM_DATASETS$ python dataset.py Downloading readme: 100%|███████████████████████████████████| 360/360 [00:00<00:00, 2.16MB/s] Downloading data: 100%|█████████████████████████████████| 65.1M/65.1M [00:19<00:00, 3.38MB/s] Downloading data: 100...
48
dataset = load_dataset("Hyperspace-Technologies/scp-wiki-text") not working ### Describe the bug ``` (datasets) mruserbox@guru-X99:/media/10TB_HHD/_LLM_DATASETS$ python dataset.py Downloading readme: 100%|███████████████████████████████████| 360/360 [00:00<00:00, 2.16MB/s] Downloading data: 100%|███████████████...
[ -0.4304260015487671, -0.008432343602180481, -0.09224650263786316, 0.2506204843521118, 0.28447407484054565, 0.03460144251585007, 0.37043696641921997, 0.33112281560897827, -0.011478707194328308, 0.15837866067886353, 0.052000924944877625, 0.12657028436660767, -0.14188843965530396, 0.345798134...
https://github.com/huggingface/datasets/issues/6400
great idea IMO this could be a `trust_remote_code=True` flag like in transformers. We could also default to loading the Parquet conversion rather than executing code (for dataset repos that have both)
Safely load datasets by disabling execution of dataset loading script
### Feature request Is there a way to disable execution of dataset loading script using `load_dataset`? This is a security vulnerability that could lead to arbitrary code execution. Any suggested workarounds are welcome as well. ### Motivation This is a security vulnerability that could lead to arbitrary code e...
31
Safely load datasets by disabling execution of dataset loading script ### Feature request Is there a way to disable execution of dataset loading script using `load_dataset`? This is a security vulnerability that could lead to arbitrary code execution. Any suggested workarounds are welcome as well. ### Motivati...
[ -0.49187153577804565, 0.24664939939975739, -0.06180678680539131, 0.07278358936309814, 0.18239806592464447, -0.06257647275924683, 0.47560447454452515, -0.0017527136951684952, 0.3500739634037018, 0.19620494544506073, -0.030168278142809868, 0.48009923100471497, -0.09626144915819168, 0.5082423...
https://github.com/huggingface/datasets/issues/6400
We added the `trust_remote_code` argument to `load_dataset()` in `datasets` 2.16: - in the future users will have to pass trust_remote_code=True to use datasets with a script - for now we just show a warning when a dataset script is used - we fallback on the Hugging Face Parquet exports when possible (to keep compat...
Safely load datasets by disabling execution of dataset loading script
### Feature request Is there a way to disable execution of dataset loading script using `load_dataset`? This is a security vulnerability that could lead to arbitrary code execution. Any suggested workarounds are welcome as well. ### Motivation This is a security vulnerability that could lead to arbitrary code e...
76
Safely load datasets by disabling execution of dataset loading script ### Feature request Is there a way to disable execution of dataset loading script using `load_dataset`? This is a security vulnerability that could lead to arbitrary code execution. Any suggested workarounds are welcome as well. ### Motivati...
[ -0.5202223062515259, 0.20792506635189056, -0.08936712890863419, 0.15276484191417694, 0.1941721886396408, -0.030967362225055695, 0.40098804235458374, 0.023934371769428253, 0.5325300097465515, 0.2042616307735443, -0.026901861652731895, 0.5292146801948547, -0.20548397302627563, 0.355250477790...
https://github.com/huggingface/datasets/issues/6396
https://github.com/huggingface/datasets-server/pull/2089#pullrequestreview-1724449532 > Yes, I understand now: they have disabled their `PyExtensionType` and we use it in `datasets` for arrays...
Issue with pyarrow 14.0.1
See https://github.com/huggingface/datasets-server/pull/2089 for reference ``` from datasets import (Array2D, Dataset, Features) feature_type = Array2D(shape=(2, 2), dtype="float32") content = [[0.0, 0.0], [0.0, 0.0]] features = Features({"col": feature_type}) dataset = Dataset.from_dict({"col": [content]}, fea...
19
Issue with pyarrow 14.0.1 See https://github.com/huggingface/datasets-server/pull/2089 for reference ``` from datasets import (Array2D, Dataset, Features) feature_type = Array2D(shape=(2, 2), dtype="float32") content = [[0.0, 0.0], [0.0, 0.0]] features = Features({"col": feature_type}) dataset = Dataset.from_...
[ -0.10018604248762131, 0.13276909291744232, -0.0065413713455200195, 0.12114937603473663, 0.33292287588119507, -0.0022718533873558044, 0.5476638078689575, 0.42727774381637573, -0.40590208768844604, -0.01682676374912262, -0.04357265681028366, 0.5135801434516907, -0.038570161908864975, 0.18583...
https://github.com/huggingface/datasets/issues/6395
We've replaced our filelock implementation with the `filelock` package, so their repo is the right place to request this feature. In the meantime, the following should work: ```python import filelock filelock.FileLock = filelock.SoftFileLock import datasets ... ```
Add ability to set lock type
### Feature request Allow setting file lock type, maybe from an environment variable Currently, it only depends on whether fnctl is available: https://github.com/huggingface/datasets/blob/12ebe695b4748c5a26e08b44ed51955f74f5801d/src/datasets/utils/filelock.py#L463-L470C16 ### Motivation In my environment...
37
Add ability to set lock type ### Feature request Allow setting file lock type, maybe from an environment variable Currently, it only depends on whether fnctl is available: https://github.com/huggingface/datasets/blob/12ebe695b4748c5a26e08b44ed51955f74f5801d/src/datasets/utils/filelock.py#L463-L470C16 ### ...
[ -0.3815123736858368, -0.05841831862926483, -0.07984399050474167, -0.3621041476726532, -0.15861397981643677, -0.08662982285022736, 0.6261725425720215, 0.16512562334537506, 0.16967499256134033, 0.08532869815826416, -0.34569329023361206, 0.21483486890792847, -0.06048876792192459, 0.0627746880...
https://github.com/huggingface/datasets/issues/6393
Adding some more information, it seems to occur more frequently with large (millions of samples) datasets.
Filter occasionally hangs
### Describe the bug A call to `.filter` occasionally hangs (after the filter is complete, according to tqdm) There is a trace produced ``` Exception ignored in: <function Dataset.__del__ at 0x7efb48130c10> Traceback (most recent call last): File "/usr/lib/python3/dist-packages/datasets/arrow_dataset.py", l...
16
Filter occasionally hangs ### Describe the bug A call to `.filter` occasionally hangs (after the filter is complete, according to tqdm) There is a trace produced ``` Exception ignored in: <function Dataset.__del__ at 0x7efb48130c10> Traceback (most recent call last): File "/usr/lib/python3/dist-packages/d...
[ -0.36410510540008545, -0.007196187973022461, -0.07324452698230743, 0.1654694676399231, 0.34145641326904297, -0.19884663820266724, 0.07616069167852402, 0.03563345596194267, 0.13033874332904816, 0.10187101364135742, 0.2923228144645691, 0.4539174735546112, -0.27917808294296265, -0.01014070212...
https://github.com/huggingface/datasets/issues/6393
More information. My code is structured as (1) load (2) map (3) filter (4) filter. It was always the second filter that failed. Combining the two filters into one seems to reliably work.
Filter occasionally hangs
### Describe the bug A call to `.filter` occasionally hangs (after the filter is complete, according to tqdm) There is a trace produced ``` Exception ignored in: <function Dataset.__del__ at 0x7efb48130c10> Traceback (most recent call last): File "/usr/lib/python3/dist-packages/datasets/arrow_dataset.py", l...
33
Filter occasionally hangs ### Describe the bug A call to `.filter` occasionally hangs (after the filter is complete, according to tqdm) There is a trace produced ``` Exception ignored in: <function Dataset.__del__ at 0x7efb48130c10> Traceback (most recent call last): File "/usr/lib/python3/dist-packages/d...
[ -0.30096906423568726, 0.02597421407699585, -0.05257953330874443, 0.20952379703521729, 0.3754666745662689, -0.19390510022640228, 0.14301031827926636, 0.054887741804122925, 0.15424728393554688, 0.05187126249074936, 0.2389901578426361, 0.46147793531417847, -0.21464863419532776, -0.03009072318...
https://github.com/huggingface/datasets/issues/6393
@lhoestq it'd be great if someone had a chance to look at this. I suspect it is impacting many users given the other issue that I linked.
Filter occasionally hangs
### Describe the bug A call to `.filter` occasionally hangs (after the filter is complete, according to tqdm) There is a trace produced ``` Exception ignored in: <function Dataset.__del__ at 0x7efb48130c10> Traceback (most recent call last): File "/usr/lib/python3/dist-packages/datasets/arrow_dataset.py", l...
27
Filter occasionally hangs ### Describe the bug A call to `.filter` occasionally hangs (after the filter is complete, according to tqdm) There is a trace produced ``` Exception ignored in: <function Dataset.__del__ at 0x7efb48130c10> Traceback (most recent call last): File "/usr/lib/python3/dist-packages/d...
[ -0.34762266278266907, 0.039869725704193115, -0.07918084412813187, 0.15619584918022156, 0.34971845149993896, -0.21276167035102844, 0.09959720075130463, 0.061069048941135406, 0.08661271631717682, 0.08014142513275146, 0.26038306951522827, 0.47290682792663574, -0.2593033015727997, 0.0341965556...
https://github.com/huggingface/datasets/issues/6393
Hi ! Sorry for the late response. Was it happening after the first or the second filter ? It looks like an issue with the garbage collector (which makes it random). Maybe datasets created with `filter` are not always handled properly ? cc @mariosasko
Filter occasionally hangs
### Describe the bug A call to `.filter` occasionally hangs (after the filter is complete, according to tqdm) There is a trace produced ``` Exception ignored in: <function Dataset.__del__ at 0x7efb48130c10> Traceback (most recent call last): File "/usr/lib/python3/dist-packages/datasets/arrow_dataset.py", l...
44
Filter occasionally hangs ### Describe the bug A call to `.filter` occasionally hangs (after the filter is complete, according to tqdm) There is a trace produced ``` Exception ignored in: <function Dataset.__del__ at 0x7efb48130c10> Traceback (most recent call last): File "/usr/lib/python3/dist-packages/d...
[ -0.29349881410598755, -0.0039390623569488525, -0.07677945494651794, 0.1827934831380844, 0.382412850856781, -0.21573856472969055, 0.056217312812805176, 0.0004232563078403473, 0.11997350305318832, 0.09813044220209122, 0.30350425839424133, 0.47646021842956543, -0.26908016204833984, -0.0152659...
https://github.com/huggingface/datasets/issues/6393
It was after the second filter (and combining the two filters into one seemingly resolved it). I obviously haven't tried all settings to know that these details are causal, but it did work for me.
Filter occasionally hangs
### Describe the bug A call to `.filter` occasionally hangs (after the filter is complete, according to tqdm) There is a trace produced ``` Exception ignored in: <function Dataset.__del__ at 0x7efb48130c10> Traceback (most recent call last): File "/usr/lib/python3/dist-packages/datasets/arrow_dataset.py", l...
35
Filter occasionally hangs ### Describe the bug A call to `.filter` occasionally hangs (after the filter is complete, according to tqdm) There is a trace produced ``` Exception ignored in: <function Dataset.__del__ at 0x7efb48130c10> Traceback (most recent call last): File "/usr/lib/python3/dist-packages/d...
[ -0.3603329062461853, 0.0368996262550354, -0.0837249830365181, 0.11418618261814117, 0.3722602427005768, -0.2295568734407425, 0.08928656578063965, 0.06284244358539581, 0.0738249123096466, 0.08880357444286346, 0.2227749228477478, 0.4795060157775879, -0.24935586750507355, 0.043282948434352875,...
https://github.com/huggingface/datasets/issues/6393
Thanks, that's good to know. The stacktrace suggests an issue when `del self._indices` is called, which happens when a filtered dataset falls out of scope. The indices are a PyArrow table memory mapped from disk, so I'm not quite sure how calling `del` on it can cause this issue. We do `del self._indices` to make su...
Filter occasionally hangs
### Describe the bug A call to `.filter` occasionally hangs (after the filter is complete, according to tqdm) There is a trace produced ``` Exception ignored in: <function Dataset.__del__ at 0x7efb48130c10> Traceback (most recent call last): File "/usr/lib/python3/dist-packages/datasets/arrow_dataset.py", l...
93
Filter occasionally hangs ### Describe the bug A call to `.filter` occasionally hangs (after the filter is complete, according to tqdm) There is a trace produced ``` Exception ignored in: <function Dataset.__del__ at 0x7efb48130c10> Traceback (most recent call last): File "/usr/lib/python3/dist-packages/d...
[ -0.2699989974498749, 0.07587841153144836, -0.05574621260166168, 0.23851166665554047, 0.3574141263961792, -0.13207729160785675, 0.14118105173110962, 0.05039813369512558, -0.031715277582407, 0.16522642970085144, 0.18220847845077515, 0.4575658440589905, -0.27161359786987305, -0.09806235134601...
https://github.com/huggingface/datasets/issues/6393
Yeah, I have a reliable repro, but it is not even close to minimal and uses a dataset I can't share. Perhaps you could try getting close to my setting. (1) make a large (~20GB) jsonl with prompt/response pairs (2) load it on a linux machine (`dataset = load_dataset(...)`) (3) map a tokenizer to it, with multiproce...
Filter occasionally hangs
### Describe the bug A call to `.filter` occasionally hangs (after the filter is complete, according to tqdm) There is a trace produced ``` Exception ignored in: <function Dataset.__del__ at 0x7efb48130c10> Traceback (most recent call last): File "/usr/lib/python3/dist-packages/datasets/arrow_dataset.py", l...
108
Filter occasionally hangs ### Describe the bug A call to `.filter` occasionally hangs (after the filter is complete, according to tqdm) There is a trace produced ``` Exception ignored in: <function Dataset.__del__ at 0x7efb48130c10> Traceback (most recent call last): File "/usr/lib/python3/dist-packages/d...
[ -0.39694449305534363, 0.12597009539604187, -0.05943926051259041, 0.1702365279197693, 0.27565693855285645, -0.29603731632232666, 0.04032320901751518, 0.13020668923854828, 0.09984563291072845, 0.04136590287089348, 0.27463430166244507, 0.5538307428359985, -0.28259706497192383, -0.001442655920...
https://github.com/huggingface/datasets/issues/6392
Hi! We made some improvements to `push_to_hub` to make it more robust a couple of weeks ago but haven't published a release in the meantime, so it would help if you could install `datasets` from `main` (`pip install https://github.com/huggingface/datasets`) and let us know if this improved version of `push_to_hub` reso...
`push_to_hub` is not robust to hub closing connection
### Describe the bug Like to #6172, `push_to_hub` will crash if Hub resets the connection and raise the following error: ``` Pushing dataset shards to the dataset hub: 32%|███▏ | 54/171 [06:38<14:23, 7.38s/it] Traceback (most recent call last): File "/admin/home-piraka9011/.virtualenvs/w2v2/lib/python3.8/...
86
`push_to_hub` is not robust to hub closing connection ### Describe the bug Like to #6172, `push_to_hub` will crash if Hub resets the connection and raise the following error: ``` Pushing dataset shards to the dataset hub: 32%|███▏ | 54/171 [06:38<14:23, 7.38s/it] Traceback (most recent call last): File ...
[ -0.5284816026687622, -0.3638116121292114, -0.06487172842025757, 0.1819303184747696, 0.3242587447166443, -0.20704427361488342, 0.17081235349178314, 0.3081148862838745, -0.03854190558195114, -0.022969290614128113, -0.19621890783309937, 0.18021035194396973, 0.1735050529241562, 0.0446992143988...
https://github.com/huggingface/datasets/issues/6392
The update helped push more data. However it still crashed a little later: ``` Traceback (most recent call last): File "/admin/home-piraka9011/.virtualenvs/w2v2/lib/python3.8/site-packages/huggingface_hub/utils/_errors.py", line 270, in hf_raise_for_status response.raise_for_status() File "/admin/home-p...
`push_to_hub` is not robust to hub closing connection
### Describe the bug Like to #6172, `push_to_hub` will crash if Hub resets the connection and raise the following error: ``` Pushing dataset shards to the dataset hub: 32%|███▏ | 54/171 [06:38<14:23, 7.38s/it] Traceback (most recent call last): File "/admin/home-piraka9011/.virtualenvs/w2v2/lib/python3.8/...
223
`push_to_hub` is not robust to hub closing connection ### Describe the bug Like to #6172, `push_to_hub` will crash if Hub resets the connection and raise the following error: ``` Pushing dataset shards to the dataset hub: 32%|███▏ | 54/171 [06:38<14:23, 7.38s/it] Traceback (most recent call last): File ...
[ -0.5284816026687622, -0.3638116121292114, -0.06487172842025757, 0.1819303184747696, 0.3242587447166443, -0.20704427361488342, 0.17081235349178314, 0.3081148862838745, -0.03854190558195114, -0.022969290614128113, -0.19621890783309937, 0.18021035194396973, 0.1735050529241562, 0.0446992143988...
https://github.com/huggingface/datasets/issues/6392
I think the previous implementation was actually better: it pushes to the hub every shard. So if it fails, as long as the shards have the same checksum, it will skip the ones that have been pushed. The implementation in `main` pushes commits at the end, so when it fails, there are no commits and therefore restarts f...
`push_to_hub` is not robust to hub closing connection
### Describe the bug Like to #6172, `push_to_hub` will crash if Hub resets the connection and raise the following error: ``` Pushing dataset shards to the dataset hub: 32%|███▏ | 54/171 [06:38<14:23, 7.38s/it] Traceback (most recent call last): File "/admin/home-piraka9011/.virtualenvs/w2v2/lib/python3.8/...
312
`push_to_hub` is not robust to hub closing connection ### Describe the bug Like to #6172, `push_to_hub` will crash if Hub resets the connection and raise the following error: ``` Pushing dataset shards to the dataset hub: 32%|███▏ | 54/171 [06:38<14:23, 7.38s/it] Traceback (most recent call last): File ...
[ -0.5284816026687622, -0.3638116121292114, -0.06487172842025757, 0.1819303184747696, 0.3242587447166443, -0.20704427361488342, 0.17081235349178314, 0.3081148862838745, -0.03854190558195114, -0.022969290614128113, -0.19621890783309937, 0.18021035194396973, 0.1735050529241562, 0.0446992143988...
https://github.com/huggingface/datasets/issues/6392
There's a new error from the hub now: ``` Pushing dataset shards to the dataset hub: 49%|████▉ | 128/261 [11:38<12:05, 5.45s/it] Traceback (most recent call last): File "/admin/home-piraka9011/.virtualenvs/w2v2/lib/python3.8/site-packages/huggingface_hub/utils/_errors.py", line 270, in hf_raise_for_status ...
`push_to_hub` is not robust to hub closing connection
### Describe the bug Like to #6172, `push_to_hub` will crash if Hub resets the connection and raise the following error: ``` Pushing dataset shards to the dataset hub: 32%|███▏ | 54/171 [06:38<14:23, 7.38s/it] Traceback (most recent call last): File "/admin/home-piraka9011/.virtualenvs/w2v2/lib/python3.8/...
225
`push_to_hub` is not robust to hub closing connection ### Describe the bug Like to #6172, `push_to_hub` will crash if Hub resets the connection and raise the following error: ``` Pushing dataset shards to the dataset hub: 32%|███▏ | 54/171 [06:38<14:23, 7.38s/it] Traceback (most recent call last): File ...
[ -0.5284816026687622, -0.3638116121292114, -0.06487172842025757, 0.1819303184747696, 0.3242587447166443, -0.20704427361488342, 0.17081235349178314, 0.3081148862838745, -0.03854190558195114, -0.022969290614128113, -0.19621890783309937, 0.18021035194396973, 0.1735050529241562, 0.0446992143988...
https://github.com/huggingface/datasets/issues/6392
> think the previous implementation was actually better: it pushes to the hub every shard. So if it fails, as long as the shards have the same checksum, it will skip the ones that have been pushed. > >The implementation in main pushes commits at the end, so when it fails, there are no commits and therefore restarts ...
`push_to_hub` is not robust to hub closing connection
### Describe the bug Like to #6172, `push_to_hub` will crash if Hub resets the connection and raise the following error: ``` Pushing dataset shards to the dataset hub: 32%|███▏ | 54/171 [06:38<14:23, 7.38s/it] Traceback (most recent call last): File "/admin/home-piraka9011/.virtualenvs/w2v2/lib/python3.8/...
178
`push_to_hub` is not robust to hub closing connection ### Describe the bug Like to #6172, `push_to_hub` will crash if Hub resets the connection and raise the following error: ``` Pushing dataset shards to the dataset hub: 32%|███▏ | 54/171 [06:38<14:23, 7.38s/it] Traceback (most recent call last): File ...
[ -0.5284816026687622, -0.3638116121292114, -0.06487172842025757, 0.1819303184747696, 0.3242587447166443, -0.20704427361488342, 0.17081235349178314, 0.3081148862838745, -0.03854190558195114, -0.022969290614128113, -0.19621890783309937, 0.18021035194396973, 0.1735050529241562, 0.0446992143988...
https://github.com/huggingface/datasets/issues/6392
But I agree that the `500 Server Error` returned by the Hub is annoying. Earlier today, I also got it on a small 5GB dataset (with 500 MB shards). @Wauplin @julien-c Is there something we can do about this?
`push_to_hub` is not robust to hub closing connection
### Describe the bug Like to #6172, `push_to_hub` will crash if Hub resets the connection and raise the following error: ``` Pushing dataset shards to the dataset hub: 32%|███▏ | 54/171 [06:38<14:23, 7.38s/it] Traceback (most recent call last): File "/admin/home-piraka9011/.virtualenvs/w2v2/lib/python3.8/...
39
`push_to_hub` is not robust to hub closing connection ### Describe the bug Like to #6172, `push_to_hub` will crash if Hub resets the connection and raise the following error: ``` Pushing dataset shards to the dataset hub: 32%|███▏ | 54/171 [06:38<14:23, 7.38s/it] Traceback (most recent call last): File ...
[ -0.5284816026687622, -0.3638116121292114, -0.06487172842025757, 0.1819303184747696, 0.3242587447166443, -0.20704427361488342, 0.17081235349178314, 0.3081148862838745, -0.03854190558195114, -0.022969290614128113, -0.19621890783309937, 0.18021035194396973, 0.1735050529241562, 0.0446992143988...
https://github.com/huggingface/datasets/issues/6392
@mariosasko can't do much if AWS raises a HTTP 500 unfortunately (we are simply pushing data to a S3 bucket). What we can do is to add a retry mechanism in the multi-part upload logic here: https://github.com/huggingface/huggingface_hub/blob/c972cba1fecb456a7b3325cdd1fdbcc425f21f94/src/huggingface_hub/lfs.py#L370 :con...
`push_to_hub` is not robust to hub closing connection
### Describe the bug Like to #6172, `push_to_hub` will crash if Hub resets the connection and raise the following error: ``` Pushing dataset shards to the dataset hub: 32%|███▏ | 54/171 [06:38<14:23, 7.38s/it] Traceback (most recent call last): File "/admin/home-piraka9011/.virtualenvs/w2v2/lib/python3.8/...
38
`push_to_hub` is not robust to hub closing connection ### Describe the bug Like to #6172, `push_to_hub` will crash if Hub resets the connection and raise the following error: ``` Pushing dataset shards to the dataset hub: 32%|███▏ | 54/171 [06:38<14:23, 7.38s/it] Traceback (most recent call last): File ...
[ -0.5284816026687622, -0.3638116121292114, -0.06487172842025757, 0.1819303184747696, 0.3242587447166443, -0.20704427361488342, 0.17081235349178314, 0.3081148862838745, -0.03854190558195114, -0.022969290614128113, -0.19621890783309937, 0.18021035194396973, 0.1735050529241562, 0.0446992143988...
https://github.com/huggingface/datasets/issues/6392
> That code already retries the request using http_backoff, no? Currently only on HTTP 503 by default. We should add 500 as well (and hope it is a transient error from AWS)
`push_to_hub` is not robust to hub closing connection
### Describe the bug Like to #6172, `push_to_hub` will crash if Hub resets the connection and raise the following error: ``` Pushing dataset shards to the dataset hub: 32%|███▏ | 54/171 [06:38<14:23, 7.38s/it] Traceback (most recent call last): File "/admin/home-piraka9011/.virtualenvs/w2v2/lib/python3.8/...
32
`push_to_hub` is not robust to hub closing connection ### Describe the bug Like to #6172, `push_to_hub` will crash if Hub resets the connection and raise the following error: ``` Pushing dataset shards to the dataset hub: 32%|███▏ | 54/171 [06:38<14:23, 7.38s/it] Traceback (most recent call last): File ...
[ -0.5284816026687622, -0.3638116121292114, -0.06487172842025757, 0.1819303184747696, 0.3242587447166443, -0.20704427361488342, 0.17081235349178314, 0.3081148862838745, -0.03854190558195114, -0.022969290614128113, -0.19621890783309937, 0.18021035194396973, 0.1735050529241562, 0.0446992143988...
https://github.com/huggingface/datasets/issues/6392
Opened a PR to retry in case S3 raises HTTP 500. Will also retry on any `ConnectionError` (connection reset by peer, connection lost,...). Hopefully this should make the upload process more robust to transient errors.
`push_to_hub` is not robust to hub closing connection
### Describe the bug Like to #6172, `push_to_hub` will crash if Hub resets the connection and raise the following error: ``` Pushing dataset shards to the dataset hub: 32%|███▏ | 54/171 [06:38<14:23, 7.38s/it] Traceback (most recent call last): File "/admin/home-piraka9011/.virtualenvs/w2v2/lib/python3.8/...
35
`push_to_hub` is not robust to hub closing connection ### Describe the bug Like to #6172, `push_to_hub` will crash if Hub resets the connection and raise the following error: ``` Pushing dataset shards to the dataset hub: 32%|███▏ | 54/171 [06:38<14:23, 7.38s/it] Traceback (most recent call last): File ...
[ -0.5284816026687622, -0.3638116121292114, -0.06487172842025757, 0.1819303184747696, 0.3242587447166443, -0.20704427361488342, 0.17081235349178314, 0.3081148862838745, -0.03854190558195114, -0.022969290614128113, -0.19621890783309937, 0.18021035194396973, 0.1735050529241562, 0.0446992143988...
https://github.com/huggingface/datasets/issues/6392
I still get the same error, using `push_to_hub`. Using `git lfs` and pushing the files solved it for me.
`push_to_hub` is not robust to hub closing connection
### Describe the bug Like to #6172, `push_to_hub` will crash if Hub resets the connection and raise the following error: ``` Pushing dataset shards to the dataset hub: 32%|███▏ | 54/171 [06:38<14:23, 7.38s/it] Traceback (most recent call last): File "/admin/home-piraka9011/.virtualenvs/w2v2/lib/python3.8/...
19
`push_to_hub` is not robust to hub closing connection ### Describe the bug Like to #6172, `push_to_hub` will crash if Hub resets the connection and raise the following error: ``` Pushing dataset shards to the dataset hub: 32%|███▏ | 54/171 [06:38<14:23, 7.38s/it] Traceback (most recent call last): File ...
[ -0.5284816026687622, -0.3638116121292114, -0.06487172842025757, 0.1819303184747696, 0.3242587447166443, -0.20704427361488342, 0.17081235349178314, 0.3081148862838745, -0.03854190558195114, -0.022969290614128113, -0.19621890783309937, 0.18021035194396973, 0.1735050529241562, 0.0446992143988...
https://github.com/huggingface/datasets/issues/6392
@BEpresent the fix has not been released yet. You can expect a release of `huggingface_hub` (with this fix) today or tomorrow :)
`push_to_hub` is not robust to hub closing connection
### Describe the bug Like to #6172, `push_to_hub` will crash if Hub resets the connection and raise the following error: ``` Pushing dataset shards to the dataset hub: 32%|███▏ | 54/171 [06:38<14:23, 7.38s/it] Traceback (most recent call last): File "/admin/home-piraka9011/.virtualenvs/w2v2/lib/python3.8/...
22
`push_to_hub` is not robust to hub closing connection ### Describe the bug Like to #6172, `push_to_hub` will crash if Hub resets the connection and raise the following error: ``` Pushing dataset shards to the dataset hub: 32%|███▏ | 54/171 [06:38<14:23, 7.38s/it] Traceback (most recent call last): File ...
[ -0.5284816026687622, -0.3638116121292114, -0.06487172842025757, 0.1819303184747696, 0.3242587447166443, -0.20704427361488342, 0.17081235349178314, 0.3081148862838745, -0.03854190558195114, -0.022969290614128113, -0.19621890783309937, 0.18021035194396973, 0.1735050529241562, 0.0446992143988...
https://github.com/huggingface/datasets/issues/6389
I managed a workaround eventually but I don't know what it was (I made a lot of changes to seq2seq). I'll try to include generating code in the future. (If I close, I don't know if you see it. Feel free to close; I'll re-open if I encounter it again (if I can)).
Index 339 out of range for dataset of size 339 <-- save_to_file()
### Describe the bug When saving out some Audio() data. The data is audio recordings with associated 'sentences'. (They use the audio 'bytes' approach because they're clips within audio files). Code is below the traceback (I can't upload the voice audio/text (it's not even me)). ``` Traceback (most recent call ...
53
Index 339 out of range for dataset of size 339 <-- save_to_file() ### Describe the bug When saving out some Audio() data. The data is audio recordings with associated 'sentences'. (They use the audio 'bytes' approach because they're clips within audio files). Code is below the traceback (I can't upload the voice ...
[ -0.20405060052871704, -0.13199637830257416, 0.041832633316516876, 0.332999050617218, 0.16036248207092285, 0.04094874858856201, 0.1847858875989914, 0.642251193523407, -0.3258969485759735, 0.2800061106681824, -0.2070016860961914, -0.11669199913740158, -0.06349034607410431, -0.200472950935363...
https://github.com/huggingface/datasets/issues/6387
Feel free to use `dataset.save_to_disk(...)`, then scp the directory containing the saved dataset and reload it on your other machine using `dataset = load_from_disk(...)`
How to load existing downloaded dataset ?
Hi @mariosasko @lhoestq @katielink Thanks for your contribution and hard work. ### Feature request First, I download a dataset as normal by: ``` from datasets import load_dataset dataset = load_dataset('username/data_name', cache_dir='data') ``` The dataset format in `data` directory will be: ``` ...
24
How to load existing downloaded dataset ? Hi @mariosasko @lhoestq @katielink Thanks for your contribution and hard work. ### Feature request First, I download a dataset as normal by: ``` from datasets import load_dataset dataset = load_dataset('username/data_name', cache_dir='data') ``` The datase...
[ -0.38434135913848877, -0.18714748322963715, 0.02825199067592621, 0.2958897352218628, 0.0300179123878479, -0.0012773796916007996, 0.03767338767647743, 0.05174326151609421, 0.2660689651966095, 0.220025897026062, -0.3922737240791321, 0.22650328278541565, -0.003428516909480095, 0.7020697593688...
https://github.com/huggingface/datasets/issues/6386
Ah I think the `line-profiler` log is off-by-one and it is in fact the `extract_batch` method that's taking forever. Will investigate further.
Formatting overhead
### Describe the bug Hi! I very recently noticed that my training time is dominated by batch formatting. Using Lightning's profilers, I located the bottleneck within `datasets.formatting.formatting` and then narrowed it down with `line-profiler`. It turns out that almost all of the overhead is due to creating new inst...
22
Formatting overhead ### Describe the bug Hi! I very recently noticed that my training time is dominated by batch formatting. Using Lightning's profilers, I located the bottleneck within `datasets.formatting.formatting` and then narrowed it down with `line-profiler`. It turns out that almost all of the overhead is du...
[ -0.17785169184207916, -0.23078474402427673, 0.02506566047668457, 0.2501610517501831, 0.35222551226615906, -0.01755823940038681, 0.33603334426879883, 0.12893426418304443, -0.38654667139053345, -0.017012380063533783, 0.265005886554718, 0.41242772340774536, -0.4074646830558777, 0.148731440305...
https://github.com/huggingface/datasets/issues/6385
The `answers.text` field in the JSON dataset needs to be a list of strings, not a string. So, here is the fixed code: ```python from huggingface_hub import notebook_login from datasets import load_dataset notebook_login("mymailadresse", "mypassword") squad = load_dataset("squad", split="train[:5000]") sq...
Get an error when i try to concatenate the squad dataset with my own dataset
### Describe the bug Hello, I'm new here and I need to concatenate the squad dataset with my own dataset i created. I find the following error when i try to do it: Traceback (most recent call last): Cell In[9], line 1 concatenated_dataset = concatenate_datasets([train_dataset, dataset1]) File ~\ana...
126
Get an error when i try to concatenate the squad dataset with my own dataset ### Describe the bug Hello, I'm new here and I need to concatenate the squad dataset with my own dataset i created. I find the following error when i try to do it: Traceback (most recent call last): Cell In[9], line 1 concate...
[ -0.22193357348442078, -0.18991602957248688, 0.010590378195047379, 0.12742571532726288, 0.24476563930511475, 0.16857588291168213, 0.3418557345867157, 0.2883613705635071, -0.3683692514896393, -0.008571326732635498, -0.11954230815172195, 0.4824563264846802, 0.25596916675567627, 0.118566699326...
https://github.com/huggingface/datasets/issues/6382
Hey @SauravMaheshkar ! Just responded to your email. _For transparency, copying part of my response here:_ I agree, it would be really great to have this and other BenchMD datasets easily accessible on the hub. I think the main limiting factor is that the ChexPert dataset is currently hosted on the Stanford AIMI...
Add CheXpert dataset for vision
### Feature request ### Name **CheXpert: A Large Chest Radiograph Dataset with Uncertainty Labels and Expert Comparison** ### Paper https://arxiv.org/abs/1901.07031 ### Data https://stanfordaimi.azurewebsites.net/datasets/8cbd9ed4-2eb9-4565-affc-111cf4f7ebe2 ### Motivation CheXpert is one of the fund...
126
Add CheXpert dataset for vision ### Feature request ### Name **CheXpert: A Large Chest Radiograph Dataset with Uncertainty Labels and Expert Comparison** ### Paper https://arxiv.org/abs/1901.07031 ### Data https://stanfordaimi.azurewebsites.net/datasets/8cbd9ed4-2eb9-4565-affc-111cf4f7ebe2 ### Motiv...
[ -0.29031553864479065, -0.3038102984428406, -0.09949086606502533, -0.07802556455135345, -0.11682909727096558, 0.011799231171607971, 0.2542484998703003, 0.46563032269477844, 0.16077250242233276, 0.38705974817276, -0.0760834664106369, 0.09143735468387604, -0.2539331912994385, 0.44283682107925...
https://github.com/huggingface/datasets/issues/6382
Hi @katielink , I would also love to be on board and contribute to this loading script/project if it is still being developed. I'm interested because I personally would like to gain access to the CheXpert dataset and am facing some weird issues, so I'd like to sort it out for me, and potentially others. Please keep me ...
Add CheXpert dataset for vision
### Feature request ### Name **CheXpert: A Large Chest Radiograph Dataset with Uncertainty Labels and Expert Comparison** ### Paper https://arxiv.org/abs/1901.07031 ### Data https://stanfordaimi.azurewebsites.net/datasets/8cbd9ed4-2eb9-4565-affc-111cf4f7ebe2 ### Motivation CheXpert is one of the fund...
66
Add CheXpert dataset for vision ### Feature request ### Name **CheXpert: A Large Chest Radiograph Dataset with Uncertainty Labels and Expert Comparison** ### Paper https://arxiv.org/abs/1901.07031 ### Data https://stanfordaimi.azurewebsites.net/datasets/8cbd9ed4-2eb9-4565-affc-111cf4f7ebe2 ### Motiv...
[ -0.18617260456085205, -0.33047229051589966, -0.07788653671741486, -0.0753679946064949, -0.021564843133091927, -0.10620618611574173, 0.06507625430822372, 0.41927891969680786, 0.10648054629564285, 0.29680293798446655, 0.09117306023836136, 0.04682594910264015, -0.33251747488975525, 0.52854597...
https://github.com/huggingface/datasets/issues/6376
I did not store it at the time but I'll try to re-do a mwe next week to get it again
Caching problem when deleting a dataset
### Describe the bug Pushing a dataset with n + m features to a repo which was deleted, but contained n features, will fail. ### Steps to reproduce the bug 1. Create a dataset with n features per row 2. `dataset.push_to_hub(YOUR_PATH, SPLIT, token=TOKEN)` 3. Go on the hub, delete the repo at `YOUR_PATH` 4. Update...
21
Caching problem when deleting a dataset ### Describe the bug Pushing a dataset with n + m features to a repo which was deleted, but contained n features, will fail. ### Steps to reproduce the bug 1. Create a dataset with n features per row 2. `dataset.push_to_hub(YOUR_PATH, SPLIT, token=TOKEN)` 3. Go on the hub...
[ -0.17563989758491516, -0.15957209467887878, -0.042584508657455444, 0.3126749098300934, 0.01576206274330616, -0.0614800751209259, 0.2875654399394989, 0.2968783676624298, 0.19769355654716492, -0.022289790213108063, 0.17500968277454376, 0.17467544972896576, 0.00816064327955246, 0.181026279926...
https://github.com/huggingface/datasets/issues/6376
I haven't managed to reproduce this issue using a [notebook](https://colab.research.google.com/drive/1m6eduYun7pFTkigrCJAFgw0BghlbvXIL?usp=sharing) that follows the steps to reproduce the bug. So, I'm closing it. But feel free to re-open it if you have a better reproducer.
Caching problem when deleting a dataset
### Describe the bug Pushing a dataset with n + m features to a repo which was deleted, but contained n features, will fail. ### Steps to reproduce the bug 1. Create a dataset with n features per row 2. `dataset.push_to_hub(YOUR_PATH, SPLIT, token=TOKEN)` 3. Go on the hub, delete the repo at `YOUR_PATH` 4. Update...
34
Caching problem when deleting a dataset ### Describe the bug Pushing a dataset with n + m features to a repo which was deleted, but contained n features, will fail. ### Steps to reproduce the bug 1. Create a dataset with n features per row 2. `dataset.push_to_hub(YOUR_PATH, SPLIT, token=TOKEN)` 3. Go on the hub...
[ -0.09146799147129059, -0.13599282503128052, -0.007731638848781586, 0.3452052175998688, -0.07417566329240799, -0.020083405077457428, 0.3198857009410858, 0.2709159255027771, 0.19415250420570374, -0.043790534138679504, 0.20972475409507751, 0.16039246320724487, -0.02975871041417122, 0.13909795...
https://github.com/huggingface/datasets/issues/6371
Indeed, setting `try_from_gcs` to `False` makes sense for `from_generator`. We plan to deprecate and remove `try_from_hf_gcs` soon, as we can use Hub for file hosting now, but this is a good temporary fix.
`Dataset.from_generator` should not try to download from HF GCS
### Describe the bug When using [`Dataset.from_generator`](https://github.com/huggingface/datasets/blob/c9c1166e1cf81d38534020f9c167b326585339e5/src/datasets/arrow_dataset.py#L1072) with `streaming=False`, the internal logic will call [`download_and_prepare`](https://github.com/huggingface/datasets/blob/main/src/datas...
33
`Dataset.from_generator` should not try to download from HF GCS ### Describe the bug When using [`Dataset.from_generator`](https://github.com/huggingface/datasets/blob/c9c1166e1cf81d38534020f9c167b326585339e5/src/datasets/arrow_dataset.py#L1072) with `streaming=False`, the internal logic will call [`download_and_pre...
[ -0.24892577528953552, -0.3623090982437134, 0.012908365577459335, 0.22331032156944275, 0.23133710026741028, -0.19252106547355652, 0.11532336473464966, 0.07888228446245193, 0.08868780732154846, 0.28646352887153625, 0.08888961374759674, 0.050307776778936386, 0.14478136599063873, -0.0412699133...
https://github.com/huggingface/datasets/issues/6370
I figured it out. I found that `Trainer` does not work with TensorDataset even though the document says it uses it. Instead, I ended up creating a dictionary and converting it to a dataset using `dataset.Dataset.from_dict()`. I will leave this post open for a while. If someone knows a better approach, please leave a...
TensorDataset format does not work with Trainer from transformers
### Describe the bug The model was built to do fine tunning on BERT model for relation extraction. trainer.train() returns an error message ```TypeError: vars() argument must have __dict__ attribute``` when it has `train_dataset` generated from `torch.utils.data.TensorDataset` However, in the document, the req...
55
TensorDataset format does not work with Trainer from transformers ### Describe the bug The model was built to do fine tunning on BERT model for relation extraction. trainer.train() returns an error message ```TypeError: vars() argument must have __dict__ attribute``` when it has `train_dataset` generated from `...
[ -0.03787489980459213, -0.3518880009651184, 0.05201185494661331, 0.4861849844455719, 0.4192970097064972, -0.05809000879526138, 0.6545382142066956, 0.43648281693458557, -0.1592072993516922, -0.005945086479187012, -0.25915586948394775, 0.32222020626068115, -0.13267740607261658, -0.25136822462...
https://github.com/huggingface/datasets/issues/6370
Only issues directly related to the HF datasets library should be reported here. ~So, I'm transferring this issue to the `transformers` repo.~ I'm not a `transformers` maintainer, so GitHub doesn't let me transfer it there :(. This means you need to do it manually.
TensorDataset format does not work with Trainer from transformers
### Describe the bug The model was built to do fine tunning on BERT model for relation extraction. trainer.train() returns an error message ```TypeError: vars() argument must have __dict__ attribute``` when it has `train_dataset` generated from `torch.utils.data.TensorDataset` However, in the document, the req...
44
TensorDataset format does not work with Trainer from transformers ### Describe the bug The model was built to do fine tunning on BERT model for relation extraction. trainer.train() returns an error message ```TypeError: vars() argument must have __dict__ attribute``` when it has `train_dataset` generated from `...
[ -0.03787489980459213, -0.3518880009651184, 0.05201185494661331, 0.4861849844455719, 0.4192970097064972, -0.05809000879526138, 0.6545382142066956, 0.43648281693458557, -0.1592072993516922, -0.005945086479187012, -0.25915586948394775, 0.32222020626068115, -0.13267740607261658, -0.25136822462...
https://github.com/huggingface/datasets/issues/6369
The inconsistency may be caused by the usage of "update_fingerprint" and setting "trust_remote_code" to "True." When the tokenizer employs "trust_remote_code," the behavior of the map function varies with each code execution. Even if the remote code of the tokenizer remains the same, the result of "asher.hexdigest()" ...
Multi process map did not load cache file correctly
### Describe the bug When I was training model on Multiple GPUs by DDP, the dataset is tokenized multiple times after main process. ![1698820541284](https://github.com/huggingface/datasets/assets/14285786/0b2fe054-54d8-4e00-96e6-6ca5b69e662b) ![1698820501568](https://github.com/huggingface/datasets/assets/142857...
64
Multi process map did not load cache file correctly ### Describe the bug When I was training model on Multiple GPUs by DDP, the dataset is tokenized multiple times after main process. ![1698820541284](https://github.com/huggingface/datasets/assets/14285786/0b2fe054-54d8-4e00-96e6-6ca5b69e662b) ![1698820501568]...
[ -0.3758842647075653, -0.5878843069076538, 0.032787516713142395, 0.44652554392814636, 0.30833250284194946, -0.06708855926990509, 0.3846704661846161, -0.0230244230479002, 0.11546281725168228, -0.130690798163414, 0.03220471739768982, 0.1839643120765686, 0.011365003883838654, -0.16018880903720...
https://github.com/huggingface/datasets/issues/6369
The issue may be related to problems previously discussed in GitHub issues [#3847](https://github.com/huggingface/datasets/issues/3847) and [#6318](https://github.com/huggingface/datasets/pull/6318). This arises from the fact that tokenizer.tokens_trie._tokens is an unordered set, leading to varying hash results: `v...
Multi process map did not load cache file correctly
### Describe the bug When I was training model on Multiple GPUs by DDP, the dataset is tokenized multiple times after main process. ![1698820541284](https://github.com/huggingface/datasets/assets/14285786/0b2fe054-54d8-4e00-96e6-6ca5b69e662b) ![1698820501568](https://github.com/huggingface/datasets/assets/142857...
71
Multi process map did not load cache file correctly ### Describe the bug When I was training model on Multiple GPUs by DDP, the dataset is tokenized multiple times after main process. ![1698820541284](https://github.com/huggingface/datasets/assets/14285786/0b2fe054-54d8-4e00-96e6-6ca5b69e662b) ![1698820501568]...
[ -0.3758842647075653, -0.5878843069076538, 0.032787516713142395, 0.44652554392814636, 0.30833250284194946, -0.06708855926990509, 0.3846704661846161, -0.0230244230479002, 0.11546281725168228, -0.130690798163414, 0.03220471739768982, 0.1839643120765686, 0.011365003883838654, -0.16018880903720...
https://github.com/huggingface/datasets/issues/6369
We now sort `set` and `dict` items to make their hashes deterministic (install from `main` with `pip install git+https://github.com/huggingface/datasets` to test this). Consequently, this should also make the `tokenizer.tokens_trie`'s hash deterministic. Feel free to re-open the issue if this is not the case.
Multi process map did not load cache file correctly
### Describe the bug When I was training model on Multiple GPUs by DDP, the dataset is tokenized multiple times after main process. ![1698820541284](https://github.com/huggingface/datasets/assets/14285786/0b2fe054-54d8-4e00-96e6-6ca5b69e662b) ![1698820501568](https://github.com/huggingface/datasets/assets/142857...
43
Multi process map did not load cache file correctly ### Describe the bug When I was training model on Multiple GPUs by DDP, the dataset is tokenized multiple times after main process. ![1698820541284](https://github.com/huggingface/datasets/assets/14285786/0b2fe054-54d8-4e00-96e6-6ca5b69e662b) ![1698820501568]...
[ -0.3758842647075653, -0.5878843069076538, 0.032787516713142395, 0.44652554392814636, 0.30833250284194946, -0.06708855926990509, 0.3846704661846161, -0.0230244230479002, 0.11546281725168228, -0.130690798163414, 0.03220471739768982, 0.1839643120765686, 0.011365003883838654, -0.16018880903720...
https://github.com/huggingface/datasets/issues/6364
You can use the following code to load this CSV with the list values preserved: ```python from datasets import load_dataset import ast converters = { "contexts" : ast.literal_eval, "ground_truths" : ast.literal_eval, } ds = load_dataset("csv", data_files="golden_dataset.csv", converters=converters) ...
ArrowNotImplementedError: Unsupported cast from string to list using function cast_list
Hi, I am trying to load a local csv dataset(similar to explodinggradients_fiqa) using load_dataset. When I try to pass features, I am facing the mentioned issue. CSV Data sample(golden_dataset.csv): Question | Context | answer | groundtruth "what is abc?"...
38
ArrowNotImplementedError: Unsupported cast from string to list using function cast_list Hi, I am trying to load a local csv dataset(similar to explodinggradients_fiqa) using load_dataset. When I try to pass features, I am facing the mentioned issue. CSV Data sample(golden_dataset.csv): Question | ...
[ -0.02789292484521866, -0.23731307685375214, -0.11266941577196121, 0.09196636080741882, 0.672694742679596, -0.07634066045284271, 0.1983177810907364, 0.24670083820819855, 0.28991514444351196, 0.11488257348537445, 0.06941263377666473, 0.47068384289741516, -0.139689639210701, 0.041423432528972...
https://github.com/huggingface/datasets/issues/6363
I think the code hangs on the `accelerator.main_process_first()` context manager exit. To verify this, you can append a print statement to the end of the `accelerator.main_process_first()` block. If the problem is in `with_transform`, it would help if you could share the error stack trace printed when you interr...
dataset.transform() hangs indefinitely while finetuning the stable diffusion XL
### Describe the bug Multi-GPU fine-tuning the stable diffusion X by following https://github.com/huggingface/diffusers/blob/main/examples/text_to_image/README_sdxl.md hangs indefinitely. ### Steps to reproduce the bug accelerate launch train_text_to_image_sdxl.py --pretrained_model_name_or_path=$MODEL_NAME --...
53
dataset.transform() hangs indefinitely while finetuning the stable diffusion XL ### Describe the bug Multi-GPU fine-tuning the stable diffusion X by following https://github.com/huggingface/diffusers/blob/main/examples/text_to_image/README_sdxl.md hangs indefinitely. ### Steps to reproduce the bug accelerate...
[ -0.29093098640441895, -0.4494297206401825, 0.03063829243183136, 0.10455924272537231, 0.7972444295883179, -0.06378263235092163, 0.38594162464141846, 0.08792893588542938, -0.19314613938331604, 0.4078892171382904, 0.08595020323991776, 0.15589837729930878, 0.01676591858267784, -0.0726515948772...
https://github.com/huggingface/datasets/issues/6363
@mariosasko yes the problem seems to be to exit from accelerator.main_process_first(). Is there any known problem?
dataset.transform() hangs indefinitely while finetuning the stable diffusion XL
### Describe the bug Multi-GPU fine-tuning the stable diffusion X by following https://github.com/huggingface/diffusers/blob/main/examples/text_to_image/README_sdxl.md hangs indefinitely. ### Steps to reproduce the bug accelerate launch train_text_to_image_sdxl.py --pretrained_model_name_or_path=$MODEL_NAME --...
16
dataset.transform() hangs indefinitely while finetuning the stable diffusion XL ### Describe the bug Multi-GPU fine-tuning the stable diffusion X by following https://github.com/huggingface/diffusers/blob/main/examples/text_to_image/README_sdxl.md hangs indefinitely. ### Steps to reproduce the bug accelerate...
[ -0.29093098640441895, -0.4494297206401825, 0.03063829243183136, 0.10455924272537231, 0.7972444295883179, -0.06378263235092163, 0.38594162464141846, 0.08792893588542938, -0.19314613938331604, 0.4078892171382904, 0.08595020323991776, 0.15589837729930878, 0.01676591858267784, -0.0726515948772...
https://github.com/huggingface/datasets/issues/6363
NCCL debug info I get below output, if it helps. ``` 11/09/2023 13:36:44 - INFO - __main__ - Distributed environment: MULTI_GPU Backend: nccl Num processes: 2 Process index: 1 Local process index: 1 Device: cuda:1 Mixed precision type: fp16 Detected kernel version 5.4.0, which is below the recommended mini...
dataset.transform() hangs indefinitely while finetuning the stable diffusion XL
### Describe the bug Multi-GPU fine-tuning the stable diffusion X by following https://github.com/huggingface/diffusers/blob/main/examples/text_to_image/README_sdxl.md hangs indefinitely. ### Steps to reproduce the bug accelerate launch train_text_to_image_sdxl.py --pretrained_model_name_or_path=$MODEL_NAME --...
620
dataset.transform() hangs indefinitely while finetuning the stable diffusion XL ### Describe the bug Multi-GPU fine-tuning the stable diffusion X by following https://github.com/huggingface/diffusers/blob/main/examples/text_to_image/README_sdxl.md hangs indefinitely. ### Steps to reproduce the bug accelerate...
[ -0.29093098640441895, -0.4494297206401825, 0.03063829243183136, 0.10455924272537231, 0.7972444295883179, -0.06378263235092163, 0.38594162464141846, 0.08792893588542938, -0.19314613938331604, 0.4078892171382904, 0.08595020323991776, 0.15589837729930878, 0.01676591858267784, -0.0726515948772...
https://github.com/huggingface/datasets/issues/6363
I don't know what the issue was, but after going through the thread here I loved the issue with https://github.com/huggingface/accelerate/issues/314#issuecomment-1565259831
dataset.transform() hangs indefinitely while finetuning the stable diffusion XL
### Describe the bug Multi-GPU fine-tuning the stable diffusion X by following https://github.com/huggingface/diffusers/blob/main/examples/text_to_image/README_sdxl.md hangs indefinitely. ### Steps to reproduce the bug accelerate launch train_text_to_image_sdxl.py --pretrained_model_name_or_path=$MODEL_NAME --...
20
dataset.transform() hangs indefinitely while finetuning the stable diffusion XL ### Describe the bug Multi-GPU fine-tuning the stable diffusion X by following https://github.com/huggingface/diffusers/blob/main/examples/text_to_image/README_sdxl.md hangs indefinitely. ### Steps to reproduce the bug accelerate...
[ -0.29093098640441895, -0.4494297206401825, 0.03063829243183136, 0.10455924272537231, 0.7972444295883179, -0.06378263235092163, 0.38594162464141846, 0.08792893588542938, -0.19314613938331604, 0.4078892171382904, 0.08595020323991776, 0.15589837729930878, 0.01676591858267784, -0.0726515948772...
https://github.com/huggingface/datasets/issues/6360
This issue stems from https://github.com/huggingface/datasets/blob/6d2f2a5e0fea3827eccfd1717d8021c15fc4292a/src/datasets/table.py#L2203-L2205 I'll address it as part of https://github.com/huggingface/datasets/pull/6283. In the meantime, this should work ```python import pyarrow as pa from datasets import Ima...
Add support for `Sequence(Audio/Image)` feature in `push_to_hub`
### Feature request Allow for `Sequence` of `Image` (or `Audio`) to be embedded inside the shards. ### Motivation Currently, thanks to #3685, when `embed_external_files` is set to True (which is the default) in `push_to_hub`, features of type `Image` and `Audio` are embedded inside the arrow/parquet shards, instead ...
58
Add support for `Sequence(Audio/Image)` feature in `push_to_hub` ### Feature request Allow for `Sequence` of `Image` (or `Audio`) to be embedded inside the shards. ### Motivation Currently, thanks to #3685, when `embed_external_files` is set to True (which is the default) in `push_to_hub`, features of type `Image...
[ -0.255060076713562, -0.5069053173065186, -0.11654671281576157, 0.14346259832382202, -0.07700437307357788, -0.09264026582241058, 0.1639629453420639, 0.2216341495513916, -0.26074954867362976, 0.17967505753040314, 0.046998776495456696, 0.26361843943595886, -0.4362025260925293, 0.1619111001491...
https://github.com/huggingface/datasets/issues/6359
Most likely, the data file inference logic is the problem here. You can run the following code to verify this: ```python import time from datasets.data_files import get_data_patterns start_time = time.time() get_data_patterns("/path/to/img_dir") end_time = time.time() print(f"Elapsed time: {end_time - start_t...
Stuck in "Resolving data files..."
### Describe the bug I have an image dataset with 300k images, the size of image is 768 * 768. When I run `dataset = load_dataset("imagefolder", data_dir="/path/to/img_dir", split='train')` in second time, it takes 50 minutes to finish "Resolving data files" part, what's going on in this part? From my understa...
74
Stuck in "Resolving data files..." ### Describe the bug I have an image dataset with 300k images, the size of image is 768 * 768. When I run `dataset = load_dataset("imagefolder", data_dir="/path/to/img_dir", split='train')` in second time, it takes 50 minutes to finish "Resolving data files" part, what's going...
[ -0.23090443015098572, 0.04058697819709778, -0.10604532063007355, 0.36753690242767334, 0.2609904706478119, 0.13373884558677673, 0.34540820121765137, 0.47421953082084656, -0.10366508364677429, 0.11970315873622894, 0.16956403851509094, 0.3559846878051758, 0.1605689525604248, -0.18263012170791...
https://github.com/huggingface/datasets/issues/6359
Hi, @mariosasko, you are right; data file inference logic is extremely slow. I have done a similar test, that is I modify the source code of datasets/load.py to measure the cost of two suspicious operations: ```python def get_module(self) -> DatasetModule: base_path = Path(self.data_dir or "").expanduser(...
Stuck in "Resolving data files..."
### Describe the bug I have an image dataset with 300k images, the size of image is 768 * 768. When I run `dataset = load_dataset("imagefolder", data_dir="/path/to/img_dir", split='train')` in second time, it takes 50 minutes to finish "Resolving data files" part, what's going on in this part? From my understa...
99
Stuck in "Resolving data files..." ### Describe the bug I have an image dataset with 300k images, the size of image is 768 * 768. When I run `dataset = load_dataset("imagefolder", data_dir="/path/to/img_dir", split='train')` in second time, it takes 50 minutes to finish "Resolving data files" part, what's going...
[ -0.2094925194978714, 0.1629522740840912, -0.089963898062706, 0.4162213206291199, 0.14655184745788574, 0.10716225951910019, 0.2901347577571869, 0.44508206844329834, -0.08686953037977219, 0.10743442177772522, 0.11540091037750244, 0.3833844065666199, 0.20739655196666718, -0.2878902554512024, ...
https://github.com/huggingface/datasets/issues/6359
Furthermore, what's my current workaround about this problem? Should I save it by `save_to_disk()` and load dataset through `load_from_disk`?
Stuck in "Resolving data files..."
### Describe the bug I have an image dataset with 300k images, the size of image is 768 * 768. When I run `dataset = load_dataset("imagefolder", data_dir="/path/to/img_dir", split='train')` in second time, it takes 50 minutes to finish "Resolving data files" part, what's going on in this part? From my understa...
19
Stuck in "Resolving data files..." ### Describe the bug I have an image dataset with 300k images, the size of image is 768 * 768. When I run `dataset = load_dataset("imagefolder", data_dir="/path/to/img_dir", split='train')` in second time, it takes 50 minutes to finish "Resolving data files" part, what's going...
[ -0.2394724190235138, -0.006858140230178833, -0.12425383925437927, 0.3820543885231018, 0.29293784499168396, 0.18949246406555176, 0.26252657175064087, 0.40455177426338196, -0.054701972752809525, 0.03535822778940201, 0.22675317525863647, 0.3705781102180481, 0.1320250928401947, -0.192263275384...
https://github.com/huggingface/datasets/issues/6358
You may be able to make it work by tweaking some environment variables, such as [`HF_HOME`](https://huggingface.co/docs/huggingface_hub/main/en/package_reference/environment_variables#hfhome) or [`HF_DATASETS_CACHE`](https://huggingface.co/docs/datasets/cache#cache-directory).
Mounting datasets cache fails due to absolute paths.
### Describe the bug Creating a datasets cache and mounting this into, for example, a docker container, renders the data unreadable due to absolute paths written into the cache. ### Steps to reproduce the bug 1. Create a datasets cache by downloading some data 2. Mount the dataset folder into a docker contain...
18
Mounting datasets cache fails due to absolute paths. ### Describe the bug Creating a datasets cache and mounting this into, for example, a docker container, renders the data unreadable due to absolute paths written into the cache. ### Steps to reproduce the bug 1. Create a datasets cache by downloading some ...
[ -0.00010085105895996094, 0.028453439474105835, 0.08273055404424667, 0.13505060970783234, 0.24859587848186493, 0.023465223610401154, 0.1194758415222168, 0.21675515174865723, 0.049085840582847595, 0.09170372039079666, -0.07647693157196045, 0.08717919886112213, -0.2398337721824646, -0.3364964...
https://github.com/huggingface/datasets/issues/6358
> You may be able to make it work by tweaking some environment variables, such as [`HF_HOME`](https://huggingface.co/docs/huggingface_hub/main/en/package_reference/environment_variables#hfhome) or [`HF_DATASETS_CACHE`](https://huggingface.co/docs/datasets/cache#cache-directory). I am already doing this. The problem ...
Mounting datasets cache fails due to absolute paths.
### Describe the bug Creating a datasets cache and mounting this into, for example, a docker container, renders the data unreadable due to absolute paths written into the cache. ### Steps to reproduce the bug 1. Create a datasets cache by downloading some data 2. Mount the dataset folder into a docker contain...
78
Mounting datasets cache fails due to absolute paths. ### Describe the bug Creating a datasets cache and mounting this into, for example, a docker container, renders the data unreadable due to absolute paths written into the cache. ### Steps to reproduce the bug 1. Create a datasets cache by downloading some ...
[ 0.053802065551280975, 0.05613492429256439, 0.08096794784069061, 0.06797824800014496, 0.2428542971611023, -0.015263326466083527, 0.10982009768486023, 0.20520839095115662, 0.03088117390871048, 0.05655992776155472, -0.1255326271057129, 0.11960466206073761, -0.24449720978736877, -0.37721857428...
https://github.com/huggingface/datasets/issues/6358
I'm unable to reproduce this with the cache ```bash export HF_CACHE=$PWD/hf_cache python -c "import datasets; datasets.load_dataset('imdb')" ``` imported inside a dummy container that is built from ```bash FROM python:3.9 WORKDIR /usr/src/app RUN pip install datasets COPY ./hf_cache ./hf_cache ENV HF...
Mounting datasets cache fails due to absolute paths.
### Describe the bug Creating a datasets cache and mounting this into, for example, a docker container, renders the data unreadable due to absolute paths written into the cache. ### Steps to reproduce the bug 1. Create a datasets cache by downloading some data 2. Mount the dataset folder into a docker contain...
73
Mounting datasets cache fails due to absolute paths. ### Describe the bug Creating a datasets cache and mounting this into, for example, a docker container, renders the data unreadable due to absolute paths written into the cache. ### Steps to reproduce the bug 1. Create a datasets cache by downloading some ...
[ 0.013612456619739532, 0.28743523359298706, 0.03356702998280525, 0.12142452597618103, 0.17181266844272614, 0.09251341223716736, 0.315844863653183, 0.2463112622499466, 0.07502992451190948, 0.018458478152751923, 0.02460268884897232, 0.11198748648166656, -0.14733318984508514, -0.64198601245880...
https://github.com/huggingface/datasets/issues/6354
I am having issues as well with this. However, the error I am getting is : `RuntimeError: It appears that you are attempting to reference SparkContext from a broadcast variable, action, or transformation. SparkContext can only be used on the driver, not in code that it run on workers. For more information, see SPA...
`IterableDataset.from_spark` does not support multiple workers in pytorch `Dataloader`
### Describe the bug Looks like `IterableDataset.from_spark` does not support multiple workers in pytorch `Dataloader` if I'm not missing anything. Also, returns not consistent error messages, which probably depend on the nondeterministic order of worker executions Some exampes I've encountered: ``` File "/l...
62
`IterableDataset.from_spark` does not support multiple workers in pytorch `Dataloader` ### Describe the bug Looks like `IterableDataset.from_spark` does not support multiple workers in pytorch `Dataloader` if I'm not missing anything. Also, returns not consistent error messages, which probably depend on the nonde...
[ -0.28735220432281494, -0.16247272491455078, -0.06743581593036652, 0.5066515207290649, -0.03184495121240616, -0.3001953959465027, 0.7239212393760681, -0.012372683733701706, -0.012353334575891495, 0.030812911689281464, 0.008178908377885818, 0.445117712020874, -0.003525044769048691, 0.0159480...
https://github.com/huggingface/datasets/issues/6353
I'm using the latest datasets and fsspec , but still got this error! datasets : Version: 2.13.0 fsspec Version: 2023.10.0 ``` File "/home/guoby/app/Anaconda3-2021.05/envs/news/lib/python3.8/site-packages/datasets/load.py", line 1892, in load_from_disk return DatasetDict.load_from_disk(dataset_path, ke...
load_dataset save_to_disk load_from_disk error
### Describe the bug datasets version: 2.10.1 I `load_dataset `and `save_to_disk` sucessfully on windows10( **and I `load_from_disk(/LLM/data/wiki)` succcesfully on windows10**), and I copy the dataset `/LLM/data/wiki` into a ubuntu system, but when I `load_from_disk(/LLM/data/wiki)` on ubuntu, something weird ha...
83
load_dataset save_to_disk load_from_disk error ### Describe the bug datasets version: 2.10.1 I `load_dataset `and `save_to_disk` sucessfully on windows10( **and I `load_from_disk(/LLM/data/wiki)` succcesfully on windows10**), and I copy the dataset `/LLM/data/wiki` into a ubuntu system, but when I `load_from_dis...
[ -0.3020021915435791, 0.007541164755821228, 0.07424850016832352, 0.35470154881477356, 0.38531994819641113, 0.19839975237846375, 0.30810630321502686, 0.09688099473714828, 0.038030557334423065, 0.08450580388307571, -0.02063780650496483, 0.3757515847682953, 0.020782995969057083, 0.058200322091...
https://github.com/huggingface/datasets/issues/6352
+1 ``` Found cached dataset csv (file:///home/ubuntu/.cache/huggingface/datasets/theSquarePond___csv/theSquarePond--XXXXX-bbf0a8365d693d2c/0.0.0/eea64c71ca8b46dd3f537ed218fc9bf495d5707789152eb2764f5c78fa66d59d) --------------------------------------------------------------------------- NotImplementedError ...
Error loading wikitext data raise NotImplementedError(f"Loading a dataset cached in a {type(self._fs).__name__} is not supported.")
I was trying to load the wiki dataset, but i got this error traindata = load_dataset('wikitext', 'wikitext-2-raw-v1', split='train') File "/home/aelkordy/.conda/envs/prune_llm/lib/python3.9/site-packages/datasets/load.py", line 1804, in load_dataset ds = builder_instance.as_dataset(split=split, verific...
183
Error loading wikitext data raise NotImplementedError(f"Loading a dataset cached in a {type(self._fs).__name__} is not supported.") I was trying to load the wiki dataset, but i got this error traindata = load_dataset('wikitext', 'wikitext-2-raw-v1', split='train') File "/home/aelkordy/.conda/envs/prune...
[ -0.26059049367904663, -0.20326289534568787, -0.04139687120914459, 0.2912045121192932, 0.2911190390586853, 0.3084626793861389, 0.23816388845443726, 0.41730719804763794, 0.4236239492893219, -0.02376360446214676, 0.10076349973678589, 0.09992368519306183, -0.06720319390296936, -0.0004816353321...
https://github.com/huggingface/datasets/issues/6352
+1 ``` Found cached dataset csv ([file://C:/Users/Shady/.cache/huggingface/datasets/knkarthick___csv/knkarthick--dialogsum-cd36827d3490488d/0.0.0/6954658bab30a358235fa864b05cf819af0e179325c740e4bc853bcc7ec513e1](file:///C:/Users/Shady/.cache/huggingface/datasets/knkarthick___csv/knkarthick--dialogsum-cd36827d349048...
Error loading wikitext data raise NotImplementedError(f"Loading a dataset cached in a {type(self._fs).__name__} is not supported.")
I was trying to load the wiki dataset, but i got this error traindata = load_dataset('wikitext', 'wikitext-2-raw-v1', split='train') File "/home/aelkordy/.conda/envs/prune_llm/lib/python3.9/site-packages/datasets/load.py", line 1804, in load_dataset ds = builder_instance.as_dataset(split=split, verific...
175
Error loading wikitext data raise NotImplementedError(f"Loading a dataset cached in a {type(self._fs).__name__} is not supported.") I was trying to load the wiki dataset, but i got this error traindata = load_dataset('wikitext', 'wikitext-2-raw-v1', split='train') File "/home/aelkordy/.conda/envs/prune...
[ -0.22216784954071045, -0.21753501892089844, -0.05269908159971237, 0.3149651288986206, 0.2889442443847656, 0.3062695264816284, 0.2462485432624817, 0.4402695596218109, 0.4070063829421997, -0.022959955036640167, 0.08359038084745407, 0.12875205278396606, -0.07824663817882538, 0.017831921577453...
https://github.com/huggingface/datasets/issues/6352
This error stems from a breaking change in `fsspec`. It has been fixed in the latest `datasets` release (`2.14.6`). Updating the installation with `pip install -U datasets` should fix the issue.
Error loading wikitext data raise NotImplementedError(f"Loading a dataset cached in a {type(self._fs).__name__} is not supported.")
I was trying to load the wiki dataset, but i got this error traindata = load_dataset('wikitext', 'wikitext-2-raw-v1', split='train') File "/home/aelkordy/.conda/envs/prune_llm/lib/python3.9/site-packages/datasets/load.py", line 1804, in load_dataset ds = builder_instance.as_dataset(split=split, verific...
31
Error loading wikitext data raise NotImplementedError(f"Loading a dataset cached in a {type(self._fs).__name__} is not supported.") I was trying to load the wiki dataset, but i got this error traindata = load_dataset('wikitext', 'wikitext-2-raw-v1', split='train') File "/home/aelkordy/.conda/envs/prune...
[ -0.36869052052497864, -0.10481400042772293, 0.001288868486881256, 0.1946215033531189, 0.27150219678878784, 0.0910446047782898, 0.062023088335990906, 0.3505300283432007, 0.38558250665664673, 0.03684042766690254, 0.15493622422218323, 0.20195260643959045, -0.050123684108257294, 0.288501232862...
https://github.com/huggingface/datasets/issues/6350
`load_dataset` returns a `DatasetDict` object unless `split` is defined, in which case it returns a `Dataset` (or a list of datasets if `split` is a list). We've discussed dropping `DatasetDict` from the API in https://github.com/huggingface/datasets/issues/5189 to always return the same type in `load_dataset` and sup...
Different objects are returned from calls that should be returning the same kind of object.
### Describe the bug 1. dataset = load_dataset("togethercomputer/RedPajama-Data-1T-Sample", cache_dir=training_args.cache_dir, split='train[:1%]') 2. dataset = load_dataset("togethercomputer/RedPajama-Data-1T-Sample", cache_dir=training_args.cache_dir) The only difference I would expect these cal...
82
Different objects are returned from calls that should be returning the same kind of object. ### Describe the bug 1. dataset = load_dataset("togethercomputer/RedPajama-Data-1T-Sample", cache_dir=training_args.cache_dir, split='train[:1%]') 2. dataset = load_dataset("togethercomputer/RedPajama-Data-...
[ 0.065458744764328, -0.35832369327545166, -0.047362830489873886, 0.541680097579956, 0.15500470995903015, -0.27208226919174194, 0.2597043514251709, 0.172197625041008, 0.10662220418453217, 0.15586790442466736, -0.15613983571529388, 0.31378409266471863, -0.04439716413617134, 0.3652834892272949...
https://github.com/huggingface/datasets/issues/6350
> `load_dataset` returns a `DatasetDict` object unless `split` is defined, in which case it returns a `Dataset` (or a list of datasets if `split` is a list). We've discussed dropping `DatasetDict` from the API in #5189 to always return the same type in `load_dataset` and support datasets without (explicit) splits. IIRC...
Different objects are returned from calls that should be returning the same kind of object.
### Describe the bug 1. dataset = load_dataset("togethercomputer/RedPajama-Data-1T-Sample", cache_dir=training_args.cache_dir, split='train[:1%]') 2. dataset = load_dataset("togethercomputer/RedPajama-Data-1T-Sample", cache_dir=training_args.cache_dir) The only difference I would expect these cal...
153
Different objects are returned from calls that should be returning the same kind of object. ### Describe the bug 1. dataset = load_dataset("togethercomputer/RedPajama-Data-1T-Sample", cache_dir=training_args.cache_dir, split='train[:1%]') 2. dataset = load_dataset("togethercomputer/RedPajama-Data-...
[ 0.0666290819644928, -0.18669439852237701, -0.08955486118793488, 0.5604851841926575, 0.058909911662340164, -0.2986440658569336, 0.28418415784835815, 0.21379944682121277, 0.12141849100589752, 0.16319358348846436, -0.017049366608262062, 0.4322671890258789, -0.04269466549158096, 0.269218504428...
https://github.com/huggingface/datasets/issues/6349
I'm unable to reproduce this error. The server hosting the files may have been down temporarily, so try again.
Can't load ds = load_dataset("imdb")
### Describe the bug I did `from datasets import load_dataset, load_metric` and then `ds = load_dataset("imdb")` and it gave me the error: ExpectedMoreDownloadedFiles: {'http://ai.stanford.edu/~amaas/data/sentiment/aclImdb_v1.tar.gz'} I tried doing `ds = load_dataset("imdb",download_mode="force_redownload")` as we...
19
Can't load ds = load_dataset("imdb") ### Describe the bug I did `from datasets import load_dataset, load_metric` and then `ds = load_dataset("imdb")` and it gave me the error: ExpectedMoreDownloadedFiles: {'http://ai.stanford.edu/~amaas/data/sentiment/aclImdb_v1.tar.gz'} I tried doing `ds = load_dataset("imdb",d...
[ -0.3720935881137848, 0.034979745745658875, -0.05936967581510544, 0.5213468670845032, 0.41254252195358276, 0.1490505039691925, 0.34063592553138733, 0.1281067281961441, 0.6659700274467468, 0.07853162288665771, -0.1272823065519333, 0.14698487520217896, -0.03618531674146652, 0.0683325678110122...
https://github.com/huggingface/datasets/issues/6347
This was fixed in https://github.com/huggingface/datasets/pull/6247. You can find the fix in the `main` version of the docs
Incorrect example code in 'Create a dataset' docs
### Describe the bug On [this](https://huggingface.co/docs/datasets/create_dataset) page, the example code for loading in images and audio is incorrect. Currently, examples are: ``` python from datasets import ImageFolder dataset = load_dataset("imagefolder", data_dir="/path/to/pokemon") ``` and ``` python...
17
Incorrect example code in 'Create a dataset' docs ### Describe the bug On [this](https://huggingface.co/docs/datasets/create_dataset) page, the example code for loading in images and audio is incorrect. Currently, examples are: ``` python from datasets import ImageFolder dataset = load_dataset("imagefolder",...
[ 0.10520012676715851, -0.3185540735721588, -0.02481408417224884, 0.494010865688324, 0.09755562245845795, -0.02400396764278412, 0.3480840027332306, -0.08572138100862503, 0.09643727540969849, 0.09243978559970856, -0.1433936357498169, 0.17106789350509644, -0.07363775372505188, 0.32302072644233...
https://github.com/huggingface/datasets/issues/6330
I also encountered a similar error below. Appreciate the team could shed some light on this issue. ``` --------------------------------------------------------------------------- NotImplementedError Traceback (most recent call last) [/home/ubuntu/work/EveryDream2trainer/prepare_dataset.ipyn...
Latest fsspec==2023.10.0 issue with streaming datasets
### Describe the bug Loading a streaming dataset with this version of fsspec fails with the following error: `NotImplementedError: Loading a streaming dataset cached in a LocalFileSystem is not supported yet.` I suspect the issue is with this PR https://github.com/fsspec/filesystem_spec/pull/1381 ### Steps ...
588
Latest fsspec==2023.10.0 issue with streaming datasets ### Describe the bug Loading a streaming dataset with this version of fsspec fails with the following error: `NotImplementedError: Loading a streaming dataset cached in a LocalFileSystem is not supported yet.` I suspect the issue is with this PR https:...
[ -0.3962833285331726, -0.05229148268699646, 0.0895015299320221, 0.22059676051139832, 0.20769593119621277, -0.055554941296577454, 0.11532916873693466, 0.1682538241147995, -0.1815401017665863, 0.2266596555709839, -0.04434163123369217, 0.2020154893398285, -0.05888967961072922, 0.46524885296821...
https://github.com/huggingface/datasets/issues/6330
Thanks for reporting and for the investigation, @ZachNagengast! :hugs: We are investigating the root cause of the issue. In the meantime, we are going to pin fsspec < 2023.10.0.
Latest fsspec==2023.10.0 issue with streaming datasets
### Describe the bug Loading a streaming dataset with this version of fsspec fails with the following error: `NotImplementedError: Loading a streaming dataset cached in a LocalFileSystem is not supported yet.` I suspect the issue is with this PR https://github.com/fsspec/filesystem_spec/pull/1381 ### Steps ...
29
Latest fsspec==2023.10.0 issue with streaming datasets ### Describe the bug Loading a streaming dataset with this version of fsspec fails with the following error: `NotImplementedError: Loading a streaming dataset cached in a LocalFileSystem is not supported yet.` I suspect the issue is with this PR https:...
[ -0.44810670614242554, -0.09368070214986801, 0.07924748212099075, 0.2137065976858139, 0.21421100199222565, -0.08358780294656754, 0.07893026620149612, 0.21099360287189484, -0.11809461563825607, 0.2367476224899292, -0.023853646591305733, 0.15875737369060516, -0.07089647650718689, 0.4641267061...
https://github.com/huggingface/datasets/issues/6330
You can also update `datasets`: ``` pip install -U datasets ``` It will also update `fsspec` to use the right version
Latest fsspec==2023.10.0 issue with streaming datasets
### Describe the bug Loading a streaming dataset with this version of fsspec fails with the following error: `NotImplementedError: Loading a streaming dataset cached in a LocalFileSystem is not supported yet.` I suspect the issue is with this PR https://github.com/fsspec/filesystem_spec/pull/1381 ### Steps ...
21
Latest fsspec==2023.10.0 issue with streaming datasets ### Describe the bug Loading a streaming dataset with this version of fsspec fails with the following error: `NotImplementedError: Loading a streaming dataset cached in a LocalFileSystem is not supported yet.` I suspect the issue is with this PR https:...
[ -0.48434990644454956, -0.09176057577133179, 0.06091972813010216, 0.20084726810455322, 0.20000611245632172, -0.06782572716474533, 0.050473786890506744, 0.2100011259317398, -0.0857999175786972, 0.21352194249629974, -0.019064685329794884, 0.21504412591457367, -0.056027695536613464, 0.46829530...
https://github.com/huggingface/datasets/issues/6327
You can clone the `togethercomputer/RedPajama-Data-1T-Sample` repo and load the dataset with `load_dataset("path/to/cloned_repo")` to use it offline.
FileNotFoundError when trying to load the downloaded dataset with `load_dataset(..., streaming=True)`
### Describe the bug Hi, I'm trying to load the dataset `togethercomputer/RedPajama-Data-1T-Sample` with `load_dataset` in streaming mode, i.e., `streaming=True`, but `FileNotFoundError` occurs. ### Steps to reproduce the bug I've downloaded the dataset and save it to the cache dir in advance. My hope is loadi...
16
FileNotFoundError when trying to load the downloaded dataset with `load_dataset(..., streaming=True)` ### Describe the bug Hi, I'm trying to load the dataset `togethercomputer/RedPajama-Data-1T-Sample` with `load_dataset` in streaming mode, i.e., `streaming=True`, but `FileNotFoundError` occurs. ### Steps to re...
[ -0.33369070291519165, -0.10643325001001358, 0.10457661002874374, 0.33565500378608704, 0.1996665745973587, 0.007022194564342499, 0.20944146811962128, 0.36828163266181946, 0.0730251893401146, 0.0012076348066329956, 0.15766292810440063, 0.28006261587142944, -0.20167627930641174, -0.1230023205...
https://github.com/huggingface/datasets/issues/6327
@mariosasko Thank you for your kind reply! I'll try it as a workaround. Does that mean that currently it's not supported to simply load with a short name?
FileNotFoundError when trying to load the downloaded dataset with `load_dataset(..., streaming=True)`
### Describe the bug Hi, I'm trying to load the dataset `togethercomputer/RedPajama-Data-1T-Sample` with `load_dataset` in streaming mode, i.e., `streaming=True`, but `FileNotFoundError` occurs. ### Steps to reproduce the bug I've downloaded the dataset and save it to the cache dir in advance. My hope is loadi...
28
FileNotFoundError when trying to load the downloaded dataset with `load_dataset(..., streaming=True)` ### Describe the bug Hi, I'm trying to load the dataset `togethercomputer/RedPajama-Data-1T-Sample` with `load_dataset` in streaming mode, i.e., `streaming=True`, but `FileNotFoundError` occurs. ### Steps to re...
[ -0.33369070291519165, -0.10643325001001358, 0.10457661002874374, 0.33565500378608704, 0.1996665745973587, 0.007022194564342499, 0.20944146811962128, 0.36828163266181946, 0.0730251893401146, 0.0012076348066329956, 0.15766292810440063, 0.28006261587142944, -0.20167627930641174, -0.1230023205...
https://github.com/huggingface/datasets/issues/6327
It is, but manually downloading repo files to the cache can easily lead to failure (the HF cache is not meant to be modified by a user besides deleting the files 🙂), as in your case. Hence, the clone + `load_dataset("path/to/cloned_repo")` workflow should be used instead.
FileNotFoundError when trying to load the downloaded dataset with `load_dataset(..., streaming=True)`
### Describe the bug Hi, I'm trying to load the dataset `togethercomputer/RedPajama-Data-1T-Sample` with `load_dataset` in streaming mode, i.e., `streaming=True`, but `FileNotFoundError` occurs. ### Steps to reproduce the bug I've downloaded the dataset and save it to the cache dir in advance. My hope is loadi...
46
FileNotFoundError when trying to load the downloaded dataset with `load_dataset(..., streaming=True)` ### Describe the bug Hi, I'm trying to load the dataset `togethercomputer/RedPajama-Data-1T-Sample` with `load_dataset` in streaming mode, i.e., `streaming=True`, but `FileNotFoundError` occurs. ### Steps to re...
[ -0.33369070291519165, -0.10643325001001358, 0.10457661002874374, 0.33565500378608704, 0.1996665745973587, 0.007022194564342499, 0.20944146811962128, 0.36828163266181946, 0.0730251893401146, 0.0012076348066329956, 0.15766292810440063, 0.28006261587142944, -0.20167627930641174, -0.1230023205...
https://github.com/huggingface/datasets/issues/6324
Unlike Pandas, Arrow is strict with types, so converting the problematic strings to ints (or ints to strings) to ensure all the values have the same type is the only fix. JSON support has been requested in Arrow [here](https://github.com/apache/arrow/issues/32538), but I don't expect this to be implemented soon. ...
Conversion to Arrow fails due to wrong type heuristic
### Describe the bug I have a list of dictionaries with valid/JSON-serializable values. One key is the denominator for a paragraph. In 99.9% of cases its a number, but there are some occurences of '1a', '2b' and so on. If trying to convert this list to a dataset with `Dataset.from_list()`, I always get `ArrowI...
97
Conversion to Arrow fails due to wrong type heuristic ### Describe the bug I have a list of dictionaries with valid/JSON-serializable values. One key is the denominator for a paragraph. In 99.9% of cases its a number, but there are some occurences of '1a', '2b' and so on. If trying to convert this list to a d...
[ -0.13682855665683746, -0.08490268886089325, 0.047282736748456955, 0.18112041056156158, 0.4715591371059418, -0.02025952935218811, 0.2181384414434433, 0.3003539741039276, -0.2322429120540619, -0.14563122391700745, 0.018224947154521942, 0.5438294410705566, -0.030470455065369606, -0.0194185748...
https://github.com/huggingface/datasets/issues/6324
> Unlike Pandas, Arrow is strict with types, so converting the problematic strings to ints (or ints to strings) to ensure all the values have the same type is the only fix. > > JSON support has been requested in Arrow [here](https://github.com/apache/arrow/issues/32538), but I don't expect this to be implemented soo...
Conversion to Arrow fails due to wrong type heuristic
### Describe the bug I have a list of dictionaries with valid/JSON-serializable values. One key is the denominator for a paragraph. In 99.9% of cases its a number, but there are some occurences of '1a', '2b' and so on. If trying to convert this list to a dataset with `Dataset.from_list()`, I always get `ArrowI...
121
Conversion to Arrow fails due to wrong type heuristic ### Describe the bug I have a list of dictionaries with valid/JSON-serializable values. One key is the denominator for a paragraph. In 99.9% of cases its a number, but there are some occurences of '1a', '2b' and so on. If trying to convert this list to a d...
[ -0.1248508095741272, -0.08872330188751221, 0.048213206231594086, 0.19750455021858215, 0.49781298637390137, -0.03315145522356033, 0.23462632298469543, 0.31393522024154663, -0.23804914951324463, -0.1408153772354126, 0.0366579070687294, 0.5494917035102844, -0.01877778209745884, -0.01878361031...
https://github.com/huggingface/datasets/issues/6320
The expression "train+test" concatenates the splits. The individual splits as separate datasets can be obtained as follows: ```python train_ds, test_ds = load_dataset("<dataset_name>", split=["train", "test"]) train_10pct_ds, test_10pct_ds = load_dataset("<dataset_name>", split=["train[:10%]", "test[:%10]"]) ```
Dataset slice splits can't load training and validation at the same time
### Describe the bug According to the [documentation](https://huggingface.co/docs/datasets/v2.14.5/loading#slice-splits) is should be possible to run the following command: `train_test_ds = datasets.load_dataset("bookcorpus", split="train+test")` to load the train and test sets from the dataset. However ex...
31
Dataset slice splits can't load training and validation at the same time ### Describe the bug According to the [documentation](https://huggingface.co/docs/datasets/v2.14.5/loading#slice-splits) is should be possible to run the following command: `train_test_ds = datasets.load_dataset("bookcorpus", split="train+...
[ -0.12745431065559387, -0.10093505680561066, -0.006819732487201691, 0.26014244556427, 0.14704033732414246, 0.07430852949619293, 0.36141237616539, 0.24388569593429565, -0.18307682871818542, -0.059660300612449646, -0.3403538465499878, 0.1565178781747818, -0.009768269956111908, 0.5907857418060...
https://github.com/huggingface/datasets/issues/6319
Hi! Instead of processing a single example at a time, you should use the batched `map` for the best performance (with `num_proc=1`) - the fast tokenizers can process a batch's samples in parallel in that scenario. E.g., the following code in Colab takes an hour to complete: ```python # !pip install datasets transf...
Datasets.map is severely broken
### Describe the bug Regardless of how many cores I used, I have 16 or 32 threads, map slows down to a crawl at around 80% done, lingers maybe until 97% extremely slowly and NEVER finishes the job. It just hangs. After watching this for 27 hours I control-C out of it. Until the end one process appears to be doing s...
73
Datasets.map is severely broken ### Describe the bug Regardless of how many cores I used, I have 16 or 32 threads, map slows down to a crawl at around 80% done, lingers maybe until 97% extremely slowly and NEVER finishes the job. It just hangs. After watching this for 27 hours I control-C out of it. Until the end...
[ -0.6238995790481567, 0.08061352372169495, 0.04081980884075165, 0.1330558806657791, 0.1382634937763214, 0.09867337346076965, 0.2976037859916687, 0.24050694704055786, -0.1358392834663391, -0.11661169677972794, -0.06945439428091049, 0.7371479272842407, -0.08465425670146942, -0.326397061347961...
https://github.com/huggingface/datasets/issues/6319
Batched is far worse. A single batch of 1000 took hours and that was only 1% On Thu, Oct 19, 2023, 2:26 PM Mario Šaško ***@***.***> wrote: > Hi! You should use the batched map for the best performance (with > num_proc=1) - the fast tokenizers can process a batch's samples in > parallel. > > E.g., the follow...
Datasets.map is severely broken
### Describe the bug Regardless of how many cores I used, I have 16 or 32 threads, map slows down to a crawl at around 80% done, lingers maybe until 97% extremely slowly and NEVER finishes the job. It just hangs. After watching this for 27 hours I control-C out of it. Until the end one process appears to be doing s...
125
Datasets.map is severely broken ### Describe the bug Regardless of how many cores I used, I have 16 or 32 threads, map slows down to a crawl at around 80% done, lingers maybe until 97% extremely slowly and NEVER finishes the job. It just hangs. After watching this for 27 hours I control-C out of it. Until the end...
[ -0.6426334381103516, 0.04849633574485779, 0.041576087474823, 0.15280376374721527, 0.12931455671787262, 0.12522882223129272, 0.26164349913597107, 0.26831841468811035, -0.09941567480564117, -0.1265753209590912, -0.08454795181751251, 0.7016130089759827, -0.09337520599365234, -0.37069922685623...
https://github.com/huggingface/datasets/issues/6319
Which specific version of datasets are you using? What is the architecture of your colab setup? Ram? Cores? OS? On Thu, Oct 19, 2023, 2:27 PM pensive introvert ***@***.***> wrote: > Batched is far worse. A single batch of 1000 took hours and that was only > 1% > > > On Thu, Oct 19, 2023, 2:26 PM Mario Š...
Datasets.map is severely broken
### Describe the bug Regardless of how many cores I used, I have 16 or 32 threads, map slows down to a crawl at around 80% done, lingers maybe until 97% extremely slowly and NEVER finishes the job. It just hangs. After watching this for 27 hours I control-C out of it. Until the end one process appears to be doing s...
163
Datasets.map is severely broken ### Describe the bug Regardless of how many cores I used, I have 16 or 32 threads, map slows down to a crawl at around 80% done, lingers maybe until 97% extremely slowly and NEVER finishes the job. It just hangs. After watching this for 27 hours I control-C out of it. Until the end...
[ -0.6481049060821533, 0.10938793420791626, 0.024797581136226654, 0.10106520354747772, 0.0879831314086914, 0.12319760024547577, 0.27200183272361755, 0.2651056945323944, -0.1283535659313202, -0.057041190564632416, -0.055256787687540054, 0.6183364987373352, -0.12706303596496582, -0.23674380779...
https://github.com/huggingface/datasets/issues/6319
from functools import partial import transformers from datasets import load_dataset, concatenate_datasets, load_from_disk model_name_or_path="/opt/data/data/daryl149/llama-2-7b-chat-hf" output_dir="/opt/data/data/LongLoRA/checkpoints" cache_dir="/opt/data/data/LongLoRA/cache" model_max_length=16384 IGNORE_IN...
Datasets.map is severely broken
### Describe the bug Regardless of how many cores I used, I have 16 or 32 threads, map slows down to a crawl at around 80% done, lingers maybe until 97% extremely slowly and NEVER finishes the job. It just hangs. After watching this for 27 hours I control-C out of it. Until the end one process appears to be doing s...
170
Datasets.map is severely broken ### Describe the bug Regardless of how many cores I used, I have 16 or 32 threads, map slows down to a crawl at around 80% done, lingers maybe until 97% extremely slowly and NEVER finishes the job. It just hangs. After watching this for 27 hours I control-C out of it. Until the end...
[ -0.6168144345283508, 0.07999947667121887, 0.05563335493206978, 0.20444941520690918, 0.2958584725856781, 0.12437372654676437, 0.2487582415342331, 0.3236621022224426, -0.10938513278961182, -0.13184724748134613, -0.19389095902442932, 0.6695704460144043, -0.06927616894245148, -0.41678601503372...
https://github.com/huggingface/datasets/issues/6319
I changed the tokenizer to one without "Fast suffix, and something changed. The fraction, although still slowed a lot at 80% was able to get over the finish line of 100% I have to do more testng, see if the whole set can be processed On Thu, Oct 19, 2023 at 3:03 PM pensive introvert < ***@***.***> wrote: ...
Datasets.map is severely broken
### Describe the bug Regardless of how many cores I used, I have 16 or 32 threads, map slows down to a crawl at around 80% done, lingers maybe until 97% extremely slowly and NEVER finishes the job. It just hangs. After watching this for 27 hours I control-C out of it. Until the end one process appears to be doing s...
290
Datasets.map is severely broken ### Describe the bug Regardless of how many cores I used, I have 16 or 32 threads, map slows down to a crawl at around 80% done, lingers maybe until 97% extremely slowly and NEVER finishes the job. It just hangs. After watching this for 27 hours I control-C out of it. Until the end...
[ -0.5819017887115479, 0.12894117832183838, 0.033899616450071335, 0.16532237827777863, 0.20738404989242554, 0.16037096083164215, 0.23931464552879333, 0.29810354113578796, -0.06317737698554993, -0.18398520350456238, -0.10214298963546753, 0.6148468255996704, -0.09060730785131454, -0.3609420061...
https://github.com/huggingface/datasets/issues/6319
So, using LlamaTokenizerFast was the problem. Changing it to LlamaTokenizer fixed things, On Thu, Oct 19, 2023 at 4:04 PM pensive introvert < ***@***.***> wrote: > I changed the tokenizer to one without "Fast suffix, and something > changed. The fraction, although still slowed a lot at 80% was able to get > o...
Datasets.map is severely broken
### Describe the bug Regardless of how many cores I used, I have 16 or 32 threads, map slows down to a crawl at around 80% done, lingers maybe until 97% extremely slowly and NEVER finishes the job. It just hangs. After watching this for 27 hours I control-C out of it. Until the end one process appears to be doing s...
327
Datasets.map is severely broken ### Describe the bug Regardless of how many cores I used, I have 16 or 32 threads, map slows down to a crawl at around 80% done, lingers maybe until 97% extremely slowly and NEVER finishes the job. It just hangs. After watching this for 27 hours I control-C out of it. Until the end...
[ -0.5650825500488281, 0.09585857391357422, 0.027538053691387177, 0.1539461612701416, 0.21590092778205872, 0.12963418662548065, 0.16759265959262848, 0.3100394010543823, -0.05893123894929886, -0.16628892719745636, -0.09866967797279358, 0.6297377347946167, -0.07801172137260437, -0.366154909133...
https://github.com/huggingface/datasets/issues/6319
Indeed, the tokenizer is super slow. Perhaps @ArthurZucker knows the reason why. ([This](https://colab.research.google.com/drive/1VgeurX-4Fl2X6aBQTwh_X4kuQKZ6K9L1?usp=sharing) simplified Colab can be used to reproduce the behavior)
Datasets.map is severely broken
### Describe the bug Regardless of how many cores I used, I have 16 or 32 threads, map slows down to a crawl at around 80% done, lingers maybe until 97% extremely slowly and NEVER finishes the job. It just hangs. After watching this for 27 hours I control-C out of it. Until the end one process appears to be doing s...
22
Datasets.map is severely broken ### Describe the bug Regardless of how many cores I used, I have 16 or 32 threads, map slows down to a crawl at around 80% done, lingers maybe until 97% extremely slowly and NEVER finishes the job. It just hangs. After watching this for 27 hours I control-C out of it. Until the end...
[ -0.5410916805267334, 0.13084936141967773, 0.02908453717827797, 0.1009187251329422, 0.17967413365840912, 0.05452781915664673, 0.2067749798297882, 0.2939973771572113, -0.11034294962882996, -0.12849867343902588, -0.00783967599272728, 0.5993674397468567, -0.07884661853313446, -0.32366651296615...
https://github.com/huggingface/datasets/issues/6319
same issue here sample to reproduce: https://github.com/philschmid/document-ai-transformers/blob/main/training/donut_sroie.ipynb with following map line https://github.com/philschmid/document-ai-transformers/blob/main/training/donut_sroie.ipynb If I directly iterate over the dataset and call the mapping method, i...
Datasets.map is severely broken
### Describe the bug Regardless of how many cores I used, I have 16 or 32 threads, map slows down to a crawl at around 80% done, lingers maybe until 97% extremely slowly and NEVER finishes the job. It just hangs. After watching this for 27 hours I control-C out of it. Until the end one process appears to be doing s...
65
Datasets.map is severely broken ### Describe the bug Regardless of how many cores I used, I have 16 or 32 threads, map slows down to a crawl at around 80% done, lingers maybe until 97% extremely slowly and NEVER finishes the job. It just hangs. After watching this for 27 hours I control-C out of it. Until the end...
[ -0.5472961068153381, 0.10284584760665894, 0.04873661696910858, 0.10942556709051132, 0.15640074014663696, 0.11435630172491074, 0.3405649960041046, 0.33139339089393616, -0.005230031907558441, -0.20240147411823273, -0.09187188744544983, 0.5996783971786499, 0.02987278252840042, -0.368177473545...
https://github.com/huggingface/datasets/issues/6319
@ewfian > If I directly iterate over the dataset and call the mapping method, it is very fast `Dataset.map` must also convert the images into bytes to write them to an Arrow file (the write itself takes some time, too). You can make the `map` faster by manually converting the images into an "arrow-compatible...
Datasets.map is severely broken
### Describe the bug Regardless of how many cores I used, I have 16 or 32 threads, map slows down to a crawl at around 80% done, lingers maybe until 97% extremely slowly and NEVER finishes the job. It just hangs. After watching this for 27 hours I control-C out of it. Until the end one process appears to be doing s...
111
Datasets.map is severely broken ### Describe the bug Regardless of how many cores I used, I have 16 or 32 threads, map slows down to a crawl at around 80% done, lingers maybe until 97% extremely slowly and NEVER finishes the job. It just hangs. After watching this for 27 hours I control-C out of it. Until the end...
[ -0.5075655579566956, 0.08682578802108765, 0.024711772799491882, 0.14701423048973083, 0.13376814126968384, 0.17092423141002655, 0.23593086004257202, 0.4097411036491394, -0.0756518766283989, -0.1975446343421936, -0.001470722258090973, 0.6998304724693298, -0.0015343837440013885, -0.3838915228...
https://github.com/huggingface/datasets/issues/6319
The problem I had was to do with map using fork and copying locks from the parent process in acquired state. I ended up changing the context to use forkserver instead. On Wed, Nov 29, 2023, 10:04 PM Mario Šaško ***@***.***> wrote: > @ewfian <https://github.com/ewfian> > > If I directly iterate over the data...
Datasets.map is severely broken
### Describe the bug Regardless of how many cores I used, I have 16 or 32 threads, map slows down to a crawl at around 80% done, lingers maybe until 97% extremely slowly and NEVER finishes the job. It just hangs. After watching this for 27 hours I control-C out of it. Until the end one process appears to be doing s...
208
Datasets.map is severely broken ### Describe the bug Regardless of how many cores I used, I have 16 or 32 threads, map slows down to a crawl at around 80% done, lingers maybe until 97% extremely slowly and NEVER finishes the job. It just hangs. After watching this for 27 hours I control-C out of it. Until the end...
[ -0.5738916993141174, 0.08227047324180603, 0.01397879421710968, 0.18170377612113953, 0.16029129922389984, 0.11835987865924835, 0.19462914764881134, 0.38573387265205383, -0.01956336945295334, -0.21568940579891205, -0.02014840953052044, 0.6655286550521851, -0.05103820562362671, -0.41035270690...
https://github.com/huggingface/datasets/issues/6317
We have opened an issue in the corresponding Hub dataset: https://huggingface.co/datasets/sentiment140/discussions/3 Let's continue the discussion there.
sentiment140 dataset unavailable
### Describe the bug loading the dataset using load_dataset("sentiment140") returns the following error ConnectionError: Couldn't reach http://cs.stanford.edu/people/alecmgo/trainingandtestdata.zip (error 403) ### Steps to reproduce the bug Run the following code (version should not matter). ``` from data...
16
sentiment140 dataset unavailable ### Describe the bug loading the dataset using load_dataset("sentiment140") returns the following error ConnectionError: Couldn't reach http://cs.stanford.edu/people/alecmgo/trainingandtestdata.zip (error 403) ### Steps to reproduce the bug Run the following code (version sh...
[ -0.36286142468452454, 0.09124556183815002, -0.0619184784591198, 0.3383810520172119, 0.035611383616924286, 0.05036894977092743, 0.03202589601278305, 0.31903377175331116, 0.07944399118423462, 0.15020966529846191, -0.2229425013065338, -0.07228529453277588, 0.09144529700279236, 0.2631198763847...