html_url stringlengths 51 51 | comments stringlengths 67 24.7k | title stringlengths 6 280 | body stringlengths 51 36.2k | comment_length int64 16 1.45k | text stringlengths 190 38.3k | embeddings list |
|---|---|---|---|---|---|---|
https://github.com/huggingface/datasets/issues/6442 | I reproduced too:
- root: metadata file is ignored (https://huggingface.co/datasets/severo/doc-image-3)
- data/ dir: metadata file is ignored (https://huggingface.co/datasets/severo/doc-image-4)
- train/ dir: works (https://huggingface.co/datasets/severo/doc-image-5) | Trouble loading image folder with additional features - metadata file ignored | ### Describe the bug
Loading image folder with a caption column using `load_dataset(<image_folder_path>)` doesn't load the captions.
When loading a local image folder with captions using `datasets==2.13.0`
```
from datasets import load_dataset
data = load_dataset(<image_folder_path>)
data.column_names
```
... | 23 | Trouble loading image folder with additional features - metadata file ignored
### Describe the bug
Loading image folder with a caption column using `load_dataset(<image_folder_path>)` doesn't load the captions.
When loading a local image folder with captions using `datasets==2.13.0`
```
from datasets import... | [
-0.005256466567516327,
0.07954807579517365,
0.025517955422401428,
0.6268149018287659,
0.4523826241493225,
0.11177997291088104,
0.42082440853118896,
0.046340543776750565,
0.2590237557888031,
0.058994755148887634,
-0.07752669602632523,
0.4900439381599426,
-0.20371121168136597,
0.149850949645... |
https://github.com/huggingface/datasets/issues/6441 | > Also when they try to click the url link for the dataset they get a 404 error.
This seems to be a Hub error then (cc @SBrandeis) | Trouble Loading a Gated Dataset For User with Granted Permission | ### Describe the bug
I have granted permissions to several users to access a gated huggingface dataset. The users accepted the invite and when trying to load the dataset using their access token they get
`FileNotFoundError: Couldn't find a dataset script at .....` . Also when they try to click the url link for the d... | 28 | Trouble Loading a Gated Dataset For User with Granted Permission
### Describe the bug
I have granted permissions to several users to access a gated huggingface dataset. The users accepted the invite and when trying to load the dataset using their access token they get
`FileNotFoundError: Couldn't find a dataset sc... | [
-0.29226887226104736,
-0.23708491027355194,
0.059749752283096313,
0.5143930912017822,
-0.0005606310442090034,
0.010804161429405212,
0.4375983774662018,
0.01528296247124672,
0.49582868814468384,
0.22912351787090302,
-0.24911551177501678,
-0.1277315467596054,
-0.11217568814754486,
0.11388504... |
https://github.com/huggingface/datasets/issues/6441 | Could you report this to https://discuss.huggingface.co/c/hub/23, providing the URL of the dataset, or at least if the dataset is public or private? | Trouble Loading a Gated Dataset For User with Granted Permission | ### Describe the bug
I have granted permissions to several users to access a gated huggingface dataset. The users accepted the invite and when trying to load the dataset using their access token they get
`FileNotFoundError: Couldn't find a dataset script at .....` . Also when they try to click the url link for the d... | 22 | Trouble Loading a Gated Dataset For User with Granted Permission
### Describe the bug
I have granted permissions to several users to access a gated huggingface dataset. The users accepted the invite and when trying to load the dataset using their access token they get
`FileNotFoundError: Couldn't find a dataset sc... | [
-0.1947937309741974,
-0.27182096242904663,
0.054994285106658936,
0.5652323961257935,
-0.0011484064161777496,
0.013401776552200317,
0.4236222505569458,
-0.05076004937291145,
0.40111419558525085,
0.14345966279506683,
-0.2048477679491043,
-0.15554077923297882,
-0.1591770350933075,
0.200249552... |
https://github.com/huggingface/datasets/issues/6440 | Tried to upgrade Python to 3.11 - still get this message. A partial solution is to NOT use `num_proc` at all. It will be considerably longer to finish the job. | `.map` not hashing under python 3.9 | ### Describe the bug
The `.map` function cannot hash under python 3.9. Tried to use [the solution here](https://github.com/huggingface/datasets/issues/4521#issuecomment-1205166653), but still get the same message:
`Parameter 'function'=<function map_to_pred at 0x7fa0b49ead30> of the transform datasets.arrow_data... | 30 | `.map` not hashing under python 3.9
### Describe the bug
The `.map` function cannot hash under python 3.9. Tried to use [the solution here](https://github.com/huggingface/datasets/issues/4521#issuecomment-1205166653), but still get the same message:
`Parameter 'function'=<function map_to_pred at 0x7fa0b49ead30... | [
-0.23779386281967163,
-0.5131184458732605,
0.039723578840494156,
-0.02358236536383629,
0.3392046391963959,
-0.013328582048416138,
0.28753432631492615,
0.25054922699928284,
0.08202031254768372,
0.08945101499557495,
0.005709212273359299,
0.4939461648464203,
-0.38324224948883057,
0.0680404901... |
https://github.com/huggingface/datasets/issues/6440 | Hi! The `model = torch.compile(model)` line is problematic for our hashing logic. We would have to merge https://github.com/huggingface/datasets/pull/5867 to support hashing `torch.compile`-ed models/functions.
I've started refactoring the hashing logic and plan to incorporate a fix for `torch.compile` as part of i... | `.map` not hashing under python 3.9 | ### Describe the bug
The `.map` function cannot hash under python 3.9. Tried to use [the solution here](https://github.com/huggingface/datasets/issues/4521#issuecomment-1205166653), but still get the same message:
`Parameter 'function'=<function map_to_pred at 0x7fa0b49ead30> of the transform datasets.arrow_data... | 52 | `.map` not hashing under python 3.9
### Describe the bug
The `.map` function cannot hash under python 3.9. Tried to use [the solution here](https://github.com/huggingface/datasets/issues/4521#issuecomment-1205166653), but still get the same message:
`Parameter 'function'=<function map_to_pred at 0x7fa0b49ead30... | [
-0.23779386281967163,
-0.5131184458732605,
0.039723578840494156,
-0.02358236536383629,
0.3392046391963959,
-0.013328582048416138,
0.28753432631492615,
0.25054922699928284,
0.08202031254768372,
0.08945101499557495,
0.005709212273359299,
0.4939461648464203,
-0.38324224948883057,
0.0680404901... |
https://github.com/huggingface/datasets/issues/6438 | Thank you, @severo ! I would be more than happy to help in any way I can. I am not familiar with this repo's codebase, but I would be eager to contribute. :)
For the preview in Datasets Hub, I think it makes sense to just display the geospatial column as text. If there were a dataset loader, though, I think it shoul... | Support GeoParquet | ### Feature request
Support the GeoParquet format
### Motivation
GeoParquet (https://geoparquet.org/) is a common format for sharing vectorial geospatial data on the cloud, along with "traditional" data columns.
It would be nice to be able to load this format with datasets, and more generally, in the Datasets Hub... | 168 | Support GeoParquet
### Feature request
Support the GeoParquet format
### Motivation
GeoParquet (https://geoparquet.org/) is a common format for sharing vectorial geospatial data on the cloud, along with "traditional" data columns.
It would be nice to be able to load this format with datasets, and more generally... | [
-0.18842795491218567,
-0.007291898131370544,
-0.004543256014585495,
0.08613798022270203,
0.20921705663204193,
-0.2451484054327011,
0.10604145377874374,
0.4123835861682892,
0.17052997648715973,
0.05875986069440842,
-0.04821088910102844,
0.4327125549316406,
-0.19047911465168,
0.6723259687423... |
https://github.com/huggingface/datasets/issues/6438 | Just hitting into this same issue too showing GeoParquet files in Datasets Viewer. I tried to implement a custom reader for GeoParquet in https://huggingface.co/datasets/weiji14/clay_vector_embeddings/discussions/1, but it seems like HuggingFace has disabled datasets with custom loading scripts from using the dataset v... | Support GeoParquet | ### Feature request
Support the GeoParquet format
### Motivation
GeoParquet (https://geoparquet.org/) is a common format for sharing vectorial geospatial data on the cloud, along with "traditional" data columns.
It would be nice to be able to load this format with datasets, and more generally, in the Datasets Hub... | 132 | Support GeoParquet
### Feature request
Support the GeoParquet format
### Motivation
GeoParquet (https://geoparquet.org/) is a common format for sharing vectorial geospatial data on the cloud, along with "traditional" data columns.
It would be nice to be able to load this format with datasets, and more generally... | [
-0.131269171833992,
-0.04446607828140259,
0.0009181350469589233,
0.1977737843990326,
0.22396178543567657,
-0.0637836903333664,
0.05105629935860634,
0.25741830468177795,
0.1787053346633911,
0.09568889439105988,
-0.19080108404159546,
0.42543378472328186,
-0.35302814841270447,
0.6135354042053... |
https://github.com/huggingface/datasets/issues/6438 | Update: It looks like renaming the GeoParquet file to have a file extension of `*.parquet` works (see https://huggingface.co/datasets/weiji14/clay_vector_embeddings). HuggingFace's default parquet reader is able to read the GeoParquet file, though the geometry column is of an unknown type:
 is a common format for sharing vectorial geospatial data on the cloud, along with "traditional" data columns.
It would be nice to be able to load this format with datasets, and more generally, in the Datasets Hub... | 68 | Support GeoParquet
### Feature request
Support the GeoParquet format
### Motivation
GeoParquet (https://geoparquet.org/) is a common format for sharing vectorial geospatial data on the cloud, along with "traditional" data columns.
It would be nice to be able to load this format with datasets, and more generally... | [
-0.1853790283203125,
-0.15920162200927734,
-0.017098691314458847,
0.17653638124465942,
0.25480347871780396,
-0.14688898622989655,
-0.037101730704307556,
0.3081713020801544,
0.15877442061901093,
0.08432714641094208,
-0.02328629419207573,
0.36783090233802795,
-0.24566690623760223,
0.64769935... |
https://github.com/huggingface/datasets/issues/6438 | @joshuasundance-swca, @weiji14, If I'm understanding this correctly, the code below wouldn't be recommended to due to dependency headaches? If that's the case, what solution would there be to see the geometry features for .gpq files in huggingfaceHub?
code for dataset_loader.py
```
import geopandas as gpd
# ... ... | Support GeoParquet | ### Feature request
Support the GeoParquet format
### Motivation
GeoParquet (https://geoparquet.org/) is a common format for sharing vectorial geospatial data on the cloud, along with "traditional" data columns.
It would be nice to be able to load this format with datasets, and more generally, in the Datasets Hub... | 157 | Support GeoParquet
### Feature request
Support the GeoParquet format
### Motivation
GeoParquet (https://geoparquet.org/) is a common format for sharing vectorial geospatial data on the cloud, along with "traditional" data columns.
It would be nice to be able to load this format with datasets, and more generally... | [
-0.2090664505958557,
-0.26573559641838074,
0.04979145526885986,
0.31051427125930786,
0.17626544833183289,
-0.16471150517463684,
0.10059749335050583,
0.3131271004676819,
0.41674306988716125,
0.03970740735530853,
-0.1878364086151123,
0.3662434220314026,
-0.1826562136411667,
0.592523694038391... |
https://github.com/huggingface/datasets/issues/6437 | `split_dataset_by_node` doesn't give the exact same number of examples to each node in the case of iterable datasets, though it tries to be as equal as possible. In particular if your dataset is sharded and you have a number of shards that is a factor of the number of workers, then the shards will be evenly distributed... | Problem in training iterable dataset | ### Describe the bug
I am using PyTorch DDP (Distributed Data Parallel) to train my model. Since the data is too large to load into memory at once, I am using load_dataset to read the data as an iterable dataset. I have used datasets.distributed.split_dataset_by_node to distribute the dataset. However, I have notice... | 129 | Problem in training iterable dataset
### Describe the bug
I am using PyTorch DDP (Distributed Data Parallel) to train my model. Since the data is too large to load into memory at once, I am using load_dataset to read the data as an iterable dataset. I have used datasets.distributed.split_dataset_by_node to distrib... | [
-0.3585677742958069,
-0.43623197078704834,
-0.06505583226680756,
0.244727224111557,
0.15072189271450043,
0.0003087371587753296,
0.22049580514431,
0.10808394849300385,
-0.03405902534723282,
0.03095584362745285,
0.07104901969432831,
0.4075590670108795,
-0.1007930338382721,
-0.148665875196456... |
https://github.com/huggingface/datasets/issues/6432 | You should use `datasets.load_dataset` instead of `nlp.load_dataset`, as the `nlp` package is outdated.
If switching to `datasets.load_dataset` doesn't fix the issue, sharing the JSON file (feel free to replace the data with dummy data) would be nice so that we can reproduce it ourselves. | load_dataset does not load all of the data in my input file | ### Describe the bug
I have 127 elements in my input dataset. When I do a len on the dataset after loaded, it is only 124 elements.
### Steps to reproduce the bug
train_dataset = nlp.load_dataset(data_args.dataset_path, name=data_args.qg_format, split=nlp.Split.TRAIN)
valid_dataset = nlp.load_dataset(data_... | 44 | load_dataset does not load all of the data in my input file
### Describe the bug
I have 127 elements in my input dataset. When I do a len on the dataset after loaded, it is only 124 elements.
### Steps to reproduce the bug
train_dataset = nlp.load_dataset(data_args.dataset_path, name=data_args.qg_format, spli... | [
-0.22444385290145874,
-0.252759724855423,
0.0006521642208099365,
0.5328270196914673,
0.18958520889282227,
-0.0704435408115387,
0.3324061632156372,
0.4615911543369293,
-0.04259912669658661,
-0.11367630958557129,
0.13742810487747192,
0.3046746850013733,
0.016711439937353134,
0.24982538819313... |
https://github.com/huggingface/datasets/issues/6422 | We have a config variable that controls the batch size in `save_to_disk`:
```python
import datasets
datasets.config.DEFAULT_MAX_BATCH_SIZE = <smaller_batch_size>
...
ds.save_to_disk(...)
``` | Allow to choose the `writer_batch_size` when using `save_to_disk` | ### Feature request
Add an argument in `save_to_disk` regarding batch size, which would be passed to `shard` and other methods.
### Motivation
The `Dataset.save_to_disk` method currently calls `shard` without passing a `writer_batch_size` argument, thus implicitly using the default value (1000). This can result in R... | 21 | Allow to choose the `writer_batch_size` when using `save_to_disk`
### Feature request
Add an argument in `save_to_disk` regarding batch size, which would be passed to `shard` and other methods.
### Motivation
The `Dataset.save_to_disk` method currently calls `shard` without passing a `writer_batch_size` argument, ... | [
-0.4314618706703186,
-0.1805671751499176,
-0.051898352801799774,
-0.018587402999401093,
0.021444931626319885,
-0.026462405920028687,
0.3796961307525635,
0.23060111701488495,
-0.11458665132522583,
0.40544524788856506,
0.2738927900791168,
-0.00015321001410484314,
-0.3835754096508026,
0.35399... |
https://github.com/huggingface/datasets/issues/6422 | Thank you for your answer!
From what I am reading in `https://github.com/huggingface/datasets/blob/2.14.5/src/datasets/arrow_dataset.py`, every function involved (`select`, `shard`, ...) has a default hardcoded batch size of 1000, as such:
```python
def select(
self,
indices: Iterable,
k... | Allow to choose the `writer_batch_size` when using `save_to_disk` | ### Feature request
Add an argument in `save_to_disk` regarding batch size, which would be passed to `shard` and other methods.
### Motivation
The `Dataset.save_to_disk` method currently calls `shard` without passing a `writer_batch_size` argument, thus implicitly using the default value (1000). This can result in R... | 134 | Allow to choose the `writer_batch_size` when using `save_to_disk`
### Feature request
Add an argument in `save_to_disk` regarding batch size, which would be passed to `shard` and other methods.
### Motivation
The `Dataset.save_to_disk` method currently calls `shard` without passing a `writer_batch_size` argument, ... | [
-0.4530668258666992,
-0.10187280178070068,
0.004923049360513687,
0.05907120928168297,
0.0542483814060688,
-0.092894047498703,
0.3411109447479248,
0.13401763141155243,
-0.17516355216503143,
0.2381095290184021,
0.20967498421669006,
0.07316283881664276,
-0.35989850759506226,
0.142907813191413... |
https://github.com/huggingface/datasets/issues/6417 | Very strange: `datasets-cli env`
>
> Copy-and-paste the text below in your GitHub issue.
>
> - `datasets` version: 2.9.0
> - Platform: macOS-14.0-arm64-arm-64bit
> - Python version: 3.9.13
> - PyArrow version: 8.0.0
> - Pandas version: 1.3.5
After updating datasets and pyarrow on base environment, although... | Bug: LayoutLMv3 finetuning on FUNSD Notebook; Arrow Error | ### Describe the bug
Arrow issues when running the example Notebook laptop locally on Mac with M1. Works on Google Collab.
**Notebook**: https://github.com/NielsRogge/Transformers-Tutorials/blob/master/LayoutLMv3/Fine_tune_LayoutLMv3_on_FUNSD_(HuggingFace_Trainer).ipynb
**Error**: `ValueError: Arrow type extensi... | 95 | Bug: LayoutLMv3 finetuning on FUNSD Notebook; Arrow Error
### Describe the bug
Arrow issues when running the example Notebook laptop locally on Mac with M1. Works on Google Collab.
**Notebook**: https://github.com/NielsRogge/Transformers-Tutorials/blob/master/LayoutLMv3/Fine_tune_LayoutLMv3_on_FUNSD_(HuggingFace... | [
-0.12534523010253906,
-0.696060299873352,
0.06675893068313599,
0.3161044716835022,
0.36773860454559326,
-0.21727141737937927,
0.570938229560852,
0.08089077472686768,
-0.18083691596984863,
-0.04032056778669357,
-0.07140608876943588,
0.6541163921356201,
-0.07819987833499908,
0.20544581115245... |
https://github.com/huggingface/datasets/issues/6417 | Hi! The latest (patch) release (published a few hours ago) includes a fix for this [PyArrow security issue](https://github.com/advisories/GHSA-5wvp-7f3h-6wmm). To install it, run `pip install -U datasets`. | Bug: LayoutLMv3 finetuning on FUNSD Notebook; Arrow Error | ### Describe the bug
Arrow issues when running the example Notebook laptop locally on Mac with M1. Works on Google Collab.
**Notebook**: https://github.com/NielsRogge/Transformers-Tutorials/blob/master/LayoutLMv3/Fine_tune_LayoutLMv3_on_FUNSD_(HuggingFace_Trainer).ipynb
**Error**: `ValueError: Arrow type extensi... | 26 | Bug: LayoutLMv3 finetuning on FUNSD Notebook; Arrow Error
### Describe the bug
Arrow issues when running the example Notebook laptop locally on Mac with M1. Works on Google Collab.
**Notebook**: https://github.com/NielsRogge/Transformers-Tutorials/blob/master/LayoutLMv3/Fine_tune_LayoutLMv3_on_FUNSD_(HuggingFace... | [
-0.12534523010253906,
-0.696060299873352,
0.06675893068313599,
0.3161044716835022,
0.36773860454559326,
-0.21727141737937927,
0.570938229560852,
0.08089077472686768,
-0.18083691596984863,
-0.04032056778669357,
-0.07140608876943588,
0.6541163921356201,
-0.07819987833499908,
0.20544581115245... |
https://github.com/huggingface/datasets/issues/6417 | > Hi! The latest (patch) release (published a few hours ago) includes a fix for this [PyArrow security issue](https://github.com/advisories/GHSA-5wvp-7f3h-6wmm). To install it, run `pip install -U datasets`.
Thanks for the info and the latest release, it seems this has also solved my issue. First run after the updat... | Bug: LayoutLMv3 finetuning on FUNSD Notebook; Arrow Error | ### Describe the bug
Arrow issues when running the example Notebook laptop locally on Mac with M1. Works on Google Collab.
**Notebook**: https://github.com/NielsRogge/Transformers-Tutorials/blob/master/LayoutLMv3/Fine_tune_LayoutLMv3_on_FUNSD_(HuggingFace_Trainer).ipynb
**Error**: `ValueError: Arrow type extensi... | 60 | Bug: LayoutLMv3 finetuning on FUNSD Notebook; Arrow Error
### Describe the bug
Arrow issues when running the example Notebook laptop locally on Mac with M1. Works on Google Collab.
**Notebook**: https://github.com/NielsRogge/Transformers-Tutorials/blob/master/LayoutLMv3/Fine_tune_LayoutLMv3_on_FUNSD_(HuggingFace... | [
-0.12534523010253906,
-0.696060299873352,
0.06675893068313599,
0.3161044716835022,
0.36773860454559326,
-0.21727141737937927,
0.570938229560852,
0.08089077472686768,
-0.18083691596984863,
-0.04032056778669357,
-0.07140608876943588,
0.6541163921356201,
-0.07819987833499908,
0.20544581115245... |
https://github.com/huggingface/datasets/issues/6412 | Indeed, this is not a good practice. I've opened a PR that removes the token value from the (deprecation) warning. | User token is printed out! | This line prints user token on command line! Is it safe?
https://github.com/huggingface/datasets/blob/12ebe695b4748c5a26e08b44ed51955f74f5801d/src/datasets/load.py#L2091 | 20 | User token is printed out!
This line prints user token on command line! Is it safe?
https://github.com/huggingface/datasets/blob/12ebe695b4748c5a26e08b44ed51955f74f5801d/src/datasets/load.py#L2091
Indeed, this is not a good practice. I've opened a PR that removes the token value from the (deprecation) warning. | [
-0.0031762495636940002,
-0.1586783230304718,
-0.1245645135641098,
-0.01705431565642357,
0.22019343078136444,
0.023465313017368317,
0.41891172528266907,
-0.0682024210691452,
0.007502242922782898,
0.07574683427810669,
-0.003919064998626709,
0.028364498168230057,
-0.009011231362819672,
0.2457... |
https://github.com/huggingface/datasets/issues/6410 | Hi! You can avoid the error by requesting only the `jsonl` files. `dataset = load_dataset("ai4privacy/pii-masking-200k", data_files=["*.jsonl"])`.
Our data file inference does not filter out (incompatible) `json` files because `json` and `jsonl` use the same builder. Still, I think the inference should differentiat... | Datasets does not load HuggingFace Repository properly | ### Describe the bug
Dear Datasets team,
We just have published a dataset on Huggingface:
https://huggingface.co/ai4privacy
However, when trying to read it using the Dataset library we get an error. As I understand jsonl files are compatible, could you please clarify how we can solve the issue? Please let me ... | 60 | Datasets does not load HuggingFace Repository properly
### Describe the bug
Dear Datasets team,
We just have published a dataset on Huggingface:
https://huggingface.co/ai4privacy
However, when trying to read it using the Dataset library we get an error. As I understand jsonl files are compatible, could you ... | [
-0.20247381925582886,
-0.10110732913017273,
-0.026749910786747932,
0.7637389898300171,
0.29011261463165283,
-0.03652459383010864,
-0.00203618872910738,
0.24976053833961487,
0.019404388964176178,
0.02537638321518898,
-0.2364252805709839,
0.2252340018749237,
-0.05940572917461395,
0.331941127... |
https://github.com/huggingface/datasets/issues/6410 | Raising an error if there is a mix of json and jsonl in the builder makes sense yea | Datasets does not load HuggingFace Repository properly | ### Describe the bug
Dear Datasets team,
We just have published a dataset on Huggingface:
https://huggingface.co/ai4privacy
However, when trying to read it using the Dataset library we get an error. As I understand jsonl files are compatible, could you please clarify how we can solve the issue? Please let me ... | 18 | Datasets does not load HuggingFace Repository properly
### Describe the bug
Dear Datasets team,
We just have published a dataset on Huggingface:
https://huggingface.co/ai4privacy
However, when trying to read it using the Dataset library we get an error. As I understand jsonl files are compatible, could you ... | [
-0.20247381925582886,
-0.10110732913017273,
-0.026749910786747932,
0.7637389898300171,
0.29011261463165283,
-0.03652459383010864,
-0.00203618872910738,
0.24976053833961487,
0.019404388964176178,
0.02537638321518898,
-0.2364252805709839,
0.2252340018749237,
-0.05940572917461395,
0.331941127... |
https://github.com/huggingface/datasets/issues/6405 | The viewer is working now.
Based on the repo commit history, the bug was due to the incorrect format of the `features` field in the README YAML (`Value` requires `dtype`, e.g., `Value("string")`, but it was not specified) | ConfigNamesError on a simple CSV file | See https://huggingface.co/datasets/Nguyendo1999/mmath/discussions/1
```
Error code: ConfigNamesError
Exception: TypeError
Message: __init__() missing 1 required positional argument: 'dtype'
Traceback: Traceback (most recent call last):
File "/src/services/worker/src/worker/job_runn... | 37 | ConfigNamesError on a simple CSV file
See https://huggingface.co/datasets/Nguyendo1999/mmath/discussions/1
```
Error code: ConfigNamesError
Exception: TypeError
Message: __init__() missing 1 required positional argument: 'dtype'
Traceback: Traceback (most recent call last):
File "... | [
-0.37286028265953064,
-0.2837359309196472,
-0.09386485815048218,
0.07334684580564499,
0.42209455370903015,
0.25863489508628845,
0.5016807913780212,
0.33707672357559204,
-0.02036479488015175,
0.2753065824508667,
0.42774254083633423,
-0.016560789197683334,
0.06351710855960846,
0.372148096561... |
https://github.com/huggingface/datasets/issues/6403 | You are most likely using an outdated version of `datasets` in the notebook, which can be verified with the `!datasets-cli env` command. You can run `!pip install -U datasets` to update the installation. | Cannot import datasets on google colab (python 3.10.12) | ### Describe the bug
I'm trying A full colab demo notebook of zero-shot-distillation from https://github.com/huggingface/transformers/tree/main/examples/research_projects/zero-shot-distillation but i got this type of error when importing datasets on my google colab (python version is 3.10.12)

### Describe the bug
I'm trying A full colab demo notebook of zero-shot-distillation from https://github.com/huggingface/transformers/tree/main/examples/research_projects/zero-shot-distillation but i got this type of error when importing datasets on my google ... | [
-0.22408875823020935,
-0.2938651144504547,
0.08976825326681137,
0.43063440918922424,
0.21170759201049805,
-0.08203870803117752,
0.3200898766517639,
0.04798250272870064,
0.08357463032007217,
0.14291954040527344,
-0.5043290257453918,
0.30278438329696655,
-0.17893865704536438,
0.2081073820590... |
https://github.com/huggingface/datasets/issues/6401 | Seems like it's a problem with the dataset, since in the [README](https://huggingface.co/datasets/Hyperspace-Technologies/scp-wiki-text/blob/main/README.md) the validation is not specified. Try cloning the dataset, removing the README (or validation split), and loading it locally/ | dataset = load_dataset("Hyperspace-Technologies/scp-wiki-text") not working | ### Describe the bug
```
(datasets) mruserbox@guru-X99:/media/10TB_HHD/_LLM_DATASETS$ python dataset.py
Downloading readme: 100%|███████████████████████████████████| 360/360 [00:00<00:00, 2.16MB/s]
Downloading data: 100%|█████████████████████████████████| 65.1M/65.1M [00:19<00:00, 3.38MB/s]
Downloading data: 100... | 31 | dataset = load_dataset("Hyperspace-Technologies/scp-wiki-text") not working
### Describe the bug
```
(datasets) mruserbox@guru-X99:/media/10TB_HHD/_LLM_DATASETS$ python dataset.py
Downloading readme: 100%|███████████████████████████████████| 360/360 [00:00<00:00, 2.16MB/s]
Downloading data: 100%|███████████████... | [
-0.4304260015487671,
-0.008432343602180481,
-0.09224650263786316,
0.2506204843521118,
0.28447407484054565,
0.03460144251585007,
0.37043696641921997,
0.33112281560897827,
-0.011478707194328308,
0.15837866067886353,
0.052000924944877625,
0.12657028436660767,
-0.14188843965530396,
0.345798134... |
https://github.com/huggingface/datasets/issues/6401 | @VarunNSrivastava thanks brother, working beautiful now
```
C:\_Work\_datasets>py dataset.py
Downloading data files: 100%|████████████████████████████████████████████████████████████████████| 3/3 [00:00<?, ?it/s]
Extracting data files: 100%|████████████████████████████████████████████████████████████| 3/3 [00:00<... | dataset = load_dataset("Hyperspace-Technologies/scp-wiki-text") not working | ### Describe the bug
```
(datasets) mruserbox@guru-X99:/media/10TB_HHD/_LLM_DATASETS$ python dataset.py
Downloading readme: 100%|███████████████████████████████████| 360/360 [00:00<00:00, 2.16MB/s]
Downloading data: 100%|█████████████████████████████████| 65.1M/65.1M [00:19<00:00, 3.38MB/s]
Downloading data: 100... | 48 | dataset = load_dataset("Hyperspace-Technologies/scp-wiki-text") not working
### Describe the bug
```
(datasets) mruserbox@guru-X99:/media/10TB_HHD/_LLM_DATASETS$ python dataset.py
Downloading readme: 100%|███████████████████████████████████| 360/360 [00:00<00:00, 2.16MB/s]
Downloading data: 100%|███████████████... | [
-0.4304260015487671,
-0.008432343602180481,
-0.09224650263786316,
0.2506204843521118,
0.28447407484054565,
0.03460144251585007,
0.37043696641921997,
0.33112281560897827,
-0.011478707194328308,
0.15837866067886353,
0.052000924944877625,
0.12657028436660767,
-0.14188843965530396,
0.345798134... |
https://github.com/huggingface/datasets/issues/6400 | great idea IMO
this could be a `trust_remote_code=True` flag like in transformers. We could also default to loading the Parquet conversion rather than executing code (for dataset repos that have both) | Safely load datasets by disabling execution of dataset loading script | ### Feature request
Is there a way to disable execution of dataset loading script using `load_dataset`? This is a security vulnerability that could lead to arbitrary code execution.
Any suggested workarounds are welcome as well.
### Motivation
This is a security vulnerability that could lead to arbitrary code e... | 31 | Safely load datasets by disabling execution of dataset loading script
### Feature request
Is there a way to disable execution of dataset loading script using `load_dataset`? This is a security vulnerability that could lead to arbitrary code execution.
Any suggested workarounds are welcome as well.
### Motivati... | [
-0.49187153577804565,
0.24664939939975739,
-0.06180678680539131,
0.07278358936309814,
0.18239806592464447,
-0.06257647275924683,
0.47560447454452515,
-0.0017527136951684952,
0.3500739634037018,
0.19620494544506073,
-0.030168278142809868,
0.48009923100471497,
-0.09626144915819168,
0.5082423... |
https://github.com/huggingface/datasets/issues/6400 | We added the `trust_remote_code` argument to `load_dataset()` in `datasets` 2.16:
- in the future users will have to pass trust_remote_code=True to use datasets with a script
- for now we just show a warning when a dataset script is used
- we fallback on the Hugging Face Parquet exports when possible (to keep compat... | Safely load datasets by disabling execution of dataset loading script | ### Feature request
Is there a way to disable execution of dataset loading script using `load_dataset`? This is a security vulnerability that could lead to arbitrary code execution.
Any suggested workarounds are welcome as well.
### Motivation
This is a security vulnerability that could lead to arbitrary code e... | 76 | Safely load datasets by disabling execution of dataset loading script
### Feature request
Is there a way to disable execution of dataset loading script using `load_dataset`? This is a security vulnerability that could lead to arbitrary code execution.
Any suggested workarounds are welcome as well.
### Motivati... | [
-0.5202223062515259,
0.20792506635189056,
-0.08936712890863419,
0.15276484191417694,
0.1941721886396408,
-0.030967362225055695,
0.40098804235458374,
0.023934371769428253,
0.5325300097465515,
0.2042616307735443,
-0.026901861652731895,
0.5292146801948547,
-0.20548397302627563,
0.355250477790... |
https://github.com/huggingface/datasets/issues/6396 | https://github.com/huggingface/datasets-server/pull/2089#pullrequestreview-1724449532
> Yes, I understand now: they have disabled their `PyExtensionType` and we use it in `datasets` for arrays... | Issue with pyarrow 14.0.1 | See https://github.com/huggingface/datasets-server/pull/2089 for reference
```
from datasets import (Array2D, Dataset, Features)
feature_type = Array2D(shape=(2, 2), dtype="float32")
content = [[0.0, 0.0], [0.0, 0.0]]
features = Features({"col": feature_type})
dataset = Dataset.from_dict({"col": [content]}, fea... | 19 | Issue with pyarrow 14.0.1
See https://github.com/huggingface/datasets-server/pull/2089 for reference
```
from datasets import (Array2D, Dataset, Features)
feature_type = Array2D(shape=(2, 2), dtype="float32")
content = [[0.0, 0.0], [0.0, 0.0]]
features = Features({"col": feature_type})
dataset = Dataset.from_... | [
-0.10018604248762131,
0.13276909291744232,
-0.0065413713455200195,
0.12114937603473663,
0.33292287588119507,
-0.0022718533873558044,
0.5476638078689575,
0.42727774381637573,
-0.40590208768844604,
-0.01682676374912262,
-0.04357265681028366,
0.5135801434516907,
-0.038570161908864975,
0.18583... |
https://github.com/huggingface/datasets/issues/6395 | We've replaced our filelock implementation with the `filelock` package, so their repo is the right place to request this feature.
In the meantime, the following should work:
```python
import filelock
filelock.FileLock = filelock.SoftFileLock
import datasets
...
``` | Add ability to set lock type | ### Feature request
Allow setting file lock type, maybe from an environment variable
Currently, it only depends on whether fnctl is available:
https://github.com/huggingface/datasets/blob/12ebe695b4748c5a26e08b44ed51955f74f5801d/src/datasets/utils/filelock.py#L463-L470C16
### Motivation
In my environment... | 37 | Add ability to set lock type
### Feature request
Allow setting file lock type, maybe from an environment variable
Currently, it only depends on whether fnctl is available:
https://github.com/huggingface/datasets/blob/12ebe695b4748c5a26e08b44ed51955f74f5801d/src/datasets/utils/filelock.py#L463-L470C16
### ... | [
-0.3815123736858368,
-0.05841831862926483,
-0.07984399050474167,
-0.3621041476726532,
-0.15861397981643677,
-0.08662982285022736,
0.6261725425720215,
0.16512562334537506,
0.16967499256134033,
0.08532869815826416,
-0.34569329023361206,
0.21483486890792847,
-0.06048876792192459,
0.0627746880... |
https://github.com/huggingface/datasets/issues/6393 | Adding some more information, it seems to occur more frequently with large (millions of samples) datasets. | Filter occasionally hangs | ### Describe the bug
A call to `.filter` occasionally hangs (after the filter is complete, according to tqdm)
There is a trace produced
```
Exception ignored in: <function Dataset.__del__ at 0x7efb48130c10>
Traceback (most recent call last):
File "/usr/lib/python3/dist-packages/datasets/arrow_dataset.py", l... | 16 | Filter occasionally hangs
### Describe the bug
A call to `.filter` occasionally hangs (after the filter is complete, according to tqdm)
There is a trace produced
```
Exception ignored in: <function Dataset.__del__ at 0x7efb48130c10>
Traceback (most recent call last):
File "/usr/lib/python3/dist-packages/d... | [
-0.36410510540008545,
-0.007196187973022461,
-0.07324452698230743,
0.1654694676399231,
0.34145641326904297,
-0.19884663820266724,
0.07616069167852402,
0.03563345596194267,
0.13033874332904816,
0.10187101364135742,
0.2923228144645691,
0.4539174735546112,
-0.27917808294296265,
-0.01014070212... |
https://github.com/huggingface/datasets/issues/6393 | More information. My code is structured as (1) load (2) map (3) filter (4) filter. It was always the second filter that failed. Combining the two filters into one seems to reliably work. | Filter occasionally hangs | ### Describe the bug
A call to `.filter` occasionally hangs (after the filter is complete, according to tqdm)
There is a trace produced
```
Exception ignored in: <function Dataset.__del__ at 0x7efb48130c10>
Traceback (most recent call last):
File "/usr/lib/python3/dist-packages/datasets/arrow_dataset.py", l... | 33 | Filter occasionally hangs
### Describe the bug
A call to `.filter` occasionally hangs (after the filter is complete, according to tqdm)
There is a trace produced
```
Exception ignored in: <function Dataset.__del__ at 0x7efb48130c10>
Traceback (most recent call last):
File "/usr/lib/python3/dist-packages/d... | [
-0.30096906423568726,
0.02597421407699585,
-0.05257953330874443,
0.20952379703521729,
0.3754666745662689,
-0.19390510022640228,
0.14301031827926636,
0.054887741804122925,
0.15424728393554688,
0.05187126249074936,
0.2389901578426361,
0.46147793531417847,
-0.21464863419532776,
-0.03009072318... |
https://github.com/huggingface/datasets/issues/6393 | @lhoestq it'd be great if someone had a chance to look at this. I suspect it is impacting many users given the other issue that I linked. | Filter occasionally hangs | ### Describe the bug
A call to `.filter` occasionally hangs (after the filter is complete, according to tqdm)
There is a trace produced
```
Exception ignored in: <function Dataset.__del__ at 0x7efb48130c10>
Traceback (most recent call last):
File "/usr/lib/python3/dist-packages/datasets/arrow_dataset.py", l... | 27 | Filter occasionally hangs
### Describe the bug
A call to `.filter` occasionally hangs (after the filter is complete, according to tqdm)
There is a trace produced
```
Exception ignored in: <function Dataset.__del__ at 0x7efb48130c10>
Traceback (most recent call last):
File "/usr/lib/python3/dist-packages/d... | [
-0.34762266278266907,
0.039869725704193115,
-0.07918084412813187,
0.15619584918022156,
0.34971845149993896,
-0.21276167035102844,
0.09959720075130463,
0.061069048941135406,
0.08661271631717682,
0.08014142513275146,
0.26038306951522827,
0.47290682792663574,
-0.2593033015727997,
0.0341965556... |
https://github.com/huggingface/datasets/issues/6393 | Hi ! Sorry for the late response. Was it happening after the first or the second filter ?
It looks like an issue with the garbage collector (which makes it random). Maybe datasets created with `filter` are not always handled properly ? cc @mariosasko | Filter occasionally hangs | ### Describe the bug
A call to `.filter` occasionally hangs (after the filter is complete, according to tqdm)
There is a trace produced
```
Exception ignored in: <function Dataset.__del__ at 0x7efb48130c10>
Traceback (most recent call last):
File "/usr/lib/python3/dist-packages/datasets/arrow_dataset.py", l... | 44 | Filter occasionally hangs
### Describe the bug
A call to `.filter` occasionally hangs (after the filter is complete, according to tqdm)
There is a trace produced
```
Exception ignored in: <function Dataset.__del__ at 0x7efb48130c10>
Traceback (most recent call last):
File "/usr/lib/python3/dist-packages/d... | [
-0.29349881410598755,
-0.0039390623569488525,
-0.07677945494651794,
0.1827934831380844,
0.382412850856781,
-0.21573856472969055,
0.056217312812805176,
0.0004232563078403473,
0.11997350305318832,
0.09813044220209122,
0.30350425839424133,
0.47646021842956543,
-0.26908016204833984,
-0.0152659... |
https://github.com/huggingface/datasets/issues/6393 | It was after the second filter (and combining the two filters into one seemingly resolved it). I obviously haven't tried all settings to know that these details are causal, but it did work for me. | Filter occasionally hangs | ### Describe the bug
A call to `.filter` occasionally hangs (after the filter is complete, according to tqdm)
There is a trace produced
```
Exception ignored in: <function Dataset.__del__ at 0x7efb48130c10>
Traceback (most recent call last):
File "/usr/lib/python3/dist-packages/datasets/arrow_dataset.py", l... | 35 | Filter occasionally hangs
### Describe the bug
A call to `.filter` occasionally hangs (after the filter is complete, according to tqdm)
There is a trace produced
```
Exception ignored in: <function Dataset.__del__ at 0x7efb48130c10>
Traceback (most recent call last):
File "/usr/lib/python3/dist-packages/d... | [
-0.3603329062461853,
0.0368996262550354,
-0.0837249830365181,
0.11418618261814117,
0.3722602427005768,
-0.2295568734407425,
0.08928656578063965,
0.06284244358539581,
0.0738249123096466,
0.08880357444286346,
0.2227749228477478,
0.4795060157775879,
-0.24935586750507355,
0.043282948434352875,... |
https://github.com/huggingface/datasets/issues/6393 | Thanks, that's good to know.
The stacktrace suggests an issue when `del self._indices` is called, which happens when a filtered dataset falls out of scope. The indices are a PyArrow table memory mapped from disk, so I'm not quite sure how calling `del` on it can cause this issue. We do `del self._indices` to make su... | Filter occasionally hangs | ### Describe the bug
A call to `.filter` occasionally hangs (after the filter is complete, according to tqdm)
There is a trace produced
```
Exception ignored in: <function Dataset.__del__ at 0x7efb48130c10>
Traceback (most recent call last):
File "/usr/lib/python3/dist-packages/datasets/arrow_dataset.py", l... | 93 | Filter occasionally hangs
### Describe the bug
A call to `.filter` occasionally hangs (after the filter is complete, according to tqdm)
There is a trace produced
```
Exception ignored in: <function Dataset.__del__ at 0x7efb48130c10>
Traceback (most recent call last):
File "/usr/lib/python3/dist-packages/d... | [
-0.2699989974498749,
0.07587841153144836,
-0.05574621260166168,
0.23851166665554047,
0.3574141263961792,
-0.13207729160785675,
0.14118105173110962,
0.05039813369512558,
-0.031715277582407,
0.16522642970085144,
0.18220847845077515,
0.4575658440589905,
-0.27161359786987305,
-0.09806235134601... |
https://github.com/huggingface/datasets/issues/6393 | Yeah, I have a reliable repro, but it is not even close to minimal and uses a dataset I can't share. Perhaps you could try getting close to my setting.
(1) make a large (~20GB) jsonl with prompt/response pairs
(2) load it on a linux machine (`dataset = load_dataset(...)`)
(3) map a tokenizer to it, with multiproce... | Filter occasionally hangs | ### Describe the bug
A call to `.filter` occasionally hangs (after the filter is complete, according to tqdm)
There is a trace produced
```
Exception ignored in: <function Dataset.__del__ at 0x7efb48130c10>
Traceback (most recent call last):
File "/usr/lib/python3/dist-packages/datasets/arrow_dataset.py", l... | 108 | Filter occasionally hangs
### Describe the bug
A call to `.filter` occasionally hangs (after the filter is complete, according to tqdm)
There is a trace produced
```
Exception ignored in: <function Dataset.__del__ at 0x7efb48130c10>
Traceback (most recent call last):
File "/usr/lib/python3/dist-packages/d... | [
-0.39694449305534363,
0.12597009539604187,
-0.05943926051259041,
0.1702365279197693,
0.27565693855285645,
-0.29603731632232666,
0.04032320901751518,
0.13020668923854828,
0.09984563291072845,
0.04136590287089348,
0.27463430166244507,
0.5538307428359985,
-0.28259706497192383,
-0.001442655920... |
https://github.com/huggingface/datasets/issues/6392 | Hi! We made some improvements to `push_to_hub` to make it more robust a couple of weeks ago but haven't published a release in the meantime, so it would help if you could install `datasets` from `main` (`pip install https://github.com/huggingface/datasets`) and let us know if this improved version of `push_to_hub` reso... | `push_to_hub` is not robust to hub closing connection | ### Describe the bug
Like to #6172, `push_to_hub` will crash if Hub resets the connection and raise the following error:
```
Pushing dataset shards to the dataset hub: 32%|███▏ | 54/171 [06:38<14:23, 7.38s/it]
Traceback (most recent call last):
File "/admin/home-piraka9011/.virtualenvs/w2v2/lib/python3.8/... | 86 | `push_to_hub` is not robust to hub closing connection
### Describe the bug
Like to #6172, `push_to_hub` will crash if Hub resets the connection and raise the following error:
```
Pushing dataset shards to the dataset hub: 32%|███▏ | 54/171 [06:38<14:23, 7.38s/it]
Traceback (most recent call last):
File ... | [
-0.5284816026687622,
-0.3638116121292114,
-0.06487172842025757,
0.1819303184747696,
0.3242587447166443,
-0.20704427361488342,
0.17081235349178314,
0.3081148862838745,
-0.03854190558195114,
-0.022969290614128113,
-0.19621890783309937,
0.18021035194396973,
0.1735050529241562,
0.0446992143988... |
https://github.com/huggingface/datasets/issues/6392 | The update helped push more data.
However it still crashed a little later:
```
Traceback (most recent call last):
File "/admin/home-piraka9011/.virtualenvs/w2v2/lib/python3.8/site-packages/huggingface_hub/utils/_errors.py", line 270, in hf_raise_for_status
response.raise_for_status()
File "/admin/home-p... | `push_to_hub` is not robust to hub closing connection | ### Describe the bug
Like to #6172, `push_to_hub` will crash if Hub resets the connection and raise the following error:
```
Pushing dataset shards to the dataset hub: 32%|███▏ | 54/171 [06:38<14:23, 7.38s/it]
Traceback (most recent call last):
File "/admin/home-piraka9011/.virtualenvs/w2v2/lib/python3.8/... | 223 | `push_to_hub` is not robust to hub closing connection
### Describe the bug
Like to #6172, `push_to_hub` will crash if Hub resets the connection and raise the following error:
```
Pushing dataset shards to the dataset hub: 32%|███▏ | 54/171 [06:38<14:23, 7.38s/it]
Traceback (most recent call last):
File ... | [
-0.5284816026687622,
-0.3638116121292114,
-0.06487172842025757,
0.1819303184747696,
0.3242587447166443,
-0.20704427361488342,
0.17081235349178314,
0.3081148862838745,
-0.03854190558195114,
-0.022969290614128113,
-0.19621890783309937,
0.18021035194396973,
0.1735050529241562,
0.0446992143988... |
https://github.com/huggingface/datasets/issues/6392 | I think the previous implementation was actually better: it pushes to the hub every shard. So if it fails, as long as the shards have the same checksum, it will skip the ones that have been pushed.
The implementation in `main` pushes commits at the end, so when it fails, there are no commits and therefore restarts f... | `push_to_hub` is not robust to hub closing connection | ### Describe the bug
Like to #6172, `push_to_hub` will crash if Hub resets the connection and raise the following error:
```
Pushing dataset shards to the dataset hub: 32%|███▏ | 54/171 [06:38<14:23, 7.38s/it]
Traceback (most recent call last):
File "/admin/home-piraka9011/.virtualenvs/w2v2/lib/python3.8/... | 312 | `push_to_hub` is not robust to hub closing connection
### Describe the bug
Like to #6172, `push_to_hub` will crash if Hub resets the connection and raise the following error:
```
Pushing dataset shards to the dataset hub: 32%|███▏ | 54/171 [06:38<14:23, 7.38s/it]
Traceback (most recent call last):
File ... | [
-0.5284816026687622,
-0.3638116121292114,
-0.06487172842025757,
0.1819303184747696,
0.3242587447166443,
-0.20704427361488342,
0.17081235349178314,
0.3081148862838745,
-0.03854190558195114,
-0.022969290614128113,
-0.19621890783309937,
0.18021035194396973,
0.1735050529241562,
0.0446992143988... |
https://github.com/huggingface/datasets/issues/6392 | There's a new error from the hub now:
```
Pushing dataset shards to the dataset hub: 49%|████▉ | 128/261 [11:38<12:05, 5.45s/it]
Traceback (most recent call last):
File "/admin/home-piraka9011/.virtualenvs/w2v2/lib/python3.8/site-packages/huggingface_hub/utils/_errors.py", line 270, in hf_raise_for_status
... | `push_to_hub` is not robust to hub closing connection | ### Describe the bug
Like to #6172, `push_to_hub` will crash if Hub resets the connection and raise the following error:
```
Pushing dataset shards to the dataset hub: 32%|███▏ | 54/171 [06:38<14:23, 7.38s/it]
Traceback (most recent call last):
File "/admin/home-piraka9011/.virtualenvs/w2v2/lib/python3.8/... | 225 | `push_to_hub` is not robust to hub closing connection
### Describe the bug
Like to #6172, `push_to_hub` will crash if Hub resets the connection and raise the following error:
```
Pushing dataset shards to the dataset hub: 32%|███▏ | 54/171 [06:38<14:23, 7.38s/it]
Traceback (most recent call last):
File ... | [
-0.5284816026687622,
-0.3638116121292114,
-0.06487172842025757,
0.1819303184747696,
0.3242587447166443,
-0.20704427361488342,
0.17081235349178314,
0.3081148862838745,
-0.03854190558195114,
-0.022969290614128113,
-0.19621890783309937,
0.18021035194396973,
0.1735050529241562,
0.0446992143988... |
https://github.com/huggingface/datasets/issues/6392 | > think the previous implementation was actually better: it pushes to the hub every shard. So if it fails, as long as the shards have the same checksum, it will skip the ones that have been pushed.
>
>The implementation in main pushes commits at the end, so when it fails, there are no commits and therefore restarts ... | `push_to_hub` is not robust to hub closing connection | ### Describe the bug
Like to #6172, `push_to_hub` will crash if Hub resets the connection and raise the following error:
```
Pushing dataset shards to the dataset hub: 32%|███▏ | 54/171 [06:38<14:23, 7.38s/it]
Traceback (most recent call last):
File "/admin/home-piraka9011/.virtualenvs/w2v2/lib/python3.8/... | 178 | `push_to_hub` is not robust to hub closing connection
### Describe the bug
Like to #6172, `push_to_hub` will crash if Hub resets the connection and raise the following error:
```
Pushing dataset shards to the dataset hub: 32%|███▏ | 54/171 [06:38<14:23, 7.38s/it]
Traceback (most recent call last):
File ... | [
-0.5284816026687622,
-0.3638116121292114,
-0.06487172842025757,
0.1819303184747696,
0.3242587447166443,
-0.20704427361488342,
0.17081235349178314,
0.3081148862838745,
-0.03854190558195114,
-0.022969290614128113,
-0.19621890783309937,
0.18021035194396973,
0.1735050529241562,
0.0446992143988... |
https://github.com/huggingface/datasets/issues/6392 | But I agree that the `500 Server Error` returned by the Hub is annoying. Earlier today, I also got it on a small 5GB dataset (with 500 MB shards).
@Wauplin @julien-c Is there something we can do about this? | `push_to_hub` is not robust to hub closing connection | ### Describe the bug
Like to #6172, `push_to_hub` will crash if Hub resets the connection and raise the following error:
```
Pushing dataset shards to the dataset hub: 32%|███▏ | 54/171 [06:38<14:23, 7.38s/it]
Traceback (most recent call last):
File "/admin/home-piraka9011/.virtualenvs/w2v2/lib/python3.8/... | 39 | `push_to_hub` is not robust to hub closing connection
### Describe the bug
Like to #6172, `push_to_hub` will crash if Hub resets the connection and raise the following error:
```
Pushing dataset shards to the dataset hub: 32%|███▏ | 54/171 [06:38<14:23, 7.38s/it]
Traceback (most recent call last):
File ... | [
-0.5284816026687622,
-0.3638116121292114,
-0.06487172842025757,
0.1819303184747696,
0.3242587447166443,
-0.20704427361488342,
0.17081235349178314,
0.3081148862838745,
-0.03854190558195114,
-0.022969290614128113,
-0.19621890783309937,
0.18021035194396973,
0.1735050529241562,
0.0446992143988... |
https://github.com/huggingface/datasets/issues/6392 | @mariosasko can't do much if AWS raises a HTTP 500 unfortunately (we are simply pushing data to a S3 bucket).
What we can do is to add a retry mechanism in the multi-part upload logic here: https://github.com/huggingface/huggingface_hub/blob/c972cba1fecb456a7b3325cdd1fdbcc425f21f94/src/huggingface_hub/lfs.py#L370 :con... | `push_to_hub` is not robust to hub closing connection | ### Describe the bug
Like to #6172, `push_to_hub` will crash if Hub resets the connection and raise the following error:
```
Pushing dataset shards to the dataset hub: 32%|███▏ | 54/171 [06:38<14:23, 7.38s/it]
Traceback (most recent call last):
File "/admin/home-piraka9011/.virtualenvs/w2v2/lib/python3.8/... | 38 | `push_to_hub` is not robust to hub closing connection
### Describe the bug
Like to #6172, `push_to_hub` will crash if Hub resets the connection and raise the following error:
```
Pushing dataset shards to the dataset hub: 32%|███▏ | 54/171 [06:38<14:23, 7.38s/it]
Traceback (most recent call last):
File ... | [
-0.5284816026687622,
-0.3638116121292114,
-0.06487172842025757,
0.1819303184747696,
0.3242587447166443,
-0.20704427361488342,
0.17081235349178314,
0.3081148862838745,
-0.03854190558195114,
-0.022969290614128113,
-0.19621890783309937,
0.18021035194396973,
0.1735050529241562,
0.0446992143988... |
https://github.com/huggingface/datasets/issues/6392 | > That code already retries the request using http_backoff, no?
Currently only on HTTP 503 by default. We should add 500 as well (and hope it is a transient error from AWS) | `push_to_hub` is not robust to hub closing connection | ### Describe the bug
Like to #6172, `push_to_hub` will crash if Hub resets the connection and raise the following error:
```
Pushing dataset shards to the dataset hub: 32%|███▏ | 54/171 [06:38<14:23, 7.38s/it]
Traceback (most recent call last):
File "/admin/home-piraka9011/.virtualenvs/w2v2/lib/python3.8/... | 32 | `push_to_hub` is not robust to hub closing connection
### Describe the bug
Like to #6172, `push_to_hub` will crash if Hub resets the connection and raise the following error:
```
Pushing dataset shards to the dataset hub: 32%|███▏ | 54/171 [06:38<14:23, 7.38s/it]
Traceback (most recent call last):
File ... | [
-0.5284816026687622,
-0.3638116121292114,
-0.06487172842025757,
0.1819303184747696,
0.3242587447166443,
-0.20704427361488342,
0.17081235349178314,
0.3081148862838745,
-0.03854190558195114,
-0.022969290614128113,
-0.19621890783309937,
0.18021035194396973,
0.1735050529241562,
0.0446992143988... |
https://github.com/huggingface/datasets/issues/6392 | Opened a PR to retry in case S3 raises HTTP 500. Will also retry on any `ConnectionError` (connection reset by peer, connection lost,...). Hopefully this should make the upload process more robust to transient errors. | `push_to_hub` is not robust to hub closing connection | ### Describe the bug
Like to #6172, `push_to_hub` will crash if Hub resets the connection and raise the following error:
```
Pushing dataset shards to the dataset hub: 32%|███▏ | 54/171 [06:38<14:23, 7.38s/it]
Traceback (most recent call last):
File "/admin/home-piraka9011/.virtualenvs/w2v2/lib/python3.8/... | 35 | `push_to_hub` is not robust to hub closing connection
### Describe the bug
Like to #6172, `push_to_hub` will crash if Hub resets the connection and raise the following error:
```
Pushing dataset shards to the dataset hub: 32%|███▏ | 54/171 [06:38<14:23, 7.38s/it]
Traceback (most recent call last):
File ... | [
-0.5284816026687622,
-0.3638116121292114,
-0.06487172842025757,
0.1819303184747696,
0.3242587447166443,
-0.20704427361488342,
0.17081235349178314,
0.3081148862838745,
-0.03854190558195114,
-0.022969290614128113,
-0.19621890783309937,
0.18021035194396973,
0.1735050529241562,
0.0446992143988... |
https://github.com/huggingface/datasets/issues/6392 | I still get the same error, using `push_to_hub`. Using `git lfs` and pushing the files solved it for me. | `push_to_hub` is not robust to hub closing connection | ### Describe the bug
Like to #6172, `push_to_hub` will crash if Hub resets the connection and raise the following error:
```
Pushing dataset shards to the dataset hub: 32%|███▏ | 54/171 [06:38<14:23, 7.38s/it]
Traceback (most recent call last):
File "/admin/home-piraka9011/.virtualenvs/w2v2/lib/python3.8/... | 19 | `push_to_hub` is not robust to hub closing connection
### Describe the bug
Like to #6172, `push_to_hub` will crash if Hub resets the connection and raise the following error:
```
Pushing dataset shards to the dataset hub: 32%|███▏ | 54/171 [06:38<14:23, 7.38s/it]
Traceback (most recent call last):
File ... | [
-0.5284816026687622,
-0.3638116121292114,
-0.06487172842025757,
0.1819303184747696,
0.3242587447166443,
-0.20704427361488342,
0.17081235349178314,
0.3081148862838745,
-0.03854190558195114,
-0.022969290614128113,
-0.19621890783309937,
0.18021035194396973,
0.1735050529241562,
0.0446992143988... |
https://github.com/huggingface/datasets/issues/6392 | @BEpresent the fix has not been released yet. You can expect a release of `huggingface_hub` (with this fix) today or tomorrow :) | `push_to_hub` is not robust to hub closing connection | ### Describe the bug
Like to #6172, `push_to_hub` will crash if Hub resets the connection and raise the following error:
```
Pushing dataset shards to the dataset hub: 32%|███▏ | 54/171 [06:38<14:23, 7.38s/it]
Traceback (most recent call last):
File "/admin/home-piraka9011/.virtualenvs/w2v2/lib/python3.8/... | 22 | `push_to_hub` is not robust to hub closing connection
### Describe the bug
Like to #6172, `push_to_hub` will crash if Hub resets the connection and raise the following error:
```
Pushing dataset shards to the dataset hub: 32%|███▏ | 54/171 [06:38<14:23, 7.38s/it]
Traceback (most recent call last):
File ... | [
-0.5284816026687622,
-0.3638116121292114,
-0.06487172842025757,
0.1819303184747696,
0.3242587447166443,
-0.20704427361488342,
0.17081235349178314,
0.3081148862838745,
-0.03854190558195114,
-0.022969290614128113,
-0.19621890783309937,
0.18021035194396973,
0.1735050529241562,
0.0446992143988... |
https://github.com/huggingface/datasets/issues/6389 | I managed a workaround eventually but I don't know what it was (I made a lot of changes to seq2seq). I'll try to include generating code in the future. (If I close, I don't know if you see it. Feel free to close; I'll re-open if I encounter it again (if I can)). | Index 339 out of range for dataset of size 339 <-- save_to_file() | ### Describe the bug
When saving out some Audio() data.
The data is audio recordings with associated 'sentences'.
(They use the audio 'bytes' approach because they're clips within audio files).
Code is below the traceback (I can't upload the voice audio/text (it's not even me)).
```
Traceback (most recent call ... | 53 | Index 339 out of range for dataset of size 339 <-- save_to_file()
### Describe the bug
When saving out some Audio() data.
The data is audio recordings with associated 'sentences'.
(They use the audio 'bytes' approach because they're clips within audio files).
Code is below the traceback (I can't upload the voice ... | [
-0.20405060052871704,
-0.13199637830257416,
0.041832633316516876,
0.332999050617218,
0.16036248207092285,
0.04094874858856201,
0.1847858875989914,
0.642251193523407,
-0.3258969485759735,
0.2800061106681824,
-0.2070016860961914,
-0.11669199913740158,
-0.06349034607410431,
-0.200472950935363... |
https://github.com/huggingface/datasets/issues/6387 | Feel free to use `dataset.save_to_disk(...)`, then scp the directory containing the saved dataset and reload it on your other machine using `dataset = load_from_disk(...)` | How to load existing downloaded dataset ? | Hi @mariosasko @lhoestq @katielink
Thanks for your contribution and hard work.
### Feature request
First, I download a dataset as normal by:
```
from datasets import load_dataset
dataset = load_dataset('username/data_name', cache_dir='data')
```
The dataset format in `data` directory will be:
```
... | 24 | How to load existing downloaded dataset ?
Hi @mariosasko @lhoestq @katielink
Thanks for your contribution and hard work.
### Feature request
First, I download a dataset as normal by:
```
from datasets import load_dataset
dataset = load_dataset('username/data_name', cache_dir='data')
```
The datase... | [
-0.38434135913848877,
-0.18714748322963715,
0.02825199067592621,
0.2958897352218628,
0.0300179123878479,
-0.0012773796916007996,
0.03767338767647743,
0.05174326151609421,
0.2660689651966095,
0.220025897026062,
-0.3922737240791321,
0.22650328278541565,
-0.003428516909480095,
0.7020697593688... |
https://github.com/huggingface/datasets/issues/6386 | Ah I think the `line-profiler` log is off-by-one and it is in fact the `extract_batch` method that's taking forever. Will investigate further. | Formatting overhead | ### Describe the bug
Hi! I very recently noticed that my training time is dominated by batch formatting. Using Lightning's profilers, I located the bottleneck within `datasets.formatting.formatting` and then narrowed it down with `line-profiler`. It turns out that almost all of the overhead is due to creating new inst... | 22 | Formatting overhead
### Describe the bug
Hi! I very recently noticed that my training time is dominated by batch formatting. Using Lightning's profilers, I located the bottleneck within `datasets.formatting.formatting` and then narrowed it down with `line-profiler`. It turns out that almost all of the overhead is du... | [
-0.17785169184207916,
-0.23078474402427673,
0.02506566047668457,
0.2501610517501831,
0.35222551226615906,
-0.01755823940038681,
0.33603334426879883,
0.12893426418304443,
-0.38654667139053345,
-0.017012380063533783,
0.265005886554718,
0.41242772340774536,
-0.4074646830558777,
0.148731440305... |
https://github.com/huggingface/datasets/issues/6385 | The `answers.text` field in the JSON dataset needs to be a list of strings, not a string.
So, here is the fixed code:
```python
from huggingface_hub import notebook_login
from datasets import load_dataset
notebook_login("mymailadresse", "mypassword")
squad = load_dataset("squad", split="train[:5000]")
sq... | Get an error when i try to concatenate the squad dataset with my own dataset | ### Describe the bug
Hello,
I'm new here and I need to concatenate the squad dataset with my own dataset i created. I find the following error when i try to do it: Traceback (most recent call last):
Cell In[9], line 1
concatenated_dataset = concatenate_datasets([train_dataset, dataset1])
File ~\ana... | 126 | Get an error when i try to concatenate the squad dataset with my own dataset
### Describe the bug
Hello,
I'm new here and I need to concatenate the squad dataset with my own dataset i created. I find the following error when i try to do it: Traceback (most recent call last):
Cell In[9], line 1
concate... | [
-0.22193357348442078,
-0.18991602957248688,
0.010590378195047379,
0.12742571532726288,
0.24476563930511475,
0.16857588291168213,
0.3418557345867157,
0.2883613705635071,
-0.3683692514896393,
-0.008571326732635498,
-0.11954230815172195,
0.4824563264846802,
0.25596916675567627,
0.118566699326... |
https://github.com/huggingface/datasets/issues/6382 | Hey @SauravMaheshkar ! Just responded to your email.
_For transparency, copying part of my response here:_
I agree, it would be really great to have this and other BenchMD datasets easily accessible on the hub.
I think the main limiting factor is that the ChexPert dataset is currently hosted on the Stanford AIMI... | Add CheXpert dataset for vision | ### Feature request
### Name
**CheXpert: A Large Chest Radiograph Dataset with Uncertainty Labels and Expert Comparison**
### Paper
https://arxiv.org/abs/1901.07031
### Data
https://stanfordaimi.azurewebsites.net/datasets/8cbd9ed4-2eb9-4565-affc-111cf4f7ebe2
### Motivation
CheXpert is one of the fund... | 126 | Add CheXpert dataset for vision
### Feature request
### Name
**CheXpert: A Large Chest Radiograph Dataset with Uncertainty Labels and Expert Comparison**
### Paper
https://arxiv.org/abs/1901.07031
### Data
https://stanfordaimi.azurewebsites.net/datasets/8cbd9ed4-2eb9-4565-affc-111cf4f7ebe2
### Motiv... | [
-0.29031553864479065,
-0.3038102984428406,
-0.09949086606502533,
-0.07802556455135345,
-0.11682909727096558,
0.011799231171607971,
0.2542484998703003,
0.46563032269477844,
0.16077250242233276,
0.38705974817276,
-0.0760834664106369,
0.09143735468387604,
-0.2539331912994385,
0.44283682107925... |
https://github.com/huggingface/datasets/issues/6382 | Hi @katielink , I would also love to be on board and contribute to this loading script/project if it is still being developed. I'm interested because I personally would like to gain access to the CheXpert dataset and am facing some weird issues, so I'd like to sort it out for me, and potentially others. Please keep me ... | Add CheXpert dataset for vision | ### Feature request
### Name
**CheXpert: A Large Chest Radiograph Dataset with Uncertainty Labels and Expert Comparison**
### Paper
https://arxiv.org/abs/1901.07031
### Data
https://stanfordaimi.azurewebsites.net/datasets/8cbd9ed4-2eb9-4565-affc-111cf4f7ebe2
### Motivation
CheXpert is one of the fund... | 66 | Add CheXpert dataset for vision
### Feature request
### Name
**CheXpert: A Large Chest Radiograph Dataset with Uncertainty Labels and Expert Comparison**
### Paper
https://arxiv.org/abs/1901.07031
### Data
https://stanfordaimi.azurewebsites.net/datasets/8cbd9ed4-2eb9-4565-affc-111cf4f7ebe2
### Motiv... | [
-0.18617260456085205,
-0.33047229051589966,
-0.07788653671741486,
-0.0753679946064949,
-0.021564843133091927,
-0.10620618611574173,
0.06507625430822372,
0.41927891969680786,
0.10648054629564285,
0.29680293798446655,
0.09117306023836136,
0.04682594910264015,
-0.33251747488975525,
0.52854597... |
https://github.com/huggingface/datasets/issues/6376 | I did not store it at the time but I'll try to re-do a mwe next week to get it again | Caching problem when deleting a dataset | ### Describe the bug
Pushing a dataset with n + m features to a repo which was deleted, but contained n features, will fail.
### Steps to reproduce the bug
1. Create a dataset with n features per row
2. `dataset.push_to_hub(YOUR_PATH, SPLIT, token=TOKEN)`
3. Go on the hub, delete the repo at `YOUR_PATH`
4. Update... | 21 | Caching problem when deleting a dataset
### Describe the bug
Pushing a dataset with n + m features to a repo which was deleted, but contained n features, will fail.
### Steps to reproduce the bug
1. Create a dataset with n features per row
2. `dataset.push_to_hub(YOUR_PATH, SPLIT, token=TOKEN)`
3. Go on the hub... | [
-0.17563989758491516,
-0.15957209467887878,
-0.042584508657455444,
0.3126749098300934,
0.01576206274330616,
-0.0614800751209259,
0.2875654399394989,
0.2968783676624298,
0.19769355654716492,
-0.022289790213108063,
0.17500968277454376,
0.17467544972896576,
0.00816064327955246,
0.181026279926... |
https://github.com/huggingface/datasets/issues/6376 | I haven't managed to reproduce this issue using a [notebook](https://colab.research.google.com/drive/1m6eduYun7pFTkigrCJAFgw0BghlbvXIL?usp=sharing) that follows the steps to reproduce the bug. So, I'm closing it.
But feel free to re-open it if you have a better reproducer. | Caching problem when deleting a dataset | ### Describe the bug
Pushing a dataset with n + m features to a repo which was deleted, but contained n features, will fail.
### Steps to reproduce the bug
1. Create a dataset with n features per row
2. `dataset.push_to_hub(YOUR_PATH, SPLIT, token=TOKEN)`
3. Go on the hub, delete the repo at `YOUR_PATH`
4. Update... | 34 | Caching problem when deleting a dataset
### Describe the bug
Pushing a dataset with n + m features to a repo which was deleted, but contained n features, will fail.
### Steps to reproduce the bug
1. Create a dataset with n features per row
2. `dataset.push_to_hub(YOUR_PATH, SPLIT, token=TOKEN)`
3. Go on the hub... | [
-0.09146799147129059,
-0.13599282503128052,
-0.007731638848781586,
0.3452052175998688,
-0.07417566329240799,
-0.020083405077457428,
0.3198857009410858,
0.2709159255027771,
0.19415250420570374,
-0.043790534138679504,
0.20972475409507751,
0.16039246320724487,
-0.02975871041417122,
0.13909795... |
https://github.com/huggingface/datasets/issues/6371 | Indeed, setting `try_from_gcs` to `False` makes sense for `from_generator`.
We plan to deprecate and remove `try_from_hf_gcs` soon, as we can use Hub for file hosting now, but this is a good temporary fix.
| `Dataset.from_generator` should not try to download from HF GCS | ### Describe the bug
When using [`Dataset.from_generator`](https://github.com/huggingface/datasets/blob/c9c1166e1cf81d38534020f9c167b326585339e5/src/datasets/arrow_dataset.py#L1072) with `streaming=False`, the internal logic will call [`download_and_prepare`](https://github.com/huggingface/datasets/blob/main/src/datas... | 33 | `Dataset.from_generator` should not try to download from HF GCS
### Describe the bug
When using [`Dataset.from_generator`](https://github.com/huggingface/datasets/blob/c9c1166e1cf81d38534020f9c167b326585339e5/src/datasets/arrow_dataset.py#L1072) with `streaming=False`, the internal logic will call [`download_and_pre... | [
-0.24892577528953552,
-0.3623090982437134,
0.012908365577459335,
0.22331032156944275,
0.23133710026741028,
-0.19252106547355652,
0.11532336473464966,
0.07888228446245193,
0.08868780732154846,
0.28646352887153625,
0.08888961374759674,
0.050307776778936386,
0.14478136599063873,
-0.0412699133... |
https://github.com/huggingface/datasets/issues/6370 | I figured it out. I found that `Trainer` does not work with TensorDataset even though the document says it uses it. Instead, I ended up creating a dictionary and converting it to a dataset using `dataset.Dataset.from_dict()`.
I will leave this post open for a while. If someone knows a better approach, please leave a... | TensorDataset format does not work with Trainer from transformers | ### Describe the bug
The model was built to do fine tunning on BERT model for relation extraction.
trainer.train() returns an error message ```TypeError: vars() argument must have __dict__ attribute``` when it has `train_dataset` generated from `torch.utils.data.TensorDataset`
However, in the document, the req... | 55 | TensorDataset format does not work with Trainer from transformers
### Describe the bug
The model was built to do fine tunning on BERT model for relation extraction.
trainer.train() returns an error message ```TypeError: vars() argument must have __dict__ attribute``` when it has `train_dataset` generated from `... | [
-0.03787489980459213,
-0.3518880009651184,
0.05201185494661331,
0.4861849844455719,
0.4192970097064972,
-0.05809000879526138,
0.6545382142066956,
0.43648281693458557,
-0.1592072993516922,
-0.005945086479187012,
-0.25915586948394775,
0.32222020626068115,
-0.13267740607261658,
-0.25136822462... |
https://github.com/huggingface/datasets/issues/6370 | Only issues directly related to the HF datasets library should be reported here. ~So, I'm transferring this issue to the `transformers` repo.~ I'm not a `transformers` maintainer, so GitHub doesn't let me transfer it there :(. This means you need to do it manually. | TensorDataset format does not work with Trainer from transformers | ### Describe the bug
The model was built to do fine tunning on BERT model for relation extraction.
trainer.train() returns an error message ```TypeError: vars() argument must have __dict__ attribute``` when it has `train_dataset` generated from `torch.utils.data.TensorDataset`
However, in the document, the req... | 44 | TensorDataset format does not work with Trainer from transformers
### Describe the bug
The model was built to do fine tunning on BERT model for relation extraction.
trainer.train() returns an error message ```TypeError: vars() argument must have __dict__ attribute``` when it has `train_dataset` generated from `... | [
-0.03787489980459213,
-0.3518880009651184,
0.05201185494661331,
0.4861849844455719,
0.4192970097064972,
-0.05809000879526138,
0.6545382142066956,
0.43648281693458557,
-0.1592072993516922,
-0.005945086479187012,
-0.25915586948394775,
0.32222020626068115,
-0.13267740607261658,
-0.25136822462... |
https://github.com/huggingface/datasets/issues/6369 | The inconsistency may be caused by the usage of "update_fingerprint" and setting "trust_remote_code" to "True."
When the tokenizer employs "trust_remote_code," the behavior of the map function varies with each code execution. Even if the remote code of the tokenizer remains the same, the result of "asher.hexdigest()" ... | Multi process map did not load cache file correctly | ### Describe the bug
When I was training model on Multiple GPUs by DDP, the dataset is tokenized multiple times after main process.


![1698820501568]... | [
-0.3758842647075653,
-0.5878843069076538,
0.032787516713142395,
0.44652554392814636,
0.30833250284194946,
-0.06708855926990509,
0.3846704661846161,
-0.0230244230479002,
0.11546281725168228,
-0.130690798163414,
0.03220471739768982,
0.1839643120765686,
0.011365003883838654,
-0.16018880903720... |
https://github.com/huggingface/datasets/issues/6369 | The issue may be related to problems previously discussed in GitHub issues [#3847](https://github.com/huggingface/datasets/issues/3847) and [#6318](https://github.com/huggingface/datasets/pull/6318).
This arises from the fact that tokenizer.tokens_trie._tokens is an unordered set, leading to varying hash results:
`v... | Multi process map did not load cache file correctly | ### Describe the bug
When I was training model on Multiple GPUs by DDP, the dataset is tokenized multiple times after main process.


![1698820501568]... | [
-0.3758842647075653,
-0.5878843069076538,
0.032787516713142395,
0.44652554392814636,
0.30833250284194946,
-0.06708855926990509,
0.3846704661846161,
-0.0230244230479002,
0.11546281725168228,
-0.130690798163414,
0.03220471739768982,
0.1839643120765686,
0.011365003883838654,
-0.16018880903720... |
https://github.com/huggingface/datasets/issues/6369 | We now sort `set` and `dict` items to make their hashes deterministic (install from `main` with `pip install git+https://github.com/huggingface/datasets` to test this). Consequently, this should also make the `tokenizer.tokens_trie`'s hash deterministic. Feel free to re-open the issue if this is not the case. | Multi process map did not load cache file correctly | ### Describe the bug
When I was training model on Multiple GPUs by DDP, the dataset is tokenized multiple times after main process.


![1698820501568]... | [
-0.3758842647075653,
-0.5878843069076538,
0.032787516713142395,
0.44652554392814636,
0.30833250284194946,
-0.06708855926990509,
0.3846704661846161,
-0.0230244230479002,
0.11546281725168228,
-0.130690798163414,
0.03220471739768982,
0.1839643120765686,
0.011365003883838654,
-0.16018880903720... |
https://github.com/huggingface/datasets/issues/6364 | You can use the following code to load this CSV with the list values preserved:
```python
from datasets import load_dataset
import ast
converters = {
"contexts" : ast.literal_eval,
"ground_truths" : ast.literal_eval,
}
ds = load_dataset("csv", data_files="golden_dataset.csv", converters=converters)
... | ArrowNotImplementedError: Unsupported cast from string to list using function cast_list | Hi,
I am trying to load a local csv dataset(similar to explodinggradients_fiqa) using load_dataset. When I try to pass features, I am facing the mentioned issue.
CSV Data sample(golden_dataset.csv):
Question | Context | answer | groundtruth
"what is abc?"... | 38 | ArrowNotImplementedError: Unsupported cast from string to list using function cast_list
Hi,
I am trying to load a local csv dataset(similar to explodinggradients_fiqa) using load_dataset. When I try to pass features, I am facing the mentioned issue.
CSV Data sample(golden_dataset.csv):
Question | ... | [
-0.02789292484521866,
-0.23731307685375214,
-0.11266941577196121,
0.09196636080741882,
0.672694742679596,
-0.07634066045284271,
0.1983177810907364,
0.24670083820819855,
0.28991514444351196,
0.11488257348537445,
0.06941263377666473,
0.47068384289741516,
-0.139689639210701,
0.041423432528972... |
https://github.com/huggingface/datasets/issues/6363 | I think the code hangs on the `accelerator.main_process_first()` context manager exit. To verify this, you can append a print statement to the end of the `accelerator.main_process_first()` block.
If the problem is in `with_transform`, it would help if you could share the error stack trace printed when you interr... | dataset.transform() hangs indefinitely while finetuning the stable diffusion XL | ### Describe the bug
Multi-GPU fine-tuning the stable diffusion X by following https://github.com/huggingface/diffusers/blob/main/examples/text_to_image/README_sdxl.md hangs indefinitely.
### Steps to reproduce the bug
accelerate launch train_text_to_image_sdxl.py --pretrained_model_name_or_path=$MODEL_NAME --... | 53 | dataset.transform() hangs indefinitely while finetuning the stable diffusion XL
### Describe the bug
Multi-GPU fine-tuning the stable diffusion X by following https://github.com/huggingface/diffusers/blob/main/examples/text_to_image/README_sdxl.md hangs indefinitely.
### Steps to reproduce the bug
accelerate... | [
-0.29093098640441895,
-0.4494297206401825,
0.03063829243183136,
0.10455924272537231,
0.7972444295883179,
-0.06378263235092163,
0.38594162464141846,
0.08792893588542938,
-0.19314613938331604,
0.4078892171382904,
0.08595020323991776,
0.15589837729930878,
0.01676591858267784,
-0.0726515948772... |
https://github.com/huggingface/datasets/issues/6363 | @mariosasko yes the problem seems to be to exit from accelerator.main_process_first(). Is there any known problem? | dataset.transform() hangs indefinitely while finetuning the stable diffusion XL | ### Describe the bug
Multi-GPU fine-tuning the stable diffusion X by following https://github.com/huggingface/diffusers/blob/main/examples/text_to_image/README_sdxl.md hangs indefinitely.
### Steps to reproduce the bug
accelerate launch train_text_to_image_sdxl.py --pretrained_model_name_or_path=$MODEL_NAME --... | 16 | dataset.transform() hangs indefinitely while finetuning the stable diffusion XL
### Describe the bug
Multi-GPU fine-tuning the stable diffusion X by following https://github.com/huggingface/diffusers/blob/main/examples/text_to_image/README_sdxl.md hangs indefinitely.
### Steps to reproduce the bug
accelerate... | [
-0.29093098640441895,
-0.4494297206401825,
0.03063829243183136,
0.10455924272537231,
0.7972444295883179,
-0.06378263235092163,
0.38594162464141846,
0.08792893588542938,
-0.19314613938331604,
0.4078892171382904,
0.08595020323991776,
0.15589837729930878,
0.01676591858267784,
-0.0726515948772... |
https://github.com/huggingface/datasets/issues/6363 | NCCL debug info I get below output, if it helps.
```
11/09/2023 13:36:44 - INFO - __main__ - Distributed environment: MULTI_GPU Backend: nccl
Num processes: 2
Process index: 1
Local process index: 1
Device: cuda:1
Mixed precision type: fp16
Detected kernel version 5.4.0, which is below the recommended mini... | dataset.transform() hangs indefinitely while finetuning the stable diffusion XL | ### Describe the bug
Multi-GPU fine-tuning the stable diffusion X by following https://github.com/huggingface/diffusers/blob/main/examples/text_to_image/README_sdxl.md hangs indefinitely.
### Steps to reproduce the bug
accelerate launch train_text_to_image_sdxl.py --pretrained_model_name_or_path=$MODEL_NAME --... | 620 | dataset.transform() hangs indefinitely while finetuning the stable diffusion XL
### Describe the bug
Multi-GPU fine-tuning the stable diffusion X by following https://github.com/huggingface/diffusers/blob/main/examples/text_to_image/README_sdxl.md hangs indefinitely.
### Steps to reproduce the bug
accelerate... | [
-0.29093098640441895,
-0.4494297206401825,
0.03063829243183136,
0.10455924272537231,
0.7972444295883179,
-0.06378263235092163,
0.38594162464141846,
0.08792893588542938,
-0.19314613938331604,
0.4078892171382904,
0.08595020323991776,
0.15589837729930878,
0.01676591858267784,
-0.0726515948772... |
https://github.com/huggingface/datasets/issues/6363 | I don't know what the issue was, but after going through the thread here I loved the issue with https://github.com/huggingface/accelerate/issues/314#issuecomment-1565259831 | dataset.transform() hangs indefinitely while finetuning the stable diffusion XL | ### Describe the bug
Multi-GPU fine-tuning the stable diffusion X by following https://github.com/huggingface/diffusers/blob/main/examples/text_to_image/README_sdxl.md hangs indefinitely.
### Steps to reproduce the bug
accelerate launch train_text_to_image_sdxl.py --pretrained_model_name_or_path=$MODEL_NAME --... | 20 | dataset.transform() hangs indefinitely while finetuning the stable diffusion XL
### Describe the bug
Multi-GPU fine-tuning the stable diffusion X by following https://github.com/huggingface/diffusers/blob/main/examples/text_to_image/README_sdxl.md hangs indefinitely.
### Steps to reproduce the bug
accelerate... | [
-0.29093098640441895,
-0.4494297206401825,
0.03063829243183136,
0.10455924272537231,
0.7972444295883179,
-0.06378263235092163,
0.38594162464141846,
0.08792893588542938,
-0.19314613938331604,
0.4078892171382904,
0.08595020323991776,
0.15589837729930878,
0.01676591858267784,
-0.0726515948772... |
https://github.com/huggingface/datasets/issues/6360 | This issue stems from https://github.com/huggingface/datasets/blob/6d2f2a5e0fea3827eccfd1717d8021c15fc4292a/src/datasets/table.py#L2203-L2205
I'll address it as part of https://github.com/huggingface/datasets/pull/6283.
In the meantime, this should work
```python
import pyarrow as pa
from datasets import Ima... | Add support for `Sequence(Audio/Image)` feature in `push_to_hub` | ### Feature request
Allow for `Sequence` of `Image` (or `Audio`) to be embedded inside the shards.
### Motivation
Currently, thanks to #3685, when `embed_external_files` is set to True (which is the default) in `push_to_hub`, features of type `Image` and `Audio` are embedded inside the arrow/parquet shards, instead ... | 58 | Add support for `Sequence(Audio/Image)` feature in `push_to_hub`
### Feature request
Allow for `Sequence` of `Image` (or `Audio`) to be embedded inside the shards.
### Motivation
Currently, thanks to #3685, when `embed_external_files` is set to True (which is the default) in `push_to_hub`, features of type `Image... | [
-0.255060076713562,
-0.5069053173065186,
-0.11654671281576157,
0.14346259832382202,
-0.07700437307357788,
-0.09264026582241058,
0.1639629453420639,
0.2216341495513916,
-0.26074954867362976,
0.17967505753040314,
0.046998776495456696,
0.26361843943595886,
-0.4362025260925293,
0.1619111001491... |
https://github.com/huggingface/datasets/issues/6359 | Most likely, the data file inference logic is the problem here.
You can run the following code to verify this:
```python
import time
from datasets.data_files import get_data_patterns
start_time = time.time()
get_data_patterns("/path/to/img_dir")
end_time = time.time()
print(f"Elapsed time: {end_time - start_t... | Stuck in "Resolving data files..." | ### Describe the bug
I have an image dataset with 300k images, the size of image is 768 * 768.
When I run `dataset = load_dataset("imagefolder", data_dir="/path/to/img_dir", split='train')` in second time, it takes 50 minutes to finish "Resolving data files" part, what's going on in this part?
From my understa... | 74 | Stuck in "Resolving data files..."
### Describe the bug
I have an image dataset with 300k images, the size of image is 768 * 768.
When I run `dataset = load_dataset("imagefolder", data_dir="/path/to/img_dir", split='train')` in second time, it takes 50 minutes to finish "Resolving data files" part, what's going... | [
-0.23090443015098572,
0.04058697819709778,
-0.10604532063007355,
0.36753690242767334,
0.2609904706478119,
0.13373884558677673,
0.34540820121765137,
0.47421953082084656,
-0.10366508364677429,
0.11970315873622894,
0.16956403851509094,
0.3559846878051758,
0.1605689525604248,
-0.18263012170791... |
https://github.com/huggingface/datasets/issues/6359 | Hi, @mariosasko, you are right; data file inference logic is extremely slow.
I have done a similar test, that is I modify the source code of datasets/load.py to measure the cost of two suspicious operations:
```python
def get_module(self) -> DatasetModule:
base_path = Path(self.data_dir or "").expanduser(... | Stuck in "Resolving data files..." | ### Describe the bug
I have an image dataset with 300k images, the size of image is 768 * 768.
When I run `dataset = load_dataset("imagefolder", data_dir="/path/to/img_dir", split='train')` in second time, it takes 50 minutes to finish "Resolving data files" part, what's going on in this part?
From my understa... | 99 | Stuck in "Resolving data files..."
### Describe the bug
I have an image dataset with 300k images, the size of image is 768 * 768.
When I run `dataset = load_dataset("imagefolder", data_dir="/path/to/img_dir", split='train')` in second time, it takes 50 minutes to finish "Resolving data files" part, what's going... | [
-0.2094925194978714,
0.1629522740840912,
-0.089963898062706,
0.4162213206291199,
0.14655184745788574,
0.10716225951910019,
0.2901347577571869,
0.44508206844329834,
-0.08686953037977219,
0.10743442177772522,
0.11540091037750244,
0.3833844065666199,
0.20739655196666718,
-0.2878902554512024,
... |
https://github.com/huggingface/datasets/issues/6359 | Furthermore, what's my current workaround about this problem? Should I save it by `save_to_disk()` and load dataset through `load_from_disk`? | Stuck in "Resolving data files..." | ### Describe the bug
I have an image dataset with 300k images, the size of image is 768 * 768.
When I run `dataset = load_dataset("imagefolder", data_dir="/path/to/img_dir", split='train')` in second time, it takes 50 minutes to finish "Resolving data files" part, what's going on in this part?
From my understa... | 19 | Stuck in "Resolving data files..."
### Describe the bug
I have an image dataset with 300k images, the size of image is 768 * 768.
When I run `dataset = load_dataset("imagefolder", data_dir="/path/to/img_dir", split='train')` in second time, it takes 50 minutes to finish "Resolving data files" part, what's going... | [
-0.2394724190235138,
-0.006858140230178833,
-0.12425383925437927,
0.3820543885231018,
0.29293784499168396,
0.18949246406555176,
0.26252657175064087,
0.40455177426338196,
-0.054701972752809525,
0.03535822778940201,
0.22675317525863647,
0.3705781102180481,
0.1320250928401947,
-0.192263275384... |
https://github.com/huggingface/datasets/issues/6358 | You may be able to make it work by tweaking some environment variables, such as [`HF_HOME`](https://huggingface.co/docs/huggingface_hub/main/en/package_reference/environment_variables#hfhome) or [`HF_DATASETS_CACHE`](https://huggingface.co/docs/datasets/cache#cache-directory). | Mounting datasets cache fails due to absolute paths. | ### Describe the bug
Creating a datasets cache and mounting this into, for example, a docker container, renders the data unreadable due to absolute paths written into the cache.
### Steps to reproduce the bug
1. Create a datasets cache by downloading some data
2. Mount the dataset folder into a docker contain... | 18 | Mounting datasets cache fails due to absolute paths.
### Describe the bug
Creating a datasets cache and mounting this into, for example, a docker container, renders the data unreadable due to absolute paths written into the cache.
### Steps to reproduce the bug
1. Create a datasets cache by downloading some ... | [
-0.00010085105895996094,
0.028453439474105835,
0.08273055404424667,
0.13505060970783234,
0.24859587848186493,
0.023465223610401154,
0.1194758415222168,
0.21675515174865723,
0.049085840582847595,
0.09170372039079666,
-0.07647693157196045,
0.08717919886112213,
-0.2398337721824646,
-0.3364964... |
https://github.com/huggingface/datasets/issues/6358 | > You may be able to make it work by tweaking some environment variables, such as [`HF_HOME`](https://huggingface.co/docs/huggingface_hub/main/en/package_reference/environment_variables#hfhome) or [`HF_DATASETS_CACHE`](https://huggingface.co/docs/datasets/cache#cache-directory).
I am already doing this. The problem ... | Mounting datasets cache fails due to absolute paths. | ### Describe the bug
Creating a datasets cache and mounting this into, for example, a docker container, renders the data unreadable due to absolute paths written into the cache.
### Steps to reproduce the bug
1. Create a datasets cache by downloading some data
2. Mount the dataset folder into a docker contain... | 78 | Mounting datasets cache fails due to absolute paths.
### Describe the bug
Creating a datasets cache and mounting this into, for example, a docker container, renders the data unreadable due to absolute paths written into the cache.
### Steps to reproduce the bug
1. Create a datasets cache by downloading some ... | [
0.053802065551280975,
0.05613492429256439,
0.08096794784069061,
0.06797824800014496,
0.2428542971611023,
-0.015263326466083527,
0.10982009768486023,
0.20520839095115662,
0.03088117390871048,
0.05655992776155472,
-0.1255326271057129,
0.11960466206073761,
-0.24449720978736877,
-0.37721857428... |
https://github.com/huggingface/datasets/issues/6358 | I'm unable to reproduce this with the cache
```bash
export HF_CACHE=$PWD/hf_cache
python -c "import datasets; datasets.load_dataset('imdb')"
```
imported inside a dummy container that is built from
```bash
FROM python:3.9
WORKDIR /usr/src/app
RUN pip install datasets
COPY ./hf_cache ./hf_cache
ENV HF... | Mounting datasets cache fails due to absolute paths. | ### Describe the bug
Creating a datasets cache and mounting this into, for example, a docker container, renders the data unreadable due to absolute paths written into the cache.
### Steps to reproduce the bug
1. Create a datasets cache by downloading some data
2. Mount the dataset folder into a docker contain... | 73 | Mounting datasets cache fails due to absolute paths.
### Describe the bug
Creating a datasets cache and mounting this into, for example, a docker container, renders the data unreadable due to absolute paths written into the cache.
### Steps to reproduce the bug
1. Create a datasets cache by downloading some ... | [
0.013612456619739532,
0.28743523359298706,
0.03356702998280525,
0.12142452597618103,
0.17181266844272614,
0.09251341223716736,
0.315844863653183,
0.2463112622499466,
0.07502992451190948,
0.018458478152751923,
0.02460268884897232,
0.11198748648166656,
-0.14733318984508514,
-0.64198601245880... |
https://github.com/huggingface/datasets/issues/6354 | I am having issues as well with this.
However, the error I am getting is :
`RuntimeError: It appears that you are attempting to reference SparkContext from a broadcast variable, action, or transformation. SparkContext can only be used on the driver, not in code that it run on workers. For more information, see SPA... | `IterableDataset.from_spark` does not support multiple workers in pytorch `Dataloader` | ### Describe the bug
Looks like `IterableDataset.from_spark` does not support multiple workers in pytorch `Dataloader` if I'm not missing anything.
Also, returns not consistent error messages, which probably depend on the nondeterministic order of worker executions
Some exampes I've encountered:
```
File "/l... | 62 | `IterableDataset.from_spark` does not support multiple workers in pytorch `Dataloader`
### Describe the bug
Looks like `IterableDataset.from_spark` does not support multiple workers in pytorch `Dataloader` if I'm not missing anything.
Also, returns not consistent error messages, which probably depend on the nonde... | [
-0.28735220432281494,
-0.16247272491455078,
-0.06743581593036652,
0.5066515207290649,
-0.03184495121240616,
-0.3001953959465027,
0.7239212393760681,
-0.012372683733701706,
-0.012353334575891495,
0.030812911689281464,
0.008178908377885818,
0.445117712020874,
-0.003525044769048691,
0.0159480... |
https://github.com/huggingface/datasets/issues/6353 | I'm using the latest datasets and fsspec , but still got this error!
datasets : Version: 2.13.0
fsspec Version: 2023.10.0
```
File "/home/guoby/app/Anaconda3-2021.05/envs/news/lib/python3.8/site-packages/datasets/load.py", line 1892, in load_from_disk
return DatasetDict.load_from_disk(dataset_path, ke... | load_dataset save_to_disk load_from_disk error | ### Describe the bug
datasets version: 2.10.1
I `load_dataset `and `save_to_disk` sucessfully on windows10( **and I `load_from_disk(/LLM/data/wiki)` succcesfully on windows10**), and I copy the dataset `/LLM/data/wiki`
into a ubuntu system, but when I `load_from_disk(/LLM/data/wiki)` on ubuntu, something weird ha... | 83 | load_dataset save_to_disk load_from_disk error
### Describe the bug
datasets version: 2.10.1
I `load_dataset `and `save_to_disk` sucessfully on windows10( **and I `load_from_disk(/LLM/data/wiki)` succcesfully on windows10**), and I copy the dataset `/LLM/data/wiki`
into a ubuntu system, but when I `load_from_dis... | [
-0.3020021915435791,
0.007541164755821228,
0.07424850016832352,
0.35470154881477356,
0.38531994819641113,
0.19839975237846375,
0.30810630321502686,
0.09688099473714828,
0.038030557334423065,
0.08450580388307571,
-0.02063780650496483,
0.3757515847682953,
0.020782995969057083,
0.058200322091... |
https://github.com/huggingface/datasets/issues/6352 | +1
```
Found cached dataset csv (file:///home/ubuntu/.cache/huggingface/datasets/theSquarePond___csv/theSquarePond--XXXXX-bbf0a8365d693d2c/0.0.0/eea64c71ca8b46dd3f537ed218fc9bf495d5707789152eb2764f5c78fa66d59d)
---------------------------------------------------------------------------
NotImplementedError ... | Error loading wikitext data raise NotImplementedError(f"Loading a dataset cached in a {type(self._fs).__name__} is not supported.") | I was trying to load the wiki dataset, but i got this error
traindata = load_dataset('wikitext', 'wikitext-2-raw-v1', split='train')
File "/home/aelkordy/.conda/envs/prune_llm/lib/python3.9/site-packages/datasets/load.py", line 1804, in load_dataset
ds = builder_instance.as_dataset(split=split, verific... | 183 | Error loading wikitext data raise NotImplementedError(f"Loading a dataset cached in a {type(self._fs).__name__} is not supported.")
I was trying to load the wiki dataset, but i got this error
traindata = load_dataset('wikitext', 'wikitext-2-raw-v1', split='train')
File "/home/aelkordy/.conda/envs/prune... | [
-0.26059049367904663,
-0.20326289534568787,
-0.04139687120914459,
0.2912045121192932,
0.2911190390586853,
0.3084626793861389,
0.23816388845443726,
0.41730719804763794,
0.4236239492893219,
-0.02376360446214676,
0.10076349973678589,
0.09992368519306183,
-0.06720319390296936,
-0.0004816353321... |
https://github.com/huggingface/datasets/issues/6352 | +1
```
Found cached dataset csv ([file://C:/Users/Shady/.cache/huggingface/datasets/knkarthick___csv/knkarthick--dialogsum-cd36827d3490488d/0.0.0/6954658bab30a358235fa864b05cf819af0e179325c740e4bc853bcc7ec513e1](file:///C:/Users/Shady/.cache/huggingface/datasets/knkarthick___csv/knkarthick--dialogsum-cd36827d349048... | Error loading wikitext data raise NotImplementedError(f"Loading a dataset cached in a {type(self._fs).__name__} is not supported.") | I was trying to load the wiki dataset, but i got this error
traindata = load_dataset('wikitext', 'wikitext-2-raw-v1', split='train')
File "/home/aelkordy/.conda/envs/prune_llm/lib/python3.9/site-packages/datasets/load.py", line 1804, in load_dataset
ds = builder_instance.as_dataset(split=split, verific... | 175 | Error loading wikitext data raise NotImplementedError(f"Loading a dataset cached in a {type(self._fs).__name__} is not supported.")
I was trying to load the wiki dataset, but i got this error
traindata = load_dataset('wikitext', 'wikitext-2-raw-v1', split='train')
File "/home/aelkordy/.conda/envs/prune... | [
-0.22216784954071045,
-0.21753501892089844,
-0.05269908159971237,
0.3149651288986206,
0.2889442443847656,
0.3062695264816284,
0.2462485432624817,
0.4402695596218109,
0.4070063829421997,
-0.022959955036640167,
0.08359038084745407,
0.12875205278396606,
-0.07824663817882538,
0.017831921577453... |
https://github.com/huggingface/datasets/issues/6352 | This error stems from a breaking change in `fsspec`. It has been fixed in the latest `datasets` release (`2.14.6`). Updating the installation with `pip install -U datasets` should fix the issue.
| Error loading wikitext data raise NotImplementedError(f"Loading a dataset cached in a {type(self._fs).__name__} is not supported.") | I was trying to load the wiki dataset, but i got this error
traindata = load_dataset('wikitext', 'wikitext-2-raw-v1', split='train')
File "/home/aelkordy/.conda/envs/prune_llm/lib/python3.9/site-packages/datasets/load.py", line 1804, in load_dataset
ds = builder_instance.as_dataset(split=split, verific... | 31 | Error loading wikitext data raise NotImplementedError(f"Loading a dataset cached in a {type(self._fs).__name__} is not supported.")
I was trying to load the wiki dataset, but i got this error
traindata = load_dataset('wikitext', 'wikitext-2-raw-v1', split='train')
File "/home/aelkordy/.conda/envs/prune... | [
-0.36869052052497864,
-0.10481400042772293,
0.001288868486881256,
0.1946215033531189,
0.27150219678878784,
0.0910446047782898,
0.062023088335990906,
0.3505300283432007,
0.38558250665664673,
0.03684042766690254,
0.15493622422218323,
0.20195260643959045,
-0.050123684108257294,
0.288501232862... |
https://github.com/huggingface/datasets/issues/6350 | `load_dataset` returns a `DatasetDict` object unless `split` is defined, in which case it returns a `Dataset` (or a list of datasets if `split` is a list). We've discussed dropping `DatasetDict` from the API in https://github.com/huggingface/datasets/issues/5189 to always return the same type in `load_dataset` and sup... | Different objects are returned from calls that should be returning the same kind of object. | ### Describe the bug
1. dataset = load_dataset("togethercomputer/RedPajama-Data-1T-Sample", cache_dir=training_args.cache_dir, split='train[:1%]')
2. dataset = load_dataset("togethercomputer/RedPajama-Data-1T-Sample", cache_dir=training_args.cache_dir)
The only difference I would expect these cal... | 82 | Different objects are returned from calls that should be returning the same kind of object.
### Describe the bug
1. dataset = load_dataset("togethercomputer/RedPajama-Data-1T-Sample", cache_dir=training_args.cache_dir, split='train[:1%]')
2. dataset = load_dataset("togethercomputer/RedPajama-Data-... | [
0.065458744764328,
-0.35832369327545166,
-0.047362830489873886,
0.541680097579956,
0.15500470995903015,
-0.27208226919174194,
0.2597043514251709,
0.172197625041008,
0.10662220418453217,
0.15586790442466736,
-0.15613983571529388,
0.31378409266471863,
-0.04439716413617134,
0.3652834892272949... |
https://github.com/huggingface/datasets/issues/6350 | > `load_dataset` returns a `DatasetDict` object unless `split` is defined, in which case it returns a `Dataset` (or a list of datasets if `split` is a list). We've discussed dropping `DatasetDict` from the API in #5189 to always return the same type in `load_dataset` and support datasets without (explicit) splits. IIRC... | Different objects are returned from calls that should be returning the same kind of object. | ### Describe the bug
1. dataset = load_dataset("togethercomputer/RedPajama-Data-1T-Sample", cache_dir=training_args.cache_dir, split='train[:1%]')
2. dataset = load_dataset("togethercomputer/RedPajama-Data-1T-Sample", cache_dir=training_args.cache_dir)
The only difference I would expect these cal... | 153 | Different objects are returned from calls that should be returning the same kind of object.
### Describe the bug
1. dataset = load_dataset("togethercomputer/RedPajama-Data-1T-Sample", cache_dir=training_args.cache_dir, split='train[:1%]')
2. dataset = load_dataset("togethercomputer/RedPajama-Data-... | [
0.0666290819644928,
-0.18669439852237701,
-0.08955486118793488,
0.5604851841926575,
0.058909911662340164,
-0.2986440658569336,
0.28418415784835815,
0.21379944682121277,
0.12141849100589752,
0.16319358348846436,
-0.017049366608262062,
0.4322671890258789,
-0.04269466549158096,
0.269218504428... |
https://github.com/huggingface/datasets/issues/6349 | I'm unable to reproduce this error. The server hosting the files may have been down temporarily, so try again. | Can't load ds = load_dataset("imdb") | ### Describe the bug
I did `from datasets import load_dataset, load_metric` and then `ds = load_dataset("imdb")` and it gave me the error:
ExpectedMoreDownloadedFiles: {'http://ai.stanford.edu/~amaas/data/sentiment/aclImdb_v1.tar.gz'}
I tried doing `ds = load_dataset("imdb",download_mode="force_redownload")` as we... | 19 | Can't load ds = load_dataset("imdb")
### Describe the bug
I did `from datasets import load_dataset, load_metric` and then `ds = load_dataset("imdb")` and it gave me the error:
ExpectedMoreDownloadedFiles: {'http://ai.stanford.edu/~amaas/data/sentiment/aclImdb_v1.tar.gz'}
I tried doing `ds = load_dataset("imdb",d... | [
-0.3720935881137848,
0.034979745745658875,
-0.05936967581510544,
0.5213468670845032,
0.41254252195358276,
0.1490505039691925,
0.34063592553138733,
0.1281067281961441,
0.6659700274467468,
0.07853162288665771,
-0.1272823065519333,
0.14698487520217896,
-0.03618531674146652,
0.0683325678110122... |
https://github.com/huggingface/datasets/issues/6347 | This was fixed in https://github.com/huggingface/datasets/pull/6247. You can find the fix in the `main` version of the docs | Incorrect example code in 'Create a dataset' docs | ### Describe the bug
On [this](https://huggingface.co/docs/datasets/create_dataset) page, the example code for loading in images and audio is incorrect.
Currently, examples are:
``` python
from datasets import ImageFolder
dataset = load_dataset("imagefolder", data_dir="/path/to/pokemon")
```
and
``` python... | 17 | Incorrect example code in 'Create a dataset' docs
### Describe the bug
On [this](https://huggingface.co/docs/datasets/create_dataset) page, the example code for loading in images and audio is incorrect.
Currently, examples are:
``` python
from datasets import ImageFolder
dataset = load_dataset("imagefolder",... | [
0.10520012676715851,
-0.3185540735721588,
-0.02481408417224884,
0.494010865688324,
0.09755562245845795,
-0.02400396764278412,
0.3480840027332306,
-0.08572138100862503,
0.09643727540969849,
0.09243978559970856,
-0.1433936357498169,
0.17106789350509644,
-0.07363775372505188,
0.32302072644233... |
https://github.com/huggingface/datasets/issues/6330 | I also encountered a similar error below.
Appreciate the team could shed some light on this issue.
```
---------------------------------------------------------------------------
NotImplementedError Traceback (most recent call last)
[/home/ubuntu/work/EveryDream2trainer/prepare_dataset.ipyn... | Latest fsspec==2023.10.0 issue with streaming datasets | ### Describe the bug
Loading a streaming dataset with this version of fsspec fails with the following error:
`NotImplementedError: Loading a streaming dataset cached in a LocalFileSystem is not supported yet.`
I suspect the issue is with this PR
https://github.com/fsspec/filesystem_spec/pull/1381
### Steps ... | 588 | Latest fsspec==2023.10.0 issue with streaming datasets
### Describe the bug
Loading a streaming dataset with this version of fsspec fails with the following error:
`NotImplementedError: Loading a streaming dataset cached in a LocalFileSystem is not supported yet.`
I suspect the issue is with this PR
https:... | [
-0.3962833285331726,
-0.05229148268699646,
0.0895015299320221,
0.22059676051139832,
0.20769593119621277,
-0.055554941296577454,
0.11532916873693466,
0.1682538241147995,
-0.1815401017665863,
0.2266596555709839,
-0.04434163123369217,
0.2020154893398285,
-0.05888967961072922,
0.46524885296821... |
https://github.com/huggingface/datasets/issues/6330 | Thanks for reporting and for the investigation, @ZachNagengast! :hugs:
We are investigating the root cause of the issue. In the meantime, we are going to pin fsspec < 2023.10.0. | Latest fsspec==2023.10.0 issue with streaming datasets | ### Describe the bug
Loading a streaming dataset with this version of fsspec fails with the following error:
`NotImplementedError: Loading a streaming dataset cached in a LocalFileSystem is not supported yet.`
I suspect the issue is with this PR
https://github.com/fsspec/filesystem_spec/pull/1381
### Steps ... | 29 | Latest fsspec==2023.10.0 issue with streaming datasets
### Describe the bug
Loading a streaming dataset with this version of fsspec fails with the following error:
`NotImplementedError: Loading a streaming dataset cached in a LocalFileSystem is not supported yet.`
I suspect the issue is with this PR
https:... | [
-0.44810670614242554,
-0.09368070214986801,
0.07924748212099075,
0.2137065976858139,
0.21421100199222565,
-0.08358780294656754,
0.07893026620149612,
0.21099360287189484,
-0.11809461563825607,
0.2367476224899292,
-0.023853646591305733,
0.15875737369060516,
-0.07089647650718689,
0.4641267061... |
https://github.com/huggingface/datasets/issues/6330 | You can also update `datasets`:
```
pip install -U datasets
```
It will also update `fsspec` to use the right version | Latest fsspec==2023.10.0 issue with streaming datasets | ### Describe the bug
Loading a streaming dataset with this version of fsspec fails with the following error:
`NotImplementedError: Loading a streaming dataset cached in a LocalFileSystem is not supported yet.`
I suspect the issue is with this PR
https://github.com/fsspec/filesystem_spec/pull/1381
### Steps ... | 21 | Latest fsspec==2023.10.0 issue with streaming datasets
### Describe the bug
Loading a streaming dataset with this version of fsspec fails with the following error:
`NotImplementedError: Loading a streaming dataset cached in a LocalFileSystem is not supported yet.`
I suspect the issue is with this PR
https:... | [
-0.48434990644454956,
-0.09176057577133179,
0.06091972813010216,
0.20084726810455322,
0.20000611245632172,
-0.06782572716474533,
0.050473786890506744,
0.2100011259317398,
-0.0857999175786972,
0.21352194249629974,
-0.019064685329794884,
0.21504412591457367,
-0.056027695536613464,
0.46829530... |
https://github.com/huggingface/datasets/issues/6327 | You can clone the `togethercomputer/RedPajama-Data-1T-Sample` repo and load the dataset with `load_dataset("path/to/cloned_repo")` to use it offline. | FileNotFoundError when trying to load the downloaded dataset with `load_dataset(..., streaming=True)` | ### Describe the bug
Hi, I'm trying to load the dataset `togethercomputer/RedPajama-Data-1T-Sample` with `load_dataset` in streaming mode, i.e., `streaming=True`, but `FileNotFoundError` occurs.
### Steps to reproduce the bug
I've downloaded the dataset and save it to the cache dir in advance. My hope is loadi... | 16 | FileNotFoundError when trying to load the downloaded dataset with `load_dataset(..., streaming=True)`
### Describe the bug
Hi, I'm trying to load the dataset `togethercomputer/RedPajama-Data-1T-Sample` with `load_dataset` in streaming mode, i.e., `streaming=True`, but `FileNotFoundError` occurs.
### Steps to re... | [
-0.33369070291519165,
-0.10643325001001358,
0.10457661002874374,
0.33565500378608704,
0.1996665745973587,
0.007022194564342499,
0.20944146811962128,
0.36828163266181946,
0.0730251893401146,
0.0012076348066329956,
0.15766292810440063,
0.28006261587142944,
-0.20167627930641174,
-0.1230023205... |
https://github.com/huggingface/datasets/issues/6327 | @mariosasko Thank you for your kind reply! I'll try it as a workaround.
Does that mean that currently it's not supported to simply load with a short name? | FileNotFoundError when trying to load the downloaded dataset with `load_dataset(..., streaming=True)` | ### Describe the bug
Hi, I'm trying to load the dataset `togethercomputer/RedPajama-Data-1T-Sample` with `load_dataset` in streaming mode, i.e., `streaming=True`, but `FileNotFoundError` occurs.
### Steps to reproduce the bug
I've downloaded the dataset and save it to the cache dir in advance. My hope is loadi... | 28 | FileNotFoundError when trying to load the downloaded dataset with `load_dataset(..., streaming=True)`
### Describe the bug
Hi, I'm trying to load the dataset `togethercomputer/RedPajama-Data-1T-Sample` with `load_dataset` in streaming mode, i.e., `streaming=True`, but `FileNotFoundError` occurs.
### Steps to re... | [
-0.33369070291519165,
-0.10643325001001358,
0.10457661002874374,
0.33565500378608704,
0.1996665745973587,
0.007022194564342499,
0.20944146811962128,
0.36828163266181946,
0.0730251893401146,
0.0012076348066329956,
0.15766292810440063,
0.28006261587142944,
-0.20167627930641174,
-0.1230023205... |
https://github.com/huggingface/datasets/issues/6327 | It is, but manually downloading repo files to the cache can easily lead to failure (the HF cache is not meant to be modified by a user besides deleting the files 🙂), as in your case. Hence, the clone + `load_dataset("path/to/cloned_repo")` workflow should be used instead. | FileNotFoundError when trying to load the downloaded dataset with `load_dataset(..., streaming=True)` | ### Describe the bug
Hi, I'm trying to load the dataset `togethercomputer/RedPajama-Data-1T-Sample` with `load_dataset` in streaming mode, i.e., `streaming=True`, but `FileNotFoundError` occurs.
### Steps to reproduce the bug
I've downloaded the dataset and save it to the cache dir in advance. My hope is loadi... | 46 | FileNotFoundError when trying to load the downloaded dataset with `load_dataset(..., streaming=True)`
### Describe the bug
Hi, I'm trying to load the dataset `togethercomputer/RedPajama-Data-1T-Sample` with `load_dataset` in streaming mode, i.e., `streaming=True`, but `FileNotFoundError` occurs.
### Steps to re... | [
-0.33369070291519165,
-0.10643325001001358,
0.10457661002874374,
0.33565500378608704,
0.1996665745973587,
0.007022194564342499,
0.20944146811962128,
0.36828163266181946,
0.0730251893401146,
0.0012076348066329956,
0.15766292810440063,
0.28006261587142944,
-0.20167627930641174,
-0.1230023205... |
https://github.com/huggingface/datasets/issues/6324 | Unlike Pandas, Arrow is strict with types, so converting the problematic strings to ints (or ints to strings) to ensure all the values have the same type is the only fix.
JSON support has been requested in Arrow [here](https://github.com/apache/arrow/issues/32538), but I don't expect this to be implemented soon.
... | Conversion to Arrow fails due to wrong type heuristic | ### Describe the bug
I have a list of dictionaries with valid/JSON-serializable values.
One key is the denominator for a paragraph. In 99.9% of cases its a number, but there are some occurences of '1a', '2b' and so on.
If trying to convert this list to a dataset with `Dataset.from_list()`, I always get
`ArrowI... | 97 | Conversion to Arrow fails due to wrong type heuristic
### Describe the bug
I have a list of dictionaries with valid/JSON-serializable values.
One key is the denominator for a paragraph. In 99.9% of cases its a number, but there are some occurences of '1a', '2b' and so on.
If trying to convert this list to a d... | [
-0.13682855665683746,
-0.08490268886089325,
0.047282736748456955,
0.18112041056156158,
0.4715591371059418,
-0.02025952935218811,
0.2181384414434433,
0.3003539741039276,
-0.2322429120540619,
-0.14563122391700745,
0.018224947154521942,
0.5438294410705566,
-0.030470455065369606,
-0.0194185748... |
https://github.com/huggingface/datasets/issues/6324 | > Unlike Pandas, Arrow is strict with types, so converting the problematic strings to ints (or ints to strings) to ensure all the values have the same type is the only fix.
>
> JSON support has been requested in Arrow [here](https://github.com/apache/arrow/issues/32538), but I don't expect this to be implemented soo... | Conversion to Arrow fails due to wrong type heuristic | ### Describe the bug
I have a list of dictionaries with valid/JSON-serializable values.
One key is the denominator for a paragraph. In 99.9% of cases its a number, but there are some occurences of '1a', '2b' and so on.
If trying to convert this list to a dataset with `Dataset.from_list()`, I always get
`ArrowI... | 121 | Conversion to Arrow fails due to wrong type heuristic
### Describe the bug
I have a list of dictionaries with valid/JSON-serializable values.
One key is the denominator for a paragraph. In 99.9% of cases its a number, but there are some occurences of '1a', '2b' and so on.
If trying to convert this list to a d... | [
-0.1248508095741272,
-0.08872330188751221,
0.048213206231594086,
0.19750455021858215,
0.49781298637390137,
-0.03315145522356033,
0.23462632298469543,
0.31393522024154663,
-0.23804914951324463,
-0.1408153772354126,
0.0366579070687294,
0.5494917035102844,
-0.01877778209745884,
-0.01878361031... |
https://github.com/huggingface/datasets/issues/6320 | The expression "train+test" concatenates the splits.
The individual splits as separate datasets can be obtained as follows:
```python
train_ds, test_ds = load_dataset("<dataset_name>", split=["train", "test"])
train_10pct_ds, test_10pct_ds = load_dataset("<dataset_name>", split=["train[:10%]", "test[:%10]"])
``` | Dataset slice splits can't load training and validation at the same time | ### Describe the bug
According to the [documentation](https://huggingface.co/docs/datasets/v2.14.5/loading#slice-splits) is should be possible to run the following command:
`train_test_ds = datasets.load_dataset("bookcorpus", split="train+test")`
to load the train and test sets from the dataset.
However ex... | 31 | Dataset slice splits can't load training and validation at the same time
### Describe the bug
According to the [documentation](https://huggingface.co/docs/datasets/v2.14.5/loading#slice-splits) is should be possible to run the following command:
`train_test_ds = datasets.load_dataset("bookcorpus", split="train+... | [
-0.12745431065559387,
-0.10093505680561066,
-0.006819732487201691,
0.26014244556427,
0.14704033732414246,
0.07430852949619293,
0.36141237616539,
0.24388569593429565,
-0.18307682871818542,
-0.059660300612449646,
-0.3403538465499878,
0.1565178781747818,
-0.009768269956111908,
0.5907857418060... |
https://github.com/huggingface/datasets/issues/6319 | Hi! Instead of processing a single example at a time, you should use the batched `map` for the best performance (with `num_proc=1`) - the fast tokenizers can process a batch's samples in parallel in that scenario.
E.g., the following code in Colab takes an hour to complete:
```python
# !pip install datasets transf... | Datasets.map is severely broken | ### Describe the bug
Regardless of how many cores I used, I have 16 or 32 threads, map slows down to a crawl at around 80% done, lingers maybe until 97% extremely slowly and NEVER finishes the job. It just hangs.
After watching this for 27 hours I control-C out of it. Until the end one process appears to be doing s... | 73 | Datasets.map is severely broken
### Describe the bug
Regardless of how many cores I used, I have 16 or 32 threads, map slows down to a crawl at around 80% done, lingers maybe until 97% extremely slowly and NEVER finishes the job. It just hangs.
After watching this for 27 hours I control-C out of it. Until the end... | [
-0.6238995790481567,
0.08061352372169495,
0.04081980884075165,
0.1330558806657791,
0.1382634937763214,
0.09867337346076965,
0.2976037859916687,
0.24050694704055786,
-0.1358392834663391,
-0.11661169677972794,
-0.06945439428091049,
0.7371479272842407,
-0.08465425670146942,
-0.326397061347961... |
https://github.com/huggingface/datasets/issues/6319 | Batched is far worse. A single batch of 1000 took hours and that was only 1%
On Thu, Oct 19, 2023, 2:26 PM Mario Šaško ***@***.***> wrote:
> Hi! You should use the batched map for the best performance (with
> num_proc=1) - the fast tokenizers can process a batch's samples in
> parallel.
>
> E.g., the follow... | Datasets.map is severely broken | ### Describe the bug
Regardless of how many cores I used, I have 16 or 32 threads, map slows down to a crawl at around 80% done, lingers maybe until 97% extremely slowly and NEVER finishes the job. It just hangs.
After watching this for 27 hours I control-C out of it. Until the end one process appears to be doing s... | 125 | Datasets.map is severely broken
### Describe the bug
Regardless of how many cores I used, I have 16 or 32 threads, map slows down to a crawl at around 80% done, lingers maybe until 97% extremely slowly and NEVER finishes the job. It just hangs.
After watching this for 27 hours I control-C out of it. Until the end... | [
-0.6426334381103516,
0.04849633574485779,
0.041576087474823,
0.15280376374721527,
0.12931455671787262,
0.12522882223129272,
0.26164349913597107,
0.26831841468811035,
-0.09941567480564117,
-0.1265753209590912,
-0.08454795181751251,
0.7016130089759827,
-0.09337520599365234,
-0.37069922685623... |
https://github.com/huggingface/datasets/issues/6319 | Which specific version of datasets are you using?
What is the architecture of your colab setup? Ram? Cores? OS?
On Thu, Oct 19, 2023, 2:27 PM pensive introvert ***@***.***>
wrote:
> Batched is far worse. A single batch of 1000 took hours and that was only
> 1%
>
>
> On Thu, Oct 19, 2023, 2:26 PM Mario Š... | Datasets.map is severely broken | ### Describe the bug
Regardless of how many cores I used, I have 16 or 32 threads, map slows down to a crawl at around 80% done, lingers maybe until 97% extremely slowly and NEVER finishes the job. It just hangs.
After watching this for 27 hours I control-C out of it. Until the end one process appears to be doing s... | 163 | Datasets.map is severely broken
### Describe the bug
Regardless of how many cores I used, I have 16 or 32 threads, map slows down to a crawl at around 80% done, lingers maybe until 97% extremely slowly and NEVER finishes the job. It just hangs.
After watching this for 27 hours I control-C out of it. Until the end... | [
-0.6481049060821533,
0.10938793420791626,
0.024797581136226654,
0.10106520354747772,
0.0879831314086914,
0.12319760024547577,
0.27200183272361755,
0.2651056945323944,
-0.1283535659313202,
-0.057041190564632416,
-0.055256787687540054,
0.6183364987373352,
-0.12706303596496582,
-0.23674380779... |
https://github.com/huggingface/datasets/issues/6319 | from functools import partial
import transformers
from datasets import load_dataset, concatenate_datasets, load_from_disk
model_name_or_path="/opt/data/data/daryl149/llama-2-7b-chat-hf"
output_dir="/opt/data/data/LongLoRA/checkpoints"
cache_dir="/opt/data/data/LongLoRA/cache"
model_max_length=16384
IGNORE_IN... | Datasets.map is severely broken | ### Describe the bug
Regardless of how many cores I used, I have 16 or 32 threads, map slows down to a crawl at around 80% done, lingers maybe until 97% extremely slowly and NEVER finishes the job. It just hangs.
After watching this for 27 hours I control-C out of it. Until the end one process appears to be doing s... | 170 | Datasets.map is severely broken
### Describe the bug
Regardless of how many cores I used, I have 16 or 32 threads, map slows down to a crawl at around 80% done, lingers maybe until 97% extremely slowly and NEVER finishes the job. It just hangs.
After watching this for 27 hours I control-C out of it. Until the end... | [
-0.6168144345283508,
0.07999947667121887,
0.05563335493206978,
0.20444941520690918,
0.2958584725856781,
0.12437372654676437,
0.2487582415342331,
0.3236621022224426,
-0.10938513278961182,
-0.13184724748134613,
-0.19389095902442932,
0.6695704460144043,
-0.06927616894245148,
-0.41678601503372... |
https://github.com/huggingface/datasets/issues/6319 | I changed the tokenizer to one without "Fast suffix, and something changed.
The fraction, although still slowed a lot at 80% was able to get over the
finish line of 100%
I have to do more testng, see if the whole set can be processed
On Thu, Oct 19, 2023 at 3:03 PM pensive introvert <
***@***.***> wrote:
... | Datasets.map is severely broken | ### Describe the bug
Regardless of how many cores I used, I have 16 or 32 threads, map slows down to a crawl at around 80% done, lingers maybe until 97% extremely slowly and NEVER finishes the job. It just hangs.
After watching this for 27 hours I control-C out of it. Until the end one process appears to be doing s... | 290 | Datasets.map is severely broken
### Describe the bug
Regardless of how many cores I used, I have 16 or 32 threads, map slows down to a crawl at around 80% done, lingers maybe until 97% extremely slowly and NEVER finishes the job. It just hangs.
After watching this for 27 hours I control-C out of it. Until the end... | [
-0.5819017887115479,
0.12894117832183838,
0.033899616450071335,
0.16532237827777863,
0.20738404989242554,
0.16037096083164215,
0.23931464552879333,
0.29810354113578796,
-0.06317737698554993,
-0.18398520350456238,
-0.10214298963546753,
0.6148468255996704,
-0.09060730785131454,
-0.3609420061... |
https://github.com/huggingface/datasets/issues/6319 | So, using LlamaTokenizerFast was the problem. Changing it to LlamaTokenizer
fixed things,
On Thu, Oct 19, 2023 at 4:04 PM pensive introvert <
***@***.***> wrote:
> I changed the tokenizer to one without "Fast suffix, and something
> changed. The fraction, although still slowed a lot at 80% was able to get
> o... | Datasets.map is severely broken | ### Describe the bug
Regardless of how many cores I used, I have 16 or 32 threads, map slows down to a crawl at around 80% done, lingers maybe until 97% extremely slowly and NEVER finishes the job. It just hangs.
After watching this for 27 hours I control-C out of it. Until the end one process appears to be doing s... | 327 | Datasets.map is severely broken
### Describe the bug
Regardless of how many cores I used, I have 16 or 32 threads, map slows down to a crawl at around 80% done, lingers maybe until 97% extremely slowly and NEVER finishes the job. It just hangs.
After watching this for 27 hours I control-C out of it. Until the end... | [
-0.5650825500488281,
0.09585857391357422,
0.027538053691387177,
0.1539461612701416,
0.21590092778205872,
0.12963418662548065,
0.16759265959262848,
0.3100394010543823,
-0.05893123894929886,
-0.16628892719745636,
-0.09866967797279358,
0.6297377347946167,
-0.07801172137260437,
-0.366154909133... |
https://github.com/huggingface/datasets/issues/6319 | Indeed, the tokenizer is super slow. Perhaps @ArthurZucker knows the reason why.
([This](https://colab.research.google.com/drive/1VgeurX-4Fl2X6aBQTwh_X4kuQKZ6K9L1?usp=sharing) simplified Colab can be used to reproduce the behavior) | Datasets.map is severely broken | ### Describe the bug
Regardless of how many cores I used, I have 16 or 32 threads, map slows down to a crawl at around 80% done, lingers maybe until 97% extremely slowly and NEVER finishes the job. It just hangs.
After watching this for 27 hours I control-C out of it. Until the end one process appears to be doing s... | 22 | Datasets.map is severely broken
### Describe the bug
Regardless of how many cores I used, I have 16 or 32 threads, map slows down to a crawl at around 80% done, lingers maybe until 97% extremely slowly and NEVER finishes the job. It just hangs.
After watching this for 27 hours I control-C out of it. Until the end... | [
-0.5410916805267334,
0.13084936141967773,
0.02908453717827797,
0.1009187251329422,
0.17967413365840912,
0.05452781915664673,
0.2067749798297882,
0.2939973771572113,
-0.11034294962882996,
-0.12849867343902588,
-0.00783967599272728,
0.5993674397468567,
-0.07884661853313446,
-0.32366651296615... |
https://github.com/huggingface/datasets/issues/6319 | same issue here
sample to reproduce: https://github.com/philschmid/document-ai-transformers/blob/main/training/donut_sroie.ipynb
with following map line
https://github.com/philschmid/document-ai-transformers/blob/main/training/donut_sroie.ipynb
If I directly iterate over the dataset and call the mapping method, i... | Datasets.map is severely broken | ### Describe the bug
Regardless of how many cores I used, I have 16 or 32 threads, map slows down to a crawl at around 80% done, lingers maybe until 97% extremely slowly and NEVER finishes the job. It just hangs.
After watching this for 27 hours I control-C out of it. Until the end one process appears to be doing s... | 65 | Datasets.map is severely broken
### Describe the bug
Regardless of how many cores I used, I have 16 or 32 threads, map slows down to a crawl at around 80% done, lingers maybe until 97% extremely slowly and NEVER finishes the job. It just hangs.
After watching this for 27 hours I control-C out of it. Until the end... | [
-0.5472961068153381,
0.10284584760665894,
0.04873661696910858,
0.10942556709051132,
0.15640074014663696,
0.11435630172491074,
0.3405649960041046,
0.33139339089393616,
-0.005230031907558441,
-0.20240147411823273,
-0.09187188744544983,
0.5996783971786499,
0.02987278252840042,
-0.368177473545... |
https://github.com/huggingface/datasets/issues/6319 | @ewfian
> If I directly iterate over the dataset and call the mapping method, it is very fast
`Dataset.map` must also convert the images into bytes to write them to an Arrow file (the write itself takes some time, too).
You can make the `map` faster by manually converting the images into an "arrow-compatible... | Datasets.map is severely broken | ### Describe the bug
Regardless of how many cores I used, I have 16 or 32 threads, map slows down to a crawl at around 80% done, lingers maybe until 97% extremely slowly and NEVER finishes the job. It just hangs.
After watching this for 27 hours I control-C out of it. Until the end one process appears to be doing s... | 111 | Datasets.map is severely broken
### Describe the bug
Regardless of how many cores I used, I have 16 or 32 threads, map slows down to a crawl at around 80% done, lingers maybe until 97% extremely slowly and NEVER finishes the job. It just hangs.
After watching this for 27 hours I control-C out of it. Until the end... | [
-0.5075655579566956,
0.08682578802108765,
0.024711772799491882,
0.14701423048973083,
0.13376814126968384,
0.17092423141002655,
0.23593086004257202,
0.4097411036491394,
-0.0756518766283989,
-0.1975446343421936,
-0.001470722258090973,
0.6998304724693298,
-0.0015343837440013885,
-0.3838915228... |
https://github.com/huggingface/datasets/issues/6319 | The problem I had was to do with map using fork and copying locks from the
parent process in acquired state. I ended up changing the context to use
forkserver instead.
On Wed, Nov 29, 2023, 10:04 PM Mario Šaško ***@***.***> wrote:
> @ewfian <https://github.com/ewfian>
>
> If I directly iterate over the data... | Datasets.map is severely broken | ### Describe the bug
Regardless of how many cores I used, I have 16 or 32 threads, map slows down to a crawl at around 80% done, lingers maybe until 97% extremely slowly and NEVER finishes the job. It just hangs.
After watching this for 27 hours I control-C out of it. Until the end one process appears to be doing s... | 208 | Datasets.map is severely broken
### Describe the bug
Regardless of how many cores I used, I have 16 or 32 threads, map slows down to a crawl at around 80% done, lingers maybe until 97% extremely slowly and NEVER finishes the job. It just hangs.
After watching this for 27 hours I control-C out of it. Until the end... | [
-0.5738916993141174,
0.08227047324180603,
0.01397879421710968,
0.18170377612113953,
0.16029129922389984,
0.11835987865924835,
0.19462914764881134,
0.38573387265205383,
-0.01956336945295334,
-0.21568940579891205,
-0.02014840953052044,
0.6655286550521851,
-0.05103820562362671,
-0.41035270690... |
https://github.com/huggingface/datasets/issues/6317 | We have opened an issue in the corresponding Hub dataset: https://huggingface.co/datasets/sentiment140/discussions/3
Let's continue the discussion there. | sentiment140 dataset unavailable | ### Describe the bug
loading the dataset using load_dataset("sentiment140") returns the following error
ConnectionError: Couldn't reach http://cs.stanford.edu/people/alecmgo/trainingandtestdata.zip (error 403)
### Steps to reproduce the bug
Run the following code (version should not matter).
```
from data... | 16 | sentiment140 dataset unavailable
### Describe the bug
loading the dataset using load_dataset("sentiment140") returns the following error
ConnectionError: Couldn't reach http://cs.stanford.edu/people/alecmgo/trainingandtestdata.zip (error 403)
### Steps to reproduce the bug
Run the following code (version sh... | [
-0.36286142468452454,
0.09124556183815002,
-0.0619184784591198,
0.3383810520172119,
0.035611383616924286,
0.05036894977092743,
0.03202589601278305,
0.31903377175331116,
0.07944399118423462,
0.15020966529846191,
-0.2229425013065338,
-0.07228529453277588,
0.09144529700279236,
0.2631198763847... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.