html_url
stringlengths
51
51
comments
stringlengths
67
24.7k
title
stringlengths
6
280
body
stringlengths
51
36.2k
comment_length
int64
16
1.45k
text
stringlengths
190
38.3k
embeddings
list
https://github.com/huggingface/datasets/issues/6597
IIUC, this could also be "fixed" by `create_repo("dataset_name")` not defaulting to `create_repo("user/dataset_name")` (when the user's token is available), which would be consistent with the rest of the `HfApi` ops used in the `push_to_hub` implementation. This is a (small) breaking change for `huggingface_hub`, but j...
Dataset.push_to_hub of a canonical dataset creates an additional dataset under the user namespace
While using `Dataset.push_to_hub` of a canonical dataset, an additional dataset was created under my user namespace. ## Steps to reproduce the bug The command: ```python commit_info = ds.push_to_hub( "caner", config_name="default", commit_message="Convert dataset to Parquet", commit_descriptio...
50
Dataset.push_to_hub of a canonical dataset creates an additional dataset under the user namespace While using `Dataset.push_to_hub` of a canonical dataset, an additional dataset was created under my user namespace. ## Steps to reproduce the bug The command: ```python commit_info = ds.push_to_hub( "caner", ...
[ 0.015438482165336609, -0.24919451773166656, 0.0909704715013504, 0.20266947150230408, 0.17012590169906616, -0.09327994287014008, 0.33426424860954285, 0.1293419897556305, 0.10203634202480316, 0.2504817545413971, -0.17379999160766602, 0.16911724209785461, 0.13417616486549377, 0.18486499786376...
https://github.com/huggingface/datasets/issues/6597
Hmm, creating repo with implicit namespace (e.g. `create_repo("dataset_name")`) is a convenient feature used in a lot of integrations. It is not consistent with other HfApi methods specifically because it is the method to create repos. Once the repo is created, the return value provides the explicit repo_id (`namespace...
Dataset.push_to_hub of a canonical dataset creates an additional dataset under the user namespace
While using `Dataset.push_to_hub` of a canonical dataset, an additional dataset was created under my user namespace. ## Steps to reproduce the bug The command: ```python commit_info = ds.push_to_hub( "caner", config_name="default", commit_message="Convert dataset to Parquet", commit_descriptio...
162
Dataset.push_to_hub of a canonical dataset creates an additional dataset under the user namespace While using `Dataset.push_to_hub` of a canonical dataset, an additional dataset was created under my user namespace. ## Steps to reproduce the bug The command: ```python commit_info = ds.push_to_hub( "caner", ...
[ -0.05977807939052582, -0.03659529983997345, 0.11248317360877991, 0.10316121578216553, 0.3251781463623047, -0.11299433559179306, 0.43455737829208374, 0.2440657913684845, 0.14013831317424774, 0.32332760095596313, -0.33410531282424927, 0.3199589252471924, 0.16297756135463715, 0.12308022379875...
https://github.com/huggingface/datasets/issues/6597
As canonical datasets are going to disappear in the following couple of months, I would not make any effort on their support. I propose reverting #6519, so that the behavior of `push_to_hub` is aligned with the one described in its dosctring: "Also accepts `<dataset_name>`, which will default to the namespace of the...
Dataset.push_to_hub of a canonical dataset creates an additional dataset under the user namespace
While using `Dataset.push_to_hub` of a canonical dataset, an additional dataset was created under my user namespace. ## Steps to reproduce the bug The command: ```python commit_info = ds.push_to_hub( "caner", config_name="default", commit_message="Convert dataset to Parquet", commit_descriptio...
58
Dataset.push_to_hub of a canonical dataset creates an additional dataset under the user namespace While using `Dataset.push_to_hub` of a canonical dataset, an additional dataset was created under my user namespace. ## Steps to reproduce the bug The command: ```python commit_info = ds.push_to_hub( "caner", ...
[ -0.11891815066337585, -0.05135495960712433, 0.05666367709636688, 0.08923321962356567, 0.2523309886455536, -0.1649923175573349, 0.24353288114070892, 0.13927210867404938, -0.03309636563062668, 0.23708900809288025, 0.011932600289583206, 0.4353119730949402, -0.10320337861776352, 0.060031294822...
https://github.com/huggingface/datasets/issues/6595
Hi ! I think the issue comes from the "float16" features that are not supported yet in Parquet Feel free to open an issue in `pyarrow` about this. In the meantime, I'd encourage you to use "float32" for your "pooled_prompt_embeds" and "prompt_embeds" features. You can cast them to "float32" using ```python fr...
Loading big dataset raises pyarrow.lib.ArrowNotImplementedError 2
### Describe the bug I'm aware of the issue #5695 . I'm using a modified SDXL trainer: https://github.com/kopyl/diffusers/blob/5e70f604155aeecee254a5c63c5e4236ad4a0d3d/examples/text_to_image/train_text_to_image_sdxl.py#L1027C16-L1027C16 So i 1. Map dataset 2. Save to disk 3. Try to upload: ``` import data...
64
Loading big dataset raises pyarrow.lib.ArrowNotImplementedError 2 ### Describe the bug I'm aware of the issue #5695 . I'm using a modified SDXL trainer: https://github.com/kopyl/diffusers/blob/5e70f604155aeecee254a5c63c5e4236ad4a0d3d/examples/text_to_image/train_text_to_image_sdxl.py#L1027C16-L1027C16 So i ...
[ -0.28692787885665894, -0.0044410377740859985, 0.16262230277061462, 0.20641998946666718, 0.36014455556869507, -0.21503600478172302, 0.30898481607437134, 0.37914708256721497, -0.34722834825515747, 0.15772300958633423, 0.015633590519428253, 0.6490374207496643, -0.33155137300491333, 0.27174314...
https://github.com/huggingface/datasets/issues/6595
@lhoestq hm. Thank you very much. Do you think it won't have any impact on the training? That it won't break it or the quality won't degrade because of this? I need to use it for [SDXL training](https://github.com/huggingface/diffusers/blob/main/examples/text_to_image/train_text_to_image_sdxl.py)
Loading big dataset raises pyarrow.lib.ArrowNotImplementedError 2
### Describe the bug I'm aware of the issue #5695 . I'm using a modified SDXL trainer: https://github.com/kopyl/diffusers/blob/5e70f604155aeecee254a5c63c5e4236ad4a0d3d/examples/text_to_image/train_text_to_image_sdxl.py#L1027C16-L1027C16 So i 1. Map dataset 2. Save to disk 3. Try to upload: ``` import data...
38
Loading big dataset raises pyarrow.lib.ArrowNotImplementedError 2 ### Describe the bug I'm aware of the issue #5695 . I'm using a modified SDXL trainer: https://github.com/kopyl/diffusers/blob/5e70f604155aeecee254a5c63c5e4236ad4a0d3d/examples/text_to_image/train_text_to_image_sdxl.py#L1027C16-L1027C16 So i ...
[ -0.28692787885665894, -0.0044410377740859985, 0.16262230277061462, 0.20641998946666718, 0.36014455556869507, -0.21503600478172302, 0.30898481607437134, 0.37914708256721497, -0.34722834825515747, 0.15772300958633423, 0.015633590519428253, 0.6490374207496643, -0.33155137300491333, 0.27174314...
https://github.com/huggingface/datasets/issues/6595
Increasing the precision should not degrade training (it only increases the precision), but make sure that it doesn't break your pytorch code (e.g. if it expects a float16 instead of a float32 somewhere)
Loading big dataset raises pyarrow.lib.ArrowNotImplementedError 2
### Describe the bug I'm aware of the issue #5695 . I'm using a modified SDXL trainer: https://github.com/kopyl/diffusers/blob/5e70f604155aeecee254a5c63c5e4236ad4a0d3d/examples/text_to_image/train_text_to_image_sdxl.py#L1027C16-L1027C16 So i 1. Map dataset 2. Save to disk 3. Try to upload: ``` import data...
33
Loading big dataset raises pyarrow.lib.ArrowNotImplementedError 2 ### Describe the bug I'm aware of the issue #5695 . I'm using a modified SDXL trainer: https://github.com/kopyl/diffusers/blob/5e70f604155aeecee254a5c63c5e4236ad4a0d3d/examples/text_to_image/train_text_to_image_sdxl.py#L1027C16-L1027C16 So i ...
[ -0.28692787885665894, -0.0044410377740859985, 0.16262230277061462, 0.20641998946666718, 0.36014455556869507, -0.21503600478172302, 0.30898481607437134, 0.37914708256721497, -0.34722834825515747, 0.15772300958633423, 0.015633590519428253, 0.6490374207496643, -0.33155137300491333, 0.27174314...
https://github.com/huggingface/datasets/issues/6595
@lhoestq just fyi pyarrow 15.0.0 (just released) supports float16 as the underlying parquetcpp does as well now :)
Loading big dataset raises pyarrow.lib.ArrowNotImplementedError 2
### Describe the bug I'm aware of the issue #5695 . I'm using a modified SDXL trainer: https://github.com/kopyl/diffusers/blob/5e70f604155aeecee254a5c63c5e4236ad4a0d3d/examples/text_to_image/train_text_to_image_sdxl.py#L1027C16-L1027C16 So i 1. Map dataset 2. Save to disk 3. Try to upload: ``` import data...
18
Loading big dataset raises pyarrow.lib.ArrowNotImplementedError 2 ### Describe the bug I'm aware of the issue #5695 . I'm using a modified SDXL trainer: https://github.com/kopyl/diffusers/blob/5e70f604155aeecee254a5c63c5e4236ad4a0d3d/examples/text_to_image/train_text_to_image_sdxl.py#L1027C16-L1027C16 So i ...
[ -0.28692787885665894, -0.0044410377740859985, 0.16262230277061462, 0.20641998946666718, 0.36014455556869507, -0.21503600478172302, 0.30898481607437134, 0.37914708256721497, -0.34722834825515747, 0.15772300958633423, 0.015633590519428253, 0.6490374207496643, -0.33155137300491333, 0.27174314...
https://github.com/huggingface/datasets/issues/6595
Oh that's amazing ! (and great timing ^^) @kopyl can you try to update `pyarrow` and try again ? Btw @assignUser there seems to be some casting implementations missing with float16 in 15.0.0, e.g. ``` ArrowNotImplementedError: Unsupported cast from int64 to halffloat using function cast_half_float ``` ```...
Loading big dataset raises pyarrow.lib.ArrowNotImplementedError 2
### Describe the bug I'm aware of the issue #5695 . I'm using a modified SDXL trainer: https://github.com/kopyl/diffusers/blob/5e70f604155aeecee254a5c63c5e4236ad4a0d3d/examples/text_to_image/train_text_to_image_sdxl.py#L1027C16-L1027C16 So i 1. Map dataset 2. Save to disk 3. Try to upload: ``` import data...
58
Loading big dataset raises pyarrow.lib.ArrowNotImplementedError 2 ### Describe the bug I'm aware of the issue #5695 . I'm using a modified SDXL trainer: https://github.com/kopyl/diffusers/blob/5e70f604155aeecee254a5c63c5e4236ad4a0d3d/examples/text_to_image/train_text_to_image_sdxl.py#L1027C16-L1027C16 So i ...
[ -0.28692787885665894, -0.0044410377740859985, 0.16262230277061462, 0.20641998946666718, 0.36014455556869507, -0.21503600478172302, 0.30898481607437134, 0.37914708256721497, -0.34722834825515747, 0.15772300958633423, 0.015633590519428253, 0.6490374207496643, -0.33155137300491333, 0.27174314...
https://github.com/huggingface/datasets/issues/6595
Ah you are right casting is not implemented yet, it's even mentioned in the docs. This pr references the relevant issues if you'd like to track them https://github.com/apache/arrow/pull/38494
Loading big dataset raises pyarrow.lib.ArrowNotImplementedError 2
### Describe the bug I'm aware of the issue #5695 . I'm using a modified SDXL trainer: https://github.com/kopyl/diffusers/blob/5e70f604155aeecee254a5c63c5e4236ad4a0d3d/examples/text_to_image/train_text_to_image_sdxl.py#L1027C16-L1027C16 So i 1. Map dataset 2. Save to disk 3. Try to upload: ``` import data...
28
Loading big dataset raises pyarrow.lib.ArrowNotImplementedError 2 ### Describe the bug I'm aware of the issue #5695 . I'm using a modified SDXL trainer: https://github.com/kopyl/diffusers/blob/5e70f604155aeecee254a5c63c5e4236ad4a0d3d/examples/text_to_image/train_text_to_image_sdxl.py#L1027C16-L1027C16 So i ...
[ -0.28692787885665894, -0.0044410377740859985, 0.16262230277061462, 0.20641998946666718, 0.36014455556869507, -0.21503600478172302, 0.30898481607437134, 0.37914708256721497, -0.34722834825515747, 0.15772300958633423, 0.015633590519428253, 0.6490374207496643, -0.33155137300491333, 0.27174314...
https://github.com/huggingface/datasets/issues/6595
@lhoestq i just recently found out that it's supported in 15.0.0, but wanted to try it first before telling you... Trying this right now and it seemingly works (although i need to wait till the end to make sure there is nothing wrong). Will update you when it's finished. <img width="918" alt="image" src="https://...
Loading big dataset raises pyarrow.lib.ArrowNotImplementedError 2
### Describe the bug I'm aware of the issue #5695 . I'm using a modified SDXL trainer: https://github.com/kopyl/diffusers/blob/5e70f604155aeecee254a5c63c5e4236ad4a0d3d/examples/text_to_image/train_text_to_image_sdxl.py#L1027C16-L1027C16 So i 1. Map dataset 2. Save to disk 3. Try to upload: ``` import data...
87
Loading big dataset raises pyarrow.lib.ArrowNotImplementedError 2 ### Describe the bug I'm aware of the issue #5695 . I'm using a modified SDXL trainer: https://github.com/kopyl/diffusers/blob/5e70f604155aeecee254a5c63c5e4236ad4a0d3d/examples/text_to_image/train_text_to_image_sdxl.py#L1027C16-L1027C16 So i ...
[ -0.28692787885665894, -0.0044410377740859985, 0.16262230277061462, 0.20641998946666718, 0.36014455556869507, -0.21503600478172302, 0.30898481607437134, 0.37914708256721497, -0.34722834825515747, 0.15772300958633423, 0.015633590519428253, 0.6490374207496643, -0.33155137300491333, 0.27174314...
https://github.com/huggingface/datasets/issues/6595
@lhoestq also it's strange that there was no error for a dataset with the same features, same data type, but smaller (much smaller). Altho i'm not sure about this, but chances are the dataset was loaded directly, not `load_from_disk`.... Maybe because of this.
Loading big dataset raises pyarrow.lib.ArrowNotImplementedError 2
### Describe the bug I'm aware of the issue #5695 . I'm using a modified SDXL trainer: https://github.com/kopyl/diffusers/blob/5e70f604155aeecee254a5c63c5e4236ad4a0d3d/examples/text_to_image/train_text_to_image_sdxl.py#L1027C16-L1027C16 So i 1. Map dataset 2. Save to disk 3. Try to upload: ``` import data...
43
Loading big dataset raises pyarrow.lib.ArrowNotImplementedError 2 ### Describe the bug I'm aware of the issue #5695 . I'm using a modified SDXL trainer: https://github.com/kopyl/diffusers/blob/5e70f604155aeecee254a5c63c5e4236ad4a0d3d/examples/text_to_image/train_text_to_image_sdxl.py#L1027C16-L1027C16 So i ...
[ -0.28692787885665894, -0.0044410377740859985, 0.16262230277061462, 0.20641998946666718, 0.36014455556869507, -0.21503600478172302, 0.30898481607437134, 0.37914708256721497, -0.34722834825515747, 0.15772300958633423, 0.015633590519428253, 0.6490374207496643, -0.33155137300491333, 0.27174314...
https://github.com/huggingface/datasets/issues/6595
> What does that missing casting implementation mean for my specific case and what does it mean in general? Nothing for you, just that casting to float16 using `.cast_column("my_column_name", Value("float16"))` raises an error > Do you know how to push_to_hub with multiple processes? It's not possible (yet ?)....
Loading big dataset raises pyarrow.lib.ArrowNotImplementedError 2
### Describe the bug I'm aware of the issue #5695 . I'm using a modified SDXL trainer: https://github.com/kopyl/diffusers/blob/5e70f604155aeecee254a5c63c5e4236ad4a0d3d/examples/text_to_image/train_text_to_image_sdxl.py#L1027C16-L1027C16 So i 1. Map dataset 2. Save to disk 3. Try to upload: ``` import data...
89
Loading big dataset raises pyarrow.lib.ArrowNotImplementedError 2 ### Describe the bug I'm aware of the issue #5695 . I'm using a modified SDXL trainer: https://github.com/kopyl/diffusers/blob/5e70f604155aeecee254a5c63c5e4236ad4a0d3d/examples/text_to_image/train_text_to_image_sdxl.py#L1027C16-L1027C16 So i ...
[ -0.28692787885665894, -0.0044410377740859985, 0.16262230277061462, 0.20641998946666718, 0.36014455556869507, -0.21503600478172302, 0.30898481607437134, 0.37914708256721497, -0.34722834825515747, 0.15772300958633423, 0.015633590519428253, 0.6490374207496643, -0.33155137300491333, 0.27174314...
https://github.com/huggingface/datasets/issues/6595
@lhoestq thank you very much. That would be amazing, I need to create a feature request for this :) By the way, in short, how does hf_transfer improves the upload speed under the hood?
Loading big dataset raises pyarrow.lib.ArrowNotImplementedError 2
### Describe the bug I'm aware of the issue #5695 . I'm using a modified SDXL trainer: https://github.com/kopyl/diffusers/blob/5e70f604155aeecee254a5c63c5e4236ad4a0d3d/examples/text_to_image/train_text_to_image_sdxl.py#L1027C16-L1027C16 So i 1. Map dataset 2. Save to disk 3. Try to upload: ``` import data...
34
Loading big dataset raises pyarrow.lib.ArrowNotImplementedError 2 ### Describe the bug I'm aware of the issue #5695 . I'm using a modified SDXL trainer: https://github.com/kopyl/diffusers/blob/5e70f604155aeecee254a5c63c5e4236ad4a0d3d/examples/text_to_image/train_text_to_image_sdxl.py#L1027C16-L1027C16 So i ...
[ -0.28692787885665894, -0.0044410377740859985, 0.16262230277061462, 0.20641998946666718, 0.36014455556869507, -0.21503600478172302, 0.30898481607437134, 0.37914708256721497, -0.34722834825515747, 0.15772300958633423, 0.015633590519428253, 0.6490374207496643, -0.33155137300491333, 0.27174314...
https://github.com/huggingface/datasets/issues/6595
@lhoestq i was just able to successfully upload without the dataset with the new pyarrow update and without increasing the precision :)
Loading big dataset raises pyarrow.lib.ArrowNotImplementedError 2
### Describe the bug I'm aware of the issue #5695 . I'm using a modified SDXL trainer: https://github.com/kopyl/diffusers/blob/5e70f604155aeecee254a5c63c5e4236ad4a0d3d/examples/text_to_image/train_text_to_image_sdxl.py#L1027C16-L1027C16 So i 1. Map dataset 2. Save to disk 3. Try to upload: ``` import data...
22
Loading big dataset raises pyarrow.lib.ArrowNotImplementedError 2 ### Describe the bug I'm aware of the issue #5695 . I'm using a modified SDXL trainer: https://github.com/kopyl/diffusers/blob/5e70f604155aeecee254a5c63c5e4236ad4a0d3d/examples/text_to_image/train_text_to_image_sdxl.py#L1027C16-L1027C16 So i ...
[ -0.28692787885665894, -0.0044410377740859985, 0.16262230277061462, 0.20641998946666718, 0.36014455556869507, -0.21503600478172302, 0.30898481607437134, 0.37914708256721497, -0.34722834825515747, 0.15772300958633423, 0.015633590519428253, 0.6490374207496643, -0.33155137300491333, 0.27174314...
https://github.com/huggingface/datasets/issues/6592
Hi! `tqdm` doesn't work well in non-interactive environments, so there isn't much we can do about this. It's best to [disable it](https://huggingface.co/docs/datasets/v2.16.1/en/package_reference/utilities#datasets.disable_progress_bars) in such environments and instead use logging to track progress.
Logs are delayed when doing .map when `docker logs`
### Describe the bug When I run my SD training in a Docker image and then listen to logs like `docker logs train -f`, the progress bar is delayed. It's updating every few percent. When you have a large dataset that has to be mapped (like 1+ million samples), it's crucial to see the updates in real-time, not every co...
32
Logs are delayed when doing .map when `docker logs` ### Describe the bug When I run my SD training in a Docker image and then listen to logs like `docker logs train -f`, the progress bar is delayed. It's updating every few percent. When you have a large dataset that has to be mapped (like 1+ million samples), it's...
[ -0.5378693342208862, -0.46852540969848633, -0.023110635578632355, -0.14820663630962372, 0.2111946940422058, -0.05372349172830582, 0.33600133657455444, 0.13271595537662506, -0.030799970030784607, 0.34662723541259766, -0.05821090191602707, 0.3205311596393585, -0.1769188642501831, 0.224839180...
https://github.com/huggingface/datasets/issues/6591
Hi! Indeed, Dropbox is not a reliable host. I've just merged https://huggingface.co/datasets/PolyAI/minds14/discussions/24 to fix this by hosting the data files inside the repo.
The datasets models housed in Dropbox can't support a lot of users downloading them
### Describe the bug I'm using the datasets ``` from datasets import load_dataset, Audio dataset = load_dataset("PolyAI/minds14", name="en-US", split="train") ``` And it seems that sometimes when I imagine a lot of users are accessing the same resources, the Dropbox host fails: `raise ConnectionError(...
23
The datasets models housed in Dropbox can't support a lot of users downloading them ### Describe the bug I'm using the datasets ``` from datasets import load_dataset, Audio dataset = load_dataset("PolyAI/minds14", name="en-US", split="train") ``` And it seems that sometimes when I imagine a lot of user...
[ -0.34906262159347534, 0.30475425720214844, -0.036654695868492126, 0.5930954217910767, 0.2086874395608902, -0.12326231598854065, 0.3543434739112854, 0.08312855660915375, 0.21345342695713043, 0.14639872312545776, -0.38722750544548035, -0.04541517421603203, -0.1008751392364502, 0.193244695663...
https://github.com/huggingface/datasets/issues/6585
Hi ! This issue comes from the fact that `map()` with `num_proc>1` shards the dataset in multiple chunks to be processed (one per process) and merges them. The DatasetInfos of each chunk are then merged together, but for some fields like `dataset_name` it's not been implemented and default to None. The DatasetInfo m...
losing DatasetInfo in Dataset.map when num_proc > 1
### Describe the bug Hello and thanks for developing this package! When I process a Dataset with the map function using multiple processors some set attributes of the DatasetInfo get lost and are None in the resulting Dataset. ### Steps to reproduce the bug ```python from datasets import Dataset, DatasetInfo...
65
losing DatasetInfo in Dataset.map when num_proc > 1 ### Describe the bug Hello and thanks for developing this package! When I process a Dataset with the map function using multiple processors some set attributes of the DatasetInfo get lost and are None in the resulting Dataset. ### Steps to reproduce the bug...
[ -0.35594192147254944, -0.10743378102779388, -0.007776990532875061, 0.6413049101829529, 0.2363981008529663, 0.20442619919776917, 0.16280147433280945, 0.22665756940841675, 0.17084787786006927, 0.3607620298862457, 0.25110965967178345, 0.6130760908126831, 0.121337890625, 0.14003291726112366, ...
https://github.com/huggingface/datasets/issues/6584
@lhoestq ``` Traceback (most recent call last): File "/home/dongzf/miniconda3/envs/dataset_ai/lib/python3.11/runpy.py", line 198, in _run_module_as_main return _run_code(code, main_globals, None, ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/home/dongzf/miniconda3/envs/dataset_ai/lib/python3.11/r...
np.fromfile not supported
How to do np.fromfile to use it like np.load ```python def xnumpy_fromfile(filepath_or_buffer, *args, download_config: Optional[DownloadConfig] = None, **kwargs): import numpy as np if hasattr(filepath_or_buffer, "read"): return np.fromfile(filepath_or_buffer, *args, **kwargs) else: ...
105
np.fromfile not supported How to do np.fromfile to use it like np.load ```python def xnumpy_fromfile(filepath_or_buffer, *args, download_config: Optional[DownloadConfig] = None, **kwargs): import numpy as np if hasattr(filepath_or_buffer, "read"): return np.fromfile(filepath_or_buffer, *ar...
[ -0.17793172597885132, -0.15880124270915985, -0.05569717288017273, 0.12712755799293518, 0.36935845017433167, -0.17062661051750183, 0.31794315576553345, 0.3332071900367737, 0.26488226652145386, 0.1581835001707077, -0.04350850731134415, 0.40969303250312805, 0.08633576333522797, 0.339933156967...
https://github.com/huggingface/datasets/issues/6584
I used this method to read point cloud data in the script ```python with open(velodyne_filepath,"rb") as obj: velodyne_data = numpy.frombuffer(obj.read(), dtype=numpy.float32).reshape([-1, 4]) ```
np.fromfile not supported
How to do np.fromfile to use it like np.load ```python def xnumpy_fromfile(filepath_or_buffer, *args, download_config: Optional[DownloadConfig] = None, **kwargs): import numpy as np if hasattr(filepath_or_buffer, "read"): return np.fromfile(filepath_or_buffer, *args, **kwargs) else: ...
23
np.fromfile not supported How to do np.fromfile to use it like np.load ```python def xnumpy_fromfile(filepath_or_buffer, *args, download_config: Optional[DownloadConfig] = None, **kwargs): import numpy as np if hasattr(filepath_or_buffer, "read"): return np.fromfile(filepath_or_buffer, *ar...
[ -0.26546746492385864, -0.19931796193122864, -0.07483597099781036, 0.15513744950294495, 0.2779790163040161, -0.2699272930622101, 0.21601557731628418, 0.22307102382183075, 0.18922102451324463, 0.20724502205848694, 0.07521598786115646, 0.3033961057662964, 0.22715897858142853, 0.43538713455200...
https://github.com/huggingface/datasets/issues/6579
Hi @haok1402, I have created an issue in the Discussion tab of the corresponding dataset: https://huggingface.co/datasets/eli5/discussions/7 Let's continue the discussion there!
Unable to load `eli5` dataset with streaming
### Describe the bug Unable to load `eli5` dataset with streaming. ### Steps to reproduce the bug This fails with FileNotFoundError: https://files.pushshift.io/reddit/submissions ``` from datasets import load_dataset load_dataset("eli5", streaming=True) ``` This works correctly. ``` from datasets import lo...
21
Unable to load `eli5` dataset with streaming ### Describe the bug Unable to load `eli5` dataset with streaming. ### Steps to reproduce the bug This fails with FileNotFoundError: https://files.pushshift.io/reddit/submissions ``` from datasets import load_dataset load_dataset("eli5", streaming=True) ``` This...
[ -0.3034294843673706, -0.37919026613235474, 0.0074720680713653564, 0.35745134949684143, 0.3565647602081299, -0.05574629455804825, 0.20355500280857086, 0.06192155182361603, 0.0340673103928566, 0.1009824275970459, -0.22593586146831512, 0.18525493144989014, 0.04847179725766182, 0.3414548337459...
https://github.com/huggingface/datasets/issues/6577
Hi! We should be able to avoid this error by retrying to read the data when it happens. I'll open a PR in `huggingface_hub` to address this.
502 Server Errors when streaming large dataset
### Describe the bug When streaming a [large ASR dataset](https://huggingface.co/datasets/sanchit-gandhi/concatenated-train-set) from the Hug (~3TB) I often encounter 502 Server Errors seemingly randomly during streaming: ``` huggingface_hub.utils._errors.HfHubHTTPError: 502 Server Error: Bad Gateway for url: http...
27
502 Server Errors when streaming large dataset ### Describe the bug When streaming a [large ASR dataset](https://huggingface.co/datasets/sanchit-gandhi/concatenated-train-set) from the Hug (~3TB) I often encounter 502 Server Errors seemingly randomly during streaming: ``` huggingface_hub.utils._errors.HfHubHTTPE...
[ -0.22672490775585175, -0.5951704978942871, 0.15816134214401245, 0.31269171833992004, 0.21198102831840515, -0.14130067825317383, 0.1209697276353836, 0.03074195235967636, -0.4138365089893341, -0.19666787981987, -0.00406932458281517, -0.03385841101408005, -0.0011482946574687958, 0.14881344139...
https://github.com/huggingface/datasets/issues/6577
Thanks for the fix @mariosasko! Just wondering whether "500 error" should also be excluded? I got these errors overnight: ``` huggingface_hub.utils._errors.HfHubHTTPError: 500 Server Error: Internal Server Error for url: https://huggingface.co/da tasets/sanchit-gandhi/concatenated-train-set-label-length-256/resolv...
502 Server Errors when streaming large dataset
### Describe the bug When streaming a [large ASR dataset](https://huggingface.co/datasets/sanchit-gandhi/concatenated-train-set) from the Hug (~3TB) I often encounter 502 Server Errors seemingly randomly during streaming: ``` huggingface_hub.utils._errors.HfHubHTTPError: 502 Server Error: Bad Gateway for url: http...
49
502 Server Errors when streaming large dataset ### Describe the bug When streaming a [large ASR dataset](https://huggingface.co/datasets/sanchit-gandhi/concatenated-train-set) from the Hug (~3TB) I often encounter 502 Server Errors seemingly randomly during streaming: ``` huggingface_hub.utils._errors.HfHubHTTPE...
[ -0.22672490775585175, -0.5951704978942871, 0.15816134214401245, 0.31269171833992004, 0.21198102831840515, -0.14130067825317383, 0.1209697276353836, 0.03074195235967636, -0.4138365089893341, -0.19666787981987, -0.00406932458281517, -0.03385841101408005, -0.0011482946574687958, 0.14881344139...
https://github.com/huggingface/datasets/issues/6568
Seems like I just used the old code which did not have `keep_in_memory=True` argument, sorry. Although i encountered a different problem – at 97% my python process just hung for around 11 minutes with no logs (when running dataset.map without `keep_in_memory=True` over around 3 million of dataset samples)...
keep_in_memory=True does not seem to work
UPD: [Fixed](https://github.com/huggingface/datasets/issues/6568#issuecomment-1880817794) . But a new issue came up :(
48
keep_in_memory=True does not seem to work UPD: [Fixed](https://github.com/huggingface/datasets/issues/6568#issuecomment-1880817794) . But a new issue came up :( Seems like I just used the old code which did not have `keep_in_memory=True` argument, sorry. Although i encountered a different problem – at 97% my pyt...
[ -0.5037835240364075, -0.29181116819381714, -0.12347058951854706, 0.014672212302684784, 0.3082585036754608, -0.032749563455581665, 0.06944156438112259, 0.22768092155456543, 0.14811402559280396, 0.21746718883514404, -0.13731825351715088, 0.2877124547958374, -0.2045990377664566, 0.29882973432...
https://github.com/huggingface/datasets/issues/6568
Can you open a new issue and provide a bit more details ? What kind of map operations did you run ?
keep_in_memory=True does not seem to work
UPD: [Fixed](https://github.com/huggingface/datasets/issues/6568#issuecomment-1880817794) . But a new issue came up :(
22
keep_in_memory=True does not seem to work UPD: [Fixed](https://github.com/huggingface/datasets/issues/6568#issuecomment-1880817794) . But a new issue came up :( Can you open a new issue and provide a bit more details ? What kind of map operations did you run ?
[ -0.3748677968978882, -0.49990132451057434, -0.11895636469125748, 0.2844051420688629, 0.3289105296134949, -0.2569744288921356, 0.14168044924736023, 0.1253788322210312, 0.10821300745010376, 0.3469023108482361, -0.14063313603401184, 0.25066202878952026, -0.0918361097574234, 0.2976199984550476...
https://github.com/huggingface/datasets/issues/6568
Hey. I will try to find some free time to describe it. (can't do it now, cause i need to reproduce it myself to be sure about everything, which requires spinning a new Azuree VM, copying a huge dataset to drive from network disk for a long time etc...)
keep_in_memory=True does not seem to work
UPD: [Fixed](https://github.com/huggingface/datasets/issues/6568#issuecomment-1880817794) . But a new issue came up :(
49
keep_in_memory=True does not seem to work UPD: [Fixed](https://github.com/huggingface/datasets/issues/6568#issuecomment-1880817794) . But a new issue came up :( Hey. I will try to find some free time to describe it. (can't do it now, cause i need to reproduce it myself to be sure about everything, which requires...
[ -0.30082494020462036, -0.45631641149520874, -0.1075349673628807, 0.14783304929733276, 0.19342222809791565, -0.2760157883167267, 0.03683233633637428, 0.1468944251537323, -0.1389954388141632, 0.28641635179519653, -0.086149200797081, 0.16647785902023315, -0.02673282101750374, 0.30877977609634...
https://github.com/huggingface/datasets/issues/6568
@lhoestq loading dataset like this does not spawn 50 python processes: ``` datasets.load_dataset("/preprocessed_2256k/train", num_proc=50) ``` I have 64 vCPU so i hoped it could speed up the dataset loading... My dataset onlly has images and metadata.csv with text column alongside image file path column
keep_in_memory=True does not seem to work
UPD: [Fixed](https://github.com/huggingface/datasets/issues/6568#issuecomment-1880817794) . But a new issue came up :(
44
keep_in_memory=True does not seem to work UPD: [Fixed](https://github.com/huggingface/datasets/issues/6568#issuecomment-1880817794) . But a new issue came up :( @lhoestq loading dataset like this does not spawn 50 python processes: ``` datasets.load_dataset("/preprocessed_2256k/train", num_proc=50) ``` I ha...
[ -0.4864200949668884, -0.2658992111682892, -0.0638510063290596, 0.3380093276500702, 0.3865876793861389, -0.06995498389005661, 0.20234602689743042, 0.1355111449956894, 0.1304621696472168, 0.05938532203435898, -0.12590380012989044, 0.1253167688846588, -0.19034166634082794, 0.3952617347240448,...
https://github.com/huggingface/datasets/issues/6568
now noticed ``` 'Setting num_proc from 50 back to 1 for the train split to disable multiprocessing as it only contains one shard ``` Any way to work around this?
keep_in_memory=True does not seem to work
UPD: [Fixed](https://github.com/huggingface/datasets/issues/6568#issuecomment-1880817794) . But a new issue came up :(
30
keep_in_memory=True does not seem to work UPD: [Fixed](https://github.com/huggingface/datasets/issues/6568#issuecomment-1880817794) . But a new issue came up :( now noticed ``` 'Setting num_proc from 50 back to 1 for the train split to disable multiprocessing as it only contains one shard ``` Any way to work ...
[ -0.3578808009624481, -0.48597973585128784, -0.05335165560245514, 0.24324099719524384, 0.22478020191192627, -0.3293018937110901, 0.29327523708343506, 0.05787957087159157, -0.25964412093162537, 0.37969741225242615, -0.09009426832199097, -0.006218463182449341, -0.20374412834644318, 0.60562366...
https://github.com/huggingface/datasets/issues/6567
I think you are reporting an issue with the `transformers` library. Note this is the repository of the `datasets` library. I recommend that you open an issue in their repository: https://github.com/huggingface/transformers/issues EDIT: I have not the rights to transfer the issue ~~I am transferring your issue to th...
AttributeError: 'str' object has no attribute 'to'
### Describe the bug ``` -------------------------------------------------------------------------- AttributeError Traceback (most recent call last) [<ipython-input-6-80c6086794e8>](https://localhost:8080/#) in <cell line: 10>() 8 report_to="wandb") 9 ---> 10 trainer =...
49
AttributeError: 'str' object has no attribute 'to' ### Describe the bug ``` -------------------------------------------------------------------------- AttributeError Traceback (most recent call last) [<ipython-input-6-80c6086794e8>](https://localhost:8080/#) in <cell line: 10>() ...
[ -0.08758515119552612, -0.32732489705085754, 0.08826373517513275, 0.04437369853258133, 0.5676831007003784, 0.01807337999343872, 0.5729138255119324, 0.4033966362476349, -0.17493602633476257, 0.32560089230537415, -0.011736348271369934, 0.22309105098247528, -0.11257587373256683, -0.03418769687...
https://github.com/huggingface/datasets/issues/6567
Thanks, I hope someone from transformers library addresses this issue. On Mon, Jan 8, 2024 at 15:29 Albert Villanova del Moral < ***@***.***> wrote: > I think you are reporting an issue with the transformers library. Note > this is the repository of the datasets library. I am transferring your > issue to their...
AttributeError: 'str' object has no attribute 'to'
### Describe the bug ``` -------------------------------------------------------------------------- AttributeError Traceback (most recent call last) [<ipython-input-6-80c6086794e8>](https://localhost:8080/#) in <cell line: 10>() 8 report_to="wandb") 9 ---> 10 trainer =...
91
AttributeError: 'str' object has no attribute 'to' ### Describe the bug ``` -------------------------------------------------------------------------- AttributeError Traceback (most recent call last) [<ipython-input-6-80c6086794e8>](https://localhost:8080/#) in <cell line: 10>() ...
[ -0.021199770271778107, -0.24903526902198792, 0.0938526839017868, 0.019450323656201363, 0.608004629611969, 0.023181691765785217, 0.6019617319107056, 0.42540740966796875, -0.18105585873126984, 0.31553900241851807, 0.025728940963745117, 0.1877838373184204, -0.1374649852514267, -0.151730209589...
https://github.com/huggingface/datasets/issues/6567
@andysingal, I recommend that you open an issue in their repository: https://github.com/huggingface/transformers/issues I don't have the rights to transfer this issue to their repo.
AttributeError: 'str' object has no attribute 'to'
### Describe the bug ``` -------------------------------------------------------------------------- AttributeError Traceback (most recent call last) [<ipython-input-6-80c6086794e8>](https://localhost:8080/#) in <cell line: 10>() 8 report_to="wandb") 9 ---> 10 trainer =...
24
AttributeError: 'str' object has no attribute 'to' ### Describe the bug ``` -------------------------------------------------------------------------- AttributeError Traceback (most recent call last) [<ipython-input-6-80c6086794e8>](https://localhost:8080/#) in <cell line: 10>() ...
[ 0.0757463276386261, -0.4929776191711426, 0.0982719212770462, 0.06737048178911209, 0.5980656743049622, -0.017362095415592194, 0.5761122107505798, 0.3846137523651123, -0.1268261820077896, 0.3702743947505951, -0.019784869626164436, 0.16745971143245697, -0.11638899147510529, -0.063405297696590...
https://github.com/huggingface/datasets/issues/6566
I also see the same error and get passed it by casting that line to float. so `for x in obj.detach().cpu().numpy()` becomes `for x in obj.detach().to(torch.float).cpu().numpy()` I got the idea from [this ](https://github.com/kohya-ss/sd-webui-additional-networks/pull/128/files) PR where someone was facing a sim...
I train controlnet_sdxl in bf16 datatype, got unsupported ERROR in datasets
### Describe the bug ``` Traceback (most recent call last): File "train_controlnet_sdxl.py", line 1252, in <module> main(args) File "train_controlnet_sdxl.py", line 1013, in main train_dataset = train_dataset.map(compute_embeddings_fn, batched=True, new_fingerprint=new_fingerprint) File "/home/mini...
51
I train controlnet_sdxl in bf16 datatype, got unsupported ERROR in datasets ### Describe the bug ``` Traceback (most recent call last): File "train_controlnet_sdxl.py", line 1252, in <module> main(args) File "train_controlnet_sdxl.py", line 1013, in main train_dataset = train_dataset.map(compute_emb...
[ -0.5037292242050171, -0.22553563117980957, -0.013837732374668121, 0.18832528591156006, 0.5082089304924011, 0.14244911074638367, 0.488095223903656, 0.20407439768314362, 0.4085872173309326, 0.15492910146713257, 0.07041168957948685, 0.09321921318769455, -0.4428427815437317, 0.2823957800865173...
https://github.com/huggingface/datasets/issues/6565
My current workaround this issue is to return `None` in the second element and then filter out samples which have `None` in them. ```python def merge_samples(batch): if len(batch['a']) == 1: batch['c'] = [batch['a'][0]] batch['d'] = [None] else: batch['c'] = [batch['a'][0]] ...
`drop_last_batch=True` for IterableDataset map function is ignored with multiprocessing DataLoader
### Describe the bug Scenario: - Interleaving two iterable datasets of unequal lengths (`all_exhausted`), followed by a batch mapping with batch size 2 to effectively merge the two datasets and get a sample from each dataset in a single batch, with `drop_last_batch=True` to skip the last batch in case it doesn't ha...
59
`drop_last_batch=True` for IterableDataset map function is ignored with multiprocessing DataLoader ### Describe the bug Scenario: - Interleaving two iterable datasets of unequal lengths (`all_exhausted`), followed by a batch mapping with batch size 2 to effectively merge the two datasets and get a sample from e...
[ -0.29564857482910156, -0.06441396474838257, -0.031954340636730194, 0.10420383512973785, -0.08406788110733032, -0.06830085813999176, 0.658607006072998, 0.2046472430229187, 0.2394963800907135, 0.4104592502117157, -0.027381081134080887, 0.37076422572135925, -0.1618267297744751, -0.24033448100...
https://github.com/huggingface/datasets/issues/6563
<del>Installing `datasets` from `main` did the trick so I guess it will be fixed in the next release. NVM https://github.com/huggingface/datasets/blob/d26abadce0b884db32382b92422d8a6aa997d40a/src/datasets/utils/info_utils.py#L5
`ImportError`: cannot import name 'insecure_hashlib' from 'huggingface_hub.utils' (.../huggingface_hub/utils/__init__.py)
### Describe the bug Yep its not [there](https://github.com/huggingface/huggingface_hub/blob/main/src/huggingface_hub/utils/__init__.py) anymore. ```text + python /home/trainer/sft_train.py --model_name cognitivecomputations/dolphin-2.2.1-mistral-7b --dataset_name wasertech/OneOS --load_in_4bit --use_peft --batch_...
20
`ImportError`: cannot import name 'insecure_hashlib' from 'huggingface_hub.utils' (.../huggingface_hub/utils/__init__.py) ### Describe the bug Yep its not [there](https://github.com/huggingface/huggingface_hub/blob/main/src/huggingface_hub/utils/__init__.py) anymore. ```text + python /home/trainer/sft_train.py -...
[ -0.05503574013710022, -0.570375382900238, -0.049042604863643646, 0.1578817367553711, 0.3562760055065155, -0.001315489411354065, 0.1583208292722702, 0.356046199798584, 0.1492595672607422, 0.24469077587127686, -0.12720763683319092, 0.24285480380058289, -0.30560654401779175, 0.235667914152145...
https://github.com/huggingface/datasets/issues/6563
Ha yes I had pinned `tokenizers` to an old version so it downgraded `huggingface_hub`. Note to myself keep HuggingFace modules relatively close together chronologically release wise.
`ImportError`: cannot import name 'insecure_hashlib' from 'huggingface_hub.utils' (.../huggingface_hub/utils/__init__.py)
### Describe the bug Yep its not [there](https://github.com/huggingface/huggingface_hub/blob/main/src/huggingface_hub/utils/__init__.py) anymore. ```text + python /home/trainer/sft_train.py --model_name cognitivecomputations/dolphin-2.2.1-mistral-7b --dataset_name wasertech/OneOS --load_in_4bit --use_peft --batch_...
26
`ImportError`: cannot import name 'insecure_hashlib' from 'huggingface_hub.utils' (.../huggingface_hub/utils/__init__.py) ### Describe the bug Yep its not [there](https://github.com/huggingface/huggingface_hub/blob/main/src/huggingface_hub/utils/__init__.py) anymore. ```text + python /home/trainer/sft_train.py -...
[ -0.05503574013710022, -0.570375382900238, -0.049042604863643646, 0.1578817367553711, 0.3562760055065155, -0.001315489411354065, 0.1583208292722702, 0.356046199798584, 0.1492595672607422, 0.24469077587127686, -0.12720763683319092, 0.24285480380058289, -0.30560654401779175, 0.235667914152145...
https://github.com/huggingface/datasets/issues/6561
In particular, I would like to have an example of how to replace the following configuration (from https://huggingface.co/docs/hub/datasets-manual-configuration#splits) ``` --- configs: - config_name: default data_files: - split: train path: "data/*.csv" - split: test path: "holdout/*.csv" ---...
Document YAML configuration with "data_dir"
See https://huggingface.co/datasets/uonlp/CulturaX/discussions/15#6597e83f185db94370d6bf50 for reference
41
Document YAML configuration with "data_dir" See https://huggingface.co/datasets/uonlp/CulturaX/discussions/15#6597e83f185db94370d6bf50 for reference In particular, I would like to have an example of how to replace the following configuration (from https://huggingface.co/docs/hub/datasets-manual-configuration#splits...
[ -0.25297069549560547, -0.1603081226348877, 0.0808432325720787, 0.018189026042819023, 0.23896105587482452, 0.40161192417144775, 0.35298261046409607, 0.054908160120248795, 0.022724732756614685, 0.16462454199790955, 0.032079875469207764, 0.03859062120318413, 0.06359618902206421, 0.42989626526...
https://github.com/huggingface/datasets/issues/6559
Hi ! The "allenai--c4" config doesn't exist (this naming schema comes from old versions of `datasets`) You can load it this way instead: ```python from datasets import load_dataset cache_dir = 'path/to/your/cache/directory' dataset = load_dataset('allenai/c4', data_files={'train': 'en/c4-train.00000-of-01024.j...
Latest version 2.16.1, when load dataset error occurs. ValueError: BuilderConfig 'allenai--c4' not found. Available: ['default']
### Describe the bug python script is: ``` from datasets import load_dataset cache_dir = 'path/to/your/cache/directory' dataset = load_dataset('allenai/c4','allenai--c4', data_files={'train': 'en/c4-train.00000-of-01024.json.gz'}, split='train', use_auth_token=False, cache_dir=cache_dir) ``` the script su...
39
Latest version 2.16.1, when load dataset error occurs. ValueError: BuilderConfig 'allenai--c4' not found. Available: ['default'] ### Describe the bug python script is: ``` from datasets import load_dataset cache_dir = 'path/to/your/cache/directory' dataset = load_dataset('allenai/c4','allenai--c4', data_files=...
[ -0.5463299751281738, 0.45560142397880554, -0.02869529463350773, 0.3143913149833679, 0.14552779495716095, 0.10748502612113953, 0.31146010756492615, 0.3529205918312073, 0.18226969242095947, 0.20120511949062347, 0.2688750624656677, 0.6102460026741028, -0.1364494264125824, -0.1731400191783905,...
https://github.com/huggingface/datasets/issues/6559
> Hi ! The "allenai--c4" config doesn't exist (this naming schema comes from old versions of `datasets`) > > You can load it this way instead: > > ```python > from datasets import load_dataset > cache_dir = 'path/to/your/cache/directory' > dataset = load_dataset('allenai/c4', data_files={'train': 'en/c4-train....
Latest version 2.16.1, when load dataset error occurs. ValueError: BuilderConfig 'allenai--c4' not found. Available: ['default']
### Describe the bug python script is: ``` from datasets import load_dataset cache_dir = 'path/to/your/cache/directory' dataset = load_dataset('allenai/c4','allenai--c4', data_files={'train': 'en/c4-train.00000-of-01024.json.gz'}, split='train', use_auth_token=False, cache_dir=cache_dir) ``` the script su...
57
Latest version 2.16.1, when load dataset error occurs. ValueError: BuilderConfig 'allenai--c4' not found. Available: ['default'] ### Describe the bug python script is: ``` from datasets import load_dataset cache_dir = 'path/to/your/cache/directory' dataset = load_dataset('allenai/c4','allenai--c4', data_files=...
[ -0.5519501566886902, 0.45018383860588074, -0.026410909369587898, 0.3139788806438446, 0.1435113549232483, 0.11366935074329376, 0.30293846130371094, 0.3482833206653595, 0.18301022052764893, 0.20341356098651886, 0.2682672142982483, 0.6109302043914795, -0.1382366269826889, -0.18821798264980316...
https://github.com/huggingface/datasets/issues/6558
You can add ```python from PIL import ImageFile ImageFile.LOAD_TRUNCATED_IMAGES = True ``` after the imports to be able to read truncated images.
OSError: image file is truncated (1 bytes not processed) #28323
### Describe the bug ``` --------------------------------------------------------------------------- OSError Traceback (most recent call last) Cell In[24], line 28 23 return example 25 # Filter the dataset 26 # filtered_dataset = dataset.filter(contains_number...
22
OSError: image file is truncated (1 bytes not processed) #28323 ### Describe the bug ``` --------------------------------------------------------------------------- OSError Traceback (most recent call last) Cell In[24], line 28 23 return example 25 # Filter the da...
[ -0.3276703357696533, -0.31357821822166443, -0.26014402508735657, 0.10543324053287506, 0.24383752048015594, 0.03358208388090134, 0.23733654618263245, 0.46688112616539, 0.08176564425230026, 0.19741684198379517, 0.054364994168281555, 0.06374165415763855, -0.06677894294261932, 0.07968498766422...
https://github.com/huggingface/datasets/issues/6554
I don't think this bug is a thing ? Do you have some code that leads to this issue ?
Parquet exports are used even if revision is passed
We should not used Parquet exports if `revision` is passed. I think this is a regression.
20
Parquet exports are used even if revision is passed We should not used Parquet exports if `revision` is passed. I think this is a regression. I don't think this bug is a thing ? Do you have some code that leads to this issue ?
[ 0.00247994065284729, -0.33710503578186035, -0.11629632860422134, 0.24757957458496094, 0.059504974633455276, -0.4397578537464142, 0.18885020911693573, 0.13291427493095398, -0.1428072452545166, 0.35665690898895264, 0.3908851742744446, 0.49041202664375305, 0.31143370270729065, 0.0302693918347...
https://github.com/huggingface/datasets/issues/6552
This bug comes from the `huggingface_hub` library, see: https://github.com/huggingface/huggingface_hub/issues/1952 A fix is provided at https://github.com/huggingface/huggingface_hub/pull/1953. Feel free to install `huggingface_hub` from this PR, or wait for it to be merged and the new version of `huggingface_hub` t...
Loading a dataset from Google Colab hangs at "Resolving data files".
### Describe the bug Hello, I'm trying to load a dataset from Google Colab but the process hangs at `Resolving data files`: ![image](https://github.com/huggingface/datasets/assets/99779/7175ad85-e571-46ed-9f87-92653985777d) It is happening when the `_get_origin_metadata` definition is invoked: ```python d...
39
Loading a dataset from Google Colab hangs at "Resolving data files". ### Describe the bug Hello, I'm trying to load a dataset from Google Colab but the process hangs at `Resolving data files`: ![image](https://github.com/huggingface/datasets/assets/99779/7175ad85-e571-46ed-9f87-92653985777d) It is happening...
[ -0.19339069724082947, -0.11780481040477753, 0.030356813222169876, 0.40891021490097046, 0.09424534440040588, -0.0761006623506546, 0.26371243596076965, 0.04111425578594208, 0.033734604716300964, 0.24775028228759766, -0.11173553764820099, 0.4329826235771179, 0.03821532055735588, -0.0427216887...
https://github.com/huggingface/datasets/issues/6549
Maybe we can add a helper message like `Maybe try again using "hf://path/without/resolve"` if the path contains `/resolve/` ? e.g. ``` FileNotFoundError: Unable to find 'hf://datasets/HuggingFaceTB/eval_data/resolve/main/eval_data_context_and_answers.json' It looks like you used parts of the URL of the file fro...
Loading from hf hub with clearer error message
### Feature request Shouldn't this kinda work ? ``` Dataset.from_json("hf://datasets/HuggingFaceTB/eval_data/resolve/main/eval_data_context_and_answers.json") ``` I got an error ``` File ~/miniconda3/envs/datatrove/lib/python3.10/site-packages/datasets/data_files.py:380, in resolve_pattern(pattern, base_path, al...
86
Loading from hf hub with clearer error message ### Feature request Shouldn't this kinda work ? ``` Dataset.from_json("hf://datasets/HuggingFaceTB/eval_data/resolve/main/eval_data_context_and_answers.json") ``` I got an error ``` File ~/miniconda3/envs/datatrove/lib/python3.10/site-packages/datasets/data_files....
[ 0.2066141963005066, -0.25089403986930847, 0.036134351044893265, 0.32898765802383423, 0.2908268868923187, -0.08821091055870056, 0.07552556693553925, 0.5636864900588989, 0.05773176997900009, 0.033721961081027985, -0.17770984768867493, 0.16017961502075195, 0.1350436508655548, 0.17397338151931...
https://github.com/huggingface/datasets/issues/6548
It looks like a transient DNS issue. It should work fine now if you try again. There is no parameter in load_dataset to skip failed downloads. In your case it would have skipped every single subsequent download until the DNS issue was resolved anyway.
Skip if a dataset has issues
### Describe the bug Hello everyone, I'm using **load_datasets** from **huggingface** to download the datasets and I'm facing an issue, the download starts but it reaches some state and then fails with the following error: Couldn't reach https://huggingface.co/datasets/wikimedia/wikipedia/resolve/4cb9b0d719291f1a10...
44
Skip if a dataset has issues ### Describe the bug Hello everyone, I'm using **load_datasets** from **huggingface** to download the datasets and I'm facing an issue, the download starts but it reaches some state and then fails with the following error: Couldn't reach https://huggingface.co/datasets/wikimedia/wikip...
[ -0.13599760830402374, -0.5338322520256042, 0.022321894764900208, 0.3548619747161865, 0.3999386727809906, 0.1268061101436615, -0.14067818224430084, 0.225484237074852, 0.08940277248620987, 0.32972007989883423, 0.442242830991745, 0.13631854951381683, 0.05722906067967415, -0.06314912438392639,...
https://github.com/huggingface/datasets/issues/6542
Hi ! We now recommend using the `wikimedia/wikipedia` dataset, can you try loading this one instead ? ```python wiki_dataset = load_dataset("wikimedia/wikipedia", "20231101.en") ```
Datasets : wikipedia 20220301.en error
### Describe the bug When I used load_dataset to download this data set, the following error occurred. The main problem was that the target data did not exist. ### Steps to reproduce the bug 1.I tried downloading directly. ```python wiki_dataset = load_dataset("wikipedia", "20220301.en") ``` An exception occurre...
23
Datasets : wikipedia 20220301.en error ### Describe the bug When I used load_dataset to download this data set, the following error occurred. The main problem was that the target data did not exist. ### Steps to reproduce the bug 1.I tried downloading directly. ```python wiki_dataset = load_dataset("wikipedia",...
[ -0.018569722771644592, 0.21296685934066772, -0.01566591113805771, 0.33132097125053406, 0.11043636500835419, 0.19831033051013947, 0.29937219619750977, 0.35208413004875183, 0.21328392624855042, 0.07991998642683029, 0.12571120262145996, 0.11823579668998718, 0.016085289418697357, -0.0669093132...
https://github.com/huggingface/datasets/issues/6542
This bug has been fixed in `2.16.1` thanks to https://github.com/huggingface/datasets/pull/6544, feel free to update `datasets` and re-run your code :) ``` pip install -U datasets ```
Datasets : wikipedia 20220301.en error
### Describe the bug When I used load_dataset to download this data set, the following error occurred. The main problem was that the target data did not exist. ### Steps to reproduce the bug 1.I tried downloading directly. ```python wiki_dataset = load_dataset("wikipedia", "20220301.en") ``` An exception occurre...
26
Datasets : wikipedia 20220301.en error ### Describe the bug When I used load_dataset to download this data set, the following error occurred. The main problem was that the target data did not exist. ### Steps to reproduce the bug 1.I tried downloading directly. ```python wiki_dataset = load_dataset("wikipedia",...
[ -0.08289985358715057, 0.20689281821250916, -0.015040706843137741, 0.33663633465766907, 0.11554808169603348, 0.1788325011730194, 0.28936833143234253, 0.349710077047348, 0.2294771373271942, 0.09866957366466522, 0.10366464406251907, 0.15296542644500732, 0.018877964466810226, -0.00778213143348...
https://github.com/huggingface/datasets/issues/6541
This is a problem with your environment. You should be able to fix it by upgrading `numpy` based on [this](https://github.com/numpy/numpy/issues/23570) issue.
Dataset not loading successfully.
### Describe the bug When I run down the below code shows this error: AttributeError: module 'numpy' has no attribute '_no_nep50_warning' I also added this issue in transformers library please check out: [link](https://github.com/huggingface/transformers/issues/28099) ### Steps to reproduce the bug ## Reproduction ...
21
Dataset not loading successfully. ### Describe the bug When I run down the below code shows this error: AttributeError: module 'numpy' has no attribute '_no_nep50_warning' I also added this issue in transformers library please check out: [link](https://github.com/huggingface/transformers/issues/28099) ### Steps to...
[ -0.23438847064971924, -0.3350293040275574, 0.06285360455513, 0.4334274232387543, 0.47662264108657837, -0.10106334835290909, 0.43459445238113403, 0.1095186322927475, 0.12539657950401306, 0.29031631350517273, -0.3058722913265228, 0.24882371723651886, -0.28929731249809265, -0.0038884356617927...
https://github.com/huggingface/datasets/issues/6541
Then, this shouldn't throw an error on your machine: ```python import numpy numpy._no_nep50_warning ``` If it does, run `python -m pip install numpy` to ensure the correct `pip` is used for the package installation.
Dataset not loading successfully.
### Describe the bug When I run down the below code shows this error: AttributeError: module 'numpy' has no attribute '_no_nep50_warning' I also added this issue in transformers library please check out: [link](https://github.com/huggingface/transformers/issues/28099) ### Steps to reproduce the bug ## Reproduction ...
34
Dataset not loading successfully. ### Describe the bug When I run down the below code shows this error: AttributeError: module 'numpy' has no attribute '_no_nep50_warning' I also added this issue in transformers library please check out: [link](https://github.com/huggingface/transformers/issues/28099) ### Steps to...
[ -0.23438847064971924, -0.3350293040275574, 0.06285360455513, 0.4334274232387543, 0.47662264108657837, -0.10106334835290909, 0.43459445238113403, 0.1095186322927475, 0.12539657950401306, 0.29031631350517273, -0.3058722913265228, 0.24882371723651886, -0.28929731249809265, -0.0038884356617927...
https://github.com/huggingface/datasets/issues/6541
Your suggestion to run `python -m pip install numpy` proved to be successful, and my issue has been resolved. I am grateful for your assistance, @mariosasko
Dataset not loading successfully.
### Describe the bug When I run down the below code shows this error: AttributeError: module 'numpy' has no attribute '_no_nep50_warning' I also added this issue in transformers library please check out: [link](https://github.com/huggingface/transformers/issues/28099) ### Steps to reproduce the bug ## Reproduction ...
26
Dataset not loading successfully. ### Describe the bug When I run down the below code shows this error: AttributeError: module 'numpy' has no attribute '_no_nep50_warning' I also added this issue in transformers library please check out: [link](https://github.com/huggingface/transformers/issues/28099) ### Steps to...
[ -0.23438847064971924, -0.3350293040275574, 0.06285360455513, 0.4334274232387543, 0.47662264108657837, -0.10106334835290909, 0.43459445238113403, 0.1095186322927475, 0.12539657950401306, 0.29031631350517273, -0.3058722913265228, 0.24882371723651886, -0.28929731249809265, -0.0038884356617927...
https://github.com/huggingface/datasets/issues/6540
Concatenating datasets doesn't create any indices mapping - so flattening indices is not needed (unless you shuffle the dataset). Can you share the snippet of code you are using to merge your datasets and save them to disk ?
Extreme inefficiency for `save_to_disk` when merging datasets
### Describe the bug Hi, I tried to merge in total 22M sequences of data, where each sequence is of maximum length 2000. I found that merging these datasets and then `save_to_disk` is extremely slow because of flattening the indices. Wondering if you have any suggestions or guidance on this. Thank you very much! ###...
39
Extreme inefficiency for `save_to_disk` when merging datasets ### Describe the bug Hi, I tried to merge in total 22M sequences of data, where each sequence is of maximum length 2000. I found that merging these datasets and then `save_to_disk` is extremely slow because of flattening the indices. Wondering if you have...
[ -0.1920756697654724, -0.4081778824329376, 0.06094733253121376, 0.3518158197402954, -0.021621473133563995, 0.42556124925613403, -0.05339167267084122, 0.3826778531074524, -0.1905861794948578, -0.006987802684307098, 0.09002067148685455, 0.11778099834918976, -0.2164076864719391, 0.049476526677...
https://github.com/huggingface/datasets/issues/6538
Hi ! Are you sure you have `datasets` 2.16 ? I just checked and on 2.16 I can run `from datasets.arrow_writer import SchemaInferenceError` without error
ImportError: cannot import name 'SchemaInferenceError' from 'datasets.arrow_writer' (/opt/conda/lib/python3.10/site-packages/datasets/arrow_writer.py)
### Describe the bug While importing from packages getting the error Code: ``` import os import torch from datasets import load_dataset, Dataset from transformers import ( AutoModelForCausalLM, AutoTokenizer, BitsAndBytesConfig, HfArgumentParser, TrainingArguments, pipeline, ...
25
ImportError: cannot import name 'SchemaInferenceError' from 'datasets.arrow_writer' (/opt/conda/lib/python3.10/site-packages/datasets/arrow_writer.py) ### Describe the bug While importing from packages getting the error Code: ``` import os import torch from datasets import load_dataset, Dataset from transf...
[ -0.256902813911438, -0.06662493944168091, -0.05580161139369011, 0.6121095418930054, 0.2748967707157135, 0.06340034306049347, 0.24404215812683105, 0.24720130860805511, -0.019955575466156006, 0.11987867951393127, -0.13812042772769928, 0.38939008116722107, -0.1408829241991043, 0.1007615476846...
https://github.com/huggingface/datasets/issues/6538
I have the same issue - using with datasets version 2.16.1. Also this is on a kaggle notebook - other people with the same issue also seem to be having it on kaggle?
ImportError: cannot import name 'SchemaInferenceError' from 'datasets.arrow_writer' (/opt/conda/lib/python3.10/site-packages/datasets/arrow_writer.py)
### Describe the bug While importing from packages getting the error Code: ``` import os import torch from datasets import load_dataset, Dataset from transformers import ( AutoModelForCausalLM, AutoTokenizer, BitsAndBytesConfig, HfArgumentParser, TrainingArguments, pipeline, ...
33
ImportError: cannot import name 'SchemaInferenceError' from 'datasets.arrow_writer' (/opt/conda/lib/python3.10/site-packages/datasets/arrow_writer.py) ### Describe the bug While importing from packages getting the error Code: ``` import os import torch from datasets import load_dataset, Dataset from transf...
[ -0.256902813911438, -0.06662493944168091, -0.05580161139369011, 0.6121095418930054, 0.2748967707157135, 0.06340034306049347, 0.24404215812683105, 0.24720130860805511, -0.019955575466156006, 0.11987867951393127, -0.13812042772769928, 0.38939008116722107, -0.1408829241991043, 0.1007615476846...
https://github.com/huggingface/datasets/issues/6538
> Hi ! Are you sure you have `datasets` 2.16 ? I just checked and on 2.16 I can run `from datasets.arrow_writer import SchemaInferenceError` without error Yes, I am sure ``` !pip show datasets Name: datasets Version: 2.16.1 Summary: HuggingFace community-driven open-source library of datasets Home-page: http...
ImportError: cannot import name 'SchemaInferenceError' from 'datasets.arrow_writer' (/opt/conda/lib/python3.10/site-packages/datasets/arrow_writer.py)
### Describe the bug While importing from packages getting the error Code: ``` import os import torch from datasets import load_dataset, Dataset from transformers import ( AutoModelForCausalLM, AutoTokenizer, BitsAndBytesConfig, HfArgumentParser, TrainingArguments, pipeline, ...
76
ImportError: cannot import name 'SchemaInferenceError' from 'datasets.arrow_writer' (/opt/conda/lib/python3.10/site-packages/datasets/arrow_writer.py) ### Describe the bug While importing from packages getting the error Code: ``` import os import torch from datasets import load_dataset, Dataset from transf...
[ -0.256902813911438, -0.06662493944168091, -0.05580161139369011, 0.6121095418930054, 0.2748967707157135, 0.06340034306049347, 0.24404215812683105, 0.24720130860805511, -0.019955575466156006, 0.11987867951393127, -0.13812042772769928, 0.38939008116722107, -0.1408829241991043, 0.1007615476846...
https://github.com/huggingface/datasets/issues/6538
> I have the same issue - using with datasets version 2.16.1. Also this is on a kaggle notebook - other people with the same issue also seem to be having it on kaggle? Don't know about other people. But I am having this issue whose solution I can't find anywhere. And this issue still persists.
ImportError: cannot import name 'SchemaInferenceError' from 'datasets.arrow_writer' (/opt/conda/lib/python3.10/site-packages/datasets/arrow_writer.py)
### Describe the bug While importing from packages getting the error Code: ``` import os import torch from datasets import load_dataset, Dataset from transformers import ( AutoModelForCausalLM, AutoTokenizer, BitsAndBytesConfig, HfArgumentParser, TrainingArguments, pipeline, ...
56
ImportError: cannot import name 'SchemaInferenceError' from 'datasets.arrow_writer' (/opt/conda/lib/python3.10/site-packages/datasets/arrow_writer.py) ### Describe the bug While importing from packages getting the error Code: ``` import os import torch from datasets import load_dataset, Dataset from transf...
[ -0.256902813911438, -0.06662493944168091, -0.05580161139369011, 0.6121095418930054, 0.2748967707157135, 0.06340034306049347, 0.24404215812683105, 0.24720130860805511, -0.019955575466156006, 0.11987867951393127, -0.13812042772769928, 0.38939008116722107, -0.1408829241991043, 0.1007615476846...
https://github.com/huggingface/datasets/issues/6538
> I have the same issue now and didn't have this problem around 2 weeks ago. Same here
ImportError: cannot import name 'SchemaInferenceError' from 'datasets.arrow_writer' (/opt/conda/lib/python3.10/site-packages/datasets/arrow_writer.py)
### Describe the bug While importing from packages getting the error Code: ``` import os import torch from datasets import load_dataset, Dataset from transformers import ( AutoModelForCausalLM, AutoTokenizer, BitsAndBytesConfig, HfArgumentParser, TrainingArguments, pipeline, ...
18
ImportError: cannot import name 'SchemaInferenceError' from 'datasets.arrow_writer' (/opt/conda/lib/python3.10/site-packages/datasets/arrow_writer.py) ### Describe the bug While importing from packages getting the error Code: ``` import os import torch from datasets import load_dataset, Dataset from transf...
[ -0.256902813911438, -0.06662493944168091, -0.05580161139369011, 0.6121095418930054, 0.2748967707157135, 0.06340034306049347, 0.24404215812683105, 0.24720130860805511, -0.019955575466156006, 0.11987867951393127, -0.13812042772769928, 0.38939008116722107, -0.1408829241991043, 0.1007615476846...
https://github.com/huggingface/datasets/issues/6538
I was having the same issue but the datasets version was 2.6.1, after I updated it to latest(2.16), error is gone while importing.
ImportError: cannot import name 'SchemaInferenceError' from 'datasets.arrow_writer' (/opt/conda/lib/python3.10/site-packages/datasets/arrow_writer.py)
### Describe the bug While importing from packages getting the error Code: ``` import os import torch from datasets import load_dataset, Dataset from transformers import ( AutoModelForCausalLM, AutoTokenizer, BitsAndBytesConfig, HfArgumentParser, TrainingArguments, pipeline, ...
23
ImportError: cannot import name 'SchemaInferenceError' from 'datasets.arrow_writer' (/opt/conda/lib/python3.10/site-packages/datasets/arrow_writer.py) ### Describe the bug While importing from packages getting the error Code: ``` import os import torch from datasets import load_dataset, Dataset from transf...
[ -0.256902813911438, -0.06662493944168091, -0.05580161139369011, 0.6121095418930054, 0.2748967707157135, 0.06340034306049347, 0.24404215812683105, 0.24720130860805511, -0.019955575466156006, 0.11987867951393127, -0.13812042772769928, 0.38939008116722107, -0.1408829241991043, 0.1007615476846...
https://github.com/huggingface/datasets/issues/6538
> I was having the same issue but the datasets version was 2.6.1, after I updated it to latest(2.16), error is gone while importing. I also have datasets version 2.16, but the error is still there.
ImportError: cannot import name 'SchemaInferenceError' from 'datasets.arrow_writer' (/opt/conda/lib/python3.10/site-packages/datasets/arrow_writer.py)
### Describe the bug While importing from packages getting the error Code: ``` import os import torch from datasets import load_dataset, Dataset from transformers import ( AutoModelForCausalLM, AutoTokenizer, BitsAndBytesConfig, HfArgumentParser, TrainingArguments, pipeline, ...
36
ImportError: cannot import name 'SchemaInferenceError' from 'datasets.arrow_writer' (/opt/conda/lib/python3.10/site-packages/datasets/arrow_writer.py) ### Describe the bug While importing from packages getting the error Code: ``` import os import torch from datasets import load_dataset, Dataset from transf...
[ -0.256902813911438, -0.06662493944168091, -0.05580161139369011, 0.6121095418930054, 0.2748967707157135, 0.06340034306049347, 0.24404215812683105, 0.24720130860805511, -0.019955575466156006, 0.11987867951393127, -0.13812042772769928, 0.38939008116722107, -0.1408829241991043, 0.1007615476846...
https://github.com/huggingface/datasets/issues/6538
> > Can you try re-installing `datasets` ? > > I tried re-installing. Still getting the same error. In kaggle I used: - `%pip install -U datasets` and then restarted runtime and then everything works fine.
ImportError: cannot import name 'SchemaInferenceError' from 'datasets.arrow_writer' (/opt/conda/lib/python3.10/site-packages/datasets/arrow_writer.py)
### Describe the bug While importing from packages getting the error Code: ``` import os import torch from datasets import load_dataset, Dataset from transformers import ( AutoModelForCausalLM, AutoTokenizer, BitsAndBytesConfig, HfArgumentParser, TrainingArguments, pipeline, ...
36
ImportError: cannot import name 'SchemaInferenceError' from 'datasets.arrow_writer' (/opt/conda/lib/python3.10/site-packages/datasets/arrow_writer.py) ### Describe the bug While importing from packages getting the error Code: ``` import os import torch from datasets import load_dataset, Dataset from transf...
[ -0.256902813911438, -0.06662493944168091, -0.05580161139369011, 0.6121095418930054, 0.2748967707157135, 0.06340034306049347, 0.24404215812683105, 0.24720130860805511, -0.019955575466156006, 0.11987867951393127, -0.13812042772769928, 0.38939008116722107, -0.1408829241991043, 0.1007615476846...
https://github.com/huggingface/datasets/issues/6538
> > > Can you try re-installing `datasets` ? > > > > > > I tried re-installing. Still getting the same error. > > In kaggle I used: > > * `%pip install -U datasets` > and then restarted runtime and then everything works fine. Yes, this is working. When I restart the runtime after installing packages, i...
ImportError: cannot import name 'SchemaInferenceError' from 'datasets.arrow_writer' (/opt/conda/lib/python3.10/site-packages/datasets/arrow_writer.py)
### Describe the bug While importing from packages getting the error Code: ``` import os import torch from datasets import load_dataset, Dataset from transformers import ( AutoModelForCausalLM, AutoTokenizer, BitsAndBytesConfig, HfArgumentParser, TrainingArguments, pipeline, ...
78
ImportError: cannot import name 'SchemaInferenceError' from 'datasets.arrow_writer' (/opt/conda/lib/python3.10/site-packages/datasets/arrow_writer.py) ### Describe the bug While importing from packages getting the error Code: ``` import os import torch from datasets import load_dataset, Dataset from transf...
[ -0.256902813911438, -0.06662493944168091, -0.05580161139369011, 0.6121095418930054, 0.2748967707157135, 0.06340034306049347, 0.24404215812683105, 0.24720130860805511, -0.019955575466156006, 0.11987867951393127, -0.13812042772769928, 0.38939008116722107, -0.1408829241991043, 0.1007615476846...
https://github.com/huggingface/datasets/issues/6538
> > > > Can you try re-installing `datasets` ? > > > > > > > > > I tried re-installing. Still getting the same error. > > > > > > In kaggle I used: > > > > * `%pip install -U datasets` > > and then restarted runtime and then everything works fine. > > Yes, this is working. When I restart the runtime ...
ImportError: cannot import name 'SchemaInferenceError' from 'datasets.arrow_writer' (/opt/conda/lib/python3.10/site-packages/datasets/arrow_writer.py)
### Describe the bug While importing from packages getting the error Code: ``` import os import torch from datasets import load_dataset, Dataset from transformers import ( AutoModelForCausalLM, AutoTokenizer, BitsAndBytesConfig, HfArgumentParser, TrainingArguments, pipeline, ...
98
ImportError: cannot import name 'SchemaInferenceError' from 'datasets.arrow_writer' (/opt/conda/lib/python3.10/site-packages/datasets/arrow_writer.py) ### Describe the bug While importing from packages getting the error Code: ``` import os import torch from datasets import load_dataset, Dataset from transf...
[ -0.256902813911438, -0.06662493944168091, -0.05580161139369011, 0.6121095418930054, 0.2748967707157135, 0.06340034306049347, 0.24404215812683105, 0.24720130860805511, -0.019955575466156006, 0.11987867951393127, -0.13812042772769928, 0.38939008116722107, -0.1408829241991043, 0.1007615476846...
https://github.com/huggingface/datasets/issues/6538
> > > > > Can you try re-installing `datasets` ? > > > > > > > > > > > > I tried re-installing. Still getting the same error. > > > > > > > > > In kaggle I used: > > > > > > * `%pip install -U datasets` > > > and then restarted runtime and then everything works fine. > > > > > > Yes, this is workin...
ImportError: cannot import name 'SchemaInferenceError' from 'datasets.arrow_writer' (/opt/conda/lib/python3.10/site-packages/datasets/arrow_writer.py)
### Describe the bug While importing from packages getting the error Code: ``` import os import torch from datasets import load_dataset, Dataset from transformers import ( AutoModelForCausalLM, AutoTokenizer, BitsAndBytesConfig, HfArgumentParser, TrainingArguments, pipeline, ...
157
ImportError: cannot import name 'SchemaInferenceError' from 'datasets.arrow_writer' (/opt/conda/lib/python3.10/site-packages/datasets/arrow_writer.py) ### Describe the bug While importing from packages getting the error Code: ``` import os import torch from datasets import load_dataset, Dataset from transf...
[ -0.256902813911438, -0.06662493944168091, -0.05580161139369011, 0.6121095418930054, 0.2748967707157135, 0.06340034306049347, 0.24404215812683105, 0.24720130860805511, -0.019955575466156006, 0.11987867951393127, -0.13812042772769928, 0.38939008116722107, -0.1408829241991043, 0.1007615476846...
https://github.com/huggingface/datasets/issues/6538
Closing this issue as it is not related to the datasets library; rather, it's linked to platform-related issues.
ImportError: cannot import name 'SchemaInferenceError' from 'datasets.arrow_writer' (/opt/conda/lib/python3.10/site-packages/datasets/arrow_writer.py)
### Describe the bug While importing from packages getting the error Code: ``` import os import torch from datasets import load_dataset, Dataset from transformers import ( AutoModelForCausalLM, AutoTokenizer, BitsAndBytesConfig, HfArgumentParser, TrainingArguments, pipeline, ...
18
ImportError: cannot import name 'SchemaInferenceError' from 'datasets.arrow_writer' (/opt/conda/lib/python3.10/site-packages/datasets/arrow_writer.py) ### Describe the bug While importing from packages getting the error Code: ``` import os import torch from datasets import load_dataset, Dataset from transf...
[ -0.256902813911438, -0.06662493944168091, -0.05580161139369011, 0.6121095418930054, 0.2748967707157135, 0.06340034306049347, 0.24404215812683105, 0.24720130860805511, -0.019955575466156006, 0.11987867951393127, -0.13812042772769928, 0.38939008116722107, -0.1408829241991043, 0.1007615476846...
https://github.com/huggingface/datasets/issues/6537
Conceptually, we can use xarray to load the netCDF file, then xarray -> pandas -> pyarrow.
Adding support for netCDF (*.nc) files
### Feature request netCDF (*.nc) is a file format for storing multidimensional scientific data, which is used by packages like `xarray` (labelled multi-dimensional arrays in Python). It would be nice to have native support for netCDF in `datasets`. ### Motivation When uploading *.nc files onto Huggingface Hub throu...
16
Adding support for netCDF (*.nc) files ### Feature request netCDF (*.nc) is a file format for storing multidimensional scientific data, which is used by packages like `xarray` (labelled multi-dimensional arrays in Python). It would be nice to have native support for netCDF in `datasets`. ### Motivation When upload...
[ -0.38227832317352295, -0.14507780969142914, -0.011451058089733124, 0.0008691772818565369, -0.07568608224391937, 0.011976070702075958, -0.2338274121284485, 0.40313783288002014, -0.12558530271053314, 0.2797532379627228, -0.4577072858810425, 0.04672252759337425, -0.45220503211021423, 0.532621...
https://github.com/huggingface/datasets/issues/6537
I'd still need to verify that such a conversion would be lossless, especially for multi-dimensional data.
Adding support for netCDF (*.nc) files
### Feature request netCDF (*.nc) is a file format for storing multidimensional scientific data, which is used by packages like `xarray` (labelled multi-dimensional arrays in Python). It would be nice to have native support for netCDF in `datasets`. ### Motivation When uploading *.nc files onto Huggingface Hub throu...
16
Adding support for netCDF (*.nc) files ### Feature request netCDF (*.nc) is a file format for storing multidimensional scientific data, which is used by packages like `xarray` (labelled multi-dimensional arrays in Python). It would be nice to have native support for netCDF in `datasets`. ### Motivation When upload...
[ -0.42485103011131287, -0.14924445748329163, -0.01656416431069374, -0.053587038069963455, -0.08428335934877396, 0.03142045438289642, -0.2771828770637512, 0.3843265771865845, -0.10386538505554199, 0.3364311754703522, -0.42337337136268616, 0.005743648856878281, -0.5121448636054993, 0.57444715...
https://github.com/huggingface/datasets/issues/6536
Hi ! Thanks for reporting This is a bug in 2.16.0 for some datasets when `cache_dir` is a relative path. I opened https://github.com/huggingface/datasets/pull/6543 to fix this
datasets.load_dataset raises FileNotFoundError for datasets==2.16.0
### Describe the bug Seems `datasets.load_dataset` raises FileNotFoundError for some hub datasets with the latest `datasets==2.16.0` ### Steps to reproduce the bug For example `pip install datasets==2.16.0` then ```python import datasets datasets.load_dataset("wentingzhao/anthropic-hh-first-prompt", cache_di...
26
datasets.load_dataset raises FileNotFoundError for datasets==2.16.0 ### Describe the bug Seems `datasets.load_dataset` raises FileNotFoundError for some hub datasets with the latest `datasets==2.16.0` ### Steps to reproduce the bug For example `pip install datasets==2.16.0` then ```python import datasets d...
[ -0.7334121465682983, -0.07373972982168198, 0.06582644581794739, 0.4895383417606354, 0.32654696702957153, 0.035552725195884705, 0.39158129692077637, 0.41822516918182373, 0.18511217832565308, 0.0822291374206543, -0.021300405263900757, 0.38214045763015747, -0.23672914505004883, -0.12617044150...
https://github.com/huggingface/datasets/issues/6535
This is surely the same issue as https://discuss.huggingface.co/t/indexerror-invalid-key-16-is-out-of-bounds-for-size-0/14298/25 that comes from the `transformers` `Trainer`. You should add `remove_unused_columns=False` to `TrainingArguments` Also check your logs: the `Trainer` should log the length of your dataset ...
IndexError: Invalid key: 47682 is out of bounds for size 0 while using PEFT
### Describe the bug I am trying to fine-tune the t5 model on the paraphrasing task. While running the same code without- model = get_peft_model(model, config) the model trains without any issues. However, using the model returned from get_peft_model raises the following error due to datasets- IndexError: Inv...
41
IndexError: Invalid key: 47682 is out of bounds for size 0 while using PEFT ### Describe the bug I am trying to fine-tune the t5 model on the paraphrasing task. While running the same code without- model = get_peft_model(model, config) the model trains without any issues. However, using the model returned from...
[ -0.21287056803703308, -0.5975072383880615, -0.021511007100343704, 0.5354565382003784, 0.28637468814849854, -0.0014908388257026672, 0.16571849584579468, 0.513259768486023, 0.18716344237327576, 0.06414628028869629, -0.05213195085525513, -0.19202271103858948, 0.00895208865404129, 0.4490537345...
https://github.com/huggingface/datasets/issues/6532
You can simply use a python dict as index: ```python >>> from datasets import load_dataset >>> ds = load_dataset("BeIR/dbpedia-entity", "corpus", split="corpus") >>> index = {key: idx for idx, key in enumerate(ds["_id"])} >>> ds[index["<dbpedia:Pikachu>"]] {'_id': '<dbpedia:Pikachu>', 'title': 'Pikachu', 't...
[Feature request] Indexing datasets by a customly-defined id field to enable random access dataset items via the id
### Feature request Some datasets may contain an id-like field, for example the `id` field in [wikimedia/wikipedia](https://huggingface.co/datasets/wikimedia/wikipedia) and the `_id` field in [BeIR/dbpedia-entity](https://huggingface.co/datasets/BeIR/dbpedia-entity). HF datasets support efficient random access via r...
87
[Feature request] Indexing datasets by a customly-defined id field to enable random access dataset items via the id ### Feature request Some datasets may contain an id-like field, for example the `id` field in [wikimedia/wikipedia](https://huggingface.co/datasets/wikimedia/wikipedia) and the `_id` field in [BeIR/d...
[ -0.18381522595882416, 0.18622073531150818, -0.08400578796863556, 0.07549107074737549, -0.17324230074882507, 0.13344812393188477, 0.4271538257598877, 0.2126123309135437, 0.41409701108932495, 0.08223025500774384, -0.1450173705816269, 0.4513753652572632, -0.13390713930130005, 0.13514842092990...
https://github.com/huggingface/datasets/issues/6532
Thanks for your reply. Yes, I can do that, but it is time-consuming to do that every time I launch the program (some datasets are extremely big). HF Datasets has a nice feature to support instant data loading and efficient random access via row ids. I'm curious if this beneficial feature could be further extended to cu...
[Feature request] Indexing datasets by a customly-defined id field to enable random access dataset items via the id
### Feature request Some datasets may contain an id-like field, for example the `id` field in [wikimedia/wikipedia](https://huggingface.co/datasets/wikimedia/wikipedia) and the `_id` field in [BeIR/dbpedia-entity](https://huggingface.co/datasets/BeIR/dbpedia-entity). HF datasets support efficient random access via r...
59
[Feature request] Indexing datasets by a customly-defined id field to enable random access dataset items via the id ### Feature request Some datasets may contain an id-like field, for example the `id` field in [wikimedia/wikipedia](https://huggingface.co/datasets/wikimedia/wikipedia) and the `_id` field in [BeIR/d...
[ -0.22591283917427063, 0.24693582952022552, -0.0808679610490799, 0.09924040734767914, -0.2066027671098709, 0.0893334224820137, 0.42412909865379333, 0.1967884749174118, 0.39071765542030334, 0.07601094245910645, -0.1207163855433464, 0.4350595772266388, -0.10443057864904404, 0.1416603475809097...
https://github.com/huggingface/datasets/issues/6530
I solved it with `train_dataset.with_format(None)` But then faced some more issues (which i later solved too). Huggingface does not seem to care, so I do. Here is an updated training script which saves a pre-processed (mapped) dataset to your local directory if you specify `--save_precomputed_data_dir=DIR_NAME`. Th...
Impossible to save a mapped dataset to disk
### Describe the bug I want to play around with different hyperparameters when training but don't want to re-map my dataset with 3 million samples each time for tens of hours when I [fully fine-tune SDXL](https://github.com/huggingface/diffusers/blob/main/examples/text_to_image/train_text_to_image_sdxl.py). After...
140
Impossible to save a mapped dataset to disk ### Describe the bug I want to play around with different hyperparameters when training but don't want to re-map my dataset with 3 million samples each time for tens of hours when I [fully fine-tune SDXL](https://github.com/huggingface/diffusers/blob/main/examples/text_t...
[ -0.25928619503974915, -0.19739599525928497, 0.07933446764945984, 0.14602430164813995, 0.6141032576560974, 0.16964267194271088, 0.29752078652381897, 0.24852529168128967, 0.1337583214044571, 0.26897141337394714, -0.04194672778248787, 0.41905495524406433, -0.3757166862487793, 0.10039237886667...
https://github.com/huggingface/datasets/issues/6529
This feature has been proposed for a long time. I'm looking forward to the implementation. On clusters `streaming=True` is not an option since we do not have Internet on compute nodes. See: https://github.com/huggingface/datasets/discussions/1896#discussioncomment-2359593
Impossible to only download a test split
I've spent a significant amount of time trying to locate the split object inside my _split_generators() custom function. Then after diving [in the code](https://github.com/huggingface/datasets/blob/5ff3670c18ed34fa8ddfa70a9aa403ae6cc9ad54/src/datasets/load.py#L2558) I realized that `download_and_prepare` is executed b...
33
Impossible to only download a test split I've spent a significant amount of time trying to locate the split object inside my _split_generators() custom function. Then after diving [in the code](https://github.com/huggingface/datasets/blob/5ff3670c18ed34fa8ddfa70a9aa403ae6cc9ad54/src/datasets/load.py#L2558) I realize...
[ -0.6835596561431885, -0.05650244653224945, -0.03289652615785599, 0.080228790640831, -0.04810931161046028, -0.08031348884105682, 0.21359404921531677, 0.40247952938079834, 0.051799848675727844, 0.33564817905426025, -0.35279375314712524, 0.023413434624671936, 0.07212822139263153, 0.5051767826...
https://github.com/huggingface/datasets/issues/6524
Hello @FelixLabelle, As you can see in the Community tab of the corresponding dataset, it is a known issue: https://huggingface.co/datasets/EleutherAI/pile/discussions/15 The data has been taken down due to reported copyright infringement. Feel free to continue the discussion there.
Streaming the Pile: Missing Files
### Describe the bug The pile does not stream, a "File not Found error" is returned. It looks like the Pile's files have been moved. ### Steps to reproduce the bug To reproduce run the following code: ``` from datasets import load_dataset dataset = load_dataset('EleutherAI/pile', 'en', split='train', streamin...
38
Streaming the Pile: Missing Files ### Describe the bug The pile does not stream, a "File not Found error" is returned. It looks like the Pile's files have been moved. ### Steps to reproduce the bug To reproduce run the following code: ``` from datasets import load_dataset dataset = load_dataset('EleutherAI/...
[ -0.010011270642280579, -0.0032875090837478638, 0.022553563117980957, 0.16257640719413757, 0.1053948923945427, 0.04117190092802048, 0.24604432284832, 0.34792575240135193, 0.0357474759221077, 0.07959195971488953, 0.07477795332670212, 0.17447461187839508, -0.15698198974132538, 0.0474701300263...
https://github.com/huggingface/datasets/issues/6521
After investigation, I think the issue was introduced by the use of the Parquet export: - #6448 I am proposing a fix. CC: @lhoestq
The order of the splits is not preserved
We had a regression and the order of the splits is not preserved. They are alphabetically sorted, instead of preserving original "train", "validation", "test" order. Check: In branch "main" ```python In [9]: dataset = load_dataset("adversarial_qa", '"adversarialQA") In [10]: dataset Out[10]: DatasetDict({ ...
24
The order of the splits is not preserved We had a regression and the order of the splits is not preserved. They are alphabetically sorted, instead of preserving original "train", "validation", "test" order. Check: In branch "main" ```python In [9]: dataset = load_dataset("adversarial_qa", '"adversarialQA") In...
[ 0.06278710812330246, -0.09385667741298676, -0.06140150874853134, 0.17789337038993835, 0.11431249976158142, -0.1434212028980255, 0.1864895522594452, 0.014443803578615189, 0.056878626346588135, 0.07198087126016617, 0.036952368915081024, 0.29690518975257874, 0.035352230072021484, 0.1410665214...
https://github.com/huggingface/datasets/issues/6506
As this is a specific issue of the "glue" dataset, I have transferred it to the dataset Discussion page: https://huggingface.co/datasets/glue/discussions/15 Let's continue the discussion there!
Incorrect test set labels for RTE and CoLA datasets via load_dataset
### Describe the bug The test set labels for the RTE and CoLA datasets when loading via datasets load_dataset are all -1. Edit: It appears this is also the case for every other dataset except for MRPC (stsb, sst2, qqp, mnli (both matched and mismatched), qnli, wnli, ax). Is this intended behavior to safeguard the t...
25
Incorrect test set labels for RTE and CoLA datasets via load_dataset ### Describe the bug The test set labels for the RTE and CoLA datasets when loading via datasets load_dataset are all -1. Edit: It appears this is also the case for every other dataset except for MRPC (stsb, sst2, qqp, mnli (both matched and mis...
[ -0.11327069252729416, -0.12450703233480453, -0.012101970613002777, 0.332303524017334, 0.026471160352230072, 0.14775629341602325, 0.48433735966682434, 0.11779926717281342, 0.2665036916732788, 0.07214049249887466, -0.2263564169406891, 0.3953258991241455, -0.02414749562740326, 0.4550586044788...
https://github.com/huggingface/datasets/issues/6505
I ran into the same problem when I used a server cluster (Slurm system managed) that couldn't load any of the huggingface datasets or models, but it worked on my laptop. I suspected some system configuration-related problem, but I had no idea. My problems are consistent with [issue #2618](https://github.com/huggingfa...
Got stuck when I trying to load a dataset
### Describe the bug Hello, everyone. I met a problem when I am trying to load a data file using load_dataset method on a Debian 10 system. The data file is not very large, only 1.63MB with 600 records. Here is my code: from datasets import load_dataset dataset = load_dataset('json', data_files='mypath/oaast_r...
59
Got stuck when I trying to load a dataset ### Describe the bug Hello, everyone. I met a problem when I am trying to load a data file using load_dataset method on a Debian 10 system. The data file is not very large, only 1.63MB with 600 records. Here is my code: from datasets import load_dataset dataset = loa...
[ -0.14756226539611816, -0.29957014322280884, -0.06535666435956955, 0.3141990900039673, 0.3323673605918884, 0.08560338616371155, 0.42914897203445435, 0.07434725016355515, 0.19302590191364288, 0.11410264670848846, -0.14673006534576416, 0.46443262696266174, 0.2579033672809601, -0.0864412710070...
https://github.com/huggingface/datasets/issues/6505
> I ran into the same problem when I used a server cluster (Slurm system managed) that couldn't load any of the huggingface datasets or models, but it worked on my laptop. I suspected some system configuration-related problem, but I had no idea. My problems are consistent with [issue #2618](https://github.com/huggingfa...
Got stuck when I trying to load a dataset
### Describe the bug Hello, everyone. I met a problem when I am trying to load a data file using load_dataset method on a Debian 10 system. The data file is not very large, only 1.63MB with 600 records. Here is my code: from datasets import load_dataset dataset = load_dataset('json', data_files='mypath/oaast_r...
89
Got stuck when I trying to load a dataset ### Describe the bug Hello, everyone. I met a problem when I am trying to load a data file using load_dataset method on a Debian 10 system. The data file is not very large, only 1.63MB with 600 records. Here is my code: from datasets import load_dataset dataset = loa...
[ -0.14756226539611816, -0.29957014322280884, -0.06535666435956955, 0.3141990900039673, 0.3323673605918884, 0.08560338616371155, 0.42914897203445435, 0.07434725016355515, 0.19302590191364288, 0.11410264670848846, -0.14673006534576416, 0.46443262696266174, 0.2579033672809601, -0.0864412710070...
https://github.com/huggingface/datasets/issues/6505
I am having the same issue on a computing cluster but this works on my laptop as well. I instead have this error: `/home/.conda/envs/py10/lib/python3.10/site-packages/filelock/_unix.py", line 43, in _acquire fcntl.flock(fd, fcntl.LOCK_EX | fcntl.LOCK_NB) OSError: [Errno 5] Input/output error` the load_datase...
Got stuck when I trying to load a dataset
### Describe the bug Hello, everyone. I met a problem when I am trying to load a data file using load_dataset method on a Debian 10 system. The data file is not very large, only 1.63MB with 600 records. Here is my code: from datasets import load_dataset dataset = load_dataset('json', data_files='mypath/oaast_r...
58
Got stuck when I trying to load a dataset ### Describe the bug Hello, everyone. I met a problem when I am trying to load a data file using load_dataset method on a Debian 10 system. The data file is not very large, only 1.63MB with 600 records. Here is my code: from datasets import load_dataset dataset = loa...
[ -0.14756226539611816, -0.29957014322280884, -0.06535666435956955, 0.3141990900039673, 0.3323673605918884, 0.08560338616371155, 0.42914897203445435, 0.07434725016355515, 0.19302590191364288, 0.11410264670848846, -0.14673006534576416, 0.46443262696266174, 0.2579033672809601, -0.0864412710070...
https://github.com/huggingface/datasets/issues/6496
I transferred from datasets-server, since the issue is more about `datasets` and the integration with `huggingface_hub`.
Error when writing a dataset to HF Hub: A commit has happened since. Please refresh and try again.
**Describe the bug** Getting a `412 Client Error: Precondition Failed` when trying to write a dataset to the HF hub. ``` huggingface_hub.utils._errors.HfHubHTTPError: 412 Client Error: Precondition Failed for url: https://huggingface.co/api/datasets/GLorr/test-dask/commit/main (Request ID: Root=1-657ae26f-3bd92b...
16
Error when writing a dataset to HF Hub: A commit has happened since. Please refresh and try again. **Describe the bug** Getting a `412 Client Error: Precondition Failed` when trying to write a dataset to the HF hub. ``` huggingface_hub.utils._errors.HfHubHTTPError: 412 Client Error: Precondition Failed for url...
[ -0.04964565485715866, -0.4572036862373352, 0.11370263248682022, 0.05369827151298523, 0.1025749146938324, -0.17956051230430603, -0.017055967822670937, 0.25558993220329285, -0.07975412160158157, 0.08024415373802185, -0.24189510941505432, -0.2631165385246277, 0.17090658843517303, 0.2232543230...
https://github.com/huggingface/datasets/issues/6490
Also, once the README.md exists in the python environment it is used when loading another dataset in the same format (e.g. json) since it always resolves the path to the same directory. The consequence here is any other dataset won't load because of infos mismatch. To reproduce this aspect: 1. Do a `load_datasets(.....
`load_dataset(...,save_infos=True)` not working without loading script
### Describe the bug It seems that saving a dataset infos back into the card file is not working for datasets without a loading script. After tracking the problem a bit it looks like saving the infos uses `Builder.get_imported_module_dir()` as its destination directory. Internally this is a call to `inspect.getfil...
81
`load_dataset(...,save_infos=True)` not working without loading script ### Describe the bug It seems that saving a dataset infos back into the card file is not working for datasets without a loading script. After tracking the problem a bit it looks like saving the infos uses `Builder.get_imported_module_dir()` as...
[ -0.4919634759426117, 0.22427958250045776, 0.13527773320674896, 0.4852668046951294, 0.47176915407180786, 0.11189945042133331, 0.3482663333415985, 0.15804512798786163, 0.22043925523757935, 0.08041888475418091, 0.14106851816177368, 0.44441938400268555, -0.011690443381667137, 0.195881903171539...
https://github.com/huggingface/datasets/issues/6488
I'm getting a similar issue even though I've already downloaded the dataset πŸ˜… ``` huggingface_hub.utils._errors.HfHubHTTPError: 429 Client Error: Too Many Requests for url: https://huggingface.co/api/datasets/HuggingFaceM4/WebSight ```
429 Client Error
Hello, I was downloading the following dataset and after 20% of data was downloaded, I started getting error 429. It is not resolved since a few days. How should I resolve it? Thanks Dataset: https://huggingface.co/datasets/cerebras/SlimPajama-627B Error: `requests.exceptions.HTTPError: 429 Client Error: Too M...
25
429 Client Error Hello, I was downloading the following dataset and after 20% of data was downloaded, I started getting error 429. It is not resolved since a few days. How should I resolve it? Thanks Dataset: https://huggingface.co/datasets/cerebras/SlimPajama-627B Error: `requests.exceptions.HTTPError: 429 ...
[ 0.022115163505077362, -0.29604384303092957, -0.005823560059070587, 0.3850882947444916, 0.11384467035531998, -0.001700758934020996, -0.05616357922554016, 0.3764071464538574, -0.20437362790107727, 0.0015221387147903442, -0.19259780645370483, -0.2489210069179535, 0.17384164035320282, -0.04846...
https://github.com/huggingface/datasets/issues/6483
You can get the expected result by fixing typos in the snippet :) ```python from datasets import load_dataset # load LS in streaming mode dataset = load_dataset("librispeech_asr", "clean", split="validation", streaming=True) # check original features dataset_features = dataset.features.keys() print("Original...
Iterable Dataset: rename column clashes with remove column
### Describe the bug Suppose I have a two iterable datasets, one with the features: * `{"audio", "text", "column_a"}` And the other with the features: * `{"audio", "sentence", "column_b"}` I want to combine both datasets using `interleave_datasets`, which requires me to unify the column names. I would typic...
75
Iterable Dataset: rename column clashes with remove column ### Describe the bug Suppose I have a two iterable datasets, one with the features: * `{"audio", "text", "column_a"}` And the other with the features: * `{"audio", "sentence", "column_b"}` I want to combine both datasets using `interleave_datasets...
[ 0.1475961059331894, 0.20639097690582275, 0.013281568884849548, 0.04070236533880234, 0.21933552622795105, 0.34094297885894775, 0.47970396280288696, 0.06645643711090088, -0.09174070507287979, 0.19398123025894165, -0.1917896866798401, 0.30293503403663635, 0.1443752646446228, -0.03058172762393...
https://github.com/huggingface/datasets/issues/6483
Fixed code: ```python from datasets import load_dataset # load LS in streaming mode dataset = load_dataset("librispeech_asr", "clean", split="validation", streaming=True) # check original features dataset_features = dataset.features.keys() print("Original features: ", dataset_features) #Β rename "text" -...
Iterable Dataset: rename column clashes with remove column
### Describe the bug Suppose I have a two iterable datasets, one with the features: * `{"audio", "text", "column_a"}` And the other with the features: * `{"audio", "sentence", "column_b"}` I want to combine both datasets using `interleave_datasets`, which requires me to unify the column names. I would typic...
67
Iterable Dataset: rename column clashes with remove column ### Describe the bug Suppose I have a two iterable datasets, one with the features: * `{"audio", "text", "column_a"}` And the other with the features: * `{"audio", "sentence", "column_b"}` I want to combine both datasets using `interleave_datasets...
[ 0.1475961059331894, 0.20639097690582275, 0.013281568884849548, 0.04070236533880234, 0.21933552622795105, 0.34094297885894775, 0.47970396280288696, 0.06645643711090088, -0.09174070507287979, 0.19398123025894165, -0.1917896866798401, 0.30293503403663635, 0.1443752646446228, -0.03058172762393...
https://github.com/huggingface/datasets/issues/6484
Hello @kenfus, this is meant to be possible to do yes. Let me ping @lhoestq or @mariosasko from the `datasets` team (`huggingface_hub` is only the underlying library to download files from the Hub but here it looks more like a `datasets` problem).
[Feature Request] Dataset versioning
**Is your feature request related to a problem? Please describe.** I am working on a project, where I would like to test different preprocessing methods for my ML-data. Thus, I would like to work a lot with revisions and compare them. Currently, I was not able to make it work with the revision keyword because it was n...
42
[Feature Request] Dataset versioning **Is your feature request related to a problem? Please describe.** I am working on a project, where I would like to test different preprocessing methods for my ML-data. Thus, I would like to work a lot with revisions and compare them. Currently, I was not able to make it work wit...
[ -0.272832989692688, -0.24783575534820557, -0.0026081427931785583, -0.07376807183027267, -0.22250692546367645, -0.015726059675216675, -0.05707693472504616, 0.34137123823165894, -0.17344817519187927, -0.03136356174945831, -0.1561669409275055, 0.23501451313495636, -0.11500318348407745, 0.3179...
https://github.com/huggingface/datasets/issues/6478
You can create a `pandas` DataFrame following [this](https://lakefs.io/data-version-control/dvc-using-python/) tutorial, and then convert this DataFrame to a `Dataset` with `datasets.Dataset.from_pandas`. For larger datasets (to memory map them), you can use `Dataset.from_generator` with a generator function that reads...
How to load data from lakefs
My dataset is stored on the company's lakefs server. How can I write code to load the dataset? It would be great if I could provide code examples or provide some references
40
How to load data from lakefs My dataset is stored on the company's lakefs server. How can I write code to load the dataset? It would be great if I could provide code examples or provide some references You can create a `pandas` DataFrame following [this](https://lakefs.io/data-version-control/dvc-using-python/) t...
[ 0.02660747617483139, -0.14837107062339783, -0.030263463035225868, 0.4822250306606293, -0.07394763827323914, 0.07551722228527069, 0.08751226961612701, 0.09296516329050064, 0.4828351140022278, -0.19326013326644897, -0.27409911155700684, 0.4478367865085602, -0.0011054053902626038, 0.332675278...
https://github.com/huggingface/datasets/issues/6478
@mariosasko hello, This can achieve and https://huggingface.co/datasets Does the same effect apply to the dataset? For example, downloading while using
How to load data from lakefs
My dataset is stored on the company's lakefs server. How can I write code to load the dataset? It would be great if I could provide code examples or provide some references
20
How to load data from lakefs My dataset is stored on the company's lakefs server. How can I write code to load the dataset? It would be great if I could provide code examples or provide some references @mariosasko hello, This can achieve and https://huggingface.co/datasets Does the same effect apply to the data...
[ -0.09849373996257782, -0.08236931264400482, -0.0687437504529953, 0.47989392280578613, -0.06994661688804626, 0.12994585931301117, -0.031182490289211273, 0.050826296210289, 0.6055763959884644, -0.12141266465187073, -0.35074567794799805, 0.2735244929790497, 0.12763109803199768, 0.358489453792...
https://github.com/huggingface/datasets/issues/6475
~~You will see this error if the cache dir filepath contains relative `..` paths. Use `os.path.realpath(_CACHE_DIR)` before passing it to the `load_dataset` function.~~
laion2B-en failed to load on Windows with PrefetchVirtualMemory failed
### Describe the bug I have downloaded laion2B-en, and I'm receiving the following error trying to load it: ``` Resolving data files: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 128/128 [00:00<00:00, 1173.79it/s] Traceback (most recent call last): File "D:\Art-Workspace\src\artworkspace\tokeneval\compute_frequencies.py", line 31, in <mo...
23
laion2B-en failed to load on Windows with PrefetchVirtualMemory failed ### Describe the bug I have downloaded laion2B-en, and I'm receiving the following error trying to load it: ``` Resolving data files: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 128/128 [00:00<00:00, 1173.79it/s] Traceback (most recent call last): File "D:\Art-Work...
[ -0.10396334528923035, -0.16585659980773926, -0.06416866183280945, 0.3617136478424072, 0.04591068997979164, -0.18339619040489197, -0.012238699942827225, 0.26278233528137207, -0.2864779233932495, -0.08145007491111755, -0.024002086371183395, 0.08407759666442871, -0.14902670681476593, 0.149506...
https://github.com/huggingface/datasets/issues/6475
Based on the StackOverflow answer, this causes the error to go away: ```diff diff --git a/table.py b/table.py --- a/table.py +++ b/table.py (date 1701824849806) @@ -47,7 +47,7 @@ def _memory_mapped_record_batch_reader_from_file(filename: str) -> pa.RecordBatchStreamReader: - memory_mapped_stream = pa....
laion2B-en failed to load on Windows with PrefetchVirtualMemory failed
### Describe the bug I have downloaded laion2B-en, and I'm receiving the following error trying to load it: ``` Resolving data files: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 128/128 [00:00<00:00, 1173.79it/s] Traceback (most recent call last): File "D:\Art-Workspace\src\artworkspace\tokeneval\compute_frequencies.py", line 31, in <mo...
56
laion2B-en failed to load on Windows with PrefetchVirtualMemory failed ### Describe the bug I have downloaded laion2B-en, and I'm receiving the following error trying to load it: ``` Resolving data files: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 128/128 [00:00<00:00, 1173.79it/s] Traceback (most recent call last): File "D:\Art-Work...
[ -0.10396334528923035, -0.16585659980773926, -0.06416866183280945, 0.3617136478424072, 0.04591068997979164, -0.18339619040489197, -0.012238699942827225, 0.26278233528137207, -0.2864779233932495, -0.08145007491111755, -0.024002086371183395, 0.08407759666442871, -0.14902670681476593, 0.149506...
https://github.com/huggingface/datasets/issues/6475
I don't really comprehend what it is that `datasets` gave me when it downloaded the laion2B-en dataset, because nothing can seemingly read these 1024 .arrow files it is retrieving. Not `polars`, not `pyarrow`, it's not an `ipc` file, it's not a `parquet` file...
laion2B-en failed to load on Windows with PrefetchVirtualMemory failed
### Describe the bug I have downloaded laion2B-en, and I'm receiving the following error trying to load it: ``` Resolving data files: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 128/128 [00:00<00:00, 1173.79it/s] Traceback (most recent call last): File "D:\Art-Workspace\src\artworkspace\tokeneval\compute_frequencies.py", line 31, in <mo...
43
laion2B-en failed to load on Windows with PrefetchVirtualMemory failed ### Describe the bug I have downloaded laion2B-en, and I'm receiving the following error trying to load it: ``` Resolving data files: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 128/128 [00:00<00:00, 1173.79it/s] Traceback (most recent call last): File "D:\Art-Work...
[ -0.10396334528923035, -0.16585659980773926, -0.06416866183280945, 0.3617136478424072, 0.04591068997979164, -0.18339619040489197, -0.012238699942827225, 0.26278233528137207, -0.2864779233932495, -0.08145007491111755, -0.024002086371183395, 0.08407759666442871, -0.14902670681476593, 0.149506...
https://github.com/huggingface/datasets/issues/6475
Hi! Instead of generating one (potentially large) Arrow file, we shard the generated data into 500 MB shards because memory-mapping large Arrow files can be problematic on some systems. Maybe deleting the dataset's cache and increasing the shard size (controlled with the `datasets.config.MAX_SHARD_SIZE` variable; e...
laion2B-en failed to load on Windows with PrefetchVirtualMemory failed
### Describe the bug I have downloaded laion2B-en, and I'm receiving the following error trying to load it: ``` Resolving data files: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 128/128 [00:00<00:00, 1173.79it/s] Traceback (most recent call last): File "D:\Art-Workspace\src\artworkspace\tokeneval\compute_frequencies.py", line 31, in <mo...
130
laion2B-en failed to load on Windows with PrefetchVirtualMemory failed ### Describe the bug I have downloaded laion2B-en, and I'm receiving the following error trying to load it: ``` Resolving data files: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 128/128 [00:00<00:00, 1173.79it/s] Traceback (most recent call last): File "D:\Art-Work...
[ -0.10396334528923035, -0.16585659980773926, -0.06416866183280945, 0.3617136478424072, 0.04591068997979164, -0.18339619040489197, -0.012238699942827225, 0.26278233528137207, -0.2864779233932495, -0.08145007491111755, -0.024002086371183395, 0.08407759666442871, -0.14902670681476593, 0.149506...
https://github.com/huggingface/datasets/issues/6475
Hmm, I have a feeling this works fine on Linux, and is a real bug for however `datasets` is doing the sharding on Windows. I will follow up, but I think this is a real bug.
laion2B-en failed to load on Windows with PrefetchVirtualMemory failed
### Describe the bug I have downloaded laion2B-en, and I'm receiving the following error trying to load it: ``` Resolving data files: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 128/128 [00:00<00:00, 1173.79it/s] Traceback (most recent call last): File "D:\Art-Workspace\src\artworkspace\tokeneval\compute_frequencies.py", line 31, in <mo...
36
laion2B-en failed to load on Windows with PrefetchVirtualMemory failed ### Describe the bug I have downloaded laion2B-en, and I'm receiving the following error trying to load it: ``` Resolving data files: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 128/128 [00:00<00:00, 1173.79it/s] Traceback (most recent call last): File "D:\Art-Work...
[ -0.10396334528923035, -0.16585659980773926, -0.06416866183280945, 0.3617136478424072, 0.04591068997979164, -0.18339619040489197, -0.012238699942827225, 0.26278233528137207, -0.2864779233932495, -0.08145007491111755, -0.024002086371183395, 0.08407759666442871, -0.14902670681476593, 0.149506...
https://github.com/huggingface/datasets/issues/6467
We will publish it soon (we usually do it in intervals of 1-2 months, so probably next week)
New version release request
### Feature request Hi! I am using `datasets` in library `xtuner` and am highly interested in the features introduced since v2.15.0. To avoid installation from source in our pypi wheels, we are eagerly waiting for the new release. So, Does your team have a new release plan for v2.15.1 and could you please share ...
18
New version release request ### Feature request Hi! I am using `datasets` in library `xtuner` and am highly interested in the features introduced since v2.15.0. To avoid installation from source in our pypi wheels, we are eagerly waiting for the new release. So, Does your team have a new release plan for v2.15...
[ -0.44579827785491943, 0.29189836978912354, -0.0761459693312645, -0.06915633380413055, 0.025643251836299896, -0.18820524215698242, -0.005974992178380489, 0.2926267385482788, -0.3280101716518402, 0.20499148964881897, 0.21357974410057068, 0.24996279180049896, -0.37874549627304077, 0.436384916...
https://github.com/huggingface/datasets/issues/6466
Friendly bump, I would be happy to work on this issue once I get the go-ahead from the dev team.
Can't align optional features of struct
### Describe the bug Hello! I'm currently experiencing an issue where I can't concatenate datasets if an inner field of a Feature is Optional. I have a column named `speaker`, and this holds some information about a speaker. ```python @dataclass class Speaker: name: str email: Optional[str] ``` ...
20
Can't align optional features of struct ### Describe the bug Hello! I'm currently experiencing an issue where I can't concatenate datasets if an inner field of a Feature is Optional. I have a column named `speaker`, and this holds some information about a speaker. ```python @dataclass class Speaker: ...
[ 0.05887938290834427, -0.3263806700706482, 0.1339399814605713, 0.23355942964553833, 0.21374757587909698, 0.14838311076164246, 0.1363564133644104, 0.030548855662345886, -0.12449514865875244, -0.018588125705718994, 0.08468344062566757, 0.022713661193847656, 0.10085195302963257, 0.246346056461...
https://github.com/huggingface/datasets/issues/6466
Thanks for the PR! I'm struggling with this as well and would love to see this PR merged. My case is slightly different, with keys completely missing rather than being `None`: ``` ds = Dataset.from_dict({'speaker': [{'name': 'Ben'}]}) ds2 = Dataset.from_dict({'speaker': [{'name': 'Fred', 'email': 'abc@aol.com'}...
Can't align optional features of struct
### Describe the bug Hello! I'm currently experiencing an issue where I can't concatenate datasets if an inner field of a Feature is Optional. I have a column named `speaker`, and this holds some information about a speaker. ```python @dataclass class Speaker: name: str email: Optional[str] ``` ...
98
Can't align optional features of struct ### Describe the bug Hello! I'm currently experiencing an issue where I can't concatenate datasets if an inner field of a Feature is Optional. I have a column named `speaker`, and this holds some information about a speaker. ```python @dataclass class Speaker: ...
[ 0.05887938290834427, -0.3263806700706482, 0.1339399814605713, 0.23355942964553833, 0.21374757587909698, 0.14838311076164246, 0.1363564133644104, 0.030548855662345886, -0.12449514865875244, -0.018588125705718994, 0.08468344062566757, 0.022713661193847656, 0.10085195302963257, 0.246346056461...
https://github.com/huggingface/datasets/issues/6460
Hi @serenalotreck, We use Apache Arrow `pyarrow` to read jsonlines and it throws an error when trying to load your data files: ```python In [1]: import pyarrow as pa In [2]: data = pa.json.read_json("train.jsonl") --------------------------------------------------------------------------- ArrowInvalid ...
jsonlines files don't load with `load_dataset`
### Describe the bug While [the docs](https://huggingface.co/docs/datasets/upload_dataset#upload-dataset) seem to state that `.jsonl` is a supported extension for `datasets`, loading the dataset results in a `JSONDecodeError`. ### Steps to reproduce the bug Code: ``` from datasets import load_dataset dset = load_...
171
jsonlines files don't load with `load_dataset` ### Describe the bug While [the docs](https://huggingface.co/docs/datasets/upload_dataset#upload-dataset) seem to state that `.jsonl` is a supported extension for `datasets`, loading the dataset results in a `JSONDecodeError`. ### Steps to reproduce the bug Code: ```...
[ -0.2301376760005951, 0.12285783886909485, -0.048369962722063065, 0.44046539068222046, 0.25598087906837463, 0.032171137630939484, 0.19228532910346985, 0.19167578220367432, 0.2536109685897827, -0.10145919024944305, 0.21212723851203918, 0.3888658881187439, -0.1920369565486908, 0.3685352802276...
https://github.com/huggingface/datasets/issues/6460
Hi @albertvillanova, Thanks for the explanation! To the best of my knowledge, arrays in a json [can contain multiple data types](https://docs.actian.com/ingres/11.2/index.html#page/SQLRef/Data_Types.htm), and I'm able to read these files with the `jsonlines` package. Is the requirement for arrays to only have one d...
jsonlines files don't load with `load_dataset`
### Describe the bug While [the docs](https://huggingface.co/docs/datasets/upload_dataset#upload-dataset) seem to state that `.jsonl` is a supported extension for `datasets`, loading the dataset results in a `JSONDecodeError`. ### Steps to reproduce the bug Code: ``` from datasets import load_dataset dset = load_...
85
jsonlines files don't load with `load_dataset` ### Describe the bug While [the docs](https://huggingface.co/docs/datasets/upload_dataset#upload-dataset) seem to state that `.jsonl` is a supported extension for `datasets`, loading the dataset results in a `JSONDecodeError`. ### Steps to reproduce the bug Code: ```...
[ -0.2301376760005951, 0.12285783886909485, -0.048369962722063065, 0.44046539068222046, 0.25598087906837463, 0.032171137630939484, 0.19228532910346985, 0.19167578220367432, 0.2536109685897827, -0.10145919024944305, 0.21212723851203918, 0.3888658881187439, -0.1920369565486908, 0.3685352802276...
https://github.com/huggingface/datasets/issues/6460
Hi again @serenalotreck, Yes, it is specific to PyArrow: as far as I know, Arrow does not support arrays with multiple data types. As this is related specifically to your dataset structure (and not the `datasets` library), I have created a dedicated issue in your dataset page: https://huggingface.co/datasets/slot...
jsonlines files don't load with `load_dataset`
### Describe the bug While [the docs](https://huggingface.co/docs/datasets/upload_dataset#upload-dataset) seem to state that `.jsonl` is a supported extension for `datasets`, loading the dataset results in a `JSONDecodeError`. ### Steps to reproduce the bug Code: ``` from datasets import load_dataset dset = load_...
54
jsonlines files don't load with `load_dataset` ### Describe the bug While [the docs](https://huggingface.co/docs/datasets/upload_dataset#upload-dataset) seem to state that `.jsonl` is a supported extension for `datasets`, loading the dataset results in a `JSONDecodeError`. ### Steps to reproduce the bug Code: ```...
[ -0.2301376760005951, 0.12285783886909485, -0.048369962722063065, 0.44046539068222046, 0.25598087906837463, 0.032171137630939484, 0.19228532910346985, 0.19167578220367432, 0.2536109685897827, -0.10145919024944305, 0.21212723851203918, 0.3888658881187439, -0.1920369565486908, 0.3685352802276...
https://github.com/huggingface/datasets/issues/6460
> Hi again @serenalotreck, > > Yes, it is specific to PyArrow: as far as I know, Arrow does not support arrays with multiple data types. > > As this is related specifically to your dataset structure (and not the `datasets` library), I have created a dedicated issue in your dataset page: https://huggingface.co/dat...
jsonlines files don't load with `load_dataset`
### Describe the bug While [the docs](https://huggingface.co/docs/datasets/upload_dataset#upload-dataset) seem to state that `.jsonl` is a supported extension for `datasets`, loading the dataset results in a `JSONDecodeError`. ### Steps to reproduce the bug Code: ``` from datasets import load_dataset dset = load_...
116
jsonlines files don't load with `load_dataset` ### Describe the bug While [the docs](https://huggingface.co/docs/datasets/upload_dataset#upload-dataset) seem to state that `.jsonl` is a supported extension for `datasets`, loading the dataset results in a `JSONDecodeError`. ### Steps to reproduce the bug Code: ```...
[ -0.2301376760005951, 0.12285783886909485, -0.048369962722063065, 0.44046539068222046, 0.25598087906837463, 0.032171137630939484, 0.19228532910346985, 0.19167578220367432, 0.2536109685897827, -0.10145919024944305, 0.21212723851203918, 0.3888658881187439, -0.1920369565486908, 0.3685352802276...
https://github.com/huggingface/datasets/issues/6457
> Maybe this should go in datasets directly... anyways you can easily fix this error by updating datasets>=2.15.1.dev0. @lhoestq @mariosasko for what I understand this is a bug fixed in `datasets` already, right? No need to do anything in `huggingface_hub`?
`TypeError`: huggingface_hub.hf_file_system.HfFileSystem.find() got multiple values for keyword argument 'maxdepth'
### Describe the bug Please see https://github.com/huggingface/huggingface_hub/issues/1872 ### Steps to reproduce the bug Please see https://github.com/huggingface/huggingface_hub/issues/1872 ### Expected behavior Please see https://github.com/huggingface/huggingface_hub/issues/1872 ### Environment info Please s...
40
`TypeError`: huggingface_hub.hf_file_system.HfFileSystem.find() got multiple values for keyword argument 'maxdepth' ### Describe the bug Please see https://github.com/huggingface/huggingface_hub/issues/1872 ### Steps to reproduce the bug Please see https://github.com/huggingface/huggingface_hub/issues/1872 ### Ex...
[ -0.03594100847840309, -0.7556374073028564, -0.036174602806568146, 0.3993551731109619, 0.2019290328025818, -0.03666580468416214, -0.06197826936841011, 0.3705710172653198, 0.15028253197669983, 0.22910165786743164, -0.3483662009239197, -0.07723751664161682, -0.17496560513973236, 0.24227726459...
https://github.com/huggingface/datasets/issues/6446
You can use `.align_labels_with_mapping` on the dataset to align the labels with the model config. Regarding the number of labels, only the special `_silence_` label corresponding to noise is missing, which is consistent with the model paper (reports training on 35 labels). You can run a `.filter` to drop it. PS...
Speech Commands v2 dataset doesn't match AST-v2 config
### Describe the bug [According](https://huggingface.co/MIT/ast-finetuned-speech-commands-v2) to `MIT/ast-finetuned-speech-commands-v2`, the model was trained on the Speech Commands v2 dataset. However, while the model config says the model should have 35 class labels, the dataset itself has 36 class labels. Moreover,...
68
Speech Commands v2 dataset doesn't match AST-v2 config ### Describe the bug [According](https://huggingface.co/MIT/ast-finetuned-speech-commands-v2) to `MIT/ast-finetuned-speech-commands-v2`, the model was trained on the Speech Commands v2 dataset. However, while the model config says the model should have 35 class ...
[ -0.18747246265411377, -0.4025111794471741, -0.002486366778612137, 0.5321387052536011, 0.23339498043060303, 0.21374253928661346, 0.22850120067596436, 0.29725173115730286, -0.3711091876029968, -0.07165176421403885, -0.11736663430929184, 0.2715616226196289, -0.22551605105400085, -0.1073668897...
https://github.com/huggingface/datasets/issues/6446
Thanks, will keep that in mind. But I tried running `dataset_aligned = dataset.align_labels_with_mapping(model.config.id2label, 'label')`, and received this error: ``` Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/Users/victor/anaconda3/envs/transformers-v2/lib/python3.9/site-...
Speech Commands v2 dataset doesn't match AST-v2 config
### Describe the bug [According](https://huggingface.co/MIT/ast-finetuned-speech-commands-v2) to `MIT/ast-finetuned-speech-commands-v2`, the model was trained on the Speech Commands v2 dataset. However, while the model config says the model should have 35 class labels, the dataset itself has 36 class labels. Moreover,...
101
Speech Commands v2 dataset doesn't match AST-v2 config ### Describe the bug [According](https://huggingface.co/MIT/ast-finetuned-speech-commands-v2) to `MIT/ast-finetuned-speech-commands-v2`, the model was trained on the Speech Commands v2 dataset. However, while the model config says the model should have 35 class ...
[ -0.18747246265411377, -0.4025111794471741, -0.002486366778612137, 0.5321387052536011, 0.23339498043060303, 0.21374253928661346, 0.22850120067596436, 0.29725173115730286, -0.3711091876029968, -0.07165176421403885, -0.11736663430929184, 0.2715616226196289, -0.22551605105400085, -0.1073668897...
https://github.com/huggingface/datasets/issues/6446
Replacing `model.config.id2label` with `model.config.label2id` should fix the issue. So, the full code to align the labels with the model config is as follows: ```python from datasets import load_dataset from transformers import AutoFeatureExtractor, AutoModelForAudioClassification # extractor = AutoFeatureExt...
Speech Commands v2 dataset doesn't match AST-v2 config
### Describe the bug [According](https://huggingface.co/MIT/ast-finetuned-speech-commands-v2) to `MIT/ast-finetuned-speech-commands-v2`, the model was trained on the Speech Commands v2 dataset. However, while the model config says the model should have 35 class labels, the dataset itself has 36 class labels. Moreover,...
57
Speech Commands v2 dataset doesn't match AST-v2 config ### Describe the bug [According](https://huggingface.co/MIT/ast-finetuned-speech-commands-v2) to `MIT/ast-finetuned-speech-commands-v2`, the model was trained on the Speech Commands v2 dataset. However, while the model config says the model should have 35 class ...
[ -0.18747246265411377, -0.4025111794471741, -0.002486366778612137, 0.5321387052536011, 0.23339498043060303, 0.21374253928661346, 0.22850120067596436, 0.29725173115730286, -0.3711091876029968, -0.07165176421403885, -0.11736663430929184, 0.2715616226196289, -0.22551605105400085, -0.1073668897...
https://github.com/huggingface/datasets/issues/6443
There is a typo in one of the file names - `data/edf.csv` should be renamed to `data/def.csv` πŸ™‚.
Trouble loading files defined in YAML explicitly
Look at https://huggingface.co/datasets/severo/doc-yaml-2 It's a reproduction of the example given in the docs at https://huggingface.co/docs/hub/datasets-manual-configuration ``` You can select multiple files per split using a list of paths: my_dataset_repository/ β”œβ”€β”€ README.md β”œβ”€β”€ data/ β”‚ β”œβ”€β”€ abc.csv ...
18
Trouble loading files defined in YAML explicitly Look at https://huggingface.co/datasets/severo/doc-yaml-2 It's a reproduction of the example given in the docs at https://huggingface.co/docs/hub/datasets-manual-configuration ``` You can select multiple files per split using a list of paths: my_dataset_repos...
[ -0.057436682283878326, -0.3854188323020935, 0.0579201877117157, 0.3542958199977875, 0.27944862842559814, 0.10894966870546341, 0.41074347496032715, 0.30620720982551575, 0.45095789432525635, 0.07631495594978333, -0.2228856384754181, -0.10569396615028381, 0.04646443575620651, 0.23681929707527...
https://github.com/huggingface/datasets/issues/6443
wow, I reviewed it twice to avoid being ashamed like that, but... I didn't notice the typo. --- Besides this: do you think we would be able to improve the error message to make this clearer?
Trouble loading files defined in YAML explicitly
Look at https://huggingface.co/datasets/severo/doc-yaml-2 It's a reproduction of the example given in the docs at https://huggingface.co/docs/hub/datasets-manual-configuration ``` You can select multiple files per split using a list of paths: my_dataset_repository/ β”œβ”€β”€ README.md β”œβ”€β”€ data/ β”‚ β”œβ”€β”€ abc.csv ...
36
Trouble loading files defined in YAML explicitly Look at https://huggingface.co/datasets/severo/doc-yaml-2 It's a reproduction of the example given in the docs at https://huggingface.co/docs/hub/datasets-manual-configuration ``` You can select multiple files per split using a list of paths: my_dataset_repos...
[ -0.057436682283878326, -0.3854188323020935, 0.0579201877117157, 0.3542958199977875, 0.27944862842559814, 0.10894966870546341, 0.41074347496032715, 0.30620720982551575, 0.45095789432525635, 0.07631495594978333, -0.2228856384754181, -0.10569396615028381, 0.04646443575620651, 0.23681929707527...