Dataset Viewer
Auto-converted to Parquet Duplicate
html_url
stringlengths
48
51
title
stringlengths
5
268
comments
stringlengths
70
51.8k
body
stringlengths
0
29.8k
comment_length
int64
16
1.52k
text
stringlengths
164
54.1k
embeddings
sequence
https://github.com/huggingface/datasets/issues/2945
Protect master branch
@lhoestq now the 2 are implemented. Please note that for the the second protection, finally I have chosen to protect the master branch only from **merge commits** (see update comment above), so no need to disable/re-enable the protection on each release (direct commits, different from merge commits, can be pushed to...
After accidental merge commit (91c55355b634d0dc73350a7ddee1a6776dbbdd69) into `datasets` master branch, all commits present in the feature branch were permanently added to `datasets` master branch history, as e.g.: - 00cc036fea7c7745cfe722360036ed306796a3f2 - 13ae8c98602bbad8197de3b9b425f4c78f582af1 - ... I propo...
64
Protect master branch After accidental merge commit (91c55355b634d0dc73350a7ddee1a6776dbbdd69) into `datasets` master branch, all commits present in the feature branch were permanently added to `datasets` master branch history, as e.g.: - 00cc036fea7c7745cfe722360036ed306796a3f2 - 13ae8c98602bbad8197de3b9b425f4c78f...
[ -0.15531975030899048, -0.10023094713687897, -0.07032088935375214, -0.07981614768505096, -0.104254350066185, -0.18798762559890747, 0.01040300726890564, 0.27286356687545776, -0.009848978370428085, -0.08414009213447571, 0.2926146686077118, -0.07787517458200455, -0.13245651125907898, 0.2158902...
https://github.com/huggingface/datasets/issues/2943
Backwards compatibility broken for cached datasets that use `.filter()`
Hi ! I guess the caching mechanism should have considered the new `filter` to be different from the old one, and don't use cached results from the old `filter`. To avoid other users from having this issue we could make the caching differentiate the two, what do you think ?
## Describe the bug After upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with `ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='in...
50
Backwards compatibility broken for cached datasets that use `.filter()` ## Describe the bug After upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with `ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=No...
[ -0.30499589443206787, 0.12438860535621643, -0.04655841737985611, 0.23689737915992737, 0.15172843635082245, -0.0666598454117775, -0.0069412775337696075, 0.32853978872299194, 0.17373839020729065, 0.02938103675842285, -0.2313872128725052, 0.3581932783126831, -0.15033014118671417, 0.2527365386...
https://github.com/huggingface/datasets/issues/2943
Backwards compatibility broken for cached datasets that use `.filter()`
If it's easy enough to implement, then yes please 😄 But this issue can be low-priority, since I've only encountered it in a couple of `transformers` CI tests.
## Describe the bug After upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with `ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='in...
28
Backwards compatibility broken for cached datasets that use `.filter()` ## Describe the bug After upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with `ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=No...
[ -0.30499589443206787, 0.12438860535621643, -0.04655841737985611, 0.23689737915992737, 0.15172843635082245, -0.0666598454117775, -0.0069412775337696075, 0.32853978872299194, 0.17373839020729065, 0.02938103675842285, -0.2313872128725052, 0.3581932783126831, -0.15033014118671417, 0.2527365386...
https://github.com/huggingface/datasets/issues/2943
Backwards compatibility broken for cached datasets that use `.filter()`
Well it can cause issue with anyone that updates `datasets` and re-run some code that uses filter, so I'm creating a PR
## Describe the bug After upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with `ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='in...
22
Backwards compatibility broken for cached datasets that use `.filter()` ## Describe the bug After upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with `ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=No...
[ -0.30499589443206787, 0.12438860535621643, -0.04655841737985611, 0.23689737915992737, 0.15172843635082245, -0.0666598454117775, -0.0069412775337696075, 0.32853978872299194, 0.17373839020729065, 0.02938103675842285, -0.2313872128725052, 0.3581932783126831, -0.15033014118671417, 0.2527365386...
https://github.com/huggingface/datasets/issues/2943
Backwards compatibility broken for cached datasets that use `.filter()`
I just merged a fix, let me know if you're still having this kind of issues :) We'll do a release soon to make this fix available
## Describe the bug After upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with `ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='in...
27
Backwards compatibility broken for cached datasets that use `.filter()` ## Describe the bug After upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with `ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=No...
[ -0.30499589443206787, 0.12438860535621643, -0.04655841737985611, 0.23689737915992737, 0.15172843635082245, -0.0666598454117775, -0.0069412775337696075, 0.32853978872299194, 0.17373839020729065, 0.02938103675842285, -0.2313872128725052, 0.3581932783126831, -0.15033014118671417, 0.2527365386...
https://github.com/huggingface/datasets/issues/2937
load_dataset using default cache on Windows causes PermissionError: [WinError 5] Access is denied
Hi @daqieq, thanks for reporting. Unfortunately, I was not able to reproduce this bug: ```ipython In [1]: from datasets import load_dataset ...: ds = load_dataset('wiki_bio') Downloading: 7.58kB [00:00, 26.3kB/s] Downloading: 2.71kB [00:00, ?B/s] Using custom data configuration default Downloading and prep...
## Describe the bug Standard process to download and load the wiki_bio dataset causes PermissionError in Windows 10 and 11. ## Steps to reproduce the bug ```python from datasets import load_dataset ds = load_dataset('wiki_bio') ``` ## Expected results It is expected that the dataset downloads without any er...
109
load_dataset using default cache on Windows causes PermissionError: [WinError 5] Access is denied ## Describe the bug Standard process to download and load the wiki_bio dataset causes PermissionError in Windows 10 and 11. ## Steps to reproduce the bug ```python from datasets import load_dataset ds = load_datas...
[ -0.22893105447292328, 0.38345402479171753, 0.040345486253499985, 0.2550400197505951, -0.012111496180295944, 0.26225730776786804, 0.5036689639091492, 0.13295315206050873, 0.3442804515361786, 0.1357182413339615, -0.10897494852542877, 0.0772843211889267, 0.07315035164356232, -0.00569866783916...
https://github.com/huggingface/datasets/issues/2937
load_dataset using default cache on Windows causes PermissionError: [WinError 5] Access is denied
Thanks @albertvillanova for looking at it! I tried on my personal Windows machine and it downloaded just fine. Running on my work machine and on a colleague's machine it is consistently hitting this error. It's not a write access issue because the `.incomplete` directory is written just fine. It just won't rename an...
## Describe the bug Standard process to download and load the wiki_bio dataset causes PermissionError in Windows 10 and 11. ## Steps to reproduce the bug ```python from datasets import load_dataset ds = load_dataset('wiki_bio') ``` ## Expected results It is expected that the dataset downloads without any er...
194
load_dataset using default cache on Windows causes PermissionError: [WinError 5] Access is denied ## Describe the bug Standard process to download and load the wiki_bio dataset causes PermissionError in Windows 10 and 11. ## Steps to reproduce the bug ```python from datasets import load_dataset ds = load_datas...
[ -0.22893105447292328, 0.38345402479171753, 0.040345486253499985, 0.2550400197505951, -0.012111496180295944, 0.26225730776786804, 0.5036689639091492, 0.13295315206050873, 0.3442804515361786, 0.1357182413339615, -0.10897494852542877, 0.0772843211889267, 0.07315035164356232, -0.00569866783916...
https://github.com/huggingface/datasets/issues/2934
to_tf_dataset keeps a reference to the open data somewhere, causing issues on windows
I did some investigation and, as it seems, the bug stems from [this line](https://github.com/huggingface/datasets/blob/8004d7c3e1d74b29c3e5b0d1660331cd26758363/src/datasets/arrow_dataset.py#L325). The lifecycle of the dataset from the linked line is bound to one of the returned `tf.data.Dataset`. So my (hacky) solution...
To reproduce: ```python import datasets as ds import weakref import gc d = ds.load_dataset("mnist", split="train") ref = weakref.ref(d._data.table) tfd = d.to_tf_dataset("image", batch_size=1, shuffle=False, label_cols="label") del tfd, d gc.collect() assert ref() is None, "Error: there is at least one refe...
99
to_tf_dataset keeps a reference to the open data somewhere, causing issues on windows To reproduce: ```python import datasets as ds import weakref import gc d = ds.load_dataset("mnist", split="train") ref = weakref.ref(d._data.table) tfd = d.to_tf_dataset("image", batch_size=1, shuffle=False, label_cols="lab...
[ 0.04569007828831673, 0.33450788259506226, 0.11488880217075348, 0.08045118302106857, 0.2304304838180542, 0.1188877746462822, 0.3988851010799408, 0.2548750042915344, -0.10991644859313965, 0.26743894815444946, -0.33538562059402466, 0.4040720462799072, -0.16448089480400085, -0.1074621528387069...
https://github.com/huggingface/datasets/issues/2924
"File name too long" error for file locks
"Hi, the filename here is less than 255\r\n```python\r\n>>> len(\"_home_garrett_.cache_huggingface_d(...TRUNCATED)
"## Describe the bug\r\n\r\nGetting the following error when calling `load_dataset(\"gar1t/test\")`:(...TRUNCATED)
39
"\"File name too long\" error for file locks \n ## Describe the bug\r\n\r\nGetting the following err(...TRUNCATED)
[0.049338411539793015,0.09200134873390198,-0.07149402797222137,0.40184879302978516,0.415167033672332(...TRUNCATED)
https://github.com/huggingface/datasets/issues/2924
"File name too long" error for file locks
"Yes, you're right! I need to get you more info here. Either there's something going with the name i(...TRUNCATED)
"## Describe the bug\r\n\r\nGetting the following error when calling `load_dataset(\"gar1t/test\")`:(...TRUNCATED)
67
"\"File name too long\" error for file locks \n ## Describe the bug\r\n\r\nGetting the following err(...TRUNCATED)
[0.049338411539793015,0.09200134873390198,-0.07149402797222137,0.40184879302978516,0.415167033672332(...TRUNCATED)
End of preview. Expand in Data Studio

Dataset Card for "another_dummy_dataset"

More Information needed

Downloads last month
2