html_url stringlengths 48 51 | title stringlengths 5 268 | comments stringlengths 70 51.8k | body stringlengths 0 29.8k | comment_length int64 16 1.52k | text stringlengths 164 54.1k | embeddings sequence |
|---|---|---|---|---|---|---|
https://github.com/huggingface/datasets/issues/2945 | Protect master branch | @lhoestq now the 2 are implemented.
Please note that for the the second protection, finally I have chosen to protect the master branch only from **merge commits** (see update comment above), so no need to disable/re-enable the protection on each release (direct commits, different from merge commits, can be pushed to... | After accidental merge commit (91c55355b634d0dc73350a7ddee1a6776dbbdd69) into `datasets` master branch, all commits present in the feature branch were permanently added to `datasets` master branch history, as e.g.:
- 00cc036fea7c7745cfe722360036ed306796a3f2
- 13ae8c98602bbad8197de3b9b425f4c78f582af1
- ...
I propo... | 64 | Protect master branch
After accidental merge commit (91c55355b634d0dc73350a7ddee1a6776dbbdd69) into `datasets` master branch, all commits present in the feature branch were permanently added to `datasets` master branch history, as e.g.:
- 00cc036fea7c7745cfe722360036ed306796a3f2
- 13ae8c98602bbad8197de3b9b425f4c78f... | [
-0.15531975030899048,
-0.10023094713687897,
-0.07032088935375214,
-0.07981614768505096,
-0.104254350066185,
-0.18798762559890747,
0.01040300726890564,
0.27286356687545776,
-0.009848978370428085,
-0.08414009213447571,
0.2926146686077118,
-0.07787517458200455,
-0.13245651125907898,
0.2158902... |
https://github.com/huggingface/datasets/issues/2943 | Backwards compatibility broken for cached datasets that use `.filter()` | Hi ! I guess the caching mechanism should have considered the new `filter` to be different from the old one, and don't use cached results from the old `filter`.
To avoid other users from having this issue we could make the caching differentiate the two, what do you think ? | ## Describe the bug
After upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with
`ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='in... | 50 | Backwards compatibility broken for cached datasets that use `.filter()`
## Describe the bug
After upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with
`ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=No... | [
-0.30499589443206787,
0.12438860535621643,
-0.04655841737985611,
0.23689737915992737,
0.15172843635082245,
-0.0666598454117775,
-0.0069412775337696075,
0.32853978872299194,
0.17373839020729065,
0.02938103675842285,
-0.2313872128725052,
0.3581932783126831,
-0.15033014118671417,
0.2527365386... |
https://github.com/huggingface/datasets/issues/2943 | Backwards compatibility broken for cached datasets that use `.filter()` | If it's easy enough to implement, then yes please 😄 But this issue can be low-priority, since I've only encountered it in a couple of `transformers` CI tests. | ## Describe the bug
After upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with
`ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='in... | 28 | Backwards compatibility broken for cached datasets that use `.filter()`
## Describe the bug
After upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with
`ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=No... | [
-0.30499589443206787,
0.12438860535621643,
-0.04655841737985611,
0.23689737915992737,
0.15172843635082245,
-0.0666598454117775,
-0.0069412775337696075,
0.32853978872299194,
0.17373839020729065,
0.02938103675842285,
-0.2313872128725052,
0.3581932783126831,
-0.15033014118671417,
0.2527365386... |
https://github.com/huggingface/datasets/issues/2943 | Backwards compatibility broken for cached datasets that use `.filter()` | Well it can cause issue with anyone that updates `datasets` and re-run some code that uses filter, so I'm creating a PR | ## Describe the bug
After upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with
`ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='in... | 22 | Backwards compatibility broken for cached datasets that use `.filter()`
## Describe the bug
After upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with
`ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=No... | [
-0.30499589443206787,
0.12438860535621643,
-0.04655841737985611,
0.23689737915992737,
0.15172843635082245,
-0.0666598454117775,
-0.0069412775337696075,
0.32853978872299194,
0.17373839020729065,
0.02938103675842285,
-0.2313872128725052,
0.3581932783126831,
-0.15033014118671417,
0.2527365386... |
https://github.com/huggingface/datasets/issues/2943 | Backwards compatibility broken for cached datasets that use `.filter()` | I just merged a fix, let me know if you're still having this kind of issues :)
We'll do a release soon to make this fix available | ## Describe the bug
After upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with
`ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='in... | 27 | Backwards compatibility broken for cached datasets that use `.filter()`
## Describe the bug
After upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with
`ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=No... | [
-0.30499589443206787,
0.12438860535621643,
-0.04655841737985611,
0.23689737915992737,
0.15172843635082245,
-0.0666598454117775,
-0.0069412775337696075,
0.32853978872299194,
0.17373839020729065,
0.02938103675842285,
-0.2313872128725052,
0.3581932783126831,
-0.15033014118671417,
0.2527365386... |
https://github.com/huggingface/datasets/issues/2937 | load_dataset using default cache on Windows causes PermissionError: [WinError 5] Access is denied | Hi @daqieq, thanks for reporting.
Unfortunately, I was not able to reproduce this bug:
```ipython
In [1]: from datasets import load_dataset
...: ds = load_dataset('wiki_bio')
Downloading: 7.58kB [00:00, 26.3kB/s]
Downloading: 2.71kB [00:00, ?B/s]
Using custom data configuration default
Downloading and prep... | ## Describe the bug
Standard process to download and load the wiki_bio dataset causes PermissionError in Windows 10 and 11.
## Steps to reproduce the bug
```python
from datasets import load_dataset
ds = load_dataset('wiki_bio')
```
## Expected results
It is expected that the dataset downloads without any er... | 109 | load_dataset using default cache on Windows causes PermissionError: [WinError 5] Access is denied
## Describe the bug
Standard process to download and load the wiki_bio dataset causes PermissionError in Windows 10 and 11.
## Steps to reproduce the bug
```python
from datasets import load_dataset
ds = load_datas... | [
-0.22893105447292328,
0.38345402479171753,
0.040345486253499985,
0.2550400197505951,
-0.012111496180295944,
0.26225730776786804,
0.5036689639091492,
0.13295315206050873,
0.3442804515361786,
0.1357182413339615,
-0.10897494852542877,
0.0772843211889267,
0.07315035164356232,
-0.00569866783916... |
https://github.com/huggingface/datasets/issues/2937 | load_dataset using default cache on Windows causes PermissionError: [WinError 5] Access is denied | Thanks @albertvillanova for looking at it! I tried on my personal Windows machine and it downloaded just fine.
Running on my work machine and on a colleague's machine it is consistently hitting this error. It's not a write access issue because the `.incomplete` directory is written just fine. It just won't rename an... | ## Describe the bug
Standard process to download and load the wiki_bio dataset causes PermissionError in Windows 10 and 11.
## Steps to reproduce the bug
```python
from datasets import load_dataset
ds = load_dataset('wiki_bio')
```
## Expected results
It is expected that the dataset downloads without any er... | 194 | load_dataset using default cache on Windows causes PermissionError: [WinError 5] Access is denied
## Describe the bug
Standard process to download and load the wiki_bio dataset causes PermissionError in Windows 10 and 11.
## Steps to reproduce the bug
```python
from datasets import load_dataset
ds = load_datas... | [
-0.22893105447292328,
0.38345402479171753,
0.040345486253499985,
0.2550400197505951,
-0.012111496180295944,
0.26225730776786804,
0.5036689639091492,
0.13295315206050873,
0.3442804515361786,
0.1357182413339615,
-0.10897494852542877,
0.0772843211889267,
0.07315035164356232,
-0.00569866783916... |
https://github.com/huggingface/datasets/issues/2934 | to_tf_dataset keeps a reference to the open data somewhere, causing issues on windows | I did some investigation and, as it seems, the bug stems from [this line](https://github.com/huggingface/datasets/blob/8004d7c3e1d74b29c3e5b0d1660331cd26758363/src/datasets/arrow_dataset.py#L325). The lifecycle of the dataset from the linked line is bound to one of the returned `tf.data.Dataset`. So my (hacky) solution... | To reproduce:
```python
import datasets as ds
import weakref
import gc
d = ds.load_dataset("mnist", split="train")
ref = weakref.ref(d._data.table)
tfd = d.to_tf_dataset("image", batch_size=1, shuffle=False, label_cols="label")
del tfd, d
gc.collect()
assert ref() is None, "Error: there is at least one refe... | 99 | to_tf_dataset keeps a reference to the open data somewhere, causing issues on windows
To reproduce:
```python
import datasets as ds
import weakref
import gc
d = ds.load_dataset("mnist", split="train")
ref = weakref.ref(d._data.table)
tfd = d.to_tf_dataset("image", batch_size=1, shuffle=False, label_cols="lab... | [
0.04569007828831673,
0.33450788259506226,
0.11488880217075348,
0.08045118302106857,
0.2304304838180542,
0.1188877746462822,
0.3988851010799408,
0.2548750042915344,
-0.10991644859313965,
0.26743894815444946,
-0.33538562059402466,
0.4040720462799072,
-0.16448089480400085,
-0.1074621528387069... |
https://github.com/huggingface/datasets/issues/2924 | "File name too long" error for file locks | "Hi, the filename here is less than 255\r\n```python\r\n>>> len(\"_home_garrett_.cache_huggingface_d(...TRUNCATED) | "## Describe the bug\r\n\r\nGetting the following error when calling `load_dataset(\"gar1t/test\")`:(...TRUNCATED) | 39 | "\"File name too long\" error for file locks \n ## Describe the bug\r\n\r\nGetting the following err(...TRUNCATED) | [0.049338411539793015,0.09200134873390198,-0.07149402797222137,0.40184879302978516,0.415167033672332(...TRUNCATED) |
https://github.com/huggingface/datasets/issues/2924 | "File name too long" error for file locks | "Yes, you're right! I need to get you more info here. Either there's something going with the name i(...TRUNCATED) | "## Describe the bug\r\n\r\nGetting the following error when calling `load_dataset(\"gar1t/test\")`:(...TRUNCATED) | 67 | "\"File name too long\" error for file locks \n ## Describe the bug\r\n\r\nGetting the following err(...TRUNCATED) | [0.049338411539793015,0.09200134873390198,-0.07149402797222137,0.40184879302978516,0.415167033672332(...TRUNCATED) |
End of preview. Expand in Data Studio
- Downloads last month
- 2