Dataset Viewer
The dataset viewer is not available for this split.
Cannot load the dataset split (in streaming mode) to extract the first rows.
Error code: StreamingRowsError
Exception: EntryNotFoundError
Message: 404 Client Error. (Request ID: Root=1-69d668cf-799a34312791365d5e6cd722;6dca4c39-d8a4-4912-958f-46e981cb936e)
Entry Not Found for url: https://huggingface.co/datasets/wd41/isruc-sleep/resolve/85bae4c4457a20070f018be8b4d952009e99b6ab/sourcedata/braindecode/dataset.zarr/recording_0/data/zarr.json.
Traceback: Traceback (most recent call last):
File "/usr/local/lib/python3.12/site-packages/huggingface_hub/utils/_http.py", line 409, in hf_raise_for_status
response.raise_for_status()
File "/usr/local/lib/python3.12/site-packages/requests/models.py", line 1026, in raise_for_status
raise HTTPError(http_error_msg, response=self)
requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/datasets/wd41/isruc-sleep/resolve/85bae4c4457a20070f018be8b4d952009e99b6ab/sourcedata/braindecode/dataset.zarr/recording_0/data/zarr.json
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/src/services/worker/src/worker/utils.py", line 99, in get_rows_or_raise
return get_rows(
^^^^^^^^^
File "/src/libs/libcommon/src/libcommon/utils.py", line 272, in decorator
return func(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^
File "/src/services/worker/src/worker/utils.py", line 77, in get_rows
rows_plus_one = list(itertools.islice(ds, rows_max_number + 1))
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/iterable_dataset.py", line 2690, in __iter__
for key, example in ex_iterable:
^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/iterable_dataset.py", line 2227, in __iter__
for key, pa_table in self._iter_arrow():
^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/iterable_dataset.py", line 2251, in _iter_arrow
for key, pa_table in self.ex_iterable._iter_arrow():
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/iterable_dataset.py", line 494, in _iter_arrow
for key, pa_table in iterator:
^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/iterable_dataset.py", line 384, in _iter_arrow
for key, pa_table in self.generate_tables_fn(**gen_kwags):
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/packaged_modules/json/json.py", line 187, in _generate_tables
batch = f.read(self.config.chunksize)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/utils/file_utils.py", line 844, in read_with_retries
out = read(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/huggingface_hub/hf_file_system.py", line 1015, in read
return super().read(length)
^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/fsspec/spec.py", line 1846, in read
out = self.cache._fetch(self.loc, self.loc + length)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/fsspec/caching.py", line 189, in _fetch
self.cache = self.fetcher(start, end) # new block replaces old
^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/huggingface_hub/hf_file_system.py", line 976, in _fetch_range
hf_raise_for_status(r)
File "/usr/local/lib/python3.12/site-packages/huggingface_hub/utils/_http.py", line 420, in hf_raise_for_status
raise _format(EntryNotFoundError, message, response) from e
huggingface_hub.errors.EntryNotFoundError: 404 Client Error. (Request ID: Root=1-69d668cf-799a34312791365d5e6cd722;6dca4c39-d8a4-4912-958f-46e981cb936e)
Entry Not Found for url: https://huggingface.co/datasets/wd41/isruc-sleep/resolve/85bae4c4457a20070f018be8b4d952009e99b6ab/sourcedata/braindecode/dataset.zarr/recording_0/data/zarr.json.Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.
EEG Dataset
This dataset was created using braindecode, a deep learning library for EEG/MEG/ECoG signals.
Dataset Information
| Property | Value |
|---|---|
| Recordings | 110 |
| Type | Windowed (from Epochs object) |
| Channels | 6 |
| Sampling frequency | 200 Hz |
| Total duration | 34 days, 4:31:47 |
| Windows/samples | 98,480 |
| Size | 1.12 MB |
| Format | zarr |
Quick Start
from braindecode.datasets import BaseConcatDataset
# Load from Hugging Face Hub
dataset = BaseConcatDataset.pull_from_hub("username/dataset-name")
# Access a sample
X, y, metainfo = dataset[0]
# X: EEG data [n_channels, n_times]
# y: target label
# metainfo: window indices
Training with PyTorch
from torch.utils.data import DataLoader
loader = DataLoader(dataset, batch_size=32, shuffle=True, num_workers=4)
for X, y, metainfo in loader:
# X: [batch_size, n_channels, n_times]
# y: [batch_size]
pass # Your training code
BIDS-inspired Structure
This dataset uses a BIDS-inspired organization. Metadata files follow BIDS conventions, while data is stored in Zarr format for efficient deep learning.
BIDS-style metadata:
dataset_description.json- Dataset informationparticipants.tsv- Subject metadata*_events.tsv- Trial/window events*_channels.tsv- Channel information*_eeg.json- Recording parameters
Data storage:
dataset.zarr/- Zarr format (optimized for random access)
sourcedata/braindecode/
βββ dataset_description.json
βββ participants.tsv
βββ dataset.zarr/
βββ sub-<label>/
βββ eeg/
βββ *_events.tsv
βββ *_channels.tsv
βββ *_eeg.json
Accessing Metadata
# Participants info
if hasattr(dataset, "participants"):
print(dataset.participants)
# Events for a recording
if hasattr(dataset.datasets[0], "bids_events"):
print(dataset.datasets[0].bids_events)
# Channel info
if hasattr(dataset.datasets[0], "bids_channels"):
print(dataset.datasets[0].bids_channels)
Created with braindecode
- Downloads last month
- 4,274