Dataset Viewer
Duplicate
The dataset viewer is not available for this split.
Cannot load the dataset split (in streaming mode) to extract the first rows.
Error code:   StreamingRowsError
Exception:    EntryNotFoundError
Message:      404 Client Error. (Request ID: Root=1-69d668cf-799a34312791365d5e6cd722;6dca4c39-d8a4-4912-958f-46e981cb936e)

Entry Not Found for url: https://huggingface.co/datasets/wd41/isruc-sleep/resolve/85bae4c4457a20070f018be8b4d952009e99b6ab/sourcedata/braindecode/dataset.zarr/recording_0/data/zarr.json.
Traceback:    Traceback (most recent call last):
                File "/usr/local/lib/python3.12/site-packages/huggingface_hub/utils/_http.py", line 409, in hf_raise_for_status
                  response.raise_for_status()
                File "/usr/local/lib/python3.12/site-packages/requests/models.py", line 1026, in raise_for_status
                  raise HTTPError(http_error_msg, response=self)
              requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/datasets/wd41/isruc-sleep/resolve/85bae4c4457a20070f018be8b4d952009e99b6ab/sourcedata/braindecode/dataset.zarr/recording_0/data/zarr.json
              
              The above exception was the direct cause of the following exception:
              
              Traceback (most recent call last):
                File "/src/services/worker/src/worker/utils.py", line 99, in get_rows_or_raise
                  return get_rows(
                         ^^^^^^^^^
                File "/src/libs/libcommon/src/libcommon/utils.py", line 272, in decorator
                  return func(*args, **kwargs)
                         ^^^^^^^^^^^^^^^^^^^^^
                File "/src/services/worker/src/worker/utils.py", line 77, in get_rows
                  rows_plus_one = list(itertools.islice(ds, rows_max_number + 1))
                                  ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/iterable_dataset.py", line 2690, in __iter__
                  for key, example in ex_iterable:
                                      ^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/iterable_dataset.py", line 2227, in __iter__
                  for key, pa_table in self._iter_arrow():
                                       ^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/iterable_dataset.py", line 2251, in _iter_arrow
                  for key, pa_table in self.ex_iterable._iter_arrow():
                                       ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/iterable_dataset.py", line 494, in _iter_arrow
                  for key, pa_table in iterator:
                                       ^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/iterable_dataset.py", line 384, in _iter_arrow
                  for key, pa_table in self.generate_tables_fn(**gen_kwags):
                                       ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/packaged_modules/json/json.py", line 187, in _generate_tables
                  batch = f.read(self.config.chunksize)
                          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/utils/file_utils.py", line 844, in read_with_retries
                  out = read(*args, **kwargs)
                        ^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/huggingface_hub/hf_file_system.py", line 1015, in read
                  return super().read(length)
                         ^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/fsspec/spec.py", line 1846, in read
                  out = self.cache._fetch(self.loc, self.loc + length)
                        ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/fsspec/caching.py", line 189, in _fetch
                  self.cache = self.fetcher(start, end)  # new block replaces old
                               ^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/huggingface_hub/hf_file_system.py", line 976, in _fetch_range
                  hf_raise_for_status(r)
                File "/usr/local/lib/python3.12/site-packages/huggingface_hub/utils/_http.py", line 420, in hf_raise_for_status
                  raise _format(EntryNotFoundError, message, response) from e
              huggingface_hub.errors.EntryNotFoundError: 404 Client Error. (Request ID: Root=1-69d668cf-799a34312791365d5e6cd722;6dca4c39-d8a4-4912-958f-46e981cb936e)
              
              Entry Not Found for url: https://huggingface.co/datasets/wd41/isruc-sleep/resolve/85bae4c4457a20070f018be8b4d952009e99b6ab/sourcedata/braindecode/dataset.zarr/recording_0/data/zarr.json.

Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.

EEG Dataset

This dataset was created using braindecode, a deep learning library for EEG/MEG/ECoG signals.

Dataset Information

Property Value
Recordings 110
Type Windowed (from Epochs object)
Channels 6
Sampling frequency 200 Hz
Total duration 34 days, 4:31:47
Windows/samples 98,480
Size 1.12 MB
Format zarr

Quick Start

from braindecode.datasets import BaseConcatDataset

# Load from Hugging Face Hub
dataset = BaseConcatDataset.pull_from_hub("username/dataset-name")

# Access a sample
X, y, metainfo = dataset[0]
# X: EEG data [n_channels, n_times]
# y: target label
# metainfo: window indices

Training with PyTorch

from torch.utils.data import DataLoader

loader = DataLoader(dataset, batch_size=32, shuffle=True, num_workers=4)

for X, y, metainfo in loader:
    # X: [batch_size, n_channels, n_times]
    # y: [batch_size]
    pass  # Your training code

BIDS-inspired Structure

This dataset uses a BIDS-inspired organization. Metadata files follow BIDS conventions, while data is stored in Zarr format for efficient deep learning.

BIDS-style metadata:

  • dataset_description.json - Dataset information
  • participants.tsv - Subject metadata
  • *_events.tsv - Trial/window events
  • *_channels.tsv - Channel information
  • *_eeg.json - Recording parameters

Data storage:

  • dataset.zarr/ - Zarr format (optimized for random access)
sourcedata/braindecode/
β”œβ”€β”€ dataset_description.json
β”œβ”€β”€ participants.tsv
β”œβ”€β”€ dataset.zarr/
└── sub-<label>/
    └── eeg/
        β”œβ”€β”€ *_events.tsv
        β”œβ”€β”€ *_channels.tsv
        └── *_eeg.json

Accessing Metadata

# Participants info
if hasattr(dataset, "participants"):
    print(dataset.participants)

# Events for a recording
if hasattr(dataset.datasets[0], "bids_events"):
    print(dataset.datasets[0].bids_events)

# Channel info
if hasattr(dataset.datasets[0], "bids_channels"):
    print(dataset.datasets[0].bids_channels)

Created with braindecode

Downloads last month
4,274