Dataset Viewer
Auto-converted to Parquet Duplicate
The dataset viewer is not available for this split.
Job manager crashed while running this job (missing heartbeats).
Error code:   JobManagerCrashedError

Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.

OtoSpeech Turn-Taking

Processed version of the OtoSpeech corpus with per-frame turn-taking labels and Mimi speech codec features. Each row is one full conversation.

Splits

Split Sessions
train 900
val 112
test 113

80/10/10 split, seed=42.

Features

Column Shape dtype Description
session_id str Unique session identifier
dataset str Source corpus name
duration_s float Conversation duration (seconds)
codes_ch0 [T, 8] int Mimi RVQ codes, speaker 0
codes_ch1 [T, 8] int Mimi RVQ codes, speaker 1
mimi_feat_ch0 [T, 512] float Mimi continuous embeddings, speaker 0
mimi_feat_ch1 [T, 512] float Mimi continuous embeddings, speaker 1
vad_ch0 [T] float Voice activity (0/1), speaker 0
vad_ch1 [T] float Voice activity (0/1), speaker 1
eot_ch0 [T] int End-of-Turn label, speaker 0
eot_ch1 [T] int End-of-Turn label, speaker 1
hold_ch0 [T] int Hold (no handover) label, speaker 0
hold_ch1 [T] int Hold (no handover) label, speaker 1
bot_ch0 [T] int Beginning-of-Turn label, speaker 0
bot_ch1 [T] int Beginning-of-Turn label, speaker 1
bc_ch0 [T] int Backchannel label, speaker 0
bc_ch1 [T] int Backchannel label, speaker 1
fvad_ch0 [T, 4] float Fine-grained VAD logits (4 heads), speaker 0
fvad_ch1 [T, 4] float Fine-grained VAD logits (4 heads), speaker 1

Frame rate: 12.5 Hz — 1 frame = 80 ms. Event labels (eot, hold, bot, bc) are sparse binary: 0 everywhere except at event frames.

Splits file

splits.json in the repo root maps every session ID to its split. Useful for reproducing the split or processing the raw audio yourself:

from huggingface_hub import hf_hub_download
import json

path = hf_hub_download("anyreach-ai/dualturn-otospeech-turn-taking", "splits.json", repo_type="dataset")
with open(path) as f:
    splits = json.load(f)

print(splits["split_counts"])
# e.g. {'train': 900, 'val': 112, 'test': 113}

Loading

import numpy as np
from datasets import load_dataset

ds = load_dataset("anyreach-ai/dualturn-otospeech-turn-taking")

session = ds["val"][0]
T = session["num_frames"]

# 2D arrays are stored flat — reshape to recover original shape
codes = np.array(session["codes_ch0"]).reshape(T, 8)      # (T, 8)   int
feats = np.array(session["mimi_feat_ch0"]).reshape(T, 512) # (T, 512) float
fvad  = np.array(session["fvad_ch0"]).reshape(T, 4)        # (T, 4)   float

# 1D arrays — use directly
vad = np.array(session["vad_ch0"])   # (T,) float
eot = np.array(session["eot_ch0"])   # (T,) int

PyTorch windowed loader

import numpy as np
import torch
from torch.utils.data import DataLoader
from datasets import load_dataset

LABEL_KEYS = ["eot", "hold", "bot", "bc"]

def collate_windows(sessions, window_frames=125, hop_frames=25):
    """Slice each session into fixed-length windows and collate into a batch."""
    windows = []
    for s in sessions:
        T = s["num_frames"]
        codes = np.array(s["codes_ch0"]).reshape(T, 8)
        for start in range(0, T - window_frames + 1, hop_frames):
            end = start + window_frames
            w = {
                "codes_ch0": torch.tensor(np.array(s["codes_ch0"]).reshape(T, 8)[start:end],  dtype=torch.long),
                "codes_ch1": torch.tensor(np.array(s["codes_ch1"]).reshape(T, 8)[start:end],  dtype=torch.long),
                "vad_ch0":   torch.tensor(np.array(s["vad_ch0"])[start:end],                  dtype=torch.float),
                "vad_ch1":   torch.tensor(np.array(s["vad_ch1"])[start:end],                  dtype=torch.float),
            }
            for name in LABEL_KEYS:
                for ch in ["ch0", "ch1"]:
                    key = f"{name}_{ch}"
                    w[key] = torch.tensor(np.array(s[key])[start:end], dtype=torch.float)
            windows.append(w)
    return {k: torch.stack([w[k] for w in windows]) for k in windows[0]}

ds     = load_dataset("anyreach-ai/dualturn-otospeech-turn-taking")
loader = DataLoader(ds["train"], batch_size=8, shuffle=True,
                    collate_fn=lambda b: collate_windows(b, window_frames=125, hop_frames=25))

batch = next(iter(loader))
print(batch["codes_ch0"].shape)   # [N_windows, 125, 8]
print(batch["eot_ch0"].shape)     # [N_windows, 125]

Label definitions

Label Meaning
EOT End-of-Turn: speaker yields the floor
HOLD Speaker keeps the floor (no handover)
BOT Beginning-of-Turn: other speaker takes the floor
BC Backchannel: short acknowledgement, no floor claim
VAD Voice Activity Detection (1 = speech)

DualTurn Model & Code

The following will be released soon:

  • Trained model checkpoint — on HuggingFace at anyreach-ai
  • Training code — model architecture, training loop, and configs
  • Evaluation code — benchmarks and metrics used in the paper

Authors

Citation

This dataset was used for training and evaluation in the DualTurn paper (submitted to Interspeech 2026). splits.json contains the exact train/val/test splits from the official dataset used for all experiments in the paper.

Paper: DualTurn: Learning Turn-Taking from Dual-Channel Generative Speech Pretraining

@misc{rajaa2026dualturnlearningturntakingdualchannel,
      title={DualTurn: Learning Turn-Taking from Dual-Channel Generative Speech Pretraining},
      author={Shangeth Rajaa},
      year={2026},
      eprint={2603.08216},
      archivePrefix={arXiv},
      primaryClass={eess.AS},
      url={https://arxiv.org/abs/2603.08216},
}

If you use this dataset, please cite - otoearth/otoSpeech-full-duplex-280h:

@misc{otoSpeech-full-duplex-280h,
  title        = {otoSpeech-full-duplex-280h: Full-Duplex Conversational Speech Dataset},
  author       = {otoearth},
  year         = {2025},
  howpublished = {\url{https://huggingface.co/datasets/otoearth/otoSpeech-full-duplex-280h}},
  note         = {License: CC BY 4.0}
}
Downloads last month
148

Models trained or fine-tuned on anyreach-ai/dualturn-otospeech-turn-taking

Collection including anyreach-ai/dualturn-otospeech-turn-taking

Paper for anyreach-ai/dualturn-otospeech-turn-taking