File size: 2,617 Bytes
08cd15b
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
14e3475
 
 
 
 
 
 
 
08cd15b
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
14e3475
 
08cd15b
a452deb
08cd15b
 
 
99a7511
08cd15b
 
 
14e3475
 
08cd15b
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
---
license: apache-2.0
task_categories:
  - video-text-to-text
  - visual-question-answering
  - image-text-to-text
language:
  - en
tags:
  - llava
  - multimodal
  - video
  - spatial-reasoning
size_categories:
  - 10M<n<100M
configs:
  - config_name: viewer_caption_30s
    data_files:
      - split: preview
        path: viewer/caption_30s.parquet
  - config_name: viewer_caption_60s
    data_files:
      - split: preview
        path: viewer/caption_60s.parquet
  - config_name: viewer_caption_180s
    data_files:
      - split: preview
        path: viewer/caption_180s.parquet
  - config_name: viewer_caption_gt10min
    data_files:
      - split: preview
        path: viewer/caption_gt10min.parquet
  - config_name: viewer_spatial
    data_files:
      - split: preview
        path: viewer/spatial.parquet
---

# LLaVA-OneVision-2-Data

Training data for the LLaVA-OneVision-2 multimodal model family, covering large-scale video and spatial reasoning corpora used in mid-training.

## Dataset Composition

| Subset | Format | Description |
|---|---|---|
| `mid_training_video/60s_rest/` | WebDataset (`.tar`) | 10,809 shards of ~60s video clips |
| `mid_training_video/caption_v0/split_30s.jsonl` | JSONL | Captions for 30-second video clips |
| `mid_training_video/caption_v0/split_60s.jsonl` | JSONL | Captions for 60-second video clips |
| `mid_training_video/caption_v0/split_180s.jsonl` | JSONL | Captions for 180-second video clips |
| `mid_training_video/caption_v0/split_gt10min.jsonl` | JSONL | Captions for >10-minute video clips |
| `spatial/` | WebDataset (`.tar`) | 84 shards of spatial reasoning data (refcoco, visual genome, pointing, 3D, etc.) |
| `mid_training_video/mapping/mapping_{5s,10s,30s,60s,180s,gt10min}.csv` | CSV | Maps each video clip's `dst_path` to its source `youtube_id` and `[start_time, end_time]` window |

## Preview Configs

The `viewer_*` configs above expose small Parquet samples so the Hugging Face Dataset Viewer can render the data directly in the browser:

- **`viewer_caption_30s`** — 5 caption samples from 30-second clips
- **`viewer_caption_60s`** — 5 caption samples from 60-second clips
- **`viewer_caption_180s`** — 3 caption samples from 180-second clips
- **`viewer_caption_gt10min`** — 1 caption sample from >10-minute clips
- **`viewer_spatial`** — 10 spatial-reasoning samples with embedded thumbnail images, mixed across tasks (refcoco, visual genome, pointing, ca1m, osd, crosspoint, erqa, roborefer)

These previews are intended for **schema inspection only**. For training, use the full `mid_training_video/` and `spatial/` shards.