File size: 1,883 Bytes
c8f66d2
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7766d12
 
 
 
 
c8f66d2
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
---
license: apache-2.0
language:
- en
pretty_name: Molmo2-ER RefSpatial
tags:
- embodied-reasoning
- molmo2
- molmo2-er
- vlm-training-data
---

# Molmo2-ER · JingkunAn/RefSpatial

2.5M spatial-referring corpus (web + indoor + simulated) covering 31 spatial relations.

This is a re-hosted, **loader-ready subset** of the upstream dataset, used to train [`allenai/Molmo2-ER-4B`](https://huggingface.co/allenai/Molmo2-ER-4B). Files mirror the upstream layout; nothing in the data has been modified.

## Upstream source

- **Original dataset:** [JingkunAn/RefSpatial](https://huggingface.co/datasets/JingkunAn/RefSpatial)
- **Paper:** *RoboRefer: Towards Spatial Referring with Reasoning in Vision-Language Models for Robotics* ([arXiv:2506.04308](https://arxiv.org/abs/2506.04308))
- **License:** `apache-2.0` (inherits from upstream)

If you use this data, please cite the original authors:

```bibtex
@misc{zhou2026roboreferspatialreferringreasoning,
  title={RoboRefer: Towards Spatial Referring with Reasoning in Vision-Language Models for Robotics},
  author={Enshen Zhou and Jingkun An and Cheng Chi and others},
  year={2026},
  eprint={2506.04308},
  archivePrefix={arXiv}
}
```

## Extracting before training

This release ships archives. Extract them in-place before pointing `SPATIAL_DATA_HOME` at this directory:

```bash
# Reassemble multipart archives, then extract
cat 2D/depth/depth.tar.gz.part_* > 2D/depth/depth.tar.gz
cat 2D/image/image.tar.gz.part_* > 2D/image/image.tar.gz
cat 3D/image_visual_choice/image_visual_choice.tar.gz.part_* > 3D/image_visual_choice/image_visual_choice.tar.gz
find . -name '*.tar.gz' -execdir tar -xzf {} \;
```

## Usage in Molmo2-ER

See the [`allenai/molmo2`](https://github.com/allenai/molmo2) repository for the data loader and training recipe. The relevant loader class for this dataset lives in `olmo/data/spatial_datasets.py`.