File size: 2,020 Bytes
79c8d10 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 | ---
license: apache-2.0
task_categories:
- video-text-to-text
tags:
- 3d
- spatial-intelligence
- vlm
- visual-grounding
---
# SpaceSpan Dataset
SpaceSpan is a large-scale dataset curated for the training and evaluation of 3D vision-language models (VLMs), specifically introduced in the paper [Proxy3D: Efficient 3D Representations for Vision-Language Models via Semantic Clustering and Alignment](https://huggingface.co/papers/2605.08064).
[**Project Page**](https://wzzheng.net/Proxy3D) | [**GitHub Repository**](https://github.com/Spacedreamer2384/Proxy3D)
## Dataset Description
The SpaceSpan dataset is designed to help VLMs develop spatial intelligence through 3D proxy representations. It incorporates heterogeneous visual information with a unified data format, enabling multi-stage training for skills ranging from simple image-text alignment to complex 3D spatial reasoning, 3D visual question answering (VQA), and visual grounding.
The dataset includes approximately **318K samples** used across four progressive training stages.
## Dataset Structure
The repository typically includes the following components used for the Proxy3D training pipeline:
- **Training Instructions**: JSON files for stages 1 through 4 (e.g., `stage_4_train_318K.json`).
- **Embeddings**: Pre-computed vision embeddings for efficiency.
- **Geometric Data**: Pointmaps and camera poses for 3D reconstruction and scene representation.
For evaluation annotations, please refer to the [Proxy3D-annotations](https://huggingface.co/datasets/Spacewanderer8263/Proxy3D-annotations) repository.
## Citation
If you use this dataset in your research, please cite the following paper:
```bibtex
@article{proxy3d2026,
title={Proxy3D: Efficient 3D Representations for Vision-Language Models via Semantic Clustering and Alignment},
author={Jiang, Jerry and Sun, Haowen and Gudovskiy, Denis and Nakata, Yohei and Okuno, Tomoyuki and Keutzer, Kurt and Zheng Wenzhao},
journal={arXiv preprint arXiv:2605.08064},
year={2026}
}
``` |