| --- |
| license: apache-2.0 |
| task_categories: |
| - video-text-to-text |
| tags: |
| - 3d |
| - spatial-intelligence |
| - vlm |
| - visual-grounding |
| --- |
| |
| # SpaceSpan Dataset |
|
|
| SpaceSpan is a large-scale dataset curated for the training and evaluation of 3D vision-language models (VLMs), specifically introduced in the paper [Proxy3D: Efficient 3D Representations for Vision-Language Models via Semantic Clustering and Alignment](https://huggingface.co/papers/2605.08064). |
|
|
| [**Project Page**](https://wzzheng.net/Proxy3D) | [**GitHub Repository**](https://github.com/Spacedreamer2384/Proxy3D) |
|
|
| ## Dataset Description |
|
|
| The SpaceSpan dataset is designed to help VLMs develop spatial intelligence through 3D proxy representations. It incorporates heterogeneous visual information with a unified data format, enabling multi-stage training for skills ranging from simple image-text alignment to complex 3D spatial reasoning, 3D visual question answering (VQA), and visual grounding. |
|
|
| The dataset includes approximately **318K samples** used across four progressive training stages. |
|
|
| ## Dataset Structure |
|
|
| The repository typically includes the following components used for the Proxy3D training pipeline: |
| - **Training Instructions**: JSON files for stages 1 through 4 (e.g., `stage_4_train_318K.json`). |
| - **Embeddings**: Pre-computed vision embeddings for efficiency. |
| - **Geometric Data**: Pointmaps and camera poses for 3D reconstruction and scene representation. |
|
|
| For evaluation annotations, please refer to the [Proxy3D-annotations](https://huggingface.co/datasets/Spacewanderer8263/Proxy3D-annotations) repository. |
|
|
| ## Citation |
|
|
| If you use this dataset in your research, please cite the following paper: |
|
|
| ```bibtex |
| @article{proxy3d2026, |
| title={Proxy3D: Efficient 3D Representations for Vision-Language Models via Semantic Clustering and Alignment}, |
| author={Jiang, Jerry and Sun, Haowen and Gudovskiy, Denis and Nakata, Yohei and Okuno, Tomoyuki and Keutzer, Kurt and Zheng Wenzhao}, |
| journal={arXiv preprint arXiv:2605.08064}, |
| year={2026} |
| } |
| ``` |