Add dataset card and metadata for SpaceSpan
Browse filesHi! I'm Niels from the Hugging Face community science team. This PR improves the dataset card for the SpaceSpan dataset (associated with the Proxy3D paper). It adds the appropriate task categories, tags, and links to the paper, project page, and code repository to make the dataset more discoverable and well-documented.
README.md
CHANGED
|
@@ -1,3 +1,54 @@
|
|
| 1 |
-
---
|
| 2 |
-
license: apache-2.0
|
| 3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: apache-2.0
|
| 3 |
+
task_categories:
|
| 4 |
+
- video-text-to-text
|
| 5 |
+
tags:
|
| 6 |
+
- 3D
|
| 7 |
+
- vision-language
|
| 8 |
+
- spatial-intelligence
|
| 9 |
+
---
|
| 10 |
+
|
| 11 |
+
# SpaceSpan Dataset
|
| 12 |
+
|
| 13 |
+
SpaceSpan is a large-scale dataset curated for aligning 3D proxy representations with Vision-Language Models (VLMs), introduced in the paper [Proxy3D: Efficient 3D Representations for Vision-Language Models via Semantic Clustering and Alignment](https://huggingface.co/papers/2605.08064).
|
| 14 |
+
|
| 15 |
+
The dataset incorporates heterogeneous visual information into a unified format to support multi-stage training for developing spatial intelligence. It enables models to progress from simple image-text alignment to complex 3D reasoning tasks, such as 3D visual question answering (VQA) and visual grounding.
|
| 16 |
+
|
| 17 |
+
[**Project Page**](https://wzzheng.net/Proxy3D) | [**GitHub**](https://github.com/Spacedreamer2384/Proxy3D) | [**Paper**](https://huggingface.co/papers/2605.08064)
|
| 18 |
+
|
| 19 |
+
## Dataset Description
|
| 20 |
+
|
| 21 |
+
The SpaceSpan dataset (specifically the SpaceSpan-318K version) supports four progressive training stages:
|
| 22 |
+
- **Stage 1**: Initial spatial alignment.
|
| 23 |
+
- **Stage 2-3**: Intermediate spatial reasoning development.
|
| 24 |
+
- **Stage 4**: Full-scale 3D reasoning.
|
| 25 |
+
|
| 26 |
+
### Directory Structure
|
| 27 |
+
|
| 28 |
+
Based on the official repository, the dataset is typically organized as follows:
|
| 29 |
+
|
| 30 |
+
```bash
|
| 31 |
+
data/ # Training and inference data
|
| 32 |
+
βββ icon_image_embeds_qwen25.pt
|
| 33 |
+
βββ number_image_embeds_qwen25.pt
|
| 34 |
+
βββ stage_1_train.json
|
| 35 |
+
βββ stage_2_train.json
|
| 36 |
+
βββ stage_3_train.json
|
| 37 |
+
βββ stage_4_train_318K.json
|
| 38 |
+
βββ pointmaps_wo_markers
|
| 39 |
+
βββ poses
|
| 40 |
+
βββ ...
|
| 41 |
+
```
|
| 42 |
+
|
| 43 |
+
## Citation
|
| 44 |
+
|
| 45 |
+
If you find this dataset useful for your research, please cite the following paper:
|
| 46 |
+
|
| 47 |
+
```bibtex
|
| 48 |
+
@article{proxy3d2026,
|
| 49 |
+
title={Proxy3D: Efficient 3D Representations for Vision-Language Models via Semantic Clustering and Alignment},
|
| 50 |
+
author={Jiang, Jerry and Sun, Haowen and Gudovskiy, Denis and Nakata, Yohei and Okuno, Tomoyuki and Keutzer, Kurt and Zheng Wenzhao},
|
| 51 |
+
journal={arXiv preprint arXiv:2605.08064},
|
| 52 |
+
year={2026}
|
| 53 |
+
}
|
| 54 |
+
```
|