nielsr HF Staff commited on
Commit
e2daf0b
Β·
verified Β·
1 Parent(s): bce42d7

Add dataset card and metadata for SpaceSpan

Browse files

Hi! I'm Niels from the Hugging Face community science team. This PR improves the dataset card for the SpaceSpan dataset (associated with the Proxy3D paper). It adds the appropriate task categories, tags, and links to the paper, project page, and code repository to make the dataset more discoverable and well-documented.

Files changed (1) hide show
  1. README.md +54 -3
README.md CHANGED
@@ -1,3 +1,54 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ task_categories:
4
+ - video-text-to-text
5
+ tags:
6
+ - 3D
7
+ - vision-language
8
+ - spatial-intelligence
9
+ ---
10
+
11
+ # SpaceSpan Dataset
12
+
13
+ SpaceSpan is a large-scale dataset curated for aligning 3D proxy representations with Vision-Language Models (VLMs), introduced in the paper [Proxy3D: Efficient 3D Representations for Vision-Language Models via Semantic Clustering and Alignment](https://huggingface.co/papers/2605.08064).
14
+
15
+ The dataset incorporates heterogeneous visual information into a unified format to support multi-stage training for developing spatial intelligence. It enables models to progress from simple image-text alignment to complex 3D reasoning tasks, such as 3D visual question answering (VQA) and visual grounding.
16
+
17
+ [**Project Page**](https://wzzheng.net/Proxy3D) | [**GitHub**](https://github.com/Spacedreamer2384/Proxy3D) | [**Paper**](https://huggingface.co/papers/2605.08064)
18
+
19
+ ## Dataset Description
20
+
21
+ The SpaceSpan dataset (specifically the SpaceSpan-318K version) supports four progressive training stages:
22
+ - **Stage 1**: Initial spatial alignment.
23
+ - **Stage 2-3**: Intermediate spatial reasoning development.
24
+ - **Stage 4**: Full-scale 3D reasoning.
25
+
26
+ ### Directory Structure
27
+
28
+ Based on the official repository, the dataset is typically organized as follows:
29
+
30
+ ```bash
31
+ data/ # Training and inference data
32
+ β”œβ”€β”€ icon_image_embeds_qwen25.pt
33
+ β”œβ”€β”€ number_image_embeds_qwen25.pt
34
+ β”œβ”€β”€ stage_1_train.json
35
+ β”œβ”€β”€ stage_2_train.json
36
+ β”œβ”€β”€ stage_3_train.json
37
+ β”œβ”€β”€ stage_4_train_318K.json
38
+ β”œβ”€β”€ pointmaps_wo_markers
39
+ β”œβ”€β”€ poses
40
+ └── ...
41
+ ```
42
+
43
+ ## Citation
44
+
45
+ If you find this dataset useful for your research, please cite the following paper:
46
+
47
+ ```bibtex
48
+ @article{proxy3d2026,
49
+ title={Proxy3D: Efficient 3D Representations for Vision-Language Models via Semantic Clustering and Alignment},
50
+ author={Jiang, Jerry and Sun, Haowen and Gudovskiy, Denis and Nakata, Yohei and Okuno, Tomoyuki and Keutzer, Kurt and Zheng Wenzhao},
51
+ journal={arXiv preprint arXiv:2605.08064},
52
+ year={2026}
53
+ }
54
+ ```