File size: 5,515 Bytes
2116b11
97292a8
2116b11
97292a8
2116b11
97292a8
2116b11
97292a8
2116b11
97292a8
2116b11
97292a8
2116b11
 
97292a8
2116b11
97292a8
2116b11
97292a8
2116b11
97292a8
 
2116b11
 
 
 
 
a00ec58
87a022f
a00ec58
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
# FastUMI Pro Dataset

## Project Description

FastUMI Pro is the upgraded enterprise version of FastUMI, designed for streamlined, end-to-end data acquisition and transformation systems for corporate users.

FastUMI (Fast Universal Manipulation Interface) is a dataset and interface framework for universal robot manipulation tasks, supporting hardware-agnostic, scalable, and efficient data collection and model training. The project provides physical prototype systems, complete data collection code, standardized data formats, and utility tools to facilitate real-world manipulation learning research.

## Dataset Overview

FastUMI Pro builds upon FastUMI with enhanced features:
- Higher precision trajectory data
- Support for more diverse robot embodiments, truly enabling "one-brain-multi-form" applications
- Comprehensive data leadership in the field

The original FastUMI open-sourced FastUMI-150K containing approximately 150,000 real-world manipulation trajectories, which was first provided to selected research partners for training large-scale VLA (Vision-Language-Action) models.

## Quick Start

### Download Example Data

```bash
# Original command (may be slow in some regions)
huggingface-cli download FastUMIPro/example_data_fastumi_pro_raw --repo-type dataset --local-dir ~/fastumi_data/

# Mirror acceleration solution
export HF_ENDPOINT=https://hf-mirror.com
huggingface-cli download --repo-type dataset --resume-download FastUMIPro/example_data_fastumi_pro_raw --local-dir ~/fastumi_data/
```

## Data Structure
FastUMI PRO uses raw format containing various types of raw sensor data, which can be easily converted to other formats. The raw format facilitates querying and validating original sensor outputs for rapid problem identification.
DATA/
└── device_label_xv_serial/
    └── session_timestamp/
        ├── RGB_Images/
        │   ├── timestamps.csv
        │   └── Frames/
        │       ├── frame_000001.jpg
        │       ├── frame_000002.jpg
        │       └── ...
        ├── SLAM_Poses/
        │   └── slam_raw.txt
        ├── Vive_Poses/
        │   └── vive_data_tum.txt
        ├── ToF_PointClouds/
        │   ├── timestamps.csv
        │   └── PointClouds/
        │       ├── pointcloud_000001.pcd
        │       ├── pointcloud_000002.pcd
        │       └── ...
        ├── Clamp_Data/
        │   └── clamp_data_tum.txt
        └── Merged_Trajectory/
            ├── merged_trajectory.txt
            └── merge_stats.txt

### Directory Descriptions
session_xxx: Individual data collection session

RGB_Images: Frame images supporting multiple viewpoints; supports both Images and Videos

SLAM_Poses: UMI pose data

Vive_Poses: Vive tracking system pose data

ToF_PointClouds: Time-of-Flight point cloud raw data (depth)

Merged_Trajectory: Trajectory data

### Data Specifications

Attributes
sim:

False: Real environment data

True: Simulation data

Observations
observations/images/: Camera image data

Default camera name: front

Shape: (frames, 1920, 1080, 3)

Data type: uint8

Compression: gzip (level 4)

observations/qpos:

Type: Floating point dataset

Shape: (timesteps, 7)

Meaning: Robot end-effector position + quaternion orientation

Order: [Pos X, Pos Y, Pos Z, Q_X, Q_Y, Q_Z, Q_W]

Actions
Type: Floating point dataset

Shape: (timesteps, 7)

Meaning: Actions (same structure as qpos, typically mirroring qpos)

Data Conversion
Supports one-click export to specific formats via web toolchain, or conversion between formats using tools like:

Any4lerobot: GitHub - Tavish9/any4lerobot

Conversion paths supported:

hdf5 → lerobot v3.0

hdf5 → lerobot(Pi0) v2.0

hdf5 → rlds

Model Performance
Preliminary experiments show that models trained on this dataset demonstrate significant multi-task generalization capabilities in universal manipulation tasks:

VLA Models: Including PI-O models with language understanding and action planning capabilities, exhibiting excellent generalization and execution stability in multi-task language-conditioned control

VA Models: Classical visual control architectures like ACT, DP also show significant improvements, particularly in complex operation sequences, viewpoint perturbations, and fine motion tracking with enhanced robustness

Related Links
Project Homepage: https://fastumi.com/pro/

FastUMI Project: https://fastumi.com

Hugging Face Dataset: https://huggingface.co/datasets/IPE...

Research Paper: [2409.19499] FastUMI: A Scalable and...

Open Source Toolchain:

Demo Replay: GitHub - Loki-Lu/FastUMI_replay_sin...

Dual-arm Demo: GitHub - Loki-Lu/FastUMI_replay_du...

Hardware SDK: GitHub - FastUMIRobotics/FastUMI_...

Monitoring Tools: GitHub - FastUMIRobotics/FastUMI_...

Data Collection Tools: GitHub - FastUMIRobotics/FastUMI_...

Related Research
[2508.10538] MLM: Learning Multi-ta...

PIO (FastUMI Lightweight Adaptation, Version V0) Full Tutorial: PIO (FastUMI数据轻量级适配,版本V0)…

Citation
If you use this dataset in your research, please cite the relevant papers:

bibtex
@article{fastumi2024,
  title={FastUMI: A Scalable and Hardware-Agnostic Framework for Robot Manipulation Learning},
  author={FastUMI Team},
  journal={arXiv preprint},
  year={2024}
}
Contact
For any questions or suggestions, please contact the development team:

Lead: [Name]

Email: [Email Address]

WeChat: [WeChat ID]