Datasets:

Languages:
English
ArXiv:
License:
danaaubakirova HF Staff commited on
Commit
a3dc2b0
·
verified ·
1 Parent(s): 5a6d9aa

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +294 -0
README.md ADDED
@@ -0,0 +1,294 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ tags:
4
+ - robotics
5
+ - community
6
+ - so100
7
+ - so101
8
+ - manipulation
9
+ - smolvla
10
+ - lerobot community
11
+ - vision-language-action
12
+ - embodied-ai
13
+ - cross-embodiment
14
+ task_categories:
15
+ - robotics
16
+ language:
17
+ - en
18
+ size_categories:
19
+ - 10M<n<100M
20
+ pretty_name: Community Dataset v3
21
+ ---
22
+
23
+ # Community Dataset v3 - A Cross-Embodiment Pretraining Dataset
24
+
25
+ A large-scale robotics dataset for vision-language-action learning, featuring **791 datasets** across **46 robot types**, enabling cross-embodiment pretraining for generalist robot policies.
26
+
27
+ ![Community Dataset v3](./comm_v3.png)
28
+
29
+ ## Overview
30
+
31
+ This is a **crowdsourced, open-source dataset** compiled from **235 community contributors** worldwide. Building upon the pretraining datasets used for [SmolVLA](https://huggingface.co/blog/smolvla), [Community Datasets v1](https://huggingface.co/datasets/HuggingFaceVLA/community_dataset_v1) and [v2](https://huggingface.co/datasets/HuggingFaceVLA/community_dataset_v2), this cleaned and organized version opens the door for **cross-embodiment training** on another completely new batch of community-contributed data.
32
+
33
+ The dataset spans **46+ robot embodiments** including single-arm, bimanual, mobile manipulation, and a few humanoid robots. All data was collected using the [LeRobot](https://github.com/huggingface/lerobot) framework and is compatible with the [VLAb](https://github.com/huggingface/VLAb) pretraining framework.
34
+
35
+
36
+ ## 📊 Dataset Statistics
37
+
38
+ | Metric | Value |
39
+ |--------|-------|
40
+ | **Total Datasets** | 791 |
41
+ | **Total Episodes** | 50,622 |
42
+ | **Total Frames** | 25,971,082 |
43
+ | **Total Duration** | 240.47 hours (10.02 days) |
44
+ | **Contributors** | 235 |
45
+ | **Robot Types** | 46 different embodiments |
46
+ | **Action Dimensions** | 12 different configurations |
47
+ | **Average Hours/Dataset** | 0.30 |
48
+
49
+ ## 🤖 Robot Distribution
50
+
51
+ ### By Category
52
+ - **Single-arm manipulators**: 72% (571 datasets)
53
+ - **Bimanual systems**: 12% (95 datasets)
54
+ - **Mobile manipulation**: 8% (63 datasets)
55
+ - **Humanoid platforms**: 1% (8 datasets)
56
+ - **Other configurations**: 7% (54 datasets)
57
+
58
+ ### Top 10 Robot Types
59
+
60
+ | Robot Type | Datasets | % | Category |
61
+ |------------|----------|---|----------|
62
+ | **so100** | 248 | 31.4% | Single-arm |
63
+ | **so101_follower** | 124 | 15.7% | Single-arm |
64
+ | **so100_follower** | 121 | 15.3% | Single-arm |
65
+ | **so101** | 82 | 10.4% | Single-arm |
66
+ | **arx5** | 43 | 5.4% | Single-arm |
67
+ | **koch** | 38 | 4.8% | Single-arm |
68
+ | **trossen_ai_mobile** | 25 | 3.2% | Mobile |
69
+ | **bi_xarm6_follower** | 16 | 2.0% | Bimanual |
70
+ | **so100_bimanual** | 12 | 1.5% | Bimanual |
71
+ | **koch_follower** | 8 | 1.0% | Single-arm |
72
+
73
+ ## 🎮 Action Space Distribution
74
+
75
+ | Action Shape | Datasets | % | Description |
76
+ |--------------|----------|---|-------------|
77
+ | **(6,)** | 607 | 76.7% | Single-arm + gripper (standard) |
78
+ | **(14,)** | 60 | 7.6% | Bimanual robots |
79
+ | **(7,)** | 58 | 7.3% | Single-arm extended DoF |
80
+ | **(16,)** | 26 | 3.3% | Complex bimanual |
81
+ | **(12,)** | 19 | 2.4% | Bimanual + grippers |
82
+ | Others (7 shapes) | 21 | 2.7% | Specialized configs |
83
+
84
+ ## 🗂️ Dataset Structure
85
+
86
+ ```
87
+ community_dataset_v3_clean/
88
+ ├── contributor1/
89
+ │ ├── dataset_name_1/
90
+ │ │ ├── data/ # Parquet files with observations
91
+ │ │ │ ├── episode_000000.parquet
92
+ │ │ │ ├── episode_000001.parquet
93
+ │ │ │ └── ...
94
+ │ │ ├── videos/ # MP4 recordings (multi-view)
95
+ │ │ │ ├── episode_000000_image.mp4
96
+ │ │ │ └── ...
97
+ │ │ └── meta/ # Metadata
98
+ │ │ └── info.json
99
+ │ └── dataset_name_2/
100
+ ├── contributor2/
101
+ └── ...
102
+ ```
103
+
104
+ ## 🚀 Usage
105
+
106
+ ### Prerequisites
107
+
108
+ Install LeRobot and authenticate:
109
+
110
+ ```bash
111
+ # Create conda environment
112
+ conda create -y -n lerobot python=3.10
113
+ conda activate lerobot
114
+
115
+ # Install ffmpeg and LeRobot
116
+ conda install ffmpeg -c conda-forge
117
+ git clone https://github.com/huggingface/lerobot.git
118
+ cd lerobot
119
+ pip install -e .
120
+
121
+ # Authenticate with Hugging Face
122
+ huggingface-cli login
123
+ # Get your token from https://huggingface.co/settings/tokens
124
+ ```
125
+
126
+ ### Download the Dataset
127
+
128
+ ```bash
129
+ huggingface-cli download HuggingFaceVLA/community_dataset_v3_clean \
130
+ --repo-type=dataset \
131
+ --local-dir ./community_dataset_v3_clean
132
+ ```
133
+
134
+ ### Load Individual Datasets
135
+
136
+ ```python
137
+ from lerobot.common.datasets.lerobot_dataset import LeRobotDataset
138
+ import os
139
+
140
+ # Browse available datasets
141
+ for contributor in os.listdir("./community_dataset_v3_clean"):
142
+ contributor_path = f"./community_dataset_v3_clean/{contributor}"
143
+ if os.path.isdir(contributor_path):
144
+ for dataset in os.listdir(contributor_path):
145
+ print(f"📁 {contributor}/{dataset}")
146
+
147
+ # Load a specific dataset
148
+ dataset = LeRobotDataset(
149
+ repo_id="local",
150
+ root="./community_dataset_v3/contributor_name/dataset_name"
151
+ )
152
+
153
+ # Access data
154
+ print(f"Episodes: {len(dataset.episode_data_index)}")
155
+ print(f"Total frames: {len(dataset)}")
156
+ ```
157
+
158
+ ### Train with VLAb
159
+
160
+ This dataset is designed for cross-embodiment VLA training using [VLAb](https://github.com/huggingface/VLAb):
161
+
162
+ ```bash
163
+ accelerate launch --config_file accelerate_configs/multi_gpu.yaml \
164
+ src/lerobot/scripts/train.py \
165
+ --policy.type=smolvla2 \
166
+ --policy.repo_id=HuggingFaceTB/SmolVLM2-500M-Video-Instruct \
167
+ --dataset.repo_id="community_dataset_v3/contributor1/dataset1,community_dataset_v3/contributor2/dataset2" \
168
+ --dataset.root="./community_dataset_v3" \
169
+ --dataset.video_backend=pyav \
170
+ --dataset.features_version=2 \
171
+ --output_dir="./outputs/training" \
172
+ --batch_size=8 \
173
+ --steps=200000 \
174
+ --wandb.enable=true \
175
+ --wandb.project="smolvla2-cross-embodiment"
176
+ ```
177
+
178
+ ## 🔧 Training Challenges with Cross-Embodiment Data
179
+
180
+ ### The Reality of Community-Contributed Data
181
+
182
+ This dataset includes 791 datasets recorded by community members under different conditions worldwide, creating an authentic in-the-wild setup. While this diversity is valuable for cross-embodiment learning, it comes with real challenges: varying data quality, inconsistent recording setups, and heterogeneous robot configurations. Using these datasets out-of-the-box will likely result in random collate errors and warnings during training.
183
+
184
+ ### What We Encountered During Data Cleaning
185
+
186
+ Starting with 851 datasets, we systematically debugged and cleaned the collection. Here's what we found:
187
+
188
+ #### Missing Video Files (Primary Removal Reason)
189
+ Some datasets had incomplete episode recordings where video files were missing:
190
+ ```
191
+ ERROR Failed to load video for key 'observation.images.image' at episode X:
192
+ [Errno 2] No such file or directory: '/path/to/episode_XXXXXX.mp4'
193
+ ```
194
+ **Impact:** Training crashes when these episodes were sampled
195
+ **Action:** Removed ~15-20 datasets with missing files
196
+
197
+ #### Data Type Incompatibilities
198
+ Certain datasets returned inconsistent data types during batch formation:
199
+ ```
200
+ RuntimeError: Could not infer dtype of dict
201
+ AttributeError: 'list' object has no attribute 'device'
202
+ ```
203
+ **Impact:** Random crashes during forward pass
204
+ **Action:** Removed ~10-15 problematic datasets, implemented resilient batch collation
205
+
206
+ #### Multi-Camera Configuration Issues
207
+ Different datasets had varying numbers of camera views, causing tensor shape mismatches:
208
+
209
+ **Root cause:** The `max_num_images` parameter wasn't properly propagated in the codebase, leading to inconsistent image tensor shapes when datasets had different numbers of cameras (some had 2, others had 4+ views).
210
+
211
+ **Impact:** Thousands of dimension/channel erros for the datasets with more than 3 images.
212
+ **Action:** Set `config.max_num_images = 3` to standardize input. This number balances multi-view information (essential for spatial reasoning) while being compatible with most datasets in the collection - the majority of community datasets use 2-3 camera views for manipulation tasks.
213
+
214
+ #### Video Timing Misalignments
215
+ Frame timestamps occasionally violated tolerance thresholds:
216
+ ```
217
+ Some query timestamps violate tolerance (tensor([2.0667]) > tolerance_s=0.0001)
218
+ ```
219
+ **Impact:** Minor temporal inconsistency, but training continued
220
+ **Action:** Automatic fallback to closest frames
221
+
222
+ #### Final Dataset Cleaning Results
223
+ - **Original datasets:** 851
224
+ - **Datasets with missing files:** ~15-20 (removed)
225
+ - **Datasets with data type issues:** ~10-15 (removed)
226
+ - **Datasets with conversion failures:** 16 (fixed and reprocessed)
227
+ - **Datasets with different FPS values:** Many datasets remain valid but have varying frame rates (some recorded at different fps than the standard 30fps)
228
+ - **Final clean dataset:** 791 datasets
229
+
230
+ ## 🎯 Intended Use
231
+
232
+ This dataset enables:
233
+
234
+ - **Cross-embodiment VLA training** - Learn policies that generalize across robot types
235
+ - **Multi-task manipulation** - Pick & place, sorting, assembly, bimanual tasks
236
+ - **Transfer learning** - Leverage diverse demonstrations for new robots
237
+ - **Imitation learning research** - Large-scale behavior cloning
238
+ - **Generalist robot policies** - Train models that work on multiple platforms
239
+ - **Mobile manipulation** - Navigation + manipulation tasks
240
+ - **Embodied AI research** - Vision-motor coordination
241
+
242
+
243
+ ## 🏆 Top Contributors
244
+
245
+ | Contributor | Datasets | % |
246
+ |-------------|----------|---|
247
+ | **shuohsuan** | 57 | 7.2% |
248
+ | **villekuosmanen** | 47 | 5.9% |
249
+ | **LeRobot-worldwide-hackathon** | 31 | 3.9% |
250
+ | **lt-s** | 27 | 3.4% |
251
+ | **Qipei** | 23 | 2.9% |
252
+ | **bjb7** | 18 | 2.3% |
253
+ | **kumarhans** | 18 | 2.3% |
254
+ | **Ryosei2** | 17 | 2.1% |
255
+ | **kyomangold** | 16 | 2.0% |
256
+ | **psg777** | 16 | 2.0% |
257
+ | Others (225) | 521 | 65.9% |
258
+
259
+ ## 🤝 Contributing
260
+
261
+ Future contributions should follow:
262
+ - LeRobot dataset format (v2.1+)
263
+ - Consistent naming for features and camera views
264
+ - Quality validation checks
265
+ - Precise task descriptions
266
+ - Robot type and action space metadata
267
+
268
+ See the [LeRobot dataset guide](https://huggingface.co/blog/lerobot-datasets) for best practices.
269
+
270
+ Please acknowledge all individual contributors who created the original datasets.
271
+
272
+ ## 📄 License
273
+
274
+ Released under **Apache 2.0 license**. Individual datasets may have additional attribution requirements.
275
+
276
+ When using this dataset:
277
+ - ✅ Cite the dataset and VLAb framework
278
+ - ✅ Acknowledge community contributors
279
+ - ✅ Follow Apache 2.0 license terms
280
+ - ✅ Consider contributing your own data
281
+
282
+ ## 🔗 Related Resources
283
+
284
+ - [VLAb Framework](https://github.com/huggingface/VLAb) - Large-scale pre-training
285
+ - [SmolVLA Model](https://huggingface.co/lerobot/smolvla_base) - Pre-trained VLA
286
+ - [SmolVLA Blog](https://huggingface.co/blog/smolvla) - Introduction and tutorials
287
+ - [SmolVLA Paper](https://huggingface.co/papers/2506.01844) - Technical details
288
+ - [LeRobot Docs](https://huggingface.co/docs/lerobot) - Complete documentation
289
+ - [Dataset Guide](https://huggingface.co/blog/lerobot-datasets) - Best practices
290
+ - [Community Dataset v2](https://huggingface.co/datasets/HuggingFaceVLA/community_dataset_v2) - Previous Dataset
291
+ - [Community Dataset v1](https://huggingface.co/datasets/HuggingFaceVLA/community_dataset_v1) - First release
292
+
293
+
294
+ *Built with ❤️ by the LeRobot Community and SmolVLA Team*