willow-cuhk commited on
Commit
01eab7d
·
verified ·
1 Parent(s): acef5d1

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +146 -26
README.md CHANGED
@@ -1,31 +1,151 @@
1
  ---
2
- license: cc-by-nc-4.0
3
  task_categories:
4
- - video-classification
5
- - text-generation
6
- modalities:
7
- - depth
8
- - infrared
9
- - thermal
10
- - imu
11
- - radar
12
- - skeleton
13
  language:
14
- - en
15
  tags:
16
- - human-activity-recognition
17
- - multimodal
18
- - sensor-data
19
- - IMU
20
- - depth
21
- - infrared
22
- - thermal
23
- - skeleton
24
- - radar
25
- - mmwave
26
- - HAR
27
- - privacy-preserving
28
- pretty_name: CUHK-S
29
  size_categories:
30
- - 100M<n<1B
31
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ license: cc-by-4.0
3
  task_categories:
4
+ - video-classification
5
+ - text-generation
 
 
 
 
 
 
 
6
  language:
7
+ - en
8
  tags:
9
+ - human-activity-recognition
10
+ - multimodal
11
+ - sensor-data
12
+ - privacy-preserving
13
+ - IMU
14
+ - depth
15
+ - infrared
16
+ - thermal
17
+ - skeleton
18
+ - radar
19
+ - mmwave
20
+ - HAR
21
+ pretty_name: "CUHK-S"
22
  size_categories:
23
+ - 100K<n<1M
24
+ ---
25
+
26
+ # CUHK-S: A Privacy-Preserving Multimodal Dataset for Human Action Recognition
27
+
28
+ [![Paper](https://img.shields.io/badge/Paper-arXiv:2512.07136-red)](https://www.arxiv.org/abs/2512.07136)
29
+ [![Project Page](https://img.shields.io/badge/Project-Page-blue)](https://siyang-jiang.github.io/CUHK-X/)
30
+
31
+ ## Dataset Description
32
+
33
+ CUHK-S is a **privacy-preserving subset** of the [CUHK-X](https://siyang-jiang.github.io/CUHK-X/) dataset, a large-scale multimodal benchmark for Human Action Recognition (HAR), Understanding (HAU), and Reasoning (HARn). CUHK-X was accepted at **MobiSys 2026**.
34
+
35
+ Compared to the full CUHK-X dataset, CUHK-S:
36
+ - **Removes all RGB video** to prevent facial identification
37
+ - **Downscales** all visual modalities to 320 × 240
38
+ - **Selects 18 out of 30** participants while preserving full action coverage (40 categories)
39
+
40
+ ## Dataset Summary
41
+
42
+ | Attribute | Value |
43
+ |-------------------|------------------------------------------------|
44
+ | Participants | 18 (selected from 30 in CUHK-X) |
45
+ | Action Categories | 40 |
46
+ | Modalities | 6 (Depth, IR, Thermal, IMU, Radar, Skeleton) |
47
+ | Visual Resolution | 320 × 240 |
48
+ | Total Size | ~146 GB (18 zip files, one per participant) |
49
+
50
+ ## Modalities
51
+
52
+ | Modality | Format | Description |
53
+ |------------|-------------|-------------------------------------------------|
54
+ | Depth | PNG (color) | Colorized depth maps from Vzense NYX 650 |
55
+ | IR | PNG | Infrared images, robust to lighting changes |
56
+ | Thermal | PNG | Heat signature from thermal camera |
57
+ | IMU | CSV | 5-sensor accelerometer/gyroscope/magnetometer |
58
+ | Radar | Binary | mmWave radar point cloud (TI Radar) |
59
+ | Skeleton | JSON/CSV | 3D joint positions from pose estimation |
60
+
61
+ > **Note**: RGB video is intentionally excluded from CUHK-S to protect participant privacy.
62
+
63
+ ## Dataset Structure
64
+
65
+ Each participant's data is packaged as a zip file: `CUHK-S_userN-userN.zip`
66
+
67
+ ```
68
+ CUHK-S/
69
+ ├── HAR/ # Human Action Recognition (per-action organized)
70
+ │ ├── data/
71
+ │ │ ├── depth_color/
72
+ │ │ ├── ir/
73
+ │ │ ├── thermal/
74
+ │ │ ├── imu/
75
+ │ │ ├── radar/
76
+ │ │ └── skeleton/
77
+ │ └── GT/ # Ground truth annotations
78
+ ├── HARn/ # Human Action Reasoning (per-modality organized)
79
+ │ ├── data/
80
+ │ └── GT/
81
+ ├── HAU/ # Human Action Understanding (per-modality organized)
82
+ │ ├── data/
83
+ │ └── GT/
84
+ └── source_data/ # Raw source data
85
+ ├── data/
86
+ └── GT/
87
+ ```
88
+
89
+ - **HAR**: Singular well-defined actions for traditional classification tasks
90
+ - **HAU**: Sequential actions for temporal and contextual understanding
91
+ - **HARn**: Sequential actions for next-action reasoning and prediction
92
+ - **source_data**: Raw unprocessed sensor data
93
+
94
+ ## IMU Sensor Layout
95
+
96
+ Five IMU sensors are placed on the body:
97
+
98
+ | Sensor | Position | Channels (per sensor) |
99
+ |--------|------------|-------------------------------------------|
100
+ | WTLA | Left Arm | Acc(X/Y/Z), Gyro(X/Y/Z), Mag(X/Y/Z) |
101
+ | WTC | Chest | Acc(X/Y/Z), Gyro(X/Y/Z), Mag(X/Y/Z) |
102
+ | WTRA | Right Arm | Acc(X/Y/Z), Gyro(X/Y/Z), Mag(X/Y/Z) |
103
+ | WTRL | Right Leg | Acc(X/Y/Z), Gyro(X/Y/Z), Mag(X/Y/Z) |
104
+ | WTLL | Left Leg | Acc(X/Y/Z), Gyro(X/Y/Z), Mag(X/Y/Z) |
105
+
106
+ ## Benchmarks & Tasks
107
+
108
+ | Task | Type | Metrics |
109
+ |-------------------------|-----------------|----------------------------------|
110
+ | Action Recognition | Classification | Accuracy, F1, Precision, Recall |
111
+ | Action Selection | Multiple Choice | Accuracy |
112
+ | Action Captioning | Generation | BLEU, METEOR |
113
+ | Emotion Analysis | Classification | Accuracy |
114
+ | Sequential Reordering | Ordering | Accuracy |
115
+ | Next Action Reasoning | Reasoning | Accuracy |
116
+
117
+ ## Citation
118
+
119
+ If you use CUHK-S in your research, please cite:
120
+
121
+ ```bibtex
122
+ @inproceedings{jiang2026cuhkx,
123
+ title={CUHK-X: A Large-Scale Multimodal Dataset and Benchmark for Human Action Recognition, Understanding and Reasoning},
124
+ author={Jiang, Siyang and others},
125
+ booktitle={Proceedings of ACM MobiSys},
126
+ year={2026}
127
+ }
128
+ ```
129
+
130
+ ## Ethics & Privacy
131
+
132
+ We obtained approval from an Institutional Review Board (IRB) to conduct this study and collect data from human subjects.
133
+
134
+ **Privacy measures in CUHK-S:**
135
+ - No RGB video is included to prevent facial identification
136
+ - All visual modalities are downscaled to 320 × 240
137
+ - Participants are identified only by numeric IDs (e.g., user1, user2)
138
+ - No personally identifiable information is linked to individual records
139
+ - IMU, Radar, and Skeleton modalities do not capture visual appearance
140
+
141
+ ## License
142
+
143
+ Code is released under the MIT License. The dataset is available for non-commercial research under a Data Use Agreement (DUA) and is not redistributable. Our derived annotations/splits are released under CC BY 4.0.
144
+
145
+ **Note**: This dataset is designed for research and educational purposes. Please ensure compliance with your institution's ethics guidelines when using human activity data.
146
+
147
+ ## Contact
148
+
149
+ - **Email**: syjiang [AT] ie.cuhk.edu.hk
150
+ - **Project Page**: [https://siyang-jiang.github.io/CUHK-X/](https://siyang-jiang.github.io/CUHK-X/)
151
+ - **Lab**: [CUHK AIoT Lab](https://aiot.ie.cuhk.edu.hk)