eamonn-zh commited on
Commit
0488400
·
0 Parent(s):

Initial clean commit

Browse files
.gitattributes ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.avro filter=lfs diff=lfs merge=lfs -text
4
+ *.bin filter=lfs diff=lfs merge=lfs -text
5
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
6
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
7
+ *.ftz filter=lfs diff=lfs merge=lfs -text
8
+ *.gz filter=lfs diff=lfs merge=lfs -text
9
+ *.h5 filter=lfs diff=lfs merge=lfs -text
10
+ *.joblib filter=lfs diff=lfs merge=lfs -text
11
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
12
+ *.lz4 filter=lfs diff=lfs merge=lfs -text
13
+ *.mds filter=lfs diff=lfs merge=lfs -text
14
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
15
+ *.model filter=lfs diff=lfs merge=lfs -text
16
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
17
+ *.npy filter=lfs diff=lfs merge=lfs -text
18
+ *.npz filter=lfs diff=lfs merge=lfs -text
19
+ *.onnx filter=lfs diff=lfs merge=lfs -text
20
+ *.ot filter=lfs diff=lfs merge=lfs -text
21
+ *.parquet filter=lfs diff=lfs merge=lfs -text
22
+ *.pb filter=lfs diff=lfs merge=lfs -text
23
+ *.pickle filter=lfs diff=lfs merge=lfs -text
24
+ *.pkl filter=lfs diff=lfs merge=lfs -text
25
+ *.pt filter=lfs diff=lfs merge=lfs -text
26
+ *.pth filter=lfs diff=lfs merge=lfs -text
27
+ *.rar filter=lfs diff=lfs merge=lfs -text
28
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
29
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
30
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
31
+ *.tar filter=lfs diff=lfs merge=lfs -text
32
+ *.tflite filter=lfs diff=lfs merge=lfs -text
33
+ *.tgz filter=lfs diff=lfs merge=lfs -text
34
+ *.wasm filter=lfs diff=lfs merge=lfs -text
35
+ *.xz filter=lfs diff=lfs merge=lfs -text
36
+ *.zip filter=lfs diff=lfs merge=lfs -text
37
+ *.zst filter=lfs diff=lfs merge=lfs -text
38
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
39
+ # Audio files - uncompressed
40
+ *.pcm filter=lfs diff=lfs merge=lfs -text
41
+ *.sam filter=lfs diff=lfs merge=lfs -text
42
+ *.raw filter=lfs diff=lfs merge=lfs -text
43
+ # Audio files - compressed
44
+ *.aac filter=lfs diff=lfs merge=lfs -text
45
+ *.flac filter=lfs diff=lfs merge=lfs -text
46
+ *.mp3 filter=lfs diff=lfs merge=lfs -text
47
+ *.ogg filter=lfs diff=lfs merge=lfs -text
48
+ *.wav filter=lfs diff=lfs merge=lfs -text
49
+ # Image files - uncompressed
50
+ *.bmp filter=lfs diff=lfs merge=lfs -text
51
+ *.gif filter=lfs diff=lfs merge=lfs -text
52
+ *.png filter=lfs diff=lfs merge=lfs -text
53
+ *.tiff filter=lfs diff=lfs merge=lfs -text
54
+ # Image files - compressed
55
+ *.jpg filter=lfs diff=lfs merge=lfs -text
56
+ *.jpeg filter=lfs diff=lfs merge=lfs -text
57
+ *.webp filter=lfs diff=lfs merge=lfs -text
58
+ # Video files - compressed
59
+ *.mp4 filter=lfs diff=lfs merge=lfs -text
60
+ *.webm filter=lfs diff=lfs merge=lfs -text
16_frame/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:435960ec3339f6be87161e1e143eca6fd027923f1d19976a35396ec34fea435d
3
+ size 147029
32_frame/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:86c70c4c52d9d0d8ed6dd5cb619ea61359d2be17a83f3a0a0fd2e843846dffdf
3
+ size 209977
64_frame/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dbd7a5a1f2069fe7c6c4dc34016e5cb604c4b139c93c94d2b993a1a51f54c6bc
3
+ size 231397
README.md ADDED
@@ -0,0 +1,389 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ dataset_info:
3
+ - config_name: 16_frame
4
+ features:
5
+ - name: id
6
+ dtype: int64
7
+ - name: dataset
8
+ dtype: string
9
+ - name: scene_id
10
+ dtype: string
11
+ - name: question_type
12
+ dtype: string
13
+ - name: question
14
+ dtype: string
15
+ - name: ground_truth
16
+ dtype: string
17
+ - name: options
18
+ sequence: string
19
+ - name: num_frames
20
+ dtype: string
21
+ - name: queried_object_ids
22
+ sequence: int64
23
+ splits:
24
+ - name: test
25
+ num_bytes: 1211030
26
+ num_examples: 4568
27
+ download_size: 147029
28
+ dataset_size: 1211030
29
+ - config_name: 32_frame
30
+ features:
31
+ - name: id
32
+ dtype: int64
33
+ - name: dataset
34
+ dtype: string
35
+ - name: scene_id
36
+ dtype: string
37
+ - name: question_type
38
+ dtype: string
39
+ - name: question
40
+ dtype: string
41
+ - name: ground_truth
42
+ dtype: string
43
+ - name: options
44
+ sequence: string
45
+ - name: num_frames
46
+ dtype: string
47
+ - name: queried_object_ids
48
+ sequence: int64
49
+ splits:
50
+ - name: test
51
+ num_bytes: 1769552
52
+ num_examples: 6158
53
+ download_size: 209977
54
+ dataset_size: 1769552
55
+ - config_name: 64_frame
56
+ features:
57
+ - name: id
58
+ dtype: int64
59
+ - name: dataset
60
+ dtype: string
61
+ - name: scene_id
62
+ dtype: string
63
+ - name: question_type
64
+ dtype: string
65
+ - name: question
66
+ dtype: string
67
+ - name: ground_truth
68
+ dtype: string
69
+ - name: options
70
+ sequence: string
71
+ - name: num_frames
72
+ dtype: string
73
+ - name: queried_object_ids
74
+ sequence: int64
75
+ splits:
76
+ - name: test
77
+ num_bytes: 1931345
78
+ num_examples: 6616
79
+ download_size: 231397
80
+ dataset_size: 1931345
81
+ - config_name: all_frame
82
+ features:
83
+ - name: id
84
+ dtype: int64
85
+ - name: dataset
86
+ dtype: string
87
+ - name: scene_id
88
+ dtype: string
89
+ - name: question_type
90
+ dtype: string
91
+ - name: question
92
+ dtype: string
93
+ - name: ground_truth
94
+ dtype: string
95
+ - name: options
96
+ sequence: string
97
+ - name: num_frames
98
+ dtype: string
99
+ - name: queried_object_ids
100
+ sequence: int64
101
+ splits:
102
+ - name: test
103
+ num_bytes: 2010779
104
+ num_examples: 6808
105
+ download_size: 239453
106
+ dataset_size: 2010779
107
+ configs:
108
+ - config_name: 16_frame
109
+ data_files:
110
+ - split: test
111
+ path: 16_frame/test-*
112
+ - config_name: 32_frame
113
+ data_files:
114
+ - split: test
115
+ path: 32_frame/test-*
116
+ - config_name: 64_frame
117
+ data_files:
118
+ - split: test
119
+ path: 64_frame/test-*
120
+ - config_name: all_frame
121
+ data_files:
122
+ - split: test
123
+ path: all_frame/test-*
124
+ default: true
125
+ task_categories:
126
+ - visual-question-answering
127
+ language:
128
+ - en
129
+ size_categories:
130
+ - 1K<n<10K
131
+ license: apache-2.0
132
+ tags:
133
+ - Spatial Intelligence
134
+ - Vision Language Models
135
+ ---
136
+
137
+ <div align="center">
138
+ <img src="metadata/revsi.png" width="350">
139
+
140
+ <a href="https://github.com/eamonn-zh">Yiming Zhang</a><sup>1*</sup>,
141
+ <a href="https://jcchen.me/">Jiacheng Chen</a><sup>1*</sup>,
142
+ <a href="https://christinatan0704.github.io/mysite/">Jiaqi Tan</a><sup>1</sup>,
143
+ <a href="https://sammaoys.github.io/">Yongsen Mao</a><sup>2</sup>,
144
+ <a href="https://wenhuchen.github.io/">Wenhu Chen</a><sup>3</sup>,
145
+ <a href="https://angelxuanchang.github.io/">Angel X. Chang</a><sup>1,4</sup>
146
+ <br>
147
+ <sup>1</sup> Simon Fraser University &nbsp;&nbsp;
148
+ <sup>2</sup> Hong Kong University of Science and Technology
149
+ <br>
150
+ <sup>3</sup> University of Waterloo &nbsp;&nbsp;
151
+ <sup>4</sup> Alberta Machine Intelligence Institute (Amii)
152
+
153
+ <a href="https://3dlg-hcvc.github.io/revsi/">
154
+ <img src="https://img.shields.io/badge/Project%20Page-84C0B8?style=for-the-badge">
155
+ </a>
156
+ <a href="https://github.com/3dlg-hcvc/revsi">
157
+ <img src="https://img.shields.io/badge/github-%23121011.svg?style=for-the-badge&logo=github&logoColor=white">
158
+ </a>
159
+ <a href="https://arxiv.org/abs/xxxx.xxxx">
160
+ <img src="https://img.shields.io/badge/arXiv-xxxx.xxxx-b31b1b.svg?style=for-the-badge">
161
+ </a>
162
+ <a href="https://revsi.site/">
163
+ <img src="https://img.shields.io/badge/Visualizer-84C0B8?style=for-the-badge&logo=eye&logoColor=white">
164
+ </a>
165
+ </div>
166
+
167
+
168
+ This repository contains the <span style="color:#84C0B8;"><b>ReVSI</b></span> benchmark and dataset, introduced in [ReVSI: Rebuilding Visual Spatial Intelligence Evaluation for Accurate Assessment of VLM 3D Reasoning](https://3dlg-hcvc.github.io/revsi/).
169
+
170
+
171
+
172
+ ## Data Subsets
173
+ <span style="color:#84C0B8;"><b>ReVSI</b></span> provides multiple data subsets corresponding to different video frame budgets:
174
+ - all-frame
175
+ - 64-frame
176
+ - 32-frame
177
+ - 16-frame
178
+
179
+ Use the following command to load a specific subset:
180
+ ```python
181
+ from datasets import load_dataset
182
+ revsi_dataset = load_dataset("3dlg-hcvc/ReVSI", "64_frame", split="test") # load the 64-frame subset
183
+ ```
184
+
185
+ > [!NOTE]
186
+ > **How video subsets are constructed:**
187
+ >
188
+ > The **all-frame** subset contains the full processed video sequence for each scene, with standardized resolution and frame rate:
189
+ > 1. **ScanNet v2 / ScanNetPP v2 / MultiScan**
190
+ > *640 × 480 · 10 FPS*
191
+ >
192
+ > 2. **ARKitScenes**
193
+ > *640 × 480 / 480 × 640 · 10 FPS (all videos have been rotated to sky-up orientation)*
194
+ >
195
+ > 3. **3RScan**
196
+ > *360 × 640 · 4 FPS*
197
+ >
198
+ > The fixed-budget subsets are constructed via hierarchical uniform sampling:
199
+ > 1. Uniformly sample **64 frames** from **all-frame**
200
+ > 2. Uniformly subsample **32 frames** from the **64-frame** set
201
+ > 3. Uniformly subsample **16 frames** from the **32-frame** set
202
+ >
203
+ > This produces a nested structure: **16-frame** ⊂ **32-frame** ⊂ **64-frame** ⊂ **all-frame**.
204
+ > For each video, all subsets cover the same time span, and each sampled frame keeps the same timestamp across subsets. This guarantees consistent timestamps for models with frame timestamp encoding.
205
+
206
+ ## Data Fields
207
+ Each entry in <span style="color:#84C0B8;"><b>ReVSI</b></span> dataset contains the following fields:
208
+ | Field Name | Type | Description |
209
+ | :--------- | :--- | :---------- |
210
+ | `id` | int64 | Unique identifier for each sample |
211
+ | `dataset` | string | Source dataset of the video |
212
+ | `scene_id` | string | Identifier of the scene (video) associated with the sample |
213
+ | `question_type` | string | Category of the question |
214
+ | `question` | string | Natural language question grounded in the video |
215
+ | `options` | list[string] | List of answer choices (only for multiple-choice questions) |
216
+ | `ground_truth` | string | Ground-truth answer to the question |
217
+ | `num_frames` | string | Frame budget used for evaluation (e.g., 16, 32, 64, all) |
218
+ | `queried_object_ids` | list[int64] | List of object instance IDs referenced in the question |
219
+
220
+ ## Evaluation
221
+ > [!WARNING]
222
+ > Please avoid using PyTorch 2.9, as a known cuDNN issue can lead to significant performance degradation for QwenVL models (see [details](https://github.com/pytorch/pytorch/issues/166122)).
223
+
224
+ <span style="color:#84C0B8;"><b>ReVSI</b></span> supports inference / evaluation with the following frameworks:
225
+ - [LMMs-Eval](https://github.com/eamonn-zh/lmms-eval) (inference + evaluation)
226
+ ```bash
227
+ # example 1: evaluate Qwen3-VL-8B-Instruct on ReVSI 64-frame subset (with huggingface transformers backend on 4 GPUs)
228
+ accelerate launch \
229
+ --num_processes=4 \
230
+ -m lmms_eval \
231
+ --model qwen3_vl \
232
+ --model_args=pretrained=Qwen/Qwen3-VL-8B-Instruct,attn_implementation=flash_attention_2,max_num_frames=64 \
233
+ --tasks revsi_64_frame \
234
+ --batch_size 8
235
+
236
+ # example 2: evaluate Qwen3-VL-8B-Instruct on ReVSI all-frame subset using 2 fps sampling rate (with vllm backend)
237
+ python -m lmms_eval \
238
+ --model vllm \
239
+ --model_args "model=Qwen/Qwen3-VL-8B-Instruct,fps=2" \
240
+ --tasks revsi_all_frame
241
+ ```
242
+
243
+ - [VLMEvalKit](https://github.com/eamonn-zh/VLMEvalKit) (inference + evaluation)
244
+ ```bash
245
+ # example 1: evaluate Qwen3-VL-8B-Instruct on ReVSI 32-frame subset (with vllm backend)
246
+ python run.py --data revsi_32_frame --model Qwen3-VL-8B-Instruct
247
+ ```
248
+
249
+ - [ModelScope SWIFT](https://github.com/modelscope/ms-swift) (inference-only, check [ReVSI GitHub repo](https://github.com/3dlg-hcvc/revsi) for data registration)
250
+ ```bash
251
+ # example 1: infer Qwen3-VL-8B-Instruct on ReVSI 64-frame subset (with huggingface transformers backend on 4 GPUs)
252
+ NPROC_PER_NODE=4 swift infer \
253
+ --model Qwen/Qwen3-VL-8B-Instruct \
254
+ --model_kwargs '{"fps_min_frames": 64, "fps_max_frames": 64}' \
255
+ --val_dataset 3dlg-hcvc/ReVSI:64_frame \
256
+ --infer_backend transformers \
257
+ --custom_register_path ./ms_swift_register/revsi_register.py \
258
+ --use_hf true \
259
+ --torch_dtype bfloat16 \
260
+ --attn_impl flash_attention_2 \
261
+ --strict true \
262
+ --max_batch_size 8 \
263
+ --temperature 0
264
+ ```
265
+
266
+ - [TorchMetrics Extension](https://github.com/eamonn-zh/torchmetrics_ext) (evaluation-only)
267
+ ```python
268
+ # example 1: evaluate existing predictions on ReVSI all-frame subset using TorchMetrics Extension evaluator
269
+ from torchmetrics_ext.metrics.vqa import ReVSIMetric
270
+
271
+ metric = ReVSIMetric(subset=all_frame)
272
+ predictions = {0: "2", 1: "4", ..., 1000: "A"} # predictions should be a dict following the format {question_id: response}
273
+ results = metric(pred_dict)
274
+ ```
275
+
276
+ ## Metadata Files
277
+ We provide several metadata files used in constructing <span style="color:#84C0B8;"><b>ReVSI</b></span>:
278
+
279
+ - [metadata/3d_annotation.json](https://huggingface.co/datasets/3dlg-hcvc/ReVSI/blob/main/metadata/3d_annotation.json): 3D annotations for each scene, including object names, oriented bounding boxes and scene area polygons. The schema is as follows:
280
+ ```json
281
+ [
282
+ {
283
+ "scene_id": # scene ID from the source dataset
284
+ "dataset": # source dataset name
285
+ "scene_area_2d_polygon": # list of 2D boundary points (x, y) defining the scene area polygon, shape (N, 2)
286
+ "scene_area_type": # scene area annotation type (single_room or multiple_room)
287
+ "objects": [
288
+ {
289
+ "id": # object id within the scene
290
+ "name": # open-vocabulary object name
291
+ "obb": {
292
+ "center": # center of the object oriented bounding boxes, shape (3, )
293
+ "extent": # extent of the object oriented bounding boxes, shape (3, )
294
+ "rotation": # rotation matrix of the object oriented bounding boxes, shape (3, 3)
295
+ }
296
+ },
297
+ ...
298
+ ]
299
+ },
300
+ ...
301
+ ]
302
+ ```
303
+
304
+ - [metadata/sampled_video_frame_idx.json](https://huggingface.co/datasets/3dlg-hcvc/ReVSI/blob/main/metadata/sampled_video_frame_idx.json): indices of sampled frames for the 16/32/64-frame subsets. The scehema is as follows:
305
+ ```json
306
+ {
307
+ "<scene_id>": {
308
+ "64-frame": # list of sampled frame indices from the all-frame video, shape (64, )
309
+ "32-frame": # list of sampled frame indices from the all-frame video, shape (32, )
310
+ "16-frame": # list of sampled frame indices from the all-frame video, shape (16, )
311
+ }
312
+ ...
313
+ }
314
+ ```
315
+
316
+ - [metadata/obj_visibility.json](https://huggingface.co/datasets/3dlg-hcvc/ReVSI/blob/main/metadata/obj_visibility.json): Object visibility under different video frame budgets. The schema is as follows:
317
+ ```json
318
+ {
319
+ "<scene_id>": [
320
+ {
321
+ "object_id": # object id within the scene (consistent with metadata/3d_annotation.json)
322
+ "object_name": # open-vocabulary object name (consistent with metadata/3d_annotation.json)
323
+ "visibility_16": # visibility under the 16-frame budget
324
+ "visibility_32": # visibility under the 32-frame budget
325
+ "visibility_64": # visibility under the 64-frame budget
326
+ },
327
+ ...
328
+ ],
329
+ ...
330
+ }
331
+ ```
332
+
333
+ - [metadata/tiny_set_question_ids.txt](https://huggingface.co/datasets/3dlg-hcvc/ReVSI/blob/main/metadata/tiny_set_question_ids.txt): The sampled question ids of `tiny` set for proprietary model evaluations.
334
+
335
+ ## Citation
336
+ If you find <span style="color:#84C0B8;"><b>ReVSI</b></span> useful for your research, please consider citing:
337
+ ```bibtex
338
+ <TODO>
339
+ ```
340
+
341
+ <span style="color:#84C0B8;"><b>ReVSI</b></span> builds upon the following 3D scene datasets and the VSI-Bench benchmark, please also consider citing:
342
+ ```bibtex
343
+ @inproceedings{dai2017scannet,
344
+ title={Scannet: Richly-annotated 3d reconstructions of indoor scenes},
345
+ author={Dai, Angela and Chang, Angel X and Savva, Manolis and Halber, Maciej and Funkhouser, Thomas and Nie{\ss}ner, Matthias},
346
+ booktitle={Proceedings of the IEEE conference on computer vision and pattern recognition},
347
+ pages={5828--5839},
348
+ year={2017}
349
+ }
350
+
351
+ @inproceedings{yeshwanth2023scannet++,
352
+ title={Scannet++: A high-fidelity dataset of 3d indoor scenes},
353
+ author={Yeshwanth, Chandan and Liu, Yueh-Cheng and Nie{\ss}ner, Matthias and Dai, Angela},
354
+ booktitle={Proceedings of the IEEE/CVF International Conference on Computer Vision},
355
+ pages={12--22},
356
+ year={2023}
357
+ }
358
+
359
+ @inproceedings{baruch1arkitscenes,
360
+ title={ARKitScenes: A Diverse Real-World Dataset For 3D Indoor Scene Understanding Using Mobile RGB-D Data},
361
+ author={Baruch, Gilad and Chen, Zhuoyuan and Dehghan, Afshin and Feigin, Yuri and Fu, Peter and Gebauer, Thomas and Kurz, Daniel and Dimry, Tal and Joffe, Brandon and Schwartz, Arik and others},
362
+ booktitle={Thirty-fifth Conference on Neural Information Processing Systems Datasets and Benchmarks Track (Round 1)}
363
+ }
364
+
365
+ @inproceedings{wald2019rio,
366
+ title={Rio: 3d object instance re-localization in changing indoor environments},
367
+ author={Wald, Johanna and Avetisyan, Armen and Navab, Nassir and Tombari, Federico and Nie{\ss}ner, Matthias},
368
+ booktitle={Proceedings of the IEEE/CVF International Conference on Computer Vision},
369
+ pages={7658--7667},
370
+ year={2019}
371
+ }
372
+
373
+ @article{mao2022multiscan,
374
+ title={Multiscan: Scalable rgbd scanning for 3d environments with articulated objects},
375
+ author={Mao, Yongsen and Zhang, Yiming and Jiang, Hanxiao and Chang, Angel and Savva, Manolis},
376
+ journal={Advances in neural information processing systems},
377
+ volume={35},
378
+ pages={9058--9071},
379
+ year={2022}
380
+ }
381
+
382
+ @inproceedings{yang2025thinking,
383
+ title={Thinking in space: How multimodal large language models see, remember, and recall spaces},
384
+ author={Yang, Jihan and Yang, Shusheng and Gupta, Anjali W and Han, Rilyn and Fei-Fei, Li and Xie, Saining},
385
+ booktitle={Proceedings of the Computer Vision and Pattern Recognition Conference},
386
+ pages={10632--10643},
387
+ year={2025}
388
+ }
389
+ ```
all_frame/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d7b1daf6b4ab3cbb6191b775ee0d300745fa270439417c21dec034e27658b9e4
3
+ size 239453
metadata/3d_annotation.json ADDED
The diff for this file is too large to render. See raw diff
 
metadata/obj_visibility.json ADDED
The diff for this file is too large to render. See raw diff
 
metadata/revsi.png ADDED

Git LFS Details

  • SHA256: 8cf1711a68784ff8fcd482e41a63d29d59aa0db15c6ac04b142ecbe5e68be0b3
  • Pointer size: 131 Bytes
  • Size of remote file: 440 kB
metadata/sampled_video_frame_idx.json ADDED
The diff for this file is too large to render. See raw diff
 
metadata/tiny_set_question_ids.txt ADDED
@@ -0,0 +1,1093 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 1
2
+ 14
3
+ 15
4
+ 28
5
+ 41
6
+ 42
7
+ 52
8
+ 54
9
+ 66
10
+ 71
11
+ 79
12
+ 89
13
+ 95
14
+ 97
15
+ 98
16
+ 103
17
+ 107
18
+ 120
19
+ 124
20
+ 140
21
+ 149
22
+ 170
23
+ 186
24
+ 187
25
+ 188
26
+ 189
27
+ 203
28
+ 211
29
+ 212
30
+ 213
31
+ 217
32
+ 218
33
+ 223
34
+ 228
35
+ 238
36
+ 242
37
+ 246
38
+ 253
39
+ 273
40
+ 275
41
+ 277
42
+ 281
43
+ 288
44
+ 304
45
+ 315
46
+ 332
47
+ 334
48
+ 335
49
+ 360
50
+ 379
51
+ 389
52
+ 392
53
+ 395
54
+ 401
55
+ 414
56
+ 415
57
+ 424
58
+ 425
59
+ 426
60
+ 441
61
+ 444
62
+ 446
63
+ 448
64
+ 475
65
+ 541
66
+ 550
67
+ 552
68
+ 563
69
+ 566
70
+ 579
71
+ 581
72
+ 590
73
+ 597
74
+ 599
75
+ 609
76
+ 612
77
+ 623
78
+ 625
79
+ 640
80
+ 648
81
+ 650
82
+ 654
83
+ 662
84
+ 671
85
+ 685
86
+ 687
87
+ 700
88
+ 707
89
+ 708
90
+ 712
91
+ 726
92
+ 731
93
+ 733
94
+ 749
95
+ 769
96
+ 779
97
+ 784
98
+ 785
99
+ 800
100
+ 816
101
+ 832
102
+ 834
103
+ 843
104
+ 861
105
+ 869
106
+ 881
107
+ 886
108
+ 891
109
+ 894
110
+ 903
111
+ 908
112
+ 916
113
+ 919
114
+ 921
115
+ 924
116
+ 932
117
+ 941
118
+ 953
119
+ 960
120
+ 968
121
+ 976
122
+ 997
123
+ 998
124
+ 1002
125
+ 1003
126
+ 1009
127
+ 1025
128
+ 1035
129
+ 1039
130
+ 1045
131
+ 1049
132
+ 1050
133
+ 1063
134
+ 1065
135
+ 1067
136
+ 1094
137
+ 1099
138
+ 1103
139
+ 1108
140
+ 1116
141
+ 1119
142
+ 1122
143
+ 1125
144
+ 1126
145
+ 1128
146
+ 1131
147
+ 1132
148
+ 1133
149
+ 1138
150
+ 1139
151
+ 1140
152
+ 1141
153
+ 1147
154
+ 1149
155
+ 1151
156
+ 1163
157
+ 1168
158
+ 1174
159
+ 1179
160
+ 1185
161
+ 1189
162
+ 1195
163
+ 1199
164
+ 1201
165
+ 1202
166
+ 1205
167
+ 1209
168
+ 1212
169
+ 1214
170
+ 1218
171
+ 1223
172
+ 1226
173
+ 1227
174
+ 1230
175
+ 1231
176
+ 1238
177
+ 1239
178
+ 1246
179
+ 1247
180
+ 1258
181
+ 1259
182
+ 1264
183
+ 1265
184
+ 1269
185
+ 1274
186
+ 1278
187
+ 1280
188
+ 1281
189
+ 1283
190
+ 1284
191
+ 1287
192
+ 1290
193
+ 1291
194
+ 1292
195
+ 1293
196
+ 1300
197
+ 1301
198
+ 1304
199
+ 1307
200
+ 1313
201
+ 1315
202
+ 1319
203
+ 1334
204
+ 1338
205
+ 1349
206
+ 1350
207
+ 1357
208
+ 1359
209
+ 1366
210
+ 1367
211
+ 1368
212
+ 1378
213
+ 1381
214
+ 1382
215
+ 1384
216
+ 1388
217
+ 1391
218
+ 1396
219
+ 1402
220
+ 1407
221
+ 1413
222
+ 1415
223
+ 1419
224
+ 1424
225
+ 1425
226
+ 1429
227
+ 1431
228
+ 1433
229
+ 1437
230
+ 1443
231
+ 1448
232
+ 1455
233
+ 1457
234
+ 1467
235
+ 1471
236
+ 1474
237
+ 1475
238
+ 1476
239
+ 1483
240
+ 1484
241
+ 1485
242
+ 1489
243
+ 1500
244
+ 1507
245
+ 1508
246
+ 1515
247
+ 1523
248
+ 1524
249
+ 1526
250
+ 1527
251
+ 1530
252
+ 1531
253
+ 1534
254
+ 1542
255
+ 1543
256
+ 1545
257
+ 1546
258
+ 1551
259
+ 1555
260
+ 1558
261
+ 1560
262
+ 1563
263
+ 1564
264
+ 1570
265
+ 1572
266
+ 1573
267
+ 1577
268
+ 1578
269
+ 1588
270
+ 1597
271
+ 1602
272
+ 1603
273
+ 1610
274
+ 1614
275
+ 1615
276
+ 1620
277
+ 1622
278
+ 1631
279
+ 1633
280
+ 1636
281
+ 1640
282
+ 1649
283
+ 1652
284
+ 1661
285
+ 1664
286
+ 1665
287
+ 1668
288
+ 1669
289
+ 1674
290
+ 1675
291
+ 1676
292
+ 1678
293
+ 1680
294
+ 1688
295
+ 1690
296
+ 1695
297
+ 1696
298
+ 1700
299
+ 1708
300
+ 1709
301
+ 1714
302
+ 1716
303
+ 1721
304
+ 1722
305
+ 1725
306
+ 1728
307
+ 1736
308
+ 1739
309
+ 1740
310
+ 1742
311
+ 1744
312
+ 1746
313
+ 1748
314
+ 1750
315
+ 1752
316
+ 1755
317
+ 1756
318
+ 1757
319
+ 1761
320
+ 1762
321
+ 1765
322
+ 1773
323
+ 1779
324
+ 1784
325
+ 1786
326
+ 1787
327
+ 1791
328
+ 1792
329
+ 1796
330
+ 1798
331
+ 1803
332
+ 1806
333
+ 1809
334
+ 1812
335
+ 1813
336
+ 1815
337
+ 1820
338
+ 1827
339
+ 1828
340
+ 1829
341
+ 1832
342
+ 1834
343
+ 1836
344
+ 1840
345
+ 1850
346
+ 1851
347
+ 1853
348
+ 1857
349
+ 1859
350
+ 1863
351
+ 1865
352
+ 1870
353
+ 1877
354
+ 1882
355
+ 1883
356
+ 1884
357
+ 1892
358
+ 1897
359
+ 1899
360
+ 1913
361
+ 1916
362
+ 1917
363
+ 1921
364
+ 1922
365
+ 1926
366
+ 1930
367
+ 1938
368
+ 1941
369
+ 1948
370
+ 1952
371
+ 1953
372
+ 1954
373
+ 1963
374
+ 1965
375
+ 1972
376
+ 1976
377
+ 1990
378
+ 1996
379
+ 2001
380
+ 2002
381
+ 2022
382
+ 2044
383
+ 2050
384
+ 2052
385
+ 2058
386
+ 2062
387
+ 2065
388
+ 2071
389
+ 2079
390
+ 2083
391
+ 2086
392
+ 2087
393
+ 2092
394
+ 2100
395
+ 2103
396
+ 2104
397
+ 2110
398
+ 2112
399
+ 2133
400
+ 2139
401
+ 2144
402
+ 2148
403
+ 2149
404
+ 2153
405
+ 2157
406
+ 2160
407
+ 2162
408
+ 2165
409
+ 2173
410
+ 2180
411
+ 2182
412
+ 2183
413
+ 2187
414
+ 2188
415
+ 2211
416
+ 2216
417
+ 2220
418
+ 2224
419
+ 2230
420
+ 2233
421
+ 2240
422
+ 2249
423
+ 2251
424
+ 2261
425
+ 2262
426
+ 2265
427
+ 2266
428
+ 2268
429
+ 2269
430
+ 2272
431
+ 2273
432
+ 2280
433
+ 2281
434
+ 2283
435
+ 2286
436
+ 2294
437
+ 2304
438
+ 2307
439
+ 2312
440
+ 2313
441
+ 2318
442
+ 2320
443
+ 2321
444
+ 2322
445
+ 2324
446
+ 2336
447
+ 2339
448
+ 2345
449
+ 2346
450
+ 2351
451
+ 2352
452
+ 2353
453
+ 2359
454
+ 2363
455
+ 2366
456
+ 2367
457
+ 2369
458
+ 2370
459
+ 2372
460
+ 2381
461
+ 2386
462
+ 2391
463
+ 2393
464
+ 2400
465
+ 2401
466
+ 2403
467
+ 2409
468
+ 2414
469
+ 2417
470
+ 2418
471
+ 2420
472
+ 2422
473
+ 2425
474
+ 2432
475
+ 2435
476
+ 2440
477
+ 2441
478
+ 2444
479
+ 2448
480
+ 2451
481
+ 2455
482
+ 2457
483
+ 2460
484
+ 2466
485
+ 2472
486
+ 2474
487
+ 2476
488
+ 2477
489
+ 2481
490
+ 2487
491
+ 2492
492
+ 2493
493
+ 2500
494
+ 2505
495
+ 2507
496
+ 2513
497
+ 2514
498
+ 2516
499
+ 2522
500
+ 2524
501
+ 2533
502
+ 2549
503
+ 2550
504
+ 2562
505
+ 2569
506
+ 2572
507
+ 2579
508
+ 2588
509
+ 2605
510
+ 2611
511
+ 2612
512
+ 2616
513
+ 2617
514
+ 2621
515
+ 2625
516
+ 2628
517
+ 2634
518
+ 2639
519
+ 2640
520
+ 2643
521
+ 2647
522
+ 2657
523
+ 2658
524
+ 2659
525
+ 2685
526
+ 2686
527
+ 2687
528
+ 2690
529
+ 2692
530
+ 2694
531
+ 2696
532
+ 2702
533
+ 2705
534
+ 2707
535
+ 2710
536
+ 2719
537
+ 2721
538
+ 2722
539
+ 2733
540
+ 2735
541
+ 2736
542
+ 2738
543
+ 2742
544
+ 2744
545
+ 2746
546
+ 2749
547
+ 2750
548
+ 2752
549
+ 2754
550
+ 2755
551
+ 2756
552
+ 2762
553
+ 2767
554
+ 2773
555
+ 2781
556
+ 2792
557
+ 2794
558
+ 2806
559
+ 2812
560
+ 2814
561
+ 2823
562
+ 2826
563
+ 2830
564
+ 2831
565
+ 2838
566
+ 2859
567
+ 2860
568
+ 2865
569
+ 2866
570
+ 2872
571
+ 2873
572
+ 2876
573
+ 2877
574
+ 2880
575
+ 2882
576
+ 2883
577
+ 2885
578
+ 2887
579
+ 2892
580
+ 2895
581
+ 2898
582
+ 2904
583
+ 2910
584
+ 2916
585
+ 2926
586
+ 2927
587
+ 2942
588
+ 2943
589
+ 2944
590
+ 2945
591
+ 2947
592
+ 2950
593
+ 2959
594
+ 2962
595
+ 2965
596
+ 2970
597
+ 2971
598
+ 2977
599
+ 2979
600
+ 2982
601
+ 2983
602
+ 2988
603
+ 2989
604
+ 2994
605
+ 2997
606
+ 2999
607
+ 3002
608
+ 3006
609
+ 3007
610
+ 3009
611
+ 3018
612
+ 3020
613
+ 3021
614
+ 3024
615
+ 3027
616
+ 3036
617
+ 3047
618
+ 3048
619
+ 3058
620
+ 3063
621
+ 3065
622
+ 3068
623
+ 3072
624
+ 3073
625
+ 3074
626
+ 3076
627
+ 3082
628
+ 3088
629
+ 3096
630
+ 3100
631
+ 3101
632
+ 3103
633
+ 3104
634
+ 3117
635
+ 3119
636
+ 3121
637
+ 3123
638
+ 3125
639
+ 3129
640
+ 3132
641
+ 3133
642
+ 3141
643
+ 3142
644
+ 3145
645
+ 3147
646
+ 3151
647
+ 3155
648
+ 3160
649
+ 3161
650
+ 3172
651
+ 3174
652
+ 3176
653
+ 3180
654
+ 3183
655
+ 3186
656
+ 3189
657
+ 3190
658
+ 3193
659
+ 3194
660
+ 3198
661
+ 3207
662
+ 3212
663
+ 3215
664
+ 3219
665
+ 3222
666
+ 3230
667
+ 3231
668
+ 3234
669
+ 3235
670
+ 3238
671
+ 3239
672
+ 3240
673
+ 3244
674
+ 3245
675
+ 3249
676
+ 3251
677
+ 3252
678
+ 3254
679
+ 3261
680
+ 3263
681
+ 3264
682
+ 3267
683
+ 3277
684
+ 3281
685
+ 3284
686
+ 3290
687
+ 3295
688
+ 3300
689
+ 3301
690
+ 3305
691
+ 3307
692
+ 3309
693
+ 3310
694
+ 3313
695
+ 3316
696
+ 3324
697
+ 3333
698
+ 3334
699
+ 3343
700
+ 3347
701
+ 3350
702
+ 3359
703
+ 3360
704
+ 3362
705
+ 3369
706
+ 3371
707
+ 3377
708
+ 3378
709
+ 3380
710
+ 3389
711
+ 3390
712
+ 3398
713
+ 3399
714
+ 3401
715
+ 3408
716
+ 3411
717
+ 3412
718
+ 3415
719
+ 3420
720
+ 3422
721
+ 3423
722
+ 3424
723
+ 3427
724
+ 3432
725
+ 3435
726
+ 3456
727
+ 3474
728
+ 3479
729
+ 3499
730
+ 3511
731
+ 3518
732
+ 3532
733
+ 3546
734
+ 3555
735
+ 3562
736
+ 3577
737
+ 3586
738
+ 3588
739
+ 3597
740
+ 3609
741
+ 3613
742
+ 3637
743
+ 3655
744
+ 3657
745
+ 3659
746
+ 3660
747
+ 3668
748
+ 3669
749
+ 3674
750
+ 3680
751
+ 3685
752
+ 3694
753
+ 3698
754
+ 3703
755
+ 3710
756
+ 3726
757
+ 3727
758
+ 3728
759
+ 3745
760
+ 3754
761
+ 3760
762
+ 3762
763
+ 3768
764
+ 3799
765
+ 3812
766
+ 3814
767
+ 3818
768
+ 3828
769
+ 3829
770
+ 3831
771
+ 3836
772
+ 3841
773
+ 3845
774
+ 3849
775
+ 3851
776
+ 3865
777
+ 3867
778
+ 3880
779
+ 3897
780
+ 3898
781
+ 3903
782
+ 3905
783
+ 3917
784
+ 3920
785
+ 3923
786
+ 3925
787
+ 3927
788
+ 3931
789
+ 3949
790
+ 3975
791
+ 4000
792
+ 4002
793
+ 4014
794
+ 4015
795
+ 4029
796
+ 4033
797
+ 4039
798
+ 4040
799
+ 4043
800
+ 4046
801
+ 4054
802
+ 4056
803
+ 4064
804
+ 4076
805
+ 4094
806
+ 4095
807
+ 4097
808
+ 4114
809
+ 4123
810
+ 4125
811
+ 4151
812
+ 4155
813
+ 4157
814
+ 4158
815
+ 4177
816
+ 4184
817
+ 4192
818
+ 4197
819
+ 4210
820
+ 4224
821
+ 4225
822
+ 4226
823
+ 4229
824
+ 4230
825
+ 4236
826
+ 4256
827
+ 4271
828
+ 4275
829
+ 4277
830
+ 4286
831
+ 4294
832
+ 4298
833
+ 4299
834
+ 4302
835
+ 4307
836
+ 4310
837
+ 4334
838
+ 4339
839
+ 4387
840
+ 4398
841
+ 4402
842
+ 4409
843
+ 4412
844
+ 4421
845
+ 4426
846
+ 4430
847
+ 4445
848
+ 4451
849
+ 4462
850
+ 4466
851
+ 4469
852
+ 4481
853
+ 4488
854
+ 4489
855
+ 4491
856
+ 4500
857
+ 4515
858
+ 4523
859
+ 4546
860
+ 4549
861
+ 4554
862
+ 4555
863
+ 4576
864
+ 4580
865
+ 4581
866
+ 4595
867
+ 4597
868
+ 4598
869
+ 4608
870
+ 4625
871
+ 4636
872
+ 4638
873
+ 4645
874
+ 4650
875
+ 4654
876
+ 4656
877
+ 4659
878
+ 4676
879
+ 4677
880
+ 4683
881
+ 4687
882
+ 4714
883
+ 4731
884
+ 4739
885
+ 4741
886
+ 4743
887
+ 4754
888
+ 4755
889
+ 4759
890
+ 4769
891
+ 4779
892
+ 4780
893
+ 4783
894
+ 4800
895
+ 4881
896
+ 4889
897
+ 4895
898
+ 4913
899
+ 4918
900
+ 4942
901
+ 4948
902
+ 4952
903
+ 4955
904
+ 4957
905
+ 4964
906
+ 4979
907
+ 4991
908
+ 5018
909
+ 5019
910
+ 5026
911
+ 5033
912
+ 5041
913
+ 5051
914
+ 5065
915
+ 5081
916
+ 5092
917
+ 5105
918
+ 5118
919
+ 5188
920
+ 5208
921
+ 5237
922
+ 5250
923
+ 5265
924
+ 5267
925
+ 5273
926
+ 5290
927
+ 5302
928
+ 5366
929
+ 5405
930
+ 5420
931
+ 5435
932
+ 5453
933
+ 5459
934
+ 5476
935
+ 5487
936
+ 5490
937
+ 5496
938
+ 5497
939
+ 5540
940
+ 5545
941
+ 5553
942
+ 5567
943
+ 5578
944
+ 5580
945
+ 5594
946
+ 5618
947
+ 5624
948
+ 5649
949
+ 5656
950
+ 5663
951
+ 5673
952
+ 5687
953
+ 5688
954
+ 5710
955
+ 5720
956
+ 5757
957
+ 5780
958
+ 5784
959
+ 5820
960
+ 5880
961
+ 5884
962
+ 5895
963
+ 5908
964
+ 5996
965
+ 6002
966
+ 6005
967
+ 6008
968
+ 6010
969
+ 6063
970
+ 6075
971
+ 6086
972
+ 6095
973
+ 6099
974
+ 6113
975
+ 6119
976
+ 6161
977
+ 6183
978
+ 6190
979
+ 6194
980
+ 6197
981
+ 6198
982
+ 6199
983
+ 6206
984
+ 6212
985
+ 6213
986
+ 6225
987
+ 6233
988
+ 6234
989
+ 6238
990
+ 6248
991
+ 6257
992
+ 6261
993
+ 6265
994
+ 6272
995
+ 6287
996
+ 6291
997
+ 6293
998
+ 6299
999
+ 6304
1000
+ 6306
1001
+ 6309
1002
+ 6312
1003
+ 6315
1004
+ 6316
1005
+ 6318
1006
+ 6321
1007
+ 6323
1008
+ 6324
1009
+ 6325
1010
+ 6330
1011
+ 6335
1012
+ 6341
1013
+ 6346
1014
+ 6347
1015
+ 6349
1016
+ 6350
1017
+ 6352
1018
+ 6353
1019
+ 6359
1020
+ 6367
1021
+ 6368
1022
+ 6373
1023
+ 6378
1024
+ 6387
1025
+ 6390
1026
+ 6391
1027
+ 6392
1028
+ 6398
1029
+ 6402
1030
+ 6405
1031
+ 6410
1032
+ 6411
1033
+ 6415
1034
+ 6419
1035
+ 6421
1036
+ 6427
1037
+ 6428
1038
+ 6450
1039
+ 6452
1040
+ 6455
1041
+ 6459
1042
+ 6460
1043
+ 6462
1044
+ 6466
1045
+ 6473
1046
+ 6474
1047
+ 6479
1048
+ 6482
1049
+ 6487
1050
+ 6489
1051
+ 6495
1052
+ 6497
1053
+ 6499
1054
+ 6501
1055
+ 6503
1056
+ 6504
1057
+ 6507
1058
+ 6517
1059
+ 6523
1060
+ 6538
1061
+ 6558
1062
+ 6571
1063
+ 6578
1064
+ 6593
1065
+ 6600
1066
+ 6602
1067
+ 6606
1068
+ 6626
1069
+ 6629
1070
+ 6630
1071
+ 6646
1072
+ 6650
1073
+ 6656
1074
+ 6658
1075
+ 6660
1076
+ 6665
1077
+ 6669
1078
+ 6683
1079
+ 6695
1080
+ 6699
1081
+ 6700
1082
+ 6710
1083
+ 6715
1084
+ 6725
1085
+ 6726
1086
+ 6732
1087
+ 6765
1088
+ 6767
1089
+ 6768
1090
+ 6775
1091
+ 6792
1092
+ 6803
1093
+ 6804
video.zip ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7d10f8a538b1a03e83c8c01b9c11a595db6b49d661e22f2dc513358f8eabb9fe
3
+ size 4866738106