Yiwei-Ou commited on
Commit
6bb11b4
·
verified ·
1 Parent(s): f711652

Upload Urban ImageNet_Croissant Metadata.json

Browse files
Urban ImageNet_Croissant Metadata.json ADDED
@@ -0,0 +1,886 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "@context": {
3
+ "@language": "en",
4
+ "@vocab": "https://schema.org/",
5
+ "sc": "https://schema.org/",
6
+ "cr": "http://mlcommons.org/croissant/",
7
+ "rai": "http://mlcommons.org/croissant/RAI/",
8
+ "dct": "http://purl.org/dc/terms/",
9
+ "prov": "http://www.w3.org/ns/prov#",
10
+ "citeAs": "cr:citeAs",
11
+ "column": "cr:column",
12
+ "conformsTo": "dct:conformsTo",
13
+ "containedIn": "cr:containedIn",
14
+ "data": {
15
+ "@id": "cr:data",
16
+ "@type": "@json"
17
+ },
18
+ "dataType": {
19
+ "@id": "cr:dataType",
20
+ "@type": "@vocab"
21
+ },
22
+ "examples": {
23
+ "@id": "cr:examples",
24
+ "@type": "@json"
25
+ },
26
+ "extract": "cr:extract",
27
+ "field": "cr:field",
28
+ "fileObject": "cr:fileObject",
29
+ "fileProperty": "cr:fileProperty",
30
+ "fileSet": "cr:fileSet",
31
+ "format": "cr:format",
32
+ "includes": "cr:includes",
33
+ "isLiveDataset": "cr:isLiveDataset",
34
+ "jsonPath": "cr:jsonPath",
35
+ "key": "cr:key",
36
+ "md5": "cr:md5",
37
+ "recordSet": "cr:recordSet",
38
+ "references": "cr:references",
39
+ "regex": "cr:regex",
40
+ "repeated": "cr:repeated",
41
+ "source": "cr:source",
42
+ "subField": "cr:subField",
43
+ "transform": "cr:transform",
44
+ "value": "cr:value"
45
+ },
46
+ "@type": "sc:Dataset",
47
+ "conformsTo": "http://mlcommons.org/croissant/1.1",
48
+ "name": "Urban-ImageNet",
49
+ "alternateName": "Urban-ImageNet: A Large-Scale Multi-Modal Dataset for Urban Space Perception Benchmarking",
50
+ "url": "https://huggingface.co/datasets/Yiwei-Ou/Urban-ImageNet",
51
+ "sameAs": [
52
+ "https://huggingface.co/datasets/Yiwei-Ou/Urban-ImageNet"
53
+ ],
54
+ "license": "https://creativecommons.org/licenses/by-nc/4.0/",
55
+ "usageInfo": "For non-commercial academic research only. Prohibited uses include re-identification, account reconstruction, surveillance, facial recognition, social scoring, law-enforcement targeting, and commercial profiling. Users must respect platform policies and the dataset data-use agreement.",
56
+ "version": "1.0",
57
+ "description": "Urban-ImageNet is a large-scale multimodal dataset and benchmark for urban commercial space perception. It contains over 2 million public Weibo image-text pairs collected from 61 commercial sites in 24 Chinese cities across 2019-2025. The dataset is organized by the HUSIC 10-class taxonomy and supports urban scene classification, category-level and post-level image-text retrieval, instance segmentation, and scaling studies through 1K, 10K, 100K, and full 2M variants.",
58
+ "dateCreated": "2019-01-01",
59
+ "datePublished": "2026-05-05",
60
+ "dateModified": "2026-05-05",
61
+ "inLanguage": [
62
+ "zh",
63
+ "en"
64
+ ],
65
+ "keywords": [
66
+ "urban perception",
67
+ "image classification",
68
+ "image-text retrieval",
69
+ "instance segmentation",
70
+ "social media",
71
+ "Weibo",
72
+ "Chinese cities",
73
+ "urban commercial spaces",
74
+ "multimodal learning",
75
+ "computational urban studies"
76
+ ],
77
+ "creator": [
78
+ {
79
+ "@type": "sc:Organization",
80
+ "name": "Urban-ImageNet Research Team"
81
+ }
82
+ ],
83
+ "publisher": {
84
+ "@type": "sc:Organization",
85
+ "name": "Urban-ImageNet Research Team"
86
+ },
87
+ "spatialCoverage": {
88
+ "@type": "sc:Place",
89
+ "name": "24 Chinese cities and 61 urban commercial sites",
90
+ "geo": {
91
+ "@type": "sc:GeoShape",
92
+ "box": "18.0 73.0 54.0 135.0"
93
+ }
94
+ },
95
+ "temporalCoverage": "2019-01-01/2025-12-31",
96
+ "isAccessibleForFree": true,
97
+ "isLiveDataset": false,
98
+ "distribution": [
99
+ {
100
+ "@type": "cr:FileObject",
101
+ "@id": "1k-sample-zip",
102
+ "name": "1K Dataset-Sample.zip",
103
+ "description": "Compressed 1K sample package for quick inspection and validator-friendly access.",
104
+ "contentUrl": "https://huggingface.co/datasets/Yiwei-Ou/Urban-ImageNet/resolve/main/1K%20Dataset-Sample.zip",
105
+ "encodingFormat": "application/zip",
106
+ "contentSize": "71744171 B",
107
+ "sha256": "ed0dd2c1087ad56c7d5098a3990cb6b871dff0418fd542a4cbe07bff39c6ee2a"
108
+ },
109
+ {
110
+ "@type": "cr:FileSet",
111
+ "@id": "1k-t1-image-files",
112
+ "name": "1K Dataset Task 1 image files",
113
+ "description": "ImageFolder-style JPG files under train/val/test and HUSIC class folders.",
114
+ "contentUrl": "https://huggingface.co/datasets/Yiwei-Ou/Urban-ImageNet/tree/main/1K%20Dataset/01%20Images%20with%20labels",
115
+ "includes": [
116
+ "train/*/*.jpg",
117
+ "val/*/*.jpg",
118
+ "test/*/*.jpg"
119
+ ],
120
+ "encodingFormat": "image/jpeg"
121
+ },
122
+ {
123
+ "@type": "cr:FileSet",
124
+ "@id": "1k-t2-spreadsheets",
125
+ "name": "1K Dataset Task 2 text-image spreadsheets",
126
+ "description": "Excel files train.xlsx, val.xlsx, and test.xlsx with image-text pair metadata.",
127
+ "contentUrl": "https://huggingface.co/datasets/Yiwei-Ou/Urban-ImageNet/tree/main/1K%20Dataset/02%20Text-Image%20Pairs",
128
+ "includes": [
129
+ "train.xlsx",
130
+ "val.xlsx",
131
+ "test.xlsx"
132
+ ],
133
+ "encodingFormat": "application/vnd.openxmlformats-officedocument.spreadsheetml.sheet"
134
+ },
135
+ {
136
+ "@type": "cr:FileSet",
137
+ "@id": "1k-t3-json",
138
+ "name": "1K Dataset Task 3 instance segmentation JSON files",
139
+ "description": "COCO-style JSON annotation files train.json, val.json, and test.json.",
140
+ "contentUrl": "https://huggingface.co/datasets/Yiwei-Ou/Urban-ImageNet/tree/main/1K%20Dataset/03%20Instance%20Segmentation",
141
+ "includes": [
142
+ "train.json",
143
+ "val.json",
144
+ "test.json"
145
+ ],
146
+ "encodingFormat": "application/json"
147
+ },
148
+ {
149
+ "@type": "cr:FileSet",
150
+ "@id": "10k-t1-image-files",
151
+ "name": "10K Dataset Task 1 image files",
152
+ "description": "ImageFolder-style JPG files under train/val/test and HUSIC class folders.",
153
+ "contentUrl": "https://huggingface.co/datasets/Yiwei-Ou/Urban-ImageNet/tree/main/10K%20Dataset/01%20Images%20with%20labels",
154
+ "includes": [
155
+ "train/*/*.jpg",
156
+ "val/*/*.jpg",
157
+ "test/*/*.jpg"
158
+ ],
159
+ "encodingFormat": "image/jpeg"
160
+ },
161
+ {
162
+ "@type": "cr:FileSet",
163
+ "@id": "10k-t2-spreadsheets",
164
+ "name": "10K Dataset Task 2 text-image spreadsheets",
165
+ "description": "Excel files train.xlsx, val.xlsx, and test.xlsx with image-text pair metadata.",
166
+ "contentUrl": "https://huggingface.co/datasets/Yiwei-Ou/Urban-ImageNet/tree/main/10K%20Dataset/02%20Text-Image%20Pairs",
167
+ "includes": [
168
+ "train.xlsx",
169
+ "val.xlsx",
170
+ "test.xlsx"
171
+ ],
172
+ "encodingFormat": "application/vnd.openxmlformats-officedocument.spreadsheetml.sheet"
173
+ },
174
+ {
175
+ "@type": "cr:FileSet",
176
+ "@id": "10k-t3-json",
177
+ "name": "10K Dataset Task 3 instance segmentation JSON files",
178
+ "description": "COCO-style JSON annotation files train.json, val.json, and test.json.",
179
+ "contentUrl": "https://huggingface.co/datasets/Yiwei-Ou/Urban-ImageNet/tree/main/10K%20Dataset/03%20Instance%20Segmentation",
180
+ "includes": [
181
+ "train.json",
182
+ "val.json",
183
+ "test.json"
184
+ ],
185
+ "encodingFormat": "application/json"
186
+ },
187
+ {
188
+ "@type": "cr:FileSet",
189
+ "@id": "100k-t1-image-files",
190
+ "name": "100K Dataset Task 1 image files",
191
+ "description": "ImageFolder-style JPG files under train/val/test and HUSIC class folders.",
192
+ "contentUrl": "https://huggingface.co/datasets/Yiwei-Ou/Urban-ImageNet/tree/main/100K%20Dataset/01%20Images%20with%20labels",
193
+ "includes": [
194
+ "train/*/*.jpg",
195
+ "val/*/*.jpg",
196
+ "test/*/*.jpg"
197
+ ],
198
+ "encodingFormat": "image/jpeg"
199
+ },
200
+ {
201
+ "@type": "cr:FileSet",
202
+ "@id": "100k-t2-spreadsheets",
203
+ "name": "100K Dataset Task 2 text-image spreadsheets",
204
+ "description": "Excel files train.xlsx, val.xlsx, and test.xlsx with image-text pair metadata.",
205
+ "contentUrl": "https://huggingface.co/datasets/Yiwei-Ou/Urban-ImageNet/tree/main/100K%20Dataset/02%20Text-Image%20Pairs",
206
+ "includes": [
207
+ "train.xlsx",
208
+ "val.xlsx",
209
+ "test.xlsx"
210
+ ],
211
+ "encodingFormat": "application/vnd.openxmlformats-officedocument.spreadsheetml.sheet"
212
+ },
213
+ {
214
+ "@type": "cr:FileSet",
215
+ "@id": "100k-t3-json",
216
+ "name": "100K Dataset Task 3 instance segmentation JSON files",
217
+ "description": "COCO-style JSON annotation files train.json, val.json, and test.json.",
218
+ "contentUrl": "https://huggingface.co/datasets/Yiwei-Ou/Urban-ImageNet/tree/main/100K%20Dataset/03%20Instance%20Segmentation",
219
+ "includes": [
220
+ "train.json",
221
+ "val.json",
222
+ "test.json"
223
+ ],
224
+ "encodingFormat": "application/json"
225
+ },
226
+ {
227
+ "@type": "cr:FileSet",
228
+ "@id": "full-2m-image-files",
229
+ "name": "Full Dataset-2M image files",
230
+ "description": "Flat directory containing all privacy-protected 512 px images in the full unbalanced corpus.",
231
+ "contentUrl": "https://huggingface.co/datasets/Yiwei-Ou/Urban-ImageNet/tree/main/Full%20Dataset-2M/Images",
232
+ "includes": [
233
+ "*.jpg",
234
+ "*.jpeg"
235
+ ],
236
+ "encodingFormat": "image/jpeg"
237
+ },
238
+ {
239
+ "@type": "cr:FileSet",
240
+ "@id": "full-2m-label-files",
241
+ "name": "Full Dataset-2M label files",
242
+ "description": "CSV label files for semantic classification, text-image pairs, and instance segmentation metadata.",
243
+ "contentUrl": "https://huggingface.co/datasets/Yiwei-Ou/Urban-ImageNet/tree/main/Full%20Dataset-2M/Labels",
244
+ "includes": [
245
+ "01 Semantic classification labels.CSV",
246
+ "02 Text-Image Pairs.CSV",
247
+ "03 Instance Segmentation labels.CSV"
248
+ ],
249
+ "encodingFormat": "text/csv"
250
+ }
251
+ ],
252
+ "recordSet": [
253
+ {
254
+ "@type": "cr:RecordSet",
255
+ "@id": "dataset-splits",
256
+ "name": "DatasetSplitRecords",
257
+ "description": "Inline summary of the four Urban-ImageNet scale variants.",
258
+ "key": {
259
+ "@id": "dataset-splits/split_id"
260
+ },
261
+ "field": [
262
+ {
263
+ "@type": "cr:Field",
264
+ "@id": "dataset-splits/split_id",
265
+ "name": "split_id",
266
+ "description": "Stable split identifier.",
267
+ "dataType": "sc:Text"
268
+ },
269
+ {
270
+ "@type": "cr:Field",
271
+ "@id": "dataset-splits/name",
272
+ "name": "name",
273
+ "description": "Human-readable split name.",
274
+ "dataType": "sc:Text"
275
+ },
276
+ {
277
+ "@type": "cr:Field",
278
+ "@id": "dataset-splits/image_count",
279
+ "name": "image_count",
280
+ "description": "Primary image count.",
281
+ "dataType": "sc:Integer"
282
+ },
283
+ {
284
+ "@type": "cr:Field",
285
+ "@id": "dataset-splits/balanced",
286
+ "name": "balanced",
287
+ "description": "Whether the 10 classes are balanced.",
288
+ "dataType": "sc:Boolean"
289
+ },
290
+ {
291
+ "@type": "cr:Field",
292
+ "@id": "dataset-splits/predefined_split",
293
+ "name": "predefined_split",
294
+ "description": "Whether train/validation/test partitions are provided.",
295
+ "dataType": "sc:Boolean"
296
+ },
297
+ {
298
+ "@type": "cr:Field",
299
+ "@id": "dataset-splits/train_count",
300
+ "name": "train_count",
301
+ "description": "Training image count when predefined.",
302
+ "dataType": "sc:Integer"
303
+ },
304
+ {
305
+ "@type": "cr:Field",
306
+ "@id": "dataset-splits/validation_count",
307
+ "name": "validation_count",
308
+ "description": "Validation image count when predefined.",
309
+ "dataType": "sc:Integer"
310
+ },
311
+ {
312
+ "@type": "cr:Field",
313
+ "@id": "dataset-splits/test_count",
314
+ "name": "test_count",
315
+ "description": "Test image count when predefined.",
316
+ "dataType": "sc:Integer"
317
+ },
318
+ {
319
+ "@type": "cr:Field",
320
+ "@id": "dataset-splits/local_bytes",
321
+ "name": "local_bytes",
322
+ "description": "Local processed package size in bytes when measured.",
323
+ "dataType": "sc:Integer"
324
+ },
325
+ {
326
+ "@type": "cr:Field",
327
+ "@id": "dataset-splits/notes",
328
+ "name": "notes",
329
+ "description": "Usage notes for the split.",
330
+ "dataType": "sc:Text"
331
+ }
332
+ ],
333
+ "data": [
334
+ {
335
+ "dataset-splits/split_id": "1k",
336
+ "dataset-splits/name": "1K Dataset",
337
+ "dataset-splits/image_count": 1000,
338
+ "dataset-splits/balanced": true,
339
+ "dataset-splits/predefined_split": true,
340
+ "dataset-splits/train_count": 800,
341
+ "dataset-splits/validation_count": 100,
342
+ "dataset-splits/test_count": 100,
343
+ "dataset-splits/local_bytes": 91533081,
344
+ "dataset-splits/notes": "Balanced 10-class subset with 100 images per class; includes T1, T2, and T3 annotations."
345
+ },
346
+ {
347
+ "dataset-splits/split_id": "10k",
348
+ "dataset-splits/name": "10K Dataset",
349
+ "dataset-splits/image_count": 10000,
350
+ "dataset-splits/balanced": true,
351
+ "dataset-splits/predefined_split": true,
352
+ "dataset-splits/train_count": 8000,
353
+ "dataset-splits/validation_count": 1000,
354
+ "dataset-splits/test_count": 1000,
355
+ "dataset-splits/local_bytes": 734595582,
356
+ "dataset-splits/notes": "Balanced 10-class subset with 1,000 images per class; includes T1, T2, and T3 annotations."
357
+ },
358
+ {
359
+ "dataset-splits/split_id": "100k",
360
+ "dataset-splits/name": "100K Dataset",
361
+ "dataset-splits/image_count": 100000,
362
+ "dataset-splits/balanced": true,
363
+ "dataset-splits/predefined_split": true,
364
+ "dataset-splits/train_count": 80000,
365
+ "dataset-splits/validation_count": 10000,
366
+ "dataset-splits/test_count": 10000,
367
+ "dataset-splits/local_bytes": 6264915204,
368
+ "dataset-splits/notes": "Balanced benchmark subset with 10,000 images per class; primary split for reported baselines."
369
+ },
370
+ {
371
+ "dataset-splits/split_id": "full-2m",
372
+ "dataset-splits/name": "Full Dataset-2M",
373
+ "dataset-splits/image_count": 2000000,
374
+ "dataset-splits/balanced": false,
375
+ "dataset-splits/predefined_split": false,
376
+ "dataset-splits/train_count": null,
377
+ "dataset-splits/validation_count": null,
378
+ "dataset-splits/test_count": null,
379
+ "dataset-splits/local_bytes": null,
380
+ "dataset-splits/notes": "Full unbalanced corpus with flat Images and Labels folders; users create task-specific splits."
381
+ }
382
+ ]
383
+ },
384
+ {
385
+ "@type": "cr:RecordSet",
386
+ "@id": "husic-classes",
387
+ "name": "HUSICClassRecords",
388
+ "description": "The 10-class HUSIC taxonomy used by all tasks.",
389
+ "key": {
390
+ "@id": "husic-classes/class_id"
391
+ },
392
+ "field": [
393
+ {
394
+ "@type": "cr:Field",
395
+ "@id": "husic-classes/class_id",
396
+ "name": "class_id",
397
+ "description": "Integer HUSIC class identifier.",
398
+ "dataType": "sc:Integer"
399
+ },
400
+ {
401
+ "@type": "cr:Field",
402
+ "@id": "husic-classes/label",
403
+ "name": "label",
404
+ "description": "Full HUSIC label.",
405
+ "dataType": "sc:Text"
406
+ },
407
+ {
408
+ "@type": "cr:Field",
409
+ "@id": "husic-classes/group",
410
+ "name": "group",
411
+ "description": "High-level semantic group.",
412
+ "dataType": "sc:Text"
413
+ },
414
+ {
415
+ "@type": "cr:Field",
416
+ "@id": "husic-classes/definition",
417
+ "name": "definition",
418
+ "description": "Annotation definition.",
419
+ "dataType": "sc:Text"
420
+ }
421
+ ],
422
+ "data": [
423
+ {
424
+ "husic-classes/class_id": 0,
425
+ "husic-classes/label": "Exterior urban spaces with people",
426
+ "husic-classes/group": "Exterior",
427
+ "husic-classes/definition": "Outdoor urban commercial spaces with visible human presence."
428
+ },
429
+ {
430
+ "husic-classes/class_id": 1,
431
+ "husic-classes/label": "Exterior urban spaces without people",
432
+ "husic-classes/group": "Exterior",
433
+ "husic-classes/definition": "Outdoor commercial architecture or public-realm views without visible people."
434
+ },
435
+ {
436
+ "husic-classes/class_id": 2,
437
+ "husic-classes/label": "Interior urban spaces with people",
438
+ "husic-classes/group": "Interior",
439
+ "husic-classes/definition": "Public or semi-public commercial interiors with occupants, shoppers, workers, or event participants."
440
+ },
441
+ {
442
+ "husic-classes/class_id": 3,
443
+ "husic-classes/label": "Interior urban spaces without people",
444
+ "husic-classes/group": "Interior",
445
+ "husic-classes/definition": "Interior commercial spaces focused on spatial design, circulation, fixtures, or displays without people."
446
+ },
447
+ {
448
+ "husic-classes/class_id": 4,
449
+ "husic-classes/label": "Hotel or commercial lodging spaces",
450
+ "husic-classes/group": "Accommodation",
451
+ "husic-classes/definition": "Hotel rooms, serviced apartments, and commercial lodging environments."
452
+ },
453
+ {
454
+ "husic-classes/class_id": 5,
455
+ "husic-classes/label": "Private home interiors",
456
+ "husic-classes/group": "Accommodation",
457
+ "husic-classes/definition": "Private residential interiors connected to the broader commercial-district social-media corpus."
458
+ },
459
+ {
460
+ "husic-classes/class_id": 6,
461
+ "husic-classes/label": "Food or drink items",
462
+ "husic-classes/group": "Consumption",
463
+ "husic-classes/definition": "Food, drinks, dining-table scenes, and restaurant consumption content."
464
+ },
465
+ {
466
+ "husic-classes/class_id": 7,
467
+ "husic-classes/label": "Retail products and merchandise",
468
+ "husic-classes/group": "Consumption",
469
+ "husic-classes/definition": "Product, merchandise, retail shelf, shopping, and display-window content."
470
+ },
471
+ {
472
+ "husic-classes/class_id": 8,
473
+ "husic-classes/label": "Human-centered portrait",
474
+ "husic-classes/group": "Portrait",
475
+ "husic-classes/definition": "Selfies, group photos, portraits, and other images where specific people dominate the composition."
476
+ },
477
+ {
478
+ "husic-classes/class_id": 9,
479
+ "husic-classes/label": "Other non-spatial content",
480
+ "husic-classes/group": "Miscellaneous",
481
+ "husic-classes/definition": "Advertisements, screenshots, memes, maps, QR-code-like content, animals, or other non-spatial material."
482
+ }
483
+ ]
484
+ },
485
+ {
486
+ "@type": "cr:RecordSet",
487
+ "@id": "task-schemas",
488
+ "name": "BenchmarkTaskRecords",
489
+ "description": "Task definitions and ground-truth semantics for T1, T2, and T3.",
490
+ "key": {
491
+ "@id": "task-schemas/task_id"
492
+ },
493
+ "field": [
494
+ {
495
+ "@type": "cr:Field",
496
+ "@id": "task-schemas/task_id",
497
+ "name": "task_id",
498
+ "description": "Task identifier.",
499
+ "dataType": "sc:Text"
500
+ },
501
+ {
502
+ "@type": "cr:Field",
503
+ "@id": "task-schemas/name",
504
+ "name": "name",
505
+ "description": "Task name.",
506
+ "dataType": "sc:Text"
507
+ },
508
+ {
509
+ "@type": "cr:Field",
510
+ "@id": "task-schemas/input",
511
+ "name": "input",
512
+ "description": "Input modality or modalities.",
513
+ "dataType": "sc:Text"
514
+ },
515
+ {
516
+ "@type": "cr:Field",
517
+ "@id": "task-schemas/output",
518
+ "name": "output",
519
+ "description": "Output label, retrieval target, or annotation.",
520
+ "dataType": "sc:Text"
521
+ },
522
+ {
523
+ "@type": "cr:Field",
524
+ "@id": "task-schemas/ground_truth",
525
+ "name": "ground_truth",
526
+ "description": "Ground-truth construction.",
527
+ "dataType": "sc:Text"
528
+ },
529
+ {
530
+ "@type": "cr:Field",
531
+ "@id": "task-schemas/format",
532
+ "name": "format",
533
+ "description": "Released file format.",
534
+ "dataType": "sc:Text"
535
+ }
536
+ ],
537
+ "data": [
538
+ {
539
+ "task-schemas/task_id": "T1",
540
+ "task-schemas/name": "Urban scene semantic classification",
541
+ "task-schemas/input": "Image",
542
+ "task-schemas/output": "One of 10 HUSIC class labels",
543
+ "task-schemas/ground_truth": "Class folder name and Image Label column.",
544
+ "task-schemas/format": "ImageFolder train/val/test directories."
545
+ },
546
+ {
547
+ "task-schemas/task_id": "T2-A",
548
+ "task-schemas/name": "Category-level image-text retrieval",
549
+ "task-schemas/input": "Image or HUSIC label text",
550
+ "task-schemas/output": "Matching HUSIC label text or images in the class",
551
+ "task-schemas/ground_truth": "Image Label column and HUSIC class definitions.",
552
+ "task-schemas/format": "Excel rows joined to images by Image Filename."
553
+ },
554
+ {
555
+ "task-schemas/task_id": "T2-B",
556
+ "task-schemas/name": "Post-level image-text retrieval",
557
+ "task-schemas/input": "Image or original Chinese Post Text",
558
+ "task-schemas/output": "Images attached to the same public post or the corresponding post text",
559
+ "task-schemas/ground_truth": "One-to-many grouping by anonymized post and Image Filename stem.",
560
+ "task-schemas/format": "Excel rows with Post Text and Image Filename join key."
561
+ },
562
+ {
563
+ "task-schemas/task_id": "T3",
564
+ "task-schemas/name": "Instance segmentation",
565
+ "task-schemas/input": "Image",
566
+ "task-schemas/output": "Bounding boxes, COCO RLE masks, detected labels, detection scores",
567
+ "task-schemas/ground_truth": "Grounding DINO + SAM2 pseudo-labels with quality filtering and review.",
568
+ "task-schemas/format": "COCO-style JSON with img_info/images, categories, and annotations."
569
+ }
570
+ ]
571
+ },
572
+ {
573
+ "@type": "cr:RecordSet",
574
+ "@id": "t2-columns",
575
+ "name": "TextImagePairColumnRecords",
576
+ "description": "Column-level schema for train.xlsx, val.xlsx, test.xlsx, and full-corpus CSV text-image pair files.",
577
+ "key": {
578
+ "@id": "t2-columns/column_name"
579
+ },
580
+ "field": [
581
+ {
582
+ "@type": "cr:Field",
583
+ "@id": "t2-columns/column_name",
584
+ "name": "column_name",
585
+ "description": "Excel/CSV column name.",
586
+ "dataType": "sc:Text"
587
+ },
588
+ {
589
+ "@type": "cr:Field",
590
+ "@id": "t2-columns/description",
591
+ "name": "description",
592
+ "description": "Column meaning and use.",
593
+ "dataType": "sc:Text"
594
+ },
595
+ {
596
+ "@type": "cr:Field",
597
+ "@id": "t2-columns/task_role",
598
+ "name": "task_role",
599
+ "description": "Task or metadata role.",
600
+ "dataType": "sc:Text"
601
+ },
602
+ {
603
+ "@type": "cr:Field",
604
+ "@id": "t2-columns/example_value",
605
+ "name": "example_value",
606
+ "description": "English illustrative example; released Post Text remains original Chinese.",
607
+ "dataType": "sc:Text"
608
+ }
609
+ ],
610
+ "data": [
611
+ {
612
+ "t2-columns/column_name": "Image Label",
613
+ "t2-columns/description": "HUSIC class label for the image; used for category-level image-text alignment and T1 labels.",
614
+ "t2-columns/task_role": "T1/T2",
615
+ "t2-columns/example_value": "Exterior urban spaces with people"
616
+ },
617
+ {
618
+ "t2-columns/column_name": "Image Filename",
619
+ "t2-columns/description": "Image stem used to join spreadsheet rows to image files; follows UserID_PostDate_Index.",
620
+ "t2-columns/task_role": "Join key",
621
+ "t2-columns/example_value": "2668383_2020-01-21_0"
622
+ },
623
+ {
624
+ "t2-columns/column_name": "Post ID",
625
+ "t2-columns/description": "Anonymized numerical post identifier.",
626
+ "t2-columns/task_role": "Metadata",
627
+ "t2-columns/example_value": "4469639217818281"
628
+ },
629
+ {
630
+ "t2-columns/column_name": "User ID",
631
+ "t2-columns/description": "Anonymized numerical user identifier; original usernames are not released.",
632
+ "t2-columns/task_role": "Metadata",
633
+ "t2-columns/example_value": "2668383"
634
+ },
635
+ {
636
+ "t2-columns/column_name": "Post Time",
637
+ "t2-columns/description": "Original post timestamp in Chinese date-time format.",
638
+ "t2-columns/task_role": "Metadata",
639
+ "t2-columns/example_value": "2020-01-21 01:00"
640
+ },
641
+ {
642
+ "t2-columns/column_name": "Post Text",
643
+ "t2-columns/description": "Original Chinese Weibo text; used for post-level image-text retrieval.",
644
+ "t2-columns/task_role": "T2",
645
+ "t2-columns/example_value": "Synthetic English illustration: A lively evening walk near a commercial plaza; the released dataset keeps the original Chinese text."
646
+ },
647
+ {
648
+ "t2-columns/column_name": "City",
649
+ "t2-columns/description": "City associated with the location hashtag or commercial site.",
650
+ "t2-columns/task_role": "Metadata",
651
+ "t2-columns/example_value": "Chengdu"
652
+ },
653
+ {
654
+ "t2-columns/column_name": "Place Tag",
655
+ "t2-columns/description": "Location hashtag or place tag used for collection.",
656
+ "t2-columns/task_role": "Metadata",
657
+ "t2-columns/example_value": "#Chengdu Taikoo Li"
658
+ },
659
+ {
660
+ "t2-columns/column_name": "Posting Tool",
661
+ "t2-columns/description": "Client or posting-source string after metadata minimization.",
662
+ "t2-columns/task_role": "Metadata",
663
+ "t2-columns/example_value": "mobile client"
664
+ },
665
+ {
666
+ "t2-columns/column_name": "Mentioned Users",
667
+ "t2-columns/description": "Anonymized or removed mentioned-user field.",
668
+ "t2-columns/task_role": "Metadata",
669
+ "t2-columns/example_value": ""
670
+ },
671
+ {
672
+ "t2-columns/column_name": "Extracted Topics",
673
+ "t2-columns/description": "Topic or hashtag terms extracted from the post text.",
674
+ "t2-columns/task_role": "Metadata",
675
+ "t2-columns/example_value": "city life; shopping district"
676
+ },
677
+ {
678
+ "t2-columns/column_name": "Extracted Locations",
679
+ "t2-columns/description": "Location mentions extracted from the post text.",
680
+ "t2-columns/task_role": "Metadata",
681
+ "t2-columns/example_value": "Chengdu"
682
+ },
683
+ {
684
+ "t2-columns/column_name": "Like Count",
685
+ "t2-columns/description": "Public engagement count at collection time.",
686
+ "t2-columns/task_role": "Metadata",
687
+ "t2-columns/example_value": "52"
688
+ },
689
+ {
690
+ "t2-columns/column_name": "Repost Count",
691
+ "t2-columns/description": "Public repost count at collection time.",
692
+ "t2-columns/task_role": "Metadata",
693
+ "t2-columns/example_value": "13"
694
+ },
695
+ {
696
+ "t2-columns/column_name": "Comment Count",
697
+ "t2-columns/description": "Public comment count at collection time.",
698
+ "t2-columns/task_role": "Metadata",
699
+ "t2-columns/example_value": "19"
700
+ }
701
+ ]
702
+ },
703
+ {
704
+ "@type": "cr:RecordSet",
705
+ "@id": "segmentation-prompts",
706
+ "name": "SegmentationPromptRecords",
707
+ "description": "Per-class prompt vocabularies used to produce Grounding DINO + SAM2 pseudo-labels.",
708
+ "key": {
709
+ "@id": "segmentation-prompts/class_id"
710
+ },
711
+ "field": [
712
+ {
713
+ "@type": "cr:Field",
714
+ "@id": "segmentation-prompts/class_id",
715
+ "name": "class_id",
716
+ "description": "HUSIC class identifier.",
717
+ "dataType": "sc:Integer"
718
+ },
719
+ {
720
+ "@type": "cr:Field",
721
+ "@id": "segmentation-prompts/label",
722
+ "name": "label",
723
+ "description": "HUSIC label.",
724
+ "dataType": "sc:Text"
725
+ },
726
+ {
727
+ "@type": "cr:Field",
728
+ "@id": "segmentation-prompts/prompt",
729
+ "name": "prompt",
730
+ "description": "Per-class Grounding DINO prompt vocabulary.",
731
+ "dataType": "sc:Text"
732
+ }
733
+ ],
734
+ "data": [
735
+ {
736
+ "segmentation-prompts/class_id": 0,
737
+ "segmentation-prompts/label": "Exterior urban spaces with people",
738
+ "segmentation-prompts/prompt": "person . crowd . pedestrian . building facade . lawn . street lamp . glass curtain wall . sky . tree . shrub . fence . road . water . river . vehicle . sculpture . installation . pavement . street signage . fountain"
739
+ },
740
+ {
741
+ "segmentation-prompts/class_id": 1,
742
+ "segmentation-prompts/label": "Exterior urban spaces without people",
743
+ "segmentation-prompts/prompt": "building facade . glass curtain wall . wooden facade . tree . shrub . lawn . sky . pavement . road . water . river . lantern . sculpture . installation . street lamp . signage . fence . bridge . water feature . fountain"
744
+ },
745
+ {
746
+ "segmentation-prompts/class_id": 2,
747
+ "segmentation-prompts/label": "Interior urban spaces with people",
748
+ "segmentation-prompts/prompt": "person . shopper . crowd . retail shelf . escalator . elevator . ceiling . floor tile . glass partition . display case . door . indoor plant . wall . window . handrail . column"
749
+ },
750
+ {
751
+ "segmentation-prompts/class_id": 3,
752
+ "segmentation-prompts/label": "Interior urban spaces without people",
753
+ "segmentation-prompts/prompt": "retail shelf . escalator . indoor corridor . ceiling . floor tile . marble floor . glass partition . display case . wall . column . indoor plant . elevator . door . window . lighting fixture . handrail"
754
+ },
755
+ {
756
+ "segmentation-prompts/class_id": 4,
757
+ "segmentation-prompts/label": "Hotel or commercial lodging spaces",
758
+ "segmentation-prompts/prompt": "hotel bed . furniture . sofa . carpet . marble floor . tile floor . wooden floor . ceiling . bathroom . window . curtain . lamp"
759
+ },
760
+ {
761
+ "segmentation-prompts/class_id": 5,
762
+ "segmentation-prompts/label": "Private home interiors",
763
+ "segmentation-prompts/prompt": "sofa . bed . dining table . floor . ceiling . kitchen . bookshelf . wardrobe . window . lamp . carpet . wall"
764
+ },
765
+ {
766
+ "segmentation-prompts/class_id": 6,
767
+ "segmentation-prompts/label": "Food or drink items",
768
+ "segmentation-prompts/prompt": "food dish . meal plate . dessert . beverage cup . coffee . drink bottle . bowl . chopsticks . spoon . dining table . person . restaurant interior"
769
+ },
770
+ {
771
+ "segmentation-prompts/class_id": 7,
772
+ "segmentation-prompts/label": "Retail products and merchandise",
773
+ "segmentation-prompts/prompt": "fashion clothing . shoes . cosmetics . product package . merchandise . retail shelf . bag . jewelry . electronics . store window . mannequin . person . floor . wall"
774
+ },
775
+ {
776
+ "segmentation-prompts/class_id": 8,
777
+ "segmentation-prompts/label": "Human-centered portrait",
778
+ "segmentation-prompts/prompt": "person . face . group photo . building facade . sky . tree . floor . food . animal . vehicle . indoor background"
779
+ },
780
+ {
781
+ "segmentation-prompts/class_id": 9,
782
+ "segmentation-prompts/label": "Other non-spatial content",
783
+ "segmentation-prompts/prompt": "animal . person . vehicle . advertisement poster . text . QR code . screenshot . sculpture . meme . sky . plant . signage . graphic design . logo . map . infographic . chat record"
784
+ }
785
+ ]
786
+ }
787
+ ],
788
+ "rai:dataLimitations": [
789
+ "Urban-ImageNet is geographically concentrated in China and should not be interpreted as a universal global urban dataset.",
790
+ "The source is public Weibo posts, so the corpus reflects platform demographics, posting norms, commercial popularity, hashtag practices, and visibility biases rather than a representative sample of all city users.",
791
+ "The 1K, 10K, and 100K subsets are intentionally class-balanced for benchmarking; the full 2M corpus is unbalanced and has no predefined train/validation/test split.",
792
+ "Post-level image-text pairing is naturally loose: one post may contain up to nine images, and the post text may describe an event, emotion, route, or commercial activity rather than a literal caption of every image.",
793
+ "Task 3 instance masks are model-generated pseudo-labels from Grounding DINO and SAM2 with quality filtering, not exhaustive human pixel annotations.",
794
+ "The dataset is not recommended for individual identification, face recognition, account reconstruction, surveillance, social scoring, commercial profiling, law-enforcement targeting, or demographic inference about specific users."
795
+ ],
796
+ "rai:dataBiases": [
797
+ "Selection bias arises because only public social-media posts with location-specific commercial-site hashtags were collected.",
798
+ "Population bias is likely because Weibo users over-represent digitally active, urban, and younger groups compared with all city users.",
799
+ "Geographic and cultural bias is present because all 24 cities are in China and the text is primarily original Chinese social-media language.",
800
+ "Visual bias is present because users tend to post photogenic, socially meaningful, popular, or commercial scenes rather than mundane or private experiences.",
801
+ "The full 2M corpus is class-imbalanced; balanced subsets are useful for model comparison but do not reflect the natural class distribution of the source platform.",
802
+ "Segmentation pseudo-labels may inherit Grounding DINO and SAM2 biases, including missed small objects, errors on commercial interiors, and prompt vocabulary effects."
803
+ ],
804
+ "rai:personalSensitiveInformation": [
805
+ "The source posts were public, but social-media images and text can still contain contextual personal information. The release therefore uses privacy-protected derivatives rather than raw data.",
806
+ "Original usernames and account names are removed. Post ID and User ID are released only as opaque numerical identifiers after anonymization/pseudonymization and are not intended to reconstruct original accounts.",
807
+ "Images are resized to a maximum long edge of 512 px. Faces, license plates, QR-code-like regions, and other sensitive visual regions are blurred with automated detectors followed by human spot checks.",
808
+ "The raw high-resolution corpus, larger than 4 TB, is not publicly released and is retained securely. The public release is substantially reduced in resolution and metadata detail.",
809
+ "The dataset does not intentionally provide health, financial, political, religious, biometric, or exact personal identity attributes. Users must not attempt re-identification or linkage to original social-media accounts."
810
+ ],
811
+ "rai:dataUseCases": [
812
+ "Validated use cases include urban scene semantic classification under the 10-class HUSIC taxonomy, category-level image-text retrieval, post-level Chinese image-text retrieval with one-to-many matching, instance segmentation benchmarking, and data-scale behavior analysis across 1K/10K/100K/2M variants.",
813
+ "The dataset is intended for non-commercial academic research in urban perception, computational urban studies, multimodal learning, image classification, cross-modal retrieval, segmentation, and reproducible benchmarking.",
814
+ "Construct validity: the dataset is designed to represent user-generated visual and textual traces of urban commercial space perception, not the full demographic distribution of urban residents.",
815
+ "Uses not validated include production deployment, safety-critical navigation, individual-level behavior prediction, demographic profiling, policing, surveillance, or inference about private individuals."
816
+ ],
817
+ "rai:dataSocialImpact": "Potential positive impacts include supporting evidence-based urban planning, helping researchers understand how people perceive and document commercial spaces, improving non-Western and Chinese-language multimodal benchmarks, and enabling public-good research for more inclusive urban design. Potential negative impacts include privacy leakage, re-identification attempts, surveillance-oriented use, account reconstruction, and commercial profiling. Mitigations include using only public posts, anonymizing identifiers, removing usernames, blurring faces and license plates, resizing images, minimizing metadata, restricting use to non-commercial academic research, and explicitly prohibiting re-identification, surveillance, face recognition, and profiling.",
818
+ "rai:hasSyntheticData": false,
819
+ "rai:dataCollection": "Urban-ImageNet was collected from public Sina Weibo posts using location-specific hashtags for 61 commercial sites across 24 Chinese cities from 2019-01-01 to 2025-12-31. For each public post, the pipeline retained image attachments, original Chinese post text, timestamp, city/site metadata, place tag, and lightweight engagement metadata.",
820
+ "rai:dataCollectionType": [
821
+ "Web Scraping",
822
+ "User-generated content data",
823
+ "Software Collection",
824
+ "Manual Human Curation"
825
+ ],
826
+ "rai:dataCollectionRawData": "The raw corpus consisted of public Weibo image-text posts from commercial districts. The raw high-resolution corpus exceeds 4 TB and is not publicly released; only privacy-protected, resized, anonymized derivatives are distributed.",
827
+ "rai:dataPreprocessingProtocol": [
828
+ "Perceptual-hash near-duplicate removal, minimum-resolution filtering, automated NSFW filtering, repeated-advertisement/spam removal, metadata minimization, image resizing to max 512 px, face/license-plate/QR-sensitive-region blurring, and human spot-check review.",
829
+ "Original usernames and direct account identifiers were removed; released Post ID and User ID fields are opaque numerical identifiers."
830
+ ],
831
+ "rai:dataAnnotationProtocol": "The HUSIC 10-class labels were manually annotated by trained researchers using standardized guidelines and calibration. Task 3 pseudo-labels were generated with class-specific Grounding DINO prompts followed by SAM2 mask prediction, non-maximum suppression, area filtering, and quality review.",
832
+ "rai:dataAnnotationPlatform": [
833
+ "Internal researcher annotation workflow",
834
+ "Grounding DINO",
835
+ "SAM2"
836
+ ],
837
+ "rai:dataAnnotationAnalysis": "The 100K benchmark set was manually labeled by three trained researchers after calibration; paper documentation reports high inter-rater agreement. Segmentation labels are marked as model-generated pseudo-labels and should be treated differently from exhaustive human pixel annotations.",
838
+ "rai:annotationsPerItem": "Each benchmark image has one HUSIC class label. Text-image rows provide post-level metadata. Instance segmentation files contain zero or more object annotations per image, depending on model detections and filtering.",
839
+ "rai:machineAnnotationTools": [
840
+ "Grounding DINO",
841
+ "SAM2",
842
+ "pHash deduplication",
843
+ "face/license-plate blurring detectors"
844
+ ],
845
+ "rai:dataReleaseMaintenancePlan": "The dataset is released as version 1.0 for NeurIPS 2026 submission. Future updates should document checksum changes, privacy-processing changes, annotation revisions, and deprecation of superseded files.",
846
+ "prov:wasDerivedFrom": [
847
+ {
848
+ "@id": "https://weibo.com"
849
+ },
850
+ {
851
+ "@id": "https://huggingface.co/datasets/Yiwei-Ou/Urban-ImageNet"
852
+ }
853
+ ],
854
+ "prov:wasGeneratedBy": [
855
+ {
856
+ "@type": "prov:Activity",
857
+ "@id": "public-weibo-collection",
858
+ "name": "Public Weibo Commercial-Site Collection",
859
+ "prov:type": "DataCollection",
860
+ "prov:startedAtTime": "2019-01-01",
861
+ "prov:endedAtTime": "2025-12-31",
862
+ "description": "Collected public posts associated with location-specific hashtags for 61 commercial sites in 24 Chinese cities."
863
+ },
864
+ {
865
+ "@type": "prov:Activity",
866
+ "@id": "privacy-preserving-preprocessing",
867
+ "name": "Privacy-Preserving Preprocessing",
868
+ "prov:type": "DataProcessing",
869
+ "description": "Applied deduplication, filtering, metadata minimization, anonymization, resizing, face/license-plate/QR-sensitive-region blurring, and manual spot checks."
870
+ },
871
+ {
872
+ "@type": "prov:Activity",
873
+ "@id": "husic-manual-annotation",
874
+ "name": "HUSIC Manual Annotation",
875
+ "prov:type": "DataAnnotation",
876
+ "description": "Trained researchers annotated the balanced benchmark subsets using the 10-class HUSIC taxonomy and calibration guidelines."
877
+ },
878
+ {
879
+ "@type": "prov:Activity",
880
+ "@id": "groundingdino-sam2-pseudolabeling",
881
+ "name": "Grounding DINO + SAM2 Instance Segmentation Pseudo-Labeling",
882
+ "prov:type": "DataAnnotation",
883
+ "description": "Generated boxes, detected labels, confidence scores, and COCO RLE masks using class-specific prompt vocabularies, NMS, area filtering, and review."
884
+ }
885
+ ]
886
+ }