liuchang666 nielsr HF Staff commited on
Commit
74c5b85
·
1 Parent(s): 7089780

Add task category, update license and add project links (#1)

Browse files

- Add task category, update license and add project links (c5ee9fd05e24fc83efc123f696c995228d7057b9)


Co-authored-by: Niels Rogge <nielsr@users.noreply.huggingface.co>

Files changed (1) hide show
  1. README.md +6 -78
README.md CHANGED
@@ -1,6 +1,8 @@
1
  ---
2
- license: other
3
  pretty_name: KubriCount
 
 
4
  tags:
5
  - image
6
  - synthetic
@@ -13,6 +15,8 @@ tags:
13
 
14
  # KubriCount
15
 
 
 
16
  KubriCount is a large-scale synthetic benchmark for **multi-grained visual counting**, built for the research project **Count Anything at Any Granularity**.
17
 
18
  The dataset targets open-world counting settings where the intended counting granularity must be explicit. A query may ask for a specific identity, an attribute variant, a category, an instance type, or a broader concept. KubriCount provides controlled distractors and dense instance-level supervision for training and evaluation.
@@ -109,8 +113,6 @@ The tar shards in this release contain only scenes that passed the automatic qua
109
  └── extracted_metadata.json
110
  ```
111
 
112
- The release intentionally does **not** include `metadata/dataset_stats.json` or per-split `vlm_filter_results.json` files.
113
-
114
  ## Files Inside Each Scene
115
 
116
  The image folders are stored inside tar shards. Each tar preserves the split/level/timestamp/scene structure:
@@ -183,51 +185,6 @@ A typical annotation item is:
183
  }
184
  ```
185
 
186
- Field meanings:
187
-
188
- - `image_id`: relative path to the edited image after shard extraction.
189
- - `count`: number of target objects.
190
- - `category`: target category or target phrase.
191
- - `box_examples_coordinates`: target-object 2D boxes represented by four corner points.
192
- - `points`: target-object center points.
193
- - `H`, `W`: image height and width.
194
- - `metadata.level`: counting granularity level.
195
- - `metadata.split`: dataset split.
196
- - `negative_category`: distractor category or phrase, when applicable.
197
- - `negative_count`: number of distractor objects.
198
- - `negative_box_examples_coordinates`: distractor-object 2D boxes.
199
- - `negative_points`: distractor-object center points.
200
-
201
- ## Manifest Format
202
-
203
- Each line in `metadata/all_pass_scenes.jsonl` describes one released scene and where it is stored:
204
-
205
- ```json
206
- {
207
- "split": "testA",
208
- "scene": "level1/20260205_132725/scene_0001",
209
- "path_in_dataset": "testA/level1/20260205_132725/scene_0001",
210
- "shard": "shards/testA/testA-000000.tar",
211
- "num_files": 4,
212
- "files": [
213
- {
214
- "path": "testA/level1/20260205_132725/scene_0001/edited_00000.png",
215
- "name": "edited_00000.png",
216
- "size_bytes": 1562567
217
- }
218
- ]
219
- }
220
- ```
221
-
222
- Important fields:
223
-
224
- - `split`: dataset split.
225
- - `scene`: scene path relative to the split folder.
226
- - `path_in_dataset`: scene path after extraction.
227
- - `shard`: tar shard containing this scene.
228
- - `num_files`: number of files in this scene.
229
- - `files`: files stored for this scene.
230
-
231
  ## Download
232
 
233
  ```python
@@ -292,23 +249,6 @@ for split in splits:
292
  print(f"Restored dataset to: {restore_dir}")
293
  ```
294
 
295
- After extraction:
296
-
297
- ```text
298
- KubriCount_restored/
299
- ├── train/
300
- │ ├── extracted_metadata.json
301
- │ └── level1/
302
- ├── testA/
303
- │ ├── extracted_metadata.json
304
- │ └── level1/
305
- ├── testB/
306
- │ ├── extracted_metadata.json
307
- │ └── level1/
308
- ├── merged_train_metadata.json
309
- └── merged_test_metadata.json
310
- ```
311
-
312
  ## Read Images Directly From Tar Shards
313
 
314
  ```python
@@ -327,18 +267,6 @@ for tar_path in sorted((repo_dir / "shards").glob("*/*.tar")):
327
  break
328
  ```
329
 
330
- To find the shard for a specific scene, use `metadata/all_pass_scenes.jsonl`.
331
-
332
- ## Companion Code
333
-
334
- The generation pipeline is released separately at [Verg-Avesta/KubriCount](https://github.com/Verg-Avesta/KubriCount).
335
-
336
- That repository contains the Kubric-based rendering code, asset preprocessing utilities, image-editing scripts, and VLM-filtering scripts used to construct KubriCount. It is only needed if you want to reproduce or extend the data generation pipeline.
337
-
338
- ## Paper
339
-
340
- The paper is available at [arXiv](https://arxiv.org/abs/2605.10887).
341
-
342
  ## Citation
343
 
344
  If you find this dataset useful, please cite:
@@ -358,4 +286,4 @@ KubriCount builds on the [Kubric](https://github.com/google-research/kubric) dat
358
 
359
  ## Contact
360
 
361
- For questions, please contact liuchang666@sjtu.edu.cn.
 
1
  ---
2
+ license: apache-2.0
3
  pretty_name: KubriCount
4
+ task_categories:
5
+ - object-detection
6
  tags:
7
  - image
8
  - synthetic
 
15
 
16
  # KubriCount
17
 
18
+ [Project Page](https://verg-avesta.github.io/KubriCount/) | [Paper](https://arxiv.org/abs/2605.10887) | [Code](https://github.com/Verg-Avesta/KubriCount)
19
+
20
  KubriCount is a large-scale synthetic benchmark for **multi-grained visual counting**, built for the research project **Count Anything at Any Granularity**.
21
 
22
  The dataset targets open-world counting settings where the intended counting granularity must be explicit. A query may ask for a specific identity, an attribute variant, a category, an instance type, or a broader concept. KubriCount provides controlled distractors and dense instance-level supervision for training and evaluation.
 
113
  └── extracted_metadata.json
114
  ```
115
 
 
 
116
  ## Files Inside Each Scene
117
 
118
  The image folders are stored inside tar shards. Each tar preserves the split/level/timestamp/scene structure:
 
185
  }
186
  ```
187
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
188
  ## Download
189
 
190
  ```python
 
249
  print(f"Restored dataset to: {restore_dir}")
250
  ```
251
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
252
  ## Read Images Directly From Tar Shards
253
 
254
  ```python
 
267
  break
268
  ```
269
 
 
 
 
 
 
 
 
 
 
 
 
 
270
  ## Citation
271
 
272
  If you find this dataset useful, please cite:
 
286
 
287
  ## Contact
288
 
289
+ For questions, please contact liuchang666@sjtu.edu.cn.