nielsr HF Staff commited on
Commit
c5ee9fd
·
verified ·
1 Parent(s): 7089780

Add task category, update license and add project links

Browse files

Hi, I'm Niels from the Hugging Face community science team! This PR improves the dataset card for KubriCount by:
- Adding the `object-detection` task category.
- Updating the license to `apache-2.0` based on the GitHub repository.
- Adding links to the project page and code repository at the top of the README.
- Ensuring the paper is linked within the Markdown content.

Files changed (1) hide show
  1. README.md +6 -78
README.md CHANGED
@@ -1,6 +1,8 @@
1
  ---
2
- license: other
3
  pretty_name: KubriCount
 
 
4
  tags:
5
  - image
6
  - synthetic
@@ -13,6 +15,8 @@ tags:
13
 
14
  # KubriCount
15
 
 
 
16
  KubriCount is a large-scale synthetic benchmark for **multi-grained visual counting**, built for the research project **Count Anything at Any Granularity**.
17
 
18
  The dataset targets open-world counting settings where the intended counting granularity must be explicit. A query may ask for a specific identity, an attribute variant, a category, an instance type, or a broader concept. KubriCount provides controlled distractors and dense instance-level supervision for training and evaluation.
@@ -109,8 +113,6 @@ The tar shards in this release contain only scenes that passed the automatic qua
109
  └── extracted_metadata.json
110
  ```
111
 
112
- The release intentionally does **not** include `metadata/dataset_stats.json` or per-split `vlm_filter_results.json` files.
113
-
114
  ## Files Inside Each Scene
115
 
116
  The image folders are stored inside tar shards. Each tar preserves the split/level/timestamp/scene structure:
@@ -183,51 +185,6 @@ A typical annotation item is:
183
  }
184
  ```
185
 
186
- Field meanings:
187
-
188
- - `image_id`: relative path to the edited image after shard extraction.
189
- - `count`: number of target objects.
190
- - `category`: target category or target phrase.
191
- - `box_examples_coordinates`: target-object 2D boxes represented by four corner points.
192
- - `points`: target-object center points.
193
- - `H`, `W`: image height and width.
194
- - `metadata.level`: counting granularity level.
195
- - `metadata.split`: dataset split.
196
- - `negative_category`: distractor category or phrase, when applicable.
197
- - `negative_count`: number of distractor objects.
198
- - `negative_box_examples_coordinates`: distractor-object 2D boxes.
199
- - `negative_points`: distractor-object center points.
200
-
201
- ## Manifest Format
202
-
203
- Each line in `metadata/all_pass_scenes.jsonl` describes one released scene and where it is stored:
204
-
205
- ```json
206
- {
207
- "split": "testA",
208
- "scene": "level1/20260205_132725/scene_0001",
209
- "path_in_dataset": "testA/level1/20260205_132725/scene_0001",
210
- "shard": "shards/testA/testA-000000.tar",
211
- "num_files": 4,
212
- "files": [
213
- {
214
- "path": "testA/level1/20260205_132725/scene_0001/edited_00000.png",
215
- "name": "edited_00000.png",
216
- "size_bytes": 1562567
217
- }
218
- ]
219
- }
220
- ```
221
-
222
- Important fields:
223
-
224
- - `split`: dataset split.
225
- - `scene`: scene path relative to the split folder.
226
- - `path_in_dataset`: scene path after extraction.
227
- - `shard`: tar shard containing this scene.
228
- - `num_files`: number of files in this scene.
229
- - `files`: files stored for this scene.
230
-
231
  ## Download
232
 
233
  ```python
@@ -292,23 +249,6 @@ for split in splits:
292
  print(f"Restored dataset to: {restore_dir}")
293
  ```
294
 
295
- After extraction:
296
-
297
- ```text
298
- KubriCount_restored/
299
- ├── train/
300
- │ ├── extracted_metadata.json
301
- │ └── level1/
302
- ├── testA/
303
- │ ├── extracted_metadata.json
304
- │ └── level1/
305
- ├── testB/
306
- │ ├── extracted_metadata.json
307
- │ └── level1/
308
- ├── merged_train_metadata.json
309
- └── merged_test_metadata.json
310
- ```
311
-
312
  ## Read Images Directly From Tar Shards
313
 
314
  ```python
@@ -327,18 +267,6 @@ for tar_path in sorted((repo_dir / "shards").glob("*/*.tar")):
327
  break
328
  ```
329
 
330
- To find the shard for a specific scene, use `metadata/all_pass_scenes.jsonl`.
331
-
332
- ## Companion Code
333
-
334
- The generation pipeline is released separately at [Verg-Avesta/KubriCount](https://github.com/Verg-Avesta/KubriCount).
335
-
336
- That repository contains the Kubric-based rendering code, asset preprocessing utilities, image-editing scripts, and VLM-filtering scripts used to construct KubriCount. It is only needed if you want to reproduce or extend the data generation pipeline.
337
-
338
- ## Paper
339
-
340
- The paper is available at [arXiv](https://arxiv.org/abs/2605.10887).
341
-
342
  ## Citation
343
 
344
  If you find this dataset useful, please cite:
@@ -358,4 +286,4 @@ KubriCount builds on the [Kubric](https://github.com/google-research/kubric) dat
358
 
359
  ## Contact
360
 
361
- For questions, please contact liuchang666@sjtu.edu.cn.
 
1
  ---
2
+ license: apache-2.0
3
  pretty_name: KubriCount
4
+ task_categories:
5
+ - object-detection
6
  tags:
7
  - image
8
  - synthetic
 
15
 
16
  # KubriCount
17
 
18
+ [Project Page](https://verg-avesta.github.io/KubriCount/) | [Paper](https://arxiv.org/abs/2605.10887) | [Code](https://github.com/Verg-Avesta/KubriCount)
19
+
20
  KubriCount is a large-scale synthetic benchmark for **multi-grained visual counting**, built for the research project **Count Anything at Any Granularity**.
21
 
22
  The dataset targets open-world counting settings where the intended counting granularity must be explicit. A query may ask for a specific identity, an attribute variant, a category, an instance type, or a broader concept. KubriCount provides controlled distractors and dense instance-level supervision for training and evaluation.
 
113
  └── extracted_metadata.json
114
  ```
115
 
 
 
116
  ## Files Inside Each Scene
117
 
118
  The image folders are stored inside tar shards. Each tar preserves the split/level/timestamp/scene structure:
 
185
  }
186
  ```
187
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
188
  ## Download
189
 
190
  ```python
 
249
  print(f"Restored dataset to: {restore_dir}")
250
  ```
251
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
252
  ## Read Images Directly From Tar Shards
253
 
254
  ```python
 
267
  break
268
  ```
269
 
 
 
 
 
 
 
 
 
 
 
 
 
270
  ## Citation
271
 
272
  If you find this dataset useful, please cite:
 
286
 
287
  ## Contact
288
 
289
+ For questions, please contact liuchang666@sjtu.edu.cn.