| --- |
| pretty_name: BagBuddy Dataset |
| size_categories: |
| - n<1K |
| task_categories: |
| - image-classification |
| - object-detection |
| configs: |
| - config_name: default |
| data_files: |
| - split: train |
| path: data/train-* |
| dataset_info: |
| features: |
| - name: image |
| dtype: image |
| - name: caption |
| dtype: string |
| - name: bucket |
| dtype: string |
| - name: sample_id |
| dtype: string |
| - name: labels |
| list: string |
| - name: annotation_indices |
| list: int64 |
| - name: annotation_x |
| list: int64 |
| - name: annotation_y |
| list: int64 |
| - name: annotation_count |
| dtype: int64 |
| splits: |
| - name: train |
| num_bytes: 12169999 |
| num_examples: 40 |
| download_size: 12169290 |
| dataset_size: 12169999 |
| --- |
| |
| # LLM-Pack: Grocery Detection Dataset |
|
|
| A small object detection and scene understanding dataset containing tabletop grocery scenes with annotated item names and object locations. |
| The dataset consists of 40 images with varying object counts, designed for evaluating object detection, counting, and multimodal reasoning systems in cluttered grocery scenarios. |
|
|
| ## Dataset Overview |
|
|
| - **Total scenes:** 40 |
| - **Object counts per scene:** 6, 8, 10, 12, 14, 16, 18, or 20 items |
| - **Samples per object-count category:** 5 |
| - **Annotations:** Object names + object center coordinates |
| - **Image resolution:** 1920×1080 |
| - **Task type:** Object detection / scene understanding / counting |
|
|
| ## Dataset Structure |
|
|
| Each sample contains: |
|
|
| ```python |
| { |
| "image": PIL.Image, |
| "caption": str, |
| "bucket": str, # number of items on the image |
| "sample_id": str, |
| |
| "labels": List[str], |
| "annotation_indices": List[int], |
| "annotation_x": List[int], |
| "annotation_y": List[int], |
| |
| "annotation_count": int |
| } |
| ``` |
|
|
| ## Annotation Format |
|
|
| Object annotations are stored as aligned lists. |
|
|
| Example: |
|
|
| ```python |
| { |
| "labels": [ |
| "Glass Beer Bottle", |
| "Apples", |
| "Noodles in Plastic Bag" |
| ], |
| |
| "annotation_x": [1480, 1251, 1123], |
| "annotation_y": [445, 822, 810] |
| } |
| ``` |
|
|
| Each `(annotation_x[i], annotation_y[i])` pair corresponds to the center position of `labels[i]` in the image. |
|
|
| ## Usage |
|
|
| ```python |
| from datasets import load_dataset |
| |
| dataset = load_dataset( |
| "Yannik019/llm_pack_detection", |
| split="train" |
| ) |
| |
| print(dataset) |
| ``` |
|
|
| ## Example |
|
|
| A full usage example is available here: |
|
|
| - [example.py](https://huggingface.co/datasets/Yannik019/llm_pack_detection/blob/main/example.py) |
|
|
| ## Intended Use |
|
|
| This dataset is intended for: |
|
|
| - Object detection benchmarking |
| - Vision-language model evaluation |
| - Scene understanding research |
| - Tabletop grocery perception |
| - Referring object localization |
|
|
| ## Citation |
|
|
| ```bibtex |
| @misc{blei2025llmpackintuitivegroceryhandling, |
| title={LLM-Pack: Intuitive Grocery Handling for Logistics Applications}, |
| author={Yannik Blei and Michael Krawez and Tobias Jülg and Pierre Krack and Florian Walter and Wolfram Burgard}, |
| year={2025}, |
| eprint={2503.08445}, |
| archivePrefix={arXiv}, |
| primaryClass={cs.RO}, |
| url={https://arxiv.org/abs/2503.08445}, |
| } |
| ``` |
|
|
|
|