chenhaoguan commited on
Commit
fc779a7
·
verified ·
1 Parent(s): 3917e90

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +66 -59
README.md CHANGED
@@ -1,80 +1,87 @@
1
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2
  license: cc-by-4.0
3
  task_categories:
4
- - visual-question-answering
5
  language:
6
- - en
7
  tags:
8
- - multi-image
9
- - hallucination
10
- - benchmark
11
- - vision-language-model
12
- - multimodal
13
  size_categories:
14
- - 1K<n<10K
15
  ---
16
 
17
- # MIHBench: Multi-Image Hallucination Benchmark
18
 
19
- **Paper**: [MIHBench: Can Multi-modal Large Language Models Understand Multi-Image Inputs?](https://arxiv.org/abs/2505.xxxxx) | ACM Multimedia 2025
20
 
21
- ## Overview
22
 
23
- MIHBench is a comprehensive benchmark for evaluating multi-image understanding and hallucination in Multi-modal Large Language Models (MLLMs). It contains **3,200 samples** across **4 tasks** (800 samples each), with each sample containing 2-4 images.
 
 
 
 
 
 
 
 
 
24
 
25
  ## Tasks
26
 
27
  | Task | # Images | Description |
28
  |------|----------|-------------|
29
- | **Count** | 2 | Determine whether the same number of a target object appears in both images. 400 samples include injected distracting objects. |
30
- | **Existence (Adversarial)** | 3 | Determine whether a target object exists in all images, with adversarially selected objects (rare, confusing). |
31
- | **Existence (Popular)** | 3 | Determine whether a target object exists in all images, using commonly known objects. |
32
- | **Existence (Random)** | 3 | Determine whether a target object exists in all images, using randomly selected objects. |
33
-
34
- > **Note**: A 5th task (ID Consistency) will be added in a future update.
35
-
36
- ## Dataset Schema
37
-
38
- ### Common columns (all tasks)
39
-
40
- | Column | Type | Description |
41
- |--------|------|-------------|
42
- | `images` | `list[image]` | 2-4 images (PIL Image objects) |
43
- | `question` | `str` | Natural language question about the images |
44
- | `label` | `str` | Ground truth answer (`"yes"` or `"no"`) |
45
- | `task` | `str` | Task identifier |
46
- | `num_images` | `int` | Number of images in the sample |
47
- | `image_names` | `list[str]` | Source image filenames |
48
-
49
- ### Additional columns (Count task only)
50
-
51
- | Column | Type | Description |
52
- |--------|------|-------------|
53
- | `injected` | `bool` | Whether distracting objects were injected into the question |
54
- | `object_counts` | `str` | JSON string mapping image identifiers to object counts (e.g., `'{"A": 1, "B": 1}'`) |
55
 
56
- ## Data Splits
57
 
58
- Each task is a separate configuration/split with 800 samples (400 `"yes"`, 400 `"no"`).
59
-
60
- ## Image Sources
61
-
62
- - **Tasks 1-4**: COCO (Common Objects in Context) dataset
63
- - **Task 5** (ID Consistency, coming soon): CO3D dataset
64
-
65
- ## Citation
66
-
67
- If you use MIHBench in your research, please cite:
68
-
69
- ```bibtex
70
- @inproceedings{mihbench2025,
71
- title={MIHBench: Can Multi-modal Large Language Models Understand Multi-Image Inputs?},
72
- author={},
73
- booktitle={Proceedings of the ACM Multimedia 2025},
74
- year={2025}
75
- }
76
  ```
77
 
78
- ## License
79
 
80
- This dataset is released under the [CC-BY-4.0](https://creativecommons.org/licenses/by/4.0/) license. The underlying images are from COCO and CO3D, which have their own licenses.
 
1
  ---
2
+ dataset_info:
3
+ features:
4
+ - name: images
5
+ sequence: image
6
+ - name: question
7
+ dtype: string
8
+ - name: label
9
+ dtype: string
10
+ - name: task
11
+ dtype: string
12
+ - name: num_images
13
+ dtype: int64
14
+ - name: image_names
15
+ sequence: string
16
+ - name: injected
17
+ dtype: bool
18
+ - name: object_counts
19
+ dtype: string
20
+ configs:
21
+ - config_name: count
22
+ data_files:
23
+ - split: train
24
+ path: data/count.parquet
25
+ - config_name: existence_adversarial
26
+ data_files:
27
+ - split: train
28
+ path: data/existence_adversarial.parquet
29
+ - config_name: existence_popular
30
+ data_files:
31
+ - split: train
32
+ path: data/existence_popular.parquet
33
+ - config_name: existence_random
34
+ data_files:
35
+ - split: train
36
+ path: data/existence_random.parquet
37
  license: cc-by-4.0
38
  task_categories:
39
+ - visual-question-answering
40
  language:
41
+ - en
42
  tags:
43
+ - multi-image
44
+ - hallucination
45
+ - benchmark
46
+ - vision-language-model
47
+ - multimodal
48
  size_categories:
49
+ - 1K<n<10K
50
  ---
51
 
52
+ # MIHBench
53
 
54
+ Multi-Image Hallucination Benchmark for evaluating multi-image understanding in MLLMs. 3,200 samples across 4 tasks (800 each), with each sample containing 2-4 images from COCO.
55
 
56
+ ## Fields
57
 
58
+ | Field | Description |
59
+ |-------|-------------|
60
+ | images | 2-4 images (list) |
61
+ | question | Natural language question about the images |
62
+ | label | Ground truth: "yes" or "no" |
63
+ | task | Task identifier |
64
+ | num_images | Number of images in the sample |
65
+ | image_names | Source image filenames |
66
+
67
+ Additional fields for **count** task: `injected` (bool), `object_counts` (JSON string).
68
 
69
  ## Tasks
70
 
71
  | Task | # Images | Description |
72
  |------|----------|-------------|
73
+ | count | 2 | Same number of target object in both images? |
74
+ | existence_adversarial | 3 | Target object exists in all images? (rare/confusing objects) |
75
+ | existence_popular | 3 | Target object exists in all images? (common objects) |
76
+ | existence_random | 3 | Target object exists in all images? (random objects) |
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
77
 
78
+ ## Evaluation
79
 
80
+ ```
81
+ metrics: Accuracy, Precision, Recall, F1
82
+ parser: yes/no binary
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
83
  ```
84
 
85
+ ## Source
86
 
87
+ Original data from [MIHBench](https://arxiv.org/abs/2505.xxxxx) (ACM Multimedia 2025).