chenhaoguan commited on
Commit
71e01c9
·
verified ·
1 Parent(s): 1b4c5c7

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +80 -0
README.md ADDED
@@ -0,0 +1,80 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: cc-by-4.0
3
+ task_categories:
4
+ - visual-question-answering
5
+ language:
6
+ - en
7
+ tags:
8
+ - multi-image
9
+ - hallucination
10
+ - benchmark
11
+ - vision-language-model
12
+ - multimodal
13
+ size_categories:
14
+ - 1K<n<10K
15
+ ---
16
+
17
+ # MIHBench: Multi-Image Hallucination Benchmark
18
+
19
+ **Paper**: [MIHBench: Can Multi-modal Large Language Models Understand Multi-Image Inputs?](https://arxiv.org/abs/2505.xxxxx) | ACM Multimedia 2025
20
+
21
+ ## Overview
22
+
23
+ MIHBench is a comprehensive benchmark for evaluating multi-image understanding and hallucination in Multi-modal Large Language Models (MLLMs). It contains **3,200 samples** across **4 tasks** (800 samples each), with each sample containing 2-4 images.
24
+
25
+ ## Tasks
26
+
27
+ | Task | # Images | Description |
28
+ |------|----------|-------------|
29
+ | **Count** | 2 | Determine whether the same number of a target object appears in both images. 400 samples include injected distracting objects. |
30
+ | **Existence (Adversarial)** | 3 | Determine whether a target object exists in all images, with adversarially selected objects (rare, confusing). |
31
+ | **Existence (Popular)** | 3 | Determine whether a target object exists in all images, using commonly known objects. |
32
+ | **Existence (Random)** | 3 | Determine whether a target object exists in all images, using randomly selected objects. |
33
+
34
+ > **Note**: A 5th task (ID Consistency) will be added in a future update.
35
+
36
+ ## Dataset Schema
37
+
38
+ ### Common columns (all tasks)
39
+
40
+ | Column | Type | Description |
41
+ |--------|------|-------------|
42
+ | `images` | `list[image]` | 2-4 images (PIL Image objects) |
43
+ | `question` | `str` | Natural language question about the images |
44
+ | `label` | `str` | Ground truth answer (`"yes"` or `"no"`) |
45
+ | `task` | `str` | Task identifier |
46
+ | `num_images` | `int` | Number of images in the sample |
47
+ | `image_names` | `list[str]` | Source image filenames |
48
+
49
+ ### Additional columns (Count task only)
50
+
51
+ | Column | Type | Description |
52
+ |--------|------|-------------|
53
+ | `injected` | `bool` | Whether distracting objects were injected into the question |
54
+ | `object_counts` | `str` | JSON string mapping image identifiers to object counts (e.g., `'{"A": 1, "B": 1}'`) |
55
+
56
+ ## Data Splits
57
+
58
+ Each task is a separate configuration/split with 800 samples (400 `"yes"`, 400 `"no"`).
59
+
60
+ ## Image Sources
61
+
62
+ - **Tasks 1-4**: COCO (Common Objects in Context) dataset
63
+ - **Task 5** (ID Consistency, coming soon): CO3D dataset
64
+
65
+ ## Citation
66
+
67
+ If you use MIHBench in your research, please cite:
68
+
69
+ ```bibtex
70
+ @inproceedings{mihbench2025,
71
+ title={MIHBench: Can Multi-modal Large Language Models Understand Multi-Image Inputs?},
72
+ author={},
73
+ booktitle={Proceedings of the ACM Multimedia 2025},
74
+ year={2025}
75
+ }
76
+ ```
77
+
78
+ ## License
79
+
80
+ This dataset is released under the [CC-BY-4.0](https://creativecommons.org/licenses/by/4.0/) license. The underlying images are from COCO and CO3D, which have their own licenses.