LIQIIIII commited on
Commit
25601eb
·
verified ·
1 Parent(s): 269da3a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +171 -1
README.md CHANGED
@@ -28,4 +28,174 @@ configs:
28
  data_files:
29
  - split: eval
30
  path: metadata/vimu_ss.jsonl
31
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
28
  data_files:
29
  - split: eval
30
  path: metadata/vimu_ss.jsonl
31
+ ---
32
+
33
+ <div align="center">
34
+
35
+ <img src="assets/overall.png" width="100%"/>
36
+
37
+ # ViMU: Benchmarking Video Metaphorical Understanding
38
+
39
+ [![Project Page](https://img.shields.io/badge/Project-Page-blue?style=for-the-badge&logo=googlechrome&logoColor=white)](https://liqiiiii.github.io/Video-Metaphorical-Understanding/)
40
+ [![arXiv](https://img.shields.io/badge/arXiv-Paper-b31b1b?style=for-the-badge&logo=arxiv&logoColor=white)](https://arxiv.org/abs/2605.14607)
41
+ [![Hugging Face](https://img.shields.io/badge/HuggingFace-Dataset-yellow?style=for-the-badge&logo=huggingface&logoColor=black)](https://huggingface.co/datasets/LIQIIIII/ViMU)
42
+
43
+
44
+ [Qi Li](https://liqiiiii.github.io/), [Xinchao Wang](https://sites.google.com/site/sitexinchaowang/)<sup>*</sup>
45
+
46
+ <sup>*</sup>Corresponding author
47
+
48
+ [xML Lab](https://sites.google.com/view/xml-nus), National University of Singapore
49
+
50
+ </div>
51
+
52
+ This repository contains the evaluation scripts for ViMU, a benchmark for video metaphorical understanding. The code evaluates multimodal models on four tasks:
53
+
54
+ 1. Open-ended interpretation (OE)
55
+ 2. Evidence grounding (EG)
56
+ 3. Rhetoric mechanism identification (RM)
57
+ 4. Social value signal identification (SV)
58
+
59
+ ## Directory Structure
60
+
61
+ Expected project structure:
62
+
63
+ ```text
64
+ ViMU/
65
+ ├── videos/
66
+ │ ├── vimu_000001.mp4
67
+ │ └── ...
68
+ ├── metadata/
69
+ │ ├── vimu_oe.jsonl
70
+ │ ├── vimu_eg.jsonl
71
+ │ ├── vimu_ss.jsonl
72
+ │ ├── video_evidence.jsonl
73
+ │ └── cache/
74
+ ├── scripts/
75
+ │ ├── 00-vimu_oe.py
76
+ │ ├── 01-vimu_oe_judge.py
77
+ │ ├── 02-vimu_oe_score.py
78
+ │ ├── 10-vimu_eg.py
79
+ │ ├── 11-vimu_eg_score.py
80
+ │ ├── 20-vimu_ss.py
81
+ │ ├── 21-vimu_ss_score.py
82
+ │ └── utils.py
83
+ └── output/
84
+ ````
85
+
86
+ ## Setup
87
+
88
+ Install dependencies:
89
+
90
+ ```bash
91
+ pip install openai requests numpy pandas tqdm
92
+ ```
93
+
94
+ Depending on the models used, additional API keys may be required.
95
+
96
+ Set API keys:
97
+
98
+ ```bash
99
+ export OPENAI_API_KEY="your_openai_key"
100
+ export OPENROUTER_API_KEY="your_openrouter_key"
101
+ export GOOGLE_API_KEY="your_google_key"
102
+ ```
103
+
104
+ Not all keys are required if you only run a subset of models.
105
+
106
+ ## Path Configuration
107
+
108
+ Before running, edit each script and set:
109
+
110
+ ```python
111
+ PROJECT_ROOT = "/Your/Path/To/ViMU"
112
+ ```
113
+
114
+ ## Recommended Running Order
115
+
116
+ For a full evaluation, run:
117
+
118
+ ```bash
119
+ # Open-ended interpretation
120
+ python scripts/00-vimu_oe.py
121
+ python scripts/01-vimu_oe_judge.py
122
+ python scripts/02-vimu_oe_score.py
123
+
124
+ # Evidence grounding
125
+ python scripts/10-vimu_eg.py
126
+ python scripts/11-vimu_eg_score.py
127
+
128
+ # Structured subtext tasks without guidance
129
+ python scripts/20-vimu_ss.py --prompt_mode without_guidance
130
+ python scripts/21-vimu_ss_score.py --prompt_mode without_guidance
131
+
132
+ # Structured subtext tasks with guidance
133
+ python scripts/20-vimu_ss.py --prompt_mode with_guidance
134
+ python scripts/21-vimu_ss_score.py --prompt_mode with_guidance
135
+ ```
136
+
137
+ ## Model Configuration
138
+
139
+ Models are configured in the `MODEL_SPECS` list inside the inference scripts.
140
+
141
+ To enable or disable a model, edit:
142
+
143
+ ```python
144
+ "enabled": True
145
+ ```
146
+
147
+ or
148
+
149
+ ```python
150
+ "enabled": False
151
+ ```
152
+
153
+ For OpenRouter models, make sure the model ID and API key are valid.
154
+
155
+ ## Output Files
156
+
157
+ The main output files are:
158
+
159
+ ```text
160
+ output/vimu_oe_summary.json
161
+ output/vimu_eg_summary.json
162
+ output/vimu_ss_without_guidance_summary.json
163
+ output/vimu_ss_with_guidance_summary.json
164
+ ```
165
+
166
+ These files contain aggregated evaluation results.
167
+
168
+ ## Scoring Rules
169
+
170
+ ### Open-ended Interpretation
171
+
172
+ Open-ended answers are evaluated using an LLM-as-a-judge protocol. The judge scores semantic understanding based on:
173
+
174
+ ```text
175
+ core intent
176
+ implicit signal
177
+ target or social meaning
178
+ hallucination penalty
179
+ literal-only penalty
180
+ ```
181
+
182
+ Evidence grounding is scored as a multi-label prediction problem. If the prediction contains any incorrect option, the score is 0. Otherwise, if the prediction is a subset of the gold answer, the score is: `score = number of correctly selected options / number of gold options`. Rhetoric and social value tasks use the same multi-label scoring rule. If no incorrect option is selected; otherwise: `score = 0`.
183
+
184
+ ## Notes
185
+
186
+ The dataset contains socially sensitive video memes. The benchmark is intended for research use only.
187
+
188
+ ## Citation
189
+
190
+ If you finding our work interesting or helpful to you, please cite as follows:
191
+
192
+ ```
193
+ @article{li2026vimu,
194
+ title={ViMU: Benchmarking Video Metaphorical Understanding},
195
+ author={Li, Qi and Wang, Xinchao},
196
+ journal={arXiv preprint arXiv:2605.14607},
197
+ year={2026}
198
+ }
199
+ ```
200
+
201
+