Crysun commited on
Commit
ca79220
·
verified ·
1 Parent(s): 5fc1ad7

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +109 -0
README.md ADDED
@@ -0,0 +1,109 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ task_categories:
4
+ - visual-question-answering
5
+ - image-to-text
6
+ language:
7
+ - en
8
+ tags:
9
+ - benchmark
10
+ - multimodal
11
+ - reasoning
12
+ - visual-grounding
13
+ - mllm-evaluation
14
+ pretty_name: DailyClue
15
+ size_categories:
16
+ - n<1K
17
+ ---
18
+
19
+ # DailyClue Dataset
20
+
21
+ **Seek-and-Solve: Benchmarking MLLMs for Visual Clue-Driven Reasoning in Daily Scenarios**
22
+
23
+ [![Paper](https://img.shields.io/badge/Paper-arXiv-red)](https://arxiv.org/abs/2604.14041)
24
+ [![Code](https://img.shields.io/badge/Code-GitHub-black)](https://github.com/your-org/DailyClue)
25
+ [![License: MIT](https://img.shields.io/badge/License-MIT-blue.svg)](https://opensource.org/licenses/MIT)
26
+
27
+ ## Dataset Summary
28
+
29
+ DailyClue is a benchmark for evaluating **visual clue-driven reasoning** in Multimodal Large Language Models (MLLMs). Unlike benchmarks that test pre-existing knowledge, DailyClue requires models to actively identify decisive visual clues from images before producing answers.
30
+
31
+ The dataset spans **4 major domains** and **16 distinct subtasks**, with **666 total questions**.
32
+
33
+ ## Dataset Structure
34
+
35
+ ```
36
+ DailyClue/
37
+ ├── daily_life/ # Images for Daily Commonsense Reasoning
38
+ ├── location/ # Images for Location Identification
39
+ ├── science/ # Images for Scientific Commonsense
40
+ ├── spatial/ # Images for Spatial Reasoning
41
+ ├── daily_life.json
42
+ ├── location.json
43
+ ├── science.json
44
+ └── spatial.json
45
+ ```
46
+
47
+ ## Statistics
48
+
49
+ | Category | # Questions | Formats |
50
+ |---|---|---|
51
+ | Daily Commonsense Reasoning | 180 | Multiple Choice, Yes/No, Open-ended |
52
+ | Location Identification | 200 | Open-ended |
53
+ | Spatial Reasoning | 163 | Multiple Choice, Yes/No |
54
+ | Scientific Commonsense | 123 | Multiple Choice, Yes/No, Open-ended |
55
+ | **Total** | **666** | |
56
+
57
+ ## Data Fields
58
+
59
+ Each JSON entry contains:
60
+
61
+ | Field | Type | Description |
62
+ |---|---|---|
63
+ | `image` | `list[str]` | Image filename(s) within the category subfolder |
64
+ | `question` | `str` | The question posed to the model |
65
+ | `clues` | `str` | Human-annotated ground-truth visual clues (see note below) |
66
+ | `ground_truth` | `str` | The correct answer |
67
+ | `format` | `str` | `"Multiple choice"`, `"Yes or no"`, or `"Open-ended"` |
68
+ | `category_1` | `str` | Primary domain (one of the four above) |
69
+ | `category_2` | `str` | Subtask within the primary domain |
70
+ | `language` | `str` | `"English"` |
71
+
72
+ > **Note on `clues`**: This field contains human-annotated ground-truth visual clues. It is used in ablation experiments (injecting GT clues to probe the impact on model accuracy) and in the Rigorous Evaluation Protocol (checking whether model-predicted clues semantically align with GT clues). It is **not** fed to the model during standard inference.
73
+
74
+ ## Usage
75
+
76
+ ### Download
77
+
78
+ ```bash
79
+ # via git
80
+ git clone https://huggingface.co/datasets/Crysun/DailyClue
81
+
82
+ # via huggingface_hub
83
+ from huggingface_hub import snapshot_download
84
+ snapshot_download(repo_id="Crysun/DailyClue", repo_type="dataset", local_dir="./dataset")
85
+ ```
86
+
87
+ ### Run Inference
88
+
89
+ After downloading, point the inference script to the local directory:
90
+
91
+ ```bash
92
+ python infer/inference.py \
93
+ --dataset ./dataset \
94
+ --model_names "gpt-4o" \
95
+ --prompt_mode "b"
96
+ ```
97
+
98
+ See the [GitHub repository](https://github.com/your-org/DailyClue) for the full evaluation pipeline.
99
+
100
+ ## Citation
101
+
102
+ ```bibtex
103
+ @article{dailyclue2026,
104
+ title={Seek-and-Solve: Benchmarking MLLMs for Visual Clue-Driven Reasoning in Daily Scenarios},
105
+ author={Li, Xiaomin and Wang, Tala and Zhong, Zichen and Zhang, Ying and Zheng, Zirui and Isobe, Takashi and Li, Dezhuang and Lu, Huchuan and He, You and Jia, Xu},
106
+ journal={arXiv preprint arXiv:2604.14041},
107
+ year={2026}
108
+ }
109
+ ```