| --- |
| license: cc-by-nc-sa-4.0 |
| pretty_name: INTERCHART |
| tags: |
| - charts |
| - visualization |
| - vqa |
| - multimodal |
| - question-answering |
| - reasoning |
| - benchmarking |
| - evaluation |
| task_categories: |
| - question-answering |
| - visual-question-answering |
| task_ids: |
| - visual-question-answering |
| language: |
| - en |
| dataset_info: |
| features: |
| - name: id |
| dtype: string |
| - name: subset |
| dtype: string |
| - name: context_format |
| dtype: string |
| - name: question |
| dtype: string |
| - name: answer |
| dtype: string |
| - name: images |
| sequence: string |
| - name: metadata |
| dtype: json |
| pretty_description: > |
| INTERCHART is a diagnostic benchmark for multi-chart visual reasoning across three tiers: |
| DECAF (decomposed single-entity charts), SPECTRA (synthetic paired charts for correlated trends), |
| and STORM (real-world chart pairs). The dataset includes chart images and question–answer pairs |
| designed to stress-test cross-chart reasoning, trend correlation, and abstract numerical inference. |
| --- |
| |
| # INTERCHART: Benchmarking Visual Reasoning Across Decomposed and Distributed Chart Information |
|
|
| [](https://coral-lab-asu.github.io/interchart/) |
| [](https://arxiv.org/abs/2508.07630v1) |
| [](https://creativecommons.org/licenses/by-nc-sa/4.0/) |
|
|
| --- |
|
|
| ## 🧩 Overview |
|
|
| **INTERCHART** is a multi-tier benchmark that evaluates how well **vision-language models (VLMs)** reason across **multiple related charts**, a crucial skill for real-world applications like scientific reports, financial analyses, and policy dashboards. |
| Unlike single-chart benchmarks, INTERCHART challenges models to integrate information across **decomposed**, **synthetic**, and **real-world** chart contexts. |
|
|
| > **Paper:** [INTERCHART: Benchmarking Visual Reasoning Across Decomposed and Distributed Chart Information](https://arxiv.org/abs/2508.07630v1) |
|
|
| --- |
|
|
| ## 📂 Dataset Structure |
|
|
| ``` |
| |
| INTERCHART/ |
| ├── DECAF |
| │ ├── combined # Multi-chart combined images (stitched) |
| │ ├── original # Original compound charts |
| │ ├── questions # QA pairs for decomposed single-variable charts |
| │ └── simple # Simplified decomposed charts |
| ├── SPECTRA |
| │ ├── combined # Synthetic chart pairs (shared axes) |
| │ ├── questions # QA pairs for correlated and independent reasoning |
| │ └── simple # Individual charts rendered from synthetic tables |
| ├── STORM |
| │ ├── combined # Real-world chart pairs (stitched) |
| │ ├── images # Original Our World in Data charts |
| │ ├── meta-data # Extracted metadata and semantic pairings |
| │ ├── questions # QA pairs for temporal, cross-domain reasoning |
| │ └── tables # Structured table representations (optional) |
| |
| ```` |
|
|
| Each subset targets a different **level of reasoning complexity** and visual diversity. |
|
|
| --- |
|
|
| ## 🧠 Subset Descriptions |
|
|
| ### **1️⃣ DECAF** — *Decomposed Elementary Charts with Answerable Facts* |
| - Focus: **Factual lookup** and **comparative reasoning** on simplified single-variable charts. |
| - Sources: Derived from ChartQA, ChartLlama, ChartInfo, DVQA. |
| - Content: 1,188 decomposed charts and 2,809 QA pairs. |
| - Tasks: Identify, compare, or extract values across clean, minimal visuals. |
|
|
| --- |
|
|
| ### **2️⃣ SPECTRA** — *Synthetic Plots for Event-based Correlated Trend Reasoning and Analysis* |
| - Focus: **Trend correlation** and **scenario-based inference** between synthetic chart pairs. |
| - Construction: Generated via Gemini 1.5 Pro + human validation to preserve shared axes and realism. |
| - Content: 870 unique charts, 1,717 QA pairs across 333 contexts. |
| - Tasks: Analyze multi-variable relationships, infer trends, and reason about co-evolving variables. |
|
|
| --- |
|
|
| ### **3️⃣ STORM** — *Sequential Temporal Reasoning Over Real-world Multi-domain Charts* |
| - Focus: **Multi-step reasoning**, **temporal analysis**, and **semantic alignment** across real-world charts. |
| - Source: Curated from *Our World in Data* with metadata-driven semantic pairing. |
| - Content: 648 charts across 324 validated contexts, 768 QA pairs. |
| - Tasks: Align mismatched domains, estimate ranges, and reason about evolving trends. |
|
|
| --- |
|
|
| ## ⚙️ Evaluation & Methodology |
|
|
| INTERCHART supports both **visual** and **table-based** evaluation modes. |
|
|
| - **Visual Inputs:** |
| - *Combined:* Charts stitched into a unified image. |
| - *Interleaved:* Charts provided sequentially. |
|
|
| - **Structured Table Inputs:** |
| Models can extract tables using tools like **DePlot** or **Gemini Title Extraction**, followed by **table-based QA**. |
|
|
| - **Prompting Strategies:** |
| - Zero-Shot |
| - Zero-Shot Chain-of-Thought (CoT) |
| - Few-Shot CoT with Directives (CoTD) |
|
|
| - **Evaluation Pipeline:** |
| Multi-LLM *semantic judging* (Gemini 1.5 Flash, Phi-4, Qwen2.5) with **majority voting** to evaluate semantic correctness. |
|
|
| --- |
|
|
| ## 📊 Dataset Statistics |
|
|
| | Subset | Charts | Contexts | QA Pairs | Reasoning Type Examples | |
| |----------|---------|-----------|-----------|--------------------------| |
| | **DECAF** | 1,188 | 355 | 2,809 | Factual lookup, comparison | |
| | **SPECTRA** | 870 | 333 | 1,717 | Trend correlation, event reasoning | |
| | **STORM** | 648 | 324 | 768 | Temporal reasoning, abstract numerical inference | |
| | **Total** | 2,706 | 1,012 | **5,214** | — | |
|
|
| --- |
|
|
| ## 🚀 Usage |
|
|
| ### 🔐 Access & Download Instructions |
|
|
| Use an **access token** as your Git credential when cloning or pushing to the repository. |
|
|
| 1. **Install Git LFS** |
| Download and install from [https://git-lfs.com](https://git-lfs.com). |
| Then run: |
| ``` |
| |
| git lfs install |
| |
| ``` |
|
|
| 2. **Clone the dataset repository** |
| When prompted for a password, use your **Hugging Face access token** with *write permissions*. |
| You can generate one here: [https://huggingface.co/settings/tokens](https://huggingface.co/settings/tokens) |
|
|
| ``` |
| |
| git clone [https://huggingface.co/datasets/interchart/Interchart](https://huggingface.co/datasets/interchart/Interchart) |
| |
| ``` |
|
|
| 3. **Clone without large files (LFS pointers only)** |
| If you only want lightweight clones without downloading all image data: |
| ``` |
| |
| GIT_LFS_SKIP_SMUDGE=1 git clone [https://huggingface.co/datasets/interchart/Interchart](https://huggingface.co/datasets/interchart/Interchart) |
| |
| ``` |
|
|
| 4. **Alternative: use the Hugging Face CLI** |
| Make sure the CLI is installed: |
| ``` |
| |
| pip install -U "huggingface_hub[cli]" |
| |
| ``` |
|
|
| Then download directly: |
| ``` |
| |
| hf download interchart/Interchart --repo-type=dataset |
| |
| ``` |
|
|
|
|
|
|
| --- |
|
|
| ## 🔍 Citation |
|
|
| If you use this dataset, please cite: |
|
|
| ``` |
| @article{iyengar2025interchart, |
| title={INTERCHART: Benchmarking Visual Reasoning Across Decomposed and Distributed Chart Information}, |
| author={Anirudh Iyengar Kaniyar Narayana Iyengar and Srija Mukhopadhyay and Adnan Qidwai and Shubhankar Singh and Dan Roth and Vivek Gupta}, |
| journal={arXiv preprint arXiv:2508.07630}, |
| year={2025} |
| } |
| ``` |
|
|
| --- |
|
|
| ## 🔗 Links |
|
|
| * 📘 **Paper:** [arXiv:2508.07630v1](https://arxiv.org/abs/2508.07630v1) |
| * 🌐 **Website:** [https://coral-lab-asu.github.io/interchart/](https://coral-lab-asu.github.io/interchart/) |
| * 🧠 **Explore Dataset:** [Interactive Evaluation Portal](https://coral-lab-asu.github.io/interchart/explore.html) |