Wild_OmniDocBench / README.md
VirtualLUO's picture
Upload README.md with huggingface_hub
7c1dd17 verified
# Wild-OmniDocBench
**A Real-World Captured Document Parsing Benchmark for Robustness Evaluation**
<p align="center">
<a href="https://huggingface.co/datasets/VirtualLUO/Wild_OmniDocBench/blob/main/README_ZH.md">中文版</a>
<a href="https://arxiv.org/abs/2603.23885">Paper</a>
<a href="https://github.com/VirtualLUOUCAS/Wild_OmniDocBench">GitHub</a>
<a href="https://huggingface.co/datasets/VirtualLUO/Wild_OmniDocBench">HuggingFace</a>
</p>
## Overview
**Wild-OmniDocBench** is a benchmark for evaluating document parsing robustness under real-world captured conditions. It is derived from [OmniDocBench](https://github.com/opendatalab/OmniDocBench) by converting scanned/digital documents into naturally captured images through controlled physical simulation, including printing, deformation, and photography under diverse lighting conditions.
Unlike standard benchmarks that rely on clean scanned or digital-born pages, Wild-OmniDocBench introduces realistic artifacts such as:
- **Geometric distortions** (perspective shifts, bends, wrinkles)
- **Illumination variations** (directional, uneven, low-light)
- **Screen capture artifacts** (moire patterns, reflections)
- **Environmental interference** (background overlays, shadows)
> **Note:** The current release of Wild-OmniDocBench corresponds to **OmniDocBench v1.5**. We are currently processing the extended portions for v1.6 and will release them in a future update.
<p align="center">
<img src="assets/overview.png" width="90%" alt="Wild-OmniDocBench Construction">
</p>
## Benchmark Statistics
| Item | Details |
|------|---------|
| Total Images | 1,350 |
| Source | Real-world captured variant of OmniDocBench |
| Document Types | Books, Textbooks, Papers, PPTs, Newspapers, Notes, Exams, Magazines, Financial Reports, etc. |
| Capture Methods | (i) Print + physical deformation + photography; (ii) Screen display + re-capture |
| Annotations | Inherited from OmniDocBench (full structural and reading-order annotations) |
## Data Format
### Directory Structure
```
Wild_OmniDocBench/
├── README.md # English README
├── README_ZH.md # Chinese README
├── wild_omnidocbench.zip # Benchmark images (1,350 JPGs)
└── assets/
└── overview.png # Overview figure
```
### Images
After unzipping `wild_omnidocbench.zip`, images are named following the OmniDocBench convention:
```
{doc_type}_{language}_{source}_{page}.jpg
```
For example: `book_en_A.Concise.Introduction.to.Linear.Algebra_page_065.jpg`
## Evaluation
Wild-OmniDocBench uses the same annotation format and evaluation protocol as [OmniDocBench](https://github.com/opendatalab/OmniDocBench). To evaluate on Wild-OmniDocBench:
1. **Obtain annotations and evaluation scripts** from the official OmniDocBench repository:
```
https://github.com/opendatalab/OmniDocBench
```
2. **Replace the image source** with Wild-OmniDocBench images (from `wild_omnidocbench.zip`).
3. **Run evaluation** following the OmniDocBench protocol. Metrics include:
- **Overall Score** (↑)
- **Text Edit Distance** (↓)
- **Formula CDM** (↑)
- **Table TEDS** (↑)
- **Reading Order Edit Distance** (↓)
## Key Results
Performance degradation from OmniDocBench to Wild-OmniDocBench (from the DocHumming paper):
| Model | Type | Overall (Origin) | Overall (Wild) | Degradation |
|-------|------|:-:|:-:|:-:|
| DocHumming (1B) | End2End | 93.75 | 87.03 | −6.72 |
| dots.ocr (3B) | End2End | 88.41 | 78.01 | −10.40 |
| Qwen3-VL (235B) | General | 89.15 | 79.69 | −9.46 |
| MinerU2.5 (1.2B) | Modular | 90.67 | 70.91 | −19.76 |
| PaddleOCR-VL (0.9B) | Modular | 91.93 | 72.19 | −19.74 |
End-to-end models exhibit significantly less degradation than modular cascaded pipelines under real-world capture conditions.
## Citation
```bibtex
@misc{li2026towardsrealworlddocument,
title={Towards Real-World Document Parsing via Realistic Scene Synthesis and Document-Aware Training},
author={Gengluo Li and Pengyuan Lyu and Chengquan Zhang and Huawen Shen and Liang Wu and Xingyu Wan and Gangyan Zeng and Han Hu and Can Ma and Yu Zhou},
year={2026},
journal={arXiv preprint arXiv:2603.23885},
url={https://arxiv.org/abs/2603.23885},
}
```
## Acknowledgements
Wild-OmniDocBench is built upon [OmniDocBench](https://github.com/opendatalab/OmniDocBench). We thank the OmniDocBench team for providing the original annotations and evaluation framework.
## License
This benchmark is released for **research purposes only**.