File size: 4,718 Bytes
7c1dd17 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 | # Wild-OmniDocBench
**A Real-World Captured Document Parsing Benchmark for Robustness Evaluation**
<p align="center">
<a href="https://huggingface.co/datasets/VirtualLUO/Wild_OmniDocBench/blob/main/README_ZH.md">中文版</a> •
<a href="https://arxiv.org/abs/2603.23885">Paper</a> •
<a href="https://github.com/VirtualLUOUCAS/Wild_OmniDocBench">GitHub</a> •
<a href="https://huggingface.co/datasets/VirtualLUO/Wild_OmniDocBench">HuggingFace</a>
</p>
## Overview
**Wild-OmniDocBench** is a benchmark for evaluating document parsing robustness under real-world captured conditions. It is derived from [OmniDocBench](https://github.com/opendatalab/OmniDocBench) by converting scanned/digital documents into naturally captured images through controlled physical simulation, including printing, deformation, and photography under diverse lighting conditions.
Unlike standard benchmarks that rely on clean scanned or digital-born pages, Wild-OmniDocBench introduces realistic artifacts such as:
- **Geometric distortions** (perspective shifts, bends, wrinkles)
- **Illumination variations** (directional, uneven, low-light)
- **Screen capture artifacts** (moire patterns, reflections)
- **Environmental interference** (background overlays, shadows)
> **Note:** The current release of Wild-OmniDocBench corresponds to **OmniDocBench v1.5**. We are currently processing the extended portions for v1.6 and will release them in a future update.
<p align="center">
<img src="assets/overview.png" width="90%" alt="Wild-OmniDocBench Construction">
</p>
## Benchmark Statistics
| Item | Details |
|------|---------|
| Total Images | 1,350 |
| Source | Real-world captured variant of OmniDocBench |
| Document Types | Books, Textbooks, Papers, PPTs, Newspapers, Notes, Exams, Magazines, Financial Reports, etc. |
| Capture Methods | (i) Print + physical deformation + photography; (ii) Screen display + re-capture |
| Annotations | Inherited from OmniDocBench (full structural and reading-order annotations) |
## Data Format
### Directory Structure
```
Wild_OmniDocBench/
├── README.md # English README
├── README_ZH.md # Chinese README
├── wild_omnidocbench.zip # Benchmark images (1,350 JPGs)
└── assets/
└── overview.png # Overview figure
```
### Images
After unzipping `wild_omnidocbench.zip`, images are named following the OmniDocBench convention:
```
{doc_type}_{language}_{source}_{page}.jpg
```
For example: `book_en_A.Concise.Introduction.to.Linear.Algebra_page_065.jpg`
## Evaluation
Wild-OmniDocBench uses the same annotation format and evaluation protocol as [OmniDocBench](https://github.com/opendatalab/OmniDocBench). To evaluate on Wild-OmniDocBench:
1. **Obtain annotations and evaluation scripts** from the official OmniDocBench repository:
```
https://github.com/opendatalab/OmniDocBench
```
2. **Replace the image source** with Wild-OmniDocBench images (from `wild_omnidocbench.zip`).
3. **Run evaluation** following the OmniDocBench protocol. Metrics include:
- **Overall Score** (↑)
- **Text Edit Distance** (↓)
- **Formula CDM** (↑)
- **Table TEDS** (↑)
- **Reading Order Edit Distance** (↓)
## Key Results
Performance degradation from OmniDocBench to Wild-OmniDocBench (from the DocHumming paper):
| Model | Type | Overall (Origin) | Overall (Wild) | Degradation |
|-------|------|:-:|:-:|:-:|
| DocHumming (1B) | End2End | 93.75 | 87.03 | −6.72 |
| dots.ocr (3B) | End2End | 88.41 | 78.01 | −10.40 |
| Qwen3-VL (235B) | General | 89.15 | 79.69 | −9.46 |
| MinerU2.5 (1.2B) | Modular | 90.67 | 70.91 | −19.76 |
| PaddleOCR-VL (0.9B) | Modular | 91.93 | 72.19 | −19.74 |
End-to-end models exhibit significantly less degradation than modular cascaded pipelines under real-world capture conditions.
## Citation
```bibtex
@misc{li2026towardsrealworlddocument,
title={Towards Real-World Document Parsing via Realistic Scene Synthesis and Document-Aware Training},
author={Gengluo Li and Pengyuan Lyu and Chengquan Zhang and Huawen Shen and Liang Wu and Xingyu Wan and Gangyan Zeng and Han Hu and Can Ma and Yu Zhou},
year={2026},
journal={arXiv preprint arXiv:2603.23885},
url={https://arxiv.org/abs/2603.23885},
}
```
## Acknowledgements
Wild-OmniDocBench is built upon [OmniDocBench](https://github.com/opendatalab/OmniDocBench). We thank the OmniDocBench team for providing the original annotations and evaluation framework.
## License
This benchmark is released for **research purposes only**.
|