File size: 4,718 Bytes
7c1dd17
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
# Wild-OmniDocBench

**A Real-World Captured Document Parsing Benchmark for Robustness Evaluation**

<p align="center">
  <a href="https://huggingface.co/datasets/VirtualLUO/Wild_OmniDocBench/blob/main/README_ZH.md">中文版</a><a href="https://arxiv.org/abs/2603.23885">Paper</a><a href="https://github.com/VirtualLUOUCAS/Wild_OmniDocBench">GitHub</a><a href="https://huggingface.co/datasets/VirtualLUO/Wild_OmniDocBench">HuggingFace</a>
</p>

## Overview

**Wild-OmniDocBench** is a benchmark for evaluating document parsing robustness under real-world captured conditions. It is derived from [OmniDocBench](https://github.com/opendatalab/OmniDocBench) by converting scanned/digital documents into naturally captured images through controlled physical simulation, including printing, deformation, and photography under diverse lighting conditions.

Unlike standard benchmarks that rely on clean scanned or digital-born pages, Wild-OmniDocBench introduces realistic artifacts such as:
- **Geometric distortions** (perspective shifts, bends, wrinkles)
- **Illumination variations** (directional, uneven, low-light)
- **Screen capture artifacts** (moire patterns, reflections)
- **Environmental interference** (background overlays, shadows)

> **Note:** The current release of Wild-OmniDocBench corresponds to **OmniDocBench v1.5**. We are currently processing the extended portions for v1.6 and will release them in a future update.

<p align="center">
  <img src="assets/overview.png" width="90%" alt="Wild-OmniDocBench Construction">
</p>

## Benchmark Statistics

| Item | Details |
|------|---------|
| Total Images | 1,350 |
| Source | Real-world captured variant of OmniDocBench |
| Document Types | Books, Textbooks, Papers, PPTs, Newspapers, Notes, Exams, Magazines, Financial Reports, etc. |
| Capture Methods | (i) Print + physical deformation + photography; (ii) Screen display + re-capture |
| Annotations | Inherited from OmniDocBench (full structural and reading-order annotations) |

## Data Format

### Directory Structure

```

Wild_OmniDocBench/

├── README.md                   # English README

├── README_ZH.md                # Chinese README

├── wild_omnidocbench.zip       # Benchmark images (1,350 JPGs)

└── assets/

    └── overview.png            # Overview figure

```

### Images

After unzipping `wild_omnidocbench.zip`, images are named following the OmniDocBench convention:

```

{doc_type}_{language}_{source}_{page}.jpg

```

For example: `book_en_A.Concise.Introduction.to.Linear.Algebra_page_065.jpg`

## Evaluation

Wild-OmniDocBench uses the same annotation format and evaluation protocol as [OmniDocBench](https://github.com/opendatalab/OmniDocBench). To evaluate on Wild-OmniDocBench:

1. **Obtain annotations and evaluation scripts** from the official OmniDocBench repository:
   ```

   https://github.com/opendatalab/OmniDocBench

   ```

2. **Replace the image source** with Wild-OmniDocBench images (from `wild_omnidocbench.zip`).

3. **Run evaluation** following the OmniDocBench protocol. Metrics include:
   - **Overall Score** (↑)
   - **Text Edit Distance** (↓)
   - **Formula CDM** (↑)
   - **Table TEDS** (↑)
   - **Reading Order Edit Distance** (↓)

## Key Results

Performance degradation from OmniDocBench to Wild-OmniDocBench (from the DocHumming paper):

| Model | Type | Overall (Origin) | Overall (Wild) | Degradation |
|-------|------|:-:|:-:|:-:|
| DocHumming (1B) | End2End | 93.75 | 87.03 | −6.72 |
| dots.ocr (3B) | End2End | 88.41 | 78.01 | −10.40 |
| Qwen3-VL (235B) | General | 89.15 | 79.69 | −9.46 |
| MinerU2.5 (1.2B) | Modular | 90.67 | 70.91 | −19.76 |
| PaddleOCR-VL (0.9B) | Modular | 91.93 | 72.19 | −19.74 |

End-to-end models exhibit significantly less degradation than modular cascaded pipelines under real-world capture conditions.

## Citation

```bibtex

@misc{li2026towardsrealworlddocument,

      title={Towards Real-World Document Parsing via Realistic Scene Synthesis and Document-Aware Training},

      author={Gengluo Li and Pengyuan Lyu and Chengquan Zhang and Huawen Shen and Liang Wu and Xingyu Wan and Gangyan Zeng and Han Hu and Can Ma and Yu Zhou},

      year={2026},

      journal={arXiv preprint arXiv:2603.23885},

      url={https://arxiv.org/abs/2603.23885},

}

```

## Acknowledgements

Wild-OmniDocBench is built upon [OmniDocBench](https://github.com/opendatalab/OmniDocBench). We thank the OmniDocBench team for providing the original annotations and evaluation framework.

## License

This benchmark is released for **research purposes only**.