VirtualLUO commited on
Commit
7c1dd17
·
verified ·
1 Parent(s): bad1a58

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +111 -0
README.md ADDED
@@ -0,0 +1,111 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Wild-OmniDocBench
2
+
3
+ **A Real-World Captured Document Parsing Benchmark for Robustness Evaluation**
4
+
5
+ <p align="center">
6
+ <a href="https://huggingface.co/datasets/VirtualLUO/Wild_OmniDocBench/blob/main/README_ZH.md">中文版</a> •
7
+ <a href="https://arxiv.org/abs/2603.23885">Paper</a> •
8
+ <a href="https://github.com/VirtualLUOUCAS/Wild_OmniDocBench">GitHub</a> •
9
+ <a href="https://huggingface.co/datasets/VirtualLUO/Wild_OmniDocBench">HuggingFace</a>
10
+ </p>
11
+
12
+ ## Overview
13
+
14
+ **Wild-OmniDocBench** is a benchmark for evaluating document parsing robustness under real-world captured conditions. It is derived from [OmniDocBench](https://github.com/opendatalab/OmniDocBench) by converting scanned/digital documents into naturally captured images through controlled physical simulation, including printing, deformation, and photography under diverse lighting conditions.
15
+
16
+ Unlike standard benchmarks that rely on clean scanned or digital-born pages, Wild-OmniDocBench introduces realistic artifacts such as:
17
+ - **Geometric distortions** (perspective shifts, bends, wrinkles)
18
+ - **Illumination variations** (directional, uneven, low-light)
19
+ - **Screen capture artifacts** (moire patterns, reflections)
20
+ - **Environmental interference** (background overlays, shadows)
21
+
22
+ > **Note:** The current release of Wild-OmniDocBench corresponds to **OmniDocBench v1.5**. We are currently processing the extended portions for v1.6 and will release them in a future update.
23
+
24
+ <p align="center">
25
+ <img src="assets/overview.png" width="90%" alt="Wild-OmniDocBench Construction">
26
+ </p>
27
+
28
+ ## Benchmark Statistics
29
+
30
+ | Item | Details |
31
+ |------|---------|
32
+ | Total Images | 1,350 |
33
+ | Source | Real-world captured variant of OmniDocBench |
34
+ | Document Types | Books, Textbooks, Papers, PPTs, Newspapers, Notes, Exams, Magazines, Financial Reports, etc. |
35
+ | Capture Methods | (i) Print + physical deformation + photography; (ii) Screen display + re-capture |
36
+ | Annotations | Inherited from OmniDocBench (full structural and reading-order annotations) |
37
+
38
+ ## Data Format
39
+
40
+ ### Directory Structure
41
+
42
+ ```
43
+ Wild_OmniDocBench/
44
+ ├── README.md # English README
45
+ ├── README_ZH.md # Chinese README
46
+ ├── wild_omnidocbench.zip # Benchmark images (1,350 JPGs)
47
+ └── assets/
48
+ └── overview.png # Overview figure
49
+ ```
50
+
51
+ ### Images
52
+
53
+ After unzipping `wild_omnidocbench.zip`, images are named following the OmniDocBench convention:
54
+
55
+ ```
56
+ {doc_type}_{language}_{source}_{page}.jpg
57
+ ```
58
+
59
+ For example: `book_en_A.Concise.Introduction.to.Linear.Algebra_page_065.jpg`
60
+
61
+ ## Evaluation
62
+
63
+ Wild-OmniDocBench uses the same annotation format and evaluation protocol as [OmniDocBench](https://github.com/opendatalab/OmniDocBench). To evaluate on Wild-OmniDocBench:
64
+
65
+ 1. **Obtain annotations and evaluation scripts** from the official OmniDocBench repository:
66
+ ```
67
+ https://github.com/opendatalab/OmniDocBench
68
+ ```
69
+
70
+ 2. **Replace the image source** with Wild-OmniDocBench images (from `wild_omnidocbench.zip`).
71
+
72
+ 3. **Run evaluation** following the OmniDocBench protocol. Metrics include:
73
+ - **Overall Score** (↑)
74
+ - **Text Edit Distance** (↓)
75
+ - **Formula CDM** (↑)
76
+ - **Table TEDS** (↑)
77
+ - **Reading Order Edit Distance** (↓)
78
+
79
+ ## Key Results
80
+
81
+ Performance degradation from OmniDocBench to Wild-OmniDocBench (from the DocHumming paper):
82
+
83
+ | Model | Type | Overall (Origin) | Overall (Wild) | Degradation |
84
+ |-------|------|:-:|:-:|:-:|
85
+ | DocHumming (1B) | End2End | 93.75 | 87.03 | −6.72 |
86
+ | dots.ocr (3B) | End2End | 88.41 | 78.01 | −10.40 |
87
+ | Qwen3-VL (235B) | General | 89.15 | 79.69 | −9.46 |
88
+ | MinerU2.5 (1.2B) | Modular | 90.67 | 70.91 | −19.76 |
89
+ | PaddleOCR-VL (0.9B) | Modular | 91.93 | 72.19 | −19.74 |
90
+
91
+ End-to-end models exhibit significantly less degradation than modular cascaded pipelines under real-world capture conditions.
92
+
93
+ ## Citation
94
+
95
+ ```bibtex
96
+ @misc{li2026towardsrealworlddocument,
97
+ title={Towards Real-World Document Parsing via Realistic Scene Synthesis and Document-Aware Training},
98
+ author={Gengluo Li and Pengyuan Lyu and Chengquan Zhang and Huawen Shen and Liang Wu and Xingyu Wan and Gangyan Zeng and Han Hu and Can Ma and Yu Zhou},
99
+ year={2026},
100
+ journal={arXiv preprint arXiv:2603.23885},
101
+ url={https://arxiv.org/abs/2603.23885},
102
+ }
103
+ ```
104
+
105
+ ## Acknowledgements
106
+
107
+ Wild-OmniDocBench is built upon [OmniDocBench](https://github.com/opendatalab/OmniDocBench). We thank the OmniDocBench team for providing the original annotations and evaluation framework.
108
+
109
+ ## License
110
+
111
+ This benchmark is released for **research purposes only**.