Codesota · OCR · Benchmark · OmniDocBench47 scored runs · 34 distinct modelsUpdated 2026-04-20
§ 00 · Opening

Every PDF element, scored honestly.

OmniDocBench is the Shanghai AI Laboratory benchmark for end-to-end PDF document parsing. It grades text extraction, table structure, formula recognition and layout on the same page — which is why it decides who ships into an accounts-payable pipeline and who does not.

§ 01 · Leaderboard · Composite score

Composite score, ranked.

The headline OmniDocBench score: ((1 − TextEditDist)·100 + TableTEDS + FormulaCDM) / 3. (higher is better)

#ModelComposite scoreVerifiedSource
01GLM-OCR
Fetched from CodeSOTA API on 2026-04-20
94.62codesota-api
02PaddleOCR-VL-1.5
Fetched from CodeSOTA API on 2026-04-20
94.50codesota-api
03Qianfan-OCR
Fetched from CodeSOTA API on 2026-04-20
93.12codesota-api
04paddleocr-vl
Fetched from CodeSOTA API on 2026-04-20
92.86codesota-api
05paddleocr-vl-0.9b
Fetched from CodeSOTA API on 2026-04-20
92.56codesota-api
06mineru-2.5
Fetched from CodeSOTA API on 2026-04-20
90.67codesota-api
07Gemini 3 Pro
Fetched from CodeSOTA API on 2026-04-20
90.33codesota-api
08Dolphin-v2
Fetched from CodeSOTA API on 2026-04-20
89.78codesota-api
09qwen3-vl-235b
Fetched from CodeSOTA API on 2026-04-20
89.15codesota-api
10monkeyocr-pro-3b
Fetched from CodeSOTA API on 2026-04-20
88.85codesota-api
11ocrverse-4b
Fetched from CodeSOTA API on 2026-04-20
88.56codesota-api
12dots-ocr-3b
Fetched from CodeSOTA API on 2026-04-20
88.41codesota-api
13gemini-25-pro
Fetched from CodeSOTA API on 2026-04-20
88.03codesota-api
14MonkeyOCR-3B
Fetched from CodeSOTA API on 2026-04-20
87.13codesota-api
15qwen25-vl
Fetched from CodeSOTA API on 2026-04-20
87.02codesota-api
16MonkeyOCR-pro-1.2B
Fetched from CodeSOTA API on 2026-04-20
86.96codesota-api
17PP-StructureV3
Fetched from CodeSOTA API on 2026-04-20
86.73codesota-api
18DeepSeek-OCR
Fetched from CodeSOTA API on 2026-04-20
86.46codesota-api
19Nanonets-OCR-s
Fetched from CodeSOTA API on 2026-04-20
85.59codesota-api
20MinerU2-VLM
Fetched from CodeSOTA API on 2026-04-20
85.56codesota-api
21Dolphin-1.5
Fetched from CodeSOTA API on 2026-04-20
85.06codesota-api
22InternVL3.5-241B
Fetched from CodeSOTA API on 2026-04-20
82.67codesota-api
23olmOCR-7B
Fetched from CodeSOTA API on 2026-04-20
81.79codesota-api
24POINTS-Reader
Fetched from CodeSOTA API on 2026-04-20
80.98codesota-api
25InternVL3-76B
Fetched from CodeSOTA API on 2026-04-20
80.33codesota-api
26mistral-ocr-3
Fetched from CodeSOTA API on 2026-04-20
79.75yescodesota-api
27mistral-ocr-2512
Fetched from CodeSOTA API on 2026-04-20
79.75yescodesota-api
28MinerU2-pipeline
Fetched from CodeSOTA API on 2026-04-20
75.51codesota-api
29GPT-4o
Fetched from CodeSOTA API on 2026-04-20
75.02codesota-api
30OCRFlux-3B
Fetched from CodeSOTA API on 2026-04-20
74.82codesota-api
31Dolphin
Fetched from CodeSOTA API on 2026-04-20
74.67codesota-api
32Marker 1.8.2
Fetched from CodeSOTA API on 2026-04-20
71.30codesota-api
33clearocr-teamquest
Fetched from CodeSOTA API on 2026-04-20
31.70yescodesota-api
Fig · 33 results on Composite score. Rows sourced from benchmarks.json; shaded row marks current SOTA.
§ 02 · Leaderboard · Table TEDS

Table TEDS, ranked.

Tree Edit Distance score on table structure recognition. (higher is better)

#ModelTable TEDSVerifiedSource
01paddleocr-vl
Fetched from CodeSOTA API on 2026-04-20
93.52codesota-api
02Qianfan-OCR
Fetched from CodeSOTA API on 2026-04-20
91.02codesota-api
03mistral-ocr-3
Fetched from CodeSOTA API on 2026-04-20
70.88yescodesota-api
04clearocr-teamquest
Fetched from CodeSOTA API on 2026-04-20
0.800yescodesota-api
Fig · 4 results on Table TEDS. Rows sourced from benchmarks.json; shaded row marks current SOTA.
§ 03 · Leaderboard · Formula CDM

Formula CDM, ranked.

Character-level match score on formula recognition. (higher is better)

#ModelFormula CDMVerifiedSource
01Qianfan-OCR
Fetched from CodeSOTA API on 2026-04-20
92.43codesota-api
Fig · 1 result on Formula CDM. Rows sourced from benchmarks.json; shaded row marks current SOTA.
§ 04 · Leaderboard · Layout mAP

Layout mAP, ranked.

Mean Average Precision on layout detection. (higher is better)

#ModelLayout mAPVerifiedSource
01mineru-2.5
Fetched from CodeSOTA API on 2026-04-20
97.50codesota-api
Fig · 1 result on Layout mAP. Rows sourced from benchmarks.json; shaded row marks current SOTA.
§ 05 · Leaderboard · Reading-order score

Reading-order score, ranked.

Agreement with ground-truth reading order of page regions. (higher is better)

#ModelReading-order scoreVerifiedSource
01mistral-ocr-3
Fetched from CodeSOTA API on 2026-04-20
91.63yescodesota-api
02clearocr-teamquest
Fetched from CodeSOTA API on 2026-04-20
86.04yescodesota-api
Fig · 2 results on Reading-order score. Rows sourced from benchmarks.json; shaded row marks current SOTA.
§ 06 · Leaderboard · Text edit distance

Text edit distance, ranked.

Normalised edit distance on full-page text extraction. (lower is better)

#ModelText edit distanceVerifiedSource
01mistral-ocr-3
Fetched from CodeSOTA API on 2026-04-20
0.099yescodesota-api
02clearocr-teamquest
Fetched from CodeSOTA API on 2026-04-20
0.154yescodesota-api
Fig · 2 results on Text edit distance. Rows sourced from benchmarks.json; shaded row marks current SOTA.
§ 07 · Leaderboard · OCR edit distance

OCR edit distance, ranked.

Character-level edit distance for raw OCR. (lower is better)

#ModelOCR edit distanceVerifiedSource
01gpt-4o
Fetched from CodeSOTA API on 2026-04-20
0.020codesota-api
Fig · 1 result on OCR edit distance. Rows sourced from benchmarks.json; shaded row marks current SOTA.
§ 08 · Leaderboard · Formula edit distance

Formula edit distance, ranked.

Edit distance on formula recognition. (lower is better)

#ModelFormula edit distanceVerifiedSource
01mistral-ocr-3
Fetched from CodeSOTA API on 2026-04-20
0.218yescodesota-api
02clearocr-teamquest
Fetched from CodeSOTA API on 2026-04-20
0.902yescodesota-api
Fig · 2 results on Formula edit distance. Rows sourced from benchmarks.json; shaded row marks current SOTA.
§ 09 · Leaderboard · Text edit (vendor variant)

Text edit (vendor variant), ranked.

Edit-distance variant reported by some vendor submissions. (lower is better)

#ModelText edit (vendor variant)VerifiedSource
01Qianfan-OCR
Fetched from CodeSOTA API on 2026-04-20
0.041codesota-api
Fig · 1 result on Text edit (vendor variant). Rows sourced from benchmarks.json; shaded row marks current SOTA.
§ What it measures

Composite = text + table + formula.

The headline OmniDocBench score is a composite defined by its authors as ((1 − TextEditDist) · 100 + TableTEDS + FormulaCDM) / 3. Each component tracks a different axis of parsing accuracy, so a model that aces tables but mangles formulas cannot win on the headline.

Secondary metrics tracked here: table-teds (tree edit distance on table recognition), formula-cdm (formula character-level match), layout-map (mean average precision on layout detection), and text-edit-distance / ocr-edit-distance (lower is better for edit distances). Every reported score is preserved verbatim from the submission.

§ Dataset details

A cross-section of real document wild-type.

OmniDocBench was released by Shanghai AI Laboratory as a comprehensive benchmark for evaluating PDF document parsing across diverse document types with multi-level annotations. It is the reference bench most open-source document parsers and vendor OCR APIs now report on.

The upstream leaderboard and test split live on alphaXiv. Per-row sources on the tables above link back to the submitting paper or vendor statement.

§ How scores are verified

Reported, then reproduced.

Every row above is imported from the canonical benchmarks.json. Open-weight models are re-executed against the OmniDocBench test split through the CodeSOTA harness; closed APIs are run through the vendor endpoint with the model version and access date recorded. Rows marked “verified” have been independently reproduced, not taken from press release.

For the full reproduction policy, see the Codesota methodology.

§ Final · Related OCR benchmarks

Cross-links, sibling leaderboards.