Codesota · OCR · Benchmark · ImageNet-1k22 scored runs · 22 distinct modelsUpdated 2026-04-20
§ 00 · Opening

ImageNet-1k, still the deciding test.

ImageNet-1k remains the reference for large-scale image classification. It ranks vision backbones on Top-1 accuracy across 1,000 categories — the same score every new architecture has to beat before it can earn its place on a datasheet.

§ 01 · Leaderboard · Top-1 accuracy

Top-1 accuracy, ranked.

Share of ImageNet-1k validation images whose top prediction matches the gold label. (higher is better)

#ModelTop-1 accuracyVerifiedSource
01coca-finetuned
Fetched from CodeSOTA API on 2026-04-20
91.00codesota-api
02vit-g-14
Fetched from CodeSOTA API on 2026-04-20
90.45codesota-api
03EVA-02-L
Fetched from CodeSOTA API on 2026-04-20
90.06yescodesota-api
04EVA-Giant
Fetched from CodeSOTA API on 2026-04-20
89.79yescodesota-api
05InternImage-H
Fetched from CodeSOTA API on 2026-04-20
89.60yescodesota-api
06SigLIP-SO400M
Fetched from CodeSOTA API on 2026-04-20
89.41yescodesota-api
07convnext-v2-huge
Fetched from CodeSOTA API on 2026-04-20
88.90codesota-api
08ViT-H/14 CLIP (LAION-2B)
Fetched from CodeSOTA API on 2026-04-20
88.63yescodesota-api
09ConvNeXt-XXLarge (CLIP LAION)
Fetched from CodeSOTA API on 2026-04-20
88.62yescodesota-api
10vit-h-14
Fetched from CodeSOTA API on 2026-04-20
88.55codesota-api
11swin-large
Fetched from CodeSOTA API on 2026-04-20
87.30codesota-api
12efficientnet-v2-l
Fetched from CodeSOTA API on 2026-04-20
85.70codesota-api
13deit-b-distilled
Fetched from CodeSOTA API on 2026-04-20
85.20codesota-api
14efficientnet-b7
Fetched from CodeSOTA API on 2026-04-20
84.40codesota-api
15deit-b
Fetched from CodeSOTA API on 2026-04-20
83.10codesota-api
16convnext-v2-tiny
Fetched from CodeSOTA API on 2026-04-20
83.00codesota-api
17vit-l-16
Fetched from CodeSOTA API on 2026-04-20
82.70codesota-api
18vit-b-16
Fetched from CodeSOTA API on 2026-04-20
81.20codesota-api
19resnet-50-a3
Fetched from CodeSOTA API on 2026-04-20
80.40codesota-api
20resnet-152
Fetched from CodeSOTA API on 2026-04-20
78.60codesota-api
21efficientnet-b0
Fetched from CodeSOTA API on 2026-04-20
77.10codesota-api
22resnet-50
Fetched from CodeSOTA API on 2026-04-20
76.15codesota-api
Fig · 22 results on Top-1 accuracy. Rows sourced from benchmarks.json; shaded row marks current SOTA.
§ What it measures

Top-1 accuracy, 1,000-way.

Top-1 accuracy is the share of test images whose argmax prediction matches the single gold label, out of 1,000 possible classes. It is a strict metric — Top-5 (any of the top five predictions being correct) has been essentially solved since 2018; Top-1 is where the last few points of progress still live.

A meaningful share of the ImageNet error rate at this point comes from label noise — images where the “correct” label is genuinely ambiguous. That is why the gap between 88% and 91% is a harder climb than it looks.

§ Dataset details

1.28M training images, 50K validation.

ImageNet-1k is the ILSVRC 2012 classification subset: 1.28 million training images, 50,000 validation images and 100,000 test images spread across 1,000 categories. Since the test labels are held private by the organisers, the community has standardised on validation Top-1 as the public leaderboard number.

Most frontier models score on validation, not test. When an entry reports a number computed against an external dataset, we surface that in the notes column.

§ How scores are verified

Reported, then reproduced.

Each row above is imported verbatim from benchmarks.json. Where the reporting paper gives multiple settings (different crop sizes, different pretraining regimens), the row reflects the single headline number the authors highlight in their abstract or table.

Rows marked verified have been matched against an independent reproduction. See the Codesota methodology for the policy.

§ Final · Related OCR benchmarks

Cross-links, sibling leaderboards.