CIFAR-100

Unknown

60K 32x32 color images in 100 fine-grained classes grouped into 20 superclasses. More challenging than CIFAR-10.

Benchmark Stats

Models13
Papers13
Metrics1

SOTA History

accuracy

accuracy

Higher is better

RankModelSourceScoreYearPaper
1EVA-02-L

EVA-02-L/14+ fine-tuned on CIFAR-100. Pre-trained with EVA-CLIP on Objects365 + ImageNet-21K. State-of-the-art as of 2023. arXiv Mar 2023.

Editorial97.152026Source
2CoAtNet-7

CoAtNet-7 (2.4B params) fine-tuned on CIFAR-100. Pre-trained on ImageNet-21K. arXiv Jun 2021, NeurIPS 2021.

Editorial96.382026Source
3ConvNeXt V2-H

ConvNeXt V2-H fine-tuned on CIFAR-100 after FCMAE pre-training on ImageNet-22K. arXiv Jan 2023, CVPR 2023.

Editorial96.172026Source
4MAE ViT-H/14

ViT-H/14 fine-tuned on CIFAR-100 after MAE pre-training on ImageNet-1K. arXiv Nov 2021, CVPR 2022.

Editorial96.082026Source
5SwinV2-G

SwinV2-G (3B params) fine-tuned on CIFAR-100. Pre-trained on ImageNet-21K with resolution 192^2. arXiv Nov 2021, CVPR 2022.

Editorial96.012026Source
6DeiT III-H/14

DeiT III ViT-H/14 fine-tuned on CIFAR-100. Improved training recipe for ViTs. arXiv Apr 2022, ECCV 2022.

Editorial95.942026Source
7InternImage-XL

InternImage-XL fine-tuned on CIFAR-100. Uses deformable convolutions as core operator. arXiv Nov 2022, CVPR 2023.

Editorial95.772026Source
8FasterViT-6

FasterViT-6 fine-tuned on CIFAR-100. Hierarchical ViT with carrier tokens for high-resolution efficiency. arXiv Jun 2023.

Editorial95.722026Source
9vit-h-14

Fine-tuned from ImageNet pretraining.

Editorial94.552025Source
10ViT-L/16 (IN-21K)

Vision Transformer ViT-L/16, pretrained on ImageNet-21K and finetuned on CIFAR-100. 93.25% reported in ViT paper (Table 5). Paper: Dosovitskiy et al. 2021, arxiv:2010.11929.

Community93.252026Source
11efficientnet-b7

Transfer learning from ImageNet.

Editorial91.72025Source
12vit-b-16

Fine-tuned from ImageNet-21K.

Editorial91.482025Source
13resnet-50

With Cutout augmentation.

Editorial78.042025Source

Submit a Result