TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/BEiT: BERT Pre-Training of Image Transformers

BEiT: BERT Pre-Training of Image Transformers

Hangbo Bao, Li Dong, Songhao Piao, Furu Wei

2021-06-15ICLR 2022 4Document Layout AnalysisSelf-Supervised Image ClassificationImage ClassificationSemantic SegmentationDocument Image Classification
PaperPDFCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCode(official)CodeCodeCode

Abstract

We introduce a self-supervised vision representation model BEiT, which stands for Bidirectional Encoder representation from Image Transformers. Following BERT developed in the natural language processing area, we propose a masked image modeling task to pretrain vision Transformers. Specifically, each image has two views in our pre-training, i.e, image patches (such as 16x16 pixels), and visual tokens (i.e., discrete tokens). We first "tokenize" the original image into visual tokens. Then we randomly mask some image patches and fed them into the backbone Transformer. The pre-training objective is to recover the original visual tokens based on the corrupted image patches. After pre-training BEiT, we directly fine-tune the model parameters on downstream tasks by appending task layers upon the pretrained encoder. Experimental results on image classification and semantic segmentation show that our model achieves competitive results with previous pre-training methods. For example, base-size BEiT achieves 83.2% top-1 accuracy on ImageNet-1K, significantly outperforming from-scratch DeiT training (81.8%) with the same setup. Moreover, large-size BEiT obtains 86.3% only using ImageNet-1K, even outperforming ViT-L with supervised pre-training on ImageNet-22K (85.2%). The code and pretrained models are available at https://aka.ms/beit.

Results

TaskDatasetMetricValueModel
Semantic SegmentationADE20K valmIoU57BEiT-L (ViT+UperNet, ImageNet-22k pretrain)
Semantic SegmentationADE20KValidation mIoU57BEiT-L (ViT+UperNet)
Image ClassificationOmniBenchmarkAverage Top-1 Accuracy30.1BeiT
Document Layout AnalysisPubLayNet valFigure0.957BEiT-B
Document Layout AnalysisPubLayNet valList0.924BEiT-B
Document Layout AnalysisPubLayNet valOverall0.931BEiT-B
Document Layout AnalysisPubLayNet valTable0.973BEiT-B
Document Layout AnalysisPubLayNet valText0.934BEiT-B
Document Layout AnalysisPubLayNet valTitle0.866BEiT-B
10-shot image generationADE20K valmIoU57BEiT-L (ViT+UperNet, ImageNet-22k pretrain)
10-shot image generationADE20KValidation mIoU57BEiT-L (ViT+UperNet)

Related Papers

SeC: Advancing Complex Video Object Segmentation via Progressive Concept Construction2025-07-21Automatic Classification and Segmentation of Tunnel Cracks Based on Deep Learning and Visual Explanations2025-07-18Adversarial attacks to image classification systems using evolutionary algorithms2025-07-17Efficient Adaptation of Pre-trained Vision Transformer underpinned by Approximately Orthogonal Fine-Tuning Strategy2025-07-17Federated Learning for Commercial Image Sources2025-07-17MUPAX: Multidimensional Problem Agnostic eXplainable AI2025-07-17DiffOSeg: Omni Medical Image Segmentation via Multi-Expert Collaboration Diffusion Model2025-07-17SCORE: Scene Context Matters in Open-Vocabulary Remote Sensing Instance Segmentation2025-07-17