TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/VOLO: Vision Outlooker for Visual Recognition

VOLO: Vision Outlooker for Visual Recognition

Li Yuan, Qibin Hou, Zihang Jiang, Jiashi Feng, Shuicheng Yan

2021-06-24Image ClassificationDomain GeneralizationSemantic Segmentation
PaperPDFCode(official)CodeCodeCodeCode(official)CodeCode

Abstract

Visual recognition has been dominated by convolutional neural networks (CNNs) for years. Though recently the prevailing vision transformers (ViTs) have shown great potential of self-attention based models in ImageNet classification, their performance is still inferior to that of the latest SOTA CNNs if no extra data are provided. In this work, we try to close the performance gap and demonstrate that attention-based models are indeed able to outperform CNNs. We find a major factor limiting the performance of ViTs for ImageNet classification is their low efficacy in encoding fine-level features into the token representations. To resolve this, we introduce a novel outlook attention and present a simple and general architecture, termed Vision Outlooker (VOLO). Unlike self-attention that focuses on global dependency modeling at a coarse level, the outlook attention efficiently encodes finer-level features and contexts into tokens, which is shown to be critically beneficial to recognition performance but largely ignored by the self-attention. Experiments show that our VOLO achieves 87.1% top-1 accuracy on ImageNet-1K classification, which is the first model exceeding 87% accuracy on this competitive benchmark, without using any extra training data In addition, the pre-trained VOLO transfers well to downstream tasks, such as semantic segmentation. We achieve 84.3% mIoU score on the cityscapes validation set and 54.3% on the ADE20K validation set. Code is available at \url{https://github.com/sail-sg/volo}.

Results

TaskDatasetMetricValueModel
Domain AdaptationVizWiz-ClassificationAccuracy - All Images57.2VOLO-D5
Domain AdaptationVizWiz-ClassificationAccuracy - Clean Images59.7VOLO-D5
Domain AdaptationVizWiz-ClassificationAccuracy - Corrupted Images51.8VOLO-D5
Semantic SegmentationGraz-02Pixel Accuracy85VOLO-D5
Semantic SegmentationCityscapes valmIoU84.3VOLO-D4 (MS, ImageNet1k pretrain)
Semantic SegmentationADE20KValidation mIoU54.3VOLO-D5
Image ClassificationImageNet V2Top 1 Accuracy78VOLO-D5
Image ClassificationImageNet V2Top 1 Accuracy77.8VOLO-D4
Image ClassificationVizWiz-ClassificationAccuracy57.2VOLO-D5
Image ClassificationImageNetGFLOPs412VOLO-D5
Image ClassificationImageNetGFLOPs197VOLO-D4
Image ClassificationImageNetGFLOPs67.9VOLO-D3
Domain GeneralizationVizWiz-ClassificationAccuracy - All Images57.2VOLO-D5
Domain GeneralizationVizWiz-ClassificationAccuracy - Clean Images59.7VOLO-D5
Domain GeneralizationVizWiz-ClassificationAccuracy - Corrupted Images51.8VOLO-D5
10-shot image generationGraz-02Pixel Accuracy85VOLO-D5
10-shot image generationCityscapes valmIoU84.3VOLO-D4 (MS, ImageNet1k pretrain)
10-shot image generationADE20KValidation mIoU54.3VOLO-D5

Related Papers

SeC: Advancing Complex Video Object Segmentation via Progressive Concept Construction2025-07-21Automatic Classification and Segmentation of Tunnel Cracks Based on Deep Learning and Visual Explanations2025-07-18Adversarial attacks to image classification systems using evolutionary algorithms2025-07-17Efficient Adaptation of Pre-trained Vision Transformer underpinned by Approximately Orthogonal Fine-Tuning Strategy2025-07-17Federated Learning for Commercial Image Sources2025-07-17MUPAX: Multidimensional Problem Agnostic eXplainable AI2025-07-17Simulate, Refocus and Ensemble: An Attention-Refocusing Scheme for Domain Generalization2025-07-17GLAD: Generalizable Tuning for Vision-Language Models2025-07-17