TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Sequencer: Deep LSTM for Image Classification

Sequencer: Deep LSTM for Image Classification

Yuki Tatsunami, Masato Taki

2022-05-04Image ClassificationDomain Generalization
PaperPDFCode(official)CodeCodeCodeCode(official)

Abstract

In recent computer vision research, the advent of the Vision Transformer (ViT) has rapidly revolutionized various architectural design efforts: ViT achieved state-of-the-art image classification performance using self-attention found in natural language processing, and MLP-Mixer achieved competitive performance using simple multi-layer perceptrons. In contrast, several studies have also suggested that carefully redesigned convolutional neural networks (CNNs) can achieve advanced performance comparable to ViT without resorting to these new ideas. Against this background, there is growing interest in what inductive bias is suitable for computer vision. Here we propose Sequencer, a novel and competitive architecture alternative to ViT that provides a new perspective on these issues. Unlike ViTs, Sequencer models long-range dependencies using LSTMs rather than self-attention layers. We also propose a two-dimensional version of Sequencer module, where an LSTM is decomposed into vertical and horizontal LSTMs to enhance performance. Despite its simplicity, several experiments demonstrate that Sequencer performs impressively well: Sequencer2D-L, with 54M parameters, realizes 84.6% top-1 accuracy on only ImageNet-1K. Not only that, we show that it has good transferability and the robust resolution adaptability on double resolution-band.

Results

TaskDatasetMetricValueModel
Domain AdaptationImageNet-RTop-1 Error Rate51.9Sequencer2D-L
Domain AdaptationImageNet-ATop-1 accuracy %35.5Sequencer2D-L
Domain AdaptationImageNet-Cmean Corruption Error (mCE)48.9Sequencer2D-L
Domain AdaptationImageNet-SketchTop-1 accuracy35.8Sequencer2D-L
Image ClassificationImageNet V2Top 1 Accuracy73.4Sequencer2D-L
Image ClassificationImageNet ReaLAccuracy87.9Sequencer2D-L
Image ClassificationImageNetGFLOPs50.7Sequencer2D-L↑392
Image ClassificationImageNetGFLOPs16.6Sequencer2D-L
Image ClassificationImageNetGFLOPs11.1Sequencer2D-M
Image ClassificationImageNetGFLOPs8.4Sequencer2D-S
Domain GeneralizationImageNet-RTop-1 Error Rate51.9Sequencer2D-L
Domain GeneralizationImageNet-ATop-1 accuracy %35.5Sequencer2D-L
Domain GeneralizationImageNet-Cmean Corruption Error (mCE)48.9Sequencer2D-L
Domain GeneralizationImageNet-SketchTop-1 accuracy35.8Sequencer2D-L

Related Papers

Automatic Classification and Segmentation of Tunnel Cracks Based on Deep Learning and Visual Explanations2025-07-18Adversarial attacks to image classification systems using evolutionary algorithms2025-07-17Efficient Adaptation of Pre-trained Vision Transformer underpinned by Approximately Orthogonal Fine-Tuning Strategy2025-07-17Federated Learning for Commercial Image Sources2025-07-17MUPAX: Multidimensional Problem Agnostic eXplainable AI2025-07-17Simulate, Refocus and Ensemble: An Attention-Refocusing Scheme for Domain Generalization2025-07-17GLAD: Generalizable Tuning for Vision-Language Models2025-07-17MoTM: Towards a Foundation Model for Time Series Imputation based on Continuous Modeling2025-07-17