TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Methods/Attention-augmented Convolution

Attention-augmented Convolution

GeneralIntroduced 20003 papers
Source Paper

Description

Attention-augmented Convolution is a type of convolution with a two-dimensional relative self-attention mechanism that can replace convolutions as a stand-alone computational primitive for image classification. It employs scaled-dot product attention and multi-head attention as with Transformers.

It works by concatenating convolutional and attentional feature map. To see this, consider an original convolution operator with kernel size kkk, F_inF\_{in}F_in input filters and F_outF\_{out}F_out output filters. The corresponding attention augmented convolution can be written as"

AAConv(X)=Concat[Conv(X),MHA(X)]\text{AAConv}\left(X\right) = \text{Concat}\left[\text{Conv}(X), \text{MHA}(X)\right] AAConv(X)=Concat[Conv(X),MHA(X)]

XXX originates from an input tensor of shape (H,W,F_in)\left(H, W, F\_{in}\right)(H,W,F_in). This is flattened to become X∈RHW×F_inX \in \mathbb{R}^{HW \times F\_{in}}X∈RHW×F_in which is passed into a multi-head attention module, as well as a convolution (see above).

Similarly to the convolution, the attention augmented convolution 1) is equivariant to translation and 2) can readily operate on inputs of different spatial dimensions.

Papers Using This Method

RISurConv: Rotation Invariant Surface Attention-Augmented Convolutions for 3D Point Cloud Classification and Segmentation2024-08-12MuSLCAT: Multi-Scale Multi-Level Convolutional Attention Transformer for Discriminative Music Modeling on Raw Waveforms2021-04-06Attention Augmented Convolutional Networks2019-04-22