GPSA

Gated Positional Self-Attention

GeneralIntroduced 20003 papers

Description

Gated Positional Self-Attention (GPSA) is a self-attention module for vision transformers, used in the ConViT architecture, that can be initialized as a convolutional layer -- helping a ViT learn inductive biases about locality.

Papers Using This Method