Sinkhorn Transformer

Natural Language ProcessingIntroduced 20001 papers

Description

The Sinkhorn Transformer is a type of transformer that uses Sparse Sinkhorn Attention as a building block. This component is a plug-in replacement for dense fully-connected attention (as well as local attention, and sparse attention alternatives), and allows for reduced memory complexity as well as sparse attention.

Papers Using This Method