Fastformer

Natural Language ProcessingIntroduced 20004 papers

Description

Fastformer is an type of Transformer which uses additive attention as a building block. Instead of modeling the pair-wise interactions between tokens, additive attention is used to model global contexts, and then each token representation is further transformed based on its interaction with global context representations.

Papers Using This Method