Tsinghua U & Microsoft Propose Fastformer: An Additive Attention Based Transformer With Linear Complexity | Synced

A team from Tsinghua University and Microsoft Research Asia proposes Fastformer, an efficient Transformer variant based on additive attention that achieves effective context modelling with linear c...

By · · 1 min read

Source: Synced | AI Technology & Industry Review

A team from Tsinghua University and Microsoft Research Asia proposes Fastformer, an efficient Transformer variant based on additive attention that achieves effective context modelling with linear complexity.