Tsinghua U & Microsoft Propose Fastformer: An Additive Attention Based Transformer With Linear Complexity | Synced
A team from Tsinghua University and Microsoft Research Asia proposes Fastformer, an efficient Transformer variant based on additive attention that achieves effective context modelling with linear c...
Source: Synced | AI Technology & Industry Review
A team from Tsinghua University and Microsoft Research Asia proposes Fastformer, an efficient Transformer variant based on additive attention that achieves effective context modelling with linear complexity.