[1] MLP-Mixer: An all-MLP Architecture for Vision - Google Research
[2] Beyond Self-attention: External Attention using Two Linear Layers for Visual Tasks - 清华大学
[3] Do You Even Need Attention? A Stack of Feed-Forward Layers Does Surprisingly Well on ImageNet - 牛津大学
[4] Are Pre-trained Convolutions Better than Pre-trained Transformers? - Google Research
[5] ResMLP: Feedforward networks for image classification with data-efficient training - Facebook AI
[6] FNet: Mixing Tokens with Fourier Transforms - Google Research
[7] Pay Attention to MLPs - Google Research
对于以上论文的浅谈:也来盘点一些最近的非Transformer工作
MLP替代Transformer浅谈
最新推荐文章于 2024-05-02 22:50:44 发布