https://arxiv.org/abs/2105.02723
Do You Even Need Attention? A Stack of Feed-Forward Layers Does Surprisingly Well on ImageNet (Luke Melas-Kyriazi)
luke melas의 mixer와 거의 완전히 똑같은 mlp 기반 image classification. 재밌네요.
#mlp #backbone
https://arxiv.org/abs/2105.02723
Do You Even Need Attention? A Stack of Feed-Forward Layers Does Surprisingly Well on ImageNet (Luke Melas-Kyriazi)
luke melas의 mixer와 거의 완전히 똑같은 mlp 기반 image classification. 재밌네요.
#mlp #backbone