ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2105.01601
  4. Cited By
MLP-Mixer: An all-MLP Architecture for Vision

MLP-Mixer: An all-MLP Architecture for Vision

4 May 2021
Ilya O. Tolstikhin
N. Houlsby
Alexander Kolesnikov
Lucas Beyer
Xiaohua Zhai
Thomas Unterthiner
Jessica Yung
Andreas Steiner
Daniel Keysers
Jakob Uszkoreit
Mario Lucic
Alexey Dosovitskiy
ArXivPDFHTML

Papers citing "MLP-Mixer: An all-MLP Architecture for Vision"

50 / 1,119 papers shown
Title
Global Filter Networks for Image Classification
Global Filter Networks for Image Classification
Yongming Rao
Wenliang Zhao
Zheng Zhu
Jiwen Lu
Jie Zhou
ViT
14
450
0
01 Jul 2021
ResViT: Residual vision transformers for multi-modal medical image
  synthesis
ResViT: Residual vision transformers for multi-modal medical image synthesis
Onat Dalmaz
Mahmut Yurt
Tolga Çukur
ViT
MedIm
32
338
0
30 Jun 2021
Multi-Exit Vision Transformer for Dynamic Inference
Multi-Exit Vision Transformer for Dynamic Inference
Arian Bakhtiarnia
Qi Zhang
Alexandros Iosifidis
28
26
0
29 Jun 2021
Rethinking Token-Mixing MLP for MLP-based Vision Backbone
Rethinking Token-Mixing MLP for MLP-based Vision Backbone
Tan Yu
Xu Li
Yunfeng Cai
Mingming Sun
Ping Li
40
26
0
28 Jun 2021
Exploring Corruption Robustness: Inductive Biases in Vision Transformers
  and MLP-Mixers
Exploring Corruption Robustness: Inductive Biases in Vision Transformers and MLP-Mixers
Katelyn Morrison
B. Gilby
Colton Lipchak
Adam Mattioli
Adriana Kovashka
ViT
20
17
0
24 Jun 2021
Mix and Mask Actor-Critic Methods
Mix and Mask Actor-Critic Methods
Dom Huh
11
1
0
24 Jun 2021
Co-advise: Cross Inductive Bias Distillation
Co-advise: Cross Inductive Bias Distillation
Sucheng Ren
Zhengqi Gao
Tianyu Hua
Zihui Xue
Yonglong Tian
Shengfeng He
Hang Zhao
42
53
0
23 Jun 2021
Vision Permutator: A Permutable MLP-Like Architecture for Visual
  Recognition
Vision Permutator: A Permutable MLP-Like Architecture for Visual Recognition
Qibin Hou
Zihang Jiang
Li-xin Yuan
Mingg-Ming Cheng
Shuicheng Yan
Jiashi Feng
ViT
MLLM
24
205
0
23 Jun 2021
Towards Biologically Plausible Convolutional Networks
Towards Biologically Plausible Convolutional Networks
Roman Pogodin
Yash Mehta
Timothy Lillicrap
P. Latham
26
22
0
22 Jun 2021
TokenLearner: What Can 8 Learned Tokens Do for Images and Videos?
TokenLearner: What Can 8 Learned Tokens Do for Images and Videos?
Michael S. Ryoo
A. Piergiovanni
Anurag Arnab
Mostafa Dehghani
A. Angelova
ViT
21
127
0
21 Jun 2021
How to train your ViT? Data, Augmentation, and Regularization in Vision
  Transformers
How to train your ViT? Data, Augmentation, and Regularization in Vision Transformers
Andreas Steiner
Alexander Kolesnikov
Xiaohua Zhai
Ross Wightman
Jakob Uszkoreit
Lucas Beyer
ViT
34
613
0
18 Jun 2021
Efficient Self-supervised Vision Transformers for Representation
  Learning
Efficient Self-supervised Vision Transformers for Representation Learning
Chunyuan Li
Jianwei Yang
Pengchuan Zhang
Mei Gao
Bin Xiao
Xiyang Dai
Lu Yuan
Jianfeng Gao
ViT
30
209
0
17 Jun 2021
XCiT: Cross-Covariance Image Transformers
XCiT: Cross-Covariance Image Transformers
Alaaeldin El-Nouby
Hugo Touvron
Mathilde Caron
Piotr Bojanowski
Matthijs Douze
...
Ivan Laptev
Natalia Neverova
Gabriel Synnaeve
Jakob Verbeek
Hervé Jégou
ViT
23
497
0
17 Jun 2021
Revisiting the Calibration of Modern Neural Networks
Revisiting the Calibration of Modern Neural Networks
Matthias Minderer
Josip Djolonga
Rob Romijnders
F. Hubis
Xiaohua Zhai
N. Houlsby
Dustin Tran
Mario Lucic
UQCV
22
356
0
15 Jun 2021
MLP Singer: Towards Rapid Parallel Korean Singing Voice Synthesis
MLP Singer: Towards Rapid Parallel Korean Singing Voice Synthesis
Jaesung Tae
Hyeongju Kim
Younggun Lee
6
14
0
15 Jun 2021
Improved Transformer for High-Resolution GANs
Improved Transformer for High-Resolution GANs
Long Zhao
Zizhao Zhang
Ting Chen
Dimitris N. Metaxas
Han Zhang
ViT
21
95
0
14 Jun 2021
S$^2$-MLP: Spatial-Shift MLP Architecture for Vision
S2^22-MLP: Spatial-Shift MLP Architecture for Vision
Tan Yu
Xu Li
Yunfeng Cai
Mingming Sun
Ping Li
39
186
0
14 Jun 2021
Styleformer: Transformer based Generative Adversarial Networks with
  Style Vector
Styleformer: Transformer based Generative Adversarial Networks with Style Vector
Jeeseung Park
Younggeun Kim
ViT
21
48
0
13 Jun 2021
NeRF in detail: Learning to sample for view synthesis
NeRF in detail: Learning to sample for view synthesis
Relja Arandjelović
Andrew Zisserman
27
41
0
09 Jun 2021
Knowledge distillation: A good teacher is patient and consistent
Knowledge distillation: A good teacher is patient and consistent
Lucas Beyer
Xiaohua Zhai
Amelie Royer
L. Markeeva
Rohan Anil
Alexander Kolesnikov
VLM
18
287
0
09 Jun 2021
On the Connection between Local Attention and Dynamic Depth-wise
  Convolution
On the Connection between Local Attention and Dynamic Depth-wise Convolution
Qi Han
Zejia Fan
Qi Dai
Lei-huan Sun
Ming-Ming Cheng
Jiaying Liu
Jingdong Wang
ViT
16
105
0
08 Jun 2021
On Improving Adversarial Transferability of Vision Transformers
On Improving Adversarial Transferability of Vision Transformers
Muzammal Naseer
Kanchana Ranasinghe
Salman Khan
F. Khan
Fatih Porikli
ViT
21
93
0
08 Jun 2021
A Lightweight and Gradient-Stable Neural Layer
A Lightweight and Gradient-Stable Neural Layer
Yueyao Yu
Yin Zhang
21
0
0
08 Jun 2021
Graph-MLP: Node Classification without Message Passing in Graph
Graph-MLP: Node Classification without Message Passing in Graph
Yang Hu
Haoxuan You
Zhecan Wang
Zhicheng Wang
Erjin Zhou
Yue Gao
14
107
0
08 Jun 2021
ViTAE: Vision Transformer Advanced by Exploring Intrinsic Inductive Bias
ViTAE: Vision Transformer Advanced by Exploring Intrinsic Inductive Bias
Yufei Xu
Qiming Zhang
Jing Zhang
Dacheng Tao
ViT
48
329
0
07 Jun 2021
Vision Transformers with Hierarchical Attention
Vision Transformers with Hierarchical Attention
Yun-Hai Liu
Yu-Huan Wu
Guolei Sun
Le Zhang
Ajad Chhatkuli
Luc Van Gool
ViT
30
32
0
06 Jun 2021
Exploring the Limits of Out-of-Distribution Detection
Exploring the Limits of Out-of-Distribution Detection
Stanislav Fort
Jie Jessie Ren
Balaji Lakshminarayanan
16
323
0
06 Jun 2021
When Vision Transformers Outperform ResNets without Pre-training or
  Strong Data Augmentations
When Vision Transformers Outperform ResNets without Pre-training or Strong Data Augmentations
Xiangning Chen
Cho-Jui Hsieh
Boqing Gong
ViT
13
320
0
03 Jun 2021
Container: Context Aggregation Network
Container: Context Aggregation Network
Peng Gao
Jiasen Lu
Hongsheng Li
Roozbeh Mottaghi
Aniruddha Kembhavi
ViT
11
69
0
02 Jun 2021
Can Attention Enable MLPs To Catch Up With CNNs?
Can Attention Enable MLPs To Catch Up With CNNs?
Meng-Hao Guo
Zheng-Ning Liu
Tai-Jiang Mu
Dun Liang
Ralph Robert Martin
Shimin Hu
AAML
16
17
0
31 May 2021
A remark on a paper of Krotov and Hopfield [arXiv:2008.06996]
A remark on a paper of Krotov and Hopfield [arXiv:2008.06996]
Fei Tang
Michael K Kopp
11
11
0
31 May 2021
Choose a Transformer: Fourier or Galerkin
Choose a Transformer: Fourier or Galerkin
Shuhao Cao
34
220
0
31 May 2021
MixerGAN: An MLP-Based Architecture for Unpaired Image-to-Image
  Translation
MixerGAN: An MLP-Based Architecture for Unpaired Image-to-Image Translation
George Cazenavette
Manuel Ladron de Guevara
14
17
0
28 May 2021
An Attention Free Transformer
An Attention Free Transformer
Shuangfei Zhai
Walter A. Talbott
Nitish Srivastava
Chen Huang
Hanlin Goh
Ruixiang Zhang
J. Susskind
ViT
19
127
0
28 May 2021
On the Bias Against Inductive Biases
On the Bias Against Inductive Biases
George Cazenavette
Simon Lucey
SSL
18
1
0
28 May 2021
Pay Attention to MLPs
Pay Attention to MLPs
Hanxiao Liu
Zihang Dai
David R. So
Quoc V. Le
AI4CE
37
651
0
17 May 2021
Brain Inspired Face Recognition: A Computational Framework
Brain Inspired Face Recognition: A Computational Framework
P. Chowdhury
Angad Wadhwa
Nikhil Tyagi
CVBM
15
4
0
15 May 2021
FNet: Mixing Tokens with Fourier Transforms
FNet: Mixing Tokens with Fourier Transforms
James Lee-Thorp
Joshua Ainslie
Ilya Eckstein
Santiago Ontanon
13
516
0
09 May 2021
ResMLP: Feedforward networks for image classification with
  data-efficient training
ResMLP: Feedforward networks for image classification with data-efficient training
Hugo Touvron
Piotr Bojanowski
Mathilde Caron
Matthieu Cord
Alaaeldin El-Nouby
...
Gautier Izacard
Armand Joulin
Gabriel Synnaeve
Jakob Verbeek
Hervé Jégou
VLM
16
655
0
07 May 2021
RepMLP: Re-parameterizing Convolutions into Fully-connected Layers for
  Image Recognition
RepMLP: Re-parameterizing Convolutions into Fully-connected Layers for Image Recognition
Xiaohan Ding
Chunlong Xia
X. Zhang
Xiaojie Chu
Jungong Han
Guiguang Ding
15
92
0
05 May 2021
Sifting out the features by pruning: Are convolutional networks the
  winning lottery ticket of fully connected ones?
Sifting out the features by pruning: Are convolutional networks the winning lottery ticket of fully connected ones?
Franco Pellegrini
Giulio Biroli
44
6
0
27 Apr 2021
ImageNet-21K Pretraining for the Masses
ImageNet-21K Pretraining for the Masses
T. Ridnik
Emanuel Ben-Baruch
Asaf Noy
Lihi Zelnik-Manor
SSeg
VLM
CLIP
173
686
0
22 Apr 2021
All Tokens Matter: Token Labeling for Training Better Vision
  Transformers
All Tokens Matter: Token Labeling for Training Better Vision Transformers
Zihang Jiang
Qibin Hou
Li-xin Yuan
Daquan Zhou
Yujun Shi
Xiaojie Jin
Anran Wang
Jiashi Feng
ViT
12
203
0
22 Apr 2021
Cloth Interactive Transformer for Virtual Try-On
Cloth Interactive Transformer for Virtual Try-On
Bin Ren
Hao Tang
Fanyang Meng
Runwei Ding
Philip H. S. Torr
N. Sebe
ViT
30
32
0
12 Apr 2021
On the Adversarial Robustness of Vision Transformers
On the Adversarial Robustness of Vision Transformers
Rulin Shao
Zhouxing Shi
Jinfeng Yi
Pin-Yu Chen
Cho-Jui Hsieh
ViT
25
137
0
29 Mar 2021
A Practical Survey on Faster and Lighter Transformers
A Practical Survey on Faster and Lighter Transformers
Quentin Fournier
G. Caron
Daniel Aloise
6
93
0
26 Mar 2021
Swin Transformer: Hierarchical Vision Transformer using Shifted Windows
Swin Transformer: Hierarchical Vision Transformer using Shifted Windows
Ze Liu
Yutong Lin
Yue Cao
Han Hu
Yixuan Wei
Zheng-Wei Zhang
Stephen Lin
B. Guo
ViT
113
20,677
0
25 Mar 2021
ConViT: Improving Vision Transformers with Soft Convolutional Inductive
  Biases
ConViT: Improving Vision Transformers with Soft Convolutional Inductive Biases
Stéphane dÁscoli
Hugo Touvron
Matthew L. Leavitt
Ari S. Morcos
Giulio Biroli
Levent Sagun
ViT
29
803
0
19 Mar 2021
Understanding Invariance via Feedforward Inversion of Discriminatively
  Trained Classifiers
Understanding Invariance via Feedforward Inversion of Discriminatively Trained Classifiers
Piotr Teterwak
Chiyuan Zhang
Dilip Krishnan
Michael C. Mozer
23
9
0
15 Mar 2021
Pyramid Vision Transformer: A Versatile Backbone for Dense Prediction
  without Convolutions
Pyramid Vision Transformer: A Versatile Backbone for Dense Prediction without Convolutions
Wenhai Wang
Enze Xie
Xiang Li
Deng-Ping Fan
Kaitao Song
Ding Liang
Tong Lu
Ping Luo
Ling Shao
ViT
274
3,622
0
24 Feb 2021
Previous
123...212223
Next