Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2105.01601
Cited By
MLP-Mixer: An all-MLP Architecture for Vision
4 May 2021
Ilya O. Tolstikhin
N. Houlsby
Alexander Kolesnikov
Lucas Beyer
Xiaohua Zhai
Thomas Unterthiner
Jessica Yung
Andreas Steiner
Daniel Keysers
Jakob Uszkoreit
Mario Lucic
Alexey Dosovitskiy
Re-assign community
ArXiv
PDF
HTML
Papers citing
"MLP-Mixer: An all-MLP Architecture for Vision"
50 / 1,119 papers shown
Title
Global Filter Networks for Image Classification
Yongming Rao
Wenliang Zhao
Zheng Zhu
Jiwen Lu
Jie Zhou
ViT
14
450
0
01 Jul 2021
ResViT: Residual vision transformers for multi-modal medical image synthesis
Onat Dalmaz
Mahmut Yurt
Tolga Çukur
ViT
MedIm
32
338
0
30 Jun 2021
Multi-Exit Vision Transformer for Dynamic Inference
Arian Bakhtiarnia
Qi Zhang
Alexandros Iosifidis
28
26
0
29 Jun 2021
Rethinking Token-Mixing MLP for MLP-based Vision Backbone
Tan Yu
Xu Li
Yunfeng Cai
Mingming Sun
Ping Li
40
26
0
28 Jun 2021
Exploring Corruption Robustness: Inductive Biases in Vision Transformers and MLP-Mixers
Katelyn Morrison
B. Gilby
Colton Lipchak
Adam Mattioli
Adriana Kovashka
ViT
20
17
0
24 Jun 2021
Mix and Mask Actor-Critic Methods
Dom Huh
11
1
0
24 Jun 2021
Co-advise: Cross Inductive Bias Distillation
Sucheng Ren
Zhengqi Gao
Tianyu Hua
Zihui Xue
Yonglong Tian
Shengfeng He
Hang Zhao
42
53
0
23 Jun 2021
Vision Permutator: A Permutable MLP-Like Architecture for Visual Recognition
Qibin Hou
Zihang Jiang
Li-xin Yuan
Mingg-Ming Cheng
Shuicheng Yan
Jiashi Feng
ViT
MLLM
24
205
0
23 Jun 2021
Towards Biologically Plausible Convolutional Networks
Roman Pogodin
Yash Mehta
Timothy Lillicrap
P. Latham
26
22
0
22 Jun 2021
TokenLearner: What Can 8 Learned Tokens Do for Images and Videos?
Michael S. Ryoo
A. Piergiovanni
Anurag Arnab
Mostafa Dehghani
A. Angelova
ViT
21
127
0
21 Jun 2021
How to train your ViT? Data, Augmentation, and Regularization in Vision Transformers
Andreas Steiner
Alexander Kolesnikov
Xiaohua Zhai
Ross Wightman
Jakob Uszkoreit
Lucas Beyer
ViT
34
613
0
18 Jun 2021
Efficient Self-supervised Vision Transformers for Representation Learning
Chunyuan Li
Jianwei Yang
Pengchuan Zhang
Mei Gao
Bin Xiao
Xiyang Dai
Lu Yuan
Jianfeng Gao
ViT
30
209
0
17 Jun 2021
XCiT: Cross-Covariance Image Transformers
Alaaeldin El-Nouby
Hugo Touvron
Mathilde Caron
Piotr Bojanowski
Matthijs Douze
...
Ivan Laptev
Natalia Neverova
Gabriel Synnaeve
Jakob Verbeek
Hervé Jégou
ViT
23
497
0
17 Jun 2021
Revisiting the Calibration of Modern Neural Networks
Matthias Minderer
Josip Djolonga
Rob Romijnders
F. Hubis
Xiaohua Zhai
N. Houlsby
Dustin Tran
Mario Lucic
UQCV
22
356
0
15 Jun 2021
MLP Singer: Towards Rapid Parallel Korean Singing Voice Synthesis
Jaesung Tae
Hyeongju Kim
Younggun Lee
6
14
0
15 Jun 2021
Improved Transformer for High-Resolution GANs
Long Zhao
Zizhao Zhang
Ting Chen
Dimitris N. Metaxas
Han Zhang
ViT
21
95
0
14 Jun 2021
S
2
^2
2
-MLP: Spatial-Shift MLP Architecture for Vision
Tan Yu
Xu Li
Yunfeng Cai
Mingming Sun
Ping Li
39
186
0
14 Jun 2021
Styleformer: Transformer based Generative Adversarial Networks with Style Vector
Jeeseung Park
Younggeun Kim
ViT
21
48
0
13 Jun 2021
NeRF in detail: Learning to sample for view synthesis
Relja Arandjelović
Andrew Zisserman
27
41
0
09 Jun 2021
Knowledge distillation: A good teacher is patient and consistent
Lucas Beyer
Xiaohua Zhai
Amelie Royer
L. Markeeva
Rohan Anil
Alexander Kolesnikov
VLM
18
287
0
09 Jun 2021
On the Connection between Local Attention and Dynamic Depth-wise Convolution
Qi Han
Zejia Fan
Qi Dai
Lei-huan Sun
Ming-Ming Cheng
Jiaying Liu
Jingdong Wang
ViT
16
105
0
08 Jun 2021
On Improving Adversarial Transferability of Vision Transformers
Muzammal Naseer
Kanchana Ranasinghe
Salman Khan
F. Khan
Fatih Porikli
ViT
21
93
0
08 Jun 2021
A Lightweight and Gradient-Stable Neural Layer
Yueyao Yu
Yin Zhang
21
0
0
08 Jun 2021
Graph-MLP: Node Classification without Message Passing in Graph
Yang Hu
Haoxuan You
Zhecan Wang
Zhicheng Wang
Erjin Zhou
Yue Gao
14
107
0
08 Jun 2021
ViTAE: Vision Transformer Advanced by Exploring Intrinsic Inductive Bias
Yufei Xu
Qiming Zhang
Jing Zhang
Dacheng Tao
ViT
48
329
0
07 Jun 2021
Vision Transformers with Hierarchical Attention
Yun-Hai Liu
Yu-Huan Wu
Guolei Sun
Le Zhang
Ajad Chhatkuli
Luc Van Gool
ViT
30
32
0
06 Jun 2021
Exploring the Limits of Out-of-Distribution Detection
Stanislav Fort
Jie Jessie Ren
Balaji Lakshminarayanan
16
323
0
06 Jun 2021
When Vision Transformers Outperform ResNets without Pre-training or Strong Data Augmentations
Xiangning Chen
Cho-Jui Hsieh
Boqing Gong
ViT
13
320
0
03 Jun 2021
Container: Context Aggregation Network
Peng Gao
Jiasen Lu
Hongsheng Li
Roozbeh Mottaghi
Aniruddha Kembhavi
ViT
11
69
0
02 Jun 2021
Can Attention Enable MLPs To Catch Up With CNNs?
Meng-Hao Guo
Zheng-Ning Liu
Tai-Jiang Mu
Dun Liang
Ralph Robert Martin
Shimin Hu
AAML
16
17
0
31 May 2021
A remark on a paper of Krotov and Hopfield [arXiv:2008.06996]
Fei Tang
Michael K Kopp
11
11
0
31 May 2021
Choose a Transformer: Fourier or Galerkin
Shuhao Cao
34
220
0
31 May 2021
MixerGAN: An MLP-Based Architecture for Unpaired Image-to-Image Translation
George Cazenavette
Manuel Ladron de Guevara
14
17
0
28 May 2021
An Attention Free Transformer
Shuangfei Zhai
Walter A. Talbott
Nitish Srivastava
Chen Huang
Hanlin Goh
Ruixiang Zhang
J. Susskind
ViT
19
127
0
28 May 2021
On the Bias Against Inductive Biases
George Cazenavette
Simon Lucey
SSL
18
1
0
28 May 2021
Pay Attention to MLPs
Hanxiao Liu
Zihang Dai
David R. So
Quoc V. Le
AI4CE
37
651
0
17 May 2021
Brain Inspired Face Recognition: A Computational Framework
P. Chowdhury
Angad Wadhwa
Nikhil Tyagi
CVBM
15
4
0
15 May 2021
FNet: Mixing Tokens with Fourier Transforms
James Lee-Thorp
Joshua Ainslie
Ilya Eckstein
Santiago Ontanon
13
516
0
09 May 2021
ResMLP: Feedforward networks for image classification with data-efficient training
Hugo Touvron
Piotr Bojanowski
Mathilde Caron
Matthieu Cord
Alaaeldin El-Nouby
...
Gautier Izacard
Armand Joulin
Gabriel Synnaeve
Jakob Verbeek
Hervé Jégou
VLM
16
655
0
07 May 2021
RepMLP: Re-parameterizing Convolutions into Fully-connected Layers for Image Recognition
Xiaohan Ding
Chunlong Xia
X. Zhang
Xiaojie Chu
Jungong Han
Guiguang Ding
15
92
0
05 May 2021
Sifting out the features by pruning: Are convolutional networks the winning lottery ticket of fully connected ones?
Franco Pellegrini
Giulio Biroli
44
6
0
27 Apr 2021
ImageNet-21K Pretraining for the Masses
T. Ridnik
Emanuel Ben-Baruch
Asaf Noy
Lihi Zelnik-Manor
SSeg
VLM
CLIP
173
686
0
22 Apr 2021
All Tokens Matter: Token Labeling for Training Better Vision Transformers
Zihang Jiang
Qibin Hou
Li-xin Yuan
Daquan Zhou
Yujun Shi
Xiaojie Jin
Anran Wang
Jiashi Feng
ViT
12
203
0
22 Apr 2021
Cloth Interactive Transformer for Virtual Try-On
Bin Ren
Hao Tang
Fanyang Meng
Runwei Ding
Philip H. S. Torr
N. Sebe
ViT
30
32
0
12 Apr 2021
On the Adversarial Robustness of Vision Transformers
Rulin Shao
Zhouxing Shi
Jinfeng Yi
Pin-Yu Chen
Cho-Jui Hsieh
ViT
25
137
0
29 Mar 2021
A Practical Survey on Faster and Lighter Transformers
Quentin Fournier
G. Caron
Daniel Aloise
6
93
0
26 Mar 2021
Swin Transformer: Hierarchical Vision Transformer using Shifted Windows
Ze Liu
Yutong Lin
Yue Cao
Han Hu
Yixuan Wei
Zheng-Wei Zhang
Stephen Lin
B. Guo
ViT
113
20,677
0
25 Mar 2021
ConViT: Improving Vision Transformers with Soft Convolutional Inductive Biases
Stéphane dÁscoli
Hugo Touvron
Matthew L. Leavitt
Ari S. Morcos
Giulio Biroli
Levent Sagun
ViT
29
803
0
19 Mar 2021
Understanding Invariance via Feedforward Inversion of Discriminatively Trained Classifiers
Piotr Teterwak
Chiyuan Zhang
Dilip Krishnan
Michael C. Mozer
23
9
0
15 Mar 2021
Pyramid Vision Transformer: A Versatile Backbone for Dense Prediction without Convolutions
Wenhai Wang
Enze Xie
Xiang Li
Deng-Ping Fan
Kaitao Song
Ding Liang
Tong Lu
Ping Luo
Ling Shao
ViT
274
3,622
0
24 Feb 2021
Previous
1
2
3
...
21
22
23
Next