ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2105.01601
  4. Cited By
MLP-Mixer: An all-MLP Architecture for Vision
v1v2v3v4 (latest)

MLP-Mixer: An all-MLP Architecture for Vision

4 May 2021
Ilya O. Tolstikhin
N. Houlsby
Alexander Kolesnikov
Lucas Beyer
Xiaohua Zhai
Thomas Unterthiner
Jessica Yung
Andreas Steiner
Daniel Keysers
Jakob Uszkoreit
Mario Lucic
Alexey Dosovitskiy
ArXiv (abs)PDFHTML

Papers citing "MLP-Mixer: An all-MLP Architecture for Vision"

50 / 1,144 papers shown
Title
Critical Initialization of Wide and Deep Neural Networks through Partial
  Jacobians: General Theory and Applications
Critical Initialization of Wide and Deep Neural Networks through Partial Jacobians: General Theory and Applications
Darshil Doshi
Tianyu He
Andrey Gromov
79
10
0
23 Nov 2021
SimpleTRON: Simple Transformer with O(N) Complexity
SimpleTRON: Simple Transformer with O(N) Complexity
Uladzislau Yorsh
Alexander Kovalenko
Vojtvech Vanvcura
Daniel Vavsata
Pavel Kordík
Tomávs Mikolov
136
1
0
23 Nov 2021
GenReg: Deep Generative Method for Fast Point Cloud Registration
GenReg: Deep Generative Method for Fast Point Cloud Registration
Xiaoshui Huang
Zongyi Xu
Guofeng Mei
Sheng Li
Jian Zhang
Yifan Zuo
Yucheng Wang
3DPC
38
1
0
23 Nov 2021
MetaFormer Is Actually What You Need for Vision
MetaFormer Is Actually What You Need for Vision
Weihao Yu
Mi Luo
Pan Zhou
Chenyang Si
Yichen Zhou
Xinchao Wang
Jiashi Feng
Shuicheng Yan
173
925
0
22 Nov 2021
PointMixer: MLP-Mixer for Point Cloud Understanding
PointMixer: MLP-Mixer for Point Cloud Understanding
Jaesung Choe
Chunghyun Park
François Rameau
Jaesik Park
In So Kweon
3DPC
128
102
0
22 Nov 2021
Are Vision Transformers Robust to Patch Perturbations?
Are Vision Transformers Robust to Patch Perturbations?
Jindong Gu
Volker Tresp
Yao Qin
AAMLViT
108
64
0
20 Nov 2021
Discrete Representations Strengthen Vision Transformer Robustness
Discrete Representations Strengthen Vision Transformer Robustness
Chengzhi Mao
Lu Jiang
Mostafa Dehghani
Carl Vondrick
Rahul Sukthankar
Irfan Essa
ViT
98
43
0
20 Nov 2021
Global and Local Alignment Networks for Unpaired Image-to-Image
  Translation
Global and Local Alignment Networks for Unpaired Image-to-Image Translation
Guanglei Yang
Hao Tang
Humphrey Shi
M. Ding
N. Sebe
Radu Timofte
Luc Van Gool
Elisa Ricci
49
1
0
19 Nov 2021
UFO: A UniFied TransfOrmer for Vision-Language Representation Learning
UFO: A UniFied TransfOrmer for Vision-Language Representation Learning
Jianfeng Wang
Xiaowei Hu
Zhe Gan
Zhengyuan Yang
Xiyang Dai
Zicheng Liu
Yumao Lu
Lijuan Wang
ViT
73
57
0
19 Nov 2021
Benchmarking and scaling of deep learning models for land cover image
  classification
Benchmarking and scaling of deep learning models for land cover image classification
Ioannis Papoutsis
Nikolaos Ioannis Bountos
Angelos Zavras
Dimitrios Michail
Christos Tryfonopoulos
99
60
0
18 Nov 2021
INTERN: A New Learning Paradigm Towards General Vision
INTERN: A New Learning Paradigm Towards General Vision
Jing Shao
Siyu Chen
Yangguang Li
Kun Wang
Zhen-fei Yin
...
F. Yu
Junjie Yan
Dahua Lin
Xiaogang Wang
Yu Qiao
108
34
0
16 Nov 2021
LiT: Zero-Shot Transfer with Locked-image text Tuning
LiT: Zero-Shot Transfer with Locked-image text Tuning
Xiaohua Zhai
Tianlin Li
Basil Mustafa
Andreas Steiner
Daniel Keysers
Alexander Kolesnikov
Lucas Beyer
VLM
131
561
0
15 Nov 2021
Attention Mechanisms in Computer Vision: A Survey
Attention Mechanisms in Computer Vision: A Survey
Meng-Hao Guo
Tianhan Xu
Jiangjiang Liu
Zheng-Ning Liu
Peng-Tao Jiang
Tai-Jiang Mu
Song-Hai Zhang
Ralph Robert Martin
Ming-Ming Cheng
Shimin Hu
140
1,718
0
15 Nov 2021
A Survey on Hyperdimensional Computing aka Vector Symbolic
  Architectures, Part II: Applications, Cognitive Models, and Challenges
A Survey on Hyperdimensional Computing aka Vector Symbolic Architectures, Part II: Applications, Cognitive Models, and Challenges
Denis Kleyko
D. Rachkovskij
Evgeny Osipov
A. Rahim
104
136
0
12 Nov 2021
Are Transformers More Robust Than CNNs?
Are Transformers More Robust Than CNNs?
Yutong Bai
Jieru Mei
Alan Yuille
Cihang Xie
ViTAAML
258
267
0
10 Nov 2021
Sliced Recursive Transformer
Sliced Recursive Transformer
Zhiqiang Shen
Zechun Liu
Eric P. Xing
ViT
59
27
0
09 Nov 2021
Personalized Benchmarking with the Ludwig Benchmarking Toolkit
Personalized Benchmarking with the Ludwig Benchmarking Toolkit
A. Narayan
Piero Molino
Karan Goel
Willie Neiswanger
Christopher Ré
76
11
0
08 Nov 2021
Are we ready for a new paradigm shift? A Survey on Visual Deep MLP
Are we ready for a new paradigm shift? A Survey on Visual Deep MLP
Ruiyang Liu
Hai-Tao Zheng
Li Tao
Dun Liang
Haitao Zheng
213
100
0
07 Nov 2021
TermiNeRF: Ray Termination Prediction for Efficient Neural Rendering
TermiNeRF: Ray Termination Prediction for Efficient Neural Rendering
Martin Piala
R. Clark
81
80
0
05 Nov 2021
Efficiently Modeling Long Sequences with Structured State Spaces
Efficiently Modeling Long Sequences with Structured State Spaces
Albert Gu
Karan Goel
Christopher Ré
229
1,843
0
31 Oct 2021
Blending Anti-Aliasing into Vision Transformer
Blending Anti-Aliasing into Vision Transformer
Shengju Qian
Hao Shao
Yi Zhu
Mu Li
Jiaya Jia
109
21
0
28 Oct 2021
Arbitrary Distribution Modeling with Censorship in Real-Time Bidding
  Advertising
Arbitrary Distribution Modeling with Censorship in Real-Time Bidding Advertising
Xu Li
Michelle Ma Zhang
Youjun Tong
Zhenya Wang
35
9
0
26 Oct 2021
K-Lane: Lidar Lane Dataset and Benchmark for Urban Roads and Highways
K-Lane: Lidar Lane Dataset and Benchmark for Urban Roads and Highways
Dong-Hee Paek
Seung-Hyung Kong
Kevin Tirta Wijaya
3DV
96
22
0
21 Oct 2021
A channel attention based MLP-Mixer network for motor imagery decoding
  with EEG
A channel attention based MLP-Mixer network for motor imagery decoding with EEG
Yan-Nan He
Zhiyang Lu
Jun Wang
Jun Shi
32
9
0
21 Oct 2021
A Data-Centric Optimization Framework for Machine Learning
A Data-Centric Optimization Framework for Machine Learning
Oliver Rausch
Tal Ben-Nun
Nikoli Dryden
Andrei Ivanov
Shigang Li
Torsten Hoefler
AI4CE
45
16
0
20 Oct 2021
Ranking and Tuning Pre-trained Models: A New Paradigm for Exploiting
  Model Hubs
Ranking and Tuning Pre-trained Models: A New Paradigm for Exploiting Model Hubs
Kaichao You
Yong Liu
Ziyang Zhang
Jianmin Wang
Michael I. Jordan
Mingsheng Long
226
34
0
20 Oct 2021
Cascaded Cross MLP-Mixer GANs for Cross-View Image Translation
Cascaded Cross MLP-Mixer GANs for Cross-View Image Translation
Bin Ren
Hao Tang
N. Sebe
76
30
0
19 Oct 2021
Inductive Biases and Variable Creation in Self-Attention Mechanisms
Inductive Biases and Variable Creation in Self-Attention Mechanisms
Benjamin L. Edelman
Surbhi Goel
Sham Kakade
Cyril Zhang
97
125
0
19 Oct 2021
SCENIC: A JAX Library for Computer Vision Research and Beyond
SCENIC: A JAX Library for Computer Vision Research and Beyond
Mostafa Dehghani
A. Gritsenko
Anurag Arnab
Matthias Minderer
Yi Tay
106
68
0
18 Oct 2021
Graph-less Neural Networks: Teaching Old MLPs New Tricks via
  Distillation
Graph-less Neural Networks: Teaching Old MLPs New Tricks via Distillation
Shichang Zhang
Yozen Liu
Yizhou Sun
Neil Shah
91
185
0
17 Oct 2021
Sharpness-Aware Minimization Improves Language Model Generalization
Sharpness-Aware Minimization Improves Language Model Generalization
Dara Bahri
H. Mobahi
Yi Tay
180
104
0
16 Oct 2021
Attention-Free Keyword Spotting
Attention-Free Keyword Spotting
Mashrur M. Morshed
Ahmad Omar Ahsan
114
9
0
14 Oct 2021
Unsupervised Learning of Full-Waveform Inversion: Connecting CNN and
  Partial Differential Equation in a Loop
Unsupervised Learning of Full-Waveform Inversion: Connecting CNN and Partial Differential Equation in a Loop
Peng Jin
Xitong Zhang
Yinpeng Chen
Sharon X. Huang
Zicheng Liu
Youzuo Lin
99
48
0
14 Oct 2021
Two-argument activation functions learn soft XOR operations like
  cortical neurons
Two-argument activation functions learn soft XOR operations like cortical neurons
Kijung Yoon
Emin Orhan
Juhyeon Kim
Xaq Pitkow
MLT
58
0
0
13 Oct 2021
Open-Set Recognition: a Good Closed-Set Classifier is All You Need?
Open-Set Recognition: a Good Closed-Set Classifier is All You Need?
S. Vaze
Kai Han
Andrea Vedaldi
Andrew Zisserman
BDL
253
437
0
12 Oct 2021
Phase Collapse in Neural Networks
Phase Collapse in Neural Networks
Florentin Guth
J. Zarka
S. Mallat
48
7
0
11 Oct 2021
Vision Transformer based COVID-19 Detection using Chest X-rays
Vision Transformer based COVID-19 Detection using Chest X-rays
Koushik Sivarama Krishnan
Karthik Sivarama Krishnan
ViTMedIm
68
57
0
09 Oct 2021
Adversarial Token Attacks on Vision Transformers
Adversarial Token Attacks on Vision Transformers
Ameya Joshi
Gauri Jagatap
Chinmay Hegde
ViT
99
19
0
08 Oct 2021
UniNet: Unified Architecture Search with Convolution, Transformer, and
  MLP
UniNet: Unified Architecture Search with Convolution, Transformer, and MLP
Jihao Liu
Hongsheng Li
Guanglu Song
Xin Huang
Yu Liu
ViT
101
36
0
08 Oct 2021
MC-LCR: Multi-modal contrastive classification by locally correlated
  representations for effective face forgery detection
MC-LCR: Multi-modal contrastive classification by locally correlated representations for effective face forgery detection
Gaojian Wang
Qian Jiang
Xin Jin
Wei Li
Xiaohui Cui
CVBM
67
25
0
07 Oct 2021
Attention is All You Need? Good Embeddings with Statistics are
  enough:Large Scale Audio Understanding without Transformers/ Convolutions/
  BERTs/ Mixers/ Attention/ RNNs or ....
Attention is All You Need? Good Embeddings with Statistics are enough:Large Scale Audio Understanding without Transformers/ Convolutions/ BERTs/ Mixers/ Attention/ RNNs or ....
Prateek Verma
AI4TS
65
2
0
07 Oct 2021
Adversarial Robustness Comparison of Vision Transformer and MLP-Mixer to
  CNNs
Adversarial Robustness Comparison of Vision Transformer and MLP-Mixer to CNNs
Philipp Benz
Soomin Ham
Chaoning Zhang
Adil Karjauv
In So Kweon
AAMLViT
100
80
0
06 Oct 2021
PoNet: Pooling Network for Efficient Token Mixing in Long Sequences
PoNet: Pooling Network for Efficient Token Mixing in Long Sequences
Chao-Hong Tan
Qian Chen
Wen Wang
Qinglin Zhang
Siqi Zheng
Zhenhua Ling
ViT
42
12
0
06 Oct 2021
Exploring the Limits of Large Scale Pre-training
Exploring the Limits of Large Scale Pre-training
Samira Abnar
Mostafa Dehghani
Behnam Neyshabur
Hanie Sedghi
AI4CE
111
119
0
05 Oct 2021
Deep Instance Segmentation with Automotive Radar Detection Points
Deep Instance Segmentation with Automotive Radar Detection Points
Tao Huang
Weiyi Xiong
Liping Bai
Yu Xia
Wei Chen
Wanli Ouyang
Bing Zhu
159
54
0
05 Oct 2021
Efficient Identification of Butterfly Sparse Matrix Factorizations
Efficient Identification of Butterfly Sparse Matrix Factorizations
Léon Zheng
E. Riccietti
Rémi Gribonval
67
6
0
04 Oct 2021
ResNet strikes back: An improved training procedure in timm
ResNet strikes back: An improved training procedure in timm
Ross Wightman
Hugo Touvron
Hervé Jégou
AI4TS
276
500
0
01 Oct 2021
UFO-ViT: High Performance Linear Vision Transformer without Softmax
UFO-ViT: High Performance Linear Vision Transformer without Softmax
Jeonggeun Song
ViT
173
21
0
29 Sep 2021
Scale Efficiently: Insights from Pre-training and Fine-tuning
  Transformers
Scale Efficiently: Insights from Pre-training and Fine-tuning Transformers
Yi Tay
Mostafa Dehghani
J. Rao
W. Fedus
Samira Abnar
Hyung Won Chung
Sharan Narang
Dani Yogatama
Ashish Vaswani
Donald Metzler
283
115
0
22 Sep 2021
Sparse MLP for Image Recognition: Is Self-Attention Really Necessary?
Sparse MLP for Image Recognition: Is Self-Attention Really Necessary?
Chuanxin Tang
Yucheng Zhao
Guangting Wang
Chong Luo
Wenxuan Xie
Wenjun Zeng
MoEViT
86
100
0
12 Sep 2021
Previous
123...20212223
Next