ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1802.05751
  4. Cited By
Image Transformer
v1v2v3 (latest)

Image Transformer

15 February 2018
Niki Parmar
Ashish Vaswani
Jakob Uszkoreit
Lukasz Kaiser
Noam M. Shazeer
Alexander Ku
Dustin Tran
    ViT
ArXiv (abs)PDFHTML

Papers citing "Image Transformer"

50 / 837 papers shown
Title
Efficient Self-supervised Vision Transformers for Representation
  Learning
Efficient Self-supervised Vision Transformers for Representation Learning
Chunyuan Li
Jianwei Yang
Pengchuan Zhang
Mei Gao
Bin Xiao
Xiyang Dai
Lu Yuan
Jianfeng Gao
ViT
110
214
0
17 Jun 2021
XCiT: Cross-Covariance Image Transformers
XCiT: Cross-Covariance Image Transformers
Alaaeldin El-Nouby
Hugo Touvron
Mathilde Caron
Piotr Bojanowski
Matthijs Douze
...
Ivan Laptev
Natalia Neverova
Gabriel Synnaeve
Jakob Verbeek
Hervé Jégou
ViT
158
517
0
17 Jun 2021
Multi-head or Single-head? An Empirical Comparison for Transformer
  Training
Multi-head or Single-head? An Empirical Comparison for Transformer Training
Liyuan Liu
Jialu Liu
Jiawei Han
71
33
0
17 Jun 2021
Dual-view Molecule Pre-training
Dual-view Molecule Pre-training
Jinhua Zhu
Yingce Xia
Tao Qin
Wen-gang Zhou
Houqiang Li
Tie-Yan Liu
AI4CE
116
52
0
17 Jun 2021
Multi-Resolution Continuous Normalizing Flows
Multi-Resolution Continuous Normalizing Flows
Vikram S. Voleti
Chris Finlay
Adam M. Oberman
Christopher Pal
98
4
0
15 Jun 2021
Improved Transformer for High-Resolution GANs
Improved Transformer for High-Resolution GANs
Long Zhao
Zizhao Zhang
Ting Chen
Dimitris N. Metaxas
Han Zhang
ViT
133
96
0
14 Jun 2021
Rethinking Architecture Design for Tackling Data Heterogeneity in
  Federated Learning
Rethinking Architecture Design for Tackling Data Heterogeneity in Federated Learning
Liangqiong Qu
Yuyin Zhou
Paul Pu Liang
Yingda Xia
Feifei Wang
Ehsan Adeli
L. Fei-Fei
D. Rubin
FedMLAI4CE
114
186
0
10 Jun 2021
Space-time Mixing Attention for Video Transformer
Space-time Mixing Attention for Video Transformer
Adrian Bulat
Juan-Manuel Perez-Rua
Swathikiran Sudhakaran
Brais Martínez
Georgios Tzimiropoulos
ViT
95
127
0
10 Jun 2021
Soft Truncation: A Universal Training Technique of Score-based Diffusion
  Model for High Precision Score Estimation
Soft Truncation: A Universal Training Technique of Score-based Diffusion Model for High Precision Score Estimation
Dongjun Kim
Seung-Jae Shin
Kyungwoo Song
Wanmo Kang
Il-Chul Moon
128
97
0
10 Jun 2021
Keeping Your Eye on the Ball: Trajectory Attention in Video Transformers
Keeping Your Eye on the Ball: Trajectory Attention in Video Transformers
Mandela Patrick
Dylan Campbell
Yuki M. Asano
Ishan Misra
Ishan Misra Florian Metze
Christoph Feichtenhofer
Andrea Vedaldi
João F. Henriques
114
282
0
09 Jun 2021
Semi-Supervised 3D Hand-Object Poses Estimation with Interactions in
  Time
Semi-Supervised 3D Hand-Object Poses Estimation with Interactions in Time
Shao-Wei Liu
Hanwen Jiang
Jiarui Xu
Sifei Liu
Xiaolong Wang
3DH
130
165
0
09 Jun 2021
Multi-Facet Clustering Variational Autoencoders
Multi-Facet Clustering Variational Autoencoders
Fabian Falck
Haoting Zhang
M. Willetts
G. Nicholson
C. Yau
Chris Holmes
DRL
80
44
0
09 Jun 2021
Densely connected normalizing flows
Densely connected normalizing flows
Matej Grcić
Ivan Grubišić
Sinisa Segvic
TPM
103
59
0
08 Jun 2021
A Survey of Transformers
A Survey of Transformers
Tianyang Lin
Yuxin Wang
Xiangyang Liu
Xipeng Qiu
ViT
208
1,150
0
08 Jun 2021
Chasing Sparsity in Vision Transformers: An End-to-End Exploration
Chasing Sparsity in Vision Transformers: An End-to-End Exploration
Tianlong Chen
Yu Cheng
Zhe Gan
Lu Yuan
Lei Zhang
Zhangyang Wang
ViT
70
224
0
08 Jun 2021
CAPE: Encoding Relative Positions with Continuous Augmented Positional
  Embeddings
CAPE: Encoding Relative Positions with Continuous Augmented Positional Embeddings
Tatiana Likhomanenko
Qiantong Xu
Gabriel Synnaeve
R. Collobert
A. Rogozhnikov
OODViT
90
60
0
06 Jun 2021
Learnable Fourier Features for Multi-Dimensional Spatial Positional
  Encoding
Learnable Fourier Features for Multi-Dimensional Spatial Positional Encoding
Yang Li
Si Si
Gang Li
Cho-Jui Hsieh
Samy Bengio
102
96
0
05 Jun 2021
Associating Objects with Transformers for Video Object Segmentation
Associating Objects with Transformers for Video Object Segmentation
Zongxin Yang
Yunchao Wei
Yi Yang
134
298
0
04 Jun 2021
Self-Attention Between Datapoints: Going Beyond Individual Input-Output
  Pairs in Deep Learning
Self-Attention Between Datapoints: Going Beyond Individual Input-Output Pairs in Deep Learning
Jannik Kossen
Neil Band
Clare Lyle
Aidan Gomez
Tom Rainforth
Y. Gal
OOD3DPC
133
142
0
04 Jun 2021
The Image Local Autoregressive Transformer
The Image Local Autoregressive Transformer
Chenjie Cao
Yue Hong
Xiang Li
Chengrong Wang
C. Xu
Xiangyang Xue
Yanwei Fu
82
13
0
04 Jun 2021
Few-Shot Segmentation via Cycle-Consistent Transformer
Few-Shot Segmentation via Cycle-Consistent Transformer
Gengwei Zhang
Guoliang Kang
Yi Yang
Yunchao Wei
ViT
107
187
0
04 Jun 2021
E2E-VLP: End-to-End Vision-Language Pre-training Enhanced by Visual
  Learning
E2E-VLP: End-to-End Vision-Language Pre-training Enhanced by Visual Learning
Haiyang Xu
Ming Yan
Chenliang Li
Bin Bi
Songfang Huang
Wenming Xiao
Fei Huang
VLM
118
119
0
03 Jun 2021
Luna: Linear Unified Nested Attention
Luna: Linear Unified Nested Attention
Xuezhe Ma
Xiang Kong
Sinong Wang
Chunting Zhou
Jonathan May
Hao Ma
Luke Zettlemoyer
93
113
0
03 Jun 2021
Ember: No-Code Context Enrichment via Similarity-Based Keyless Joins
Ember: No-Code Context Enrichment via Similarity-Based Keyless Joins
S. Suri
Ihab F. Ilyas
Christopher Ré
Theodoros Rekatsinas
51
22
0
02 Jun 2021
On the Distribution, Sparsity, and Inference-time Quantization of
  Attention Values in Transformers
On the Distribution, Sparsity, and Inference-time Quantization of Attention Values in Transformers
Tianchu Ji
Shraddhan Jain
M. Ferdman
Peter Milder
H. Andrew Schwartz
Niranjan Balasubramanian
MQ
113
16
0
02 Jun 2021
Self-Supervised Document Similarity Ranking via Contextualized Language
  Models and Hierarchical Inference
Self-Supervised Document Similarity Ranking via Contextualized Language Models and Hierarchical Inference
Dvir Ginzburg
Itzik Malkiel
Oren Barkan
Avi Caciularu
Noam Koenigstein
RALM
75
27
0
02 Jun 2021
Omnizart: A General Toolbox for Automatic Music Transcription
Omnizart: A General Toolbox for Automatic Music Transcription
Yu-Te Wu
Yin-Jyun Luo
Tsung-Ping Chen
I-Chieh Wei
Jui-Yang Hsu
Yi-Chin Chuang
Li Su
SyDa
65
31
0
01 Jun 2021
An Attention Free Transformer
An Attention Free Transformer
Shuangfei Zhai
Walter A. Talbott
Nitish Srivastava
Chen Huang
Hanlin Goh
Ruixiang Zhang
J. Susskind
ViT
94
132
0
28 May 2021
Nested Hierarchical Transformer: Towards Accurate, Data-Efficient and
  Interpretable Visual Understanding
Nested Hierarchical Transformer: Towards Accurate, Data-Efficient and Interpretable Visual Understanding
Zizhao Zhang
Han Zhang
Long Zhao
Ting Chen
Sercan O. Arik
Tomas Pfister
ViT
102
174
0
26 May 2021
CogView: Mastering Text-to-Image Generation via Transformers
CogView: Mastering Text-to-Image Generation via Transformers
Ming Ding
Zhuoyi Yang
Wenyi Hong
Wendi Zheng
Chang Zhou
...
Junyang Lin
Xu Zou
Zhou Shao
Hongxia Yang
Jie Tang
ViTVLM
161
784
0
26 May 2021
Properties of the After Kernel
Properties of the After Kernel
Philip M. Long
75
29
0
21 May 2021
Combining Transformer Generators with Convolutional Discriminators
Combining Transformer Generators with Convolutional Discriminators
Ricard Durall
Stanislav Frolov
Jörn Hees
Federico Raue
Franz-Josef Pfreundt
Andreas Dengel
J. Keuper
ViT
68
16
0
21 May 2021
DeepCAD: A Deep Generative Network for Computer-Aided Design Models
DeepCAD: A Deep Generative Network for Computer-Aided Design Models
Rundi Wu
Chang Xiao
Changxi Zheng
3DPC
103
173
0
20 May 2021
I2C2W: Image-to-Character-to-Word Transformers for Accurate Scene Text
  Recognition
I2C2W: Image-to-Character-to-Word Transformers for Accurate Scene Text Recognition
Chuhui Xue
Jiaxing Huang
Wenqing Zhang
Shijian Lu
Changhu Wang
S. Bai
106
17
0
18 May 2021
Parallel and Flexible Sampling from Autoregressive Models via Langevin
  Dynamics
Parallel and Flexible Sampling from Autoregressive Models via Langevin Dynamics
V. Jayaram
John Thickstun
DiffM
107
25
0
17 May 2021
Vision Transformers are Robust Learners
Vision Transformers are Robust Learners
Sayak Paul
Pin-Yu Chen
ViT
86
312
0
17 May 2021
EL-Attention: Memory Efficient Lossless Attention for Generation
EL-Attention: Memory Efficient Lossless Attention for Generation
Yu Yan
Jiusheng Chen
Weizhen Qi
Nikhil Bhendawade
Yeyun Gong
Nan Duan
Ruofei Zhang
VLM
68
6
0
11 May 2021
HyperHyperNetworks for the Design of Antenna Arrays
HyperHyperNetworks for the Design of Antenna Arrays
Shahar Lutati
Lior Wolf
69
3
0
09 May 2021
FNet: Mixing Tokens with Fourier Transforms
FNet: Mixing Tokens with Fourier Transforms
James Lee-Thorp
Joshua Ainslie
Ilya Eckstein
Santiago Ontanon
149
537
0
09 May 2021
RBNN: Memory-Efficient Reconfigurable Deep Binary Neural Network with IP
  Protection for Internet of Things
RBNN: Memory-Efficient Reconfigurable Deep Binary Neural Network with IP Protection for Internet of Things
Huming Qiu
Hua Ma
Zhi-Li Zhang
Yifeng Zheng
Anmin Fu
Pan Zhou
Yansong Gao
Derek Abbott
S. Al-Sarawi
MQ
88
10
0
09 May 2021
Long-Span Summarization via Local Attention and Content Selection
Long-Span Summarization via Local Attention and Content Selection
Potsawee Manakul
Mark Gales
87
42
0
08 May 2021
ResMLP: Feedforward networks for image classification with
  data-efficient training
ResMLP: Feedforward networks for image classification with data-efficient training
Hugo Touvron
Piotr Bojanowski
Mathilde Caron
Matthieu Cord
Alaaeldin El-Nouby
...
Gautier Izacard
Armand Joulin
Gabriel Synnaeve
Jakob Verbeek
Hervé Jégou
VLM
135
675
0
07 May 2021
Handwritten Mathematical Expression Recognition with Bidirectionally
  Trained Transformer
Handwritten Mathematical Expression Recognition with Bidirectionally Trained Transformer
Wenqi Zhao
Liangcai Gao
Zuoyu Yan
Shuai Peng
Lin Du
Ziyin Zhang
ViT
181
55
0
06 May 2021
MLP-Mixer: An all-MLP Architecture for Vision
MLP-Mixer: An all-MLP Architecture for Vision
Ilya O. Tolstikhin
N. Houlsby
Alexander Kolesnikov
Lucas Beyer
Xiaohua Zhai
...
Andreas Steiner
Daniel Keysers
Jakob Uszkoreit
Mario Lucic
Alexey Dosovitskiy
538
2,722
0
04 May 2021
Moving Towards Centers: Re-ranking with Attention and Memory for
  Re-identification
Moving Towards Centers: Re-ranking with Attention and Memory for Re-identification
Yunhao Zhou
Yi Wang
Lap-Pui Chau
96
10
0
04 May 2021
AGMB-Transformer: Anatomy-Guided Multi-Branch Transformer Network for
  Automated Evaluation of Root Canal Therapy
AGMB-Transformer: Anatomy-Guided Multi-Branch Transformer Network for Automated Evaluation of Root Canal Therapy
Yunxiang Li
G. Zeng
Yifan Zhang
Jun Wang
Qianni Zhang
...
Neng Xia
Ruizi Peng
Kai Tang
Yaqi Wang
Shuai Wang
MedImAI4CE
166
29
0
02 May 2021
Perceptual Image Quality Assessment with Transformers
Perceptual Image Quality Assessment with Transformers
Manri Cheon
Sung-Jun Yoon
Byungyeon Kang
Junwoo Lee
ViT
106
117
0
30 Apr 2021
Inpainting Transformer for Anomaly Detection
Inpainting Transformer for Anomaly Detection
Jonathan Pirnay
K. Chai
ViT
211
169
0
28 Apr 2021
Autoregressive Dynamics Models for Offline Policy Evaluation and
  Optimization
Autoregressive Dynamics Models for Offline Policy Evaluation and Optimization
Michael Ruogu Zhang
T. Paine
Ofir Nachum
Cosmin Paduraru
George Tucker
Ziyun Wang
Mohammad Norouzi
OffRL
91
49
0
28 Apr 2021
Twins: Revisiting the Design of Spatial Attention in Vision Transformers
Twins: Revisiting the Design of Spatial Attention in Vision Transformers
Xiangxiang Chu
Zhi Tian
Yuqing Wang
Bo Zhang
Haibing Ren
Xiaolin K. Wei
Huaxia Xia
Chunhua Shen
ViT
118
1,034
0
28 Apr 2021
Previous
123...111213...151617
Next