ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1912.12180
  4. Cited By
Axial Attention in Multidimensional Transformers

Axial Attention in Multidimensional Transformers

20 December 2019
Jonathan Ho
Nal Kalchbrenner
Dirk Weissenborn
Tim Salimans
ArXivPDFHTML

Papers citing "Axial Attention in Multidimensional Transformers"

37 / 287 papers shown
Title
DeepViT: Towards Deeper Vision Transformer
DeepViT: Towards Deeper Vision Transformer
Daquan Zhou
Bingyi Kang
Xiaojie Jin
Linjie Yang
Xiaochen Lian
Zihang Jiang
Qibin Hou
Jiashi Feng
ViT
42
510
0
22 Mar 2021
Scalable Vision Transformers with Hierarchical Pooling
Scalable Vision Transformers with Hierarchical Pooling
Zizheng Pan
Bohan Zhuang
Jing Liu
Haoyu He
Jianfei Cai
ViT
27
126
0
19 Mar 2021
Perceiver: General Perception with Iterative Attention
Perceiver: General Perception with Iterative Attention
Andrew Jaegle
Felix Gimeno
Andrew Brock
Andrew Zisserman
Oriol Vinyals
João Carreira
VLM
ViT
MDE
91
976
0
04 Mar 2021
Random Feature Attention
Random Feature Attention
Hao Peng
Nikolaos Pappas
Dani Yogatama
Roy Schwartz
Noah A. Smith
Lingpeng Kong
36
348
0
03 Mar 2021
Generative Adversarial Transformers
Generative Adversarial Transformers
Drew A. Hudson
C. L. Zitnick
ViT
25
179
0
01 Mar 2021
DR-TANet: Dynamic Receptive Temporal Attention Network for Street Scene
  Change Detection
DR-TANet: Dynamic Receptive Temporal Attention Network for Street Scene Change Detection
Shuo Chen
Kailun Yang
Rainer Stiefelhagen
13
38
0
01 Mar 2021
Convolution-Free Medical Image Segmentation using Transformers
Convolution-Free Medical Image Segmentation using Transformers
Davood Karimi
Serge Vasylechko
Ali Gholipour
ViT
MedIm
84
121
0
26 Feb 2021
Medical Transformer: Gated Axial-Attention for Medical Image
  Segmentation
Medical Transformer: Gated Axial-Attention for Medical Image Segmentation
Jeya Maria Jose Valanarasu
Poojan Oza
I. Hacihaliloglu
Vishal M. Patel
ViT
MedIm
20
962
0
21 Feb 2021
Going Full-TILT Boogie on Document Understanding with Text-Image-Layout
  Transformer
Going Full-TILT Boogie on Document Understanding with Text-Image-Layout Transformer
Rafal Powalski
Łukasz Borchmann
Dawid Jurkiewicz
Tomasz Dwojak
Michal Pietruszka
Gabriela Pałka
ViT
36
157
0
18 Feb 2021
LambdaNetworks: Modeling Long-Range Interactions Without Attention
LambdaNetworks: Modeling Long-Range Interactions Without Attention
Irwan Bello
281
179
0
17 Feb 2021
Axial Residual Networks for CycleGAN-based Voice Conversion
Axial Residual Networks for CycleGAN-based Voice Conversion
J. You
Gyuhyeon Nam
Dalhyun Kim
Gyeongsu Chae
11
3
0
16 Feb 2021
Is Space-Time Attention All You Need for Video Understanding?
Is Space-Time Attention All You Need for Video Understanding?
Gedas Bertasius
Heng Wang
Lorenzo Torresani
ViT
283
1,984
0
09 Feb 2021
Colorization Transformer
Colorization Transformer
Manoj Kumar
Dirk Weissenborn
Nal Kalchbrenner
ViT
232
156
0
08 Feb 2021
Bottleneck Transformers for Visual Recognition
Bottleneck Transformers for Visual Recognition
A. Srinivas
Nayeon Lee
Niki Parmar
Jonathon Shlens
Pieter Abbeel
Ashish Vaswani
SLR
290
980
0
27 Jan 2021
Channelized Axial Attention for Semantic Segmentation -- Considering
  Channel Relation within Spatial Attention for Semantic Segmentation
Channelized Axial Attention for Semantic Segmentation -- Considering Channel Relation within Spatial Attention for Semantic Segmentation
Ye Huang
Di Kang
W. Jia
Xiangjian He
Liu Liu
31
35
0
19 Jan 2021
Transformers in Vision: A Survey
Transformers in Vision: A Survey
Salman Khan
Muzammal Naseer
Munawar Hayat
Syed Waqas Zamir
Fahad Shahbaz Khan
M. Shah
ViT
227
2,431
0
04 Jan 2021
Rethinking Semantic Segmentation from a Sequence-to-Sequence Perspective
  with Transformers
Rethinking Semantic Segmentation from a Sequence-to-Sequence Perspective with Transformers
Sixiao Zheng
Jiachen Lu
Hengshuang Zhao
Xiatian Zhu
Zekun Luo
...
Yanwei Fu
Jianfeng Feng
Tao Xiang
Philip Torr
Li Zhang
ViT
42
2,841
0
31 Dec 2020
3D Axial-Attention for Lung Nodule Classification
3D Axial-Attention for Lung Nodule Classification
M. Al-Shabi
Kelvin Shak
Maxine Tan
MedIm
9
18
0
28 Dec 2020
HyperSeg: Patch-wise Hypernetwork for Real-time Semantic Segmentation
HyperSeg: Patch-wise Hypernetwork for Real-time Semantic Segmentation
Y. Nirkin
Lior Wolf
Tal Hassner
SSeg
16
177
0
21 Dec 2020
Taming Transformers for High-Resolution Image Synthesis
Taming Transformers for High-Resolution Image Synthesis
Patrick Esser
Robin Rombach
Bjorn Ommer
ViT
64
2,819
0
17 Dec 2020
MaX-DeepLab: End-to-End Panoptic Segmentation with Mask Transformers
MaX-DeepLab: End-to-End Panoptic Segmentation with Mask Transformers
Huiyu Wang
Yukun Zhu
Hartwig Adam
Alan Yuille
Liang-Chieh Chen
ViT
40
527
0
01 Dec 2020
Long Range Arena: A Benchmark for Efficient Transformers
Long Range Arena: A Benchmark for Efficient Transformers
Yi Tay
Mostafa Dehghani
Samira Abnar
Songlin Yang
Dara Bahri
Philip Pham
J. Rao
Liu Yang
Sebastian Ruder
Donald Metzler
47
693
0
08 Nov 2020
Deep coastal sea elements forecasting using U-Net based models
Deep coastal sea elements forecasting using U-Net based models
Jesús García Fernández
Ismail Alaoui Abdellaoui
S. Mehrkanoon
AI4Cl
33
27
0
06 Nov 2020
An Image is Worth 16x16 Words: Transformers for Image Recognition at
  Scale
An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale
Alexey Dosovitskiy
Lucas Beyer
Alexander Kolesnikov
Dirk Weissenborn
Xiaohua Zhai
...
Matthias Minderer
G. Heigold
Sylvain Gelly
Jakob Uszkoreit
N. Houlsby
ViT
41
39,330
0
22 Oct 2020
Is Batch Norm unique? An empirical investigation and prescription to
  emulate the best properties of common normalizers without batch dependence
Is Batch Norm unique? An empirical investigation and prescription to emulate the best properties of common normalizers without batch dependence
Vinay Rao
Jascha Narain Sohl-Dickstein
BDL
40
4
0
21 Oct 2020
Deformable DETR: Deformable Transformers for End-to-End Object Detection
Deformable DETR: Deformable Transformers for End-to-End Object Detection
Xizhou Zhu
Weijie Su
Lewei Lu
Bin Li
Xiaogang Wang
Jifeng Dai
ViT
63
4,929
0
08 Oct 2020
Global Self-Attention Networks for Image Recognition
Global Self-Attention Networks for Image Recognition
Zhuoran Shen
Irwan Bello
Raviteja Vemulapalli
Xuhui Jia
Ching-Hui Chen
ViT
22
28
0
06 Oct 2020
Deep multi-stations weather forecasting: explainable recurrent
  convolutional neural networks
Deep multi-stations weather forecasting: explainable recurrent convolutional neural networks
Ismail Alaoui Abdellaoui
S. Mehrkanoon
AI4TS
6
11
0
23 Sep 2020
Efficient Transformers: A Survey
Efficient Transformers: A Survey
Yi Tay
Mostafa Dehghani
Dara Bahri
Donald Metzler
VLM
111
1,102
0
14 Sep 2020
Latent Video Transformer
Latent Video Transformer
Ruslan Rakhimov
Denis Volkhonskiy
Alexey Artemov
Denis Zorin
Evgeny Burnaev
VGen
33
118
0
18 Jun 2020
Jukebox: A Generative Model for Music
Jukebox: A Generative Model for Music
Prafulla Dhariwal
Heewoo Jun
Christine Payne
Jong Wook Kim
Alec Radford
Ilya Sutskever
VLM
23
722
0
30 Apr 2020
Beyond 512 Tokens: Siamese Multi-depth Transformer-based Hierarchical
  Encoder for Long-Form Document Matching
Beyond 512 Tokens: Siamese Multi-depth Transformer-based Hierarchical Encoder for Long-Form Document Matching
Liu Yang
Mingyang Zhang
Cheng Li
Michael Bendersky
Marc Najork
32
86
0
26 Apr 2020
A Spatio-temporal Transformer for 3D Human Motion Prediction
A Spatio-temporal Transformer for 3D Human Motion Prediction
Emre Aksan
Manuel Kaufmann
Peng Cao
Otmar Hilliges
ViT
28
223
0
18 Apr 2020
MetNet: A Neural Weather Model for Precipitation Forecasting
MetNet: A Neural Weather Model for Precipitation Forecasting
C. Sønderby
L. Espeholt
Jonathan Heek
Mostafa Dehghani
Avital Oliver
Tim Salimans
Shreya Agrawal
Jason Hickey
Nal Kalchbrenner
AI4Cl
231
273
0
24 Mar 2020
Axial-DeepLab: Stand-Alone Axial-Attention for Panoptic Segmentation
Axial-DeepLab: Stand-Alone Axial-Attention for Panoptic Segmentation
Huiyu Wang
Yukun Zhu
Bradley Green
Hartwig Adam
Alan Yuille
Liang-Chieh Chen
3DPC
28
657
0
17 Mar 2020
A Decomposable Attention Model for Natural Language Inference
A Decomposable Attention Model for Natural Language Inference
Ankur P. Parikh
Oscar Täckström
Dipanjan Das
Jakob Uszkoreit
213
1,367
0
06 Jun 2016
Pixel Recurrent Neural Networks
Pixel Recurrent Neural Networks
Aaron van den Oord
Nal Kalchbrenner
Koray Kavukcuoglu
SSeg
GAN
251
2,550
0
25 Jan 2016
Previous
123456