Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2211.16368
Cited By
DBA: Efficient Transformer with Dynamic Bilinear Low-Rank Attention
24 November 2022
Bosheng Qin
Juncheng Li
Siliang Tang
Yueting Zhuang
Re-assign community
ArXiv
PDF
HTML
Papers citing
"DBA: Efficient Transformer with Dynamic Bilinear Low-Rank Attention"
5 / 5 papers shown
Title
Transformer Quality in Linear Time
Weizhe Hua
Zihang Dai
Hanxiao Liu
Quoc V. Le
78
222
0
21 Feb 2022
Improving Transformers with Probabilistic Attention Keys
Tam Nguyen
T. Nguyen
Dung D. Le
Duy Khuong Nguyen
Viet-Anh Tran
Richard G. Baraniuk
Nhat Ho
Stanley J. Osher
50
32
0
16 Oct 2021
Train Short, Test Long: Attention with Linear Biases Enables Input Length Extrapolation
Ofir Press
Noah A. Smith
M. Lewis
253
695
0
27 Aug 2021
Big Bird: Transformers for Longer Sequences
Manzil Zaheer
Guru Guruganesh
Kumar Avinava Dubey
Joshua Ainslie
Chris Alberti
...
Philip Pham
Anirudh Ravula
Qifan Wang
Li Yang
Amr Ahmed
VLM
282
2,015
0
28 Jul 2020
Efficient Content-Based Sparse Attention with Routing Transformers
Aurko Roy
M. Saffar
Ashish Vaswani
David Grangier
MoE
243
580
0
12 Mar 2020
1