Hydra Attention: Efficient Attention with Many Heads

Hydra Attention: Efficient Attention with Many Heads

Papers citing "Hydra Attention: Efficient Attention with Many Heads"

15 / 15 papers shown
Title
Linformer: Self-Attention with Linear Complexity
Linformer: Self-Attention with Linear Complexity
Sinong Wang
Belinda Z. Li
Madian Khabsa
Han Fang
Hao Ma
179
1,678
0
08 Jun 2020