RATTENTION: Towards the Minimal Sliding Window Size in Local-Global Attention Models

RATTENTION: Towards the Minimal Sliding Window Size in Local-Global Attention Models

Bailin Wang
Chang Lan
Chong Wang
Ruoming Pang

Papers citing "RATTENTION: Towards the Minimal Sliding Window Size in Local-Global Attention Models"

Title
No papers