20
0

Gated Rotary-Enhanced Linear Attention for Long-term Sequential Recommendation

Main:20 Pages
11 Figures
Bibliography:4 Pages
6 Tables
Abstract

In Sequential Recommendation Systems (SRSs), Transformer models show remarkable performance but face computation cost challenges when modeling long-term user behavior sequences due to the quadratic complexity of the dot-product attention mechanism. By approximating the dot-product attention, linear attention provides an efficient option with linear complexity. However, existing linear attention methods face two limitations: 1) they often use learnable position encodings, which incur extra computational costs in long-term sequence scenarios, and 2) they may not consider the user's fine-grained local preferences and confuse these with the actual change of long-term interests. To remedy these drawbacks, we propose a long-term sequential Recommendation model with Gated Rotary Enhanced Linear Attention (RecGRELA). Specifically, we first propose a Rotary-Enhanced Linear Attention (RELA) module to model long-range dependency within the user's historical information using rotary position encodings. We then introduce a local short operation to incorporate local preferences and demonstrate the theoretical insight. We further introduce a SiLU-based Gated mechanism for RELA (GRELA) to help the model determine whether a user's behavior indicates local interest or a genuine shift in long-term preferences. Experimental results on four public datasets demonstrate that our RecGRELA achieves state-of-the-art performance compared to existing SRSs while maintaining low memory overhead.

View on arXiv
@article{hu2025_2506.13315,
  title={ Gated Rotary-Enhanced Linear Attention for Long-term Sequential Recommendation },
  author={ Juntao Hu and Wei Zhou and Huayi Shen and Xiao Du and Jie Liao and Junhao Wen and Min Gao },
  journal={arXiv preprint arXiv:2506.13315},
  year={ 2025 }
}
Comments on this paper