ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2311.00684
  4. Cited By
Attention Alignment and Flexible Positional Embeddings Improve
  Transformer Length Extrapolation

Attention Alignment and Flexible Positional Embeddings Improve Transformer Length Extrapolation

1 November 2023
Ta-Chung Chi
Ting-Han Fan
Alexander I. Rudnicky
ArXivPDFHTML

Papers citing "Attention Alignment and Flexible Positional Embeddings Improve Transformer Length Extrapolation"

2 / 2 papers shown
Title
GLM-130B: An Open Bilingual Pre-trained Model
GLM-130B: An Open Bilingual Pre-trained Model
Aohan Zeng
Xiao Liu
Zhengxiao Du
Zihan Wang
Hanyu Lai
...
Jidong Zhai
Wenguang Chen
Peng Zhang
Yuxiao Dong
Jie Tang
BDL
LRM
275
1,077
0
05 Oct 2022
Train Short, Test Long: Attention with Linear Biases Enables Input
  Length Extrapolation
Train Short, Test Long: Attention with Linear Biases Enables Input Length Extrapolation
Ofir Press
Noah A. Smith
M. Lewis
253
710
0
27 Aug 2021
1