ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2501.19399
  4. Cited By
Scalable-Softmax Is Superior for Attention

Scalable-Softmax Is Superior for Attention

31 January 2025
Ken M. Nakanishi
ArXiv (abs)PDFHTML

Papers citing "Scalable-Softmax Is Superior for Attention"

3 / 3 papers shown
Title
Scale-invariant Attention
Scale-invariant Attention
Ben Anson
Xi Wang
Laurence Aitchison
LRM
105
0
0
20 May 2025
Continuity and Isolation Lead to Doubts or Dilemmas in Large Language Models
Continuity and Isolation Lead to Doubts or Dilemmas in Large Language Models
Hector Pasten
Felipe Urrutia
Hector Jimenez
Cristian B. Calderon
Cristóbal Rojas
Alexander Kozachinskiy
117
0
0
15 May 2025
Multi-Token Attention
Multi-Token Attention
O. Yu. Golovneva
Tianlu Wang
Jason Weston
Sainbayar Sukhbaatar
89
1
0
01 Apr 2025
1