ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2404.00672
  4. Cited By
A General and Efficient Training for Transformer via Token Expansion

A General and Efficient Training for Transformer via Token Expansion

31 March 2024
Wenxuan Huang
Yunhang Shen
Jiao Xie
Baochang Zhang
Gaoqi He
Ke Li
Xing Sun
Shaohui Lin
ArXivPDFHTML

Papers citing "A General and Efficient Training for Transformer via Token Expansion"

2 / 2 papers shown
Title
Efficient On-device Training via Gradient Filtering
Efficient On-device Training via Gradient Filtering
Yuedong Yang
Guihong Li
R. Marculescu
39
18
0
01 Jan 2023
Deduplicating Training Data Makes Language Models Better
Deduplicating Training Data Makes Language Models Better
Katherine Lee
Daphne Ippolito
A. Nystrom
Chiyuan Zhang
Douglas Eck
Chris Callison-Burch
Nicholas Carlini
SyDa
242
595
0
14 Jul 2021
1