ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2406.18485
  4. Cited By
LoongTrain: Efficient Training of Long-Sequence LLMs with Head-Context
  Parallelism

LoongTrain: Efficient Training of Long-Sequence LLMs with Head-Context Parallelism

26 June 2024
Diandian Gu
Peng Sun
Qinghao Hu
Ting Huang
Xun Chen
Yingtong Xiong
Guoteng Wang
Qiaoling Chen
Shangchun Zhao
Jiarui Fang
Yonggang Wen
Tianwei Zhang
Xin Jin
Xuanzhe Liu
    LRM
ArXivPDFHTML

Papers citing "LoongTrain: Efficient Training of Long-Sequence LLMs with Head-Context Parallelism"

3 / 3 papers shown
Title
Varuna: Scalable, Low-cost Training of Massive Deep Learning Models
Varuna: Scalable, Low-cost Training of Massive Deep Learning Models
Sanjith Athlur
Nitika Saran
Muthian Sivathanu
Ramachandran Ramjee
Nipun Kwatra
GNN
33
80
0
07 Nov 2021
Survey: Transformer based Video-Language Pre-training
Survey: Transformer based Video-Language Pre-training
Ludan Ruan
Qin Jin
VLM
ViT
72
44
0
21 Sep 2021
ZeRO-Offload: Democratizing Billion-Scale Model Training
ZeRO-Offload: Democratizing Billion-Scale Model Training
Jie Ren
Samyam Rajbhandari
Reza Yazdani Aminabadi
Olatunji Ruwase
Shuangyang Yang
Minjia Zhang
Dong Li
Yuxiong He
MoE
177
416
0
18 Jan 2021
1