Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2406.18485
Cited By
LoongTrain: Efficient Training of Long-Sequence LLMs with Head-Context Parallelism
26 June 2024
Diandian Gu
Peng Sun
Qinghao Hu
Ting Huang
Xun Chen
Yingtong Xiong
Guoteng Wang
Qiaoling Chen
Shangchun Zhao
Jiarui Fang
Yonggang Wen
Tianwei Zhang
Xin Jin
Xuanzhe Liu
LRM
Re-assign community
ArXiv
PDF
HTML
Papers citing
"LoongTrain: Efficient Training of Long-Sequence LLMs with Head-Context Parallelism"
3 / 3 papers shown
Title
Varuna: Scalable, Low-cost Training of Massive Deep Learning Models
Sanjith Athlur
Nitika Saran
Muthian Sivathanu
Ramachandran Ramjee
Nipun Kwatra
GNN
31
80
0
07 Nov 2021
Survey: Transformer based Video-Language Pre-training
Ludan Ruan
Qin Jin
VLM
ViT
72
44
0
21 Sep 2021
ZeRO-Offload: Democratizing Billion-Scale Model Training
Jie Ren
Samyam Rajbhandari
Reza Yazdani Aminabadi
Olatunji Ruwase
Shuangyang Yang
Minjia Zhang
Dong Li
Yuxiong He
MoE
177
414
0
18 Jan 2021
1