Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2403.08837
Cited By
Cyclic Data Parallelism for Efficient Parallelism of Deep Neural Networks
13 March 2024
Louis Fournier
Edouard Oyallon
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Cyclic Data Parallelism for Efficient Parallelism of Deep Neural Networks"
3 / 3 papers shown
Title
ZeRO++: Extremely Efficient Collective Communication for Giant Model Training
Guanhua Wang
Heyang Qin
S. A. Jacobs
Connor Holmes
Samyam Rajbhandari
Olatunji Ruwase
Feng Yan
Lei Yang
Yuxiong He
VLM
59
57
0
16 Jun 2023
ZeRO-Offload: Democratizing Billion-Scale Model Training
Jie Ren
Samyam Rajbhandari
Reza Yazdani Aminabadi
Olatunji Ruwase
Shuangyang Yang
Minjia Zhang
Dong Li
Yuxiong He
MoE
177
414
0
18 Jan 2021
Megatron-LM: Training Multi-Billion Parameter Language Models Using Model Parallelism
M. Shoeybi
M. Patwary
Raul Puri
P. LeGresley
Jared Casper
Bryan Catanzaro
MoE
245
1,821
0
17 Sep 2019
1