ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2504.14960
  4. Cited By
MoE Parallel Folding: Heterogeneous Parallelism Mappings for Efficient Large-Scale MoE Model Training with Megatron Core

MoE Parallel Folding: Heterogeneous Parallelism Mappings for Efficient Large-Scale MoE Model Training with Megatron Core

21 April 2025
Dennis Liu
Zijie Yan
Xin Yao
Tong Liu
V. Korthikanti
Evan Wu
Shiqing Fan
Gao Deng
Hongxiao Bai
Jianbin Chang
Ashwath Aithal
M. Andersch
M. Shoeybi
Jiajie Yao
Chandler Zhou
David Wu
Xipeng Li
J. Yang
    MoE
ArXivPDFHTML

Papers citing "MoE Parallel Folding: Heterogeneous Parallelism Mappings for Efficient Large-Scale MoE Model Training with Megatron Core"

Title
No papers