ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2505.22296
  4. Cited By
360-LLaMA-Factory: Plug & Play Sequence Parallelism for Long Post-Training

360-LLaMA-Factory: Plug & Play Sequence Parallelism for Long Post-Training

28 May 2025
Haosheng Zou
Xiaowei Lv
Shousheng Jia
Xiangzheng Zhang
    SyDa
    LRM
ArXivPDFHTML

Papers citing "360-LLaMA-Factory: Plug & Play Sequence Parallelism for Long Post-Training"

4 / 4 papers shown
Title
SWIFT:A Scalable lightWeight Infrastructure for Fine-Tuning
SWIFT:A Scalable lightWeight Infrastructure for Fine-Tuning
Yuze Zhao
Jintao Huang
Jinghan Hu
Xingjun Wang
Yunlin Mao
...
Zhikai Wu
Baole Ai
Ang Wang
Wenmeng Zhou
Yingda Chen
55
36
0
10 Aug 2024
USP: A Unified Sequence Parallelism Approach for Long Context Generative
  AI
USP: A Unified Sequence Parallelism Approach for Long Context Generative AI
Jiarui Fang
Shangchun Zhao
40
17
0
13 May 2024
RoFormer: Enhanced Transformer with Rotary Position Embedding
RoFormer: Enhanced Transformer with Rotary Position Embedding
Jianlin Su
Yu Lu
Shengfeng Pan
Ahmed Murtadha
Bo Wen
Yunfeng Liu
70
2,307
0
20 Apr 2021
Megatron-LM: Training Multi-Billion Parameter Language Models Using
  Model Parallelism
Megatron-LM: Training Multi-Billion Parameter Language Models Using Model Parallelism
Mohammad Shoeybi
M. Patwary
Raul Puri
P. LeGresley
Jared Casper
Bryan Catanzaro
MoE
268
1,861
0
17 Sep 2019
1