ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2506.14646
5
0

GuiLoMo: Allocating Expert Number and Rank for LoRA-MoE via Bilevel Optimization with GuidedSelection Vectors

17 June 2025
Hengyuan Zhang
Xinrong Chen
Yingmin Qiu
Xiao Liang
Ziyue Li
Guanyu Wang
Weiping Li
Tong Mo
Wenyue Li
Hayden Kwok-Hay So
Ngai Wong
    MoEALM
ArXiv (abs)PDFHTML
Main:7 Pages
8 Figures
Bibliography:6 Pages
9 Tables
Appendix:3 Pages
Abstract

Parameter-efficient fine-tuning (PEFT) methods, particularly Low-Rank Adaptation (LoRA), offer an efficient way to adapt large language models with reduced computational costs. However, their performance is limited by the small number of trainable parameters. Recent work combines LoRA with the Mixture-of-Experts (MoE), i.e., LoRA-MoE, to enhance capacity, but two limitations remain in hindering the full exploitation of its potential: 1) the influence of downstream tasks when assigning expert numbers, and 2) the uniform rank assignment across all LoRA experts, which restricts representational diversity. To mitigate these gaps, we propose GuiLoMo, a fine-grained layer-wise expert numbers and ranks allocation strategy with GuidedSelection Vectors (GSVs). GSVs are learned via a prior bilevel optimization process to capture both model- and task-specific needs, and are then used to allocate optimal expert numbers and ranks. Experiments on three backbone models across diverse benchmarks show that GuiLoMo consistently achieves superior or comparable performance to all baselines. Further analysis offers key insights into how expert numbers and ranks vary across layers and tasks, highlighting the benefits of adaptive expert configuration. Our code is available atthis https URL.

View on arXiv
@article{zhang2025_2506.14646,
  title={ GuiLoMo: Allocating Expert Number and Rank for LoRA-MoE via Bilevel Optimization with GuidedSelection Vectors },
  author={ Hengyuan Zhang and Xinrong Chen and Yingmin Qiu and Xiao Liang and Ziyue Li and Guanyu Wang and Weiping Li and Tong Mo and Wenyue Li and Hayden Kwok-Hay So and Ngai Wong },
  journal={arXiv preprint arXiv:2506.14646},
  year={ 2025 }
}
Comments on this paper