Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2404.05089
Cited By
SEER-MoE: Sparse Expert Efficiency through Regularization for Mixture-of-Experts
7 April 2024
Alexandre Muzio
Alex Sun
Churan He
MoE
Re-assign community
ArXiv
PDF
HTML
Papers citing
"SEER-MoE: Sparse Expert Efficiency through Regularization for Mixture-of-Experts"
3 / 3 papers shown
Title
QoS-Efficient Serving of Multiple Mixture-of-Expert LLMs Using Partial Runtime Reconfiguration
HamidReza Imani
Jiaxin Peng
Peiman Mohseni
Abdolah Amirany
Tarek A. El-Ghazawi
MoE
31
0
0
10 May 2025
Faster MoE LLM Inference for Extremely Large Models
Haoqi Yang
Luohe Shi
Qiwei Li
Zuchao Li
Ping Wang
Bo Du
Mengjia Shen
Hai Zhao
MoE
63
0
0
06 May 2025
Scaling Laws for Neural Language Models
Jared Kaplan
Sam McCandlish
T. Henighan
Tom B. Brown
B. Chess
R. Child
Scott Gray
Alec Radford
Jeff Wu
Dario Amodei
261
4,489
0
23 Jan 2020
1