ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2502.05633
50
0

Mol-MoE: Training Preference-Guided Routers for Molecule Generation

8 February 2025
Diego Calanzone
P. DÓro
Pierre-Luc Bacon
ArXivPDFHTML
Abstract

Recent advances in language models have enabled framing molecule generation as sequence modeling. However, existing approaches often rely on single-objective reinforcement learning, limiting their applicability to real-world drug design, where multiple competing properties must be optimized. Traditional multi-objective reinforcement learning (MORL) methods require costly retraining for each new objective combination, making rapid exploration of trade-offs impractical. To overcome these limitations, we introduce Mol-MoE, a mixture-of-experts (MoE) architecture that enables efficient test-time steering of molecule generation without retraining. Central to our approach is a preference-based router training objective that incentivizes the router to combine experts in a way that aligns with user-specified trade-offs. This provides improved flexibility in exploring the chemical property space at test time, facilitating rapid trade-off exploration. Benchmarking against state-of-the-art methods, we show that Mol-MoE achieves superior sample quality and steerability.

View on arXiv
@article{calanzone2025_2502.05633,
  title={ Mol-MoE: Training Preference-Guided Routers for Molecule Generation },
  author={ Diego Calanzone and Pierluca DÓro and Pierre-Luc Bacon },
  journal={arXiv preprint arXiv:2502.05633},
  year={ 2025 }
}
Comments on this paper