ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2412.16378
91
1

REFA: Reference Free Alignment for multi-preference optimization

20 December 2024
Taneesh Gupta
Rahul Madhavan
Xuchao Zhang
Chetan Bansal
Saravan Rajmohan
ArXivPDFHTML
Abstract

We introduce REFA\textbf{REFA}REFA, a family of reference-free alignment methods that optimize over multiple user preferences while enforcing fine-grained length control. Our approach integrates deviation-based weighting to emphasize high-quality responses, length normalization to prevent trivial short-response solutions, and an EOS-probability regularizer to mitigate dataset-induced brevity biases. Theoretically, we show that under the Uncertainty Reduction with Sequence Length Assertion (URSLA) framework, naive length normalization can still incentivize length-based shortcuts. In contrast, REFA corrects these subtle incentives, guiding models toward genuinely more informative and higher-quality outputs. Empirically, REFA achieves a new state-of-the-art\textbf{state-of-the-art}state-of-the-art among reference-free alignment methods, generating richer responses that align more closely with human preferences. Notably, REFA improves performance on the AlpacaEval2 benchmark, achieving a \textbf{26.6%} Length-Controlled Win Rate (LC-WR) and \textbf{24.2%} Win Rate (WR).

View on arXiv
@article{gupta2025_2412.16378,
  title={ REFA: Reference Free Alignment for multi-preference optimization },
  author={ Taneesh Gupta and Rahul Madhavan and Xuchao Zhang and Chetan Bansal and Saravan Rajmohan },
  journal={arXiv preprint arXiv:2412.16378},
  year={ 2025 }
}
Comments on this paper