ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2410.21533
145
0

L3Ms -- Lagrange Large Language Models

28 October 2024
Guneet S. Dhillon
Xingjian Shi
Yee Whye Teh
Alex Smola
ArXivPDFHTML
Abstract

Supervised fine-tuning (SFT) and alignment of large language models (LLMs) are key steps in providing a good user experience. However, the concept of an appropriate alignment is inherently application-dependent, and current methods often rely on heuristic choices to drive optimization. In this work, we formulate SFT and alignment as a constrained optimization problem: the LLM is fine-tuned on a task while being required to meet application-specific requirements, without resorting to heuristics. To solve this, we propose Lagrange Large Language Models (L3Ms), which employ logarithmic barriers to enforce the constraints. This approach allows for the customization of L3Ms across diverse applications while avoiding heuristic-driven processes. We experimentally demonstrate the versatility and efficacy of L3Ms in achieving tailored alignments for various applications.

View on arXiv
@article{dhillon2025_2410.21533,
  title={ L3Ms -- Lagrange Large Language Models },
  author={ Guneet S. Dhillon and Xingjian Shi and Yee Whye Teh and Alex Smola },
  journal={arXiv preprint arXiv:2410.21533},
  year={ 2025 }
}
Comments on this paper