SGDPO: Self-Guided Direct Preference Optimization for Language Model Alignment

Direct Preference Optimization (DPO) is broadly utilized for aligning Large Language Models (LLMs) with human values because of its flexibility. Despite its effectiveness, it has been observed that the capability of DPO to generate human-preferred response is limited and the results of DPO are far from resilient. To address these limitations, in this paper we propose a novel Self-Guided Direct Preference Optimization algorithm, i.e., SGDPO, which incorporates a pilot term to steer the gradient flow during the optimization process, allowing for fine-grained control over the updates of chosen and rejected rewards. We provide a detailed theoretical analysis of our proposed method and elucidate its operational mechanism. Furthermore, we conduct comprehensive experiments on various models and benchmarks. The extensive experimental results demonstrate the consistency between the empirical results and our theoretical analysis and confirm the effectiveness of our proposed approach (up to 9.19% higher score).
View on arXiv@article{zhu2025_2505.12435, title={ SGDPO: Self-Guided Direct Preference Optimization for Language Model Alignment }, author={ Wenqiao Zhu and Ji Liu and Lulu Wang and Jun Wu and Yulun Zhang }, journal={arXiv preprint arXiv:2505.12435}, year={ 2025 } }