ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2505.19176
64
0
v1v2 (latest)

Assistant-Guided Mitigation of Teacher Preference Bias in LLM-as-a-Judge

25 May 2025
Zhuo Liu
Moxin Li
Xun Deng
Qifan Wang
Fuli Feng
    ELM
ArXiv (abs)PDFHTML
Main:8 Pages
7 Figures
Bibliography:2 Pages
15 Tables
Appendix:10 Pages
Abstract

LLM-as-a-Judge employs large language models (LLMs), such as GPT-4, to evaluate the quality of LLM-generated responses, gaining popularity for its cost-effectiveness and strong alignment with human evaluations. However, training proxy judge models using evaluation data generated by powerful teacher models introduces a critical yet previously overlooked issue: teacher preference bias, where the proxy judge model learns a biased preference for responses from the teacher model. To tackle this problem, we propose a novel setting that incorporates an additional assistant model, which is not biased toward the teacher model's responses, to complement the training data. Building on this setup, we introduce AGDe-Judge, a three-stage framework designed to debias from both the labels and feedbacks in the training data. Extensive experiments demonstrate that AGDe-Judge effectively reduces teacher preference bias while maintaining strong performance across six evaluation benchmarks. Code is available at this https URL.

View on arXiv
@article{liu2025_2505.19176,
  title={ Assistant-Guided Mitigation of Teacher Preference Bias in LLM-as-a-Judge },
  author={ Zhuo Liu and Moxin Li and Xun Deng and Qifan Wang and Fuli Feng },
  journal={arXiv preprint arXiv:2505.19176},
  year={ 2025 }
}
Comments on this paper