ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2503.00355
45
1

Structured Reasoning for Fairness: A Multi-Agent Approach to Bias Detection in Textual Data

1 March 2025
Tianyi Huang
Elsa Fan
ArXivPDFHTML
Abstract

From disinformation spread by AI chatbots to AI recommendations that inadvertently reinforce stereotypes, textual bias poses a significant challenge to the trustworthiness of large language models (LLMs). In this paper, we propose a multi-agent framework that systematically identifies biases by disentangling each statement as fact or opinion, assigning a bias intensity score, and providing concise, factual justifications. Evaluated on 1,500 samples from the WikiNPOV dataset, the framework achieves 84.9% accuracy\unicodex2014\unicode{x2014}\unicodex2014an improvement of 13.0% over the zero-shot baseline\unicodex2014\unicode{x2014}\unicodex2014demonstrating the efficacy of explicitly modeling fact versus opinion prior to quantifying bias intensity. By combining enhanced detection accuracy with interpretable explanations, this approach sets a foundation for promoting fairness and accountability in modern language models.

View on arXiv
@article{huang2025_2503.00355,
  title={ Structured Reasoning for Fairness: A Multi-Agent Approach to Bias Detection in Textual Data },
  author={ Tianyi Huang and Elsa Fan },
  journal={arXiv preprint arXiv:2503.00355},
  year={ 2025 }
}
Comments on this paper