ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2506.14912
5
0

CrEst: Credibility Estimation for Contexts in LLMs via Weak Supervision

17 June 2025
Dyah Adila
Shuai Zhang
Boran Han
Bonan Min
Yuyang Wang
ArXiv (abs)PDFHTML
Main:7 Pages
8 Figures
Bibliography:3 Pages
3 Tables
Appendix:1 Pages
Abstract

The integration of contextual information has significantly enhanced the performance of large language models (LLMs) on knowledge-intensive tasks. However, existing methods often overlook a critical challenge: the credibility of context documents can vary widely, potentially leading to the propagation of unreliable information. In this paper, we introduce CrEst, a novel weakly supervised framework for assessing the credibility of context documents during LLM inference--without requiring manual annotations. Our approach is grounded in the insight that credible documents tend to exhibit higher semantic coherence with other credible documents, enabling automated credibility estimation through inter-document agreement. To incorporate credibility into LLM inference, we propose two integration strategies: a black-box approach for models without access to internal weights or activations, and a white-box method that directly modifies attention mechanisms. Extensive experiments across three model architectures and five datasets demonstrate that CrEst consistently outperforms strong baselines, achieving up to a 26.86% improvement in accuracy and a 3.49% increase in F1 score. Further analysis shows that CrEst maintains robust performance even under high-noise conditions.

View on arXiv
@article{adila2025_2506.14912,
  title={ CrEst: Credibility Estimation for Contexts in LLMs via Weak Supervision },
  author={ Dyah Adila and Shuai Zhang and Boran Han and Bonan Min and Yuyang Wang },
  journal={arXiv preprint arXiv:2506.14912},
  year={ 2025 }
}
Comments on this paper