ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2506.07309
13
0

ConfQA: Answer Only If You Are Confident

8 June 2025
Yin Huang
Yifan Ethan Xu
Kai Sun
Vera Yan
Alicia Sun
Haidar Khan
Jimmy Nguyen
Mohammad Kachuee
Zhaojiang Lin
Yue Liu
Aaron Colak
Anuj Kumar
Wen-tau Yih
Xin Luna Dong
    HILM
ArXiv (abs)PDFHTML
Main:15 Pages
4 Figures
Bibliography:1 Pages
8 Tables
Appendix:4 Pages
Abstract

Can we teach Large Language Models (LLMs) to refrain from hallucinating factual statements? In this paper we present a fine-tuning strategy that we call ConfQA, which can reduce hallucination rate from 20-40% to under 5% across multiple factuality benchmarks. The core idea is simple: when the LLM answers a question correctly, it is trained to continue with the answer; otherwise, it is trained to admit "I am unsure". But there are two key factors that make the training highly effective. First, we introduce a dampening prompt "answer only if you are confident" to explicitly guide the behavior, without which hallucination remains high as 15%-25%. Second, we leverage simple factual statements, specifically attribute values from knowledge graphs, to help LLMs calibrate the confidence, resulting in robust generalization across domains and question types. Building on this insight, we propose the Dual Neural Knowledge framework, which seamlessly select between internally parameterized neural knowledge and externally recorded symbolic knowledge based on ConfQA's confidence. The framework enables potential accuracy gains to beyond 95%, while reducing unnecessary external retrievals by over 30%.

View on arXiv
@article{huang2025_2506.07309,
  title={ ConfQA: Answer Only If You Are Confident },
  author={ Yin Huang and Yifan Ethan Xu and Kai Sun and Vera Yan and Alicia Sun and Haidar Khan and Jimmy Nguyen and Mohammad Kachuee and Zhaojiang Lin and Yue Liu and Aaron Colak and Anuj Kumar and Wen-tau Yih and Xin Luna Dong },
  journal={arXiv preprint arXiv:2506.07309},
  year={ 2025 }
}
Comments on this paper