ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2409.18786
27
4

A Survey on the Honesty of Large Language Models

27 September 2024
Siheng Li
Cheng Yang
Taiqiang Wu
Chufan Shi
Yuji Zhang
Xinyu Zhu
Zesen Cheng
Deng Cai
Mo Yu
Lemao Liu
Jie Zhou
Yujiu Yang
Ngai Wong
Xixin Wu
Wai Lam
    HILM
ArXivPDFHTML
Abstract

Honesty is a fundamental principle for aligning large language models (LLMs) with human values, requiring these models to recognize what they know and don't know and be able to faithfully express their knowledge. Despite promising, current LLMs still exhibit significant dishonest behaviors, such as confidently presenting wrong answers or failing to express what they know. In addition, research on the honesty of LLMs also faces challenges, including varying definitions of honesty, difficulties in distinguishing between known and unknown knowledge, and a lack of comprehensive understanding of related research. To address these issues, we provide a survey on the honesty of LLMs, covering its clarification, evaluation approaches, and strategies for improvement. Moreover, we offer insights for future research, aiming to inspire further exploration in this important area.

View on arXiv
Comments on this paper