ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2310.12558
  4. Cited By
Large Language Models Help Humans Verify Truthfulness -- Except When
  They Are Convincingly Wrong

Large Language Models Help Humans Verify Truthfulness -- Except When They Are Convincingly Wrong

19 October 2023
Chenglei Si
Navita Goyal
Sherry Tongshuang Wu
Chen Zhao
Shi Feng
Hal Daumé
Jordan L. Boyd-Graber
    LRM
ArXivPDFHTML

Papers citing "Large Language Models Help Humans Verify Truthfulness -- Except When They Are Convincingly Wrong"

27 / 27 papers shown
Title
Improving LLM Personas via Rationalization with Psychological Scaffolds
Improving LLM Personas via Rationalization with Psychological Scaffolds
Brihi Joshi
Xiang Ren
Swabha Swayamdipta
Rik Koncel-Kedziorski
Tim Paek
73
0
0
25 Apr 2025
Fostering Appropriate Reliance on Large Language Models: The Role of Explanations, Sources, and Inconsistencies
Fostering Appropriate Reliance on Large Language Models: The Role of Explanations, Sources, and Inconsistencies
Sunnie S. Y. Kim
J. Vaughan
Q. V. Liao
Tania Lombrozo
Olga Russakovsky
102
5
0
12 Feb 2025
Plan-Then-Execute: An Empirical Study of User Trust and Team Performance When Using LLM Agents As A Daily Assistant
Plan-Then-Execute: An Empirical Study of User Trust and Team Performance When Using LLM Agents As A Daily Assistant
Gaole He
Gianluca Demartini
U. Gadiraju
LLMAG
65
8
0
03 Feb 2025
Personalized Help for Optimizing Low-Skilled Users' Strategy
Personalized Help for Optimizing Low-Skilled Users' Strategy
Feng Gu
Wichayaporn Wongkamjan
Jordan Lee Boyd-Graber
Jonathan K. Kummerfeld
Denis Peskoff
Jonathan May
31
0
0
14 Nov 2024
Contrastive Explanations That Anticipate Human Misconceptions Can Improve Human Decision-Making Skills
Contrastive Explanations That Anticipate Human Misconceptions Can Improve Human Decision-Making Skills
Zana Buçinca
S. Swaroop
Amanda E. Paluch
Finale Doshi-Velez
Krzysztof Z. Gajos
48
2
0
05 Oct 2024
Wait, but Tylenol is Acetaminophen... Investigating and Improving
  Language Models' Ability to Resist Requests for Misinformation
Wait, but Tylenol is Acetaminophen... Investigating and Improving Language Models' Ability to Resist Requests for Misinformation
Shan Chen
Mingye Gao
Kuleen Sasse
Thomas Hartvigsen
Brian Anthony
Lizhou Fan
Hugo J. W. L. Aerts
Jack Gallifant
Danielle S. Bitterman
LM&MA
33
0
0
30 Sep 2024
Claim Verification in the Age of Large Language Models: A Survey
Claim Verification in the Age of Large Language Models: A Survey
A. Dmonte
Roland Oruche
Marcos Zampieri
Prasad Calyam
Isabelle Augenstein
44
8
0
26 Aug 2024
CHECKWHY: Causal Fact Verification via Argument Structure
CHECKWHY: Causal Fact Verification via Argument Structure
Jiasheng Si
Yibo Zhao
Yingjie Zhu
Haiyang Zhu
Wenpeng Lu
Deyu Zhou
CML
HILM
LRM
40
1
0
20 Aug 2024
STORYSUMM: Evaluating Faithfulness in Story Summarization
STORYSUMM: Evaluating Faithfulness in Story Summarization
Melanie Subbiah
Faisal Ladhak
Akankshya Mishra
Griffin Adams
Lydia B. Chilton
Kathleen McKeown
48
4
0
09 Jul 2024
One Thousand and One Pairs: A "novel" challenge for long-context
  language models
One Thousand and One Pairs: A "novel" challenge for long-context language models
Marzena Karpinska
Katherine Thai
Kyle Lo
Tanya Goyal
Mohit Iyyer
LRM
41
40
0
24 Jun 2024
Persuasiveness of Generated Free-Text Rationales in Subjective
  Decisions: A Case Study on Pairwise Argument Ranking
Persuasiveness of Generated Free-Text Rationales in Subjective Decisions: A Case Study on Pairwise Argument Ranking
Mohamed S. Elaraby
Diane Litman
Xiang Lorraine Li
Ahmed Magooda
LRM
34
2
0
20 Jun 2024
Learning to Generate Answers with Citations via Factual Consistency
  Models
Learning to Generate Answers with Citations via Factual Consistency Models
Rami Aly
Zhiqiang Tang
Samson Tan
George Karypis
HILM
34
4
0
19 Jun 2024
WildVision: Evaluating Vision-Language Models in the Wild with Human
  Preferences
WildVision: Evaluating Vision-Language Models in the Wild with Human Preferences
Yujie Lu
Dongfu Jiang
Wenhu Chen
William Yang Wang
Yejin Choi
Bill Yuchen Lin
VLM
45
26
0
16 Jun 2024
Scenarios and Approaches for Situated Natural Language Explanations
Scenarios and Approaches for Situated Natural Language Explanations
Pengshuo Qiu
Frank Rudzicz
Zining Zhu
LRM
38
0
0
07 Jun 2024
More Victories, Less Cooperation: Assessing Cicero's Diplomacy Play
More Victories, Less Cooperation: Assessing Cicero's Diplomacy Play
Wichayaporn Wongkamjan
Feng Gu
Yanze Wang
Ulf Hermjakob
Jonathan May
Brandon M. Stewart
Jonathan K. Kummerfeld
Denis Peskoff
Jordan L. Boyd-Graber
45
3
0
07 Jun 2024
LLM-Generated Black-box Explanations Can Be Adversarially Helpful
LLM-Generated Black-box Explanations Can Be Adversarially Helpful
R. Ajwani
Shashidhar Reddy Javaji
Frank Rudzicz
Zining Zhu
AAML
40
6
0
10 May 2024
Claim Check-Worthiness Detection: How Well do LLMs Grasp Annotation
  Guidelines?
Claim Check-Worthiness Detection: How Well do LLMs Grasp Annotation Guidelines?
Laura Majer
Jan Snajder
39
3
0
18 Apr 2024
FOFO: A Benchmark to Evaluate LLMs' Format-Following Capability
FOFO: A Benchmark to Evaluate LLMs' Format-Following Capability
Congying Xia
Chen Xing
Jiangshu Du
Xinyi Yang
Yihao Feng
Ran Xu
Wenpeng Yin
Caiming Xiong
ALM
30
41
0
28 Feb 2024
LLMAuditor: A Framework for Auditing Large Language Models Using
  Human-in-the-Loop
LLMAuditor: A Framework for Auditing Large Language Models Using Human-in-the-Loop
Maryam Amirizaniani
Jihan Yao
Adrian Lavergne
Elizabeth Snell Okada
Aman Chadha
Tanya Roosta
Chirag Shah
HILM
34
2
0
14 Feb 2024
Faithfulness vs. Plausibility: On the (Un)Reliability of Explanations
  from Large Language Models
Faithfulness vs. Plausibility: On the (Un)Reliability of Explanations from Large Language Models
Chirag Agarwal
Sree Harsha Tanneru
Himabindu Lakkaraju
LRM
39
35
0
07 Feb 2024
A Study on the Calibration of In-context Learning
A Study on the Calibration of In-context Learning
Hanlin Zhang
Yi-Fan Zhang
Yaodong Yu
Dhruv Madeka
Dean Phillips Foster
Eric Xing
Hima Lakkaraju
Sham Kakade
26
15
0
07 Dec 2023
Conceptual Engineering Using Large Language Models
Conceptual Engineering Using Large Language Models
Bradley Paul Allen
29
0
0
01 Dec 2023
Human Feedback is not Gold Standard
Human Feedback is not Gold Standard
Tom Hosking
Phil Blunsom
Max Bartolo
ALM
26
49
0
28 Sep 2023
Getting MoRE out of Mixture of Language Model Reasoning Experts
Getting MoRE out of Mixture of Language Model Reasoning Experts
Chenglei Si
Weijia Shi
Chen Zhao
Luke Zettlemoyer
Jordan L. Boyd-Graber
LRM
24
24
0
24 May 2023
On the Risk of Misinformation Pollution with Large Language Models
On the Risk of Misinformation Pollution with Large Language Models
Yikang Pan
Liangming Pan
Wenhu Chen
Preslav Nakov
Min-Yen Kan
Luu Anh Tuan
DeLMO
192
110
0
23 May 2023
Re-Examining Calibration: The Case of Question Answering
Re-Examining Calibration: The Case of Question Answering
Chenglei Si
Chen Zhao
Sewon Min
Jordan L. Boyd-Graber
61
30
0
25 May 2022
Human Interpretation of Saliency-based Explanation Over Text
Human Interpretation of Saliency-based Explanation Over Text
Hendrik Schuff
Alon Jacovi
Heike Adel
Yoav Goldberg
Ngoc Thang Vu
MILM
XAI
FAtt
144
39
0
27 Jan 2022
1