Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2305.14552
Cited By
Sources of Hallucination by Large Language Models on Inference Tasks
23 May 2023
Nick McKenna
Tianyi Li
Liang Cheng
Mohammad Javad Hosseini
Mark Johnson
Mark Steedman
LRM
HILM
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Sources of Hallucination by Large Language Models on Inference Tasks"
50 / 119 papers shown
Title
MiniCheck: Efficient Fact-Checking of LLMs on Grounding Documents
Liyan Tang
Philippe Laban
Greg Durrett
HILM
SyDa
43
76
0
16 Apr 2024
Improving the Capabilities of Large Language Model Based Marketing Analytics Copilots With Semantic Search And Fine-Tuning
Yilin Gao
Arava Sai Kumar
Yancheng Li
James W. Snyder
AI4MH
40
2
0
16 Apr 2024
RLHF Deciphered: A Critical Analysis of Reinforcement Learning from Human Feedback for LLMs
Shreyas Chaudhari
Pranjal Aggarwal
Vishvak Murahari
Tanmay Rajpurohit
Ashwin Kalyan
Karthik Narasimhan
Ameet Deshpande
Bruno Castro da Silva
29
34
0
12 Apr 2024
Know When To Stop: A Study of Semantic Drift in Text Generation
Ava Spataru
Eric Hambro
Elena Voita
Nicola Cancedda
37
3
0
08 Apr 2024
Humanizing Machine-Generated Content: Evading AI-Text Detection through Adversarial Attack
Ying Zhou
Xianpei Han
Le Sun
DeLMO
AAML
38
8
0
02 Apr 2024
Truth-Aware Context Selection: Mitigating Hallucinations of Large Language Models Being Misled by Untruthful Contexts
Tian Yu
Shaolei Zhang
Yang Feng
HILM
36
7
0
12 Mar 2024
Tuning-Free Accountable Intervention for LLM Deployment -- A Metacognitive Approach
Zhen Tan
Jie Peng
Tianlong Chen
Huan Liu
37
6
0
08 Mar 2024
RAGged Edges: The Double-Edged Sword of Retrieval-Augmented Chatbots
Philip G. Feldman
James R. Foulds
Shimei Pan
SILM
35
13
0
02 Mar 2024
Navigating Hallucinations for Reasoning of Unintentional Activities
Shresth Grover
Vibhav Vineet
Yogesh S Rawat
LRM
52
1
0
29 Feb 2024
Imagine, Initialize, and Explore: An Effective Exploration Method in Multi-Agent Reinforcement Learning
Zeyang Liu
Lipeng Wan
Xinrui Yang
Zhuoran Chen
Xingyu Chen
Xuguang Lan
23
3
0
28 Feb 2024
Fine-Grained Natural Language Inference Based Faithfulness Evaluation for Diverse Summarisation Tasks
Huajian Zhang
Yumo Xu
Laura Perez-Beltrachini
HILM
32
9
0
27 Feb 2024
Large Language Models: A Survey
Shervin Minaee
Tomáš Mikolov
Narjes Nikzad
M. Asgari-Chenaghlu
R. Socher
Xavier Amatriain
Jianfeng Gao
ALM
LM&MA
ELM
134
371
0
09 Feb 2024
Rethinking Interpretability in the Era of Large Language Models
Chandan Singh
J. Inala
Michel Galley
Rich Caruana
Jianfeng Gao
LRM
AI4CE
77
62
0
30 Jan 2024
Learning to Trust Your Feelings: Leveraging Self-awareness in LLMs for Hallucination Mitigation
Yuxin Liang
Zhuoyang Song
Hao Wang
Jiaxing Zhang
HILM
40
30
0
27 Jan 2024
Can Large Language Model Summarizers Adapt to Diverse Scientific Communication Goals?
Marcio Fonseca
Shay B. Cohen
39
10
0
18 Jan 2024
Risk Taxonomy, Mitigation, and Assessment Benchmarks of Large Language Model Systems
Tianyu Cui
Yanling Wang
Chuanpu Fu
Yong Xiao
Sijia Li
...
Junwu Xiong
Xinyu Kong
Zujie Wen
Ke Xu
Qi Li
60
56
0
11 Jan 2024
keqing: knowledge-based question answering is a nature chain-of-thought mentor of LLM
Chaojie Wang
Yishi Xu
Zhong Peng
Chenxi Zhang
Bo Chen
Xinrun Wang
Lei Feng
Bo An
74
18
0
31 Dec 2023
LLM-Assist: Enhancing Closed-Loop Planning with Language-Based Reasoning
S. P. Sharan
Francesco Pittaluga
G. VijayKumarB.
Manmohan Chandraker
LRM
30
52
0
30 Dec 2023
Alleviating Hallucinations of Large Language Models through Induced Hallucinations
Yue Zhang
Leyang Cui
Wei Bi
Shuming Shi
HILM
42
50
0
25 Dec 2023
Artificial Neural Nets and the Representation of Human Concepts
Timo Freiesleben
NAI
24
1
0
08 Dec 2023
Understanding Your Agent: Leveraging Large Language Models for Behavior Explanation
Xijia Zhang
Yue (Sophie) Guo
Simon Stepputtis
Katia P. Sycara
Joseph Campbell
LLMAG
LM&Ro
28
1
0
29 Nov 2023
Clinical Risk Prediction Using Language Models: Benefits And Considerations
Angeela Acharya
Sulabh Shrestha
Anyi Chen
Joseph Conte
Sanja Avramovic
S. Sikdar
Antonios Anastasopoulos
Sanmay Das
26
6
0
29 Nov 2023
Beyond Hallucinations: Enhancing LVLMs through Hallucination-Aware Direct Preference Optimization
Zhiyuan Zhao
Bin Wang
Linke Ouyang
Xiao-wen Dong
Jiaqi Wang
Conghui He
MLLM
VLM
32
106
0
28 Nov 2023
On the Calibration of Large Language Models and Alignment
Chiwei Zhu
Benfeng Xu
Quan Wang
Yongdong Zhang
Zhendong Mao
77
32
0
22 Nov 2023
Examining LLMs' Uncertainty Expression Towards Questions Outside Parametric Knowledge
Genglin Liu
Xingyao Wang
Lifan Yuan
Yangyi Chen
Hao Peng
29
16
0
16 Nov 2023
Does Pre-trained Language Model Actually Infer Unseen Links in Knowledge Graph Completion?
Yusuke Sakai
Hidetaka Kamigaito
Katsuhiko Hayashi
Taro Watanabe
26
1
0
15 Nov 2023
How are Prompts Different in Terms of Sensitivity?
Sheng Lu
Hendrik Schuff
Iryna Gurevych
40
18
0
13 Nov 2023
A Survey on Hallucination in Large Language Models: Principles, Taxonomy, Challenges, and Open Questions
Lei Huang
Weijiang Yu
Weitao Ma
Weihong Zhong
Zhangyin Feng
...
Qianglong Chen
Weihua Peng
Xiaocheng Feng
Bing Qin
Ting Liu
LRM
HILM
44
732
0
09 Nov 2023
Beyond Words: A Mathematical Framework for Interpreting Large Language Models
Javier González
Aditya V. Nori
11
1
0
06 Nov 2023
MILL: Mutual Verification with Large Language Models for Zero-Shot Query Expansion
Pengyue Jia
Yiding Liu
Xiangyu Zhao
Xiaopeng Li
Changying Hao
Shuaiqiang Wang
Dawei Yin
29
16
0
29 Oct 2023
HI-TOM: A Benchmark for Evaluating Higher-Order Theory of Mind Reasoning in Large Language Models
Yinghui He
Yufan Wu
Yilin Jia
Rada Mihalcea
Yulong Chen
Naihao Deng
LRM
LLMAG
38
22
0
25 Oct 2023
Why LLMs Hallucinate, and How to Get (Evidential) Closure: Perceptual, Intensional, and Extensional Learning for Faithful Natural Language Generation
Adam Bouyamourn
108
15
0
23 Oct 2023
An LLM can Fool Itself: A Prompt-Based Adversarial Attack
Xilie Xu
Keyi Kong
Ning Liu
Li-zhen Cui
Di Wang
Jingfeng Zhang
Mohan Kankanhalli
AAML
SILM
33
68
0
20 Oct 2023
Core Building Blocks: Next Gen Geo Spatial GPT Application
Ashley Fernandez
Swaraj Dube
24
5
0
17 Oct 2023
Large Language Models can Learn Rules
Zhaocheng Zhu
Yuan Xue
Xinyun Chen
Denny Zhou
Jian Tang
Dale Schuurmans
Hanjun Dai
LRM
ReLM
41
63
0
10 Oct 2023
A New Benchmark and Reverse Validation Method for Passage-level Hallucination Detection
Shiping Yang
Renliang Sun
Xiao-Yi Wan
HILM
40
41
0
10 Oct 2023
Towards Mitigating Hallucination in Large Language Models via Self-Reflection
Ziwei Ji
Tiezheng Yu
Yan Xu
Nayeon Lee
Etsuko Ishii
Pascale Fung
HILM
11
57
0
10 Oct 2023
CoQuest: Exploring Research Question Co-Creation with an LLM-based Agent
Yiren Liu
Si Chen
Haocong Cheng
Mengxia Yu
Xiao Ran
Andrew Mo
Yiliu Tang
Yun Huang
LLMAG
41
46
0
09 Oct 2023
LLM4VV: Developing LLM-Driven Testsuite for Compiler Validation
Christian Munley
Aaron Jarmusch
Sunita Chandrasekaran
27
16
0
08 Oct 2023
Improving the Reliability of Large Language Models by Leveraging Uncertainty-Aware In-Context Learning
Yuchen Yang
Houqiang Li
Yanfeng Wang
Yu Wang
11
25
0
07 Oct 2023
Chain of Natural Language Inference for Reducing Large Language Model Ungrounded Hallucinations
Deren Lei
Yaxi Li
Mengya Hu
Mingyu Wang
Vincent Yun
Emily Ching
Eslam Kamal
HILM
LRM
24
40
0
06 Oct 2023
Fine-tune Language Models to Approximate Unbiased In-context Learning
Timothy Chu
Zhao Song
Chiwun Yang
27
15
0
05 Oct 2023
LLM Lies: Hallucinations are not Bugs, but Features as Adversarial Examples
Jia-Yu Yao
Kun-Peng Ning
Zhen-Hui Liu
Munan Ning
Li Yuan
HILM
LRM
AAML
26
175
0
02 Oct 2023
UPAR: A Kantian-Inspired Prompting Framework for Enhancing Large Language Model Capabilities
Hejia Geng
Boxun Xu
Peng Li
ELM
LRM
ReLM
41
1
0
30 Sep 2023
AutoHall: Automated Hallucination Dataset Generation for Large Language Models
Zouying Cao
Yifei Yang
Hai Zhao
HILM
18
8
0
30 Sep 2023
Suspicion-Agent: Playing Imperfect Information Games with Theory of Mind Aware GPT-4
Jiaxian Guo
Bo Yang
Paul D. Yoo
Bill Yuchen Lin
Yusuke Iwasawa
Yutaka Matsuo
LLMAG
21
41
0
29 Sep 2023
Probing the Moral Development of Large Language Models through Defining Issues Test
Kumar Tanmay
Aditi Khandelwal
Utkarsh Agarwal
Monojit Choudhury
LRM
8
14
0
23 Sep 2023
Foundation Metrics for Evaluating Effectiveness of Healthcare Conversations Powered by Generative AI
Mahyar Abbasian
Elahe Khatibi
Iman Azimi
David Oniani
Zahra Shakeri Hossein Abad
...
Bryant Lin
Olivier Gevaert
Li-Jia Li
Ramesh C. Jain
Amir M. Rahmani
LM&MA
ELM
AI4MH
40
66
0
21 Sep 2023
Generative AI vs. AGI: The Cognitive Strengths and Weaknesses of Modern LLMs
Ben Goertzel
38
13
0
19 Sep 2023
Explaining Agent Behavior with Large Language Models
Xijia Zhang
Yue (Sophie) Guo
Simon Stepputtis
Katia P. Sycara
Joseph Campbell
LM&Ro
LLMAG
38
6
0
19 Sep 2023
Previous
1
2
3
Next