Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2309.13788
Cited By
v1
v2
v3
v4
v5 (latest)
Can LLM-Generated Misinformation Be Detected?
25 September 2023
Canyu Chen
Kai Shu
DeLMO
Re-assign community
ArXiv (abs)
PDF
HTML
Papers citing
"Can LLM-Generated Misinformation Be Detected?"
43 / 43 papers shown
Title
The Compositional Architecture of Regret in Large Language Models
Xiangxiang Cui
Shu Yang
Tianjin Huang
Wanyu Lin
Lijie Hu
Di Wang
27
0
0
18 Jun 2025
GradEscape: A Gradient-Based Evader Against AI-Generated Text Detectors
Wenlong Meng
Shuguo Fan
Chengkun Wei
Min Chen
Yuwei Li
Yuanchao Zhang
Zhikun Zhang
Wenzhi Chen
17
0
0
09 Jun 2025
SoK: Are Watermarks in LLMs Ready for Deployment?
Kieu Dang
Phung Lai
Nhathai Phan
Yelong Shen
Ruoming Jin
Abdallah Khreishah
My T. Thai
27
0
0
05 Jun 2025
MultiHoax: A Dataset of Multi-hop False-Premise Questions
Mohammadamin Shafiei
Hamidreza Saffari
Nafise Sadat Moosavi
LRM
34
0
0
30 May 2025
CMIE: Combining MLLM Insights with External Evidence for Explainable Out-of-Context Misinformation Detection
Fanxiao Li
Jiaying Wu
Canyuan He
Wei Zhou
LRM
12
0
0
29 May 2025
T^2Agent A Tool-augmented Multimodal Misinformation Detection Agent with Monte Carlo Tree Search
Xing Cui
Yueying Zou
Zekun Li
Peipei Li
Xinyuan Xu
Xuannan Liu
Huaibo Huang
Ran He
233
0
0
26 May 2025
GUARDIAN: Safeguarding LLM Multi-Agent Collaborations with Temporal Graph Modeling
Jialong Zhou
L. Wang
Xiao Yang
LLMAG
56
0
0
25 May 2025
Evidence-Grounded Multimodal Misinformation Detection with Attention-Based GNNs
Sharad Duwal
Mir Nafis Sharear Shopnil
Abhishek Tyagi
Adiba Mahbub Proma
194
0
0
23 May 2025
Seeing Through Deception: Uncovering Misleading Creator Intent in Multimodal News with Vision-Language Models
Jiaying Wu
Fanxiao Li
Min-Yen Kan
Bryan Hooi
86
0
0
21 May 2025
Unified Attacks to Large Language Model Watermarks: Spoofing and Scrubbing in Unauthorized Knowledge Distillation
Xin Yi
Shunfan Zhengc
Linlin Wanga
Xiaoling Wang
Liang He
Liang He
AAML
491
0
0
24 Apr 2025
Reinforcement Learning from Human Feedback
Nathan Lambert
OffRL
AI4CE
112
22
0
16 Apr 2025
From Misleading Queries to Accurate Answers: A Three-Stage Fine-Tuning Method for LLMs
Guocong Li
Weize Liu
Yihang Wu
Ping Wang
Shuaihan Huang
Hongxia Xu
Jian Wu
KELM
HILM
117
0
0
15 Apr 2025
Firm or Fickle? Evaluating Large Language Models Consistency in Sequential Interactions
Yubo Li
Yidi Miao
Xueying Ding
Ramayya Krishnan
R. Padman
122
0
0
28 Mar 2025
Tricking Retrievers with Influential Tokens: An Efficient Black-Box Corpus Poisoning Attack
Cheng Wang
Yiwei Wang
Yujun Cai
Bryan Hooi
AAML
75
1
0
27 Mar 2025
How to Protect Yourself from 5G Radiation? Investigating LLM Responses to Implicit Misinformation
Ruohao Guo
Wei Xu
Alan Ritter
106
3
0
12 Mar 2025
When Claims Evolve: Evaluating and Enhancing the Robustness of Embedding Models Against Misinformation Edits
Jabez Magomere
Emanuele La Malfa
Manuel Tonneau
Ashkan Kazemi
Scott A. Hale
KELM
171
1
0
05 Mar 2025
SafeVLA: Towards Safety Alignment of Vision-Language-Action Model via Constrained Learning
Borong Zhang
Yuhao Zhang
Yalan Qin
Yingshan Lei
Josef Dai
Yuanpei Chen
Yaodong Yang
126
4
0
05 Mar 2025
FACT-AUDIT: An Adaptive Multi-Agent Framework for Dynamic Fact-Checking Evaluation of Large Language Models
Hongzhan Lin
Yang Deng
Yuxuan Gu
Wenxuan Zhang
Jing Ma
See-Kiong Ng
Tat-Seng Chua
LLMAG
KELM
HILM
142
1
0
25 Feb 2025
MindAligner: Explicit Brain Functional Alignment for Cross-Subject Visual Decoding from Limited fMRI Data
Yuqin Dai
Zhouheng Yao
Chunfeng Song
Qihao Zheng
Weijian Mai
Kunyu Peng
Shuai Lu
Wanli Ouyang
Jian Yang
Jiamin Wu
483
2
0
07 Feb 2025
Fake News Detection After LLM Laundering: Measurement and Explanation
Rupak Kumar Das
Jonathan Dodge
187
1
0
29 Jan 2025
Data-Free Model-Related Attacks: Unleashing the Potential of Generative AI
Dayong Ye
Tianqing Zhu
Shang Wang
B. Liu
Lefei Zhang
Wanlei Zhou
Yanmei Zhang
AAML
SILM
105
0
0
28 Jan 2025
Mitigating GenAI-powered Evidence Pollution for Out-of-Context Multimodal Misinformation Detection
Zehong Yan
Peng Qi
Wynne Hsu
Mong Li Lee
85
0
0
24 Jan 2025
Can AI-Generated Text be Reliably Detected?
Vinu Sankar Sadasivan
Aounon Kumar
S. Balasubramanian
Wenxiao Wang
Soheil Feizi
DeLMO
287
389
0
20 Jan 2025
Surveying Attitudinal Alignment Between Large Language Models Vs. Humans Towards 17 Sustainable Development Goals
Qingyang Wu
Ying Xu
Tingsong Xiao
Yunze Xiao
Yitong Li
...
Yichi Zhang
Shanghai Zhong
Yuwei Zhang
Wei Lu
Yifan Yang
186
3
0
17 Jan 2025
DEFAME: Dynamic Evidence-based FAct-checking with Multimodal Experts
Tobias Braun
Mark Rothermel
Marcus Rohrbach
Anna Rohrbach
206
6
0
13 Dec 2024
Retrieval-Augmented Generation with Estimation of Source Reliability
Jeongyeon Hwang
Junyoung Park
Hyejin Park
Dongwoo Kim
Sangdon Park
Jungseul Ok
RALM
98
1
0
30 Oct 2024
Online Detection of LLM-Generated Texts via Sequential Hypothesis Testing by Betting
Can Chen
Jun-Kun Wang
DeLMO
162
0
0
29 Oct 2024
Analysing the Residual Stream of Language Models Under Knowledge Conflicts
Yu Zhao
Xiaotang Du
Giwon Hong
Aryo Pradipta Gema
Alessio Devoto
Hongru Wang
Xuanli He
Kam-Fai Wong
Pasquale Minervini
KELM
168
1
0
21 Oct 2024
Steering Knowledge Selection Behaviours in LLMs via SAE-Based Representation Engineering
Yu Zhao
Alessio Devoto
Giwon Hong
Xiaotang Du
Aryo Pradipta Gema
Hongru Wang
Xuanli He
Kam-Fai Wong
Pasquale Minervini
KELM
LLMSV
132
28
0
21 Oct 2024
Unveiling Large Language Models Generated Texts: A Multi-Level Fine-Grained Detection Framework
Zhen Tao
Zhiyu Li
Runyu Chen
Dinghao Xi
Wei Xu
DeLMO
77
2
0
18 Oct 2024
Recent Advances in Attack and Defense Approaches of Large Language Models
Jing Cui
Yishi Xu
Zhewei Huang
Shuchang Zhou
Jianbin Jiao
Junge Zhang
PILM
AAML
121
2
0
05 Sep 2024
On Large Language Models in National Security Applications
William N. Caballero
Phillip R. Jenkins
ELM
90
10
0
03 Jul 2024
Detecting AI-Generated Text: Factors Influencing Detectability with Current Methods
Kathleen C. Fraser
Hillary Dawkins
S. Kiritchenko
DeLMO
150
13
0
21 Jun 2024
MMFakeBench: A Mixed-Source Multimodal Misinformation Detection Benchmark for LVLMs
Xuannan Liu
Zekun Li
Peipei Li
Shuhan Xia
Xing Cui
Linzhi Huang
Huaibo Huang
Weihong Deng
Zhaofeng He
139
23
0
13 Jun 2024
Navigating LLM Ethics: Advancements, Challenges, and Future Directions
Junfeng Jiao
S. Afroogh
Yiming Xu
Connor Phillips
AILaw
134
23
0
14 May 2024
Hallucination Detection in Foundation Models for Decision-Making: A Flexible Definition and Review of the State of the Art
Neeloy Chakraborty
Melkior Ornik
Katherine Driggs-Campbell
LRM
241
12
0
25 Mar 2024
TELEClass: Taxonomy Enrichment and LLM-Enhanced Hierarchical Text Classification with Minimal Supervision
Yunyi Zhang
Ruozhen Yang
Xueqiang Xu
Rui Li
Jinfeng Xiao
Jiaming Shen
Jiawei Han
139
17
0
29 Feb 2024
Large Language Models are Vulnerable to Bait-and-Switch Attacks for Generating Harmful Content
Federico Bianchi
James Zou
74
5
0
21 Feb 2024
DELL: Generating Reactions and Explanations for LLM-Based Misinformation Detection
Herun Wan
Shangbin Feng
Zhaoxuan Tan
Heng Wang
Yulia Tsvetkov
Minnan Luo
137
34
0
16 Feb 2024
Detecting Multimedia Generated by Large AI Models: A Survey
Li Lin
Neeraj Gupta
Yue Zhang
Hainan Ren
Chun-Hao Liu
Feng Ding
Xin Eric Wang
Xin Li
Luisa Verdoliva
Shu Hu
211
64
0
22 Jan 2024
Baichuan 2: Open Large-scale Language Models
Ai Ming Yang
Bin Xiao
Bingning Wang
Borong Zhang
Ce Bian
...
Youxin Jiang
Yuchen Gao
Yupeng Zhang
Guosheng Dong
Zhiying Wu
ELM
LRM
328
755
0
19 Sep 2023
Certifying LLM Safety against Adversarial Prompting
Aounon Kumar
Chirag Agarwal
Suraj Srinivas
Aaron Jiaxun Li
Soheil Feizi
Himabindu Lakkaraju
AAML
149
197
0
06 Sep 2023
HowkGPT: Investigating the Detection of ChatGPT-generated University Student Homework through Context-Aware Perplexity Analysis
Christoforos Vasilatos
Manaar Alam
Talal Rahwan
Yasir Zaki
Michail Maniatakos
DeLMO
142
36
0
26 May 2023
1