Papers
Communities
Organizations
Events
Blog
Pricing
Search
Open menu
Home
Papers
2501.01336
Cited By
Aligning Large Language Models for Faithful Integrity Against Opposing Argument
3 January 2025
Yong Zhao
Yang Deng
See-Kiong Ng
Tat-Seng Chua
Re-assign community
ArXiv (abs)
PDF
HTML
Papers citing
"Aligning Large Language Models for Faithful Integrity Against Opposing Argument"
2 / 2 papers shown
Title
Reasoning Models Are More Easily Gaslighted Than You Think
B. Zhu
Hailong Yin
Jingjing Chen
Yu Jiang
LRM
80
0
0
11 Jun 2025
Calling a Spade a Heart: Gaslighting Multimodal Large Language Models via Negation
Bin Zhu
Hui yan Qi
Yinxuan Gui
Jingjing Chen
Chong-Wah Ngo
Ee-Peng Lim
483
2
0
31 Jan 2025
1