Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2503.17682
Cited By
Safe RLHF-V: Safe Reinforcement Learning from Human Feedback in Multimodal Large Language Models
22 March 2025
Yalan Qin
Xiuying Chen
Rui Pan
Han Zhu
C. Zhang
J. Li
Chongye Guo
Boyuan Chen
Jiayi Zhou
Kaile Wang
Juntao Dai
Chi-Min Chan
Sirui Han
Yike Guo
Yiran Yang
OffRL
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Safe RLHF-V: Safe Reinforcement Learning from Human Feedback in Multimodal Large Language Models"
4 / 4 papers shown
Title
SafeVid: Toward Safety Aligned Video Large Multimodal Models
Yixu Wang
Jiaxin Song
Yifeng Gao
Xin Wang
Yang Yao
Yan Teng
Xingjun Ma
Yingchun Wang
Yu-Gang Jiang
2
0
0
17 May 2025
GuardReasoner-VL: Safeguarding VLMs via Reinforced Reasoning
Yao Liu
Shengfang Zhai
Mingzhe Du
Yulin Chen
Tri Cao
...
Xinfeng Li
Kun Wang
Junfeng Fang
Jiaheng Zhang
Bryan Hooi
OffRL
LRM
7
0
0
16 May 2025
Safety in Large Reasoning Models: A Survey
Cheng Wang
Yao Liu
Yangqiu Song
Duzhen Zhang
ZeLin Li
Junfeng Fang
Bryan Hooi
LRM
165
1
0
24 Apr 2025
Optimizing Safe and Aligned Language Generation: A Multi-Objective GRPO Approach
Xuying Li
Zhuo Li
Yuji Kosuga
Victor Bian
47
3
0
26 Mar 2025
1