ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2503.17682
  4. Cited By
Safe RLHF-V: Safe Reinforcement Learning from Human Feedback in Multimodal Large Language Models

Safe RLHF-V: Safe Reinforcement Learning from Human Feedback in Multimodal Large Language Models

22 March 2025
Yalan Qin
Xiuying Chen
Rui Pan
Han Zhu
C. Zhang
J. Li
Chongye Guo
Boyuan Chen
Jiayi Zhou
Kaile Wang
Juntao Dai
Chi-Min Chan
Sirui Han
Yike Guo
Yiran Yang
    OffRL
ArXivPDFHTML

Papers citing "Safe RLHF-V: Safe Reinforcement Learning from Human Feedback in Multimodal Large Language Models"

4 / 4 papers shown
Title
SafeVid: Toward Safety Aligned Video Large Multimodal Models
SafeVid: Toward Safety Aligned Video Large Multimodal Models
Yixu Wang
Jiaxin Song
Yifeng Gao
Xin Wang
Yang Yao
Yan Teng
Xingjun Ma
Yingchun Wang
Yu-Gang Jiang
2
0
0
17 May 2025
GuardReasoner-VL: Safeguarding VLMs via Reinforced Reasoning
GuardReasoner-VL: Safeguarding VLMs via Reinforced Reasoning
Yao Liu
Shengfang Zhai
Mingzhe Du
Yulin Chen
Tri Cao
...
Xinfeng Li
Kun Wang
Junfeng Fang
Jiaheng Zhang
Bryan Hooi
OffRL
LRM
7
0
0
16 May 2025
Safety in Large Reasoning Models: A Survey
Safety in Large Reasoning Models: A Survey
Cheng Wang
Yao Liu
Yangqiu Song
Duzhen Zhang
ZeLin Li
Junfeng Fang
Bryan Hooi
LRM
165
1
0
24 Apr 2025
Optimizing Safe and Aligned Language Generation: A Multi-Objective GRPO Approach
Optimizing Safe and Aligned Language Generation: A Multi-Objective GRPO Approach
Xuying Li
Zhuo Li
Yuji Kosuga
Victor Bian
47
3
0
26 Mar 2025
1