ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2203.02155
  4. Cited By
Training language models to follow instructions with human feedback

Training language models to follow instructions with human feedback

4 March 2022
Long Ouyang
Jeff Wu
Xu Jiang
Diogo Almeida
Carroll L. Wainwright
Pamela Mishkin
Chong Zhang
Sandhini Agarwal
Katarina Slama
Alex Ray
John Schulman
Jacob Hilton
Fraser Kelton
Luke E. Miller
Maddie Simens
Amanda Askell
Peter Welinder
Paul Christiano
Jan Leike
Ryan J. Lowe
    OSLMALM
ArXiv (abs)PDFHTML

Papers citing "Training language models to follow instructions with human feedback"

50 / 6,370 papers shown
Title
OmniHuman-1: Rethinking the Scaling-Up of One-Stage Conditioned Human Animation Models
OmniHuman-1: Rethinking the Scaling-Up of One-Stage Conditioned Human Animation Models
Gaojie Lin
Jianwen Jiang
Jiaqi Yang
Zerong Zheng
Chao Liang
DiffMVGen
391
29
0
01 Jul 2025
Think Before You Segment: High-Quality Reasoning Segmentation with GPT Chain of Thoughts
Think Before You Segment: High-Quality Reasoning Segmentation with GPT Chain of Thoughts
Shiu-hong Kao
Yu-Wing Tai
Chi-Keung Tang
MLLMLRM
283
1
0
01 Jul 2025
ReasonGRM: Enhancing Generative Reward Models through Large Reasoning Models
ReasonGRM: Enhancing Generative Reward Models through Large Reasoning Models
Bin Chen
Xinzge Gao
Chuanrui Hu
Penghang Yu
Hua Zhang
Bing-Kun Bao
ReLMLRM
19
0
0
20 Jun 2025
The Role of Model Confidence on Bias Effects in Measured Uncertainties
The Role of Model Confidence on Bias Effects in Measured Uncertainties
Xinyi Liu
Weiguang Wang
Hangfeng He
21
0
0
20 Jun 2025
Reward-Agnostic Prompt Optimization for Text-to-Image Diffusion Models
Reward-Agnostic Prompt Optimization for Text-to-Image Diffusion Models
Semin Kim
Yeonwoo Cha
Jaehoon Yoo
Seunghoon Hong
EGVM
32
0
0
20 Jun 2025
Arch-Router: Aligning LLM Routing with Human Preferences
Arch-Router: Aligning LLM Routing with Human Preferences
Co Tran
Salman Paracha
Adil Hafeez
Shuguang Chen
29
0
0
19 Jun 2025
Can structural correspondences ground real world representational content in Large Language Models?
Can structural correspondences ground real world representational content in Large Language Models?
Iwan Williams
22
0
0
19 Jun 2025
StoryWriter: A Multi-Agent Framework for Long Story Generation
StoryWriter: A Multi-Agent Framework for Long Story Generation
Haotian Xia
Hao Peng
Yunjia Qi
Xiaozhi Wang
Bin Xu
Lei Hou
Juanzi Li
VGen
13
0
0
19 Jun 2025
Reranking-based Generation for Unbiased Perspective Summarization
Reranking-based Generation for Unbiased Perspective Summarization
Narutatsu Ri
Nicholas Deas
Kathleen McKeown
OffRL
24
0
0
19 Jun 2025
Probing the Robustness of Large Language Models Safety to Latent Perturbations
Probing the Robustness of Large Language Models Safety to Latent Perturbations
Tianle Gu
Kexin Huang
Zongqi Wang
Yixu Wang
Jie Li
Yuanqi Yao
Yang Yao
Yujiu Yang
Yan Teng
Yingchun Wang
AAMLLLMSV
35
0
0
19 Jun 2025
Probe before You Talk: Towards Black-box Defense against Backdoor Unalignment for Large Language Models
Probe before You Talk: Towards Black-box Defense against Backdoor Unalignment for Large Language Models
Biao Yi
Tiansheng Huang
Sishuo Chen
Tong Li
Zheli Liu
Zhixuan Chu
Yiming Li
AAML
37
9
0
19 Jun 2025
AutoV: Learning to Retrieve Visual Prompt for Large Vision-Language Models
AutoV: Learning to Retrieve Visual Prompt for Large Vision-Language Models
Yuan Zhang
Chun-Kai Fan
Tao Huang
Ming Lu
Sicheng Yu
Junwen Pan
Kuan Cheng
Qi She
Shanghang Zhang
VLMLRM
21
0
0
19 Jun 2025
GRPO-CARE: Consistency-Aware Reinforcement Learning for Multimodal Reasoning
GRPO-CARE: Consistency-Aware Reinforcement Learning for Multimodal Reasoning
Yi Chen
Yuying Ge
Rui Wang
Yixiao Ge
Junhao Cheng
Ying Shan
Xihui Liu
OffRLVLMLRM
32
0
0
19 Jun 2025
From General to Targeted Rewards: Surpassing GPT-4 in Open-Ended Long-Context Generation
From General to Targeted Rewards: Surpassing GPT-4 in Open-Ended Long-Context Generation
Zhihan Guo
Jiele Wu
Wenqian Cui
Yifei Zhang
Minda Hu
Yufei Wang
Irwin King
ALMLRM
22
0
0
19 Jun 2025
RePCS: Diagnosing Data Memorization in LLM-Powered Retrieval-Augmented Generation
RePCS: Diagnosing Data Memorization in LLM-Powered Retrieval-Augmented Generation
Le Vu Anh
Nguyen Viet Anh
Mehmet Dik
Luong Van Nghia
30
0
0
18 Jun 2025
video-SALMONN 2: Captioning-Enhanced Audio-Visual Large Language Models
video-SALMONN 2: Captioning-Enhanced Audio-Visual Large Language Models
Changli Tang
Yixuan Li
Yudong Yang
Jimin Zhuang
Guangzhi Sun
Wei Li
Zejun Ma
Chao Zhang
25
0
0
18 Jun 2025
Steering Your Diffusion Policy with Latent Space Reinforcement Learning
Steering Your Diffusion Policy with Latent Space Reinforcement Learning
Andrew Wagenmaker
Mitsuhiko Nakamoto
Yunchu Zhang
S. Park
Waleed Yagoub
Anusha Nagabandi
Abhishek Gupta
Sergey Levine
OffRL
37
0
0
18 Jun 2025
AgentGroupChat-V2: Divide-and-Conquer Is What LLM-Based Multi-Agent System Need
AgentGroupChat-V2: Divide-and-Conquer Is What LLM-Based Multi-Agent System Need
Zhouhong Gu
Xiaoxuan Zhu
Yin Cai
Hao Shen
Xingzhou Chen
...
Hongwei Feng
Yanghua Xiao
Zheyu Ye
Yao Hu
Shaosheng Cao
LLMAG
22
0
0
18 Jun 2025
AutoRule: Reasoning Chain-of-thought Extracted Rule-based Rewards Improve Preference Learning
AutoRule: Reasoning Chain-of-thought Extracted Rule-based Rewards Improve Preference Learning
Tevin Wang
Chenyan Xiong
LRM
37
0
0
18 Jun 2025
SANSKRITI: A Comprehensive Benchmark for Evaluating Language Models' Knowledge of Indian Culture
SANSKRITI: A Comprehensive Benchmark for Evaluating Language Models' Knowledge of Indian Culture
Arijit Maji
Raghvendra Kumar
Akash Ghosh
Anushka
Sriparna Saha
ELM
25
0
0
18 Jun 2025
Revisiting Compositional Generalization Capability of Large Language Models Considering Instruction Following Ability
Revisiting Compositional Generalization Capability of Large Language Models Considering Instruction Following Ability
Yusuke Sakai
Hidetaka Kamigaito
Taro Watanabe
LRM
31
0
0
18 Jun 2025
Learning-Time Encoding Shapes Unlearning in LLMs
Learning-Time Encoding Shapes Unlearning in LLMs
Ruihan Wu
Konstantin Garov
Kamalika Chaudhuri
MU
24
0
0
18 Jun 2025
Lessons from Training Grounded LLMs with Verifiable Rewards
Lessons from Training Grounded LLMs with Verifiable Rewards
Shang Hong Sim
Tej Deep Pala
Vernon Y.H. Toh
Hai Leong Chieu
Amir Zadeh
Chuan Li
Navonil Majumder
Soujanya Poria
OffRLRALMLRM
20
0
0
18 Jun 2025
Modeling the One-to-Many Property in Open-Domain Dialogue with LLMs
Modeling the One-to-Many Property in Open-Domain Dialogue with LLMs
Jing Yang Lee
Kong-Aik Lee
Woon-Seng Gan
40
0
0
18 Jun 2025
Sysformer: Safeguarding Frozen Large Language Models with Adaptive System Prompts
Sysformer: Safeguarding Frozen Large Language Models with Adaptive System Prompts
Kartik Sharma
Yiqiao Jin
Vineeth Rakesh
Yingtong Dou
Menghai Pan
Mahashweta Das
Srijan Kumar
AAML
18
0
0
18 Jun 2025
LoX: Low-Rank Extrapolation Robustifies LLM Safety Against Fine-tuning
LoX: Low-Rank Extrapolation Robustifies LLM Safety Against Fine-tuning
Gabrel J. Perin
Runjin Chen
Xuxi Chen
Nina S. T. Hirata
Zhangyang Wang
Junyuan Hong
AAML
41
0
0
18 Jun 2025
From LLMs to MLLMs to Agents: A Survey of Emerging Paradigms in Jailbreak Attacks and Defenses within LLM Ecosystem
From LLMs to MLLMs to Agents: A Survey of Emerging Paradigms in Jailbreak Attacks and Defenses within LLM Ecosystem
Yanxu Mao
Tiehan Cui
Peipei Liu
Datao You
Hongsong Zhu
AAML
17
0
0
18 Jun 2025
Reward Models in Deep Reinforcement Learning: A Survey
Reward Models in Deep Reinforcement Learning: A Survey
Rui Yu
Shenghua Wan
Yucen Wang
Chen-Xiao Gao
Le Gan
Zongzhang Zhang
De-Chuan Zhan
OffRL
15
0
0
18 Jun 2025
Intelligent Assistants for the Semiconductor Failure Analysis with LLM-Based Planning Agents
Intelligent Assistants for the Semiconductor Failure Analysis with LLM-Based Planning Agents
Aline Dobrovsky
Konstantin Schekotihin
Christian Burmer
LLMAG
22
0
0
18 Jun 2025
GRAM: A Generative Foundation Reward Model for Reward Generalization
GRAM: A Generative Foundation Reward Model for Reward Generalization
Chenglong Wang
Yang Gan
Yifu Huo
Yongyu Mu
Qiaozhi He
...
Bei Li
Tong Xiao
Chunliang Zhang
Tongran Liu
Jingbo Zhu
ALMOffRLLRM
57
0
0
17 Jun 2025
Adaptive Accompaniment with ReaLchords
Adaptive Accompaniment with ReaLchords
Yusong Wu
Tim Cooijmans
Kyle Kastner
Adam Roberts
Ian Simon
...
Shayegan Omidshafiei
Aaron Courville
Pablo Samuel Castro
Natasha Jaques
Cheng-Zhi Anna Huang
19
0
0
17 Jun 2025
Reinforcement Learning with Verifiable Rewards Implicitly Incentivizes Correct Reasoning in Base LLMs
Reinforcement Learning with Verifiable Rewards Implicitly Incentivizes Correct Reasoning in Base LLMs
Xumeng Wen
Zihan Liu
Shun Zheng
Zhijian Xu
Shengyu Ye
...
Yang Wang
Junjie Li
Ziming Miao
Jiang Bian
Mao Yang
LRM
34
0
0
17 Jun 2025
MDBench: A Synthetic Multi-Document Reasoning Benchmark Generated with Knowledge Guidance
Joseph Peper
Wenzhao Qiu
Ali Payani
Lu Wang
20
0
0
17 Jun 2025
AviationLLM: An LLM-based Knowledge System for Aviation Training
AviationLLM: An LLM-based Knowledge System for Aviation Training
Jiaáng Wan
Feng Shen
Fujuan Li
Yanjin Sun
Yan Li
Shiwen Zhang
25
0
0
17 Jun 2025
FORTRESS: Frontier Risk Evaluation for National Security and Public Safety
FORTRESS: Frontier Risk Evaluation for National Security and Public Safety
Christina Q. Knight
Kaustubh Deshpande
Ved Sirdeshmukh
Meher Mankikar
Scale Red Team
SEAL Research Team
Julian Michael
AAMLELM
39
0
0
17 Jun 2025
Probabilistic Aggregation and Targeted Embedding Optimization for Collective Moral Reasoning in Large Language Models
Probabilistic Aggregation and Targeted Embedding Optimization for Collective Moral Reasoning in Large Language Models
Chenchen Yuan
Zheyu Zhang
Shuo Yang
Bardh Prenkaj
Gjergji Kasneci
36
0
0
17 Jun 2025
SFT-GO: Supervised Fine-Tuning with Group Optimization for Large Language Models
SFT-GO: Supervised Fine-Tuning with Group Optimization for Large Language Models
Gyuhak Kim
Sumiran Thakur
Su Min Park
Wei Wei
Yujia Bao
20
0
0
17 Jun 2025
Reasoning with Exploration: An Entropy Perspective
Reasoning with Exploration: An Entropy Perspective
Daixuan Cheng
Shaohan Huang
Xuekai Zhu
Bo Dai
Wayne Xin Zhao
Zhenliang Zhang
Furu Wei
LRM
32
0
0
17 Jun 2025
Expectation Confirmation Preference Optimization for Multi-Turn Conversational Recommendation Agent
Expectation Confirmation Preference Optimization for Multi-Turn Conversational Recommendation Agent
Xueyang Feng
Jingsen Zhang
Jiakai Tang
Wei Li
Guohao Cai
X. Chen
Quanyu Dai
Y. Zhu
Zhenhua Dong
27
0
0
17 Jun 2025
TGDPO: Harnessing Token-Level Reward Guidance for Enhancing Direct Preference Optimization
TGDPO: Harnessing Token-Level Reward Guidance for Enhancing Direct Preference Optimization
Mingkang Zhu
Xi Chen
Zhongdao Wang
Bei Yu
Hengshuang Zhao
Jiaya Jia
19
0
0
17 Jun 2025
ASMR: Augmenting Life Scenario using Large Generative Models for Robotic Action Reflection
ASMR: Augmenting Life Scenario using Large Generative Models for Robotic Action Reflection
Shang-Chi Tsai
Seiya Kawano
Angel García Contreras
Koichiro Yoshino
Yun-Nung Chen
LM&Ro
40
2
0
16 Jun 2025
Language Agents for Hypothesis-driven Clinical Decision Making with Reinforcement Learning
Language Agents for Hypothesis-driven Clinical Decision Making with Reinforcement Learning
David Bani-Harouni
Chantal Pellegrini
Ege Özsoy
Matthias Keicher
Nassir Navab
LLMAGLM&MA
22
0
0
16 Jun 2025
AutoVLA: A Vision-Language-Action Model for End-to-End Autonomous Driving with Adaptive Reasoning and Reinforcement Fine-Tuning
AutoVLA: A Vision-Language-Action Model for End-to-End Autonomous Driving with Adaptive Reasoning and Reinforcement Fine-Tuning
Zewei Zhou
Tianhui Cai
Seth Z. Zhao
Yun Zhang
Zhiyu Huang
Bolei Zhou
Jiaqi Ma
LRMVLM
22
0
0
16 Jun 2025
Discrete Diffusion in Large Language and Multimodal Models: A Survey
Discrete Diffusion in Large Language and Multimodal Models: A Survey
Runpeng Yu
Qi Li
Xinchao Wang
DiffMAI4CE
42
0
0
16 Jun 2025
Decompositional Reasoning for Graph Retrieval with Large Language Models
Decompositional Reasoning for Graph Retrieval with Large Language Models
Valentin Six
Evan Dufraisse
Gaël de Chalendar
ReLMLRM
25
0
0
16 Jun 2025
Document-Level Tabular Numerical Cross-Checking: A Coarse-to-Fine Approach
Document-Level Tabular Numerical Cross-Checking: A Coarse-to-Fine Approach
Chaoxu Pang
Yixuan Cao
Ganbin Zhou
Hongwei Bran Li
Ping Luo
LMTD
42
0
0
16 Jun 2025
Flexible-length Text Infilling for Discrete Diffusion Models
Flexible-length Text Infilling for Discrete Diffusion Models
Andrew Zhang
Anushka Sivakumar
Chiawei Tang
Chris Thomas
DiffM
28
0
0
16 Jun 2025
Balancing Knowledge Delivery and Emotional Comfort in Healthcare Conversational Systems
Balancing Knowledge Delivery and Emotional Comfort in Healthcare Conversational Systems
Shang-Chi Tsai
Yun-Nung Chen
AI4MH
31
1
0
16 Jun 2025
Prefix-Tuning+: Modernizing Prefix-Tuning by Decoupling the Prefix from Attention
Prefix-Tuning+: Modernizing Prefix-Tuning by Decoupling the Prefix from Attention
Haonan Wang
Brian K Chen
Siquan Li
Xinhe Liang
Hwee Kuan Lee
Kenji Kawaguchi
Tianyang Hu
23
0
0
16 Jun 2025
Attribution-guided Pruning for Compression, Circuit Discovery, and Targeted Correction in LLMs
Attribution-guided Pruning for Compression, Circuit Discovery, and Targeted Correction in LLMs
Sayed Mohammad Vakilzadeh Hatefi
Maximilian Dreyer
Reduan Achtibat
Patrick Kahardipraja
Thomas Wiegand
Wojciech Samek
Sebastian Lapuschkin
31
0
0
16 Jun 2025
1234...126127128
Next