Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2203.02155
Cited By
Training language models to follow instructions with human feedback
4 March 2022
Long Ouyang
Jeff Wu
Xu Jiang
Diogo Almeida
Carroll L. Wainwright
Pamela Mishkin
Chong Zhang
Sandhini Agarwal
Katarina Slama
Alex Ray
John Schulman
Jacob Hilton
Fraser Kelton
Luke E. Miller
Maddie Simens
Amanda Askell
Peter Welinder
Paul Christiano
Jan Leike
Ryan J. Lowe
OSLM
ALM
Re-assign community
ArXiv (abs)
PDF
HTML
Papers citing
"Training language models to follow instructions with human feedback"
50 / 6,390 papers shown
Title
SegAgent: Exploring Pixel Understanding Capabilities in MLLMs by Imitating Human Annotator Trajectories
Muzhi Zhu
Yuzhuo Tian
Hao Chen
Chunluan Zhou
Qingpei Guo
Yongxu Liu
M. Yang
Chunhua Shen
MLLM
VLM
126
1
0
11 Mar 2025
LangTime: A Language-Guided Unified Model for Time Series Forecasting with Proximal Policy Optimization
Wenzhe Niu
Zongxia Xie
Yanru Sun
Wei He
Man Xu
Chao Hao
AI4TS
198
1
0
11 Mar 2025
Group Preference Alignment: Customized LLM Response Generation from In-Situ Conversations
Ishani Mondal
Jack W. Stokes
S. Jauhar
Longqi Yang
Mengting Wan
Xiaofeng Xu
Xia Song
Jennifer Neville
107
0
0
11 Mar 2025
Towards Large-scale Chemical Reaction Image Parsing via a Multimodal Large Language Model
Yufan Chen
Ching Ting Leung
Jianwei Sun
Yong Huang
Linyan Li
Hao Chen
Hanyu Gao
98
1
0
11 Mar 2025
Exploiting Instruction-Following Retrievers for Malicious Information Retrieval
Parishad BehnamGhader
Nicholas Meade
Siva Reddy
145
1
0
11 Mar 2025
GTR: Guided Thought Reinforcement Prevents Thought Collapse in RL-based VLM Agent Training
Tong Wei
Yijun Yang
Junliang Xing
Yuanchun Shi
Zongqing Lu
Deheng Ye
OffRL
LRM
95
2
0
11 Mar 2025
Near-Optimal Sample Complexity for Iterated CVaR Reinforcement Learning with a Generative Model
Zilong Deng
Simon Khan
Shaofeng Zou
179
1
0
11 Mar 2025
Dialogue Injection Attack: Jailbreaking LLMs through Context Manipulation
Wenlong Meng
Fan Zhang
Wendao Yao
Zhenyuan Guo
Yongqian Li
Chengkun Wei
Wenzhi Chen
AAML
120
5
0
11 Mar 2025
Generating Robot Constitutions & Benchmarks for Semantic Safety
P. Sermanet
Anirudha Majumdar
A. Irpan
Dmitry Kalashnikov
Vikas Sindhwani
LM&Ro
170
3
0
11 Mar 2025
PromptLNet: Region-Adaptive Aesthetic Enhancement via Prompt Guidance in Low-Light Enhancement Net
Jun Yin
Yangfan He
Miao Zhang
Pengyu Zeng
Tianyi Wang
Shuai Lu
Xueqian Wang
DiffM
148
7
0
11 Mar 2025
Chemical reasoning in LLMs unlocks steerable synthesis planning and reaction mechanism elucidation
Andres M Bran
Theo A Neukomm
Daniel Armstrong
Zlatko Joncev
Philippe Schwaller
LRM
121
3
0
11 Mar 2025
Regulatory DNA sequence Design with Reinforcement Learning
Zhao-Qing Yang
Fuchun Sun
Chuan Cao
Ji-Rong Wen
101
1
0
11 Mar 2025
A Cascading Cooperative Multi-agent Framework for On-ramp Merging Control Integrating Large Language Models
Miao Zhang
Zhenlong Fang
Tianyi Wang
Qin Zhang
Shuai Lu
Junfeng Jiao
Tianyu Shi
AI4CE
123
5
0
11 Mar 2025
Robust Multi-Objective Controlled Decoding of Large Language Models
Seongho Son
William Bankes
Sangwoong Yoon
Shyam Sundhar Ramesh
Xiaohang Tang
Ilija Bogunovic
129
2
0
11 Mar 2025
Backtracking for Safety
Bilgehan Sel
Dingcheng Li
Phillip Wallis
Vaishakh Keshava
Ming Jin
Siddhartha Reddy Jonnalagadda
KELM
96
0
0
11 Mar 2025
AuthorMist: Evading AI Text Detectors with Reinforcement Learning
Isaac David
Arthur Gervais
DeLMO
78
0
0
10 Mar 2025
Graphormer-Guided Task Planning: Beyond Static Rules with LLM Safety Perception
Wanjing Huang
Tongjie Pan
Yalan Ye
98
2
0
10 Mar 2025
Navigating Motion Agents in Dynamic and Cluttered Environments through LLM Reasoning
Yubo Zhao
Qi Wu
Yifan Wang
Yu-Wing Tai
Chi-Keung Tang
LLMAG
LRM
512
0
0
10 Mar 2025
Towards Large Language Models that Benefit for All: Benchmarking Group Fairness in Reward Models
Kefan Song
Jin Yao
Runnan Jiang
Rohan Chandra
Shangtong Zhang
ALM
93
0
0
10 Mar 2025
Painting with Words: Elevating Detailed Image Captioning with Benchmark and Alignment Learning
Qinghao Ye
Xianhan Zeng
Fu Li
Chong Li
Haoqi Fan
CoGe
116
5
0
10 Mar 2025
DistiLLM-2: A Contrastive Approach Boosts the Distillation of LLMs
Jongwoo Ko
Tianyi Chen
Sungnyun Kim
Tianyu Ding
Luming Liang
Ilya Zharkov
Se-Young Yun
VLM
464
2
0
10 Mar 2025
Boosting the Generalization and Reasoning of Vision Language Models with Curriculum Reinforcement Learning
Huilin Deng
Ding Zou
Rui Ma
Hongchen Luo
Yang Cao
Yu Kang
LRM
VLM
117
22
0
10 Mar 2025
From Idea to Implementation: Evaluating the Influence of Large Language Models in Software Development -- An Opinion Paper
Sargam Yadav
Asifa Mehmood Qureshi
Abhishek Kaushik
Shubham Sharma
Roisin Loughran
...
. Nikhil Singh
Padraic O'Hara
Pranay Jaiswal
Roshan Chandru
David Lillis
164
1
0
10 Mar 2025
VLRMBench: A Comprehensive and Challenging Benchmark for Vision-Language Reward Models
Jiacheng Ruan
Wenzhen Yuan
Xian Gao
Ye Guo
Daoxin Zhang
Zhe Xu
Yao Hu
Ting Liu
Yuzhuo Fu
LRM
VLM
167
6
0
10 Mar 2025
Utilizing Jailbreak Probability to Attack and Safeguard Multimodal LLMs
Wenzhuo Xu
Zhipeng Wei
Xiongtao Sun
Deyue Zhang
Dongdong Yang
Quanchen Zou
Xinming Zhang
AAML
92
0
0
10 Mar 2025
Uncovering Cross-Domain Recommendation Ability of Large Language Models
Xinyi Liu
Ruijie Wang
Dachun Sun
Dilek Hakkani-Tur
Tarek Abdelzaher
455
1
0
10 Mar 2025
Unleashing the Potential of Large Language Models for Text-to-Image Generation through Autoregressive Representation Alignment
Xing Xie
Jiawei Liu
Ziyue Lin
Huijie Fan
Zhi Han
Yandong Tang
Liangqiong Qu
117
0
0
10 Mar 2025
LMM-R1: Empowering 3B LMMs with Strong Reasoning Abilities Through Two-Stage Rule-Based RL
Yingzhe Peng
Gongrui Zhang
Miaosen Zhang
Zhiyuan You
Jie Liu
Qipeng Zhu
Kai Yang
Xingzhong Xu
Xin Geng
Xu Yang
LRM
ReLM
252
88
0
10 Mar 2025
AlphaDrive: Unleashing the Power of VLMs in Autonomous Driving via Reinforcement Learning and Reasoning
Bo Jiang
Shaoyu Chen
Qian Zhang
Wenyu Liu
Xinggang Wang
OffRL
LRM
VLM
161
12
0
10 Mar 2025
XIFBench: Evaluating Large Language Models on Multilingual Instruction Following
Zhiyu Li
Kehai Chen
Yunfei Long
X. Bai
Yaoyin Zhang
Xuchen Wei
Junlin Li
Min Zhang
ELM
86
1
0
10 Mar 2025
RePO: ReLU-based Preference Optimization
Junkang Wu
Kexin Huang
Xue Wang
Jinyang Gao
Bolin Ding
Jiancan Wu
Xiangnan He
Xiang Wang
110
1
0
10 Mar 2025
Hierarchical Balance Packing: Towards Efficient Supervised Fine-tuning for Long-Context LLM
Yongqiang Yao
Jingru Tan
Kaihuan Liang
Feizhao Zhang
Yazhe Niu
Jiahao Hu
Ruihao Gong
Dahua Lin
Ningyi Xu
100
0
0
10 Mar 2025
Safety Guardrails for LLM-Enabled Robots
Zachary Ravichandran
Alexander Robey
Vijay Kumar
George Pappas
Hamed Hassani
124
5
0
10 Mar 2025
Mitigating Preference Hacking in Policy Optimization with Pessimism
Dhawal Gupta
Adam Fisch
Christoph Dann
Alekh Agarwal
123
1
0
10 Mar 2025
UC-MOA: Utility-Conditioned Multi-Objective Alignment for Distributional Pareto-Optimality
Zelei Cheng
Xin-Qiang Cai
Yuting Tang
Pushi Zhang
Boming Yang
Masashi Sugiyama
Xinyu Xing
157
0
0
10 Mar 2025
VisRL: Intention-Driven Visual Perception via Reinforced Reasoning
Zhangquan Chen
Xufang Luo
Dongsheng Li
OffRL
LRM
152
3
0
10 Mar 2025
Alignment for Efficient Tool Calling of Large Language Models
Hongshen Xu
Zihan Wang
Zichen Zhu
Lei Pan
Xingyu Chen
Lu Chen
Kai Yu
92
1
0
09 Mar 2025
Less is More: Adaptive Program Repair with Bug Localization and Preference Learning
Zhenlong Dai
Bingrui Chen
Zhuoluo Zhao
Xiu Tang
Sai Wu
Chang Yao
Zhipeng Gao
Jingyuan Chen
KELM
116
4
0
09 Mar 2025
Evaluating and Aligning Human Economic Risk Preferences in LLMs
Qingbin Liu
Yi Yang
Kar Yan Tam
125
0
0
09 Mar 2025
Life-Cycle Routing Vulnerabilities of LLM Router
Qiqi Lin
Xiaoyang Ji
Shengfang Zhai
Qingni Shen
Zhi-Li Zhang
Yuejian Fang
Yansong Gao
AAML
90
1
0
09 Mar 2025
PerturboLLaVA: Reducing Multimodal Hallucinations with Perturbative Visual Training
Cong Chen
Mingyu Liu
Chenchen Jing
Y. Zhou
Fengyun Rao
Hao Chen
Bo Zhang
Chunhua Shen
MLLM
AAML
VLM
132
5
0
09 Mar 2025
PFDial: A Structured Dialogue Instruction Fine-tuning Method Based on UML Flowcharts
Ming Zhang
Yansen Wang
Yujiong Shen
Tingyi Yang
Changhao Jiang
...
Tao Liang
Guojun Ma
Qi Zhang
Tao Gui
Xuanjing Huang
115
0
0
09 Mar 2025
From Captions to Rewards (CAREVL): Leveraging Large Language Model Experts for Enhanced Reward Modeling in Large Vision-Language Models
Muzhi Dai
Jiashuo Sun
Zhiyuan Zhao
Shixuan Liu
Rui Li
Junyu Gao
Xuelong Li
VLM
98
1
0
08 Mar 2025
Research on Superalignment Should Advance Now with Parallel Optimization of Competence and Conformity
HyunJin Kim
Xiaoyuan Yi
Jing Yao
Muhua Huang
Jinyeong Bak
James Evans
Xing Xie
97
0
0
08 Mar 2025
Language Model Personalization via Reward Factorization
Idan Shenfeld
Felix Faltings
Pulkit Agrawal
Aldo Pacchiano
109
1
0
08 Mar 2025
Intent-Aware Self-Correction for Mitigating Social Biases in Large Language Models
Panatchakorn Anantaprayoon
Masahiro Kaneko
Naoaki Okazaki
LRM
KELM
153
0
0
08 Mar 2025
Dynamically evolving segment anything model with continuous learning for medical image segmentation
Zhaori Liu
Mengyang Li
Hu Han
Enli Zhang
Shiguang Shan
Zhiming Zhao
VLM
86
0
0
08 Mar 2025
From Style to Facts: Mapping the Boundaries of Knowledge Injection with Finetuning
Eric Zhao
Pranjal Awasthi
Nika Haghtalab
81
0
0
07 Mar 2025
Knowledge Updating? No More Model Editing! Just Selective Contextual Reasoning
Guoxiu He
Xin Song
Aixin Sun
KELM
131
5
0
07 Mar 2025
Extracting and Emulsifying Cultural Explanation to Improve Multilingual Capability of LLMs
Hamin Koo
Jaehyung Kim
109
0
0
07 Mar 2025
Previous
1
2
3
...
23
24
25
...
126
127
128
Next