ResearchTrend.AI
  • Papers
  • Communities
  • Organizations
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2203.02155
  4. Cited By
Training language models to follow instructions with human feedback

Training language models to follow instructions with human feedback

4 March 2022
Long Ouyang
Jeff Wu
Xu Jiang
Diogo Almeida
Carroll L. Wainwright
Pamela Mishkin
Chong Zhang
Sandhini Agarwal
Katarina Slama
Alex Ray
John Schulman
Jacob Hilton
Fraser Kelton
Luke E. Miller
Maddie Simens
Amanda Askell
Peter Welinder
Paul Christiano
Jan Leike
Ryan J. Lowe
    OSLMALM
ArXiv (abs)PDFHTML

Papers citing "Training language models to follow instructions with human feedback"

50 / 6,404 papers shown
Title
Generative Reward Models
Generative Reward Models
Dakota Mahan
Duy Phung
Rafael Rafailov
Chase Blagden
Nathan Lile
Louis Castricato
Jan-Philipp Fränken
Chelsea Finn
Alon Albalak
VLMSyDaOffRL
88
42
0
02 Oct 2024
Investigating on RLHF methodology
Investigating on RLHF methodology
Alexey Kutalev
Sergei Markoff
45
0
0
02 Oct 2024
Open-RAG: Enhanced Retrieval-Augmented Reasoning with Open-Source Large
  Language Models
Open-RAG: Enhanced Retrieval-Augmented Reasoning with Open-Source Large Language Models
Shayekh Bin Islam
Md Asib Rahman
K S M Tozammel Hossain
Enamul Hoque
Shafiq Joty
Md. Rizwan Parvez
RALMAIFinLRMVLM
86
16
0
02 Oct 2024
PreND: Enhancing Intrinsic Motivation in Reinforcement Learning through
  Pre-trained Network Distillation
PreND: Enhancing Intrinsic Motivation in Reinforcement Learning through Pre-trained Network Distillation
Mohammadamin Davoodabadi
Negin Hashemi Dijujin
M. Baghshah
63
0
0
02 Oct 2024
Evaluating Robustness of Reward Models for Mathematical Reasoning
Evaluating Robustness of Reward Models for Mathematical Reasoning
Sunghwan Kim
Dongjin Kang
Taeyoon Kwon
Hyungjoo Chae
Jungsoo Won
Dongha Lee
Jinyoung Yeo
81
7
0
02 Oct 2024
FactAlign: Long-form Factuality Alignment of Large Language Models
FactAlign: Long-form Factuality Alignment of Large Language Models
Chao-Wei Huang
Yun-Nung Chen
HILM
79
4
0
02 Oct 2024
Upcycling Instruction Tuning from Dense to Mixture-of-Experts via
  Parameter Merging
Upcycling Instruction Tuning from Dense to Mixture-of-Experts via Parameter Merging
Tingfeng Hui
Zhenyu Zhang
Shuohuan Wang
Yu Sun
Hua Wu
Sen Su
MoE
84
0
0
02 Oct 2024
Integrative Decoding: Improve Factuality via Implicit Self-consistency
Integrative Decoding: Improve Factuality via Implicit Self-consistency
Yi Cheng
Xiao Liang
Yeyun Gong
Wen Xiao
Song Wang
...
Wenjie Li
Jian Jiao
Qi Chen
Peng Cheng
Wayne Xiong
HILM
208
3
0
02 Oct 2024
Can We Further Elicit Reasoning in LLMs? Critic-Guided Planning with
  Retrieval-Augmentation for Solving Challenging Tasks
Can We Further Elicit Reasoning in LLMs? Critic-Guided Planning with Retrieval-Augmentation for Solving Challenging Tasks
Xingxuan Li
Weiwen Xu
Ruochen Zhao
Fangkai Jiao
Shafiq Joty
Lidong Bing
LRM
119
11
0
02 Oct 2024
FlipAttack: Jailbreak LLMs via Flipping
FlipAttack: Jailbreak LLMs via Flipping
Yue Liu
Xiaoxin He
Miao Xiong
Jinlan Fu
Shumin Deng
Bryan Hooi
AAML
105
17
0
02 Oct 2024
FlashMask: Efficient and Rich Mask Extension of FlashAttention
FlashMask: Efficient and Rich Mask Extension of FlashAttention
Guoxia Wang
Jinle Zeng
Xiyuan Xiao
Siming Wu
Jiabin Yang
Lujing Zheng
Zeyu Chen
Jiang Bian
Dianhai Yu
Haifeng Wang
390
3
0
02 Oct 2024
Moral Alignment for LLM Agents
Moral Alignment for LLM Agents
Elizaveta Tennant
Stephen Hailes
Mirco Musolesi
145
8
0
02 Oct 2024
Reasoning Elicitation in Language Models via Counterfactual Feedback
Reasoning Elicitation in Language Models via Counterfactual Feedback
Alihan Hüyük
Xinnuo Xu
Jacqueline R. M. A. Maasch
Aditya V. Nori
Javier González
ReLMLRM
449
3
0
02 Oct 2024
Step-by-Step Reasoning for Math Problems via Twisted Sequential Monte Carlo
Step-by-Step Reasoning for Math Problems via Twisted Sequential Monte Carlo
Shengyu Feng
Xiang Kong
Shuang Ma
Aonan Zhang
Dong Yin
Chong-Jun Wang
Ruoming Pang
Yiming Yang
LRM
120
2
0
02 Oct 2024
Seeing Eye to AI: Human Alignment via Gaze-Based Response Rewards for Large Language Models
Seeing Eye to AI: Human Alignment via Gaze-Based Response Rewards for Large Language Models
Angela Lopez-Cardona
Carlos Segura
Alexandros Karatzoglou
Sergi Abadal
Ioannis Arapakis
ALM
177
4
0
02 Oct 2024
Endless Jailbreaks with Bijection Learning
Endless Jailbreaks with Bijection Learning
Brian R. Y. Huang
Maximilian Li
Leonard Tang
AAML
185
8
0
02 Oct 2024
LASeR: Learning to Adaptively Select Reward Models with Multi-Armed Bandits
LASeR: Learning to Adaptively Select Reward Models with Multi-Armed Bandits
Duy Nguyen
Archiki Prasad
Elias Stengel-Eskin
Joey Tianyi Zhou
51
3
0
02 Oct 2024
Frozen Large Language Models Can Perceive Paralinguistic Aspects of Speech
Frozen Large Language Models Can Perceive Paralinguistic Aspects of Speech
Wonjune Kang
Junteng Jia
Chunyang Wu
Wei Zhou
Egor Lakomkin
...
Leda Sari
Suyoun Kim
Ke Li
Jay Mahadeokar
Ozlem Kalinli
AuLLM
131
6
0
02 Oct 2024
Auction-Based Regulation for Artificial Intelligence
Auction-Based Regulation for Artificial Intelligence
Marco Bornstein
Zora Che
Suhas Julapalli
Abdirisak Mohamed
Amrit Singh Bedi
Furong Huang
113
1
0
02 Oct 2024
Bridging Context Gaps: Leveraging Coreference Resolution for Long Contextual Understanding
Bridging Context Gaps: Leveraging Coreference Resolution for Long Contextual Understanding
Yanming Liu
Xinyue Peng
Jiannan Cao
Shi Bo
Yanxin Shen
Tianyu Du
Sheng Cheng
Xun Wang
Jianwei Yin
Xuhong Zhang
156
9
0
02 Oct 2024
Mixing It Up: The Cocktail Effect of Multi-Task Fine-Tuning on LLM
  Performance -- A Case Study in Finance
Mixing It Up: The Cocktail Effect of Multi-Task Fine-Tuning on LLM Performance -- A Case Study in Finance
Meni Brief
Oded Ovadia
Gil Shenderovitz
Noga Ben Yoash
Rachel Lemberg
Eitam Sheetrit
87
4
0
01 Oct 2024
Interactive Explainable Anomaly Detection for Industrial Settings
Interactive Explainable Anomaly Detection for Industrial Settings
Daniel Gramelt
Timon Höfer
Ute Schmid
AAMLHAI
133
1
0
01 Oct 2024
FlipGuard: Defending Preference Alignment against Update Regression with
  Constrained Optimization
FlipGuard: Defending Preference Alignment against Update Regression with Constrained Optimization
Mingye Zhu
Yi Liu
Quan Wang
Junbo Guo
Zhendong Mao
65
1
0
01 Oct 2024
AlignSum: Data Pyramid Hierarchical Fine-tuning for Aligning with Human
  Summarization Preference
AlignSum: Data Pyramid Hierarchical Fine-tuning for Aligning with Human Summarization Preference
Yang Han
Yiming Wang
Rui Wang
Lu Chen
Kai Yu
AI4TSALM
65
2
0
01 Oct 2024
AHA: A Vision-Language-Model for Detecting and Reasoning Over Failures
  in Robotic Manipulation
AHA: A Vision-Language-Model for Detecting and Reasoning Over Failures in Robotic Manipulation
Jiafei Duan
Wilbert Pumacay
Nishanth Kumar
Yi Ru Wang
Shulin Tian
Wentao Yuan
Ranjay Krishna
Dieter Fox
Ajay Mandlekar
Yijie Guo
VLMLRM
123
29
0
01 Oct 2024
FedPT: Federated Proxy-Tuning of Large Language Models on
  Resource-Constrained Edge Devices
FedPT: Federated Proxy-Tuning of Large Language Models on Resource-Constrained Edge Devices
Zhidong Gao
Yu Zhang
Zhenxiao Zhang
Yanmin Gong
Yuanxiong Guo
71
1
0
01 Oct 2024
PclGPT: A Large Language Model for Patronizing and Condescending
  Language Detection
PclGPT: A Large Language Model for Patronizing and Condescending Language Detection
Hongbo Wang
Mingda Li
Junyu Lu
Hebin Xia
Liang Yang
Bo Xu
Ruizhu Liu
Hongfei Lin
68
0
0
01 Oct 2024
A Taxonomy of Loss Functions for Stochastic Optimal Control
A Taxonomy of Loss Functions for Stochastic Optimal Control
Carles Domingo-Enrich
86
4
0
01 Oct 2024
Uncertainty-aware Reward Model: Teaching Reward Models to Know What is Unknown
Uncertainty-aware Reward Model: Teaching Reward Models to Know What is Unknown
Xingzhou Lou
Dong Yan
Wei Shen
Yuzi Yan
Jian Xie
Junge Zhang
220
28
0
01 Oct 2024
Scaling Offline Model-Based RL via Jointly-Optimized World-Action Model Pretraining
Scaling Offline Model-Based RL via Jointly-Optimized World-Action Model Pretraining
Jie Cheng
Ruixi Qiao
Gang Xiong
Binhua Li
Yingwei Ma
Binhua Li
Yongbin Li
Yisheng Lv
OffRLOnRLLM&Ro
151
4
0
01 Oct 2024
PrivTuner with Homomorphic Encryption and LoRA: A P3EFT Scheme for Privacy-Preserving Parameter-Efficient Fine-Tuning of AI Foundation Models
PrivTuner with Homomorphic Encryption and LoRA: A P3EFT Scheme for Privacy-Preserving Parameter-Efficient Fine-Tuning of AI Foundation Models
Yang Li
Wenhan Yu
Jun Zhao
104
3
0
01 Oct 2024
DoPAMine: Domain-specific Pre-training Adaptation from seed-guided data
  Mining
DoPAMine: Domain-specific Pre-training Adaptation from seed-guided data Mining
Vinayak Arannil
Neha Narwal
Sourav Sanjukta Bhabesh
Sai Nikhil Thirandas
Darren Yow-Bang Wang
Graham Horwood
Alex Anto Chirayath
Gouri Pandeshwar
121
0
0
30 Sep 2024
ACE: All-round Creator and Editor Following Instructions via Diffusion
  Transformer
ACE: All-round Creator and Editor Following Instructions via Diffusion Transformer
Zhen Han
Zeyinzi Jiang
Yulin Pan
Jingfeng Zhang
Chaojie Mao
Chenwei Xie
Yu Liu
Jingren Zhou
DiffM
108
21
0
30 Sep 2024
Unsupervised Human Preference Learning
Unsupervised Human Preference Learning
Sumuk Shashidhar
Abhinav Chinta
Vaibhav Sahai
Dilek Hakkani Tur
LRM
113
2
0
30 Sep 2024
Wait, but Tylenol is Acetaminophen... Investigating and Improving
  Language Models' Ability to Resist Requests for Misinformation
Wait, but Tylenol is Acetaminophen... Investigating and Improving Language Models' Ability to Resist Requests for Misinformation
Shan Chen
Mingye Gao
Kuleen Sasse
Thomas Hartvigsen
Brian Anthony
Lizhou Fan
Hugo J. W. L. Aerts
Jack Gallifant
Danielle S. Bitterman
LM&MA
89
1
0
30 Sep 2024
The Perfect Blend: Redefining RLHF with Mixture of Judges
The Perfect Blend: Redefining RLHF with Mixture of Judges
Tengyu Xu
Eryk Helenowski
Karthik Abinav Sankararaman
Di Jin
Kaiyan Peng
...
Gabriel Cohen
Yuandong Tian
Hao Ma
Sinong Wang
Han Fang
144
14
0
30 Sep 2024
Do Influence Functions Work on Large Language Models?
Do Influence Functions Work on Large Language Models?
Zhe Li
Wei Zhao
Yige Li
Jun Sun
TDI
94
3
0
30 Sep 2024
CONTESTS: a Framework for Consistency Testing of Span Probabilities in
  Language Models
CONTESTS: a Framework for Consistency Testing of Span Probabilities in Language Models
Eitan Wagner
Yuli Slavutsky
Omri Abend
88
1
0
30 Sep 2024
HDMoLE: Mixture of LoRA Experts with Hierarchical Routing and Dynamic Thresholds for Fine-Tuning LLM-based ASR Models
HDMoLE: Mixture of LoRA Experts with Hierarchical Routing and Dynamic Thresholds for Fine-Tuning LLM-based ASR Models
Bingshen Mu
Kun Wei
Qijie Shao
Yong Xu
Lei Xie
MoE
120
2
0
30 Sep 2024
Task-Agnostic Pre-training and Task-Guided Fine-tuning for Versatile Diffusion Planner
Task-Agnostic Pre-training and Task-Guided Fine-tuning for Versatile Diffusion Planner
Chenyou Fan
Chenjia Bai
Zhao Shan
Haoran He
Yang Zhang
Zhen Wang
120
3
0
30 Sep 2024
Can Models Learn Skill Composition from Examples?
Can Models Learn Skill Composition from Examples?
Haoyu Zhao
Simran Kaur
Dingli Yu
Anirudh Goyal
Sanjeev Arora
CoGeMoE
139
8
0
29 Sep 2024
Mitigating the Negative Impact of Over-association for Conversational
  Query Production
Mitigating the Negative Impact of Over-association for Conversational Query Production
Ante Wang
Linfeng Song
Zijun Min
Ge Xu
Xiaoli Wang
Junfeng Yao
Jinsong Su
128
1
0
29 Sep 2024
A Critical Look at Meta-evaluating Summarisation Evaluation Metrics
A Critical Look at Meta-evaluating Summarisation Evaluation Metrics
Xiang Dai
Sarvnaz Karimi
Biaoyan Fang
71
0
0
29 Sep 2024
MedHalu: Hallucinations in Responses to Healthcare Queries by Large
  Language Models
MedHalu: Hallucinations in Responses to Healthcare Queries by Large Language Models
Vibhor Agarwal
Yiqiao Jin
Mohit Chandra
Munmun De Choudhury
Srijan Kumar
Nishanth R. Sastry
HILMLM&MA
123
7
0
29 Sep 2024
Can Large Language Models Analyze Graphs like Professionals? A Benchmark, Datasets and Models
Can Large Language Models Analyze Graphs like Professionals? A Benchmark, Datasets and Models
Xin Sky Li
Weize Chen
Qizhi Chu
Haopeng Li
Zhaojun Sun
...
Yiwei Wei
Zhiyuan Liu
Chuan Shi
Maosong Sun
Cheng Yang
129
6
0
29 Sep 2024
HybridFlow: A Flexible and Efficient RLHF Framework
HybridFlow: A Flexible and Efficient RLHF Framework
Guangming Sheng
Chi Zhang
Zilingfeng Ye
Xibin Wu
Wang Zhang
Ru Zhang
Size Zheng
Haibin Lin
Chuan Wu
AI4CE
241
240
0
28 Sep 2024
Revisiting the Superficial Alignment Hypothesis
Revisiting the Superficial Alignment Hypothesis
Mohit Raghavendra
Vaskar Nath
Sean Hendryx
LRM
51
3
0
27 Sep 2024
Responsible AI in Open Ecosystems: Reconciling Innovation with Risk
  Assessment and Disclosure
Responsible AI in Open Ecosystems: Reconciling Innovation with Risk Assessment and Disclosure
Mahasweta Chakraborti
Bert Joseph Prestoza
Nicholas Vincent
Seth Frey
88
1
0
27 Sep 2024
Outlining the Borders for LLM Applications in Patient Education:
  Developing an Expert-in-the-Loop LLM-Powered Chatbot for Prostate Cancer
  Patient Education
Outlining the Borders for LLM Applications in Patient Education: Developing an Expert-in-the-Loop LLM-Powered Chatbot for Prostate Cancer Patient Education
Yuexing Hao
J. Holmes
Mark Waddle
N. Yu
Kirstin Vickers
...
Corinna E. Löckenhoff
Aditya Vashistha
Marzyeh Ghassemi
Saleh Kalantari
Wei Liu
46
2
0
27 Sep 2024
Ruler: A Model-Agnostic Method to Control Generated Length for Large
  Language Models
Ruler: A Model-Agnostic Method to Control Generated Length for Large Language Models
Jiaming Li
Lei Zhang
Yunshui Li
Ziqiang Liu
Yuelin Bai
Run Luo
Longze Chen
Min Yang
ALM
49
0
0
27 Sep 2024
Previous
123...484950...127128129
Next