ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2203.02155
  4. Cited By
Training language models to follow instructions with human feedback

Training language models to follow instructions with human feedback

4 March 2022
Long Ouyang
Jeff Wu
Xu Jiang
Diogo Almeida
Carroll L. Wainwright
Pamela Mishkin
Chong Zhang
Sandhini Agarwal
Katarina Slama
Alex Ray
John Schulman
Jacob Hilton
Fraser Kelton
Luke E. Miller
Maddie Simens
Amanda Askell
Peter Welinder
Paul Christiano
Jan Leike
Ryan J. Lowe
    OSLM
    ALM
ArXivPDFHTML

Papers citing "Training language models to follow instructions with human feedback"

50 / 7,311 papers shown
Title
PFDial: A Structured Dialogue Instruction Fine-tuning Method Based on UML Flowcharts
Ming Zhang
Yansen Wang
Yujiong Shen
Tingyi Yang
Changhao Jiang
...
Tao Liang
Guojun Ma
Qi Zhang
Tao Gui
Xuanjing Huang
65
0
0
09 Mar 2025
Less is More: Adaptive Program Repair with Bug Localization and Preference Learning
Zhenlong Dai
Bingrui Chen
Zhuoluo Zhao
Xiu Tang
Sai Wu
Chang Yao
Zhipeng Gao
Jingyuan Chen
KELM
59
2
0
09 Mar 2025
From Captions to Rewards (CAREVL): Leveraging Large Language Model Experts for Enhanced Reward Modeling in Large Vision-Language Models
Muzhi Dai
Jiashuo Sun
Zhiyuan Zhao
Shixuan Liu
Rui Li
Junyu Gao
Xuelong Li
VLM
58
1
0
08 Mar 2025
Intent-Aware Self-Correction for Mitigating Social Biases in Large Language Models
Panatchakorn Anantaprayoon
Masahiro Kaneko
Naoaki Okazaki
LRM
KELM
60
0
0
08 Mar 2025
Language Model Personalization via Reward Factorization
Idan Shenfeld
Felix Faltings
Pulkit Agrawal
Aldo Pacchiano
52
1
0
08 Mar 2025
Research on Superalignment Should Advance Now with Parallel Optimization of Competence and Conformity
HyunJin Kim
Xiaoyuan Yi
Jing Yao
Muhua Huang
Jinyeong Bak
James Evans
Xing Xie
49
0
0
08 Mar 2025
Dynamically evolving segment anything model with continuous learning for medical image segmentation
Zhaori Liu
Mengyang Li
Hu Han
Enli Zhang
Shiguang Shan
Zhiming Zhao
VLM
57
0
0
08 Mar 2025
SANDWiCH: Semantical Analysis of Neighbours for Disambiguating Words in Context ad Hoc
Daniel Guzman-Olivares
Lara Quijano-Sanchez
Federico Liberatore
41
0
0
07 Mar 2025
Every FLOP Counts: Scaling a 300B Mixture-of-Experts LING LLM without Premium GPUs
Ling Team
B. Zeng
Chenyu Huang
Chao Zhang
Changxin Tian
...
Zhaoxin Huan
Zujie Wen
Zhenhang Sun
Zhuoxuan Du
Z. He
MoE
ALM
111
2
0
07 Mar 2025
Superintelligence Strategy: Expert Version
Superintelligence Strategy: Expert Version
Dan Hendrycks
Eric Schmidt
Alexandr Wang
69
1
0
07 Mar 2025
Knowledge Updating? No More Model Editing! Just Selective Contextual Reasoning
Guoxiu He
Xin Song
Aixin Sun
KELM
77
3
0
07 Mar 2025
Similarity-Based Domain Adaptation with LLMs
Jie He
Wendi Zhou
Xiang Li
Jeff Z. Pan
58
0
0
07 Mar 2025
Dynamic Knowledge Integration for Evidence-Driven Counter-Argument Generation with Large Language Models
Anar Yeginbergen
Maite Oronoz
Rodrigo Agerri
65
0
0
07 Mar 2025
Conformal Prediction for Image Segmentation Using Morphological Prediction Sets
Luca Mossina
Corentin Friedrich
MedIm
69
2
0
07 Mar 2025
Revitalizing Saturated Benchmarks: A Weighted Metric Approach for Differentiating Large Language Model Performance
Bryan Etzine
Masoud Hashemi
Nishanth Madhusudhan
Sagar Davasam
Roshnee Sharma
Sathwik Tejaswi Madhusudhan
Vikas Yadav
44
0
0
07 Mar 2025
Adversarial Policy Optimization for Offline Preference-based Reinforcement Learning
Hyungkyu Kang
Min-hwan Oh
OffRL
55
0
0
07 Mar 2025
Extracting and Emulsifying Cultural Explanation to Improve Multilingual Capability of LLMs
Hamin Koo
Jaehyung Kim
53
0
0
07 Mar 2025
Soft Policy Optimization: Online Off-Policy RL for Sequence Models
Taco Cohen
David W. Zhang
Kunhao Zheng
Yunhao Tang
Rémi Munos
Gabriel Synnaeve
OffRL
88
0
0
07 Mar 2025
From Style to Facts: Mapping the Boundaries of Knowledge Injection with Finetuning
Eric Zhao
Pranjal Awasthi
Nika Haghtalab
58
0
0
07 Mar 2025
DiffPO: Diffusion-styled Preference Optimization for Efficient Inference-Time Alignment of Large Language Models
Ruizhe Chen
Wenhao Chai
Zhifei Yang
Xiaotian Zhang
Qiufeng Wang
Tony Q.S. Quek
Soujanya Poria
Zuozhu Liu
55
0
0
06 Mar 2025
Implicit Cross-Lingual Rewarding for Efficient Multilingual Preference Alignment
Wen Yang
Junhong Wu
Chen Wang
Chengqing Zong
J.N. Zhang
77
1
0
06 Mar 2025
Enhancing SAM with Efficient Prompting and Preference Optimization for Semi-supervised Medical Image Segmentation
Aishik Konwer
Zhijian Yang
Erhan Bas
Cao Xiao
Prateek Prasanna
Parminder Bhatia
Taha A. Kass-Hout
MedIm
VLM
78
0
0
06 Mar 2025
Mixed Likelihood Variational Gaussian Processes
Kaiwen Wu
Craig Sanders
Benjamin Letham
Phillip Guan
82
0
0
06 Mar 2025
An Empirical Study on Eliciting and Improving R1-like Reasoning Models
Z. Chen
Yingqian Min
Beichen Zhang
Jie Chen
Jinhao Jiang
...
Xu Miao
Yaojie Lu
Lei Fang
Zhongyuan Wang
Zhicheng Dou
ReLM
OffRL
LRM
83
20
0
06 Mar 2025
Lost in Literalism: How Supervised Training Shapes Translationese in LLMs
Yafu Li
Ronghao Zhang
Zhilin Wang
Huajian Zhang
Leyang Cui
Yongjing Yin
Tong Xiao
Yue Zhang
78
0
0
06 Mar 2025
Towards Autonomous Reinforcement Learning for Real-World Robotic Manipulation with Large Language Models
Niccolò Turcato
Matteo Iovino
Aris Synodinos
Alberto Dalla Libera
R. Carli
Pietro Falco
LM&Ro
45
0
0
06 Mar 2025
TIMER: Temporal Instruction Modeling and Evaluation for Longitudinal Clinical Records
Hejie Cui
Alyssa Unell
Bowen Chen
Jason Alan Fries
Emily Alsentzer
Sanmi Koyejo
N. Shah
86
0
0
06 Mar 2025
Adding Alignment Control to Language Models
Wenhong Zhu
Weinan Zhang
Rui Wang
65
0
0
06 Mar 2025
IFIR: A Comprehensive Benchmark for Evaluating Instruction-Following in Expert-Domain Information Retrieval
Tingyu Song
Guo Gan
Mingsheng Shang
Yilun Zhao
VLM
73
1
0
06 Mar 2025
Underlying Semantic Diffusion for Effective and Efficient In-Context Learning
Zhong Ji
Weilong Cao
Yan Zhang
Yanwei Pang
Jungong Han
Xuelong Li
DiffM
VLM
52
0
0
06 Mar 2025
Robust Data Watermarking in Language Models by Injecting Fictitious Knowledge
Xinyue Cui
Johnny Tian-Zheng Wei
Swabha Swayamdipta
Robin Jia
WaLM
96
1
0
06 Mar 2025
Talking Back -- human input and explanations to interactive AI systems
Alan Dix
Tommaso Turchi
Ben Wilson
Anna Monreale
Matt Roach
56
1
0
06 Mar 2025
Uncovering Gaps in How Humans and LLMs Interpret Subjective Language
Erik Jones
Arjun Patrawala
Jacob Steinhardt
49
0
0
06 Mar 2025
High-Precision Transformer-Based Visual Servoing for Humanoid Robots in Aligning Tiny Objects
Jialong Xue
Wei Gao
Yu Wang
Chao Ji
Dongdong Zhao
Shi Yan
Shiwu Zhang
47
0
0
06 Mar 2025
CodeIF-Bench: Evaluating Instruction-Following Capabilities of Large Language Models in Interactive Code Generation
CodeIF-Bench: Evaluating Instruction-Following Capabilities of Large Language Models in Interactive Code Generation
Peiding Wang
Lihe Zhang
Fang Liu
Lin Shi
Minxiao Li
Bo Shen
An Fu
ELM
LRM
211
1
0
05 Mar 2025
Improving LLM Safety Alignment with Dual-Objective Optimization
Xuandong Zhao
Will Cai
Tianneng Shi
David Huang
Licong Lin
Song Mei
Dawn Song
AAML
MU
72
1
0
05 Mar 2025
Unified Mind Model: Reimagining Autonomous Agents in the LLM Era
Pengbo Hu
Xiang Ying
LLMAG
LM&Ro
AI4CE
72
0
0
05 Mar 2025
Human Implicit Preference-Based Policy Fine-tuning for Multi-Agent Reinforcement Learning in USV Swarm
Haksub Kim
Kanghoon Lee
J. Park
Jiachen Li
Jinkyoo Park
64
1
0
05 Mar 2025
Addressing Overprescribing Challenges: Fine-Tuning Large Language Models for Medication Recommendation Tasks
Zihao Zhao
Chenxiao Fan
Chongming Gao
Fuli Feng
Xiangnan He
LM&MA
AI4MH
80
0
0
05 Mar 2025
MAS-GPT: Training LLMs to Build LLM-based Multi-Agent Systems
Guangyi Liu
Shuo Tang
Rui Ge
Yaxin Du
Zhenfei Yin
Tian Jin
Jing Shao
LLMAG
95
1
0
05 Mar 2025
SafeVLA: Towards Safety Alignment of Vision-Language-Action Model via Safe Reinforcement Learning
Borong Zhang
Yuhao Zhang
Yalan Qin
Yingshan Lei
Josef Dai
Yuanpei Chen
Yaodong Yang
66
4
0
05 Mar 2025
Extrapolation Merging: Keep Improving With Extrapolation and Merging
Yiguan Lin
Bin Xu
Yinghao Li
Yang Gao
MoMe
59
1
0
05 Mar 2025
Visualising Policy-Reward Interplay to Inform Zeroth-Order Preference Optimisation of Large Language Models
Alessio Galatolo
Zhenbang Dai
Katie Winkle
Meriem Beloucif
60
0
0
05 Mar 2025
Deep Causal Behavioral Policy Learning: Applications to Healthcare
Jonas Knecht
Anna Zink
Jonathan Kolstad
Maya Petersen
CML
88
0
0
05 Mar 2025
FANS -- Formal Answer Selection for Natural Language Math Reasoning Using Lean4
Jiarui Yao
Ruida Wang
Tong Zhang
LRM
67
0
0
05 Mar 2025
Can Frontier LLMs Replace Annotators in Biomedical Text Mining? Analyzing Challenges and Exploring Solutions
Can Frontier LLMs Replace Annotators in Biomedical Text Mining? Analyzing Challenges and Exploring Solutions
Yichong Zhao
Susumu Goto
67
0
0
05 Mar 2025
Token-Level Privacy in Large Language Models
Reém Harel
Niv Gilboa
Yuval Pinter
53
0
0
05 Mar 2025
Adversarial Training for Multimodal Large Language Models against Jailbreak Attacks
Adversarial Training for Multimodal Large Language Models against Jailbreak Attacks
Liming Lu
Shuchao Pang
Siyuan Liang
Haotian Zhu
Xiyu Zeng
Aishan Liu
Yunhuai Liu
Yongbin Zhou
AAML
58
2
0
05 Mar 2025
Rebalanced Multimodal Learning with Data-aware Unimodal Sampling
Qingyuan Jiang
Zhouyang Chi
Xiao Ma
Qirong Mao
Yang Yang
Jinhui Tang
67
0
0
05 Mar 2025
Generative Artificial Intelligence in Robotic Manipulation: A Survey
Kun Zhang
Peng Yun
Jun Cen
Junhao Cai
DiDi Zhu
...
Qifeng Chen
Jia Pan
Wei Zhang
Bo Yang
Hua Chen
64
1
0
05 Mar 2025
Previous
123...141516...145146147
Next