ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2009.01325
  4. Cited By
Learning to summarize from human feedback

Learning to summarize from human feedback

2 September 2020
Nisan Stiennon
Long Ouyang
Jeff Wu
Daniel M. Ziegler
Ryan J. Lowe
Chelsea Voss
Alec Radford
Dario Amodei
Paul Christiano
    ALM
ArXivPDFHTML

Papers citing "Learning to summarize from human feedback"

50 / 1,443 papers shown
Title
ICE-GRT: Instruction Context Enhancement by Generative Reinforcement
  based Transformers
ICE-GRT: Instruction Context Enhancement by Generative Reinforcement based Transformers
Chen Zheng
Ke Sun
Da Tang
Yukun Ma
Yuyu Zhang
Chenguang Xi
Xun Zhou
LRM
LLMAG
58
2
0
04 Jan 2024
A Mechanistic Understanding of Alignment Algorithms: A Case Study on DPO
  and Toxicity
A Mechanistic Understanding of Alignment Algorithms: A Case Study on DPO and Toxicity
Andrew Lee
Xiaoyan Bai
Itamar Pres
Martin Wattenberg
Jonathan K. Kummerfeld
Rada Mihalcea
77
104
0
03 Jan 2024
Theoretical guarantees on the best-of-n alignment policy
Theoretical guarantees on the best-of-n alignment policy
Ahmad Beirami
Alekh Agarwal
Jonathan Berant
Alex DÁmour
Jacob Eisenstein
Chirag Nagpal
A. Suresh
50
44
0
03 Jan 2024
Self-Play Fine-Tuning Converts Weak Language Models to Strong Language
  Models
Self-Play Fine-Tuning Converts Weak Language Models to Strong Language Models
Zixiang Chen
Yihe Deng
Huizhuo Yuan
Kaixuan Ji
Quanquan Gu
SyDa
53
285
0
02 Jan 2024
Uncertainty-Penalized Reinforcement Learning from Human Feedback with
  Diverse Reward LoRA Ensembles
Uncertainty-Penalized Reinforcement Learning from Human Feedback with Diverse Reward LoRA Ensembles
Yuanzhao Zhai
Han Zhang
Yu Lei
Yue Yu
Kele Xu
Dawei Feng
Bo Ding
Huaimin Wang
AI4CE
81
33
0
30 Dec 2023
Is Knowledge All Large Language Models Needed for Causal Reasoning?
Is Knowledge All Large Language Models Needed for Causal Reasoning?
Hengrui Cai
Shengjie Liu
Rui Song
LRM
ELM
30
10
0
30 Dec 2023
Olapa-MCoT: Enhancing the Chinese Mathematical Reasoning Capability of
  LLMs
Olapa-MCoT: Enhancing the Chinese Mathematical Reasoning Capability of LLMs
Shaojie Zhu
Zhaobin Wang
Chengxiang Zhuo
Hui Lu
Bo Hu
Zang Li
LRM
35
0
0
29 Dec 2023
How Far Are LLMs from Believable AI? A Benchmark for Evaluating the
  Believability of Human Behavior Simulation
How Far Are LLMs from Believable AI? A Benchmark for Evaluating the Believability of Human Behavior Simulation
Yang Xiao
Yi Cheng
Jinlan Fu
Jiashuo Wang
Wenjie Li
Pengfei Liu
LLMAG
62
4
0
28 Dec 2023
LLM4EDA: Emerging Progress in Large Language Models for Electronic
  Design Automation
LLM4EDA: Emerging Progress in Large Language Models for Electronic Design Automation
Ruizhe Zhong
Xingbo Du
Shixiong Kai
Zhentao Tang
Siyuan Xu
Hui-Ling Zhen
Jianye Hao
Qiang Xu
M. Yuan
Junchi Yan
39
35
0
28 Dec 2023
Rethinking Tabular Data Understanding with Large Language Models
Rethinking Tabular Data Understanding with Large Language Models
Tianyang Liu
Fei Wang
Muhao Chen
ReLM
LMTD
LRM
37
13
0
27 Dec 2023
Some things are more CRINGE than others: Iterative Preference
  Optimization with the Pairwise Cringe Loss
Some things are more CRINGE than others: Iterative Preference Optimization with the Pairwise Cringe Loss
Jing Xu
Andrew Lee
Sainbayar Sukhbaatar
Jason Weston
34
87
0
27 Dec 2023
Adapting Large Language Models for Education: Foundational Capabilities,
  Potentials, and Challenges
Adapting Large Language Models for Education: Foundational Capabilities, Potentials, and Challenges
Qingyao Li
Lingyue Fu
Weiming Zhang
Xianyu Chen
Jingwei Yu
Wei Xia
Weinan Zhang
Ruiming Tang
Yong Yu
AI4Ed
ELM
51
18
0
27 Dec 2023
Preference as Reward, Maximum Preference Optimization with Importance
  Sampling
Preference as Reward, Maximum Preference Optimization with Importance Sampling
Zaifan Jiang
Xing Huang
Chao Wei
36
2
0
27 Dec 2023
AutoTask: Executing Arbitrary Voice Commands by Exploring and Learning
  from Mobile GUI
AutoTask: Executing Arbitrary Voice Commands by Exploring and Learning from Mobile GUI
Lihang Pan
Bowen Wang
Chun Yu
Yuxuan Chen
Xiangyu Zhang
Yuanchun Shi
47
3
0
26 Dec 2023
Aligning Large Language Models with Human Preferences through
  Representation Engineering
Aligning Large Language Models with Human Preferences through Representation Engineering
Tianlong Li
Xiaohua Wang
Muling Wu
Changze Lv
Changze Lv
Zixuan Ling
Jianhao Zhu
Cenyuan Zhang
Xiaoqing Zheng
Xuanjing Huang
16
33
0
26 Dec 2023
Typhoon: Thai Large Language Models
Typhoon: Thai Large Language Models
Kunat Pipatanakul
Phatrasek Jirabovonvisut
Potsawee Manakul
Sittipong Sripaisarnmongkol
Ruangsak Patomwong
Pathomporn Chokchainant
Kasima Tharnpipitchai
52
16
0
21 Dec 2023
OpenRL: A Unified Reinforcement Learning Framework
OpenRL: A Unified Reinforcement Learning Framework
Shiyu Huang
Wentse Chen
Yiwen Sun
Fuqing Bie
Weijuan Tu
53
3
0
20 Dec 2023
InstructVideo: Instructing Video Diffusion Models with Human Feedback
InstructVideo: Instructing Video Diffusion Models with Human Feedback
Hangjie Yuan
Shiwei Zhang
Xiang Wang
Yujie Wei
Tao Feng
Yining Pan
Yingya Zhang
Ziwei Liu
Samuel Albanie
Dong Ni
VGen
42
42
0
19 Dec 2023
Instruct-SCTG: Guiding Sequential Controlled Text Generation through
  Instructions
Instruct-SCTG: Guiding Sequential Controlled Text Generation through Instructions
Yinhong Liu
Yixuan Su
Ehsan Shareghi
Nigel Collier
35
1
0
19 Dec 2023
HuTuMotion: Human-Tuned Navigation of Latent Motion Diffusion Models
  with Minimal Feedback
HuTuMotion: Human-Tuned Navigation of Latent Motion Diffusion Models with Minimal Feedback
Gaoge Han
Shaoli Huang
Biwei Huang
Jinglei Tang
VGen
39
2
0
19 Dec 2023
An Adaptive Placement and Parallelism Framework for Accelerating RLHF
  Training
An Adaptive Placement and Parallelism Framework for Accelerating RLHF Training
Youshao Xiao
Weichang Wu
Zhenglei Zhou
Fagui Mao
Shangchun Zhao
Lin Ju
Lei Liang
Xiaolu Zhang
Jun Zhou
34
5
0
19 Dec 2023
Explore 3D Dance Generation via Reward Model from Automatically-Ranked
  Demonstrations
Explore 3D Dance Generation via Reward Model from Automatically-Ranked Demonstrations
Zilin Wang
Hao-Wen Zhuang
Lu Li
Yinmin Zhang
Junjie Zhong
Jun Chen
Yu Yang
Boshi Tang
Zhiyong Wu
53
3
0
18 Dec 2023
Silkie: Preference Distillation for Large Visual Language Models
Silkie: Preference Distillation for Large Visual Language Models
Lei Li
Zhihui Xie
Mukai Li
Shunian Chen
Peiyi Wang
Liang Chen
Yazheng Yang
Benyou Wang
Lingpeng Kong
MLLM
117
69
0
17 Dec 2023
Policy Optimization in RLHF: The Impact of Out-of-preference Data
Policy Optimization in RLHF: The Impact of Out-of-preference Data
Ziniu Li
Tian Xu
Yang Yu
39
30
0
17 Dec 2023
Weak-to-Strong Generalization: Eliciting Strong Capabilities With Weak
  Supervision
Weak-to-Strong Generalization: Eliciting Strong Capabilities With Weak Supervision
Collin Burns
Pavel Izmailov
Jan Hendrik Kirchner
Bowen Baker
Leo Gao
...
Adrien Ecoffet
Manas Joglekar
Jan Leike
Ilya Sutskever
Jeff Wu
ELM
52
265
0
14 Dec 2023
LLM-MARS: Large Language Model for Behavior Tree Generation and
  NLP-enhanced Dialogue in Multi-Agent Robot Systems
LLM-MARS: Large Language Model for Behavior Tree Generation and NLP-enhanced Dialogue in Multi-Agent Robot Systems
Artem Lykov
Maria Dronova
Nikolay Naglov
Mikhail Litvinov
Sergei Satsevich
Artem Bazhenov
Vladimir Berman
Aleksei Shcherbak
Dzmitry Tsetserukou
LLMAG
LM&Ro
34
14
0
14 Dec 2023
Self-Evaluation Improves Selective Generation in Large Language Models
Self-Evaluation Improves Selective Generation in Large Language Models
Jie Jessie Ren
Yao-Min Zhao
Tu Vu
Peter J. Liu
Balaji Lakshminarayanan
ELM
36
34
0
14 Dec 2023
Helping or Herding? Reward Model Ensembles Mitigate but do not Eliminate
  Reward Hacking
Helping or Herding? Reward Model Ensembles Mitigate but do not Eliminate Reward Hacking
Jacob Eisenstein
Chirag Nagpal
Alekh Agarwal
Ahmad Beirami
Alex DÁmour
...
Katherine Heller
Stephen R. Pfohl
Deepak Ramachandran
Peter Shaw
Jonathan Berant
32
85
0
14 Dec 2023
Distributional Preference Learning: Understanding and Accounting for
  Hidden Context in RLHF
Distributional Preference Learning: Understanding and Accounting for Hidden Context in RLHF
Anand Siththaranjan
Cassidy Laidlaw
Dylan Hadfield-Menell
39
58
0
13 Dec 2023
Personalized Decision Supports based on Theory of Mind Modeling and
  Explainable Reinforcement Learning
Personalized Decision Supports based on Theory of Mind Modeling and Explainable Reinforcement Learning
Huao Li
Yao Fan
Keyang Zheng
Michael Lewis
Katia Sycara
30
0
0
13 Dec 2023
An Invitation to Deep Reinforcement Learning
An Invitation to Deep Reinforcement Learning
Bernhard Jaeger
Andreas Geiger
OffRL
OOD
80
5
0
13 Dec 2023
On Diversified Preferences of Large Language Model Alignment
On Diversified Preferences of Large Language Model Alignment
Dun Zeng
Yong Dai
Pengyu Cheng
Longyue Wang
Tianhao Hu
Wanshun Chen
Nan Du
Zenglin Xu
ALM
43
16
0
12 Dec 2023
Exploring Large Language Models to Facilitate Variable Autonomy for
  Human-Robot Teaming
Exploring Large Language Models to Facilitate Variable Autonomy for Human-Robot Teaming
Younes Lakhnati
Max Pascher
Jens Gerken
LLMAG
LM&Ro
42
3
0
12 Dec 2023
Frugal LMs Trained to Invoke Symbolic Solvers Achieve
  Parameter-Efficient Arithmetic Reasoning
Frugal LMs Trained to Invoke Symbolic Solvers Achieve Parameter-Efficient Arithmetic Reasoning
Subhabrata Dutta
Joykirat Singh
Ishan Pandey
Sunny Manchanda
Soumen Chakrabarti
Tanmoy Chakraborty
ReLM
LRM
34
4
0
09 Dec 2023
Is Feedback All You Need? Leveraging Natural Language Feedback in
  Goal-Conditioned Reinforcement Learning
Is Feedback All You Need? Leveraging Natural Language Feedback in Goal-Conditioned Reinforcement Learning
Sabrina McCallum
Max Taylor-Davies
Stefano V. Albrecht
Alessandro Suglia
26
1
0
07 Dec 2023
Language Model Alignment with Elastic Reset
Language Model Alignment with Elastic Reset
Michael Noukhovitch
Samuel Lavoie
Florian Strub
Aaron Courville
KELM
102
25
0
06 Dec 2023
Mitigating Open-Vocabulary Caption Hallucinations
Mitigating Open-Vocabulary Caption Hallucinations
Assaf Ben-Kish
Moran Yanuka
Morris Alper
Raja Giryes
Hadar Averbuch-Elor
MLLM
VLM
30
6
0
06 Dec 2023
PneumoLLM: Harnessing the Power of Large Language Model for
  Pneumoconiosis Diagnosis
PneumoLLM: Harnessing the Power of Large Language Model for Pneumoconiosis Diagnosis
Meiyue Song
Zhihua Yu
Weiwen Zhang
Jiarui Wang
Yuting Lu
...
Nikolaos I. Kanellakis
Jiangfeng Liu
Jing Wang
Binglu Wang
Juntao Yang
LM&MA
33
0
0
06 Dec 2023
How should the advent of large language models affect the practice of
  science?
How should the advent of large language models affect the practice of science?
Marcel Binz
Stephan Alaniz
Adina Roskies
B. Aczel
Carl T. Bergstrom
...
Emily M. Bender
M. Marelli
Matthew M. Botvinick
Zeynep Akata
Eric Schulz
39
9
0
05 Dec 2023
ULMA: Unified Language Model Alignment with Human Demonstration and
  Point-wise Preference
ULMA: Unified Language Model Alignment with Human Demonstration and Point-wise Preference
Tianchi Cai
Xierui Song
Jiyan Jiang
Fei Teng
Jinjie Gu
Guannan Zhang
ALM
21
4
0
05 Dec 2023
BEDD: The MineRL BASALT Evaluation and Demonstrations Dataset for
  Training and Benchmarking Agents that Solve Fuzzy Tasks
BEDD: The MineRL BASALT Evaluation and Demonstrations Dataset for Training and Benchmarking Agents that Solve Fuzzy Tasks
Stephanie Milani
Anssi Kanervisto
Karolis Ramanauskas
Sander Schulhoff
Brandon Houghton
Rohin Shah
25
7
0
05 Dec 2023
Axiomatic Preference Modeling for Longform Question Answering
Axiomatic Preference Modeling for Longform Question Answering
Corby Rosset
Guoqing Zheng
Victor C. Dibia
Ahmed Hassan Awadallah
Paul Bennett
SyDa
32
3
0
02 Dec 2023
RLHF and IIA: Perverse Incentives
RLHF and IIA: Perverse Incentives
Wanqiao Xu
Shi Dong
Xiuyuan Lu
Grace Lam
Zheng Wen
Benjamin Van Roy
37
2
0
02 Dec 2023
Nash Learning from Human Feedback
Nash Learning from Human Feedback
Rémi Munos
Michal Valko
Daniele Calandriello
M. G. Azar
Mark Rowland
...
Nikola Momchev
Olivier Bachem
D. Mankowitz
Doina Precup
Bilal Piot
49
126
0
01 Dec 2023
SeaLLMs -- Large Language Models for Southeast Asia
SeaLLMs -- Large Language Models for Southeast Asia
Xuan-Phi Nguyen
Wenxuan Zhang
Xin Li
Mahani Aljunied
Zhiqiang Hu
...
Yue Deng
Sen Yang
Chaoqun Liu
Hang Zhang
Li Bing
LRM
42
74
0
01 Dec 2023
Sample Efficient Preference Alignment in LLMs via Active Exploration
Sample Efficient Preference Alignment in LLMs via Active Exploration
Viraj Mehta
Vikramjeet Das
Ojash Neopane
Yijia Dai
Ilija Bogunovic
Ilija Bogunovic
Willie Neiswanger
Stefano Ermon
Jeff Schneider
Willie Neiswanger
OffRL
50
12
0
01 Dec 2023
AlignBench: Benchmarking Chinese Alignment of Large Language Models
AlignBench: Benchmarking Chinese Alignment of Large Language Models
Xiao Liu
Xuanyu Lei
Sheng-Ping Wang
Yue Huang
Zhuoer Feng
...
Hongning Wang
Jing Zhang
Minlie Huang
Yuxiao Dong
Jie Tang
ELM
LM&MA
ALM
125
43
0
30 Nov 2023
LMRL Gym: Benchmarks for Multi-Turn Reinforcement Learning with Language
  Models
LMRL Gym: Benchmarks for Multi-Turn Reinforcement Learning with Language Models
Marwa Abdulhai
Isadora White
Charles Burton Snell
Charles Sun
Joey Hong
Yuexiang Zhai
Kelvin Xu
Sergey Levine
LLMAG
OffRL
LRM
39
31
0
30 Nov 2023
Unveiling the Implicit Toxicity in Large Language Models
Unveiling the Implicit Toxicity in Large Language Models
Jiaxin Wen
Pei Ke
Hao Sun
Zhexin Zhang
Chengfei Li
Jinfeng Bai
Minlie Huang
42
26
0
29 Nov 2023
MoDS: Model-oriented Data Selection for Instruction Tuning
MoDS: Model-oriented Data Selection for Instruction Tuning
Qianlong Du
Chengqing Zong
Jiajun Zhang
ALM
28
78
0
27 Nov 2023
Previous
123...171819...272829
Next