ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2009.01325
  4. Cited By
Learning to summarize from human feedback

Learning to summarize from human feedback

2 September 2020
Nisan Stiennon
Long Ouyang
Jeff Wu
Daniel M. Ziegler
Ryan J. Lowe
Chelsea Voss
Alec Radford
Dario Amodei
Paul Christiano
    ALM
ArXivPDFHTML

Papers citing "Learning to summarize from human feedback"

50 / 1,443 papers shown
Title
Pedagogical Alignment of Large Language Models
Pedagogical Alignment of Large Language Models
Shashank Sonkar
Kangqi Ni
Sapana Chaudhary
Richard G. Baraniuk
AI4Ed
18
7
0
07 Feb 2024
Direct Language Model Alignment from Online AI Feedback
Direct Language Model Alignment from Online AI Feedback
Shangmin Guo
Biao Zhang
Tianlin Liu
Tianqi Liu
Misha Khalman
...
Thomas Mesnard
Yao-Min Zhao
Bilal Piot
Johan Ferret
Mathieu Blondel
ALM
42
134
0
07 Feb 2024
TransLLaMa: LLM-based Simultaneous Translation System
TransLLaMa: LLM-based Simultaneous Translation System
Roman Koshkin
Katsuhito Sudoh
Satoshi Nakamura
29
21
0
07 Feb 2024
MusicRL: Aligning Music Generation to Human Preferences
MusicRL: Aligning Music Generation to Human Preferences
Geoffrey Cideron
Sertan Girgin
Mauro Verzetti
Damien Vincent
Matej Kastelic
...
Olivier Pietquin
Matthieu Geist
Léonard Hussenot
Neil Zeghidour
A. Agostinelli
45
17
0
06 Feb 2024
Harnessing the Plug-and-Play Controller by Prompting
Harnessing the Plug-and-Play Controller by Prompting
Hao Wang
Lei Sha
34
3
0
06 Feb 2024
Personalized Language Modeling from Personalized Human Feedback
Personalized Language Modeling from Personalized Human Feedback
Xinyu Li
Zachary C. Lipton
Liu Leqi
ALM
76
48
0
06 Feb 2024
Learning to Generate Explainable Stock Predictions using Self-Reflective
  Large Language Models
Learning to Generate Explainable Stock Predictions using Self-Reflective Large Language Models
Kelvin J.L. Koa
Yunshan Ma
Ritchie Ng
Tat-Seng Chua
AIFin
LLMAG
47
25
0
06 Feb 2024
Toward Human-AI Alignment in Large-Scale Multi-Player Games
Toward Human-AI Alignment in Large-Scale Multi-Player Games
Sugandha Sharma
Guy Davidson
Khimya Khetarpal
Anssi Kanervisto
Udit Arora
Katja Hofmann
Ida Momennejad
35
0
0
05 Feb 2024
Nevermind: Instruction Override and Moderation in Large Language Models
Nevermind: Instruction Override and Moderation in Large Language Models
Edward Kim
ALM
26
0
0
05 Feb 2024
Preference-Conditioned Language-Guided Abstraction
Preference-Conditioned Language-Guided Abstraction
Andi Peng
Andreea Bobu
Belinda Z. Li
T. Sumers
Ilia Sucholutsky
Nishanth Kumar
Thomas Griffiths
Julie A. Shah
42
12
0
05 Feb 2024
Decoding-time Realignment of Language Models
Decoding-time Realignment of Language Models
Tianlin Liu
Shangmin Guo
Leonardo Bianco
Daniele Calandriello
Quentin Berthet
Felipe Llinares-López
Jessica Hoffmann
Lucas Dixon
Michal Valko
Mathieu Blondel
AI4CE
54
37
0
05 Feb 2024
IllusionX: An LLM-powered mixed reality personal companion
IllusionX: An LLM-powered mixed reality personal companion
Ramez Yousri
Zeyad Essam
Yehia Kareem
Youstina Sherief
Sherry Gamil
Soha Safwat
29
3
0
04 Feb 2024
BRAIn: Bayesian Reward-conditioned Amortized Inference for natural
  language generation from feedback
BRAIn: Bayesian Reward-conditioned Amortized Inference for natural language generation from feedback
Gaurav Pandey
Yatin Nandwani
Tahira Naseem
Mayank Mishra
Guangxuan Xu
Dinesh Raghu
Sachindra Joshi
Asim Munawar
Ramón Fernández Astudillo
BDL
44
3
0
04 Feb 2024
Diversity Measurement and Subset Selection for Instruction Tuning
  Datasets
Diversity Measurement and Subset Selection for Instruction Tuning Datasets
Peiqi Wang
Songlin Yang
Zhen Guo
Matt Stallone
Yoon Kim
Polina Golland
Yikang Shen
31
9
0
04 Feb 2024
Preference Poisoning Attacks on Reward Model Learning
Preference Poisoning Attacks on Reward Model Learning
Junlin Wu
Jiong Wang
Chaowei Xiao
Chenguang Wang
Ning Zhang
Yevgeniy Vorobeychik
AAML
32
5
0
02 Feb 2024
The RL/LLM Taxonomy Tree: Reviewing Synergies Between Reinforcement
  Learning and Large Language Models
The RL/LLM Taxonomy Tree: Reviewing Synergies Between Reinforcement Learning and Large Language Models
M. Pternea
Prerna Singh
Abir Chakraborty
Y. Oruganti
M. Milletarí
Sayli Bapat
Kebei Jiang
OffRL
36
7
0
02 Feb 2024
Foundation Model Sherpas: Guiding Foundation Models through Knowledge
  and Reasoning
Foundation Model Sherpas: Guiding Foundation Models through Knowledge and Reasoning
D. Bhattacharjya
Junkyu Lee
Don Joven Agravante
Balaji Ganesan
Radu Marinescu
LLMAG
38
1
0
02 Feb 2024
Rethinking the Role of Proxy Rewards in Language Model Alignment
Rethinking the Role of Proxy Rewards in Language Model Alignment
Sungdong Kim
Minjoon Seo
SyDa
ALM
31
0
0
02 Feb 2024
KTO: Model Alignment as Prospect Theoretic Optimization
KTO: Model Alignment as Prospect Theoretic Optimization
Kawin Ethayarajh
Winnie Xu
Niklas Muennighoff
Dan Jurafsky
Douwe Kiela
182
463
0
02 Feb 2024
DTS-SQL: Decomposed Text-to-SQL with Small Large Language Models
DTS-SQL: Decomposed Text-to-SQL with Small Large Language Models
Mohammadreza Pourreza
Davood Rafiei
23
24
0
02 Feb 2024
Plan-Grounded Large Language Models for Dual Goal Conversational
  Settings
Plan-Grounded Large Language Models for Dual Goal Conversational Settings
Diogo Glória-Silva
Rafael Ferreira
Diogo Tavares
David Semedo
João Magalhães
LLMAG
50
4
0
01 Feb 2024
Towards Efficient Exact Optimization of Language Model Alignment
Towards Efficient Exact Optimization of Language Model Alignment
Haozhe Ji
Cheng Lu
Yilin Niu
Pei Ke
Hongning Wang
Jun Zhu
Jie Tang
Minlie Huang
63
12
0
01 Feb 2024
Dense Reward for Free in Reinforcement Learning from Human Feedback
Dense Reward for Free in Reinforcement Learning from Human Feedback
Alex J. Chan
Hao Sun
Samuel Holt
M. Schaar
28
32
0
01 Feb 2024
Transforming and Combining Rewards for Aligning Large Language Models
Transforming and Combining Rewards for Aligning Large Language Models
Zihao Wang
Chirag Nagpal
Jonathan Berant
Jacob Eisenstein
Alex DÁmour
Oluwasanmi Koyejo
Victor Veitch
27
11
0
01 Feb 2024
Efficient Exploration for LLMs
Efficient Exploration for LLMs
Vikranth Dwaracherla
S. Asghari
Botao Hao
Benjamin Van Roy
LLMAG
28
20
0
01 Feb 2024
Iterative Data Smoothing: Mitigating Reward Overfitting and
  Overoptimization in RLHF
Iterative Data Smoothing: Mitigating Reward Overfitting and Overoptimization in RLHF
Banghua Zhu
Michael I. Jordan
Jiantao Jiao
36
25
0
29 Jan 2024
Mapping the Design Space of Teachable Social Media Feed Experiences
Mapping the Design Space of Teachable Social Media Feed Experiences
K. J. Kevin Feng
Xander Koo
Lawrence Tan
Amy Bruckman
David W. McDonald
Amy X. Zhang
32
15
0
25 Jan 2024
Towards Consistent Natural-Language Explanations via
  Explanation-Consistency Finetuning
Towards Consistent Natural-Language Explanations via Explanation-Consistency Finetuning
Yanda Chen
Chandan Singh
Xiaodong Liu
Simiao Zuo
Bin-Xia Yu
He He
Jianfeng Gao
LRM
32
13
0
25 Jan 2024
Instruction Fine-Tuning: Does Prompt Loss Matter?
Instruction Fine-Tuning: Does Prompt Loss Matter?
Mathew Huerta-Enochian
Seung Yong Ko
27
6
0
24 Jan 2024
Can AI Assistants Know What They Don't Know?
Can AI Assistants Know What They Don't Know?
Qinyuan Cheng
Tianxiang Sun
Xiangyang Liu
Wenwei Zhang
Zhangyue Yin
Shimin Li
Linyang Li
Zhengfu He
Kai Chen
Xipeng Qiu
47
24
0
24 Jan 2024
ARGS: Alignment as Reward-Guided Search
ARGS: Alignment as Reward-Guided Search
Maxim Khanov
Jirayu Burapacheep
Yixuan Li
40
48
0
23 Jan 2024
Improving Machine Translation with Human Feedback: An Exploration of
  Quality Estimation as a Reward Model
Improving Machine Translation with Human Feedback: An Exploration of Quality Estimation as a Reward Model
Zhiwei He
Xing Wang
Wenxiang Jiao
ZhuoSheng Zhang
Rui Wang
Shuming Shi
Zhaopeng Tu
ALM
39
24
0
23 Jan 2024
WARM: On the Benefits of Weight Averaged Reward Models
WARM: On the Benefits of Weight Averaged Reward Models
Alexandre Ramé
Nino Vieillard
Léonard Hussenot
Robert Dadashi
Geoffrey Cideron
Olivier Bachem
Johan Ferret
123
95
0
22 Jan 2024
Mastering Text-to-Image Diffusion: Recaptioning, Planning, and
  Generating with Multimodal LLMs
Mastering Text-to-Image Diffusion: Recaptioning, Planning, and Generating with Multimodal LLMs
Ling Yang
Zhaochen Yu
Chenlin Meng
Minkai Xu
Stefano Ermon
Tengjiao Wang
CoGe
DiffM
52
116
0
22 Jan 2024
Linear Alignment: A Closed-form Solution for Aligning Human Preferences
  without Tuning and Feedback
Linear Alignment: A Closed-form Solution for Aligning Human Preferences without Tuning and Feedback
Songyang Gao
Qiming Ge
Wei Shen
Shihan Dou
Junjie Ye
...
Yicheng Zou
Zhi Chen
Hang Yan
Qi Zhang
Dahua Lin
57
11
0
21 Jan 2024
Reinforcement learning for question answering in programming domain
  using public community scoring as a human feedback
Reinforcement learning for question answering in programming domain using public community scoring as a human feedback
Alexey Gorbatovski
Sergey Kovalchuk
25
3
0
19 Jan 2024
Self-Rewarding Language Models
Self-Rewarding Language Models
Weizhe Yuan
Richard Yuanzhe Pang
Kyunghyun Cho
Xian Li
Sainbayar Sukhbaatar
Jing Xu
Jason Weston
ReLM
SyDa
ALM
LRM
244
304
0
18 Jan 2024
Tuning Language Models by Proxy
Tuning Language Models by Proxy
Alisa Liu
Xiaochuang Han
Yizhong Wang
Yulia Tsvetkov
Yejin Choi
Noah A. Smith
ALM
43
46
0
16 Jan 2024
EmoLLMs: A Series of Emotional Large Language Models and Annotation
  Tools for Comprehensive Affective Analysis
EmoLLMs: A Series of Emotional Large Language Models and Annotation Tools for Comprehensive Affective Analysis
Zhiwei Liu
Kailai Yang
Tianlin Zhang
Qianqian Xie
Sophia Ananiadou
49
40
0
16 Jan 2024
Beyond Sparse Rewards: Enhancing Reinforcement Learning with Language
  Model Critique in Text Generation
Beyond Sparse Rewards: Enhancing Reinforcement Learning with Language Model Critique in Text Generation
Meng Cao
Lei Shu
Lei Yu
Yun Zhu
Nevan Wichers
Yinxiao Liu
Lei Meng
OffRL
ALM
27
4
0
14 Jan 2024
Small Language Model Can Self-correct
Small Language Model Can Self-correct
Haixia Han
Jiaqing Liang
Jie Shi
Qi He
Yanghua Xiao
LRM
SyDa
ReLM
KELM
40
11
0
14 Jan 2024
Quantized Side Tuning: Fast and Memory-Efficient Tuning of Quantized
  Large Language Models
Quantized Side Tuning: Fast and Memory-Efficient Tuning of Quantized Large Language Models
Zhengxin Zhang
Dan Zhao
Xupeng Miao
Gabriele Oliaro
Qing Li
Yong-jia Jiang
Zhihao Jia
MQ
51
7
0
13 Jan 2024
Secrets of RLHF in Large Language Models Part II: Reward Modeling
Secrets of RLHF in Large Language Models Part II: Reward Modeling
Bing Wang
Rui Zheng
Luyao Chen
Yan Liu
Shihan Dou
...
Qi Zhang
Xipeng Qiu
Xuanjing Huang
Zuxuan Wu
Yuanyuan Jiang
ALM
47
96
0
11 Jan 2024
Agent Alignment in Evolving Social Norms
Agent Alignment in Evolving Social Norms
Shimin Li
Tianxiang Sun
Qinyuan Cheng
Xipeng Qiu
LLMAG
43
8
0
09 Jan 2024
TechGPT-2.0: A large language model project to solve the task of
  knowledge graph construction
TechGPT-2.0: A large language model project to solve the task of knowledge graph construction
Jiaqi Wang
Yuying Chang
Zhong Li
Ning An
Qi Ma
Lei Hei
Haibo Luo
Yifei Lu
Feiliang Ren
21
6
0
09 Jan 2024
A Minimaximalist Approach to Reinforcement Learning from Human Feedback
A Minimaximalist Approach to Reinforcement Learning from Human Feedback
Gokul Swamy
Christoph Dann
Rahul Kidambi
Zhiwei Steven Wu
Alekh Agarwal
OffRL
51
96
0
08 Jan 2024
An Exploratory Study on Automatic Identification of Assumptions in the
  Development of Deep Learning Frameworks
An Exploratory Study on Automatic Identification of Assumptions in the Development of Deep Learning Frameworks
Chen Yang
Peng Liang
Zinan Ma
32
0
0
08 Jan 2024
MLLM-Protector: Ensuring MLLM's Safety without Hurting Performance
MLLM-Protector: Ensuring MLLM's Safety without Hurting Performance
Renjie Pi
Tianyang Han
Jianshu Zhang
Yueqi Xie
Rui Pan
Qing Lian
Hanze Dong
Jipeng Zhang
Tong Zhang
AAML
34
61
0
05 Jan 2024
Hyperparameter-Free Approach for Faster Minimum Bayes Risk Decoding
Hyperparameter-Free Approach for Faster Minimum Bayes Risk Decoding
Yuu Jinnai
Kaito Ariu
34
8
0
05 Jan 2024
Blending Is All You Need: Cheaper, Better Alternative to
  Trillion-Parameters LLM
Blending Is All You Need: Cheaper, Better Alternative to Trillion-Parameters LLM
Xiaoding Lu
Zongyi Liu
Adian Liusie
Vyas Raina
Vineet Mudupalli
Yuwen Zhang
W. Beauchamp
35
15
0
04 Jan 2024
Previous
123...161718...272829
Next