Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2009.01325
Cited By
v1
v2
v3 (latest)
Learning to summarize from human feedback
2 September 2020
Nisan Stiennon
Long Ouyang
Jeff Wu
Daniel M. Ziegler
Ryan J. Lowe
Chelsea Voss
Alec Radford
Dario Amodei
Paul Christiano
ALM
Re-assign community
ArXiv (abs)
PDF
HTML
Papers citing
"Learning to summarize from human feedback"
50 / 1,548 papers shown
Title
Argumentative Reward Learning: Reasoning About Human Preferences
Francis Rhys Ward
Francesco Belardinelli
Francesca Toni
HAI
147
2
0
28 Sep 2022
Defining and Characterizing Reward Hacking
Joar Skalse
Nikolaus H. R. Howe
Dmitrii Krasheninnikov
David M. Krueger
133
61
0
27 Sep 2022
Adaptive Natural Language Generation for Task-oriented Dialogue via Reinforcement Learning
Atsumoto Ohashi
Ryuichiro Higashinaka
OffRL
81
6
0
16 Sep 2022
Law Informs Code: A Legal Informatics Approach to Aligning Artificial Intelligence with Humans
John J. Nay
ELM
AILaw
190
29
0
14 Sep 2022
The Alignment Problem from a Deep Learning Perspective
Richard Ngo
Lawrence Chan
Sören Mindermann
149
193
0
30 Aug 2022
Z-Code++: A Pre-trained Language Model Optimized for Abstractive Summarization
Pengcheng He
Baolin Peng
Liyang Lu
Song Wang
Jie Mei
...
Chenguang Zhu
Wayne Xiong
Michael Zeng
Jianfeng Gao
Xuedong Huang
119
47
0
21 Aug 2022
Abstractive Meeting Summarization: A Survey
Virgile Rennard
Guokan Shang
Julie Hunter
Michalis Vazirgiannis
101
16
0
08 Aug 2022
Efficient Training of Language Models to Fill in the Middle
Mohammad Bavarian
Heewoo Jun
Nikolas Tezak
John Schulman
C. McLeavey
Jerry Tworek
Mark Chen
94
197
0
28 Jul 2022
MAD for Robust Reinforcement Learning in Machine Translation
Domenic Donato
Lei Yu
Wang Ling
Chris Dyer
MoE
56
7
0
18 Jul 2022
SummScore: A Comprehensive Evaluation Metric for Summary Quality Based on Cross-Encoder
Wuhang Lin
Shasha Li
Chen Zhang
Bing Ji
Jie Yu
Jun Ma
Zibo Yi
49
6
0
11 Jul 2022
Conditional Generation with a Question-Answering Blueprint
Shashi Narayan
Joshua Maynez
Reinald Kim Amplayo
Kuzman Ganchev
Annie Louis
Fantine Huot
Anders Sandholm
Dipanjan Das
Mirella Lapata
124
49
0
01 Jul 2022
Mapping the Design Space of Human-AI Interaction in Text Summarization
Ruijia Cheng
Alison Smith-Renner
Kecheng Zhang
Joel R. Tetreault
A. Jaimes
100
32
0
29 Jun 2022
Know your audience: specializing grounded language models with listener subtraction
Aaditya K. Singh
David Ding
Andrew M. Saxe
Felix Hill
Andrew Kyle Lampinen
76
2
0
16 Jun 2022
'John ate 5 apples' != 'John ate some apples': Self-Supervised Paraphrase Quality Detection for Algebraic Word Problems
Rishabh Gupta
Venktesh V
Mukesh Mohania
Vikram Goyal
AIMat
16
2
0
16 Jun 2022
An Exploration of Post-Editing Effectiveness in Text Summarization
Vivian Lai
Alison Smith-Renner
Ke Zhang
Ruijia Cheng
Wenjuan Zhang
Joel R. Tetreault
Alejandro Jaimes
53
1
0
13 Jun 2022
Human-AI Interaction Design in Machine Teaching
Karan Taneja
Harsh Sikka
Ashok K. Goel
41
2
0
10 Jun 2022
Offline RL for Natural Language Generation with Implicit Language Q Learning
Charles Burton Snell
Ilya Kostrikov
Yi Su
Mengjiao Yang
Sergey Levine
OffRL
236
115
0
05 Jun 2022
On Reinforcement Learning and Distribution Matching for Fine-Tuning Language Models with no Catastrophic Forgetting
Tomasz Korbak
Hady ElSahar
Germán Kruszewski
Marc Dymetman
CLL
105
57
0
01 Jun 2022
Teaching Models to Express Their Uncertainty in Words
Stephanie C. Lin
Jacob Hilton
Owain Evans
OOD
144
425
0
28 May 2022
Quark: Controllable Text Generation with Reinforced Unlearning
Ximing Lu
Sean Welleck
Jack Hessel
Liwei Jiang
Lianhui Qin
Peter West
Prithviraj Ammanabrolu
Yejin Choi
MU
179
220
0
26 May 2022
Multimodal Knowledge Alignment with Reinforcement Learning
Youngjae Yu
Jiwan Chung
Heeseung Yun
Jack Hessel
Jinho Park
...
Prithviraj Ammanabrolu
Rowan Zellers
Ronan Le Bras
Gunhee Kim
Yejin Choi
VLM
163
37
0
25 May 2022
RL with KL penalties is better viewed as Bayesian inference
Tomasz Korbak
Ethan Perez
Christopher L. Buckley
OffRL
96
77
0
23 May 2022
Efficient Unsupervised Sentence Compression by Fine-tuning Transformers with Reinforcement Learning
D. Ghalandari
Chris Hokamp
Georgiana Ifrim
69
20
0
17 May 2022
Training Language Models with Language Feedback
Jérémy Scheurer
Jon Ander Campos
Jun Shern Chan
Angelica Chen
Kyunghyun Cho
Ethan Perez
ALM
126
51
0
29 Apr 2022
A Framework for Interactive Knowledge-Aided Machine Teaching
Karan Taneja
Harsh Sikka
Ashok K. Goel
HAI
46
4
0
21 Apr 2022
A Survey on Neural Abstractive Summarization Methods and Factual Consistency of Summarization
Meng Cao
67
6
0
20 Apr 2022
Text Revision by On-the-Fly Representation Optimization
Jingjing Li
Zichao Li
Tao Ge
Irwin King
Michael R. Lyu
BDL
93
18
0
15 Apr 2022
GPT-NeoX-20B: An Open-Source Autoregressive Language Model
Sid Black
Stella Biderman
Eric Hallahan
Quentin G. Anthony
Leo Gao
...
Shivanshu Purohit
Laria Reynolds
J. Tow
Benqi Wang
Samuel Weinbach
189
841
0
14 Apr 2022
Causal Confusion and Reward Misidentification in Preference-Based Reward Learning
J. Tien
Jerry Zhi-Yang He
Zackory M. Erickson
Anca Dragan
Daniel S. Brown
CML
105
43
0
13 Apr 2022
ASQA: Factoid Questions Meet Long-Form Answers
Ivan Stelmakh
Yi Luan
Bhuwan Dhingra
Ming-Wei Chang
90
178
0
12 Apr 2022
Make The Most of Prior Data: A Solution for Interactive Text Summarization with Preference Feedback
Duy-Hung Nguyen
Nguyen-Viet-Dung Nghiem
Bao-Sinh Nguyen
Dung Tien Le
Shahab Sabahi
Minh Le Nguyen
Hung Le
75
13
0
12 Apr 2022
Active Learning with Label Comparisons
G. Yona
Shay Moran
G. Elidan
Amir Globerson
74
6
0
10 Apr 2022
Using Interactive Feedback to Improve the Accuracy and Explainability of Question Answering Systems Post-Deployment
Zichao Li
Prakhar Sharma
Xing Han Lù
Jackie C.K. Cheung
Siva Reddy
HAI
71
26
0
06 Apr 2022
Teaching language models to support answers with verified quotes
Jacob Menick
Maja Trebacz
Vladimir Mikulik
John Aslanides
Francis Song
...
Mia Glaese
Susannah Young
Lucy Campbell-Gillingham
G. Irving
Nat McAleese
ELM
RALM
321
267
0
21 Mar 2022
Simulating Bandit Learning from User Feedback for Extractive Question Answering
Ge Gao
Eunsol Choi
Yoav Artzi
85
14
0
18 Mar 2022
SURF: Semi-supervised Reward Learning with Data Augmentation for Feedback-efficient Preference-based Reinforcement Learning
Jongjin Park
Younggyo Seo
Jinwoo Shin
Honglak Lee
Pieter Abbeel
Kimin Lee
60
82
0
18 Mar 2022
Invariance in Policy Optimisation and Partial Identifiability in Reward Learning
Joar Skalse
Matthew Farrugia-Roberts
Stuart J. Russell
Alessandro Abate
Adam Gleave
84
48
0
14 Mar 2022
Uncertainty Estimation for Language Reward Models
Adam Gleave
G. Irving
UQLM
86
34
0
14 Mar 2022
Active Evaluation: Efficient NLG Evaluation with Few Pairwise Comparisons
Akash Kumar Mohankumar
Mitesh M. Khapra
ELM
AAML
70
7
0
11 Mar 2022
Training language models to follow instructions with human feedback
Long Ouyang
Jeff Wu
Xu Jiang
Diogo Almeida
Carroll L. Wainwright
...
Amanda Askell
Peter Welinder
Paul Christiano
Jan Leike
Ryan J. Lowe
OSLM
ALM
1.3K
13,290
0
04 Mar 2022
Capturing Failures of Large Language Models via Human Cognitive Biases
Erik Jones
Jacob Steinhardt
83
93
0
24 Feb 2022
CAISE: Conversational Agent for Image Search and Editing
Hyounghun Kim
Doo Soon Kim
Seunghyun Yoon
Franck Dernoncourt
Trung Bui
Joey Tianyi Zhou
49
6
0
24 Feb 2022
Reward Modeling for Mitigating Toxicity in Transformer-based Language Models
Farshid Faal
K. Schmitt
Jia Yuan Yu
83
25
0
19 Feb 2022
A data-driven approach for learning to control computers
Peter C. Humphreys
David Raposo
Tobias Pohlen
Gregory Thornton
Rachita Chhaparia
...
Josh Abramson
Petko Georgiev
Alex Goldin
Adam Santoro
Timothy Lillicrap
113
103
0
16 Feb 2022
Repairing the Cracked Foundation: A Survey of Obstacles in Evaluation Practices for Generated Text
Sebastian Gehrmann
Elizabeth Clark
Thibault Sellam
ELM
AI4CE
157
193
0
14 Feb 2022
Red Teaming Language Models with Language Models
Ethan Perez
Saffron Huang
Francis Song
Trevor Cai
Roman Ring
John Aslanides
Amelia Glaese
Nat McAleese
G. Irving
AAML
243
672
0
07 Feb 2022
Safe Deep RL in 3D Environments using Human Feedback
Matthew Rahtz
Vikrant Varma
Ramana Kumar
Zachary Kenton
Shane Legg
Jan Leike
94
4
0
20 Jan 2022
A Survey of Controllable Text Generation using Transformer-based Pre-trained Language Models
Hanqing Zhang
Haolin Song
Shaoyu Li
Ming Zhou
Dawei Song
143
230
0
14 Jan 2022
The Effects of Reward Misspecification: Mapping and Mitigating Misaligned Models
Alexander Pan
Kush S. Bhatia
Jacob Steinhardt
131
184
0
10 Jan 2022
Beyond modeling: NLP Pipeline for efficient environmental policy analysis
J. Planas
Daniel Firebanks-Quevedo
G. Naydenova
Ramansh Sharma
Cristina Taylor
Kathleen Buckingham
Rong Fang
22
4
0
08 Jan 2022
Previous
1
2
3
...
29
30
31
Next