ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2009.01325
  4. Cited By
Learning to summarize from human feedback

Learning to summarize from human feedback

2 September 2020
Nisan Stiennon
Long Ouyang
Jeff Wu
Daniel M. Ziegler
Ryan J. Lowe
Chelsea Voss
Alec Radford
Dario Amodei
Paul Christiano
    ALM
ArXivPDFHTML

Papers citing "Learning to summarize from human feedback"

50 / 1,443 papers shown
Title
Llama 2: Open Foundation and Fine-Tuned Chat Models
Llama 2: Open Foundation and Fine-Tuned Chat Models
Hugo Touvron
Louis Martin
Kevin R. Stone
Peter Albert
Amjad Almahairi
...
Sharan Narang
Aurelien Rodriguez
Robert Stojnic
Sergey Edunov
Thomas Scialom
AI4MH
ALM
135
11,144
0
18 Jul 2023
Latent Jailbreak: A Benchmark for Evaluating Text Safety and Output
  Robustness of Large Language Models
Latent Jailbreak: A Benchmark for Evaluating Text Safety and Output Robustness of Large Language Models
Huachuan Qiu
Shuai Zhang
Anqi Li
Hongliang He
Zhenzhong Lan
ALM
44
48
0
17 Jul 2023
On the application of Large Language Models for language teaching and
  assessment technology
On the application of Large Language Models for language teaching and assessment technology
Andrew Caines
Luca Benedetto
Shiva Taslimipoor
Christopher Davis
Yuan Gao
...
Marek Rei
H. Yannakoudakis
Andrew Mullooly
D. Nicholls
P. Buttery
ELM
29
43
0
17 Jul 2023
Measuring Faithfulness in Chain-of-Thought Reasoning
Measuring Faithfulness in Chain-of-Thought Reasoning
Tamera Lanham
Anna Chen
Ansh Radhakrishnan
Benoit Steiner
Carson E. Denison
...
Zac Hatfield-Dodds
Jared Kaplan
J. Brauner
Sam Bowman
Ethan Perez
ReLM
LRM
30
169
0
17 Jul 2023
Dialogue Agents 101: A Beginner's Guide to Critical Ingredients for
  Designing Effective Conversational Systems
Dialogue Agents 101: A Beginner's Guide to Critical Ingredients for Designing Effective Conversational Systems
Shivani Kumar
S. Bhatia
Milan Aggarwal
Tanmoy Chakraborty
29
1
0
14 Jul 2023
Secrets of RLHF in Large Language Models Part I: PPO
Secrets of RLHF in Large Language Models Part I: PPO
Rui Zheng
Shihan Dou
Songyang Gao
Yuan Hua
Wei Shen
...
Hang Yan
Tao Gui
Qi Zhang
Xipeng Qiu
Xuanjing Huang
ALM
OffRL
55
160
0
11 Jul 2023
BeaverTails: Towards Improved Safety Alignment of LLM via a
  Human-Preference Dataset
BeaverTails: Towards Improved Safety Alignment of LLM via a Human-Preference Dataset
Jiaming Ji
Mickel Liu
Juntao Dai
Xuehai Pan
Chi Zhang
Ce Bian
Chi Zhang
Ruiyang Sun
Yizhou Wang
Yaodong Yang
ALM
30
413
0
10 Jul 2023
Improving Factuality of Abstractive Summarization via Contrastive Reward
  Learning
Improving Factuality of Abstractive Summarization via Contrastive Reward Learning
Ethan Chern
Zhiruo Wang
Sanjan Das
Bhavuk Sharma
Pengfei Liu
Graham Neubig
HILM
20
14
0
10 Jul 2023
TIM: Teaching Large Language Models to Translate with Comparison
TIM: Teaching Large Language Models to Translate with Comparison
Jiali Zeng
Fandong Meng
Yongjing Yin
Jie Zhou
37
55
0
10 Jul 2023
Advancements in Scientific Controllable Text Generation Methods
Advancements in Scientific Controllable Text Generation Methods
Arnav Goel
Medha Hira
Avinash Anand
Siddhesh Bangar
R. Shah
27
7
0
08 Jul 2023
Provably Efficient Iterated CVaR Reinforcement Learning with Function
  Approximation and Human Feedback
Provably Efficient Iterated CVaR Reinforcement Learning with Function Approximation and Human Feedback
Yu Chen
Yihan Du
Pihe Hu
Si-Yi Wang
De-hui Wu
Longbo Huang
24
6
0
06 Jul 2023
Censored Sampling of Diffusion Models Using 3 Minutes of Human Feedback
Censored Sampling of Diffusion Models Using 3 Minutes of Human Feedback
Taeho Yoon
Kibeom Myoung
Keon Lee
Jaewoong Cho
Albert No
Ernest K. Ryu
23
8
0
06 Jul 2023
Jailbroken: How Does LLM Safety Training Fail?
Jailbroken: How Does LLM Safety Training Fail?
Alexander Wei
Nika Haghtalab
Jacob Steinhardt
127
856
0
05 Jul 2023
Natural Language Generation and Understanding of Big Code for
  AI-Assisted Programming: A Review
Natural Language Generation and Understanding of Big Code for AI-Assisted Programming: A Review
M. Wong
Shangxin Guo
Ching Nam Hang
Siu-Wai Ho
C. Tan
47
78
0
04 Jul 2023
The Inner Sentiments of a Thought
The Inner Sentiments of a Thought
Christian Gagné
Peter Dayan
33
4
0
04 Jul 2023
Let Me Teach You: Pedagogical Foundations of Feedback for Language
  Models
Let Me Teach You: Pedagogical Foundations of Feedback for Language Models
Beatriz Borges
Niket Tandon
Tanja Käser
Antoine Bosselut
31
4
0
01 Jul 2023
Preference Ranking Optimization for Human Alignment
Preference Ranking Optimization for Human Alignment
Feifan Song
Yu Bowen
Minghao Li
Haiyang Yu
Fei Huang
Yongbin Li
Houfeng Wang
ALM
31
240
0
30 Jun 2023
On the Exploitability of Instruction Tuning
On the Exploitability of Instruction Tuning
Manli Shu
Jiong Wang
Chen Zhu
Jonas Geiping
Chaowei Xiao
Tom Goldstein
SILM
47
92
0
28 Jun 2023
Towards Measuring the Representation of Subjective Global Opinions in
  Language Models
Towards Measuring the Representation of Subjective Global Opinions in Language Models
Esin Durmus
Karina Nyugen
Thomas I. Liao
Nicholas Schiefer
Amanda Askell
...
Alex Tamkin
Janel Thamkul
Jared Kaplan
Jack Clark
Deep Ganguli
46
213
0
28 Jun 2023
System-Level Natural Language Feedback
System-Level Natural Language Feedback
Weizhe Yuan
Kyunghyun Cho
Jason Weston
41
5
0
23 Jun 2023
A Survey on Multimodal Large Language Models
A Survey on Multimodal Large Language Models
Shukang Yin
Chaoyou Fu
Sirui Zhao
Ke Li
Xing Sun
Tong Xu
Enhong Chen
MLLM
LRM
62
562
0
23 Jun 2023
Opportunities and Risks of LLMs for Scalable Deliberation with Polis
Opportunities and Risks of LLMs for Scalable Deliberation with Polis
Christopher T. Small
Ivan Vendrov
Esin Durmus
Hadjar Homaei
Elizabeth Barry
Julien Cornebise
Ted Suzman
Deep Ganguli
Colin Megill
35
27
0
20 Jun 2023
Learning to Generate Better Than Your LLM
Learning to Generate Better Than Your LLM
Jonathan D. Chang
Kianté Brantley
Rajkumar Ramamurthy
Dipendra Kumar Misra
Wen Sun
27
42
0
20 Jun 2023
Learning Profitable NFT Image Diffusions via Multiple Visual-Policy
  Guided Reinforcement Learning
Learning Profitable NFT Image Diffusions via Multiple Visual-Policy Guided Reinforcement Learning
Huiguo He
Tianfu Wang
Huan Yang
Jianlong Fu
N. Yuan
Jian Yin
Hongyang Chao
Qi Zhang
EGVM
37
9
0
20 Jun 2023
Snowman: A Million-scale Chinese Commonsense Knowledge Graph Distilled
  from Foundation Model
Snowman: A Million-scale Chinese Commonsense Knowledge Graph Distilled from Foundation Model
Jiaan Wang
Jianfeng Qu
Yunlong Liang
Zhixu Li
An Liu
Guanfeng Liu
Xin Zheng
30
2
0
17 Jun 2023
Aligning Synthetic Medical Images with Clinical Knowledge using Human
  Feedback
Aligning Synthetic Medical Images with Clinical Knowledge using Human Feedback
Shenghuan Sun
Gregory M. Goldgof
A. Butte
Ahmed Alaa
MedIm
27
12
0
16 Jun 2023
Fairness in Preference-based Reinforcement Learning
Fairness in Preference-based Reinforcement Learning
Umer Siddique
Abhinav Sinha
Yongcan Cao
19
4
0
16 Jun 2023
ClinicalGPT: Large Language Models Finetuned with Diverse Medical Data
  and Comprehensive Evaluation
ClinicalGPT: Large Language Models Finetuned with Diverse Medical Data and Comprehensive Evaluation
Guangyu Wang
Guoxing Yang
Zongxin Du
Longjun Fan
Xiaohu Li
LM&MA
ELM
AI4MH
25
81
0
16 Jun 2023
Inverse Scaling: When Bigger Isn't Better
Inverse Scaling: When Bigger Isn't Better
I. R. McKenzie
Alexander Lyzhov
Michael Pieler
Alicia Parrish
Aaron Mueller
...
Yuhui Zhang
Zhengping Zhou
Najoung Kim
Sam Bowman
Ethan Perez
41
128
0
15 Jun 2023
WebGLM: Towards An Efficient Web-Enhanced Question Answering System with
  Human Preferences
WebGLM: Towards An Efficient Web-Enhanced Question Answering System with Human Preferences
Xiao Liu
Hanyu Lai
Hao Yu
Yifan Xu
Aohan Zeng
Zhengxiao Du
Peng Zhang
Yuxiao Dong
Jie Tang
17
97
0
13 Jun 2023
A Markovian Formalism for Active Querying
A Markovian Formalism for Active Querying
Sid Ijju
26
1
0
13 Jun 2023
When Do Annotator Demographics Matter? Measuring the Influence of
  Annotator Demographics with the POPQUORN Dataset
When Do Annotator Demographics Matter? Measuring the Influence of Annotator Demographics with the POPQUORN Dataset
Jiaxin Pei
David Jurgens
40
32
0
12 Jun 2023
Multi-Source Test-Time Adaptation as Dueling Bandits for Extractive
  Question Answering
Multi-Source Test-Time Adaptation as Dueling Bandits for Extractive Question Answering
Hai Ye
Qizhe Xie
Hwee Tou Ng
53
8
0
11 Jun 2023
Reliability Check: An Analysis of GPT-3's Response to Sensitive Topics
  and Prompt Wording
Reliability Check: An Analysis of GPT-3's Response to Sensitive Topics and Prompt Wording
Aisha Khatun
Daniel Brown
KELM
18
12
0
09 Jun 2023
Towards a Robust Detection of Language Model Generated Text: Is ChatGPT
  that Easy to Detect?
Towards a Robust Detection of Language Model Generated Text: Is ChatGPT that Easy to Detect?
Wissam Antoun
Virginie Mouilleron
Benoît Sagot
Djamé Seddah
DeLMO
30
33
0
09 Jun 2023
Prefer to Classify: Improving Text Classifiers via Auxiliary Preference
  Learning
Prefer to Classify: Improving Text Classifiers via Auxiliary Preference Learning
Jaehyung Kim
Jinwoo Shin
Dongyeop Kang
19
2
0
08 Jun 2023
Absformer: Transformer-based Model for Unsupervised Multi-Document
  Abstractive Summarization
Absformer: Transformer-based Model for Unsupervised Multi-Document Abstractive Summarization
M. Trabelsi
H. Uzunalioglu
19
1
0
07 Jun 2023
Cross-attention learning enables real-time nonuniform rotational
  distortion correction in OCT
Cross-attention learning enables real-time nonuniform rotational distortion correction in OCT
Haoran Zhang
Jianlong Yang
Jingqian Zhang
Shiqing Zhao
Aili Zhang
184
8
0
07 Jun 2023
Rewarded soups: towards Pareto-optimal alignment by interpolating
  weights fine-tuned on diverse rewards
Rewarded soups: towards Pareto-optimal alignment by interpolating weights fine-tuned on diverse rewards
Alexandre Ramé
Guillaume Couairon
Mustafa Shukor
Corentin Dancette
Jean-Baptiste Gaya
Laure Soulier
Matthieu Cord
MoMe
35
136
0
07 Jun 2023
GPT Self-Supervision for a Better Data Annotator
GPT Self-Supervision for a Better Data Annotator
Xiaohuan Pei
Yanxi Li
Chang Xu
30
7
0
07 Jun 2023
PokemonChat: Auditing ChatGPT for Pokémon Universe Knowledge
PokemonChat: Auditing ChatGPT for Pokémon Universe Knowledge
Laura Cabello
Jiaang Li
Ilias Chalkidis
ELM
AI4MH
LRM
24
2
0
05 Jun 2023
Fine-Tuning Language Models with Advantage-Induced Policy Alignment
Fine-Tuning Language Models with Advantage-Induced Policy Alignment
Banghua Zhu
Hiteshi Sharma
Felipe Vieira Frujeri
Shi Dong
Chenguang Zhu
Michael I. Jordan
Jiantao Jiao
OSLM
36
39
0
04 Jun 2023
Fine-Grained Human Feedback Gives Better Rewards for Language Model
  Training
Fine-Grained Human Feedback Gives Better Rewards for Language Model Training
Zeqiu Wu
Yushi Hu
Weijia Shi
Nouha Dziri
Alane Suhr
Prithviraj Ammanabrolu
Noah A. Smith
Mari Ostendorf
Hannaneh Hajishirzi
ALM
53
305
0
02 Jun 2023
ChatGPT for Zero-shot Dialogue State Tracking: A Solution or an
  Opportunity?
ChatGPT for Zero-shot Dialogue State Tracking: A Solution or an Opportunity?
Michael Heck
Nurul Lubis
Benjamin Ruppik
Renato Vukovic
Shutong Feng
Christian Geishauser
Hsien-chin Lin
Carel van Niekerk
Milica Gavsić
24
43
0
02 Jun 2023
Hybrid Long Document Summarization using C2F-FAR and ChatGPT: A
  Practical Study
Hybrid Long Document Summarization using C2F-FAR and ChatGPT: A Practical Study
Guang Lu
Sylvia B. Larcher
Tu-Anh Tran
31
9
0
01 Jun 2023
The ethical ambiguity of AI data enrichment: Measuring gaps in research
  ethics norms and practices
The ethical ambiguity of AI data enrichment: Measuring gaps in research ethics norms and practices
Will Hawkins
Brent Mittelstadt
57
10
0
01 Jun 2023
Identifiability and Generalizability in Constrained Inverse
  Reinforcement Learning
Identifiability and Generalizability in Constrained Inverse Reinforcement Learning
Andreas Schlaginhaufen
Maryam Kamgarpour
29
10
0
01 Jun 2023
Preference-grounded Token-level Guidance for Language Model Fine-tuning
Preference-grounded Token-level Guidance for Language Model Fine-tuning
Shentao Yang
Shujian Zhang
Congying Xia
Yihao Feng
Caiming Xiong
Mi Zhou
29
23
0
01 Jun 2023
Factually Consistent Summarization via Reinforcement Learning with
  Textual Entailment Feedback
Factually Consistent Summarization via Reinforcement Learning with Textual Entailment Feedback
Paul Roit
Johan Ferret
Lior Shani
Roee Aharoni
Geoffrey Cideron
...
Olivier Bachem
G. Elidan
Avinatan Hassidim
Olivier Pietquin
Idan Szpektor
HILM
28
79
0
31 May 2023
Let's Verify Step by Step
Let's Verify Step by Step
Hunter Lightman
V. Kosaraju
Yura Burda
Harrison Edwards
Bowen Baker
Teddy Lee
Jan Leike
John Schulman
Ilya Sutskever
K. Cobbe
ALM
OffRL
LRM
50
945
0
31 May 2023
Previous
123...222324...272829
Next