ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2009.01325
  4. Cited By
Learning to summarize from human feedback
v1v2v3 (latest)

Learning to summarize from human feedback

2 September 2020
Nisan Stiennon
Long Ouyang
Jeff Wu
Daniel M. Ziegler
Ryan J. Lowe
Chelsea Voss
Alec Radford
Dario Amodei
Paul Christiano
    ALM
ArXiv (abs)PDFHTML

Papers citing "Learning to summarize from human feedback"

50 / 1,548 papers shown
Title
MoCa: Measuring Human-Language Model Alignment on Causal and Moral
  Judgment Tasks
MoCa: Measuring Human-Language Model Alignment on Causal and Moral Judgment Tasks
Allen Nie
Yuhui Zhang
Atharva Amdekar
Chris Piech
Tatsunori Hashimoto
Tobias Gerstenberg
84
40
0
30 Oct 2023
Leveraging generative artificial intelligence to simulate student
  learning behavior
Leveraging generative artificial intelligence to simulate student learning behavior
Songlin Xu
Xinyu Zhang
SyDaAI4Ed
52
13
0
30 Oct 2023
Reward Finetuning for Faster and More Accurate Unsupervised Object
  Discovery
Reward Finetuning for Faster and More Accurate Unsupervised Object Discovery
Katie Z Luo
Zhenzhen Liu
Xiangyu Chen
Yurong You
Sagie Benaim
Cheng Perng Phoo
Mark E. Campbell
Wen Sun
B. Hariharan
Kilian Q. Weinberger
OffRL
94
11
0
29 Oct 2023
Ever Evolving Evaluator (EV3): Towards Flexible and Reliable
  Meta-Optimization for Knowledge Distillation
Ever Evolving Evaluator (EV3): Towards Flexible and Reliable Meta-Optimization for Knowledge Distillation
Li Ding
M. Zoghi
Guy Tennenholtz
Maryam Karimzadehgan
90
0
0
29 Oct 2023
Fine-Tuning Language Models Using Formal Methods Feedback
Fine-Tuning Language Models Using Formal Methods Feedback
Yunhao Yang
N. Bhatt
Tyler Ingebrand
William Ward
Steven Carr
Zhangyang Wang
Ufuk Topcu
69
9
0
27 Oct 2023
Using State-of-the-Art Speech Models to Evaluate Oral Reading Fluency in
  Ghana
Using State-of-the-Art Speech Models to Evaluate Oral Reading Fluency in Ghana
Owen Henkel
Hannah Horne-Robinson
Libby Hills
Bill Roberts
Joshua A. McGrane
48
0
0
26 Oct 2023
Can LLMs Grade Short-Answer Reading Comprehension Questions : An
  Empirical Study with a Novel Dataset
Can LLMs Grade Short-Answer Reading Comprehension Questions : An Empirical Study with a Novel Dataset
Owen Henkel
Libby Hills
Bill Roberts
Joshua A. McGrane
AI4Ed
76
1
0
26 Oct 2023
Beyond MLE: Convex Learning for Text Generation
Beyond MLE: Convex Learning for Text Generation
Chenze Shao
Zhengrui Ma
Min Zhang
Yang Feng
61
3
0
26 Oct 2023
Counterfactual-Augmented Importance Sampling for Semi-Offline Policy
  Evaluation
Counterfactual-Augmented Importance Sampling for Semi-Offline Policy Evaluation
Shengpu Tang
Jenna Wiens
OffRLCML
94
4
0
26 Oct 2023
Controlled Decoding from Language Models
Controlled Decoding from Language Models
Sidharth Mudgal
Jong Lee
H. Ganapathy
Yaguang Li
Tao Wang
...
Michael Collins
Trevor Strohman
Jilin Chen
Alex Beutel
Ahmad Beirami
153
91
0
25 Oct 2023
The Data Provenance Initiative: A Large Scale Audit of Dataset Licensing
  & Attribution in AI
The Data Provenance Initiative: A Large Scale Audit of Dataset Licensing & Attribution in AI
Shayne Longpre
Robert Mahari
Anthony Chen
Naana Obeng-Marnu
Damien Sileo
...
K. Bollacker
Tongshuang Wu
Luis Villa
Sandy Pentland
Sara Hooker
98
65
0
25 Oct 2023
BabyStories: Can Reinforcement Learning Teach Baby Language Models to
  Write Better Stories?
BabyStories: Can Reinforcement Learning Teach Baby Language Models to Write Better Stories?
Xingmeng Zhao
Tongnian Wang
Sheri Osborn
Anthony Rios
60
6
0
25 Oct 2023
BLESS: Benchmarking Large Language Models on Sentence Simplification
BLESS: Benchmarking Large Language Models on Sentence Simplification
Tannon Kew
Alison Chi
Laura Vásquez-Rodríguez
Sweta Agrawal
Dennis Aumiller
Fernando Alva-Manchego
Teven Le Scao
97
26
0
24 Oct 2023
COPR: Continual Learning Human Preference through Optimal Policy
  Regularization
COPR: Continual Learning Human Preference through Optimal Policy Regularization
Han Zhang
Lin Gui
Yuanzhao Zhai
Hui Wang
Yu Lei
Ruifeng Xu
CLL
87
0
0
24 Oct 2023
Retrieval-based Knowledge Transfer: An Effective Approach for Extreme
  Large Language Model Compression
Retrieval-based Knowledge Transfer: An Effective Approach for Extreme Large Language Model Compression
Jiduan Liu
Jiahao Liu
Qifan Wang
Jingang Wang
Xunliang Cai
Dongyan Zhao
Ran Wang
Rui Yan
73
4
0
24 Oct 2023
Specialist or Generalist? Instruction Tuning for Specific NLP Tasks
Specialist or Generalist? Instruction Tuning for Specific NLP Tasks
Chufan Shi
Yixuan Su
Cheng Yang
Yujiu Yang
Deng Cai
120
18
0
23 Oct 2023
Large Language Models are Visual Reasoning Coordinators
Large Language Models are Visual Reasoning Coordinators
Liangyu Chen
Bo Li
Sheng Shen
Jingkang Yang
Chunyuan Li
Kurt Keutzer
Trevor Darrell
Ziwei Liu
VLMLRM
130
58
0
23 Oct 2023
Unleashing the potential of prompt engineering in Large Language Models:
  a comprehensive review
Unleashing the potential of prompt engineering in Large Language Models: a comprehensive review
Banghao Chen
Zhaofeng Zhang
Nicolas Langrené
Shengxin Zhu
LLMAG
122
89
0
23 Oct 2023
PRCA: Fitting Black-Box Large Language Models for Retrieval Question
  Answering via Pluggable Reward-Driven Contextual Adapter
PRCA: Fitting Black-Box Large Language Models for Retrieval Question Answering via Pluggable Reward-Driven Contextual Adapter
Haoyan Yang
Zhitao Li
Yong Zhang
Jianzong Wang
Ning Cheng
Ming Li
Jing Xiao
RALM
47
31
0
23 Oct 2023
Active teacher selection for reinforcement learning from human feedback
Active teacher selection for reinforcement learning from human feedback
Rachel Freedman
Justin Svegliato
K. H. Wray
Stuart J. Russell
193
6
0
23 Oct 2023
Which Prompts Make The Difference? Data Prioritization For Efficient
  Human LLM Evaluation
Which Prompts Make The Difference? Data Prioritization For Efficient Human LLM Evaluation
M. Boubdir
Edward Kim
Beyza Ermis
Marzieh Fadaee
Sara Hooker
ALM
90
19
0
22 Oct 2023
Language Model Unalignment: Parametric Red-Teaming to Expose Hidden
  Harms and Biases
Language Model Unalignment: Parametric Red-Teaming to Expose Hidden Harms and Biases
Rishabh Bhardwaj
Soujanya Poria
ALM
118
18
0
22 Oct 2023
Revisiting Instruction Fine-tuned Model Evaluation to Guide Industrial
  Applications
Revisiting Instruction Fine-tuned Model Evaluation to Guide Industrial Applications
Manuel Faysse
Gautier Viaud
C´eline Hudelot
Pierre Colombo
84
11
0
21 Oct 2023
Contrastive Preference Learning: Learning from Human Feedback without RL
Contrastive Preference Learning: Learning from Human Feedback without RL
Joey Hejna
Rafael Rafailov
Harshit S. Sikchi
Chelsea Finn
S. Niekum
W. B. Knox
Dorsa Sadigh
OffRL
127
55
0
20 Oct 2023
Teaching Language Models to Self-Improve through Interactive
  Demonstrations
Teaching Language Models to Self-Improve through Interactive Demonstrations
Xiao Yu
Baolin Peng
Michel Galley
Jianfeng Gao
Zhou Yu
LRMReLM
104
22
0
20 Oct 2023
POSQA: Probe the World Models of LLMs with Size Comparisons
POSQA: Probe the World Models of LLMs with Size Comparisons
Chang Shu
Paul Burgess
Fangyu Liu
Ehsan Shareghi
Nigel Collier
54
2
0
20 Oct 2023
Safe RLHF: Safe Reinforcement Learning from Human Feedback
Safe RLHF: Safe Reinforcement Learning from Human Feedback
Josef Dai
Xuehai Pan
Ruiyang Sun
Jiaming Ji
Xinbo Xu
Mickel Liu
Yizhou Wang
Yaodong Yang
147
364
0
19 Oct 2023
Privacy Preserving Large Language Models: ChatGPT Case Study Based
  Vision and Framework
Privacy Preserving Large Language Models: ChatGPT Case Study Based Vision and Framework
Imdad Ullah
Najm Hassan
S. Gill
Basem Suleiman
T. Ahanger
Zawar Shah
Junaid Qadir
S. Kanhere
97
17
0
19 Oct 2023
Attack Prompt Generation for Red Teaming and Defending Large Language
  Models
Attack Prompt Generation for Red Teaming and Defending Large Language Models
Boyi Deng
Wenjie Wang
Fuli Feng
Yang Deng
Qifan Wang
Xiangnan He
AAML
78
57
0
19 Oct 2023
Quality Diversity through Human Feedback: Towards Open-Ended
  Diversity-Driven Optimization
Quality Diversity through Human Feedback: Towards Open-Ended Diversity-Driven Optimization
Lijie Ding
Jenny Zhang
Jeff Clune
Lee Spector
Joel Lehman
EGVM
117
9
0
18 Oct 2023
A General Theoretical Paradigm to Understand Learning from Human
  Preferences
A General Theoretical Paradigm to Understand Learning from Human Preferences
M. G. Azar
Mark Rowland
Bilal Piot
Daniel Guo
Daniele Calandriello
Michal Valko
Rémi Munos
292
648
0
18 Oct 2023
Improving Generalization of Alignment with Human Preferences through
  Group Invariant Learning
Improving Generalization of Alignment with Human Preferences through Group Invariant Learning
Rui Zheng
Wei Shen
Yuan Hua
Wenbin Lai
Shihan Dou
...
Xiao Wang
Haoran Huang
Tao Gui
Qi Zhang
Xuanjing Huang
113
17
0
18 Oct 2023
Personalized Soups: Personalized Large Language Model Alignment via
  Post-hoc Parameter Merging
Personalized Soups: Personalized Large Language Model Alignment via Post-hoc Parameter Merging
Joel Jang
Seungone Kim
Bill Yuchen Lin
Yizhong Wang
Jack Hessel
Luke Zettlemoyer
Hannaneh Hajishirzi
Yejin Choi
Prithviraj Ammanabrolu
MoMe
131
153
0
17 Oct 2023
An Empirical Study of Translation Hypothesis Ensembling with Large
  Language Models
An Empirical Study of Translation Hypothesis Ensembling with Large Language Models
António Farinhas
José G. C. de Souza
André F. T. Martins
84
11
0
17 Oct 2023
H2O Open Ecosystem for State-of-the-art Large Language Models
H2O Open Ecosystem for State-of-the-art Large Language Models
Arno Candel
Jon McKinney
Philipp Singer
Pascal Pfeiffer
Maximilian Jeblick
Chun Ming Lee
Marcos V. Conde
VLM
55
4
0
17 Oct 2023
Survey of Vulnerabilities in Large Language Models Revealed by
  Adversarial Attacks
Survey of Vulnerabilities in Large Language Models Revealed by Adversarial Attacks
Erfan Shayegani
Md Abdullah Al Mamun
Yu Fu
Pedram Zaree
Yue Dong
Nael B. Abu-Ghazaleh
AAML
247
164
0
16 Oct 2023
Sample Complexity of Preference-Based Nonparametric Off-Policy
  Evaluation with Deep Networks
Sample Complexity of Preference-Based Nonparametric Off-Policy Evaluation with Deep Networks
Zihao Li
Xiang Ji
Minshuo Chen
Mengdi Wang
OffRL
86
0
0
16 Oct 2023
ReMax: A Simple, Effective, and Efficient Reinforcement Learning Method
  for Aligning Large Language Models
ReMax: A Simple, Effective, and Efficient Reinforcement Learning Method for Aligning Large Language Models
Ziniu Li
Tian Xu
Yushun Zhang
Zhihang Lin
Yang Yu
Ruoyu Sun
Zhimin Luo
142
79
0
16 Oct 2023
Harnessing the Power of LLMs: Evaluating Human-AI Text Co-Creation
  through the Lens of News Headline Generation
Harnessing the Power of LLMs: Evaluating Human-AI Text Co-Creation through the Lens of News Headline Generation
Zijian Ding
Alison Smith-Renner
Wenjuan Zhang
Joel R. Tetreault
Alejandro Jaimes
100
26
0
16 Oct 2023
Privacy in Large Language Models: Attacks, Defenses and Future
  Directions
Privacy in Large Language Models: Attacks, Defenses and Future Directions
Haoran Li
Yulin Chen
Jinglong Luo
Yan Kang
Xiaojin Zhang
Qi Hu
Chunkit Chan
Yangqiu Song
PILM
118
45
0
16 Oct 2023
Legal NLP Meets MiCAR: Advancing the Analysis of Crypto White Papers
Legal NLP Meets MiCAR: Advancing the Analysis of Crypto White Papers
C. Camassa
AILaw
71
1
0
16 Oct 2023
Multi-Stage Pre-training Enhanced by ChatGPT for Multi-Scenario
  Multi-Domain Dialogue Summarization
Multi-Stage Pre-training Enhanced by ChatGPT for Multi-Scenario Multi-Domain Dialogue Summarization
Weixiao Zhou
Gengyao Li
Xianfu Cheng
Xinnian Liang
Junnan Zhu
Feifei Zhai
Zhoujun Li
51
6
0
16 Oct 2023
Verbosity Bias in Preference Labeling by Large Language Models
Verbosity Bias in Preference Labeling by Large Language Models
Keita Saito
Akifumi Wachi
Koki Wataoka
Youhei Akimoto
ALM
100
39
0
16 Oct 2023
Calibrating Likelihoods towards Consistency in Summarization Models
Calibrating Likelihoods towards Consistency in Summarization Models
Polina Zablotskaia
Misha Khalman
Rishabh Joshi
Livio Baldini Soares
Shoshana Jakobovits
Joshua Maynez
Shashi Narayan
49
4
0
12 Oct 2023
Eliciting Model Steering Interactions from Users via Data and Visual
  Design Probes
Eliciting Model Steering Interactions from Users via Data and Visual Design Probes
Anamaria Crisan
Maddie Shang
Eric Brochu
76
3
0
12 Oct 2023
Octopus: Embodied Vision-Language Programmer from Environmental Feedback
Octopus: Embodied Vision-Language Programmer from Environmental Feedback
Jingkang Yang
Yuhao Dong
Shuai Liu
Yue Liu
Ziyue Wang
...
Haoran Tan
Jiamu Kang
Yuanhan Zhang
Kaiyang Zhou
Ziwei Liu
LM&Ro
92
49
0
12 Oct 2023
Towards Better Evaluation of Instruction-Following: A Case-Study in
  Summarization
Towards Better Evaluation of Instruction-Following: A Case-Study in Summarization
Ondrej Skopek
Rahul Aralikatte
Sian Gooding
Victor Carbune
ELM
102
19
0
12 Oct 2023
Receive, Reason, and React: Drive as You Say with Large Language Models
  in Autonomous Vehicles
Receive, Reason, and React: Drive as You Say with Large Language Models in Autonomous Vehicles
Can Cui
Yunsheng Ma
Xu Cao
Wenqian Ye
Ziran Wang
97
91
0
12 Oct 2023
What Matters to You? Towards Visual Representation Alignment for Robot
  Learning
What Matters to You? Towards Visual Representation Alignment for Robot Learning
Ran Tian
Chenfeng Xu
Masayoshi Tomizuka
Jitendra Malik
Andrea V. Bajcsy
80
10
0
11 Oct 2023
The Past, Present and Better Future of Feedback Learning in Large
  Language Models for Subjective Human Preferences and Values
The Past, Present and Better Future of Feedback Learning in Large Language Models for Subjective Human Preferences and Values
Hannah Rose Kirk
Andrew M. Bean
Bertie Vidgen
Paul Röttger
Scott A. Hale
ALM
119
50
0
11 Oct 2023
Previous
123...212223...293031
Next