ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2505.16245
  4. Cited By
Diverse, not Short: A Length-Controlled Self-Learning Framework for Improving Response Diversity of Language Models
v1v2 (latest)

Diverse, not Short: A Length-Controlled Self-Learning Framework for Improving Response Diversity of Language Models

22 May 2025
Vijeta Deshpande
Debasmita Ghose
John D. Patterson
Roger Beaty
Anna Rumshisky
ArXiv (abs)PDFHTML

Papers citing "Diverse, not Short: A Length-Controlled Self-Learning Framework for Improving Response Diversity of Language Models"

32 / 32 papers shown
Title
Base Models Beat Aligned Models at Randomness and Creativity
Base Models Beat Aligned Models at Randomness and Creativity
Peter West
Christopher Potts
453
4
0
30 Apr 2025
Modifying Large Language Model Post-Training for Diverse Creative Writing
Modifying Large Language Model Post-Training for Diverse Creative Writing
John Joon Young Chung
Vishakh Padmakumar
Melissa Roemmele
Yuqian Sun
Max Kreminski
MoMe
75
2
0
21 Mar 2025
Echoes in AI: Quantifying Lack of Plot Diversity in LLM Outputs
Weijia Xu
Nebojsa Jojic
Sudha Rao
Chris Brockett
Bill Dolan
65
7
0
03 Jan 2025
Can AI writing be salvaged? Mitigating Idiosyncrasies and Improving Human-AI Alignment in the Writing Process through Edits
Can AI writing be salvaged? Mitigating Idiosyncrasies and Improving Human-AI Alignment in the Writing Process through Edits
Tuhin Chakrabarty
Philippe Laban
Chien-Sheng Wu
88
13
0
22 Sep 2024
Benchmarking Language Model Creativity: A Case Study on Code Generation
Benchmarking Language Model Creativity: A Case Study on Code Generation
Yining Lu
Dixuan Wang
Tianjian Li
Dongwei Jiang
Daniel Khashabi
Meng Jiang
Daniel Khashabi
LRM
100
15
0
12 Jul 2024
Detection and Measurement of Syntactic Templates in Generated Text
Detection and Measurement of Syntactic Templates in Generated Text
Chantal Shaib
Yanai Elazar
Junyi Jessy Li
Byron C. Wallace
85
20
0
28 Jun 2024
Scaling Synthetic Data Creation with 1,000,000,000 Personas
Scaling Synthetic Data Creation with 1,000,000,000 Personas
Tao Ge
Xin Chan
Dian Yu
Haitao Mi
Dong Yu
Dong Yu
SyDa
199
150
0
28 Jun 2024
Interpretable Preferences via Multi-Objective Reward Modeling and
  Mixture-of-Experts
Interpretable Preferences via Multi-Objective Reward Modeling and Mixture-of-Experts
Haoxiang Wang
Wei Xiong
Tengyang Xie
Han Zhao
Tong Zhang
105
179
0
18 Jun 2024
Toward Self-Improvement of LLMs via Imagination, Searching, and
  Criticizing
Toward Self-Improvement of LLMs via Imagination, Searching, and Criticizing
Ye Tian
Baolin Peng
Linfeng Song
Lifeng Jin
Dian Yu
Haitao Mi
Dong Yu
LRMReLM
91
85
0
18 Apr 2024
How Bad is Training on Synthetic Data? A Statistical Analysis of
  Language Model Collapse
How Bad is Training on Synthetic Data? A Statistical Analysis of Language Model Collapse
M. Seddik
Suei-Wen Chen
Soufiane Hayou
Pierre Youssef
Merouane Debbah
91
36
0
07 Apr 2024
Emergent Abilities in Reduced-Scale Generative Language Models
Emergent Abilities in Reduced-Scale Generative Language Models
Sherin Muckatira
Vijeta Deshpande
Vladislav Lialin
Anna Rumshisky
ReLMELMLRM
46
5
0
02 Apr 2024
Standardizing the Measurement of Text Diversity: A Tool and a Comparative Analysis of Scores
Standardizing the Measurement of Text Diversity: A Tool and a Comparative Analysis of Scores
Chantal Shaib
Joe Barrow
Jiuding Sun
Alexa F. Siu
Byron C. Wallace
A. Nenkova
134
38
0
01 Mar 2024
Homogenization Effects of Large Language Models on Human Creative
  Ideation
Homogenization Effects of Large Language Models on Human Creative Ideation
Barrett R Anderson
Jash Hemant Shah
Max Kreminski
67
82
0
02 Feb 2024
The Curious Decline of Linguistic Diversity: Training Language Models on
  Synthetic Text
The Curious Decline of Linguistic Diversity: Training Language Models on Synthetic Text
Yanzhu Guo
Guokan Shang
Michalis Vazirgiannis
Chloé Clavel
71
58
0
16 Nov 2023
Safe RLHF: Safe Reinforcement Learning from Human Feedback
Safe RLHF: Safe Reinforcement Learning from Human Feedback
Josef Dai
Xuehai Pan
Ruiyang Sun
Jiaming Ji
Xinbo Xu
Mickel Liu
Yizhou Wang
Yaodong Yang
121
361
0
19 Oct 2023
A Confederacy of Models: a Comprehensive Evaluation of LLMs on Creative
  Writing
A Confederacy of Models: a Comprehensive Evaluation of LLMs on Creative Writing
Carlos Gómez-Rodríguez
Paul Williams
66
82
0
12 Oct 2023
Understanding the Effects of RLHF on LLM Generalisation and Diversity
Understanding the Effects of RLHF on LLM Generalisation and Diversity
Robert Kirk
Ishita Mediratta
Christoforos Nalmpantis
Jelena Luketina
Eric Hambro
Edward Grefenstette
Roberta Raileanu
AI4CEALM
174
149
0
10 Oct 2023
Creativity Support in the Age of Large Language Models: An Empirical
  Study Involving Emerging Writers
Creativity Support in the Age of Large Language Models: An Empirical Study Involving Emerging Writers
Tuhin Chakrabarty
Vishakh Padmakumar
Faeze Brahman
Smaranda Muresan
87
40
0
22 Sep 2023
Textbooks Are All You Need II: phi-1.5 technical report
Textbooks Are All You Need II: phi-1.5 technical report
Yuan-Fang Li
Sébastien Bubeck
Ronen Eldan
Allison Del Giorno
Suriya Gunasekar
Yin Tat Lee
ALMLRM
161
481
0
11 Sep 2023
Does Writing with Language Models Reduce Content Diversity?
Does Writing with Language Models Reduce Content Diversity?
Vishakh Padmakumar
He He
79
91
0
11 Sep 2023
Increasing Diversity While Maintaining Accuracy: Text Data Generation
  with Large Language Models and Human Interventions
Increasing Diversity While Maintaining Accuracy: Text Data Generation with Large Language Models and Human Interventions
John Joon Young Chung
Ece Kamar
Saleema Amershi
ALM
96
121
0
07 Jun 2023
Direct Preference Optimization: Your Language Model is Secretly a Reward
  Model
Direct Preference Optimization: Your Language Model is Secretly a Reward Model
Rafael Rafailov
Archit Sharma
E. Mitchell
Stefano Ermon
Christopher D. Manning
Chelsea Finn
ALM
389
4,139
0
29 May 2023
The Curse of Recursion: Training on Generated Data Makes Models Forget
The Curse of Recursion: Training on Generated Data Makes Models Forget
Ilia Shumailov
Zakhar Shumaylov
Yiren Zhao
Y. Gal
Nicolas Papernot
Ross J. Anderson
DiffM
79
298
0
27 May 2023
Honey, I Shrunk the Language: Language Model Behavior at Reduced Scale
Honey, I Shrunk the Language: Language Model Behavior at Reduced Scale
Vijeta Deshpande
Dan Pechi
Shree Thatte
Vladislav Lialin
Anna Rumshisky
111
8
0
26 May 2023
QLoRA: Efficient Finetuning of Quantized LLMs
QLoRA: Efficient Finetuning of Quantized LLMs
Tim Dettmers
Artidoro Pagnoni
Ari Holtzman
Luke Zettlemoyer
ALM
152
2,602
0
23 May 2023
The Flan Collection: Designing Data and Methods for Effective
  Instruction Tuning
The Flan Collection: Designing Data and Methods for Effective Instruction Tuning
Shayne Longpre
Le Hou
Tu Vu
Albert Webson
Hyung Won Chung
...
Denny Zhou
Quoc V. Le
Barret Zoph
Jason W. Wei
Adam Roberts
ALM
112
677
0
31 Jan 2023
Training a Helpful and Harmless Assistant with Reinforcement Learning
  from Human Feedback
Training a Helpful and Harmless Assistant with Reinforcement Learning from Human Feedback
Yuntao Bai
Andy Jones
Kamal Ndousse
Amanda Askell
Anna Chen
...
Jack Clark
Sam McCandlish
C. Olah
Benjamin Mann
Jared Kaplan
256
2,611
0
12 Apr 2022
Training language models to follow instructions with human feedback
Training language models to follow instructions with human feedback
Long Ouyang
Jeff Wu
Xu Jiang
Diogo Almeida
Carroll L. Wainwright
...
Amanda Askell
Peter Welinder
Paul Christiano
Jan Leike
Ryan J. Lowe
OSLMALM
883
13,176
0
04 Mar 2022
LoRA: Low-Rank Adaptation of Large Language Models
LoRA: Low-Rank Adaptation of Large Language Models
J. E. Hu
Yelong Shen
Phillip Wallis
Zeyuan Allen-Zhu
Yuanzhi Li
Shean Wang
Lu Wang
Weizhu Chen
OffRLAI4TSAI4CEALMAIMat
490
10,496
0
17 Jun 2021
Proximal Policy Optimization Algorithms
Proximal Policy Optimization Algorithms
John Schulman
Filip Wolski
Prafulla Dhariwal
Alec Radford
Oleg Klimov
OffRL
535
19,265
0
20 Jul 2017
Learning the Preferences of Ignorant, Inconsistent Agents
Learning the Preferences of Ignorant, Inconsistent Agents
Owain Evans
Andreas Stuhlmuller
Noah D. Goodman
64
128
0
18 Dec 2015
A Diversity-Promoting Objective Function for Neural Conversation Models
A Diversity-Promoting Objective Function for Neural Conversation Models
Jiwei Li
Michel Galley
Chris Brockett
Jianfeng Gao
W. Dolan
145
2,402
0
11 Oct 2015
1