Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2205.12548
Cited By
RLPrompt: Optimizing Discrete Text Prompts with Reinforcement Learning
25 May 2022
Mingkai Deng
Jianyu Wang
Cheng-Ping Hsieh
Yihan Wang
Han Guo
Tianmin Shu
Meng Song
Eric P. Xing
Zhiting Hu
Re-assign community
ArXiv
PDF
HTML
Papers citing
"RLPrompt: Optimizing Discrete Text Prompts with Reinforcement Learning"
50 / 69 papers shown
Title
PLHF: Prompt Optimization with Few-Shot Human Feedback
Chun-Pai Yang
Kan Zheng
Shou-De Lin
21
0
0
11 May 2025
CAPO: Cost-Aware Prompt Optimization
Tom Zehle
Moritz Schlager
Timo Heiß
Matthias Feurer
VLM
52
0
0
22 Apr 2025
Prompt Optimization with Logged Bandit Data
Haruka Kiyohara
Daniel Yiming Cao
Yuta Saito
Thorsten Joachims
64
0
0
03 Apr 2025
ProAPO: Progressively Automatic Prompt Optimization for Visual Classification
Xiangyan Qu
Gaopeng Gou
Jiamin Zhuang
Jing Yu
Kun Song
Qihao Wang
Yili Li
Gang Xiong
VLM
83
0
0
13 Mar 2025
Oreo: A Plug-in Context Reconstructor to Enhance Retrieval-Augmented Generation
Sha Li
Naren Ramakrishnan
RALM
KELM
149
1
0
18 Feb 2025
RePrompt: Planning by Automatic Prompt Engineering for Large Language Models Agents
Weizhe Chen
Sven Koenig
B. Dilkina
LLMAG
104
8
0
17 Feb 2025
Commercial LLM Agents Are Already Vulnerable to Simple Yet Dangerous Attacks
Ang Li
Yin Zhou
Vethavikashini Chithrra Raghuram
Tom Goldstein
Micah Goldblum
AAML
76
7
0
12 Feb 2025
When LLM Meets DRL: Advancing Jailbreaking Efficiency via DRL-guided Search
Xuan Chen
Yuzhou Nie
Wenbo Guo
Xiangyu Zhang
110
9
0
28 Jan 2025
Unleashing the Potential of Large Language Models as Prompt Optimizers: Analogical Analysis with Gradient-based Model Optimizers
Xinyu Tang
Xiaolei Wang
Wayne Xin Zhao
Siyuan Lu
Yaliang Li
Ji-Rong Wen
LRM
56
13
0
28 Jan 2025
The potential -- and the pitfalls -- of using pre-trained language models as cognitive science theories
Raj Sanjay Shah
Sashank Varma
LRM
89
0
0
22 Jan 2025
QualityFlow: An Agentic Workflow for Program Synthesis Controlled by LLM Quality Checks
Yaojie Hu
Qiang Zhou
Qihong Chen
Xiaopeng Li
Linbo Liu
Dejiao Zhang
Amit Kachroo
Talha Oz
Omer Tripp
66
4
0
20 Jan 2025
TAPO: Task-Referenced Adaptation for Prompt Optimization
Wenxin Luo
W. Wang
Xiaopeng Li
Weibo Zhou
Pengyue Jia
Xiangyu Zhao
47
0
0
12 Jan 2025
TestAgent: A Framework for Domain-Adaptive Evaluation of LLMs via Dynamic Benchmark Construction and Exploratory Interaction
Wanying Wang
Zeyu Ma
Pengfei Liu
Mingang Chen
LLMAG
45
1
0
15 Oct 2024
Survival of the Safest: Towards Secure Prompt Optimization through Interleaved Multi-Objective Evolution
Ankita Sinha
Wendi Cui
Kamalika Das
Jiaxin Zhang
AAML
28
2
0
12 Oct 2024
StablePrompt: Automatic Prompt Tuning using Reinforcement Learning for Large Language Models
Minchan Kwon
Gaeun Kim
Jongsuk Kim
Haeil Lee
Junmo Kim
OffRL
LRM
LLMAG
21
2
0
10 Oct 2024
Detecting Training Data of Large Language Models via Expectation Maximization
Gyuwan Kim
Yang Li
Evangelia Spiliopoulou
Jie Ma
Miguel Ballesteros
William Yang Wang
MIALM
95
3
2
10 Oct 2024
Retrieval Augmented Generation (RAG) and Beyond: A Comprehensive Survey on How to Make your LLMs use External Data More Wisely
Siyun Zhao
Yuqing Yang
Zilong Wang
Zhiyuan He
Luna Qiu
Lili Qiu
SyDa
RALM
3DV
44
34
0
23 Sep 2024
Prompt Obfuscation for Large Language Models
David Pape
Thorsten Eisenhofer
Thorsten Eisenhofer
Lea Schönherr
AAML
33
2
0
17 Sep 2024
Explaining Datasets in Words: Statistical Models with Natural Language Parameters
Ruiqi Zhong
Heng Wang
Dan Klein
Jacob Steinhardt
35
6
0
13 Sep 2024
Hard Prompts Made Interpretable: Sparse Entropy Regularization for Prompt Tuning with RL
Yunseon Choi
Sangmin Bae
Seonghyun Ban
Minchan Jeong
Chuheng Zhang
Lei Song
Li Zhao
Jiang Bian
Kee-Eung Kim
VLM
AAML
34
3
0
20 Jul 2024
Prompting Techniques for Secure Code Generation: A Systematic Investigation
Catherine Tony
Nicolás E. Díaz Ferreyra
Markus Mutas
Salem Dhiff
Riccardo Scandariato
SILM
71
9
0
09 Jul 2024
MAPO: Boosting Large Language Model Performance with Model-Adaptive Prompt Optimization
Yuyan Chen
Zhihao Wen
Ge Fan
Zhengyu Chen
Wei Yu Wu
Dayiheng Liu
Zhixu Li
Bang Liu
Yanghua Xiao
31
18
0
04 Jul 2024
Few-shot Personalization of LLMs with Mis-aligned Responses
Jaehyung Kim
Yiming Yang
47
7
0
26 Jun 2024
On the Transformations across Reward Model, Parameter Update, and In-Context Prompt
Deng Cai
Huayang Li
Tingchen Fu
Siheng Li
Weiwen Xu
...
Leyang Cui
Yan Wang
Lemao Liu
Taro Watanabe
Shuming Shi
KELM
30
2
0
24 Jun 2024
Demonstration Notebook: Finding the Most Suited In-Context Learning Example from Interactions
Yiming Tang
Bin Dong
36
0
0
16 Jun 2024
Concentrate Attention: Towards Domain-Generalizable Prompt Optimization for Language Models
Chengzhengxu Li
Xiaoming Liu
Zhaohan Zhang
Yichen Wang
Chen Liu
Y. Lan
Chao Shen
48
2
0
15 Jun 2024
Evaluation of Few-Shot Learning for Classification Tasks in the Polish Language
Tsimur Hadeliya
D. Kajtoch
43
0
0
27 Apr 2024
Examining the robustness of LLM evaluation to the distributional assumptions of benchmarks
Melissa Ailem
Katerina Marazopoulou
Charlotte Siska
James Bono
56
14
0
25 Apr 2024
PURPLE: Making a Large Language Model a Better SQL Writer
Tonghui Ren
Yuankai Fan
Zhenying He
Ren Huang
Jiaqi Dai
Can Huang
Yinan Jing
Kai Zhang
Yifan Yang
Xiaoyang Sean Wang
23
21
0
29 Mar 2024
Monotonic Paraphrasing Improves Generalization of Language Model Prompting
Qin Liu
Fei Wang
Nan Xu
Tianyi Yan
Tao Meng
Muhao Chen
LRM
41
7
0
24 Mar 2024
Can LLMs Separate Instructions From Data? And What Do We Even Mean By That?
Egor Zverev
Sahar Abdelnabi
Soroush Tabesh
Mario Fritz
Christoph H. Lampert
48
19
0
11 Mar 2024
MORL-Prompt: An Empirical Analysis of Multi-Objective Reinforcement Learning for Discrete Prompt Optimization
Yasaman Jafari
Dheeraj Mekala
Rose Yu
Taylor Berg-Kirkpatrick
53
7
0
18 Feb 2024
The RL/LLM Taxonomy Tree: Reviewing Synergies Between Reinforcement Learning and Large Language Models
M. Pternea
Prerna Singh
Abir Chakraborty
Y. Oruganti
M. Milletarí
Sayli Bapat
Kebei Jiang
OffRL
18
7
0
02 Feb 2024
Gradient-Based Language Model Red Teaming
Nevan Wichers
Carson E. Denison
Ahmad Beirami
14
25
0
30 Jan 2024
Black-Box Access is Insufficient for Rigorous AI Audits
Stephen Casper
Carson Ezell
Charlotte Siegmann
Noam Kolt
Taylor Lynn Curtis
...
Michael Gerovitch
David Bau
Max Tegmark
David M. Krueger
Dylan Hadfield-Menell
AAML
34
78
0
25 Jan 2024
AutoTask: Executing Arbitrary Voice Commands by Exploring and Learning from Mobile GUI
Lihang Pan
Bowen Wang
Chun Yu
Yuxuan Chen
Xiangyu Zhang
Yuanchun Shi
21
3
0
26 Dec 2023
DP-OPT: Make Large Language Model Your Privacy-Preserving Prompt Engineer
Junyuan Hong
Jiachen T. Wang
Chenhui Zhang
Zhangheng Li
Bo-wen Li
Zhangyang Wang
42
29
0
27 Nov 2023
Reinforcement Learning for Generative AI: A Survey
Yuanjiang Cao
Quan.Z Sheng
Julian McAuley
Lina Yao
SyDa
46
10
0
28 Aug 2023
Manipulating Embeddings of Stable Diffusion Prompts
Niklas Deckers
Julia Peters
Martin Potthast
DiffM
37
9
0
23 Aug 2023
PACE: Improving Prompt with Actor-Critic Editing for Large Language Model
Yihong Dong
Kangcheng Luo
Xue Jiang
Zhi Jin
Ge Li
LRM
KELM
20
9
0
19 Aug 2023
AutoHint: Automatic Prompt Optimization with Hint Generation
Hong Sun
Xue Li
Yi Xu
Youkow Homma
Qinhao Cao
Min-man Wu
Jian Jiao
Denis Xavier Charles
31
23
0
13 Jul 2023
Better Zero-Shot Reasoning with Self-Adaptive Prompting
Xingchen Wan
Ruoxi Sun
H. Dai
Sercan Ö. Arik
Tomas Pfister
ReLM
OffRL
LRM
16
48
0
23 May 2023
Flatness-Aware Prompt Selection Improves Accuracy and Sample Efficiency
Lingfeng Shen
Weiting Tan
Boyuan Zheng
Daniel Khashabi
VLM
39
6
0
18 May 2023
Reprompting: Automated Chain-of-Thought Prompt Inference Through Gibbs Sampling
Weijia Xu
Andrzej Banburski-Fahey
Nebojsa Jojic
ReLM
LRM
21
32
0
17 May 2023
Prompt-Tuning Decision Transformer with Preference Ranking
Shengchao Hu
Li Shen
Ya-Qin Zhang
Dacheng Tao
OffRL
26
14
0
16 May 2023
VPGTrans: Transfer Visual Prompt Generator across LLMs
Ao Zhang
Hao Fei
Yuan Yao
Wei Ji
Li Li
Zhiyuan Liu
Tat-Seng Chua
MLLM
VLM
27
85
0
02 May 2023
Evaluating the Robustness of Discrete Prompts
Yoichi Ishibashi
Danushka Bollegala
Katsuhito Sudoh
Satoshi Nakamura
21
18
0
11 Feb 2023
Multimodality Helps Unimodality: Cross-Modal Few-Shot Learning with Multimodal Models
Zhiqiu Lin
Samuel Yu
Zhiyi Kuang
Deepak Pathak
Deva Ramana
VLM
15
100
0
16 Jan 2023
Parameter-Efficient Fine-Tuning Design Spaces
Jiaao Chen
Aston Zhang
Xingjian Shi
Mu Li
Alexander J. Smola
Diyi Yang
31
59
0
04 Jan 2023
Self-Adaptive In-Context Learning: An Information Compression Perspective for In-Context Example Selection and Ordering
Zhiyong Wu
Yaoxiang Wang
Jiacheng Ye
Lingpeng Kong
39
119
0
20 Dec 2022
1
2
Next