ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2310.17022
  4. Cited By
Controlled Decoding from Language Models
v1v2v3 (latest)

Controlled Decoding from Language Models

25 October 2023
Sidharth Mudgal
Jong Lee
H. Ganapathy
Yaguang Li
Tao Wang
Yanping Huang
Zhifeng Chen
Heng-Tze Cheng
Michael Collins
Trevor Strohman
Jilin Chen
Alex Beutel
Ahmad Beirami
ArXiv (abs)PDFHTML

Papers citing "Controlled Decoding from Language Models"

50 / 62 papers shown
Title
$\texttt{SPECS}$: Faster Test-Time Scaling through Speculative Drafts
SPECS\texttt{SPECS}SPECS: Faster Test-Time Scaling through Speculative Drafts
Mert Cemri
Nived Rajaraman
Rishabh Tiwari
Xiaoxuan Liu
Kurt Keutzer
Ion Stoica
Kannan Ramchandran
Ahmad Beirami
Ziteng Sun
LRM
19
0
0
15 Jun 2025
Tokenized Bandit for LLM Decoding and Alignment
Tokenized Bandit for LLM Decoding and Alignment
Suho Shin
Chenghao Yang
Haifeng Xu
Mohammad T. Hajiaghayi
28
0
0
08 Jun 2025
Guided Speculative Inference for Efficient Test-Time Alignment of LLMs
Guided Speculative Inference for Efficient Test-Time Alignment of LLMs
Jonathan Geuter
Youssef Mroueh
David Alvarez-Melis
103
1
0
04 Jun 2025
Advancing Decoding Strategies: Enhancements in Locally Typical Sampling for LLMs
Advancing Decoding Strategies: Enhancements in Locally Typical Sampling for LLMs
Jaydip Sen
Saptarshi Sengupta
S. Dasgupta
50
0
0
03 Jun 2025
Soft Reasoning: Navigating Solution Spaces in Large Language Models through Controlled Embedding Exploration
Soft Reasoning: Navigating Solution Spaces in Large Language Models through Controlled Embedding Exploration
Qinglin Zhu
Runcong Zhao
Hanqi Yan
Yulan He
Yudong Chen
Lin Gui
LRM
33
0
0
30 May 2025
EnsemW2S: Enhancing Weak-to-Strong Generalization with Large Language Model Ensembles
EnsemW2S: Enhancing Weak-to-Strong Generalization with Large Language Model Ensembles
Aakriti Agrawal
Mucong Ding
Zora Che
Chenghao Deng
Anirudh Satheesh
Bang An
Bayan Bruss
John Langford
Furong Huang
74
0
0
28 May 2025
Efficient Controllable Diffusion via Optimal Classifier Guidance
Efficient Controllable Diffusion via Optimal Classifier Guidance
Owen Oertell
Shikun Sun
Yiding Chen
Jin Peng Zhou
Zhiyong Wang
Wen Sun
48
0
0
27 May 2025
Diffusion Blend: Inference-Time Multi-Preference Alignment for Diffusion Models
Diffusion Blend: Inference-Time Multi-Preference Alignment for Diffusion Models
Min Cheng
Fatemeh Doudi
D. Kalathil
Mohammad Ghavamzadeh
P. R. Kumar
57
0
0
24 May 2025
Value-Guided Search for Efficient Chain-of-Thought Reasoning
Value-Guided Search for Efficient Chain-of-Thought Reasoning
Kaiwen Wang
Jin Peng Zhou
Jonathan D. Chang
Zhaolin Gao
Nathan Kallus
Kianté Brantley
Wen Sun
LRM
90
1
0
23 May 2025
Incremental Sequence Classification with Temporal Consistency
Incremental Sequence Classification with Temporal Consistency
Lucas Maystre
Gabriel Barello
Tudor Berariu
Aleix Cambray
Rares Dolga
Alvaro Ortega Gonzalez
Andrei Nica
David Barber
CLL
57
0
0
22 May 2025
Guided Search Strategies in Non-Serializable Environments with Applications to Software Engineering Agents
Guided Search Strategies in Non-Serializable Environments with Applications to Software Engineering Agents
Karina Zainullina
Alexander Golubev
Maria Trofimova
Sergei Polezhaev
Ibragim Badertdinov
...
Filipp Fisin
Sergei Skvortsov
Maksim Nekrashevich
Anton Shevtsov
Boris Yangel
62
0
0
19 May 2025
TRACE Back from the Future: A Probabilistic Reasoning Approach to Controllable Language Generation
TRACE Back from the Future: A Probabilistic Reasoning Approach to Controllable Language Generation
Gwen Yidou Weng
Benjie Wang
Guy Van den Broeck
BDL
423
0
0
25 Apr 2025
Antidistillation Sampling
Antidistillation Sampling
Yash Savani
Asher Trockman
Zhili Feng
Avi Schwarzschild
Alexander Robey
Marc Finzi
J. Zico Kolter
127
3
0
17 Apr 2025
Syntactic and Semantic Control of Large Language Models via Sequential Monte Carlo
Syntactic and Semantic Control of Large Language Models via Sequential Monte Carlo
João Loula
Benjamin LeBrun
Li Du
Ben Lipkin
Clemente Pasti
...
Ryan Cotterel
Vikash K. Mansinghka
Alexander K. Lew
Tim Vieira
Timothy J. O'Donnell
163
8
0
17 Apr 2025
Bridging the Gap Between Preference Alignment and Machine Unlearning
Bridging the Gap Between Preference Alignment and Machine Unlearning
Xiaohua Feng
Yuyuan Li
Huwei Ji
Jiaming Zhang
Lulu Zhang
Tianyu Du
Chaochao Chen
MU
93
0
0
09 Apr 2025
A Survey on Personalized and Pluralistic Preference Alignment in Large Language Models
A Survey on Personalized and Pluralistic Preference Alignment in Large Language Models
Zhouhang Xie
Junda Wu
Yiran Shen
Yu Xia
Xintong Li
...
Sachin Kumar
Bodhisattwa Prasad Majumder
Jingbo Shang
Prithviraj Ammanabrolu
Julian McAuley
158
1
0
09 Apr 2025
On the Role of Feedback in Test-Time Scaling of Agentic AI Workflows
On the Role of Feedback in Test-Time Scaling of Agentic AI Workflows
Souradip Chakraborty
Mohammadreza Pourreza
Ruoxi Sun
Yiwen Song
Nino Scherrer
...
Furong Huang
Amrit Singh Bedi
Ahmad Beirami
Hamid Palangi
Tomas Pfister
129
2
0
02 Apr 2025
reWordBench: Benchmarking and Improving the Robustness of Reward Models with Transformed Inputs
reWordBench: Benchmarking and Improving the Robustness of Reward Models with Transformed Inputs
Zhaofeng Wu
Michihiro Yasunaga
Andrew Cohen
Yoon Kim
Asli Celikyilmaz
Marjan Ghazvininejad
90
3
0
14 Mar 2025
Robust Multi-Objective Controlled Decoding of Large Language Models
Seongho Son
William Bankes
Sangwoong Yoon
Shyam Sundhar Ramesh
Xiaohang Tang
Ilija Bogunovic
127
2
0
11 Mar 2025
Language Model Personalization via Reward Factorization
Idan Shenfeld
Felix Faltings
Pulkit Agrawal
Aldo Pacchiano
109
1
0
08 Mar 2025
DiffPO: Diffusion-styled Preference Optimization for Efficient Inference-Time Alignment of Large Language Models
DiffPO: Diffusion-styled Preference Optimization for Efficient Inference-Time Alignment of Large Language Models
Ruizhe Chen
Wenhao Chai
Zhifei Yang
Xiaotian Zhang
Qiufeng Wang
Tony Q.S. Quek
Soujanya Poria
Zuozhu Liu
133
1
0
06 Mar 2025
Dynamic Search for Inference-Time Alignment in Diffusion Models
Dynamic Search for Inference-Time Alignment in Diffusion Models
Xiner Li
Masatoshi Uehara
Xingyu Su
Gabriele Scalia
Tommaso Biancalani
Aviv Regev
Sergey Levine
Shuiwang Ji
96
4
0
03 Mar 2025
Sampling-Efficient Test-Time Scaling: Self-Estimating the Best-of-N Sampling in Early Decoding
Yun Wang
Pei Zhang
Siyuan Huang
Baosong Yang
Zizhuo Zhang
Fei Huang
Rui Wang
BDLLRM
139
11
0
03 Mar 2025
Constrained Decoding with Speculative Lookaheads
Constrained Decoding with Speculative Lookaheads
Nishanth Nakshatri
Shamik Roy
Rajarshi Das
Suthee Chaidaroon
Leonid Boytsov
Rashmi Gangadharaiah
199
0
0
09 Dec 2024
Time-Reversal Provides Unsupervised Feedback to LLMs
Time-Reversal Provides Unsupervised Feedback to LLMs
Yerram Varun
Rahul Madhavan
Sravanti Addepalli
A. Suggala
Karthikeyan Shanmugam
Prateek Jain
LRMSyDa
107
0
0
03 Dec 2024
Synthesize, Partition, then Adapt: Eliciting Diverse Samples from
  Foundation Models
Synthesize, Partition, then Adapt: Eliciting Diverse Samples from Foundation Models
Yeming Wen
Swarat Chaudhuri
85
0
0
11 Nov 2024
Fast Best-of-N Decoding via Speculative Rejection
Fast Best-of-N Decoding via Speculative Rejection
Hanshi Sun
Momin Haider
Ruiqi Zhang
Huitao Yang
Jiahao Qiu
Ming Yin
Mengdi Wang
Peter L. Bartlett
Andrea Zanette
BDL
117
52
0
26 Oct 2024
TreeBoN: Enhancing Inference-Time Alignment with Speculative Tree-Search
  and Best-of-N Sampling
TreeBoN: Enhancing Inference-Time Alignment with Speculative Tree-Search and Best-of-N Sampling
Jiahao Qiu
Yifu Lu
Yifan Zeng
Jiacheng Guo
Jiayi Geng
Huazheng Wang
Kaixuan Huang
Yue Wu
Mengdi Wang
112
29
0
18 Oct 2024
Controllable Safety Alignment: Inference-Time Adaptation to Diverse Safety Requirements
Controllable Safety Alignment: Inference-Time Adaptation to Diverse Safety Requirements
Jingyu Zhang
Ahmed Elgohary
Ahmed Magooda
Daniel Khashabi
Benjamin Van Durme
473
8
0
11 Oct 2024
GenARM: Reward Guided Generation with Autoregressive Reward Model for Test-time Alignment
GenARM: Reward Guided Generation with Autoregressive Reward Model for Test-time Alignment
Yuancheng Xu
Udari Madhushani Sehwag
Alec Koppel
Sicheng Zhu
Bang An
Furong Huang
Sumitra Ganesh
161
14
0
10 Oct 2024
Guaranteed Generation from Large Language Models
Guaranteed Generation from Large Language Models
Minbeom Kim
Thibaut Thonet
Jos Rozen
Hwaran Lee
Kyomin Jung
Marc Dymetman
89
1
0
09 Oct 2024
EVOLvE: Evaluating and Optimizing LLMs For Exploration
EVOLvE: Evaluating and Optimizing LLMs For Exploration
Allen Nie
Yi Su
Bo Chang
Jonathan N. Lee
Ed H. Chi
Quoc V. Le
Minmin Chen
64
14
0
08 Oct 2024
EnsemW2S: Can an Ensemble of LLMs be Leveraged to Obtain a Stronger LLM?
EnsemW2S: Can an Ensemble of LLMs be Leveraged to Obtain a Stronger LLM?
Aakriti Agrawal
Mucong Ding
Zora Che
Chenghao Deng
Anirudh Satheesh
John Langford
Furong Huang
95
8
0
06 Oct 2024
AIME: AI System Optimization via Multiple LLM Evaluators
AIME: AI System Optimization via Multiple LLM Evaluators
Bhrij Patel
Souradip Chakraborty
Wesley A Suttle
Mengdi Wang
Amrit Singh Bedi
Dinesh Manocha
65
15
0
04 Oct 2024
Step-by-Step Reasoning for Math Problems via Twisted Sequential Monte Carlo
Step-by-Step Reasoning for Math Problems via Twisted Sequential Monte Carlo
Shengyu Feng
Xiang Kong
Shuang Ma
Aonan Zhang
Dong Yin
Chong-Jun Wang
Ruoming Pang
Yiming Yang
LRM
120
2
0
02 Oct 2024
Towards a Unified View of Preference Learning for Large Language Models:
  A Survey
Towards a Unified View of Preference Learning for Large Language Models: A Survey
Bofei Gao
Feifan Song
Yibo Miao
Zefan Cai
Zhiyong Yang
...
Houfeng Wang
Zhifang Sui
Peiyi Wang
Baobao Chang
Baobao Chang
155
14
0
04 Sep 2024
Derivative-Free Guidance in Continuous and Discrete Diffusion Models
  with Soft Value-Based Decoding
Derivative-Free Guidance in Continuous and Discrete Diffusion Models with Soft Value-Based Decoding
Xiner Li
Yulai Zhao
Chenyu Wang
Gabriele Scalia
Gökçen Eraslan
Surag Nair
Tommaso Biancalani
Aviv Regev
Sergey Levine
Masatoshi Uehara
119
37
0
15 Aug 2024
Variational Best-of-N Alignment
Variational Best-of-N Alignment
Afra Amini
Tim Vieira
Ryan Cotterell
Ryan Cotterell
BDL
109
23
0
08 Jul 2024
Decoding-Time Language Model Alignment with Multiple Objectives
Decoding-Time Language Model Alignment with Multiple Objectives
Ruizhe Shi
Yifang Chen
Yushi Hu
Alisa Liu
Hannaneh Hajishirzi
Noah A. Smith
Simon Du
140
43
0
27 Jun 2024
Cascade Reward Sampling for Efficient Decoding-Time Alignment
Cascade Reward Sampling for Efficient Decoding-Time Alignment
Bolian Li
Yifan Wang
A. Grama
Ruqi Zhang
Ruqi Zhang
AI4TS
143
15
0
24 Jun 2024
Aligning Large Language Models with Representation Editing: A Control
  Perspective
Aligning Large Language Models with Representation Editing: A Control Perspective
Lingkai Kong
Haorui Wang
Wenhao Mu
Yuanqi Du
Yuchen Zhuang
Yifei Zhou
Yue Song
Rongzhi Zhang
Kai Wang
Chao Zhang
94
26
0
10 Jun 2024
Safety Alignment Should Be Made More Than Just a Few Tokens Deep
Safety Alignment Should Be Made More Than Just a Few Tokens Deep
Xiangyu Qi
Ashwinee Panda
Kaifeng Lyu
Xiao Ma
Subhrajit Roy
Ahmad Beirami
Prateek Mittal
Peter Henderson
120
142
0
10 Jun 2024
Information Theoretic Guarantees For Policy Alignment In Large Language
  Models
Information Theoretic Guarantees For Policy Alignment In Large Language Models
Youssef Mroueh
97
8
0
09 Jun 2024
BoNBoN Alignment for Large Language Models and the Sweetness of
  Best-of-n Sampling
BoNBoN Alignment for Large Language Models and the Sweetness of Best-of-n Sampling
Lin Gui
Cristina Garbacea
Victor Veitch
BDLLM&MA
112
49
0
02 Jun 2024
Transfer Q Star: Principled Decoding for LLM Alignment
Transfer Q Star: Principled Decoding for LLM Alignment
Souradip Chakraborty
Soumya Suvra Ghosal
Ming Yin
Dinesh Manocha
Mengdi Wang
Amrit Singh Bedi
Furong Huang
120
33
0
30 May 2024
Weak-to-Strong Search: Align Large Language Models via Searching over
  Small Language Models
Weak-to-Strong Search: Align Large Language Models via Searching over Small Language Models
Zhanhui Zhou
Zhixuan Liu
Jie Liu
Zhichen Dong
Chao Yang
Yu Qiao
ALM
109
27
0
29 May 2024
Robust Preference Optimization through Reward Model Distillation
Robust Preference Optimization through Reward Model Distillation
Adam Fisch
Jacob Eisenstein
Vicky Zayats
Alekh Agarwal
Ahmad Beirami
Chirag Nagpal
Peter Shaw
Jonathan Berant
154
37
0
29 May 2024
Impact of Non-Standard Unicode Characters on Security and Comprehension
  in Large Language Models
Impact of Non-Standard Unicode Characters on Security and Comprehension in Large Language Models
Johan S Daniel
Anand Pal
55
1
0
23 May 2024
Value Augmented Sampling for Language Model Alignment and
  Personalization
Value Augmented Sampling for Language Model Alignment and Personalization
Seungwook Han
Idan Shenfeld
Akash Srivastava
Yoon Kim
Pulkit Agrawal
OffRL
88
29
0
10 May 2024
Probabilistic Inference in Language Models via Twisted Sequential Monte
  Carlo
Probabilistic Inference in Language Models via Twisted Sequential Monte Carlo
Stephen Zhao
Rob Brekelmans
Alireza Makhzani
Roger C. Grosse
89
41
0
26 Apr 2024
12
Next