ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2406.16838
  4. Cited By
From Decoding to Meta-Generation: Inference-time Algorithms for Large
  Language Models

From Decoding to Meta-Generation: Inference-time Algorithms for Large Language Models

24 June 2024
Sean Welleck
Amanda Bertsch
Matthew Finlayson
Hailey Schoelkopf
Alex Xie
Graham Neubig
Ilia Kulikov
Zaid Harchaoui
ArXiv (abs)PDFHTML

Papers citing "From Decoding to Meta-Generation: Inference-time Algorithms for Large Language Models"

26 / 26 papers shown
Title
DynScaling: Efficient Verifier-free Inference Scaling via Dynamic and Integrated Sampling
DynScaling: Efficient Verifier-free Inference Scaling via Dynamic and Integrated Sampling
Fei Wang
Xingchen Wan
Ruoxi Sun
Jiefeng Chen
Sercan Ö. Arık
LRM
26
0
0
19 Jun 2025
Learning a Continue-Thinking Token for Enhanced Test-Time Scaling
Learning a Continue-Thinking Token for Enhanced Test-Time Scaling
Liran Ringel
Elad Tolochinsky
Yaniv Romano
LRM
25
0
0
12 Jun 2025
Overclocking LLM Reasoning: Monitoring and Controlling Thinking Path Lengths in LLMs
Overclocking LLM Reasoning: Monitoring and Controlling Thinking Path Lengths in LLMs
Roy Eisenstadt
Itamar Zimerman
Lior Wolf
LRM
21
0
0
08 Jun 2025
Tokenized Bandit for LLM Decoding and Alignment
Tokenized Bandit for LLM Decoding and Alignment
Suho Shin
Chenghao Yang
Haifeng Xu
Mohammad T. Hajiaghayi
28
0
0
08 Jun 2025
EPiC: Towards Lossless Speedup for Reasoning Training through Edge-Preserving CoT Condensation
EPiC: Towards Lossless Speedup for Reasoning Training through Edge-Preserving CoT Condensation
Jinghan Jia
Hadi Reisizadeh
Chongyu Fan
Nathalie Baracaldo
Mingyi Hong
Sijia Liu
LRM
135
0
0
04 Jun 2025
When Models Reason in Your Language: Controlling Thinking Trace Language Comes at the Cost of Accuracy
When Models Reason in Your Language: Controlling Thinking Trace Language Comes at the Cost of Accuracy
Jirui Qi
Shan Chen
Zidi Xiong
Raquel Fernández
Danielle S. Bitterman
Arianna Bisazza
LRM
97
0
0
28 May 2025
To CoT or To Loop? A Formal Comparison Between Chain-of-Thought and Looped Transformers
To CoT or To Loop? A Formal Comparison Between Chain-of-Thought and Looped Transformers
Kevin Xu
Issei Sato
LRM
68
0
0
25 May 2025
Don't Look Only Once: Towards Multimodal Interactive Reasoning with Selective Visual Revisitation
Don't Look Only Once: Towards Multimodal Interactive Reasoning with Selective Visual Revisitation
Jiwan Chung
Junhyeok Kim
Siyeol Kim
Jaeyoung Lee
Min Soo Kim
Youngjae Yu
LRM
95
0
0
24 May 2025
First Finish Search: Efficient Test-Time Scaling in Large Language Models
Aradhye Agarwal
Ayan Sengupta
Tanmoy Chakraborty
ReLMRALMALMLRM
111
0
0
23 May 2025
The Unreasonable Effectiveness of Entropy Minimization in LLM Reasoning
The Unreasonable Effectiveness of Entropy Minimization in LLM Reasoning
Shivam Agarwal
Zimin Zhang
Lifan Yuan
Jiawei Han
Hao Peng
167
8
0
21 May 2025
A Mathematical Philosophy of Explanations in Mechanistic Interpretability -- The Strange Science Part I.i
A Mathematical Philosophy of Explanations in Mechanistic Interpretability -- The Strange Science Part I.i
Kola Ayonrinde
Louis Jaburi
MILM
175
1
0
01 May 2025
COSMOS: Predictable and Cost-Effective Adaptation of LLMs
COSMOS: Predictable and Cost-Effective Adaptation of LLMs
Jiayu Wang
Aws Albarghouthi
Frederic Sala
101
0
0
30 Apr 2025
TTRL: Test-Time Reinforcement Learning
TTRL: Test-Time Reinforcement Learning
Yuxin Zuo
Kaiyan Zhang
Li Sheng
Li Sheng
Xuekai Zhu
...
Youbang Sun
Zhiyuan Ma
Lifan Yuan
Ning Ding
Bowen Zhou
OffRL
421
31
0
22 Apr 2025
Large Language Models as Span Annotators
Large Language Models as Span Annotators
Zdeněk Kasner
Vilém Zouhar
Patrícia Schmidtová
Ivan Kartáč
Kristýna Onderková
Ondřej Plátek
Dimitra Gkatzia
Saad Mahamood
Ondrej Dusek
Simone Balloccu
ALM
124
0
0
11 Apr 2025
Inference-Time Computations for LLM Reasoning and Planning: A Benchmark and Insights
Inference-Time Computations for LLM Reasoning and Planning: A Benchmark and Insights
Shubham Parashar
Blake Olson
Sambhav Khurana
Eric Li
Hongyi Ling
James Caverlee
Shuiwang Ji
LRMReLM
153
13
0
18 Feb 2025
PEA: Enhancing LLM Performance on Computational-Reasoning Tasks
PEA: Enhancing LLM Performance on Computational-Reasoning Tasks
Zi Wang
Shiwei Weng
Mohannad J. Alhanahnah
S. Jha
Tom Reps
LRMReLM
109
0
0
16 Feb 2025
Examining False Positives under Inference Scaling for Mathematical Reasoning
Examining False Positives under Inference Scaling for Mathematical Reasoning
Yu Guang Wang
Nan Yang
Liang Wang
Furu Wei
LRM
144
4
0
10 Feb 2025
Optimizing Temperature for Language Models with Multi-Sample Inference
Optimizing Temperature for Language Models with Multi-Sample Inference
Weihua Du
Yiming Yang
Sean Welleck
166
4
0
07 Feb 2025
Inference Scaling fLaws: The Limits of LLM Resampling with Imperfect
  Verifiers
Inference Scaling fLaws: The Limits of LLM Resampling with Imperfect Verifiers
Benedikt Stroebl
Sayash Kapoor
Arvind Narayanan
LRM
149
18
0
26 Nov 2024
Think Carefully and Check Again! Meta-Generation Unlocking LLMs for Low-Resource Cross-Lingual Summarization
Think Carefully and Check Again! Meta-Generation Unlocking LLMs for Low-Resource Cross-Lingual Summarization
Zhecheng Li
Yijiao Wang
Bryan Hooi
Yujun Cai
Naifan Cheung
Nanyun Peng
Kai-Wei Chang
202
1
0
26 Oct 2024
Understanding Layer Significance in LLM Alignment
Understanding Layer Significance in LLM Alignment
Guangyuan Shi
Zexin Lu
Xiaoyu Dong
Wenlong Zhang
Xuanyu Zhang
Yujie Feng
Xiao-Ming Wu
147
3
0
23 Oct 2024
Better Instruction-Following Through Minimum Bayes Risk
Better Instruction-Following Through Minimum Bayes Risk
Ian Wu
Patrick Fernandes
Amanda Bertsch
Seungone Kim
Sina Pakazad
Graham Neubig
141
11
0
03 Oct 2024
QServe: W4A8KV4 Quantization and System Co-design for Efficient LLM Serving
QServe: W4A8KV4 Quantization and System Co-design for Efficient LLM Serving
Chengyue Wu
Haotian Tang
Shang Yang
Zhekai Zhang
Guangxuan Xiao
Chuang Gan
Song Han
172
98
0
07 May 2024
Theoretical guarantees on the best-of-n alignment policy
Theoretical guarantees on the best-of-n alignment policy
Ahmad Beirami
Alekh Agarwal
Jonathan Berant
Alex DÁmour
Jacob Eisenstein
Chirag Nagpal
A. Suresh
126
61
0
03 Jan 2024
Fast Transformer Decoding: One Write-Head is All You Need
Fast Transformer Decoding: One Write-Head is All You Need
Noam M. Shazeer
174
479
0
06 Nov 2019
Beam Search Strategies for Neural Machine Translation
Beam Search Strategies for Neural Machine Translation
Markus Freitag
Yaser Al-Onaizan
129
396
0
06 Feb 2017
1