Soft Reasoning: Navigating Solution Spaces in Large Language Models through Controlled Embedding Exploration
- LRM

Main:8 Pages
9 Figures
Bibliography:4 Pages
15 Tables
Appendix:9 Pages
Abstract
Large Language Models (LLMs) struggle with complex reasoning due to limited diversity and inefficient search. We propose Soft Reasoning, an embedding-based search framework that optimises the embedding of the first token to guide generation. It combines (1) embedding perturbation for controlled exploration and (2) Bayesian optimisation to refine embeddings via a verifier-guided objective, balancing exploration and exploitation. This approach improves reasoning accuracy and coherence while avoiding reliance on heuristic search. Experiments demonstrate superior correctness with minimal computation, making it a scalable, model-agnostic solution.
View on arXiv@article{zhu2025_2505.24688, title={ Soft Reasoning: Navigating Solution Spaces in Large Language Models through Controlled Embedding Exploration }, author={ Qinglin Zhu and Runcong Zhao and Hanqi Yan and Yulan He and Yudong Chen and Lin Gui }, journal={arXiv preprint arXiv:2505.24688}, year={ 2025 } }
Comments on this paper