Soft Reasoning: Navigating Solution Spaces in Large Language Models through Controlled Embedding Exploration
8citations
8
Citations
#200
in ICML 2025
of 3340 papers
6
Authors
1
Data Points
Abstract
Large Language Models (LLMs) struggle with complex reasoning due to limited diversity and inefficient search. We propose Soft Reasoning, an embedding-based search framework that optimises the embedding of the first token to guide generation. It combines (1) embedding perturbation for controlled exploration and (2) Bayesian optimisation to refine embeddings via a verifier-guided objective, balancing exploration and exploitation. This approach improves reasoning accuracy and coherence while avoiding reliance on heuristic search. Experiments demonstrate superior correctness with minimal computation, making it a scalable, model-agnostic solution.
Citation History
Jan 28, 2026
8