Guided Search Strategies in Non-Serializable Environments with Applications to Software Engineering Agents

Large language models (LLMs) have recently achieved remarkable results in complex multi-step tasks, such as mathematical reasoning and agentic software engineering. However, they often struggle to maintain consistent performance across multiple solution attempts. One effective approach to narrow the gap between average-case and best-case performance is guided test-time search, which explores multiple solution paths to identify the most promising one. Unfortunately, effective search techniques (e.g. MCTS) are often unsuitable for non-serializable RL environments, such as Docker containers, where intermediate environment states cannot be easily saved and restored. We investigate two complementary search strategies applicable to such environments: 1-step lookahead and trajectory selection, both guided by a learned action-value function estimator. On the SWE-bench Verified benchmark, a key testbed for agentic software engineering, we find these methods to double the average success rate of a fine-tuned Qwen-72B model, achieving 40.8%, the new state-of-the-art for open-weights models. Additionally, we show that these techniques are transferable to more advanced closed models, yielding similar improvements with GPT-4o.
View on arXiv@article{zainullina2025_2505.13652, title={ Guided Search Strategies in Non-Serializable Environments with Applications to Software Engineering Agents }, author={ Karina Zainullina and Alexander Golubev and Maria Trofimova and Sergei Polezhaev and Ibragim Badertdinov and Daria Litvintseva and Simon Karasik and Filipp Fisin and Sergei Skvortsov and Maksim Nekrashevich and Anton Shevtsov and Boris Yangel }, journal={arXiv preprint arXiv:2505.13652}, year={ 2025 } }