44
v1v2 (latest)

Search More, Think Less: Rethinking Long-Horizon Agentic Search for Efficiency and Generalization

Qianben Chen
Tianrui Qin
King Zhu
Qiexiang Wang
Chengjun Yu
Shu Xu
Jiaqi Wu
Jiayu Zhang
Xinpeng Liu
Xin Gui
Jingyi Cao
Piaohong Wang
Dingfeng Shi
He Zhu
Tiannan Wang
Yuqing Wang
Maojia Song
Tianyu Zheng
Ge Zhang
Jian Yang
Jiaheng Liu
Minghao Liu
Yuchen Eleanor Jiang
Wangchunshu Zhou
Main:13 Pages
5 Figures
Bibliography:3 Pages
2 Tables
Appendix:14 Pages
Abstract

Recent deep research agents primarily improve performance by scaling reasoning depth, but this leads to high inference cost and latency in search-intensive scenarios. Moreover, generalization across heterogeneous research settings remains challenging. In this work, we propose \emph{Search More, Think Less} (SMTL), a framework for long-horizon agentic search that targets both efficiency and generalization. SMTL replaces sequential reasoning with parallel evidence acquisition, enabling efficient context management under constrained context budgets. To support generalization across task types, we further introduce a unified data synthesis pipeline that constructs search tasks spanning both deterministic question answering and open-ended research scenarios with task appropriate evaluation metrics. We train an end-to-end agent using supervised fine-tuning and reinforcement learning, achieving strong and often state of the art performance across benchmarks including BrowseComp (48.6\%), GAIA (75.7\%), Xbench (82.0\%), and DeepResearch Bench (45.9\%). Compared to Mirothinker-v1.0, SMTL with maximum 100 interaction steps reduces the average number of reasoning steps on BrowseComp by 70.7\%, while improving accuracy.

View on arXiv
Comments on this paper