7
0

Revisiting Overthinking in Long Chain-of-Thought from the Perspective of Self-Doubt

Abstract

Reasoning Large Language Models (RLLMs) have demonstrated impressive performance on complex tasks, largely due to the adoption of Long Chain-of-Thought (Long CoT) reasoning. However, they often exhibit overthinking -- performing unnecessary reasoning steps even after arriving at the correct answer. Prior work has largely focused on qualitative analyses of overthinking through sample-based observations of long CoTs. In contrast, we present a quantitative analysis of overthinking from the perspective of self-doubt, characterized by excessive token usage devoted to re-verifying already-correct answer. We find that self-doubt significantly contributes to overthinking. In response, we introduce a simple and effective prompting method to reduce the model's over-reliance on input questions, thereby avoiding self-doubt. Specifically, we first prompt the model to question the validity of the input question, and then respond concisely based on the outcome of that evaluation. Experiments on three mathematical reasoning tasks and four datasets with missing premises demonstrate that our method substantially reduces answer length and yields significant improvements across nearly all datasets upon 4 widely-used RLLMs. Further analysis demonstrates that our method effectively minimizes the number of reasoning steps and reduces self-doubt.

View on arXiv
@article{peng2025_2505.23480,
  title={ Revisiting Overthinking in Long Chain-of-Thought from the Perspective of Self-Doubt },
  author={ Keqin Peng and Liang Ding and Yuanxin Ouyang and Meng Fang and Dacheng Tao },
  journal={arXiv preprint arXiv:2505.23480},
  year={ 2025 }
}
Comments on this paper