ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2506.15894
20
0

Language Models can perform Single-Utterance Self-Correction of Perturbed Reasoning

18 June 2025
Sam Silver
Jimin Sun
Ivan Zhang
Sara Hooker
Eddie Kim
    KELMReLMLRM
ArXiv (abs)PDFHTML
Main:5 Pages
23 Figures
Bibliography:3 Pages
3 Tables
Appendix:20 Pages
Abstract

Large Language Models (LLMs) have demonstrated impressive mathematical reasoning capabilities, yet their performance remains brittle to minor variations in problem description and prompting strategy. Furthermore, reasoning is vulnerable to sampling-induced errors which autoregressive models must primarily address using self-correction via additionally-generated tokens. To better understand self-correction capabilities of recent models, we conduct experiments measuring models' ability to self-correct synthetic perturbations introduced into their Chain of Thought (CoT) reasoning. We observe robust single-utterance intrinsic self-correction behavior across a range of open-weight models and datasets, ranging from subtle, implicit corrections to explicit acknowledgments and corrections of errors. Our findings suggest that LLMs, including those not finetuned for long CoT, may possess stronger intrinsic self-correction capabilities than commonly shown in the literature. The presence of this ability suggests that recent "reasoning" model work involves amplification of traits already meaningfully present in models.

View on arXiv
@article{silver2025_2506.15894,
  title={ Language Models can perform Single-Utterance Self-Correction of Perturbed Reasoning },
  author={ Sam Silver and Jimin Sun and Ivan Zhang and Sara Hooker and Eddie Kim },
  journal={arXiv preprint arXiv:2506.15894},
  year={ 2025 }
}
Comments on this paper