ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2503.14495
58
0

Temporal Consistency for LLM Reasoning Process Error Identification

18 March 2025
Jiacheng Guo
Yue Wu
Jiahao Qiu
Kaixuan Huang
Xinzhe Juan
L. Yang
Mengdi Wang
    LRM
ArXivPDFHTML
Abstract

Verification is crucial for effective mathematical reasoning. We present a new temporal consistency method where verifiers iteratively refine their judgments based on the previous assessment. Unlike one-round verification or multi-model debate approaches, our method leverages consistency in a sequence of self-reflection actions to improve verification accuracy. Empirical evaluations across diverse mathematical process error identification benchmarks (Mathcheck, ProcessBench, and PRM800K) show consistent performance improvements over baseline methods. When applied to the recent DeepSeek R1 distilled models, our method demonstrates strong performance, enabling 7B/8B distilled models to outperform all 70B/72B models and GPT-4o on ProcessBench. Notably, the distilled 14B model with our method achieves performance comparable to Deepseek-R1. Our codes are available atthis https URL

View on arXiv
@article{guo2025_2503.14495,
  title={ Temporal Consistency for LLM Reasoning Process Error Identification },
  author={ Jiacheng Guo and Yue Wu and Jiahao Qiu and Kaixuan Huang and Xinzhe Juan and Ling Yang and Mengdi Wang },
  journal={arXiv preprint arXiv:2503.14495},
  year={ 2025 }
}
Comments on this paper