ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2408.17017
22
7

Reasoning Aware Self-Consistency: Leveraging Reasoning Paths for Efficient LLM Sampling

30 August 2024
Guangya Wan
Yuqi Wu
Jie Chen
Sheng Li
    LRM
ArXivPDFHTML
Abstract

Self-Consistency mitigates hallucinations in Large Language Models (LLMs) by sampling multiple reasoning paths,but it lacks a systematic approach to determine the optimal number of samples or select the most faithful rationale. To address this limitation, we introduce Reasoning-Aware Self-Consistency (RASC), a novel framework that enhances sampling efficiency and reasoning faithfulness by dynamically evaluating both outputs and rationales. RASC assesses the quality of reasoning and the consistency of answers for each generated sample, using these assessments to guide early stopping decisions and rationale selection. The framework employs criteria-based stopping and weighted majority voting, enabling more informed choices on when to halt sampling and which rationale to select. Our comprehensive experiments across diverse question-answering datasets demonstrate that RASC outperforms existing methods, reducing sample usage by approximately 70% while maintaining accuracy. Moreover, RASC facilitates the selection of high-fidelity rationales, thereby improving the faithfulness of LLM outputs. Our approach effectively addresses the efficiency-accuracy trade-off in LLM reasoning tasks, offering a new perspective for more nuanced, faithful, and effective utilization of LLMs in resource-constrained environments.

View on arXiv
@article{wan2025_2408.17017,
  title={ Reasoning Aware Self-Consistency: Leveraging Reasoning Paths for Efficient LLM Sampling },
  author={ Guangya Wan and Yuqi Wu and Jie Chen and Sheng Li },
  journal={arXiv preprint arXiv:2408.17017},
  year={ 2025 }
}
Comments on this paper