ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2503.13208
53
0

Improving Complex Reasoning with Dynamic Prompt Corruption: A soft prompt Optimization Approach

17 March 2025
Sinan Fan
Liang Xie
Chen Shen
Ge Teng
Xiaosong Yuan
Xiaofeng Zhang
Chenxi Huang
Wenxiao Wang
Xiaofei He
Jieping Ye
    AAML
    LRM
ArXivPDFHTML
Abstract

Prompt-tuning (PT) for large language models (LLMs) can facilitate the performance on various conventional NLP tasks with significantly fewer trainable parameters. However, our investigation reveals that PT provides limited improvement and may even degrade the primitive performance of LLMs on complex reasoning tasks. Such a phenomenon suggests that soft prompts can positively impact certain instances while negatively affecting others, particularly during the later phases of reasoning. To address these challenges, We first identify an information accumulation within the soft prompts. Through detailed analysis, we demonstrate that this phenomenon is often accompanied by erroneous information flow patterns in the deeper layers of the model, which ultimately lead to incorrect reasoning outcomes. we propose a novel method called Dynamic Prompt Corruption (DPC) to take better advantage of soft prompts in complex reasoning tasks, which dynamically adjusts the influence of soft prompts based on their impact on the reasoning process. Specifically, DPC consists of two stages: Dynamic Trigger and Dynamic Corruption. First, Dynamic Trigger measures the impact of soft prompts, identifying whether beneficial or detrimental. Then, Dynamic Corruption mitigates the negative effects of soft prompts by selectively masking key tokens that interfere with the reasoning process. We validate the proposed approach through extensive experiments on various LLMs and reasoning tasks, including GSM8K, MATH, and AQuA. Experimental results demonstrate that DPC can consistently enhance the performance of PT, achieving 4%-8% accuracy gains compared to vanilla prompt tuning, highlighting the effectiveness of our approach and its potential to enhance complex reasoning in LLMs.

View on arXiv
@article{fan2025_2503.13208,
  title={ Improving Complex Reasoning with Dynamic Prompt Corruption: A soft prompt Optimization Approach },
  author={ Sinan Fan and Liang Xie and Chen Shen and Ge Teng and Xiaosong Yuan and Xiaofeng Zhang and Chenxi Huang and Wenxiao Wang and Xiaofei He and Jieping Ye },
  journal={arXiv preprint arXiv:2503.13208},
  year={ 2025 }
}
Comments on this paper