ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2411.11932
88
1
v1v2 (latest)

Reviving Dormant Memories: Investigating Catastrophic Forgetting in Language Models through Rationale-Guidance Difficulty

18 November 2024
Huashan Sun
Yizhe Yang
    CLL
ArXiv (abs)PDFHTML
Main:9 Pages
12 Figures
Bibliography:3 Pages
10 Tables
Appendix:5 Pages
Abstract

Although substantial efforts have been made to mitigate catastrophic forgetting in continual learning, the intrinsic mechanisms are not well understood. In this paper, we discover that when a forgetting model passively receives an externally provided partial appropriate rationale, its performance on the forgotten task can be restored. Furthermore, by simply adding a task-agnostic prefix to the original instruction, the forgetting model can actively generate an appropriate rationale to reach the correct answer. These findings suggest that the model does not actually ``forget'' the task knowledge; instead, the degraded performance can be attributed to the failure of the original instructions in guiding the model to generate the appropriate rationales. Based on this insight, we propose the Rationale-Guidance Difficulty metric to evaluate how effectively a given instruction guides the model in generating appropriate rationales. We apply this metric to optimize the allocation of replay data in replay-based continual learning algorithm. Experimental results demonstrate that our data allocation method effectively mitigates catastrophic forgetting and maintains better model plasticity simultaneously across models.

View on arXiv
@article{sun2025_2411.11932,
  title={ Unveiling and Addressing Pseudo Forgetting in Large Language Models },
  author={ Huashan Sun and Yizhe Yang and Yinghao Li and Jiawei Li and Yang Gao },
  journal={arXiv preprint arXiv:2411.11932},
  year={ 2025 }
}
Comments on this paper