ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2505.15467
17
0

Joint Flashback Adaptation for Forgetting-Resistant Instruction Tuning

21 May 2025
Yukun Zhao
Lingyong Yan
Zhenyang Li
Shuaiqiang Wang
Zhumin Chen
Zhaochun Ren
Dawei Yin
    CLL
    KELM
    VLM
    LRM
ArXivPDFHTML
Abstract

Large language models have achieved remarkable success in various tasks. However, it is challenging for them to learn new tasks incrementally due to catastrophic forgetting. Existing approaches rely on experience replay, optimization constraints, or task differentiation, which encounter strict limitations in real-world scenarios. To address these issues, we propose Joint Flashback Adaptation. We first introduce flashbacks -- a limited number of prompts from old tasks -- when adapting to new tasks and constrain the deviations of the model outputs compared to the original one. We then interpolate latent tasks between flashbacks and new tasks to enable jointly learning relevant latent tasks, new tasks, and flashbacks, alleviating data sparsity in flashbacks and facilitating knowledge sharing for smooth adaptation. Our method requires only a limited number of flashbacks without access to the replay data and is task-agnostic. We conduct extensive experiments on state-of-the-art large language models across 1000+ instruction-following tasks, arithmetic reasoning tasks, and general reasoning tasks. The results demonstrate the superior performance of our method in improving generalization on new tasks and reducing forgetting in old tasks.

View on arXiv
@article{zhao2025_2505.15467,
  title={ Joint Flashback Adaptation for Forgetting-Resistant Instruction Tuning },
  author={ Yukun Zhao and Lingyong Yan and Zhenyang Li and Shuaiqiang Wang and Zhumin Chen and Zhaochun Ren and Dawei Yin },
  journal={arXiv preprint arXiv:2505.15467},
  year={ 2025 }
}
Comments on this paper