84
0
v1v2 (latest)

Improved Supervised Fine-Tuning for Large Language Models to Mitigate Catastrophic Forgetting

Main:5 Pages
1 Figures
Bibliography:2 Pages
3 Tables
Appendix:1 Pages
Abstract

Supervised Fine-Tuning (SFT) is a critical step for enhancing the instruction-following capabilities of Large Language Models (LLMs) and adapting them to specialized domains. However, SFT often leads to a degradation of the model's general abilities, a phenomenon known as catastrophic forgetting. This problem is exacerbated when third-party practitioners fine-tune open-source models, as the original SFT data is typically not available. To address this challenge, we propose a novel and cost-effective SFT method that effectively mitigates catastrophic forgetting without requiring access to the original SFT data. Our approach first reconstructs the likely instruction distribution of the base model. It then employs a multi-model generation and filtering pipeline to synthesize a high-quality general-purpose dataset. This synthetic dataset is mixed with new, domain-specific data for fine-tuning. Experimental results show that our method not only preserves the model's capabilities in general domains but also improves task-specific performance, outperforming baselines that use publicly available SFT datasets.

View on arXiv
@article{ding2025_2506.09428,
  title={ Improved Supervised Fine-Tuning for Large Language Models to Mitigate Catastrophic Forgetting },
  author={ Fei Ding and Baiqiao Wang },
  journal={arXiv preprint arXiv:2506.09428},
  year={ 2025 }
}
Comments on this paper