Gradual Fine-Tuning for Low-Resource Domain Adaptation
- CLL
Main:3 Pages
4 Figures
Bibliography:4 Pages
5 Tables
Appendix:1 Pages
Abstract
Fine-tuning is known to improve NLP models by adapting an initial model trained on more plentiful but less domain-salient examples to data in a target domain. Such domain adaptation is typically done using one stage of fine-tuning. We demonstrate that gradually fine-tuning in a multi-stage process can yield substantial further gains and can be applied without modifying the model or learning objective.
View on arXivComments on this paper
