115
7

Finetuning Is a Surprisingly Effective Domain Adaptation Baseline in Handwriting Recognition

Main:15 Pages
11 Figures
Bibliography:3 Pages
3 Tables
Abstract

In many machine learning tasks, a large general dataset and a small specialized dataset are available. In such situations, various domain adaptation methods can be used to adapt a general model to the target dataset. We show that in the case of neural networks trained for handwriting recognition using CTC, simple finetuning with data augmentation works surprisingly well in such scenarios and that it is resistant to overfitting even for very small target domain datasets. We evaluated the behavior of finetuning with respect to augmentation, training data size, and quality of the pre-trained network, both in writer-dependent and writer-independent settings. On a large real-world dataset, finetuning provided an average relative CER improvement of 25 % with 16 text lines for new writers and 50 % for 256 text lines.

View on arXiv
@article{kohút2025_2302.06308,
  title={ Fine-tuning Is a Surprisingly Effective Domain Adaptation Baseline in Handwriting Recognition },
  author={ Jan Kohút and Michal Hradiš },
  journal={arXiv preprint arXiv:2302.06308},
  year={ 2025 }
}
Comments on this paper

We use cookies and other tracking technologies to improve your browsing experience on our website, to show you personalized content and targeted ads, to analyze our website traffic, and to understand where our visitors are coming from. See our policy.