29
0

Mitigating covariate shift in non-colocated data with learned parameter priors

Abstract

When training data are distributed across{ time or space,} covariate shift across fragments of training data biases cross-validation, compromising model selection and assessment. We present \textit{Fragmentation-Induced covariate-shift Remediation} (FIcsRFIcsR), which minimizes an ff-divergence between a fragment's covariate distribution and that of the standard cross-validation baseline. We s{how} an equivalence with popular importance-weighting methods. {The method}'s numerical solution poses a computational challenge owing to the overparametrized nature of a neural network, and we derive a Fisher Information approximation. When accumulated over fragments, this provides a global estimate of the amount of shift remediation thus far needed, and we incorporate that as a prior via the minimization objective. In the paper, we run extensive classification experiments on multiple data classes, over 4040 datasets, and with data batched over multiple sequence lengths. We extend the study to the kk-fold cross-validation setting through a similar set of experiments. An ablation study exposes the method to varying amounts of shift and demonstrates slower degradation with FIcsRFIcsR in place. The results are promising under all these conditions; with improved accuracy against batch and fold state-of-the-art by more than 5%5\% and 10%10\%, respectively.

View on arXiv
Comments on this paper