28
8

A surrogate loss function for optimization of FβF_β score in binary classification with imbalanced data

Abstract

The FβF_\beta score is a commonly used measure of classification performance, which plays crucial roles in classification tasks with imbalanced data sets. However, the FβF_\beta score cannot be used as a loss function by gradient-based learning algorithms for optimizing neural network parameters due to its non-differentiability. On the other hand, commonly used loss functions such as the binary cross-entropy (BCE) loss are not directly related to performance measures such as the FβF_\beta score, so that neural networks optimized by using the loss functions may not yield optimal performance measures. In this study, we investigate a relationship between classification performance measures and loss functions in terms of the gradients with respect to the model parameters. Then, we propose a differentiable surrogate loss function for the optimization of the FβF_\beta score. We show that the gradient paths of the proposed surrogate FβF_\beta loss function approximate the gradient paths of the large sample limit of the FβF_\beta score. Through numerical experiments using ResNets and benchmark image data sets, it is demonstrated that the proposed surrogate FβF_\beta loss function is effective for optimizing FβF_\beta scores under class imbalances in binary classification tasks compared with other loss functions.

View on arXiv
Comments on this paper

We use cookies and other tracking technologies to improve your browsing experience on our website, to show you personalized content and targeted ads, to analyze our website traffic, and to understand where our visitors are coming from. See our policy.