88
2

The Exploration of Neural Collapse under Imbalanced Data

Abstract

Neural collapse, a newly identified characteristic, describes a property of solutions during model training. In this paper, we explore neural collapse in the context of imbalanced data. We consider the LL-extended unconstrained feature model with a bias term and provide a theoretical analysis of global minimizer. Our findings include: (1) Features within the same class converge to their class mean, similar to both the balanced case and the imbalanced case without bias. (2) The geometric structure is mainly on the left orthonormal transformation of the product of LL linear classifiers and the right transformation of the class-mean matrix. (3) Some rows of the left orthonormal transformation of the product of LL linear classifiers collapse to zeros and others are orthogonal, which relies on the singular values of Y^=(IK1/Nn1K)D\hat Y=(I_K-1/N\mathbf{n}1^\top_K)D, where KK is class size, n\mathbf{n} is the vector of sample size for each class, DD is the diagonal matrix whose diagonal entries are given by n\sqrt{\mathbf{n}}. Similar results are for the columns of the right orthonormal transformation of the product of class-mean matrix and DD. (4) The ii-th row of the left orthonormal transformation of the product of LL linear classifiers aligns with the ii-th column of the right orthonormal transformation of the product of class-mean matrix and DD. (5) We provide the estimation of singular values about Y^\hat Y. Our numerical experiments support these theoretical findings.

View on arXiv
Comments on this paper

We use cookies and other tracking technologies to improve your browsing experience on our website, to show you personalized content and targeted ads, to analyze our website traffic, and to understand where our visitors are coming from. See our policy.