Identifiability of Complete Dictionary Learning

Sparse component analysis (SCA), also known as complete dictionary learning, is the following problem: Given an input matrix and an integer , find a dictionary with columns and a matrix with -sparse columns (that is, each column of has at most non-zero entries) such that . A key issue in SCA is identifiability, that is, characterizing the conditions under which and are essentially unique (that is, they are unique up to permutation and scaling of the columns of and rows of ). Although SCA has been vastly investigated in the last two decades, only a few works have tackled this issue in the deterministic scenario, and no work provides reasonable bounds in the minimum number of samples (that is, columns of ) that leads to identifiability. In this work, we provide new results in the deterministic scenario when the data has a low-rank structure, that is, when is (under)complete. While previous bounds feature a combinatorial term , we exhibit a sufficient condition involving samples that yields an essentially unique decomposition, as long as these data points are well spread among the subspaces spanned by columns of . We also exhibit a necessary lower bound on the number of samples that contradicts previous results in the literature when equals . Our bounds provide a drastic improvement compared to the state of the art, and imply for example that for a fixed proportion of zeros (constant and independent of , e.g., 10\% of zero entries in ), one only requires data points to guarantee identifiability.
View on arXiv