ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2503.23832
40
0

An extrapolated and provably convergent algorithm for nonlinear matrix decomposition with the ReLU function

31 March 2025
Nicolas Gillis
M. Porcelli
Giovanni Seraghiti
ArXivPDFHTML
Abstract

Nonlinear matrix decomposition (NMD) with the ReLU function, denoted ReLU-NMD, is the following problem: given a sparse, nonnegative matrix XXX and a factorization rank rrr, identify a rank-rrr matrix Θ\ThetaΘ such that X≈max⁡(0,Θ)X\approx \max(0,\Theta)X≈max(0,Θ). This decomposition finds application in data compression, matrix completion with entries missing not at random, and manifold learning. The standard ReLU-NMD model minimizes the least squares error, that is, ∥X−max⁡(0,Θ)∥F2\|X - \max(0,\Theta)\|_F^2∥X−max(0,Θ)∥F2​. The corresponding optimization problem is nondifferentiable and highly nonconvex. This motivated Saul to propose an alternative model, Latent-ReLU-NMD, where a latent variable ZZZ is introduced and satisfies max⁡(0,Z)=X\max(0,Z)=Xmax(0,Z)=X while minimizing ∥Z−Θ∥F2\|Z - \Theta\|_F^2∥Z−Θ∥F2​ (``A nonlinear matrix decomposition for mining the zeros of sparse data'', SIAM J. Math. Data Sci., 2022). Our first contribution is to show that the two formulations may yield different low-rank solutions Θ\ThetaΘ; in particular, we show that Latent-ReLU-NMD can be ill-posed when ReLU-NMD is not, meaning that there are instances in which the infimum of Latent-ReLU-NMD is not attained while that of ReLU-NMD is. We also consider another alternative model, called 3B-ReLU-NMD, which parameterizes Θ=WH\Theta=WHΘ=WH, where WWW has rrr columns and HHH has rrr rows, allowing one to get rid of the rank constraint in Latent-ReLU-NMD. Our second contribution is to prove the convergence of a block coordinate descent (BCD) applied to 3B-ReLU-NMD and referred to as BCD-NMD. Our third contribution is a novel extrapolated variant of BCD-NMD, dubbed eBCD-NMD, which we prove is also convergent under mild assumptions. We illustrate the significant acceleration effect of eBCD-NMD compared to BCD-NMD, and also show that eBCD-NMD performs well against the state of the art on synthetic and real-world data sets.

View on arXiv
@article{gillis2025_2503.23832,
  title={ An extrapolated and provably convergent algorithm for nonlinear matrix decomposition with the ReLU function },
  author={ Nicolas Gillis and Margherita Porcelli and Giovanni Seraghiti },
  journal={arXiv preprint arXiv:2503.23832},
  year={ 2025 }
}
Comments on this paper