ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2401.07874
22
0

Do stable neural networks exist for classification problems? -- A new view on stability in AI

15 January 2024
Z. N. D. Liu
A. C. Hansen
ArXivPDFHTML
Abstract

In deep learning (DL) the instability phenomenon is widespread and well documented, most commonly using the classical measure of stability, the Lipschitz constant. While a small Lipchitz constant is traditionally viewed as guarantying stability, it does not capture the instability phenomenon in DL for classification well. The reason is that a classification function -- which is the target function to be approximated -- is necessarily discontinuous, thus having an ínfinite' Lipchitz constant. As a result, the classical approach will deem every classification function unstable, yet basic classification functions a la ís there a cat in the image?' will typically be locally very 'flat' -- and thus locally stable -- except at the decision boundary. The lack of an appropriate measure of stability hinders a rigorous theory for stability in DL, and consequently, there are no proper approximation theoretic results that can guarantee the existence of stable networks for classification functions. In this paper we introduce a novel stability measure S(f)\mathscr{S}(f)S(f), for any classification function fff, appropriate to study the stability of discontinuous functions and their approximations. We further prove two approximation theorems: First, for any ϵ>0\epsilon > 0ϵ>0 and any classification function fff on a \emph{compact set}, there is a neural network (NN) ψ\psiψ, such that ψ−f≠0\psi - f \neq 0ψ−f=0 only on a set of measure <ϵ< \epsilon<ϵ, moreover, S(ψ)≥S(f)−ϵ\mathscr{S}(\psi) \geq \mathscr{S}(f) - \epsilonS(ψ)≥S(f)−ϵ (as accurate and stable as fff up to ϵ\epsilonϵ). Second, for any classification function fff and ϵ>0\epsilon > 0ϵ>0, there exists a NN ψ\psiψ such that ψ=f\psi = fψ=f on the set of points that are at least ϵ\epsilonϵ away from the decision boundary.

View on arXiv
Comments on this paper