ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2506.14899
5
0

Optimal Convergence Rates of Deep Neural Network Classifiers

17 June 2025
Zihan Zhang
Lei Shi
Ding-Xuan Zhou
ArXiv (abs)PDFHTML
Main:65 Pages
1 Figures
Bibliography:2 Pages
Abstract

In this paper, we study the binary classification problem on [0,1]d[0,1]^d[0,1]d under the Tsybakov noise condition (with exponent s∈[0,∞]s \in [0,\infty]s∈[0,∞]) and the compositional assumption. This assumption requires the conditional class probability function of the data distribution to be the composition of q+1q+1q+1 vector-valued multivariate functions, where each component function is either a maximum value function or a Hölder-β\betaβ smooth function that depends only on d∗d_*d∗​ of its input variables. Notably, d∗d_*d∗​ can be significantly smaller than the input dimension ddd. We prove that, under these conditions, the optimal convergence rate for the excess 0-1 risk of classifiers is \left( \frac{1}{n} \right)^{\frac{\beta\cdot(1\wedge\beta)^q}{{\frac{d_*}{s+1}+(1+\frac{1}{s+1})\cdot\beta\cdot(1\wedge\beta)^q}}}\;\;\;, which is independent of the input dimension ddd. Additionally, we demonstrate that ReLU deep neural networks (DNNs) trained with hinge loss can achieve this optimal convergence rate up to a logarithmic factor. This result provides theoretical justification for the excellent performance of ReLU DNNs in practical classification tasks, particularly in high-dimensional settings. The technique used to establish these results extends the oracle inequality presented in our previous work. The generalized approach is of independent interest.

View on arXiv
@article{zhang2025_2506.14899,
  title={ Optimal Convergence Rates of Deep Neural Network Classifiers },
  author={ Zihan Zhang and Lei Shi and Ding-Xuan Zhou },
  journal={arXiv preprint arXiv:2506.14899},
  year={ 2025 }
}
Comments on this paper