ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2401.09902
23
7

Interplay between depth and width for interpolation in neural ODEs

18 January 2024
Antonio Álvarez-López
Arselane Hadj Slimane
Enrique Zuazua
ArXivPDFHTML
Abstract

Neural ordinary differential equations (neural ODEs) have emerged as a natural tool for supervised learning from a control perspective, yet a complete understanding of their optimal architecture remains elusive. In this work, we examine the interplay between their width ppp and number of layer transitions LLL (effectively the depth L+1L+1L+1). Specifically, we assess the model expressivity in terms of its capacity to interpolate either a finite dataset DDD comprising NNN pairs of points or two probability measures in Rd\mathbb{R}^dRd within a Wasserstein error margin ε>0\varepsilon>0ε>0. Our findings reveal a balancing trade-off between ppp and LLL, with LLL scaling as O(1+N/p)O(1+N/p)O(1+N/p) for dataset interpolation, and L=O(1+(pεd)−1)L=O\left(1+(p\varepsilon^d)^{-1}\right)L=O(1+(pεd)−1) for measure interpolation. In the autonomous case, where L=0L=0L=0, a separate study is required, which we undertake focusing on dataset interpolation. We address the relaxed problem of ε\varepsilonε-approximate controllability and establish an error decay of ε∼O(log⁡(p)p−1/d)\varepsilon\sim O(\log(p)p^{-1/d})ε∼O(log(p)p−1/d). This decay rate is a consequence of applying a universal approximation theorem to a custom-built Lipschitz vector field that interpolates DDD. In the high-dimensional setting, we further demonstrate that p=O(N)p=O(N)p=O(N) neurons are likely sufficient to achieve exact control.

View on arXiv
Comments on this paper