27
0
v1v2 (latest)

SensLI: Sensitivity-Based Layer Insertion for Neural Networks

Main:34 Pages
23 Figures
Bibliography:3 Pages
6 Tables
Appendix:1 Pages
Abstract

The training of neural networks requires tedious and often manual tuning of the network architecture. We propose a systematic approach to inserting new layers during the training process. Our method eliminates the need to choose a fixed network size before training, is numerically inexpensive to execute and applicable to various architectures including fully connected feedforward networks, ResNets and CNNs. Our technique borrows ideas from constrained optimization and is based on first-order sensitivity information of the loss function with respect to the virtual parameters that additional layers, if inserted, would offer. In numerical experiments, our proposed sensitivity-based layer insertion technique (SensLI) exhibits improved performance on training loss and test error, compared to training on a fixed architecture, and reduced computational effort in comparison to training the extended architecture from the beginning. Our code is available onthis https URL.

View on arXiv
@article{kreis2025_2311.15995,
  title={ SensLI: Sensitivity-Based Layer Insertion for Neural Networks },
  author={ Leonie Kreis and Evelyn Herberg and Frederik Köhne and Anton Schiela and Roland Herzog },
  journal={arXiv preprint arXiv:2311.15995},
  year={ 2025 }
}
Comments on this paper

We use cookies and other tracking technologies to improve your browsing experience on our website, to show you personalized content and targeted ads, to analyze our website traffic, and to understand where our visitors are coming from. See our policy.