SensLI: Sensitivity-Based Layer Insertion for Neural Networks

The training of neural networks requires tedious and often manual tuning of the network architecture. We propose a systematic approach to inserting new layers during the training process. Our method eliminates the need to choose a fixed network size before training, is numerically inexpensive to execute and applicable to various architectures including fully connected feedforward networks, ResNets and CNNs. Our technique borrows ideas from constrained optimization and is based on first-order sensitivity information of the loss function with respect to the virtual parameters that additional layers, if inserted, would offer. In numerical experiments, our proposed sensitivity-based layer insertion technique (SensLI) exhibits improved performance on training loss and test error, compared to training on a fixed architecture, and reduced computational effort in comparison to training the extended architecture from the beginning. Our code is available onthis https URL.
View on arXiv@article{kreis2025_2311.15995, title={ SensLI: Sensitivity-Based Layer Insertion for Neural Networks }, author={ Leonie Kreis and Evelyn Herberg and Frederik Köhne and Anton Schiela and Roland Herzog }, journal={arXiv preprint arXiv:2311.15995}, year={ 2025 } }