30
0

Smooth Model Compression without Fine-Tuning

Main:9 Pages
9 Figures
Bibliography:4 Pages
3 Tables
Appendix:4 Pages
Abstract

Compressing and pruning large machine learning models has become a critical step towards their deployment in real-world applications. Standard pruning and compression techniques are typically designed without taking the structure of the network's weights into account, limiting their effectiveness. We explore the impact of smooth regularization on neural network training and model compression. By applying nuclear norm, first- and second-order derivative penalties of the weights during training, we encourage structured smoothness while preserving predictive performance on par with non-smooth models. We find that standard pruning methods often perform better when applied to these smooth models. Building on this observation, we apply a Singular-Value-Decomposition-based compression method that exploits the underlying smooth structure and approximates the model's weight tensors by smaller low-rank tensors. Our approach enables state-of-the-art compression without any fine-tuning - reaching up to 91%91\% accuracy on a smooth ResNet-18 on CIFAR-10 with 70%70\% fewer parameters.

View on arXiv
@article{runkel2025_2505.24469,
  title={ Smooth Model Compression without Fine-Tuning },
  author={ Christina Runkel and Natacha Kuete Meli and Jovita Lukasik and Ander Biguri and Carola-Bibiane Schönlieb and Michael Moeller },
  journal={arXiv preprint arXiv:2505.24469},
  year={ 2025 }
}
Comments on this paper