SGD as Free Energy Minimization: A Thermodynamic View on Neural Network Training

We present a thermodynamic interpretation of the stationary behavior of stochastic gradient descent (SGD) under fixed learning rates (LRs) in neural network training. We show that SGD implicitly minimizes a free energy function , balancing training loss and the entropy of the weights distribution , with temperature determined by the LR. This perspective offers a new lens on why high LRs prevent training from converging to the loss minima and how different LRs lead to stabilization at different loss levels. We empirically validate the free energy framework on both underparameterized (UP) and overparameterized (OP) models. UP models consistently follow free energy minimization, with temperature increasing monotonically with LR, while for OP models, the temperature effectively drops to zero at low LRs, causing SGD to minimize the loss directly and converge to an optimum. We attribute this mismatch to differences in the signal-to-noise ratio of stochastic gradients near optima, supported by both a toy example and neural network experiments.
View on arXiv@article{sadrtdinov2025_2505.23489, title={ SGD as Free Energy Minimization: A Thermodynamic View on Neural Network Training }, author={ Ildus Sadrtdinov and Ivan Klimov and Ekaterina Lobacheva and Dmitry Vetrov }, journal={arXiv preprint arXiv:2505.23489}, year={ 2025 } }