609

Revisiting the Moment Accountant Method for DP-SGD

International Conference on Machine Learning (ICML), 2021
Abstract

In order to provide differential privacy, Gaussian noise with standard deviation σ\sigma is added to local SGD updates after performing a clipping operation in Differential Private SGD (DP-SGD). By non-trivially improving the account method we prove a simple and easy to evaluate closed form (ϵ,δ)(\epsilon,\delta)-DP guarantee: DP-SGD is (ϵ,δ)(\epsilon,\delta)-DP if σ=2(ϵ+ln(1/δ))/ϵ\sigma=\sqrt{2(\epsilon +\ln(1/\delta))/\epsilon} and TT is at least 2k2/ϵ\approx 2k^2/\epsilon, where TT is the total number of rounds, and K=kNK=kN is the total number of gradient computations where kk measures KK in number of epochs of size NN of the local data set. We prove that our expression is close to tight in that if TT is more than a constant factor 8\approx 8 smaller than the lower bound 2k2/ϵ\approx 2k^2/\epsilon, then the (ϵ,δ)(\epsilon,\delta)-DP guarantee is violated. Choosing the smallest possible value T2k2/ϵT\approx 2k^2/\epsilon not only leads to a close to tight DP guarantee, but also minimizes the total number of communicated updates and this means that the least amount of noise is aggregated into the global model and accuracy is optimized as confirmed by simulations. In addition this minimizes round communication.

View on arXiv
Comments on this paper