627

On the Tightness of the Moment Accountant for DP-SGD

International Conference on Machine Learning (ICML), 2021
Abstract

In order to provide differential privacy, Gaussian noise with standard deviation σ\sigma is added to local SGD updates after performing a clipping operation in Differential Private SGD (DP-SGD). By non-trivially improving the moment account method we prove a closed form (ϵ,δ)(\epsilon,\delta)-DP guarantee: DP-SGD is (ϵ1/2,δ=1/N)(\epsilon\leq 1/2,\delta=1/N)-DP if σ=2(ϵ+ln(1/δ))/ϵ\sigma=\sqrt{2(\epsilon +\ln(1/\delta))/\epsilon} with TT at least 2k2/ϵ\approx 2k^2/\epsilon and (2/e)2k21/2ln(N)(2/e)^2k^2-1/2\geq \ln(N), where TT is the total number of rounds, and K=kNK=kN is the total number of gradient computations where kk measures KK in number of epochs of size NN of the local data set. We prove that our expression is close to tight in that if TT is more than a constant factor 8\approx 8 smaller than the lower bound 2k2/ϵ\approx 2k^2/\epsilon, then the (ϵ,δ)(\epsilon,\delta)-DP guarantee is violated. Choosing the smallest possible value T2k2/ϵT\approx 2k^2/\epsilon not only leads to a close to tight DP guarantee, but also minimizes the total number of communicated updates and this means that the least amount of noise is aggregated into the global model and in addition accuracy is optimized as confirmed by simulations.

View on arXiv
Comments on this paper