ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2205.03014
20
16

Differentially Private Generalized Linear Models Revisited

6 May 2022
R. Arora
Raef Bassily
Cristóbal Guzmán
Michael Menart
Enayat Ullah
    FedML
ArXivPDFHTML
Abstract

We study the problem of (ϵ,δ)(\epsilon,\delta)(ϵ,δ)-differentially private learning of linear predictors with convex losses. We provide results for two subclasses of loss functions. The first case is when the loss is smooth and non-negative but not necessarily Lipschitz (such as the squared loss). For this case, we establish an upper bound on the excess population risk of O~(∥w∗∥n+min⁡{∥w∗∥2(nϵ)2/3,d∥w∗∥2nϵ})\tilde{O}\left(\frac{\Vert w^*\Vert}{\sqrt{n}} + \min\left\{\frac{\Vert w^* \Vert^2}{(n\epsilon)^{2/3}},\frac{\sqrt{d}\Vert w^*\Vert^2}{n\epsilon}\right\}\right)O~(n​∥w∗∥​+min{(nϵ)2/3∥w∗∥2​,nϵd​∥w∗∥2​}), where nnn is the number of samples, ddd is the dimension of the problem, and w∗w^*w∗ is the minimizer of the population risk. Apart from the dependence on ∥w∗∥\Vert w^\ast\Vert∥w∗∥, our bound is essentially tight in all parameters. In particular, we show a lower bound of Ω~(1n+min⁡{∥w∗∥4/3(nϵ)2/3,d∥w∗∥nϵ})\tilde{\Omega}\left(\frac{1}{\sqrt{n}} + {\min\left\{\frac{\Vert w^*\Vert^{4/3}}{(n\epsilon)^{2/3}}, \frac{\sqrt{d}\Vert w^*\Vert}{n\epsilon}\right\}}\right)Ω~(n​1​+min{(nϵ)2/3∥w∗∥4/3​,nϵd​∥w∗∥​}). We also revisit the previously studied case of Lipschitz losses [SSTT20]. For this case, we close the gap in the existing work and show that the optimal rate is (up to log factors) Θ(∥w∗∥n+min⁡{∥w∗∥nϵ,rank∥w∗∥nϵ})\Theta\left(\frac{\Vert w^*\Vert}{\sqrt{n}} + \min\left\{\frac{\Vert w^*\Vert}{\sqrt{n\epsilon}},\frac{\sqrt{\text{rank}}\Vert w^*\Vert}{n\epsilon}\right\}\right)Θ(n​∥w∗∥​+min{nϵ​∥w∗∥​,nϵrank​∥w∗∥​}), where rank\text{rank}rank is the rank of the design matrix. This improves over existing work in the high privacy regime. Finally, our algorithms involve a private model selection approach that we develop to enable attaining the stated rates without a-priori knowledge of ∥w∗∥\Vert w^*\Vert∥w∗∥.

View on arXiv
Comments on this paper