ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1411.5417
22
56

Private Empirical Risk Minimization Beyond the Worst Case: The Effect of the Constraint Set Geometry

20 November 2014
Kunal Talwar
Abhradeep Thakurta
Li Zhang
ArXivPDFHTML
Abstract

Empirical Risk Minimization (ERM) is a standard technique in machine learning, where a model is selected by minimizing a loss function over constraint set. When the training dataset consists of private information, it is natural to use a differentially private ERM algorithm, and this problem has been the subject of a long line of work started with Chaudhuri and Monteleoni 2008. A private ERM algorithm outputs an approximate minimizer of the loss function and its error can be measured as the difference from the optimal value of the loss function. When the constraint set is arbitrary, the required error bounds are fairly well understood \cite{BassilyST14}. In this work, we show that the geometric properties of the constraint set can be used to derive significantly better results. Specifically, we show that a differentially private version of Mirror Descent leads to error bounds of the form O~(GC/n)\tilde{O}(G_{\mathcal{C}}/n)O~(GC​/n) for a lipschitz loss function, improving on the O~(p/n)\tilde{O}(\sqrt{p}/n)O~(p​/n) bounds in Bassily, Smith and Thakurta 2014. Here ppp is the dimensionality of the problem, nnn is the number of data points in the training set, and GCG_{\mathcal{C}}GC​ denotes the Gaussian width of the constraint set that we optimize over. We show similar improvements for strongly convex functions, and for smooth functions. In addition, we show that when the loss function is Lipschitz with respect to the ℓ1\ell_1ℓ1​ norm and C\mathcal{C}C is ℓ1\ell_1ℓ1​-bounded, a differentially private version of the Frank-Wolfe algorithm gives error bounds of the form O~(n−2/3)\tilde{O}(n^{-2/3})O~(n−2/3). This captures the important and common case of sparse linear regression (LASSO), when the data xix_ixi​ satisfies ∣xi∣∞≤1|x_i|_{\infty} \leq 1∣xi​∣∞​≤1 and we optimize over the ℓ1\ell_1ℓ1​ ball. We show new lower bounds for this setting, that together with known bounds, imply that all our upper bounds are tight.

View on arXiv
Comments on this paper