ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2205.13709
33
21

DP-PCA: Statistically Optimal and Differentially Private PCA

27 May 2022
Xiyang Liu
Weihao Kong
Prateek Jain
Sewoong Oh
ArXivPDFHTML
Abstract

We study the canonical statistical task of computing the principal component from nnn i.i.d.~data in ddd dimensions under (ε,δ)(\varepsilon,\delta)(ε,δ)-differential privacy. Although extensively studied in literature, existing solutions fall short on two key aspects: (iii) even for Gaussian data, existing private algorithms require the number of samples nnn to scale super-linearly with ddd, i.e., n=Ω(d3/2)n=\Omega(d^{3/2})n=Ω(d3/2), to obtain non-trivial results while non-private PCA requires only n=O(d)n=O(d)n=O(d), and (iiiiii) existing techniques suffer from a non-vanishing error even when the randomness in each data point is arbitrarily small. We propose DP-PCA, which is a single-pass algorithm that overcomes both limitations. It is based on a private minibatch gradient ascent method that relies on {\em private mean estimation}, which adds minimal noise required to ensure privacy by adapting to the variance of a given minibatch of gradients. For sub-Gaussian data, we provide nearly optimal statistical error rates even for n=O~(d)n=\tilde O(d)n=O~(d). Furthermore, we provide a lower bound showing that sub-Gaussian style assumption is necessary in obtaining the optimal error rate.

View on arXiv
Comments on this paper