ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1701.07953
14
94

The Price of Differential Privacy For Online Learning

27 January 2017
Naman Agarwal
Karan Singh
    FedML
ArXivPDFHTML
Abstract

We design differentially private algorithms for the problem of online linear optimization in the full information and bandit settings with optimal O~(T)\tilde{O}(\sqrt{T})O~(T​) regret bounds. In the full-information setting, our results demonstrate that ϵ\epsilonϵ-differential privacy may be ensured for free -- in particular, the regret bounds scale as O(T)+O~(1ϵ)O(\sqrt{T})+\tilde{O}\left(\frac{1}{\epsilon}\right)O(T​)+O~(ϵ1​). For bandit linear optimization, and as a special case, for non-stochastic multi-armed bandits, the proposed algorithm achieves a regret of O~(1ϵT)\tilde{O}\left(\frac{1}{\epsilon}\sqrt{T}\right)O~(ϵ1​T​), while the previously known best regret bound was O~(1ϵT23)\tilde{O}\left(\frac{1}{\epsilon}T^{\frac{2}{3}}\right)O~(ϵ1​T32​).

View on arXiv
Comments on this paper