ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2306.05275
23
11

Federated Linear Contextual Bandits with User-level Differential Privacy

8 June 2023
Ruiquan Huang
Huanyu Zhang
Luca Melis
Milan Shen
Meisam Hajzinia
J. Yang
    FedML
ArXivPDFHTML
Abstract

This paper studies federated linear contextual bandits under the notion of user-level differential privacy (DP). We first introduce a unified federated bandits framework that can accommodate various definitions of DP in the sequential decision-making setting. We then formally introduce user-level central DP (CDP) and local DP (LDP) in the federated bandits framework, and investigate the fundamental trade-offs between the learning regrets and the corresponding DP guarantees in a federated linear contextual bandits model. For CDP, we propose a federated algorithm termed as ROBIN\texttt{ROBIN}ROBIN and show that it is near-optimal in terms of the number of clients MMM and the privacy budget ε\varepsilonε by deriving nearly-matching upper and lower regret bounds when user-level DP is satisfied. For LDP, we obtain several lower bounds, indicating that learning under user-level (ε,δ)(\varepsilon,\delta)(ε,δ)-LDP must suffer a regret blow-up factor at least min⁡{1/ε,M}\min\{1/\varepsilon,M\}min{1/ε,M} or min⁡{1/ε,M}\min\{1/\sqrt{\varepsilon},\sqrt{M}\}min{1/ε​,M​} under different conditions.

View on arXiv
Comments on this paper