We analyze the problem of private learning in generalized linear contextual bandits. Our approach is based on a novel method of re-weighted regression, yielding an efficient algorithm with regret of order and in the joint and local model of -privacy, respectively. Further, we provide near-optimal private procedures that achieve dimension-independent rates in private linear models and linear contextual bandits. In particular, our results imply that joint privacy is almost "for free" in all the settings we consider, partially addressing the open problem posed by Azize and Basu (2024).
View on arXiv@article{chen2025_2502.13115, title={ Near-Optimal Private Learning in Linear Contextual Bandits }, author={ Fan Chen and Jiachun Li and Alexander Rakhlin and David Simchi-Levi }, journal={arXiv preprint arXiv:2502.13115}, year={ 2025 } }