ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1906.08947
25
96

Randomized Exploration in Generalized Linear Bandits

21 June 2019
Branislav Kveton
Manzil Zaheer
Csaba Szepesvári
Lihong Li
Mohammad Ghavamzadeh
Craig Boutilier
ArXivPDFHTML
Abstract

We study two randomized algorithms for generalized linear bandits. The first, GLM-TSL, samples a generalized linear model (GLM) from the Laplace approximation to the posterior distribution. The second, GLM-FPL, fits a GLM to a randomly perturbed history of past rewards. We analyze both algorithms and derive O~(dnlog⁡K)\tilde{O}(d \sqrt{n \log K})O~(dnlogK​) upper bounds on their nnn-round regret, where ddd is the number of features and KKK is the number of arms. The former improves on prior work while the latter is the first for Gaussian noise perturbations in non-linear models. We empirically evaluate both GLM-TSL and GLM-FPL in logistic bandits, and apply GLM-FPL to neural network bandits. Our work showcases the role of randomization, beyond posterior sampling, in exploration.

View on arXiv
Comments on this paper