ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2310.02975
24
0

(ε,u)(ε, u)(ε,u)-Adaptive Regret Minimization in Heavy-Tailed Bandits

4 October 2023
Gianmarco Genalti
Lupo Marsigli
Nicola Gatti
Alberto Maria Metelli
ArXivPDFHTML
Abstract

Heavy-tailed distributions naturally arise in several settings, from finance to telecommunications. While regret minimization under subgaussian or bounded rewards has been widely studied, learning with heavy-tailed distributions only gained popularity over the last decade. In this paper, we consider the setting in which the reward distributions have finite absolute raw moments of maximum order 1+ϵ1+\epsilon1+ϵ, uniformly bounded by a constant u<+∞u<+\inftyu<+∞, for some ϵ∈(0,1]\epsilon \in (0,1]ϵ∈(0,1]. In this setting, we study the regret minimization problem when ϵ\epsilonϵ and uuu are unknown to the learner and it has to adapt. First, we show that adaptation comes at a cost and derive two negative results proving that the same regret guarantees of the non-adaptive case cannot be achieved with no further assumptions. Then, we devise and analyze a fully data-driven trimmed mean estimator and propose a novel adaptive regret minimization algorithm, AdaR-UCB, that leverages such an estimator. Finally, we show that AdaR-UCB is the first algorithm that, under a known distributional assumption, enjoys regret guarantees nearly matching those of the non-adaptive heavy-tailed case.

View on arXiv
Comments on this paper