ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2206.00873
22
24

Nearly Optimal Best-of-Both-Worlds Algorithms for Online Learning with Feedback Graphs

2 June 2022
Shinji Ito
Taira Tsuchiya
Junya Honda
ArXivPDFHTML
Abstract

This study considers online learning with general directed feedback graphs. For this problem, we present best-of-both-worlds algorithms that achieve nearly tight regret bounds for adversarial environments as well as poly-logarithmic regret bounds for stochastic environments. As Alon et al. [2015] have shown, tight regret bounds depend on the structure of the feedback graph: strongly observable graphs yield minimax regret of Θ~(α1/2T1/2)\tilde{\Theta}( \alpha^{1/2} T^{1/2} )Θ~(α1/2T1/2), while weakly observable graphs induce minimax regret of Θ~(δ1/3T2/3)\tilde{\Theta}( \delta^{1/3} T^{2/3} )Θ~(δ1/3T2/3), where α\alphaα and δ\deltaδ, respectively, represent the independence number of the graph and the domination number of a certain portion of the graph. Our proposed algorithm for strongly observable graphs has a regret bound of O~(α1/2T1/2)\tilde{O}( \alpha^{1/2} T^{1/2} ) O~(α1/2T1/2) for adversarial environments, as well as of O(α(ln⁡T)3Δmin⁡) {O} ( \frac{\alpha (\ln T)^3 }{\Delta_{\min}} ) O(Δmin​α(lnT)3​) for stochastic environments, where Δmin⁡\Delta_{\min}Δmin​ expresses the minimum suboptimality gap. This result resolves an open question raised by Erez and Koren [2021]. We also provide an algorithm for weakly observable graphs that achieves a regret bound of O~(δ1/3T2/3)\tilde{O}( \delta^{1/3}T^{2/3} )O~(δ1/3T2/3) for adversarial environments and poly-logarithmic regret for stochastic environments. The proposed algorithms are based on the follow-the-regularized-leader approach combined with newly designed update rules for learning rates.

View on arXiv
Comments on this paper