ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2106.07836
19
4

Improved Regret Bounds for Online Submodular Maximization

15 June 2021
Omid Sadeghi
P. Raut
Maryam Fazel
ArXivPDFHTML
Abstract

In this paper, we consider an online optimization problem over TTT rounds where at each step t∈[T]t\in[T]t∈[T], the algorithm chooses an action xtx_txt​ from the fixed convex and compact domain set K\mathcal{K}K. A utility function ft(⋅)f_t(\cdot)ft​(⋅) is then revealed and the algorithm receives the payoff ft(xt)f_t(x_t)ft​(xt​). This problem has been previously studied under the assumption that the utilities are adversarially chosen monotone DR-submodular functions and O(T)\mathcal{O}(\sqrt{T})O(T​) regret bounds have been derived. We first characterize the class of strongly DR-submodular functions and then, we derive regret bounds for the following new online settings: (1)(1)(1) {ft}t=1T\{f_t\}_{t=1}^T{ft​}t=1T​ are monotone strongly DR-submodular and chosen adversarially, (2)(2)(2) {ft}t=1T\{f_t\}_{t=1}^T{ft​}t=1T​ are monotone submodular (while the average 1T∑t=1Tft\frac{1}{T}\sum_{t=1}^T f_tT1​∑t=1T​ft​ is strongly DR-submodular) and chosen by an adversary but they arrive in a uniformly random order, (3)(3)(3) {ft}t=1T\{f_t\}_{t=1}^T{ft​}t=1T​ are drawn i.i.d. from some unknown distribution ft∼Df_t\sim \mathcal{D}ft​∼D where the expected function f(⋅)=Eft∼D[ft(⋅)]f(\cdot)=\mathbb{E}_{f_t\sim\mathcal{D}}[f_t(\cdot)]f(⋅)=Eft​∼D​[ft​(⋅)] is monotone DR-submodular. For (1)(1)(1), we obtain the first logarithmic regret bounds. In terms of the second framework, we show that it is possible to obtain similar logarithmic bounds with high probability. Finally, for the i.i.d. model, we provide algorithms with O~(T)\tilde{\mathcal{O}}(\sqrt{T})O~(T​) stochastic regret bound, both in expectation and with high probability. Experimental results demonstrate that our algorithms outperform the previous techniques in the aforementioned three settings.

View on arXiv
Comments on this paper