Stochastic -Submodular Bandits with Full Bandit Feedback

In this paper, we present the first sublinear -regret bounds for online -submodular optimization problems with full-bandit feedback, where is a corresponding offline approximation ratio. Specifically, we propose online algorithms for multiple -submodular stochastic combinatorial multi-armed bandit problems, including (i) monotone functions and individual size constraints, (ii) monotone functions with matroid constraints, (iii) non-monotone functions with matroid constraints, (iv) non-monotone functions without constraints, and (v) monotone functions without constraints. We transform approximation algorithms for offline -submodular maximization problems into online algorithms through the offline-to-online framework proposed by Nie et al. (2023a). A key contribution of our work is analyzing the robustness of the offline algorithms.
View on arXiv