ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2407.02601
25
0

Linear Submodular Maximization with Bandit Feedback

2 July 2024
Wenjing Chen
Victoria G. Crawford
ArXiv (abs)PDFHTML
Abstract

Submodular optimization with bandit feedback has recently been studied in a variety of contexts. In a number of real-world applications such as diversified recommender systems and data summarization, the submodular function exhibits additional linear structure. We consider developing approximation algorithms for the maximization of a submodular objective function f:2U→R≥0f:2^U\to\mathbb{R}_{\geq 0}f:2U→R≥0​, where f=∑i=1dwiFif=\sum_{i=1}^dw_iF_{i}f=∑i=1d​wi​Fi​. It is assumed that we have value oracle access to the functions FiF_iFi​, but the coefficients wiw_iwi​ are unknown, and fff can only be accessed via noisy queries. We develop algorithms for this setting inspired by adaptive allocation algorithms in the best-arm identification for linear bandit, with approximation guarantees arbitrarily close to the setting where we have value oracle access to fff. Finally, we empirically demonstrate that our algorithms make vast improvements in terms of sample efficiency compared to algorithms that do not exploit the linear structure of fff on instances of move recommendation.

View on arXiv
Comments on this paper