Finite Continuum-Armed Bandits

We consider a situation where an agent has ressources to be allocated to a larger number of actions. Each action can be completed at most once and results in a stochastic reward with unknown mean. The goal of the agent is to maximize her cumulative reward. Non trivial strategies are possible when side information on the actions is available, for example in the form of covariates. Focusing on a nonparametric setting, where the mean reward is an unknown function of a one-dimensional covariate, we propose an optimal strategy for this problem. Under natural assumptions on the reward function, we prove that the optimal regret scales as up to poly-logarithmic factors when the budget is proportional to the number of actions . When becomes small compared to , a smooth transition occurs. When the ratio decreases from a constant to , the regret increases progressively up to the rate encountered in continuum-armed bandits.
View on arXiv