Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
1505.04627
Cited By
Simple regret for infinitely many armed bandits
18 May 2015
Alexandra Carpentier
Michal Valko
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Simple regret for infinitely many armed bandits"
7 / 7 papers shown
Title
Fast and Robust: Task Sampling with Posterior and Diversity Synergies for Adaptive Decision-Makers in Randomized Environments
Yun Qu
Wenjie Wang
Yixiu Mao
Yiqin Lv
Xiangyang Ji
TTA
112
0
0
27 Apr 2025
Online Bandit Learning with Offline Preference Data for Improved RLHF
Akhil Agnihotri
Rahul Jain
Deepak Ramachandran
Zheng Wen
OffRL
104
2
0
13 Jun 2024
Online Stochastic Optimization under Correlated Bandit Feedback
M. G. Azar
A. Lazaric
Emma Brunskill
OffRL
46
54
0
04 Feb 2014
lil' UCB : An Optimal Exploration Algorithm for Multi-Armed Bandits
Kevin Jamieson
Matthew Malloy
Robert D. Nowak
Sébastien Bubeck
63
411
0
27 Dec 2013
Adaptive and minimax optimal estimation of the tail coefficient
Alexandra Carpentier
Arlene K. H. Kim
49
17
0
10 Sep 2013
Challenging the empirical mean and empirical variance: a deviation study
O. Catoni
107
462
0
10 Sep 2010
Multi-Armed Bandits in Metric Spaces
Robert D. Kleinberg
Aleksandrs Slivkins
E. Upfal
209
468
0
29 Sep 2008
1