ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1505.04627
  4. Cited By
Simple regret for infinitely many armed bandits

Simple regret for infinitely many armed bandits

18 May 2015
Alexandra Carpentier
Michal Valko
ArXivPDFHTML

Papers citing "Simple regret for infinitely many armed bandits"

7 / 7 papers shown
Title
Fast and Robust: Task Sampling with Posterior and Diversity Synergies for Adaptive Decision-Makers in Randomized Environments
Fast and Robust: Task Sampling with Posterior and Diversity Synergies for Adaptive Decision-Makers in Randomized Environments
Yun Qu
Wenjie Wang
Yixiu Mao
Yiqin Lv
Xiangyang Ji
TTA
112
0
0
27 Apr 2025
Online Bandit Learning with Offline Preference Data for Improved RLHF
Online Bandit Learning with Offline Preference Data for Improved RLHF
Akhil Agnihotri
Rahul Jain
Deepak Ramachandran
Zheng Wen
OffRL
104
2
0
13 Jun 2024
Online Stochastic Optimization under Correlated Bandit Feedback
Online Stochastic Optimization under Correlated Bandit Feedback
M. G. Azar
A. Lazaric
Emma Brunskill
OffRL
46
54
0
04 Feb 2014
lil' UCB : An Optimal Exploration Algorithm for Multi-Armed Bandits
lil' UCB : An Optimal Exploration Algorithm for Multi-Armed Bandits
Kevin Jamieson
Matthew Malloy
Robert D. Nowak
Sébastien Bubeck
63
411
0
27 Dec 2013
Adaptive and minimax optimal estimation of the tail coefficient
Adaptive and minimax optimal estimation of the tail coefficient
Alexandra Carpentier
Arlene K. H. Kim
49
17
0
10 Sep 2013
Challenging the empirical mean and empirical variance: a deviation study
Challenging the empirical mean and empirical variance: a deviation study
O. Catoni
107
462
0
10 Sep 2010
Multi-Armed Bandits in Metric Spaces
Multi-Armed Bandits in Metric Spaces
Robert D. Kleinberg
Aleksandrs Slivkins
E. Upfal
209
468
0
29 Sep 2008
1