Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
1405.4758
Cited By
Lipschitz Bandits: Regret Lower Bounds and Optimal Algorithms
19 May 2014
Stefan Magureanu
Richard Combes
Alexandre Proutiere
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Lipschitz Bandits: Regret Lower Bounds and Optimal Algorithms"
27 / 27 papers shown
Title
Cost-Aware Optimal Pairwise Pure Exploration
Di Wu
Chengshuai Shi
Ruida Zhou
Cong Shen
41
0
0
10 Mar 2025
Bandit Optimal Transport
Lorenzo Croissant
84
0
0
11 Feb 2025
A Complete Characterization of Learnability for Stochastic Noisy Bandits
Steve Hanneke
Kun Wang
45
0
0
20 Jan 2025
Pretraining Decision Transformers with Reward Prediction for In-Context Multi-task Structured Bandit Learning
Subhojyoti Mukherjee
Josiah P. Hanna
Qiaomin Xie
Robert Nowak
89
2
0
07 Jun 2024
Batched Stochastic Bandit for Nondegenerate Functions
Yu Liu
Yunlu Shu
Tianyu Wang
52
0
0
09 May 2024
Causally Abstracted Multi-armed Bandits
Fabio Massimo Zennaro
Nicholas Bishop
Joel Dyer
Yorgos Felekis
Anisoara Calinescu
Michael Wooldridge
Theodoros Damoulas
38
2
0
26 Apr 2024
Robust Lipschitz Bandits to Adversarial Corruptions
Yue Kang
Cho-Jui Hsieh
T. C. Lee
AAML
34
8
0
29 May 2023
Tight Guarantees for Interactive Decision Making with the Decision-Estimation Coefficient
Dylan J. Foster
Noah Golowich
Yanjun Han
OffRL
33
29
0
19 Jan 2023
On Elimination Strategies for Bandit Fixed-Confidence Identification
Andrea Tirinzoni
Rémy Degenne
47
7
0
22 May 2022
Measurement-based Admission Control in Sliced Networks: A Best Arm Identification Approach
Simon Lindstaahl
Alexandre Proutiere
A. Johnsson
8
9
0
14 Apr 2022
Dealing With Misspecification In Fixed-Confidence Linear Top-m Identification
Clémence Réda
Andrea Tirinzoni
Rémy Degenne
33
9
0
02 Nov 2021
Max-Utility Based Arm Selection Strategy For Sequential Query Recommendations
S. P. Parambath
Christos Anagnostopoulos
R. Murray-Smith
Sean MacAvaney
E. Zervas
30
5
0
31 Aug 2021
Policy Optimization as Online Learning with Mediator Feedback
Alberto Maria Metelli
Matteo Papini
P. DÓro
Marcello Restelli
OffRL
27
10
0
15 Dec 2020
Dual-Mandate Patrols: Multi-Armed Bandits for Green Security
Lily Xu
Elizabeth Bondi-Kelly
Fei Fang
Andrew Perrault
Kai Wang
Milind Tambe
21
22
0
14 Sep 2020
Crush Optimism with Pessimism: Structured Bandits Beyond Asymptotic Optimality
Kwang-Sung Jun
Chicheng Zhang
31
10
0
15 Jun 2020
Multi-Armed Bandits with Correlated Arms
Samarth Gupta
Shreyas Chaudhari
Gauri Joshi
Osman Yağan
22
50
0
06 Nov 2019
Sequential Controlled Sensing for Composite Multihypothesis Testing
Aditya Deshmukh
S. Bhashyam
V. Veeravalli
14
13
0
24 Oct 2019
Mixture Martingales Revisited with Applications to Sequential Tests and Confidence Intervals
E. Kaufmann
Wouter M. Koolen
31
117
0
28 Nov 2018
Exploration in Structured Reinforcement Learning
Jungseul Ok
Alexandre Proutiere
Damianos Tranos
33
62
0
03 Jun 2018
Multi-Armed Bandits on Partially Revealed Unit Interval Graphs
Xiao Xu
Sattar Vakili
Qing Zhao
A. Swami
18
5
0
12 Feb 2018
Online Learning: A Comprehensive Survey
Guosheng Lin
Doyen Sahoo
Jing Lu
P. Zhao
OffRL
31
636
0
08 Feb 2018
Thresholding Bandit for Dose-ranging: The Impact of Monotonicity
Aurélien Garivier
Pierre Ménard
Laurent Rossi
Pierre Menard
22
27
0
13 Nov 2017
Minimal Exploration in Structured Stochastic Bandits
Richard Combes
Stefan Magureanu
Alexandre Proutiere
44
115
0
01 Nov 2017
Learning the distribution with largest mean: two bandit frameworks
E. Kaufmann
Aurélien Garivier
30
19
0
31 Jan 2017
Online Learning with Gaussian Payoffs and Side Observations
Yifan Wu
András Gyorgy
Csaba Szepesvári
15
44
0
27 Oct 2015
Unimodal Bandits without Smoothness
Richard Combes
Alexandre Proutiere
38
18
0
28 Jun 2014
Bandits and Experts in Metric Spaces
Robert D. Kleinberg
Aleksandrs Slivkins
E. Upfal
49
120
0
04 Dec 2013
1