ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1405.4758
  4. Cited By
Lipschitz Bandits: Regret Lower Bounds and Optimal Algorithms

Lipschitz Bandits: Regret Lower Bounds and Optimal Algorithms

19 May 2014
Stefan Magureanu
Richard Combes
Alexandre Proutiere
ArXivPDFHTML

Papers citing "Lipschitz Bandits: Regret Lower Bounds and Optimal Algorithms"

27 / 27 papers shown
Title
Cost-Aware Optimal Pairwise Pure Exploration
Di Wu
Chengshuai Shi
Ruida Zhou
Cong Shen
41
0
0
10 Mar 2025
Bandit Optimal Transport
Bandit Optimal Transport
Lorenzo Croissant
84
0
0
11 Feb 2025
A Complete Characterization of Learnability for Stochastic Noisy Bandits
A Complete Characterization of Learnability for Stochastic Noisy Bandits
Steve Hanneke
Kun Wang
45
0
0
20 Jan 2025
Pretraining Decision Transformers with Reward Prediction for In-Context Multi-task Structured Bandit Learning
Pretraining Decision Transformers with Reward Prediction for In-Context Multi-task Structured Bandit Learning
Subhojyoti Mukherjee
Josiah P. Hanna
Qiaomin Xie
Robert Nowak
89
2
0
07 Jun 2024
Batched Stochastic Bandit for Nondegenerate Functions
Batched Stochastic Bandit for Nondegenerate Functions
Yu Liu
Yunlu Shu
Tianyu Wang
52
0
0
09 May 2024
Causally Abstracted Multi-armed Bandits
Causally Abstracted Multi-armed Bandits
Fabio Massimo Zennaro
Nicholas Bishop
Joel Dyer
Yorgos Felekis
Anisoara Calinescu
Michael Wooldridge
Theodoros Damoulas
38
2
0
26 Apr 2024
Robust Lipschitz Bandits to Adversarial Corruptions
Robust Lipschitz Bandits to Adversarial Corruptions
Yue Kang
Cho-Jui Hsieh
T. C. Lee
AAML
34
8
0
29 May 2023
Tight Guarantees for Interactive Decision Making with the
  Decision-Estimation Coefficient
Tight Guarantees for Interactive Decision Making with the Decision-Estimation Coefficient
Dylan J. Foster
Noah Golowich
Yanjun Han
OffRL
33
29
0
19 Jan 2023
On Elimination Strategies for Bandit Fixed-Confidence Identification
On Elimination Strategies for Bandit Fixed-Confidence Identification
Andrea Tirinzoni
Rémy Degenne
47
7
0
22 May 2022
Measurement-based Admission Control in Sliced Networks: A Best Arm
  Identification Approach
Measurement-based Admission Control in Sliced Networks: A Best Arm Identification Approach
Simon Lindstaahl
Alexandre Proutiere
A. Johnsson
8
9
0
14 Apr 2022
Dealing With Misspecification In Fixed-Confidence Linear Top-m
  Identification
Dealing With Misspecification In Fixed-Confidence Linear Top-m Identification
Clémence Réda
Andrea Tirinzoni
Rémy Degenne
33
9
0
02 Nov 2021
Max-Utility Based Arm Selection Strategy For Sequential Query
  Recommendations
Max-Utility Based Arm Selection Strategy For Sequential Query Recommendations
S. P. Parambath
Christos Anagnostopoulos
R. Murray-Smith
Sean MacAvaney
E. Zervas
30
5
0
31 Aug 2021
Policy Optimization as Online Learning with Mediator Feedback
Policy Optimization as Online Learning with Mediator Feedback
Alberto Maria Metelli
Matteo Papini
P. DÓro
Marcello Restelli
OffRL
27
10
0
15 Dec 2020
Dual-Mandate Patrols: Multi-Armed Bandits for Green Security
Dual-Mandate Patrols: Multi-Armed Bandits for Green Security
Lily Xu
Elizabeth Bondi-Kelly
Fei Fang
Andrew Perrault
Kai Wang
Milind Tambe
21
22
0
14 Sep 2020
Crush Optimism with Pessimism: Structured Bandits Beyond Asymptotic
  Optimality
Crush Optimism with Pessimism: Structured Bandits Beyond Asymptotic Optimality
Kwang-Sung Jun
Chicheng Zhang
31
10
0
15 Jun 2020
Multi-Armed Bandits with Correlated Arms
Multi-Armed Bandits with Correlated Arms
Samarth Gupta
Shreyas Chaudhari
Gauri Joshi
Osman Yağan
22
50
0
06 Nov 2019
Sequential Controlled Sensing for Composite Multihypothesis Testing
Sequential Controlled Sensing for Composite Multihypothesis Testing
Aditya Deshmukh
S. Bhashyam
V. Veeravalli
14
13
0
24 Oct 2019
Mixture Martingales Revisited with Applications to Sequential Tests and
  Confidence Intervals
Mixture Martingales Revisited with Applications to Sequential Tests and Confidence Intervals
E. Kaufmann
Wouter M. Koolen
31
117
0
28 Nov 2018
Exploration in Structured Reinforcement Learning
Exploration in Structured Reinforcement Learning
Jungseul Ok
Alexandre Proutiere
Damianos Tranos
33
62
0
03 Jun 2018
Multi-Armed Bandits on Partially Revealed Unit Interval Graphs
Multi-Armed Bandits on Partially Revealed Unit Interval Graphs
Xiao Xu
Sattar Vakili
Qing Zhao
A. Swami
18
5
0
12 Feb 2018
Online Learning: A Comprehensive Survey
Online Learning: A Comprehensive Survey
Guosheng Lin
Doyen Sahoo
Jing Lu
P. Zhao
OffRL
31
636
0
08 Feb 2018
Thresholding Bandit for Dose-ranging: The Impact of Monotonicity
Thresholding Bandit for Dose-ranging: The Impact of Monotonicity
Aurélien Garivier
Pierre Ménard
Laurent Rossi
Pierre Menard
22
27
0
13 Nov 2017
Minimal Exploration in Structured Stochastic Bandits
Minimal Exploration in Structured Stochastic Bandits
Richard Combes
Stefan Magureanu
Alexandre Proutiere
44
115
0
01 Nov 2017
Learning the distribution with largest mean: two bandit frameworks
Learning the distribution with largest mean: two bandit frameworks
E. Kaufmann
Aurélien Garivier
30
19
0
31 Jan 2017
Online Learning with Gaussian Payoffs and Side Observations
Online Learning with Gaussian Payoffs and Side Observations
Yifan Wu
András Gyorgy
Csaba Szepesvári
15
44
0
27 Oct 2015
Unimodal Bandits without Smoothness
Unimodal Bandits without Smoothness
Richard Combes
Alexandre Proutiere
38
18
0
28 Jun 2014
Bandits and Experts in Metric Spaces
Bandits and Experts in Metric Spaces
Robert D. Kleinberg
Aleksandrs Slivkins
E. Upfal
49
120
0
04 Dec 2013
1