Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
1711.00400
Cited By
Minimal Exploration in Structured Stochastic Bandits
1 November 2017
Richard Combes
Stefan Magureanu
Alexandre Proutiere
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Minimal Exploration in Structured Stochastic Bandits"
28 / 28 papers shown
Title
Greedy Algorithm for Structured Bandits: A Sharp Characterization of Asymptotic Success / Failure
Aleksandrs Slivkins
Yunzong Xu
Shiliang Zuo
86
1
0
06 Mar 2025
A Complete Characterization of Learnability for Stochastic Noisy Bandits
Steve Hanneke
Kun Wang
40
0
0
20 Jan 2025
Matroid Semi-Bandits in Sublinear Time
Ruo-Chun Tzeng
Naoto Ohsaka
Kaito Ariu
37
0
0
28 May 2024
Causally Abstracted Multi-armed Bandits
Fabio Massimo Zennaro
Nicholas Bishop
Joel Dyer
Yorgos Felekis
Anisoara Calinescu
Michael Wooldridge
Theodoros Damoulas
38
2
0
26 Apr 2024
Quantum contextual bandits and recommender systems for quantum data
Shrigyan Brahmachari
Josep Lumbreras
Marco Tomamichel
37
3
0
31 Jan 2023
SPEED: Experimental Design for Policy Evaluation in Linear Heteroscedastic Bandits
Subhojyoti Mukherjee
Qiaomin Xie
Josiah P. Hanna
R. Nowak
OffRL
55
5
0
29 Jan 2023
Interactive Recommendations for Optimal Allocations in Markets with Constraints
Yigit Efe Erginbas
Soham R. Phade
Kannan Ramchandran
23
1
0
08 Jul 2022
Near-Optimal Collaborative Learning in Bandits
Clémence Réda
Sattar Vakili
E. Kaufmann
FedML
30
21
0
31 May 2022
Truncated LinUCB for Stochastic Linear Bandits
Yanglei Song
Meng zhou
52
0
0
23 Feb 2022
Fast online inference for nonlinear contextual bandit based on Generative Adversarial Network
Yun-Da Tsai
Shou-De Lin
51
5
0
17 Feb 2022
Breaking the Moments Condition Barrier: No-Regret Algorithm for Bandits with Super Heavy-Tailed Payoffs
Han Zhong
Jiayi Huang
Lin F. Yang
Liwei Wang
24
7
0
26 Oct 2021
Multi-armed Bandit Algorithm against Strategic Replication
Suho Shin
Seungjoon Lee
Jungseul Ok
30
4
0
23 Oct 2021
Fair Exploration via Axiomatic Bargaining
Jackie Baek
Vivek F. Farias
FaML
18
28
0
04 Jun 2021
Information Directed Sampling for Sparse Linear Bandits
Botao Hao
Tor Lattimore
Wei Deng
25
19
0
29 May 2021
On the Suboptimality of Thompson Sampling in High Dimensions
Raymond Zhang
Richard Combes
16
4
0
10 Feb 2021
TSEC: a framework for online experimentation under experimental constraints
Simon Mak
Yuanshuo Zhou
Lavonne Hoang
C. F. J. Wu
20
2
0
17 Jan 2021
Policy Optimization as Online Learning with Mediator Feedback
Alberto Maria Metelli
Matteo Papini
P. DÓro
Marcello Restelli
OffRL
27
10
0
15 Dec 2020
Multi-Armed Bandits with Dependent Arms
Rahul Singh
Fang Liu
Yin Sun
Ness B. Shroff
21
11
0
13 Oct 2020
Optimal Best-arm Identification in Linear Bandits
Yassir Jedra
Alexandre Proutiere
19
75
0
29 Jun 2020
Crush Optimism with Pessimism: Structured Bandits Beyond Asymptotic Optimality
Kwang-Sung Jun
Chicheng Zhang
31
10
0
15 Jun 2020
Categorized Bandits
Matthieu Jedor
Jonathan Louëdec
Vianney Perchet
25
11
0
04 May 2020
Bounded Regret for Finitely Parameterized Multi-Armed Bandits
Kishan Panaganti
D. Kalathil
18
1
0
03 Mar 2020
Multi-Armed Bandits with Correlated Arms
Samarth Gupta
Shreyas Chaudhari
Gauri Joshi
Osman Yağan
22
50
0
06 Nov 2019
Adaptive Exploration in Linear Contextual Bandit
Botao Hao
Tor Lattimore
Csaba Szepesvári
27
74
0
15 Oct 2019
Beating Stochastic and Adversarial Semi-bandits Optimally and Simultaneously
Julian Zimmert
Haipeng Luo
Chen-Yu Wei
11
79
0
25 Jan 2019
Mixture Martingales Revisited with Applications to Sequential Tests and Confidence Intervals
E. Kaufmann
Wouter M. Koolen
26
117
0
28 Nov 2018
Exploration in Structured Reinforcement Learning
Jungseul Ok
Alexandre Proutiere
Damianos Tranos
27
62
0
03 Jun 2018
Multi-Armed Bandits on Partially Revealed Unit Interval Graphs
Xiao Xu
Sattar Vakili
Qing Zhao
A. Swami
13
5
0
12 Feb 2018
1