ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1711.00400
  4. Cited By
Minimal Exploration in Structured Stochastic Bandits

Minimal Exploration in Structured Stochastic Bandits

1 November 2017
Richard Combes
Stefan Magureanu
Alexandre Proutiere
ArXivPDFHTML

Papers citing "Minimal Exploration in Structured Stochastic Bandits"

28 / 28 papers shown
Title
Greedy Algorithm for Structured Bandits: A Sharp Characterization of Asymptotic Success / Failure
Greedy Algorithm for Structured Bandits: A Sharp Characterization of Asymptotic Success / Failure
Aleksandrs Slivkins
Yunzong Xu
Shiliang Zuo
86
1
0
06 Mar 2025
A Complete Characterization of Learnability for Stochastic Noisy Bandits
A Complete Characterization of Learnability for Stochastic Noisy Bandits
Steve Hanneke
Kun Wang
40
0
0
20 Jan 2025
Matroid Semi-Bandits in Sublinear Time
Matroid Semi-Bandits in Sublinear Time
Ruo-Chun Tzeng
Naoto Ohsaka
Kaito Ariu
37
0
0
28 May 2024
Causally Abstracted Multi-armed Bandits
Causally Abstracted Multi-armed Bandits
Fabio Massimo Zennaro
Nicholas Bishop
Joel Dyer
Yorgos Felekis
Anisoara Calinescu
Michael Wooldridge
Theodoros Damoulas
38
2
0
26 Apr 2024
Quantum contextual bandits and recommender systems for quantum data
Quantum contextual bandits and recommender systems for quantum data
Shrigyan Brahmachari
Josep Lumbreras
Marco Tomamichel
37
3
0
31 Jan 2023
SPEED: Experimental Design for Policy Evaluation in Linear
  Heteroscedastic Bandits
SPEED: Experimental Design for Policy Evaluation in Linear Heteroscedastic Bandits
Subhojyoti Mukherjee
Qiaomin Xie
Josiah P. Hanna
R. Nowak
OffRL
53
5
0
29 Jan 2023
Interactive Recommendations for Optimal Allocations in Markets with
  Constraints
Interactive Recommendations for Optimal Allocations in Markets with Constraints
Yigit Efe Erginbas
Soham R. Phade
Kannan Ramchandran
23
1
0
08 Jul 2022
Near-Optimal Collaborative Learning in Bandits
Near-Optimal Collaborative Learning in Bandits
Clémence Réda
Sattar Vakili
E. Kaufmann
FedML
30
21
0
31 May 2022
Truncated LinUCB for Stochastic Linear Bandits
Truncated LinUCB for Stochastic Linear Bandits
Yanglei Song
Meng zhou
52
0
0
23 Feb 2022
Fast online inference for nonlinear contextual bandit based on
  Generative Adversarial Network
Fast online inference for nonlinear contextual bandit based on Generative Adversarial Network
Yun-Da Tsai
Shou-De Lin
51
5
0
17 Feb 2022
Breaking the Moments Condition Barrier: No-Regret Algorithm for Bandits
  with Super Heavy-Tailed Payoffs
Breaking the Moments Condition Barrier: No-Regret Algorithm for Bandits with Super Heavy-Tailed Payoffs
Han Zhong
Jiayi Huang
Lin F. Yang
Liwei Wang
21
7
0
26 Oct 2021
Multi-armed Bandit Algorithm against Strategic Replication
Multi-armed Bandit Algorithm against Strategic Replication
Suho Shin
Seungjoon Lee
Jungseul Ok
30
4
0
23 Oct 2021
Fair Exploration via Axiomatic Bargaining
Fair Exploration via Axiomatic Bargaining
Jackie Baek
Vivek F. Farias
FaML
18
28
0
04 Jun 2021
Information Directed Sampling for Sparse Linear Bandits
Information Directed Sampling for Sparse Linear Bandits
Botao Hao
Tor Lattimore
Wei Deng
25
19
0
29 May 2021
On the Suboptimality of Thompson Sampling in High Dimensions
On the Suboptimality of Thompson Sampling in High Dimensions
Raymond Zhang
Richard Combes
16
4
0
10 Feb 2021
TSEC: a framework for online experimentation under experimental
  constraints
TSEC: a framework for online experimentation under experimental constraints
Simon Mak
Yuanshuo Zhou
Lavonne Hoang
C. F. J. Wu
20
2
0
17 Jan 2021
Policy Optimization as Online Learning with Mediator Feedback
Policy Optimization as Online Learning with Mediator Feedback
Alberto Maria Metelli
Matteo Papini
P. DÓro
Marcello Restelli
OffRL
27
10
0
15 Dec 2020
Multi-Armed Bandits with Dependent Arms
Multi-Armed Bandits with Dependent Arms
Rahul Singh
Fang Liu
Yin Sun
Ness B. Shroff
21
11
0
13 Oct 2020
Optimal Best-arm Identification in Linear Bandits
Optimal Best-arm Identification in Linear Bandits
Yassir Jedra
Alexandre Proutiere
13
75
0
29 Jun 2020
Crush Optimism with Pessimism: Structured Bandits Beyond Asymptotic
  Optimality
Crush Optimism with Pessimism: Structured Bandits Beyond Asymptotic Optimality
Kwang-Sung Jun
Chicheng Zhang
31
10
0
15 Jun 2020
Categorized Bandits
Categorized Bandits
Matthieu Jedor
Jonathan Louëdec
Vianney Perchet
25
11
0
04 May 2020
Bounded Regret for Finitely Parameterized Multi-Armed Bandits
Bounded Regret for Finitely Parameterized Multi-Armed Bandits
Kishan Panaganti
D. Kalathil
18
1
0
03 Mar 2020
Multi-Armed Bandits with Correlated Arms
Multi-Armed Bandits with Correlated Arms
Samarth Gupta
Shreyas Chaudhari
Gauri Joshi
Osman Yağan
22
50
0
06 Nov 2019
Adaptive Exploration in Linear Contextual Bandit
Adaptive Exploration in Linear Contextual Bandit
Botao Hao
Tor Lattimore
Csaba Szepesvári
24
74
0
15 Oct 2019
Beating Stochastic and Adversarial Semi-bandits Optimally and
  Simultaneously
Beating Stochastic and Adversarial Semi-bandits Optimally and Simultaneously
Julian Zimmert
Haipeng Luo
Chen-Yu Wei
11
79
0
25 Jan 2019
Mixture Martingales Revisited with Applications to Sequential Tests and
  Confidence Intervals
Mixture Martingales Revisited with Applications to Sequential Tests and Confidence Intervals
E. Kaufmann
Wouter M. Koolen
21
117
0
28 Nov 2018
Exploration in Structured Reinforcement Learning
Exploration in Structured Reinforcement Learning
Jungseul Ok
Alexandre Proutiere
Damianos Tranos
25
62
0
03 Jun 2018
Multi-Armed Bandits on Partially Revealed Unit Interval Graphs
Multi-Armed Bandits on Partially Revealed Unit Interval Graphs
Xiao Xu
Sattar Vakili
Qing Zhao
A. Swami
13
5
0
12 Feb 2018
1