ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1903.09132
  4. Cited By
Perturbed-History Exploration in Stochastic Linear Bandits

Perturbed-History Exploration in Stochastic Linear Bandits

21 March 2019
Branislav Kveton
Csaba Szepesvári
Mohammad Ghavamzadeh
Craig Boutilier
ArXivPDFHTML

Papers citing "Perturbed-History Exploration in Stochastic Linear Bandits"

13 / 13 papers shown
Title
Improved Regret of Linear Ensemble Sampling
Improved Regret of Linear Ensemble Sampling
Harin Lee
Min-hwan Oh
42
1
0
06 Nov 2024
Random Latent Exploration for Deep Reinforcement Learning
Random Latent Exploration for Deep Reinforcement Learning
Srinath Mahankali
Zhang-Wei Hong
Ayush Sekhari
Alexander Rakhlin
Pulkit Agrawal
38
3
0
18 Jul 2024
Graph Neural Thompson Sampling
Graph Neural Thompson Sampling
Shuang Wu
Arash A. Amini
56
0
0
15 Jun 2024
Zero-Inflated Bandits
Zero-Inflated Bandits
Haoyu Wei
Runzhe Wan
Lei Shi
Rui Song
44
0
0
25 Dec 2023
Ensemble sampling for linear bandits: small ensembles suffice
Ensemble sampling for linear bandits: small ensembles suffice
David Janz
A. Litvak
Csaba Szepesvári
38
1
0
14 Nov 2023
Multiplier Bootstrap-based Exploration
Multiplier Bootstrap-based Exploration
Runzhe Wan
Haoyu Wei
Branislav Kveton
R. Song
21
3
0
03 Feb 2023
An Analysis of Ensemble Sampling
An Analysis of Ensemble Sampling
Chao Qin
Zheng Wen
Xiuyuan Lu
Benjamin Van Roy
34
21
0
02 Mar 2022
Anti-Concentrated Confidence Bonuses for Scalable Exploration
Anti-Concentrated Confidence Bonuses for Scalable Exploration
Jordan T. Ash
Cyril Zhang
Surbhi Goel
A. Krishnamurthy
Sham Kakade
45
6
0
21 Oct 2021
An Efficient Algorithm for Deep Stochastic Contextual Bandits
An Efficient Algorithm for Deep Stochastic Contextual Bandits
Tan Zhu
Guannan Liang
Chunjiang Zhu
HaiNing Li
J. Bi
42
1
0
12 Apr 2021
An Efficient Algorithm For Generalized Linear Bandit: Online Stochastic
  Gradient Descent and Thompson Sampling
An Efficient Algorithm For Generalized Linear Bandit: Online Stochastic Gradient Descent and Thompson Sampling
Qin Ding
Cho-Jui Hsieh
James Sharpnack
25
37
0
07 Jun 2020
Neural Contextual Bandits with UCB-based Exploration
Neural Contextual Bandits with UCB-based Exploration
Dongruo Zhou
Lihong Li
Quanquan Gu
36
15
0
11 Nov 2019
Stochastic Bandits with Context Distributions
Stochastic Bandits with Context Distributions
Johannes Kirschner
Andreas Krause
29
30
0
06 Jun 2019
Garbage In, Reward Out: Bootstrapping Exploration in Multi-Armed Bandits
Garbage In, Reward Out: Bootstrapping Exploration in Multi-Armed Bandits
Branislav Kveton
Csaba Szepesvári
Sharan Vaswani
Zheng Wen
Mohammad Ghavamzadeh
Tor Lattimore
10
69
0
13 Nov 2018
1