ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2002.03221
  4. Cited By
Improved Algorithms for Conservative Exploration in Bandits

Improved Algorithms for Conservative Exploration in Bandits

8 February 2020
Evrard Garcelon
Mohammad Ghavamzadeh
A. Lazaric
Matteo Pirotta
ArXivPDFHTML

Papers citing "Improved Algorithms for Conservative Exploration in Bandits"

7 / 7 papers shown
Title
Conservative Exploration for Policy Optimization via Off-Policy Policy
  Evaluation
Conservative Exploration for Policy Optimization via Off-Policy Policy Evaluation
Paul Daoudi
Mathias Formoso
Othman Gaizi
Achraf Azize
Evrard Garcelon
OffRL
31
0
0
24 Dec 2023
Near-optimal Conservative Exploration in Reinforcement Learning under
  Episode-wise Constraints
Near-optimal Conservative Exploration in Reinforcement Learning under Episode-wise Constraints
Donghao Li
Ruiquan Huang
Cong Shen
Jing Yang
45
3
0
09 Jun 2023
Safety Aware Changepoint Detection for Piecewise i.i.d. Bandits
Safety Aware Changepoint Detection for Piecewise i.i.d. Bandits
Subhojyoti Mukherjee
21
1
0
27 May 2022
Learning Infinite-Horizon Average-Reward Markov Decision Processes with
  Constraints
Learning Infinite-Horizon Average-Reward Markov Decision Processes with Constraints
Liyu Chen
R. Jain
Haipeng Luo
72
25
0
31 Jan 2022
Bandit Algorithms for Precision Medicine
Bandit Algorithms for Precision Medicine
Yangyi Lu
Ziping Xu
Ambuj Tewari
66
11
0
10 Aug 2021
Learning Policies with Zero or Bounded Constraint Violation for
  Constrained MDPs
Learning Policies with Zero or Bounded Constraint Violation for Constrained MDPs
Tao-Wen Liu
Ruida Zhou
D. Kalathil
P. R. Kumar
Chao Tian
44
78
0
04 Jun 2021
Exploration-Exploitation in Constrained MDPs
Exploration-Exploitation in Constrained MDPs
Yonathan Efroni
Shie Mannor
Matteo Pirotta
33
171
0
04 Mar 2020
1