Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2205.13566
Cited By
Exploration, Exploitation, and Engagement in Multi-Armed Bandits with Abandonment
26 May 2022
Zixi Yang
Xin Liu
Lei Ying
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Exploration, Exploitation, and Engagement in Multi-Armed Bandits with Abandonment"
12 / 12 papers shown
Title
Reinforcement Learning from Human Feedback without Reward Inference: Model-Free Algorithm and Instance-Dependent Analysis
Qining Zhang
Honghao Wei
Lei Ying
OffRL
108
2
0
11 Jun 2024
Modeling Attrition in Recommender Systems with Departing Bandits
Omer Ben-Porat
Lee Cohen
Liu Leqi
Zachary Chase Lipton
Yishay Mansour
64
12
0
25 Mar 2022
Implicit Finite-Horizon Approximation and Efficient Optimal Algorithms for Stochastic Shortest Path
Liyu Chen
Mehdi Jafarnia-Jahromi
R. Jain
Haipeng Luo
61
25
0
15 Jun 2021
Regret Bounds for Stochastic Shortest Path Problems with Linear Function Approximation
Daniel Vial
Advait Parulekar
Sanjay Shakkottai
R. Srikant
72
15
0
04 May 2021
Stochastic Shortest Path: Minimax, Parameter-Free and Towards Horizon-Free Regret
Jean Tarbouriech
Runlong Zhou
S. Du
Matteo Pirotta
M. Valko
A. Lazaric
98
36
0
22 Apr 2021
Minimax Regret for Stochastic Shortest Path
Alon Cohen
Yonathan Efroni
Yishay Mansour
Aviv A. Rosenberg
61
28
0
24 Mar 2021
Q
Q
Q
-learning with Logarithmic Regret
Kunhe Yang
Lin F. Yang
S. Du
74
59
0
16 Jun 2020
Dynamic Learning of Sequential Choice Bandit Problem under Marketing Fatigue
Junyu Cao
Wei-Ju Sun
28
18
0
19 Mar 2019
Concentration Inequalities for the Empirical Distribution
Jay Mardia
Jiantao Jiao
Ervin Tánczos
Robert D. Nowak
Tsachy Weissman
48
53
0
18 Sep 2018
Learning with Abandonment
Ramesh Johari
S. Schmit
40
10
0
23 Feb 2018
Adaptive Exploration-Exploitation Tradeoff for Opportunistic Bandits
Huasen Wu
Xueying Guo
Xin Liu
44
29
0
12 Sep 2017
The KL-UCB Algorithm for Bounded Stochastic Bandits and Beyond
Aurélien Garivier
Olivier Cappé
174
612
0
12 Feb 2011
1