ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1910.06508
  4. Cited By
Understanding the Curse of Horizon in Off-Policy Evaluation via
  Conditional Importance Sampling

Understanding the Curse of Horizon in Off-Policy Evaluation via Conditional Importance Sampling

15 October 2019
Yao Liu
Pierre-Luc Bacon
Emma Brunskill
    OffRL
ArXivPDFHTML

Papers citing "Understanding the Curse of Horizon in Off-Policy Evaluation via Conditional Importance Sampling"

13 / 13 papers shown
Title
DOLCE: Decomposing Off-Policy Evaluation/Learning into Lagged and Current Effects
DOLCE: Decomposing Off-Policy Evaluation/Learning into Lagged and Current Effects
Shu Tamano
Masanori Nojima
OffRL
37
0
0
02 May 2025
Conservative Exploration for Policy Optimization via Off-Policy Policy
  Evaluation
Conservative Exploration for Policy Optimization via Off-Policy Policy Evaluation
Paul Daoudi
Mathias Formoso
Othman Gaizi
Achraf Azize
Evrard Garcelon
OffRL
26
0
0
24 Dec 2023
Stackelberg Batch Policy Learning
Stackelberg Batch Policy Learning
Wenzhuo Zhou
Annie Qu
OffRL
35
0
0
28 Sep 2023
Estimating Time-Varying Direct and Indirect Causal Excursion Effects
  with Longitudinal Binary Outcomes
Estimating Time-Varying Direct and Indirect Causal Excursion Effects with Longitudinal Binary Outcomes
Jieru Shi
Zhanghua Wu
Walter Dempsey
CML
16
1
0
02 Dec 2022
Flexible Option Learning
Flexible Option Learning
Martin Klissarov
Doina Precup
OffRL
41
26
0
06 Dec 2021
SOPE: Spectrum of Off-Policy Estimators
SOPE: Spectrum of Off-Policy Estimators
C. J. Yuan
Yash Chandak
S. Giguere
Philip S. Thomas
S. Niekum
OffRL
50
5
0
06 Nov 2021
Adaptive Importance Sampling meets Mirror Descent: a Bias-variance
  tradeoff
Adaptive Importance Sampling meets Mirror Descent: a Bias-variance tradeoff
Anna Korba
Franccois Portier
28
12
0
29 Oct 2021
Universal Off-Policy Evaluation
Universal Off-Policy Evaluation
Yash Chandak
S. Niekum
Bruno C. da Silva
Erik Learned-Miller
Emma Brunskill
Philip S. Thomas
OffRL
ELM
34
52
0
26 Apr 2021
Instabilities of Offline RL with Pre-Trained Neural Representation
Instabilities of Offline RL with Pre-Trained Neural Representation
Ruosong Wang
Yifan Wu
Ruslan Salakhutdinov
Sham Kakade
OffRL
20
42
0
08 Mar 2021
CoinDICE: Off-Policy Confidence Interval Estimation
CoinDICE: Off-Policy Confidence Interval Estimation
Bo Dai
Ofir Nachum
Yinlam Chow
Lihong Li
Csaba Szepesvári
Dale Schuurmans
OffRL
27
84
0
22 Oct 2020
Off-policy Evaluation in Infinite-Horizon Reinforcement Learning with
  Latent Confounders
Off-policy Evaluation in Infinite-Horizon Reinforcement Learning with Latent Confounders
Andrew Bennett
Nathan Kallus
Lihong Li
Ali Mousavi
OffRL
35
43
0
27 Jul 2020
Minimax Value Interval for Off-Policy Evaluation and Policy Optimization
Minimax Value Interval for Off-Policy Evaluation and Policy Optimization
Nan Jiang
Jiawei Huang
OffRL
41
17
0
06 Feb 2020
Double Reinforcement Learning for Efficient Off-Policy Evaluation in
  Markov Decision Processes
Double Reinforcement Learning for Efficient Off-Policy Evaluation in Markov Decision Processes
Nathan Kallus
Masatoshi Uehara
OffRL
38
183
0
22 Aug 2019
1