ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2202.01562
  4. Cited By
Doubly Robust Off-Policy Evaluation for Ranking Policies under the
  Cascade Behavior Model

Doubly Robust Off-Policy Evaluation for Ranking Policies under the Cascade Behavior Model

3 February 2022
Haruka Kiyohara
Yuta Saito
Tatsuya Matsuhiro
Yusuke Narita
N. Shimizu
Yasuo Yamamoto
    OffRL
ArXivPDFHTML

Papers citing "Doubly Robust Off-Policy Evaluation for Ranking Policies under the Cascade Behavior Model"

23 / 23 papers shown
Title
DOLCE: Decomposing Off-Policy Evaluation/Learning into Lagged and Current Effects
DOLCE: Decomposing Off-Policy Evaluation/Learning into Lagged and Current Effects
Shu Tamano
Masanori Nojima
OffRL
37
0
0
02 May 2025
Effective Off-Policy Evaluation and Learning in Contextual Combinatorial
  Bandits
Effective Off-Policy Evaluation and Learning in Contextual Combinatorial Bandits
Tatsuhiro Shimizu
Koichi Tanaka
Ren Kishimoto
Haruka Kiyohara
Masahiro Nomura
Yuta Saito
CML
OffRL
39
0
0
20 Aug 2024
Long-term Off-Policy Evaluation and Learning
Long-term Off-Policy Evaluation and Learning
Yuta Saito
Himan Abdollahpouri
Jesse Anderton
Ben Carterette
M. Lalmas
OffRL
29
5
0
24 Apr 2024
Hyperparameter Optimization Can Even be Harmful in Off-Policy Learning
  and How to Deal with It
Hyperparameter Optimization Can Even be Harmful in Off-Policy Learning and How to Deal with It
Yuta Saito
Masahiro Nomura
OffRL
47
2
0
23 Apr 2024
POTEC: Off-Policy Learning for Large Action Spaces via Two-Stage Policy
  Decomposition
POTEC: Off-Policy Learning for Large Action Spaces via Two-Stage Policy Decomposition
Yuta Saito
Jihan Yao
Thorsten Joachims
OffRL
24
6
0
09 Feb 2024
Off-Policy Evaluation of Slate Bandit Policies via Optimizing
  Abstraction
Off-Policy Evaluation of Slate Bandit Policies via Optimizing Abstraction
Haruka Kiyohara
Masahiro Nomura
Yuta Saito
22
5
0
03 Feb 2024
Towards Assessing and Benchmarking Risk-Return Tradeoff of Off-Policy
  Evaluation
Towards Assessing and Benchmarking Risk-Return Tradeoff of Off-Policy Evaluation
Haruka Kiyohara
Ren Kishimoto
K. Kawakami
Ken Kobayashi
Kazuhide Nakata
Yuta Saito
OffRL
27
9
0
30 Nov 2023
SCOPE-RL: A Python Library for Offline Reinforcement Learning and
  Off-Policy Evaluation
SCOPE-RL: A Python Library for Offline Reinforcement Learning and Off-Policy Evaluation
Haruka Kiyohara
Ren Kishimoto
K. Kawakami
Ken Kobayashi
Kazuhide Nakata
Yuta Saito
OffRL
ELM
34
4
0
30 Nov 2023
Distributional Off-Policy Evaluation for Slate Recommendations
Distributional Off-Policy Evaluation for Slate Recommendations
Shreyas Chaudhari
David Arbour
Georgios Theocharous
N. Vlassis
OffRL
44
0
0
27 Aug 2023
On (Normalised) Discounted Cumulative Gain as an Off-Policy Evaluation
  Metric for Top-$n$ Recommendation
On (Normalised) Discounted Cumulative Gain as an Off-Policy Evaluation Metric for Top-nnn Recommendation
Olivier Jeunen
Ivan Potapov
Aleksei Ustimenko
ELM
OffRL
27
11
0
27 Jul 2023
Off-Policy Evaluation of Ranking Policies under Diverse User Behavior
Off-Policy Evaluation of Ranking Policies under Diverse User Behavior
Haruka Kiyohara
Masatoshi Uehara
Yusuke Narita
N. Shimizu
Yasuo Yamamoto
Yuta Saito
OffRL
CML
28
8
0
26 Jun 2023
Unified Off-Policy Learning to Rank: a Reinforcement Learning
  Perspective
Unified Off-Policy Learning to Rank: a Reinforcement Learning Perspective
Zeyu Zhang
Yi-Hsun Su
Hui Yuan
Yiran Wu
R. Balasubramanian
Qingyun Wu
Huazheng Wang
Mengdi Wang
OffRL
CML
36
4
0
13 Jun 2023
Off-Policy Evaluation for Large Action Spaces via Conjunct Effect
  Modeling
Off-Policy Evaluation for Large Action Spaces via Conjunct Effect Modeling
Yuta Saito
Qingyang Ren
Thorsten Joachims
CML
OffRL
19
22
0
14 May 2023
Recent Advances in the Foundations and Applications of Unbiased Learning
  to Rank
Recent Advances in the Foundations and Applications of Unbiased Learning to Rank
Shashank Gupta
Philipp Hager
Jin Huang
Ali Vardasbi
Harrie Oosterhuis
OffRL
35
5
0
04 May 2023
Practical Bandits: An Industry Perspective
Practical Bandits: An Industry Perspective
Bram van den Akker
Olivier Jeunen
Ying Li
Ben London
Zahra Nazari
Devesh Parekh
21
5
0
02 Feb 2023
An Instrumental Variable Approach to Confounded Off-Policy Evaluation
An Instrumental Variable Approach to Confounded Off-Policy Evaluation
Yang Xu
Jin Zhu
C. Shi
S. Luo
R. Song
OffRL
21
14
0
29 Dec 2022
Policy-Adaptive Estimator Selection for Off-Policy Evaluation
Policy-Adaptive Estimator Selection for Off-Policy Evaluation
Takuma Udagawa
Haruka Kiyohara
Yusuke Narita
Yuta Saito
Keisuke Tateno
OffRL
27
23
0
25 Nov 2022
Pessimistic Off-Policy Optimization for Learning to Rank
Pessimistic Off-Policy Optimization for Learning to Rank
Matej Cief
B. Kveton
Michal Kompan
OffRL
22
3
0
06 Jun 2022
Doubly-Robust Estimation for Correcting Position-Bias in Click Feedback
  for Unbiased Learning to Rank
Doubly-Robust Estimation for Correcting Position-Bias in Click Feedback for Unbiased Learning to Rank
Harrie Oosterhuis
CML
25
27
0
31 Mar 2022
Off-Policy Evaluation for Large Action Spaces via Embeddings
Off-Policy Evaluation for Large Action Spaces via Embeddings
Yuta Saito
Thorsten Joachims
OffRL
25
43
0
13 Feb 2022
Counterfactual Evaluation of Slate Recommendations with Sequential
  Reward Interactions
Counterfactual Evaluation of Slate Recommendations with Sequential Reward Interactions
James McInerney
B. Brost
Praveen Chandar
Rishabh Mehrotra
Ben Carterette
BDL
CML
OffRL
118
55
0
25 Jul 2020
Offline Reinforcement Learning: Tutorial, Review, and Perspectives on
  Open Problems
Offline Reinforcement Learning: Tutorial, Review, and Perspectives on Open Problems
Sergey Levine
Aviral Kumar
George Tucker
Justin Fu
OffRL
GP
340
1,960
0
04 May 2020
Towards Resolving Propensity Contradiction in Offline Recommender
  Learning
Towards Resolving Propensity Contradiction in Offline Recommender Learning
Yuta Saito
Masahiro Nomura
OffRL
25
12
0
16 Oct 2019
1