ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1911.05815
  4. Cited By
Kinematic State Abstraction and Provably Efficient Rich-Observation
  Reinforcement Learning

Kinematic State Abstraction and Provably Efficient Rich-Observation Reinforcement Learning

13 November 2019
Dipendra Kumar Misra
Mikael Henaff
A. Krishnamurthy
John Langford
ArXivPDFHTML

Papers citing "Kinematic State Abstraction and Provably Efficient Rich-Observation Reinforcement Learning"

50 / 52 papers shown
Title
Learning a Fast Mixing Exogenous Block MDP using a Single Trajectory
Learning a Fast Mixing Exogenous Block MDP using a Single Trajectory
Alexander Levine
Peter Stone
Amy Zhang
OffRL
41
0
0
03 Oct 2024
The Central Role of the Loss Function in Reinforcement Learning
The Central Role of the Loss Function in Reinforcement Learning
Kaiwen Wang
Nathan Kallus
Wen Sun
OffRL
62
7
0
19 Sep 2024
A Note on Loss Functions and Error Compounding in Model-based
  Reinforcement Learning
A Note on Loss Functions and Error Compounding in Model-based Reinforcement Learning
Nan Jiang
32
5
0
15 Apr 2024
Exploration is Harder than Prediction: Cryptographically Separating
  Reinforcement Learning from Supervised Learning
Exploration is Harder than Prediction: Cryptographically Separating Reinforcement Learning from Supervised Learning
Noah Golowich
Ankur Moitra
Dhruv Rohatgi
OffRL
35
4
0
04 Apr 2024
When is Agnostic Reinforcement Learning Statistically Tractable?
When is Agnostic Reinforcement Learning Statistically Tractable?
Zeyu Jia
Gene Li
Alexander Rakhlin
Ayush Sekhari
Nathan Srebro
OffRL
32
5
0
09 Oct 2023
Policy Finetuning in Reinforcement Learning via Design of Experiments
  using Offline Data
Policy Finetuning in Reinforcement Learning via Design of Experiments using Offline Data
Ruiqi Zhang
Andrea Zanette
OffRL
OnRL
40
7
0
10 Jul 2023
Provably Efficient Representation Learning with Tractable Planning in
  Low-Rank POMDP
Provably Efficient Representation Learning with Tractable Planning in Low-Rank POMDP
Jiacheng Guo
Zihao Li
Huazheng Wang
Mengdi Wang
Zhuoran Yang
Xuezhou Zhang
32
5
0
21 Jun 2023
Accelerating exploration and representation learning with offline
  pre-training
Accelerating exploration and representation learning with offline pre-training
Bogdan Mazoure
Jake Bruce
Doina Precup
Rob Fergus
Ankit Anand
OffRL
34
5
0
31 Mar 2023
Improved Sample Complexity for Reward-free Reinforcement Learning under
  Low-rank MDPs
Improved Sample Complexity for Reward-free Reinforcement Learning under Low-rank MDPs
Yuan Cheng
Ruiquan Huang
J. Yang
Yitao Liang
OffRL
41
8
0
20 Mar 2023
Fast exploration and learning of latent graphs with aliased observations
Fast exploration and learning of latent graphs with aliased observations
Miguel Lazaro-Gredilla
Ishani Deshpande
Siva K. Swaminathan
Meet Dave
Dileep George
28
3
0
13 Mar 2023
Distributional Offline Policy Evaluation with Predictive Error
  Guarantees
Distributional Offline Policy Evaluation with Predictive Error Guarantees
Runzhe Wu
Masatoshi Uehara
Wen Sun
OffRL
38
13
0
19 Feb 2023
Near-Minimax-Optimal Risk-Sensitive Reinforcement Learning with CVaR
Near-Minimax-Optimal Risk-Sensitive Reinforcement Learning with CVaR
Kaiwen Wang
Nathan Kallus
Wen Sun
107
18
0
07 Feb 2023
Understanding the Complexity Gains of Single-Task RL with a Curriculum
Understanding the Complexity Gains of Single-Task RL with a Curriculum
Qiyang Li
Yuexiang Zhai
Yi Ma
Sergey Levine
37
14
0
24 Dec 2022
Discrete Factorial Representations as an Abstraction for Goal
  Conditioned Reinforcement Learning
Discrete Factorial Representations as an Abstraction for Goal Conditioned Reinforcement Learning
Riashat Islam
Hongyu Zang
Anirudh Goyal
Alex Lamb
Kenji Kawaguchi
Xin-hui Li
Romain Laroche
Yoshua Bengio
Rémi Tachet des Combes
OffRL
AI4CE
23
9
0
01 Nov 2022
Planning to the Information Horizon of BAMDPs via Epistemic State
  Abstraction
Planning to the Information Horizon of BAMDPs via Epistemic State Abstraction
Dilip Arumugam
Satinder Singh
26
3
0
30 Oct 2022
Hardness in Markov Decision Processes: Theory and Practice
Hardness in Markov Decision Processes: Theory and Practice
Michelangelo Conserva
Paulo E. Rauber
37
3
0
24 Oct 2022
Spectral Decomposition Representation for Reinforcement Learning
Spectral Decomposition Representation for Reinforcement Learning
Tongzheng Ren
Tianjun Zhang
Lisa Lee
Joseph E. Gonzalez
Dale Schuurmans
Bo Dai
OffRL
40
27
0
19 Aug 2022
Contrastive UCB: Provably Efficient Contrastive Self-Supervised Learning
  in Online Reinforcement Learning
Contrastive UCB: Provably Efficient Contrastive Self-Supervised Learning in Online Reinforcement Learning
Shuang Qiu
Lingxiao Wang
Chenjia Bai
Zhuoran Yang
Zhaoran Wang
SSL
OffRL
26
32
0
29 Jul 2022
Guaranteed Discovery of Control-Endogenous Latent States with Multi-Step
  Inverse Models
Guaranteed Discovery of Control-Endogenous Latent States with Multi-Step Inverse Models
Alex Lamb
Riashat Islam
Yonathan Efroni
Aniket Didolkar
Dipendra Kumar Misra
Dylan J. Foster
Lekan Molu
Rajan Chari
A. Krishnamurthy
John Langford
46
24
0
17 Jul 2022
Making Linear MDPs Practical via Contrastive Representation Learning
Making Linear MDPs Practical via Contrastive Representation Learning
Tianjun Zhang
Tongzheng Ren
Mengjiao Yang
Joseph E. Gonzalez
Dale Schuurmans
Bo Dai
25
44
0
14 Jul 2022
Provably Efficient Reinforcement Learning in Partially Observable
  Dynamical Systems
Provably Efficient Reinforcement Learning in Partially Observable Dynamical Systems
Masatoshi Uehara
Ayush Sekhari
Jason D. Lee
Nathan Kallus
Wen Sun
OffRL
49
32
0
24 Jun 2022
On the Statistical Efficiency of Reward-Free Exploration in Non-Linear
  RL
On the Statistical Efficiency of Reward-Free Exploration in Non-Linear RL
Jinglin Chen
Aditya Modi
A. Krishnamurthy
Nan Jiang
Alekh Agarwal
38
25
0
21 Jun 2022
Model-based RL with Optimistic Posterior Sampling: Structural Conditions
  and Sample Complexity
Model-based RL with Optimistic Posterior Sampling: Structural Conditions and Sample Complexity
Alekh Agarwal
Tong Zhang
47
22
0
15 Jun 2022
Sample-Efficient Reinforcement Learning in the Presence of Exogenous
  Information
Sample-Efficient Reinforcement Learning in the Presence of Exogenous Information
Yonathan Efroni
Dylan J. Foster
Dipendra Kumar Misra
A. Krishnamurthy
John Langford
OffRL
31
25
0
09 Jun 2022
Uniqueness and Complexity of Inverse MDP Models
Uniqueness and Complexity of Inverse MDP Models
Marcus Hutter
Steven Hansen
25
4
0
02 Jun 2022
Provable Benefits of Representational Transfer in Reinforcement Learning
Provable Benefits of Representational Transfer in Reinforcement Learning
Alekh Agarwal
Yuda Song
Wen Sun
Kaiwen Wang
Mengdi Wang
Xuezhou Zhang
OffRL
23
33
0
29 May 2022
Chemoreception and chemotaxis of a three-sphere swimmer
Chemoreception and chemotaxis of a three-sphere swimmer
S. Paz
R. Ausas
J. P. Carbajal
G. Buscaglia
13
4
0
05 May 2022
When Is Partially Observable Reinforcement Learning Not Scary?
When Is Partially Observable Reinforcement Learning Not Scary?
Qinghua Liu
Alan Chung
Csaba Szepesvári
Chi Jin
22
93
0
19 Apr 2022
Efficient Reinforcement Learning in Block MDPs: A Model-free
  Representation Learning Approach
Efficient Reinforcement Learning in Block MDPs: A Model-free Representation Learning Approach
Xuezhou Zhang
Yuda Song
Masatoshi Uehara
Mengdi Wang
Alekh Agarwal
Wen Sun
OffRL
29
57
0
31 Jan 2022
A Free Lunch from the Noise: Provable and Practical Exploration for
  Representation Learning
A Free Lunch from the Noise: Provable and Practical Exploration for Representation Learning
Tongzheng Ren
Tianjun Zhang
Csaba Szepesvári
Bo Dai
27
19
0
22 Nov 2021
Representation Learning for Online and Offline RL in Low-rank MDPs
Representation Learning for Online and Offline RL in Low-rank MDPs
Masatoshi Uehara
Xuezhou Zhang
Wen Sun
OffRL
62
127
0
09 Oct 2021
The Information Geometry of Unsupervised Reinforcement Learning
The Information Geometry of Unsupervised Reinforcement Learning
Benjamin Eysenbach
Ruslan Salakhutdinov
Sergey Levine
SSL
OffRL
61
31
0
06 Oct 2021
Explore and Control with Adversarial Surprise
Explore and Control with Adversarial Surprise
Arnaud Fickinger
Natasha Jaques
Samyak Parajuli
Michael Chang
Nicholas Rhinehart
Glen Berseth
Stuart J. Russell
Sergey Levine
40
8
0
12 Jul 2021
Provably Efficient Representation Selection in Low-rank Markov Decision
  Processes: From Online to Offline RL
Provably Efficient Representation Selection in Low-rank Markov Decision Processes: From Online to Offline RL
Weitong Zhang
Jiafan He
Dongruo Zhou
Amy Zhang
Quanquan Gu
OffRL
22
11
0
22 Jun 2021
On the Sample Complexity and Metastability of Heavy-tailed Policy Search
  in Continuous Control
On the Sample Complexity and Metastability of Heavy-tailed Policy Search in Continuous Control
Amrit Singh Bedi
Anjaly Parayil
Junyu Zhang
Mengdi Wang
Alec Koppel
38
15
0
15 Jun 2021
Which Mutual-Information Representation Learning Objectives are
  Sufficient for Control?
Which Mutual-Information Representation Learning Objectives are Sufficient for Control?
Kate Rakelly
Abhishek Gupta
Carlos Florensa
Sergey Levine
SSL
26
38
0
14 Jun 2021
Learning Markov State Abstractions for Deep Reinforcement Learning
Learning Markov State Abstractions for Deep Reinforcement Learning
Cameron Allen
Neev Parikh
Omer Gottesman
George Konidaris
BDL
OffRL
31
36
0
08 Jun 2021
Rapid Exploration for Open-World Navigation with Latent Goal Models
Rapid Exploration for Open-World Navigation with Latent Goal Models
Dhruv Shah
Benjamin Eysenbach
G. Kahn
Nicholas Rhinehart
Sergey Levine
32
70
0
12 Apr 2021
An Exponential Lower Bound for Linearly-Realizable MDPs with Constant
  Suboptimality Gap
An Exponential Lower Bound for Linearly-Realizable MDPs with Constant Suboptimality Gap
Yuanhao Wang
Ruosong Wang
Sham Kakade
OffRL
39
43
0
23 Mar 2021
Return-Based Contrastive Representation Learning for Reinforcement
  Learning
Return-Based Contrastive Representation Learning for Reinforcement Learning
Guoqing Liu
Chuheng Zhang
Li Zhao
Tao Qin
Jinhua Zhu
Jian Li
Nenghai Yu
Tie-Yan Liu
SSL
OffRL
19
47
0
22 Feb 2021
Model-Invariant State Abstractions for Model-Based Reinforcement
  Learning
Model-Invariant State Abstractions for Model-Based Reinforcement Learning
Manan Tomar
Amy Zhang
Roberto Calandra
Matthew E. Taylor
Joelle Pineau
24
24
0
19 Feb 2021
Model-free Representation Learning and Exploration in Low-rank MDPs
Model-free Representation Learning and Exploration in Low-rank MDPs
Aditya Modi
Jinglin Chen
A. Krishnamurthy
Nan Jiang
Alekh Agarwal
OffRL
102
78
0
14 Feb 2021
Improved Variance-Aware Confidence Sets for Linear Bandits and Linear
  Mixture MDP
Improved Variance-Aware Confidence Sets for Linear Bandits and Linear Mixture MDP
Zihan Zhang
Jiaqi Yang
Xiangyang Ji
S. Du
71
36
0
29 Jan 2021
Randomized Value Functions via Posterior State-Abstraction Sampling
Randomized Value Functions via Posterior State-Abstraction Sampling
Dilip Arumugam
Benjamin Van Roy
OffRL
31
7
0
05 Oct 2020
Contrastive learning, multi-view redundancy, and linear models
Contrastive learning, multi-view redundancy, and linear models
Christopher Tosh
A. Krishnamurthy
Daniel J. Hsu
SSL
35
163
0
24 Aug 2020
Provably Efficient Reward-Agnostic Navigation with Linear Value
  Iteration
Provably Efficient Reward-Agnostic Navigation with Linear Value Iteration
Andrea Zanette
A. Lazaric
Mykel J. Kochenderfer
Emma Brunskill
20
64
0
18 Aug 2020
FLAMBE: Structural Complexity and Representation Learning of Low Rank
  MDPs
FLAMBE: Structural Complexity and Representation Learning of Low Rank MDPs
Alekh Agarwal
Sham Kakade
A. Krishnamurthy
Wen Sun
OffRL
41
223
0
18 Jun 2020
$Q$-learning with Logarithmic Regret
QQQ-learning with Logarithmic Regret
Kunhe Yang
Lin F. Yang
S. Du
43
59
0
16 Jun 2020
Preference-based Reinforcement Learning with Finite-Time Guarantees
Preference-based Reinforcement Learning with Finite-Time Guarantees
Yichong Xu
Ruosong Wang
Lin F. Yang
Aarti Singh
A. Dubrawski
23
53
0
16 Jun 2020
Deep Reinforcement and InfoMax Learning
Deep Reinforcement and InfoMax Learning
Bogdan Mazoure
Rémi Tachet des Combes
T. Doan
Philip Bachman
R. Devon Hjelm
AI4CE
25
108
0
12 Jun 2020
12
Next