ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1906.08253
  4. Cited By
When to Trust Your Model: Model-Based Policy Optimization

When to Trust Your Model: Model-Based Policy Optimization

19 June 2019
Michael Janner
Justin Fu
Marvin Zhang
Sergey Levine
    OffRL
ArXivPDFHTML

Papers citing "When to Trust Your Model: Model-Based Policy Optimization"

50 / 242 papers shown
Title
Model-based Trajectory Stitching for Improved Offline Reinforcement
  Learning
Model-based Trajectory Stitching for Improved Offline Reinforcement Learning
Charles A. Hepburn
Giovanni Montana
OffRL
34
13
0
21 Nov 2022
Q-Ensemble for Offline RL: Don't Scale the Ensemble, Scale the Batch
  Size
Q-Ensemble for Offline RL: Don't Scale the Ensemble, Scale the Batch Size
Alexander Nikulin
Vladislav Kurenkov
Denis Tarasov
Dmitry Akimov
Sergey Kolesnikov
OffRL
36
14
0
20 Nov 2022
Model Based Residual Policy Learning with Applications to Antenna
  Control
Model Based Residual Policy Learning with Applications to Antenna Control
Viktor Eriksson Mollerstedt
Alessio Russo
Maxime Bouton
OffRL
31
3
0
16 Nov 2022
Learning Modular Robot Locomotion from Demonstrations
Learning Modular Robot Locomotion from Demonstrations
Julian Whitman
Howie Choset
31
1
0
31 Oct 2022
Adaptive Behavior Cloning Regularization for Stable Offline-to-Online
  Reinforcement Learning
Adaptive Behavior Cloning Regularization for Stable Offline-to-Online Reinforcement Learning
Yi Zhao
Rinu Boney
Alexander Ilin
Arno Solin
Joni Pajarinen
OffRL
OnRL
28
39
0
25 Oct 2022
On Many-Actions Policy Gradient
On Many-Actions Policy Gradient
Michal Nauman
Marek Cygan
19
0
0
24 Oct 2022
Deep Reinforcement Learning for Stabilization of Large-scale
  Probabilistic Boolean Networks
Deep Reinforcement Learning for Stabilization of Large-scale Probabilistic Boolean Networks
S. Moschoyiannis
Evangelos Chatzaroulas
Vytenis Sliogeris
Yuhu Wu
BDL
OffRL
AI4CE
24
7
0
21 Oct 2022
MoCoDA: Model-based Counterfactual Data Augmentation
MoCoDA: Model-based Counterfactual Data Augmentation
Silviu Pitis
Elliot Creager
Ajay Mandlekar
Animesh Garg
OffRL
48
33
0
20 Oct 2022
Safe Policy Improvement in Constrained Markov Decision Processes
Safe Policy Improvement in Constrained Markov Decision Processes
Luigi Berducci
Radu Grosu
OffRL
36
2
0
20 Oct 2022
Integrated Decision and Control for High-Level Automated Vehicles by
  Mixed Policy Gradient and Its Experiment Verification
Integrated Decision and Control for High-Level Automated Vehicles by Mixed Policy Gradient and Its Experiment Verification
Yang Guan
Liye Tang
Chuanxiao Li
Shengbo Eben Li
Yangang Ren
Junqing Wei
Bo Zhang
Ke Li
23
0
0
19 Oct 2022
Planning for Sample Efficient Imitation Learning
Planning for Sample Efficient Imitation Learning
Zhao-Heng Yin
Weirui Ye
Qifeng Chen
Yang Gao
OffRL
33
21
0
18 Oct 2022
On Uncertainty in Deep State Space Models for Model-Based Reinforcement
  Learning
On Uncertainty in Deep State Space Models for Model-Based Reinforcement Learning
P. Becker
Gerhard Neumann
30
9
0
17 Oct 2022
Limited or Biased: Modeling Sub-Rational Human Investors in Financial
  Markets
Limited or Biased: Modeling Sub-Rational Human Investors in Financial Markets
Penghang Liu
Kshama Dwarakanath
Svitlana Vyetrenko
Tucker Balch
AIFin
34
5
0
16 Oct 2022
When to Update Your Model: Constrained Model-based Reinforcement
  Learning
When to Update Your Model: Constrained Model-based Reinforcement Learning
Tianying Ji
Yu-Juan Luo
Gang Hua
Mingxuan Jing
Fengxiang He
Wen-bing Huang
24
18
0
15 Oct 2022
Visual Reinforcement Learning with Self-Supervised 3D Representations
Visual Reinforcement Learning with Self-Supervised 3D Representations
Yanjie Ze
Nicklas Hansen
Yinbo Chen
Mohit Jain
Xiaolong Wang
SSL
32
49
0
13 Oct 2022
ControlVAE: Model-Based Learning of Generative Controllers for
  Physics-Based Characters
ControlVAE: Model-Based Learning of Generative Controllers for Physics-Based Characters
Heyuan Yao
Zhenhua Song
Bin Chen
Libin Liu
DRL
VGen
16
41
0
12 Oct 2022
Exploration via Planning for Information about the Optimal Trajectory
Exploration via Planning for Information about the Optimal Trajectory
Viraj Mehta
I. Char
J. Abbate
R. Conlin
M. Boyer
Stefano Ermon
J. Schneider
Willie Neiswanger
OffRL
27
6
0
06 Oct 2022
S2P: State-conditioned Image Synthesis for Data Augmentation in Offline
  Reinforcement Learning
S2P: State-conditioned Image Synthesis for Data Augmentation in Offline Reinforcement Learning
Daesol Cho
D. Shim
H. J. Kim
OffRL
45
11
0
30 Sep 2022
Mastering the Unsupervised Reinforcement Learning Benchmark from Pixels
Mastering the Unsupervised Reinforcement Learning Benchmark from Pixels
Sai Rajeswar
Pietro Mazzaglia
Tim Verbelen
Alexandre Piché
Bart Dhoedt
Rameswar Panda
Alexandre Lacoste
SSL
28
21
0
24 Sep 2022
Modern Machine Learning Tools for Monitoring and Control of Industrial
  Processes: A Survey
Modern Machine Learning Tools for Monitoring and Control of Industrial Processes: A Survey
R. Bhushan Gopaluni
Aditya Tulsyan
Benoît Chachuat
Biao Huang
J. M. Lee
Faraz Amjad
S. Damarla
Jong Woo Kim
Nathan P. Lawrence
AI4CE
21
38
0
22 Sep 2022
Masked Imitation Learning: Discovering Environment-Invariant Modalities
  in Multimodal Demonstrations
Masked Imitation Learning: Discovering Environment-Invariant Modalities in Multimodal Demonstrations
Yilun Hao
Ruinan Wang
Zhangjie Cao
Zihan Wang
Yuchen Cui
Dorsa Sadigh
33
2
0
16 Sep 2022
Model-based Reinforcement Learning with Multi-step Plan Value Estimation
Model-based Reinforcement Learning with Multi-step Plan Value Estimation
Hao-Chu Lin
Yihao Sun
Jiajin Zhang
Yang Yu
OffRL
37
7
0
12 Sep 2022
Q-learning Decision Transformer: Leveraging Dynamic Programming for
  Conditional Sequence Modelling in Offline RL
Q-learning Decision Transformer: Leveraging Dynamic Programming for Conditional Sequence Modelling in Offline RL
Taku Yamagata
Ahmed Khalil
Raúl Santos-Rodríguez
OffRL
160
72
0
08 Sep 2022
Variational Inference for Model-Free and Model-Based Reinforcement
  Learning
Variational Inference for Model-Free and Model-Based Reinforcement Learning
Felix Leibfried
OffRL
23
0
0
04 Sep 2022
Backward Imitation and Forward Reinforcement Learning via Bi-directional
  Model Rollouts
Backward Imitation and Forward Reinforcement Learning via Bi-directional Model Rollouts
Yuxin Pan
Fangzhen Lin
OffRL
25
3
0
04 Aug 2022
Skill-based Model-based Reinforcement Learning
Skill-based Model-based Reinforcement Learning
Lu Shi
Joseph J. Lim
Youngwoon Lee
32
45
0
15 Jul 2022
Scalable Model-based Policy Optimization for Decentralized Networked
  Systems
Scalable Model-based Policy Optimization for Decentralized Networked Systems
Yali Du
Chengdong Ma
Yuchen Liu
Runji Lin
Hao Dong
Jun Wang
Yaodong Yang
34
8
0
13 Jul 2022
Robust Reinforcement Learning in Continuous Control Tasks with
  Uncertainty Set Regularization
Robust Reinforcement Learning in Continuous Control Tasks with Uncertainty Set Regularization
Yuan Zhang
Jianhong Wang
Joschka Boedecker
38
3
0
05 Jul 2022
Masked World Models for Visual Control
Masked World Models for Visual Control
Younggyo Seo
Danijar Hafner
Hao Liu
Fangchen Liu
Stephen James
Kimin Lee
Pieter Abbeel
OffRL
93
147
0
28 Jun 2022
Generalized Policy Improvement Algorithms with Theoretically Supported
  Sample Reuse
Generalized Policy Improvement Algorithms with Theoretically Supported Sample Reuse
James Queeney
I. Paschalidis
Christos G. Cassandras
OffRL
32
2
0
28 Jun 2022
A Survey on Model-based Reinforcement Learning
A Survey on Model-based Reinforcement Learning
Fan Luo
Tian Xu
Hang Lai
Xiong-Hui Chen
Weinan Zhang
Yang Yu
OffRL
LRM
53
101
0
19 Jun 2022
Bootstrapped Transformer for Offline Reinforcement Learning
Bootstrapped Transformer for Offline Reinforcement Learning
Kerong Wang
Hanye Zhao
Xufang Luo
Kan Ren
Weinan Zhang
Dongsheng Li
OffRL
16
37
0
17 Jun 2022
Relative Policy-Transition Optimization for Fast Policy Transfer
Relative Policy-Transition Optimization for Fast Policy Transfer
Jiawei Xu
Cheng Zhou
Yizheng Zhang
Zhengyou Zhang
Lei Han
21
0
0
13 Jun 2022
Model-based Offline Imitation Learning with Non-expert Data
Model-based Offline Imitation Learning with Non-expert Data
Jeongwon Park
Lin F. Yang
OffRL
37
1
0
11 Jun 2022
Multifidelity Reinforcement Learning with Control Variates
Multifidelity Reinforcement Learning with Control Variates
Sami Khairy
Prasanna Balaprakash
OffRL
36
5
0
10 Jun 2022
Offline Policy Comparison with Confidence: Benchmarks and Baselines
Offline Policy Comparison with Confidence: Benchmarks and Baselines
Anurag Koul
Mariano Phielipp
Alan Fern
OffRL
28
0
0
22 May 2022
BATS: Best Action Trajectory Stitching
BATS: Best Action Trajectory Stitching
I. Char
Viraj Mehta
Adam R. Villaflor
John M. Dolan
J. Schneider
OffRL
32
8
0
26 Apr 2022
Mingling Foresight with Imagination: Model-Based Cooperative Multi-Agent
  Reinforcement Learning
Mingling Foresight with Imagination: Model-Based Cooperative Multi-Agent Reinforcement Learning
Zhiwei Xu
Dapeng Li
Bin Zhang
Yuan Zhan
Yunru Bai
Guoliang Fan
OffRL
27
7
0
20 Apr 2022
Gradient-Based Trajectory Optimization With Learned Dynamics
Gradient-Based Trajectory Optimization With Learned Dynamics
Bhavya Sukhija
Nathanael Kohler
Miguel Zamora
Simon Zimmermann
Sebastian Curi
Andreas Krause
Stelian Coros
30
9
0
09 Apr 2022
Revisiting Model-based Value Expansion
Revisiting Model-based Value Expansion
Daniel Palenicek
M. Lutter
Jan Peters
27
2
0
28 Mar 2022
Investigating Compounding Prediction Errors in Learned Dynamics Models
Investigating Compounding Prediction Errors in Learned Dynamics Models
Nathan Lambert
K. Pister
Roberto Calandra
AI4CE
22
27
0
17 Mar 2022
Temporal Difference Learning for Model Predictive Control
Temporal Difference Learning for Model Predictive Control
Nicklas Hansen
Xiaolong Wang
H. Su
PINN
MU
38
226
0
09 Mar 2022
Pessimistic Bootstrapping for Uncertainty-Driven Offline Reinforcement
  Learning
Pessimistic Bootstrapping for Uncertainty-Driven Offline Reinforcement Learning
Chenjia Bai
Lingxiao Wang
Zhuoran Yang
Zhihong Deng
Animesh Garg
Peng Liu
Zhaoran Wang
OffRL
40
132
0
23 Feb 2022
Saute RL: Almost Surely Safe Reinforcement Learning Using State
  Augmentation
Saute RL: Almost Surely Safe Reinforcement Learning Using State Augmentation
Aivar Sootla
Alexander I. Cowen-Rivers
Taher Jafferjee
Ziyan Wang
D. Mguni
Jun Wang
Haitham Bou-Ammar
32
54
0
14 Feb 2022
Deconstructing the Inductive Biases of Hamiltonian Neural Networks
Deconstructing the Inductive Biases of Hamiltonian Neural Networks
Nate Gruver
Marc Finzi
Samuel Stanton
A. Wilson
AI4CE
26
40
0
10 Feb 2022
Model-Based Offline Meta-Reinforcement Learning with Regularization
Model-Based Offline Meta-Reinforcement Learning with Regularization
Sen Lin
Jialin Wan
Tengyu Xu
Yingbin Liang
Junshan Zhang
OffRL
33
17
0
07 Feb 2022
Bellman Meets Hawkes: Model-Based Reinforcement Learning via Temporal
  Point Processes
Bellman Meets Hawkes: Model-Based Reinforcement Learning via Temporal Point Processes
Chao Qu
Xiaoyu Tan
Siqiao Xue
Xiaoming Shi
James Y. Zhang
Hongyuan Mei
OffRL
30
17
0
29 Jan 2022
MOORe: Model-based Offline-to-Online Reinforcement Learning
MOORe: Model-based Offline-to-Online Reinforcement Learning
Yihuan Mao
Chao Wang
Bin Wang
Chongjie Zhang
OffRL
OnRL
39
14
0
25 Jan 2022
Reinforcement Learning for Personalized Drug Discovery and Design for
  Complex Diseases: A Systems Pharmacology Perspective
Reinforcement Learning for Personalized Drug Discovery and Design for Complex Diseases: A Systems Pharmacology Perspective
Ryan K. Tan
Yang Liu
Lei Xie
44
2
0
21 Jan 2022
Automated Reinforcement Learning (AutoRL): A Survey and Open Problems
Automated Reinforcement Learning (AutoRL): A Survey and Open Problems
Jack Parker-Holder
Raghunandan Rajan
Xingyou Song
André Biedenkapp
Yingjie Miao
...
Vu-Linh Nguyen
Roberto Calandra
Aleksandra Faust
Frank Hutter
Marius Lindauer
AI4CE
33
100
0
11 Jan 2022
Previous
12345
Next