Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2502.06491
Cited By
Model-Based Offline Reinforcement Learning with Reliability-Guaranteed Sequence Modeling
10 February 2025
Shenghong He
OffRL
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Model-Based Offline Reinforcement Learning with Reliability-Guaranteed Sequence Modeling"
34 / 34 papers shown
Title
Offline Imitation Learning with Model-based Reverse Augmentation
Jie-Jing Shao
Hao-Sen Shi
Lan-Zhe Guo
Yu-Feng Li
OffRL
45
5
0
18 Jun 2024
DiffStitch: Boosting Offline Reinforcement Learning with Diffusion-based Trajectory Stitching
Guanghe Li
Yixiang Shan
Zhengbang Zhu
Ting Long
Weinan Zhang
OffRL
44
12
0
04 Feb 2024
Conditional Variational Autoencoder for Sign Language Translation with Cross-Modal Alignment
Rui Zhao
Liang Zhang
Biao Fu
Cong Hu
Jinsong Su
Yidong Chen
SLR
67
13
0
25 Dec 2023
Data-Efficient Task Generalization via Probabilistic Model-based Meta Reinforcement Learning
Arjun Bhardwaj
Jonas Rothfuss
Bhavya Sukhija
Yarden As
Marco Hutter
Stelian Coros
Andreas Krause
44
5
0
13 Nov 2023
ACT: Empowering Decision Transformer with Dynamic Programming via Advantage Conditioning
Chenxiao Gao
Chenyang Wu
Mingjun Cao
Rui Kong
Zongzhang Zhang
Yang Yu
OffRL
55
14
0
12 Sep 2023
Benchmarking Large Language Models in Retrieval-Augmented Generation
Jiawei Chen
Hongyu Lin
Xianpei Han
Le Sun
3DV
RALM
29
282
0
04 Sep 2023
Waypoint Transformer: Reinforcement Learning via Supervised Learning with Intermediate Targets
Anirudhan Badrinath
Yannis Flet-Berliac
Allen Nie
Emma Brunskill
OffRL
54
17
0
24 Jun 2023
Optimal Exploration for Model-Based RL in Nonlinear Systems
Andrew Wagenmaker
Guanya Shi
Kevin Jamieson
52
15
0
15 Jun 2023
Double Pessimism is Provably Efficient for Distributionally Robust Offline Reinforcement Learning: Generic Algorithm and Robust Partial Coverage
Jose H. Blanchet
Miao Lu
Tong Zhang
Han Zhong
OffRL
78
32
0
16 May 2023
Uncertainty-driven Trajectory Truncation for Data Augmentation in Offline Reinforcement Learning
Junjie Zhang
Jiafei Lyu
Xiaoteng Ma
Jiangpeng Yan
Jun Yang
Le Wan
Xiu Li
OffRL
50
6
0
10 Apr 2023
Synthetic Experience Replay
Cong Lu
Philip J. Ball
Yee Whye Teh
Jack Parker-Holder
OffRL
107
73
0
12 Mar 2023
Model-Based Uncertainty in Value Functions
Carlos E. Luis
A. Bottero
Julia Vinogradska
Felix Berkenkamp
Jan Peters
49
15
0
24 Feb 2023
Is Model Ensemble Necessary? Model-based RL via a Single Model with Lipschitz Regularized Value Function
Ruijie Zheng
Xiyao Wang
Huazhe Xu
Furong Huang
63
14
0
02 Feb 2023
Q-learning Decision Transformer: Leveraging Dynamic Programming for Conditional Sequence Modelling in Offline RL
Taku Yamagata
Ahmed Khalil
Raúl Santos-Rodríguez
OffRL
172
75
0
08 Sep 2022
Double Check Your State Before Trusting It: Confidence-Aware Bidirectional Offline Model-Based Imagination
Jiafei Lyu
Xiu Li
Zongqing Lu
OffRL
44
25
0
16 Jun 2022
Bridging the Data Gap between Training and Inference for Unsupervised Neural Machine Translation
Zhiwei He
Xing Wang
Rui Wang
Shuming Shi
Zhaopeng Tu
54
12
0
16 Mar 2022
UMBRELLA: Uncertainty-Aware Model-Based Offline Reinforcement Learning Leveraging Planning
Christopher P. Diehl
Timo Sievernich
Martin Krüger
F. Hoffmann
Torsten Bertram
OffRL
69
26
0
22 Nov 2021
Offline Reinforcement Learning with Implicit Q-Learning
Ilya Kostrikov
Ashvin Nair
Sergey Levine
OffRL
249
874
0
12 Oct 2021
Offline Reinforcement Learning with Reverse Model-based Imagination
Jianhao Wang
Wenzhe Li
Haozhe Jiang
Guangxiang Zhu
Siyuan Li
Chongjie Zhang
OffRL
145
61
0
01 Oct 2021
Offline Reinforcement Learning as One Big Sequence Modeling Problem
Michael Janner
Qiyang Li
Sergey Levine
OffRL
95
665
0
03 Jun 2021
Decision Transformer: Reinforcement Learning via Sequence Modeling
Lili Chen
Kevin Lu
Aravind Rajeswaran
Kimin Lee
Aditya Grover
Michael Laskin
Pieter Abbeel
A. Srinivas
Igor Mordatch
OffRL
77
1,608
0
02 Jun 2021
Model-Based Offline Planning with Trajectory Pruning
Xianyuan Zhan
Xiangyu Zhu
Haoran Xu
OffRL
65
36
0
16 May 2021
COMBO: Conservative Offline Model-Based Policy Optimization
Tianhe Yu
Aviral Kumar
Rafael Rafailov
Aravind Rajeswaran
Sergey Levine
Chelsea Finn
OffRL
263
425
0
16 Feb 2021
Overcoming Model Bias for Robust Offline Deep Reinforcement Learning
Phillip Swazinna
Steffen Udluft
Thomas Runkler
OffRL
43
83
0
12 Aug 2020
Conservative Q-Learning for Offline Reinforcement Learning
Aviral Kumar
Aurick Zhou
George Tucker
Sergey Levine
OffRL
OnRL
94
1,780
0
08 Jun 2020
MOPO: Model-based Offline Policy Optimization
Tianhe Yu
G. Thomas
Lantao Yu
Stefano Ermon
James Zou
Sergey Levine
Chelsea Finn
Tengyu Ma
OffRL
65
759
0
27 May 2020
MOReL : Model-Based Offline Reinforcement Learning
Rahul Kidambi
Aravind Rajeswaran
Praneeth Netrapalli
Thorsten Joachims
OffRL
67
662
0
12 May 2020
D4RL: Datasets for Deep Data-Driven Reinforcement Learning
Justin Fu
Aviral Kumar
Ofir Nachum
George Tucker
Sergey Levine
GP
OffRL
177
1,338
0
15 Apr 2020
CTRL: A Conditional Transformer Language Model for Controllable Generation
N. Keskar
Bryan McCann
Lav Varshney
Caiming Xiong
R. Socher
AI4CE
99
1,239
0
11 Sep 2019
When to Trust Your Model: Model-Based Policy Optimization
Michael Janner
Justin Fu
Marvin Zhang
Sergey Levine
OffRL
55
939
0
19 Jun 2019
Off-Policy Deep Reinforcement Learning without Exploration
Scott Fujimoto
David Meger
Doina Precup
OffRL
BDL
154
1,586
0
07 Dec 2018
Behavioral Cloning from Observation
F. Torabi
Garrett A. Warnell
Peter Stone
OffRL
100
715
0
04 May 2018
Lipschitz Continuity in Model-based Reinforcement Learning
Kavosh Asadi
Dipendra Kumar Misra
Michael L. Littman
KELM
60
152
0
19 Apr 2018
Self-Correcting Models for Model-Based Reinforcement Learning
Erik Talvitie
LRM
55
94
0
19 Dec 2016
1