ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2502.06491
  4. Cited By
Model-Based Offline Reinforcement Learning with Reliability-Guaranteed Sequence Modeling

Model-Based Offline Reinforcement Learning with Reliability-Guaranteed Sequence Modeling

10 February 2025
Shenghong He
    OffRL
ArXivPDFHTML

Papers citing "Model-Based Offline Reinforcement Learning with Reliability-Guaranteed Sequence Modeling"

34 / 34 papers shown
Title
Offline Imitation Learning with Model-based Reverse Augmentation
Offline Imitation Learning with Model-based Reverse Augmentation
Jie-Jing Shao
Hao-Sen Shi
Lan-Zhe Guo
Yu-Feng Li
OffRL
45
5
0
18 Jun 2024
DiffStitch: Boosting Offline Reinforcement Learning with Diffusion-based
  Trajectory Stitching
DiffStitch: Boosting Offline Reinforcement Learning with Diffusion-based Trajectory Stitching
Guanghe Li
Yixiang Shan
Zhengbang Zhu
Ting Long
Weinan Zhang
OffRL
44
12
0
04 Feb 2024
Conditional Variational Autoencoder for Sign Language Translation with
  Cross-Modal Alignment
Conditional Variational Autoencoder for Sign Language Translation with Cross-Modal Alignment
Rui Zhao
Liang Zhang
Biao Fu
Cong Hu
Jinsong Su
Yidong Chen
SLR
67
13
0
25 Dec 2023
Data-Efficient Task Generalization via Probabilistic Model-based Meta
  Reinforcement Learning
Data-Efficient Task Generalization via Probabilistic Model-based Meta Reinforcement Learning
Arjun Bhardwaj
Jonas Rothfuss
Bhavya Sukhija
Yarden As
Marco Hutter
Stelian Coros
Andreas Krause
44
5
0
13 Nov 2023
ACT: Empowering Decision Transformer with Dynamic Programming via
  Advantage Conditioning
ACT: Empowering Decision Transformer with Dynamic Programming via Advantage Conditioning
Chenxiao Gao
Chenyang Wu
Mingjun Cao
Rui Kong
Zongzhang Zhang
Yang Yu
OffRL
55
14
0
12 Sep 2023
Benchmarking Large Language Models in Retrieval-Augmented Generation
Benchmarking Large Language Models in Retrieval-Augmented Generation
Jiawei Chen
Hongyu Lin
Xianpei Han
Le Sun
3DV
RALM
29
282
0
04 Sep 2023
Waypoint Transformer: Reinforcement Learning via Supervised Learning
  with Intermediate Targets
Waypoint Transformer: Reinforcement Learning via Supervised Learning with Intermediate Targets
Anirudhan Badrinath
Yannis Flet-Berliac
Allen Nie
Emma Brunskill
OffRL
54
17
0
24 Jun 2023
Optimal Exploration for Model-Based RL in Nonlinear Systems
Optimal Exploration for Model-Based RL in Nonlinear Systems
Andrew Wagenmaker
Guanya Shi
Kevin Jamieson
52
15
0
15 Jun 2023
Double Pessimism is Provably Efficient for Distributionally Robust
  Offline Reinforcement Learning: Generic Algorithm and Robust Partial Coverage
Double Pessimism is Provably Efficient for Distributionally Robust Offline Reinforcement Learning: Generic Algorithm and Robust Partial Coverage
Jose H. Blanchet
Miao Lu
Tong Zhang
Han Zhong
OffRL
78
32
0
16 May 2023
Uncertainty-driven Trajectory Truncation for Data Augmentation in
  Offline Reinforcement Learning
Uncertainty-driven Trajectory Truncation for Data Augmentation in Offline Reinforcement Learning
Junjie Zhang
Jiafei Lyu
Xiaoteng Ma
Jiangpeng Yan
Jun Yang
Le Wan
Xiu Li
OffRL
50
6
0
10 Apr 2023
Synthetic Experience Replay
Synthetic Experience Replay
Cong Lu
Philip J. Ball
Yee Whye Teh
Jack Parker-Holder
OffRL
107
73
0
12 Mar 2023
Model-Based Uncertainty in Value Functions
Model-Based Uncertainty in Value Functions
Carlos E. Luis
A. Bottero
Julia Vinogradska
Felix Berkenkamp
Jan Peters
49
15
0
24 Feb 2023
Is Model Ensemble Necessary? Model-based RL via a Single Model with
  Lipschitz Regularized Value Function
Is Model Ensemble Necessary? Model-based RL via a Single Model with Lipschitz Regularized Value Function
Ruijie Zheng
Xiyao Wang
Huazhe Xu
Furong Huang
63
14
0
02 Feb 2023
Q-learning Decision Transformer: Leveraging Dynamic Programming for
  Conditional Sequence Modelling in Offline RL
Q-learning Decision Transformer: Leveraging Dynamic Programming for Conditional Sequence Modelling in Offline RL
Taku Yamagata
Ahmed Khalil
Raúl Santos-Rodríguez
OffRL
172
75
0
08 Sep 2022
Double Check Your State Before Trusting It: Confidence-Aware
  Bidirectional Offline Model-Based Imagination
Double Check Your State Before Trusting It: Confidence-Aware Bidirectional Offline Model-Based Imagination
Jiafei Lyu
Xiu Li
Zongqing Lu
OffRL
44
25
0
16 Jun 2022
Bridging the Data Gap between Training and Inference for Unsupervised
  Neural Machine Translation
Bridging the Data Gap between Training and Inference for Unsupervised Neural Machine Translation
Zhiwei He
Xing Wang
Rui Wang
Shuming Shi
Zhaopeng Tu
54
12
0
16 Mar 2022
UMBRELLA: Uncertainty-Aware Model-Based Offline Reinforcement Learning
  Leveraging Planning
UMBRELLA: Uncertainty-Aware Model-Based Offline Reinforcement Learning Leveraging Planning
Christopher P. Diehl
Timo Sievernich
Martin Krüger
F. Hoffmann
Torsten Bertram
OffRL
69
26
0
22 Nov 2021
Offline Reinforcement Learning with Implicit Q-Learning
Offline Reinforcement Learning with Implicit Q-Learning
Ilya Kostrikov
Ashvin Nair
Sergey Levine
OffRL
249
874
0
12 Oct 2021
Offline Reinforcement Learning with Reverse Model-based Imagination
Offline Reinforcement Learning with Reverse Model-based Imagination
Jianhao Wang
Wenzhe Li
Haozhe Jiang
Guangxiang Zhu
Siyuan Li
Chongjie Zhang
OffRL
145
61
0
01 Oct 2021
Offline Reinforcement Learning as One Big Sequence Modeling Problem
Offline Reinforcement Learning as One Big Sequence Modeling Problem
Michael Janner
Qiyang Li
Sergey Levine
OffRL
95
665
0
03 Jun 2021
Decision Transformer: Reinforcement Learning via Sequence Modeling
Decision Transformer: Reinforcement Learning via Sequence Modeling
Lili Chen
Kevin Lu
Aravind Rajeswaran
Kimin Lee
Aditya Grover
Michael Laskin
Pieter Abbeel
A. Srinivas
Igor Mordatch
OffRL
77
1,608
0
02 Jun 2021
Model-Based Offline Planning with Trajectory Pruning
Model-Based Offline Planning with Trajectory Pruning
Xianyuan Zhan
Xiangyu Zhu
Haoran Xu
OffRL
65
36
0
16 May 2021
COMBO: Conservative Offline Model-Based Policy Optimization
COMBO: Conservative Offline Model-Based Policy Optimization
Tianhe Yu
Aviral Kumar
Rafael Rafailov
Aravind Rajeswaran
Sergey Levine
Chelsea Finn
OffRL
263
425
0
16 Feb 2021
Overcoming Model Bias for Robust Offline Deep Reinforcement Learning
Overcoming Model Bias for Robust Offline Deep Reinforcement Learning
Phillip Swazinna
Steffen Udluft
Thomas Runkler
OffRL
43
83
0
12 Aug 2020
Conservative Q-Learning for Offline Reinforcement Learning
Conservative Q-Learning for Offline Reinforcement Learning
Aviral Kumar
Aurick Zhou
George Tucker
Sergey Levine
OffRL
OnRL
94
1,780
0
08 Jun 2020
MOPO: Model-based Offline Policy Optimization
MOPO: Model-based Offline Policy Optimization
Tianhe Yu
G. Thomas
Lantao Yu
Stefano Ermon
James Zou
Sergey Levine
Chelsea Finn
Tengyu Ma
OffRL
65
759
0
27 May 2020
MOReL : Model-Based Offline Reinforcement Learning
MOReL : Model-Based Offline Reinforcement Learning
Rahul Kidambi
Aravind Rajeswaran
Praneeth Netrapalli
Thorsten Joachims
OffRL
67
662
0
12 May 2020
D4RL: Datasets for Deep Data-Driven Reinforcement Learning
D4RL: Datasets for Deep Data-Driven Reinforcement Learning
Justin Fu
Aviral Kumar
Ofir Nachum
George Tucker
Sergey Levine
GP
OffRL
177
1,338
0
15 Apr 2020
CTRL: A Conditional Transformer Language Model for Controllable
  Generation
CTRL: A Conditional Transformer Language Model for Controllable Generation
N. Keskar
Bryan McCann
Lav Varshney
Caiming Xiong
R. Socher
AI4CE
99
1,239
0
11 Sep 2019
When to Trust Your Model: Model-Based Policy Optimization
When to Trust Your Model: Model-Based Policy Optimization
Michael Janner
Justin Fu
Marvin Zhang
Sergey Levine
OffRL
55
939
0
19 Jun 2019
Off-Policy Deep Reinforcement Learning without Exploration
Off-Policy Deep Reinforcement Learning without Exploration
Scott Fujimoto
David Meger
Doina Precup
OffRL
BDL
154
1,586
0
07 Dec 2018
Behavioral Cloning from Observation
Behavioral Cloning from Observation
F. Torabi
Garrett A. Warnell
Peter Stone
OffRL
100
715
0
04 May 2018
Lipschitz Continuity in Model-based Reinforcement Learning
Lipschitz Continuity in Model-based Reinforcement Learning
Kavosh Asadi
Dipendra Kumar Misra
Michael L. Littman
KELM
60
152
0
19 Apr 2018
Self-Correcting Models for Model-Based Reinforcement Learning
Self-Correcting Models for Model-Based Reinforcement Learning
Erik Talvitie
LRM
55
94
0
19 Dec 2016
1