ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2010.13611
  4. Cited By
OPAL: Offline Primitive Discovery for Accelerating Offline Reinforcement
  Learning

OPAL: Offline Primitive Discovery for Accelerating Offline Reinforcement Learning

26 October 2020
Anurag Ajay
Aviral Kumar
Pulkit Agrawal
Sergey Levine
Ofir Nachum
    OffRL
    OnRL
ArXivPDFHTML

Papers citing "OPAL: Offline Primitive Discovery for Accelerating Offline Reinforcement Learning"

50 / 127 papers shown
Title
Inverse Dynamics Pretraining Learns Good Representations for Multitask
  Imitation
Inverse Dynamics Pretraining Learns Good Representations for Multitask Imitation
David Brandfonbrener
Ofir Nachum
Joan Bruna
AI4CE
26
21
0
26 May 2023
Future-conditioned Unsupervised Pretraining for Decision Transformer
Future-conditioned Unsupervised Pretraining for Decision Transformer
Zhihui Xie
Zichuan Lin
Deheng Ye
Qiang Fu
Wei Yang
Shuai Li
OffRL
OnRL
45
22
0
26 May 2023
Beyond Reward: Offline Preference-guided Policy Optimization
Beyond Reward: Offline Preference-guided Policy Optimization
Yachen Kang
Dingxu Shi
Jinxin Liu
Li He
Donglin Wang
OffRL
26
31
0
25 May 2023
Policy Representation via Diffusion Probability Model for Reinforcement
  Learning
Policy Representation via Diffusion Probability Model for Reinforcement Learning
Long Yang
Zhixiong Huang
Fenghao Lei
Yucun Zhong
Yiming Yang
Cong Fang
Shiting Wen
Binbin Zhou
Zhouchen Lin
DiffM
28
40
0
22 May 2023
CRISP: Curriculum inducing Primitive Informed Subgoal Prediction
CRISP: Curriculum inducing Primitive Informed Subgoal Prediction
Utsav Singh
Vinay P. Namboodiri
31
3
0
07 Apr 2023
Boosting Reinforcement Learning and Planning with Demonstrations: A
  Survey
Boosting Reinforcement Learning and Planning with Demonstrations: A Survey
Tongzhou Mu
H. Su
OffRL
35
1
0
23 Mar 2023
Cal-QL: Calibrated Offline RL Pre-Training for Efficient Online
  Fine-Tuning
Cal-QL: Calibrated Offline RL Pre-Training for Efficient Online Fine-Tuning
Mitsuhiko Nakamoto
Yuexiang Zhai
Anika Singh
Max Sobol Mark
Yi Ma
Chelsea Finn
Aviral Kumar
Sergey Levine
OffRL
OnRL
112
108
0
09 Mar 2023
Foundation Models for Decision Making: Problems, Methods, and
  Opportunities
Foundation Models for Decision Making: Problems, Methods, and Opportunities
Sherry Yang
Ofir Nachum
Yilun Du
Jason W. Wei
Pieter Abbeel
Dale Schuurmans
LM&Ro
OffRL
LRM
AI4CE
92
155
0
07 Mar 2023
Hierarchical Reinforcement Learning in Complex 3D Environments
Hierarchical Reinforcement Learning in Complex 3D Environments
Bernardo Avila-Pires
Feryal M. P. Behbahani
Hubert Soyer
Kyriacos Nikiforou
Thomas Keck
Satinder Singh
OffRL
16
0
0
28 Feb 2023
Predictable MDP Abstraction for Unsupervised Model-Based RL
Predictable MDP Abstraction for Unsupervised Model-Based RL
Seohong Park
Sergey Levine
16
9
0
08 Feb 2023
Skill Decision Transformer
Skill Decision Transformer
Shyam Sudhakaran
S. Risi
OffRL
21
5
0
31 Jan 2023
Cross-Domain Transfer via Semantic Skill Imitation
Cross-Domain Transfer via Semantic Skill Imitation
Karl Pertsch
Ruta Desai
Vikash Kumar
Franziska Meier
Joseph J. Lim
Dhruv Batra
Akshara Rai
LM&Ro
13
18
0
14 Dec 2022
Learning Options via Compression
Learning Options via Compression
Yiding Jiang
E. Liu
Benjamin Eysenbach
Zico Kolter
Chelsea Finn
OffRL
25
13
0
08 Dec 2022
Flow to Control: Offline Reinforcement Learning with Lossless Primitive
  Discovery
Flow to Control: Offline Reinforcement Learning with Lossless Primitive Discovery
Yiqin Yang
Haotian Hu
Wenzhe Li
Siyuan Li
Jun Yang
Qianchuan Zhao
Chongjie Zhang
OffRL
28
9
0
02 Dec 2022
A Unified Algorithm Framework for Unsupervised Discovery of Skills based
  on Determinantal Point Process
A Unified Algorithm Framework for Unsupervised Discovery of Skills based on Determinantal Point Process
Jiayu Chen
Vaneet Aggarwal
Tian-Shing Lan
16
1
0
01 Dec 2022
Is Conditional Generative Modeling all you need for Decision-Making?
Is Conditional Generative Modeling all you need for Decision-Making?
Anurag Ajay
Yilun Du
Abhi Gupta
J. Tenenbaum
Tommi Jaakkola
Pulkit Agrawal
DiffM
47
360
0
28 Nov 2022
A System for Morphology-Task Generalization via Unified Representation
  and Behavior Distillation
A System for Morphology-Task Generalization via Unified Representation and Behavior Distillation
Hiroki Furuta
Yusuke Iwasawa
Yutaka Matsuo
S. Gu
20
14
0
25 Nov 2022
SkillS: Adaptive Skill Sequencing for Efficient Temporally-Extended
  Exploration
SkillS: Adaptive Skill Sequencing for Efficient Temporally-Extended Exploration
Giulia Vezzani
Dhruva Tirumala
Markus Wulfmeier
Dushyant Rao
A. Abdolmaleki
...
Tim Hertweck
Thomas Lampe
Fereshteh Sadeghi
N. Heess
Martin Riedmiller
OffRL
35
6
0
24 Nov 2022
Robot Learning on the Job: Human-in-the-Loop Autonomy and Learning
  During Deployment
Robot Learning on the Job: Human-in-the-Loop Autonomy and Learning During Deployment
Huihan Liu
Soroush Nasiriany
Lance Zhang
Zhiyao Bao
Yuke Zhu
38
56
0
15 Nov 2022
Pretraining in Deep Reinforcement Learning: A Survey
Pretraining in Deep Reinforcement Learning: A Survey
Zhihui Xie
Zichuan Lin
Junyou Li
Shuai Li
Deheng Ye
OffRL
OnRL
AI4CE
26
23
0
08 Nov 2022
Leveraging Demonstrations with Latent Space Priors
Leveraging Demonstrations with Latent Space Priors
Jonas Gehring
Deepak Gopinath
Jungdam Won
Andreas Krause
Gabriel Synnaeve
Nicolas Usunier
41
4
0
26 Oct 2022
Adaptive Behavior Cloning Regularization for Stable Offline-to-Online
  Reinforcement Learning
Adaptive Behavior Cloning Regularization for Stable Offline-to-Online Reinforcement Learning
Yi Zhao
Rinu Boney
Alexander Ilin
Arno Solin
Joni Pajarinen
OffRL
OnRL
26
39
0
25 Oct 2022
Dichotomy of Control: Separating What You Can Control from What You
  Cannot
Dichotomy of Control: Separating What You Can Control from What You Cannot
Mengjiao Yang
Dale Schuurmans
Pieter Abbeel
Ofir Nachum
OffRL
22
42
0
24 Oct 2022
LEAGUE: Guided Skill Learning and Abstraction for Long-Horizon
  Manipulation
LEAGUE: Guided Skill Learning and Abstraction for Long-Horizon Manipulation
Shuo Cheng
Danfei Xu
54
37
0
23 Oct 2022
H-SAUR: Hypothesize, Simulate, Act, Update, and Repeat for Understanding
  Object Articulations from Interactions
H-SAUR: Hypothesize, Simulate, Act, Update, and Repeat for Understanding Object Articulations from Interactions
Keita Ota
H. Tung
Kevin A. Smith
A. Cherian
Tim K. Marks
Alan Sullivan
Asako Kanezaki
J. Tenenbaum
25
3
0
22 Oct 2022
STAP: Sequencing Task-Agnostic Policies
STAP: Sequencing Task-Agnostic Policies
Christopher Agia
Toki Migimatsu
Jiajun Wu
Jeannette Bohg
43
19
0
21 Oct 2022
Learning and Retrieval from Prior Data for Skill-based Imitation
  Learning
Learning and Retrieval from Prior Data for Skill-based Imitation Learning
Soroush Nasiriany
Tian Gao
Ajay Mandlekar
Yuke Zhu
SSL
41
47
0
20 Oct 2022
ASPiRe:Adaptive Skill Priors for Reinforcement Learning
ASPiRe:Adaptive Skill Priors for Reinforcement Learning
Mengda Xu
Manuela Veloso
Shuran Song
CLL
OffRL
19
10
0
30 Sep 2022
DCE: Offline Reinforcement Learning With Double Conservative Estimates
DCE: Offline Reinforcement Learning With Double Conservative Estimates
Chen Zhao
K. Huang
Chun yuan
OffRL
35
1
0
27 Sep 2022
Latent Plans for Task-Agnostic Offline Reinforcement Learning
Latent Plans for Task-Agnostic Offline Reinforcement Learning
Erick Rosete-Beas
Oier Mees
Gabriel Kalweit
Joschka Boedecker
Wolfram Burgard
OffRL
30
81
0
19 Sep 2022
MO2: Model-Based Offline Options
MO2: Model-Based Offline Options
Sasha Salter
Markus Wulfmeier
Dhruva Tirumala
N. Heess
Martin Riedmiller
R. Hadsell
Dushyant Rao
OffRL
16
13
0
05 Sep 2022
Spectral Decomposition Representation for Reinforcement Learning
Spectral Decomposition Representation for Reinforcement Learning
Tongzheng Ren
Tianjun Zhang
Lisa Lee
Joseph E. Gonzalez
Dale Schuurmans
Bo Dai
OffRL
40
27
0
19 Aug 2022
Learning Dynamic Manipulation Skills from Haptic-Play
Learning Dynamic Manipulation Skills from Haptic-Play
Taeyoon Lee
D. Sung
Kyoung-Whan Choi
Choong-Keun Lee
Changwoo Park
Keunjun Choi
37
3
0
28 Jul 2022
Watch and Match: Supercharging Imitation with Regularized Optimal
  Transport
Watch and Match: Supercharging Imitation with Regularized Optimal Transport
Siddhant Haldar
Vaibhav Mathur
Denis Yarats
Lerrel Pinto
48
62
0
30 Jun 2022
Learning Neuro-Symbolic Skills for Bilevel Planning
Learning Neuro-Symbolic Skills for Bilevel Planning
Tom Silver
Ashay Athalye
J. Tenenbaum
Tomas Lozano-Perez
L. Kaelbling
37
59
0
21 Jun 2022
Beyond Rewards: a Hierarchical Perspective on Offline Multiagent
  Behavioral Analysis
Beyond Rewards: a Hierarchical Perspective on Offline Multiagent Behavioral Analysis
Shayegan Omidshafiei
A. Kapishnikov
Yannick Assogba
Lucas Dixon
Been Kim
OffRL
36
5
0
17 Jun 2022
Mildly Conservative Q-Learning for Offline Reinforcement Learning
Mildly Conservative Q-Learning for Offline Reinforcement Learning
Jiafei Lyu
Xiaoteng Ma
Xiu Li
Zongqing Lu
OffRL
34
101
0
09 Jun 2022
A Mixture-of-Expert Approach to RL-based Dialogue Management
A Mixture-of-Expert Approach to RL-based Dialogue Management
Yinlam Chow
Azamat Tulepbergenov
Ofir Nachum
Moonkyung Ryu
Mohammad Ghavamzadeh
Craig Boutilier
MoE
17
14
0
31 May 2022
Why So Pessimistic? Estimating Uncertainties for Offline RL through
  Ensembles, and Why Their Independence Matters
Why So Pessimistic? Estimating Uncertainties for Offline RL through Ensembles, and Why Their Independence Matters
Seyed Kamyar Seyed Ghasemipour
S. Gu
Ofir Nachum
OffRL
25
69
0
27 May 2022
SFP: State-free Priors for Exploration in Off-Policy Reinforcement
  Learning
SFP: State-free Priors for Exploration in Off-Policy Reinforcement Learning
Marco Bagatella
Sammy Christen
Otmar Hilliges
OffRL
24
5
0
26 May 2022
Hierarchical Planning Through Goal-Conditioned Offline Reinforcement
  Learning
Hierarchical Planning Through Goal-Conditioned Offline Reinforcement Learning
Jinning Li
Chen Tang
M. Tomizuka
Wei Zhan
OffRL
53
57
0
24 May 2022
Efficient Reinforcement Learning from Demonstration Using Local Ensemble
  and Reparameterization with Split and Merge of Expert Policies
Efficient Reinforcement Learning from Demonstration Using Local Ensemble and Reparameterization with Split and Merge of Expert Policies
Yu Wang
Fang Liu
16
0
0
23 May 2022
Chain of Thought Imitation with Procedure Cloning
Chain of Thought Imitation with Procedure Cloning
Mengjiao Yang
Dale Schuurmans
Pieter Abbeel
Ofir Nachum
OffRL
30
29
0
22 May 2022
User-Interactive Offline Reinforcement Learning
User-Interactive Offline Reinforcement Learning
Phillip Swazinna
Steffen Udluft
Thomas Runkler
OffRL
25
11
0
21 May 2022
No More Pesky Hyperparameters: Offline Hyperparameter Tuning for RL
No More Pesky Hyperparameters: Offline Hyperparameter Tuning for RL
Han Wang
Archit Sakhadeo
Adam White
James Bell
Vincent Liu
Xutong Zhao
Puer Liu
Tadashi Kozuno
Alona Fyshe
Martha White
OffRL
OnRL
17
7
0
18 May 2022
Skill-based Meta-Reinforcement Learning
Skill-based Meta-Reinforcement Learning
Taewook Nam
Shao-Hua Sun
Karl Pertsch
Sung Ju Hwang
Joseph J. Lim
OffRL
26
45
0
25 Apr 2022
Unsupervised Learning of Temporal Abstractions with Slot-based
  Transformers
Unsupervised Learning of Temporal Abstractions with Slot-based Transformers
Anand Gopalakrishnan
Kazuki Irie
Jürgen Schmidhuber
Sjoerd van Steenkiste
OffRL
26
16
0
25 Mar 2022
Latent-Variable Advantage-Weighted Policy Optimization for Offline RL
Latent-Variable Advantage-Weighted Policy Optimization for Offline RL
Xi Chen
Ali Ghadirzadeh
Tianhe Yu
Yuan Gao
Jianhao Wang
Wenzhe Li
Bin Liang
Chelsea Finn
Chongjie Zhang
OffRL
32
14
0
16 Mar 2022
Pessimistic Bootstrapping for Uncertainty-Driven Offline Reinforcement
  Learning
Pessimistic Bootstrapping for Uncertainty-Driven Offline Reinforcement Learning
Chenjia Bai
Lingxiao Wang
Zhuoran Yang
Zhihong Deng
Animesh Garg
Peng Liu
Zhaoran Wang
OffRL
31
132
0
23 Feb 2022
SAFER: Data-Efficient and Safe Reinforcement Learning via Skill
  Acquisition
SAFER: Data-Efficient and Safe Reinforcement Learning via Skill Acquisition
Dylan Slack
Yinlam Chow
Bo Dai
Nevan Wichers
OffRL
24
7
0
10 Feb 2022
Previous
123
Next