ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2110.13625
  4. Cited By
Landmark-Guided Subgoal Generation in Hierarchical Reinforcement
  Learning

Landmark-Guided Subgoal Generation in Hierarchical Reinforcement Learning

26 October 2021
Junsup Kim
Younggyo Seo
Jinwoo Shin
ArXivPDFHTML

Papers citing "Landmark-Guided Subgoal Generation in Hierarchical Reinforcement Learning"

43 / 43 papers shown
Title
OGBench: Benchmarking Offline Goal-Conditioned RL
OGBench: Benchmarking Offline Goal-Conditioned RL
Seohong Park
Kevin Frans
Benjamin Eysenbach
Sergey Levine
OffRL
87
18
0
26 Oct 2024
HG2P: Hippocampus-inspired High-reward Graph and Model-Free Q-Gradient Penalty for Path Planning and Motion Control
HG2P: Hippocampus-inspired High-reward Graph and Model-Free Q-Gradient Penalty for Path Planning and Motion Control
Haoran Wang
Yaoru Sun
Zeshen Tang
Haibo Shi
Chenyuan Jiao
59
0
0
12 Oct 2024
Looking Backward: Retrospective Backward Synthesis for Goal-Conditioned GFlowNets
Looking Backward: Retrospective Backward Synthesis for Goal-Conditioned GFlowNets
Haoran He
C. Chang
Huazhe Xu
Ling Pan
104
7
0
03 Jun 2024
Subgoal Search For Complex Reasoning Tasks
Subgoal Search For Complex Reasoning Tasks
K. Czechowski
Tomasz Odrzygó'zd'z
Marek Zbysiñski
Michał Zawalski
Krzysztof Olejnik
Yuhuai Wu
Lukasz Kuciñski
Piotr Milo's
ReLM
LRM
44
36
0
25 Aug 2021
State Entropy Maximization with Random Encoders for Efficient
  Exploration
State Entropy Maximization with Random Encoders for Efficient Exploration
Younggyo Seo
Lili Chen
Jinwoo Shin
Honglak Lee
Pieter Abbeel
Kimin Lee
32
123
0
18 Feb 2021
World Model as a Graph: Learning Latent Landmarks for Planning
World Model as a Graph: Learning Latent Landmarks for Planning
Lunjun Zhang
Ge Yang
Bradly C. Stadie
DRL
33
73
0
25 Nov 2020
Task-Agnostic Exploration via Policy Gradient of a Non-Parametric State
  Entropy Estimate
Task-Agnostic Exploration via Policy Gradient of a Non-Parametric State Entropy Estimate
Mirco Mutti
Lorenzo Pratissoli
Marcello Restelli
38
19
0
09 Jul 2020
Generating Adjacency-Constrained Subgoals in Hierarchical Reinforcement
  Learning
Generating Adjacency-Constrained Subgoals in Hierarchical Reinforcement Learning
Tianren Zhang
Shangqi Guo
Tian Tan
Xiaolin Hu
Feng Chen
35
84
0
20 Jun 2020
Planning to Explore via Self-Supervised World Models
Planning to Explore via Self-Supervised World Models
Ramanan Sekar
Oleh Rybkin
Kostas Daniilidis
Pieter Abbeel
Danijar Hafner
Deepak Pathak
SSL
43
403
0
12 May 2020
Agent57: Outperforming the Atari Human Benchmark
Agent57: Outperforming the Atari Human Benchmark
Adria Puigdomenech Badia
Bilal Piot
Steven Kapturowski
Pablo Sprechmann
Alex Vitvitskyi
Daniel Guo
Charles Blundell
OffRL
38
510
0
30 Mar 2020
Sparse Graphical Memory for Robust Planning
Sparse Graphical Memory for Robust Planning
Scott Emmons
Ajay Jain
Michael Laskin
Thanard Kurutach
Pieter Abbeel
Deepak Pathak
38
50
0
13 Mar 2020
Hallucinative Topological Memory for Zero-Shot Visual Planning
Hallucinative Topological Memory for Zero-Shot Visual Planning
Kara Liu
Thanard Kurutach
Christine Tung
Pieter Abbeel
Aviv Tamar
41
48
0
27 Feb 2020
Planning with Goal-Conditioned Policies
Planning with Goal-Conditioned Policies
Soroush Nasiriany
Vitchyr H. Pong
Steven Lin
Sergey Levine
OffRL
89
218
0
19 Nov 2019
Hierarchical Reinforcement Learning with Advantage-Based Auxiliary
  Rewards
Hierarchical Reinforcement Learning with Advantage-Based Auxiliary Rewards
Siyuan Li
Rui Wang
Minxue Tang
Chongjie Zhang
38
82
0
10 Oct 2019
Hierarchical Foresight: Self-Supervised Learning of Long-Horizon Tasks
  via Visual Subgoal Generation
Hierarchical Foresight: Self-Supervised Learning of Long-Horizon Tasks via Visual Subgoal Generation
Suraj Nair
Chelsea Finn
VGen
40
138
0
12 Sep 2019
Mapping State Space using Landmarks for Universal Goal Reaching
Mapping State Space using Landmarks for Universal Goal Reaching
Zhiao Huang
Fangchen Liu
Hao Su
30
67
0
15 Aug 2019
Learning World Graphs to Accelerate Hierarchical Reinforcement Learning
Learning World Graphs to Accelerate Hierarchical Reinforcement Learning
Wenling Shang
Alexander R. Trott
Stephan Zheng
Caiming Xiong
R. Socher
46
18
0
01 Jul 2019
Efficient Exploration via State Marginal Matching
Efficient Exploration via State Marginal Matching
Lisa Lee
Benjamin Eysenbach
Emilio Parisotto
Eric Xing
Sergey Levine
Ruslan Salakhutdinov
86
242
0
12 Jun 2019
Search on the Replay Buffer: Bridging Planning and Reinforcement
  Learning
Search on the Replay Buffer: Bridging Planning and Reinforcement Learning
Benjamin Eysenbach
Ruslan Salakhutdinov
Sergey Levine
OffRL
37
289
0
12 Jun 2019
Provably Efficient Maximum Entropy Exploration
Provably Efficient Maximum Entropy Exploration
Elad Hazan
Sham Kakade
Karan Singh
A. V. Soest
40
295
0
06 Dec 2018
Learning Goal Embeddings via Self-Play for Hierarchical Reinforcement
  Learning
Learning Goal Embeddings via Self-Play for Hierarchical Reinforcement Learning
Sainbayar Sukhbaatar
Emily L. Denton
Arthur Szlam
Rob Fergus
SSL
24
43
0
22 Nov 2018
Learning Actionable Representations with Goal-Conditioned Policies
Learning Actionable Representations with Goal-Conditioned Policies
Dibya Ghosh
Abhishek Gupta
Sergey Levine
51
109
0
19 Nov 2018
Exploration by Random Network Distillation
Exploration by Random Network Distillation
Yuri Burda
Harrison Edwards
Amos Storkey
Oleg Klimov
74
1,310
0
30 Oct 2018
Near-Optimal Representation Learning for Hierarchical Reinforcement
  Learning
Near-Optimal Representation Learning for Hierarchical Reinforcement Learning
Ofir Nachum
S. Gu
Honglak Lee
Sergey Levine
49
208
0
02 Oct 2018
QT-Opt: Scalable Deep Reinforcement Learning for Vision-Based Robotic
  Manipulation
QT-Opt: Scalable Deep Reinforcement Learning for Vision-Based Robotic Manipulation
Dmitry Kalashnikov
A. Irpan
P. Pastor
Julian Ibarz
Alexander Herzog
...
Deirdre Quillen
E. Holly
Mrinal Kalakrishnan
Vincent Vanhoucke
Sergey Levine
84
1,454
0
27 Jun 2018
Deep Reinforcement Learning in a Handful of Trials using Probabilistic
  Dynamics Models
Deep Reinforcement Learning in a Handful of Trials using Probabilistic Dynamics Models
Kurtland Chua
Roberto Calandra
R. McAllister
Sergey Levine
BDL
103
1,263
0
30 May 2018
Data-Efficient Hierarchical Reinforcement Learning
Data-Efficient Hierarchical Reinforcement Learning
Ofir Nachum
S. Gu
Honglak Lee
Sergey Levine
OffRL
80
803
0
21 May 2018
Subgoal Discovery for Hierarchical Dialogue Policy Learning
Subgoal Discovery for Hierarchical Dialogue Policy Learning
Da Tang
Xiujun Li
Jianfeng Gao
Chong-Jun Wang
Lihong Li
Tony Jebara
32
50
0
20 Apr 2018
Learning to Adapt in Dynamic, Real-World Environments Through
  Meta-Reinforcement Learning
Learning to Adapt in Dynamic, Real-World Environments Through Meta-Reinforcement Learning
Anusha Nagabandi
I. Clavera
Simin Liu
R. Fearing
Pieter Abbeel
Sergey Levine
Chelsea Finn
84
540
0
30 Mar 2018
Unsupervised Learning of Goal Spaces for Intrinsically Motivated Goal
  Exploration
Unsupervised Learning of Goal Spaces for Intrinsically Motivated Goal Exploration
Alexandre Péré
Sébastien Forestier
Olivier Sigaud
Pierre-Yves Oudeyer
SSL
DRL
18
95
0
02 Mar 2018
Semi-parametric Topological Memory for Navigation
Semi-parametric Topological Memory for Navigation
Nikolay Savinov
Alexey Dosovitskiy
V. Koltun
31
379
0
01 Mar 2018
Composable Planning with Attributes
Composable Planning with Attributes
Amy Zhang
Adam Lerer
Sainbayar Sukhbaatar
Rob Fergus
Arthur Szlam
53
64
0
01 Mar 2018
Addressing Function Approximation Error in Actor-Critic Methods
Addressing Function Approximation Error in Actor-Critic Methods
Scott Fujimoto
H. V. Hoof
David Meger
OffRL
122
5,121
0
26 Feb 2018
Feature Control as Intrinsic Motivation for Hierarchical Reinforcement
  Learning
Feature Control as Intrinsic Motivation for Hierarchical Reinforcement Learning
Nat Dilokthanakul
Christos Kaplanis
Nick Pawlowski
Murray Shanahan
35
92
0
18 May 2017
Curiosity-driven Exploration by Self-supervised Prediction
Curiosity-driven Exploration by Self-supervised Prediction
Deepak Pathak
Pulkit Agrawal
Alexei A. Efros
Trevor Darrell
LRM
SSL
89
2,416
0
15 May 2017
Stochastic Neural Networks for Hierarchical Reinforcement Learning
Stochastic Neural Networks for Hierarchical Reinforcement Learning
Carlos Florensa
Yan Duan
Pieter Abbeel
BDL
56
360
0
10 Apr 2017
Count-Based Exploration with Neural Density Models
Count-Based Exploration with Neural Density Models
Georg Ostrovski
Marc G. Bellemare
Aaron van den Oord
Rémi Munos
61
616
0
03 Mar 2017
FeUdal Networks for Hierarchical Reinforcement Learning
FeUdal Networks for Hierarchical Reinforcement Learning
A. Vezhnevets
Simon Osindero
Tom Schaul
N. Heess
Max Jaderberg
David Silver
Koray Kavukcuoglu
FedML
60
902
0
03 Mar 2017
#Exploration: A Study of Count-Based Exploration for Deep Reinforcement
  Learning
#Exploration: A Study of Count-Based Exploration for Deep Reinforcement Learning
Haoran Tang
Rein Houthooft
Davis Foote
Adam Stooke
Xi Chen
Yan Duan
John Schulman
F. Turck
Pieter Abbeel
OffRL
71
764
0
15 Nov 2016
Unifying Count-Based Exploration and Intrinsic Motivation
Unifying Count-Based Exploration and Intrinsic Motivation
Marc G. Bellemare
S. Srinivasan
Georg Ostrovski
Tom Schaul
D. Saxton
Rémi Munos
145
1,465
0
06 Jun 2016
Benchmarking Deep Reinforcement Learning for Continuous Control
Benchmarking Deep Reinforcement Learning for Continuous Control
Yan Duan
Xi Chen
Rein Houthooft
John Schulman
Pieter Abbeel
OffRL
53
1,689
0
22 Apr 2016
Hierarchical Deep Reinforcement Learning: Integrating Temporal
  Abstraction and Intrinsic Motivation
Hierarchical Deep Reinforcement Learning: Integrating Temporal Abstraction and Intrinsic Motivation
Tejas D. Kulkarni
Karthik Narasimhan
A. Saeedi
J. Tenenbaum
36
1,130
0
20 Apr 2016
Adam: A Method for Stochastic Optimization
Adam: A Method for Stochastic Optimization
Diederik P. Kingma
Jimmy Ba
ODL
288
149,474
0
22 Dec 2014
1