ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1809.03447
  4. Cited By
Expert-augmented actor-critic for ViZDoom and Montezumas Revenge

Expert-augmented actor-critic for ViZDoom and Montezumas Revenge

10 September 2018
Michal Garmulewicz
Henryk Michalewski
Piotr Milos
ArXiv (abs)PDFHTML

Papers citing "Expert-augmented actor-critic for ViZDoom and Montezumas Revenge"

19 / 19 papers shown
Title
Learning Montezuma's Revenge from a Single Demonstration
Learning Montezuma's Revenge from a Single Demonstration
Tim Salimans
Richard J. Chen
110
139
0
08 Dec 2018
Semi-parametric Topological Memory for Navigation
Semi-parametric Topological Memory for Navigation
Nikolay Savinov
Alexey Dosovitskiy
V. Koltun
73
383
0
01 Mar 2018
Pretraining Deep Actor-Critic Reinforcement Learning Algorithms With
  Expert Demonstrations
Pretraining Deep Actor-Critic Reinforcement Learning Algorithms With Expert Demonstrations
Xiaoqin Zhang
Huimin Ma
OffRL
79
38
0
31 Jan 2018
CARLA: An Open Urban Driving Simulator
CARLA: An Open Urban Driving Simulator
Alexey Dosovitskiy
G. Ros
Felipe Codevilla
Antonio M. López
V. Koltun
VLM
137
5,199
0
10 Nov 2017
Learning Complex Dexterous Manipulation with Deep Reinforcement Learning
  and Demonstrations
Learning Complex Dexterous Manipulation with Deep Reinforcement Learning and Demonstrations
Aravind Rajeswaran
Vikash Kumar
Abhishek Gupta
Giulia Vezzani
John Schulman
E. Todorov
Sergey Levine
142
1,101
0
28 Sep 2017
Scalable trust-region method for deep reinforcement learning using
  Kronecker-factored approximation
Scalable trust-region method for deep reinforcement learning using Kronecker-factored approximation
Yuhuai Wu
Elman Mansimov
Shun Liao
Roger C. Grosse
Jimmy Ba
OffRL
57
630
0
17 Aug 2017
Hindsight Experience Replay
Hindsight Experience Replay
Marcin Andrychowicz
Dwight Crow
Alex Ray
Jonas Schneider
Rachel Fong
Peter Welinder
Bob McGrew
Joshua Tobin
Pieter Abbeel
Wojciech Zaremba
OffRL
271
2,337
0
05 Jul 2017
Curiosity-driven Exploration by Self-supervised Prediction
Curiosity-driven Exploration by Self-supervised Prediction
Deepak Pathak
Pulkit Agrawal
Alexei A. Efros
Trevor Darrell
LRMSSL
113
2,449
0
15 May 2017
Beating Atari with Natural Language Guided Reinforcement Learning
Beating Atari with Natural Language Guided Reinforcement Learning
Russell Kaplan
Chris Sauer
A. Sosa
LM&Ro
64
69
0
18 Apr 2017
FeUdal Networks for Hierarchical Reinforcement Learning
FeUdal Networks for Hierarchical Reinforcement Learning
A. Vezhnevets
Simon Osindero
Tom Schaul
N. Heess
Max Jaderberg
David Silver
Koray Kavukcuoglu
FedML
96
907
0
03 Mar 2017
RL$^2$: Fast Reinforcement Learning via Slow Reinforcement Learning
RL2^22: Fast Reinforcement Learning via Slow Reinforcement Learning
Yan Duan
John Schulman
Xi Chen
Peter L. Bartlett
Ilya Sutskever
Pieter Abbeel
OffRL
99
1,027
0
09 Nov 2016
Learning to Act by Predicting the Future
Learning to Act by Predicting the Future
Alexey Dosovitskiy
V. Koltun
152
281
0
06 Nov 2016
Deep Successor Reinforcement Learning
Deep Successor Reinforcement Learning
Tejas D. Kulkarni
A. Saeedi
Simanta Gautam
S. Gershman
69
209
0
08 Jun 2016
Unifying Count-Based Exploration and Intrinsic Motivation
Unifying Count-Based Exploration and Intrinsic Motivation
Marc G. Bellemare
S. Srinivasan
Georg Ostrovski
Tom Schaul
D. Saxton
Rémi Munos
176
1,483
0
06 Jun 2016
ViZDoom: A Doom-based AI Research Platform for Visual Reinforcement
  Learning
ViZDoom: A Doom-based AI Research Platform for Visual Reinforcement Learning
Michal Kempka
Marek Wydmuch
Grzegorz Runc
Jakub Toczek
Wojciech Ja'skowski
80
700
0
06 May 2016
Asynchronous Methods for Deep Reinforcement Learning
Asynchronous Methods for Deep Reinforcement Learning
Volodymyr Mnih
Adria Puigdomenech Badia
M. Berk Mirza
Alex Graves
Timothy Lillicrap
Tim Harley
David Silver
Koray Kavukcuoglu
204
8,875
0
04 Feb 2016
High-Dimensional Continuous Control Using Generalized Advantage
  Estimation
High-Dimensional Continuous Control Using Generalized Advantage Estimation
John Schulman
Philipp Moritz
Sergey Levine
Michael I. Jordan
Pieter Abbeel
OffRL
106
3,434
0
08 Jun 2015
The Arcade Learning Environment: An Evaluation Platform for General
  Agents
The Arcade Learning Environment: An Evaluation Platform for General Agents
Marc G. Bellemare
Yavar Naddaf
J. Veness
Michael Bowling
120
3,020
0
19 Jul 2012
A Reduction of Imitation Learning and Structured Prediction to No-Regret
  Online Learning
A Reduction of Imitation Learning and Structured Prediction to No-Regret Online Learning
Stéphane Ross
Geoffrey J. Gordon
J. Andrew Bagnell
OffRL
236
3,232
0
02 Nov 2010
1