Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
1801.00690
Cited By
DeepMind Control Suite
2 January 2018
Yuval Tassa
Yotam Doron
Alistair Muldal
Tom Erez
Yazhe Li
Diego de Las Casas
David Budden
A. Abdolmaleki
J. Merel
Andrew Lefrancq
Timothy Lillicrap
Martin Riedmiller
ELM
LM&Ro
BDL
Re-assign community
ArXiv
PDF
HTML
Papers citing
"DeepMind Control Suite"
50 / 791 papers shown
Title
Continuous Control Reinforcement Learning: Distributed Distributional DrQ Algorithms
Zehao Zhou
OffRL
29
0
0
16 Apr 2024
Adversarial Imitation Learning via Boosting
Jonathan D. Chang
Dhruv Sreenivas
Yingbing Huang
Kianté Brantley
Wen Sun
29
3
0
12 Apr 2024
AI-MOLE: Autonomous Iterative Motion Learning for Unknown Nonlinear Dynamics with Extensive Experimental Validation
Michael Meindl
Simon Bachhuber
Thomas Seel
41
4
0
09 Apr 2024
SENSOR: Imitate Third-Person Expert's Behaviors via Active Sensoring
Kaichen Huang
Minghao Shao
Shenghua Wan
Hai-Hang Sun
Shuai Feng
Le Gan
De-Chuan Zhan
40
0
0
04 Apr 2024
Decision Transformer as a Foundation Model for Partially Observable Continuous Control
Xiangyuan Zhang
Weichao Mao
Haoran Qiu
Tamer Basar
OffRL
AI4CE
32
6
0
03 Apr 2024
Learning Off-policy with Model-based Intrinsic Motivation For Active Online Exploration
Yibo Wang
Jiang Zhao
OffRL
OnRL
30
0
0
31 Mar 2024
Simple Ingredients for Offline Reinforcement Learning
Edoardo Cetin
Andrea Tirinzoni
Matteo Pirotta
A. Lazaric
Yann Ollivier
Ahmed Touati
OffRL
42
2
0
19 Mar 2024
Dreaming of Many Worlds: Learning Contextual World Models Aids Zero-Shot Generalization
Sai Prasanna
Karim Farid
Raghu Rajan
André Biedenkapp
60
2
0
16 Mar 2024
HumanoidBench: Simulated Humanoid Benchmark for Whole-Body Locomotion and Manipulation
Carmelo Sferrazza
Dun-Ming Huang
Xingyu Lin
Youngwoon Lee
Pieter Abbeel
57
37
0
15 Mar 2024
AD3: Implicit Action is the Key for World Models to Distinguish the Diverse Visual Distractors
Yucen Wang
Shenghua Wan
Le Gan
Shuai Feng
De-Chuan Zhan
VGen
27
4
0
15 Mar 2024
BEHAVIOR-1K: A Human-Centered, Embodied AI Benchmark with 1,000 Everyday Activities and Realistic Simulation
Chengshu Li
Ruohan Zhang
J. Wong
Cem Gokmen
S. Srivastava
...
Silvio Savarese
H. Gweon
Chenxi Liu
Jiajun Wu
Fei-Fei Li
VGen
LM&Ro
VLM
37
37
0
14 Mar 2024
Spatiotemporal Predictive Pre-training for Robotic Motor Control
Jiange Yang
Bei Liu
Jianlong Fu
Bocheng Pan
Gangshan Wu
Limin Wang
53
10
0
08 Mar 2024
Reset & Distill: A Recipe for Overcoming Negative Transfer in Continual Reinforcement Learning
Hongjoon Ahn
Jinu Hyeon
Youngmin Oh
Bosun Hwang
Taesup Moon
CLL
OnRL
37
2
0
08 Mar 2024
Mastering Memory Tasks with World Models
Mohammad Reza Samsami
Artem Zholus
Janarthanan Rajendran
Sarath Chandar
CLL
OffRL
34
23
0
07 Mar 2024
World Models for Autonomous Driving: An Initial Survey
Yanchen Guan
Haicheng Liao
Zhenning Li
Jia Hu
Runze Yuan
Yunjian Li
Guohui Zhang
Chengzhong Xu
40
33
0
05 Mar 2024
A Case for Validation Buffer in Pessimistic Actor-Critic
Michal Nauman
M. Ostaszewski
Marek Cygan
34
0
0
01 Mar 2024
EfficientZero V2: Mastering Discrete and Continuous Control with Limited Data
Shengjie Wang
Shaohuai Liu
Weirui Ye
Jiacheng You
Yang Gao
OffRL
29
13
0
01 Mar 2024
Overestimation, Overfitting, and Plasticity in Actor-Critic: the Bitter Lesson of Reinforcement Learning
Michal Nauman
Michal Bortkiewicz
Piotr Milo's
Tomasz Trzciñski
M. Ostaszewski
Marek Cygan
OffRL
35
17
0
01 Mar 2024
A Model-Based Approach for Improving Reinforcement Learning Efficiency Leveraging Expert Observations
E. C. Ozcan
Vittorio Giammarino
James Queeney
I. Paschalidis
OffRL
42
0
0
29 Feb 2024
Sample-Efficient Preference-based Reinforcement Learning with Dynamics Aware Rewards
Katherine Metcalf
Miguel Sarabia
Natalie Mackraz
B. Theobald
42
6
0
28 Feb 2024
RIME: Robust Preference-based Reinforcement Learning with Noisy Preferences
Jie Cheng
Gang Xiong
Xingyuan Dai
Qinghai Miao
Yisheng Lv
Fei-Yue Wang
33
15
0
27 Feb 2024
Unsupervised Zero-Shot Reinforcement Learning via Functional Reward Encodings
Kevin Frans
Seohong Park
Pieter Abbeel
Sergey Levine
OffRL
48
11
0
27 Feb 2024
Foundation Policies with Hilbert Representations
Seohong Park
Tobias Kreiman
Sergey Levine
SSL
OffRL
50
21
0
23 Feb 2024
Stable Neural Stochastic Differential Equations in Analyzing Irregular Time Series Data
YongKyung Oh
Dongyoung Lim
Sungil Kim
AI4TS
43
13
0
22 Feb 2024
ACE : Off-Policy Actor-Critic with Causality-Aware Entropy Regularization
Tianying Ji
Yongyuan Liang
Yan Zeng
Yu-Juan Luo
Guowei Xu
Jiawei Guo
Ruijie Zheng
Furong Huang
Gang Hua
Huazhe Xu
CML
55
11
0
22 Feb 2024
Enhancing Reinforcement Learning Agents with Local Guides
Paul Daoudi
Bogdan Robu
Christophe Prieur
Ludovic Dos Santos
M. Barlier
OnRL
31
3
0
21 Feb 2024
Revisiting Data Augmentation in Deep Reinforcement Learning
Jianshu Hu
Yunpeng Jiang
Paul Weng
OffRL
37
6
0
19 Feb 2024
Jack of All Trades, Master of Some, a Multi-Purpose Transformer Agent
Quentin Gallouedec
E. Beeching
Clément Romac
Emmanuel Dellandrea
26
11
0
15 Feb 2024
Task-conditioned adaptation of visual features in multi-task policy learning
Pierre Marza
L. Matignon
Olivier Simonin
Christian Wolf
53
2
0
12 Feb 2024
Real-World Robot Applications of Foundation Models: A Review
Kento Kawaharazuka
T. Matsushima
Andrew Gambardella
Jiaxian Guo
Chris Paxton
Andy Zeng
OffRL
VLM
LM&Ro
51
47
0
08 Feb 2024
DiffTORI: Differentiable Trajectory Optimization for Deep Reinforcement and Imitation Learning
Weikang Wan
Ziyu Wang
Zackory M. Erickson
David Held
David Held
34
4
0
08 Feb 2024
Reinforcement Learning from Bagged Reward
Yuting Tang
Xin-Qiang Cai
Yao-Xiang Ding
Qiyu Wu
Guoqing Liu
Masashi Sugiyama
OffRL
36
0
0
06 Feb 2024
Transductive Reward Inference on Graph
B. Qu
Xiaofeng Cao
Qing Guo
Yi Chang
Ivor W. Tsang
Chengqi Zhang
OffRL
38
0
0
06 Feb 2024
A Multi-step Loss Function for Robust Learning of the Dynamics in Model-based Reinforcement Learning
Abdelhakim Benechehab
Albert Thomas
Giuseppe Paolo
Maurizio Filippone
Balázs Kégl
NoLa
28
1
0
05 Feb 2024
Integrating Human Expertise in Continuous Spaces: A Novel Interactive Bayesian Optimization Framework with Preference Expected Improvement
Nikolaus Feith
Elmar Rueckert
37
1
0
23 Jan 2024
Bridging Evolutionary Algorithms and Reinforcement Learning: A Comprehensive Survey on Hybrid Algorithms
Pengyi Li
Jianye Hao
Hongyao Tang
Xian Fu
Yan Zheng
Ke Tang
39
9
0
22 Jan 2024
Sharing Knowledge in Multi-Task Deep Reinforcement Learning
Carlo DÉramo
Davide Tateo
Andrea Bonarini
Marcello Restelli
Jan Peters
59
124
0
17 Jan 2024
Crowd-PrefRL: Preference-Based Reward Learning from Crowds
David Chhan
Ellen R. Novoseller
Vernon J. Lawhern
40
5
0
17 Jan 2024
Personalized Reinforcement Learning with a Budget of Policies
Dmitry Ivanov
Omer Ben-Porat
OffRL
6
2
0
12 Jan 2024
A Minimaximalist Approach to Reinforcement Learning from Human Feedback
Gokul Swamy
Christoph Dann
Rahul Kidambi
Zhiwei Steven Wu
Alekh Agarwal
OffRL
41
96
0
08 Jan 2024
Self-supervised Pretraining for Decision Foundation Model: Formulation, Pipeline and Challenges
Xiaoqian Liu
Jianbin Jiao
Junge Zhang
OffRL
LRM
46
2
0
29 Dec 2023
Gradient-based Planning with World Models
V. JyothirS
Siddhartha Jalagam
Yann LeCun
Vlad Sobal
34
4
0
28 Dec 2023
Generalizable Visual Reinforcement Learning with Segment Anything Model
Ziyu Wang
Yanjie Ze
Yifei Sun
Zhecheng Yuan
Huazhe Xu
VLM
33
8
0
28 Dec 2023
Visual Spatial Attention and Proprioceptive Data-Driven Reinforcement Learning for Robust Peg-in-Hole Task Under Variable Conditions
André Yuji Yasutomi
Hideyuki Ichiwara
Hiroshi Ito
Hiroki Mori
Tetsuya Ogata
9
19
0
27 Dec 2023
Efficient Reinforcement Learning via Decoupling Exploration and Utilization
Jingpu Yang
Helin Wang
Qirui Zhao
Zhecheng Shi
Zirui Song
Miao Fang
24
0
0
26 Dec 2023
MaDi: Learning to Mask Distractions for Generalization in Visual Deep Reinforcement Learning
Bram Grooten
Tristan Tomilin
Gautham Vasan
Matthew E. Taylor
A. R. Mahmood
Meng Fang
Mykola Pechenizkiy
Decebal Constantin Mocanu
32
8
0
23 Dec 2023
CUDC: A Curiosity-Driven Unsupervised Data Collection Method with Adaptive Temporal Distances for Offline Reinforcement Learning
Chenyu Sun
Hangwei Qian
Chunyan Miao
OffRL
40
1
0
19 Dec 2023
Colored Noise in PPO: Improved Exploration and Performance through Correlated Action Sampling
Jakob J. Hollenstein
Georg Martius
J. Piater
22
3
0
18 Dec 2023
Toward General-Purpose Robots via Foundation Models: A Survey and Meta-Analysis
Yafei Hu
Quanting Xie
Vidhi Jain
Jonathan M Francis
Jay Patrikar
...
Xiaolong Wang
Sebastian A. Scherer
Z. Kira
Fei Xia
Yonatan Bisk
LM&Ro
AI4CE
50
64
0
14 Dec 2023
Beyond Expected Return: Accounting for Policy Reproducibility when Evaluating Reinforcement Learning Algorithms
Manon Flageat
Bryan Lim
Antoine Cully
OffRL
25
3
0
12 Dec 2023
Previous
1
2
3
4
5
...
14
15
16
Next