ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2206.08332
  4. Cited By
BYOL-Explore: Exploration by Bootstrapped Prediction

BYOL-Explore: Exploration by Bootstrapped Prediction

16 June 2022
Z. Guo
S. Thakoor
Miruna Pislar
Bernardo Avila-Pires
Florent Altché
Corentin Tallec
Alaa Saade
Daniele Calandriello
Jean-Bastien Grill
Yunhao Tang
Michal Valko
Rémi Munos
M. G. Azar
Bilal Piot
ArXiv (abs)PDFHTML

Papers citing "BYOL-Explore: Exploration by Bootstrapped Prediction"

50 / 67 papers shown
Title
seq-JEPA: Autoregressive Predictive Learning of Invariant-Equivariant World Models
seq-JEPA: Autoregressive Predictive Learning of Invariant-Equivariant World Models
Hafez Ghaemi
Eilif Muller
Shahab Bakhtiari
152
0
0
06 May 2025
Contextual Similarity Distillation: Ensemble Uncertainties with a Single Model
Contextual Similarity Distillation: Ensemble Uncertainties with a Single Model
Moritz A. Zanger
Pascal R. van der Vaart
Wendelin Bohmer
M. Spaan
UQCVBDL
497
2
0
14 Mar 2025
Leveraging Skills from Unlabeled Prior Data for Efficient Online Exploration
Leveraging Skills from Unlabeled Prior Data for Efficient Online Exploration
Max Wilcoxson
Qiyang Li
Kevin Frans
Sergey Levine
SSLOffRLOnRL
172
0
0
23 Oct 2024
RLeXplore: Accelerating Research in Intrinsically-Motivated Reinforcement Learning
RLeXplore: Accelerating Research in Intrinsically-Motivated Reinforcement Learning
Mingqi Yuan
Roger Creus Castanyer
Bo Li
Xin Jin
Glen Berseth
Wenjun Zeng
160
0
0
29 May 2024
Inference via Interpolation: Contrastive Representations Provably Enable Planning and Inference
Inference via Interpolation: Contrastive Representations Provably Enable Planning and Inference
Benjamin Eysenbach
Vivek Myers
Ruslan Salakhutdinov
Sergey Levine
AI4TS
128
12
0
06 Mar 2024
The Mechanism of Prediction Head in Non-contrastive Self-supervised
  Learning
The Mechanism of Prediction Head in Non-contrastive Self-supervised Learning
Zixin Wen
Yuanzhi Li
SSL
103
35
0
12 May 2022
Reward-Free Model-Based Reinforcement Learning with Linear Function
  Approximation
Reward-Free Model-Based Reinforcement Learning with Linear Function Approximation
Weitong Zhang
Dongruo Zhou
Quanquan Gu
OffRL
60
28
0
12 Oct 2021
Is Curiosity All You Need? On the Utility of Emergent Behaviours from
  Curious Exploration
Is Curiosity All You Need? On the Utility of Emergent Behaviours from Curious Exploration
Oliver Groth
Markus Wulfmeier
Giulia Vezzani
Vibhavari Dasagi
Tim Hertweck
Roland Hafner
N. Heess
Martin Riedmiller
LRM
68
20
0
17 Sep 2021
Podracer architectures for scalable Reinforcement Learning
Podracer architectures for scalable Reinforcement Learning
Matteo Hessel
M. Kroiss
Aidan Clark
Iurii Kemaev
John Quan
Thomas Keck
Fabio Viola
H. V. Hasselt
54
39
0
13 Apr 2021
Broaden Your Views for Self-Supervised Video Learning
Broaden Your Views for Self-Supervised Video Learning
Adrià Recasens
Pauline Luc
Jean-Baptiste Alayrac
Luyu Wang
Ross Hemsley
...
Florent Altché
M. Valko
Jean-Bastien Grill
Aaron van den Oord
Andrew Zisserman
SSLAI4TS
101
128
0
30 Mar 2021
Understanding self-supervised Learning Dynamics without Contrastive
  Pairs
Understanding self-supervised Learning Dynamics without Contrastive Pairs
Yuandong Tian
Xinlei Chen
Surya Ganguli
SSL
213
286
0
12 Feb 2021
Large-Scale Representation Learning on Graphs via Bootstrapping
Large-Scale Representation Learning on Graphs via Bootstrapping
S. Thakoor
Corentin Tallec
M. G. Azar
Mehdi Azabou
Eva L. Dyer
Rémi Munos
Petar Velivcković
Michal Valko
SSL
74
228
0
12 Feb 2021
Geometric Entropic Exploration
Geometric Entropic Exploration
Z. Guo
M. G. Azar
Alaa Saade
S. Thakoor
Bilal Piot
Bernardo Avila-Pires
Michal Valko
Thomas Mesnard
Tor Lattimore
Rémi Munos
75
32
0
06 Jan 2021
BYOL works even without batch statistics
BYOL works even without batch statistics
Pierre Harvey Richemond
Jean-Bastien Grill
Florent Altché
Corentin Tallec
Florian Strub
...
Samuel L. Smith
Soham De
Razvan Pascanu
Bilal Piot
Michal Valko
SSL
297
115
0
20 Oct 2020
Causal Curiosity: RL Agents Discovering Self-supervised Experiments for
  Causal Representation Learning
Causal Curiosity: RL Agents Discovering Self-supervised Experiments for Causal Representation Learning
Sumedh Anand Sontakke
Arash Mehrjou
Laurent Itti
Bernhard Schölkopf
CML
79
63
0
07 Oct 2020
Provably Efficient Reward-Agnostic Navigation with Linear Value
  Iteration
Provably Efficient Reward-Agnostic Navigation with Linear Value Iteration
Andrea Zanette
A. Lazaric
Mykel J. Kochenderfer
Emma Brunskill
75
64
0
18 Aug 2020
Fast active learning for pure exploration in reinforcement learning
Fast active learning for pure exploration in reinforcement learning
Pierre Ménard
O. D. Domingues
Anders Jonsson
E. Kaufmann
Edouard Leurent
Michal Valko
56
97
0
27 Jul 2020
Data-Efficient Reinforcement Learning with Self-Predictive
  Representations
Data-Efficient Reinforcement Learning with Self-Predictive Representations
Max Schwarzer
Ankesh Anand
Rishab Goel
R. Devon Hjelm
Aaron Courville
Philip Bachman
99
321
0
12 Jul 2020
The NetHack Learning Environment
The NetHack Learning Environment
Heinrich Küttler
Nantas Nardelli
Alexander H. Miller
Roberta Raileanu
Marco Selvatici
Edward Grefenstette
Tim Rocktaschel
84
181
0
24 Jun 2020
On Reward-Free Reinforcement Learning with Linear Function Approximation
On Reward-Free Reinforcement Learning with Linear Function Approximation
Ruosong Wang
S. Du
Lin F. Yang
Ruslan Salakhutdinov
OffRL
73
107
0
19 Jun 2020
Automatic Curriculum Learning through Value Disagreement
Automatic Curriculum Learning through Value Disagreement
Yunzhi Zhang
Pieter Abbeel
Lerrel Pinto
70
109
0
17 Jun 2020
Bootstrap your own latent: A new approach to self-supervised Learning
Bootstrap your own latent: A new approach to self-supervised Learning
Jean-Bastien Grill
Florian Strub
Florent Altché
Corentin Tallec
Pierre Harvey Richemond
...
M. G. Azar
Bilal Piot
Koray Kavukcuoglu
Rémi Munos
Michal Valko
SSL
398
6,844
0
13 Jun 2020
Adaptive Reward-Free Exploration
Adaptive Reward-Free Exploration
E. Kaufmann
Pierre Ménard
O. D. Domingues
Anders Jonsson
Edouard Leurent
Michal Valko
58
82
0
11 Jun 2020
Planning to Explore via Self-Supervised World Models
Planning to Explore via Self-Supervised World Models
Ramanan Sekar
Oleh Rybkin
Kostas Daniilidis
Pieter Abbeel
Danijar Hafner
Deepak Pathak
SSL
77
412
0
12 May 2020
Bootstrap Latent-Predictive Representations for Multitask Reinforcement
  Learning
Bootstrap Latent-Predictive Representations for Multitask Reinforcement Learning
Z. Guo
Bernardo Avila-Pires
Bilal Piot
Jean-Bastien Grill
Florent Altché
Rémi Munos
M. G. Azar
BDLDRLSSL
180
143
0
30 Apr 2020
First return, then explore
First return, then explore
Adrien Ecoffet
Joost Huizinga
Joel Lehman
Kenneth O. Stanley
Jeff Clune
82
363
0
27 Apr 2020
CURL: Contrastive Unsupervised Representations for Reinforcement
  Learning
CURL: Contrastive Unsupervised Representations for Reinforcement Learning
A. Srinivas
Michael Laskin
Pieter Abbeel
SSLDRLOffRL
100
1,092
0
08 Apr 2020
Agent57: Outperforming the Atari Human Benchmark
Agent57: Outperforming the Atari Human Benchmark
Adria Puigdomenech Badia
Bilal Piot
Steven Kapturowski
Pablo Sprechmann
Alex Vitvitskyi
Daniel Guo
Charles Blundell
OffRL
88
521
0
30 Mar 2020
Active Model Estimation in Markov Decision Processes
Active Model Estimation in Markov Decision Processes
Jean Tarbouriech
S. Shekhar
Matteo Pirotta
Mohammad Ghavamzadeh
A. Lazaric
72
25
0
06 Mar 2020
Never Give Up: Learning Directed Exploration Strategies
Never Give Up: Learning Directed Exploration Strategies
Adria Puigdomenech Badia
Pablo Sprechmann
Alex Vitvitskyi
Daniel Guo
Bilal Piot
...
O. Tieleman
Martín Arjovsky
Alexander Pritzel
Andew Bolt
Charles Blundell
72
299
0
14 Feb 2020
Reward-Free Exploration for Reinforcement Learning
Reward-Free Exploration for Reinforcement Learning
Chi Jin
A. Krishnamurthy
Max Simchowitz
Tiancheng Yu
OffRL
172
197
0
07 Feb 2020
Dream to Control: Learning Behaviors by Latent Imagination
Dream to Control: Learning Behaviors by Latent Imagination
Danijar Hafner
Timothy Lillicrap
Jimmy Ba
Mohammad Norouzi
VLM
130
1,374
0
03 Dec 2019
V-MPO: On-Policy Maximum a Posteriori Policy Optimization for Discrete
  and Continuous Control
V-MPO: On-Policy Maximum a Posteriori Policy Optimization for Discrete and Continuous Control
H. F. Song
A. Abdolmaleki
Jost Tobias Springenberg
Aidan Clark
Hubert Soyer
...
Dhruva Tirumala
N. Heess
Dan Belov
Martin Riedmiller
M. Botvinick
91
125
0
26 Sep 2019
Making Efficient Use of Demonstrations to Solve Hard Exploration
  Problems
Making Efficient Use of Demonstrations to Solve Hard Exploration Problems
T. Paine
Çağlar Gülçehre
Bobak Shahriari
Misha Denil
Matt Hoffman
...
Duncan Williams
Gabriel Barth-Maron
Ziyun Wang
Nando de Freitas
Worlds Team
67
84
0
03 Sep 2019
Dynamical Distance Learning for Semi-Supervised and Unsupervised Skill
  Discovery
Dynamical Distance Learning for Semi-Supervised and Unsupervised Skill Discovery
Kristian Hartikainen
Xinyang Geng
Tuomas Haarnoja
Sergey Levine
SSL
70
82
0
18 Jul 2019
Shaping Belief States with Generative Environment Models for RL
Shaping Belief States with Generative Environment Models for RL
Karol Gregor
Danilo Jimenez Rezende
F. Besse
Yan Wu
Hamza Merzic
Aaron van den Oord
OffRLAI4CE
107
119
0
21 Jun 2019
Self-Supervised Exploration via Disagreement
Self-Supervised Exploration via Disagreement
Deepak Pathak
Dhiraj Gandhi
Abhinav Gupta
SSL
83
384
0
10 Jun 2019
Maximum Entropy-Regularized Multi-Goal Reinforcement Learning
Maximum Entropy-Regularized Multi-Goal Reinforcement Learning
Rui Zhao
Xudong Sun
Volker Tresp
63
83
0
21 May 2019
Skew-Fit: State-Covering Self-Supervised Reinforcement Learning
Skew-Fit: State-Covering Self-Supervised Reinforcement Learning
Vitchyr H. Pong
Murtaza Dalal
Steven Lin
Ashvin Nair
Shikhar Bahl
Sergey Levine
OffRLSSL
100
277
0
08 Mar 2019
Active Exploration in Markov Decision Processes
Active Exploration in Markov Decision Processes
Jean Tarbouriech
A. Lazaric
55
50
0
28 Feb 2019
World Discovery Models
World Discovery Models
M. G. Azar
Bilal Piot
Bernardo Avila-Pires
Jean-Bastien Grill
Florent Altché
Rémi Munos
74
26
0
20 Feb 2019
Go-Explore: a New Approach for Hard-Exploration Problems
Go-Explore: a New Approach for Hard-Exploration Problems
Adrien Ecoffet
Joost Huizinga
Joel Lehman
Kenneth O. Stanley
Jeff Clune
AI4TS
97
370
0
30 Jan 2019
Provably Efficient Maximum Entropy Exploration
Provably Efficient Maximum Entropy Exploration
Elad Hazan
Sham Kakade
Karan Singh
A. V. Soest
81
304
0
06 Dec 2018
Unsupervised Control Through Non-Parametric Discriminative Rewards
Unsupervised Control Through Non-Parametric Discriminative Rewards
David Warde-Farley
T. Wiele
Tejas D. Kulkarni
Catalin Ionescu
Steven Hansen
Volodymyr Mnih
DRLOffRLSSL
96
178
0
28 Nov 2018
Learning Latent Dynamics for Planning from Pixels
Learning Latent Dynamics for Planning from Pixels
Danijar Hafner
Timothy Lillicrap
Ian S. Fischer
Ruben Villegas
David R Ha
Honglak Lee
James Davidson
BDL
92
1,448
0
12 Nov 2018
Exploration by Random Network Distillation
Exploration by Random Network Distillation
Yuri Burda
Harrison Edwards
Amos Storkey
Oleg Klimov
161
1,345
0
30 Oct 2018
Multi-task Deep Reinforcement Learning with PopArt
Multi-task Deep Reinforcement Learning with PopArt
Matteo Hessel
Hubert Soyer
L. Espeholt
Wojciech M. Czarnecki
Simon Schmitt
H. V. Hasselt
135
319
0
12 Sep 2018
Large-Scale Study of Curiosity-Driven Learning
Large-Scale Study of Curiosity-Driven Learning
Yuri Burda
Harrison Edwards
Deepak Pathak
Amos Storkey
Trevor Darrell
Alexei A. Efros
LRM
72
707
0
13 Aug 2018
Visual Reinforcement Learning with Imagined Goals
Visual Reinforcement Learning with Imagined Goals
Ashvin Nair
Vitchyr H. Pong
Murtaza Dalal
Shikhar Bahl
Steven Lin
Sergey Levine
SSL
86
544
0
12 Jul 2018
World Models
World Models
David R Ha
Jürgen Schmidhuber
SyDa
150
1,101
0
27 Mar 2018
12
Next