ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1905.06893
  4. Cited By
Leveraging exploration in off-policy algorithms via normalizing flows

Leveraging exploration in off-policy algorithms via normalizing flows

16 May 2019
Bogdan Mazoure
T. Doan
A. Durand
R. Devon Hjelm
Joelle Pineau
    OnRL
ArXivPDFHTML

Papers citing "Leveraging exploration in off-policy algorithms via normalizing flows"

14 / 14 papers shown
Title
S$^2$AC: Energy-Based Reinforcement Learning with Stein Soft Actor
  Critic
S2^22AC: Energy-Based Reinforcement Learning with Stein Soft Actor Critic
Safa Messaoud
Billel Mokeddem
Zhenghai Xue
Linsey Pang
Bo An
Haipeng Chen
Sanjay Chawla
51
3
0
02 May 2024
Learning Sampling Distributions for Model Predictive Control
Learning Sampling Distributions for Model Predictive Control
Jacob Sacks
Byron Boots
25
21
0
05 Dec 2022
CEIP: Combining Explicit and Implicit Priors for Reinforcement Learning
  with Demonstrations
CEIP: Combining Explicit and Implicit Priors for Reinforcement Learning with Demonstrations
Kai Yan
Alex Schwing
Yu-xiong Wang
OffRL
30
2
0
18 Oct 2022
Efficiently Learning Small Policies for Locomotion and Manipulation
Efficiently Learning Small Policies for Locomotion and Manipulation
Shashank Hegde
Gaurav Sukhatme
40
3
0
30 Sep 2022
Action Noise in Off-Policy Deep Reinforcement Learning: Impact on
  Exploration and Performance
Action Noise in Off-Policy Deep Reinforcement Learning: Impact on Exploration and Performance
Jakob J. Hollenstein
Sayantan Auddy
Matteo Saveriano
Erwan Renaudo
J. Piater
46
17
0
08 Jun 2022
AUTM Flow: Atomic Unrestricted Time Machine for Monotonic Normalizing
  Flows
AUTM Flow: Atomic Unrestricted Time Machine for Monotonic Normalizing Flows
Difeng Cai
Yuliang Ji
Huan He
Qiang Ye
Yuanzhe Xi
TPM
36
4
0
05 Jun 2022
Flow-based Recurrent Belief State Learning for POMDPs
Flow-based Recurrent Belief State Learning for POMDPs
Xiaoyu Chen
Yao Mu
Ping Luo
Sheng Li
Jianyu Chen
61
19
0
23 May 2022
TO-FLOW: Efficient Continuous Normalizing Flows with Temporal
  Optimization adjoint with Moving Speed
TO-FLOW: Efficient Continuous Normalizing Flows with Temporal Optimization adjoint with Moving Speed
Shian Du
Yihong Luo
Wei Chen
Jian Xu
Delu Zeng
37
7
0
19 Mar 2022
Transflower: probabilistic autoregressive dance generation with
  multimodal attention
Transflower: probabilistic autoregressive dance generation with multimodal attention
Guillermo Valle Pérez
G. Henter
Jonas Beskow
A. Holzapfel
Pierre-Yves Oudeyer
Simon Alexanderson
41
43
0
25 Jun 2021
Bellman: A Toolbox for Model-Based Reinforcement Learning in TensorFlow
Bellman: A Toolbox for Model-Based Reinforcement Learning in TensorFlow
John Mcleod
Hrvoje Stojić
Vincent Adam
Dongho Kim
Jordi Grau-Moya
Peter Vrancx
Felix Leibfried
OffRL
23
2
0
26 Mar 2021
Augmented Normalizing Flows: Bridging the Gap Between Generative Flows
  and Latent Variable Models
Augmented Normalizing Flows: Bridging the Gap Between Generative Flows and Latent Variable Models
Chin-Wei Huang
Laurent Dinh
Aaron Courville
DRL
31
87
0
17 Feb 2020
Representation of Reinforcement Learning Policies in Reproducing Kernel
  Hilbert Spaces
Representation of Reinforcement Learning Policies in Reproducing Kernel Hilbert Spaces
Bogdan Mazoure
T. Doan
Tianyu Li
V. Makarenkov
Joelle Pineau
Doina Precup
Guillaume Rabusseau
OffRL
21
1
0
07 Feb 2020
Discrete and Continuous Action Representation for Practical RL in Video
  Games
Discrete and Continuous Action Representation for Practical RL in Video Games
Olivier Delalleau
Maxim Peter
Eloi Alonso
Adrien Logut
25
52
0
23 Dec 2019
Dropout as a Bayesian Approximation: Representing Model Uncertainty in
  Deep Learning
Dropout as a Bayesian Approximation: Representing Model Uncertainty in Deep Learning
Y. Gal
Zoubin Ghahramani
UQCV
BDL
289
9,167
0
06 Jun 2015
1