ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2110.05038
  4. Cited By
Recurrent Model-Free RL Can Be a Strong Baseline for Many POMDPs

Recurrent Model-Free RL Can Be a Strong Baseline for Many POMDPs

11 October 2021
Tianwei Ni
Benjamin Eysenbach
Ruslan Salakhutdinov
ArXivPDFHTML

Papers citing "Recurrent Model-Free RL Can Be a Strong Baseline for Many POMDPs"

19 / 69 papers shown
Title
Filter-Aware Model-Predictive Control
Filter-Aware Model-Predictive Control
Baris Kayalibay
Atanas Mirchev
Ahmed Agha
Patrick van der Smagt
Justin Bayer
44
0
0
20 Apr 2023
Habits and goals in synergy: a variational Bayesian framework for
  behavior
Habits and goals in synergy: a variational Bayesian framework for behavior
Dongqi Han
Kenji Doya
Dongsheng Li
Jun Tani
BDL
23
220
0
11 Apr 2023
Active hypothesis testing in unknown environments using recurrent neural
  networks and model free reinforcement learning
Active hypothesis testing in unknown environments using recurrent neural networks and model free reinforcement learning
George Stamatelis
N. Kalouptsidis
24
2
0
19 Mar 2023
Structured State Space Models for In-Context Reinforcement Learning
Structured State Space Models for In-Context Reinforcement Learning
Chris Xiaoxuan Lu
Yannick Schroecker
Albert Gu
Emilio Parisotto
Jakob N. Foerster
Satinder Singh
Feryal M. P. Behbahani
AI4TS
97
83
0
07 Mar 2023
Seq2Seq Imitation Learning for Tactile Feedback-based Manipulation
Seq2Seq Imitation Learning for Tactile Feedback-based Manipulation
Wenyan Yang
A. Angleraud
R. Pieters
Joni Pajarinen
Joni-Kristian Kämäräinen
32
6
0
05 Mar 2023
POPGym: Benchmarking Partially Observable Reinforcement Learning
POPGym: Benchmarking Partially Observable Reinforcement Learning
Steven D. Morad
Ryan Kortvelesy
Matteo Bettini
Stephan Liwicki
Amanda Prorok
OffRL
22
38
0
03 Mar 2023
Graph schemas as abstractions for transfer learning, inference, and
  planning
Graph schemas as abstractions for transfer learning, inference, and planning
J. S. Guntupalli
Rajkumar Vasudeva Raju
Shrinu Kushagra
Carter Wendelken
Daniel P. Sawyer
Ishani Deshpande
Guangyao Zhou
Miguel Lazaro-Gredilla
Dileep George
37
9
0
14 Feb 2023
Don't Watch Me: A Spatio-Temporal Trojan Attack on
  Deep-Reinforcement-Learning-Augment Autonomous Driving
Don't Watch Me: A Spatio-Temporal Trojan Attack on Deep-Reinforcement-Learning-Augment Autonomous Driving
Yinbo Yu
Jiajia Liu
21
1
0
22 Nov 2022
Leveraging Fully Observable Policies for Learning under Partial
  Observability
Leveraging Fully Observable Policies for Learning under Partial Observability
Hai V. Nguyen
Andrea Baisero
Dian Wang
Chris Amato
Robert W. Platt
OffRL
30
19
0
03 Nov 2022
In-context Reinforcement Learning with Algorithm Distillation
In-context Reinforcement Learning with Algorithm Distillation
Michael Laskin
Luyu Wang
Junhyuk Oh
Emilio Parisotto
Stephen Spencer
...
Ethan A. Brooks
Maxime Gazeau
Himanshu Sahni
Satinder Singh
Volodymyr Mnih
OffRL
29
121
0
25 Oct 2022
Time-Varying Propensity Score to Bridge the Gap between the Past and
  Present
Time-Varying Propensity Score to Bridge the Gap between the Past and Present
Rasool Fakoor
Jonas W. Mueller
Zachary Chase Lipton
Pratik Chaudhari
Alexander J. Smola
OOD
AI4TS
32
3
0
04 Oct 2022
Meta-Learning Parameterized Skills
Meta-Learning Parameterized Skills
Haotian Fu
Shangqun Yu
Saket Tiwari
Michael Littman
George Konidaris
38
6
0
07 Jun 2022
Learning in Observable POMDPs, without Computationally Intractable
  Oracles
Learning in Observable POMDPs, without Computationally Intractable Oracles
Noah Golowich
Ankur Moitra
Dhruv Rohatgi
32
26
0
07 Jun 2022
Deep Transformer Q-Networks for Partially Observable Reinforcement
  Learning
Deep Transformer Q-Networks for Partially Observable Reinforcement Learning
Kevin Esslinger
Robert W. Platt
Chris Amato
OffRL
32
35
0
02 Jun 2022
Disentangling Abstraction from Statistical Pattern Matching in Human and
  Machine Learning
Disentangling Abstraction from Statistical Pattern Matching in Human and Machine Learning
Sreejan Kumar
Ishita Dasgupta
Nathaniel D. Daw
Jonathan D. Cohen
Thomas L. Griffiths
32
10
0
04 Apr 2022
Robust Reinforcement Learning on State Observations with Learned Optimal
  Adversary
Robust Reinforcement Learning on State Observations with Learned Optimal Adversary
Huan Zhang
Hongge Chen
Duane S. Boning
Cho-Jui Hsieh
67
162
0
21 Jan 2021
Learning Guidance Rewards with Trajectory-space Smoothing
Learning Guidance Rewards with Trajectory-space Smoothing
Tanmay Gangwani
Yuanshuo Zhou
Jian Peng
26
33
0
23 Oct 2020
Soft Actor-Critic for Discrete Action Settings
Soft Actor-Critic for Discrete Action Settings
Petros Christodoulou
OffRL
104
292
0
16 Oct 2019
Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks
Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks
Chelsea Finn
Pieter Abbeel
Sergey Levine
OOD
365
11,700
0
09 Mar 2017
Previous
12