ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2410.12197
  4. Cited By
Potential-Based Intrinsic Motivation: Preserving Optimality With
  Complex, Non-Markovian Shaping Rewards

Potential-Based Intrinsic Motivation: Preserving Optimality With Complex, Non-Markovian Shaping Rewards

16 October 2024
Grant C. Forbes
Leonardo Villalobos-Arias
Jianxun Wang
Arnav Jhala
David L. Roberts
ArXiv (abs)PDFHTML

Papers citing "Potential-Based Intrinsic Motivation: Preserving Optimality With Complex, Non-Markovian Shaping Rewards"

12 / 12 papers shown
Title
Potential-Based Reward Shaping For Intrinsic Motivation
Potential-Based Reward Shaping For Intrinsic Motivation
Grant C. Forbes
Nitish Gupta
Leonardo Villalobos-Arias
Colin M. Potts
Arnav Jhala
David L. Roberts
10
5
0
12 Feb 2024
Redeeming Intrinsic Rewards via Constrained Optimization
Redeeming Intrinsic Rewards via Constrained Optimization
Eric Chen
Zhang-Wei Hong
Joni Pajarinen
Pulkit Agrawal
OnRL
77
26
0
14 Nov 2022
RIDE: Rewarding Impact-Driven Exploration for Procedurally-Generated
  Environments
RIDE: Rewarding Impact-Driven Exploration for Procedurally-Generated Environments
Roberta Raileanu
Tim Rocktaschel
81
174
0
27 Feb 2020
Dota 2 with Large Scale Deep Reinforcement Learning
Dota 2 with Large Scale Deep Reinforcement Learning
OpenAI OpenAI
:
Christopher Berner
Greg Brockman
Brooke Chan
...
Szymon Sidor
Ilya Sutskever
Jie Tang
Filip Wolski
Susan Zhang
GNNVLMCLLAI4CELRM
169
1,838
0
13 Dec 2019
Optimal Policies Tend to Seek Power
Optimal Policies Tend to Seek Power
Alexander Matt Turner
Logan Smith
Rohin Shah
Andrew Critch
Prasad Tadepalli
79
71
0
03 Dec 2019
Using Natural Language for Reward Shaping in Reinforcement Learning
Using Natural Language for Reward Shaping in Reinforcement Learning
Prasoon Goyal
S. Niekum
Raymond J. Mooney
LM&Ro
97
183
0
05 Mar 2019
Proximal Policy Optimization Algorithms
Proximal Policy Optimization Algorithms
John Schulman
Filip Wolski
Prafulla Dhariwal
Alec Radford
Oleg Klimov
OffRL
547
19,296
0
20 Jul 2017
Hindsight Experience Replay
Hindsight Experience Replay
Marcin Andrychowicz
Dwight Crow
Alex Ray
Jonas Schneider
Rachel Fong
Peter Welinder
Bob McGrew
Joshua Tobin
Pieter Abbeel
Wojciech Zaremba
OffRL
282
2,339
0
05 Jul 2017
Curiosity-driven Exploration by Self-supervised Prediction
Curiosity-driven Exploration by Self-supervised Prediction
Deepak Pathak
Pulkit Agrawal
Alexei A. Efros
Trevor Darrell
LRMSSL
125
2,453
0
15 May 2017
Concrete Problems in AI Safety
Concrete Problems in AI Safety
Dario Amodei
C. Olah
Jacob Steinhardt
Paul Christiano
John Schulman
Dandelion Mané
253
2,405
0
21 Jun 2016
Unifying Count-Based Exploration and Intrinsic Motivation
Unifying Count-Based Exploration and Intrinsic Motivation
Marc G. Bellemare
S. Srinivasan
Georg Ostrovski
Tom Schaul
D. Saxton
Rémi Munos
184
1,484
0
06 Jun 2016
Potential-Based Shaping and Q-Value Initialization are Equivalent
Potential-Based Shaping and Q-Value Initialization are Equivalent
Eric Wiewiora
OffRL
76
179
0
26 Jun 2011
1