ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2408.02295
39
0

Generalized Gaussian Temporal Difference Error for Uncertainty-aware Reinforcement Learning

5 August 2024
Seyeon Kim
Joonhun Lee
Namhoon Cho
Sungjun Han
Seungeon Baek
ArXivPDFHTML
Abstract

Conventional uncertainty-aware temporal difference (TD) learning often assumes a zero-mean Gaussian distribution for TD errors, leading to inaccurate error representations and compromised uncertainty estimation. We introduce a novel framework for generalized Gaussian error modeling in deep reinforcement learning to enhance the flexibility of error distribution modeling by incorporating additional higher-order moment, particularly kurtosis, thereby improving the estimation and mitigation of data-dependent aleatoric uncertainty. We examine the influence of the shape parameter of the generalized Gaussian distribution (GGD) on aleatoric uncertainty and provide a closed-form expression that demonstrates an inverse relationship between uncertainty and the shape parameter. Additionally, we propose a theoretically grounded weighting scheme to address epistemic uncertainty by fully leveraging the GGD. We refine batch inverse variance weighting with bias reduction and kurtosis considerations, enhancing robustness. Experiments with policy gradient algorithms demonstrate significant performance gains.

View on arXiv
@article{kim2025_2408.02295,
  title={ Generalized Gaussian Temporal Difference Error for Uncertainty-aware Reinforcement Learning },
  author={ Seyeon Kim and Joonhun Lee and Namhoon Cho and Sungjun Han and Wooseop Hwang },
  journal={arXiv preprint arXiv:2408.02295},
  year={ 2025 }
}
Comments on this paper