ResearchTrend.AI
  • Papers
  • Communities
  • Organizations
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2010.01404
428
0
v1v2v3 (latest)

Direct Expected Quadratic Utility Maximization for Mean-Variance Controlled Reinforcement Learning

3 October 2020
Masahiro Kato
Kei Nakagawa
ArXiv (abs)PDFHTML
Abstract

Risk management is critical in decision-making, and mean-variance (MV) trade-off is one of the most common criteria. However, in reinforcement learning (RL) under a dynamic environment, MV control is not as easy as that under a static environment owing to computational difficulties. For MV controlled RL, this paper proposes direct expected quadratic utility maximization (EQUM), where a mean-variance efficient agent is given as its solution. This approach does not only avoid computational difficulties but also improves empirical performances. In experiments, we demonstrate the effectiveness of the proposed EQUM with benchmark settings.

View on arXiv
Comments on this paper