29
0

When Maximum Entropy Misleads Policy Optimization

Main:8 Pages
15 Figures
Bibliography:3 Pages
4 Tables
Appendix:11 Pages
Abstract

The Maximum Entropy Reinforcement Learning (MaxEnt RL) framework is a leading approach for achieving efficient learning and robust performance across many RL tasks. However, MaxEnt methods have also been shown to struggle with performance-critical control problems in practice, where non-MaxEnt algorithms can successfully learn. In this work, we analyze how the trade-off between robustness and optimality affects the performance of MaxEnt algorithms in complex control tasks: while entropy maximization enhances exploration and robustness, it can also mislead policy optimization, leading to failure in tasks that require precise, low-entropy policies. Through experiments on a variety of control problems, we concretely demonstrate this misleading effect. Our analysis leads to better understanding of how to balance reward design and entropy maximization in challenging control problems.

View on arXiv
@article{zhang2025_2506.05615,
  title={ When Maximum Entropy Misleads Policy Optimization },
  author={ Ruipeng Zhang and Ya-Chien Chang and Sicun Gao },
  journal={arXiv preprint arXiv:2506.05615},
  year={ 2025 }
}
Comments on this paper