ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2405.11432
35
4

On Robust Reinforcement Learning with Lipschitz-Bounded Policy Networks

19 May 2024
Nicholas H. Barbara
Ruigang Wang
I. Manchester
ArXivPDFHTML
Abstract

This paper presents a study of robust policy networks in deep reinforcement learning. We investigate the benefits of policy parameterizations that naturally satisfy constraints on their Lipschitz bound, analyzing their empirical performance and robustness on two representative problems: pendulum swing-up and Atari Pong. We illustrate that policy networks with smaller Lipschitz bounds are more robust to disturbances, random noise, and targeted adversarial attacks than unconstrained policies composed of vanilla multi-layer perceptrons or convolutional neural networks. However, the structure of the Lipschitz layer is important. We find that the widely-used method of spectral normalization is too conservative and severely impacts clean performance, whereas more expressive Lipschitz layers such as the recently-proposed Sandwich layer can achieve improved robustness without sacrificing clean performance.

View on arXiv
@article{barbara2025_2405.11432,
  title={ On Robust Reinforcement Learning with Lipschitz-Bounded Policy Networks },
  author={ Nicholas H. Barbara and Ruigang Wang and Ian R. Manchester },
  journal={arXiv preprint arXiv:2405.11432},
  year={ 2025 }
}
Comments on this paper