30
0

Optimizing Sensory Neurons: Nonlinear Attention Mechanisms for Accelerated Convergence in Permutation-Invariant Neural Networks for Reinforcement Learning

Main:9 Pages
4 Figures
Bibliography:2 Pages
6 Tables
Abstract

Training reinforcement learning (RL) agents often requires significant computational resources and prolonged training durations. To address this challenge, we build upon prior work that introduced a neural architecture with permutation-invariant sensory processing. We propose a modified attention mechanism that applies a non-linear transformation to the key vectors (K), producing enriched representations (K') through a custom mapping function. This Nonlinear Attention (NLA) mechanism enhances the representational capacity of the attention layer, enabling the agent to learn more expressive feature interactions. As a result, our model achieves significantly faster convergence and improved training efficiency, while maintaining performance on par with the baseline. These results highlight the potential of nonlinear attention mechanisms to accelerate reinforcement learning without sacrificing effectiveness.

View on arXiv
@article{muzaffar2025_2506.00691,
  title={ Optimizing Sensory Neurons: Nonlinear Attention Mechanisms for Accelerated Convergence in Permutation-Invariant Neural Networks for Reinforcement Learning },
  author={ Junaid Muzaffar and Khubaib Ahmed and Ingo Frommholz and Zeeshan Pervez and Ahsan ul Haq },
  journal={arXiv preprint arXiv:2506.00691},
  year={ 2025 }
}
Comments on this paper