56
0

Multi-Agent Reinforcement Learning-based Cooperative Autonomous Driving in Smart Intersections

Abstract

Unsignalized intersections pose significant safety and efficiency challenges due to complex traffic flows. This paper proposes a novel roadside unit (RSU)-centric cooperative driving system leveraging global perception and vehicle-to-infrastructure (V2I) communication. The core of the system is an RSU-based decision-making module using a two-stage hybrid reinforcement learning (RL) framework. At first, policies are pre-trained offline using conservative Q-learning (CQL) combined with behavior cloning (BC) on collected dataset. Subsequently, these policies are fine-tuned in the simulation using multi-agent proximal policy optimization (MAPPO), aligned with a self-attention mechanism to effectively solve inter-agent dependencies. RSUs perform real-time inference based on the trained models to realize vehicle control via V2I communications. Extensive experiments in CARLA environment demonstrate high effectiveness of the proposed system, by: \textit{(i)} achieving failure rates below 0.03\% in coordinating three connected and autonomous vehicles (CAVs) through complex intersection scenarios, significantly outperforming the traditional Autoware control method, and \textit{(ii)} exhibiting strong robustness across varying numbers of controlled agents and shows promising generalization capabilities on other maps.

View on arXiv
@article{yu2025_2505.04231,
  title={ Multi-Agent Reinforcement Learning-based Cooperative Autonomous Driving in Smart Intersections },
  author={ Taoyuan Yu and Kui Wang and Zongdian Li and Tao Yu and Kei Sakaguchi },
  journal={arXiv preprint arXiv:2505.04231},
  year={ 2025 }
}
Comments on this paper