Mean-Field Sampling for Cooperative Multi-Agent Reinforcement Learning

Abstract
Designing efficient algorithms for multi-agent reinforcement learning (MARL) is fundamentally challenging because the size of the joint state and action spaces grows exponentially in the number of agents. These difficulties are exacerbated when balancing sequential global decision-making with local agent interactions. In this work, we propose a new algorithm (-ean-ield--learning) and a decentralized randomized policy for a system with agents. For any , our algorithm learns a policy for the system in time polynomial in . We prove that this learned policy converges to the optimal policy on the order of as the number of subsampled agents increases. In particular, this bound is independent of the number of agents .
View on arXiv@article{anand2025_2412.00661, title={ Mean-Field Sampling for Cooperative Multi-Agent Reinforcement Learning }, author={ Emile Anand and Ishani Karmarkar and Guannan Qu }, journal={arXiv preprint arXiv:2412.00661}, year={ 2025 } }
Comments on this paper