5
0

MEAL: A Benchmark for Continual Multi-Agent Reinforcement Learning

Tristan Tomilin
Luka van den Boogaard
Samuel Garcin
Bram Grooten
Meng Fang
Mykola Pechenizkiy
Main:8 Pages
10 Figures
Bibliography:3 Pages
5 Tables
Appendix:7 Pages
Abstract

Benchmarks play a crucial role in the development and analysis of reinforcement learning (RL) algorithms, with environment availability strongly impacting research. One particularly underexplored intersection is continual learning (CL) in cooperative multi-agent settings. To remedy this, we introduce MEAL (Multi-agent Environments for Adaptive Learning), the first benchmark tailored for continual multi-agent reinforcement learning (CMARL). Existing CL benchmarks run environments on the CPU, leading to computational bottlenecks and limiting the length of task sequences. MEAL leverages JAX for GPU acceleration, enabling continual learning across sequences of 100 tasks on a standard desktop PC in a few hours. We show that naively combining popular CL and MARL methods yields strong performance on simple environments, but fails to scale to more complex settings requiring sustained coordination and adaptation. Our ablation study identifies architectural and algorithmic features critical for CMARL on MEAL.

View on arXiv
@article{tomilin2025_2506.14990,
  title={ MEAL: A Benchmark for Continual Multi-Agent Reinforcement Learning },
  author={ Tristan Tomilin and Luka van den Boogaard and Samuel Garcin and Bram Grooten and Meng Fang and Mykola Pechenizkiy },
  journal={arXiv preprint arXiv:2506.14990},
  year={ 2025 }
}
Comments on this paper