MEAL: A Benchmark for Continual Multi-Agent Reinforcement Learning
Tristan Tomilin,
Luka Boogaard,
Samuel Garcin,
Constantin Ruhdorfer,
Bram Grooten,
Fabrice Kusters,
Yali Du,
Andreas Bulling,
Mykola Pechenizkiy,
Meng Fang
Proc. International Conference on Machine Learning (ICML),
2026.
Abstract
BibTeX
Project
Benchmarks play a central role in reinforcement learning (RL) research, yet their computational constraints often shape what is studied. Despite the motivation of lifelong learning, most continual RL papers consider only 3-10 sequential tasks, as CPU-bound environments make longer sequences impractical. Meanwhile, continual learning in cooperative multi-agent settings remains largely unexplored. To address these gaps, we introduce MEAL (Multi-agent Environments for Adaptive Learning), the first benchmark for continual multi-agent RL. By leveraging JAX and GPU acceleration, MEAL enables training on sequences of 100 tasks on a single GPU in a few hours. We find that long task sequences reveal failure modes that do not appear at smaller scales.
@inproceedings{tomilin26_icml,
title = {MEAL: A Benchmark for Continual Multi-Agent Reinforcement Learning},
author = {Tomilin, Tristan and van den Boogaard, Luka and Garcin, Samuel and Ruhdorfer, Constantin and Grooten, Bram and Kusters, Fabrice and Du, Yali and Bulling, Andreas and Pechenizkiy, Mykola and Fang, Meng},
year = {2026},
booktitle = {Proc. International Conference on Machine Learning (ICML)}
}