Population-aware Online Mirror Descent for Mean-Field Games with Common Noise by Deep Reinforcement Learning

Mean Field Games (MFGs) offer a powerful framework for studying large-scale multi-agent systems. Yet, learning Nash equilibria in MFGs remains a challenging problem, particularly when the initial distribution is unknown or when the population is subject to common noise. In this paper, we introduce a...

Full description

Saved in:
Bibliographic Details
Main Authors Wu, Zida, Lauriere, Mathieu, Geist, Matthieu, Pietquin, Olivier, Mehta, Ankur
Format Journal Article
LanguageEnglish
Published 03.09.2025
Subjects
Online AccessGet full text
DOI10.48550/arxiv.2509.03030

Cover

More Information
Summary:Mean Field Games (MFGs) offer a powerful framework for studying large-scale multi-agent systems. Yet, learning Nash equilibria in MFGs remains a challenging problem, particularly when the initial distribution is unknown or when the population is subject to common noise. In this paper, we introduce an efficient deep reinforcement learning (DRL) algorithm designed to achieve population-dependent Nash equilibria without relying on averaging or historical sampling, inspired by Munchausen RL and Online Mirror Descent. The resulting policy is adaptable to various initial distributions and sources of common noise. Through numerical experiments on seven canonical examples, we demonstrate that our algorithm exhibits superior convergence properties compared to state-of-the-art algorithms, particularly a DRL version of Fictitious Play for population-dependent policies. The performance in the presence of common noise underscores the robustness and adaptability of our approach.
DOI:10.48550/arxiv.2509.03030