Generalization in Mean Field Games by ...
Type de document :
Pré-publication ou Document de travail
Titre :
Generalization in Mean Field Games by Learning Master Policies
Auteur(s) :
Perrin, Sarah [Auteur]
Scool [Scool]
Laurière, Mathieu [Auteur]
Pérolat, Julien [Auteur]
Élie, Romuald [Auteur]
Geist, Matthieu [Auteur]
Pietquin, Olivier [Auteur]
Scool [Scool]
Laurière, Mathieu [Auteur]
Pérolat, Julien [Auteur]
Élie, Romuald [Auteur]
Geist, Matthieu [Auteur]
Pietquin, Olivier [Auteur]

Discipline(s) HAL :
Informatique [cs]/Intelligence artificielle [cs.AI]
Informatique [cs]/Informatique et théorie des jeux [cs.GT]
Informatique [cs]/Système multi-agents [cs.MA]
Informatique [cs]/Réseau de neurones [cs.NE]
Informatique [cs]/Informatique et théorie des jeux [cs.GT]
Informatique [cs]/Système multi-agents [cs.MA]
Informatique [cs]/Réseau de neurones [cs.NE]
Résumé en anglais : [en]
Mean Field Games (MFGs) can potentially scale multi-agent systems to extremely large populations of agents. Yet, most of the literature assumes a single initial distribution for the agents, which limits the practical ...
Lire la suite >Mean Field Games (MFGs) can potentially scale multi-agent systems to extremely large populations of agents. Yet, most of the literature assumes a single initial distribution for the agents, which limits the practical applications of MFGs. Machine Learning has the potential to solve a wider diversity of MFG problems thanks to generalizations capacities. We study how to leverage these generalization properties to learn policies enabling a typical agent to behave optimally against any population distribution. In reference to the Master equation in MFGs, we coin the term ``Master policies'' to describe them and we prove that a single Master policy provides a Nash equilibrium, whatever the initial distribution. We propose a method to learn such Master policies. Our approach relies on three ingredients: adding the current population distribution as part of the observation, approximating Master policies with neural networks, and training via Reinforcement Learning and Fictitious Play. We illustrate on numerical examples not only the efficiency of the learned Master policy but also its generalization capabilities beyond the distributions used for training.Lire moins >
Lire la suite >Mean Field Games (MFGs) can potentially scale multi-agent systems to extremely large populations of agents. Yet, most of the literature assumes a single initial distribution for the agents, which limits the practical applications of MFGs. Machine Learning has the potential to solve a wider diversity of MFG problems thanks to generalizations capacities. We study how to leverage these generalization properties to learn policies enabling a typical agent to behave optimally against any population distribution. In reference to the Master equation in MFGs, we coin the term ``Master policies'' to describe them and we prove that a single Master policy provides a Nash equilibrium, whatever the initial distribution. We propose a method to learn such Master policies. Our approach relies on three ingredients: adding the current population distribution as part of the observation, approximating Master policies with neural networks, and training via Reinforcement Learning and Fictitious Play. We illustrate on numerical examples not only the efficiency of the learned Master policy but also its generalization capabilities beyond the distributions used for training.Lire moins >
Langue :
Anglais
Collections :
Source :
Fichiers
- http://arxiv.org/pdf/2109.09717
- Accès libre
- Accéder au document
- 2109.09717
- Accès libre
- Accéder au document