On the Convergence of Model Free Learning in Mean Field Games

被引:0
|
作者
Elie, Romuald [1 ]
Perolat, Julien [2 ]
Lauriere, Mathieu [3 ]
Geist, Matthieu [4 ]
Pietquin, Olivier [4 ]
机构
[1] Univ Paris Est, Paris, France
[2] Deepmind, London, England
[3] Princeton Univ, ORFE, Princeton, NJ 08544 USA
[4] Google Res, Brain Team, New York, NY USA
关键词
RATES;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Learning by experience in Multi-Agent Systems (MAS) is a difficult and exciting task, due to the lack of stationarity of the environment, whose dynamics evolves as the population learns. In order to design scalable algorithms for systems with a large population of interacting agents (e.g., swarms), this paper focuses on Mean Field MAS, where the number of agents is asymptotically infinite. Recently, a very active burgeoning field studies the effects of diverse reinforcement learning algorithms for agents with no prior information on a stationary Mean Field Game (MFG) and learn their policy through repeated experience. We adopt a high perspective on this problem and analyze in full generality the convergence of a fictitious iterative scheme using any single agent learning algorithm at each step. We quantify the quality of the computed approximate Nash equilibrium, in terms of the accumulated errors arising at each learning iteration step. Notably, we show for the first time convergence of model free learning algorithms towards non-stationary MFG equilibria, relying only on classical assumptions on the MFG dynamics. We illustrate our theoretical results with a numerical experiment in a continuous action-space environment, where the approximate best response of the iterative fictitious play scheme is computed with a deep RL algorithm.
引用
收藏
页码:7143 / 7150
页数:8
相关论文
共 50 条
  • [31] Reinforcement Learning in Stationary Mean-field Games
    Subramanian, Jayakumar
    Mahajan, Aditya
    AAMAS '19: PROCEEDINGS OF THE 18TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS, 2019, : 251 - 259
  • [32] Reinforcement Learning for Mean Field Games with Strategic Complementarities
    Lee, Kiyeob
    Rengarajan, Desik
    Kalathil, Dileep
    Shakkottai, Srinivas
    24TH INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS (AISTATS), 2021, 130
  • [33] A Machine Learning Method for Stackelberg Mean Field Games
    Dayanikli, Gokce
    Lauriere, Mathieu
    MATHEMATICS OF OPERATIONS RESEARCH, 2024,
  • [34] A General Framework for Learning Mean-Field Games
    Guo, Xin
    Hu, Anran
    Xu, Renyuan
    Zhang, Junzi
    MATHEMATICS OF OPERATIONS RESEARCH, 2023, 48 (02) : 656 - 686
  • [35] Independent Learning and Subjectivity in Mean-Field Games
    Yongacoglu, Bora
    Arslan, Gürdal
    Yuksel, Serdar
    2022 IEEE 61ST CONFERENCE ON DECISION AND CONTROL (CDC), 2022, : 2845 - 2850
  • [36] Generalization in Mean Field Games by Learning Master Policies
    Perrin, Sarah
    Lauriere, Mathieu
    Perolat, Julien
    Elie, Romuald
    Geist, Matthieu
    Pietquin, Olivier
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 9413 - 9421
  • [37] CONVERGENCE ANALYSIS OF MACHINE LEARNING ALGORITHMS FOR THE NUMERICAL SOLUTION OF MEAN FIELD CONTROL AND GAMES I: THE ERGODIC CASE
    Carmona, Rene
    Lauriere, Mathieu
    SIAM JOURNAL ON NUMERICAL ANALYSIS, 2021, 59 (03) : 1455 - 1485
  • [38] Mean field games
    Lasry, Jean-Michel
    Lions, Pierre-Louis
    JAPANESE JOURNAL OF MATHEMATICS, 2007, 2 (01): : 229 - 260
  • [39] Mean field games
    Jean-Michel Lasry
    Pierre-Louis Lions
    Japanese Journal of Mathematics, 2007, 2 : 229 - 260
  • [40] A Quadratic Mean Field Games Model for the Langevin Equation
    Camilli, Fabio
    AXIOMS, 2021, 10 (02)