On the Convergence of Model Free Learning in Mean Field Games

被引:0
|
作者
Elie, Romuald [1 ]
Perolat, Julien [2 ]
Lauriere, Mathieu [3 ]
Geist, Matthieu [4 ]
Pietquin, Olivier [4 ]
机构
[1] Univ Paris Est, Paris, France
[2] Deepmind, London, England
[3] Princeton Univ, ORFE, Princeton, NJ 08544 USA
[4] Google Res, Brain Team, New York, NY USA
关键词
RATES;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Learning by experience in Multi-Agent Systems (MAS) is a difficult and exciting task, due to the lack of stationarity of the environment, whose dynamics evolves as the population learns. In order to design scalable algorithms for systems with a large population of interacting agents (e.g., swarms), this paper focuses on Mean Field MAS, where the number of agents is asymptotically infinite. Recently, a very active burgeoning field studies the effects of diverse reinforcement learning algorithms for agents with no prior information on a stationary Mean Field Game (MFG) and learn their policy through repeated experience. We adopt a high perspective on this problem and analyze in full generality the convergence of a fictitious iterative scheme using any single agent learning algorithm at each step. We quantify the quality of the computed approximate Nash equilibrium, in terms of the accumulated errors arising at each learning iteration step. Notably, we show for the first time convergence of model free learning algorithms towards non-stationary MFG equilibria, relying only on classical assumptions on the MFG dynamics. We illustrate our theoretical results with a numerical experiment in a continuous action-space environment, where the approximate best response of the iterative fictitious play scheme is computed with a deep RL algorithm.
引用
收藏
页码:7143 / 7150
页数:8
相关论文
共 50 条
  • [41] Correlated Equilibria and Mean Field Games: A Simple Model
    Campi, Luciano
    Fischer, Markus
    MATHEMATICS OF OPERATIONS RESEARCH, 2022, 47 (03) : 1 - 20
  • [42] Convergence of one-dimensional stationary mean field games with vanishing potential
    Cai, Yiru
    Qi, Haobo
    Su, Xifeng
    Tan, Yi
    PARTIAL DIFFERENTIAL EQUATIONS AND APPLICATIONS, 2025, 6 (02):
  • [43] Learning Regularized Monotone Graphon Mean-Field Games
    Zhang, Fengzhuo
    Tan, Vincent Y. F.
    Wang, Zhaoran
    Yang, Zhuoran
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [44] Q-Learning in Regularized Mean-field Games
    Berkay Anahtarci
    Can Deha Kariksiz
    Naci Saldi
    Dynamic Games and Applications, 2023, 13 : 89 - 117
  • [45] Q-Learning in Regularized Mean-field Games
    Anahtarci, Berkay
    Kariksiz, Can Deha
    Saldi, Naci
    DYNAMIC GAMES AND APPLICATIONS, 2023, 13 (01) : 89 - 117
  • [46] Learning Mean-Field Games with Discounted and Average Costs
    Anahtarci, Berkay
    Kariksiz, Can Deha
    Saldi, Naci
    JOURNAL OF MACHINE LEARNING RESEARCH, 2023, 24
  • [47] Scalable Deep Reinforcement Learning Algorithms for Mean Field Games
    Lauriere, Mathieu
    Perrin, Sarah
    Girgin, Sertan
    Muller, Paul
    Jain, Ayush
    Cabannes, Theophile
    Piliouras, Georgios
    Perolat, Julien
    Elie, Romuald
    Pietquin, Olivier
    Geist, Matthieu
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [48] Generalized Conditional Gradient and Learning in Potential Mean Field Games
    Lavigne P.
    Pfeiffer L.
    Applied Mathematics and Optimization, 2023, 88 (03):
  • [49] Model-free Computation Method in First-order Linear Quadratic Mean Field Games
    Xu, Zhenhui
    Shen, Tielong
    IFAC PAPERSONLINE, 2023, 56 (02): : 888 - 893
  • [50] CONVERGENCE ANALYSIS OF MACHINE LEARNING ALGORITHMS FOR THE NUMERICAL SOLUTION OF MEAN FIELD CONTROL AND GAMES: II-THE FINITE HORIZON CASE
    Carmona, Rene
    Lauriere, Mathieu
    ANNALS OF APPLIED PROBABILITY, 2022, 32 (06): : 4065 - 4105