On the Convergence of Model Free Learning in Mean Field Games

被引:0
|
作者
Elie, Romuald [1 ]
Perolat, Julien [2 ]
Lauriere, Mathieu [3 ]
Geist, Matthieu [4 ]
Pietquin, Olivier [4 ]
机构
[1] Univ Paris Est, Paris, France
[2] Deepmind, London, England
[3] Princeton Univ, ORFE, Princeton, NJ 08544 USA
[4] Google Res, Brain Team, New York, NY USA
关键词
RATES;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Learning by experience in Multi-Agent Systems (MAS) is a difficult and exciting task, due to the lack of stationarity of the environment, whose dynamics evolves as the population learns. In order to design scalable algorithms for systems with a large population of interacting agents (e.g., swarms), this paper focuses on Mean Field MAS, where the number of agents is asymptotically infinite. Recently, a very active burgeoning field studies the effects of diverse reinforcement learning algorithms for agents with no prior information on a stationary Mean Field Game (MFG) and learn their policy through repeated experience. We adopt a high perspective on this problem and analyze in full generality the convergence of a fictitious iterative scheme using any single agent learning algorithm at each step. We quantify the quality of the computed approximate Nash equilibrium, in terms of the accumulated errors arising at each learning iteration step. Notably, we show for the first time convergence of model free learning algorithms towards non-stationary MFG equilibria, relying only on classical assumptions on the MFG dynamics. We illustrate our theoretical results with a numerical experiment in a continuous action-space environment, where the approximate best response of the iterative fictitious play scheme is computed with a deep RL algorithm.
引用
收藏
页码:7143 / 7150
页数:8
相关论文
共 50 条
  • [1] Model-Free Reinforcement Learning for Mean Field Games
    Mishra, Rajesh
    Vasal, Deepanshu
    Vishwanath, Sriram
    IEEE TRANSACTIONS ON CONTROL OF NETWORK SYSTEMS, 2023, 10 (04): : 2141 - 2151
  • [2] Learning While Playing in Mean-Field Games: Convergence and Optimality
    Xie, Qiaomin
    Yang, Zhuoran
    Wang, Zhaoran
    Minca, Andreea
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [3] Model-free Reinforcement Learning for Non-stationary Mean Field Games
    Mishra, Rajesh K.
    Vasal, Deepanshu
    Vishwanath, Sriram
    2020 59TH IEEE CONFERENCE ON DECISION AND CONTROL (CDC), 2020, : 1032 - 1037
  • [4] Mean field stochastic games: convergence, Q/H-learning and optimality
    Tembine, Hamidou
    2011 AMERICAN CONTROL CONFERENCE, 2011, : 2423 - 2428
  • [5] ON THE CONVERGENCE PROBLEM IN MEAN FIELD GAMES: A TWO STATE MODEL WITHOUT UNIQUENESS
    Cecchin, Alekos
    Pra, Paolo Dai
    Fischer, Markus
    Pelino, Guglielmo
    SIAM JOURNAL ON CONTROL AND OPTIMIZATION, 2019, 57 (04) : 2443 - 2466
  • [6] On the convergence of finite state mean-field games through Γ-convergence
    Ferreira, Rita
    Comes, Diogo A.
    JOURNAL OF MATHEMATICAL ANALYSIS AND APPLICATIONS, 2014, 418 (01) : 211 - 230
  • [7] Learning Mean-Field Games
    Guo, Xin
    Hu, Anran
    Xu, Renyuan
    Zhang, Junzi
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [8] Learning in Mean-Field Games
    Yin, Huibing
    Mehta, Prashant G.
    Meyn, Sean P.
    Shanbhag, Uday V.
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2014, 59 (03) : 629 - 644
  • [9] MEAN FIELD GAMES OF CONTROLS: ON THE CONVERGENCE OF NASH EQUILIBRIA
    Djete, Mao Fabrice
    ANNALS OF APPLIED PROBABILITY, 2023, 33 (04): : 2824 - 2862
  • [10] MEAN FIELD GAMES: CONVERGENCE OF A FINITE DIFFERENCE METHOD
    Achdou, Yves
    Camilli, Fabio
    Capuzzo-Dolcetta, Italo
    SIAM JOURNAL ON NUMERICAL ANALYSIS, 2013, 51 (05) : 2585 - 2612