On the Convergence of Model Free Learning in Mean Field Games

被引:0
|
作者
Elie, Romuald [1 ]
Perolat, Julien [2 ]
Lauriere, Mathieu [3 ]
Geist, Matthieu [4 ]
Pietquin, Olivier [4 ]
机构
[1] Univ Paris Est, Paris, France
[2] Deepmind, London, England
[3] Princeton Univ, ORFE, Princeton, NJ 08544 USA
[4] Google Res, Brain Team, New York, NY USA
关键词
RATES;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Learning by experience in Multi-Agent Systems (MAS) is a difficult and exciting task, due to the lack of stationarity of the environment, whose dynamics evolves as the population learns. In order to design scalable algorithms for systems with a large population of interacting agents (e.g., swarms), this paper focuses on Mean Field MAS, where the number of agents is asymptotically infinite. Recently, a very active burgeoning field studies the effects of diverse reinforcement learning algorithms for agents with no prior information on a stationary Mean Field Game (MFG) and learn their policy through repeated experience. We adopt a high perspective on this problem and analyze in full generality the convergence of a fictitious iterative scheme using any single agent learning algorithm at each step. We quantify the quality of the computed approximate Nash equilibrium, in terms of the accumulated errors arising at each learning iteration step. Notably, we show for the first time convergence of model free learning algorithms towards non-stationary MFG equilibria, relying only on classical assumptions on the MFG dynamics. We illustrate our theoretical results with a numerical experiment in a continuous action-space environment, where the approximate best response of the iterative fictitious play scheme is computed with a deep RL algorithm.
引用
收藏
页码:7143 / 7150
页数:8
相关论文
共 50 条
  • [21] MODEL-FREE MEAN-FIELD REINFORCEMENT LEARNING: MEAN-FIELD MDP AND MEAN-FIELD Q-LEARNING
    Carmona, Rene
    Lauriere, Mathieu
    Tan, Zongjun
    ANNALS OF APPLIED PROBABILITY, 2023, 33 (6B): : 5334 - 5381
  • [22] Finite Mean Field Games: Fictitious play and convergence to a first order continuous mean field game
    Hadikhanloo, Saeed
    Silva, Francisco J.
    JOURNAL DE MATHEMATIQUES PURES ET APPLIQUEES, 2019, 132 : 369 - 397
  • [23] A Mean Field Games Model for Cryptocurrency Mining
    Li Z.
    Reppen A.M.
    Sircar R.
    Management Science, 2024, 70 (04) : 2188 - 2208
  • [24] A Mean Field Games Model for Cryptocurrency Mining
    Li, Zongxi
    Reppen, A. Max
    Sircar, Ronnie
    MANAGEMENT SCIENCE, 2023,
  • [25] Mean Field LQG Games with Model Uncertainty
    Huang, Jianhui
    Huang, Minyi
    2013 IEEE 52ND ANNUAL CONFERENCE ON DECISION AND CONTROL (CDC), 2013, : 3103 - 3108
  • [26] A MODEL PROBLEM FOR MEAN FIELD GAMES ON NETWORKS
    Camilli, Fabio
    Carlini, Elisabetta
    Marchi, Claudio
    DISCRETE AND CONTINUOUS DYNAMICAL SYSTEMS, 2015, 35 (09) : 4173 - 4192
  • [27] Quantitative Convergence for Displacement Monotone Mean Field Games with Controlled Volatility
    Jackson, Joe
    Tangpib, Ludovic
    MATHEMATICS OF OPERATIONS RESEARCH, 2024, 49 (04) : 2527 - 2564
  • [28] Convergence of some Mean Field Games systems to aggregation and flocking models
    Bardi, Martino
    Cardaliaguet, Pierre
    NONLINEAR ANALYSIS-THEORY METHODS & APPLICATIONS, 2021, 204
  • [29] CLOSED-LOOP CONVERGENCE FOR MEAN FIELD GAMES WITH COMMON NOISE
    Lacker, Daniel
    Le Flem, Luc
    ANNALS OF APPLIED PROBABILITY, 2023, 33 (04): : 2681 - 2733
  • [30] Rates of convergence for the policy iteration method for Mean Field Games systems
    Camilli, Fabio
    Tang, Qing
    JOURNAL OF MATHEMATICAL ANALYSIS AND APPLICATIONS, 2022, 512 (01)