Online Policy Iteration Solution for Dynamic Graphical Games

被引:0
|
作者
Abouheaf, Mohammed I. [1 ]
Mahmoud, Magdi S. [1 ]
机构
[1] King Fahd Univ Petr & Minerals, Dept Syst Engn, Dhahran, Saudi Arabia
关键词
Dynamic Games; Optimal Control; Game Theory; Cooperative Control; ADAPTIVE LEARNING SOLUTION; CONSENSUS; SYNCHRONIZATION; SYSTEMS;
D O I
暂无
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
The dynamic graphical game is a special class of the standard dynamic game and explicitly captures the structure of a communication graph, where the information flow between the agents is governed by the communication graph topology. A novel online adaptive learning (policy iteration) solution for the graphical game is given in terms of the solution to a set of coupled graphical game Hamiltonian and Bellman equations. The policy iteration solution is developed to learn Nash solution for the dynamic graphical game online in real-time. Policy iteration convergence proof for the dynamic graphical game is given under mild condition about the graph interconnectivity properties. Critic neural network structures are used to implement the online policy iteration solution. Only partial knowledge of the dynamics is required and the tuning is done in a distributed fashion in terms of the local information available to each agent.
引用
收藏
页码:787 / 797
页数:11
相关论文
共 50 条
  • [21] Discrete-time dynamic graphical games: model-free reinforcement learning solution
    Abouheaf M.I.
    Lewis F.L.
    Mahmoud M.S.
    Mikulski D.G.
    Control theory technol., 1 (55-69): : 55 - 69
  • [22] Discrete-time dynamic graphical games:model-free reinforcement learning solution
    Mohammed I.ABOUHEAF
    Frank L.LEWIS
    Magdi S.MAHMOUD
    Dariusz G.MIKULSKI
    Control Theory and Technology, 2015, 13 (01) : 55 - 69
  • [23] Solving multichain stochastic games with mean payoff by policy iteration
    Akian, Marianne
    Cochet-Terrasson, Jean
    Detournay, Sylvie
    Gaubert, Stephane
    2013 IEEE 52ND ANNUAL CONFERENCE ON DECISION AND CONTROL (CDC), 2013, : 1834 - 1841
  • [24] Solving Common-Payoff Games with Approximate Policy Iteration
    Sokota, Samuel
    Lockhart, Edward
    Timbers, Finbarr
    Davoodi, Elnaz
    D'Orazio, Ryan
    Burch, Neil
    Schmid, Martin
    Bowling, Michael
    Lanctot, Marc
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 9695 - 9703
  • [25] Smooth policy iteration for zero-sum Markov Games
    Lyu, Yao
    Wang, Wenxuan
    Li, Shengbo Eben
    Li, Zeyang
    Duan, Jingliang
    NEUROCOMPUTING, 2025, 630
  • [26] Policy Iteration Adaptive Dynamic Programming for Optimal Control of Multi-Player Stackelberg-Nash Games
    Lin, Mingduo
    Zhao, Bo
    Liu, Derong
    Zhang, Yongwei
    2022 41ST CHINESE CONTROL CONFERENCE (CCC), 2022, : 2393 - 2397
  • [27] Empirical Policy Iteration for Approximate Dynamic Programming
    Haskell, William B.
    Jain, Rahul
    Kalathil, Dileep
    2014 IEEE 53RD ANNUAL CONFERENCE ON DECISION AND CONTROL (CDC), 2014, : 6573 - 6578
  • [28] Distributed Adaptive Nash Equilibrium Solution for Differential Graphical Games
    Qian, Yang-Yang
    Liu, Mushuang
    Wan, Yan
    Lewis, Frank L.
    Davoudi, Ali
    IEEE TRANSACTIONS ON CYBERNETICS, 2023, 53 (04) : 2275 - 2287
  • [29] Solution of 3 x 3 games using graphical method
    Nair, KGK
    Ranjith, G
    EUROPEAN JOURNAL OF OPERATIONAL RESEARCH, 1999, 112 (02) : 472 - 478
  • [30] An accelerated value/policy iteration scheme for optimal control problems and games
    Alla, Alessandro
    Falcone, Maurizio
    Kalise, Dante
    Lecture Notes in Computational Science and Engineering, 2015, 103 : 489 - 497