Cooperative Multi-Agent Q-Learning Using Distributed MPC

被引:2
|
作者
Esfahani, Hossein Nejatbakhsh [1 ]
Velni, Javad Mohammadpour [1 ]
机构
[1] Clemson Univ, Dept Mech Engn, Clemson, SC 29634 USA
来源
基金
美国国家科学基金会;
关键词
Q-learning; Approximation algorithms; Couplings; Costs; Predictive control; Multi-agent systems; Linear programming; Multi-agent Q-Learning; distributed MPC; cooperative control;
D O I
10.1109/LCSYS.2024.3407632
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this letter, we propose a cooperative Multi-Agent Reinforcement Learning (MARL) approach based on Distributed Model Predictive Control (DMPC). In the proposed framework, the local MPC schemes are formulated based on the dual decomposition method in the context of DMPC and will be used to derive the local state (and action) value functions required in a cooperative Q-learning algorithm. We further show that the DMPC scheme can yield a framework based on the Value Function Decomposition (VFD) principle so that the global state (and action) value functions can be decomposed into several local state (and action) value functions captured from the local MPCs. In the proposed cooperative MARL, the coordination between individual agents is then achieved based on the multiplier-sharing step, a.k.a inter-agent negotiation in the DMPC scheme.
引用
收藏
页码:2193 / 2198
页数:6
相关论文
共 50 条
  • [21] Continuous Q-Learning for Multi-Agent Cooperation
    Hwang, Kao-Shing
    Jiang, Wei-Cheng
    Lin, Yu-Hong
    Lai, Li-Hsin
    CYBERNETICS AND SYSTEMS, 2012, 43 (03) : 227 - 256
  • [22] Untangling Braids with Multi-Agent Q-Learning
    Khan, Abdullah
    Vernitski, Alexei
    Lisitsa, Alexei
    2021 23RD INTERNATIONAL SYMPOSIUM ON SYMBOLIC AND NUMERIC ALGORITHMS FOR SCIENTIFIC COMPUTING (SYNASC 2021), 2021, : 135 - 139
  • [23] Q-learning with FCMAC in multi-agent cooperation
    Hwang, Kao-Shing
    Chen, Yu-Jen
    Lin, Tzung-Feng
    ADVANCES IN NEURAL NETWORKS - ISNN 2006, PT 1, 2006, 3971 : 599 - 606
  • [24] Finite-Sample Analysis of Distributed Q-learning for Multi-Agent Networks
    Heredia, Paulo
    Ghadialy, Hasan
    Mou, Shaoshuai
    2020 AMERICAN CONTROL CONFERENCE (ACC), 2020, : 3511 - 3516
  • [25] Multi-agent Q-learning of Spectrum Access in Distributed Cognitive Radio Network
    Min Neng
    Wu Qi-hui
    Xu Yu-hua
    Ding Guo-ru
    INTERNATIONAL CONFERENCE OF CHINA COMMUNICATION (ICCC2010), 2010, : 656 - 660
  • [26] Q-learning based cooperative multi-agent system applied to coordination of overcurrent relays
    Sadeh, J.
    Rahimiyan, M.
    Journal of Applied Sciences, 2008, 8 (21) : 3924 - 3930
  • [27] Q-Learning Policies for Multi-Agent Foraging Task
    Yogeswaran, M.
    Ponnambalam, S. C.
    TRENDS IN INTELLIGENT ROBOTICS, 2010, 103 : 194 - 201
  • [28] Regularized Softmax Deep Multi-Agent Q-Learning
    Pan, Ling
    Rashid, Tabish
    Peng, Bei
    Huang, Longbo
    Whiteson, Shimon
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [29] Multi-Agent Q-Learning for Drone Base Stations
    Janji, Salim
    Kliks, Adrian
    2023 19TH INTERNATIONAL CONFERENCE ON WIRELESS AND MOBILE COMPUTING, NETWORKING AND COMMUNICATIONS, WIMOB, 2023, : 261 - 266
  • [30] Multi-agent crowdsourcing model based on Q-learning
    Fang, Xin
    Guo, Yongan
    2019 IEEE INTERNATIONAL CONFERENCE ON CONSUMER ELECTRONICS - TAIWAN (ICCE-TW), 2019,