The Gradient Convergence Bound of Federated Multi-Agent Reinforcement Learning With Efficient Communication

被引:4
|
作者
Xu, Xing [1 ]
Li, Rongpeng [1 ]
Zhao, Zhifeng [2 ,3 ]
Zhang, Honggang [2 ,3 ]
机构
[1] Zhejiang Univ, Hangzhou 310027, Peoples R China
[2] Zhejiang Lab, Hangzhou 311121, Peoples R China
[3] Zhejiang Univ, Coll Informat Sci & Elect Engn, Hangzhou 310027, Peoples R China
基金
中国国家自然科学基金;
关键词
Independent reinforcement learning; federated learning; consensus algorithm; communication overheads;
D O I
10.1109/TWC.2023.3279268
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
The paper considers independent reinforcement learning (IRL) for multi-agent collaborative decision-making in the paradigm of federated learning (FL). However, FL generates excessive communication overheads between agents and a remote central server, especially when it involves a large number of agents or iterations. Besides, due to the heterogeneity of independent learning environments, multiple agents may undergo asynchronous Markov decision processes (MDPs), which will affect the training samples and the model's convergence performance. On top of the variation-aware periodic averaging (VPA) method and the policy-based deep reinforcement learning (DRL) algorithm (i.e., proximal policy optimization (PPO)), this paper proposes two advanced optimization schemes orienting to stochastic gradient descent (SGD): 1) A decay-based scheme gradually decays the weights of a model's local gradients with the progress of successive local updates, and 2) By representing the agents as a graph, a consensus-based scheme studies the impact of exchanging a model's local gradients among nearby agents from an algebraic connectivity perspective. This paper also provides novel convergence guarantees for both developed schemes, and demonstrates their superior effectiveness and efficiency in improving the system's utility value through theoretical analyses and simulation results.
引用
收藏
页码:507 / 528
页数:22
相关论文
共 50 条
  • [31] Semantic Communication for Partial Observation Multi-agent Reinforcement Learning
    Do, Hoang Khoi
    Dinh, Thi Quynh
    Nguyen, Minh Duong
    Nguyen, Tien Hoa
    2023 IEEE STATISTICAL SIGNAL PROCESSING WORKSHOP, SSP, 2023, : 319 - 323
  • [32] Optimistic sequential multi-agent reinforcement learning with motivational communication
    Huang, Anqi
    Wang, Yongli
    Zhou, Xiaoliang
    Zou, Haochen
    Dong, Xu
    Che, Xun
    NEURAL NETWORKS, 2024, 179
  • [33] Advancing Multi-Agent Systems Integrating Federated Learning with Deep Reinforcement Learning : A Survey
    Kim, Jaemin
    Kim, Gahyun
    Hong, Seonghun
    Cho, Sungrae
    2024 FIFTEENTH INTERNATIONAL CONFERENCE ON UBIQUITOUS AND FUTURE NETWORKS, ICUFN 2024, 2024, : 55 - 57
  • [34] Deep Hierarchical Communication Graph in Multi-Agent Reinforcement Learning
    Liu, Zeyang
    Wan, Lipeng
    Sui, Xue
    Chen, Zhuoran
    Sun, Kewu
    Lan, Xuguang
    PROCEEDINGS OF THE THIRTY-SECOND INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2023, 2023, : 208 - 216
  • [35] Cooperative Multi-agent Reinforcement Learning with Hierachical Communication Architecture
    Liu, Shifan
    Yuan, Quan
    Chen, Bo
    Luo, Guiyang
    Li, Jinglin
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2022, PT II, 2022, 13530 : 14 - 25
  • [36] Learning Efficient Communication in Cooperative Multi-Agent Environment
    Zhao, Yuhang
    Ma, Xiujun
    AAMAS '19: PROCEEDINGS OF THE 18TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS, 2019, : 2321 - 2323
  • [37] Federated Dynamic Spectrum Access through Multi-Agent Deep Reinforcement Learning
    Song, Yifei
    Chang, Hao-Hsuan
    Liu, Lingjia
    2022 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM 2022), 2022, : 3466 - 3471
  • [38] Scalable order dispatching through Federated Multi-Agent Deep Reinforcement Learning
    Jing, Yao
    Guo, Bin
    Li, Nuo
    Ding, Yasan
    Liu, Yan
    Yu, Zhiwen
    EXPERT SYSTEMS WITH APPLICATIONS, 2025, 264
  • [39] Efficient multi-agent cooperation: Scalable reinforcement learning with heterogeneous graph networks and limited communication
    Li, Z.
    Yang, Y.
    Cheng, H.
    KNOWLEDGE-BASED SYSTEMS, 2024, 300
  • [40] Determining the Applicability of Advice for Efficient Multi-Agent Reinforcement Learning
    Wang, Yuchen
    Ren, Fenghui
    Zhang, Minjie
    PRICAI 2018: TRENDS IN ARTIFICIAL INTELLIGENCE, PT II, 2018, 11013 : 343 - 351