Adaptive Learning: A New Decentralized Reinforcement Learning Approach for Cooperative Multiagent Systems

被引:13
|
作者
Li, Meng-Lin [1 ]
Chen, Shaofei [1 ]
Chen, Jing [1 ]
机构
[1] Natl Univ Def Technol, Coll Intelligence Sci & Technol, Changsha 410073, Peoples R China
来源
IEEE ACCESS | 2020年 / 8卷
基金
中国国家自然科学基金;
关键词
Learning (artificial intelligence); Training; Multi-agent systems; Heuristic algorithms; Roads; Urban areas; Games; Reinforcement learning; multiagent system; intelligent control;
D O I
10.1109/ACCESS.2020.2997899
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Multiagent systems (MASs) have received extensive attention in a variety of domains, such as robotics and distributed control. This paper focuses on how independent learners (ILs, structures used in decentralized reinforcement learning) decide on their individual behaviors to achieve coherent joint behavior. To date, Reinforcement learning(RL) approaches for ILs have not guaranteed convergence to the optimal joint policy in scenarios in which communication is difficult. Especially in a decentralized algorithm, the proportion of credit for a single agent's action in a multiagent system is not distinguished, which can lead to miscoordination of joint actions. Therefore, it is highly significant to study the mechanisms of coordination between agents in MASs. Most previous coordination mechanisms have been carried out by modeling the communication mechanism and other agent policies. These methods are applicable only to a particular system, so such algorithms do not offer generalizability, especially when there are dozens or more agents. Therefore, this paper mainly focuses on the MAS contains more than a dozen agents. By combining the method of parallel computation, the experimental environment is closer to the application scene. By studying the paradigm of centralized training and decentralized execution(CTDE), a multi-agent reinforcement learning algorithm for implicit coordination based on TD error is proposed. The new algorithm can dynamically adjust the learning rate by deeply analyzing the dissonance problem in the matrix game and combining it with a multiagent environment. By adjusting the dynamic learning rate between agents, coordination of the agents' strategies can be achieved. Experimental results show that the proposed algorithm can effectively improve the coordination ability of a MAS. Moreover, the variance of the training results is more stable than that of the hysteretic Q learning(HQL) algorithm. Hence, the problem of miscoordination in a MAS can be avoided to some extent without additional communication. Our work provides a new way to solve the miscoordination problem for reinforcement learning algorithms in the scale of dozens or more number of agents. As a new IL structure algorithm, our results should be extended and further studied.
引用
收藏
页码:99404 / 99421
页数:18
相关论文
共 50 条
  • [21] A REINFORCEMENT LEARNING APPROACH FOR MULTIAGENT NAVIGATION
    Martinez-Gil, Francisco
    Barber, Fernando
    Lozano, Miguel
    Grimaldo, Francisco
    Fernandez, Fernando
    ICAART 2010: PROCEEDINGS OF THE 2ND INTERNATIONAL CONFERENCE ON AGENTS AND ARTIFICIAL INTELLIGENCE, VOL 1: ARTIFICIAL INTELLIGENCE, 2010, : 607 - 610
  • [22] Deep Decentralized Reinforcement Learning for Cooperative Control
    Koepf, Florian
    Tesfazgi, Samuel
    Flad, Michael
    Hohmann, Soeren
    IFAC PAPERSONLINE, 2020, 53 (02): : 1555 - 1562
  • [23] A survey on transfer learning for multiagent reinforcement learning systems
    Da Silva, Felipe Leno
    Reali Costa, Anna Helena
    Journal of Artificial Intelligence Research, 2019, 64 : 645 - 703
  • [24] Coordination in multiagent reinforcement learning systems
    Kamal, MAS
    Murata, J
    KNOWLEDGE-BASED INTELLIGENT INFORMATION AND ENGINEERING SYSTEMS, PT 1, PROCEEDINGS, 2004, 3213 : 1197 - 1204
  • [25] A Survey on Transfer Learning for Multiagent Reinforcement Learning Systems
    Da Silva, Felipe Leno
    Reali Costa, Anna Helena
    JOURNAL OF ARTIFICIAL INTELLIGENCE RESEARCH, 2019, 64 : 645 - 703
  • [26] Adaptive Multiagent Model Based on Reinforcement Learning for Distributed Generation Systems
    Divenyi, Daniel
    Dan, Andras
    2012 23RD INTERNATIONAL WORKSHOP ON DATABASE AND EXPERT SYSTEMS APPLICATIONS (DEXA), 2012, : 303 - 307
  • [27] V-Learning-A Simple, Efficient, Decentralized Algorithm for Multiagent Reinforcement Learning
    Jin, Chi
    Liu, Qinghua
    Wang, Yuanhao
    Yu, Tiancheng
    MATHEMATICS OF OPERATIONS RESEARCH, 2024, 49 (04) : 2295 - 2322
  • [28] CAFEEN: A Cooperative Approach for Energy-Efficient NoCs With Multiagent Reinforcement Learning
    Khan, Kamil
    Pasricha, Sudeep
    IEEE DESIGN & TEST, 2025, 42 (02) : 71 - 78
  • [29] Peer Incentive Reinforcement Learning for Cooperative Multiagent Games
    Zhang, Tianle
    Liu, Zhen
    Pu, Zhiqiang
    Yi, Jianqiang
    IEEE TRANSACTIONS ON GAMES, 2023, 15 (04) : 623 - 636
  • [30] UNMAS: Multiagent Reinforcement Learning for Unshaped Cooperative Scenarios
    Chai, Jiajun
    Li, Weifan
    Zhu, Yuanheng
    Zhao, Dongbin
    Ma, Zhe
    Sun, Kewu
    Ding, Jishiyu
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (04) : 2093 - 2104