Adaptive Learning: A New Decentralized Reinforcement Learning Approach for Cooperative Multiagent Systems

被引:13
|
作者
Li, Meng-Lin [1 ]
Chen, Shaofei [1 ]
Chen, Jing [1 ]
机构
[1] Natl Univ Def Technol, Coll Intelligence Sci & Technol, Changsha 410073, Peoples R China
来源
IEEE ACCESS | 2020年 / 8卷
基金
中国国家自然科学基金;
关键词
Learning (artificial intelligence); Training; Multi-agent systems; Heuristic algorithms; Roads; Urban areas; Games; Reinforcement learning; multiagent system; intelligent control;
D O I
10.1109/ACCESS.2020.2997899
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Multiagent systems (MASs) have received extensive attention in a variety of domains, such as robotics and distributed control. This paper focuses on how independent learners (ILs, structures used in decentralized reinforcement learning) decide on their individual behaviors to achieve coherent joint behavior. To date, Reinforcement learning(RL) approaches for ILs have not guaranteed convergence to the optimal joint policy in scenarios in which communication is difficult. Especially in a decentralized algorithm, the proportion of credit for a single agent's action in a multiagent system is not distinguished, which can lead to miscoordination of joint actions. Therefore, it is highly significant to study the mechanisms of coordination between agents in MASs. Most previous coordination mechanisms have been carried out by modeling the communication mechanism and other agent policies. These methods are applicable only to a particular system, so such algorithms do not offer generalizability, especially when there are dozens or more agents. Therefore, this paper mainly focuses on the MAS contains more than a dozen agents. By combining the method of parallel computation, the experimental environment is closer to the application scene. By studying the paradigm of centralized training and decentralized execution(CTDE), a multi-agent reinforcement learning algorithm for implicit coordination based on TD error is proposed. The new algorithm can dynamically adjust the learning rate by deeply analyzing the dissonance problem in the matrix game and combining it with a multiagent environment. By adjusting the dynamic learning rate between agents, coordination of the agents' strategies can be achieved. Experimental results show that the proposed algorithm can effectively improve the coordination ability of a MAS. Moreover, the variance of the training results is more stable than that of the hysteretic Q learning(HQL) algorithm. Hence, the problem of miscoordination in a MAS can be avoided to some extent without additional communication. Our work provides a new way to solve the miscoordination problem for reinforcement learning algorithms in the scale of dozens or more number of agents. As a new IL structure algorithm, our results should be extended and further studied.
引用
收藏
页码:99404 / 99421
页数:18
相关论文
共 50 条
  • [1] Inverse Reinforcement Learning for Decentralized Non-Cooperative Multiagent Systems
    Reddy, Tummalapalli Sudhamsh
    Gopikrishna, Vamsikrishna
    Zaruba, Gergely
    Huber, Manfred
    PROCEEDINGS 2012 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2012, : 1930 - 1935
  • [2] Decentralized Reinforcement Learning Inspired by Multiagent Systems
    Adjodah, Dhaval
    PROCEEDINGS OF THE 17TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS (AAMAS' 18), 2018, : 1729 - 1730
  • [3] Formation Tracking of Spatiotemporal Multiagent Systems: A Decentralized Reinforcement Learning Approach
    Liu, Tianrun
    Chen, Yang-Yang
    IEEE SYSTEMS MAN AND CYBERNETICS MAGAZINE, 2024, 10 (04): : 52 - 60
  • [4] The dynamics of reinforcement learning in cooperative multiagent systems
    Claus, C
    Boutilier, C
    FIFTEENTH NATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE (AAAI-98) AND TENTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICAL INTELLIGENCE (IAAI-98) - PROCEEDINGS, 1998, : 746 - 752
  • [5] Reinforcement Learning With Task Decomposition for Cooperative Multiagent Systems
    Sun, Changyin
    Liu, Wenzhang
    Dong, Lu
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2021, 32 (05) : 2054 - 2065
  • [6] Learning to Teach in Cooperative Multiagent Reinforcement Learning
    Omidshafiei, Shayegan
    Kim, Dong-Ki
    Liu, Miao
    Tesauro, Gerald
    Riemer, Matthew
    Amato, Christopher
    Campbell, Murray
    How, Jonathan P.
    THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 6128 - 6136
  • [7] Learning Cooperative Behaviours in Multiagent Reinforcement Learning
    Phon-Amnuaisuk, Somnuk
    NEURAL INFORMATION PROCESSING, PT 1, PROCEEDINGS, 2009, 5863 : 570 - 579
  • [8] Multiagent Reinforcement Social Learning toward Coordination in Cooperative Multiagent Systems
    Hao, Jianye
    Leung, Ho-Fung
    Ming, Zhong
    ACM TRANSACTIONS ON AUTONOMOUS AND ADAPTIVE SYSTEMS, 2015, 9 (04)
  • [9] A Study on Cooperative Action Selection Considering Unfairness in Decentralized Multiagent Reinforcement Learning
    Matsui, Toshihiro
    Matsuo, Hiroshi
    ICAART: PROCEEDINGS OF THE 9TH INTERNATIONAL CONFERENCE ON AGENTS AND ARTIFICIAL INTELLIGENCE, VOL 1, 2017, : 88 - 95
  • [10] The dynamics of reinforcement social learning in networked cooperative multiagent systems
    Hao, Jianye
    Huang, Dongping
    Cai, Yi
    Leung, Ho-fung
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2017, 58 : 111 - 122