Differential Advising in Multiagent Reinforcement Learning

被引:18
|
作者
Ye, Dayong [1 ,2 ]
Zhu, Tianqing [4 ]
Cheng, Zishuo [1 ,2 ]
Zhou, Wanlei [3 ]
Yu, Philip S. [5 ]
机构
[1] Univ Technol Sydney, Ctr Cyber Secur & Privacy, Ultimo, NSW 2007, Australia
[2] Univ Technol Sydney, Sch Comp Sci, Ultimo, NSW 2007, Australia
[3] City Univ Macau, Macau, Peoples R China
[4] China Univ Geosci, Sch Comp Sci, Wuhan 430000, Peoples R China
[5] Univ Illinois, Dept Comp Sci, Chicago, IL 60607 USA
基金
中国国家自然科学基金; 澳大利亚研究理事会;
关键词
Robots; Differential privacy; Privacy; Reinforcement learning; Task analysis; Sensitivity; Computer science; Agent advising; differential privacy; multiagent reinforcement learning (MARL);
D O I
10.1109/TCYB.2020.3034424
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Agent advising is one of the main approaches to improve agent learning performance by enabling agents to share advice. Existing advising methods have a common limitation that an adviser agent can offer advice to an advisee agent only if the advice is created in the same state as the advisee's state. However, in complex environments, it is a very strong requirement that two states are the same, because a state may consist of multiple dimensions and two states being the same means that all these dimensions in the two states are correspondingly identical. Therefore, this requirement may limit the applicability of existing advising methods to complex environments. In this article, inspired by the differential privacy scheme, we propose a differential advising method that relaxes this requirement by enabling agents to use advice in a state even if the advice is created in a slightly different state. Compared with the existing methods, agents using the proposed method have more opportunity to take advice from others. This article is the first to adopt the concept of differential privacy on advising to improve agent learning performance instead of addressing security issues. The experimental results demonstrate that the proposed method is more efficient in complex environments than the existing methods.
引用
收藏
页码:5508 / 5521
页数:14
相关论文
共 50 条
  • [41] PyTAG: Tabletop Games for Multiagent Reinforcement Learning
    Balla, Martin
    Long, George E. M.
    Goodman, James
    Gaina, Raluca D.
    Perez-Liebana, Diego
    IEEE TRANSACTIONS ON GAMES, 2024, 16 (04) : 993 - 1002
  • [42] Heuristic Selection of Actions in Multiagent Reinforcement Learning
    Bianchi, Reinaldo A. C.
    Ribeiro, Carlos H. C.
    Costa, Anna H. R.
    20TH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2007, : 690 - 695
  • [43] A survey and critique of multiagent deep reinforcement learning
    Pablo Hernandez-Leal
    Bilal Kartal
    Matthew E. Taylor
    Autonomous Agents and Multi-Agent Systems, 2019, 33 : 750 - 797
  • [44] Reinforcement learning for encouraging cooperation in a multiagent system
    Jiang, Wei-Cheng
    Huang, Hong-Hao
    Wang, Yu-Teng
    INFORMATION SCIENCES, 2024, 680
  • [45] Deep multiagent reinforcement learning: challenges and directions
    Annie Wong
    Thomas Bäck
    Anna V. Kononova
    Aske Plaat
    Artificial Intelligence Review, 2023, 56 : 5023 - 5056
  • [46] Resilient Multiagent Reinforcement Learning With Function Approximation
    Ye, Lintao
    Figura, Martin
    Lin, Yixuan
    Pal, Mainak
    Das, Pranoy
    Liu, Ji
    Gupta, Vijay
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2024, 69 (12) : 8497 - 8512
  • [47] Employing OLAP mining for multiagent reinforcement learning
    Alhajj, R
    Kaya, M
    DESIGN AND APPLICATION OF HYBRID INTELLIGENT SYSTEMS, 2003, 104 : 759 - 768
  • [48] Scheduling in Multiagent Systems Using Reinforcement Learning
    Minashina, I. K.
    Gorbachev, R. A.
    Zakharova, E. M.
    DOKLADY MATHEMATICS, 2022, 106 (SUPPL 1) : S70 - S78
  • [49] Multiagent Reinforcement Learning With Unshared Value Functions
    Hu, Yujing
    Gao, Yang
    An, Bo
    IEEE TRANSACTIONS ON CYBERNETICS, 2015, 45 (04) : 647 - 662
  • [50] Decentralized Reinforcement Learning Inspired by Multiagent Systems
    Adjodah, Dhaval
    PROCEEDINGS OF THE 17TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS (AAMAS' 18), 2018, : 1729 - 1730