Learning to Model Opponent Learning (Student Abstract)

被引:0
|
作者
Davies, Ian [1 ]
Tian, Zheng [1 ]
Wang, Jun [1 ]
机构
[1] UCL, Gower St, London WC1E 6BT, England
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multi-Agent Reinforcement Learning (MARL) considers settings in which a set of coexisting agents interact with one another and their environment. The adaptation and learning of other agents induces non-stationarity in the environment dynamics. This poses a great challenge for value function-based algorithms whose convergence usually relies on the assumption of a stationary environment. Policy search algorithms also struggle in multi-agent settings as the partial observability resulting from an opponent's actions not being known introduces high variance to policy training. Modelling an agent's opponent(s) is often pursued as a means of resolving the issues arising from the coexistence of learning opponents. An opponent model provides an agent with some ability to reason about other agents to aid its own decision making. Most prior works learn an opponent model by assuming the opponent is employing a stationary policy or switching between a set of stationary policies. Such an approach can reduce the variance of training signals for policy search algorithms. However, in the multi-agent setting, agents have an incentive to continually adapt and learn. This means that the assumptions concerning opponent stationarity are unrealistic. In this work, we develop a novel approach to modelling an opponent's learning dynamics which we term Learning to Model Opponent Learning (LeMOL). We show our structured opponent model is more accurate and stable than naive behaviour cloning baselines. We further show that opponent modelling can improve the performance of algorithmic agents in multi-agent settings.
引用
收藏
页码:13771 / 13772
页数:2
相关论文
共 50 条
  • [1] Reinforcement Learning Explainability via Model Transforms (Student Abstract)
    Finkelstein, Mira
    Liu, Lucy
    Kolumbus, Yoav
    Parkes, David C.
    Rosenshein, Jeffrey S.
    Keren, Sarah
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 12943 - 12944
  • [2] Student Research Abstract: Smart Learning Environments for social learning
    Pesare, Enrica
    30TH ANNUAL ACM SYMPOSIUM ON APPLIED COMPUTING, VOLS I AND II, 2015, : 273 - 274
  • [3] Knowledge Transfer via Compact Model in Federated Learning (Student Abstract)
    Pei, Jiaming
    Li, Wei
    Wang, Lukun
    THIRTY-EIGTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 21, 2024, : 23621 - 23622
  • [4] Opponent Model Selection Using Deep Learning
    Chang, Hung-Jui
    Yueh, Cheng
    Fan, Gang-Yu
    Lin, Ting-Yu
    Hsu, Tsan-sheng
    ADVANCES IN COMPUTER GAMES, ACG 2021, 2022, 13262 : 176 - 186
  • [5] Learning to Evolve on Dynamic Graphs (Student Abstract)
    Xiang, Xintao
    Huang, Tiancheng
    Wang, Donglin
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 13091 - 13092
  • [6] Learning with Opponent-Learning Awareness
    Foerster, Jakob
    Chen, Richard Y.
    Al-Shedivat, Maruan
    Whiteson, Shimon
    Abbeel, Pieter
    Mordatch, Igor
    PROCEEDINGS OF THE 17TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS (AAMAS' 18), 2018, : 122 - 130
  • [7] LGML: Logic Guided Machine Learning (Student Abstract)
    Scott, Joseph
    Panju, Maysum
    Ganesh, Vijay
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 13909 - 13910
  • [8] Explainable Earnings Call Representation Learning (Student Abstract)
    Huang, Yanlong
    Lei, Yue
    Tai, Wenxin
    Cheng, Zhangtao
    Zhong, Ting
    Zhang, Kunpeng
    THIRTY-EIGTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 21, 2024, : 23518 - 23520
  • [9] Simple Orthogonal Graph Representation Learning (Student Abstract)
    Cui, Taoyong
    Dong, Yuhan
    THIRTY-EIGTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 21, 2024, : 23462 - 23464
  • [10] ERLP: Ensembles of Reinforcement Learning Policies (Student Abstract)
    Saphal, Rohan
    Ravindran, Balaraman
    Mudigere, Dheevatsa
    Avancha, Sasikanth
    Kaul, Bharat
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 13905 - 13906