Formal contracts mitigate social dilemmas in multi-agent reinforcement learning

被引:0
|
作者
Haupt, Andreas [1 ]
Christoffersen, Phillip [1 ]
Damani, Mehul [1 ]
Hadfield-Menell, Dylan [1 ]
机构
[1] MIT, Comp Sci & Artificial Intelligence Lab, 32 Vassar St, Cambridge, MA 02139 USA
关键词
Social dilemma; Decentralized training; Formal contracts; LEVEL;
D O I
10.1007/s10458-024-09682-5
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Multi-agent Reinforcement Learning (MARL) is a powerful tool for training autonomous agents acting independently in a common environment. However, it can lead to sub-optimal behavior when individual incentives and group incentives diverge. Humans are remarkably capable at solving these social dilemmas. It is an open problem in MARL to replicate such cooperative behaviors in selfish agents. In this work, we draw upon the idea of formal contracting from economics to overcome diverging incentives between agents in MARL. We propose an augmentation to a Markov game where agents voluntarily agree to binding transfers of reward, under pre-specified conditions. Our contributions are theoretical and empirical. First, we show that this augmentation makes all subgame-perfect equilibria of all Fully Observable Markov Games exhibit socially optimal behavior, given a sufficiently rich space of contracts. Next, we show that for general contract spaces, and even under partial observability, richer contract spaces lead to higher welfare. Hence, contract space design solves an exploration-exploitation tradeoff, sidestepping incentive issues. We complement our theoretical analysis with experiments. Issues of exploration in the contracting augmentation are mitigated using a training methodology inspired by multi-objective reinforcement learning: Multi-Objective Contract Augmentation Learning. We test our methodology in static, single-move games, as well as dynamic domains that simulate traffic, pollution management, and common pool resource management.
引用
收藏
页数:38
相关论文
共 50 条
  • [1] Multi-agent Reinforcement Learning in Sequential Social Dilemmas
    Leibo, Joel Z.
    Zambaldi, Vinicius
    Lanctot, Marc
    Marecki, Janusz
    Graepel, Thore
    AAMAS'17: PROCEEDINGS OF THE 16TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS, 2017, : 464 - 473
  • [2] Evolutionary multi-agent reinforcement learning in group social dilemmas
    Mintz, B.
    Fu, F.
    CHAOS, 2025, 35 (02)
  • [3] Modeling Moral Choices in Social Dilemmas with Multi-Agent Reinforcement Learning
    Tennant, Elizaveta
    Hailes, Stephen
    Musolesi, Mirco
    PROCEEDINGS OF THE THIRTY-SECOND INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2023, 2023, : 317 - 325
  • [4] Formal Reachability Analysis for Multi-Agent Reinforcement Learning Systems
    Wang, Xiaoyan
    Peng, Jun
    Li, Shuqiu
    Li, Bing
    IEEE ACCESS, 2021, 9 : 45812 - 45821
  • [5] The evolution of cooperation in continuous dilemmas via multi-agent reinforcement learning
    Zhu, Congcong
    Ye, Dayong
    Zhu, Tianqing
    Zhou, Wanlei
    KNOWLEDGE-BASED SYSTEMS, 2025, 315
  • [6] Emergent Social Learning via Multi-agent Reinforcement Learning
    Ndousse, Kamal
    Eck, Douglas
    Levine, Sergey
    Jaques, Natasha
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [7] Multi-Agent Reinforcement Learning
    Stankovic, Milos
    2016 13TH SYMPOSIUM ON NEURAL NETWORKS AND APPLICATIONS (NEUREL), 2016, : 43 - 43
  • [8] Social satisficing: Multi-agent reinforcement learning with satisficing agents
    Uragami, Daisuke
    Sonota, Noriaki
    Takahashi, Tatsuji
    BIOSYSTEMS, 2024, 243
  • [9] Multi-Agent Cognition Difference Reinforcement Learning for Multi-Agent Cooperation
    Wang, Huimu
    Qiu, Tenghai
    Liu, Zhen
    Pu, Zhiqiang
    Yi, Jianqiang
    Yuan, Wanmai
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [10] Multi-Agent Reinforcement Learning With Distributed Targeted Multi-Agent Communication
    Xu, Chi
    Zhang, Hui
    Zhang, Ya
    2023 35TH CHINESE CONTROL AND DECISION CONFERENCE, CCDC, 2023, : 2915 - 2920