Eavesdropping Game Based on Multi-Agent Deep Reinforcement Learning

被引:0
|
作者
Guo, Delin [1 ]
Tang, Lan [1 ]
Yang, Lvxi [2 ]
Liang, Ying-Chang [2 ]
机构
[1] Nanjing Univ, Nanjing, Peoples R China
[2] Southeast Univ, Nanjing, Peoples R China
基金
国家重点研发计划; 中国国家自然科学基金;
关键词
Physical layer security; proactive eavesdropping; stochastic game; multi-agent reinforcement learning; WIRETAP CHANNEL;
D O I
10.1109/SPAWC51304.2022.9833927
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
This paper considers an adversarial scenario between a legitimate eavesdropper and a suspicious communication pair. All three nodes are equipped with multiple antennas. The eavesdropper, which operates in a full-duplex model, aims to wiretap the dubious communication pair via proactive jamming. On the other hand, the suspicious transmitter, which can send artificial noise (AN) to disturb the wiretap channel, aims to guarantee secrecy. More specifically, the eavesdropper adjusts jamming power to enhance the wiretap rate, while the suspicious transmitter jointly adapts the transmit power and noise power against the eavesdropping. Considering the partial observation and complicated interactions between the eavesdropper and the suspicious pair in unknown system dynamics, we model the problem as an imperfect-information stochastic game. To approach the Nash equilibrium solution of the eavesdropping game, we develop a multi-agent reinforcement learning (MARL) algorithm, termed neural fictitious self-play with soft actor-critic (NFSP-SAC), by combining the fictitious self-play (FSP) with a deep reinforcement learning algorithm, SAC. The introduction of SAC enables FSP to handle the problems with continuous and high dimension observation and action space. The simulation results demonstrate that the power allocation policies learned by our method empirically converge to a Nash equilibrium, while the compared reinforcement learning algorithms suffer from severe fluctuations during the learning process.
引用
收藏
页数:5
相关论文
共 50 条
  • [11] Lenient Multi-Agent Deep Reinforcement Learning
    Palmer, Gregory
    Tuyls, Karl
    Bloembergen, Daan
    Savani, Rahul
    PROCEEDINGS OF THE 17TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS (AAMAS' 18), 2018, : 443 - 451
  • [12] Multi-agent deep reinforcement learning: a survey
    Gronauer, Sven
    Diepold, Klaus
    ARTIFICIAL INTELLIGENCE REVIEW, 2022, 55 (02) : 895 - 943
  • [13] Cloud game computing offload based on Multi-Agent Reinforcement Learning
    Tian, Kaicong
    Yang, Hongwen
    Liu, Yitong
    Zheng, Qingbi
    2022 IEEE 96TH VEHICULAR TECHNOLOGY CONFERENCE (VTC2022-FALL), 2022,
  • [14] Distributed Signal Control of Multi-agent Reinforcement Learning Based on Game
    Qu Z.-W.
    Pan Z.-T.
    Chen Y.-H.
    Li H.-T.
    Wang X.
    Chen, Yong-Heng (cyh@jlu.edu.cn), 1600, Science Press (20): : 76 - 82and100
  • [15] Learning to Communicate with Deep Multi-Agent Reinforcement Learning
    Foerster, Jakob N.
    Assael, Yannis M.
    de Freitas, Nando
    Whiteson, Shimon
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 29 (NIPS 2016), 2016, 29
  • [16] Hierarchical reinforcement learning based on multi-agent cooperation game theory
    Tang H.
    Dong C.
    International Journal of Wireless and Mobile Computing, 2019, 16 (04): : 369 - 376
  • [17] A reinforcement learning scheme for a multi-agent card game
    Fujita, H
    Matsuno, Y
    Ishii, S
    2003 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN AND CYBERNETICS, VOLS 1-5, CONFERENCE PROCEEDINGS, 2003, : 4071 - 4078
  • [18] Evolutionary game theory and multi-agent reinforcement learning
    Tuyls, K
    Nowé, A
    KNOWLEDGE ENGINEERING REVIEW, 2005, 20 (01): : 63 - 90
  • [19] Multi-Agent Reinforcement Learning for a Random Access Game
    Lee, Dongwoo
    Zhao, Yu
    Seo, Jun-Bae
    Lee, Joohyun
    IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2022, 71 (08) : 9119 - 9124
  • [20] Human Strategy Learning-Based Multi-Agent Deep Reinforcement Learning for Online Team Sports Game
    Lee, Seongbeen
    Lee, Gyuhyuk
    Kim, Wongyeom
    Kim, Junoh
    Park, Jisun
    Cho, Kyungeun
    IEEE ACCESS, 2025, 13 : 15437 - 15452