Deep Q-network Based Reinforcement Learning for Distributed Dynamic Spectrum Access

被引:1
|
作者
Yadav, Manish Anand [1 ]
Li, Yuhui [1 ]
Fang, Guangjin [1 ]
Shen, Bin [1 ]
机构
[1] Chongqing Univ Posts & Telecommun CQUPT, Sch Commun & Informat Engn SCIE, Chongqing 400065, Peoples R China
关键词
dynamic spectrum access; Q-learning; deep reinforcement learning; double deep Q-network;
D O I
10.1109/CCAI55564.2022.9807797
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
To solve the problem of spectrum scarcity and spectrum under-utilization in wireless networks, we propose a double deep Q-network based reinforcement learning algorithm for distributed dynamic spectrum access. Channels in the network are either busy or idle based on the two-state Markov chain. At the start of each time slot, every secondary user (SU) performs spectrum sensing on each channel and accesses one based on the sensing result as well as the output of the Q-network of our algorithm. Over time, the Deep Reinforcement Learning (DRL) algorithm learns the spectrum environment and becomes good at modeling the behavior pattern of the primary users (PUs). Through simulation, we show that our proposed algorithm is simple to train, yet effective in reducing interference to primary as well as secondary users and achieving higher successful transmission.
引用
收藏
页码:227 / 232
页数:6
相关论文
共 50 条
  • [41] Deep Q-Network Based Dynamic Movement Strategy in a UAV-Assisted Network
    Zhong, Xukai
    Huo, Yiming
    Dong, Xiaodai
    Liang, Zhonghua
    2020 IEEE 92ND VEHICULAR TECHNOLOGY CONFERENCE (VTC2020-FALL), 2020,
  • [42] Influence on Learning of Various Conditions in Deep Q-Network
    Niitsuma, Jun
    Osana, Yuko
    2017 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2017, : 1932 - 1935
  • [43] Intelligent Dynamic Spectrum Access Using Deep Reinforcement Learning for VANETs
    Wang, Yonghua
    Li, Xueyang
    Wan, Pin
    Shao, Ruiyu
    IEEE SENSORS JOURNAL, 2021, 21 (14) : 15554 - 15563
  • [44] Listen-After-Collision Mechanism for Dynamic Spectrum Access Using Deep Q-Network With an Improved Thompson Sampling Algorithm
    He, Ming
    Jin, Ming
    Guo, Qinghua
    Xu, Weiqiang
    IEEE INTERNET OF THINGS JOURNAL, 2024, 11 (04): : 6596 - 6606
  • [45] Deep-Reinforcement-Learning-Based Distributed Dynamic Spectrum Access in Multiuser Multichannel Cognitive Radio Internet of Things Networks
    Zhang, Xiaohui
    Chen, Ze
    Zhang, Yinghui
    Liu, Yang
    Jin, Minglu
    Qiu, Tianshuang
    IEEE INTERNET OF THINGS JOURNAL, 2024, 11 (10): : 17495 - 17509
  • [46] Multiagent Learning and Coordination with Clustered Deep Q-Network
    Pageaud, Simon
    Deslandres, Veronique
    Lehoux, Vassilissa
    Hassas, Salima
    AAMAS '19: PROCEEDINGS OF THE 18TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS, 2019, : 2156 - 2158
  • [47] A Dueling Deep Recurrent Q-Network Framework for Dynamic Multichannel Access in Heterogeneous Wireless Networks
    Chen, Haitao
    Zhao, Haitao
    Zhou, Li
    Zhang, Jiao
    Liu, Yan
    Pan, Xiaoqian
    Liu, Xingguang
    Wei, Jibo
    WIRELESS COMMUNICATIONS & MOBILE COMPUTING, 2022, 2022
  • [48] A Study on Vision-based Mobile Robot Learning by Deep Q-network
    Sasaki, Hikaru
    Horiuchi, Tadashi
    Kato, Satoru
    2017 56TH ANNUAL CONFERENCE OF THE SOCIETY OF INSTRUMENT AND CONTROL ENGINEERS OF JAPAN (SICE), 2017, : 799 - 804
  • [49] Intelligent fault diagnosis for rotating machinery using deep Q-network based health state classification: A deep reinforcement learning approach
    Ding, Yu
    Ma, Liang
    Ma, Jian
    Suo, Mingliang
    Tao, Laifa
    Cheng, Yujie
    Lu, Chen
    ADVANCED ENGINEERING INFORMATICS, 2019, 42
  • [50] Deep-reinforcement learning for fair distributed dynamic spectrum access in priority buffered heterogeneous wireless networks
    Janiar, Siavash Barqi
    Pourahmadi, Vahid
    IET COMMUNICATIONS, 2021, 15 (05) : 674 - 682