Deep Reinforcement Learning-Based Multichannel Access for Industrial Wireless Networks With Dynamic Multiuser Priority

被引:12
|
作者
Liu, Xiaoyu [1 ,2 ,3 ,4 ]
Xu, Chi [1 ,2 ,3 ]
Yu, Haibin [1 ,2 ,3 ]
Zeng, Peng [1 ,2 ,3 ]
机构
[1] Chinese Acad Sci, Shenyang Inst Automat, State Key Lab Robot, Shenyang 110016, Peoples R China
[2] Chinese Acad Sci, Key Lab Networked Control Syst, Shenyang 110016, Peoples R China
[3] Chinese Acad Sci, Inst Robot & Intelligent Mfg, Shenyang 110169, Peoples R China
[4] Univ Chinese Acad Sci, Beijing 100049, Peoples R China
基金
中国国家自然科学基金; 中国博士后科学基金;
关键词
Deep reinforcement learning; dynamic priority; industrial wireless networks (IWNs); multichannel access; quality of service; SPECTRUM ACCESS; COMMUNICATION; TECHNOLOGY; ALLOCATION;
D O I
10.1109/TII.2021.3139349
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In Industry 4.0, massive heterogeneous industrial devices generate a great deal of data with different quality of service requirements, and communicate via industrial wireless networks (IWNs). However, the limited time-frequency resources of IWNs cannot well support the high concurrent access of massive industrial devices with strict real-time and reliable communication requirements. To address this problem, a deep reinforcement learning-based dynamic priority multichannel access (DRL-DPMCA) algorithm is proposed in this article. Firstly, according to the time-sensitivity of industrial data, industrial devices are assigned with different priorities, based on which their channel access probabilities are dynamically adjusted. Then, the Markov decision process is utilized to model the dynamic priority multichannel access problem. To cope with the explosion of state space caused by the multichannel access of massive industrial devices with dynamic priorities, DRL is used to establish the mapping from states to actions. Next, the long-term cumulative reward is maximized to obtain an effective policy. Especially, with joint consideration of the access reward and priority reward, a compound reward for multichannel access and dynamic priority is designed. For breaking the time correlation of training data while accelerating the convergence of DRL-DPMCA, an experience replay with experience-weight is proposed to store and sample experiences categorically. Besides, the gated recurrent unit, dueling architecture and step-by-step epsilon-greedy method are employed to make states more comprehensive and reduce model oscillation. Extensive experiments show that, compared with slotted-Aloha and deep Q network algorithms, DRL-DPMCA converges quickly, and guarantees the highest channel access probability and the minimum queuing delay for high-priority industrial devices in the context of minimum access conflict and nearly 100% channel utilization.
引用
收藏
页码:7048 / 7058
页数:11
相关论文
共 50 条
  • [31] Deep Learning-Based Dynamic Bandwidth Allocation for Future Optical Access Networks
    Hatem, John Abied
    Dhaini, Ahmad R.
    Elbassuoni, Shady
    IEEE ACCESS, 2019, 7 : 97307 - 97318
  • [32] Deep Reinforcement Learning-based Adaptive Wireless Resource Allocation Algorithm for Heterogeneous Cloud Wireless Access Network
    Chen Qianbin
    Guang Lingjin
    Li Ziyu
    Wang Zhaokun
    Yang Heng
    Tang Lun
    JOURNAL OF ELECTRONICS & INFORMATION TECHNOLOGY, 2020, 42 (06) : 1468 - 1477
  • [33] Deep Reinforcement Learning-Based Dynamic Spectrum Access for D2D Communication Underlay Cellular Networks
    Huang, Jingfei
    Yang, Yang
    He, Gang
    Xiao, Yang
    Liu, Jun
    IEEE COMMUNICATIONS LETTERS, 2021, 25 (08) : 2614 - 2618
  • [34] Multiple Access in Cell-Free Networks: Outage Performance, Dynamic Clustering, and Deep Reinforcement Learning-Based Design
    Al-Eryani, Yasser
    Akrout, Mohamed
    Hossain, Ekram
    IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, 2021, 39 (04) : 1028 - 1042
  • [35] Hybrid Deep Learning-Based Adaptive Multiple Access Schemes Underwater Wireless Networks
    Anitha, D.
    Karthika, R. A.
    INTELLIGENT AUTOMATION AND SOFT COMPUTING, 2023, 35 (02): : 2463 - 2477
  • [36] Deep-Reinforcement Learning Multiple Access for Heterogeneous Wireless Networks
    Yu, Yiding
    Wang, Taotao
    Liew, Soung Chang
    2018 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC), 2018,
  • [37] Deep-Reinforcement Learning Multiple Access for Heterogeneous Wireless Networks
    Yu, Yiding
    Wang, Taotao
    Liew, Soung Chang
    IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, 2019, 37 (06) : 1277 - 1290
  • [38] Modelling a Learning-Based Dynamic Tree Routing Model for Wireless Mesh Access Networks
    Krishnammal, N.
    Kalaiarasan, C.
    Bharathi, A.
    COMPUTER SYSTEMS SCIENCE AND ENGINEERING, 2023, 44 (02): : 1531 - 1549
  • [39] A Reinforcement Learning-based Multipath Scheduling for Heterogeneous Wireless Networks
    Nguyen, Thanh Trung
    Vu, Minh Hai
    Le Nguyen, Phi
    Do, Phan Thuan
    Nguyen, Kien
    2022 IEEE 8TH WORLD FORUM ON INTERNET OF THINGS, WF-IOT, 2022,
  • [40] Deep Reinforcement Learning for Dynamic Spectrum Access in the Multi-Channel Wireless Local Area Networks
    Bhandari, Sovit
    Ranjan, Navin
    Kim, Yeong-Chan
    Kim, Hoon
    2022 INTERNATIONAL CONFERENCE ON ELECTRONICS, INFORMATION, AND COMMUNICATION (ICEIC), 2022,