Q-learning-based algorithms for dynamic transmission control in IoT equipment

被引:3
|
作者
Malekijou, Hanieh [1 ]
Hakami, Vesal [1 ]
Javan, Nastooh Taheri [2 ]
Malekijoo, Amirhossein [3 ]
机构
[1] Iran Univ Sci & Technol, Sch Comp Engn, Tehran, Iran
[2] Imam Khomeini Int Univ, Comp Engn Dept, Qazvin, Iran
[3] Semnan Univ, Dept Elect & Comp Engn, Semnan, Iran
来源
JOURNAL OF SUPERCOMPUTING | 2023年 / 79卷 / 01期
关键词
Delay; Energy harvesting; Jitter; Transmission control; Markov decision process; Reinforcement learning; POWER ALLOCATION; ENERGY; COMPRESSION; COMMUNICATION; POLICY;
D O I
10.1007/s11227-022-04643-9
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
We investigate an energy-harvesting IoT device transmitting (delay/jitter)-sensitive data over a wireless fading channel. The sensory module on the device injects captured event packets into its transmission buffer and relies on the random supply of the energy harvested from the environment to transmit them. Given the limited harvested energy, our goal is to compute optimal transmission control policies that decide on how many packets of data should be transmitted from the buffer's head-of-line at each discrete timeslot such that a long-run criterion involving the average delay/jitter is either minimized or never exceeds a pre-specified threshold. We realistically assume that no advance knowledge is available regarding the random processes underlying the variations in the channel, captured events, or harvested energy dynamics. Instead, we utilize a suite of Q-learning-based techniques (from the reinforcement learning theory) to optimize the transmission policy in a model-free fashion. In particular, we come up with three Q-learning algorithms: a constrained Markov decision process (CMDP)-based algorithm for optimizing energy consumption under a delay constraint, an MDP-based algorithm for minimizing the average delay under the limitations imposed by the energy harvesting process, and finally, a variance-penalized MDP-based algorithm to minimize a linearly combined cost function consisting of both delay and delay variation. Extensive numerical results are presented for performance evaluation.
引用
收藏
页码:75 / 108
页数:34
相关论文
共 50 条
  • [31] Deploying SDN Control in Internet of UAVs: Q-Learning-Based Edge Scheduling
    Zhang, Chaofeng
    Dong, Mianxiong
    Ota, Kaoru
    IEEE TRANSACTIONS ON NETWORK AND SERVICE MANAGEMENT, 2021, 18 (01): : 526 - 537
  • [32] A deep q-learning-based optimization of the inventory control in a linear process chain
    Dittrich, M. -A.
    Fohlmeister, S.
    PRODUCTION ENGINEERING-RESEARCH AND DEVELOPMENT, 2021, 15 (01): : 35 - 43
  • [33] Q-Learning-Based Supervisory Control Adaptability Investigation for Hybrid Electric Vehicles
    Xu, Bin
    Tang, Xiaolin
    Hu, Xiaosong
    Lin, Xianke
    Li, Huayi
    Rathod, Dhruvang
    Wang, Zhe
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2022, 23 (07) : 6797 - 6806
  • [34] An Analysis of Double Q-Learning-Based Energy Management Strategies for TEG-Powered IoT Devices
    Prauzek, Michal
    Konecny, Jaromir
    Paterova, Tereza
    IEEE INTERNET OF THINGS JOURNAL, 2023, 10 (21) : 18919 - 18929
  • [35] A Comparative Study of Reinforcement Learning Algorithms for Distribution Network Reconfiguration With Deep Q-Learning-Based Action Sampling
    Gholizadeh, Nastaran
    Kazemi, Nazli
    Musilek, Petr
    IEEE ACCESS, 2023, 11 : 13714 - 13723
  • [36] Q-learning-based unmanned aerial vehicle path planning with dynamic obstacle avoidance
    Sonny, Amala
    Yeduri, Sreenivasa Reddy
    Cenkeramaddi, Linga Reddy
    APPLIED SOFT COMPUTING, 2023, 147
  • [37] A Q-Learning-Based Fault-Tolerance Data Routing Scheme for IoT-Enabled WSNs
    Rawal, Anand Prakash
    Chanak, Prasenjit
    IEEE INTERNET OF THINGS JOURNAL, 2024, 11 (14): : 25283 - 25293
  • [38] Deep Q-Learning-Based Dynamic Network Slicing and Task Offloading in Edge Network
    Chiang, Yao
    Hsu, Chih-Ho
    Chen, Guan-Hao
    Wei, Hung-Yu
    IEEE TRANSACTIONS ON NETWORK AND SERVICE MANAGEMENT, 2023, 20 (01): : 369 - 384
  • [39] Q-Learning-based Adaptive Power Management for IoT System-an-Chips with Embedded Power States
    Debizet, Yvan
    Lallement, Guenole
    Abouzeid, Fady
    Roche, Philippe
    Autran, Jean-Luc
    2018 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), 2018,
  • [40] A Q-Learning-Based Routing Approach for Energy Efficient Information Transmission in Wireless Sensor Network
    Su, Xing
    Ren, Yiting
    Cai, Zhi
    Liang, Yi
    Guo, Limin
    IEEE TRANSACTIONS ON NETWORK AND SERVICE MANAGEMENT, 2023, 20 (02): : 1949 - 1961