Next-gen resource optimization in NB-IoT networks: Harnessing soft actor-critic reinforcement learning

被引:1
|
作者
Anbazhagan, S. [1 ]
Mugelan, R. K. [1 ]
机构
[1] Vellore Inst Technol, Sch Elect Engn, Dept Commun Engn, Vellore 632014, Tamil Nadu, India
关键词
Narrowband Internet of Things (NB-ioT); Resource allocation; Reinforcement learning; Soft actor-critic (SAC); D2D COMMUNICATION; ALLOCATION; UPLINK; PERFORMANCE; ADAPTATION; DOWNLINK;
D O I
10.1016/j.comnet.2024.110670
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Resource allocation in Narrowband Internet of Things (NB-IoT) networks is a complex challenge due to dynamic user demands, variable channel conditions, and distance considerations. Traditional approaches often struggle to adapt to the dynamic nature of these environments. In this study, we leverage reinforcement learning (RL) to address the intricate nature of NB-IoT resource allocation. Specifically, we employ the Soft Actor-Critic (SAC) algorithm, comparing its performance against conventional RL algorithms such as Deep Q-Network (DQN) and Proximal Policy Optimization (PPO). The Soft Actor-Critic (SAC) algorithm is employed to train an agent for adaptive resource allocation, considering energy efficiency, throughput, latency, fairness, and interference constraints. The agent adeptly balances these objectives through an intricate reward structure and penalty mechanisms. Through comprehensive analysis, we present performance metrics, including total reward, energy efficiency, throughput, fairness, and latency, showcasing the efficacy of SAC when compared to DQN and PPO. Our findings underscore the efficiency of SAC in optimizing resource allocation in NBIoT networks, offering a promising solution to the complexities inherent in such dynamic environments. Resource allocation in Narrowband Internet of Things (NB-IoT) networks presents a complex challenge due to dynamic user demands, variable channel conditions, and distance considerations. Traditional approaches often struggle to adapt to these dynamic environments. This study leverages reinforcement learning (RL), specifically the Soft Actor-Critic (SAC) algorithm, to address the intricacies of NB-IoT resource allocation. We compare SAC's performance against conventional RL algorithms, including Deep Q-Network (DQN) and Proximal Policy Optimization (PPO). The SAC algorithm is utilized to train an agent for adaptive resource allocation, focusing on energy efficiency, throughput, latency, fairness, interference constraints, recovery time, and long-term performance stability. To demonstrate the scalability and effectiveness of SAC, we conducted experiments on NB-IoT networks with varying deployment types and configurations, including standard urban and suburban, high-density urban, industrial IoT, rural and low-density, and IoT service providers. To assess generalization capability, we tested SAC across applications like smart metering, smart cities, smart agriculture, and asset tracking & management. Our comprehensive analysis demonstrates that SAC significantly outperforms DQN and PPO across multiple performance metrics. Specifically, SAC improves energy efficiency by 5.60% over PPO and 10.25% over DQN. In terms of latency, SAC achieves a marginal reduction of approximately 0.0124% compared to PPO and 0.0126% compared to DQN. SAC enhances throughput by 214.98% over PPO and 15.72% over DQN. Additionally, SAC shows a substantial increase in fairness (Jain's index), improving by 358.31% over PPO and 614.46% over DQN. SAC also demonstrates superior recovery time, improving by 18.99% over PPO and 25.07% over DQN. In both deployment scenarios and diverse IoT applications, SAC consistently achieves high total rewards, minimal fluctuations, and stable performance. Energy efficiency remains constant at 7.2 bits per Joule, and latency is approximately 0.080 s. Throughput is robust across different deployments, while fairness remains high, ensuring equitable resource allocation. Recovery times are stable, enhancing operational reliability. These results underscore SAC's efficiency and robustness in optimizing resource allocation in NB-IoT networks, presenting a promising solution to the complexities of dynamic environments.
引用
收藏
页数:42
相关论文
共 50 条
  • [21] COOPERATIVE DEEP REINFORCEMENT LEARNING FOR MULTIPLE-GROUP NB-IOT NETWORKS OPTIMIZATION
    Jiang, Nan
    Deng, Yansha
    Simeone, Osvaldo
    Nallanathan, Arumugam
    2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2019, : 8424 - 8428
  • [22] Maximizing Information Usefulness in Vehicular CP Networks Using Actor-Critic Reinforcement Learning
    Ghnaya, Imed
    Ahmed, Toufik
    Mosbah, Mohamed
    Aniss, Hasnaa
    2022 18TH INTERNATIONAL CONFERENCE ON NETWORK AND SERVICE MANAGEMENT (CNSM 2022): INTELLIGENT MANAGEMENT OF DISRUPTIVE NETWORK TECHNOLOGIES AND SERVICES, 2022, : 296 - 302
  • [23] Actor-Critic Deep Reinforcement Learning for Energy Minimization in UAV-Aided Networks
    Yuan, Yaxiong
    Lei, Lei
    Vu, Thang X.
    Chatzinotas, Symeon
    Ottersten, Bjorn
    2020 EUROPEAN CONFERENCE ON NETWORKS AND COMMUNICATIONS (EUCNC 2020), 2020, : 348 - 352
  • [24] An actor-critic reinforcement learning-based resource management in mobile edge computing systems
    Fu, Fang
    Zhang, Zhicai
    Yu, Fei Richard
    Yan, Qiao
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2020, 11 (08) : 1875 - 1889
  • [25] Proactive Content Caching Based on Actor-Critic Reinforcement Learning for Mobile Edge Networks
    Jiang, Wei
    Feng, Daquan
    Sun, Yao
    Feng, Gang
    Wang, Zhenzhong
    Xia, Xiang-Gen
    IEEE TRANSACTIONS ON COGNITIVE COMMUNICATIONS AND NETWORKING, 2022, 8 (02) : 1239 - 1252
  • [26] Power Allocation in Dual Connectivity Networks Based on Actor-Critic Deep Reinforcement Learning
    Moein, Elham
    Hasibi, Ramin
    Shokri, Matin
    Rasti, Mehdi
    17TH INTERNATIONAL SYMPOSIUM ON MODELING AND OPTIMIZATION IN MOBILE, AD HOC, AND WIRELESS NETWORKS (WIOPT 2019), 2019, : 170 - 177
  • [27] An actor-critic reinforcement learning-based resource management in mobile edge computing systems
    Fang Fu
    Zhicai Zhang
    Fei Richard Yu
    Qiao Yan
    International Journal of Machine Learning and Cybernetics, 2020, 11 : 1875 - 1889
  • [28] Graph Soft Actor-Critic Reinforcement Learning for Large-Scale Distributed Multirobot Coordination
    Hu, Yifan
    Fu, Junjie
    Wen, Guanghui
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, : 1 - 12
  • [29] Dynamic Actor-critic: Reinforcement Learning based Radio Resource Scheduling For LTE-Advanced
    Tathe, Pallavi K.
    Sharma, Manish
    2018 FOURTH INTERNATIONAL CONFERENCE ON COMPUTING COMMUNICATION CONTROL AND AUTOMATION (ICCUBEA), 2018,
  • [30] A Deep Reinforcement Learning Approach for Adaptive Traffic Routing in Next-gen Networks
    Abrol, Akshita
    Mohan, Purnima Murali
    Truong-Huu, Tram
    ICC 2024 - IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS, 2024, : 465 - 471