共 33 条
Intelligent Resource Allocation for Edge-Cloud Collaborative Networks: A Hybrid DDPG-D3QN Approach
被引:23
|作者:
Hu, Han
[1
,2
]
Wu, Dingguo
[1
,2
]
Zhou, Fuhui
[3
,4
]
Zhu, Xingwu
[1
,2
]
Hu, Rose Qingyang
[5
]
Zhu, Hongbo
[1
,2
]
机构:
[1] Nanjing Univ Posts & Telecommun, Jiangsu Key Lab Wireless Commun, Nanjing 210003, Peoples R China
[2] Nanjing Univ Posts & Telecommun, Minist Educ, Engn Res Ctr Hlth Serv Syst Based Ubiquitous Wire, Nanjing 210003, Peoples R China
[3] Nanjing Univ Aeronaut & Astronaut, Coll Elect & Informat Engn, Nanjing 210000, Peoples R China
[4] Xidian Univ, State Key Lab Integrated Serv Networks, Xian 710071, Peoples R China
[5] Utah State Univ, Dept Elect & Comp Engn, Logan, UT 84322 USA
基金:
中国国家自然科学基金;
关键词:
Servers;
Resource management;
Delays;
Collaboration;
Cloud computing;
Task analysis;
Optimization;
Cooperative offloading;
deep reinforcement learning;
dynamic offloading;
mobile edge computing (MEC);
ENERGY-EFFICIENT;
ENABLED IOT;
OPTIMIZATION;
D O I:
10.1109/TVT.2023.3253905
中图分类号:
TM [电工技术];
TN [电子技术、通信技术];
学科分类号:
0808 ;
0809 ;
摘要:
To handle the ever-increasing IoT devices with computation-intensive and delay-critical applications, it is imperative to leverage the collaborative potential of edge and cloud computing. In this paper, we investigate the dynamic offloading of packets with finite block length (FBL) in an edge-cloud collaboration system consisting of multi-mobile IoT devices (MIDs) with energy harvesting (EH), multi-edge servers, and one cloud server (CS) in a dynamic environment. The optimization problem is formulated to minimize the average long-term service cost defined as the weighted sum of MID energy consumption and service delay, with the constraints of the available resource, the energy causality, the allowable service delay, and the maximum decoding error probability. To address the problem involving both discrete and continuous variables, we propose a multi-device hybrid decision-based deep reinforcement learning (DRL) solution, named DDPG-D3QN algorithm, where the deep deterministic policy gradient (DDPG) and dueling double deep Q networks (D3QN) are invoked to tackle continuous and discrete action domains, respectively. Specifically, we improve the actor-critic structure of DDPG by combining D3QN. It utilizes the actor part of DDPG to search for the optimal offloading rate and power control of local execution. Meanwhile, it combines the critic part of DDPG with D3QN to select the optimal server for offloading. Simulation results demonstrate the proposed DDPG-D3QN algorithm has better stability and faster convergence, while achieving higher rewards than the existing DRL-based methods. Furthermore, the edge-cloud collaboration approach outperforms non-collaborative schemes.
引用
收藏
页码:10696 / 10709
页数:14
相关论文