Incentive-Aware Partitioning and Offloading Scheme for Inference Services in Edge Computing

被引:0
|
作者
Kim, TaeYoung [1 ]
Kim, Chang Kyung [1 ]
Lee, Seung-seob [2 ]
Lee, Sukyoung [1 ]
机构
[1] Yonsei Univ, Dept Comp Sci, Seoul 03722, South Korea
[2] Yale Univ, New Haven, CT 06520 USA
基金
新加坡国家研究基金会;
关键词
Task analysis; Delays; Computational modeling; Servers; Edge computing; Games; Artificial neural networks; Incentive; DNN Partitioning; offloading; scheduling; utility; energy; inference delay; game theory; RESOURCE-ALLOCATION; DRIVEN; DEPLOYMENT; MECHANISM;
D O I
10.1109/TSC.2024.3359148
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Owing to remarkable improvements in deep neural networks (DNNs), various computation-intensive and delay-sensitive DNN services have been developed for smart IoT devices. However, employing these services on the devices is challenging due to their limited battery capacity and computational constraints. Although edge computing is proposed as a solution, edge devices cannot meet the performance requirements of DNN services because the majority of IoT applications require simultaneous inference services, and DNN models grow larger. To address this problem, we propose a framework that enables parallel execution of partitioned and offloaded DNN inference services over multiple distributed edge devices. Noteworthy, edge devices are reluctant to process tasks due to their energy consumption. Thus, to provide an incentive mechanism for edge devices, we model the interaction between the edge devices and DNN inference service users as a two-level Stackelberg game. Based on this model, we design the proposed framework to determine the optimal scheduling with a partitioning strategy, aiming to maximize user satisfaction while incentivizing the participation of edge devices. We further derive the Nash equilibrium points in the two levels. The simulation results show that the proposed scheme outperforms other benchmark methods in terms of user satisfaction and profits of edge devices.
引用
收藏
页码:1580 / 1592
页数:13
相关论文
共 50 条
  • [1] An Incentive-Aware Job Offloading Control Framework for Multi-Access Edge Computing
    Li, Lingxiang
    Quek, Tony Q. S.
    Ren, Ju
    Yang, Howard H.
    Chen, Zhi
    Zhang, Yaoxue
    IEEE TRANSACTIONS ON MOBILE COMPUTING, 2021, 20 (01) : 63 - 75
  • [2] Incentive-Aware Blockchain-Assisted Intelligent Edge Caching and Computation Offloading for IoT
    Wang, Qian
    Chen, Siguang
    Wu, Meng
    ENGINEERING, 2023, 31 : 127 - 138
  • [3] Delay-Optimal and Incentive-Aware Computation Offloading for Reconfigurable Intelligent Surface-Assisted Mobile Edge Computing
    Kumar, Vikas
    Mukherjee, Mithun
    Lloret, Jaime
    Zhang, Qi
    Kumari, Mamta
    IEEE Networking Letters, 2022, 4 (03): : 127 - 131
  • [4] Incentive-aware virtual machine scheduling in cloud computing
    Xu, Heyang
    Liu, Yang
    Wei, Wei
    Zhang, Wenqiang
    JOURNAL OF SUPERCOMPUTING, 2018, 74 (07): : 3016 - 3038
  • [5] A Task Partitioning and Offloading Scheme in Vehicular Edge Computing Networks
    Qi, Wen
    Xia, Xu
    Wang, Heng
    Xing, Yanxia
    2021 IEEE 94TH VEHICULAR TECHNOLOGY CONFERENCE (VTC2021-FALL), 2021,
  • [6] Incentive-aware virtual machine scheduling in cloud computing
    Heyang Xu
    Yang Liu
    Wei Wei
    Wenqiang Zhang
    The Journal of Supercomputing, 2018, 74 : 3016 - 3038
  • [7] Incentive-Aware Micro Computing Cluster Formation for Cooperative Fog Computing
    Luo, Siqi
    Chen, Xu
    Zhou, Zhi
    Chen, Xiang
    Wu, Weigang
    IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2020, 19 (04) : 2643 - 2657
  • [8] Dewing in Fog: Incentive-Aware Micro Computing Cluster Formation for Fog Computing
    Luo, Siqi
    Zhou, Zhi
    Chen, Xiang
    Wu, Weigang
    2018 IEEE 24TH INTERNATIONAL CONFERENCE ON PARALLEL AND DISTRIBUTED SYSTEMS (ICPADS 2018), 2018, : 722 - 729
  • [9] Reliability and Latency Aware Code-Partitioning Offloading in Mobile Edge Computing
    Liu, Jianhui
    Zhang, Qi
    2019 IEEE WIRELESS COMMUNICATIONS AND NETWORKING CONFERENCE (WCNC), 2019,
  • [10] Collaborative Inference Acceleration Integrating DNN Partitioning and Task Offloading in Mobile Edge Computing
    Xu, Wenxiu
    Yin, Yin
    Chen, Ningjiang
    Tu, Huan
    INTERNATIONAL JOURNAL OF SOFTWARE ENGINEERING AND KNOWLEDGE ENGINEERING, 2023, 33 (11N12) : 1835 - 1863