An Offloading Algorithm for Maximizing Inference Accuracy on Edge Device in an Edge Intelligence System

被引:5
|
作者
Fresa, Andrea [1 ]
Varma Champati, Jaya Prakash [2 ]
机构
[1] Univ Carlos III Madrid, Edge Networks Grp, IMDEA Networks Inst, Madrid, Spain
[2] IMDEA Networks Inst, Edge Networks Grp, Madrid, Spain
关键词
Approximation ratio algorithm; Machine Learning Inference; Edge Computing;
D O I
10.1145/3551659.3559044
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
With the emergence of edge computing, the problem of offloading jobs between an Edge Device (ED) and an Edge Server (ES) received significant attention in the past. Motivated by the fact that an increasing number of applications are using Machine Learning (ML) inference from the data samples collected at the EDs, we study the problem of offloading inference jobs by considering the following novel aspects: 1) in contrast to a typical computational job an inference job has accuracy measure, 2) both inference accuracy and processing time of an inference job increases with the size of the ML model, and 3) recently proposed Deep Neural Networks (DNNs) for resource-constrained EDs provide the choice of scaling down the model size by trading off the inference accuracy. Therefore, we consider a newsystem with multiple small-size ML models at the ED and a powerful large-size ML model at the ES and study the problem of offloading inference jobs with the objective of maximizing the total inference accuracy at the ED subject to a time constraint.. on the makespan. Noting that the problem is NP-hard, we propose an approximation algorithm: Accuracy Maximization using LP-Relaxation and Rounding (AMR(2)), and prove that it results in a makespan at most 2T, and achieves a total accuracy that is lower by a small constant (less than 1) from the optimal total accuracy. As proof of concept, we implemented AMR2 on a Raspberry Pi, equipped with MobileNets, that is connected via LAN to a server equipped with ResNet, and studied the total accuracy and makespan performance of AMR(2) for image classification.
引用
收藏
页码:15 / 23
页数:9
相关论文
共 50 条
  • [1] Offloading Algorithms for Maximizing Inference Accuracy on Edge Device in an Edge Intelligence System
    Fresa, Andrea
    Champati, Jaya Prakash
    IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2023, 34 (07) : 2025 - 2039
  • [2] Research on cloud-edge joint task inference algorithm in edge intelligence
    Zheng, Yaping
    Journal of Computers (Taiwan), 2021, 32 (04) : 211 - 224
  • [3] Accuracy-Based Task Offloading and Resource Allocation for Edge Intelligence in IoT
    Fan, Wenhao
    Chen, Zeyu
    Su, Yi
    Wu, Fan
    Tang, Bihua
    Liu, Yuan'an
    IEEE WIRELESS COMMUNICATIONS LETTERS, 2022, 11 (02) : 371 - 375
  • [4] Mobility and Cost Aware Inference Accelerating Algorithm for Edge Intelligence
    Yuan, Xin
    Li, Ning
    Wei, Kang
    Xu, Wenchao
    Chen, Quan
    Chen, Hao
    Guo, Song
    IEEE TRANSACTIONS ON MOBILE COMPUTING, 2025, 24 (03) : 1530 - 1549
  • [5] Microservice Replacement Algorithm in Cloud-Edge System for Edge Intelligence
    Miao, Weiwei
    Zeng, Zeng
    Li, Shihao
    Wei, Lei
    Jiang, Chengling
    Quan, Siping
    Li, Yong
    19TH IEEE INTERNATIONAL SYMPOSIUM ON PARALLEL AND DISTRIBUTED PROCESSING WITH APPLICATIONS (ISPA/BDCLOUD/SOCIALCOM/SUSTAINCOM 2021), 2021, : 1737 - 1744
  • [6] Compensational Computation Offloading for Maximizing Lifetime of Edge Networks
    Fan, Wenhao
    Chen, Jiayi
    Wu, Fan
    Tang, Bihua
    2019 IEEE 89TH VEHICULAR TECHNOLOGY CONFERENCE (VTC2019-SPRING), 2019,
  • [7] A Greedy Algorithm for Task Offloading in Mobile Edge Computing System
    Feng Wei
    Sixuan Chen
    Weixia Zou
    中国通信, 2018, 15 (11) : 149 - 157
  • [8] A Greedy Algorithm for Task Offloading in Mobile Edge Computing System
    Wei, Feng
    Chen, Sixuan
    Zou, Weixia
    CHINA COMMUNICATIONS, 2018, 15 (11) : 149 - 157
  • [9] Survey of Collaborative Inference for Edge Intelligence
    Wang R.
    Qi J.
    Chen L.
    Yang L.
    Jisuanji Yanjiu yu Fazhan/Computer Research and Development, 2023, 60 (02): : 398 - 414
  • [10] Edge Intelligence: On-Demand Deep Learning Model Co-Inference with Device-Edge Synergy
    Li, En
    Zhou, Zhi
    Chen, Xu
    MECOMM'18: PROCEEDINGS OF THE 2018 WORKSHOP ON MOBILE EDGE COMMUNICATIONS, 2018, : 31 - 36