Optimizing Resource Allocation for Joint AI Model Training and Task Inference in Edge Intelligence Systems

被引:14
|
作者
Li, Xian [1 ]
Bi, Suzhi [1 ,2 ]
Wang, Hui [1 ]
机构
[1] Shenzhen Univ, Coll Elect & Informat Engn, Shenzhen 518060, Peoples R China
[2] Peng Cheng Lab, Shenzhen 518066, Peoples R China
基金
中国国家自然科学基金;
关键词
Task analysis; Artificial intelligence; Training; Data models; Computational modeling; Resource management; Energy consumption; Edge intelligence; distributed training; resource allocation; alternating direction method of multipliers;
D O I
10.1109/LWC.2020.3036852
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This letter considers an edge intelligence system where multiple end users (EUs) collaboratively train an artificial intelligence (AI) model under the coordination of an edge server (ES) and the ES in return assists the AI inference task computation of EUs. Aiming at minimizing the energy consumption and execution latency of the EUs, we jointly consider the model training and task inference processes to optimize the local CPU frequency and task splitting ratio (i.e., the portion of task executed at the ES) of each EU, and the system bandwidth allocation. In particular, each task splitting ratio is correlated to a binary decision that represents whether downloading the trained AI model for local task inference. The problem is formulated as a hard mixed integer non-linear programming (MINLP). To tackle the combinatorial binary decisions, we propose a decomposition-oriented method by leveraging the ADMM (alternating direction method of multipliers) technique, whereby the primal MINLP problem is decomposed into multiple parallel sub-problems that can be efficiently handled. The proposed method enjoys linear complexity with the network size and simulation results show that it achieves near-optimal performance (less than 3.18% optimality gap), which significantly outperforms the considered benchmark algorithms.
引用
收藏
页码:532 / 536
页数:5
相关论文
共 50 条
  • [1] Optimizing AI Service Placement and Resource Allocation in Mobile Edge Intelligence Systems
    Lin, Zehong
    Bi, Suzhi
    Zhang, Ying-Jun Angela
    IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2021, 20 (11) : 7257 - 7271
  • [2] Online Resource Allocation for Edge Intelligence with Colocated Model Retraining and Inference
    Cai, Huaiguang
    Zhou, Zhi
    Huang, Qianyi
    IEEE INFOCOM 2024-IEEE CONFERENCE ON COMPUTER COMMUNICATIONS, 2024, : 1900 - 1909
  • [3] Joint Task offload and Resource Allocation for Cognitive Edge Computing Using AI Algorithm
    Li, Cuiling
    Deng, Xiaofang
    Qin, Huipin
    Zheng, Lin
    Qiu, Hongbing
    2021 IEEE/ACIS 21ST INTERNATIONAL FALL CONFERENCE ON COMPUTER AND INFORMATION SCIENCE (ICIS 2021-FALL), 2021, : 275 - 280
  • [4] Joint Foundation Model Caching and Inference of Generative AI Services for Edge Intelligence
    Xu, Minrui
    Niyato, Dusit
    Zhang, Hongliang
    Kang, Jiawen
    Xiong, Zehui
    Mao, Shiwen
    Han, Zhu
    IEEE CONFERENCE ON GLOBAL COMMUNICATIONS, GLOBECOM, 2023, : 3548 - 3553
  • [5] Joint Task and Resource Allocation for Mobile Edge Learning
    Abutuleb, Amr
    Sorour, Sameh
    Hassanein, Hossam S.
    2020 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM), 2020,
  • [6] Joint Task Offloading and Resource Allocation for Quality-Aware Edge-Assisted Machine Learning Task Inference
    Fan, Wenhao
    Chen, Zeyu
    Hao, Zhibo
    Wu, Fan
    Liu, Yuan'an
    IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2023, 72 (05) : 6739 - 6752
  • [7] DNN Deployment, Task Offloading, and Resource Allocation for Joint Task Inference in IIoT
    Fan, Wenhao
    Chen, Zeyu
    Hao, Zhibo
    Su, Yi
    Wu, Fan
    Tang, Bihua
    Liu, Yuan'an
    IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2023, 19 (02) : 1634 - 1646
  • [8] Research on cloud-edge joint task inference algorithm in edge intelligence
    Zheng, Yaping
    Journal of Computers (Taiwan), 2021, 32 (04) : 211 - 224
  • [9] Joint Task Offloading and Resource Allocation in Heterogeneous Edge Environments
    Liu, Yu
    Mao, Yingling
    Liu, Zhenhua
    Ye, Fan
    Yang, Yuanyuan
    IEEE TRANSACTIONS ON MOBILE COMPUTING, 2024, 23 (06) : 7318 - 7334
  • [10] Decentralized Convex Optimization for Joint Task Offloading and Resource Allocation of Vehicular Edge Computing Systems
    Tan, Kaige
    Feng, Lei
    Dan, Gyorgy
    Torngren, Martin
    IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2022, 71 (12) : 13226 - 13241