Optimizing Resource Allocation for Joint AI Model Training and Task Inference in Edge Intelligence Systems

被引:14
|
作者
Li, Xian [1 ]
Bi, Suzhi [1 ,2 ]
Wang, Hui [1 ]
机构
[1] Shenzhen Univ, Coll Elect & Informat Engn, Shenzhen 518060, Peoples R China
[2] Peng Cheng Lab, Shenzhen 518066, Peoples R China
基金
中国国家自然科学基金;
关键词
Task analysis; Artificial intelligence; Training; Data models; Computational modeling; Resource management; Energy consumption; Edge intelligence; distributed training; resource allocation; alternating direction method of multipliers;
D O I
10.1109/LWC.2020.3036852
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This letter considers an edge intelligence system where multiple end users (EUs) collaboratively train an artificial intelligence (AI) model under the coordination of an edge server (ES) and the ES in return assists the AI inference task computation of EUs. Aiming at minimizing the energy consumption and execution latency of the EUs, we jointly consider the model training and task inference processes to optimize the local CPU frequency and task splitting ratio (i.e., the portion of task executed at the ES) of each EU, and the system bandwidth allocation. In particular, each task splitting ratio is correlated to a binary decision that represents whether downloading the trained AI model for local task inference. The problem is formulated as a hard mixed integer non-linear programming (MINLP). To tackle the combinatorial binary decisions, we propose a decomposition-oriented method by leveraging the ADMM (alternating direction method of multipliers) technique, whereby the primal MINLP problem is decomposed into multiple parallel sub-problems that can be efficiently handled. The proposed method enjoys linear complexity with the network size and simulation results show that it achieves near-optimal performance (less than 3.18% optimality gap), which significantly outperforms the considered benchmark algorithms.
引用
收藏
页码:532 / 536
页数:5
相关论文
共 50 条
  • [41] Joint Task Offloading and Resource Allocation for Vehicular Edge Computing With Result Feedback Delay
    Nan, Zhaojun
    Zhou, Sheng
    Jia, Yunjian
    Niu, Zhisheng
    IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2023, 22 (10) : 6547 - 6561
  • [42] Joint Task Assignment and Wireless Resource Allocation for Cooperative Mobile-Edge Computing
    Xing, Hong
    Liu, Liang
    Xu, Jie
    Nallanathan, Arumugam
    2018 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC), 2018,
  • [43] QoS-Aware Joint Task Scheduling and Resource Allocation in Vehicular Edge Computing
    Cao, Chenhong
    Su, Meijia
    Duan, Shengyu
    Dai, Miaoling
    Li, Jiangtao
    Li, Yufeng
    SENSORS, 2022, 22 (23)
  • [44] Joint Task Offloading and Resource Allocation for Energy-Constrained Mobile Edge Computing
    Jiang, Hongbo
    Dai, Xingxia
    Xiao, Zhu
    Iyengar, Arun
    IEEE TRANSACTIONS ON MOBILE COMPUTING, 2023, 22 (07) : 4000 - 4015
  • [45] Joint Optimization of Task Caching,Computation Offloading and Resource Allocation for Mobile Edge Computing
    Zhixiong Chen
    Zhengchuan Chen
    Zhi Ren
    Liang Liang
    Wanli Wen
    Yunjian Jia
    China Communications, 2022, 19 (12) : 142 - 159
  • [46] Joint Task Offloading and Resource Allocation in UAV-Enabled Mobile Edge Computing
    Yu, Zhe
    Gong, Yanmin
    Gong, Shimin
    Guo, Yuanxiong
    IEEE INTERNET OF THINGS JOURNAL, 2020, 7 (04) : 3147 - 3159
  • [47] Joint Task Offloading and Resource Allocation in Vehicular Edge Computing Networks for Emergency Logistics
    Li, Rui
    Ling, Darong
    Wang, Yisheng
    Zhao, Shuang
    Wang, Jun
    Li, Jun
    Mathematical Problems in Engineering, 2023, 2023
  • [48] Joint Optimization of Task Caching, Computation Offloading and Resource Allocation for Mobile Edge Computing
    Chen, Zhixiong
    Chen, Zhengchuan
    Ren, Zhi
    Liang, Liang
    Wen, Wanli
    Jia, Yunjian
    CHINA COMMUNICATIONS, 2022, 19 (12) : 142 - 159
  • [49] Enabling Latency-Sensitive DNN Inference via Joint Optimization of Model Surgery and Resource Allocation in Heterogeneous Edge
    Huang, Zhaowu
    Dong, Fang
    Shen, Dian
    Wang, Huitian
    Guo, Xiaolin
    Fu, Shucun
    51ST INTERNATIONAL CONFERENCE ON PARALLEL PROCESSING, ICPP 2022, 2022,
  • [50] Optimizing Task Offloading and Resource Allocation in Vehicular Edge Computing Based on Heterogeneous Cellular Networks
    Fan, Xinggang
    Gu, Wenting
    Long, Changqing
    Gu, Chaojie
    He, Shibo
    IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2024, 73 (05) : 7175 - 7187