Evolutionary Neural Architecture Search for Transferable Networks

被引:1
|
作者
Zhou, Xun [1 ]
Liu, Songbai [2 ]
Qin, A. K. [3 ]
Tan, Kay Chen [1 ]
机构
[1] Hong Kong Polytech Univ, Dept Comp, Hung Hom, Hong Kong 999077, Peoples R China
[2] Shenzhen Univ, Coll Comp Sci & Software Engn, Shenzhen 518060, Peoples R China
[3] Swinburne Univ Technol, Dept Comp Technol, Hawthorn, Vic 3122, Australia
基金
中国国家自然科学基金; 澳大利亚研究理事会;
关键词
Evolutionary algorithm; neural architecture search; transferable architecture; TRANSFER OPTIMIZATION; GENETIC ALGORITHM;
D O I
10.1109/TETCI.2024.3427763
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The recent proliferation of edge computing has led to the deployment of deep neural networks (DNNs) on edge devices like smartphones and IoT devices to serve end users. However, developing the most suitable DNN model for each on-device task is nontrivial, due to data governance of these tasks and data heterogeneity across them. Existing approaches tackle this issue by learning task-specific models on the device, but this requires substantial computational resources, exacerbating the computational and energy demands on edge devices. This research strives to enhance the deployment efficiency of advanced models on edge devices, with a specific focus on reducing the on-device learning cost. In pursuit of this goal, we propose a category-specific but task-agnostic evolutionary neural architecture search (CSTA-ENAS) method. This method can utilize the available datasets from multiple other tasks in the same category as on-device tasks to design a transferable architecture on the server. Then, this architecture only requires light on-device fine-tuning to satisfactorily solve all different on-device tasks, significantly reducing the on-device learning time and related energy consumption. To improve the search efficiency of our method, a supernet-based partial training strategy is proposed to reduce the evaluation cost for candidate architectures. To showcase the effectiveness of CSTA-ENAS, we build transferable DNN models and evaluate their accuracies on a set of new image classification tasks. Our models demonstrate competitive performance compared to most of the existing task-specific models and transferable models while requiring fewer on-device computational resources.
引用
收藏
页数:13
相关论文
共 50 条
  • [31] Evolutionary Neural Architecture Search for Transformer in Knowledge Tracing
    Yang, Shangshang
    Yu, Xiaoshan
    Tian, Ye
    Yan, Xueming
    Ma, Haiping
    Zhang, Xingyi
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [32] BenchENAS: A Benchmarking Platform for Evolutionary Neural Architecture Search
    Xie, Xiangning
    Liu, Yuqiao
    Sun, Yanan
    Yen, Gary G.
    Xue, Bing
    Zhang, Mengjie
    IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, 2022, 26 (06) : 1473 - 1485
  • [33] Efficient evolutionary neural architecture search based on hybrid search space
    Gong, Tao
    Ma, Yongjie
    Xu, Yang
    Song, Changwei
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2024, 15 (08) : 3313 - 3326
  • [34] Hybrid Architecture-Based Evolutionary Robust Neural Architecture Search
    Yang, Shangshang
    Sun, Xiangkun
    Xu, Ke
    Liu, Yuanchao
    Tian, Ye
    Zhang, Xingyi
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2024, 8 (04): : 2919 - 2934
  • [35] Robustness Enhancement of Neural Networks via Architecture Search with Multi-Objective Evolutionary Optimization
    Chen, Haojie
    Huang, Hai
    Zuo, Xingquan
    Zhao, Xinchao
    MATHEMATICS, 2022, 10 (15)
  • [36] Hybrid evolutionary network architecture search (HyENAS) for convolution class of deep neural networks with applications
    Soniya
    Singh, Lotika
    Paul, Sandeep
    EXPERT SYSTEMS, 2023, 40 (04)
  • [37] Neural Architecture Search of SPD Manifold Networks
    Sukthanker, Rhea Sanjay
    Huang, Zhiwu
    Kumar, Suryansh
    Endsjo, Erik Goron
    Wu, Yan
    Van Gool, Luc
    PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 3002 - 3009
  • [38] Genetic Architecture Search for Binarized Neural Networks
    Chang, Yangyang
    Sobelman, Gerald E.
    Zhou, Xiaofang
    2019 IEEE 13TH INTERNATIONAL CONFERENCE ON ASIC (ASICON), 2019,
  • [39] Efficient Architecture Search for Deep Neural Networks
    Gottapu, Ram Deepak
    Dagli, Cihan H.
    COMPLEX ADAPTIVE SYSTEMS, 2020, 168 : 19 - 25
  • [40] Arch-Graph: Acyclic Architecture Relation Predictor for Task-Transferable Neural Architecture Search
    Huang, Minbin
    Huang, Zhijian
    Li, Changlin
    Chen, Xin
    Xu, Hang
    Li, Zhenguo
    Liang, Xiaodan
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 11871 - 11881