SurgeNAS: A Comprehensive Surgery on Hardware-Aware Differentiable Neural Architecture Search

被引:8
|
作者
Luo, Xiangzhong [1 ]
Liu, Di [2 ]
Kong, Hao [1 ]
Huai, Shuo [1 ]
Chen, Hui [1 ]
Liu, Weichen [1 ]
机构
[1] Nanyang Technol Univ, Sch Comp Sci & Engn, Singapore 639798, Singapore
[2] Nanyang Technol Univ, HP NTU Digital Mfg Corp Lab, Singapore 639798, Singapore
关键词
Hardware; Task analysis; Optimization; Memory management; Estimation; Graph neural networks; Computers; Hardware-aware differentiable neural architecture search; graph neural networks; hardware performance prediction;
D O I
10.1109/TC.2022.3188175
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Differentiable neural architecture search (NAS) is an emerging paradigm to automate the design of top-performing convolutional neural networks (CNNs). Nonetheless, existing differentiable NAS methods suffer from several crucial weaknesses, such as inaccurate gradient estimation, high memory consumption, search fairness, etc. In this work, we introduce a novel hardware-aware differentiable NAS framework, namely SurgeNAS, in which we leverage the one-level optimization to avoid inaccuracy in gradient estimation. To this end, we propose an effective identity mapping regularization to alleviate the over-selecting issue. Besides, to mitigate the memory bottleneck, we propose an ordered differentiable sampling approach, which significantly reduces the search memory consumption to the single-path level, thereby allowing to directly search on target tasks instead of small proxy tasks. Meanwhile, it guarantees the strict search fairness. Moreover, we introduce a graph neural networks (GNNs) based predictor to approximate the on-device latency, which is further integrated into SurgeNAS to enable the latency-aware architecture search. Finally, we analyze the resource underutilization issue, in which we propose to scale up the searched SurgeNets within Comfort Zone to balance the computation and memory access, which brings considerable accuracy improvement without deteriorating the execution efficiency. Extensive experiments are conducted on ImageNet with diverse hardware platforms, which clearly show the effectiveness of SurgeNAS in terms of accuracy, latency, and search efficiency.
引用
收藏
页码:1081 / 1094
页数:14
相关论文
共 50 条
  • [31] An Affordable Hardware-Aware Neural Architecture Search for Deploying Convolutional Neural Networks on Ultra-Low-Power Computing Platforms
    Garavagno, Andrea Mattia
    Ragusa, Edoardo
    Frisoli, Antonio
    Gastaldo, Paolo
    IEEE SENSORS LETTERS, 2024, 8 (05) : 1 - 4
  • [32] SqueezeNext: Hardware-Aware Neural Network Design
    Gholami, Amir
    Kwon, Kiseok
    Wu, Bichen
    Tai, Zizheng
    Yue, Xiangyu
    Jin, Peter
    Zhao, Sicheng
    Keutzer, Kurt
    PROCEEDINGS 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW), 2018, : 1719 - 1728
  • [33] Compression-Accuracy Co-Optimization Through Hardware-Aware Neural Architecture Search for Vibration Damage Detection
    Ragusa, Edoardo
    Zonzini, Federica
    De Marchi, Luca
    Zunino, Rodolfo
    IEEE INTERNET OF THINGS JOURNAL, 2024, 11 (19): : 31745 - 31757
  • [34] TinyOdom: Hardware-Aware Efficient Neural Inertial Navigation
    Saha, Swapnil Sayan
    Sandha, Sandeep Singh
    Garcia, Luis Antonio
    Srivastava, Mani
    PROCEEDINGS OF THE ACM ON INTERACTIVE MOBILE WEARABLE AND UBIQUITOUS TECHNOLOGIES-IMWUT, 2022, 6 (02):
  • [35] Hardware-Aware and Efficient Feature Fusion Network Search
    Guo J.-M.
    Zhang R.
    Zhi T.
    He D.-Y.
    Huang D.
    Chang M.
    Zhang X.-S.
    Guo Q.
    Jisuanji Xuebao/Chinese Journal of Computers, 2022, 45 (11): : 2420 - 2432
  • [36] Fine-grained complexity-driven latency predictor in hardware-aware neural architecture search using composite loss
    Lin, Chengmin
    Yang, Pengfei
    Li, Chengcheng
    Cheng, Fei
    Lv, Wenkai
    Wang, Zhenyi
    Wang, Quan
    INFORMATION SCIENCES, 2024, 676
  • [37] Hardware-aware approach to deep neural network optimization
    Li, Hengyi
    Meng, Lin
    NEUROCOMPUTING, 2023, 559
  • [38] Hardware-Aware Softmax Approximation for Deep Neural Networks
    Geng, Xue
    Lin, Jie
    Zhao, Bin
    Kong, Anmin
    Aly, Mohamed M. Sabry
    Chandrasekhar, Vijay
    COMPUTER VISION - ACCV 2018, PT IV, 2019, 11364 : 107 - 122
  • [39] Hardware-Aware Quantization for Multiplierless Neural Network Controllers
    Habermann, Tobias
    Kuehle, Jonas
    Kumm, Martin
    Volkova, Anastasia
    2022 IEEE ASIA PACIFIC CONFERENCE ON CIRCUITS AND SYSTEMS, APCCAS, 2022, : 541 - 545
  • [40] HASS: Hardware-Aware Sparsity Search for Dataflow DNN Accelerator
    Yu, Zhewen
    Sreeram, Sudarshan
    Agrawal, Krish
    Wu, Junyi
    Montgomerie-Corcoran, Alexander
    Zhang, Cheng
    Cheng, Jianyi
    Bouganis, Christos-Savvas
    Zhao, Yiren
    2024 34TH INTERNATIONAL CONFERENCE ON FIELD-PROGRAMMABLE LOGIC AND APPLICATIONS, FPL 2024, 2024, : 257 - 263