RelativeNAS: Relative Neural Architecture Search via Slow-Fast Learning

被引:29
|
作者
Tan, Hao [1 ]
Cheng, Ran [1 ]
Huang, Shihua [1 ]
He, Cheng [1 ]
Qiu, Changxiao [2 ]
Yang, Fan [2 ]
Luo, Ping [3 ]
机构
[1] Southern Univ Sci & Technol, Univ Key Lab Evolving Intelligent Syst Guangdong, Dept Comp Sci & Engn, Shenzhen 518055, Peoples R China
[2] Huawei Technol Co Ltd, Hisilicon Res Dept, Shenzhen 518055, Peoples R China
[3] Univ Hong Kong, Dept Comp Sci, Hong Kong, Peoples R China
基金
中国国家自然科学基金;
关键词
Computer architecture; Statistics; Sociology; Search problems; Optimization; Neural networks; Estimation; AutoML; convolutional neural network (CNN); neural architecture search (NAS); population-based search; slow-fast learning; NETWORKS;
D O I
10.1109/TNNLS.2021.3096658
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Despite the remarkable successes of convolutional neural networks (CNNs) in computer vision, it is time-consuming and error-prone to manually design a CNN. Among various neural architecture search (NAS) methods that are motivated to automate designs of high-performance CNNs, the differentiable NAS and population-based NAS are attracting increasing interests due to their unique characters. To benefit from the merits while overcoming the deficiencies of both, this work proposes a novel NAS method, RelativeNAS. As the key to efficient search, RelativeNAS performs joint learning between fast learners (i.e., decoded networks with relatively lower loss value) and slow learners in a pairwise manner. Moreover, since RelativeNAS only requires low-fidelity performance estimation to distinguish each pair of fast learner and slow learner, it saves certain computation costs for training the candidate architectures. The proposed RelativeNAS brings several unique advantages: 1) it achieves state-of-the-art performances on ImageNet with top-1 error rate of 24.88%, that is, outperforming DARTS and AmoebaNet-B by 1.82% and 1.12%, respectively; 2) it spends only 9 h with a single 1080Ti GPU to obtain the discovered cells, that is, 3.75x and 7875x faster than DARTS and AmoebaNet, respectively; and 3) it provides that the discovered cells obtained on CIFAR-10 can be directly transferred to object detection, semantic segmentation, and keypoint detection, yielding competitive results of 73.1% mAP on PASCAL VOC, 78.7% mIoU on Cityscapes, and 68.5% AP on MSCOCO, respectively. The implementation of RelativeNAS is available at https://github.com/EMI-Group/RelativeNAS.
引用
收藏
页码:475 / 489
页数:15
相关论文
共 50 条
  • [1] Slow-Fast Duffing Neural Mass Model
    Jafarian, Amirhossein
    Freestone, Dean R.
    Nesic, Dragan
    Grayden, David B.
    2019 41ST ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY (EMBC), 2019, : 142 - 145
  • [2] Sliding Vector Fields via Slow-Fast Systems
    Llibre, Jaume
    da Silva, Paulo R.
    Teixeira, Marco A.
    BULLETIN OF THE BELGIAN MATHEMATICAL SOCIETY-SIMON STEVIN, 2008, 15 (05) : 851 - 869
  • [3] Multiscale analysis of slow-fast neuronal learning models with noise
    Galtier, Mathieu
    wainrib, Gilles
    JOURNAL OF MATHEMATICAL NEUROSCIENCE, 2012, 2
  • [4] Neural mass modeling of slow-fast dynamics of seizure initiation and abortion
    Ersoz, Elif Koksal
    Modolo, Julien
    Bartolomei, Fabrice
    Wendling, Fabrice
    Wendling, Fabrice (fabrice.wendling@inserm.fr), 1600, Public Library of Science (16):
  • [5] Approximate Neural Architecture Search via Operation Distribution Learning
    Wan, Xingchen
    Ru, Binxin
    Esparanca, Pedro M.
    Carlucci, Fabio M.
    2022 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV 2022), 2022, : 3545 - 3554
  • [6] Neural mass modeling of slow-fast dynamics of seizure initiation and abortion
    Koksal Ersoz, Elif
    Modolo, Julien
    Bartolomei, Fabrice
    Wendling, Fabrice
    PLOS COMPUTATIONAL BIOLOGY, 2020, 16 (11)
  • [7] Cascades of Periodic Solutions in a Neural Circuit With Delays and Slow-Fast Dynamics
    Ferrario, Andrea
    Rankin, James
    FRONTIERS IN APPLIED MATHEMATICS AND STATISTICS, 2021, 7
  • [8] Neural architecture search via standard machine learning methodologies
    Franchini, Giorgia
    Ruggiero, Valeria
    Porta, Federica
    Zanni, Luca
    MATHEMATICS IN ENGINEERING, 2023, 5 (01): : 1 - 21
  • [9] Fast and Practical Neural Architecture Search
    Cui, Jiequan
    Chen, Pengguang
    Li, Ruiyu
    Liu, Shu
    Shen, Xiaoyong
    Jia, Jiaya
    2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 6211 - 6220
  • [10] Magnomechanically induced transparency and tunable slow-fast light via a levitated
    Bayati, Sahar
    Harouni, Malek bagheri
    Mahdifar, Ali
    OPTICS EXPRESS, 2024, 32 (09): : 14914 - 14928