ATNAS: Automatic Termination for Neural Architecture Search

被引:1
|
作者
Sakamoto, Kotaro [1 ]
Ishibashi, Hideaki [2 ]
Sato, Rei [3 ]
Shirakawa, Shinichi [4 ]
Akimoto, Youhei [5 ,6 ]
Hino, Hideitsu [1 ,6 ]
机构
[1] Inst Stat Math, 10-3 Midori Cho, Tachikawa, Tokyo 1900014, Japan
[2] Kyushu Inst Technol, 1-1 Sensui Cho,Tobata Ku, Fukuoka 8048550, Japan
[3] LINE Corp, 1-6-1 Yotsuya,Shinjuku Ku, Tokyo 1600004, Japan
[4] Yokohama Natl Univ, 79-8 Tokiwadai,Hodogaya Ku, Yokohama 2408501, Japan
[5] Univ Tsukuba, 1-1-1 Tennodai, Tsukuba, Ibaraki 3058577, Japan
[6] RIKEN, Ctr Adv Intelligence Project, 1-4-1 Nihonbashi,Chuo Ku, Tokyo 1030027, Japan
关键词
Neural Architecture Search; Deep learning;
D O I
10.1016/j.neunet.2023.07.011
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Neural architecture search (NAS) is a framework for automating the design process of a neural network structure. While the recent one-shot approaches have reduced the search cost, there still exists an inherent trade-off between cost and performance. It is important to appropriately stop the search and further reduce the high cost of NAS. Meanwhile, the differentiable architecture search (DARTS), a typical one-shot approach, is known to suffer from overfitting. Heuristic early-stopping strategies have been proposed to overcome such performance degradation. In this paper, we propose a more versatile and principled early-stopping criterion on the basis of the evaluation of a gap between expectation values of generalisation errors of the previous and current search steps with respect to the architecture parameters. The stopping threshold is automatically determined at each search epoch without cost. In numerical experiments, we demonstrate the effectiveness of the proposed method. We stop the one-shot NAS algorithms and evaluate the acquired architectures on the benchmark datasets: NASBench-201 and NATS-Bench. Our algorithm is shown to reduce the cost of the search process while maintaining a high performance. (c) 2023 The Author(s). Published by Elsevier Ltd. This is an open access article under the CC BY license (http://creativecommons.org/licenses/by/4.0/).
引用
收藏
页码:446 / 458
页数:13
相关论文
共 50 条
  • [41] An architecture entropy regularizer for differentiable neural architecture search
    Jing, Kun
    Chen, Luoyu
    Xu, Jungang
    NEURAL NETWORKS, 2023, 158 : 111 - 120
  • [42] Trends in Neural Architecture Search: Towards the Acceleration of Search
    Kim, Youngkee
    Yun, Won Joon
    Lee, Youn Kyu
    Jung, Soyi
    Kim, Joongheon
    12TH INTERNATIONAL CONFERENCE ON ICT CONVERGENCE (ICTC 2021): BEYOND THE PANDEMIC ERA WITH ICT CONVERGENCE INNOVATION, 2021, : 421 - 424
  • [43] Neural Architecture Search: A Visual Analysis
    Ochoa, Gabriela
    Veerapen, Nadarajen
    PARALLEL PROBLEM SOLVING FROM NATURE - PPSN XVII, PPSN 2022, PT I, 2022, 13398 : 603 - 615
  • [44] Neural network architecture search with AMBER
    Yi Zhang
    Yang Liu
    X. Shirley Liu
    Nature Machine Intelligence, 2021, 3 : 372 - 373
  • [45] Evolutionary approximation and neural architecture search
    Pinos, Michal
    Mrazek, Vojtech
    Sekanina, Lukas
    GENETIC PROGRAMMING AND EVOLVABLE MACHINES, 2022, 23 (03) : 351 - 374
  • [46] Neural Architecture Search for Transformers: A Survey
    Chitty-Venkata, Krishna Teja
    Emani, Murali
    Vishwanath, Venkatram
    Somani, Arun K.
    IEEE ACCESS, 2022, 10 : 108374 - 108412
  • [47] Deep Multimodal Neural Architecture Search
    Yu, Zhou
    Cui, Yuhao
    Yu, Jun
    Wang, Meng
    Tao, Dacheng
    Tian, Qi
    MM '20: PROCEEDINGS OF THE 28TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, 2020, : 3743 - 3752
  • [48] VNAS: Variational Neural Architecture Search
    Ma, Benteng
    Zhang, Jing
    Xia, Yong
    Tao, Dacheng
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2024, 132 (09) : 3689 - 3713
  • [49] A Neural Architecture Search for Automated Multimodal
    Singh, Anuraj
    Nair, Haritha
    EXPERT SYSTEMS WITH APPLICATIONS, 2022, 207
  • [50] Deep Neural Architecture Search: A Survey
    Meng Z.
    Gu X.
    Liang Y.
    Xu D.
    Wu C.
    Jisuanji Yanjiu yu Fazhan/Computer Research and Development, 2021, 58 (01): : 22 - 33