Aries: Efficient Testing of Deep Neural Networks via Labeling-Free Accuracy Estimation

被引:7
|
作者
Hu, Qiang [1 ]
Guo, Yuejun [2 ]
Xie, Xiaofei [3 ]
Cordy, Maxime [1 ]
Papadakis, Mike [1 ]
Ma, Lei [4 ,5 ]
Le Traon, Yves [1 ]
机构
[1] Univ Luxembourg, Luxembourg, Luxembourg
[2] Luxembourg Inst Sci & Technol, Luxembourg, Luxembourg
[3] Singapore Management Univ, Singapore, Singapore
[4] Univ Alberta, Edmonton, AB, Canada
[5] Univ Tokyo, Tokyo, Japan
来源
2023 IEEE/ACM 45TH INTERNATIONAL CONFERENCE ON SOFTWARE ENGINEERING, ICSE | 2023年
基金
加拿大自然科学与工程研究理事会;
关键词
deep learning testing; performance estimation; distribution shift;
D O I
10.1109/ICSE48619.2023.00152
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Deep learning (DL) plays a more and more important role in our daily life due to its competitive performance in industrial application domains. As the core of DL-enabled systems, deep neural networks (DNNs) need to be carefully evaluated to ensure the produced models match the expected requirements. In practice, the de facto standard to assess the quality of DNNs in the industry is to check their performance (accuracy) on a collected set of labeled test data. However, preparing such labeled data is often not easy partly because of the huge labeling effort, i.e., data labeling is labor-intensive, especially with the massive new incoming unlabeled data every day. Recent studies show that test selection for DNN is a promising direction that tackles this issue by selecting minimal representative data to label and using these data to assess the model. However, it still requires human effort and cannot be automatic. In this paper, we propose a novel technique, named Aries, that can estimate the performance of DNNs on new unlabeled data using only the information obtained from the original test data. The key insight behind our technique is that the model should have similar prediction accuracy on the data which have similar distances to the decision boundary. We performed a large-scale evaluation of our technique on two famous datasets, CIFAR-10 and Tiny-ImageNet, four widely studied DNN models including ResNet101 and DenseNet-121, and 13 types of data transformation methods. Results show that the estimated accuracy by Aries is only 0.03% - 2.60% off the true accuracy. Besides, Aries also outperforms the state-of-the-art labeling-free methods in 50 out of 52 cases and selection-labeling-based methods in 96 out of 128 cases.
引用
收藏
页码:1776 / 1787
页数:12
相关论文
共 50 条
  • [41] Efficient generation of valid test inputs for deep neural networks via gradient search
    Jiang, Zhouxian
    Li, Honghui
    Wang, Rui
    JOURNAL OF SOFTWARE-EVOLUTION AND PROCESS, 2024, 36 (04)
  • [42] Efficient Deep Neural Networks for Edge Computing
    Alnemari, Mohammed
    Bagherzadeh, Nader
    2019 IEEE INTERNATIONAL CONFERENCE ON EDGE COMPUTING (IEEE EDGE), 2019, : 1 - 7
  • [43] Efficient Model Averaging for Deep Neural Networks
    Opitz, Michael
    Possegger, Horst
    Bischof, Horst
    COMPUTER VISION - ACCV 2016, PT II, 2017, 10112 : 205 - 220
  • [44] Bit Efficient Quantization for Deep Neural Networks
    Nayak, Prateeth
    Zhang, David
    Chai, Sek
    FIFTH WORKSHOP ON ENERGY EFFICIENT MACHINE LEARNING AND COGNITIVE COMPUTING - NEURIPS EDITION (EMC2-NIPS 2019), 2019, : 52 - 56
  • [45] The Efficient Hedging Frontier with Deep Neural Networks
    Gong, Zheng
    Ventre, Carmine
    O'Hara, John
    ICAIF 2021: THE SECOND ACM INTERNATIONAL CONFERENCE ON AI IN FINANCE, 2021,
  • [46] Efficient Architecture Search for Deep Neural Networks
    Gottapu, Ram Deepak
    Dagli, Cihan H.
    COMPLEX ADAPTIVE SYSTEMS, 2020, 168 : 19 - 25
  • [47] Deep Neural Networks with Efficient Guaranteed Invariances
    Rath, Matthias
    Condurache, Alexandru Paul
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 206, 2023, 206
  • [48] Error detection and accuracy estimation in automatic speech recognition using deep bidirectional recurrent neural networks
    Ogawa, Atsunori
    Hori, Takaaki
    SPEECH COMMUNICATION, 2017, 89 : 70 - 83
  • [49] ADAPTIVE LABELING FOR HASH CODE LEARNING VIA NEURAL NETWORKS
    Yang, Huei-Fang
    Tu, Cheng-Hao
    Chen, Chu-Song
    2019 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2019, : 2244 - 2248
  • [50] Efficient scale estimation methods using lightweight deep convolutional neural networks for visual tracking
    Seyed Mojtaba Marvasti-Zadeh
    Hossein Ghanei-Yakhdan
    Shohreh Kasaei
    Neural Computing and Applications, 2021, 33 : 8319 - 8334