Towards Characterizing Adversarial Defects of Deep Learning Software from the Lens of Uncertainty

被引:55
|
作者
Zhang, Xiyue [1 ]
Xie, Xiaofei [2 ]
Ma, Lei [3 ]
Du, Xiaoning [2 ]
Hu, Qiang [3 ]
Liu, Yang [2 ]
Zhao, Jianjun [3 ]
Sun, Meng [1 ]
机构
[1] Peking Univ, Beijing, Peoples R China
[2] Nanyang Technol Univ, Singapore, Singapore
[3] Kyushu Univ, Fukuoka, Japan
基金
中国国家自然科学基金; 新加坡国家研究基金会;
关键词
Deep learning; uncertainty; adversarial attack; software testing; NEURAL-NETWORKS;
D O I
10.1145/3377811.3380368
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Over the past decade, deep learning (DL) has been successfully applied to many industrial domain-specific tasks. However, the current state-of-the-art DL software still suffers from quality issues, which raises great concern especially in the context of safety- and security-critical scenarios. Adversarial examples (AEs) represent a typical and important type of defects needed to be urgently addressed, on which a DL software makes incorrect decisions. Such defects occur through either intentional attack or physical-world noise perceived by input sensors, potentially hindering further industry deployment. The intrinsic uncertainty nature of deep learning decisions can be a fundamental reason for its incorrect behavior. Although some testing, adversarial attack and defense techniques have been recently proposed, it still lacks a systematic study to uncover the relationship between AEs and DL uncertainty. In this paper, we conduct a large-scale study towards bridging this gap. We first investigate the capability of multiple uncertainty metrics in differentiating benign examples (BEs) and AEs, which enables to characterize the uncertainty patterns of input data. Then, we identify and categorize the uncertainty patterns of BEs and AEs, and find that while BEs and AEs generated by existing methods do follow common uncertainty patterns, some other uncertainty patterns are largely missed. Based on this, we propose an automated testing technique to generate multiple types of uncommon AEs and BEs that are largely missed by existing techniques. Our further evaluation reveals that the uncommon data generated by our method is hard to be defended by the existing defense techniques with the average defense success rate reduced by 35%. Our results call for attention and necessity to generate more diverse data for evaluating quality assurance solutions of DL software.
引用
收藏
页码:739 / 751
页数:13
相关论文
共 50 条
  • [1] CuRTAIL: ChaRacterizing and Thwarting AdversarIal Deep Learning
    Javaheripi, Mojan
    Samragh, Mohammad
    Rouhani, Bita Darvish
    Javidi, Tara
    Koushanfar, Farinaz
    IEEE TRANSACTIONS ON DEPENDABLE AND SECURE COMPUTING, 2021, 18 (02) : 736 - 752
  • [2] Leveraging Uncertainty in Adversarial Learning to Improve Deep Learning Based Segmentation
    Javed, Mahed
    Mihaylova, Lyudmila
    2019 SYMPOSIUM ON SENSOR DATA FUSION: TRENDS, SOLUTIONS, APPLICATIONS (SDF 2019), 2019,
  • [3] DEEP ADVERSARIAL ACTIVE LEARNING WITH MODEL UNCERTAINTY FOR IMAGE CLASSIFICATION
    Zhu, Zheng
    Wang, Hongxing
    2020 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2020, : 1711 - 1715
  • [4] Certifiable Robustness to Adversarial State Uncertainty in Deep Reinforcement Learning
    Everett, Michael
    Lutjens, Bjorn
    How, Jonathan P.
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2022, 33 (09) : 4184 - 4198
  • [5] Universal Adversarial Perturbations Through the Lens of Deep Steganography: Towards a Fourier Perspective
    Zhang, Chaoning
    Benz, Philipp
    Karjauv, Adil
    Kweon, In So
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 3296 - 3304
  • [6] Deep Feature Learning to Quantitative Prediction of Software Defects
    Qiao, Lei
    Li, Guangjie
    Yu, Daohua
    Liu, Hui
    2021 IEEE 45TH ANNUAL COMPUTERS, SOFTWARE, AND APPLICATIONS CONFERENCE (COMPSAC 2021), 2021, : 1401 - 1402
  • [7] Towards Adversarial and Unintentional Collisions Detection Using Deep Learning
    Nguyen, Hai N.
    Vo-Huu, Tien
    Vo-Huu, Triet
    Noubir, Guevara
    PROCEEDINGS OF THE 2019 ACM WORKSHOP ON WIRELESS SECURITY AND MACHINE LEARNING (WISEML '19), 2019, : 22 - 24
  • [8] Robust adversarial uncertainty quantification for deep learning fine-tuning
    Ahmed, Usman
    Lin, Jerry Chun-Wei
    JOURNAL OF SUPERCOMPUTING, 2023, 79 (10): : 11355 - 11386
  • [9] Exploiting epistemic uncertainty of the deep learning models to generate adversarial samples
    Tuna, Omer Faruk
    Catak, Ferhat Ozgur
    Eskil, M. Taner
    MULTIMEDIA TOOLS AND APPLICATIONS, 2022, 81 (08) : 11479 - 11500
  • [10] Exploiting epistemic uncertainty of the deep learning models to generate adversarial samples
    Omer Faruk Tuna
    Ferhat Ozgur Catak
    M. Taner Eskil
    Multimedia Tools and Applications, 2022, 81 : 11479 - 11500