Towards interpreting vulnerability of object detection models via adversarial distillation

被引:2
|
作者
Zhang, Yaoyuan [1 ]
Tan, Yu-an [2 ]
Lu, Mingfeng [3 ]
Liu, Lu [2 ]
Wang, Dianxin [1 ]
Zhang, Quanxing [1 ]
Li, Yuanzhang [1 ]
机构
[1] Beijing Inst Technol, Sch Comp Sci & Technol, Beijing 100081, Peoples R China
[2] Beijing Inst Technol, Sch Cyberspace Sci & Technol, Beijing 100081, Peoples R China
[3] Beijing Inst Technol, Sch Informat & Elect, Beijing 100081, Peoples R China
关键词
Adversarial examples; Interpretability; Object detection; Deep learning;
D O I
10.1016/j.jisa.2022.103410
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Recent works have shown that deep learning models are highly vulnerable to adversarial examples, limiting the application of deep learning in security-critical systems. This paper aims to interpret the vulnerability of deep learning models to adversarial examples. We propose adversarial distillation to illustrate that adversarial examples are generalizable data features. Deep learning models are vulnerable to adversarial examples because models do not learn this data distribution. More specifically, we obtain adversarial features by introducing a generation and extraction mechanism. The generation mechanism generates adversarial examples, which mislead the source model trained on the original clean samples. The extraction term removes the original features and selects valid and generalizable adversarial features. Valuable adversarial features guide the model to learn the data distribution of adversarial examples and realize the model's generalization on the adversarial dataset. Extensive experimental evaluations have proved the excellent generalization performance of the adversarial distillation model. Compared with the normally trained model, the mAP has increased by 2.17% on their respective test sets, while the mAP on the opponent's test set is very low. The experimental results further prove that adversarial examples are also generalizable data features, which obey a different data distribution from the clean data.
引用
收藏
页数:7
相关论文
共 50 条
  • [21] Mitigating Adversarial Attacks in Object Detection through Conditional Diffusion Models
    Ye, Xudong
    Zhang, Qi
    Cui, Sanshuai
    Ying, Zuobin
    Sun, Jingzhang
    Du, Xia
    MATHEMATICS, 2024, 12 (19)
  • [22] Knowledge distillation vulnerability of DeiT through CNN adversarial attack
    Hong, Inpyo
    Choi, Chang
    NEURAL COMPUTING & APPLICATIONS, 2023, 37 (12): : 7721 - 7731
  • [23] Towards Robust Community Detection via Extreme Adversarial Attacks
    Chen, Chunchun
    Zhu, Wenjie
    Peng, Bo
    Lu, Huijuan
    2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 2231 - 2237
  • [24] Generative Object Detection: Erasing the Boundary via Adversarial Learning with Mask
    Jang, Heeoh
    Kim, Dongkyu
    Ahn, Wonhyuk
    Lee, Heung-Kyu
    2019 2ND IEEE INTERNATIONAL CONFERENCE ON INFORMATION COMMUNICATION AND SIGNAL PROCESSING (ICICSP), 2019, : 495 - 499
  • [25] Multimodal salient object detection via adversarial learning with collaborative generator
    Tu, Zhengzheng
    Yang, Wenfang
    Wang, Kunpeng
    Hussain, Amir
    Luo, Bin
    Li, Chenglong
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2023, 119
  • [26] ROBUST AND ACCURATE OBJECT DETECTION VIA SELF-KNOWLEDGE DISTILLATION
    Xu, Weipeng
    Chu, Pengzhi
    Xie, Renhao
    Xiao, Xiongziyan
    Huang, Hongcheng
    2022 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2022, : 91 - 95
  • [27] Accurate and Robust Object Detection via Selective Adversarial Learning With Constraints
    Chen, Jianpin
    Li, Heng
    Gao, Qi
    Liang, Junling
    Zhang, Ruipeng
    Yin, Liping
    Chai, Xinyu
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2024, 33 : 5593 - 5605
  • [28] Daedalus: Breaking Nonmaximum Suppression in Object Detection via Adversarial Examples
    Wang, Derui
    Li, Chaoran
    Wen, Sheng
    Han, Qing-Long
    Nepal, Surya
    Zhang, Xiangyu
    Xiang, Yang
    IEEE TRANSACTIONS ON CYBERNETICS, 2022, 52 (08) : 7427 - 7440
  • [29] Adversarial Attacks Assessment of Salient Object Detection via Symbolic Learning
    Olague, Gustavo
    Pineda, Roberto
    Ibarra-Vazquez, Gerardo
    Olague, Matthieu
    Martinez, Axel
    Bakshi, Sambit
    Vargas, Jonathan
    Reducindo, Isnardo
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTING, 2023, 11 (04) : 1018 - 1030
  • [30] Adversarial Attacks for Object Detection
    Xu, Bo
    Zhu, Jinlin
    Wang, Danwei
    PROCEEDINGS OF THE 39TH CHINESE CONTROL CONFERENCE, 2020, : 7281 - 7287