Towards interpreting vulnerability of object detection models via adversarial distillation

被引:2
|
作者
Zhang, Yaoyuan [1 ]
Tan, Yu-an [2 ]
Lu, Mingfeng [3 ]
Liu, Lu [2 ]
Wang, Dianxin [1 ]
Zhang, Quanxing [1 ]
Li, Yuanzhang [1 ]
机构
[1] Beijing Inst Technol, Sch Comp Sci & Technol, Beijing 100081, Peoples R China
[2] Beijing Inst Technol, Sch Cyberspace Sci & Technol, Beijing 100081, Peoples R China
[3] Beijing Inst Technol, Sch Informat & Elect, Beijing 100081, Peoples R China
关键词
Adversarial examples; Interpretability; Object detection; Deep learning;
D O I
10.1016/j.jisa.2022.103410
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Recent works have shown that deep learning models are highly vulnerable to adversarial examples, limiting the application of deep learning in security-critical systems. This paper aims to interpret the vulnerability of deep learning models to adversarial examples. We propose adversarial distillation to illustrate that adversarial examples are generalizable data features. Deep learning models are vulnerable to adversarial examples because models do not learn this data distribution. More specifically, we obtain adversarial features by introducing a generation and extraction mechanism. The generation mechanism generates adversarial examples, which mislead the source model trained on the original clean samples. The extraction term removes the original features and selects valid and generalizable adversarial features. Valuable adversarial features guide the model to learn the data distribution of adversarial examples and realize the model's generalization on the adversarial dataset. Extensive experimental evaluations have proved the excellent generalization performance of the adversarial distillation model. Compared with the normally trained model, the mAP has increased by 2.17% on their respective test sets, while the mAP on the opponent's test set is very low. The experimental results further prove that adversarial examples are also generalizable data features, which obey a different data distribution from the clean data.
引用
收藏
页数:7
相关论文
共 50 条
  • [1] Towards Interpreting Vulnerability of Object Detection Models via Adversarial Distillation
    Zhang, Yaoyuan
    Tan, Yu-an
    Lu, Mingfeng
    Liu, Lu
    Zhang, Quanxing
    Li, Yuanzhang
    Wang, Dianxin
    APPLIED CRYPTOGRAPHY AND NETWORK SECURITY WORKSHOPS, ACNS 2022, 2022, 13285 : 53 - 65
  • [2] One-stage object detection knowledge distillation via adversarial learning
    Na Dong
    Yongqiang Zhang
    Mingli Ding
    Shibiao Xu
    Yancheng Bai
    Applied Intelligence, 2022, 52 : 4582 - 4598
  • [3] One-stage object detection knowledge distillation via adversarial learning
    Dong, Na
    Zhang, Yongqiang
    Ding, Mingli
    Xu, Shibiao
    Bai, Yancheng
    APPLIED INTELLIGENCE, 2022, 52 (04) : 4582 - 4598
  • [4] Towards an Awareness of Time Series Anomaly Detection Models' Adversarial Vulnerability
    Tariq, Shahroz
    Le, Binh M.
    Woo, Simon S.
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2022, 2022, : 3534 - 3544
  • [5] Towards a physical-world adversarial patch for blinding object detection models
    Wang, Yajie
    Lv, Haoran
    Kuang, Xiaohui
    Zhao, Gang
    Tan, Yu-an
    Zhang, Quanxin
    Hu, Jingjing
    INFORMATION SCIENCES, 2021, 556 : 459 - 471
  • [6] Learning Efficient Object Detection Models with Knowledge Distillation
    Chen, Guobin
    Choi, Wongun
    Yu, Xiang
    Han, Tony
    Chandraker, Manmohan
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 30 (NIPS 2017), 2017, 30
  • [7] Robust and Accurate Object Detection via Adversarial Learning
    Chen, Xiangning
    Xie, Cihang
    Tan, Mingxing
    Zhang, Li
    Hsieh, Cho-Jui
    Gong, Boqing
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 16617 - 16626
  • [8] Boosting Accuracy and Robustness of Student Models via Adaptive Adversarial Distillation
    Huang, Bo
    Chen, Mingyang
    Wang, Yi
    Lu, Junda
    Cheng, Minhao
    Wang, Wei
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 24668 - 24677
  • [9] Learning Differentially Private Diffusion Models via Stochastic Adversarial Distillation
    Liu, Bochao
    Wang, Pengju
    Ge, Shiming
    COMPUTER VISION-ECCV 2024, PT VII, 2025, 15065 : 55 - 71
  • [10] KD-SCFNet: Towards more accurate and lightweight salient object detection via knowledge distillation
    Zhang, Jin
    Shi, Yanjiao
    Yang, Jinyu
    Guo, Qianqian
    NEUROCOMPUTING, 2024, 572