Towards interpreting vulnerability of object detection models via adversarial distillation

被引:2
|
作者
Zhang, Yaoyuan [1 ]
Tan, Yu-an [2 ]
Lu, Mingfeng [3 ]
Liu, Lu [2 ]
Wang, Dianxin [1 ]
Zhang, Quanxing [1 ]
Li, Yuanzhang [1 ]
机构
[1] Beijing Inst Technol, Sch Comp Sci & Technol, Beijing 100081, Peoples R China
[2] Beijing Inst Technol, Sch Cyberspace Sci & Technol, Beijing 100081, Peoples R China
[3] Beijing Inst Technol, Sch Informat & Elect, Beijing 100081, Peoples R China
关键词
Adversarial examples; Interpretability; Object detection; Deep learning;
D O I
10.1016/j.jisa.2022.103410
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Recent works have shown that deep learning models are highly vulnerable to adversarial examples, limiting the application of deep learning in security-critical systems. This paper aims to interpret the vulnerability of deep learning models to adversarial examples. We propose adversarial distillation to illustrate that adversarial examples are generalizable data features. Deep learning models are vulnerable to adversarial examples because models do not learn this data distribution. More specifically, we obtain adversarial features by introducing a generation and extraction mechanism. The generation mechanism generates adversarial examples, which mislead the source model trained on the original clean samples. The extraction term removes the original features and selects valid and generalizable adversarial features. Valuable adversarial features guide the model to learn the data distribution of adversarial examples and realize the model's generalization on the adversarial dataset. Extensive experimental evaluations have proved the excellent generalization performance of the adversarial distillation model. Compared with the normally trained model, the mAP has increased by 2.17% on their respective test sets, while the mAP on the opponent's test set is very low. The experimental results further prove that adversarial examples are also generalizable data features, which obey a different data distribution from the clean data.
引用
收藏
页数:7
相关论文
共 50 条
  • [31] Localization Distillation for Object Detection
    Zheng, Zhaohui
    Ye, Rongguang
    Hou, Qibin
    Ren, Dongwei
    Wang, Ping
    Zuo, Wangmeng
    Cheng, Ming-Ming
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (08) : 10070 - 10083
  • [32] Reinforcement-learning-based Adversarial Attacks Against Vulnerability Detection Models
    Chen, Si-Ran
    Wu, Jing-Zheng
    Ling, Xiang
    Luo, Tian-Yue
    Liu, Jia-Yu
    Wu, Yan-Jun
    Ruan Jian Xue Bao/Journal of Software, 2024, 35 (08): : 3647 - 3667
  • [33] Adversarial Attack on Object Detection via Object Feature-Wise Attention and Perturbation Extraction
    Xue, Wei
    Xia, Xiaoyan
    Wan, Pengcheng
    Zhong, Ping
    Zheng, Xiao
    TSINGHUA SCIENCE AND TECHNOLOGY, 2025, 30 (03): : 1174 - 1189
  • [34] Towards cost-efficient vulnerability detection with cross-modal adversarial reprogramming
    Tian, Zhenzhou
    Qiu, Rui
    Teng, Yudong
    Sun, Jiaze
    Chen, Yanping
    Chen, Lingwei
    JOURNAL OF SYSTEMS AND SOFTWARE, 2025, 223
  • [35] Towards Interpreting and Utilizing Symmetry Property in Adversarial Examples
    Mei, Shibin
    Zhao, Chenglong
    Ni, Bingbing
    Yuan, Shengchao
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 8, 2023, : 9126 - 9133
  • [36] Lightning fast video anomaly detection via multi-scale adversarial distillation
    Croitoru, Florinel-Alin
    Ristea, Nicolae-Catalin
    Dascalescu, Dana
    Ionescu, Radu Tudor
    Khan, Fahad Shahbaz
    Shah, Mubarak
    COMPUTER VISION AND IMAGE UNDERSTANDING, 2024, 247
  • [37] Towards Robust Android Malware Detection Models using Adversarial Learning
    Rathore, Hemant
    2021 IEEE INTERNATIONAL CONFERENCE ON PERVASIVE COMPUTING AND COMMUNICATIONS WORKSHOPS AND OTHER AFFILIATED EVENTS (PERCOM WORKSHOPS), 2021, : 424 - 425
  • [38] EvoAttack: An Evolutionary Search-Based Adversarial Attack for Object Detection Models
    Chan, Kenneth
    Cheng, Betty H. C.
    SEARCH-BASED SOFTWARE ENGINEERING, SSBSE 2022, 2022, 13711 : 83 - 97
  • [39] Unsupervised Anomaly Detection for Improving Adversarial Robustness of 3D Object Detection Models
    Cai, Mumuxin
    Wang, Xupeng
    Sohel, Ferdous
    Lei, Hang
    ELECTRONICS, 2025, 14 (02):
  • [40] Joint Distribution Alignment via Adversarial Learning for Domain Adaptive Object Detection
    Zhang, Bo
    Chen, Tao
    Wang, Bin
    Li, Ruoyao
    arXiv, 2021,