Explanation-based data-free model extraction attacks

被引:3
|
作者
Yan, Anli [1 ,2 ]
Hou, Ruitao [2 ]
Yan, Hongyang [2 ]
Liu, Xiaozhang [3 ]
机构
[1] Hainan Univ, Sch Cyberspace Secur, Sch Cryptol, Haikou, Peoples R China
[2] Guangzhou Univ, Inst Artificial Intelligence & Blockchain, Guangzhou, Peoples R China
[3] Hainan Univ, Sch Comp Sci & Technol, Haikou, Peoples R China
基金
中国国家自然科学基金;
关键词
Deep neural network; Model explanation; Black-box; Model extraction attack; FRAMEWORK; EFFICIENT;
D O I
10.1007/s11280-023-01150-6
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep learning (DL) has dramatically pushed the previous limits of various tasks, ranging from computer vision to natural language processing. Despite its success, the lack of model explanations thwarts the usage of these techniques in life-critical domains, e.g., medical diagnosis and self-driving systems. To date, the core technology to solve the explainable issue is explainable artificial intelligence (XAI). XAI methods have been developed to produce human-understandable explanations by leveraging intermediate results of the DL models, e.g., gradients and model parameters. While the effectiveness of XAI methods has been demonstrated in benign environments, their privacy against model extraction attacks (i.e., attacks at the model confidentially) requires to be studied. To this end, this paper proposes DMEAE, a data-free model extraction attack using explanation-guided, to explore XAI privacy threats. Compared with previous works, DMEAE does not require collecting any data and utilizes model explanation loss. Specifically, DMEAE creates synthetic data using a generative model with model explanation loss items. Extensive evaluations verify the effectiveness and efficiency of the proposed attack strategy on SVHN and CIFAR-10 datasets. We hope that our research can provide insights for the development of practical tools to trade off the relationship between privacy and model explanations.
引用
收藏
页码:3081 / 3092
页数:12
相关论文
共 50 条
  • [1] Explanation-based data-free model extraction attacks
    Anli Yan
    Ruitao Hou
    Hongyang Yan
    Xiaozhang Liu
    World Wide Web, 2023, 26 : 3081 - 3092
  • [2] Data-Free Model Extraction Attacks in the Context of Object Detection
    Shah, Harshit
    Aravindhan, G.
    Kulkarni, Pavan
    Govindarajulu, Yuvaraj
    Parmar, Manojkumar
    COMPUTER VISION SYSTEMS, ICVS 2023, 2023, 14253 : 391 - 402
  • [3] Data-Free Model Extraction
    Truong, Jean-Baptiste
    Maini, Pratyush
    Walls, Robert J.
    Papernot, Nicolas
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 4769 - 4778
  • [4] Black-Box Attacks on Sequential Recommenders via Data-Free Model Extraction
    Yue, Zhenrui
    He, Zhankui
    Zeng, Huimin
    McAuley, Julian
    15TH ACM CONFERENCE ON RECOMMENDER SYSTEMS (RECSYS 2021), 2021, : 44 - 54
  • [5] Training A Secure Model Against Data-Free Model Extraction
    Wang, Zhenyi
    Li Shen
    Guo, Junfeng
    Duan, Tiehang
    Luan, Siyu
    Liu, Tongliang
    Gao, Mingchen
    COMPUTER VISION - ECCV 2024, PT LXXIX, 2025, 15137 : 323 - 340
  • [6] Unveiling the Secrets without Data: Can Graph Neural Networks Be Exploited through Data-Free Model Extraction Attacks?
    Zhuang, Yuanxin
    Shi, Chuan
    Zhang, Mengmei
    Chen, Jinghui
    Lyu, Lingjuan
    Zhou, Pan
    Sun, Lichao
    PROCEEDINGS OF THE 33RD USENIX SECURITY SYMPOSIUM, SECURITY 2024, 2024, : 5251 - 5268
  • [7] QUDA: Query-Limited Data-Free Model Extraction
    Lin, Zijun
    Xu, Ke
    Fang, Chengfang
    Zheng, Huadi
    Ahmed, Jaheezuddin Aneez
    Shi, Jie
    PROCEEDINGS OF THE 2023 ACM ASIA CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, ASIA CCS 2023, 2023, : 913 - 924
  • [8] DisGUIDE: Disagreement-Guided Data-Free Model Extraction
    Rosenthal, Jonathan
    Enouen, Eric
    Hung Viet Pham
    Tan, Lin
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 8, 2023, : 9614 - 9622
  • [9] Explanation-based Data Augmentation for Image Classification
    Wickramanayake, Sandareka
    Lee, Mong Li
    Hsu, Wynne
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [10] Explanation-Based Auditing
    Fabbri, Daniel
    LeFevre, Kristen
    PROCEEDINGS OF THE VLDB ENDOWMENT, 2011, 5 (01): : 1 - 12