Query-efficient model extraction for text classification model in a hard label setting

被引:0
|
作者
Peng, Hao [1 ]
Guo, Shixin [1 ]
Zhao, Dandan [1 ]
Wu, Yiming [3 ]
Han, Jianming [1 ]
Wang, Zhe [1 ]
Ji, Shouling [2 ,4 ]
Zhong, Ming [1 ]
机构
[1] Zhejiang Normal Univ, Coll Comp Sci & Technol, Jinhua 321004, Peoples R China
[2] Zhejiang Univ, Coll Comp Sci & Technol, Hangzhou 310027, Zhejiang, Peoples R China
[3] Zhejiang Univ Technol, Inst Cyberspace Secur, Hangzhou 310027, Zhejiang, Peoples R China
[4] Georgia Inst Technol, Elect & Comp Engn, Atlanta, GA 30332 USA
基金
中国国家自然科学基金;
关键词
Model extraction; Language model stealing; Model privacy; Adversarial attack; Natural language processing; Performance Evaluation;
D O I
10.1016/j.jksuci.2023.02.019
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Designing a query-efficient model extraction strategy to steal models from cloud-based platforms with black-box constraints remains a challenge, especially for language models. In a more realistic setting, a lack of information about the target model's internal parameters, gradients, training data, or even confi-dence scores prevents attackers from easily copying the target model. Selecting informative and useful examples to train a substitute model is critical to query-efficient model stealing. We propose a novel model extraction framework that fine-tunes a pretrained model based on bidirectional encoder represen-tations from transformers (BERT) while improving query efficiency by utilizing an active learning selection strategy. The active learning strategy, incorporating semantic-based diversity sampling and class-balanced uncertainty sampling, builds an informative subset from the public unannotated dataset as the input for fine-tuning. We apply our method to extract deep classifiers with identical and mis-matched architectures as the substitute model under tight and moderate query budgets. Furthermore, we evaluate the transferability of adversarial examples constructed with the help of the models extracted by our method. The results show that our method achieves higher accuracy with fewer queries than existing baselines and the resulting models exhibit a high transferability success rate of adversarial examples. (c) 2023 The Author(s). Published by Elsevier B.V. on behalf of King Saud University. This is an open access article under the CC BY-NC-ND license (http://creativecommons.org/licenses/by-nc-nd/4.0/).
引用
收藏
页码:10 / 20
页数:11
相关论文
共 50 条
  • [21] Multi-Label Text Classification model integrating Label Attention and Historical Attention
    Sun, Guoying
    Cheng, Yanan
    Dong, Fangzhou
    Wang, Luhua
    Zhao, Dong
    Zhang, Zhaoxin
    Tong, Xiaojun
    KNOWLEDGE-BASED SYSTEMS, 2024, 296
  • [22] Label-text bi-attention capsule networks model for multi-label text classification
    Wang, Gang
    Du, Yajun
    Jiang, Yurui
    Liu, Jia
    Li, Xianyong
    Chen, Xiaoliang
    Gao, Hongmei
    Xie, Chunzhi
    Lee, Yan-li
    NEUROCOMPUTING, 2024, 588
  • [23] Multi-label text classification model based on semantic embedding
    Yan Danfeng
    Ke Nan
    Gu Chao
    Cui Jianfei
    Ding Yiqi
    TheJournalofChinaUniversitiesofPostsandTelecommunications, 2019, 26 (01) : 95 - 104
  • [24] Hierarchy-Aware and Label Balanced Model for Hierarchical Text Classification
    Zhang, Jun
    Li, Yubin
    Shen, Fanfan
    Xia, Chenxi
    Tan, Hai
    He, Yanxiang
    KNOWLEDGE-BASED SYSTEMS, 2024, 300
  • [25] An Interactive Fusion Model for Hierarchical Multi-label Text Classification
    Zhao, Xiuhao
    Li, Zhao
    Zhang, Xianming
    Wang, Jibin
    Chen, Tong
    Ju, Zhengyu
    Wang, Canjun
    Zhang, Chao
    Zhan, Yiming
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, NLPCC 2022, PT II, 2022, 13552 : 168 - 178
  • [26] Multi-Label Text Classification Model Integrating GAT and Head-Tail Label
    Yang, Chunxia
    Huang, Yukun
    Yan, Han
    Wu, Yalei
    Computer Engineering and Applications, 2024, 60 (15) : 150 - 160
  • [27] A Multi-task Text Classification Model Based on Label Embedding Learning
    Xu, Yuemei
    Fan, Zuwei
    Cao, Han
    CYBER SECURITY, CNCERT 2021, 2022, 1506 : 211 - 225
  • [28] Multi-label Text Classification Model Combining BiLSTM and Hypergraph Attention
    Wang, Xing
    Hu, HuiTing
    Zhu, GuoHua
    2024 4TH INTERNATIONAL CONFERENCE ON COMPUTER COMMUNICATION AND ARTIFICIAL INTELLIGENCE, CCAI 2024, 2024, : 344 - 349
  • [29] Cognitive structure learning model for hierarchical multi-label text classification
    Wang, Boyan
    Hu, Xuegang
    Li, Peipei
    Yu, Philip S.
    KNOWLEDGE-BASED SYSTEMS, 2021, 218