Query-efficient model extraction for text classification model in a hard label setting

被引:0
|
作者
Peng, Hao [1 ]
Guo, Shixin [1 ]
Zhao, Dandan [1 ]
Wu, Yiming [3 ]
Han, Jianming [1 ]
Wang, Zhe [1 ]
Ji, Shouling [2 ,4 ]
Zhong, Ming [1 ]
机构
[1] Zhejiang Normal Univ, Coll Comp Sci & Technol, Jinhua 321004, Peoples R China
[2] Zhejiang Univ, Coll Comp Sci & Technol, Hangzhou 310027, Zhejiang, Peoples R China
[3] Zhejiang Univ Technol, Inst Cyberspace Secur, Hangzhou 310027, Zhejiang, Peoples R China
[4] Georgia Inst Technol, Elect & Comp Engn, Atlanta, GA 30332 USA
基金
中国国家自然科学基金;
关键词
Model extraction; Language model stealing; Model privacy; Adversarial attack; Natural language processing; Performance Evaluation;
D O I
10.1016/j.jksuci.2023.02.019
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Designing a query-efficient model extraction strategy to steal models from cloud-based platforms with black-box constraints remains a challenge, especially for language models. In a more realistic setting, a lack of information about the target model's internal parameters, gradients, training data, or even confi-dence scores prevents attackers from easily copying the target model. Selecting informative and useful examples to train a substitute model is critical to query-efficient model stealing. We propose a novel model extraction framework that fine-tunes a pretrained model based on bidirectional encoder represen-tations from transformers (BERT) while improving query efficiency by utilizing an active learning selection strategy. The active learning strategy, incorporating semantic-based diversity sampling and class-balanced uncertainty sampling, builds an informative subset from the public unannotated dataset as the input for fine-tuning. We apply our method to extract deep classifiers with identical and mis-matched architectures as the substitute model under tight and moderate query budgets. Furthermore, we evaluate the transferability of adversarial examples constructed with the help of the models extracted by our method. The results show that our method achieves higher accuracy with fewer queries than existing baselines and the resulting models exhibit a high transferability success rate of adversarial examples. (c) 2023 The Author(s). Published by Elsevier B.V. on behalf of King Saud University. This is an open access article under the CC BY-NC-ND license (http://creativecommons.org/licenses/by-nc-nd/4.0/).
引用
收藏
页码:10 / 20
页数:11
相关论文
共 50 条
  • [41] Efficient text-based evolution algorithm to hard-label adversarial attacks on text
    Peng, Hao
    Wang, Zhe
    Zhao, Dandan
    Wu, Yiming
    Han, Jianming
    Guo, Shixin
    Ji, Shouling
    Zhong, Ming
    JOURNAL OF KING SAUD UNIVERSITY-COMPUTER AND INFORMATION SCIENCES, 2023, 35 (05)
  • [42] A Hierarchical Feature Extraction Model for Multi-Label Mechanical Patent Classification
    Hu, Jie
    Li, Shaobo
    Hu, Jianjun
    Yang, Guanci
    SUSTAINABILITY, 2018, 10 (01)
  • [43] Spatial relation extraction in natural language with multi-label classification model
    Zhou, J. (zhoujs@njnu.edu.cn), 1600, ICIC Express Letters Office, Tokai University, Kumamoto Campus, 9-1-1, Toroku, Kumamoto, 862-8652, Japan (03):
  • [44] Multigranularity Label Prediction Model for Automatic International Classification of Diseases Coding in Clinical Text
    Yu, Ying
    Qiu, Tian
    Duan, Junwen
    Wang, Jianxin
    JOURNAL OF COMPUTATIONAL BIOLOGY, 2023, 30 (08) : 900 - 911
  • [45] Labeled Bilingual Topic Model for Cross-Lingual Text Classification and Label Recommendation
    Tian, Ming-Jie
    Huang, Zheng-Hao
    Cui, Rong-Yi
    2018 5TH INTERNATIONAL CONFERENCE ON INFORMATION SCIENCE AND CONTROL ENGINEERING (ICISCE 2018), 2018, : 285 - 289
  • [46] Multi-Label Text Classification Model Combining CNN-SAM and GAT
    Yang, Chunxia
    Ma, Wenwen
    Chen, Qigang
    Gui, Qiang
    Computer Engineering and Applications, 2024, 59 (05) : 106 - 114
  • [47] A Multi-Task Text Classification Model Based on Label Embedding of Attention Mechanism
    Yuemei X.
    Zuwei F.
    Han C.
    Data Analysis and Knowledge Discovery, 2022, 6 (2-3): : 105 - 116
  • [48] Multi-Label Text Classification Based on Multidimensional Information Extraction
    Fan, Bin
    Zhu, Feng
    Ning, D. J.
    Lu, Junzhe
    20TH INT CONF ON UBIQUITOUS COMP AND COMMUNICAT (IUCC) / 20TH INT CONF ON COMP AND INFORMATION TECHNOLOGY (CIT) / 4TH INT CONF ON DATA SCIENCE AND COMPUTATIONAL INTELLIGENCE (DSCI) / 11TH INT CONF ON SMART COMPUTING, NETWORKING, AND SERV (SMARTCNS), 2021, : 474 - 483
  • [49] Incorporating keyword extraction and attention for multi-label text classification
    Zhao, Hua
    Li, Xiaoqian
    Wang, Fengling
    Zeng, Qingtian
    Diao, Xiuli
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2023, 45 (02) : 2083 - 2093
  • [50] Document-base extraction for single-label Text Classification
    Wang, Yanbo J.
    Sanderson, Robert
    Coenen, Frans
    Leng, Paul
    DATA WAREHOUSING AND KNOWLEDGE DISCOVERY, PROCEEDINGS, 2008, 5182 : 357 - +