Query-efficient model extraction for text classification model in a hard label setting

被引:0
|
作者
Peng, Hao [1 ]
Guo, Shixin [1 ]
Zhao, Dandan [1 ]
Wu, Yiming [3 ]
Han, Jianming [1 ]
Wang, Zhe [1 ]
Ji, Shouling [2 ,4 ]
Zhong, Ming [1 ]
机构
[1] Zhejiang Normal Univ, Coll Comp Sci & Technol, Jinhua 321004, Peoples R China
[2] Zhejiang Univ, Coll Comp Sci & Technol, Hangzhou 310027, Zhejiang, Peoples R China
[3] Zhejiang Univ Technol, Inst Cyberspace Secur, Hangzhou 310027, Zhejiang, Peoples R China
[4] Georgia Inst Technol, Elect & Comp Engn, Atlanta, GA 30332 USA
基金
中国国家自然科学基金;
关键词
Model extraction; Language model stealing; Model privacy; Adversarial attack; Natural language processing; Performance Evaluation;
D O I
10.1016/j.jksuci.2023.02.019
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Designing a query-efficient model extraction strategy to steal models from cloud-based platforms with black-box constraints remains a challenge, especially for language models. In a more realistic setting, a lack of information about the target model's internal parameters, gradients, training data, or even confi-dence scores prevents attackers from easily copying the target model. Selecting informative and useful examples to train a substitute model is critical to query-efficient model stealing. We propose a novel model extraction framework that fine-tunes a pretrained model based on bidirectional encoder represen-tations from transformers (BERT) while improving query efficiency by utilizing an active learning selection strategy. The active learning strategy, incorporating semantic-based diversity sampling and class-balanced uncertainty sampling, builds an informative subset from the public unannotated dataset as the input for fine-tuning. We apply our method to extract deep classifiers with identical and mis-matched architectures as the substitute model under tight and moderate query budgets. Furthermore, we evaluate the transferability of adversarial examples constructed with the help of the models extracted by our method. The results show that our method achieves higher accuracy with fewer queries than existing baselines and the resulting models exhibit a high transferability success rate of adversarial examples. (c) 2023 The Author(s). Published by Elsevier B.V. on behalf of King Saud University. This is an open access article under the CC BY-NC-ND license (http://creativecommons.org/licenses/by-nc-nd/4.0/).
引用
收藏
页码:10 / 20
页数:11
相关论文
共 50 条
  • [31] Hierarchical Sequence-to-Sequence Model for Multi-Label Text Classification
    Yang, Zhenyu
    Liu, Guojing
    IEEE ACCESS, 2019, 7 : 153012 - 153020
  • [32] Unifying Model Explainability and Robustness for Joint Text Classification and Rationale Extraction
    Li, Dongfang
    Hu, Baotian
    Chen, Qingcai
    Xu, Tujie
    Tao, Jingcong
    Zhang, Yunan
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 10947 - 10955
  • [33] An Approach for Process Model Extraction by Multi-grained Text Classification
    Qian, Chen
    Wen, Lijie
    Kumar, Akhil
    Lin, Leilei
    Lin, Li
    Zong, Zan
    Li, Shu'ang
    Wang, Jianmin
    ADVANCED INFORMATION SYSTEMS ENGINEERING, CAISE 2020, 2020, 12127 : 268 - 282
  • [34] A model and a visual query language for structured text
    Baeza-Yates, R
    Navarro, G
    Vegas, J
    de la Fuente, P
    STRING PROCESSING AND INFORMATION RETRIEVAL - PROCEEDINGS: A SOUTH AMERICAN SYMPOSIUM, 1998, : 7 - 13
  • [35] MatchXML: An Efficient Text-Label Matching Framework for Extreme Multi-Label Text Classification
    Ye, Hui
    Sunderraman, Rajshekhar
    Ji, Shihao
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2024, 36 (09) : 4781 - 4793
  • [36] Label-Embedding Bi-directional Attentive Model for Multi-label Text Classification
    Liu, Naiyin
    Wang, Qianlong
    Ren, Jiangtao
    NEURAL PROCESSING LETTERS, 2021, 53 (01) : 375 - 389
  • [37] Multi-label Text Classification with Multi-variate Bernoulli Model and Label Dependent Representation
    Alfaro A, Rodrigo
    Allende O, Hector
    REVISTA SIGNOS, 2020, 53 (104): : 549 - 567
  • [38] Label-Embedding Bi-directional Attentive Model for Multi-label Text Classification
    Naiyin Liu
    Qianlong Wang
    Jiangtao Ren
    Neural Processing Letters, 2021, 53 : 375 - 389
  • [39] Language model based query classification
    Merkel, Andreas
    Klakow, Dietrich
    ADVANCES IN INFORMATION RETRIEVAL, 2007, 4425 : 720 - +
  • [40] Fusing Label Embedding into BERT: An Efficient Improvement for Text Classification
    Xiong, Yijin
    Feng, Yukun
    Wu, Hao
    Kamigaito, Hidetaka
    Okumura, Manabu
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL-IJCNLP 2021, 2021, : 1743 - 1750