Robust scientific text classification using prompt tuning based on data augmentation with L2 regularization

被引:7
|
作者
Shi, Shijun [1 ]
Hu, Kai [1 ]
Xie, Jie [2 ,3 ]
Guo, Ya [1 ]
Wu, Huayi [4 ]
机构
[1] Jiangnan Univ, Key Lab Adv Proc Control Light Ind, Minist Educ, Wuxi 214122, Peoples R China
[2] Nanjing Normal Univ, Sch Comp & Elect Informat, Nanjing 210023, Peoples R China
[3] Nanjing Normal Univ, Sch Artificial Intelligence, Nanjing 210023, Peoples R China
[4] Wuhan Univ, State Key Lab Informat Engn Surveying Mapping & Re, Wuhan 430079, Peoples R China
基金
中国国家自然科学基金;
关键词
Scientific text classification; Pre-training model; Prompt tuning; Data augmentation; Pairwise training; L2; regularization;
D O I
10.1016/j.ipm.2023.103531
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Recently, the prompt tuning technique, which incorporates prompts into the input of the pretraining language model (like BERT, GPT), has shown promise in improving the performance of language models when facing limited annotated data. However, the equivalence of template semantics in learning is not related to the effect of prompts and the prompt tuning often exhibits unstable performance, which is more severe in the domain of the scientific domain. To address this challenge, we propose to enhance prompt tuning using data augmentation with L2 regularization. Namely, pairing-wise training for the pair of the original and transformed data is performed. Our experiments on two scientific text datasets (ACL-ARC and SciCite) demonstrate that our proposed method significantly improves both accuracy and robustness. By using 1000 samples out of 1688 in the ACL-ARC training set, our method achieved an F1 score 3.33% higher than the same model trained on all 1688-sample data. In the SciCite dataset, our method surpassed the same model with labeled data reduced by over 93%. Our method is also proved to have high robustness, reaching F1 scores from 1% to 8% higher than those models without our method after the Probability Weighted Word Saliency attack.
引用
收藏
页数:19
相关论文
共 50 条
  • [31] Chinese Text Sentiment Analysis Using LSTM Network Based on L2 and Nadam
    Wang, Jian
    Cao, Zewen
    2017 17TH IEEE INTERNATIONAL CONFERENCE ON COMMUNICATION TECHNOLOGY (ICCT 2017), 2017, : 1891 - 1895
  • [32] An Improved Variable Kernel Density Estimator Based on L2 Regularization
    Jin, Yi
    He, Yulin
    Huang, Defa
    MATHEMATICS, 2021, 9 (16)
  • [33] Blind Image Restoration Based on l1 - l2 Blur Regularization
    Xiao, Su
    ENGINEERING LETTERS, 2020, 28 (01) : 148 - 154
  • [34] A New Discriminative Sparse Representation Method for Robust Face Recognition via l2 Regularization
    Xu, Yong
    Zhong, Zuofeng
    Yang, Jian
    You, Jane
    Zhang, David
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2017, 28 (10) : 2233 - 2242
  • [35] Knowledge and separating soft verbalizer based prompt-tuning for multi-label short text classification
    Chen, Zhanwang
    Li, Peipei
    Hu, Xuegang
    APPLIED INTELLIGENCE, 2024, 54 (17-18) : 8020 - 8040
  • [36] INTERMIX: AN INTERFERENCE-BASED DATA AUGMENTATION AND REGULARIZATION TECHNIQUE FOR AUTOMATIC DEEP SOUND CLASSIFICATION
    Sawhney, Ramit
    Neerkaje, Atula Tejaswi
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 3443 - 3447
  • [37] REAL-TIME VISUAL TRACKING USING L2 NORM REGULARIZATION BASED COLLABORATIVE REPRESENTATION
    Lu, Xiusheng
    Yao, Hongxun
    Sun, Xin
    Jiang, Xuesong
    2013 20TH IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP 2013), 2013, : 3934 - 3938
  • [38] ALP: Data Augmentation Using Lexicalized PCFGs for Few-Shot Text Classification
    Kim, Hazel H.
    Woo, Daecheol
    Oh, Seong Joon
    Cha, Jeong-Won
    Han, Yo-Sub
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 10894 - 10902
  • [39] Research on Data Augmentation Techniques for Text Classification Based on Antonym Replacement and Random Swapping
    Wang, Shaoyan
    Xiang, Yu
    PROCEEDINGS OF INTERNATIONAL CONFERENCE ON MODELING, NATURAL LANGUAGE PROCESSING AND MACHINE LEARNING, CMNM 2024, 2024, : 103 - 108
  • [40] Image Reconstruction in Ultrasonic Transmission Tomography Using L1/L2 Regularization
    Li, Aoyu
    Liang, Guanghui
    Dong, Feng
    2024 IEEE INTERNATIONAL INSTRUMENTATION AND MEASUREMENT TECHNOLOGY CONFERENCE, I2MTC 2024, 2024,