Defense against adversarial attacks via textual embeddings based on semantic associative field

被引:0
|
作者
Huang, Jiacheng [1 ]
Chen, Long [1 ,2 ]
机构
[1] Chongqing Univ Posts & Telecommun, Sch Comp Sci & Technol, Chongqing 400065, Peoples R China
[2] Chongqing Univ Posts & Telecommun, Sch Cyber Secur & Informat Law, Chongqing 400065, Peoples R China
来源
NEURAL COMPUTING & APPLICATIONS | 2024年 / 36卷 / 01期
关键词
Adversarial examples; Natural language processing; Semantic associative field; Word-level;
D O I
10.1007/s00521-023-08946-7
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks are known to be vulnerable to various types of adversarial attacks, especially word-level attacks, in the field of natural language processing. In recent years, various defense methods are proposed against word-level attacks; however, most of those defense methods only focus on synonyms substitution-based attacks, while word-level attacks are not based on synonym substitution. In this paper, we propose a textual adversarial defense method against word-level adversarial attacks via textual embedding based on the semantic associative field. More specifically, we analyze the reasons why humans can read and understand textual adversarial examples and observe two crucial points: (1) There must be a relation between the original word and the perturbed word or token. (2) Such a kind of relation enables humans to infer original words, while humans have the ability to associations. Motivated by this, we introduce the concept of semantic associative field and propose a new defense method by building a robust word embedding, that is, we calculate the word vector by exerting the related word vector to it with potential function and weighted embedding sampling for simulating the semantic influence between words in same semantic field. We conduct comprehensive experiments and demonstrate that the models using the proposed method can achieve higher accuracy than the baseline defense methods under various adversarial attacks or original testing sets. Moreover, the proposed method is more universal, while it is irrelevant to model structure and will not affect the efficiency of training.
引用
收藏
页码:289 / 301
页数:13
相关论文
共 50 条
  • [21] DIFFender: Diffusion-Based Adversarial Defense Against Patch Attacks
    Kang, Caixin
    Dong, Yinpeng
    Wang, Zhengyi
    Ruan, Shouwei
    Chen, Yubo
    Su, Hang
    Wei, Xingxing
    COMPUTER VISION - ECCV 2024, PT LII, 2025, 15110 : 130 - 147
  • [22] Defense against Adversarial Attacks in Image Recognition Based on Multilayer Filters
    Wang, Mingde
    Liu, Zhijing
    APPLIED SCIENCES-BASEL, 2024, 14 (18):
  • [23] Sparsity-based Defense against Adversarial Attacks on Linear Classifiers
    Marzi, Zhinus
    Gopalakrishnan, Soorya
    Madhow, Upamanyu
    Pedarsani, Ramtin
    2018 IEEE INTERNATIONAL SYMPOSIUM ON INFORMATION THEORY (ISIT), 2018, : 31 - 35
  • [24] TENSORSHIELD: Tensor-based Defense Against Adversarial Attacks on Images
    Entezari, Negin
    Papalexakis, Evangelos E.
    2022 IEEE MILITARY COMMUNICATIONS CONFERENCE (MILCOM), 2022,
  • [25] Defense against Adversarial Patch Attacks for Aerial Image Semantic Segmentation by Robust Feature Extraction
    Wang, Zhen
    Wang, Buhong
    Zhang, Chuanlei
    Liu, Yaohui
    REMOTE SENSING, 2023, 15 (06)
  • [26] BDDR: An Effective Defense Against Textual Backdoor Attacks
    Shao, Kun
    Yang, Junan
    Ai, Yang
    Liu, Hui
    Zhang, Yu
    Shao, Kun (1608053548@qq.com), 1600, Elsevier Ltd (110):
  • [27] BDDR: An Effective Defense Against Textual Backdoor Attacks
    Shao, Kun
    Yang, Junan
    Ai, Yang
    Liu, Hui
    Zhang, Yu
    COMPUTERS & SECURITY, 2021, 110
  • [28] The Best Defense is a Good Offense: Adversarial Augmentation against Adversarial Attacks
    Frosio, Iuri
    Kautz, Jan
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR, 2023, : 4067 - 4076
  • [29] Defense Against Adversarial Attacks Using Topology Aligning Adversarial Training
    Kuang, Huafeng
    Liu, Hong
    Lin, Xianming
    Ji, Rongrong
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2024, 19 : 3659 - 3673
  • [30] Adaptive Image Reconstruction for Defense Against Adversarial Attacks
    Yang, Yanan
    Shih, Frank Y.
    Chang, I-Cheng
    INTERNATIONAL JOURNAL OF PATTERN RECOGNITION AND ARTIFICIAL INTELLIGENCE, 2022, 36 (12)