Understanding adversarial training: Increasing local stability of supervised models through robust optimization

被引:143
|
作者
Shaham, Uri [1 ]
Yamada, Yutaro [2 ]
Negahban, Sahand [2 ]
机构
[1] Yale Univ, Ctr Outcome Res, 200 Church St, New Haven, CT 06510 USA
[2] Yale Univ, Dept Stat, 24 Hillhouse St, New Haven, CT 06511 USA
关键词
Adversarial examples; Robust optimization; Non-parametric supervised models; Deep learning; NETWORKS;
D O I
10.1016/j.neucom.2018.04.027
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We show that adversarial training of supervised learning models is in fact a robust optimization procedure. To do this, we establish a general framework for increasing local stability of supervised learning models using robust optimization. The framework is general and broadly applicable to differentiable non-parametric models, e.g., Artificial Neural Networks (ANNs). Using an alternating minimization-maximization procedure, the loss of the model is minimized with respect to perturbed examples that are generated at each parameter update, rather than with respect to the original training data. Our proposed framework generalizes adversarial training, as well as previous approaches for increasing local stability of ANNs. Experimental results reveal that our approach increases the robustness of the network to existing adversarial examples, while making it harder to generate new ones. Furthermore, our algorithm improves the accuracy of the networks also on the original test data. (C) 2018 Elsevier B.V. All rights reserved.
引用
收藏
页码:195 / 204
页数:10
相关论文
共 50 条
  • [21] Efficient adversarial training with multi-fidelity optimization for robust neural network
    Wang, Zhaoxin
    Wang, Handing
    Tian, Cong
    Jin, Yaochu
    NEUROCOMPUTING, 2024, 585
  • [22] Training Robust Deep Collaborative Filtering Models via Adversarial Noise Propagation
    Chen, Hai
    Qian, Fulan
    Liu, Chang
    Zhang, Yanping
    Su, Hang
    Zhao, Shu
    ACM TRANSACTIONS ON INFORMATION SYSTEMS, 2024, 42 (01)
  • [23] Robust shortcut and disordered robustness: Improving adversarial training through adaptive smoothing
    Li, Lin
    Spratling, Michael
    PATTERN RECOGNITION, 2025, 163
  • [24] Improving the robustness and accuracy of biomedical language models through adversarial training
    Moradi, Milad
    Samwald, Matthias
    JOURNAL OF BIOMEDICAL INFORMATICS, 2022, 132
  • [25] Defending Against Local Adversarial Attacks through Empirical Gradient Optimization
    Sun, Boyang
    Ma, Xiaoxuan
    Wang, Hengyou
    TEHNICKI VJESNIK-TECHNICAL GAZETTE, 2023, 30 (06): : 1888 - 1898
  • [26] SEMI-SUPERVISED TRAINING USING ADVERSARIAL MULTI-TASK LEARNING FOR SPOKEN LANGUAGE UNDERSTANDING
    Lan, Ouyu
    Zhu, Su
    Yu, Kai
    2018 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2018, : 6049 - 6053
  • [27] Robust deep neural network surrogate models with uncertainty quantification via adversarial training
    Zhang, Lixiang
    Li, Jia
    STATISTICAL ANALYSIS AND DATA MINING, 2023, 16 (03) : 295 - 304
  • [28] P-NOC: Adversarial training of CAM generating networks for robust weakly supervised semantic segmentation priors
    David, Lucas
    Pedrini, Helio
    Dias, Zanoni
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2024, 102
  • [29] WEAK-SUPERVISED DYSARTHRIA-INVARIANT FEATURES FOR SPOKEN LANGUAGE UNDERSTANDING USING AN FHVAE AND ADVERSARIAL TRAINING
    Qi, Jinzi
    Van hamme, Hugo
    2022 IEEE SPOKEN LANGUAGE TECHNOLOGY WORKSHOP, SLT, 2022, : 375 - 381
  • [30] A continual learning framework to train robust image recognition models by adversarial training and knowledge distillation
    Chou, Ting-Chun
    Kuo, Yu-Cheng
    Huang, Jhih-Yuan
    Lee, Wei-Po
    CONNECTION SCIENCE, 2024, 36 (01)