Understanding adversarial training: Increasing local stability of supervised models through robust optimization

被引:143
|
作者
Shaham, Uri [1 ]
Yamada, Yutaro [2 ]
Negahban, Sahand [2 ]
机构
[1] Yale Univ, Ctr Outcome Res, 200 Church St, New Haven, CT 06510 USA
[2] Yale Univ, Dept Stat, 24 Hillhouse St, New Haven, CT 06511 USA
关键词
Adversarial examples; Robust optimization; Non-parametric supervised models; Deep learning; NETWORKS;
D O I
10.1016/j.neucom.2018.04.027
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We show that adversarial training of supervised learning models is in fact a robust optimization procedure. To do this, we establish a general framework for increasing local stability of supervised learning models using robust optimization. The framework is general and broadly applicable to differentiable non-parametric models, e.g., Artificial Neural Networks (ANNs). Using an alternating minimization-maximization procedure, the loss of the model is minimized with respect to perturbed examples that are generated at each parameter update, rather than with respect to the original training data. Our proposed framework generalizes adversarial training, as well as previous approaches for increasing local stability of ANNs. Experimental results reveal that our approach increases the robustness of the network to existing adversarial examples, while making it harder to generate new ones. Furthermore, our algorithm improves the accuracy of the networks also on the original test data. (C) 2018 Elsevier B.V. All rights reserved.
引用
收藏
页码:195 / 204
页数:10
相关论文
共 50 条
  • [1] Increasing the Robustness of Image Quality Assessment Models Through Adversarial Training
    Chistyakova, Anna
    Antsiferova, Anastasia
    Khrebtov, Maksim
    Lavrushkin, Sergey
    Arkhipenko, Konstantin
    Vatolin, Dmitriy
    Turdakov, Denis
    TECHNOLOGIES, 2024, 12 (11)
  • [2] Supervised contrastive learning for robust text adversarial training
    Li, Weidong
    Zhao, Bo
    An, Yang
    Shangguan, Chenhan
    Ji, Minzi
    Yuan, Anqi
    NEURAL COMPUTING & APPLICATIONS, 2023, 35 (10): : 7357 - 7368
  • [3] Supervised contrastive learning for robust text adversarial training
    Weidong Li
    Bo Zhao
    Yang An
    Chenhan Shangguan
    Minzi Ji
    Anqi Yuan
    Neural Computing and Applications, 2023, 35 : 7357 - 7368
  • [4] Towards Better Understanding of Training Certifiably Robust Models against Adversarial Examples
    Lee, Sungyoon
    Lee, Woojin
    Park, Jinseong
    Lee, Jaewook
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021,
  • [5] Understanding Robust Overfitting of Adversarial Training and Beyond
    Yu, Chaojian
    Han, Bo
    Shen, Li
    Yu, Jun
    Gong, Chen
    Gong, Mingming
    Liu, Tongliang
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [6] Understanding and Increasing Efficiency of Frank-Wolfe Adversarial Training
    Tsiligkaridis, Theodoros
    Roberts, Jay
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 50 - 59
  • [7] SafeAMC: Adversarial training for robust modulation classification models
    Maroto, Javier
    Bovet, Gerome
    Frossard, Pascal
    2022 30TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO 2022), 2022, : 1636 - 1640
  • [8] Self-Supervised Animation Synthesis Through Adversarial Training
    Yu, Cheng
    Wang, Wenmin
    Yan, Jianhao
    IEEE ACCESS, 2020, 8 : 128140 - 128151
  • [9] Bilateral Adversarial Training: Towards Fast Training of More Robust Models Against Adversarial Attacks
    Wang, Jianyu
    Zhang, Haichao
    2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 6628 - 6637
  • [10] Weakly Supervised Facial Action Unit Recognition through Adversarial Training
    Peng, Guozhu
    Wang, Shangfei
    2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 2188 - 2196