A Comprehensive Study on Robustness of Image Classification Models: Benchmarking and Rethinking

被引:6
|
作者
Liu, Chang [2 ]
Dong, Yinpeng [1 ,5 ]
Xiang, Wenzhao [3 ,7 ]
Yang, Xiao [1 ]
Su, Hang [1 ,6 ]
Zhu, Jun [1 ,5 ]
Chen, Yuefeng [4 ]
He, Yuan [4 ]
Xue, Hui [4 ]
Zheng, Shibao [2 ]
机构
[1] Chinese Acad Sci, Key Lab Intelligent Informat Proc, Inst Comp Technol ICT, Beijing 100190, Peoples R China
[2] Shanghai Jiao Tong Univ, Inst Image Commun & Networks Engn, Dept Elect Engn EE, Shanghai 200240, Peoples R China
[3] Chinese Acad Sci, Inst Comp Technol, Key Lab Intelligent Informat Proc, Beijing 100190, Peoples R China
[4] Alibaba Grp, Hangzhou 310023, Zhejiang, Peoples R China
[5] RealAI, Beijing 100085, Peoples R China
[6] Zhongguancun Lab, Beijing 100080, Peoples R China
[7] Peng Cheng Lab, Shenzhen 518000, Peoples R China
基金
中国国家自然科学基金;
关键词
Robustness benchmark; Distribution shift; Pre-training; Adversarial training; Image classification;
D O I
10.1007/s11263-024-02196-3
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The robustness of deep neural networks is frequently compromised when faced with adversarial examples, common corruptions, and distribution shifts, posing a significant research challenge in the advancement of deep learning. Although new deep learning methods and robustness improvement techniques have been constantly proposed, the robustness evaluations of existing methods are often inadequate due to their rapid development, diverse noise patterns, and simple evaluation metrics. Without thorough robustness evaluations, it is hard to understand the advances in the field and identify the effective methods. In this paper, we establish a comprehensive robustness benchmark called ARES-Bench on the image classification task. In our benchmark, we evaluate the robustness of 61 typical deep learning models on ImageNet with diverse architectures (e.g., CNNs, Transformers) and learning algorithms (e.g., normal supervised training, pre-training, adversarial training) under numerous adversarial attacks and out-of-distribution (OOD) datasets. Using robustness curves as the major evaluation criteria, we conduct large-scale experiments and draw several important findings, including: (1) there exists an intrinsic trade-off between the adversarial and natural robustness of specific noise types for the same model architecture; (2) adversarial training effectively improves adversarial robustness, especially when performed on Transformer architectures; (3) pre-training significantly enhances natural robustness by leveraging larger training datasets, incorporating multi-modal data, or employing self-supervised learning techniques. Based on ARES-Bench, we further analyze the training tricks in large-scale adversarial training on ImageNet. Through tailored training settings, we achieve a new state-of-the-art in adversarial robustness. We have made the benchmarking results and code platform publicly available.
引用
收藏
页码:567 / 589
页数:23
相关论文
共 50 条
  • [1] Is Robustness the Cost of Accuracy? - A Comprehensive Study on the Robustness of 18 Deep Image Classification Models
    Su, Dong
    Zhang, Huan
    Chen, Hongge
    Yi, Jinfeng
    Chen, Pin-Yu
    Gao, Yupeng
    COMPUTER VISION - ECCV 2018, PT XII, 2018, 11216 : 644 - 661
  • [2] Benchmarking Adversarial Robustness on Image Classification
    Dong, Yinpeng
    Fu, Qi-An
    Yang, Xiao
    Pang, Tianyu
    Su, Hang
    Xiao, Zihao
    Zhu, Jun
    2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, : 318 - 328
  • [3] Robustness of models addressing Information Disorder: A comprehensive review and benchmarking study
    Fenza, Giuseppe
    Loia, Vincenzo
    Stanzione, Claudio
    Di Gisi, Maria
    NEUROCOMPUTING, 2024, 596
  • [4] A Comprehensive Study on the Robustness of Deep Learning-Based Image Classification and Object Detection in Remote Sensing: Surveying and Benchmarking
    Mei, Shaohui
    Lian, Jiawei
    Wang, Xiaofei
    Su, Yuru
    Ma, Mingyang
    Chau, Lap-Pui
    JOURNAL OF REMOTE SENSING, 2024, 4
  • [5] Impact of Attention on Adversarial Robustness of Image Classification Models
    Agrawal, Prachi
    Punn, Narinder Singh
    Sonbhadra, Sanjay Kumar
    Agarwal, Sonali
    2021 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2021, : 3013 - 3019
  • [6] WAVES: Benchmarking the Robustness of Image Watermarks
    An, Bang
    Ding, Mucong
    Rabbani, Tahseen
    Agrawal, Aakriti
    Xu, Yuancheng
    Deng, Chenghao
    Zhu, Sicheng
    Mohamed, Abdirisak
    Wen, Yuxin
    Goldstein, Tom
    Huang, Furong
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, 2024, 235
  • [7] A Comprehensive Study on Large-Scale Graph Training: Benchmarking and Rethinking
    Duan, Keyu
    Liu, Zirui
    Wang, Peihao
    Zheng, Wenqing
    Zhou, Kaixiong
    Chen, Tianlong
    Hu, Xia
    Wang, Zhangyang
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [8] Benchmarking the Robustness of Instance Segmentation Models
    Dalva, Yusuf
    Pehlivan, Hamza
    Altindis, Said Fahri
    Dundar, Aysegul
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (12) : 17021 - 17035
  • [9] RobustMQ: benchmarking robustness of quantized models
    Yisong Xiao
    Aishan Liu
    Tianyuan Zhang
    Haotong Qin
    Jinyang Guo
    Xianglong Liu
    Visual Intelligence, 1 (1):
  • [10] Benchmarking and scaling of deep learning models for land cover image classification
    Papoutsis, Ioannis
    Bountos, Nikolaos Ioannis
    Zavras, Angelos
    Michail, Dimitrios
    Tryfonopoulos, Christos
    ISPRS JOURNAL OF PHOTOGRAMMETRY AND REMOTE SENSING, 2023, 195 : 250 - 268