Stealthiness Assessment of Adversarial Perturbation: From a Visual Perspective

被引:0
|
作者
Liu, Hangcheng [1 ]
Zhou, Yuan [2 ]
Yang, Ying [3 ,4 ]
Zhao, Qingchuan [5 ]
Zhang, Tianwei [1 ]
Xiang, Tao [6 ]
机构
[1] Nanyang Technol Univ, Coll Comp & Data Sci, Jurong West 639798, Singapore
[2] Zhejiang Sci Tech Univ, Sch Comp Sci & Technol, Hangzhou 310018, Zhejiang, Peoples R China
[3] ASTAR, Inst High Performance Comp IHPC, Singapore 138632, Singapore
[4] ASTAR, Ctr Frontier AI Res CFAR, Singapore 138632, Singapore
[5] City Univ Hong Kong, Dept Comp Sci, Kowloon Tong, Hong Kong, Peoples R China
[6] Chongqing Univ, Coll Comp Sci, Chongqing 400044, Peoples R China
基金
中国国家自然科学基金;
关键词
Perturbation methods; Measurement; Observers; Predictive models; Distortion; Noise; Feature extraction; Computer science; Visualization; Visual systems; Adversarial stealthiness assessment; adversarial attack; classification; IMAGE QUALITY ASSESSMENT; DATABASE;
D O I
10.1109/TIFS.2024.3520016
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Assessing the stealthiness of adversarial perturbations is challenging due to the lack of appropriate evaluation metrics. Existing evaluation metrics, e.g., L-p norms or Image Quality Assessment (IQA), fall short of assessing the pixel-level stealthiness of subtle adversarial perturbations since these metrics are primarily designed for traditional distortions. To bridge this gap, we present the first comprehensive study on the subjective and objective assessment of the stealthiness of adversarial perturbations from a visual perspective at a pixel level. Specifically, we propose new subjective assessment criteria for human observers to score adversarial stealthiness in a fine-grained manner. Then, we create a large-scale adversarial example dataset comprising 10586 pairs of clean and adversarial samples encompassing twelve state-of-the-art adversarial attacks. To obtain the subjective scores according to the proposed criterion, we recruit 60 human observers, and each adversarial example is evaluated by at least 15 observers. The mean opinion score of each adversarial example is utilized for labeling. Finally, we develop a three-stage objective scoring model that mimics human scoring habits to predict adversarial perturbation's stealthiness. Experimental results demonstrate that our objective model exhibits superior consistency with the human visual system, surpassing commonly employed metrics like PSNR and SSIM.
引用
收藏
页码:898 / 913
页数:16
相关论文
共 50 条
  • [31] Universal adversarial perturbation for remote sensing images
    Wang, Qingyu
    Feng, Guorui
    Yin, Zhaoxia
    Luo, Bin
    2022 IEEE 24TH INTERNATIONAL WORKSHOP ON MULTIMEDIA SIGNAL PROCESSING (MMSP), 2022,
  • [32] Adversarial perturbation in remote sensing image recognition
    Ai, Shan
    Koe, Arthur Sandor Voundi
    Huang, Teng
    APPLIED SOFT COMPUTING, 2021, 105
  • [33] Defending Against Deepfakes with Ensemble Adversarial Perturbation
    Guan, Weinan
    He, Ziwen
    Wang, Wei
    Dong, Jing
    Peng, Bo
    2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 1952 - 1958
  • [34] An Audio Watermarking Algorithm Based on Adversarial Perturbation
    Wu, Shiqiang
    Liu, Jie
    Huang, Ying
    Guan, Hu
    Zhang, Shuwu
    APPLIED SCIENCES-BASEL, 2024, 14 (16):
  • [35] Adversarial Weight Perturbation Helps Robust Generalization
    Wu, Dongxian
    Xia, Shu-Tao
    Wang, Yisen
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS (NEURIPS 2020), 2020, 33
  • [36] Protecting image privacy through adversarial perturbation
    Baoyu Liang
    Chao Tong
    Chao Lang
    Qinglong Wang
    Joel J. P. C Rodrigues
    Sergei Kozlov
    Multimedia Tools and Applications, 2022, 81 : 34759 - 34774
  • [37] A Way to Explore the Lower Bound of Adversarial Perturbation
    Han, Keji
    Bai, Yuxuan
    Li, Yun
    2020 IEEE INTERNATIONAL CONFERENCE ON BIG DATA AND SMART COMPUTING (BIGCOMP 2020), 2020, : 338 - 341
  • [38] Consistency Training with Virtual Adversarial Discrete Perturbation
    Park, Jungsoo
    Kim, Gyuwan
    Kang, Jaewoo
    NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES, 2022, : 5646 - 5656
  • [39] Towards Transferable Adversarial Attacks with Centralized Perturbation
    Wu, Shangbo
    Tan, Yu-an
    Wang, Yajie
    Ma, Ruinan
    Ma, Wencong
    Li, Yuanzhang
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 6, 2024, : 6109 - 6116
  • [40] Adversarial Perturbation Defense on Deep Neural Networks
    Zhang, Xingwei
    Zheng, Xiaolong
    Mao, Wenji
    ACM COMPUTING SURVEYS, 2021, 54 (08)