Maximum output discrepancy computation for convolutional neural network compression

被引:2
|
作者
Mo, Zihao [1 ]
Xiang, Weiming [1 ]
机构
[1] Augusta Univ, Sch Comp & Cyber Sci, 1120 15th St, Augusta, GA 30912 USA
基金
美国国家科学基金会;
关键词
Reachability analysis; Convolutional neural network; Discrepancy computation; Neural network compression; RECOGNITION;
D O I
10.1016/j.ins.2024.120367
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Network compression methods minimize the number of network parameters and computation costs while maintaining desired network performance. However, the safety assurance of many compression methods is based on a large amount of experimental data, whereas unforeseen incidents beyond the experiment data may result in unsafe consequences. In this work, we developed a discrepancy computation method for two convolutional neural networks by giving a concrete value to characterize the maximum output difference between the two networks after compression. Using Imagestar-based reachability analysis, we propose a novel method to merge the two networks to compute the difference. We illustrate reachability computation for each layer in the merged network, such as the convolution, max pooling, fully connected, and ReLU layers. We apply our method to a numerical example to prove its correctness. Furthermore, we implement our developed methods on the VGG16 model with the Quantization Aware Training (QAT) compression method; the results show that our approach can efficiently compute the accurate maximum output discrepancy between the original neural network and the compressed neural network.
引用
收藏
页数:18
相关论文
共 50 条
  • [1] Dual-Precision Acceleration of Convolutional Neural Network Computation with Mixed Input and Output Data Reuse
    Hsiao, Shen-Fu
    Wu, Pei-Hsuan
    Chen, Jien-Min
    Chen, Kun-Chih
    2019 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), 2019,
  • [2] A Convolutional Neural Network Image Compression Algorithm for UAVs
    Dai, Yongdong
    Tan, Jing
    Wang, Maofei
    Jiang, Chengling
    Li, Mingjiang
    JOURNAL OF CIRCUITS SYSTEMS AND COMPUTERS, 2024, 33 (12)
  • [3] Learning Filter Basis for Convolutional Neural Network Compression
    Li, Yawei
    Gu, Shuhang
    Van Gool, Luc
    Timofte, Radu
    2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 5622 - 5631
  • [4] Deep Convolutional Neural Network Architecture With Reconfigurable Computation Patterns
    Tu, Fengbin
    Yin, Shouyi
    Ouyang, Peng
    Tang, Shibin
    Liu, Leibo
    Wei, Shaojun
    IEEE TRANSACTIONS ON VERY LARGE SCALE INTEGRATION (VLSI) SYSTEMS, 2017, 25 (08) : 2220 - 2233
  • [5] A VLSI Convolutional Neural Network Architecture for Vanishing Point Computation
    Villemur, M.
    Di Federico, M.
    Julian, P.
    PROCEEDINGS OF THE 2015 ARGENTINE SCHOOL OF MICRO-NANOELECTRONICS, TECHNOLOGY AND APPLICATIONS (EAMTA), 2015, : 53 - 57
  • [6] Reliable identification of redundant kernels for convolutional neural network compression
    Wang, Wei
    Zhu, Liqiang
    Guo, Baoqing
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2019, 63
  • [7] Global Instance Relation Distillation for convolutional neural network compression
    Hu H.
    Zeng H.
    Xie Y.
    Shi Y.
    Zhu J.
    Chen J.
    Neural Computing and Applications, 2024, 36 (18) : 10941 - 10953
  • [8] Double JPEG compression forensics based on a convolutional neural network
    Wang Q.
    Zhang R.
    EURASIP Journal on Information Security, 2016 (1):
  • [9] Convolutional Neural Network for Image Compression with Application to JPEG Standard
    Puchala, Dariusz
    Stokfiszewski, Kamil
    2021 DATA COMPRESSION CONFERENCE (DCC 2021), 2021, : 361 - 361
  • [10] Max-Variance Convolutional Neural Network Model Compression
    Boone-Sifuentes, Tanya
    Robles-Kelly, Antonio
    Nazari, Asef
    2020 DIGITAL IMAGE COMPUTING: TECHNIQUES AND APPLICATIONS (DICTA), 2020,