Maximum output discrepancy computation for convolutional neural network compression

被引:2
|
作者
Mo, Zihao [1 ]
Xiang, Weiming [1 ]
机构
[1] Augusta Univ, Sch Comp & Cyber Sci, 1120 15th St, Augusta, GA 30912 USA
基金
美国国家科学基金会;
关键词
Reachability analysis; Convolutional neural network; Discrepancy computation; Neural network compression; RECOGNITION;
D O I
10.1016/j.ins.2024.120367
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Network compression methods minimize the number of network parameters and computation costs while maintaining desired network performance. However, the safety assurance of many compression methods is based on a large amount of experimental data, whereas unforeseen incidents beyond the experiment data may result in unsafe consequences. In this work, we developed a discrepancy computation method for two convolutional neural networks by giving a concrete value to characterize the maximum output difference between the two networks after compression. Using Imagestar-based reachability analysis, we propose a novel method to merge the two networks to compute the difference. We illustrate reachability computation for each layer in the merged network, such as the convolution, max pooling, fully connected, and ReLU layers. We apply our method to a numerical example to prove its correctness. Furthermore, we implement our developed methods on the VGG16 model with the Quantization Aware Training (QAT) compression method; the results show that our approach can efficiently compute the accurate maximum output discrepancy between the original neural network and the compressed neural network.
引用
收藏
页数:18
相关论文
共 50 条
  • [21] EDP: An Efficient Decomposition and Pruning Scheme for Convolutional Neural Network Compression
    Ruan, Xiaofeng
    Liu, Yufan
    Yuan, Chunfeng
    Li, Bing
    Hu, Weiming
    Li, Yangxi
    Maybank, Stephen
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2021, 32 (10) : 4499 - 4513
  • [22] Model-based Weight Quantization for Convolutional Neural Network Compression
    Gheorghe, Stefan
    Ivanovici, Mihai
    2021 16TH INTERNATIONAL CONFERENCE ON ENGINEERING OF MODERN ELECTRIC SYSTEMS (EMES), 2021, : 94 - 97
  • [23] Convolutional Neural Network Compression via Dynamic Parameter Rank Pruning
    Sharma, Manish
    Heard, Jamison
    Saber, Eli
    Markopoulos, Panagiotis
    IEEE ACCESS, 2025, 13 : 18441 - 18456
  • [24] TC-CNN: Trajectory Compression based on Convolutional Neural Network
    Wang, Yulong
    Tang, Jingwang
    Jia, Zhe
    PROCEEDINGS OF THE 2ND INTERNATIONAL CONFERENCE ON DEEP LEARNING THEORY AND APPLICATIONS (DELTA), 2021, : 170 - 176
  • [25] A Learning Automata-Based Compression Scheme for Convolutional Neural Network
    Feng, Shuai
    Guo, Haonan
    Yang, Jichao
    Xu, Zhengwu
    Li, Shenghong
    COMMUNICATIONS, SIGNAL PROCESSING, AND SYSTEMS, CSPS 2018, VOL III: SYSTEMS, 2020, 517 : 42 - 49
  • [26] Extended Bit-Plane Compression for Convolutional Neural Network Accelerators
    Cavigelli, Lukas
    Benini, Luca
    2019 IEEE INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE CIRCUITS AND SYSTEMS (AICAS 2019), 2019, : 279 - 283
  • [27] Deep Convolutional Neural Network Compression via Coupled Tensor Decomposition
    Sun, Weize
    Chen, Shaowu
    Huang, Lei
    So, Hing Cheung
    Xie, Min
    IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2021, 15 (03) : 603 - 616
  • [28] A Feature Map Lossless Compression Framework for Convolutional Neural Network Accelerators
    Zhang, Zekun
    Jiao, Xin
    Xu, Chengyu
    2024 IEEE 6TH INTERNATIONAL CONFERENCE ON AI CIRCUITS AND SYSTEMS, AICAS 2024, 2024, : 422 - 426
  • [29] Convolutional Neural Network Compression through Generalized Kronecker Product Decomposition
    Hameed, Marawan Gamal Abdel
    Tahaei, Marzieh S.
    Mosleh, Ali
    Nia, Vahid Partovi
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / THE TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 771 - 779
  • [30] Tensor Network Methods for Hyperparameter Optimization and Compression of Convolutional Neural Networks
    Naumov, A.
    Melnikov, A.
    Perelshtein, M.
    Melnikov, Ar.
    Abronin, V.
    Oksanichenko, F.
    APPLIED SCIENCES-BASEL, 2025, 15 (04):