Towards Efficient Verification of Quantized Neural Networks

被引:0
|
作者
Huang, Pei [1 ]
Wu, Haoze [1 ]
Yang, Yuting [2 ]
Daukantas, Ieva [3 ]
Wu, Min [1 ]
Zhang, Yedi [4 ]
Barrett, Clark [1 ]
机构
[1] Stanford Univ, Stanford, CA USA
[2] Chinese Acad Sci, Inst Comp Technol, Beijing, Peoples R China
[3] IT Univ Copenhagen, Copenhagen, Denmark
[4] Natl Univ Singapore, Singapore, Singapore
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Quantization replaces floating point arithmetic with integer arithmetic in deep neural network models, providing more efficient on-device inference with less power and memory. In this work, we propose a framework for formally verifying properties of quantized neural networks. Our baseline technique is based on integer linear programming which guarantees both soundness and completeness. We then show how efficiency can be improved by utilizing gradient-based heuristic search methods and also bound-propagation techniques. We evaluate our approach on perception networks quantized with PyTorch. Our results show that we can verify quantized networks with better scalability and efficiency than the previous state of the art.
引用
收藏
页码:21152 / 21160
页数:9
相关论文
共 50 条
  • [41] Efficient Robustness Verification of the Deep Neural Networks for Smart IoT Devices
    Zhang, Zhaodi
    Liu, Jing
    Zhang, Min
    Sun, Haiying
    COMPUTER JOURNAL, 2022, 65 (11): : 2894 - 2908
  • [42] Explore Efficient LUT-based Architecture for Quantized Convolutional Neural Networks on FPGA
    Cao, Yanpeng
    Wang, Chengcheng
    Tang, Yongming
    28TH IEEE INTERNATIONAL SYMPOSIUM ON FIELD-PROGRAMMABLE CUSTOM COMPUTING MACHINES (FCCM), 2020, : 232 - 232
  • [43] Efficient verification of neural networks based on neuron branching and LP abstraction
    Zhao, Liang
    Duan, Xinmin
    Yang, Chenglong
    Liu, Yuehao
    Dong, Yansong
    Wang, Xiaobing
    Wang, Wensheng
    NEUROCOMPUTING, 2024, 596
  • [44] Designing Efficient Shortcut Architecture for Improving the Accuracy of Fully Quantized Neural Networks Accelerator
    Li, Baoting
    Liu, Longjun
    Jin, Yanming
    Gao, Peng
    Sun, Hongbin
    Zheng, Nanning
    2020 25TH ASIA AND SOUTH PACIFIC DESIGN AUTOMATION CONFERENCE, ASP-DAC 2020, 2020, : 289 - 294
  • [45] Adaptive gradients and weight projection based on quantized neural networks for efficient image classification
    Wu, Ran
    Liu, Huanyu
    Li, Jun-Bao
    COMPUTER VISION AND IMAGE UNDERSTANDING, 2022, 223
  • [46] Towards Energy Efficient Architecture for Spaceborne Neural Networks Computation
    Wang, Shiyu
    Zhang, Shengbing
    Wang, Jihe
    Huang, Xiaoping
    ALGORITHMS AND ARCHITECTURES FOR PARALLEL PROCESSING, ICA3PP 2020, PT II, 2020, 12453 : 575 - 586
  • [47] Self-Distillation: Towards Efficient and Compact Neural Networks
    Zhang, Linfeng
    Bao, Chenglong
    Ma, Kaisheng
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (08) : 4388 - 4403
  • [48] AutoSNN: Towards Energy-Efficient Spiking Neural Networks
    Na, Byunggook
    Mok, Jisoo
    Park, Seongsik
    Lee, Dongjin
    Choe, Hyeokjun
    Yoon, Sungroh
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [49] Towards Formal Repair and Verification of Industry-scale Deep Neural Networks
    Munakata, Satoshi
    Tokumoto, Susumu
    Yamamoto, Koji
    Munakata, Kazuki
    2023 IEEE/ACM 45TH INTERNATIONAL CONFERENCE ON SOFTWARE ENGINEERING: COMPANION PROCEEDINGS, ICSE-COMPANION, 2023, : 360 - 364
  • [50] Quantized Neural Networks: Training Neural Networks with Low Precision Weights and Activations
    Hubara, Itay
    Courbariaux, Matthieu
    Soudry, Daniel
    El-Yaniv, Ran
    Bengio, Yoshua
    JOURNAL OF MACHINE LEARNING RESEARCH, 2018, 18