EfficientQ: An efficient and accurate post-training neural network quantization method for medical image segmentation

被引:1
|
作者
Zhang, Rongzhao [1 ]
Chung, Albert C. S. [1 ]
机构
[1] Hong Kong Univ Sci & Technol, Dept Comp Sci & Engn, Hong Kong, Peoples R China
关键词
Neural network quantization; Post-training quantization; Image segmentation; Deep learning; Model acceleration; Model compression;
D O I
10.1016/j.media.2024.103277
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Model quantization is a promising technique that can simultaneously compress and accelerate a deep neural network by limiting its computation bit-width, which plays a crucial role in the fast-growing AI industry. Despite model quantization's success in producing well-performing low-bit models, the quantization process itself can still be expensive, which may involve a long fine-tuning stage on a large, well-annotated training set. To make the quantization process more efficient in terms of both time and data requirements, this paper proposes a fast and accurate post-training quantization method, namely EfficientQ. We develop this new method with a layer-wise optimization strategy and leverage the powerful alternating direction method of multipliers (ADMM) algorithm to ensure fast convergence. Furthermore, a weight regularization scheme is incorporated to provide more guidance for the optimization of the discrete weights, and a self-adaptive attention mechanism is proposed to combat the class imbalance problem. Extensive comparison and ablation experiments are conducted on two publicly available medical image segmentation datasets, i.e., LiTS and BraTS2020, and the results demonstrate the superiority of the proposed method over various existing post training quantization methods in terms of both accuracy and optimization speed. Remarkably, with EfficientQ, the quantization of a practical 3D UNet only requires less than 5 min on a single GPU and one data sample. The source code is available at https://github.com/rongzhao-zhang/EfficientQ.
引用
收藏
页数:16
相关论文
共 50 条
  • [1] Toward Accurate Post-Training Quantization for Image Super Resolution
    Tu, Zhijun
    Hu, Jie
    Chen, Hanting
    Wang, Yunhe
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR, 2023, : 5856 - 5865
  • [2] Post-training Quantization of Deep Neural Network Weights
    Khayrov, E. M.
    Malsagov, M. Yu.
    Karandashev, I. M.
    ADVANCES IN NEURAL COMPUTATION, MACHINE LEARNING, AND COGNITIVE RESEARCH III, 2020, 856 : 230 - 238
  • [3] Improving the Post-Training Neural Network Quantization by Prepositive Feature Quantization
    Chu, Tianshu
    Yang, Zuopeng
    Huang, Xiaolin
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (04) : 3056 - 3060
  • [4] AE-Qdrop: Towards Accurate and Efficient Low-Bit Post-Training Quantization for A Convolutional Neural Network
    Li, Jixing
    Chen, Gang
    Jin, Min
    Mao, Wenyu
    Lu, Huaxiang
    ELECTRONICS, 2024, 13 (03)
  • [5] Lost-minimum post-training parameter quantization method for convolutional neural network
    Zhang F.
    Huang Y.
    Fang Z.
    Guo W.
    Tongxin Xuebao/Journal on Communications, 2022, 43 (04): : 114 - 122
  • [6] Post-Training Quantization for Energy Efficient Realization of Deep Neural Networks
    Latotzke, Cecilia
    Balim, Batuhan
    Gemmeke, Tobias
    2022 21ST IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS, ICMLA, 2022, : 1559 - 1566
  • [7] Towards accurate post-training quantization for reparameterized models
    Zhang, Luoming
    He, Yefei
    Fei, Wen
    Lou, Zhenyu
    Wu, Weijia
    Ying, Yangwei
    Zhou, Hong
    APPLIED INTELLIGENCE, 2025, 55 (07)
  • [8] Towards Accurate Post-Training Quantization for Vision Transformer
    Ding, Yifu
    Qin, Haotong
    Yan, Qinghua
    Chai, Zhenhua
    Liu, Junjie
    Wei, Xiaolin
    Liu, Xianglong
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2022, 2022, : 5380 - 5388
  • [9] An efficient neural network based method for medical image segmentation
    Torbati, Nima
    Ayatollahi, Ahmad
    Kermani, Ali
    COMPUTERS IN BIOLOGY AND MEDICINE, 2014, 44 : 76 - 87
  • [10] PTQD: Accurate Post-Training Quantization for Diffusion Models
    He, Yefei
    Liu, Luping
    Liu, Jing
    Wu, Weijia
    Zhou, Hong
    Zhuang, Bohan
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,