Global Channel Pruning With Self-Supervised Mask Learning

被引:1
|
作者
Ma, Ming [1 ]
Zhang, Tongzhou [1 ]
Wang, Ziming [1 ]
Wang, Yue [1 ]
Du, Taoli [1 ]
Li, Wenhui [1 ]
机构
[1] Jilin Univ, Coll Comp Sci & Technol, Changchun 130012, Peoples R China
关键词
Self-supervised learning; Training; Filters; Sparse matrices; Supervised learning; Neural networks; Circuits and systems; Accuracy; Time series analysis; Libraries; Deep neural networks; network pruning; self-supervised learning;
D O I
10.1109/TCSVT.2024.3488098
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Network pruning is widely used in model compression due to its simplicity and efficiency. Existing methods typically introduce sparse loss regularization to learn masks. However, this sparse regularization approach lacks a clear criterion for evaluating channel importance and relies on manually defined rules, leading to a decline in model performance. In this article, a Self-Supervised Mask Learning (SSML) method for global channel pruning is proposed, casting mask learning as a self-supervised binary classification task to automatically identify less important channels. Specifically, a dedicated pretext task is designed for the channelwise masks, which leverages the original network to generate pseudo-labels from the mask itself to guide mask learning. Then, a polarization mask loss function is proposed, transforming the discrete mask learning problem into a differentiable binary classification problem. The proposed loss function distinguishes the similarity between pseudo-labels and masks, clustering similar masks together in the feature space and separating dissimilar masks, ultimately allowing channels with masks of 0 to be safely removed without damaging the performance of the pruned model. In addition, SSML can train from scratch to yield a compact model. Extensive experiments on CIFAR-10, CIFAR-100 and ImageNet datasets demonstrate that SSML outperforms state-of-the-art methods. For instance, SSML prunes 52.7% of the FLOPs of ResNe34 on the ImageNet dataset with only 0.01% drop in Top-1 accuracy. Moreover, the generalization of SSML is verified on downstream tasks.
引用
收藏
页码:2013 / 2025
页数:13
相关论文
共 50 条
  • [31] MarioNette: Self-Supervised Sprite Learning
    Smirnov, Dmitriy
    Gharbi, Michael
    Fisher, Matthew
    Guizilini, Vitor
    Efros, Alexei A.
    Solomon, Justin
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [32] Self-supervised Learning: A Succinct Review
    Veenu Rani
    Syed Tufael Nabi
    Munish Kumar
    Ajay Mittal
    Krishan Kumar
    Archives of Computational Methods in Engineering, 2023, 30 : 2761 - 2775
  • [33] Self-supervised learning for outlier detection
    Diers, Jan
    Pigorsch, Christian
    STAT, 2021, 10 (01):
  • [34] Self-Supervised Learning for Multimedia Recommendation
    Tao, Zhulin
    Liu, Xiaohao
    Xia, Yewei
    Wang, Xiang
    Yang, Lifang
    Huang, Xianglin
    Chua, Tat-Seng
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 5107 - 5116
  • [35] Relational Self-Supervised Learning on Graphs
    Lee, Namkyeong
    Hyun, Dongmin
    Lee, Junseok
    Park, Chanyoung
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2022, 2022, : 1054 - 1063
  • [36] Self-Supervised Learning in Remote Sensing
    Wang, Yi
    Albrecht, Conrad M.
    Ait Ali Braham, Nassim
    Mou, Lichao
    Zhu, Xiao Xiang
    IEEE GEOSCIENCE AND REMOTE SENSING MAGAZINE, 2022, 10 (04) : 213 - 247
  • [37] Whitening for Self-Supervised Representation Learning
    Ermolov, Aleksandr
    Siarohin, Aliaksandr
    Sangineto, Enver
    Sebe, Nicu
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [38] Self-supervised Graph Learning for Recommendation
    Wu, Jiancan
    Wang, Xiang
    Feng, Fuli
    He, Xiangnan
    Chen, Liang
    Lian, Jianxun
    Xie, Xing
    SIGIR '21 - PROCEEDINGS OF THE 44TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, 2021, : 726 - 735
  • [39] COMBINING SELF-SUPERVISED AND SUPERVISED LEARNING WITH NOISY LABELS
    Zhang, Yongqi
    Zhang, Hui
    Yao, Quanming
    Wan, Jun
    2023 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2023, : 605 - 609
  • [40] Self-Supervised Learning for Videos: A Survey
    Schiappa, Madeline C.
    Rawat, Yogesh S.
    Shah, Mubarak
    ACM COMPUTING SURVEYS, 2023, 55 (13S)