SA-NET: SHUFFLE ATTENTION FOR DEEP CONVOLUTIONAL NEURAL NETWORKS

被引:597
|
作者
Zhang, Qing-Long [1 ]
Yang, Yu-Bin [1 ]
机构
[1] Nanjing Univ, State Key Lab Novel Software Technol, Nanjing, Peoples R China
关键词
spatial attention; channel attention; channel shuffle; grouped features;
D O I
10.1109/ICASSP39728.2021.9414568
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Attention mechanisms, which enable a neural network to accurately focus on all the relevant elements of the input, have become an essential component to improve the performance of deep neural networks. There are mainly two attention mechanisms widely used in computer vision studies, spatial attention and channel attention, which aim to capture the pixel-level pairwise relationship and channel dependency, respectively. Although fusing them together may achieve better performance than their individual implementations, it will inevitably increase the computational overhead. In this paper, we propose an efficient Shuffle Attention (SA) module to address this issue, which adopts Shuffle Units to combine two types of attention mechanisms effectively. Specifically, SA first groups channel dimensions into multiple sub-features before processing them in parallel. Then, for each sub-feature, SA utilizes a Shuffle Unit to depict feature dependencies in both spatial and channel dimensions. After that, all sub-features are aggregated and a "channel shuffle" operator is adopted to enable information communication between different sub-features. The proposed SA module is efficient yet effective, e.g., the parameters and computations of SA against the backbone ResNet50 are 300 vs. 25.56M and 2.76e-3 GFLOPs vs. 4.12 GFLOPs, respectively, and the performance boost is more than 1.34% in terms of Top-1 accuracy. Extensive experimental results on commonused benchmarks, including ImageNet-1k for classification, MS COCO for object detection, and instance segmentation, demonstrate that the proposed SA outperforms the current SOTA methods significantly by achieving higher accuracy while having lower model complexity.
引用
收藏
页码:2235 / 2239
页数:5
相关论文
共 50 条
  • [1] BA-Net: Bridge Attention for Deep Convolutional Neural Networks
    Zhao, Yue
    Chen, Junzhou
    Zhang, Zirui
    Zhang, Ronghui
    COMPUTER VISION, ECCV 2022, PT XXI, 2022, 13681 : 297 - 312
  • [2] Rega-Net: Retina Gabor Attention for Deep Convolutional Neural Networks
    Bao, Chun
    Cao, Jie
    Ning, Yaqian
    Cheng, Yang
    Hao, Qun
    IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2023, 20
  • [3] SA-Net: A scale-attention network for medical image segmentation
    Hu, Jingfei
    Wang, Hua
    Wang, Jie
    Wang, Yunqi
    He, Fang
    Zhang, Jicong
    PLOS ONE, 2021, 16 (04):
  • [4] SA-Net: A deep spectral analysis network for image clustering
    Wang, Jinghua
    Jiang, Jianmin
    NEUROCOMPUTING, 2020, 383 (383) : 10 - 23
  • [5] Spatial Channel Attention for Deep Convolutional Neural Networks
    Liu, Tonglai
    Luo, Ronghai
    Xu, Longqin
    Feng, Dachun
    Cao, Liang
    Liu, Shuangyin
    Guo, Jianjun
    MATHEMATICS, 2022, 10 (10)
  • [6] Spatial Pyramid Attention for Deep Convolutional Neural Networks
    Ma, Xu
    Guo, Jingda
    Sansom, Andrew
    McGuire, Mara
    Kalaani, Andrew
    Chen, Qi
    Tang, Sihai
    Yang, Qing
    Fu, Song
    IEEE TRANSACTIONS ON MULTIMEDIA, 2021, 23 : 3048 - 3058
  • [7] ShuffleBlock: Shuffle to Regularize Convolutional Neural Networks
    Kumawat, Sudhakar
    Kanojia, Gagan
    Raman, Shanmuganathan
    2022 NATIONAL CONFERENCE ON COMMUNICATIONS (NCC), 2022, : 36 - 41
  • [8] Multiscale Hybrid Convolutional Deep Neural Networks with Channel Attention
    Yang, Hua
    Yang, Ming
    He, Bitao
    Qin, Tao
    Yang, Jing
    ENTROPY, 2022, 24 (09)
  • [9] Brain Tumour Segmentation Using S-Net and SA-Net
    Roy, Sunita
    Saha, Rikan
    Sarkar, Suvarthi
    Mehera, Ranjan
    Pal, Rajat Kumar
    Bandyopadhyay, Samir Kumar
    IEEE ACCESS, 2023, 11 : 28658 - 28679
  • [10] Deep Convolutional Neural Networks
    Gonzalez, Rafael C.
    IEEE SIGNAL PROCESSING MAGAZINE, 2018, 35 (06) : 79 - 87