SSIR: Spatial shuffle multi-head self-attention for Single Image Super-Resolution

被引:14
|
作者
Zhao, Liangliang [1 ,2 ]
Gao, Junyu [1 ,2 ,3 ]
Deng, Donghu [1 ,2 ]
Li, Xuelong [1 ,2 ]
机构
[1] Northwestern Polytech Univ, Sch Artificial Intelligence OPt & Elect iOPEN, Xian 710072, Shaanxi, Peoples R China
[2] Minist Ind & Informat Technol, Key Lab Intelligent Interact & Applicat, Xian 710072, Shaanxi, Peoples R China
[3] Shanghai Artificial Intelligence Lab, Shanghai 200232, Peoples R China
关键词
Single Image Super-Resolution; Long-range attention; Vision transformer;
D O I
10.1016/j.patcog.2023.110195
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Benefiting from the development of deep convolutional neural networks, CNN-based single-image super-resolution methods have achieved remarkable reconstruction results. However, the limited perceptual field of the convolutional kernel and the use of static weights in the inference process limit the performance of CNN-based methods. Recently, a few vision transformer-based image super-resolution methods have achieved excellent performance compared to CNN-based methods. These methods contain many parameters and require vast amounts of GPU memory for training. In this paper, we propose a spatial shuffle multi-head self-attention for single-image super-resolution that can significantly model long-range pixel dependencies without additional computational consumption. A local perception module is also proposed to combine convolutional neural networks' local connectivity and translational invariance. Reconstruction results on five popular benchmarks show that the proposed method outperforms existing methods in both reconstruction accuracy and visual performance. The proposed method matches the performance of transformed-based methods but requires an inferior number of transformer blocks, which reduces the number of parameters by 40%, GPU memory by 30%, and inference time by 30% compared to transformer-based methods.
引用
收藏
页数:12
相关论文
共 50 条
  • [1] DROPOUT MULTI-HEAD ATTENTION FOR SINGLE IMAGE SUPER-RESOLUTION
    Yang, Chao
    Fan, Yong
    Lu, Cheng
    2024 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING, ICASSP 2024, 2024, : 2655 - 2659
  • [2] SRFormer: Permuted Self-Attention for Single Image Super-Resolution
    Zhou, Yupeng
    Li, Zhen
    Guo, Chun-Le
    Bai, Song
    Cheng, Ming-Ming
    Hou, Qibin
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 12734 - 12745
  • [3] Spatial and channel enhanced self-attention network for efficient single image super-resolution
    Song, Xiaogang
    Tan, Yuping
    Pang, Xinchao
    Zhang, Lei
    Lu, Xiaofeng
    Hei, Xinhong
    NEUROCOMPUTING, 2025, 620
  • [4] Optimal Deep Multi-Route Self-Attention for Single Image Super-Resolution
    Ngambenjavichaikul, Nisawan
    Chen, Sovann
    Aramvith, Supavadee
    PROCEEDINGS OF 2022 ASIA-PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE (APSIPA ASC), 2022, : 1181 - 1186
  • [5] A Dynamic Residual Self-Attention Network for Lightweight Single Image Super-Resolution
    Park, Karam
    Soh, Jae Woong
    Cho, Nam Ik
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 907 - 918
  • [6] SELF-ATTENTION FOR AUDIO SUPER-RESOLUTION
    Rakotonirina, Nathanael Carraz
    2021 IEEE 31ST INTERNATIONAL WORKSHOP ON MACHINE LEARNING FOR SIGNAL PROCESSING (MLSP), 2021,
  • [7] Image super-resolution reconstruction based on self-attention GAN
    Wang X.-S.
    Chao J.
    Cheng Y.-H.
    Kongzhi yu Juece/Control and Decision, 2021, 36 (06): : 1324 - 1332
  • [8] Multimodal Multi-Head Convolutional Attention with Various Kernel Sizes for Medical Image Super-Resolution
    Georgescu, Mariana-Iuliana
    Ionescu, Radu Tudor
    Miron, Andreea-Iuliana
    Savencu, Olivian
    Ristea, Nicolae-Catalin
    Verga, Nicolae
    Khan, Fahad Shahbaz
    2023 IEEE/CVF WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2023, : 2194 - 2204
  • [9] Multi-feature self-attention super-resolution network
    Yang, Aiping
    Wei, Zihao
    Wang, Jinbin
    Cao, Jiale
    Ji, Zhong
    Pang, Yanwei
    VISUAL COMPUTER, 2024, 40 (05): : 3473 - 3486
  • [10] Multi-feature self-attention super-resolution network
    Aiping Yang
    Zihao Wei
    Jinbin Wang
    Jiale Cao
    Zhong Ji
    Yanwei Pang
    The Visual Computer, 2024, 40 : 3473 - 3486