Residual adaptive sparse hybrid attention transformer for image super resolution

被引:0
|
作者
Huan, Hai [1 ]
Wang, Mingxuan [1 ]
机构
[1] Nanjing Univ Informat Sci & Technol, Sch Artificial Intelligence, Nanjing 210044, Peoples R China
关键词
Image super-resolution; Vision transformer; Hybrid attention; Frequency domain loss; Deep learning; MODEL;
D O I
10.1016/j.engappai.2024.108990
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Image super-resolution is a visual task that reconstructs low-resolution images into high-resolution ones. Currently, many researchers favor applying Transformer-based methods to image super-resolution tasks, which have yielded promising results. However, due to the need to capture long-range dependencies across the entire image, existing research on Vision Transformers (ViT) for super-resolution reconstruction incurs high computational costs, thereby increasing system overhead. Additionally, some researchers have proposed methods based on manually sparse attention mechanisms. However, these approaches, which acquire receptive fields in a manner similar to traditional convolutions, do not fully exploit the advantages of Transformers in extracting global information, resulting in suboptimal reconstruction performance. To leverage the Transformer's ability to capture long-range dependencies, this paper introduces a novel network called RASHAT. In RASHAT, we propose an Adaptive Sparse Hybrid Attention Block (ASHAB). This module introduces a Bi-level Routing Attention(BRA), incorporating both Channel Attention(CA) and Switch Window Multi-head Self-attention((S)W-MSA). These components are designed to capture long-range dependencies, global context, and local dependencies within the image. Additionally, the model employs an Overlapping Cross-Attention Block(OCAB) to enhance information interaction between neighboring pixels. During model training, we introduce a novel composite loss function that combines frequency domain loss with pixel loss, further improving model performance. Extensive experiments demonstrate that benefiting from the sparse attention provided by the Bi-Level Routing Attention (BRA), RASHAT achieves similar performance to the current stateof-the-art results (20.8M) with significantly fewer parameters (11.6M). These results hold across multiple commonly used datasets.
引用
收藏
页数:12
相关论文
共 50 条
  • [41] GRAN: ghost residual attention network for single image super resolution
    Niu, Axi
    Wang, Pei
    Zhu, Yu
    Sun, Jinqiu
    Yan, Qingsen
    Zhang, Yanning
    MULTIMEDIA TOOLS AND APPLICATIONS, 2024, 83 (10) : 28505 - 28522
  • [42] Residual Attribute Attention Network for Face Image Super-Resolution
    Xin, Jingwei
    Wang, Nannan
    Gao, Xinbo
    Li, Jie
    THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 9054 - 9061
  • [43] Deep Residual Attention Network for Spectral Image Super-Resolution
    Shi, Zhan
    Chen, Chang
    Xiong, Zhiwei
    Liu, Dong
    Zha, Zheng-Jun
    Wu, Feng
    COMPUTER VISION - ECCV 2018 WORKSHOPS, PT V, 2019, 11133 : 214 - 229
  • [44] Efficient residual attention network for single image super-resolution
    Hao, Fangwei
    Zhang, Taiping
    Zhao, Linchang
    Tang, Yuanyan
    APPLIED INTELLIGENCE, 2022, 52 (01) : 652 - 661
  • [45] Cascading residual-residual attention generative adversarial network for image super resolution
    Chen, Jianqiang
    Zhang, Yali
    Hu, Xiang
    Chen, Calvin Yu-Chian
    SOFT COMPUTING, 2021, 25 (14) : 9651 - 9662
  • [46] Attention-guided hybrid transformer-convolutional neural network for underwater image super-resolution
    Zhan, Zihan
    Li, Chaofeng
    Zhang, Yuqi
    JOURNAL OF ELECTRONIC IMAGING, 2024, 33 (01)
  • [47] HADT: Image super-resolution restoration using Hybrid Attention-Dense Connected Transformer Networks
    Guo, Ying
    Tian, Chang
    Liu, Jie
    Di, Chong
    Ning, Keqing
    NEUROCOMPUTING, 2025, 614
  • [48] Adaptive Nonnegative Sparse Representation for Hyperspectral Image Super-Resolution
    Li, Xuesong
    Zhang, Youqiang
    Ge, Zixian
    Cao, Guo
    Shi, Hao
    Fu, Peng
    IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, 2021, 14 : 4267 - 4283
  • [49] Hierarchical Sparse Representation with Adaptive Dictionaries for Image Super-Resolution
    Wu, Xuelian
    Deng, Daiguo
    Li, Jianhong
    Luo, Xiaonan
    Zeng, Kun
    2013 6TH INTERNATIONAL CONGRESS ON IMAGE AND SIGNAL PROCESSING (CISP), VOLS 1-3, 2013, : 272 - 276
  • [50] Image super-resolution reconstruction based on adaptive sparse representation
    Xu, Mengxi
    Yang, Yun
    Sun, Quansen
    Wu, Xiaobin
    CONCURRENCY AND COMPUTATION-PRACTICE & EXPERIENCE, 2018, 30 (24):