Residual adaptive sparse hybrid attention transformer for image super resolution

被引:0
|
作者
Huan, Hai [1 ]
Wang, Mingxuan [1 ]
机构
[1] Nanjing Univ Informat Sci & Technol, Sch Artificial Intelligence, Nanjing 210044, Peoples R China
关键词
Image super-resolution; Vision transformer; Hybrid attention; Frequency domain loss; Deep learning; MODEL;
D O I
10.1016/j.engappai.2024.108990
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Image super-resolution is a visual task that reconstructs low-resolution images into high-resolution ones. Currently, many researchers favor applying Transformer-based methods to image super-resolution tasks, which have yielded promising results. However, due to the need to capture long-range dependencies across the entire image, existing research on Vision Transformers (ViT) for super-resolution reconstruction incurs high computational costs, thereby increasing system overhead. Additionally, some researchers have proposed methods based on manually sparse attention mechanisms. However, these approaches, which acquire receptive fields in a manner similar to traditional convolutions, do not fully exploit the advantages of Transformers in extracting global information, resulting in suboptimal reconstruction performance. To leverage the Transformer's ability to capture long-range dependencies, this paper introduces a novel network called RASHAT. In RASHAT, we propose an Adaptive Sparse Hybrid Attention Block (ASHAB). This module introduces a Bi-level Routing Attention(BRA), incorporating both Channel Attention(CA) and Switch Window Multi-head Self-attention((S)W-MSA). These components are designed to capture long-range dependencies, global context, and local dependencies within the image. Additionally, the model employs an Overlapping Cross-Attention Block(OCAB) to enhance information interaction between neighboring pixels. During model training, we introduce a novel composite loss function that combines frequency domain loss with pixel loss, further improving model performance. Extensive experiments demonstrate that benefiting from the sparse attention provided by the Bi-Level Routing Attention (BRA), RASHAT achieves similar performance to the current stateof-the-art results (20.8M) with significantly fewer parameters (11.6M). These results hold across multiple commonly used datasets.
引用
收藏
页数:12
相关论文
共 50 条
  • [31] Adaptive Residual Neural Network for Image Super-Resolution
    Li, Weiwei
    Li, Xinlong
    Liu, Zhenbing
    MIPPR 2019: PARALLEL PROCESSING OF IMAGES AND OPTIMIZATION TECHNIQUES; AND MEDICAL IMAGING, 2020, 11431
  • [32] Non-local sparse attention based swin transformer V2 for image super-resolution
    Lv, Ningning
    Yuan, Min
    Xie, Yufei
    Zhan, Kun
    Lu, Fuxiang
    SIGNAL PROCESSING, 2024, 222
  • [33] Remote Sensing Image Super-Resolution via Residual-Dense Hybrid Attention Network
    Yu, Bo
    Lei, Bin
    Guo, Jiayi
    Sun, Jiande
    Li, Shengtao
    Xie, Guangshuai
    REMOTE SENSING, 2022, 14 (22)
  • [34] Lightweight image super-resolution with multiscale residual attention network
    Xiao, Cunjun
    Dong, Hui
    Li, Haibin
    Li, Yaqian
    Zhang, Wenming
    JOURNAL OF ELECTRONIC IMAGING, 2022, 31 (04)
  • [35] Dual residual attention module network for single image super resolution
    Wang, Xiumei
    Gu, Yanan
    Gao, Xinbo
    Hui, Zheng
    NEUROCOMPUTING, 2019, 364 : 269 - 279
  • [36] Image Super-resolution via Residual Block Attention Networks
    Dai, Tao
    Zha, Hua
    Jiang, Yong
    Xia, Shu-Tao
    2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW), 2019, : 3879 - 3886
  • [37] GRAN: ghost residual attention network for single image super resolution
    Axi Niu
    Pei Wang
    Yu Zhu
    Jinqiu Sun
    Qingsen Yan
    Yanning Zhang
    Multimedia Tools and Applications, 2024, 83 : 28505 - 28522
  • [38] Efficient residual attention network for single image super-resolution
    Fangwei Hao
    Taiping Zhang
    Linchang Zhao
    Yuanyan Tang
    Applied Intelligence, 2022, 52 : 652 - 661
  • [39] Channel attention and residual concatenation network for image super-resolution
    Cai T.-J.
    Peng X.-Y.
    Shi Y.-P.
    Huang J.
    Peng, Xiao-Yu (pengxy96@qq.com), 1600, Chinese Academy of Sciences (29): : 142 - 151
  • [40] Gradient residual attention network for infrared image super-resolution
    Yuan, Xilin
    Zhang, Baohui
    Zhou, Jinjie
    Lian, Cheng
    Zhang, Qian
    Yue, Jiang
    OPTICS AND LASERS IN ENGINEERING, 2024, 175