Residual adaptive sparse hybrid attention transformer for image super resolution

被引:0
|
作者
Huan, Hai [1 ]
Wang, Mingxuan [1 ]
机构
[1] Nanjing Univ Informat Sci & Technol, Sch Artificial Intelligence, Nanjing 210044, Peoples R China
关键词
Image super-resolution; Vision transformer; Hybrid attention; Frequency domain loss; Deep learning; MODEL;
D O I
10.1016/j.engappai.2024.108990
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Image super-resolution is a visual task that reconstructs low-resolution images into high-resolution ones. Currently, many researchers favor applying Transformer-based methods to image super-resolution tasks, which have yielded promising results. However, due to the need to capture long-range dependencies across the entire image, existing research on Vision Transformers (ViT) for super-resolution reconstruction incurs high computational costs, thereby increasing system overhead. Additionally, some researchers have proposed methods based on manually sparse attention mechanisms. However, these approaches, which acquire receptive fields in a manner similar to traditional convolutions, do not fully exploit the advantages of Transformers in extracting global information, resulting in suboptimal reconstruction performance. To leverage the Transformer's ability to capture long-range dependencies, this paper introduces a novel network called RASHAT. In RASHAT, we propose an Adaptive Sparse Hybrid Attention Block (ASHAB). This module introduces a Bi-level Routing Attention(BRA), incorporating both Channel Attention(CA) and Switch Window Multi-head Self-attention((S)W-MSA). These components are designed to capture long-range dependencies, global context, and local dependencies within the image. Additionally, the model employs an Overlapping Cross-Attention Block(OCAB) to enhance information interaction between neighboring pixels. During model training, we introduce a novel composite loss function that combines frequency domain loss with pixel loss, further improving model performance. Extensive experiments demonstrate that benefiting from the sparse attention provided by the Bi-Level Routing Attention (BRA), RASHAT achieves similar performance to the current stateof-the-art results (20.8M) with significantly fewer parameters (11.6M). These results hold across multiple commonly used datasets.
引用
收藏
页数:12
相关论文
共 50 条
  • [1] Residual SwinV2 transformer coordinate attention network for image super resolution
    Lei, Yushi
    Zhu, Zhengwei
    Qin, Yilin
    Zhu, Chenyang
    Zhu, Yanping
    AI COMMUNICATIONS, 2024, 37 (04) : 693 - 709
  • [2] Adaptive Residual Channel Attention Network for Single Image Super-Resolution
    Cao, Kerang
    Liu, Yuqing
    Duan, Lini
    Xie, Tian
    SCIENTIFIC PROGRAMMING, 2020, 2020
  • [3] HRAN: Hybrid Residual Attention Network for Single Image Super-Resolution
    Muqeet, Abdul
    Bin Iqbal, Md Tauhid
    Bae, Sung-Ho
    IEEE ACCESS, 2019, 7 : 137020 - 137029
  • [4] EHAT:Enhanced Hybrid Attention Transformer for Remote Sensing Image Super-Resolution
    Wang, Jian
    Xie, Zexin
    Du, Yanlin
    Song, Wei
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2024, PT VIII, 2025, 15038 : 225 - 237
  • [5] Residual Adaptive Dense Weight Attention Network for Single Image Super-Resolution
    Chen, Jiacheng
    Wang, Wanliang
    Xing, Fangsen
    Qian, Yutong
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [6] Lightweight Super-Resolution Image-Reconstruction Model with Adaptive Residual Attention
    Jiang Ming
    Xiao Qingsheng
    Yi Jianbing
    Cao Feng
    LASER & OPTOELECTRONICS PROGRESS, 2022, 59 (16)
  • [7] Efficient Dual Attention Transformer for Image Super-Resolution
    Park, Soobin
    Jeong, Yuna
    Choi, Yong Suk
    39TH ANNUAL ACM SYMPOSIUM ON APPLIED COMPUTING, SAC 2024, 2024, : 963 - 970
  • [8] A sparse lightweight attention network for image super-resolution
    Zhang, Hongao
    Fang, Jinsheng
    Hu, Siyu
    Zeng, Kun
    VISUAL COMPUTER, 2024, 40 (02): : 1261 - 1272
  • [9] A sparse lightweight attention network for image super-resolution
    Hongao Zhang
    Jinsheng Fang
    Siyu Hu
    Kun Zeng
    The Visual Computer, 2024, 40 (2) : 1261 - 1272
  • [10] Adaptive Attention Network for Image Super-resolution
    Chen Y.-M.
    Zhou D.-W.
    Zidonghua Xuebao/Acta Automatica Sinica, 2022, 48 (08): : 1950 - 1960