Dual Attention with the Self-Attention Alignment for Efficient Video Super-resolution

被引:0
|
作者
Yuezhong Chu
Yunan Qiao
Heng Liu
Jungong Han
机构
[1] Anhui University of Technology,School of Computer Science and Technology
[2] Aberystwyth University,Department of Computer Science
来源
Cognitive Computation | 2022年 / 14卷
关键词
Video super-resolution; Dual attention; Self-attention alignment; FLOPS;
D O I
暂无
中图分类号
学科分类号
摘要
By selectively enhancing the features extracted from convolution networks, the attention mechanism has shown its effectiveness for low-level visual tasks, especially for image super-resolution (SR). However, due to the spatiotemporal continuity of video sequences, simply applying image attention to a video does not seem to obtain good SR results. At present, there is still a lack of suitable attention structure to achieve efficient video SR. In this work, building upon the dual attention, i.e., position attention and channel attention, we proposed deep dual attention, underpinned by self-attention alignment (DASAA), for video SR. Specifically, we start by constructing a dual attention module (DAM) to strengthen the acquired spatiotemporal features and adopt a self-attention structure with the morphological mask to achieve attention alignment. Then, on top of the attention features, we utilize the up-sampling operation to reconstruct the super-resolved video images and introduce the LSTM (long short-time memory) network to guarantee the coherent consistency of the generated video frames both temporally and spatially. Experimental results and comparisons on the actual Youku-VESR dataset and the typical benchmark dataset-Vimeo-90 k demonstrate that our proposed approach achieves the best video SR effect while taking the least amount of computation. Specifically, in the Youku-VESR dataset, our proposed approach achieves a test PSNR of 35.290db and a SSIM of 0.939, respectively. In the Vimeo-90 k dataset, the PSNR/SSIM indexes of our approach are 32.878db and 0.774. Moreover, the FLOPS (float-point operations per second) of our approach is as low as 6.39G. The proposed DASAA method surpasses all video SR algorithms in the comparison. It is also revealed that there is no linear relationship between positional attention and channel attention. It suggests that our DASAA with LSTM coherent consistency architecture may have great potential for many low-level vision video applications.
引用
收藏
页码:1140 / 1151
页数:11
相关论文
共 50 条
  • [41] Fast Online Video Super-Resolution with Deformable Attention Pyramid
    Fuoli, Dario
    Danelljan, Martin
    Timofte, Radu
    Van Gool, Luc
    2023 IEEE/CVF WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2023, : 1735 - 1744
  • [42] A Lightweight Recurrent Grouping Attention Network for Video Super-Resolution
    Zhu, Yonggui
    Li, Guofang
    SENSORS, 2023, 23 (20)
  • [43] From Local to Global: Efficient Dual Attention Mechanism for Single Image Super-Resolution
    Zhang, Pei
    Lam, Edmund Y.
    IEEE ACCESS, 2021, 9 : 114957 - 114964
  • [44] Dynamic dual attention iterative network for image super-resolution
    Feng, Hao
    Wang, Liejun
    Cheng, Shuli
    Du, Anyu
    Li, Yongming
    APPLIED INTELLIGENCE, 2022, 52 (07) : 8189 - 8208
  • [45] Dynamic dual attention iterative network for image super-resolution
    Hao Feng
    Liejun Wang
    Shuli Cheng
    Anyu Du
    Yongming Li
    Applied Intelligence, 2022, 52 : 8189 - 8208
  • [46] Dual-Camera Super-Resolution with Aligned Attention Modules
    Wang, Tengfei
    Xie, Jiaxin
    Sun, Wenxiu
    Yan, Qiong
    Chen, Qifeng
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 1981 - 1990
  • [47] HAMSA: Hybrid attention transformer and multi-scale alignment aggregation network for video super-resolution
    Xiao, Hanguang
    Wen, Hao
    Wang, Xin
    Zuo, Kun
    Liu, Tianqi
    Wang, Wei
    Xu, Yong
    DIGITAL SIGNAL PROCESSING, 2025, 161
  • [48] LOW REDUNDANT ATTENTION NETWORK FOR EFFICIENT IMAGE SUPER-RESOLUTION
    Liu, Yican
    Li, Jiacheng
    Zeng, Delu
    2024 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING, ICASSP 2024, 2024, : 2950 - 2954
  • [49] Efficient residual attention network for single image super-resolution
    Fangwei Hao
    Taiping Zhang
    Linchang Zhao
    Yuanyan Tang
    Applied Intelligence, 2022, 52 : 652 - 661
  • [50] EFFICIENT HIERARCHICAL STRIPE ATTENTION FOR LIGHTWEIGHT IMAGE SUPER-RESOLUTION
    Chen, Xiaying
    Zhou, Yue
    2024 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING, ICASSP 2024, 2024, : 3770 - 3774