Video super-resolution based on spatial-temporal recurrent residual networks

被引:43
|
作者
Yang, Wenhan [1 ]
Feng, Jiashi [2 ]
Xie, Guosen [3 ]
Liu, Jiaying [1 ]
Guo, Zongming [1 ]
Yan, Shuicheng [4 ]
机构
[1] Peking Univ, Inst Comp Sci & Technol, Beijing 100871, Peoples R China
[2] Natl Univ Singapore, Dept Elect & Comp Engn, Singapore 117583, Singapore
[3] Chinese Acad Sci, Inst Automat, NLPR, Beijing 100190, Peoples R China
[4] Qihoo 360 Technol Co Ltd, Artificial Intelligence Inst, Beijing 100015, Peoples R China
基金
中国国家自然科学基金;
关键词
Spatial residue; Temporal residue; Video super-resolution; Inter-frame motion context; Intra-frame redundancy; IMAGE SUPERRESOLUTION; ALGORITHM;
D O I
10.1016/j.cviu.2017.09.002
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we propose a new video Super-Resolution (SR) method by jointly modeling intra-frame redundancy and inter-frame motion context in a unified deep network. Different from conventional methods, the proposed Spatial-Temporal Recurrent Residual Network (STR-ResNet) investigates both spatial and temporal residues, which are represented by the difference between a high resolution (HR) frame and its corresponding low resolution (LR) frame and the difference between adjacent HR frames, respectively. This spatial-temporal residual learning model is then utilized to connect the intra-frame and inter-frame redundancies within video sequences in a recurrent convolutional network and to predict HR temporal residues in the penultimate layer as guidance to benefit estimating the spatial residue for video SR. Extensive experiments have demonstrated that the proposed STR-ResNet is able to efficiently reconstruct videos with diversified contents and complex motions, which outperforms the existing video SR approaches and offers new state-of-the-art performances on benchmark datasets.
引用
收藏
页码:79 / 92
页数:14
相关论文
共 50 条
  • [41] Video Super-Resolution via Residual Learning
    Wang, Wenjun
    Ren, Chao
    He, Xiaohai
    Chen, Honggang
    Qing, Linbo
    IEEE ACCESS, 2018, 6 : 23767 - 23777
  • [42] Deep Residual Networks of Residual Networks for Image Super-Resolution
    Wei, Xueqi
    Yang, Fumeng
    Wu, Congzhong
    LIDAR IMAGING DETECTION AND TARGET RECOGNITION 2017, 2017, 10605
  • [43] Video Temporal Super-Resolution Based on Self-similarity
    Shimano, Mihoko
    Okabe, Takahiro
    Sato, Imari
    Sato, Yoichi
    COMPUTER VISION-ACCV 2010, PT I, 2011, 6492 : 93 - 106
  • [44] Deep Recurrent Resnet for Video Super-Resolution
    Lim, Bee
    Lee, Kyoung Mu
    2017 ASIA-PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE (APSIPA ASC 2017), 2017, : 1452 - 1455
  • [45] Frame-Recurrent Video Super-Resolution
    Sajjadi, Mehdi S. M.
    Vemulapalli, Raviteja
    Brown, Matthew
    2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 6626 - 6634
  • [46] An Efficient and Lightweight Structure for Spatial-Temporal Feature Extraction in Video Super Resolution
    He, Xiaonan
    Xia, Yukun
    Qiao, Yuansong
    Lee, Brian
    Ye, Yuhang
    ADVANCES IN COMPUTER GRAPHICS, CGI 2023, PT I, 2024, 14495 : 362 - 374
  • [47] Super-resolution Reconstruction of Video Sequences Based on Wavelet-domain Spatial and Temporal Processing
    Lee, Chang-Ming
    Lee, Chien-Jung
    Hsieh, Chia-Yung
    Lie, Wen-Nung
    2012 21ST INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR 2012), 2012, : 194 - 197
  • [48] An FPGA-Based Residual Recurrent Neural Network for Real-Time Video Super-Resolution
    Sun, Kaicong
    Koch, Maurice
    Wang, Zhe
    Jovanovic, Slavisa
    Rabah, Hassan
    Simon, Sven
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (04) : 1739 - 1750
  • [49] Video Super-Resolution Using a Grouped Residual in Residual Network
    Ashoori, MohammadHossein
    Amini, Arash
    arXiv, 2023,
  • [50] SPATIAL SUPER-RESOLUTION OF A DIFFUSION FIELD BY TEMPORAL OVERSAMPLING IN SENSOR NETWORKS
    Lu, Yue M.
    Vetterli, Martin
    2009 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, VOLS 1- 8, PROCEEDINGS, 2009, : 2249 - 2252