STD-Net: Spatio-Temporal Decomposition Network for Video Demoiring With Sparse Transformers

被引:0
|
作者
Niu, Yuzhen [1 ,2 ]
Xu, Rui [1 ,2 ]
Lin, Zhihua [3 ]
Liu, Wenxi [1 ,2 ]
机构
[1] Fuzhou Univ, Coll Comp & Data Sci, Fujian Key Lab Network Comp & Intelligent Informa, Fuzhou 350108, Peoples R China
[2] Minist Educ, Engn Res Ctr Bigdata Intelligence, Fuzhou 350108, Peoples R China
[3] Res Inst Alipay Informat Technol Co Ltd, Hangzhou 310000, Peoples R China
基金
中国国家自然科学基金;
关键词
Image restoration; video demoireing; video restoration; spatio-temporal network; sparse transformer; QUALITY ASSESSMENT; IMAGE;
D O I
10.1109/TCSVT.2024.3386604
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
The problem of video demoireing is a new challenge in video restoration. Unlike image demoireing, which involves removing static and uniform patterns, video demoireing requires tackling dynamic and varied moire patterns while maintaining video details, colors, and temporal consistency. It is particularly challenging to model moire patterns for videos with camera or object motions, where separating moire from the original video content across frames is extremely difficult. Nonetheless, we observe that the spatial distribution of moire patterns is often sparse on each frame, and their long-range temporal correlation is not significant. To fully leverage this phenomenon, a sparsity-constrained spatial self-attention scheme is proposed to concentrate on removing sparse moire efficiently for each frame without being distracted by dynamic video content. The frame-wise spatial features are then correlated and aggregated via the local temporal cross-frame-attention module to produce temporal-consistent high-quality moire-free videos. The above decoupled spatial and temporal transformers constitute the Spatio-Temporal Decomposition Network, dubbed STD-Net. For evaluation, we present a large-scale video demoireing benchmark featuring various real-life scenes, camera motions, and object motions. We demonstrate that our proposed model can effectively and efficiently achieve superior performance on video demoireing and single image demoireing tasks. The proposed dataset is released at https://github.com/FZU-N/LVDM.
引用
收藏
页码:8562 / 8575
页数:14
相关论文
共 50 条
  • [41] Spatio-Temporal Sparse Graph Convolution Network for Hand Gesture Recognition
    Ikne, Omar
    Slama, Rim
    Saoudi, Hichem
    Wannous, Hazem
    2024 IEEE 18TH INTERNATIONAL CONFERENCE ON AUTOMATIC FACE AND GESTURE RECOGNITION, FG 2024, 2024,
  • [42] DAST-Net: Dense visual attention augmented spatio-temporal network for unsupervised video anomaly detection
    Kommanduri, Rangachary
    Ghorai, Mrinmoy
    NEUROCOMPUTING, 2024, 579
  • [43] Video Segmentation with Spatio-Temporal Tubes
    Trichet, Remi
    Nevatia, Ramakant
    2013 10TH IEEE INTERNATIONAL CONFERENCE ON ADVANCED VIDEO AND SIGNAL BASED SURVEILLANCE (AVSS 2013), 2013, : 330 - 335
  • [44] Spatio-temporal segmentation for video surveillance
    Sun, HZ
    Tan, TN
    ELECTRONICS LETTERS, 2001, 37 (01) : 20 - 21
  • [45] Spatio-temporal segmentation for video surveillance
    Sun, HZ
    Feng, T
    Tan, TN
    15TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION, VOL 1, PROCEEDINGS: COMPUTER VISION AND IMAGE ANALYSIS, 2000, : 843 - 846
  • [46] VideoZoom Spatio-Temporal Video Browser
    Smith, John R.
    IEEE TRANSACTIONS ON MULTIMEDIA, 1999, 1 (02) : 157 - 171
  • [47] Spatio-temporal video contrast enhancement
    Celik, Turgay
    IET IMAGE PROCESSING, 2013, 7 (06) : 543 - 555
  • [48] Spatio-Temporal Perturbations for Video Attribution
    Li, Zhenqiang
    Wang, Weimin
    Li, Zuoyue
    Huang, Yifei
    Sato, Yoichi
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (04) : 2043 - 2056
  • [49] Spatio-temporal querying in video databases
    Köprülü, M
    Çiçekli, NK
    Yazici, A
    FLEXIBLE QUERY ANSWERING SYSTEMS, PROCEEDINGS, 2002, 2522 : 251 - 262
  • [50] Spatio-temporal querying in video databases
    Koprulu, M
    Cicekli, NK
    Yazici, A
    INFORMATION SCIENCES, 2004, 160 (1-4) : 131 - 152