Video Frame Prediction by Joint Optimization of Direct Frame Synthesis and Optical-Flow Estimation

被引:1
|
作者
Ranjan, Navin [1 ]
Bhandari, Sovit [1 ]
Kim, Yeong-Chan [1 ,2 ]
Kim, Hoon [1 ,2 ]
机构
[1] Incheon Natl Univ, Iot & Big Data Res Ctr, Incheon 22012, South Korea
[2] Incheon Natl Univ, Dept Elect Engn, Incheon 22012, South Korea
来源
CMC-COMPUTERS MATERIALS & CONTINUA | 2023年 / 75卷 / 02期
关键词
Video frame prediction; multi -step prediction; optical; -flow; prediction; delay; deep learning;
D O I
10.32604/cmc.2023.026086
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Video prediction is the problem of generating future frames by exploiting the spatiotemporal correlation from the past frame sequence. It is one of the crucial issues in computer vision and has many real-world applications, mainly focused on predicting future scenarios to avoid unde-sirable outcomes. However, modeling future image content and object is challenging due to the dynamic evolution and complexity of the scene, such as occlusions, camera movements, delay and illumination. Direct frame synthe-sis or optical-flow estimation are common approaches used by researchers. However, researchers mainly focused on video prediction using one of the approaches. Both methods have limitations, such as direct frame synthesis, usually face blurry prediction due to complex pixel distributions in the scene, and optical-flow estimation, usually produce artifacts due to large object displacements or obstructions in the clip. In this paper, we constructed a deep neural network Frame Prediction Network (FPNet-OF) with multiple -branch inputs (optical flow and original frame) to predict the future video frame by adaptively fusing the future object-motion with the future frame generator. The key idea is to jointly optimize direct RGB frame synthesis and dense optical flow estimation to generate a superior video prediction network. Using various real-world datasets, we experimentally verify that our proposed framework can produce high-level video frame compared to other state-of-the-art framework.
引用
收藏
页码:2615 / 2639
页数:25
相关论文
共 50 条
  • [31] Joint reference frame synthesis and post filter enhancement for Versatile Video Coding
    Bao, Weijie
    Zhang, Yuantong
    Jia, Jianghao
    Chen, Zhenzhong
    Liu, Shan
    Journal of Visual Communication and Image Representation, 2025, 108
  • [32] Enhanced Real-Time Intermediate Flow Estimation for Video Frame Interpolation
    Kim, Minseop
    Choi, Haechul
    JOURNAL OF WEB ENGINEERING, 2021, 20 (08): : 2413 - 2431
  • [33] Distributed Video Coding with Frame Estimation at Decoder
    Chiam, Kin Honn
    Salleh, Mohd Fadzli Mohd
    ADVANCED COMPUTER AND COMMUNICATION ENGINEERING TECHNOLOGY, 2015, 315 : 299 - 308
  • [34] Future Motion Estimation for Video Frame Extrapolation
    Woo, Seunggyun
    Park, Junheum
    Kim, Chang-Su
    2022 INTERNATIONAL CONFERENCE ON ELECTRONICS, INFORMATION, AND COMMUNICATION (ICEIC), 2022,
  • [35] Video Frame Prediction via Deep Learning
    Yilmaz, M. Akin
    Tekalp, A. Murat
    2020 28TH SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE (SIU), 2020,
  • [36] Video inter-frame forgery identification based on optical flow consistency
    Wang, Qi
    Li, Zhaohong
    Zhang, Zhenzhen
    Ma, Qinglong
    Sensors and Transducers, 2014, 166 (03): : 229 - 234
  • [37] Optical Flow Based Co-Located Reference Frame for Video Compression
    Li, Bohan
    Han, Jingning
    Xu, Yaowu
    Rose, Kenneth
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2020, 29 : 8303 - 8315
  • [38] Scalable video transformer for full-frame video prediction
    Li, Zhan
    Liu, Feng
    Computer Vision and Image Understanding, 2024, 249
  • [39] VIDEO SIGNAL-DEPENDENT NOISE ESTIMATION VIA INTER-FRAME PREDICTION
    Li, Yanhao
    Gardella, Marina
    Bammey, Quentin
    Nikoukhah, Tina
    von Gioi, Rafael Grompone
    Colom, Miguel
    Morel, Jean-Michel
    2022 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2022, : 1406 - 1410
  • [40] Joint Learned and Traditional Video Compression for P Frame
    Wang, Zhao
    Liao, Ru-Ling
    Ye, Yan
    2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW 2020), 2020, : 560 - 564