Depth-Aware Video Frame Interpolation

被引:361
|
作者
Bao, Wenbo [1 ]
Lai, Wei-Sheng [3 ]
Ma, Chao [2 ]
Zhang, Xiaoyun [1 ]
Gao, Zhiyong [1 ]
Yang, Ming-Hsuan [3 ,4 ]
机构
[1] Shanghai Jiao Tong Univ, Inst Image Commun & Network Engn, Shanghai, Peoples R China
[2] Shanghai Jiao Tong Univ, AI Inst, MoE Key Lab Artificial Intelligence, Shanghai, Peoples R China
[3] Univ Calif Merced, Merced, CA USA
[4] Google, Mountain View, CA 94043 USA
基金
上海市自然科学基金;
关键词
D O I
10.1109/CVPR.2019.00382
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Video frame interpolation aims to synthesize nonexistent frames in-between the original frames. While significant advances have been made from the recent deep convolutional neural networks, the quality of interpolation is often reduced due to large object motion or occlusion. In this work, we propose a video frame interpolation method which explicitly detects the occlusion by exploring the depth information. Specifically, we develop a depth-aware flow projection layer to synthesize intermediate flows that preferably sample closer objects than farther ones. In addition, we learn hierarchical features to gather contextual information from neighboring pixels. The proposed model then warps the input frames, depth maps, and contextual features based on the optical flow and local interpolation kernels for synthesizing the output frame. Our model is compact, efficient, and fully differentiable. Quantitative and qualitative results demonstrate that the proposed model performs favorably against state-of-the-art frame interpolation methods on a wide variety of datasets. The source code and pre-trained model are available at https://github.com/baowenbo/DAIN.
引用
收藏
页码:3698 / 3707
页数:10
相关论文
共 50 条
  • [11] Depth-Aware Panoptic Segmentation
    Tuan Nguyen
    Mehltretter, Max
    Rottensteiner, Franz
    ISPRS ANNALS OF THE PHOTOGRAMMETRY, REMOTE SENSING AND SPATIAL INFORMATION SCIENCES: VOLUME X-2-2024, 2024, : 153 - 161
  • [12] Depth-Aware Shadow Removal
    Fu, Yanping
    Gai, Zhenyu
    Zhao, Haifeng
    Zhang, Shaojie
    Shan, Ying
    Wu, Yang
    Tang, Jin
    COMPUTER GRAPHICS FORUM, 2022, 41 (07) : 455 - 464
  • [13] DEPTH-AWARE 3D VIDEO FILTERING TARGETTING MULTIVIEW VIDEO PLUS DEPTH COMPRESSION
    Aflaki, Payman
    Hannuksela, Miska M.
    Homayouni, Maryam
    Gabbouj, Moncef
    2014 3DTV-CONFERENCE: THE TRUE VISION - CAPTURE, TRANSMISSION AND DISPLAY OF 3D VIDEO (3DTV-CON), 2014,
  • [14] PolyphonicFormer: Unified Query Learning for Depth-Aware Video Panoptic Segmentation
    Yuan, Haobo
    Li, Xiangtai
    Yang, Yibo
    Cheng, Guangliang
    Zhang, Jing
    Tong, Yunhai
    Zhang, Lefei
    Tao, Dacheng
    COMPUTER VISION - ECCV 2022, PT XXVII, 2022, 13687 : 582 - 599
  • [15] Depth-Aware Generative Adversarial Network for Talking Head Video Generation
    Hong, Fa-Ting
    Zhang, Longhao
    Shen, Li
    Xu, Dan
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 3387 - 3396
  • [16] SAIN: SIMILARITY-AWARE VIDEO FRAME INTERPOLATION
    Lv, Yue
    Yang, Wenming
    Zuo, Wangmeng
    Liao, Qingmin
    Zhu, Rui
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 1920 - 1924
  • [17] Context-aware Synthesis for Video Frame Interpolation
    Niklaus, Simon
    Liu, Feng
    2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 1701 - 1710
  • [18] Stable Video Style Transfer Based on Partial Convolution with Depth-Aware Supervision
    Liu, Songhua
    Wu, Hao
    Luo, Shoutong
    Sun, Zhengxing
    MM '20: PROCEEDINGS OF THE 28TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, 2020, : 2445 - 2453
  • [19] Uni-DVPS: Unified Model for Depth-Aware Video Panoptic Segmentation
    Ji-Yeon, Kim
    Hyun-Bin, Oh
    Byung-Ki, Kwon
    Kim, Dahun
    Kwon, Yongjin
    Oh, Tae-Hyun
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2024, 9 (07): : 6186 - 6193
  • [20] Depth-Aware Image Seam Carving
    Shen, Jianbing
    Wang, Dapeng
    Li, Xuelong
    IEEE TRANSACTIONS ON CYBERNETICS, 2013, 43 (05) : 1453 - 1461