Discrete Cosine Transform Network for Guided Depth Map Super-Resolution

被引:66
|
作者
Zhao, Zixiang [1 ,2 ]
Zhang, Jiangshe [1 ]
Xu, Shuang [1 ,3 ]
Lin, Zudi [2 ]
Pfister, Hanspeter [2 ]
机构
[1] Xi An Jiao Tong Univ, Xian, Peoples R China
[2] Harvard Univ, Cambridge, MA 02138 USA
[3] Northwestern Polytech Univ, Xian, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
10.1109/CVPR52688.2022.00561
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Guided depth super-resolution (GDSR) is an essential topic in multi-modal image processing, which reconstructs high-resolution (HR) depth maps from low-resolution ones collected with suboptimal conditions with the help of HR RGB images of the same scene. To solve the challenges in interpreting the working mechanism, extracting cross-modal features and RGB texture over-transferred, we propose a novel Discrete Cosine Transform Network (DCTNet) to alleviate the problems from three aspects. First, the Discrete Cosine Transform (DCT) module reconstructs the multi-channel HR depth features by using DCT to solve the channel-wise optimization problem derived from the image domain. Second, we introduce a semi-coupled feature extraction module that uses shared convolutional kernels to extract common information and private kernels to extract modality-specific information. Third, we employ an edge attention mechanism to highlight the contours informative for guided upsampling. Extensive quantitative and qualitative evaluations demonstrate the effectiveness of our DCTNet, which outperforms previous state-of-the-art methods with a relatively small number of parameters. The code is available at https:// github.com/Zhaozixiang1228/GDSR- DCTNet.
引用
收藏
页码:5687 / 5697
页数:11
相关论文
共 50 条
  • [21] DSRNet: Depth Super-Resolution Network guided by blurry depth and clear intensity edges
    Lan, Hui
    Jung, Cheolkon
    SIGNAL PROCESSING-IMAGE COMMUNICATION, 2024, 121
  • [22] Multi-Scale Frequency Reconstruction for Guided Depth Map Super-Resolution via Deep Residual Network
    Zuo, Yifan
    Wu, Qiang
    Fang, Yuming
    An, Ping
    Huang, Liqin
    Chen, Zhifeng
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2020, 30 (02) : 297 - 306
  • [23] Learning content-aware feature fusion for guided depth map super-resolution
    Zuo, Yifan
    Wang, Hao
    Xu, Yaping
    Huang, Huimin
    Huang, Xiaoshui
    Xia, Xue
    Fang, Yuming
    SIGNAL PROCESSING-IMAGE COMMUNICATION, 2024, 126
  • [24] Edge-guided single facial depth map super-resolution using CNN
    Zhang, Fan
    Liu, Na
    Chang, Liang
    Duan, Fuqing
    Deng, Xiaoming
    IET IMAGE PROCESSING, 2020, 14 (17) : 4708 - 4716
  • [25] Depth map super-resolution based on edge-guided joint trilateral upsampling
    Yang, Shuyuan
    Cao, Ning
    Guo, Bin
    Li, Gang
    VISUAL COMPUTER, 2022, 38 (03): : 883 - 895
  • [26] Depth map super-resolution based on edge-guided joint trilateral upsampling
    Shuyuan Yang
    Ning Cao
    Bin Guo
    Gang Li
    The Visual Computer, 2022, 38 : 883 - 895
  • [27] Depth map super-resolution via low-resolution depth guided joint trilateral up-sampling
    Yuan, Liang
    Jin, Xin
    Li, Yangguang
    Yuan, Chun
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2017, 46 : 280 - 291
  • [28] Deep Depth Super-Resolution: Learning Depth Super-Resolution Using Deep Convolutional Neural Network
    Song, Xibin
    Dai, Yuchao
    Qin, Xueying
    COMPUTER VISION - ACCV 2016, PT IV, 2017, 10114 : 360 - 376
  • [29] Joint Residual Pyramid for Depth Map Super-Resolution
    Xiao, Yi
    Cao, Xiang
    Zheng, Yan
    Zhu, Xianyi
    PRICAI 2018: TRENDS IN ARTIFICIAL INTELLIGENCE, PT I, 2018, 11012 : 797 - 810
  • [30] JOINT TRILATERAL FILTERING FOR DEPTH MAP SUPER-RESOLUTION
    Lo, Kai-Han
    Wang, Yu-Chiang Frank
    Hua, Kai-Lung
    2013 IEEE INTERNATIONAL CONFERENCE ON VISUAL COMMUNICATIONS AND IMAGE PROCESSING (IEEE VCIP 2013), 2013,