Misalignment-Robust Joint Filter for Cross-Modal Image Pairs

被引:22
|
作者
Shibata, Takashi [1 ]
Tanaka, Masayuki [2 ]
Okutomi, Masatoshi [2 ]
机构
[1] NEC Corp Ltd, Tokyo, Japan
[2] Tokyo Inst Technol, Tokyo, Japan
关键词
D O I
10.1109/ICCV.2017.357
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Although several powerful joint filters for cross-modal image pairs have been proposed, the existing joint filters generate severe artifacts when there are misalignments between a target and a guidance images. Our goal is to generate an artifact-free output image even from the misaligned target and guidance images. We propose a novel misalignment-robust joint filter based on weight-volume-based image composition and joint-filter cost volume. Our proposed method first generates a set of translated guidances.Next, the joint-filter cost volume and a set of filtered images are computed from the target image and the set of the translated guidances. Then, a weight volume is obtained from the joint-filter cost volume while considering a spatial smoothness and a label-sparseness. The final output image is composed by fusing the set of the filtered images with the weight volume for the filtered images. The key is to generate the final output image directly from the set of the filtered images by weighted averaging using the weight volume that is obtained from the joint-filter cost volume. The proposed framework is widely applicable and can involve any kind of joint filter. Experimental results show that the proposed method is effective for various applications including image denosing, image up-sampling, haze removal and depth map interpolation.
引用
收藏
页码:3315 / 3324
页数:10
相关论文
共 50 条
  • [31] Learning TFIDF Enhanced Joint Embedding for Recipe-Image Cross-Modal Retrieval Service
    Xie, Zhongwei
    Liu, Ling
    Wu, Yanzhao
    Li, Lin
    Zhong, Luo
    IEEE TRANSACTIONS ON SERVICES COMPUTING, 2022, 15 (06) : 3304 - 3316
  • [32] Texture BERT for Cross-modal Texture Image Retrieval
    Xu, Zelai
    Yu, Tan
    Li, Ping
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2022, 2022, : 4610 - 4614
  • [33] An ensemble prior of image structure for cross-modal inference
    Ravela, S
    Torralba, A
    Freeman, WT
    TENTH IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION, VOLS 1 AND 2, PROCEEDINGS, 2005, : 871 - 876
  • [34] Cross-Modal Coherence for Text-to-Image Retrieval
    Alikhani, Malihe
    Han, Fangda
    Ravi, Hareesh
    Kapadia, Mubbasir
    Pavlovic, Vladimir
    Stone, Matthew
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 10427 - 10435
  • [35] Exploring and Distilling Cross-Modal Information for Image Captioning
    Liu, Fenglin
    Ren, Xuancheng
    Liu, Yuanxin
    Lei, Kai
    Sun, Xu
    PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 5095 - 5101
  • [36] Image Tagging via Cross-Modal Semantic Mapping
    Deng, Zhi-Hong
    Yu, Hongliang
    Yang, Yunlun
    MM'15: PROCEEDINGS OF THE 2015 ACM MULTIMEDIA CONFERENCE, 2015, : 1143 - 1146
  • [37] Social Image Parsing by Cross-Modal Data Refinement
    Lu, Zhiwu
    Gao, Xin
    Huang, Songfang
    Wang, Liwei
    Wen, Ji-Rong
    PROCEEDINGS OF THE TWENTY-FOURTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE (IJCAI), 2015, : 2169 - 2175
  • [38] SOUND AND IMAGE TOGETHER + FILM - CROSS-MODAL CONFIRMATION
    ANDERSON, JD
    WIDE ANGLE-A QUARTERLY JOURNAL OF FILM HISTORY THEORY CRITICISM & PRACTICE, 1993, 15 (01): : 30 - 43
  • [39] Cross-Modal Transformers for Infrared and Visible Image Fusion
    Park, Seonghyun
    Vien, An Gia
    Lee, Chul
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (02) : 770 - 785
  • [40] Explanation guided cross-modal social image clustering
    Yan, Xiaoqiang
    Mao, Yiqiao
    Ye, Yangdong
    Yu, Hui
    Wang, Fei-Yue
    INFORMATION SCIENCES, 2022, 593 : 1 - 16