Object-level Scene Deocclusion

被引:0
|
作者
Liu, Zhengzhe [1 ]
Liu, Qing [2 ]
Chang, Chirui [3 ]
Zhang, Jianming [2 ]
Pakhomov, Daniil [2 ]
Zheng, Haitian [2 ]
Lin, Zhe [2 ]
Cohen-Or, Daniel [4 ]
Fu, Chi-Wing [1 ]
机构
[1] Chinese Univ Hong Kong, Hong Kong, Peoples R China
[2] Adobe, San Jose, CA USA
[3] Univ Hong Kong, Hong Kong, Peoples R China
[4] Tel Aviv Univ, Tel Aviv, Israel
关键词
scene deocclusion; object completion; image recomposition;
D O I
10.1145/3641519.3657409
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deoccluding the hidden portions of objects in a scene is a formidable task, particularly when addressing real-world scenes. In this paper, we present a new self-supervised PArallel visible-to-COmplete diffusion framework, named PACO, a foundation model for object-level scene deocclusion. Leveraging the rich prior of pre-trained models, we first design the parallel variational autoencoder, which produces a full-view feature map that simultaneously encodes multiple complete objects, and the visible-to-complete latent generator, which learns to implicitly predict the full-view feature map from partial-view feature map and text prompts extracted from the incomplete objects in the input image. To train PACO, we create a large-scale dataset with 500k samples to enable self-supervised learning, avoiding tedious annotations of the amodal masks and occluded regions. At inference, we devise a layer-wise deocclusion strategy to improve efficiency while maintaining the deocclusion quality. Extensive experiments on COCOA and various real-world scenes demonstrate the superior capability of PACO for scene deocclusion, surpassing the state of the arts by a large margin. Our method can also be extended to cross-domain scenes and novel categories that are not covered by the training set. Further, we demonstrate the deocclusion applicability of PACO in single-view 3D scene reconstruction and object recomposition. Project page: https://liuzhengzhe.github.io/Deocclude-Any-Object.github.io/.
引用
收藏
页数:11
相关论文
共 50 条
  • [41] Fast Object-Level Change Detection for VHR Images
    Huo, Chunlei
    Zhou, Zhixin
    Lu, Hanqing
    Pan, Chunhong
    Chen, Keming
    IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2010, 7 (01) : 118 - 122
  • [42] Object-Level Motion Detection From Moving Cameras
    Chen, Tao
    Lu, Shijian
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2017, 27 (11) : 2333 - 2343
  • [43] Object-level Attention for Aesthetic Rating Distribution Prediction
    Hou, Jingwen
    Yang, Sheng
    Lin, Weisi
    MM '20: PROCEEDINGS OF THE 28TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, 2020, : 816 - 824
  • [44] Fusion plus plus : Volumetric Object-Level SLAM
    McCormac, John
    Clark, Ronald
    Bloesch, Michael
    Davison, Andrew J.
    Leutenegger, Stefan
    2018 INTERNATIONAL CONFERENCE ON 3D VISION (3DV), 2018, : 32 - 41
  • [45] Learning Semantic Correspondence Exploiting an Object-Level Prior
    Lee, Junghyup
    Kim, Dohyung
    Lee, Wonkyung
    Ponce, Jean
    Ham, Bumsub
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (03) : 1399 - 1414
  • [46] Category-Independent Object-level Saliency Detection
    Jia, Yangqing
    Han, Mei
    2013 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2013, : 1761 - 1768
  • [47] Multi-Adversarial Partial Transfer Learning With Object-Level Attention Mechanism for Unsupervised Remote Sensing Scene Classification
    Li, Peng
    Zhang, Dezheng
    Chen, Peng
    Liu, Xin
    Wulamu, Aziguli
    IEEE ACCESS, 2020, 8 (08): : 56650 - 56665
  • [48] Object-Level Semantic Map Construction for Dynamic Scenes
    Kang, Xujie
    Li, Jing
    Fan, Xiangtao
    Jian, Hongdeng
    Xu, Chen
    APPLIED SCIENCES-BASEL, 2021, 11 (02): : 1 - 20
  • [49] Confidence judgments: The monitoring of object-level and same-level performance
    Asher Koriat
    Metacognition and Learning, 2019, 14 : 463 - 478
  • [50] Confidence judgments: The monitoring of object-level and same-level performance
    Koriat, Asher
    METACOGNITION AND LEARNING, 2019, 14 (03) : 463 - 478