Object-level Scene Deocclusion

被引:0
|
作者
Liu, Zhengzhe [1 ]
Liu, Qing [2 ]
Chang, Chirui [3 ]
Zhang, Jianming [2 ]
Pakhomov, Daniil [2 ]
Zheng, Haitian [2 ]
Lin, Zhe [2 ]
Cohen-Or, Daniel [4 ]
Fu, Chi-Wing [1 ]
机构
[1] Chinese Univ Hong Kong, Hong Kong, Peoples R China
[2] Adobe, San Jose, CA USA
[3] Univ Hong Kong, Hong Kong, Peoples R China
[4] Tel Aviv Univ, Tel Aviv, Israel
关键词
scene deocclusion; object completion; image recomposition;
D O I
10.1145/3641519.3657409
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deoccluding the hidden portions of objects in a scene is a formidable task, particularly when addressing real-world scenes. In this paper, we present a new self-supervised PArallel visible-to-COmplete diffusion framework, named PACO, a foundation model for object-level scene deocclusion. Leveraging the rich prior of pre-trained models, we first design the parallel variational autoencoder, which produces a full-view feature map that simultaneously encodes multiple complete objects, and the visible-to-complete latent generator, which learns to implicitly predict the full-view feature map from partial-view feature map and text prompts extracted from the incomplete objects in the input image. To train PACO, we create a large-scale dataset with 500k samples to enable self-supervised learning, avoiding tedious annotations of the amodal masks and occluded regions. At inference, we devise a layer-wise deocclusion strategy to improve efficiency while maintaining the deocclusion quality. Extensive experiments on COCOA and various real-world scenes demonstrate the superior capability of PACO for scene deocclusion, surpassing the state of the arts by a large margin. Our method can also be extended to cross-domain scenes and novel categories that are not covered by the training set. Further, we demonstrate the deocclusion applicability of PACO in single-view 3D scene reconstruction and object recomposition. Project page: https://liuzhengzhe.github.io/Deocclude-Any-Object.github.io/.
引用
收藏
页数:11
相关论文
共 50 条
  • [31] Salient Object Detection via Fuzzy Theory and Object-Level Enhancement
    Zhou, Yuan
    Mao, Ailing
    Huo, Shuwei
    Lei, Jianjun
    Kung, Sun-Yuan
    IEEE TRANSACTIONS ON MULTIMEDIA, 2019, 21 (01) : 74 - 85
  • [32] Lightweight object-level monocular data association system
    Peng, Xinyi
    Chen, Guo
    2023 35TH CHINESE CONTROL AND DECISION CONFERENCE, CCDC, 2023, : 4331 - 4336
  • [33] Object-Level Reasoning with Logics Encoded in HOL Light
    Papapanagiotou, Petros
    Fleuriot, Jacques
    ELECTRONIC PROCEEDINGS IN THEORETICAL COMPUTER SCIENCE, 2021, (332): : 18 - 34
  • [34] Bayesian object-level change detection in grayscale imagery
    Perera, AGA
    Hoogs, A
    PROCEEDINGS OF THE 17TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION, VOL 1, 2004, : 71 - 75
  • [35] Semantic SLAM with Autonomous Object-Level Data Association
    Qian, Zhentian
    Patath, Kartik
    Fu, Jie
    Xiao, Jing
    2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2021), 2021, : 11203 - 11209
  • [36] ROLS : Robust Object-level SLAM for grape counting
    Nellithimaru, Anjana K.
    Kantor, George A.
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW 2019), 2019, : 2648 - 2656
  • [37] EXPERIMENTAL OBJECT-LEVEL STRATEGIC CONTROL WITH COOPERATING MANIPULATORS
    SCHNEIDER, SA
    CANNON, RH
    INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 1993, 12 (04): : 338 - 350
  • [38] INERTIAL PROPERTIES IN ROBOTIC MANIPULATION - AN OBJECT-LEVEL FRAMEWORK
    KHATIB, O
    INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 1995, 14 (01): : 19 - 36
  • [39] Online object-level SLAM with dual bundle adjustment
    Jiaqi Liu
    Yongbin Gao
    Xiaoyan Jiang
    Zhijun Fang
    Applied Intelligence, 2023, 53 : 25092 - 25105
  • [40] Visual Grounding for Object-Level Generalization in Reinforcement Learning
    Jiang, Haobin
    Lu, Zongqing
    COMPUTER VISION - ECCV 2024, PT XXX, 2025, 15088 : 55 - 72