NICER-SLAM: Neural Implicit Scene Encoding for RGB SLAM

被引:22
|
作者
Zhu, Zihan [1 ]
Peng, Songyou [1 ,2 ]
Larsson, Viktor [3 ]
Cui, Zhaopeng [4 ]
Oswald, Martin R. [1 ,5 ]
Geiger, Andreas [6 ]
Pollefeys, Marc [1 ,7 ]
机构
[1] Swiss Fed Inst Technol, Zurich, Switzerland
[2] MPI Intelligent Syst, Tubingen, Germany
[3] Lund Univ, Lund, Sweden
[4] Zhejiang Univ, State Key Lab CAD & CG, Hangzhou, Peoples R China
[5] Univ Amsterdam, Amsterdam, Netherlands
[6] Univ Tubingen, Tubingen AI Ctr, Tubingen, Germany
[7] Microsoft, Redmond, WA USA
关键词
D O I
10.1109/3DV62453.2024.00096
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Neural implicit representations have recently become popular in simultaneous localization and mapping (SLAM), especially in dense visual SLAM. However, existing works either rely on RGB-D sensors or require a separate monocular SLAM approach for camera tracking, and fail to produce high-fidelity 3D dense reconstructions. To address these shortcomings, we present NICER-SLAM, a dense RGB SLAM system that simultaneously optimizes for camera poses and a hierarchical neural implicit map representation, which also allows for high-quality novel view synthesis. To facilitate the optimization process for mapping, we integrate additional supervision signals including easy-to-obtain monocular geometric cues and optical flow, and also introduce a simple warping loss to further enforce geometric consistency. Moreover, to further boost performance in complex large-scale scenes, we also propose a local adaptive transformation from signed distance functions (SDFs) to density in the volume rendering equation. On multiple challenging indoor and outdoor datasets, NICER-SLAM demonstrates strong performance in dense mapping, novel view synthesis, and tracking, even competitive with recent RGB-D SLAM systems. Project page: https:// nicer-slam.github.io/
引用
收藏
页码:42 / 52
页数:11
相关论文
共 50 条
  • [21] RGB-D SLAM method of dynamic scene based on instance segmentation and optical flow
    Wang C.
    Shi J.
    Zhu H.
    Bai S.
    Sun Y.
    Lu J.
    Huang S.
    Guangxue Jingmi Gongcheng/Optics and Precision Engineering, 2024, 32 (06): : 857 - 867
  • [22] Low-Drift RGB-D SLAM with Room Reconstruction Using Scene Understanding
    Ye, Zefeng
    Jiang, Xin
    Liu, Yunhui
    2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND BIOMIMETICS (IEEE-ROBIO 2021), 2021, : 808 - 813
  • [23] Robust RGB-D SLAM Using Point and Line Features for Low Textured Scene
    Zou, Yajing
    Eldemiry, Amr
    Li, Yaxin
    Chen, Wu
    SENSORS, 2020, 20 (17) : 1 - 20
  • [24] Multi-robot collaborative SLAM and scene reconstruction based on RGB-D camera
    Ma, Tianyun
    Zhang, Tao
    Li, Shaopeng
    2020 CHINESE AUTOMATION CONGRESS (CAC 2020), 2020, : 139 - 144
  • [25] Incremental Scene Understanding on Dense SLAM
    Li, Chi
    Xiao, Han
    Tateno, Keisuke
    Tombari, Federico
    Navab, Nassir
    Hager, Gregory D.
    2016 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS 2016), 2016, : 574 - 581
  • [26] DGFlow-SLAM: A Novel Dynamic Environment RGB-D SLAM without Prior Semantic Knowledge Based on Grid Segmentation of Scene Flow
    Long, Fei
    Ding, Lei
    Li, Jianfeng
    BIOMIMETICS, 2022, 7 (04)
  • [27] PIN-SLAM: LiDAR SLAM Using a Point-Based Implicit Neural Representation for Achieving Global Map Consistency
    Pan, Yue
    Zhong, Xingguang
    Wiesmann, Louis
    Posewsky, Thorbjorn
    Behley, Jens
    Stachniss, Cyrill
    IEEE TRANSACTIONS ON ROBOTICS, 2024, 40 : 4045 - 4064
  • [28] A Combined RGB and Depth Descriptor for SLAM with Humanoids
    Sheikh, Rasha
    Osswald, Stefan
    Bennewitz, Maren
    2018 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2018, : 1718 - 1724
  • [29] Survey and Evaluation of RGB-D SLAM
    Zhang, Shishun
    Zheng, Longyu
    Tao, Wenbing
    IEEE ACCESS, 2021, 9 : 21367 - 21387
  • [30] Visual SLAM with RGB-D Cameras
    Jin, Qiongyao
    Liu, Yungang
    Man, Yongchao
    Li, Fengzhong
    PROCEEDINGS OF THE 38TH CHINESE CONTROL CONFERENCE (CCC), 2019, : 4072 - 4077