Skip to main content

Lund University Publications

LUND UNIVERSITY LIBRARIES

NICER-SLAM : Neural Implicit Scene Encoding for RGB SLAM

Zhu, Zihan ; Peng, Songyou ; Larsson, Viktor LU ; Cui, Zhaopeng ; Oswald, Martin R. ; Geiger, Andreas and Pollefeys, Marc (2024) 11th International Conference on 3D Vision, 3DV 2024 p.42-52
Abstract

Neural implicit representations have recently become popular in simultaneous localization and mapping (SLAM), especially in dense visual SLAM. However, existing works either rely on RGB-D sensors or require a separate monocular SLAM approach for camera tracking, and fail to produce high-fidelity 3D dense reconstructions. To address these shortcomings, we present NICER-SLAM, a dense RGB SLAM system that simultaneously optimizes for camera poses and a hierarchical neural implicit map representation, which also allows for high-quality novel view synthesis. To facilitate the optimization process for mapping, we integrate additional supervision signals including easy-to-obtain monocular geometric cues and optical flow, and also introduce a... (More)

Neural implicit representations have recently become popular in simultaneous localization and mapping (SLAM), especially in dense visual SLAM. However, existing works either rely on RGB-D sensors or require a separate monocular SLAM approach for camera tracking, and fail to produce high-fidelity 3D dense reconstructions. To address these shortcomings, we present NICER-SLAM, a dense RGB SLAM system that simultaneously optimizes for camera poses and a hierarchical neural implicit map representation, which also allows for high-quality novel view synthesis. To facilitate the optimization process for mapping, we integrate additional supervision signals including easy-to-obtain monocular geometric cues and optical flow, and also introduce a simple warping loss to further enforce geometric consistency. Moreover, to further boost performance in complex large-scale scenes, we also propose a local adaptive transformation from signed distance functions (SDFs) to density in the volume rendering equation. On multiple challenging indoor and outdoor datasets, NICER-SLAM demonstrates strong performance in dense mapping, novel view synthesis, and tracking, even competitive with recent RGB-D SLAM systems. Project page: https://nicer-slam.github.io/.

(Less)
Please use this url to cite or link to this publication:
author
; ; ; ; ; and
organization
publishing date
type
Chapter in Book/Report/Conference proceeding
publication status
published
subject
keywords
NeRF, neural implicit representation, SLAM
host publication
Proceedings - 2024 International Conference on 3D Vision, 3DV 2024
pages
11 pages
publisher
IEEE - Institute of Electrical and Electronics Engineers Inc.
conference name
11th International Conference on 3D Vision, 3DV 2024
conference location
Davos, Switzerland
conference dates
2024-03-18 - 2024-03-21
external identifiers
  • scopus:85193006230
ISBN
9798350362459
DOI
10.1109/3DV62453.2024.00096
language
English
LU publication?
yes
id
ab0d2da1-1b77-44e4-929c-b85983c1588c
date added to LUP
2025-01-16 10:06:52
date last changed
2025-04-10 09:29:48
@inproceedings{ab0d2da1-1b77-44e4-929c-b85983c1588c,
  abstract     = {{<p>Neural implicit representations have recently become popular in simultaneous localization and mapping (SLAM), especially in dense visual SLAM. However, existing works either rely on RGB-D sensors or require a separate monocular SLAM approach for camera tracking, and fail to produce high-fidelity 3D dense reconstructions. To address these shortcomings, we present NICER-SLAM, a dense RGB SLAM system that simultaneously optimizes for camera poses and a hierarchical neural implicit map representation, which also allows for high-quality novel view synthesis. To facilitate the optimization process for mapping, we integrate additional supervision signals including easy-to-obtain monocular geometric cues and optical flow, and also introduce a simple warping loss to further enforce geometric consistency. Moreover, to further boost performance in complex large-scale scenes, we also propose a local adaptive transformation from signed distance functions (SDFs) to density in the volume rendering equation. On multiple challenging indoor and outdoor datasets, NICER-SLAM demonstrates strong performance in dense mapping, novel view synthesis, and tracking, even competitive with recent RGB-D SLAM systems. Project page: https://nicer-slam.github.io/.</p>}},
  author       = {{Zhu, Zihan and Peng, Songyou and Larsson, Viktor and Cui, Zhaopeng and Oswald, Martin R. and Geiger, Andreas and Pollefeys, Marc}},
  booktitle    = {{Proceedings - 2024 International Conference on 3D Vision, 3DV 2024}},
  isbn         = {{9798350362459}},
  keywords     = {{NeRF; neural implicit representation; SLAM}},
  language     = {{eng}},
  pages        = {{42--52}},
  publisher    = {{IEEE - Institute of Electrical and Electronics Engineers Inc.}},
  title        = {{NICER-SLAM : Neural Implicit Scene Encoding for RGB SLAM}},
  url          = {{http://dx.doi.org/10.1109/3DV62453.2024.00096}},
  doi          = {{10.1109/3DV62453.2024.00096}},
  year         = {{2024}},
}