TY - JOUR
T1 - Structerf-SLAM
T2 - Neural implicit representation SLAM for structural environments
AU - Wang, Haocheng
AU - Cao, Yanlong
AU - Wei, Xiaoyao
AU - Shou, Yejun
AU - Shen, Lingfeng
AU - Xu, Zhijie
AU - Ren, Kai
N1 - Funding Information:
Research supported by Foundation of Science and Technology Innovation leading talent project of special support plan for high-level talents of Zhejiang Province, China ( 2022R52053 ), Open Fund of Anhui Key Laboratory of Mine Intelligent Equipment and Technology, China ( KSZN202001001 ).
Publisher Copyright:
© 2024
PY - 2024/4/1
Y1 - 2024/4/1
N2 - In recent years, research on simultaneous localization and mapping (SLAM) using neural implicit representation has shown promising outcomes due to its smooth mapping and low memory consumption, particularly suitable for structured environments with limited boundaries. However, there is currently no implicit SLAM that can effectively utilize prior structural constraints to accurately build 3D maps. In this study, we propose an RGB-D dense tracking and mapping approach, Structerf-SLAM, that combines visual odometry with neural implicit representation. Our scene representation consists of dual-layer feature grids and pre-trained decoders that decode the interpolated features into RGB and depth values. Moreover, structured planar constraints are integrated. In the tracking stage, utilizing the three-dimensional plane features under the Manhattan assumption achieves more stable and rapid data association, consequently resolving the tracking misalignment issue in textureless regions (e.g., floor, wall, etc.). In the mapping stage, by enforcing planar consistency, the depth predicted by the neural radiation field is well-fitted by a plane, resulting in smoother and more realistic map reconstruction. Experiments on synthetic and real scene datasets demonstrate competitive results of Structerf-SLAM in both mapping and tracking quality.
AB - In recent years, research on simultaneous localization and mapping (SLAM) using neural implicit representation has shown promising outcomes due to its smooth mapping and low memory consumption, particularly suitable for structured environments with limited boundaries. However, there is currently no implicit SLAM that can effectively utilize prior structural constraints to accurately build 3D maps. In this study, we propose an RGB-D dense tracking and mapping approach, Structerf-SLAM, that combines visual odometry with neural implicit representation. Our scene representation consists of dual-layer feature grids and pre-trained decoders that decode the interpolated features into RGB and depth values. Moreover, structured planar constraints are integrated. In the tracking stage, utilizing the three-dimensional plane features under the Manhattan assumption achieves more stable and rapid data association, consequently resolving the tracking misalignment issue in textureless regions (e.g., floor, wall, etc.). In the mapping stage, by enforcing planar consistency, the depth predicted by the neural radiation field is well-fitted by a plane, resulting in smoother and more realistic map reconstruction. Experiments on synthetic and real scene datasets demonstrate competitive results of Structerf-SLAM in both mapping and tracking quality.
KW - Neural Implicit Representation
KW - Visual SLAM
KW - Implicit Scene Reconstruction
KW - Structural Constraints
KW - Neural implicit representation
KW - Implicit scene reconstruction
KW - Structural constraints
UR - http://www.scopus.com/inward/record.url?scp=85186330095&partnerID=8YFLogxK
U2 - 10.1016/j.cag.2024.103893
DO - 10.1016/j.cag.2024.103893
M3 - Article
VL - 119
JO - Computers and Graphics (Pergamon)
JF - Computers and Graphics (Pergamon)
SN - 0097-8493
M1 - 103893
ER -