Tightly Coupled Semantic RGB-D Inertial Odometry for Accurate Long-Term Localization and Mapping
N. Patel, , F. Khorrami, P. Krishnamurthy, A. Tzes
Published in ICAR 2019 - December 2019
Links: ICAR 2019 page, bibtex
Abstract
In this paper, we utilize semantically enhanced feature matching and visual inertial bundle adjustment to improve the robustness of odometry especially in feature-sparse environments. A novel semantically enhanced feature matching algorithm is developed for robust: 1) medium and long-term tracking, and 2) loop-closing. Additionally, a semantic visual inertial bundle adjustment algorithm is introduced to robustly estimate pose in presence of ambiguous correspondences or in feature sparse environment. Our tightly coupled semantic RGB-D odometry approach is demonstrated on a real world indoor dataset collected using our unmanned ground vehicle (UGV). Our approach improves traditional visual odometry relying on low-level geometric features like corners, points, and planes for localization and mapping. Additionally, prior approaches are limited due to their sensitivity to scene geometry and changes in light intensity. The semantic inertial odometry is especially important to significantly reduce drifts in longer intervals.
Our proposed odometry pipeline with blocks shaded in green and purple are novel components added to the ORB-SLAM2 algorithm
Bibtex
@inproceedings{PatelKKT19,
author = {Naman Patel and
Farshad Khorrami and
Prashanth Krishnamurthy and
Anthony Tzes},
title = {Tightly Coupled Semantic RGB-D Inertial Odometry for Accurate Long-Term
Localization and Mapping},
booktitle = {Proceedings of the International Conference on Advanced Robotics, {ICAR}},
address = {Belo Horizonte, Brazil},
month = {December},
pages = {523--528},
year = {2019},
}