Abstract

This paper presents a method for stereo visual odometry and mapping that integrates VINS-Fusion-based visual odometry estimation with deep learning techniques for camera pose tracking and stereo image matching. Traditional approaches in the VINS-Fusion relied on classical methods for feature extraction and matching, which often resulted in inaccuracies in triangulation-based 3D position estimation. These inaccuracies could be mitigated by incorporating IMU-based position estimation, which yielded more accurate odometry estimates compared to using stereo camera only in three-dimensional space. Consequently, the original VINS-stereo algorithm necessitated a tightly-coupled integration of IMU sensor measurements with estimated visual odometry.
To address these challenges, our work proposes replacing the traditional feature extraction method used in VINS-Fusion, the Shi-Tomasi (Good Features to Track) technique, with feature extraction via the SuperPoint deep network. This approach has demonstrated promising experimental results. Additionally, we have applied deep learning models to the matching of feature points that project the same three-dimensional point to pixel coordinates in different images. Instead of using the KLT optical flow algorithm previously employed by VINS-Fusion, our proposed method utilizes SuperGlue, a deep graph neural network for graph matching, to improve image tracking and stereo image matching performance. The performance of the proposed algorithm is evaluated using the publicly available EuRoC dataset, providing a comparison with existing algorithms.

Details

Title
Stereo Vision SLAM with SuperPoint and SuperGlue
Author
Si-Won, Yoon 1   VIAFID ORCID Logo  ; Park, Soon-Yong 1 

 Graduate School of Electronic and Electrical Engineering, Kyungpook National University, Daegu 41566, South Korea; Graduate School of Electronic and Electrical Engineering, Kyungpook National University, Daegu 41566, South Korea 
Pages
183-188
Publication year
2024
Publication date
2024
Publisher
Copernicus GmbH
ISSN
16821750
e-ISSN
21949034
Source type
Conference Paper
Language of publication
English
ProQuest document ID
3073179730
Copyright
© 2024. This work is published under https://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.