Self-Supervised 3D Keypoint Learning for Ego-Motion EstimationShow others and affiliations
2020 (English)In: Proceedings of the 2020 Conference on Robot Learning, CoRL 2020, ML Research Press , 2020, p. 2085-2103Conference paper, Published paper (Refereed)
Abstract [en]
Detecting and matching robust viewpoint-invariant keypoints is critical for visual SLAM and Structure-from-Motion. State-of-the-art learning-based methods generate training samples via homography adaptation to create 2D synthetic views with known keypoint matches from a single image. This approach, however, does not generalize to non-planar 3D scenes with illumination variations commonly seen in real-world videos. In this work, we propose self-supervised learning of depth-aware keypoints directly from unlabeled videos. We jointly learn keypoint and depth estimation networks by combining appearance and geometric matching via a differentiable structure-from-motion module based on Procrustean residual pose correction. We describe how our self-supervised keypoints can be integrated into state-of-the-art visual odometry frameworks for robust and accurate ego-motion estimation of autonomous vehicles in real-world conditions.
Place, publisher, year, edition, pages
ML Research Press , 2020. p. 2085-2103
Keywords [en]
Keypoints, Monocular, Self-supervised-learning, Visual odometry
National Category
Computer graphics and computer vision Robotics and automation
Identifiers
URN: urn:nbn:se:kth:diva-339686Scopus ID: 2-s2.0-85175858250OAI: oai:DiVA.org:kth-339686DiVA, id: diva2:1812470
Conference
4th Conference on Robot Learning, CoRL 2020, Virtual/Online, United States of America, Nov 16 2020 - Nov 18 2020
Note
QC 20231116
2023-11-162023-11-162025-02-05Bibliographically approved