kth.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Self-Supervised 3D Keypoint Learning for Ego-Motion Estimation
KTH, School of Electrical Engineering and Computer Science (EECS), Intelligent systems, Robotics, Perception and Learning, RPL. Toyota Research Institute, Toyota Research Institute.
Toyota Research Institute, Toyota Research Institute.
Toyota Research Institute, Toyota Research Institute.
Toyota Research Institute, Toyota Research Institute.
Show others and affiliations
2020 (English)In: Proceedings of the 2020 Conference on Robot Learning, CoRL 2020, ML Research Press , 2020, p. 2085-2103Conference paper, Published paper (Refereed)
Abstract [en]

Detecting and matching robust viewpoint-invariant keypoints is critical for visual SLAM and Structure-from-Motion. State-of-the-art learning-based methods generate training samples via homography adaptation to create 2D synthetic views with known keypoint matches from a single image. This approach, however, does not generalize to non-planar 3D scenes with illumination variations commonly seen in real-world videos. In this work, we propose self-supervised learning of depth-aware keypoints directly from unlabeled videos. We jointly learn keypoint and depth estimation networks by combining appearance and geometric matching via a differentiable structure-from-motion module based on Procrustean residual pose correction. We describe how our self-supervised keypoints can be integrated into state-of-the-art visual odometry frameworks for robust and accurate ego-motion estimation of autonomous vehicles in real-world conditions.

Place, publisher, year, edition, pages
ML Research Press , 2020. p. 2085-2103
Keywords [en]
Keypoints, Monocular, Self-supervised-learning, Visual odometry
National Category
Computer graphics and computer vision Robotics and automation
Identifiers
URN: urn:nbn:se:kth:diva-339686Scopus ID: 2-s2.0-85175858250OAI: oai:DiVA.org:kth-339686DiVA, id: diva2:1812470
Conference
4th Conference on Robot Learning, CoRL 2020, Virtual/Online, United States of America, Nov 16 2020 - Nov 18 2020
Note

QC 20231116

Available from: 2023-11-16 Created: 2023-11-16 Last updated: 2025-02-05Bibliographically approved

Open Access in DiVA

No full text in DiVA

Scopus

Authority records

Tang, JiexiongJensfelt, Patric

Search in DiVA

By author/editor
Tang, JiexiongJensfelt, Patric
By organisation
Robotics, Perception and Learning, RPL
Computer graphics and computer visionRobotics and automation

Search outside of DiVA

GoogleGoogle Scholar

urn-nbn

Altmetric score

urn-nbn
Total: 42 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf