Visual-inertial odometry, mapping and re-localization through learning

<p>Precise pose information is a fundamental prerequisite for numerous applications in robotics, AI and mobile computing. Monocular cameras are the ideal sensor for this purpose - they are cheap, lightweight and ubiquitous. As such, monocular visual localization is widely regarded as a corners...

Full description

Bibliographic Details
Main Author: Clark, R
Other Authors: Markham, A
Format: Thesis
Published: 2017
_version_ 1797073553575641088
author Clark, R
author2 Markham, A
author_facet Markham, A
Clark, R
author_sort Clark, R
collection OXFORD
description <p>Precise pose information is a fundamental prerequisite for numerous applications in robotics, AI and mobile computing. Monocular cameras are the ideal sensor for this purpose - they are cheap, lightweight and ubiquitous. As such, monocular visual localization is widely regarded as a cornerstone requirement of machine perception. However, a large gap still exists between the performance that these applications require and that which is achievable through existing monocular perception algorithms.</p> <p>In this thesis we directly tackle the issue of robust egocentric visual localization and mapping through a data-centric approach. As a first major contribution we propose novel learnt models for visual odometry which form the basis of the ego-motion estimates used in later chapters. The proposed approaches are less fragile and much more robust than existing approaches. We present experimental evidence that these approaches can not only approach the accuracy of standard methods but in many cases also show major improvements in computational and memory efficiency.</p> <p>To cope with the drift inherent to the odometry methods, we then introduce a novel learnt spatio-temporal model for performing global relocalization updates. The proposed approach allows one to efficiently infer the global location of an image stream at the fraction of the time of traditional feature-based approaches with minimal loss in localization accuracy.</p> <p>Finally, we present a novel SLAM system integrating our learnt priors for creating 3D maps from monocular image sequences. The approach is designed to harness multiple input sources, including prior depth and ego-motion estimates and incorporates both loop-closure and relocalization updates. The approach, based on the well-established standard visual-inertial structure-from-motion process, allows us to perform accurate posterior inference of camera poses and scene structure to significantly boost the reconstruction robustness and fidelity.</p> <p>Through our qualitative and quantitative experimentation on a wide range of datasets, we conclude that the proposed methods can bring accurate visual localization to a wide class of consumer devices and robotic platforms.</p>
first_indexed 2024-03-06T23:23:49Z
format Thesis
id oxford-uuid:69b03c50-f315-42f8-ad41-d97cd4c9bf09
institution University of Oxford
last_indexed 2024-03-06T23:23:49Z
publishDate 2017
record_format dspace
spelling oxford-uuid:69b03c50-f315-42f8-ad41-d97cd4c9bf092022-03-26T18:52:31ZVisual-inertial odometry, mapping and re-localization through learningThesishttp://purl.org/coar/resource_type/c_db06uuid:69b03c50-f315-42f8-ad41-d97cd4c9bf09ORA Deposit2017Clark, RMarkham, ATrigoni, A<p>Precise pose information is a fundamental prerequisite for numerous applications in robotics, AI and mobile computing. Monocular cameras are the ideal sensor for this purpose - they are cheap, lightweight and ubiquitous. As such, monocular visual localization is widely regarded as a cornerstone requirement of machine perception. However, a large gap still exists between the performance that these applications require and that which is achievable through existing monocular perception algorithms.</p> <p>In this thesis we directly tackle the issue of robust egocentric visual localization and mapping through a data-centric approach. As a first major contribution we propose novel learnt models for visual odometry which form the basis of the ego-motion estimates used in later chapters. The proposed approaches are less fragile and much more robust than existing approaches. We present experimental evidence that these approaches can not only approach the accuracy of standard methods but in many cases also show major improvements in computational and memory efficiency.</p> <p>To cope with the drift inherent to the odometry methods, we then introduce a novel learnt spatio-temporal model for performing global relocalization updates. The proposed approach allows one to efficiently infer the global location of an image stream at the fraction of the time of traditional feature-based approaches with minimal loss in localization accuracy.</p> <p>Finally, we present a novel SLAM system integrating our learnt priors for creating 3D maps from monocular image sequences. The approach is designed to harness multiple input sources, including prior depth and ego-motion estimates and incorporates both loop-closure and relocalization updates. The approach, based on the well-established standard visual-inertial structure-from-motion process, allows us to perform accurate posterior inference of camera poses and scene structure to significantly boost the reconstruction robustness and fidelity.</p> <p>Through our qualitative and quantitative experimentation on a wide range of datasets, we conclude that the proposed methods can bring accurate visual localization to a wide class of consumer devices and robotic platforms.</p>
spellingShingle Clark, R
Visual-inertial odometry, mapping and re-localization through learning
title Visual-inertial odometry, mapping and re-localization through learning
title_full Visual-inertial odometry, mapping and re-localization through learning
title_fullStr Visual-inertial odometry, mapping and re-localization through learning
title_full_unstemmed Visual-inertial odometry, mapping and re-localization through learning
title_short Visual-inertial odometry, mapping and re-localization through learning
title_sort visual inertial odometry mapping and re localization through learning
work_keys_str_mv AT clarkr visualinertialodometrymappingandrelocalizationthroughlearning