Visual-inertial odometry, mapping and re-localization through learning
<p>Precise pose information is a fundamental prerequisite for numerous applications in robotics, AI and mobile computing. Monocular cameras are the ideal sensor for this purpose - they are cheap, lightweight and ubiquitous. As such, monocular visual localization is widely regarded as a corners...
Main Author: | |
---|---|
Other Authors: | |
Format: | Thesis |
Published: |
2017
|
_version_ | 1797073553575641088 |
---|---|
author | Clark, R |
author2 | Markham, A |
author_facet | Markham, A Clark, R |
author_sort | Clark, R |
collection | OXFORD |
description | <p>Precise pose information is a fundamental prerequisite for numerous applications in robotics, AI and mobile computing. Monocular cameras are the ideal sensor for this purpose - they are cheap, lightweight and ubiquitous. As such, monocular visual localization is widely regarded as a cornerstone requirement of machine perception. However, a large gap still exists between the performance that these applications require and that which is achievable through existing monocular perception algorithms.</p> <p>In this thesis we directly tackle the issue of robust egocentric visual localization and mapping through a data-centric approach. As a first major contribution we propose novel learnt models for visual odometry which form the basis of the ego-motion estimates used in later chapters. The proposed approaches are less fragile and much more robust than existing approaches. We present experimental evidence that these approaches can not only approach the accuracy of standard methods but in many cases also show major improvements in computational and memory efficiency.</p> <p>To cope with the drift inherent to the odometry methods, we then introduce a novel learnt spatio-temporal model for performing global relocalization updates. The proposed approach allows one to efficiently infer the global location of an image stream at the fraction of the time of traditional feature-based approaches with minimal loss in localization accuracy.</p> <p>Finally, we present a novel SLAM system integrating our learnt priors for creating 3D maps from monocular image sequences. The approach is designed to harness multiple input sources, including prior depth and ego-motion estimates and incorporates both loop-closure and relocalization updates. The approach, based on the well-established standard visual-inertial structure-from-motion process, allows us to perform accurate posterior inference of camera poses and scene structure to significantly boost the reconstruction robustness and fidelity.</p> <p>Through our qualitative and quantitative experimentation on a wide range of datasets, we conclude that the proposed methods can bring accurate visual localization to a wide class of consumer devices and robotic platforms.</p> |
first_indexed | 2024-03-06T23:23:49Z |
format | Thesis |
id | oxford-uuid:69b03c50-f315-42f8-ad41-d97cd4c9bf09 |
institution | University of Oxford |
last_indexed | 2024-03-06T23:23:49Z |
publishDate | 2017 |
record_format | dspace |
spelling | oxford-uuid:69b03c50-f315-42f8-ad41-d97cd4c9bf092022-03-26T18:52:31ZVisual-inertial odometry, mapping and re-localization through learningThesishttp://purl.org/coar/resource_type/c_db06uuid:69b03c50-f315-42f8-ad41-d97cd4c9bf09ORA Deposit2017Clark, RMarkham, ATrigoni, A<p>Precise pose information is a fundamental prerequisite for numerous applications in robotics, AI and mobile computing. Monocular cameras are the ideal sensor for this purpose - they are cheap, lightweight and ubiquitous. As such, monocular visual localization is widely regarded as a cornerstone requirement of machine perception. However, a large gap still exists between the performance that these applications require and that which is achievable through existing monocular perception algorithms.</p> <p>In this thesis we directly tackle the issue of robust egocentric visual localization and mapping through a data-centric approach. As a first major contribution we propose novel learnt models for visual odometry which form the basis of the ego-motion estimates used in later chapters. The proposed approaches are less fragile and much more robust than existing approaches. We present experimental evidence that these approaches can not only approach the accuracy of standard methods but in many cases also show major improvements in computational and memory efficiency.</p> <p>To cope with the drift inherent to the odometry methods, we then introduce a novel learnt spatio-temporal model for performing global relocalization updates. The proposed approach allows one to efficiently infer the global location of an image stream at the fraction of the time of traditional feature-based approaches with minimal loss in localization accuracy.</p> <p>Finally, we present a novel SLAM system integrating our learnt priors for creating 3D maps from monocular image sequences. The approach is designed to harness multiple input sources, including prior depth and ego-motion estimates and incorporates both loop-closure and relocalization updates. The approach, based on the well-established standard visual-inertial structure-from-motion process, allows us to perform accurate posterior inference of camera poses and scene structure to significantly boost the reconstruction robustness and fidelity.</p> <p>Through our qualitative and quantitative experimentation on a wide range of datasets, we conclude that the proposed methods can bring accurate visual localization to a wide class of consumer devices and robotic platforms.</p> |
spellingShingle | Clark, R Visual-inertial odometry, mapping and re-localization through learning |
title | Visual-inertial odometry, mapping and re-localization through learning |
title_full | Visual-inertial odometry, mapping and re-localization through learning |
title_fullStr | Visual-inertial odometry, mapping and re-localization through learning |
title_full_unstemmed | Visual-inertial odometry, mapping and re-localization through learning |
title_short | Visual-inertial odometry, mapping and re-localization through learning |
title_sort | visual inertial odometry mapping and re localization through learning |
work_keys_str_mv | AT clarkr visualinertialodometrymappingandrelocalizationthroughlearning |