SLAM-aware, self-supervised perception in mobile robots

Thesis: Ph. D., Massachusetts Institute of Technology, Department of Electrical Engineering and Computer Science, 2017.

Bibliographic Details
Main Author: Pillai, Sudeep
Other Authors: John J. Leonard.
Format: Thesis
Language:eng
Published: Massachusetts Institute of Technology 2018
Subjects:
Online Access:http://hdl.handle.net/1721.1/114054
_version_ 1826207883417616384
author Pillai, Sudeep
author2 John J. Leonard.
author_facet John J. Leonard.
Pillai, Sudeep
author_sort Pillai, Sudeep
collection MIT
description Thesis: Ph. D., Massachusetts Institute of Technology, Department of Electrical Engineering and Computer Science, 2017.
first_indexed 2024-09-23T13:56:28Z
format Thesis
id mit-1721.1/114054
institution Massachusetts Institute of Technology
language eng
last_indexed 2024-09-23T13:56:28Z
publishDate 2018
publisher Massachusetts Institute of Technology
record_format dspace
spelling mit-1721.1/1140542019-04-10T09:00:27Z SLAM-aware, self-supervised perception in mobile robots Simultaneous Localization and Mapping- aware, self-supervised perception in mobile robots Pillai, Sudeep John J. Leonard. Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science. Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science. Electrical Engineering and Computer Science. Thesis: Ph. D., Massachusetts Institute of Technology, Department of Electrical Engineering and Computer Science, 2017. This electronic version was submitted by the student author. The certified thesis is available in the Institute Archives and Special Collections. Cataloged from student-submitted PDF version of thesis. Includes bibliographical references (pages 152-171). Simultaneous Localization and Mapping (SLAM) is a fundamental capability in mobile robots, and has been typically considered in the context of aiding mapping and navigation tasks. In this thesis, we advocate for the use of SLAM as a supervisory signal to further the perceptual capabilities in robots. Through the concept of SLAM-supported object recognition, we develop the ability for robots equipped with a single camera to be able to leverage their SLAM-awareness (via Monocular Visual-SLAM) to better inform object recognition within its immediate environment. Additionally, by maintaining a spatially-cognizant view of the world,we find our SLAM-aware approach to be particularly amenable to few-shot object learning. We show that a SLAM-aware, few-shot object learning strategy can be especially advantageous to mobile robots, and is able to learn object detectors from a reduced set of training examples. Implicit to realizing modern visual-SLAM systems is its choice of map representation. It is imperative that the map representation is crucially utilized by multiple components in the robot's decision-making stack, while it is constantly optimized as more measurements are available. Motivated by the need for a unified map representation in vision-based mapping, navigation and planning, we develop an iterative and high-performance mesh-reconstruction algorithm for stereo imagery. We envision that in the future, these tunable mesh representations can potentially enable robots to quickly reconstruct their immediate surroundings while being able to directly plan in them and maneuver at high-speeds. While most visual-SLAM front-ends explicitly encode application-specific constraints for accurate and robust operation, we advocate for an automated solution to developing these systems. By bootstrapping the robot's ability to perform GP Saided SLAM, we develop a self-supervised visual-Slam front-end capable of performing visual ego-motion, and vision-based loop-closure recognition in mobile robots. We propose a novel, generative model solution that it is able to predict ego-motion estimates from optical flow, while also allowing for the prediction of induced scene flow conditioned on the ego-motion. Following a similar bootstrapped learning strategy, we explore the ability to self-supervise place recognition in mobile robots and cast it as a metric learning problem, with a GPS-aided SLAM solution providing the relevant supervision. Furthermore, we show that the newly learned embedding can be particularly powerful in discriminating visual scene instances from each other for the purpose of loop-closure detection. We envision that such self-supervised solutions to vision-based task learning will have far-reaching implications in several domains, especially facilitating life-long learning in autonomous systems. by Sudeep Pillai. Ph. D. 2018-03-12T18:53:10Z 2018-03-12T18:53:10Z 2017 2017 Thesis http://hdl.handle.net/1721.1/114054 1027217486 eng MIT theses are protected by copyright. They may be viewed, downloaded, or printed from this source but further reproduction or distribution in any format is prohibited without written permission. http://dspace.mit.edu/handle/1721.1/7582 171 pages application/pdf Massachusetts Institute of Technology
spellingShingle Electrical Engineering and Computer Science.
Pillai, Sudeep
SLAM-aware, self-supervised perception in mobile robots
title SLAM-aware, self-supervised perception in mobile robots
title_full SLAM-aware, self-supervised perception in mobile robots
title_fullStr SLAM-aware, self-supervised perception in mobile robots
title_full_unstemmed SLAM-aware, self-supervised perception in mobile robots
title_short SLAM-aware, self-supervised perception in mobile robots
title_sort slam aware self supervised perception in mobile robots
topic Electrical Engineering and Computer Science.
url http://hdl.handle.net/1721.1/114054
work_keys_str_mv AT pillaisudeep slamawareselfsupervisedperceptioninmobilerobots
AT pillaisudeep simultaneouslocalizationandmappingawareselfsupervisedperceptioninmobilerobots