BOR2G: Building Optimal Regularised Reconstructions with GPUs (in cubes)

<p>Robots require high-quality <em>maps</em>—internal representations of their operating workspace—to localise, path plan, and perceive their environment. Until recently, these maps were restricted to sparse, 2D representations due to computational, memory, and sensor limitations....

Szczegółowa specyfikacja

Opis bibliograficzny
1. autor: Tanner, M
Kolejni autorzy: Newman, P
Format: Praca dyplomowa
Język:English
Wydane: 2017
Hasła przedmiotowe:
_version_ 1826317371519795200
author Tanner, M
author2 Newman, P
author_facet Newman, P
Tanner, M
author_sort Tanner, M
collection OXFORD
description <p>Robots require high-quality <em>maps</em>—internal representations of their operating workspace—to localise, path plan, and perceive their environment. Until recently, these maps were restricted to sparse, 2D representations due to computational, memory, and sensor limitations. With the widespread adoption of high-quality sensors and graphics processors for parallel processing, these restrictions no longer apply: dense 3D maps are feasible to compute in real time (i.e., at the input sensor’s frame rate).</p> <p>This thesis presents the theory and system to create large-scale dense 3D maps (i.e., reconstruct continuous surface models) using only sensors found on modern autonomous automobiles: 2D laser, 3D laser, and cameras. In contrast to active RGB-D cameras, passive cameras produce noisy surface observations and must be regularised in both 2D and 3D to create accurate reconstructions. Unfortunately, straight-forward application of 3D regularisation causes undesired surface interpolation and extrapolation in regions unexplored by the robot. We propose a method to overcome this challenge by informing the regulariser of the specific subsets of 3D surfaces upon which to operate. When combined with a compressed voxel grid data structure, we demonstrate our system fusing data from both laser and camera sensors to reconstruct 7.3 km of urban environments. We evaluate the quantitative performance of our proposed method through the use of synthetic and real-world datasets—including datasets from Stanford's Burghers of Calais, University of Oxford's RobotCar, University of Oxford’s Dense Reconstruction, and Karlsruhe Institute of Technology’s KITTI—compared to ground-truth laser data. With only stereo camera inputs, our regulariser reduces the 3D reconstruction metric error between 27% to 36% with a final median accuracy ranging between 4 cm to 8 cm.</p> <p>Furthermore, by augmenting our system with object detection, we remove ephemeral objects (e.g., automobiles, bicycles, and pedestrians) from the input sensor data and target our regulariser to interpolate the occluded urban surfaces. Augmented with Kernel Conditional Density Estimation, our regulariser creates reconstructions with median errors between 5.64 cm and 9.24 cm.</p> <p>Finally, we present a machine-learning pipeline that learns, in an automatic fashion, to recognise the errors in dense reconstructions. Our system trains on image and laser data from a 3.8 km urban sequence. Using a separate 2.2 km urban sequence, our pipeline consistently identifies error-prone regions in the image-based dense reconstruction.</p>
first_indexed 2024-03-06T19:18:13Z
format Thesis
id oxford-uuid:1928c996-d913-4d7e-8ca5-cf247f90aa0f
institution University of Oxford
language English
last_indexed 2025-02-19T04:37:24Z
publishDate 2017
record_format dspace
spelling oxford-uuid:1928c996-d913-4d7e-8ca5-cf247f90aa0f2025-02-11T16:11:04ZBOR2G: Building Optimal Regularised Reconstructions with GPUs (in cubes)Thesishttp://purl.org/coar/resource_type/c_db06uuid:1928c996-d913-4d7e-8ca5-cf247f90aa0foptimisationautonomous vehiclesroboticsmachine learningartificial intelligencemappingEnglishORA Deposit2017Tanner, MNewman, P<p>Robots require high-quality <em>maps</em>—internal representations of their operating workspace—to localise, path plan, and perceive their environment. Until recently, these maps were restricted to sparse, 2D representations due to computational, memory, and sensor limitations. With the widespread adoption of high-quality sensors and graphics processors for parallel processing, these restrictions no longer apply: dense 3D maps are feasible to compute in real time (i.e., at the input sensor’s frame rate).</p> <p>This thesis presents the theory and system to create large-scale dense 3D maps (i.e., reconstruct continuous surface models) using only sensors found on modern autonomous automobiles: 2D laser, 3D laser, and cameras. In contrast to active RGB-D cameras, passive cameras produce noisy surface observations and must be regularised in both 2D and 3D to create accurate reconstructions. Unfortunately, straight-forward application of 3D regularisation causes undesired surface interpolation and extrapolation in regions unexplored by the robot. We propose a method to overcome this challenge by informing the regulariser of the specific subsets of 3D surfaces upon which to operate. When combined with a compressed voxel grid data structure, we demonstrate our system fusing data from both laser and camera sensors to reconstruct 7.3 km of urban environments. We evaluate the quantitative performance of our proposed method through the use of synthetic and real-world datasets—including datasets from Stanford's Burghers of Calais, University of Oxford's RobotCar, University of Oxford’s Dense Reconstruction, and Karlsruhe Institute of Technology’s KITTI—compared to ground-truth laser data. With only stereo camera inputs, our regulariser reduces the 3D reconstruction metric error between 27% to 36% with a final median accuracy ranging between 4 cm to 8 cm.</p> <p>Furthermore, by augmenting our system with object detection, we remove ephemeral objects (e.g., automobiles, bicycles, and pedestrians) from the input sensor data and target our regulariser to interpolate the occluded urban surfaces. Augmented with Kernel Conditional Density Estimation, our regulariser creates reconstructions with median errors between 5.64 cm and 9.24 cm.</p> <p>Finally, we present a machine-learning pipeline that learns, in an automatic fashion, to recognise the errors in dense reconstructions. Our system trains on image and laser data from a 3.8 km urban sequence. Using a separate 2.2 km urban sequence, our pipeline consistently identifies error-prone regions in the image-based dense reconstruction.</p>
spellingShingle optimisation
autonomous vehicles
robotics
machine learning
artificial intelligence
mapping
Tanner, M
BOR2G: Building Optimal Regularised Reconstructions with GPUs (in cubes)
title BOR2G: Building Optimal Regularised Reconstructions with GPUs (in cubes)
title_full BOR2G: Building Optimal Regularised Reconstructions with GPUs (in cubes)
title_fullStr BOR2G: Building Optimal Regularised Reconstructions with GPUs (in cubes)
title_full_unstemmed BOR2G: Building Optimal Regularised Reconstructions with GPUs (in cubes)
title_short BOR2G: Building Optimal Regularised Reconstructions with GPUs (in cubes)
title_sort bor2g building optimal regularised reconstructions with gpus in cubes
topic optimisation
autonomous vehicles
robotics
machine learning
artificial intelligence
mapping
work_keys_str_mv AT tannerm bor2gbuildingoptimalregularisedreconstructionswithgpusincubes