FastDepth: Fast monocular depth estimation on embedded systems

Depth sensing is a critical function for robotic tasks such as localization, mapping and obstacle detection. There has been a significant and growing interest in depth estimation from a single RGB image, due to the relatively low cost and size of monocular cameras. However, state-of-the-art single-v...

Full description

Bibliographic Details
Main Authors: Wofk, Diana, Ma, Fangchang, Yang, Tien-Ju, Karaman, Sertac, Sze, Vivienne
Other Authors: Massachusetts Institute of Technology. Department of Aeronautics and Astronautics
Format: Article
Language:English
Published: IEEE 2020
Online Access:https://hdl.handle.net/1721.1/126546
_version_ 1811087480341921792
author Wofk, Diana
Ma, Fangchang
Yang, Tien-Ju
Karaman, Sertac
Sze, Vivienne
author2 Massachusetts Institute of Technology. Department of Aeronautics and Astronautics
author_facet Massachusetts Institute of Technology. Department of Aeronautics and Astronautics
Wofk, Diana
Ma, Fangchang
Yang, Tien-Ju
Karaman, Sertac
Sze, Vivienne
author_sort Wofk, Diana
collection MIT
description Depth sensing is a critical function for robotic tasks such as localization, mapping and obstacle detection. There has been a significant and growing interest in depth estimation from a single RGB image, due to the relatively low cost and size of monocular cameras. However, state-of-the-art single-view depth estimation algorithms are based on fairly complex deep neural networks that are too slow for real-time inference on an embedded platform, for instance, mounted on a micro aerial vehicle. In this paper, we address the problem of fast depth estimation on embedded systems. We propose an efficient and lightweight encoder-decoder network architecture and apply network pruning to further reduce computational complexity and latency. In particular, we focus on the design of a low-latency decoder. Our methodology demonstrates that it is possible to achieve similar accuracy as prior work on depth estimation, but at inference speeds that are an order of magnitude faster. Our proposed network, FastDepth, runs at 178 fps on an NVIDIA Jetson TX2 GPU and at 27 fps when using only the TX2 CPU, with active power consumption under 10 W. FastDepth achieves close to state-of-the-art accuracy on the NYU Depth v2 dataset. To the best of the authors' knowledge, this paper demonstrates real-time monocular depth estimation using a deep neural network with the lowest latency and highest throughput on an embedded platform that can be carried by a micro aerial vehicle.
first_indexed 2024-09-23T13:46:46Z
format Article
id mit-1721.1/126546
institution Massachusetts Institute of Technology
language English
last_indexed 2024-09-23T13:46:46Z
publishDate 2020
publisher IEEE
record_format dspace
spelling mit-1721.1/1265462022-10-01T17:06:34Z FastDepth: Fast monocular depth estimation on embedded systems Wofk, Diana Ma, Fangchang Yang, Tien-Ju Karaman, Sertac Sze, Vivienne Massachusetts Institute of Technology. Department of Aeronautics and Astronautics Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science Massachusetts Institute of Technology. Microsystems Technology Laboratories Depth sensing is a critical function for robotic tasks such as localization, mapping and obstacle detection. There has been a significant and growing interest in depth estimation from a single RGB image, due to the relatively low cost and size of monocular cameras. However, state-of-the-art single-view depth estimation algorithms are based on fairly complex deep neural networks that are too slow for real-time inference on an embedded platform, for instance, mounted on a micro aerial vehicle. In this paper, we address the problem of fast depth estimation on embedded systems. We propose an efficient and lightweight encoder-decoder network architecture and apply network pruning to further reduce computational complexity and latency. In particular, we focus on the design of a low-latency decoder. Our methodology demonstrates that it is possible to achieve similar accuracy as prior work on depth estimation, but at inference speeds that are an order of magnitude faster. Our proposed network, FastDepth, runs at 178 fps on an NVIDIA Jetson TX2 GPU and at 27 fps when using only the TX2 CPU, with active power consumption under 10 W. FastDepth achieves close to state-of-the-art accuracy on the NYU Depth v2 dataset. To the best of the authors' knowledge, this paper demonstrates real-time monocular depth estimation using a deep neural network with the lowest latency and highest throughput on an embedded platform that can be carried by a micro aerial vehicle. 2020-08-12T17:45:42Z 2020-08-12T17:45:42Z 2019-05 2019-10-29T15:58:10Z Article http://purl.org/eprint/type/ConferencePaper https://hdl.handle.net/1721.1/126546 Wofk, Diana et al. “FastDepth: Fast monocular depth estimation on embedded systems.” Paper presented at the 2019 International Conference on Robotics and Automation (ICRA), Montreal, Canada, May 20-24, 2019, © 2019 The Author(s) en 10.1109/ICRA.2019.8794182 2019 International Conference on Robotics and Automation (ICRA) Creative Commons Attribution-Noncommercial-Share Alike http://creativecommons.org/licenses/by-nc-sa/4.0/ application/pdf IEEE arXiv
spellingShingle Wofk, Diana
Ma, Fangchang
Yang, Tien-Ju
Karaman, Sertac
Sze, Vivienne
FastDepth: Fast monocular depth estimation on embedded systems
title FastDepth: Fast monocular depth estimation on embedded systems
title_full FastDepth: Fast monocular depth estimation on embedded systems
title_fullStr FastDepth: Fast monocular depth estimation on embedded systems
title_full_unstemmed FastDepth: Fast monocular depth estimation on embedded systems
title_short FastDepth: Fast monocular depth estimation on embedded systems
title_sort fastdepth fast monocular depth estimation on embedded systems
url https://hdl.handle.net/1721.1/126546
work_keys_str_mv AT wofkdiana fastdepthfastmonoculardepthestimationonembeddedsystems
AT mafangchang fastdepthfastmonoculardepthestimationonembeddedsystems
AT yangtienju fastdepthfastmonoculardepthestimationonembeddedsystems
AT karamansertac fastdepthfastmonoculardepthestimationonembeddedsystems
AT szevivienne fastdepthfastmonoculardepthestimationonembeddedsystems