Uncertainty from Motion for DNN Monocular Depth Estimation

Deployment of deep neural networks (DNNs) for monocular depth estimation in safety-critical scenarios on resource-constrained platforms requires well-calibrated and efficient uncertainty estimates. However, many popular uncertainty estimation techniques, including state-of-the-art ensembles and p...

Full description

Bibliographic Details
Main Authors: Sze, Vivienne, Karaman, Sertac, Sudhakar, Soumya
Format: Article
Published: 2022
Online Access:https://hdl.handle.net/1721.1/141382
_version_ 1826212953341296640
author Sze, Vivienne
Karaman, Sertac
Sudhakar, Soumya
author_facet Sze, Vivienne
Karaman, Sertac
Sudhakar, Soumya
author_sort Sze, Vivienne
collection MIT
description Deployment of deep neural networks (DNNs) for monocular depth estimation in safety-critical scenarios on resource-constrained platforms requires well-calibrated and efficient uncertainty estimates. However, many popular uncertainty estimation techniques, including state-of-the-art ensembles and popular sampling-based methods, require multiple inferences per input, making them difficult to deploy in latencyconstrained or energy-constrained scenarios. We propose a new algorithm, called Uncertainty from Motion (UfM), that requires only one inference per input. UfM exploits the temporal redundancy in video inputs by merging incrementally the per-pixel depth prediction and per-pixel aleatoric uncertainty prediction of points that are seen in multiple views in the video sequence. When UfM is applied to ensembles, we show that UfM can retain the uncertainty quality of ensembles at a fraction of the energy by running only a single ensemble member at each frame and fusing the uncertainty over the sequence of frames. In a set of representative experiments using FCDenseNet and eight indistribution and out-of-distribution video sequences, UfM offers comparable uncertainty quality to an ensemble of size 10 while consuming only 11.3% of the ensemble’s energy and running 6.4× faster on a single Nvidia RTX 2080 Ti GPU, enabling near ensemble uncertainty quality for resource-constrained, real-time scenarios.
first_indexed 2024-09-23T15:41:05Z
format Article
id mit-1721.1/141382
institution Massachusetts Institute of Technology
last_indexed 2024-09-23T15:41:05Z
publishDate 2022
record_format dspace
spelling mit-1721.1/1413822022-03-30T03:04:37Z Uncertainty from Motion for DNN Monocular Depth Estimation Sze, Vivienne Karaman, Sertac Sudhakar, Soumya Deployment of deep neural networks (DNNs) for monocular depth estimation in safety-critical scenarios on resource-constrained platforms requires well-calibrated and efficient uncertainty estimates. However, many popular uncertainty estimation techniques, including state-of-the-art ensembles and popular sampling-based methods, require multiple inferences per input, making them difficult to deploy in latencyconstrained or energy-constrained scenarios. We propose a new algorithm, called Uncertainty from Motion (UfM), that requires only one inference per input. UfM exploits the temporal redundancy in video inputs by merging incrementally the per-pixel depth prediction and per-pixel aleatoric uncertainty prediction of points that are seen in multiple views in the video sequence. When UfM is applied to ensembles, we show that UfM can retain the uncertainty quality of ensembles at a fraction of the energy by running only a single ensemble member at each frame and fusing the uncertainty over the sequence of frames. In a set of representative experiments using FCDenseNet and eight indistribution and out-of-distribution video sequences, UfM offers comparable uncertainty quality to an ensemble of size 10 while consuming only 11.3% of the ensemble’s energy and running 6.4× faster on a single Nvidia RTX 2080 Ti GPU, enabling near ensemble uncertainty quality for resource-constrained, real-time scenarios. National Science Foundation (NSF) 2022-03-29T12:01:20Z 2022-03-29T12:01:20Z 2022-05-23 Article http://purl.org/eprint/type/ConferencePaper https://hdl.handle.net/1721.1/141382 Sze, Vivienne, Karaman, Sertac and Sudhakar, Soumya. 2022. "Uncertainty from Motion for DNN Monocular Depth Estimation." IEEE International Conference on Robotics and Automation (ICRA). IEEE International Conference on Robotics and Automation (ICRA) Creative Commons Attribution-Noncommercial-Share Alike http://creativecommons.org/licenses/by-nc-sa/4.0/ application/pdf Prof. Sze
spellingShingle Sze, Vivienne
Karaman, Sertac
Sudhakar, Soumya
Uncertainty from Motion for DNN Monocular Depth Estimation
title Uncertainty from Motion for DNN Monocular Depth Estimation
title_full Uncertainty from Motion for DNN Monocular Depth Estimation
title_fullStr Uncertainty from Motion for DNN Monocular Depth Estimation
title_full_unstemmed Uncertainty from Motion for DNN Monocular Depth Estimation
title_short Uncertainty from Motion for DNN Monocular Depth Estimation
title_sort uncertainty from motion for dnn monocular depth estimation
url https://hdl.handle.net/1721.1/141382
work_keys_str_mv AT szevivienne uncertaintyfrommotionfordnnmonoculardepthestimation
AT karamansertac uncertaintyfrommotionfordnnmonoculardepthestimation
AT sudhakarsoumya uncertaintyfrommotionfordnnmonoculardepthestimation