Depth Estimation From a Single Image Using Guided Deep Network

This paper addresses the problem of monocular depth estimation, which plays a key role to understand a given scene. Owing to the success of the generative model using deep neural networks, the performance of depth estimation from a single image has been significantly improved. However, most previous...

Full description

Bibliographic Details
Main Authors: Minsoo Song, Wonjun Kim
Format: Article
Language:English
Published: IEEE 2019-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/8854079/
_version_ 1818444283072479232
author Minsoo Song
Wonjun Kim
author_facet Minsoo Song
Wonjun Kim
author_sort Minsoo Song
collection DOAJ
description This paper addresses the problem of monocular depth estimation, which plays a key role to understand a given scene. Owing to the success of the generative model using deep neural networks, the performance of depth estimation from a single image has been significantly improved. However, most previous approaches still fail to accurately estimate the depth boundary and thus lead to the result of the blurry restoration. In this paper, a novel and simple method is proposed by exploiting the latent space of the depth-to-depth network, which contains useful encoded features for guiding the process of depth generation. This network, so-called guided network, simply consists of convolution layers and their corresponding deconvolution ones, and is also easily trained by only using single depth images. For efficiently learning the relationship between a color value and its related depth value in a given image, we propose to train the color-to-depth network via loss defined along with features from the latent space of our guided network (i.e., depth-to-depth network). One important advantage of the proposed method is to greatly enhance local details even under complicated background regions. Moreover, the proposed method works very fast (at 125 fps with GPU). Experimental results on various benchmark datasets show the efficiency and robustness of the proposed approach compared to state-of-the-art methods.
first_indexed 2024-12-14T19:13:28Z
format Article
id doaj.art-7559744663f940e9a440c00d48a5f684
institution Directory Open Access Journal
issn 2169-3536
language English
last_indexed 2024-12-14T19:13:28Z
publishDate 2019-01-01
publisher IEEE
record_format Article
series IEEE Access
spelling doaj.art-7559744663f940e9a440c00d48a5f6842022-12-21T22:50:40ZengIEEEIEEE Access2169-35362019-01-01714259514260610.1109/ACCESS.2019.29449378854079Depth Estimation From a Single Image Using Guided Deep NetworkMinsoo Song0Wonjun Kim1https://orcid.org/0000-0001-5121-5931Department of Electrical and Electronics Engineering, Konkuk University, Seoul, South KoreaDepartment of Electrical and Electronics Engineering, Konkuk University, Seoul, South KoreaThis paper addresses the problem of monocular depth estimation, which plays a key role to understand a given scene. Owing to the success of the generative model using deep neural networks, the performance of depth estimation from a single image has been significantly improved. However, most previous approaches still fail to accurately estimate the depth boundary and thus lead to the result of the blurry restoration. In this paper, a novel and simple method is proposed by exploiting the latent space of the depth-to-depth network, which contains useful encoded features for guiding the process of depth generation. This network, so-called guided network, simply consists of convolution layers and their corresponding deconvolution ones, and is also easily trained by only using single depth images. For efficiently learning the relationship between a color value and its related depth value in a given image, we propose to train the color-to-depth network via loss defined along with features from the latent space of our guided network (i.e., depth-to-depth network). One important advantage of the proposed method is to greatly enhance local details even under complicated background regions. Moreover, the proposed method works very fast (at 125 fps with GPU). Experimental results on various benchmark datasets show the efficiency and robustness of the proposed approach compared to state-of-the-art methods.https://ieeexplore.ieee.org/document/8854079/Monocular depth estimationdepth-to-depth networkcolor-to-depth networklatent spaceguided network
spellingShingle Minsoo Song
Wonjun Kim
Depth Estimation From a Single Image Using Guided Deep Network
IEEE Access
Monocular depth estimation
depth-to-depth network
color-to-depth network
latent space
guided network
title Depth Estimation From a Single Image Using Guided Deep Network
title_full Depth Estimation From a Single Image Using Guided Deep Network
title_fullStr Depth Estimation From a Single Image Using Guided Deep Network
title_full_unstemmed Depth Estimation From a Single Image Using Guided Deep Network
title_short Depth Estimation From a Single Image Using Guided Deep Network
title_sort depth estimation from a single image using guided deep network
topic Monocular depth estimation
depth-to-depth network
color-to-depth network
latent space
guided network
url https://ieeexplore.ieee.org/document/8854079/
work_keys_str_mv AT minsoosong depthestimationfromasingleimageusingguideddeepnetwork
AT wonjunkim depthestimationfromasingleimageusingguideddeepnetwork