Depth Estimation From a Single Image Using Guided Deep Network
This paper addresses the problem of monocular depth estimation, which plays a key role to understand a given scene. Owing to the success of the generative model using deep neural networks, the performance of depth estimation from a single image has been significantly improved. However, most previous...
Main Authors: | , |
---|---|
Format: | Article |
Language: | English |
Published: |
IEEE
2019-01-01
|
Series: | IEEE Access |
Subjects: | |
Online Access: | https://ieeexplore.ieee.org/document/8854079/ |
_version_ | 1818444283072479232 |
---|---|
author | Minsoo Song Wonjun Kim |
author_facet | Minsoo Song Wonjun Kim |
author_sort | Minsoo Song |
collection | DOAJ |
description | This paper addresses the problem of monocular depth estimation, which plays a key role to understand a given scene. Owing to the success of the generative model using deep neural networks, the performance of depth estimation from a single image has been significantly improved. However, most previous approaches still fail to accurately estimate the depth boundary and thus lead to the result of the blurry restoration. In this paper, a novel and simple method is proposed by exploiting the latent space of the depth-to-depth network, which contains useful encoded features for guiding the process of depth generation. This network, so-called guided network, simply consists of convolution layers and their corresponding deconvolution ones, and is also easily trained by only using single depth images. For efficiently learning the relationship between a color value and its related depth value in a given image, we propose to train the color-to-depth network via loss defined along with features from the latent space of our guided network (i.e., depth-to-depth network). One important advantage of the proposed method is to greatly enhance local details even under complicated background regions. Moreover, the proposed method works very fast (at 125 fps with GPU). Experimental results on various benchmark datasets show the efficiency and robustness of the proposed approach compared to state-of-the-art methods. |
first_indexed | 2024-12-14T19:13:28Z |
format | Article |
id | doaj.art-7559744663f940e9a440c00d48a5f684 |
institution | Directory Open Access Journal |
issn | 2169-3536 |
language | English |
last_indexed | 2024-12-14T19:13:28Z |
publishDate | 2019-01-01 |
publisher | IEEE |
record_format | Article |
series | IEEE Access |
spelling | doaj.art-7559744663f940e9a440c00d48a5f6842022-12-21T22:50:40ZengIEEEIEEE Access2169-35362019-01-01714259514260610.1109/ACCESS.2019.29449378854079Depth Estimation From a Single Image Using Guided Deep NetworkMinsoo Song0Wonjun Kim1https://orcid.org/0000-0001-5121-5931Department of Electrical and Electronics Engineering, Konkuk University, Seoul, South KoreaDepartment of Electrical and Electronics Engineering, Konkuk University, Seoul, South KoreaThis paper addresses the problem of monocular depth estimation, which plays a key role to understand a given scene. Owing to the success of the generative model using deep neural networks, the performance of depth estimation from a single image has been significantly improved. However, most previous approaches still fail to accurately estimate the depth boundary and thus lead to the result of the blurry restoration. In this paper, a novel and simple method is proposed by exploiting the latent space of the depth-to-depth network, which contains useful encoded features for guiding the process of depth generation. This network, so-called guided network, simply consists of convolution layers and their corresponding deconvolution ones, and is also easily trained by only using single depth images. For efficiently learning the relationship between a color value and its related depth value in a given image, we propose to train the color-to-depth network via loss defined along with features from the latent space of our guided network (i.e., depth-to-depth network). One important advantage of the proposed method is to greatly enhance local details even under complicated background regions. Moreover, the proposed method works very fast (at 125 fps with GPU). Experimental results on various benchmark datasets show the efficiency and robustness of the proposed approach compared to state-of-the-art methods.https://ieeexplore.ieee.org/document/8854079/Monocular depth estimationdepth-to-depth networkcolor-to-depth networklatent spaceguided network |
spellingShingle | Minsoo Song Wonjun Kim Depth Estimation From a Single Image Using Guided Deep Network IEEE Access Monocular depth estimation depth-to-depth network color-to-depth network latent space guided network |
title | Depth Estimation From a Single Image Using Guided Deep Network |
title_full | Depth Estimation From a Single Image Using Guided Deep Network |
title_fullStr | Depth Estimation From a Single Image Using Guided Deep Network |
title_full_unstemmed | Depth Estimation From a Single Image Using Guided Deep Network |
title_short | Depth Estimation From a Single Image Using Guided Deep Network |
title_sort | depth estimation from a single image using guided deep network |
topic | Monocular depth estimation depth-to-depth network color-to-depth network latent space guided network |
url | https://ieeexplore.ieee.org/document/8854079/ |
work_keys_str_mv | AT minsoosong depthestimationfromasingleimageusingguideddeepnetwork AT wonjunkim depthestimationfromasingleimageusingguideddeepnetwork |