Depth-Based Dynamic Sampling of Neural Radiation Fields

Although the NeRF approach can achieve outstanding view synthesis, it is limited in practical use because it requires many views (hundreds) for training. With only a few input views, the Depth-DYN NeRF that we propose can accurately match the shape. First, we adopted the ip_basic depth-completion me...

Full description

Bibliographic Details
Main Authors: Jie Wang, Jiangjian Xiao, Xiaolu Zhang, Xiaolin Xu, Tianxing Jin, Zhijia Jin
Format: Article
Language:English
Published: MDPI AG 2023-02-01
Series:Electronics
Subjects:
Online Access:https://www.mdpi.com/2079-9292/12/4/1053
_version_ 1797621328615833600
author Jie Wang
Jiangjian Xiao
Xiaolu Zhang
Xiaolin Xu
Tianxing Jin
Zhijia Jin
author_facet Jie Wang
Jiangjian Xiao
Xiaolu Zhang
Xiaolin Xu
Tianxing Jin
Zhijia Jin
author_sort Jie Wang
collection DOAJ
description Although the NeRF approach can achieve outstanding view synthesis, it is limited in practical use because it requires many views (hundreds) for training. With only a few input views, the Depth-DYN NeRF that we propose can accurately match the shape. First, we adopted the ip_basic depth-completion method, which can recover the complete depth map from sparse radar depth data. Then, we further designed the Depth-DYN MLP network architecture, which uses a dense depth prior to constraining the NeRF optimization and combines the depthloss to supervise the Depth-DYN MLP network. When compared to the color-only supervised-based NeRF, the Depth-DYN MLP network can better recover the geometric structure of the model and reduce the appearance of shadows. To further ensure that the depth depicted along the rays intersecting these 3D points is close to the measured depth, we dynamically modified the sample space based on the depth of each pixel point. Depth-DYN NeRF considerably outperforms depth NeRF and other sparse view versions when there are a few input views. Using only 10–20 photos to render high-quality images on the new view, our strategy was tested and confirmed on a variety of benchmark datasets. Compared with NeRF, we obtained better image quality (NeRF average at 22.47 dB vs. our 27.296 dB).
first_indexed 2024-03-11T08:54:17Z
format Article
id doaj.art-e20288ca76e64f999c1c718243baeba9
institution Directory Open Access Journal
issn 2079-9292
language English
last_indexed 2024-03-11T08:54:17Z
publishDate 2023-02-01
publisher MDPI AG
record_format Article
series Electronics
spelling doaj.art-e20288ca76e64f999c1c718243baeba92023-11-16T20:14:14ZengMDPI AGElectronics2079-92922023-02-01124105310.3390/electronics12041053Depth-Based Dynamic Sampling of Neural Radiation FieldsJie Wang0Jiangjian Xiao1Xiaolu Zhang2Xiaolin Xu3Tianxing Jin4Zhijia Jin5Faculty of Electrical Engineering and Computer Science, Ningbo University, Ningbo 315201, ChinaFaculty of Electrical Engineering and Computer Science, Ningbo University, Ningbo 315201, ChinaNingbo Institute of Industrial Technology, Chinese Academy of Sciences, Ningbo 315201, ChinaFaculty of Electrical Engineering and Computer Science, Ningbo University, Ningbo 315201, ChinaFaculty of Electrical Engineering and Computer Science, Ningbo University, Ningbo 315201, ChinaFaculty of Electrical Engineering and Computer Science, Ningbo University, Ningbo 315201, ChinaAlthough the NeRF approach can achieve outstanding view synthesis, it is limited in practical use because it requires many views (hundreds) for training. With only a few input views, the Depth-DYN NeRF that we propose can accurately match the shape. First, we adopted the ip_basic depth-completion method, which can recover the complete depth map from sparse radar depth data. Then, we further designed the Depth-DYN MLP network architecture, which uses a dense depth prior to constraining the NeRF optimization and combines the depthloss to supervise the Depth-DYN MLP network. When compared to the color-only supervised-based NeRF, the Depth-DYN MLP network can better recover the geometric structure of the model and reduce the appearance of shadows. To further ensure that the depth depicted along the rays intersecting these 3D points is close to the measured depth, we dynamically modified the sample space based on the depth of each pixel point. Depth-DYN NeRF considerably outperforms depth NeRF and other sparse view versions when there are a few input views. Using only 10–20 photos to render high-quality images on the new view, our strategy was tested and confirmed on a variety of benchmark datasets. Compared with NeRF, we obtained better image quality (NeRF average at 22.47 dB vs. our 27.296 dB).https://www.mdpi.com/2079-9292/12/4/1053NeRFscene representationview synthesisimage-based renderingvolume rendering
spellingShingle Jie Wang
Jiangjian Xiao
Xiaolu Zhang
Xiaolin Xu
Tianxing Jin
Zhijia Jin
Depth-Based Dynamic Sampling of Neural Radiation Fields
Electronics
NeRF
scene representation
view synthesis
image-based rendering
volume rendering
title Depth-Based Dynamic Sampling of Neural Radiation Fields
title_full Depth-Based Dynamic Sampling of Neural Radiation Fields
title_fullStr Depth-Based Dynamic Sampling of Neural Radiation Fields
title_full_unstemmed Depth-Based Dynamic Sampling of Neural Radiation Fields
title_short Depth-Based Dynamic Sampling of Neural Radiation Fields
title_sort depth based dynamic sampling of neural radiation fields
topic NeRF
scene representation
view synthesis
image-based rendering
volume rendering
url https://www.mdpi.com/2079-9292/12/4/1053
work_keys_str_mv AT jiewang depthbaseddynamicsamplingofneuralradiationfields
AT jiangjianxiao depthbaseddynamicsamplingofneuralradiationfields
AT xiaoluzhang depthbaseddynamicsamplingofneuralradiationfields
AT xiaolinxu depthbaseddynamicsamplingofneuralradiationfields
AT tianxingjin depthbaseddynamicsamplingofneuralradiationfields
AT zhijiajin depthbaseddynamicsamplingofneuralradiationfields