Towards Urban Scene Semantic Segmentation with Deep Learning from LiDAR Point Clouds: A Case Study in Baden-Württemberg, Germany

An accurate understanding of urban objects is critical for urban modeling, intelligent infrastructure planning and city management. The semantic segmentation of light detection and ranging (LiDAR) point clouds is a fundamental approach for urban scene analysis. Over the last years, several methods h...

Full description

Bibliographic Details
Main Authors: Yanling Zou, Holger Weinacker, Barbara Koch
Format: Article
Language:English
Published: MDPI AG 2021-08-01
Series:Remote Sensing
Subjects:
Online Access:https://www.mdpi.com/2072-4292/13/16/3220
_version_ 1797522151183482880
author Yanling Zou
Holger Weinacker
Barbara Koch
author_facet Yanling Zou
Holger Weinacker
Barbara Koch
author_sort Yanling Zou
collection DOAJ
description An accurate understanding of urban objects is critical for urban modeling, intelligent infrastructure planning and city management. The semantic segmentation of light detection and ranging (LiDAR) point clouds is a fundamental approach for urban scene analysis. Over the last years, several methods have been developed to segment urban furniture with point clouds. However, the traditional processing of large amounts of spatial data has become increasingly costly, both time-wise and financially. Recently, deep learning (DL) techniques have been increasingly used for 3D segmentation tasks. Yet, most of these deep neural networks (DNNs) were conducted on benchmarks. It is, therefore, arguable whether DL approaches can achieve the state-of-the-art performance of 3D point clouds segmentation in real-life scenarios. In this research, we apply an adapted DNN (ARandLA-Net) to directly process large-scale point clouds. In particular, we develop a new paradigm for training and validation, which presents a typical urban scene in central Europe (Munzingen, Freiburg, Baden-Württemberg, Germany). Our dataset consists of nearly 390 million dense points acquired by Mobile Laser Scanning (MLS), which has a rather larger quantity of sample points in comparison to existing datasets and includes meaningful object categories that are particular to applications for smart cities and urban planning. We further assess the DNN on our dataset and investigate a number of key challenges from varying aspects, such as data preparation strategies, the advantage of color information and the unbalanced class distribution in the real world. The final segmentation model achieved a mean Intersection-over-Union (mIoU) score of 54.4% and an overall accuracy score of 83.9%. Our experiments indicated that different data preparation strategies influenced the model performance. Additional RGB information yielded an approximately 4% higher mIoU score. Our results also demonstrate that the use of weighted cross-entropy with inverse square root frequency loss led to better segmentation performance than when other losses were considered.
first_indexed 2024-03-10T08:25:24Z
format Article
id doaj.art-6a1c7c375c7740159093eb06fbc661eb
institution Directory Open Access Journal
issn 2072-4292
language English
last_indexed 2024-03-10T08:25:24Z
publishDate 2021-08-01
publisher MDPI AG
record_format Article
series Remote Sensing
spelling doaj.art-6a1c7c375c7740159093eb06fbc661eb2023-11-22T09:33:59ZengMDPI AGRemote Sensing2072-42922021-08-011316322010.3390/rs13163220Towards Urban Scene Semantic Segmentation with Deep Learning from LiDAR Point Clouds: A Case Study in Baden-Württemberg, GermanyYanling Zou0Holger Weinacker1Barbara Koch2Chair of Remote Sensing and Landscape Information Systems, University of Freiburg, Tennenbacherstr. 4, 79106 Freiburg, GermanyChair of Remote Sensing and Landscape Information Systems, University of Freiburg, Tennenbacherstr. 4, 79106 Freiburg, GermanyChair of Remote Sensing and Landscape Information Systems, University of Freiburg, Tennenbacherstr. 4, 79106 Freiburg, GermanyAn accurate understanding of urban objects is critical for urban modeling, intelligent infrastructure planning and city management. The semantic segmentation of light detection and ranging (LiDAR) point clouds is a fundamental approach for urban scene analysis. Over the last years, several methods have been developed to segment urban furniture with point clouds. However, the traditional processing of large amounts of spatial data has become increasingly costly, both time-wise and financially. Recently, deep learning (DL) techniques have been increasingly used for 3D segmentation tasks. Yet, most of these deep neural networks (DNNs) were conducted on benchmarks. It is, therefore, arguable whether DL approaches can achieve the state-of-the-art performance of 3D point clouds segmentation in real-life scenarios. In this research, we apply an adapted DNN (ARandLA-Net) to directly process large-scale point clouds. In particular, we develop a new paradigm for training and validation, which presents a typical urban scene in central Europe (Munzingen, Freiburg, Baden-Württemberg, Germany). Our dataset consists of nearly 390 million dense points acquired by Mobile Laser Scanning (MLS), which has a rather larger quantity of sample points in comparison to existing datasets and includes meaningful object categories that are particular to applications for smart cities and urban planning. We further assess the DNN on our dataset and investigate a number of key challenges from varying aspects, such as data preparation strategies, the advantage of color information and the unbalanced class distribution in the real world. The final segmentation model achieved a mean Intersection-over-Union (mIoU) score of 54.4% and an overall accuracy score of 83.9%. Our experiments indicated that different data preparation strategies influenced the model performance. Additional RGB information yielded an approximately 4% higher mIoU score. Our results also demonstrate that the use of weighted cross-entropy with inverse square root frequency loss led to better segmentation performance than when other losses were considered.https://www.mdpi.com/2072-4292/13/16/3220urban scenemobile mappingdeep learningremote sensingpoint cloudssemantic segmentation
spellingShingle Yanling Zou
Holger Weinacker
Barbara Koch
Towards Urban Scene Semantic Segmentation with Deep Learning from LiDAR Point Clouds: A Case Study in Baden-Württemberg, Germany
Remote Sensing
urban scene
mobile mapping
deep learning
remote sensing
point clouds
semantic segmentation
title Towards Urban Scene Semantic Segmentation with Deep Learning from LiDAR Point Clouds: A Case Study in Baden-Württemberg, Germany
title_full Towards Urban Scene Semantic Segmentation with Deep Learning from LiDAR Point Clouds: A Case Study in Baden-Württemberg, Germany
title_fullStr Towards Urban Scene Semantic Segmentation with Deep Learning from LiDAR Point Clouds: A Case Study in Baden-Württemberg, Germany
title_full_unstemmed Towards Urban Scene Semantic Segmentation with Deep Learning from LiDAR Point Clouds: A Case Study in Baden-Württemberg, Germany
title_short Towards Urban Scene Semantic Segmentation with Deep Learning from LiDAR Point Clouds: A Case Study in Baden-Württemberg, Germany
title_sort towards urban scene semantic segmentation with deep learning from lidar point clouds a case study in baden wurttemberg germany
topic urban scene
mobile mapping
deep learning
remote sensing
point clouds
semantic segmentation
url https://www.mdpi.com/2072-4292/13/16/3220
work_keys_str_mv AT yanlingzou towardsurbanscenesemanticsegmentationwithdeeplearningfromlidarpointcloudsacasestudyinbadenwurttemberggermany
AT holgerweinacker towardsurbanscenesemanticsegmentationwithdeeplearningfromlidarpointcloudsacasestudyinbadenwurttemberggermany
AT barbarakoch towardsurbanscenesemanticsegmentationwithdeeplearningfromlidarpointcloudsacasestudyinbadenwurttemberggermany