Towards urban scene semantic segmentation with deep learning from lidar point clouds: A case study in baden-württemberg, germany

17Citations
Citations of this article
35Readers
Mendeley users who have this article in their library.

Abstract

An accurate understanding of urban objects is critical for urban modeling, intelligent infrastructure planning and city management. The semantic segmentation of light detection and ranging (LiDAR) point clouds is a fundamental approach for urban scene analysis. Over the last years, several methods have been developed to segment urban furniture with point clouds. However, the traditional processing of large amounts of spatial data has become increasingly costly, both time-wise and financially. Recently, deep learning (DL) techniques have been increasingly used for 3D segmentation tasks. Yet, most of these deep neural networks (DNNs) were conducted on benchmarks. It is, therefore, arguable whether DL approaches can achieve the state-of-the-art performance of 3D point clouds segmentation in real-life scenarios. In this research, we apply an adapted DNN (ARandLA-Net) to directly process large-scale point clouds. In particular, we develop a new paradigm for training and validation, which presents a typical urban scene in central Europe (Munzingen, Freiburg, Baden-Württemberg, Germany). Our dataset consists of nearly 390 million dense points acquired by Mobile Laser Scanning (MLS), which has a rather larger quantity of sample points in comparison to existing datasets and includes meaningful object categories that are particular to applications for smart cities and urban planning. We further assess the DNN on our dataset and investigate a number of key challenges from varying aspects, such as data preparation strategies, the advantage of color information and the unbalanced class distribution in the real world. The final segmentation model achieved a mean Intersection-over-Union (mIoU) score of 54.4% and an overall accuracy score of 83.9%. Our experiments indicated that different data preparation strategies influenced the model performance. Additional RGB information yielded an approximately 4% higher mIoU score. Our results also demonstrate that the use of weighted cross-entropy with inverse square root frequency loss led to better segmentation performance than when other losses were considered.

References Powered by Scopus

Dynamic graph Cnn for learning on point clouds

4786Citations
N/AReaders
Get full text

KPConv: Flexible and deformable convolution for point clouds

2260Citations
N/AReaders
Get full text

4D spatio-temporal convnets: Minkowski convolutional neural networks

1412Citations
N/AReaders
Get full text

Cited by Powered by Scopus

Algorithmic urban planning for smart and sustainable development: Systematic review of the literature

144Citations
N/AReaders
Get full text

LESS: Label-Efficient Semantic Segmentation for LiDAR Point Clouds

43Citations
N/AReaders
Get full text

Road-Side Individual Tree Segmentation from Urban MLS Point Clouds Using Metric Learning

12Citations
N/AReaders
Get full text

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Cite

CITATION STYLE

APA

Zou, Y., Weinacker, H., & Koch, B. (2021). Towards urban scene semantic segmentation with deep learning from lidar point clouds: A case study in baden-württemberg, germany. Remote Sensing, 13(16). https://doi.org/10.3390/rs13163220

Readers' Seniority

Tooltip

PhD / Post grad / Masters / Doc 16

80%

Researcher 2

10%

Professor / Associate Prof. 1

5%

Lecturer / Post doc 1

5%

Readers' Discipline

Tooltip

Computer Science 8

40%

Engineering 7

35%

Chemistry 3

15%

Earth and Planetary Sciences 2

10%

Save time finding and organizing research with Mendeley

Sign up for free