JOURNAL ARTICLE

Real-Time LiDAR Point Cloud Semantic Segmentation for Autonomous Driving

Xing XieLin BaiXinming Huang

Year: 2021 Journal:   Electronics Vol: 11 (1)Pages: 11-11   Publisher: Multidisciplinary Digital Publishing Institute

Abstract

LiDAR has been widely used in autonomous driving systems to provide high-precision 3D geometric information about the vehicle’s surroundings for perception, localization, and path planning. LiDAR-based point cloud semantic segmentation is an important task with a critical real-time requirement. However, most of the existing convolutional neural network (CNN) models for 3D point cloud semantic segmentation are very complex and can hardly be processed at real-time on an embedded platform. In this study, a lightweight CNN structure was proposed for projection-based LiDAR point cloud semantic segmentation with only 1.9 M parameters that gave an 87% reduction comparing to the state-of-the-art networks. When evaluated on a GPU, the processing time was 38.5 ms per frame, and it achieved a 47.9% mIoU score on Semantic-KITTI dataset. In addition, the proposed CNN is targeted on an FPGA using an NVDLA architecture, which results in a 2.74x speedup over the GPU implementation with a 46 times improvement in terms of power efficiency.

Keywords:
Point cloud Lidar Computer science Convolutional neural network Segmentation Artificial intelligence Computer vision Speedup Frame rate Cloud computing Deep learning Feature (linguistics) Real-time computing Remote sensing Geography

Metrics

28
Cited By
2.04
FWCI (Field Weighted Citation Impact)
37
Refs
0.88
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

Advanced Neural Network Applications
Physical Sciences →  Computer Science →  Computer Vision and Pattern Recognition
Remote Sensing and LiDAR Applications
Physical Sciences →  Environmental Science →  Environmental Engineering
Robotics and Sensor-Based Localization
Physical Sciences →  Engineering →  Aerospace Engineering
© 2026 ScienceGate Book Chapters — All rights reserved.