Abstract
This work aims to address the challenges in autonomous driving by focusing on the 3D perception of the environment using roadside LiDARs. We design a 3D object detection model that can detect traffic participants in roadside LiDARs in real-time. Our model uses an existing 3D detector as a baseline and improves its accuracy. To prove the effectiveness of our proposed modules, we train and evaluate the model on three different vehicle and infrastructure datasets. To show the domain adaptation ability of our detector, we train it on an infrastructure dataset from China and perform transfer learning on a different dataset recorded in Germany. We do several sets of experiments and ablation studies for each module in the detector that show that our model outperforms the baseline by a significant margin, while the inference speed is at 45 Hz (22 ms). We make a significant contribution with our LiDAR-based 3D detector that can be used for smart city applications to provide connected and automated vehicles with a far-reaching view. Vehicles that are connected to the roadside sensors can get information about other vehicles around the corner to improve their path and maneuver planning and to increase road traffic safety.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Caesar H, Bankiti V, Lang AH, Vora S, Liong VE, Xu Q, Krishnan A, Pan Y, Baldan G, Beijbom O (2020) Nuscenes: a multimodal dataset for autonomous driving. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 11,621–11,631
Cress C, Zimmer W, Strand L, Fortkord M, Dai S, Lakshminarasimhan V, Knoll A (2022) A9-dataset: multi-sensor infrastructure-based dataset for mobility research. arXiv preprint
Croce N (2021) Openlabel version 1.0.0 standardization project by asam association for standardization of automation and measuring systems. https://www.asam.net/standards/detail/openlabel/. Accessed 12 Nov 2021
Deng J, Zhou W, Zhang Y, Li H (2021) From multi-view to hollow-3d: Hallucinated hollow-3d r-cnn for 3d object detection. IEEE Trans Circuits Syst Video Technol
Dosovitskiy A, Ros G, Codevilla F, Lopez A, Koltun V (2017) Carla: an open urban driving simulator. In: Conference on robot learning. PMLR, pp 1–16
Eldar Y, Lindenbaum M, Porat M, Zeevi YY (1997) The farthest point strategy for progressive image sampling. IEEE Trans Image Process 6(9):1305–1315
Fan L, Xiong X, Wang F, Wang N, Zhang Z (2021) Rangedet: in defense of range view for lidar-based 3d object detection. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 2918–2927
Geiger A, Lenz P, Stiller C, Urtasun R (2013) Vision meets robotics: the kitti dataset. Int J Robot Res 32(11):1231–1237
Geiger A, Lenz P, Urtasun R (2012) Are we ready for autonomous driving? the kitti vision benchmark suite. In: 2012 IEEE conference on computer vision and pattern recognition. IEEE, pp 3354–3361
Graham B, Engelcke M, Van Der Maaten L (2018) 3d semantic segmentation with submanifold sparse convolutional networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 9224–9232
He C, Zeng H, Huang J, Hua X.S, Zhang L (2020) Structure aware single-stage 3d object detection from point cloud. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 11,873–11,882
He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 770–778
Hu J, Shen L, Sun G (2018) Squeeze-and-excitation networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 7132–7141
Ioffe S, Szegedy C (2015) Batch normalization: accelerating deep network training by reducing internal covariate shift. In: International conference on machine learning. PMLR, pp 448–456
Kloeker L, Kotulla C, Eckstein L (2020) Real-time point cloud fusion of multi-lidar infrastructure sensor setups with unknown spatial location and orientation. In: 2020 IEEE 23rd international conference on intelligent transportation systems (ITSC). IEEE, pp 1–8
Krämmer A, Schöller C, Gulati D, Lakshminarasimhan V, Kurz F, Rosenbaum D, Lenz C, Knoll A (2019) Providentia-a large-scale sensor system for the assistance of autonomous vehicles and its evaluation. arXiv preprint arXiv:1906.06789
Lang AH, Vora S, Caesar H, Zhou L, Yang J, Beijbom O (2019) Pointpillars: fast encoders for object detection from point clouds. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 12,697–12,705
Liang Z, Zhang M, Zhang Z, Zhao X, Pu S (2020) Rangercnn: towards fast and accurate 3d object detection with range image representation. arXiv preprint arXiv:2009.00206
Lin TY, Goyal P, Girshick R, He K, Dollár P (2017) Focal loss for dense object detection. In: Proceedings of the IEEE international conference on computer vision, pp 2980–2988
Liu B, Wang M, Foroosh H, Tappen M, Pensky M (2015) Sparse convolutional neural networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 806–814
Liu Z, Zhao X, Huang T, Hu R, Zhou Y, Bai X (2020) Tanet: robust 3d object detection from point clouds with triple attention. Proceedings of the AAAI conference on artificial intelligence 34:11677–11684
Nair V, Hinton GE (2010) Rectified linear units improve restricted boltzmann machines. In: Icml
Noh J, Lee S, Ham B (2021) Hvpr: hybrid voxel-point representation for single-stage 3d object detection. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 14,605–14,614
Paigwar A, Sierra-Gonzalez D, Erkent Ö, Laugier C (2021) Frustum-pointpillars: a multi-stage approach for 3d object detection using rgb camera and lidar. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 2926–2933
Qi CR, Su H, Mo K, Guibas LJ (2017) Pointnet: deep learning on point sets for 3d classification and segmentation. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 652–660
Qi CR, Yi L, Su H, Guibas LJ (2017) Pointnet++ deep hierarchical feature learning on point sets in a metric space. In: Proceedings of the 31st international conference on neural information processing systems, pp 5105–5114
Shi S, Guo C, Jiang L, Wang Z, Shi J, Wang X, Li H (2020) Pv-rcnn: point-voxel feature set abstraction for 3d object detection. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 10,529–10,538
Shi S, Wang X, Li H (2019) Pointrcnn: 3d object proposal generation and detection from point cloud. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 770–779
Triess LT, Dreissig M, Rist CB, Zöllner JM (2021) A survey on deep domain adaptation for lidar perception. In: 2021 IEEE intelligent vehicles symposium workshops (IV workshops), pp 350–357. IEEE
Wang H, Zhang X, Li J, Li Z, Yang L, Pan S, Deng Y (2021) Ips300+: a challenging multimodal dataset for intersection perception system. arXiv preprint arXiv:2106.02781
Yan Y, Mao Y, Li B (2018) Second: sparsely embedded convolutional detection. Sensors 18(10):3337
Yang Z, Sun Y, Liu S, Jia J (2020) 3dssd: point-based 3d single stage object detector. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 11,040–11,048
Yang Z, Sun Y, Liu S, Shen X, Jia J (2019) Std: sparse-to-dense 3d object detector for point cloud. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 1951–1960
Yu J, Jiang Y, Wang Z, Cao Z, Huang T (2016) Unitbox: an advanced object detection network. In: Proceedings of the 24th ACM international conference on multimedia, pp 516–520
Zhang W, Li W, Xu D (2021) Srdan: scale-aware and range-aware domain adaptation network for cross-dataset 3d object detection. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 6769–6779
Zheng W, Tang W, Chen S, Jiang L, Fu CW (2020) Cia-ssd: confident iou-aware single-stage object detector from point cloud. arXiv preprint arXiv:2012.03015
Zheng W, Tang W, Jiang L, Fu C.W (2021) Se-ssd: self-ensembling single-stage object detector from point cloud. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 14,494–14,503
Zhou X, Zimmer W, Erçelik E, Knoll A (2021) Real-time lidar-based 3d object detection on the highway. Master’s thesis, Technische Universität München. Unpublished thesis
Zhou Y, Tuzel O (2018) Voxelnet: end-to-end learning for point cloud based 3d object detection. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 4490–4499
Zimmer W, Rangesh A, Trivedi M (2019) 3d bat: a semi-automatic, web-based 3d annotation toolbox for full-surround, multi-modal data streams. In: 2019 IEEE intelligent vehicles symposium (IV). IEEE, pp 1816–1821
Acknowledgements
This work was funded by the Federal Ministry of Transport and Digital Infrastructure, Germany as part of the research project Providentia++ (Grant Number: 01MM19008A). The authors would like to express their gratitude to the funding agency and to the numerous students at TUM who have contributed to the creation of the first batch of the A9-Dataset.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Zimmer, W., Wu, J., Zhou, X., Knoll, A.C. (2023). Real-Time And Robust 3D Object Detection with Roadside LiDARs. In: Antoniou, C., Busch, F., Rau, A., Hariharan, M. (eds) Proceedings of the 12th International Scientific Conference on Mobility and Transport. Lecture Notes in Mobility. Springer, Singapore. https://doi.org/10.1007/978-981-19-8361-0_13
Download citation
DOI: https://doi.org/10.1007/978-981-19-8361-0_13
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-19-8360-3
Online ISBN: 978-981-19-8361-0
eBook Packages: EngineeringEngineering (R0)