Real-Time Detection of Vulnerable Road Users Using a Lightweight Object Detection Model
Keywords:
Vulnerable Road Users, Object Detection, Deep Learning, Real-Time Processing, Lightweight Model, Road SafetyAbstract
Vulnerable Road Users (VRUs), including pedestrians, cyclists, and motorcyclists, face a heightened risk in traffic scenarios. The safety enhancement for VRUs relies heavily on evolving driver assistance systems and autonomous vehicles, anchored by swift VRU detection and localisation. Object detection models in computer vision are pivotal for this. Deploying such models on edge devices, especially the NVIDIA Jetson Nano presents challenges due to computational and power constraints. Our study compares the SSD MobileNetV2 FPN-Lite 320x320 model on the Jetson Nano for VRU detection with models like YOLOv3 and Faster RCNN. Key findings indicate that the SSD MobileNetV2 FPN-Lite 320x320 model achieves a mean average precision (mAP) of 0.45, precision of 0.80, and recall of 0.65 at 25 FPS in baseline evaluations. With optimisation, the FPS improved to 32 with slight changes in other metrics. The insights also touch upon inherent challenges in VRU detection, suggesting future research directions to refine the SSD MobileNetV2 FPN-Lite model's efficiency, ultimately striving for a safer transportation ecosystem.
Downloads
References
A. Alinda et al., Global status report on road safety 2018. Geneva: World Health Organization; 2018. doi: 10.1051/matecconf/201712107005.
P. Viola and M. Jones, "Rapid object detection using a boosted cascade of simple features," Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, vol. 1, 2001, doi: 10.1109/CVPR.2001.990517.
N. Dalal and B. Triggs, "Histograms of oriented gradients for human detection," Proceedings - 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, CVPR 2005, vol. I, pp. 886–893, 2005, doi: 10.1109/CVPR.2005.177.
M. T. Emirler, H. Wang, and B. A. Güvenç, “Socially Acceptable Collision Avoidance System for Vulnerable Road Users,” IFAC-PapersOnLine, vol. 49, no. 3, pp. 436–441, 2016, doi: 10.1016/j.ifacol.2016.07.073.
A. Bighashdel and G. Dubbelman, "A Survey on Path Prediction Techniques for Vulnerable Road Users: From Traditional to Deep-Learning Approaches," 2019 IEEE Intelligent Transportation Systems Conference, ITSC 2019, no. 783190, pp. 1039–1046, 2019, doi: 10.1109/ITSC.2019.8917053.
S. Ren, K. He, R. Girshick, and J. Sun, "Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks." [Online]. Available: https://github.com/
J. Borrego-Carazo, D. Castells-Rufas, E. Biempica, and J. Carrabina, "Resource-Constrained Machine Learning for ADAS: A Systematic Review," IEEE Access, vol. 8, pp. 40573–40598, 2020, doi: 10.1109/ACCESS.2020.2976513.
J. Redmon, S. Divvala, R. Girshick, and A. Farhadi, "You Only Look Once: Unified, Real-Time Object Detection." [Online]. Available: http://pjreddie.com/yolo/
S. Gilroy, E. Jones, and M. Glavin, "Overcoming Occlusion in the Automotive Environment-A Review," IEEE Transactions on Intelligent Transportation Systems, vol. PP, pp. 1–13, 2020, doi: 10.1109/tits.2019.2956813.
W. Liu et al., "SSD: Single shot multibox detector," Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), vol. 9905 LNCS, pp. 21–37, 2016, doi: 10.1007/978-3-319-46448-0_2/FIGURES/5.
T. Wu, Z. Zhang, Y. Liu, W. Pei, and H. Chen, "A lightweight small object detection algorithm based on improved SSD," Hongwai yu Jiguang Gongcheng/Infrared and Laser Engineering, vol. 47, no. 7, Jul. 2018, doi: 10.3788/IRLA201847.0703005.
L. Jiao et al., "A survey of deep learning-based object detection," IEEE Access, vol. 7, no. 3, pp. 128837–128868, 2019, doi: 10.1109/ACCESS.2019.2939201.
S. Grigorescu, B. Trasnea, T. Cocias, and G. Macesanu, "A survey of deep learning techniques for autonomous driving," J Field Robot, vol. 37, no. 3, pp. 362–386, 2020, doi: 10.1002/rob.21918.
A. G. Howard et al., "MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications," Apr. 2017, [Online]. Available: http://arxiv.org/abs/1704.04861
M. Sandler, A. Howard, M. Zhu, A. Zhmoginov, and L.-C. Chen, "MobileNetV2: Inverted Residuals and Linear Bottlenecks."
"Jetson Nano Developer Kit | NVIDIA Developer." https://developer.nvidia.com/embedded/jetson-nano-developer-kit (accessed Apr. 23, 2023).
W. Zhou, X. Min, R. Hu, Y. Long, H. Luo, and JunYi, "FasterX: Real-Time Object Detection Based on Edge GPUs for UAV Applications," Sep. 2022, [Online]. Available: http://arxiv.org/abs/2209.03157
A. A. Suzen, B. Duman, and B. Sen, "Benchmark Analysis of Jetson TX2, Jetson Nano and Raspberry PI using Deep-CNN," HORA 2020 - 2nd International Congress on Human-Computer Interaction, Optimization and Robotic Applications, Proceedings, Jun. 2020, doi: 10.1109/HORA49412.2020.9152915.
H. H. Nguyen, D. N. N. Tran, and J. W. Jeon, "Towards Real-Time Vehicle Detection on Edge Devices with Nvidia Jetson TX2," in 2020 IEEE International Conference on Consumer Electronics - Asia, ICCE-Asia 2020, Institute of Electrical and Electronics Engineers Inc., Nov. 2020. doi: 10.1109/ICCE-Asia49877.2020.9277463.
S. Valladares, M. Toscano, R. Tufiño, P. Morillo, and D. Vallejo-Huanga, "Performance Evaluation of the Nvidia Jetson Nano Through a Real-Time Machine Learning Application," pp. 343–349, 2021, doi: 10.1007/978-3-030-68017-6_51.
"TensorFlow." https://www.tensorflow.org/ (accessed Apr. 23, 2023).
A. Geiger, P. Lenz, and R. Urtasun, "Are we ready for autonomous driving? The KITTI vision benchmark suite," Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 3354–3361, 2012, doi: 10.1109/CVPR.2012.6248074.
M. Cordts et al., "The Cityscapes Dataset for Semantic Urban Scene Understanding," Apr. 2016, [Online]. Available: http://arxiv.org/abs/1604.01685
T.-Y. Lin et al., "Microsoft COCO: Common Objects in Context," May 2014, [Online]. Available: http://arxiv.org/abs/1405.0312
Downloads
Published
How to Cite
Issue
Section
License
This work is licensed under a Creative Commons Attribution-ShareAlike 4.0 International License.
All papers should be submitted electronically. All submitted manuscripts must be original work that is not under submission at another journal or under consideration for publication in another form, such as a monograph or chapter of a book. Authors of submitted papers are obligated not to submit their paper for publication elsewhere until an editorial decision is rendered on their submission. Further, authors of accepted papers are prohibited from publishing the results in other publications that appear before the paper is published in the Journal unless they receive approval for doing so from the Editor-In-Chief.
IJISAE open access articles are licensed under a Creative Commons Attribution-ShareAlike 4.0 International License. This license lets the audience to give appropriate credit, provide a link to the license, and indicate if changes were made and if they remix, transform, or build upon the material, they must distribute contributions under the same license as the original.