Citation: | Sun G L, Wang X B. THI-YOLO: Improved non-motorized drivers helmet detection of YOLOv8[J]. Opto-Electron Eng, 2024, 51(12): 240244. doi: 10.12086/oee.2024.240244 |
[1] | 李斌, 侯德藻, 张纪升, 等. 论智能车路协同的概念与机理[J]. 公路交通科技, 2020, 37 (10): 134−141. doi: 10.3969/j.issn.1002-0268.2020.10.015 Li B, Hou D Z, Zhang J S, et al. Study on conception and mechanism of intelligent vehicle-infrastructure cooperation[J]. J Highw Transp Res Dev, 2020, 37 (10): 134−141. doi: 10.3969/j.issn.1002-0268.2020.10.015 |
[2] | 岑晏青, 宋向辉, 王东柱, 等. 智慧高速公路技术体系构建[J]. 公路交通科技, 2020, 37 (7): 111−121. doi: 10.3969/j.issn.1002-0268.2020.07.015 Cen Y Q, Song X H, Wang D Z, et al. Establishment of technology system of smart expressway[J]. J Highw Transp Res Dev, 2020, 37 (7): 111−121. doi: 10.3969/j.issn.1002-0268.2020.07.015 |
[3] | Ren S, He K, Girshick R, et al. Faster R-CNN: towards real-time object detection with region proposal networks[C]//Proceedings of the 28th International Conference on Neural Information Processing Systems, 2015, 1 : 91–99. |
[4] | Liu W, Anguelov D, Erhan D, et al. SSD: single shot MultiBox detector[C]//Proceedings of the 14th European Conference on Computer Vision (ECCV), 2016: 21–37. https://doi.org/10.1007/978-3-319-46448-0_2. |
[5] | Carion N, Massa F, Synnaeve G, et al. End-to-end object detection with transformers[C]//Proceedings of the 16th European Conference on Computer Vision, 2020: 213–229. https://doi.org/10.1007/978-3-030-58452-8_13. |
[6] | 刘超, 高健. 一种改进的YOLOv5电动车头盔佩戴检测方法[J]. 软件导刊, 2024, 23 (6): 143−149. doi: 10.11907/rjdk.231506 Liu C, Gao J. An improved YOLOv5 electric vehicle helmet wearing detection method[J]. Software Guide, 2024, 23 (6): 143−149. doi: 10.11907/rjdk.231506 |
[7] | 王茹, 刘大明, 张健. Wear-YOLO: 变电站电力人员安全装备检测方法研究[J]. 计算机工程与应用, 2024, 60 (9): 111−121. doi: 10.3778/j.issn.1002-8331.2308-0317 Wang R, Liu D M, Zhang J. Wear-YOLO: research on detection methods of safety equipment for power personnel in substations[J]. Comput Eng Appl, 2024, 60 (9): 111−121. doi: 10.3778/j.issn.1002-8331.2308-0317 |
[8] | 王鹏飞, 黄汉明, 王梦琪. 改进YOLOv5的复杂道路目标检测算法[J]. 计算机工程与应用, 2022, 58 (17): 81−92. doi: 10.3778/j.issn.1002-8331.2205-0158 Wang P F, Huang H M, Wang M Q. Complex road target detection algorithm based on improved YOLOv5[J]. Comput Eng Appl, 2022, 58 (17): 81−92. doi: 10.3778/j.issn.1002-8331.2205-0158 |
[9] | 周顺勇, 彭梓洋, 张航领, 等. Helmet-YOLO: 一种更高精度的道路安全头盔检测算法[J]. 计算机工程与应用, 2024, . doi: http://kns.cnki.net/kcms/detail/11.2127.TP.20240909.1514.018.html Zhou S Y, Peng Z Y, Zhang H L, et al. Helmet-YOLO: a higher-accuracy road safety helmet detection algorithm[J]. Comput Eng Appl, 2024, . doi: http://kns.cnki.net/kcms/detail/11.2127.TP.20240909.1514.018.html |
[10] | Li H L, Li J, Wei H B, et al. Slim-neck by GSConv: a better design paradigm of detector architectures for autonomous vehicles[Z]. arXiv: 2206.02424, 2022. https://doi.org/10.48550/arXiv.2206.02424. |
[11] | Woo S, Park J, Lee J Y, et al. CBAM: convolutional block attention module[C]//Proceedings of the 15th European Conference on Computer Vision (ECCV), 2018: 3–19. https://doi.org/10.1007/978-3-030-01234-2_1. |
[12] | Goyal A, Bochkovskiy A, Deng J, et al. Non-deep networks[C]//Proceedings of the 36th International Conference on Neural Information Processing Systems, 2022: 492. |
[13] | Chollet F. Xception: deep learning with depthwise separable convolutions[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2017: 1251–1258. https://doi.org/10.1109/CVPR.2017.195. |
[14] | Zhang H, Zhang S J. Focaler-IoU: more focused intersection over union loss[Z]. arXiv: 2401.10525, 2024. https://doi.org/10.48550/arXiv.2401.10525. |
[15] | 陈旭, 彭冬亮, 谷雨. 基于改进YOLOv5s的无人机图像实时目标检测[J]. 光电工程, 2022, 49 (3): 210372. doi: 10.12086/oee.2022.210372 Chen X, Peng D L, Gu Y. Real-time object detection for UAV images based on improved YOLOv5s[J]. Opto-Electron Eng, 2022, 49 (3): 210372. doi: 10.12086/oee.2022.210372 |
[16] | Han K, Wang Y H, Tian Q, et al. GhostNet: more features from cheap operations[C]//Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020: 1577–1586. https://doi.org/10.1109/CVPR42600.2020.00165. |
[17] | Zhang X Y, Zhou X Y, Lin M X, et al. ShuffleNet: an extremely efficient convolutional neural network for mobile devices[C]// Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2018: 6848–6856. https://doi.org/10.1109/CVPR.2018.00716. |
[18] | Zhang H, Xu C, Zhang S J. Inner-IoU: more effective intersection over union loss with auxiliary bounding box[Z]. arXiv: 2311.02877, 2023. https://doi.org/10.48550/arXiv.2311.02877. |
[19] | Hu J, Shen L, Sun G. Squeeze-and-excitation networks[C]//Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2018: 7132–7141. https://doi.org/10.1109/CVPR.2018.00745. |
[20] | Liu Y C, Shao Z R, Hoffmann N. Global attention mechanism: retain information to enhance channel-spatial interactions[Z]. arXiv: 2112.05561, 2021. https://doi.org/10.48550/arXiv.2112.05561. |
[21] | Ouyang D L, He S, Zhang G Z, et al. Efficient multi-scale attention module with cross-spatial learning[C]//ICASSP 2023–2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2023: 1–5. https://doi.org/10.1109/ICASSP49357.2023.10096516. |
[22] | Wang A, Chen H, Liu L H, et al. YOLOv10: real-time end-to-end object detection[Z]. arXiv: 2405.14458, 2024. https://doi.org/10.48550/arXiv.2405.14458. |
[23] | Lv W Y, Xu S L, Zhao Y A, et al. DETRs beat YOLOs on real-time object detection[Z]. arXiv: 2304.08069, 2023. https://doi.org/10.48550/arXiv.2304.08069. |
[24] | 熊恩杰, 张荣芬, 刘宇红, 等. 面向交通标志的Ghost-YOLOv8检测算法[J]. 计算机工程与应用, 2023, 59 (20): 200−207. doi: 10.3778/j.issn.1002-8331.2306-0032 Xiong E J, Zhang R F, Liu Y H, et al. Ghost-YOLOv8 detection algorithm for traffic signs[J]. Comput Eng Appl, 2023, 59 (20): 200−207. doi: 10.3778/j.issn.1002-8331.2306-0032 |
With the gradual emergence of computer vision in recent years, target detection algorithms are also constantly innovating and developing, and the detection of safety helmets on two-wheeled vehicles is also one of the focuses of research scholars. Two-wheeled vehicles have become the main means of transportation for citizens to travel at this stage, but the phenomena of citizens running red lights and not wearing safety helmets are still common, so it is especially urgent to design a helmet-wearing detection method for cyclists. At the present stage of safety helmet detection, there are still some difficulties, such as complex background information, the detection target exists in different scales of the diversity of changes, so the design of higher performance helmet detection algorithms needs to carry out further research. Aiming at the phenomenon that complex background often exists in non-motorized drivers' helmet detection and the diversity of detection target scales often exists in the detection scene, which in turn leads to low detection efficiency and misdetection and omission, a YOLOv8 algorithm oriented to traffic helmet detection is proposed. The alteration points of this paper mainly contain the following two blocks. The first is the C2f_BC module, the GSConv module with the idea of combining group convolution and spatial convolution is introduced, and the attention mechanism combining channel and space (convolutional block attention module, CBAM) is introduced in Bottleneck in C2f. To effectively reduce the computational complexity and enhance the extraction of local and global features, we designed the parallel multiscale feature fusion module MP-Parnet (parallel multiscale perception networks) and redesigned the Parnet (parallel networks) by using the parallel depth-separable feature fusion module with different scales. The second is a parallel depthwise separable convolution (DWConv) kernel for different scales. It is used instead of the ordinary convolution of the original module, which effectively adapts to the acquisition ability of different scales of targets. 3. Focaler-IoU is introduced into the original model, and Focaler-CIoU is designed, which effectively enhances the performance of the detection model in both classification and detection. detection performance. The experimental results show that compared with the original model, the improved YOLOv8n maintains a decrease in the number of parameters while the mAP50 and mAP50: 95 increase by 2.2% and 1.9% on the self-built dataset Helmet, and 1.8% and 1.9% on the open-source dataset TWHD, which suggests that the improved model can be better applied to the helmet detection of non-motorized drivers in the scenario.
THI-YOLO model
Structure of GSConv module
CBAM structure
Channel attention submodule
Spatial attention submodule
Diagram of the C2f_BC structure
Diagram of the Bottleneck_GC structure
Diagram of the MP-Parnet structure
Labelimg script interface
Scatterplots of length and width distribution of different categories
Comparison of box_loss before and after improvement
Display of qualitative experiment results. (a) Cases of object leakage; (b) Cases of object misdetections; (c) Comparison of detection effects for dense scenes of drivers