Ma Y J, Cao C Q, Zhang J. Steel surface defect detection based on YOLOv8-SOE[J]. Opto-Electron Eng, 2025, 52(5): 250032. doi: 10.12086/oee.2025.250032
Citation: Ma Y J, Cao C Q, Zhang J. Steel surface defect detection based on YOLOv8-SOE[J]. Opto-Electron Eng, 2025, 52(5): 250032. doi: 10.12086/oee.2025.250032

Steel surface defect detection based on YOLOv8-SOE

    Fund Project: Special Project for the Central Government to Guide Local Science and Technology Development-High-Level New R&D Institutions Fund (S202407a12020270), Anhui Province Key Research and Development Plan-High-tech Field Fund(202304a05020077)
More Information
  • In order to improve the detection capability of small target defects in steel surface inspection, an improved YOLOv8-SOE model is proposed. The model processes the P2 layer features by designing the FSCConv module. By compressing the P2 layer features and deeply fusing them with the P3 layer features, the model's sensitivity to small target features is effectively enhanced, while avoiding the computational burden caused by the introduction of additional detection layers. In order to further optimize the multi-scale feature fusion capability, cross stage partial omni-kernel (CSP-OK) module is used to optimize the multi-scale feature fusion, which improves the integration efficiency of features of different scales. The SIoU loss function is introduced to optimize the bounding box regression, which further improves the positioning accuracy. Experimental results show that the mAP of the YOLOv8-SOE model on the NEU-DET dataset achieves 80.7%, which is 5.4% higher than the baseline model, and has good generalization ability on the VOC2012 dataset. While improving the accuracy of small target detection, the model maintains a high computational efficiency and has good application prospects.
  • 加载中
  • [1] Luo Q W, Fang X X, Liu L, et al. Automated visual defect detection for flat steel surface: a survey[J]. IEEE Trans Instrum Meas, 2020, 69(3): 626−644. doi: 10.1109/TIM.2019.2963555

    CrossRef Google Scholar

    [2] Girshick R, Donahue J, Darrell T, et al. Rich feature hierarchies for accurate object detection and semantic segmentation[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2014: 580–587. https://doi.org/10.1109/CVPR.2014.81.

    Google Scholar

    [3] Girshick R. Fast R-CNN[C]//Proceedings of the IEEE International Conference on Computer Vision, 2015: 1440–1448. https://doi.org/10.1109/ICCV.2015.169.

    Google Scholar

    [4] Ren S Q, He K M, Girshick R, et al. Faster R-CNN: towards real-time object detection with region proposal networks[J]. IEEE Trans Patt Anal Mach Intell, 2017, 39(6): 1137−1149. doi: 10.1109/TPAMI.2016.2577031

    CrossRef Google Scholar

    [5] Shi X C, Zhou S K, Tai Y C, et al. An improved faster R-CNN for steel surface defect detection[C]//2022 IEEE 24th International Workshop on Multimedia Signal Processing (MMSP), 2022: 1–5. https://doi.org/10.1109/MMSP55362.2022.9949350.

    Google Scholar

    [6] Liu W, Anguelov D, Erhan D, et al. SSD: single shot multibox detector[C]//Proceedings of the 14th European Conference on Computer Vision, 2016: 21–37. https://doi.org/10.1007/978-3-319-46448-0_2.

    Google Scholar

    [7] Zhao Y A, Lv W Y, Xu S L, et al. DETRs beat YOLOs on real-time object detection[C]//Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2024: 16965–16974. https://doi.org/10.1109/CVPR52733.2024.01605.

    Google Scholar

    [8] Redmon J, Divvala S, Girshick R, et al. You only look once: Unified, real-time object detection[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2016: 779–788. https://doi.org/10.1109/CVPR.2016.91.

    Google Scholar

    [9] Redmon J, Farhadi A. YOLO9000: better, faster, stronger[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2017: 7263–7271. https://doi.org/10.1109/CVPR.2017.690.

    Google Scholar

    [10] Lin T Y, Goyal P, Girshick R, et al. Focal loss for dense object detection[C]//Proceedings of the IEEE International Conference on Computer Vision, 2017: 2980–2988. https://doi.org/10.1109/ICCV.2017.324.

    Google Scholar

    [11] Li J J, Chen M X. DEW-YOLO: an efficient algorithm for steel surface defect detection[J]. Appl Sci, 2024, 14(12): 5171. doi: 10.3390/app14125171

    CrossRef Google Scholar

    [12] Yang S X, Xie Y, Wu J Q, et al. CFE-YOLOv8s: improved YOLOv8s for steel surface defect detection[J]. Electronics, 2024, 13(14): 2771. doi: 10.3390/electronics13142771

    CrossRef Google Scholar

    [13] 彭菊红, 张弛, 高谦, 等. 基于改进的YOLOv8算法的钢材缺陷检测[J/OL]. 计算机工程, 1-9. https://doi.org/10.19678/j.issn.1000-3428.00EC0069283.

    Google Scholar

    Peng J H, Zhang C, Gao Q, et al. Steel defect detection basedon improved YOLOv8 algorithm[J/OL]. Comput Eng,1-9. https://doi.org/10.19678/j.issn.1000-3428.00EC0069283.

    Google Scholar

    [14] Zhang X R, Wang Y L, Fang H S. Steel surface defect detection algorithm based on ESI-YOLOv8[J]. Mater Res Express, 2024, 11(5): 056509. doi: 10.1088/2053-1591/ad46ec

    CrossRef Google Scholar

    [15] Huang Y, Tan W Z, Li L, et al. WFRE-YOLOv8s: a new type of defect detector for steel surfaces[J]. Coatings, 2023, 13(12): 2011. doi: 10.3390/coatings13122011

    CrossRef Google Scholar

    [16] Cui Y N, Ren W Q, Knoll A. Omni-kernel network for image restoration[C]//Proceedings of the 38th AAAI Conference on Artificial Intelligence, 2024: 1426–1434. https://doi.org/10.1609/aaai.v38i2.27907.

    Google Scholar

    [17] Gevorgyan Z. SIoU loss: more powerful learning for bounding box regression[Z]. arXiv: 2205.12740, 2022. https://arxiv.org/abs/2205.12740.

    Google Scholar

    [18] Wang C Y, Liao H Y M, Wu Y H, et al. CSPNet: a new backbone that can enhance learning capability of CNN[C]//Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2020: 390–391. https://doi.org/10.1109/CVPRW50498.2020.00203.

    Google Scholar

    [19] Zheng Z H, Wang P, Ren D W, et al. Enhancing geometric factors in model learning and inference for object detection and instance segmentation[J]. IEEE Trans Cybern, 2022, 52(8): 8574−8586. doi: 10.1109/TCYB.2021.3095305

    CrossRef Google Scholar

    [20] Tsai F J, Peng Y T, Lin Y Y, et al. Stripformer: strip transformer for fast image deblurring[C]//Proceedings of the 17th European Conference on Computer Vision, 2022: 146–162. https://doi.org/10.1007/978-3-031-19800-7_9.

    Google Scholar

    [21] Bao Y Q, Song K C, Liu J, et al. Triplet-graph reasoning network for few-shot metal generic surface defect segmentation[J]. IEEE Trans Instrum Meas, 2021, 70: 5011111. doi: 10.1109/TIM.2021.3083561

    CrossRef Google Scholar

    [22] Song K C, Yan Y H. A noise robust method based on completed local binary patterns for hot-rolled steel strip surface defects[J]. Appl Surf Sci, 2013, 285: 858−864. doi: 10.1016/j.apsusc.2013.09.002

    CrossRef Google Scholar

    [23] He Y, Song K C, Meng Q G, et al. An end-to-end steel surface defect detection approach via fusing multiple hierarchical features[J]. IEEE Trans Instrum Meas, 2020, 69(4): 1493−1504. doi: 10.1109/TIM.2019.2915404

    CrossRef Google Scholar

    [24] 梁礼明, 龙鹏威, 卢宝贺, 等. 改进GBS-YOLOv7t的钢材表面缺陷检测[J]. 光电工程, 2024, 51(5): 240044. doi: 10.12086/oee.2024.240044

    CrossRef Google Scholar

    Liang L M, Long P W, Lu B H, et al. Improvement of GBS-YOLOv7t for steel surface defect detection[J]. Opto-Electron Eng, 2024, 51(5): 240044. doi: 10.12086/oee.2024.240044

    CrossRef Google Scholar

    [25] 胡依伦, 杨俊, 许聪源, 等. PIC2f-YOLO: 金属表面缺陷检测轻量化方法[J]. 光电工程, 2025, 52(1): 240250. doi: 10.12086/oee.2025.240250

    CrossRef Google Scholar

    Hu Y L, Yang J, Xu C Y, et al. PIC2f-YOLO: a lightweight method for the detection of metal surface defects[J]. Opto-Electron Eng, 2025, 52(1): 240250. doi: 10.12086/oee.2025.240250

    CrossRef Google Scholar

    [26] 黄硕清, 黄金贵. 基于RFB和YOLOv5特征增强融合改进的钢材缺陷检测方法[J]. 计算机工程, 2025, 51(4): 249−260. doi: 10.19678/j.issn.1000-3428.0068476

    CrossRef Google Scholar

    Huang S Q, Huang J G. Improved steel defect detection method based on enhanced fusion of RFB and YOLOv5 features[J]. Comput Eng, 2025, 51(4): 249−260. doi: 10.19678/j.issn.1000-3428.0068476

    CrossRef Google Scholar

    [27] 阳丽莎, 李茂军, 胡建文, 等. 基于改进YOLOv7-tiny的带钢表面缺陷检测算法[J]. 计算机工程, 2025, 51(1): 208−215. doi: 10.19678/j.issn.1000-3428.0068397

    CrossRef Google Scholar

    Yang L S, Li M J, Hu J W, et al. Strip steel surface defect detection algorithm based on improved YOLOv7-tiny[J]. Comput Eng, 2025, 51(1): 208−215. doi: 10.19678/j.issn.1000-3428.0068397

    CrossRef Google Scholar

    [28] Everingham M, Eslami S M A, Van Gool L, et al. The pascal visual object classes challenge: A retrospective[J]. Int J Comput Vis, 2015, 111(1): 98−136. doi: 10.1007/s11263-014-0733-5

    CrossRef Google Scholar

  • In industrial applications such as steel surface defect detection, small target detection remains a challenging task due to the limited resolution of conventional detection layers, making it difficult to capture fine-grained defect details. Although YOLOv8 has shown remarkable performance in multi-scale target detection and complex environments, the model struggles with small target detection, especially when it comes to tiny defects on steel surfaces. Traditional solutions to enhance small target perception, such as adding additional detection heads like the P2 layer, often lead to increased computational overhead and inference time. To address this issue, this paper proposes an improved YOLOv8-SOE model that specifically enhances the detection performance for small defects on steel surfaces. The YOLOv8-SOE model incorporates several innovations aimed at improving both detection accuracy and computational efficiency. First, a novel feature processing module, FSCConv, is introduced to handle the P2 layer features. FSCConv leverages dilated convolutions to capture multi-scale contextual information while preserving fine details of small targets. This approach enhances small target perception without the need for additional detection layers, thus avoiding the computational burden typically associated with such modifications. Next, the processed P2 features are fused with P3 layer features, improving small target detection further without incurring significant computational costs. To optimize the feature fusion process, a cross-stage local network combined with Omni-Kernel (CSP-OK) is proposed. CSP-OK primarily leverages the CSPNet approach to reduce redundant gradient computations and integrates the Omni-Kernel to prevent the repetitive extraction of similar information at different layers, thereby improving information utilization efficiency. This optimization reduces redundant information computations and effectively utilizes inter-layer features, resulting in a more efficient and detailed integration of multi-scale information. In addition, the model uses the SIoU loss function for bounding box regression. This loss function not only considers the overlap between the predicted and ground truth boxes but also incorporates their distance, angular deviation, and shape similarity. By integrating these factors, the SIoU loss function provides a more comprehensive optimization strategy, thereby improving the accuracy of target localization. Experimental results demonstrate that the YOLOv8-SOE model achieves a mean average precision (mAP) of 80.7% on the NEU-DET dataset, a 5.4% improvement over the baseline YOLOv8 model. The model also exhibits excellent generalization ability on the VOC2012 dataset. Overall, the proposed YOLOv8-SOE model significantly enhances small target detection precision while maintaining high computational efficiency, making it a promising solution for real-world industrial defect detection applications.

  • 加载中
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Figures(10)

Tables(7)

Article Metrics

Article views() PDF downloads() Cited by()

Access History

Other Articles By Authors

Article Contents

Catalog

    /

    DownLoad:  Full-Size Img  PowerPoint