• 摘要: 针对锂电池电芯外观缺陷检测中目标尺寸小、缺陷类型复杂多样,以及实际生产环境下数据采集困难等问题,提出一种基于结构光成像与改进YOLOv8的检测方法,提升检测精度与鲁棒性。首先,在骨干网络中引入CMUNeXt Block,通过大核卷积增强空间信息交互,提升特征提取能力;其次,在颈部网络中设计C2f_RFAConv模块,引入动态参数生成机制,缓解参数共享带来的表达瓶颈,并增强特征传递中的信息保留;最后,改进检测头结构,构建ASFF_F模块,融合自适应空间特征融合策略,重构特征金字塔路径,实现新增微小目标检测分支,强化多尺度细节感知能力。实验结果表明,所提模型平均检测精度达97.98%,准确率和召回率分别为97.05%和92.67%,相较基准模型YOLOv8分别提升了4.29%、1.97%和6.46%。在NEU-DET公开数据集上的扩展实验亦验证了其良好的泛化性能,对比基准模型,平均检测精度提高了0.65%,表明该方法具备较高的工程应用价值。

       

      Abstract:
      Objective Lithium-ion batteries are critical components in modern energy systems, and their safety and reliability heavily depend on the quality of cell manufacturing. Detecting surface defects on battery cells—such as scratches, pits, wrinkles, and corrosion—is essential yet challenging due to the small size, high diversity, and complex morphology of these defects. Traditional inspection methods, including manual visual checks or conventional 2D imaging, struggle with low accuracy, high false/missed detection rates, and sensitivity to lighting variations in real-world production environments. Moreover, data acquisition under industrial conditions is often hindered by reflective surfaces (e.g., aluminum-plastic film) and inconsistent illumination. To address these issues, this study aims to develop a robust, high-precision defect detection model tailored for structured light imaging, which provides stable, high-contrast 3D-like surface information while mitigating environmental interference.
      Methods We propose YOLO-SLB, an enhanced YOLOv8 architecture specifically designed for micro-defect detection on battery cells using structured light images. The model introduces three key innovations. First, the backbone integrates CMUNeXt Blocks featuring large-kernel depthwise convolutions in an inverted bottleneck structure to strengthen spatial feature interaction and preserve fine details critical for detecting minute anomalies. Second, the neck employs a novel C2f_RFAConv module that replaces standard convolutions with receptive field attention convolution (RFAConv), which dynamically generates convolutional weights to overcome the limitations of static weight sharing in large kernels, thereby improving fusion between shallow spatial and deep semantic features. Third, the detection head is redesigned as ASFF_F—a four-branch adaptive spatial feature fusion module that not only applies learnable scale-wise weights but also adds a dedicated micro-target detection path directly utilizing high-resolution shallow features. This reconstruction enhances multi-scale perception and reduces information loss during feature pyramid operations. The model is evaluated on SLBD, a new dataset of 477 original high-resolution (1520 pixel×2000 pixel) battery cell images captured via line-scan cameras under structured light, augmented to 3147 samples covering eight common defect types.
      Results and Discussions The experimental results show that the proposed model achieves an average detection accuracy of 97.98%, with precision and recall rates of 97.05% and 92.67%, respectively, which are 4.29%, 1.97%, and 6.46% higher than those of the benchmark model YOLOv8. Extensive experiments on the NEU-DET public dataset also verified its good generalization performance, with an average detection accuracy improvement of 0.65% compared to the benchmark model, indicating that this method has high engineering application value.
      Conclusions YOLO-SLB effectively addresses the core challenges of micro-defect detection in lithium battery cells by synergizing structured light imaging with architectural innovations in feature extraction, fusion, and detection. The integration of CMUNeXt, C2f_RFAConv, and ASFF_F significantly boosts sensitivity to small anomalies without compromising robustness. The model's high accuracy, generalization capability, and real-time performance underscore its engineering value for automated quality control in battery manufacturing. Future work will explore semi-supervised learning and online model updating to enable continuous improvement using unlabeled production data.