New website getting online, testing
    • 摘要: 针对无人机航拍视角下的道路损伤图像背景复杂、目标尺度差异大的检测难题,提出了一种融合多分支混合注意力机制的道路损伤检测方法MAS-YOLOv8n。首先,设计了一种多分支混合注意力机制,并将该结构添加到C2f结构中,加强了特征的表达能力,在捕获到更为丰富的特征信息的同时,减少噪声对检测结果的影响,以解决YOLOv8n模型中残差结构易受干扰,导致信息丢失的问题。其次,针对道路损伤形态差异大导致检测效果差的问题,利用ShapeIoU对YOLOv8n模型使用的TaskAlignedAssigner标签分配算法进行改进,使其更适用于形态多变的目标,进一步提高了检测精度。将MAS-YOLOv8n模型在无人机拍摄的道路损伤数据集China-Drone上进行实验,相较于基线模型YOLOv8n,本文模型的平均精度均值提高了3.1%,且没有额外增加计算代价。为进一步验证模型通用性,在RDD2022_Chinese和RDD2022_Japanese两个数据集上进行实验,精度均有所提升。与YOLOv5n、YOLOv8n、YOLOv10n、GOLD-YOLO、Faster-RCNN、TOOD、RTMDet-Tiny、RT-DETR相比,本文模型检测精度更高,性能更为优秀,展现了其较好的泛化能力。

       

      Abstract: To address the detection challenges posed by the complex backgrounds and significant variations in target scales in road damage images captured from drone aerial perspectives, a road damage detection method called MAS-YOLOv8n, incorporating a multi-branch hybrid attention mechanism, is proposed. Firstly, to address the problem of the residual structure in the YOLOv8n model being prone to interference, resulting in information loss, a multi-branch mixed attention (MBMA) mechanism is introduced. This MBMA structure is integrated into the C2f structure, strengthening the feature representation capabilities. It not only captures richer feature information but also reduces the impact of noise on the detection results. Secondly, to address the issue of poor detection performance resulting from significant variations in road damage morphologies, the TaskAlignedAssigner label assignment algorithm used in the YOLOv8n model is improved by utilizing ShapeIoU (shape-intersection over union), making it more suitable for targets with diverse shapes and further enhancing detection accuracy. Experimental evaluations of the MAS-YOLOv8n model on the China-Drone dataset of road damages captured by drones reveal that compared to the baseline YOLOv8n model, our model achieves a 3.1% increase in mean average precision (mAP) without incurring additional computational costs. To further validate the model's generalizability, tests on the RDD2022_Chinese and RDD2022_Japanese datasets also demonstrate improved accuracy. Compared to YOLOv5n, YOLOv8n, YOLOv10n, GOLD-YOLO, Faster-RCNN, TOOD, RTMDet-Tiny, and RT-DETR, our model exhibits superior detection accuracy and performance, showcasing its robust generalization capabilities.