Xiao Z J, Zhang J H, Lin B H. Feature coordination and fine-grained perception of small targets in remote sensing images[J]. Opto-Electron Eng, 2024, 51(6): 240066. doi: 10.12086/oee.2024.240066
Citation: Xiao Z J, Zhang J H, Lin B H. Feature coordination and fine-grained perception of small targets in remote sensing images[J]. Opto-Electron Eng, 2024, 51(6): 240066. doi: 10.12086/oee.2024.240066

Feature coordination and fine-grained perception of small targets in remote sensing images

    Fund Project: Project supported by Basic Scientific Research Project of Liaoning Provincial Universities (LJKMZ20220699), and Subject Innovation Team Project of Liaoning Technical University (LNTU20TD-23)
More Information
  • Addressing the challenge of missed detection caused by many small targets and dense arrangement in remote sensing images, this study introduces a small target detection algorithm for remote sensing applications, leveraging a combination of feature synergy and micro-perception strategies. Initially, we propose a refined feature synergistic fusion strategy that optimizes the interaction and integration of features across different scales by intelligently adjusting the parameters of convolution kernels. This strategy facilitates progressive refinement of features from coarse to fine granularity. Building upon this foundation, a micro-perception unit is developed in this paper, incorporating perceptual attention mechanisms with moving inverse convolution to form an advanced detection head. This innovative approach substantially boosts the network's capability to detect very small objects. Furthermore, to augment the training efficiency of the model, we employ MPDIoU and NWD as regression loss functions, mitigating positional bias issues and expediting model convergence. Experimental evaluations on the DOTA1.0 dataset and DOTA1.5 dataset reveal that our algorithm achieves a substantial improvement in mean Average Precision (mAP) by 7.4% and 6.1% over the baseline method, which has obvious advantages over other algorithms. The results underscore the algorithm's efficacy in significantly reducing the incidence of missed detections of small targets within remote sensing imagery.
  • 加载中
  • [1] 马梁, 苟于涛, 雷涛, 等. 基于多尺度特征融合的遥感图像小目标检测[J]. 光电工程, 2022, 49(4): 210363. doi: 10.12086/oee.2022.210363

    CrossRef Google Scholar

    Ma L, Gou Y T, Lei T, et al. Small object detection based on multi-scale feature fusion using remote sensing images[J]. Opto-Electron Eng, 2022, 49(4): 210363. doi: 10.12086/oee.2022.210363

    CrossRef Google Scholar

    [2] 陈旭, 彭冬亮, 谷雨. 基于改进YOLOv5s的无人机图像实时目标检测[J]. 光电工程, 2022, 49(3): 210372. doi: 10.12086/oee.2022.210372

    CrossRef Google Scholar

    Chen X, Peng D L, Gu Y. Real-time object detection for UAV images based on improved YOLOv5s[J]. Opto-Electron Eng, 2022, 49(3): 210372. doi: 10.12086/oee.2022.210372

    CrossRef Google Scholar

    [3] 王友伟, 郭颖, 邵香迎. 基于改进级联算法的遥感图像目标检测[J]. 光学学报, 2022, 42(24): 2428004. doi: 10.3788/AOS202242.2428004

    CrossRef Google Scholar

    Wang Y W, Guo Y, Shao X Y. Target detection in remote sensing images based on improved cascade algorithm[J]. Acta Opt Sin, 2022, 42(24): 2428004. doi: 10.3788/AOS202242.2428004

    CrossRef Google Scholar

    [4] 王家宝, 程塨, 谢星星, 等. 多元信息监督的遥感图像有向目标检测[J]. 遥感学报, 2023, 27(12): 2726−2735. doi: 10.11834/jrs.20211564

    CrossRef Google Scholar

    Wang J B, Cheng G, Xie X X, et al. Multi-information supervision in optical remote sensing images[J]. Natl Remote Sens Bull, 2023, 27(12): 2726−2735. doi: 10.11834/jrs.20211564

    CrossRef Google Scholar

    [5] 张德银, 赵志恒, 谢逸戈, 等. 基于改进YOLOv8的遥感图像飞机目标检测研究[J]. 自动化应用, 2024, 65(2): 193−195,198. doi: 10.19769/j.zdhy.2024.02.060

    CrossRef Google Scholar

    Zhang D Y, Zhao Z H, Xie Y G, et al. Research on aircraft target detection in remote sensing images based on improved YOLOv8[J]. Autom Appl, 2024, 65(2): 193−195,198. doi: 10.19769/j.zdhy.2024.02.060

    CrossRef Google Scholar

    [6] Zhang Y L, Jin H Y. Detector consistency research on remote sensing object detection[J]. Remote Sens, 2023, 15(17): 4130. doi: 10.3390/rs15174130

    CrossRef Google Scholar

    [7] Lyu Z, Jin H F, Zhen T, et al. Small object recognition algorithm of grain pests based on SSD feature fusion[J]. IEEE Access, 2021, 9: 43202−43213. doi: 10.1109/ACCESS.2021.3066510

    CrossRef Google Scholar

    [8] Lin T Y, Goyal P, Girshick R, et al. Focal loss for dense object detection[J]. IEEE Trans Pattern Anal Mach Intell, 2020, 42(2): 318−327. doi: 10.1109/TPAMI.2018.2858826

    CrossRef Google Scholar

    [9] Redmon J, Divvala S, Girshick R, et al. You only look once: unified, real-time object detection[C]//Proceedings of 2016 IEEE Conference on Computer Vision and Pattern Recognition, 2016: 779–788. https://doi.org/10.1109/CVPR.2016.91.

    Google Scholar

    [10] Redmon J, Farhadi A. YOLO9000: better, faster, stronger[C]//Proceedings of 2017 IEEE Conference on Computer Vision and Pattern Recognition, 2017: 6517–6525. https://doi.org/10.1109/CVPR.2017.690.

    Google Scholar

    [11] Redmon J, Farhadi A. YOLOv3: an incremental improvement[Z]. arXiv: 1804.02767, 2018. https://doi.org/10.48550/arXiv.1804.02767.

    Google Scholar

    [12] Bochkovskiy A, Wang C Y, Liao H Y M. YOLOv4: optimal speed and accuracy of object detection[Z]. arXiv: 2004.10934, 2020. https://doi.org/10.48550/arXiv.2004.10934.

    Google Scholar

    [13] Ge Z, Liu S T, Wang F, et al. YOLOX: exceeding YOLO series in 2021[Z]. arXiv: 2107.08430, 2021. https://doi.org/10.48550/arXiv.2107.08430.

    Google Scholar

    [14] Girshick R. Fast R-CNN[C]//Proceedings of 2015 IEEE International Conference on Computer Vision, 2015: 1440–1448. https://doi.org/10.1109/ICCV.2015.169.

    Google Scholar

    [15] Dai J F, Li Y, He K M, et al. R-FCN: object detection via region-based fully convolutional networks[C]//Proceedings of the 30th International Conference on Neural Information Processing Systems, 2016: 29.

    Google Scholar

    [16] Li Y T, Fan Q S, Huang H S, et al. A modified YOLOv8 detection network for UAV aerial image recognition[J]. Drones, 2023, 7(5): 304. doi: 10.3390/drones7050304

    CrossRef Google Scholar

    [17] Zhu F Z, Wang Y Y, Cui J Y, et al. Target detection for remote sensing based on the enhanced YOLOv4 with improved BiFPN[J]. Egypt J Remote Sens Space Sci, 2023, 26(2): 351−360. doi: 10.1016/j.ejrs.2023.04.003

    CrossRef Google Scholar

    [18] Zhai X X, Huang Z H, Li T, et al. YOLO-Drone: an optimized YOLOv8 network for tiny UAV object detection[J]. Electronics, 2023, 12(17): 3664. doi: 10.3390/electronics12173664

    CrossRef Google Scholar

    [19] Zhou F Y, Deng H G, Xu Q G, et al. CNTR-YOLO: improved YOLOv5 based on ConvNext and transformer for aircraft detection in remote sensing images[J]. Electronics, 2023, 12(12): 2671. doi: 10.3390/electronics12122671

    CrossRef Google Scholar

    [20] Zhu B Y, Lv Q B, Yang Y B, et al. Gradient structure information-guided attention generative adversarial networks for remote sensing image generation[J]. Remote Sens, 2023, 15(11): 2827. doi: 10.3390/rs15112827

    CrossRef Google Scholar

    [21] Xiao J S, Guo H W, Yao Y T, et al. Multi-scale object detection with the pixel attention mechanism in a complex background[J]. Remote Sens, 2022, 14(16): 3969. doi: 10.3390/rs14163969

    CrossRef Google Scholar

    [22] Wu J J, Su L M, Lin Z W, et al. Object detection of flexible objects with arbitrary orientation based on rotation-adaptive YOLOv5[J]. Sensors, 2023, 23(10): 4925. doi: 10.3390/s23104925

    CrossRef Google Scholar

    [23] Yang X, Yan J C, Liao W L, et al. SCRDet++: detecting small, cluttered and rotated objects via instance-level feature denoising and rotation loss smoothing[J]. IEEE Trans Pattern Anal Mach Intell, 2023, 45(2): 2384−2399. doi: 10.1109/TPAMI.2022.3166956

    CrossRef Google Scholar

    [24] Zhang H Y, Liu J. Direction estimation of aerial image object based on neural network[J]. Remote Sens, 2022, 14(15): 3523. doi: 10.3390/rs14153523

    CrossRef Google Scholar

    [25] Yu H T, Tian Y J, Ye Q X, et al. Spatial transform decoupling for oriented object detection[C]//Proceedings of the 38th AAAI Conference on Artificial Intelligence, 2024: 6782–6790. https://doi.org/10.1609/aaai.v38i7.28502.

    Google Scholar

    [26] Li C, Yao A B. KernelWarehouse: towards parameter-efficient dynamic convolution[Z]. arXiv: 2308.08361, 2023. https://doi.org/10.48550/arXiv.2308.08361.

    Google Scholar

    [27] Ma S L, Xu Y. MPDIoU: a loss for efficient and accurate bounding box regression[Z]. arXiv: 2307.07662, 2023. https://doi.org/10.48550/arXiv.2307.07662.

    Google Scholar

    [28] Wang J W, Xu C, Yang W, et al. A normalized Gaussian Wasserstein distance for tiny object detection[Z]. arXiv: 2110.13389, 2021. https://doi.org/10.48550/arXiv.2110.13389.

    Google Scholar

    [29] Sandler M, Howard A, Zhu M L, et al. MobileNetV2: inverted residuals and linear bottlenecks[C]//Proceedings of 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2018: 4510–4520.

    Google Scholar

    [30] Konstantinov A V, Utkin L V, Lukashin A A, et al. Neural attention forests: transformer-based forest improvement[C]//Proceedings of the Seventh International Scientific Conference “Intelligent Information Technologies for Industry”, 2023. https://doi.org/10.1007/978-3-031-43789-2_14.

    Google Scholar

    [31] Tan M X, Le Q V. EfficientNet: rethinking model scaling for convolutional neural networks[C]//Proceedings of the 36th International Conference on Machine Learning, 2019: 6105–6114.

    Google Scholar

    [32] Law H, Deng J. CornerNet: detecting objects as paired keypoints[C]//Proceedings of the 15th European Conference on Computer Vision (ECCV), 2018: 734–750. https://doi.org/10.1007/978-3-030-01264-9_45.

    Google Scholar

    [33] 于傲泽, 魏维伟, 王平, 等. 基于分块复合注意力的无人机小目标检测算法[J]. 航空学报, 2023: 1−12. doi: 10.7527/S1000-6893.2023.29148

    CrossRef Google Scholar

    Yu A Z, Wei W W, Wang P, et al. Small target detection algorithm for UAV based on patch-wise co-attention[J]. Acta Aeronaut Astronaut Sin, 2023: 1−12. doi: 10.7527/S1000-6893.2023.29148

    CrossRef Google Scholar

    [34] Min L T, Fan Z M, Lv Q Y, et al. YOLO-DCTI: small object detection in remote sensing base on contextual transformer enhancement[J]. Remote Sens, 2023, 15(16): 3970. doi: 10.3390/rs15163970

    CrossRef Google Scholar

  • With the rapid development of remote sensing image technology, remote sensing image target detection is widely used in many important fields, including military target location and identification, natural environment protection, disaster detection, and urban planning and construction. The task of remote sensing image target detection is to accurately identify and locate the specific target in the image, and speculate its type and position. Different from targets in natural scenes, targets in remote sensing images have the characteristics of large scenes, small targets, multi-scale, complex backgrounds, overlapping occlusion, etc., so it is a challenging task to detect specific objects accurately. At present, great breakthroughs have been made in remote sensing image target detection algorithms, but the effect of small target detection is still not ideal. Small target detection faces two major difficulties: Little feature information of the target, scarce positive samples, and unbalanced classification; The target location is difficult, the background is complex, and contains a lot of redundant information, which causes serious interference to the detection. This makes it challenging to extract the edge features from aerial images and distinguish the object from the background. Therefore, the research on object detection and application in remote sensing images has important theoretical and practical significance. Addressing the challenge of missed detection caused by many small targets and dense arrangement in remote sensing images, this study introduces a small target detection algorithm for remote sensing applications, leveraging a combination of feature synergy and micro-perception strategies. Initially, we propose a refined feature synergistic fusion strategy that optimizes the interaction and integration of features across different scales by intelligently adjusting the parameters of convolution kernels. This strategy facilitates progressive refinement of features from coarse to fine granularity. Building upon this foundation, a micro-perception unit is developed in this paper, incorporating perceptual attention mechanisms with moving inverse convolution to form an advanced detection head. This innovative approach substantially boosts the network's capability to detect very small objects. Furthermore, to augment the training efficiency of the model, we employ MPDIoU and NWD as regression loss functions, mitigating positional bias issues and expediting model convergence. Experimental evaluations on the DOTA1.0 dataset and DOTA1.5 dataset reveal that our algorithm substantially improves mean Average Precision (mAP) by 7.4% and 6.1% over the baseline method, which has obvious advantages over other algorithms. The results underscore the algorithm's efficacy in significantly reducing the incidence of missed detections of small targets within remote sensing imagery.

  • 加载中
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Figures(9)

Tables(3)

Article Metrics

Article views() PDF downloads() Cited by()

Access History

Other Articles By Authors

Article Contents

Catalog

    /

    DownLoad:  Full-Size Img  PowerPoint