Welcome to Journal of Graphics share: 

Journal of Graphics ›› 2025, Vol. 46 ›› Issue (4): 719-726.DOI: 10.11996/JG.j.2095-302X.2025040719

• Image Processing and Computer Vision • Previous Articles     Next Articles

An object detection algorithm for powerline inspection based on the feature focus & diffusion network

GUO Ruidong1(), LAN Guiwen1,2(), FAN Donglin1, ZHONG Zhan1, XU Zirui1, REN Xinyue1   

  1. 1. College of Geomatics and Geoinformation, Guilin University of Technology, Guilin Guangxi 541006, China
    2. Guangxi Key Laboratory of Spatial Information and Geomatics, Guilin Guangxi 541006, China
  • Received:2024-10-09 Revised:2025-01-07 Online:2025-08-30 Published:2025-08-11
  • Contact: LAN Guiwen
  • About author:First author contact:

    GUO Ruidong (1995-), master student. His main research interests cover image processing, computer vision, etc. E-mail:13028689662@163.com

  • Supported by:
    National Natural Science Foundation of China(41861050);Guangxi Natural Science Foundation(2022GXNSFBA035637)

Abstract:

UAV images for powerline inspection usually have complex backgrounds, and often contain a lot of small targets, which may lead to a high rate of missed detections and false detections when processed by the general feature extraction networks for object detection. To address this, a feature focus & diffusion network (FFDN) was proposed for feature fusion, and an improved algorithm (YOLOv8-SFD) based on FFDN and YOLOv8 was designed for powerline component detection. Spatial-to-depth non-stride convolutions (SPDConv) were employed in the backbone network instead to preserve small-scale features and reduce feature loss caused by stride convolutions. The traditional feature pyramid network was replaced with the proposed FFDN. At the feature fusion stage, the feature focus modules in the FFDN were utilized to expand the receptive field and fuse multi-scale features, and the output feature maps by them were then diffused across different scales to enhance small target detection accuracy. Finally, the original YOLOv8 head was replaced with a dynamic detection head (DyHead) that integrates three attention mechanisms (scale, space, and task), to further enhance the performance. Experimental results demonstrated that YOLOv8-SFD achieved an accuracy rate of 76.7%, which was 7.6% higher than YOLOv8n; a recall rate of 43.0%, which was 2.0% higher; and a MAP of 48.2%, which was 3.8%. YOLOv8-SFD effectively enhanced the detection precision for small and obscured targets, and the detection speed reached 119 FPS, which satisfied real-time detection requirements.

Key words: powerline inspection, feature fusion, object detection, feature focus & diffusion network, YOLOv8

CLC Number: