欢迎访问《图学学报》 分享到:

图学学报 ›› 2025, Vol. 46 ›› Issue (4): 719-726.DOI: 10.11996/JG.j.2095-302X.2025040719

• 图像处理与计算机视觉 • 上一篇    下一篇

基于特征聚焦扩散网络的电力巡检目标检测算法

郭瑞东1(), 蓝贵文1,2(), 范冬林1, 钟展1, 徐梓睿1, 任新月1   

  1. 1.桂林理工大学测绘地理信息学院,广西 桂林 541006
    2.广西空间信息与测绘重点实验室,广西 桂林 541006
  • 收稿日期:2024-10-09 修回日期:2025-01-07 出版日期:2025-08-30 发布日期:2025-08-11
  • 通讯作者:蓝贵文(1977-),男,教授,博士。主要研究方向为地理信息系统、地理空间智能等。E-mail:2009043@glut.edu.cn
  • 第一作者:郭瑞东(1995-),男,硕士研究生。主要研究方向为图像处理、计算机视觉等。E-mail:13028689662@163.com
  • 基金资助:
    国家自然科学基金(41861050);广西自然科学基金(2022GXNSFBA035637)

An object detection algorithm for powerline inspection based on the feature focus & diffusion network

GUO Ruidong1(), LAN Guiwen1,2(), FAN Donglin1, ZHONG Zhan1, XU Zirui1, REN Xinyue1   

  1. 1. College of Geomatics and Geoinformation, Guilin University of Technology, Guilin Guangxi 541006, China
    2. Guangxi Key Laboratory of Spatial Information and Geomatics, Guilin Guangxi 541006, China
  • Received:2024-10-09 Revised:2025-01-07 Published:2025-08-30 Online:2025-08-11
  • First author:GUO Ruidong (1995-), master student. His main research interests cover image processing, computer vision, etc. E-mail:13028689662@163.com
  • Supported by:
    National Natural Science Foundation of China(41861050);Guangxi Natural Science Foundation(2022GXNSFBA035637)

摘要:

针对无人机电力巡检航拍图像通常背景复杂、待检目标尺寸较小,使用通用的特征提取网络往往导致漏检、误检率较高,提出了一种用于特征融合的特征聚焦扩散网络(FFDN),利用FFDN对YOLOv8算法进行改进,设计了基于YOLOv8的电力巡检目标检测算法YOLOv8-SFD。在主干网络采用空间到深度非跨步卷积(SPDConv)保留更多小尺度特征,减少跨步卷积造成的特征损失,避免小目标以及遮挡目标特征提取过程中的特征丢失;在特征融合阶段使用FFDN代替传统的特征金字塔网络,利用特征聚焦模块融合多尺度特征并扩大感受野,将特征聚焦模块输出的多尺度特征图扩散传递至不同尺度,提升小目标的检测精度;将原始YOLOv8的头部替换为融合了尺度、空间和任务3种注意力机制的动态检测头(DyHead),进一步提高模型检测性能。实验结果表明,YOLOv8-SFD准确率达到76.7%,召回率达到43.0%,平均精确率均值达到48.2%,分别比YOLOv8n提高了7.6%,2.0%和3.8%。YOLOv8-SFD有效提升了小目标和遮挡目标的检测精度,且检测速度达到119帧/秒,满足实时检测的需要。

关键词: 电力巡检, 特征融合, 目标检测, 特征聚焦扩散网络, YOLOv8

Abstract:

UAV images for powerline inspection usually have complex backgrounds, and often contain a lot of small targets, which may lead to a high rate of missed detections and false detections when processed by the general feature extraction networks for object detection. To address this, a feature focus & diffusion network (FFDN) was proposed for feature fusion, and an improved algorithm (YOLOv8-SFD) based on FFDN and YOLOv8 was designed for powerline component detection. Spatial-to-depth non-stride convolutions (SPDConv) were employed in the backbone network instead to preserve small-scale features and reduce feature loss caused by stride convolutions. The traditional feature pyramid network was replaced with the proposed FFDN. At the feature fusion stage, the feature focus modules in the FFDN were utilized to expand the receptive field and fuse multi-scale features, and the output feature maps by them were then diffused across different scales to enhance small target detection accuracy. Finally, the original YOLOv8 head was replaced with a dynamic detection head (DyHead) that integrates three attention mechanisms (scale, space, and task), to further enhance the performance. Experimental results demonstrated that YOLOv8-SFD achieved an accuracy rate of 76.7%, which was 7.6% higher than YOLOv8n; a recall rate of 43.0%, which was 2.0% higher; and a MAP of 48.2%, which was 3.8%. YOLOv8-SFD effectively enhanced the detection precision for small and obscured targets, and the detection speed reached 119 FPS, which satisfied real-time detection requirements.

Key words: powerline inspection, feature fusion, object detection, feature focus & diffusion network, YOLOv8

中图分类号: