欢迎访问《图学学报》 分享到:

图学学报 ›› 2023, Vol. 44 ›› Issue (5): 890-898.DOI: 10.11996/JG.j.2095-302X.2023050890

• 图像处理与计算机视觉 • 上一篇    下一篇

一种改进YOLOv8的密集行人检测算法

高昂1(), 梁兴柱1,2(), 夏晨星1, 张春炯3   

  1. 1.安徽理工大学计算机科学与工程学院,安徽 淮南 232001
    2.安徽理工大学环境友好材料与职业健康研究院(芜湖),安徽 芜湖 241003
    3.同济大学电子与信息工程学院,上海 201804
  • 收稿日期:2023-05-15 接受日期:2023-07-24 出版日期:2023-10-31 发布日期:2023-10-31
  • 通讯作者: 梁兴柱(1979-),男,副教授,硕士。主要研究方向为模式识别、计算机视觉等。E-mail:xzliang@aust.edu.cn
  • 作者简介:高昂(1999-),男,硕士研究生。主要研究方向为目标检测与图像处理。E-mail:2021201221@aust.edu.cn
  • 基金资助:
    国家自然科学基金项目(62102003);安徽理工大学环境友好材料与职业健康研究院研发专项(ALW2021YF04);芜湖市科技计划项目(2020yf48)

A dense pedestrian detection algorithm with improved YOLOv8

GAO Ang1(), LIANG Xing-zhu1,2(), XIA Chen-xing1, ZHANG Chun-jiong3   

  1. 1. School of Computer Science and Engineering, Anhui University of Science and Technology, Huainan Anhui 232001, China
    2. Institute of Environment-friendly Materials and Occupational Health, Anhui University of Science and Technology, Wuhu Anhui 241003, China
    3. College of Electronics and Information Engineering, Tongji University, Shanghai 201804, China
  • Received:2023-05-15 Accepted:2023-07-24 Online:2023-10-31 Published:2023-10-31
  • Contact: LIANG Xing-zhu (1979-), associate professor, master. His main research interests cover pattern recognition, computer vision, etc. E-mail:xzliang@aust.edu.cn
  • About author:GAO Ang (1999-), master student. His main research interests cover object detection and image processing. E-mail:2021201221@aust.edu.cn
  • Supported by:
    National Natural Science Foundation of China(62102003);Research Foundation of the Institute of Environment-Friendly Materials and Occupational Health (Wuhu), Anhui University of Science and Technology(ALW2021YF04);Science and Technology Research Project of Wuhu City(2020yf48)

摘要:

针对密集场景中小尺度的遮挡行人容易漏检的问题,提出一种改进YOLOv8检测算法。首先,针对小尺度行人特征提取问题,采用由可变形卷积改进的骨干网络增强网络对特征的提取能力,并设计遮挡感知注意力机制增强遮挡行人可见部分特征;其次,针对密集行人场景检测头定位不准的问题,设计动态解耦头增强对多尺度行人特征的关注,提高检测头的表达能力;最后,针对模型训练效率低的问题,训练时采用Wise-IoU与分布式聚焦损失结合的回归损失,提高模型的收敛能力。通过实验结果分析,改进YOLOv8算法在2个具有挑战性的密集行人数据集CrowdHuman和WiderPerson上性能优秀,AP50分别达到90.6%和92.3%,AP50:95分别达到62.5%和68.2%。相较原算法有了较大提升,且与其他先进行人检测模型进行比较时表现出了很强的竞争力。所提算法在密集行人检测任务中具有广泛的应用前景。

关键词: YOLOv8, 密集行人检测, 遮挡感知注意力, 可变形卷积, 动态解耦头

Abstract:

In response to the challenge of detecting small-scale, occluded pedestrians in dense scenes, where they are prone to being missed, we proposed an improved YOLOv8 detection algorithm. First, to address the issue of extracting features from small-scale pedestrians, a backbone network improved by deformable convolution was employed to enhance the feature extraction capability of the network, and an occlusion-aware attention mechanism was designed to enhance the visible part of the occluded pedestrian features. Second, to address imprecise localization of the detection head in dense pedestrian scenes, a dynamic decoupling head was designed to enhance attention to multi-scale pedestrian features, thereby improving the expression capability of the detection head. Finally, to address the problem of low model training efficiency, the regression loss that combined Wise-IoU with distributed focus loss was utilized for training, thereby enhancing the convergence ability of the model. Through the analysis of experimental results, the improved YOLOv8 algorithm demonstrated exceptional performance on two challenging and dense pedestrian datasets, namely CrowdHuman and WiderPerson, achieving an AP50 of 90.6% and 92.3% and an AP50:95 of 62.5% and 68.2%, respectively. In contrast to the original algorithm, the improvements were substantial, establishing robust competitiveness when compared with other state-of-the-art pedestrian detection models. The proposed algorithm exhibited a wide range of applications in dense pedestrian detection tasks.

Key words: YOLOv8, dense pedestrian detection, occlusion-aware attention, deformable convolution, dynamic decoupled head

中图分类号: