Welcome to Journal of Graphics share: 

Journal of Graphics ›› 2023, Vol. 44 ›› Issue (3): 456-464.DOI: 10.11996/JG.j.2095-302X.2023030456

Previous Articles     Next Articles

YOLO-RD-Apple orchard heterogenous image obscured fruit detection model

HAO Peng-fei(), LIU Li-qun(), GU Ren-yuan   

  1. College of Information Science and Technology, Gansu Agricultural University, Lanzhou Gansu 730070, China
  • Received:2022-10-25 Accepted:2023-01-15 Online:2023-06-30 Published:2023-06-30
  • Contact: LIU Li-qun (1982-), associate professor, master. Her main research interests cover intelligent computing and deep learning, etc. E-mail:llqhjy@126.com
  • About author:

    HAO Peng-fei (1998-), master student. His main research interests cover deep learning and image processing. E-mail:hola_Cc@163.com

  • Supported by:
    Gansu Provincial University Teacher Innovation Fund Project(2023A-051);Young Tutor Fund of Gansu Agricultural University(GAU-QDFC-2020-08);National Science Foundation of Gansu Province(20JR5RA032)

Abstract:

In order to address the challenge of robotic automated picking of highly occluded fruits in natural apple orchard environments, a YOLO-RD-Apple orchard heterogenous image occlusion fruit detection model based on dual inputs of RGB and Depth images was proposed. To reduce computational effort while ensuring the feature extraction capability, the lightweight MobileNetV2 and the lighter MobileNetV2-Lite, which was designed on the basis of MobileNetV2, were utilized as feature extractors for RGB and Depth images, respectively. Combining CSPNet with depth-separable convolution to accompany the SE attention module, the new SE-DWCSP3 module was proposed to improve the PANet structure and enhance the feature extraction capability of the network for stubby apple targets. Furthermore, the Soft NMS algorithm was introduced to replace the general NMS algorithm to address the false suppression phenomenon of the algorithm for dense targets and reduce the missed detection rate of obscured apples. The experimental results demonstrated the efficacy of this model on a natural obscured apple dataset, with an AP value of 93.1% on the test set, surpassing YOLOv4 by 1.4 percentage points, a 70% reduction in the number of parameters compared to YOLOv4, and a detection speed of 40.5 FPS on GPU (V100), which is 12.5% higher than that of YOLOv4. The proposed model exhibited improved detection accuracy and speed compared with YOLOv4, while simultaneously reducing the number of network parameters, making it more applicable to actual orchard apple picking scenarios.

Key words: target detection, heterogenous images, YOLOv4, apple picking, attention mechanism

CLC Number: