Skip to main content

ORIGINAL RESEARCH article

Front. Plant Sci.
Sec. Technical Advances in Plant Science
Volume 15 - 2024 | doi: 10.3389/fpls.2024.1447263
This article is part of the Research Topic Trends and Challenges in Plant Biomonitoring, Bioremediation and Biomining View all 6 articles

Multi-stage Tomato Fruit Recognition Method Based on Improved YOLOv8

Provisionally accepted
Yuliang Fu Yuliang Fu 1Weiheng Li Weiheng Li 1*Gang Li Gang Li 1*Yuanzhi Dong Yuanzhi Dong 1*Songlin Wang Songlin Wang 1*Zhiguang Dai Zhiguang Dai 2
  • 1 School of Water Conservancy, North China University of Water Resources and Electric Power, Zhengzhou, China
  • 2 College of Agricultural Equipment Engineering, Henan University of Science and Technology, Luoyang, China

The final, formatted version of the article will be published soon.

    To address the challenges of low recognition and localization efficiency and poor accuracy in multi-stage tomato recognition within complex environments, this study proposed a method based on an improved YOLOv8 model, dubbed YOLOv8-EA. Initially, the EfficientViT network was utilized to replace the backbone of the original YOLOv8 (You Only Look Once version 8) model, reducing the model's parameter count and enhancing its feature extraction capabilities. Subsequently, partial convolution was integrated into the C2f module, forming the C2f-Faster module, which further accelerated the model’s inference speed. The boundary box loss function was modified to SIoU, facilitating faster model convergence and improving the accuracy and precision of detections. Finally, an auxiliary detection head (Aux-Head) module was incorporated to bolster the network's learning potential. On the self-built dataset, the results show that: the accuracy, recall and average precision of the YOLOv8-EA model are 91.4%, 88.7% and 93.9% respectively, and the detection speed is 163.33 frames/s. Compared to the baseline YOLOv8n network, the model weight is increased by 2.07 MB, the accuracy, recall and average precision are improved by 10.9, 11.7 and 7.2 percentage points, the detection speed is improved by 42.1%, and the detection precision is 97.1%, 91% and 93. On the publicly available dataset, the accuracy, recall and average precision of YOLOv8-EA are 91%, 89.2% and 95.1% respectively, and the detection speed is 1.8 ms: YOLOv8-EA model accuracy, recall, and average precision are 91.4%, 88.7%, and 93.9%, respectively, and the detection speed is 163.33 frames/s. Compared to the baseline YOLOv8n network, the model weight is increased by 2.07 MB, the accuracy, recall and average precision are improved by 10.9, 11.7 and 7.2 percentage points, respectively, and the detection speed is improved by 42. 1%, and the detection accuracies for unripe, semi-ripe and ripe tomatoes are 97.1%, 91% and 93.7%, respectively; on the public dataset, the accuracy, recall and average precision of YOLOv8-EA are 91%, 89.2% and 95.1%, respectively, and the detection speed is 1.8ms.The enhanced model is capable of more efficient and accurate recognition of tomatoes at various stages, providing a technical reference for intelligent tomato harvesting.

    Keywords: Image Recognition, object detection, YOLOv8, EfficientViT, Auxiliary detection head, Tomato

    Received: 11 Jun 2024; Accepted: 01 Aug 2024.

    Copyright: © 2024 Fu, Li, Li, Dong, Wang and Dai. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

    * Correspondence:
    Weiheng Li, School of Water Conservancy, North China University of Water Resources and Electric Power, Zhengzhou, China
    Gang Li, School of Water Conservancy, North China University of Water Resources and Electric Power, Zhengzhou, China
    Yuanzhi Dong, School of Water Conservancy, North China University of Water Resources and Electric Power, Zhengzhou, China
    Songlin Wang, School of Water Conservancy, North China University of Water Resources and Electric Power, Zhengzhou, China

    Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.