- College of Engineering, Nanjing Agricultural University, Nanjing, China
Monitoring and understanding pest population dynamics is essential to greenhouse management for effectively preventing infestations and crop diseases. Image-based pest recognition approaches demonstrate the potential for real-time pest monitoring. However, the pest detection models are challenged by the tiny pest scale and complex image background. Therefore, high-quality image datasets and reliable pest detection models are required. In this study, we developed a trapping system with yellow sticky paper and LED light for automatic pest image collection, and proposed an improved YOLOv5 model with copy-pasting data augmentation for pest recognition. We evaluated the system in cherry tomato and strawberry greenhouses during 40 days of continuous monitoring. Six diverse pests, including tobacco whiteflies, leaf miners, aphids, fruit flies, thrips, and houseflies, are observed in the experiment. The results indicated that the proposed improved YOLOv5 model obtained an average recognition accuracy of 96% and demonstrated superiority in identification of nearby pests over the original YOLOv5 model. Furthermore, the two greenhouses show different pest numbers and populations dynamics, where the number of pests in the cherry tomato greenhouse was approximately 1.7 times that in the strawberry greenhouse. The developed time-series pest-monitoring system could provide insights for pest control and further applied to other greenhouses.
1 Introduction
Monitoring pest population dynamics are essential to greenhouse management for effectively predicting the potential distribution of pests and preventing infestations and crop diseases (Deng et al., 2018; Capinha et al., 2021; Preti et al., 2021). Therefore, monitoring the time series of pest numbers in greenhouse is important to support human pest control decisions (Wen et al., n.d). However, automatic identification of pest populations in greenhouses is challenging due to the small size of pests. The current manual pest detection and counting approaches are labor intensive, time-consuming, and unreliable (Xie et al., 2018). Developing a fast and reliable automatic pest identification and counting approach may reduce the workload and improve timely pest control (Rustia et al., 2021b).
Image-based automatic pest monitoring systems can replace laborious manual identification and improve reliability, thus proving to be a powerful tool with real-time pest information to facilitate agricultural management (Sun et al., 2017; Rustia et al., 2020; Wang et al., 2020a). Deep learning has shown the potential of image processing with end-to-end feature extraction patterns for automatic pest identification and counting (Liu and Wang, 2020; Liu and Wang, 2021). To build an accurate deep learning model for automatic pest identification, high-quality and large-scale image datasets are indispensable (Li and Chao, 2021; Li et al., 2021b).
Current pest image datasets are mainly acquired from the internet, crop surface, pest-trapping containers, or sticky boards (Table 1). Pest datasets collected from the internet typically contain a large number of samples and various categories. A previous study reported an IP 102 dataset containing 75,000 images of 102 pest species collected from agricultural and insect-science websites; the dataset was evaluated using machine learning models including support vector machine, K Nearest Neighbors, and deep learning methods like Faster R-CNN, and FPN (Wu et al., 2019). Another website-based dataset for crop pest recognition comprises 46,567 images from 41 classes (Wang et al., 2022). However, online datasets have limited practical pest-monitoring applications because many pest categories are not representative of greenhouse crop production and show low-level harm to common crops (Wang et al., 2021).
Field investigations have mainly focused on pest images collected from crop fields with specific illumination conditions, crops, or pest types. A current study developed of an AgriPest dataset including 14 common pests of wheat, rice, corn, and rape (Wang et al., 2021). The filed collected datasets usually demonstrated different pest distribution density, light reflection variability, crop surface backgrounds, and pest types (Xie et al., 2018; Du et al., 2022). However, pest detection models demonstrate high precision in the present datasets, but are not replicable in a greenhouse setting due to large variations in pest type, environmental conditions, and unclear image collection owing to the pest movement. This limits the application of powerful deep learning technology for pest control in specific domains such as greenhouses.
Pest images collected with trapping devices are less affected by the environmental conditions and plant types. Trapping containers adopt pheromones, toxic gas, or a lamp to attract pests, and collect the image from its baseplate (Li et al., 2021b). Capturing pest images by light trapping have been widely used for automatic monitoring of pests (Wang et al., 2021). However, the trapping containers designed based on the pest phototaxis typically use a concentrated light source, thus pests are attracted near the light source with overlapping, which is not applicable to long-term monitoring. A vibration plate and a moving conveyor belt could be adopted to disperse the pests and avoid overlapping (Yao et al., 2020). The trapping containers usually used for detecting large size pests, but not effective for tiny pests of high-density in the greenhouse.
A Sticky board with color tropism can collect clear pest images by attracting flying pests, which are widely used as trapping devices (Ding and Taylor, 2016; Sütő, 2021). It reduces pest overlapping and limb mutilation, and could be applied to pests of different sizes. Current studies reported automatic methods for detecting insect pests using sticky paper traps and achieved high counting accuracies (Li et al., 2021a; Rustia et al., 2021a). The trapping board-based systems, however, have low trapping efficiency at night, and thus could not accurately monitoring pest number changes over 24 hours. In addition, changes in light during the day and night cause significant variations in the images in the trapping board, which reduce the accuracy and reliability of pest detection. Therefore, it is essential to develop a trapping system with a uniform light source and a high trapping efficiency.
Considering the diversity of pests, automatic pest-recognition models are required. Deep learning, with the advantage of an end-to-end learning strategy, is the current state-of-the-art object detection approach. Recently, deep-learning-based pest detection models are developed for multiclass pest detection and classification, and achieved high performances (Liu et al., 2019; Wang et al., 2020b). However, most models are primarily designed for large-scale pest detection in field crops, which may not be directly applicable to the detection of small pests in greenhouses due to the limited number of pixels in their images; therefore, it is necessary to develop small object detection models for greenhouse pests. The YOLOv5 model was developed for high inference and at three different scales, enabling it to effectively tackle the challenges associated with small-object detection in images. (Zhan et al., 2022; Zhao et al., 2022). The mosaic data augmentation adopted in the YOLOv5 model enriches the dataset by random scaling, cutting, and arranging the original input (Zhao et al., 2022). Further, it improves the accuracy of small target identification, especially when the number of large or medium targets is higher.
To address the challenge of automatic pest detection and counting in greenhouses, a comprehensive experiment on pest image recognition was conducted in two greenhouses, a cherry tomato and a strawberry greenhouse. The objectives of this study were as follows: 1) design an automatic pest identification system with an LED lamp and insect-trapping board to trap pests and automatically capture images with a uniform background to improve the insect trapping efficiency; 2) develop an improved YOLOv5 model with copy-pasting data augmentation to effectively identify small-sized pests and improve recognition accuracy of unbalanced distributed pest images; and 3) Monitoring the pest population dynamics in two greenhouses and compare their changing patterns.
2 Methods
2.1 Experimental design
We developed an automatic pest image collection system in this study. An improved YOLOv5 model was proposed for detecting small-sized pests, and its performance was compared with that of the original YOLOv5 model. Pest images were collected during the spring and summer for 43 days in two greenhouses (Figure 1). We examined two greenhouses in Nanjing, China. The first one is a “Hongxi” cherry tomato glass greenhouse situated in Xuanwu District, and the second one is a “Hongyan” strawberry multi-span greenhouse located in Pukou District. For the first three days, we placed 10 insect traps in the greenhouses and collected 10 images of each greenhouse to train a pest detection model. After the first three days, we used one insect trap to collect images and monitor the pest populations daily. Images were collected regularly at 16:00. Thus, we utilized 20 original images and applied data augmentation techniques to expand the training dataset to train the pest detection model and 80 images to monitor pest populations. Twenty training images were preprocessed using a data augmentation approach, which is discussed in Section 2. Data for the first dataset was collected from June 6, 2020 to July 18, 2020 from a greenhouse for cherry tomato plants from the seedling stage to the flowering and fruit setting stages. The main pests were tobacco whiteflies, leaf miners, aphids, and fruit flies. Data for the second dataset was collected from March 1, 2021 to April 12, 2021 from a greenhouse for strawberry plants during the flowering and fruit-setting stages.
Figure 1 Experimental design of pest image collection, pest detection model and pest population monitoring.
2.2 Data collection system
An automatic pest identification and monitoring system with an LED insect trapping lamp and yellow sticky paper are established to attract pests and collect images in greenhouses. The system consists of a pest trapping device, a power supply unit, an image collecting unit, a pest data processing unit, and a visual interface (Figure 2A).
Figure 2 Automatic pest identification and monitoring system with LED trap lamp, sticky paper and image acquisition system. (A) Pest monitoring system, (B) Pest trapping device, (C) Image collecting unit, (D) Visualization interface.
The pest-trapping device consisted of a stainless-steel plate, an aluminum profile frame, a yellow insect-trapping plate, and LED lamps (Figure 2B). The size of the yellow insect-trapping board was 20 × 25 cm, with a wavelength of 575 nm ± 10 nm. The wavelength of the LED insect-trapping lamp was 365 nm, and its voltage was 12 V. Two LED light strings were fixed on both sides of the frame, with one string consisting of 10 lamp beads. A storage battery with an output voltage of 12 V was used to provide 24-h power supply for the device. By adjusting the incident angle of the trapping device, the LED lamp could simultaneously trap the insects and provide light, making the trapping board images clear and conducive to the process. The pest-image collecting device consisted of an industrial camera and a cantilever bracket (Figure 2C). A Sony IMX226 camera with 3280*2464 resolution was installed in a cantilever and aimed at a piece of sticky paper at an appropriate distance to capture clear images. The visualization interface shows pest images and their location, types, and numbers (Figure 2D). The visualization interface presented dynamic changes in pest numbers within a week. Computation and visualization were conducted using Python on a Windows PC (Intel® Core™ i7-7500U) with a RAM of 8 Gb.
2.3 Benchmark YOLOv5 model
The benchmark model adopted in this study was YOLOv5 (Figure 3), which was released in 2020 for object detection. The YOLOv5 model consists of backbone, neck, and head modules that connect the procedure for predicting bounding boxes with class labels in an end-to-end differentiable network (Ultralytics, 2020). The YOLOv5 model has demonstrated excellent performance in small object detection in previous studies (Mathew and Mahesh, 2022; Zhan et al., 2022); thus, it was selected for identification of small-sized pests in this study.
2.3.1 Mosaic data augmentation
Mosaic data augmentation was adopted in the YOLOv5 model, which enriches the dataset by stitching together four images, thereby introducing novel object locations, partial occlusion, and variations in surrounding pixels for the model to learn from. The model could simultaneously process four images in the batch normalization layer, which decreased the GPU memory usage by using a relatively small mini-batch. The workflow of mosaic data augmentation was as follows: 1) random selection of four images from the original training dataset; 2) random rotation, scaling, flipping, and adjustment of the brightness and chromaticity of the four selected images; and 3) combining the images and box layout stitching into new images (Figure 4A) (Huang et al., 2022). The augmented images were enriched with various backgrounds, as shown in Figure 4B.
Figure 4 Mosaic data augmentation adopted in the YOLOv5 model. (A) Workflow of mosaic data augmentation, and (B) Pest image after mosaic data augmentation.
2.3.2 Backbone
The backbone module extracted features from the input image and transmitted them to the neck module (Figure 4). The backbone includes the focus, CBL, Cross Stage Partial (CSP), and Spatial Pyramid Pooling (SPP) modules. The focus structure transformed the original 608×608×1 images into 304×304×32 feature maps using slicing and convolution operations, which increased the computational complexity, but retained the original features. The CBL comprises one convolutional layer, one batch normalization layer, and one Leaky ReLU layer. CSP1 was used in the backbone network, and CSP2 was used in the neck network. The SPP network outputs a feature map of fixed size with multiscale feature fusion through the 1×1, 1×5, 9×9, and 13×13 max pooling.
2.3.3 Neck
The neck module generates a feature pyramid based on a path aggregation network (PANet) (Liu et al., 2018). This enables the model to detect pests of different sizes by shortening the information path between low-level spatial features and high-level semantic features through bottom-up path augmentation. Adaptive feature pooling directly propagates the effective information at each level to subsequent subnetworks (Liu et al., 2018).
2.3.4 Head module
The head module provides detection boxes, pest categories, coordinates of the detected pests, and confidence values. The loss function in the head module includes classification loss and bounding box regression loss. The YOLOv5 model uses the complete intersection over union (CIoU) loss (Zheng et al., 2020), which improves the regression accuracy and convergence speed by considering the distance between the detection frame and target box, overlapping area, aspect ratio, and other aspects.
2.4 Improved YOLOv5 model
The mosaic data augmentation approach shows limitations on the proposed datasets where small targets comprise more than 80% of total pests. Mosaic data augmentation improves the accuracy of small target identification only if there are more large and medium targets than small targets. For example, in Figure 5A, the size of the raw image was 3280 × 2464 pixels. The images were resized to 608 × 608 pixels before being fed into the YOLOv5 network. Thus, the output feature sizes were 19 × 19, 38 × 38, and 76 × 76 pixels. The largest feature map (76 × 76) corresponded to the smallest anchor box for small target detection. Its receptive field was 8 × 8 when back-propagated to an input image of 608 × 608 pixels. When the 8 × 8 receptive field corresponded to the raw image, it was approximately 43 × 32 (Figure 5A). Therefore, insects smaller than 43 × 32 pixels in the raw image were not recognized in the YOLOv5 model. Therefore, the mosaic data augmentation approach reduces the recognition precision for tiny pests. To improve the pest identification performance of the proposed pest images collected from greenhouses, an improved YOLOv5 model adopting a copy-pasting data augmentation approach to virtually increase the number of pests is developed.
2.4.1 Improved data augmentation approach
A copy-pasting augmentation approach was developed for data augmentation to improve the identification accuracy for small objects (Figure 5B). A raw image of 3280 × 2464 pixels was cropped by setting the overlapping area. The image was horizontally cropped into six pieces, each with 600 pixels, and the overlap length was set to 64 pixels. The image was also vertically cropped into 5 pieces, each piece of size 600 pixels, and the overlap length was set to 134 pixels. Therefore, the raw images were divided into thirty smaller images with 600 × 600 pixels. Setting the overlapping area could improve the detection accuracies for pests on the segmentation lines. Finally, to reduce the number of overlapping detection boxes, a non-maximum suppression (NMS) operation was performed on the entire image.
2.4.2 Improved head module
The YOLOv5 model generated candidate anchor boxes with various sizes and shapes, but these windows supposedly to contain one object, so it is necessary to filter out the ones. NMS is adopted to remove redundant boxes when their overlaps exceed a threshold. The intersection over union (IoU) loss is commonly used in NMS, but it demonstrates poor performances for the nonoverlapping boxes. The improved distance IoU (DIoU) considers the overlap area and distance between the central points of two bounding boxes when suppressing redundant boxes, making the model more robust to occlusion objects (Zheng et al., 2020). (Figure 6).
Figure 6 Intersection over union (IoU) and distance IoU losses and three application scenarios. (Blue and red colors represent the target box and predicted box, respectively.).
Where is the ground truth, and is the predicted box, The is the distance between the center points of two boxes, and represent the center points of the anchor frame and target frame respectively, the is the distance between the two center points, and c is the diagonal distance of the smallest rectangle covering two boxes.
In the improved YOLOv5 model, DIoU was deployed in the NMS and applied to the head module to remove redundant bounding boxes and improve the detection accuracy of occluded pests (Figure 6). The DIoU-NMS approach is more robust than the original NMS used in the YOLOv5 model for removing redundant boxes. Therefore, the DIoU-NMS method adopted in the improved YOLOv5 model improved the detection accuracy of occluded pests. The DIoU-NMS was defined as follows:
where S is the confidence level of category , is the threshold of the NMS, is the box with the highest confidence level, and IOU is the intersection ratio of the anchor and target frames (Zheng et al., 2020).
All training and processes were implemented on a Python library torch 1.9.0 framework in the Pycharm platform with Python 3.6. The computations were performed on a Windows workstation with an Nvidia GeForce 940MX graphics card (NVIDIA Corporation, Santa Clara, California, United States).
2.5 Performance metrics
The model performance is evaluated by precision, recall, F1-score, and accuracy, which are shown in Eqs. (5) to (8).
Where the (true positive) represents the number of correct positive predictions, (false positive) represents the number of incorrect positive predictions, and (false negative) represents the number of incorrect negative predictions.
The mean average precision (mAP) is adopted for evaluating the object detection performance. It is calculated by taking the average of the AP scores across all classes. The AP value is obtained by calculating the area under the precision-recall curve, which measures the trade-off between precision and recall at different confidence thresholds.
2.6 Pest dataset preparation
The datasets collected from the two greenhouses included six types of pests: tobacco whiteflies, leaf miners, aphids, fruit flies, thrips, and houseflies (Figure 7). Houseflies had the largest length of 5–8 mm while thrips were the smallest, with a length of 0.5–2 mm. Tobacco whiteflies were the highest in number but were difficult to detect by the human eye because of their white color. The lengths of the leaf miners, aphids, and fruit flies were 4–6 mm, 2.2 mm, and 1.5–4 mm, respectively. The number distribution of different types of pests are unbalanced, which brings challenge for the pest detection.
Figure 7 Images of pests collected in the (A) cherry tomato greenhouse and (B) strawberry greenhouse.
We collected 20 images with size of 3280 × 2464 pixels in the two greenhouses to train the YOLOv5 model. However, after data preprocessing with mosaic data augmentation, 450 images of size 600 × 600 pixels were acquired to train the insect pest detection model. Of the 450 training images, 200 were randomly selected for copy-pasting data augmentation. In total, 1,024 tobacco whiteflies, 857 thrips, 1,092 winged aphids, 941 leaf miners, 873 fruit flies, and 1,013 houseflies were included in the training set (Table 2).
The test set contained 328 images of 12,014 pests, including 5,832 tobacco whiteflies, 2,928 thrips, 1,001 winged aphids, 460 leaf miners, 890 fruit flies, and 267 houseflies (Table 2). The test set was used to monitor the pest population dynamics over a long period; therefore, the number of pests in the test set was larger than that in the training set.
3 Results and discussion
3.1 Pest detection results
The improved YOLOv5 model obtained higher accuracy than the original YOLOv5 model by adopting copy-pasting data augmentation (Table 3). The precision-recall graph, depicted in Figure 8, compares the performance of YOLOv5 and the improved YOLOv5 models. The overall pest-detection precision improved from 64% to 96% by using the improved YOLOv5 model. The improved YOLOv5 model obtained the highest detection precision of 99% on leaf miners and fruit flies, followed by aphids and houseflies, with a precision of 98%. Thrips are very difficult to identify because of their small size. The improved YOLOv5 model obtains a precision of 83% for thrips, which is still higher than that of the YOLOv5 model (80%). The improved model achieved an F1 score of 0.99 for detecting aphids, leaf miners, fruit flies, and houseflies, and a score of 0.98 for detecting whiteflies. However, it exhibited a lower F1 score of 0.91 for detecting thrips. The primary reason for the low detection accuracy of thrips may be the varying size of trips during different growth stage. Image resolution may be another factor that some pests are confused with dust. Furthermore, the improved YOLOv5 model required more time for image processing because it adopts a copy-pasting operation and feeds the input images into the network in different batches, the recognition speed of the improved YOLOv5 model was lower than that of the original YOLOv5 model.
A confusion matrix of the detection accuracies is shown in Figure 9. The improved YOLOv5 model demonstrated the best performance for leaf miners and fruit flies, with a 99% detection accuracy. 1% of leaf miners were incorrectly recognized as thrips and 1% of fruit flies were incorrectly detected as fruit flies and aphids. The model showed 98% accuracy for aphids and houseflies, while 2% of the aphids were incorrectly recognized as thrips, 1% of houseflies were incorrectly recognized as thrips, and 1% of houseflies were incorrectly identified as leaf miners. Among tobacco whiteflies, 97% were correctly detected, and 3% were missed. The model obtained the lowest detection accuracy for thrips at 83% accuracy, with 1% being incorrectly recognized as aphids. Missed-detected pests were recognized as background.
The recognition performance of the improved YOLOv5 model is illustrated in Figure 10. The improved YOLOv5 model demonstrated high classification accuracies for both cherry tomato and strawberry greenhouses. However, the model could not recognize tobacco whiteflies that were too light in color, and some tiny thrips were incorrectly recognized as small dust particles. Houseflies and leaf miners are similar in color and shape; therefore, they are sometimes misidentified. Precision, recall and mAP after training the improved YOLOv5 model are shown in Figure 11.
Figure 10 Results of pest detection for the two greenhouses using the improved YOLOv5 model. (A) Cherry tomato greenhouse, (B) Strawberry greenhouse.
3.2 Recognition performance on adjacent pests
The improved YOLOv5 model demonstrated better identification accuracy for nearby insects and could distinguish between two adjacent whitefly insects (Figure 12). For example, the original YOLOv5 model recognized two tobacco whiteflies as one pest in the black circle, whereas the improved YOLOv5 model could distinguish two adjacent whiteflies. DIoU-NMS computes the overlapping area of pests and the central point distance between two pest boxes when suppressing redundant boxes.
Figure 12 Pest recognition performance on two adjacent tobacco whiteflies using (A) YOLOv5 and (B) improved YOLOv5 models.
3.3 Real-time pest population dynamics in the cherry tomato and strawberry greenhouses
The improved YOLOv5 model obtained reliable prediction results; the predicted number of pests showed similar trends to the manual counting results. The cherry tomato greenhouse had a total of 7619 pests after 40 days of continuous monitoring as compared to 4395 pests in the strawberry greenhouse.
The dynamic trends in the two greenhouses demonstrated different patterns at different growth stages of cherry tomatoes and strawberries (Figure 13). The number of pests in the cherry tomato greenhouse increased sharply from 37 to 325 from June 9th to 10th. Thereafter, the number of total pests decreased to 92 on June 28th, and rose rapidly in late June and early July. The number of pests was at its highest from 214 to 419, from July 1st to July 18th, although there were fluctuations. The strawberry greenhouse had a low density of pests from March 4th to March 19th, and showed two peaks on March 23rd and April 3rd. Changes in pest numbers may be related to environmental temperature, humidity, and the growth stages of greenhouse crops (Aiello et al., 2018). Understanding pest outbreaks may help identify periods of risk in greenhouses and provide decision-making support for managers.
Figure 13 Pest population dynamics in the (A) cherry tomato greenhouse and (B) strawberry greenhouse during the 40 days of monitoring.
Pest diversity and frequency varied between the two greenhouses (Figure 14). The main pest observed in the cherry tomato greenhouse was the tobacco whitefly, whereas thrips were the most prevalent pest in the strawberry greenhouse. The number of tobacco whiteflies in the cherry tomato greenhouse was approximately 2.5 times the total number of leaf miners, fruit flies, and aphids. In the strawberry greenhouse, 87% of the pests comprised thrips, fruit flies, tobacco whiteflies, and houseflies. In the cherry tomato greenhouse, the dynamic changes in tobacco whitefly, fruit fly, and aphid showed patterns similar to the overall trend of pest numbers, which decreased in June and increased in July. The population trend of leaf miners showed the opposite pattern, increasing in June and decreasing in July. In the strawberry greenhouse, the numbers of thrips and fruit flies showed the same trend of increasing in late March and early April. The numbers of tobacco whiteflies and houseflies fluctuated and showed a small increase at the same time. The dynamic trend in pest numbers provides insights into pest control during certain periods throughout the year, which is of vital importance for greenhouse management.
Figure 14 Changes in individual pest numbers in the (A) cherry tomato greenhouse and (B) strawberry greenhouse during the 40 days of monitoring.
Obtaining the dynamic changes in pest occurrence will help us to further build pest prediction models based on time series and understand the patterns of greenhouse pest occurrence, thus assisting agricultural pest control decisions (Chen et al., 2022). The present study monitored pest populations in two greenhouses for 40 days, which provided the change in number of pests at different times. This study provides insights for pest management at different plant growth stages.
4 Conclusions
This study proposed a pest detection and long-term pest number monitoring system for cherry tomato and strawberry greenhouses. To obtain high-quality images, we designed a sticky board trap that combines yellow sticky paper and LED pest-trap lamps to achieve all-weather pest-trapping effects. The LED lamp can also be used as a light supplement to increase the brightness of the trap board image and alleviate the problem of uneven illumination, thereby improving image quality. The pest image capturing system with LED traps provided clearer images compared to that adopted sticky paper only. The system was applied to cherry tomato and strawberry greenhouses, and the improved YOLOv5 model obtained an overall pest detection precision of 96% during the 40 days of monitoring. The model achieved the highest F1 score of 0.99 for the detection of four types of pests, while the lowest F1 score of 0.91 was obtained for thrips. This system provides important decision-support information for the management of pests and diseases in greenhouses. The pest-monitoring system developed in this study can be applied to other types of greenhouses for pest image collection and for building pest-detection models for a large number of common pests. The system can also be used for pest population dynamics and status prediction, considering future changes in climate and weather conditions.
Data availability statement
The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.
Author contributions
XLZ: Conceptualization, Writing – review & editing, Methodology, Writing – original draft. JB: Conceptualization, Data curation, Methodology, Software, Writing – original draft, Writing – review & editing. XXZ: Writing – original draft, Writing – review & editing. XW: Funding acquisition, Investigation, Supervision, Conceptualization, Writing – review & editing, Data curation.
Funding
This work was supported by the China National Key Research and Development Plan (grant numbers 2019YFE0125201 and 2019YFD1001902-11).
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
Aiello, G., Giovino, I., Vallone, M., Catania, P., Argento, A. (2018). A decision support system based on multisensor data fusion for sustainable greenhouse management. J. Cleaner Production 172, 4057–4065. doi: 10.1016/j.jclepro.2017.02.197
Alfarisy, A. A., Chen, Q., Guo, M. (2018). “Deep learning based classification for paddy pests & diseases recognition,” in Proceedings of 2018 International Conference on Mathematics and Artificial Intelligence, Chengdu China. 21–25 (ACM). doi: 10.1145/3208788.3208795
Capinha, C., Ceia-Hasse, A., Kramer, A. M., Meijer, C. (2021). Deep learning for supervised classification of temporal data in ecology. Ecol. Inf. 61, 101252. doi: 10.1016/j.ecoinf.2021.101252
Chen, C.-J., Li, Y.-S., Tai, C.-Y., Chen, Y.-C., Huang, Y.-M. (2022). Pest incidence forecasting based on Internet of Things and Long Short-Term Memory Network. Appl. Soft Computing 124, 108895. doi: 10.1016/j.asoc.2022.108895
Deng, L., Wang, Y., Han, Z., Yu, R. (2018). Research on insect pest image detection and recognition based on bio-inspired methods. Biosyst. Eng. 169, 139–148. doi: 10.1016/j.biosystemseng.2018.02.008
Ding, W., Taylor, G. (2016). Automatic moth detection from trap images for pest management. Comput. Electron. Agric. 123, 17–28. doi: 10.1016/j.compag.2016.02.003
Du, J., Liu, L., Li, R., Jiao, L., Xie, C., Wang, R. (2022). Towards densely clustered tiny pest detection in the wild environment. Neurocomputing 490, 400–412. doi: 10.1016/j.neucom.2021.12.012
Huang, L., Qiu, M., Xu, A., Sun, Y., Zhu, J. (2022). UAV imagery for automatic multi-element recognition and detection of road traffic elements. Aerospace 9, 198. doi: 10.3390/aerospace9040198
Kalfas, I., De Ketelaere, B., Bunkens, K., Saeys, W. (2023). Towards automatic insect monitoring on witloof chicory fields using sticky plate image analysis. Ecol. Inf. 75, 102037. doi: 10.1016/j.ecoinf.2023.102037
Li, Y., Chao, X. (2021). Toward sustainability: trade-off between data quality and quantity in crop pest recognition. Front. Plant Sci. 12. doi: 10.3389/fpls.2021.811241
Li, W., Wang, D., Li, M., Gao, Y., Wu, J., Yang, X. (2021a). Field detection of tiny pests from sticky trap images using deep learning in agricultural greenhouse. Comput. Electron. Agric. 183, 106048. doi: 10.1016/j.compag.2021.106048
Li, W., Zheng, T., Yang, Z., Li, M., Sun, C., Yang, X. (2021b). Classification and detection of insects from field images using deep learning for smart pest management: A systematic review. Ecol. Inf. 66, 101460. doi: 10.1016/j.ecoinf.2021.101460
Liu, S., Qi, L., Qin, H., Shi, J., Jia, J. (2018). “Path aggregation network for instance segmentation,” in 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT. 8759–8768 (IEEE). doi: 10.1109/CVPR.2018.00913
Liu, J., Wang, X. (2020). Tomato diseases and pests detection based on improved yolo V3 convolutional neural network. Front. Plant Sci. 11. doi: 10.3389/fpls.2020.00898
Liu, J., Wang, X. (2021). Plant diseases and pests detection based on deep learning: a review. Plant Methods 17, 22. doi: 10.1186/s13007-021-00722-9
Liu, L., Wang, R., Xie, C., Yang, P., Wang, F., Sudirman, S., et al. (2019). PestNet: an end-to-end deep learning approach for large-scale multi-class pest detection and classification. IEEE Access 7, 45301–45312. doi: 10.1109/ACCESS.2019.2909522
Mathew, M. P., Mahesh, T. Y. (2022). Leaf-based disease detection in bell pepper plant using YOLO v5. SIViP 16, 841–847. doi: 10.1007/s11760-021-02024-y
Nanni, L., Manfè, A., Maguolo, G., Lumini, A., Brahnam, S. (2022). High performing ensemble of convolutional neural networks for insect pest image detection. Ecol. Inf. 67, 101515. doi: 10.1016/j.ecoinf.2021.101515
Preti, M., Verheggen, F., Angeli, S. (2021). Insect pest monitoring with camera-equipped traps: strengths and limitations. J. Pest Sci. 94, 203–217. doi: 10.1007/s10340-020-01309-4
Rustia, D. J. A., Chao, J., Chiu, L., Wu, Y., Chung, J., Hsu, J., et al. (2021a). Automatic greenhouse insect pest detection and recognition based on a cascaded deep learning classification method. J. Appl. Entomol 145, 206–222. doi: 10.1111/jen.12834
Rustia, D. J. A., Lin, C. E., Chung, J.-Y., Zhuang, Y.-J., Hsu, J.-C., Lin, T.-T. (2020). Application of an image and environmental sensor network for automated greenhouse insect pest monitoring. J. Asia-Pacific Entomology 23, 17–28. doi: 10.1016/j.aspen.2019.11.006
Rustia, D. J. A., Lu, C.-Y., Chao, J.-J., Wu, Y.-F., Chung, J.-Y., Hsu, J.-C., et al. (2021b). Online semi-supervised learning applied to an automated insect pest monitoring system. Biosyst. Eng. 208, 28–44. doi: 10.1016/j.biosystemseng.2021.05.006
Sun, Y., Cheng, H., Cheng, Q., Zhou, H., Li, M., Fan, Y., et al. (2017). A smart-vision algorithm for counting whiteflies and thrips on sticky traps using two-dimensional Fourier transform spectrum. Biosyst. Eng. 153, 82–88. doi: 10.1016/j.biosystemseng.2016.11.001
Sun, Y., Liu, X., Yuan, M., Ren, L., Wang, J., Chen, Z. (2018). Automatic in-trap pest detection using deep learning for pheromone-based Dendroctonus valens monitoring. Biosyst. Eng. 176, 140–150. doi: 10.1016/j.biosystemseng.2018.10.012
Sütő, J. (2021). Embedded system-based sticky paper trap with deep learning-based insect-counting algorithm. Electronics 10, 1754. doi: 10.3390/electronics10151754
Teixeira, A. C., Morais, R., Sousa, J. J., Peres, E., Cunha, A. (2023). A deep learning approach for automatic counting of bedbugs and grape moth. Proc. Comput. Sci. 219, 145–152. doi: 10.1016/j.procs.2023.01.275
Ultralytics (2020) YOLOv5. Available at: https://github.com/ultralytics/yolov5.
Wang, K., Chen, K., Du, H., Liu, S., Xu, J., Zhao, J., et al. (2022). New image dataset and new negative sample judgment method for crop pest recognition based on deep learning models. Ecol. Inf. 69, 101620. doi: 10.1016/j.ecoinf.2022.101620
Wang, R., Liu, L., Xie, C., Yang, P., Li, R., Zhou, M. (2021). AgriPest: A large-scale domain-specific benchmark dataset for practical agricultural pest detection in the wild. Sensors 21, 1601. doi: 10.3390/s21051601
Wang, F., Wang, R., Xie, C., Yang, P., Liu, L. (2020a). Fusing multi-scale context-aware information representation for automatic in-field pest detection and recognition. Comput. Electron. Agric. 169, 105222. doi: 10.1016/j.compag.2020.105222
Wang, Q.-J., Zhang, S.-Y., Dong, S.-F., Zhang, G.-C., Yang, J., Li, R., et al. (2020b). Pest24: A large-scale very small object data set of agricultural pests for multi-target detection. Comput. Electron. Agric. 175, 105585. doi: 10.1016/j.compag.2020.105585
Wen, C., Chen, H., Ma, Z., Zhang, T., Yang, C., Su, H., et al. (n.d). Pest-YOLO: A model for large-scale multi-class dense and tiny pest detection and counting. Front. Plant Science. 13, 973985. doi: 10.3389/fpls.2022.973985
Wu, X., Zhan, C., Lai, Y.-K., Cheng, M.-M., Yang, J. (2019). “IP102: A large-scale benchmark dataset for insect pest recognition,” in 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA. 8779–8788 (IEEE). doi: 10.1109/CVPR.2019.00899
Xie, C., Wang, R., Zhang, J., Chen, P., Dong, W., Li, R., et al. (2018). Multi-level learning features for automatic classification of field crop pests. Comput. Electron. Agric. 152, 233–241. doi: 10.1016/j.compag.2018.07.014
Yao, Q., Feng, J., Tang, J., Xu, W., Zhu, X., Yang, B., et al. (2020). Development of an automatic monitoring system for rice light-trap pests based on machine vision. J. Integr. Agric. 19, 2500–2513. doi: 10.1016/S2095-3119(20)63168-9
Zhan, W., Sun, C., Wang, M., She, J., Zhang, Y., Zhang, Z., et al. (2022). An improved Yolov5 real-time detection method for small objects captured by UAV. Soft Comput. 26, 361–373. doi: 10.1007/s00500-021-06407-8
Zhao, Y., Shi, Y., Wang, Z. (2022). “The improved YOLOV5 algorithm and its application in small target detection,” in Intelligent Robotics and Applications Lecture Notes in Computer Science. Eds. Liu, H., Yin, Z., Liu, L., Jiang, L., Gu, G., Wu, X., et al (Cham: Springer International Publishing), 679–688. doi: 10.1007/978-3-031-13841-6_61
Keywords: tiny pest detection, improved YOLOv5, pest population dynamics, pest trapping system, greenhouse
Citation: Zhang X, Bu J, Zhou X and Wang X (2023) Automatic pest identification system in the greenhouse based on deep learning and machine vision. Front. Plant Sci. 14:1255719. doi: 10.3389/fpls.2023.1255719
Received: 09 July 2023; Accepted: 11 September 2023;
Published: 28 September 2023.
Edited by:
Jingcheng Zhang, Hangzhou Dianzi University, ChinaReviewed by:
Dongmei Chen, Hangzhou Dianzi University, ChinaPappu Kumar Yadav, University of Florida, United States
Copyright © 2023 Zhang, Bu, Zhou and Wang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Xiaochan Wang, d2FuZ3hpYW9jaGFuQG5qYXUuZWR1LmNu