- 1Key Laboratory of Tea Quality and Safety Control, Ministry of Agriculture and Rural Affairs, Tea Research Institute, Chinese Academy of Agricultural Sciences, Hangzhou, China
- 2Hangzhou Ruikun Technology Co., Ltd., Hangzhou, China
- 3Tea Station of Xinchang County, Shaoxing, China
Accurate detection of tea leaf diseases and insects is crucial for their scientific and effective prevention and control, essential for ensuring the quality and yield of tea. Traditional methods for identifying tea leaf diseases and insects primarily rely on professional technicians, which are difficult to apply in various scenarios. This study proposes a recognition method for tea leaf diseases and insects based on improved MobileNetV3. Initially, a dataset containing images of 17 different types of tea leaf diseases and insects was curated, with data augmentation techniques utilized to broaden recognition scenarios. Subsequently, the network structure of MobileNetV3 was enhanced by integrating the CA (coordinate attention) module to improve the perception of location information. Moreover, a fine-tuning transfer learning strategy was employed to optimize model training and accelerate convergence. Experimental results on the constructed dataset reveal that the initial recognition accuracy of MobileNetV3 is 94.45%, with an F1-score of 94.12%. Without transfer learning, the recognition accuracy of MobileNetV3-CA reaches 94.58%, while with transfer learning, it reaches 95.88%. Through comparative experiments, this study compares the improved algorithm with the original MobileNetV3 model and other classical image classification models (ResNet18, AlexNet, VGG16, SqueezeNet, and ShuffleNetV2). The findings show that MobileNetV3-CA based on transfer learning achieves higher accuracy in identifying tea leaf diseases and insects. Finally, a tea diseases and insects identification application was developed based on this model. The model showed strong robustness and could provide a reliable reference for intelligent diagnosis of tea diseases and insects.
1 Introduction
Tea is a significant economic crop in China, characterized by extensive cultivation and a wide variety of cultivars (Li et al., 2022). However, during its growth and cultivation, tea is susceptible to pests and diseases, which directly impact its quality and quantity (Liu and Wang, 2021). Tea production is reduced by about 20% each year due to tea leaf diseases and insects (Xue et al., 2023).The traditional approach to identifying pests and diseases in tea plants has relied heavily on the experience and visual inspection of technicians. However, this methodology is often constrained by the lack of specialized personnel and insufficient timeliness in the identification process (Huang et al., 2019). Consequently, the real-time and efficient monitoring of pest and disease conditions in tea plantations is crucial for precise pest management and the assurance of tea quality and safety.
With the advancement of machine vision technology, image processing and machine learning methods have been widely utilized in the identification of crop pests and diseases (Pal and Kumar, 2023; Pan et al., 2022). Sun et al. (2019) proposed an algorithm that combines simple linear iterative clustering with Support Vector Machine (SVM). This method effectively extracts important tea disease patterns from complex backgrounds, laying a solid foundation for further research on tea disease identification. Yousef et al. (2022) developed an apple disease recognition model using digital image processing and sparse coding, achieving an average accuracy rate of 85%. However, most of these studies are based on the identification of insects and diseases using characteristics such as color, texture, and shape, which often rely on manual selection and design. This dependency limits the adaptability of models to the environment, consequently resulting in weaker accuracy and universality in pest and disease classification.
In response to the problems of machine learning methods, more and more scholars are using deep learning models to identify tea leaf disease and insect (Chen et al., 2020; Qi et al., 2022; Liu and Zhang, 2023; Zhang and Zhang, 2023). Various models, including AlexNet, GoogLeNet, VGG, and ResNet (Krizhevsky et al., 2012; Szegedy et al., 2015; Simonyan and Zisserman, 2015), have demonstrated outstanding performance in crop disease identification. For example, an optimized Dense Convolutional Neural Network structure, presented by Waheed et al. (2020), achieved 98.06% accuracy in classifying corn leaf diseases. Li et al. (2022) integrated the SENet module into the DenseNet framework for tea disease identification using transfer learning. Sun et al. (2023) introduced TeaDiseaseNet based on YOLOv5 for detecting six tea leaf diseases, despite lower recognition rates and high-resolution image challenges. The above recognition models are improved by using a large-scale CNN, which is more computationally complex, has a slightly larger number of parameters, and imposes high requirements for deployment and application. Therefore, a lightweight recognition model with high accuracy is more valuable in actual production.
In the application of lightweight neural networks, MobileNet is often used as a basic model (Ullah et al., 2023; Peng and Li, 2023). To reduce the number of model parameters and the amount of computation, making the model more lightweight while ensuring good recognition results, researchers have improved the model structure and embedded an attention mechanism (Chen et al., 2022; Bi et al., 2022). However, in the identification of tea diseases, due to the small size of the dataset and the relatively sparse distribution of disease spots in some images, existing lightweight models struggle to achieve high-precision classification results for this problem.
Therefore, this study addresses the aforementioned issues by proposing a network model named MobileNetV3-CA, which integrates MobileNetV3 (Howard et al., 2019) with a Coordinate Attention (CA) module (Hou et al., 2021). The CA module enhances the model’s discriminative ability by expanding the local receptive field through the incorporation of attention mechanisms. Given the limited research on tea leaf diseases compared to fruit and cereal crops (Mu et al., 2023), images of tea leaf diseases were collected and augmented to construct a dataset containing 17 common types of tea leaf diseases and insects. Utilizing transfer learning, the model was pre-trained on a large-scale public dataset and then fine-tuned on a dataset of tea leaf diseases and insects to accelerate convergence and improve accuracy and robustness with limited samples. Finally, the effectiveness of the MobileNetV3-CA network model was validated through the recognition of common tea leaf diseases and insects, as well as testing within application programs.
2 Materials and methods
2.1 Construction of image data set of tea leaf diseases and insects
The images of tea leaf diseases and insects used in this study were sourced from tea-producing regions in Xinchang County, Zhejiang Province. The primary tea plant varieties involved are Longjing 43, Wuniuzao, Jiukeng, and Zhongcha 108. Data collection occurred from March to April each year, from 2021 to 2023, during the high incidence period of tea diseases and insects, facilitating the comprehensive collection of disease and insect data. This study employed various smartphones, including models from brands such as Huawei, Xiaomi, and Apple, for image capture. The majority of image data was acquired through the team-developed backend of the “Xinchang Tea Guardian” WeChat mini-program. In total, 22,380 images were collected, encompassing 17 types of tea leaf diseases and insects, such as anthracnose, tea blister blight, ectropis obliqua hypulina. Some image samples of tea leaf diseases and insects in the data set are shown in Figure 1. Considering the significant morphological differences among certain insects in their larval, pupal, and adult stages, this study categorizes them into distinct groups. For instance, ladybugs and their larvae, as well as corn earworm larvae and adults, are included in separate categories. Furthermore, due to the uncertainty in collecting crop disease images, the distribution of tea leaf disease and insect images obtained is highly uneven. The number and proportion of various tea diseases and insects are shown in Figure 2. For instance, tea blister blight accounts for approximately 3% of the total, while ladybird pupa images constitute around 11%.
Figure 2. Proportion and quantity of all kinds of diseases and insect pests in the original data set.
2.2 Data preprocessing
Based on the distribution of the statistical samples shown in Figure 2, the number of images for different categories of diseases and insects ranges from 500 to 2000. To prevent the model from overfitting due to insufficient training samples, data augmentation techniques were employed. These operations include flipping, adding Gaussian noise, adjusting contrast, rotation, shear, and adding histogram equalization, all of which were applied in random order. Flipping refers to randomly flipping pictures up and down or left and right with a probability of 0.5. Adding Gaussian noise refers to applying Gaussian noise to images with a probability of 0.5, using a Gaussian kernel with a random standard deviation sampled uniformly from the interval [0.0, 0.6]. Adjusting contrast refers to modifying the contrast of images according to 127 + alpha*(v-127), where v is a pixel value and alpha is sampled uniformly from the interval [0.75, 1.5] (once per image). Rotation refers to rotating images by -20 to 20 degrees with a probability of 0.5. Shear refers to shearing images by -20 to 20 degrees with a probability of 0.5. Adding histogram equalization refers to applying histogram equalization to input images with a probability of 0.5. Data augmentation could enhance sample diversity and simulate natural conditions for the identification of diseases and insects. After data augmentation, the number of images for each category was increased to 2000, totaling 34,000 images. For example, the data augmentation process for tea blister blight is illustrated in Figure 3. The augmentation adjusted the original images’ rotation angle, brightness, and blurriness, highlighting the local details of diseases and insects. Additionally, considering the potential differences in image format and size due to varying sample sources, all images were standardized to a uniform size of 224 pixels × 224 pixels.
2.3 Construction and improvement of disease and insect recognition model
2.3.1 An overall introduction to the improved model
Common lightweight neural network models include SqueezeNet (Iandola et al., 2016), MobileNet, EfficientNet (Tan and Le, 2019), among others. In this study, MobileNetV3 from the MobileNet series was selected. The MobileNetV3 model retains its lightweight characteristics while continuing to utilize the depthwise separable convolutions and inverted residual modules from the MobileNetV2 model (Sandler et al., 2018). It enhances the bottleneck structure by integrating the SE (Squeeze-and-Excitation) module (Hu et al., 2018), which strengthens the emphasis on significant features and suppresses less important ones. Additionally, the new hard-swish activation function is adopted to further optimize the network structure. These improvements allow the MobileNetV3 model to maintain its lightweight nature while enhancing accuracy in tasks such as image classification. The MobileNetV3 model is available in large and small versions based on resource availability, and this study employs the MobileNetV3-large model as the baseline. Its structural parameters are shown in Table 1.
2.3.2 Improvement of attention mechanisms
While incorporating the SE module into the bottleneck structure of MobileNetV3-Large has indeed improved model performance, the SE module only considers information between channels to determine the importance of each channel. However, it overlooks the crucial positional information in the visual space. As a result, the model can only capture local feature information, leading to issues such as scattered regions of interest and limited performance. To address these limitations, the ECA module improves on the SE module by avoiding dimensionality reduction and capturing cross-channel interaction information more efficiently (Wang et al., 2020). Even though the ECA module is an improvement over the SE module, it still only considers the information between channels in essence (Jia et al., 2022). Therefore, in order to improve the recognition rate of the model and enhance its ability to capture the location information of tea leaf diseases and insects, the coordinate information must be considered. In this study, we replace the SE module in the MobileNetV3 structure with the CA module to improve MobileNetV3. The overall structure of the improved MobileNetV3-CA model is shown in Figure 4. To accurately obtain the relative position information in the image of diseases and insect pests of tea leaves, the CA module was introduced into the attention module of the bottleneck structure of layers 4-6 and layers 12-16.
Figure 4. The structure of MobileNetV3-CA model. Conv stands for convolution layer; Bneck and CA-Bneck stand for a bottleneck structure and a bottleneck structure after introducing coordinate attention module, respectively; Pool represents pooling layer; BatchNorm stands for batch normalization; ReLU6/hard-swish stand for activation function.
The CA module can focus the model’s attention on the region of interest through effective positioning in the pixel coordinate system, thereby obtaining information that considers both channel and position in the tea leaf image, reducing the attention to interference information, and thus improving the feature expression ability of the model. The basic structure of the CA module is shown in Figure 5.
Figure 5. The structure of CA modules. “X/Y Avg Pool” is the average pool in X/Y direction; “Concat” stands for concatenate; “BatchNorm” stands for batch normalization; “Swish” and “Sigmoid” represent nonlinear activation functions; C is the number of channels; H is the height of the feature map; W is the width of the feature map; R is the reduction factor.
For a given characteristic graph X, the number of channels is C, the height is H, and the width is W. The CA module first pools the input X in two spatial directions, namely, height and width, to obtain feature maps in both directions. Next, it concatenates the feature maps from these two directions in spatial dimensions, and then changes the dimensions to the original C/r using a 1×1 convolution transformation. Subsequently, it applies batch normalization and Swish activation operations to obtain an intermediate feature map containing information from both directions, as shown in the formula below.
In the formula, f is the intermediate feature map obtained by encoding spatial information in two directions, δ is the activation function Swish, and F1 is the convolution transformation function of 1×1. Here, xc is the feature information of the specific position of the feature graph in channel c, h is the specific height of the feature map, and j is the width of the feature map, with the value range of [0, W]. Similarly, w is the specific width of the feature map, and i is the height of the feature map, with the value range of [0, H]. F is decomposed into two separate tensors hf and wf along the spatial dimension in two directions. Through two 1×1 convolution transformation functions, hf and wf are converted into tensors with the same number of channels as the input X. Next, the attention weights in height and width are obtained by activating the function σ. Finally, we multiply the extended attention weight with X to get the output of the CA module, as shown in the equation below.
Where yc is the output of the c-th channel, σ is the activation function Sigmoid, and Fh and Fw are convolution transformation functions in height and width.
2.4 Transfer learning
In deep learning models, a large number of parameters are typically required for training, often necessitating extensive support from large-scale datasets. However, not all tasks have access to sufficiently large datasets for training. Transfer learning offers a solution to this issue. Transfer learning facilitates the application of the same model to another research context. Given that there are often commonalities between different samples, sharing similar characteristics, leveraging models pre-trained on large datasets to retrain for new tasks can achieve effective training outcomes and rapid convergence speeds. Therefore, in order to make full use of the existing labeled data and ensure the recognition accuracy of the model on new tasks, this study adopts transfer learning to optimize the model. Specifically, the fine-tuning method involves freezing part of the convolution layers as the optimization strategy for transfer learning. First of all, the large dataset ImageNet (Szegedy et al., 2015) serves as the source domain for network pre-training. The learned model weights from this pre-training phase are then transferred to identify diseases and insects in tea leaves. Drawing on existing prior knowledge allows for efficient handling of similar recognition tasks. Subsequently, the model parameters are fine-tuned during the training process on images of tea leaf diseases and insects, ultimately producing the final tea leaf disease and insect recognition model.
3 Results and discussion
3.1 Test environment and parameter setting
To evaluate the performance of the tea leaf disease and insect recognition model MobileNetV3-CA, acquired images of tea leaf diseases and insects were used for both training and testing. The dataset was divided into training, validation, and test sets in a ratio of 7:2:1. The experiment employed the PyTorch 1.10.0 deep learning framework, programmed in Python 3.8. The development environment was set up using VSCode. The computer used for running the programs is equipped with an Intel® Core i5-1135G7 CPU @ 2.40 GHz, 32 GB of RAM, and operates on a 64-bit Windows 10 system.
The Batch Size of the experiment was set to 16. In order to improve the convergence of the model, the classified cross-entropy is used as the loss function, and the random gradient descent method (Stochastic Gradient Descent, SGD) is used to train the model. The learning rate, weight attenuation and momentum of the three training parameters are set to 0.001, 0.00001 and 0.9, respectively, and the learning rate attenuation strategy is set. Every 5 Epoch, the learning rate decays to 80% of the original.
3.2 Evaluation index
In order to comprehensively evaluate the performance of the MobileNetV3-CA model, this experiment selected four indicators to comprehensively evaluate the recognition effect of the model: Precision, Recall, F1-score and Accuracy. The calculation formulas are as follows:
In the formula, TP, FP, FN and TN are the statistics of the classification of different tea insects and diseases by the classification model in the confusion matrix respectively. Among them, TP(True Positive) represents the number of samples whose true value is positive and identified as positive, FP(False Positive) represents the number of samples whose true value is negative but identified as positive, FN(False Negative) represents the number of samples whose true value is positive but identified as negative, and TN(True Negative) represents the number of samples whose true value is negative and identified as negative. For the purpose of identifying diseases and insects, the actual number of categories of samples to be identified is regarded as the positive sample number, while the sum of all other categories is considered the negative sample number.
3.3 Comparative experiment of transfer learning training methods
There are three common ways of transfer learning: the full parameter migration method, which involves freezing all convolution layers and only training the fully connected layer; the reuse model method, which only uses the model structure but not the pre-trained parameters; and the fine-tuning method, which involves freezing part of the convolution layers. This study adopts the fine-tuning method of freezing only part of the convolution layers. To test the effectiveness of the fine-tuning method used in this study, the above three transfer learning methods were used to train three models. Experiments were conducted based on the improved MobileNetV3-CA model using the same experimental data. The experimental results are shown in Table 2, and the variation curve of training and validation accuracy of the three transfer learning methods is shown in Figure 6.
It can be seen from Table 2 that under the same experimental conditions, the accuracy of the fine-tuning method is the highest, while the accuracy of the full migration method is the lowest. Additionally, the F1 value of the fine-tuning method is also the highest among the three methods. In the full parameter migration method, all the initial parameters of the model are obtained by pre-training, and only the fully connected layer is trained, making it difficult to optimize the model’s parameters. As shown in Figure 6, in the reuse model method, the initial parameters of the model are set randomly, and it takes a long time to improve the recognition accuracy. In contrast, the fine-tuning method uses initial parameters obtained after extensive data training, rather than random settings, allowing the model to find suitable parameters more quickly. Moreover, the fine-tuning method retrains the convolution layers in the middle of the model, making the model parameters more suitable for the tea leaf disease and insect identification task. Therefore, the four indicators in the comprehensive experiment show that the fine-tuning transfer learning method used in this study is more effective than the other two transfer learning methods.
3.4 Performance analysis of MobileNetV3-CA model
In the fine-tuning transfer learning method, the loss value change curve of the MobileNetV3-CA model on the self-built training set is shown in Figure 7. During the training process, the model’s loss value decreased rapidly in the first 10 epochs and gradually slowed down after 10 epochs of training. By the time the training reached 20 epochs, the loss value curves of the model tended to flatten, indicating that the MobileNetV3-CA model had reached saturation. Notably, during the training process, the change trend of the loss curve of the MobileNetV3-CA model on both the training and validation sets was basically the same. This shows that the overall convergence trend of the model is good and there is no overfitting, verifying the effectiveness and learnability of the MobileNetV3-CA model.
In order to further verify the performance of the MobileNetV3-CA model, the classification results on the self-built test set were analyzed. The test set contained 17 species of tea leaf pests and diseases, including 200 pictures for each species, for a total of 3400 pictures. Overall, the average recognition accuracy, recall rate, and F1-score of the MobileNetV3-CA model on the test set are 95.88%, 96.19%, and 96.01%, respectively, all exceeding 95%. These experimental results demonstrate that the improved MobileNetV3-CA model can efficiently locate and extract small feature differences in tea leaf disease and insect images.
3.5 Comparative experiment on different attention mechanisms
In order to further verify the competitive advantage of introducing the CA module into the attention module, the SE attention module in the MobileNetV3-Large model was replaced by two classical attention mechanisms, namely, the ECA (Efficient Channel Attention) module (Wang et al., 2020) and the CBAM (Convolutional Block Attention Module) (Woo et al., 2018), under the same experimental conditions.
Figure 8 shows the confusion matrix of the recognition results for each model on the self-built test set. Overall, the recognition accuracy of the MobileNetV3-Large, MobileNetV3-CBAM, and MobileNetV3-CA models is 94.45%, 94.80%, and 95.88%, respectively. This demonstrates that compared with the other two models, the MobileNetV3-CA model can more accurately identify the characteristics of diseases and insects in tea leaves, effectively improving the model’s accuracy. Details in Figure 9 show that the introduction of the ECA, CBAM, and CA modules can alleviate misclassification and omission issues in the MobileNetV3-Large model to some extent, making the model more suitable for identifying diseases and insects in tea leaves. Therefore, compared with other attention mechanisms, the introduction of the CA module can better improve the recognition performance of the MobileNetV3-Large model, verifying the competitive advantage of the CA module.
3.6 Comparative test of different models
Regarding image classification, ShuffleNetV2 and Inceptionv3 are two excellent lightweight convolutional neural networks that can be easily deployed on mobile devices. Meanwhile, AlexNet, VGG16, and the ResNet series models are also representative convolutional neural networks in visual tasks, achieving excellent results in visual classification tasks. They provide valuable reference points and comparability for the tea leaf disease and pest recognition method proposed in this study. To demonstrate the effectiveness of the proposed model, it was compared with six other models using the same experimental data and training strategy. The performance of these models on the test set is shown in Table 3.
As seen in the table, ResNet18 has strong feature extraction ability and shows good recognition performance in the experiment, but it consumes a lot of memory and computing resources. The accuracies of AlexNet and VGG16 models on the test set are slightly lower than that of the ResNet18 model. Additionally, both models have a large amount of model parameters, which require more storage space. The SqueezeNet model has small parameters, but its accuracy index is low compared to other models. The ShuffleNetV2 model uses the idea of grouped convolution to reduce the number of parameters and calculations. However, its recognition accuracy is slightly poor on the tea leaf diseases and insects image dataset with insignificant feature differences, resulting in poor model stability. Although the MobileNetV3-Large model uses deep separable convolution to restrict the depth and width of the network, it still achieves excellent results in the task of disease and insect identification in tea leaves, with performance close to ResNet18 and better than ShuffleNetV2, AlexNet and SqueezeNet. Compared with other models, the MobileNetV3-CA model achieves better recognition results, with a recognition accuracy as high as 95.88%, which is 1.33, 5.66, 1.80, 7.48, 3.65, and 1.43 percentage higher than ResNet18, AlexNet, VGG16, SqueezeNet, ShuffleNetV2, and MobileNetV3-Large, respectively. In general, the MobileNetV3-CA model not only ensures the detection speed but also improves the identification efficiency of diseases and pests in tea leaves, better balancing the complexity and recognition effect of the model.
3.7 Application of the proposed recognition model
To verify the practical application effectiveness of this method and to better support actual tea production, an application program for identifying tea diseases and insects was developed using cloud and mobile terminals. Users can capture or upload images of tea leaves with diseases or insects using their mobile devices, which are then sent to the cloud for processing. The cloud-based program identifies tea leaf insects and diseases using the developed identification model and sends the identification results back to the mobile terminal. The interface of the mobile recognition process and results can be seen in Figure 10. The recognition results include the most probable category information of diseases and insects, such as the category name and the probability of the category label, accompanied by corresponding prevention and control suggestions.
The system has been implemented in tea gardens located in Xinchang County, Zhejiang Province, and Fuding County, Fujian Province, China. It has been utilized to identify and diagnose numerous common tea leaf diseases and insects, such as tea blister blight, tea anthracnose, larva of Ectropis obliqua hypulina, Chrysopa sinica, among others, achieving an average accuracy of 90.36%. According to the test results, due to the similarities among certain diseases and insects, occasional misjudgments may occur; however, the average misjudgment rate remains below 6%. Hence, this system holds practical value for application.
4 Conclusion
In summary, this study makes several significant contributions. It introduces an enhanced classification algorithm for tea leaf disease and insect recognition based on MobileNetV3, leveraging the CA attention mechanism and transfer learning to notably improve model performance.
1. The research presented an improved classification algorithm for tea leaf disease and insect recognition, named MobilenetV3-CA, which builds upon the MobileNetV3 architecture. The findings demonstrate that incorporating the CA attention mechanism into the MobileNetV3 model could enhances the performance of disease and insect recognition in tea leaves. The introduction of the CA attention mechanism enables the model to better comprehend and utilize spatial information, thereby improving its ability to identify disease locations.
2. The study also investigates the algorithm’s effectiveness in transfer learning, validating its ability to improve model performance. Through transfer learning, the model can rapidly adapt and learn when faced with new tea leaf disease data, and the accuracy rate of disease recognition is raised from 94.45% to 95.88%, resulting in an overall increase in recognition accuracy. This provides a reliable theoretical basis and experimental support for the application of the MobilenetV3-CA algorithm in actual tea leaf disease monitoring systems.
3. The advancements in this study offer promising applications in real-world tea leaf disease monitoring and management systems, introducing innovative approaches for integrating intelligent technologies into agriculture. However, the current images of tea tree leaf pests primarily focus on the adult stage of the pests, a point at which the infestation may have already spread extensively or become significantly harmful. The next step is to collect more images of pests in their earlier stages to better facilitate early identification and warning of pest outbreaks. Additionally, future research directions may include further extending the application of this algorithm to the identification of other stages of insect pests and to other areas of crop pests and diseases, thereby expanding its impact and utility in agricultural practice.
Data availability statement
The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.
Author contributions
YaL: Conceptualization, Data curation, Formal Analysis, Funding acquisition, Methodology, Writing – original draft, Writing – review & editing. YuL: Conceptualization, Investigation, Supervision, Writing – original draft, Writing – review & editing. HL: Investigation, Methodology, Resources, Supervision, Writing – review & editing. JB: Data curation, Resources, Supervision, Writing – review & editing. CY: Data curation, Resources, Writing – review & editing. HY: Resources, Supervision, Writing – review & editing. XL: Data curation, Resources, Supervision, Writing – review & editing. QX: Conceptualization, Supervision, Writing – review & editing.
Funding
The author(s) declare that financial support was received for the research, authorship, and/or publication of this article. This study was financially supported by Science and Technology Innovation Project of the Chinese Academy of Agricultural Sciences (CAAS-ASTIP-2023-TRICAAS, TRI-ZDRW-02-04), Zhejiang Provincial Natural Science Foundation of China under grant No. LTGN23C130004, the Central Public-Interest Scientific Institution Basal Research Fund (1610212021004), and the Key R&D Program of Zhejiang (2022C02052).
Acknowledgments
We would like to thank the software engineers of the team for their development of WeChat Mini Programs.
Conflict of interest
Author YuL, CY were employed by the company Hangzhou review Ruikun Technology Co., Ltd.
The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be constructed as a potential conflict of interest.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
Bi, C., Wang, J. M., Duan, Y. L. (2022). MobileNet based apple leaf diseases identification. Mobile Networks Appl. 27, 172–180. doi: 10.1007/s11036-020-01640-1
Chen, J. D., Chen, J. X., Zhang, D. F., Sun, Y. D., Nanehkaran, Y. A. (2020). Using deep transfer learning for image-based plant disease identification. Comput. Electron. Agric. 173, 105393. doi: 10.1016/j.compag.2020.105393
Chen, R. Y., Qi, H. X., Liang, Y., Yang, M. C. (2022). Identification of plant leaf diseases by deep learning based on channel attention and channel pruning. Front. Plant Sci. 13, 1023515. doi: 10.3389/fpls.2022.1023515
Hou, Q. B., Zhou, D. Q., Feng, J. S. (2021). “Coordinate Attention for efficient mobile network design,” in 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA. pp. 13713–13722. doi: 10.1109/CVPR46437.2021.01350
Howard, A., Sandler, M., Chu, G., Chen, L. C., Chen, B., Tan, M. X., et al. (2019). “Searching for mobilenetv3,” in 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Korea (South), pp. 1314–1324. doi: 10.1109/ICCV.2019.00140
Hu, J., Shen, L., Sun, G. (2018). “Squeeze-and-excitation networks,” in 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, pp. 7132–7141. doi: 10.1109/CVPR.2018.00745
Huang, F., Liu, F., Wang, Y., Luo, F. (2019). Research progress and prospect on computer vision technology application in tea production. J. Tea Sci. 39, 81–87. doi: 10.13305/j.cnki.jts.2019.01.009
Iandola, F. N., Moskewicz, M. W., Ashraf, K., Han, S., Dally, W. J., Keutzer, K. (2016). SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and<1MB model size. ArXiv. abs/1602.07360. doi: 10.48550/arXiv.1602.07360
Jia, Z. H., Zhang, Y. Y., Wang, H. T., Liang, D. (2022). Identification method of tomato disease period based on res2net and bilinear attention mechanism. Soc. Agric. Machinery Trans. Chinese. 53, 259–266. doi: 10.6041/j.issn.1000-1298.2022.07.027
Krizhevsky, A., Sutskever, I., Hinton, G. E. (2012). “ImageNet classification with deep convolutional neural networks,” in Advances in neural information processing systems. Lake Tahoe, Nevada, USA: MIT Press. 25 (2), 1097–1105. doi: 10.1145/3065386
Li, Z. M., Xu, J., Zheng, L., Tie, J., Yu, S. (2022). Small sample recognition method of tea disease based on improved DenseNet. Trans. Chin. Soc. Agric. Eng. (Transactions CSAE) 38, 182–190. doi: 10.11975/j.issn.1002-6819.2022.10.022
Liu, J., Wang, X. W. (2021). Plant diseases and pests detection based on deep learning: a review. Plant Methods 17, 22. doi: 10.1186/s13007-021-00722-9
Liu, K. C., Zhang, X. J. (2023). PiTLiD: Identification of plant disease from leaf images based on convolutional neural network. IEEE/ACM Trans. Comput. Biol. Bioinf. 20, 1278–1288. doi: 10.1109/TCBB.2022.3195291
Mu, J. L., Ma, B., Wang, Y. F., Ren, Z., Liu, S. X., Wang, J. X. (2023). Review of crop disease and pest detection algorithms based on deep learning. Trans. Chin. Soc. Agric. Machinery 54, 301–313. doi: 910.6041/j.issn.1000-1298.2023.S2.036
Pal, A., Kumar, V. (2023). AgriDet: Plant Leaf Disease severity classification using agriculture detection framework. Eng. Appl. Artif. Intell. 119, 105754. doi: 10.1016/j.engappai.2022.105754
Pan, S. Q., Qiao, J. F., Wang, R., Yu, H. L., Wang, C., Taylor, K., et al. (2022). Intelligent diagnosis of northern corn leaf blight with deep learning model. J. Integr. Agric. 21, 1094–1105. doi: 10.1016/S2095-3119(21)63707-3
Peng, Y. H., Li, S. Q. (2023). Recognizing crop leaf diseases using reparameterized MobileNetV2. Trans. Chin. Soc. Agric. Eng. (Transactions CSAE) 39, 132–140. doi: 10.11975/j.issn.1002-6819.202304241
Qi, J., Liu, X., Liu, K., Xu, F., Guo, H., Tian, X., et al. (2022). An improved YOLOv5 model based on visual attention mechanism: Application to recognition of tomato virus disease. Comput. Electron. Agric. 194, 106780. doi: 10.1016/j.compag.2022.106780
Sandler, M., Howard, A., Zhu, M. L., Zhmoginov, A., Chen, L. C. (2018). “Mobilenetv2: Inverted residuals and linear bottlenecks,” in 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, pp. 4510–4520. doi: 10.1109/CVPR.2018.00474
Simonyan, K., Zisserman, A. (2015). “Very deep convolutional networks for large-scale image recognition,” in 3rd International Conference on Learning Representations (ICLR 2015), San Diego, USA, May 7-9, 2015, Conference Track Proceedings, pp. 1–14. doi: 10.48550/arXiv.1409.1556
Sun, Y., Wu, F., Guo, H. P., Li, R., Yao, J. F., Shen, J. B. (2023). TeaDiseaseNet: multi-scale self attentive tea disease detection. Front. Plant Sci. 14, 1257212. doi: 10.3389/fpls.2023.1257212
Sun, Y. Y., Jiang, Z. H., Zhang, L. P., Dong, W., Rao, Y. (2019). SLIC_SVM based leaf diseases saliency map extraction of tea plant. Comput. Electron. Agric. 157, 102–109. doi: 10.1016/j.compag.2018.12.042
Szegedy, C., Liu, W., Jia, Y. Q., Sermanet, P., Reed, S., Anguelov, D., et al. (2015). “Going deeper with convolutions,” in 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). Boston, USA: IEEE Press. pp. 1–9. doi: 10.1109/CVPR.2015.7298594
Tan, M., Le, Q. V. (2019). EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks. In: 36th Int. Conf, Mach, Learn. ICML 2019 2019-June, pp. 10691–10700. doi: 10.48550/arXiv.1905.11946
Ullah, Z., Alsubaie, N., Jamjoom, M., Alajmani, S. H., Saleem, F. (2023). EffiMob-Net: A deep learning-based hybrid model for detection and identification of tomato diseases using leaf images. Agriculture 13, 737. doi: 10.3390/agriculture13030737
Waheed, A., Muskan, G., Gupta, D., Goyal, M., Gupta, D., Khanna, A., et al. (2020). An optimized dense convolutional neural network model for disease recognition and classification in corn leaf. Comput. Electron. Agric. 175, 105456. doi: 10.1016/j.compag.2020.105456
Wang, Q. L., Wu, B. G., Zhu, P. F., Li, P. H., Hu, Q. H. (2020). “ECA-net: efficient channel attention for deep convolutional neural networks,” in 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, pp. 11531–11539. doi: 10.1109/CVPR42600.2020.01155
Woo, S., Park, J., Lee, J. Y., In, S. K. (2018). “CBAM: Convolutional block attention module,” in 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Munich, Germany, September 8–14, 2018, Proceedings, Part VII. Springer-Verlag, Berlin, Heidelberg, 3–19. doi: 10.1007/978-3-030-01234-2_1
Xue, Z., Xu, R., Bai, D., Lin, H. (2023). YOLO-tea: A tea disease detection model improved by YOLOv5. Forests 14, 415. doi: 10.3390/f14020415
Yousef, A. G., Abdollah, A., Mahdi, D., Joe, M. M. (2022). Feasibility of using computer vision and artificial intelligence techniques in detection of some apple pests and diseases. Appl. Sci. 12, 906–920. doi: 10.3390/app12020906
Keywords: tea leaf diseases and insects, convolution neural network, transfer learning, MobileNetV3, recognition and classification
Citation: Li Y, Lu Y, Liu H, Bai J, Yang C, Yuan H, Li X and Xiao Q (2024) Tea leaf disease and insect identification based on improved MobileNetV3. Front. Plant Sci. 15:1459292. doi: 10.3389/fpls.2024.1459292
Received: 04 July 2024; Accepted: 10 September 2024;
Published: 27 September 2024.
Edited by:
Zhenghong Yu, Guangdong Polytechnic of Science and Technology, ChinaReviewed by:
Xiujun Zhang, Chinese Academy of Sciences (CAS), ChinaDr. Mohammad Shameem Al Mamun, Bangladesh Tea Research Institute, Bangladesh
Copyright © 2024 Li, Lu, Liu, Bai, Yang, Yuan, Li and Xiao. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Qiang Xiao, eHF0ZWFAdmlwLjE2My5jb20=