Skip to main content

METHODS article

Front. Plant Sci., 21 November 2024
Sec. Plant Pathogen Interactions
This article is part of the Research Topic Innovative Strategies for Enhancing Crop Resilience Against Plant Viral Diseases View all 12 articles

Plant pest and disease lightweight identification model by fusing tensor features and knowledge distillation

  • College of Artificial Intelligence, Tianjin University of Science & Technology, Tianjin, China

Plant pest and disease management is an important factor affecting the yield and quality of crops, and due to the rich variety and the diagnosis process mostly relying on experts’ experience, there are problems of low diagnosis efficiency and accuracy. For this, we proposed a Plant pest and Disease Lightweight identification Model by fusing Tensor features and Knowledge distillation (PDLM-TK). First, a Lightweight Residual Blocks based on Spatial Tensor (LRB-ST) is constructed to enhance the perception and extraction of shallow detail features of plant images by introducing spatial tensor. And the depth separable convolution is used to reduce the number of model parameters to improve the diagnosis efficiency. Secondly, a Branch Network Fusion with Graph Convolutional features (BNF-GC) is proposed to realize image super-pixel segmentation by using spanning tree clustering based on pixel features. And the graph convolution neural network is utilized to extract the correlation features to improve the diagnosis accuracy. Finally, we designed a Model Training Strategy based on knowledge Distillation (MTS-KD) to train the pest and disease diagnosis model by building a knowledge migration architecture, which fully balances the accuracy and diagnosis efficiency of the model. The experimental results show that PDLM-TK performs well in three plant pest and disease datasets such as Plant Village, with the highest classification accuracy and F1 score of 96.19% and 94.94%. Moreover, the model execution efficiency performs better compared to lightweight methods such as MobileViT, which can quickly and accurately diagnose plant diseases.

1 Introduction

The variety of crops grown in countries around the world is large and intensive, and according to the world food and agriculture statistical yearbook released by the Food and Agriculture Organization of the United Nations (FAO) in 2023, the global agricultural economy reached $3.7 trillion and employed a population of approximately 873 million people, which accounts for 27% of the global labor force (FAO, 2023). On average, more than 40% of natural losses in agricultural production are caused by plant pests and diseases each year, resulting in a global economic loss of more than 220 billion dollars (FAO, 2022). Thus, once a crop is infested with pests and diseases, it is very easy to cause widespread infection of crops, which seriously affects crop yields and restricts the development of agricultural productivity. However, traditional identification of plant pests and diseases is usually done by experienced plant pathologists or farmers to diagnose the type of leaf pests and diseases. This not only requires a lot of time and effort, but also the process is subjective and limited, which makes it difficult to accurately determine the type of disease and thus leads to the aggravation of plant pests and diseases (Sajitha et al., 2024). Therefore, how to quickly and accurately identify the types of plant pests and diseases is the crucial to ensure the safety and stability of agricultural production, and its research has important theoretical significance and application value.

With the continuous development of smart agriculture and artificial intelligence technology, scholars combined with computer vision related theories to carry out research on plant pest and disease recognition methods. Especially, the research on intelligent plant image recognition based on machine vision and deep learning has achieved better practical results (Rimal et al., 2023). Machine vision methods usually use traditional image processing algorithms or manually designed feature classifiers, which mainly rely on the distinguishing features of different pests and diseases to design the recognition scheme, and are widely used in crop pest detection and classification (Thakuria et al., 2023). However, plant images collected in the natural environment are easily interfered by factors such as light in the environment, which leads to errors in the detection results and inaccurate classification (Zheng and Zhang, 2022). At the same time, with the increasing variety of plant diseases, it is difficult to construct suitable classifiers to distinguish approximate representations by means of manual feature selection. Therefore, traditional machine vision-based pest and disease detection methods are difficult to achieve effective recognition results (Ali et al., 2024). With the continuous development of deep learning in recent years, network models represented by multilayer convolutional neural network (CNN) and attention mechanism have achieved effective results in plant pest and disease recognition (Math and Dharwadkar, 2022). Deep learning based plant pest and disease recognition technique is automated to extract global and contextual features of pest and disease images compared to traditional recognition methods mainly using supervised learning (Roxa et al., 2023). It avoids manual selection of features but extracts richer feature information through autonomous learning, which can cope with diverse crop environments (Sharma et al., 2022). And it can handle massive image data with strong robustness and high accuracy (Shamsul Kamar et al., 2023). In order to further improve the accuracy and training efficiency of the model, techniques such as KD and migration learning are usually combined to fine-tune the parameters of the model to achieve knowledge migration (Khan et al., 2022). Although the above methods can effectively promote the construction of smart agriculture, with the increasing types of plant pests and diseases, as well as the limitations of hardware equipment in the actual application scenarios, there are still some problems that affect the efficiency and accuracy of plant pests and diseases recognition (Manavalan, 2022). 1) The complex structure and large number of parameters of the deep neural network greatly affect the recognition efficiency of the model. 2) The lightweight network model is difficult to mine and fuse the key features, which leads to information loss and thus reduces the accuracy of the model. To address the above problems, we proposed a Plant pest and Disease Lightweight identification Model by fusing Tensor features and Knowledge distillation (MTS-KD) to realize accurate and efficient diagnosis of multiple plant pests and diseases. The specific contributions are as follows.

● Constructed Lightweight Residual Blocks based on Spatial Tensor (LRB-ST). Enhanced the perception and extraction of shallow detail features of plant images by introducing spatial tensor, and reduced the number of model parameters by using depth-separable convolution to improve the diagnostic efficiency.

● Proposed Branch Network Fusion with Graph Convolutional features (BNF-GC). The image super-pixel segmentation was realized by using spanning tree clustering based on pixel features, and the graph convolution neural network was used to extract correlation features to improve the diagnostic accuracy.

● Designed Model Training Strategy based on Knowledge Distillation (MTS-KD). Train the pest and disease diagnosed model by building a knowledge migration architecture, fully balancing the accuracy of the model with the diagnosis efficiency.

The sections of this paper are organized as follows, section 2 focus on introducing and analyzing the current research related to plant pest and disease diagnosis. Section 3 focus on the method proposed in this paper. Section 4 describes the qualitative and quantitative analysis of this paper’s method with other image classification methods to verify the accuracy of the method for plant pest and disease diagnosis. Section 5 discusses the convergence of the method and the configuration of each module to fully justify the method in terms of model construction and parameter selection. Finally, section 6 gives a summary and future research.

2 Related work

In order to fully utilize computer or artificial intelligence techniques to assist in pest control, machine learning and deep learning based methods have been proposed for plant image recognition (Cetiner, 2022). While improving the accuracy of more plant pest and disease recognition, it also focuses on lightweight design to improve the diagnostic efficiency. It effectively solves the problem that traditional crop pest and disease image recognition methods rely heavily on manual feature extraction and have poor generalization ability for image recognition in complex backgrounds (Rustia et al., 2022).

The application of machine learning in pest recognition has significantly enhanced the efficiency of pest control. Zhao et al. (2022) and Johari et al. (2022) proposed a multi-step plant adversity recognition method based on hyperspectral imaging and continuous wavelet analysis, used k-mean clustering and support vector machine algorithms to detect abnormal regions of tea tree leaves, and used the random forest algorithm to construct the tea tree discriminant model. Motie et al. (2023) extracted and analyzed spectral vegetation features and used support vector machine (SVM) to classify wheat diseased plants. Bhandari et al. (2023) implemented EfficientNetB5 with a tomato leaf disease (TLD) dataset without any segmentation, and the model achieved high accuracy. Catalkaya et al. (2024) established a KASP (Kompetitive Allele-Specific PCR) analysis method for plant pest identification by designing two forward primers and one reverse primer to enhance the identification accuracy of the algorithm. Pansy and Murali (2023) proposed an unmanned aerial imaging remote sensor for spatio-temporal resolution identification of mango pests and diseases using fuzzy C-mean clustering for diseased leaves and pests, respectively. Despite the deep mining of plant image features, to further enhance the recognition and accuracy, Aldakheel et al. (2024) and Subbaian et al. (2024) applied the YOLOv4 algorithm to plant leaf disease detection. And data enhancement techniques such as histogram equalization and level flipping were used to improve the dataset and effectively enhance the accuracy of plant image disease classification. Zhu et al., 2023 proposed a data evaluation method based on martingale distance and entropy to address the problem of lack of labeling data in intelligent pest identification. This method can filter high value data, thus achieving effective pest recognition performance with small data size. Chodey and Shariff (2023) proposed a Self-Improving Tephritid Swarm Optimization Algorithm (SITSA) to train a pest detection model by selecting the optimal weights and designing a grey scale covariance matrix based feature extraction method to segment the image. Nandhini and Brindha (2024) proposed a new data enhancement technique and feature fusion technique that fuses multi-scale features from global feature extraction network and visual regeneration network to improve accuracy as well as robustness. Although the efficiency of plant pest and disease recognition has been substantially improved, the recognition effect still needs to be improved. This is due to the fact that traditional machine learning methods usually need to select features manually, and their feature representation capability is relatively limited to fully capture the complex structure and information in the image.

Deep learning currently possesses robust feature representation capabilities, enabling it to automatically extract essential semantic information from plant disease and pest images. Li et al. (2024) proposed an interactive bilinear Transformer network, which utilizes fine-grained recognition techniques to realize the types of garden plant diseases. Xia et al. (2023) and Zhang et al. (2023) proposed a pest classification method based on Convolutional Neural Network (CNN) and improved Vision Transformer model, which extracts the features of the objects at different scales and fine-grains, to address the problems of low efficiency of pest classification methods, which are not adapted to large-scale environments. Wei et al. (2022) proposed a multi-scale feature fusion based crop pest classification method (MFFNet), which obtains the deep feature information of pest images through multiple convolutional operations to accurately recognize and classify crop pests. Liu et al. (2024) and Yu et al. (2022) proposed a plant pest type recognition method based on YOLOv5 by introducing modules such as hierarchical classification and attention mechanism, respectively, which effectively avoided the problems of time-consuming, laborious, and inaccurate manual classification. Xiao et al. (2022), on the other hand, combined hyperspectral imaging with deep learning, which designed a spectral feature extraction module through one-dimensional convolution and attention mechanism between spectral channels, effectively utilizing spectral information to improve detection accuracy. On this basis, Cheng et al. (2022) used to recognize the category of tomato diseases in images by fine-tuning the pre-trained model. Qiang et al. (2023) proposed a dual backbone network based pest detection method for citrus leaves in response to the problem that pests on the surface of plants are difficult to distinguish due to their small size and camouflage, which utilizes a single-shot multi-box detector improved by a dual backbone network to enhance the detection accuracy. Dai et al. (2024) and Li et al. (2024) proposed similar deep information feature fusion networks extracting and fusing relevant features from different network layers, respectively, while fusing contextual information at different scales using pyramid-squeezed attention (PSA) to produce better pixel-level attention for improved localization of plant disease areas. Ishengoma et al. (2022) proposed a hybrid convolutional neural network (CNN) model. It was also combined with Unmanned Aerial Vehicle (UAV) technology to build a parallel architecture using two separate models (i.e., VGG 16 and InceptionV3) to realize the identification of plant diseases in large areas. Shafik et al. (2023) proposed an enhanced Convolutional Neural Network (CNN) along the use of Long Short Term Memory (LSTM) using Majority Voting Integrated Classifier for plant disease and pest recognition.

Plant pest and disease identification methods based on artificial intelligence not only enhance the accuracy of disease type identification but also advance agricultural automation and intelligence, helping to mitigate losses caused by pests and diseases. Machine learning-based methods struggle to efficiently extract potential critical features and often require substantial manual labeling efforts. Although deep neural network architectures improve accuracy, they significantly increase model complexity, presenting challenges in parameter training and practical deployment (Li et al., 2023). This complexity results in longer training times and demands greater computational resources, limiting recognition efficiency, particularly in real-time applications. Additionally, the intricacies of hyperparameter tuning and the risk of overfitting pose significant obstacles for practitioners, especially in resource-constrained environments. Although lightweight network models have been proposed to address these issues, fewer parameters often struggle to capture key features in large image datasets. Additionally, noise and other interference present in the images are challenging to filter effectively. These disturbances can obscure important features that are critical for accurate identification of plant pests and diseases. Therefore, how to balance the accuracy and efficiency of plant pest and disease diagnosis methods remains to be solved.

3 Methodology

3.1 Model framework

In order to improve the efficiency and accuracy of plant pest and disease diagnosis, we constructed a classification model applied to plant pest and disease diagnosis based on a deep learning network architecture, combined with model compression and training methods to achieve accurate identification of various categories of plant pest and disease classes.

The PDLM-TK model comprises three primary components, as shown in Figure 1. First, a Lightweight Residual Blocks based on Spatial Tensor (LRB-ST) network is designed, which integrates multiple spatial tensors to extract semantic information from plant disease images progressively, from the initial to the final layers. The input plant disease and pest images undergo initial downsampling, followed by four sequential layers of LRB-ST, to capture advanced semantic information about the disease. Next, the Branch Network Fusion with Graph Convolutional features (BNF-GC) is introduced to deeply mine and fuse different levels of residual block features using graph convolution. BNF-GC applies a graph convolutional neural network to the output of each LRB-ST layer, focusing on localized pest and disease information to guide classification. Finally, a Model Training Strategy based on Knowledge Distillation (MTS-KD) is implemented. This strategy utilizes the plant pest and disease dataset to train a teacher network, enabling knowledge transfer to the student network. KD is performed during the student network’s training using the target dataset.

Figure 1
www.frontiersin.org

Figure 1. General framework diagram of PDLM-TK.

The number of network layers and parameters of PDLM-TK are presented in Table 1. PDLM-TK primarily consists of four LRB-STs and BNF-GC as the main feature extraction structure, with a fully connected multi-classification network appended at the end for pest and disease type recognition. Among these, LRB-ST is primarily responsible for extracting key features from pest and disease images, while BNF-GC focuses on mining the correlation features between image regions, thereby achieving a balance between global and local image features. Plant pest and disease images are downsampled through preliminary convolution and pooling layers. Subsequently, feature maps at various levels are deeply mined by LRB-ST and BNF-GC, respectively. It is important to note that LRB-ST utilizes Depth-Separable Convolution (DSC), whereas BNF-GC performs feature extraction using Graph Convolution Network (GCN). The number of output feature map channels for each LRB-ST corresponds to the output feature vector size of BNF-GC. MTS-KD, in turn, is based on PDLM-TK, which serves as the student network for knowledge distillation. The remaining sections of this chapter provide a detailed description of each of these modules.

Table 1
www.frontiersin.org

Table 1. Network layers and parameters of PDLM-TK.

3.2 Lightweight residual blocks based on spatial tensor

With the continuous stacking of network layers, the high-level semantic features embedded in images are continuously mined and acquired, but the shallow image semantic features are also easily lost. These features play a crucial role in constructing more complex representations, especially in plant pest and disease recognition, where the texture of some leaves and the color change of damaged areas have a great impact on the final classification results. The residual network enhances the model’s perception of small changes in the input image more effectively by enhancing the underlying semantics. Aiming at the problem that it is difficult to effectively capture and transfer feature information of different dimensions in residual networks, we propose LRB-ST. The principle is shown in Figure 2. The semantic information of different dimensions in the input feature graph is mined by defining a trainable spatial tensor. And it is fused with the underlying data as a way to enhance the feature mining ability of the classification model. Meanwhile, in order to improve the classification efficiency of the model, Depthwise Separable Convolution (DSC) is used to replace the traditional convolution operation to reduce the number of model parameters. The specific steps are as follows.

Figure 2
www.frontiersin.org

Figure 2. Schematic diagram of LRB-ST.

Firstly, the spatial tensor Pxy is defined to extract the residual block features. For the input layer i feature block Blocki, the spatial tensor Pxy is defined according to its input size, which is fused by multiplying with the feature maps of each channel of Blocki on the original residual connection. The computational procedure is shown in Equation 1. Where DSW represents the depth separable convolution and σ represents the normalization and activation operation on the features.

Blocki=σ(DSW(Blocki))+PxyBlocki(1)

Secondly, depth-separable convolution is used to reduce the number of parameters. Feature extraction is achieved by depth convolution and pointwise convolution, and the input block feature Blocki is stacked with the feature fusion result in Equation 1 on an element-by-element basis to obtain the fused feature Blocki after two depth-separable convolutions and normalized activation. After that, the feature Blocki+1 of the residual block is output through the same two depth-separable convolution and used for the calculation of the next residual block. As shown in Equation 2.

Blocki+1=Blocki+σ(DSW(Blocki))(2)

The overall structure of the LRB-ST is depicted in Figure 2, comprising two feature superposition processes and one downsampling operation. This study adopts the ResNet18 architecture to facilitate the deep feature extraction of plant pest and disease image data by superimposing four LRB-STs as the backbone network. This approach enables the model to focus on features across different levels while ensuring the effective transmission of deep semantic information. Furthermore, it effectively reduces the number of model parameters, thereby enhancing the classification efficiency of model.

3.3 Branch network fusion with graph convolutional features

Since there are factors such as background and noise in the plant image data besides the target region, and the use of deep learning-based classification model can extract key features, but it is still difficult to avoid interference by redundant information, which in turn affects the results of plant pest and disease diagnosis. In order to improve the robustness of the model and mine the correlation features of different regions in the image, we propose BNF-GC, the principle of which is shown in Figure 3.

Figure 3
www.frontiersin.org

Figure 3. Schematic diagram of BNF-GC.

BNF-GC mainly uses super-pixel segmentation to effectively reduce image complexity and extract potential regional features, which can effectively improve the classification performance of the model. Meanwhile, the graph convolutional neural network is used to mine the intrinsic correlation of different regions, which further improves the accuracy of pest and disease diagnosis.

Step1: Construct plant disease image spanning tree ST(v,e,w). Map the plant image feature data into an undirected graph G and represent it as G=(V,E). Where each pixel point represents a node in the undirected graph and two neighboring pixel points form an edge e(vi,vj). Where the weights w of the edges are jointly determined by the coordinates vi(xi,yi) and pixel values pi of the pixel points. The distance between nodes is normalized according to the size of the image M×N and fused with the normalized pixel value features as the weight of each node pixel. The calculation process is shown in Equation 3.

w=[(xixj)2+(yiyj)2]/(M2+N2)+pi/255(3)

Step2: Aggregate the nodes in the region to form a super-pixel segmented image. The nodes are aggregated to form different segmentation regions R according to the minimum spanning tree method. The maximum weight Max(wi) on the minimum spanning tree in a segmented region indicates the degree of intra-class variation of its internal nodes, while the minimum weight Min(wi) between different regions represents its inter-class variation. Therefore, the regions that meet the condition Max(wi)Min(wi) are merged, after which the inter and intra-class differences are re-compared until the expected number of segmented regions is reached and then stopped, and finally the plant pest image Seg(R1,R2,R3,Rn) containing multiple aggregated regions is obtained. The calculation process is shown in Equation 4.

Seg(Ri)={iRRi(wi)Max(wi)Min(wi)Ri(wi)other(4)

Step3: Calculate the feature terms of each segmentation region of the superpixel map Seg(s1,s2,s3,sn). Since the image in each segmentation region contains rich semantic information, this paper selects the pixel mean value pi, the region center coordinate (cxi,cyi) and the number of included pixel points ci in the segmentation region as the superpixel point feature si(pi,cxi,cyi,ci) of each segmentation region. And the feature values of each region are normalized for subsequent feature mining respectively. The calculation process is shown in Equations 57.

pi=1nj=0npjjRi(5)
(cxi,cyi)=(1mi=0mci,1ni=0ncyi)   m,nRi(6)
ci=count(Ri)(7)

Step4: Construct graph convolutional neural network GCN to mine local correlation information. According to the dimension of image features in the segmented region, construct the graph convolution neural network GCNi corresponding to the ith residual block, and then mine the regional correlation through multi-layer graph convolution computation to get the feature vector GFi. Finally, the correlation features are multiplied with the output feature map Blocki1 of the previous residual block to generate the fusion Blocki, which is used for the feature computation of the next residual block. The computation process is shown in Equations 8 and 9.

GFi=GCNi(si)(8)
Blocki=Blocki1GFi(9)

The BNF-GC module creates a graph convolutional network branching structure for deep mining of residual block features, which effectively realizes regional relevance feature extraction for plant pest and disease image data. The module enhances the classification performance of the model by enhancing the deep fusion of the underlying semantic information, which emphasizes the main features of the target region and effectively avoids the interference of the background and other information.

3.4 Model training strategy based on knowledge distillation

Although existing publicly available datasets contain more images of plant pests and diseases, the lightweight model proposed in this paper contains fewer parameters, which makes it difficult to capture the rich features of the dataset. Especially when the dataset is limited or noisy, it is difficult for the model to learn the implicit knowledge by using only the dataset training. In this regard, this paper proposes MTS-KD, as shown in Figure 4. The ResNet18 network is used as the backbone network of the teacher model, and the prediction results and soft labels of the teacher model are used to guide the training of the student model. This approach effectively enhances the model’s learning capacity and improves its generalization ability, which helps mitigate the significant bias present in the plant dataset. As a result, the model’s training efficiency is improved without compromising its high accuracy.

Figure 4
www.frontiersin.org

Figure 4. Framework diagram of MTS-KD.

During the training process of the model through MTS-KD, the plant pest and disease images were first categorized according to the training set and test set. After that, they were fed into both teacher-student models for forward computation. The Softmax classification result produced by the teacher model under high temperature T is used as soft label SL.

The student model produces the same prediction results after going through the training based on KD. In this case, the Softmax output at the same temperature T condition is Ls, while the prediction result produced at T=1 is Lh. Ls in MTS-KD uses Kullback-Leibler Divergence to calculate the relative difference between the predictions of the teacher-student models. In turn, Lh uses the cross-entropy loss function to calculate the difference between the results predicted by the student model and the real pest label. Afterwards, the losses of these two components are summed up as the total model loss for optimization and training of the parameters. Where the loss function formulas for Ls and Lh are shown in Equations 10 and 11.

Ls=i=0NsiTlog(tiT)(10)
Lh=i=0Nlilog(ti1)(11)

Where N is the number of total categories in the dataset. tiT refers to the value that the teacher model predicts as i after Softmax at temperature T. ti1 represents the true label prediction result of the teacher model at temperature 1. Similarly, siT refers to the value of T predicted by the student model at temperature i, and li represents the value of the ith true label in the total number of categories N. The two calculations are shown in Equation 12.

piT=exp(Pi/T)Nexp(PN/T)(12)

piT in Equation 12 can be used to calculate the predictions for the teacher model and the student model, respectively. Therefore, the total loss Lt is defined as shown in Equation 13. Loss weights are employed to balance the influence between soft targets and hard targets from the teacher model during the training process of the student model. In the context of plant pest and disease identification, it is particularly important to emphasize the knowledge derived from the teacher model for training the student network. Where λ is the loss weight for student model training, and (1λ) is the loss weight for prediction layer distillation.

Model parameter optimization is guided by the Knowledge Distillation (KD) training strategy, which facilitates the transfer of knowledge from the teacher model to the student model. This innovative approach allows the student model to more effectively replicate the feature extraction capabilities of the teacher model, thereby enhancing its ability to grasp complex patterns and nuances in the input images. Furthermore, this knowledge transfer not only improves the classification and recognition performance of the student model but also contributes to training efficiency. When compared to training from scratch, the student model is able to converge to optimal performance levels more quickly and requires less time and resources. Consequently, the implementation of KD not only enhances the capabilities of student models but also simplifies the overall training process, establishing it as a compelling strategy for improving the effectiveness of artificial intelligence in plant pest and disease identification.

4 Experimental analysis

4.1 Experimental environment and dataset

For the configuration of the experimental environment, the input plant pest and disease image data were randomly clipped into (3, 512, 512) size inputs. The experiments are based on PyTorch 1.11.0 deep learning framework and the operating system of the experimental environment is Ubuntu 20.04. We implement the model code using Python 3.9 programming language and the GPU hardware platform is used with 2 pieces of RTX3080.Adam is chosen as the model optimizer. The batch of input images is 8 at a time. The learning rate is initially set to 0.01, and the number of iterations for model training is set to 120. All models are trained from scratch to ensure the fairness of the comparison results. The configuration of specific parameters is shown in Table 2.

Table 2
www.frontiersin.org

Table 2. Configuration table of experimental environment and parameters.

To evaluate the effectiveness of the proposed method, two plant datasets, Ai Challenger (AI Challenge, 2018), Plant Village (Hughes and Salathé, 2015), and an insect dataset, IP102 (Wu et al., 2019), are selected in this paper. The Ai Challenger plant disease identification dataset is shown in Figure 5A, which includes 31,718 plant leaf images with 61 categories, including 10 species such as apple and 27 pests. Plant Village dataset is shown in Figure 5B, which is labelled by crop pathologists and contains 54309 images with 13 species and 26 crop disease categories. IP102 is a field-constructed large-scale dataset used for pest identification and is shown in Figure 5C. It has a total of 75,222 images containing 102 common pests with an average of 737 samples per class. Some images of plant pests and diseases in the three datasets are shown in Figure 5. The experimental datasets were divided according to the 8:2 training and testing sets.

Figure 5
www.frontiersin.org

Figure 5. Selected images from the three datasets used for the experiment. As shown in Figures (A, B) are five plant leaves from the Ai Challenger and Plant Village datasets, respectively, and the five images in Figure (C) are plant insects from the IP102 dataset.

4.2 Comparative models and evaluation metrics

In order to demonstrate the superiority of the proposed method, the experiment uses five lightweight classification models such as ShuffleNet-V2 (Ma et al., 2018), MobileNet-V3-Large (Howard et al., 2020), EfficientNet B1 (Tan and Le, 2019), MobileViT-S (Mehta and Rastegari, 2022) and Inception v3 (Szegedy et al., 2015) to carry out comparison experiments with PDLM-TK. Since this study belongs to the task of pest and disease image classification, the various models are quantitatively evaluated using Precision (P), Recall (R) and F1 score (F1). And the confusion matrix is drawn according to its evaluation results to visualize the correct classification of the models in each category and realize the visual evaluation of the model performance. Its calculation is shown in Equations 1315.

Precision=TPTP+FP(13)
Recall=TPTP+FN(14)
F1=2×Precision×RecallPrecision+Recall(15)

Where TP is the correct image pest category, the number of images that the model correctly predicts as positive instances. FP is the false positive instances of the image, the number of images that the model incorrectly categorizes as positive instances. TN is the number of images that the model correctly predicted as other pest and disease types. FN denotes the number of images that the model incorrectly predicted as other categories when categorizing the correct category.

Considering the large number of pest and disease categories in the plant and insect datasets used, in order to evaluate the diagnostic performance of the proposed methods more comprehensively, the experimental evaluation was statistically analyzed using the Top1 accuracy rate (the accuracy rate of the pest category ranked No. 1 in the model diagnostic results in accordance with the actual results) and the Top5 accuracy rate, respectively.

At the same time, in order to evaluate the diagnostic efficiency of the models, the experiments not only analyzed the performance of the above six models in the three indexes of Acc, P and F1, but also evaluated the number of parameters and the amount of floating-point calculations (GFLOPs) included in the models. In general, the larger the number of parameters included in the model, the lower its training and execution efficiency, and the larger the amount of floating point calculations, the slower the model diagnosis speed.

4.3 Quantitative effectiveness evaluation of plant pest and disease diagnostics

The results of the evaluation of Top1 accuracy of six models on three plant pest and disease datasets are shown in Table 3. The comparison shows that multiple models have the best classification results on the Plant Village dataset. The reason for this is analyzed to be due to the fact that this dataset contains fewer types of pests and diseases as compared to the other two, and therefore the models classify better on this dataset with less difference in the amount of data. The PDLM-TK method achieved the best P, exceeding the best EfficientNet B1 model by 1.82%, but slightly lower in R and F1 values. This is due to the fact that the composite scaling strategy used by EfficientNet can well utilize the feature extraction ability of the convolutional layer to form a more complete deep learning model network structure. PDLM-TK performs even better on the other two datasets, exceeding the other models by an average of 2.54%, 1.22% and 1.86% on P, R and F1. The experimental results effectively demonstrate the effectiveness of the PDLM-TK method proposed in this paper in plant pest and disease diagnosis. Specifically, the method not only achieves high classification accuracy across various datasets but also showcases superior efficiency compared to traditional approaches.

Table 3
www.frontiersin.org

Table 3. Evaluation results of Top1 accuracy for various models.

In order to evaluate the comprehensive classification performance of the methods and avoid the error that exists in using only Top1 accuracy, as shown in Table 4, the evaluation results of Top5 accuracy of various models on the three plant diseases and pests datasets are shown in Table 3. According to the experimental results, it can be seen that it is in line with the Top1 accuracy assessment, and the Top5 accuracy assessment results of each model are more accurate compared to the Top1 accuracy. Meanwhile, the PDLM-TK method performed better in classifying pests and diseases on multiple datasets compared to other methods. The highest P, R and F1 values were found on the Ai Challenger and IP102 datasets, and the R and F1 values on the Plant Village dataset were higher than the Inception v3 optimal model by 2% and 0.82, respectively. This shows that our proposed PDLM-TK can realize the diagnosis of plant pests and diseases in a more comprehensive and accurate way.

Table 4
www.frontiersin.org

Table 4. Evaluation results of Top5 accuracy for various models.

4.4 Plant pest and disease diagnosis effect visualization and analysis

As shown in Figure 6, the average accuracy histograms of methods such as PDLM-TK on the three plant pest and disease datasets are shown. Analyzing along the direction of the dataset, it can be seen from the height of the histogram that the three colors of the bar represented by PDLM-TK have the highest average on the three datasets. This proves that its combined performance is better on both Top1 and Top5 assessment methods. Meanwhile, according to the column height performance of different models, it can be seen that the heights of the columns of the three colors of PDLM-TK are closer to each other. This indicates that the method has stronger classification accuracy and stability for plant pest and disease images, and can realize more accurate pest and disease diagnosis.

Figure 6
www.frontiersin.org

Figure 6. Histogram of Top1 and Top5 average accuracy of various models on three datasets.

In order to further observe the performance of each model on the test set of plant pests and diseases in a more intuitive way, we utilize the confusion matrix to present the test set classification results of each model. The confusion matrix can be taken in the form of a matrix to summarize the real and predicted categories, and to observe the differences that exist in the prediction results for the classification of different categories. Due to the large number of labels and samples in the plant pest and disease dataset, the numbers in it are no longer displayed, and the performance of each model is only observed and analyzed through the model classification results. Each square in the horizontal and vertical axes of the confusion matrix represents the category corresponding to plant pests and diseases, and the color of each square corresponds to the number of images classified into that region. The darker the color on the diagonal represents the more images that are classified correctly.

The confusion matrix for each model’s classification prediction for the test set portion of the Ai Challenger dataset is shown in Figure 7. From the distribution of squares of different colors in the confusion matrix, it can be seen that the MobileNet-V3-Large and MobileViT-S models have more uniform diagnosis results for different categories of plant pests and diseases, which is mainly manifested by the similar colors of the squares on the diagonal line, but their classification effect is lower than that of the MobileNet-V3-Large and Inception v3 models. Inception v3 has a better classification effect on some kinds of plant pest and disease images, but there is a phenomenon of misclassification of some categories, which is mainly reflected in the lighter color of the squares in the middle of the diagonal of some categories. The confusion matrix corresponding to the PDLM-TK method has a more uniform and saturated color distribution on the diagonal, which shows that its classification effect is more stable.

Figure 7
www.frontiersin.org

Figure 7. Confusion matrix of different methods on Ai Challenger dataset.

The confusion matrix of each model for classifying images on Plant Village test set is shown in Figure 8, which is a richer dataset with fewer disease categories. As shown by the color and distribution of diagonal squares, the methods have better classification results. Especially, the two methods, Inception v3 and EfficientNet B1, have the majority of dark-colored squares on the diagonal. The MobileNet-V3-Large method has multiple light-colored squares compared to the two, indicating a partial misclassification. While PDLM-TK has the darkest colored squares compared to the other methods and all of them are distributed on diagonal. It can be fully demonstrated that PDLM-TK has strong classification performance for plant pest disease image diagnosis.

Figure 8
www.frontiersin.org

Figure 8. Confusion matrix for different methods on the Plant Village dataset.

According to the confusion matrix distribution shown in Figure 9, it can be seen that the models are slightly less effective than the other two datasets in classifying insects in the IP102 dataset. The reason is due to the large number of insect species contained in this dataset, and the lightweight classification models are difficult to balance multiple critical features. Among them, two models, EfficientNet B1 and Inception v3, have better classification effects than the other models and can basically realize the discrimination of multiple types of insects. The PDLM-TK method proposed in this paper has the best classification performance compared to the other models because it trains the model by KD, which has a strong feature capturing ability despite the small number of model parameters. In summary, the validation of the plant leaf and insect datasets demonstrates that the PDLM-TK method is able to accurately realize plant pest and disease diagnosis.

Figure 9
www.frontiersin.org

Figure 9. Confusion matrix for different methods on the IP102 dataset.

5 Discussion

5.1 Evaluation of model stability and efficiency

Plant pest and disease diagnosis should not only have high accuracy, but also the stability and efficiency of the model is equally important, which is related to the classification effect and diagnosis quality in the practical application of agriculture. Therefore, in this paper, several models are trained on Ai Challenger using the same experimental environment, and their loss functions are plotted as line graphs for analysis. At the same time, the models are ranked according to the number of parameters they contain and their actual execution efficiency, and are compared and analyzed in the form of a table.

As shown in Figure 10, the change process of loss when the six methods are trained on the IP102 dataset is shown. From the fluctuation of the folded line, it can be seen that the PDLM- TK method proposed in this paper converges rapidly in the first 20 iteration loops and decreases steadily with the growth of the number of iterations. Notably, after the 110th iteration, the models exhibited stability. While the other methods exhibit similar fold convergence processes, their fluctuations are more pronounced, and their final converged values are higher. This clearly demonstrates that the method proposed in this paper is more stable and achieves faster convergence.

Figure 10
www.frontiersin.org

Figure 10. Loss curves for various methods on the IP102 dataset.

To evaluate the actual running efficiency of each model, this paper uses GFLOPs and Params metrics to measure the execution speed of the models. As shown in Table 5, EfficientNet B1 has the smallest amount of computation, and PDLM-TK is only 0.1 behind it in the second place. As shown by the parameter count, PDLM-TK contains the least number of parameters, followed by MobileNet-V3-Large. Combining the evaluation indexes such as the accuracy of each method, the PDLM-TK method achieves higher accuracy through less time, and is able to better balance the efficiency and accuracy of the model, compared to other models both in diagnostic accuracy and efficiency.

Table 5
www.frontiersin.org

Table 5. Statistical results of parameters and computational efficiency from various models.

5.2 Assessment of KD effectiveness in teacher networks

As presented in Table 6, the evaluation results of the Top1 accuracy for PDLM-TK when different teacher networks were utilized across three plant pest and disease datasets are shown. The comparison indicates that using ResNet18 as the teacher network yielded superior overall results, surpassing the ImageNet model by 0.63, 1.14, and 0.9 in the P, R, and F1 metrics, respectively. This improvement is attributed to the structural similarities between PDLM-TK’s backbone, which is derived from the residual blocks of ResNet. It can thus be concluded that knowledge migration is more effective when the chosen teacher network has a similar model structure, enhancing the effectiveness of the KD process.

Table 6
www.frontiersin.org

Table 6. Top1 accuracy evaluation results of different teacher networks.

5.3 Network grad-CAM visualization

To better assess the PDLM-TK model’s ability to learn the characteristics of plant pests and diseases, we predicted a portion of the test set data for each disease and visualized the results using Grad-CAM. The visualization outcomes are presented in Figure 11. In this study, the last layer of the PDLM-TK model was selected as the feature visualization layer, with the heat map superimposed on the original image. The Figure 11A displays the original plant pest images, while the Figure 11B shows the weighted visualization results.

Figure 11
www.frontiersin.org

Figure 11. Plant disease and pest images used Grad-CAM visualization results, figure (A) shows the plant leaves in the dataset and figure (B) shows the results visualized using Grad-CAM.

Upon examining the visualization results, we observed that the PDLM-TK model not only accurately predicted the classification of each disease but also successfully identified the key regions corresponding to various plant disease locations or pests. Additionally, it was noted that the model paid less attention to irrelevant and complex backgrounds surrounding the diseased leaves, instead concentrating on the characterization of disease features and pests during the feature selection process.

5.4 Ablation experiment

Based on the existing plant disease and pest classification model structure we designed the PDLM-TK model, which contains several modules to realize feature extraction, mining and model training respectively. Among them, residual blocks integrating multiple LRB-STs are designed to enhance the deep extraction of semantic features of different dimensions and to improve the model classification accuracy, which extracts the semantic information of different channels in plant disease images layer by layer. For the redundant information interference in the image, we proposed BNF-GC, which deeply mines the residual block features by graph convolution in order to focus on the disease features. We utilize the plant pest and disease dataset to train the teacher network and construct MTS-KD to realize knowledge transfer to the student network. To fully evaluate the main role played by each module, we designed ablation experiments based on the architecture of unused modules (Basic). Finally, they are evaluated and analyzed on Ai Challenger using P, R and F1 respectively.

As shown in Table 7 for the results of the ablation experiments of each module. “✓” stands for the blocks being selected, in the first four sets of experiments the Basic model (ResNet18) without several modules has the lowest evaluation scores, while the effect of using MTS-KD is higher than the other two types of modules. This proves that each module plays a role in improving the model classification effect, especially the effect of MTS-KD improves the model classification performance more. From the experimental results of groups 5 to 7, it can be seen that BNF-GC and MTS-KD have the best R when they are paired together, which is higher than the effect of group 8 when multiple modules are used at the same time, but it is slightly lower in terms of P and F1 values. The reason for this is that the BNF-GC module is able to better balance the feature information between different levels, but due to the similarity of some plant pests and diseases in terms of the types of performance, it is difficult to distinguish the real categories by only emphasizing the attention to the same representations, which leads to the easy occurrence of the misclassification phenomenon. With the addition of LRB-ST enables the multilayer channel semantic information to be better mined, which is used to enhance the model’s capture of plant pest and disease characteristics for each category. Therefore, a combination of multiple evaluation indexes leads to the best classification performance of the PDLM-TK method incorporating the three modules.

Table 7
www.frontiersin.org

Table 7. Evaluation results of ablation experiments.

6 Conclusion

To address the issues of low efficiency and accuracy in plant pest and disease diagnosis, the capabilities of tensor and graph deep learning in feature mining are fully utilized. The model network structure is optimized by integrating knowledge distillation (KD) and other techniques, resulting in the proposal of a Plant Pest and Disease Lightweight Identification Model (PDLM-TK) that fuses tensor features and KD. First, a lightweight residual block based on spatial tensor is constructed to enhance the perception and extraction of shallow detail features of plant images by introducing spatial tensor, and depth separable convolution is used to reduce the number of model parameters to improve the diagnostic efficiency. Secondly, a branching network incorporating graph convolutional features is proposed to realize image super-pixel segmentation by using spanning tree clustering based on pixel features, and graph convolutional neural network is used to extract correlation features to improve diagnostic accuracy. Finally, a model training strategy based on KD is designed to train the pest and disease diagnosis model by building a knowledge migration architecture, which fully balances the accuracy and diagnosis efficiency of the model.

In order to verify the diagnostic performance of the model in plant diseases and pests, we carried out experiments on three plant disease and pest datasets, including Plant Village, and the results proved that PDLM-TK had the best performance in terms of classification accuracy and efficiency, and was able to realize fast and accurate diagnosis of plant diseases. Although the method achieves better plant disease and pest classification results, the performance of the model in the face of some small-sample disease and pest datasets remains to be examined, and in the future, we can further increase the number of disease and pest species as well as datasets with different sample set sizes for testing, so as to further enhance the model’s value for practical application. In the subsequent study, we will further integrate real farming environments within the experimental field, collect data on a variety of plant diseases and pests, and assess the model’s performance in varied deployment settings and on smaller-scale farmland.

Data availability statement

The datasets presented in this study can be found in online repositories. The names of the repository/repositories and accession number(s) can be found in the article/supplementary material.

Ethics statement

The studies involving humans were approved by Tianjin University of Science and Technology. The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their written informed consent to participate in this study. The animal studies were approved by Tianjin University of Science and Technology. The studies were conducted in accordance with the local legislation and institutional requirements. Written informed consent was obtained from the owners for the participation of their animals in this study.

Author contributions

XZ: Writing – review & editing, Conceptualization, Methodology, Project administration. KL: Writing – review & editing, Conceptualization, Methodology. YZ: Writing – original draft, Conceptualization, Methodology.

Funding

The author(s) declare that financial support was received for the research, authorship, and/or publication of this article. This work was supported in part by the National Natural Science Foundation of China (No. 62377036).

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

AI Challenge (2018). Classification data set of diseases and insect pests. Available online at: https://aistudio.baidu.com/datasetdetail/76075. (Accessed Oct. 29, 2024).

Google Scholar

Aldakheel, E. A., Zakariah, M., Alabdalall, A. H. (2024). Detection and identification of plant leaf diseases using YOLOv4. Front. Plant Sci. 15, 1355941. doi: 10.3389/fpls.2024.1355941

PubMed Abstract | Crossref Full Text | Google Scholar

Ali, M. A., Sharma, A. K., Dhanaraj, R. K. (2024). Heterogeneous features and deep learning networks fusion-based pest detection, prevention and controlling system using IoT and pest sound analytics in a vast agriculture system. Comput. Electric. Eng. 116, 109146. doi: 10.1016/j.compeleceng.2024.109146

Crossref Full Text | Google Scholar

Bhandari, M., Shahi, T. B., Neupane, A., Walsh, K. B. (2023). BotanicX-AI: identification of tomato leaf diseases using an explanation-driven deep-learning model. Joournal Imaging 9, 53. doi: 10.3390/jimaging9020053

PubMed Abstract | Crossref Full Text | Google Scholar

Catalkaya, M., Goknur, A., Devran, Z. (2024). Rapid identification of Meloidogyne hapla by KASP assay. Crop Prot. 178, 106600. doi: 10.1016/j.cropro.2024.106600

Crossref Full Text | Google Scholar

Cetiner, H. (2022). Citrus disease detection and classification using based on convolution deep neural network. Microprocess. Microsys. 95, 104687. doi: 10.1016/j.micpro.2022.104687

Crossref Full Text | Google Scholar

Cheng, H.-H., Dai, Y.-L., Lin, Y., Hsu, H., Lin, C., Huang, J., et al. (2022). Identifying tomato leaf diseases under real field conditions using convolutional neural networks and a chatbot. Comput. Electron. Agric. 202, 107365. doi: 10.1016/j.compag.2022.107365

Crossref Full Text | Google Scholar

Chodey, M. D., Shariff, C. N. (2023). Pest detection via hybrid classification model with fuzzy C-means segmentation and proposed texture feature. Biomed. Signal Process. Control 84, 104710. doi: 10.1016/j.bspc.2023.104710

Crossref Full Text | Google Scholar

Dai, G., Tian, Z., Fan, J., Sunil, C., Dewi, C. (2024). DFN-PSAN: Multi-level deep information feature fusion extraction network for interpretable plant disease classification. Comput. Electron. Agric. 216, 108481. doi: 10.1016/j.compag.2023.108481

Crossref Full Text | Google Scholar

FAO (2022). FAO’s Plant Production and Protection Division (Rome: FAO). doi: 10.4060/cc2447en

Crossref Full Text | Google Scholar

FAO (2023). World Food and Agriculture – Statistical Yearbook 2023 (Rome: FAO). doi: 10.4060/cc8166en

Crossref Full Text | Google Scholar

Howard, A., Sandler, M., Chu, G., Chen, B., Tan, M., Wang, W., et al. (2020). “Searching for mobileNetV3,” in 2019 IEEE/CVF International Conference on Computer Vision. 1314–1324 (Seoul, Korea: IEEE). doi: 10.1109/ICCV.2019.00140

Crossref Full Text | Google Scholar

Hughes, D. P., Salathé, M. (2015). An open access repository of images on plant health to enable the development of mobile disease diagnostics. 1–13. Available at: https://arxiv.org/abs/1511.08060.

Google Scholar

Ishengoma, F. S., Rai, I. A., Ngoga, S. R. (2022). Hybrid convolution neural network model for a quicker detection of infested maize plants with fall armyworms using UAV-based images. Ecol. Inf. 67, 101502. doi: 10.1016/j.ecoinf.2021.101502

Crossref Full Text | Google Scholar

Johari, S. N. A. M., Khairunniza-Bejo, S., Shariff, A. R. M., Husin, N., Brasi, M., Kamarudin, N. (2022). Identification of bagworm (Metisa plana) instar stages using hyperspectral imaging and machine learning techniques. Comput. Electron. Agric. 194, 106739. doi: 10.1016/j.compag.2022.106739

Crossref Full Text | Google Scholar

Khan, A. I., Quadri, S. M. K., Banday, S., Shah, J. (2022). Deep diagnosis: A real-time apple leaf disease detection system based on deep learning. Comput. Electron. Agric. 198, 107093. doi: 10.1016/j.compag.2022.107093

Crossref Full Text | Google Scholar

Li, D., Zhang, C., Li, J., Li, M., Huang, M., Tang, Y. (2024). MCCM: multi-scale feature extraction network for disease classification and recognition of chili leaves. Front. Plant Sci. 15, 1367738. doi: 10.3389/fpls.2024.1367738

PubMed Abstract | Crossref Full Text | Google Scholar

Li, R., Shang, Z., Zheng, C., Li, H., Liang, Q., Cui, Y. (2023). Efficient distributional reinforcement learning with kullback-leibler divergence regularization. Appl. Intel. 53, 24847–24863. doi: 10.1007/s10489-023-04867-z

Crossref Full Text | Google Scholar

Li, Y., Ma, L., Sun, N. (2024). A bilinear transformer interactive neural networks-based approach to fine-grained recognition and protection of plant diseases for gardening design. Crop Prot. 180, 106660. doi: 10.1016/j.cropro.2024.106660

Crossref Full Text | Google Scholar

Liu, Y., Li, X., Fan, Y., Liu, L., Shao, L. (2024). Classification of peanut pod rot based on improved YOLOv5s. Front. Plant Sci. 15, 1364185. doi: 10.3389/fpls.2024.1364185

PubMed Abstract | Crossref Full Text | Google Scholar

Ma, N., Zhang, X., Zheng, H.-T., Sun, J. (2018). “ShuffleNet V2: Practical guidelines for efficient cnn architecture design,” in Proceedings of the European Conference on Computer Vision (ECCV), New York. 122–138 (ACM).

Google Scholar

Manavalan, R. (2022). Towards an intelligent approaches for cotton diseases detection: A review. Comput. Electron. Agric. 200, 107255. doi: 10.1016/j.compag.2022.107255

Crossref Full Text | Google Scholar

Math, R. M., Dharwadkar, N. (2022). Early detection and identification of grape diseases using convolutional neural networks. J. Plant Dis. Prot. 129, 521–532. doi: 10.1007/s41348-022-00589-5

Crossref Full Text | Google Scholar

Mehta, S., Rastegari, M. (2022). MobileViT: Light-weight, general-purpose, and mobile-friendly vision transformer. 1–26. Available at: https://arxiv.org/abs/2110.02178.

Google Scholar

Motie, J. B., Saeidirad, M. H., Jafarian, M. (2023). Identification of Sunn-pest affected (Eurygaster Integriceps put.) wheat plants and their distribution in wheat fields using aerial imaging. Ecol. Inf. 76, 102146. doi: 10.1016/j.ecoinf.2023.102146

Crossref Full Text | Google Scholar

Nandhini, C., Brindha, M. (2024). Visual regenerative fusion network for pest recognition. Neural Comput. Appl. 36, 2867–2882. doi: 10.1007/s00521-023-09173-w

Crossref Full Text | Google Scholar

Pansy, D. L., Murali, M. (2023). UAV hyperspectral remote sensor images for mango plant disease and pest identification using MD-FCM and XCS-RBFNN. Environ. Monit. Assess. 195, 1120. doi: 10.1007/s10661-023-11678-9

PubMed Abstract | Crossref Full Text | Google Scholar

Qiang, J., Liu, W., Li, X., Guan, P., Du, Y., Liu, B., et al. (2023). Detection of citrus pests in double backbone network based on single shot multibox detector. Comput. Electron. Agric. 212, 108158. doi: 10.1016/j.compag.2023.108158

Crossref Full Text | Google Scholar

Rimal, K., Shah, K. B., Jha, A. K. (2023). Advanced multi-class deep learning convolution neural network approach for insect pest classification using TensorFlow. Int. J. Environ. Sci. Technol. 20, 4003–4016. doi: 10.1007/s13762-022-04277-7

Crossref Full Text | Google Scholar

Roxa, Z., Phoofolo, M. W., Dawuda, P. M., Molapo, S., Majoro, L. (2023). Identification and prevalence of ticks on Merino sheep in Lesotho. Trop. Anim. Health Prod. 55, 70. doi: 10.1007/s11250-023-03492-9

PubMed Abstract | Crossref Full Text | Google Scholar

Rustia, D. J. A., Chiu, L.-Y., Lu, C.-Y., Wu, Y., Chen, S. (2022). Towards intelligent and integrated pest management through an AIoT-based monitoring system. Pest Manage. Sci. 78, 4288–4302. doi: 10.1002/ps.v78.10

PubMed Abstract | Crossref Full Text | Google Scholar

Sajitha, P., Andrushia, A. D., Anand, N. (2024). A review on machine learning and deep learning image-based plant disease classification for industrial farming systems. J. Ind. Inf. Integr. 38, 100572. doi: 10.1016/j.jii.2024.100572

Crossref Full Text | Google Scholar

Shafik, W., Tufail, A., Liyanage, C. D. S., Apong, R. (2023). Using a novel convolutional neural network for plant pests detection and disease classification. J. Sci. Food Agric. 103, 5849–5861. doi: 10.1002/jsfa.v103.12

PubMed Abstract | Crossref Full Text | Google Scholar

Shamsul Kamar, N. A., Abd Rahim, S. K., Ambrose, A. A., Awing, N., Samdin, Z. (2023). Pest and disease incidence of coniferous species in Taman Saujana Hijau, Putrajaya urban park, Malaysia. J. Forest. Res. 34, 2065–2077. doi: 10.1007/s11676-023-01644-z

Crossref Full Text | Google Scholar

Sharma, R. P., Ramesh, D., Pal, P., Tripathi, S., Kumar, C. (2022). IoT-enabled IEEE 802.15.4 WSN monitoring infrastructure-driven fuzzy-logic-based crop pest prediction. IEEE Internet Things J. 9, 3037–3045. doi: 10.1109/JIOT.2021.3094198

Crossref Full Text | Google Scholar

Subbaian, S., Balasubramanian, A., Marimuthu, M., Chandrasekaran, S., Muthusaravanan, G. (2024). Detection of coconut leaf diseases using enhanced deep learning techniques. J. Intel. Fuzzy Syst. 46, 5033–5045. doi: 10.3233/JIFS-233831

Crossref Full Text | Google Scholar

Szegedy, C., Vanhoucke, V., Ioffe, S., Shlens, J., Wojna, Z. (2015). Rethinking the inception architecture for computer vision. arXiv: 1512.00567. Available at: http://arxiv.org/abs/1512.00567.

Google Scholar

Tan, M., Le, Q. V. (2019). “EfficientNet: rethinking model scaling for convolutional neural networks,” in International Conference on Machine Learning, Long Beach, CA, USA. 6105–6114, LR.

Google Scholar

Thakuria, D., Chaliha, C., Dutta, P., Sinha, S., Uzir, P. (2023). Citrus Huanglongbing (HLB): Diagnostic and management options. Physiol. Mol. Plant Pathol. 125, 102016. doi: 10.1016/j.pmpp.2023.102016

Crossref Full Text | Google Scholar

Wei, D., Chen, J., Luo, T., Wang, Z. (2022). Classification of crop pests based on multi-scale feature fusion. Comput. Electron. Agric. 194, 106736. doi: 10.1016/j.compag.2022.106736

Crossref Full Text | Google Scholar

Wu, X., Zhan, C., Lai, Y.-K., Cheng, M., Yang, J. (2019). “IP102: A large-scale benchmark dataset for insect pest recognition,” in 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). 8779–8788.

Google Scholar

Xia, W., Han, D., Li, D., Wu, Z., Han, B., Wang, J., et al. (2023). An ensemble learning integration of multiple CNN with improved vision transformer models for pest classification. Ann. Appl. Biol. 182, 144–158. doi: 10.1111/aab.v182.2

Crossref Full Text | Google Scholar

Xiao, Z., Yin, K., Geng, L., Wu, J., Zhang, F., Liu, Y., et al. (2022). Pest identification via hyperspectral image and deep learning. Signal Image Video Process. 16, 873–880. doi: 10.1007/s11760-021-02029-7

Crossref Full Text | Google Scholar

Yu, Y., Sun, T., Yan, J. (2022). Improved YOLOv5 network for agricultural pest detection. Fourteenth Int. Conf. Graphics Image Process. (ICGIP 2022). 12705, 1270507. doi: 10.1117/12.2680747

Crossref Full Text | Google Scholar

Zhang, Y., Li, M., Zhou, B., et al (2023). Automatic cotton spider mites’ damage grading algorithm for depth residual network based on transfer learning. Comput. Electron. Agric. 212, 108070. doi: 10.1016/j.compag.2023.108070

Crossref Full Text | Google Scholar

Zhao, X., Zhang, J., Huang, Y., Tian, Y., Yuan, L. (2022). Detection and discrimination of disease and insect stress of tea plants using hyperspectral imaging combined with wavelet analysis. Comput. Electron. Agric. 193, 106717. doi: 10.1016/j.compag.2022.106717

Crossref Full Text | Google Scholar

Zheng, Z., Zhang, C. (2022). Electronic noses based on metal oxide semiconductor sensors for detecting crop diseases and insect pests. Comput. Electron. Agric. 197, 106988. doi: 10.1016/j.compag.2022.106988

Crossref Full Text | Google Scholar

Zhu, L., Li, Z., Hu, S. (2023). Efficient data-centric pest images identification method based on Mahalanobis entropy for intelligent agriculture. J. Electron. Imaging 32, 1–13. doi: 10.1117/1.JEI.32.5.052406

Crossref Full Text | Google Scholar

Keywords: image classification, spatial tensor, knowledge distillation, light weighting, graph convolutional neural networks

Citation: Zhang X, Liang K and Zhang Y (2024) Plant pest and disease lightweight identification model by fusing tensor features and knowledge distillation. Front. Plant Sci. 15:1443815. doi: 10.3389/fpls.2024.1443815

Received: 04 June 2024; Accepted: 17 October 2024;
Published: 21 November 2024.

Edited by:

Chellappan Padmanabhan, USDA APHIS PPQ Science and Technology, United States

Reviewed by:

Dongmei Chen, Hangzhou Dianzi University, China
Md Ashraful Haque, Indian Council of Agricultural Research (ICAR), India
Shivaranjani Baruah, Cornell University, in collaboration with reviewer KG

Copyright © 2024 Zhang, Liang and Zhang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Kun Liang, liangkun@tust.edu.cn

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.