Skip to main content

ORIGINAL RESEARCH article

Front. Plant Sci., 27 August 2024
Sec. Technical Advances in Plant Science
This article is part of the Research Topic UAVs for Crop Protection: Remote Sensing, Prescription Mapping and Precision Spraying View all 4 articles

Residual swin transformer for classifying the types of cotton pests in complex background

Ting ZhangTing Zhang1Jikui ZhuJikui Zhu1Fengkui ZhangFengkui Zhang1Shijie ZhaoShijie Zhao1Wei LiuWei Liu2Ruohong HeRuohong He1Hongqiang DongHongqiang Dong3Qingqing HongQingqing Hong2Changwei TanChangwei Tan4Ping Li*Ping Li1*
  • 1College of Mechanical and Electrical Engineering/Modern Agricultural Engineering Key Laboratory at Universities of Education Department of Xinjiang Uygur Autonomous Region/Key Laboratory of Tarim Oasis Agriculture (Tarim University) Ministry of Education, Tarim University, Alar, China
  • 2Jiangsu Co-Innovation Center for Modern Production Technology of Grain Crops/Joint International Research Laboratory of Agriculture and Agri-Product Safety of the Ministry of Education of China/Jiangsu Province Engineering Research Center of Knowledge Management and Intelligent Service, College of Information Engineer, Yangzhou University, Yangzhou, China
  • 3Corps in Southern Xinjiang, College of Agronomy, Tarim University, Alar, China
  • 4Jiangsu Key Laboratory of Crop Genetics and Physiology/Jiangsu Key Laboratory of Crop Cultivation and Physiology, Agricultural College of Yangzhou University, Yangzhou, China

Background: Cotton pests have a major impact on cotton quality and yield during cotton production and cultivation. With the rapid development of agricultural intelligence, the accurate classification of cotton pests is a key factor in realizing the precise application of medicines by utilize unmanned aerial vehicles (UAVs), large application devices and other equipment.

Methods: In this study, a cotton insect pest classification model based on improved Swin Transformer is proposed. The model introduces the residual module, skip connection, into Swin Transformer to improve the problem that pest features are easily confused in complex backgrounds leading to poor classification accuracy, and to enhance the recognition of cotton pests. In this study, 2705 leaf images of cotton insect pests (including three insect pests, cotton aphids, cotton mirids and cotton leaf mites) were collected in the field, and after image preprocessing and data augmentation operations, model training was performed.

Results: The test results proved that the accuracy of the improved model compared to the original model increased from 94.6% to 97.4%, and the prediction time for a single image was 0.00434s. The improved Swin Transformer model was compared with seven kinds of classification models (VGG11, VGG11-bn, Resnet18, MobilenetV2, VIT, Swin Transformer small, and Swin Transformer base), and the model accuracy was increased respectively by 0.5%, 4.7%, 2.2%, 2.5%, 6.3%, 7.9%, 8.0%.

Discussion: Therefore, this study demonstrates that the improved Swin Transformer model significantly improves the accuracy and efficiency of cotton pest detection compared with other classification models, and can be deployed on edge devices such as utilize unmanned aerial vehicles (UAVs), thus providing an important technological support and theoretical basis for cotton pest control and precision drug application.

1 Introduction

Cotton is one of the most important cash crops in the world and occupies an important position in the economic development of China and the world. Cotton may be subjected to various viruses, bacteria, fungi, and insects during its growth, which can have a significant impact on its quality and yield (Chen et al., 2020; Kiruthika et al., 2022). Cotton aphid infections in the seedling stage of cotton can lead to short plants and shriveled leaves; infestation by cotton mirids bugs at the bud stage can lead to bud shedding and sparse boll setting; cotton leaf mite infestation at the boll stage causes leaf, bud, flower, and young boll abscission; causing irreparable damage to growth and yield (Islam et al., 2023; Thivya Lakshmi et al., 2024). Insect-damaged plants usually show visible signs or lesions on leaves, stems, flowers, or fruits, and each type of damaged leaf exhibits unique characteristics (Li et al., 2021; Zhu et al., 2023). Traditional identification methods are still mainly based on manual visual hand checking, the method relies on personal experience, has a strong subjectivity, and low work efficiency (Skelsey et al.,2013; Zhang et al., 2022). Therefore, the accurate identification of cotton insect pests is a key component in the prevention and treatment of cotton pests, as well as reducing the use of pesticides and promoting the development of green agriculture.

In recent years, with the rapid development of agriculture 4.0, the use of advanced information systems and Internet technology can realize a large amount of agricultural data collection, analysis, processing, and realization of providing assistance to the agricultural decision support system (Zhai et al., 2020). In early studies, (Arnal Barbedo, 2013; Manavalan, 2022; Kong et al., 2022) used traditional machine learning and image processing methods to extract pest and disease features for classification and identification. (Haoyu et al., 2010) used the support vector machine(SVM) algorithm with four kinds of kernel functions to process the spectral data and establish the diagnostic model of cucumber pests and diseases to realize the rapid and accurate diagnosis of cucumber pests and diseases, with the highest recognition rate of 98.3%. (Chaudhary et al., 2016). proposed an improved Random Forest Classifier (RFC) method, which combines the forest machine learning algorithm, attribute evaluator method, and instance filtering method to achieve the identification of peanut diseases with a classification accuracy of 97.8%. (Irfan et al., 2019) used Support Vector Machine (SVM) and C4.5 algorithm for disease and pest identification of leaves, stems, fruits of chili peppers and the study showed that the accuracy of the C4.5 based method is higher. However, such an operation requires certain basic knowledge and specialized skills in image processing, which makes it difficult to promote its application in field production.

In recent years, computer vision techniques have been widely used in agriculture, including machine learning (ML), convolutional neural networks (CNN), deep learning (DL) and transfer learning (TL). Deep learning techniques are favored by many scholars in detecting plant pests and diseases (Rezaei et al., 2024; Luo et al., 2023; Reddy et al., 2023). (Ramcharan et al., 2017) utilized cassava disease image dataset and applied transfer learning to train deep convolutional neural networks to achieve the recognition of three diseases and two insect pests in cassava, and the overall accuracy of the best model was tested to be 93%. However, image data with a single background encountered problems in practical applications in agriculture, so (Yanfen Li et al., 2020) proposed a fine-tuned GoogLeNet model for the complex background presented by farmland scenes to realize the recognition of 10 common crop pests. Since the occurrence of pests and diseases is affected by regional, climatic and other factors, there will be uneven data on pests and diseases, (Abbas et al., 2021) and other scholars addressed this problem by using Conditional Generative Inverse Networks (C-GAN) to generate synthetic images of tomato plant leaves, and realized the classification of 10 types of diseases in tomato through DenseNet121 model, with an accuracy rate of 97.11%. Deep learning-based detection methods have greater advantages in terms of detection efficiency, accuracy and application scenarios (Reder et al., 2021; Syed-Ab-Rahman et al., 2022). Deep learning-based models can provide technical support for edge devices, such as unmanned aerial vehicles (UAVs), to enable wide-area pest detection.

The above research has led to a series of successes in identifying plant pests and diseases. However, most of the models based on deep learning are dominated by convolutional neural networks, which are subject to the limitations of the convolutional network itself, such as the field of view of the convolutional neural network is limited by the size of the convolutional kernel, which can lead to the loss of most of the global information in the image (Guo et al., 2022). Recently, (Liang et al., 2021; Liu et al., 2021) applied the Transformer model in the field of natural language processing to the field of computer vision and achieved good results, and proposed a new vision converter, Swin Transformer, to provide a new direction for research in the field of vision. In this study, from the practical application of agriculture, we propose a cotton insect pest classification model suitable for the field environment by combining the image characteristics of cotton insect pests with the problem of small and unbalanced image data of cotton insect pests. In this study, three kinds of cotton insect pests (cotton aphids, cotton mirids, cotton leaf mites) are taken as the research objects, and the leaf images of cotton insect pests under the complex background of the field are collected as the dataset, and Swin Transformer model is chosen as the backbone, and the residual module and skip connection are introduced to construct the classification model of cotton insect pests, and to realize the recognition of cotton insect pests in the field environment.

The main contributions of this paper are summarized as follows:

1. An enhanced version of the standard Swin Transformer architecture, the residual Swin Transformer, is proposed, which is capable of accurately categorizing cotton pest species.

2. In Swin Transformer, 2 residual modules and skip connection are added to improve the recognition accuracy of the model while retaining the excellent design of Swin Transformer. Meanwhile, we verify the effectiveness of the residual Swin Transformer through many experiments.

The subsequent sections are structured as follows: section 2 describes the materials and methods used in this study, section 3 organizes and summarizes the experimental results, section 4 analyzes the results, and section 5 provides conclusions.

2 Materials and methods

2.1 Materials

2.1.1 Image data acquisition and dataset production

This study collected and constructed a dataset on cotton insect pests for three insect pests, including cotton aphids, cotton leaf mites, and cotton mirids. The dataset was collected in May-August 2023 during the cotton planting period at the cotton experimental field of Tarim University in Xinjiang, China, as shown in Figure 1A. The image capturing work was carried out using a mobile device paired with 108MP primary camera + 8MP wide angle camera and the resolution of the acquired image was 2928×2928 pixels. A total of 5581 images were collected under natural environments, 30 cm away from the leaf surface and perpendicular to the leaf surface. Under the guidance of cotton plant protection experts, 2705 usable images were obtained by removing the leaf images of drug-infested wounds, early senescence and other factors, including 1112 images of cotton leaves damaged by cotton aphids, 703 images of cotton leaves damaged by cotton leaf mites, and 890 images of cotton leaves damaged by cotton mirids. Cotton leaves damaged by cotton aphid will appear curled and white or black aphids will appear on the surface of the leaves, on the back and on the stems (Steckel et al., 2021). Cotton leaf mites use stinging mouthparts to suck sap on the back of cotton leaves. When the number of cotton leaf mites on the back is low, yellow and white spots appear on the surface of the leaves; as the number of cotton leaf mites increases, the spots on the surface of the leaves become reddish, and the area of the spots becomes larger and larger (Ferraz et al., 2017). Cotton mirids through the assassination type mouthparts into the cotton plant to suck sap, fresh leaves were damaged at the beginning of a small black spot, with the growth of the leaf stretching after a large number of broken, forming a “broken leaf madness” (Mithal Rind et al., 2021). The samples for each category are shown in Figure 1B.

Figure 1
www.frontiersin.org

Figure 1. (A) Collecting location information; (B) Images of cotton pests.

The organized dataset is divided into training set, validation set, and test set in the ratio of 7:2:1 as shown in Table 1. The model learns from the images of the training set, extracting features of the data from them, and improves the learning of the model through parameter updating and optimization. By interacting with the validation set, the best model and hyperparameter configurations can be selected to improve the generalization of the model. Finally, the performance of the model is objectively assessed by evaluating its performance on a test set and will be used as an indicator of the final performance of the model.

Table 1
www.frontiersin.org

Table 1. Partitioning of cotton image dataset.

2.1.2 Image resizing

During this experiment, the original images collected were too large to be used directly for model training. Therefore, the image is resized using Bicubic interpolation algorithm (Keys, 1981). The bicubic interpolation determines the value of the target pixel by calculating the weighted average of the 16 nearest neighbor pixels around the target pixel, which achieves image smoothing and at the same time is able to better preserve the details of the image. The interpolation formula is shown as follows:

f(i+u,j+v)=a=12b=12f(i+a,j+b)W(au)W(bv)    

where f is the pixel value of the pixel point and u, v are the distance between the pixel points. where W(x) is the interpolating kernel function with the expression as follows:

W(X)={(a+2)|x|3(a+3)|x|3+1   0<|x|1a|x|35a|x|2+8a|x|4a   1<|x|<20     2|x|

where a=0.5 in Equation.

The size of the processed image by bicubic interpolation is 224×224×3 and the processed image is shown in Figure 2.

Figure 2
www.frontiersin.org

Figure 2. Data preprocessing results.

2.1.3 Image augmentation

In deep learning, training requires large-scale data samples in order to help the model learn a wider range of features and patterns. The collection of cotton pest images is limited by the natural environment and weather conditions, resulting in a limited number of images. In this study, the number of training samples is increased by data augmentation. In order to simulate the growth of leaves in random directions and weather conditions such as sun exposure and cloudy days in a field environment, the data augmentation work was carried out in this study by using the technique of rotating 90°, 180°, mirroring, dimming brightness by 0.9, and brightening brightness by 1.5 images. Considering that the cotton plantation is located in the Tarim Basin, where severe weather with sandstorms often occurs, Gaussian filter image technique is added for image processing so as to simulate the interference of wind and sand. Figure 2 shows the cotton image after data augmentation. After the data augmentation operation, a total of 16,230 images of cotton pests were obtained, and the partitioning of the augmented dataset is shown in Table 2.

Table 2
www.frontiersin.org

Table 2. Partitioning of augmented cotton image dataset.

2.2 Methods

2.2.1 Swin transformer network

Originally proposed by Google in 2017, the Transformer model is mainly used for natural language processing tasks such as machine translation and language modeling (Vaswani et al., 2023). With the successful application of the Transformer model in processing natural language tasks, some scholars have begun to try to apply the model in visual tasks. In 2020, Vision Transformer (VIT) was proposed as a model mainly for image classification tasks (Dosovitskiy et al., 2021). However, VIT suffers from information loss and increased computation, a new Transformer model has been proposed. Swin Transformer improves on VIT by enabling computation in non-overlapping windows through a shifting window scheme, while also allowing cross-window connections for better scalability and efficiency (Liu et al., 2021).

The Swin Transformer model was chosen as the base model for this study. As shown in Figure 3A, Swin Transformer consists of the following parts: the Patch Partition, the Swin Transformer Backbone Block, the Classification Head. In this study, the Patch Partition section splits the cotton leaf image into a series of equal-sized image patches. The position of the split cotton image patches is encoded by Linear Embedding, and the position information is embedded into the feature vectors so that the model can perceive the relative positions of the image patches. The Swin Transformer backbone block consists of multiple Transformer encoders containing both Window-MSA and Shifted Window-MSA. Classification Head Part realizes the classification prediction by Multilayer Perceptron network (MLP), the structure is shown in Figure 3B. Aiming at the problem that the leaf features of cotton insect pests are small and easy to be confused in the complex background, this study introduces the residual module and skip connection in the Swin Transformer network, and the improved whole model structure is shown in Figure 3C.

Figure 3
www.frontiersin.org

Figure 3. (A) Swin Transformer Structure; (B) Multilayer Perceptron Structure; (C) Improved Swin Transformer structure.

2.2.2 Residual module

In deep learning, more abstract properties can be extracted as the number of network layers increases. However simply increasing the number of network layers leads to the problem of vanishing and exploding gradients (Zhang et al., 2018; Alaeddine and Jihene, 2023). Because the leaf features of cotton insect pests are relatively small, few in number, and easily confused with the background, the pest features are easily lost in the deep network. In order to solve these problems, this study builds the residual module, which is combined with Swin Transformer network.

In this study, the residual module is built to enhance the network representation by combining the input features with the output features to preserve the input features, as well as to integrate the classification features of the image. The residual structure is shown in Figure 4, including convolutional layer, Batch Normalization, Rectified Linear Unit. Suppose x is the input, σ is the Rectified Linear Unit (ReLU), BN is the Batch Normalization, and y is the output.

Figure 4
www.frontiersin.org

Figure 4. Structure of residual module.

F(x)={BN[Conv2d(x)]}
y=x+Conv2d{F[F(X)]}

In this study, the residual module is introduced at two locations respectively, as shown in Figure 3C. The first place is located after the image input and before the Patch Partition, where the residual module both preserves the input image features and incorporates classification features to improve the model’s detection of fine leaf features. The second is located between the Swin Transformer blocks and is used in conjunction with the skip connection in 3.3.

2.2.3 Skip connection

Skip connections are commonly used in deep networks to improve information transfer. In deep networks, information is passed from layer to layer, layer by layer. In deep networks, a single sequential delivery method can easily cause gradient vanishing or gradient explosion. Skip connections can be realized to cross over certain layers for information transfer, fusing the output of the previous layer with the input of one of the subsequent layers, increasing the flow of information (Zhang et al., 2023; Ding et al., 2024). Considering the problem that the characteristics of insect pest leaves are small and easy to be lost in the transfer process, this experiment introduces skip connections in stage 1 - stage 4 of the backbone network.

The skip connection can realize the fusion of the underlying features and the higher-level features to retain more high-resolution details and provide more information for the later image classification, thus improving the correct rate of classification. Experimentally, it was found that after adding skip connections, there may be a semantic gap between the two combined feature sets, which is problematic to use directly after fusion. Therefore, in this experiment, the fused data were processed again after fusion. As shown in Figure 3C, skip connections are introduced between Swin Transformer blocks, and the data are processed again after fusion by the residual module proposed in 3.2. Experiments have demonstrated that reprocessing the fused data through the residual module after skip connections can improve the whole performance of the network and increase the accuracy of the model. The structure of the original Swin Transformer model is shown in Figure 3A and the structure of the improved Swin Transformer model is shown in Figure 3C.

2.2.4 Self-attention mechanism

Self-attention is a key part of the Transformer model in visual tasks. As shown in Figure 5A, the input image is segmented into multiple patches, and after recombining them into a set of series, the correlation calculation between each position in the sequence and the other positions is realized through the self-attention. By mapping the input sequence into query (WQ), key (Wk), and value (Wv) vectors, the similarity, weighted summation, and the correlation of different positions are computed (Vaswani et al., 2023).

Figure 5
www.frontiersin.org

Figure 5. Illustration of Attention Mechanisms.

The traditional self-attention computes the attention score for the entire image, which suffers from high computational effort. The window self-attention used in this paper was proposed by scholars in 2021, which makes upgrades based on the self-attention and reduces the amount of computation (Liu et al., 2021). As shown in Figure 5B, after the image is split into multiple patches, the neighboring patches are integrated into a single window, and the attention score between each window is calculated, thus reducing the amount of computation. The computational complexity of the self-attention (MSA) and the window self-attention (W-MSA) is shown as follows:

ΩMSA=4HWC2+2(HW)2C
ΩWMSA=4HWC2+2M2HWC

In order to improve the problem of missing information interactions that can occur between different windows, the Shifted Window self-attention (SW-MSA) is introduced to realize cross-window connection reorganization (Liu et al., 2021). As shown in Figure 5C, different from the window self-attention, the shift window self-attention splits the image into 9 pieces of different sizes, which are recombined, and the result of the combination is shown in Figure 5D. The combined windows are computed to obtain localized attention.

3 Results

3.1 Installations and evaluation

Deep learning models require a large number of datasets for training. Therefore, we expanded the cropped 2705 cotton pest images into 16,230 images by data augmentation and randomly divided them into training, validation, and testing datasets in the ratio of 7:2:1. We used adaptive motion estimation (Adam) to automatically optimize the learning rate during deep learning model training with a learning rate of 0.001 and a number of training epochs of 100. To ensure minimal validation loss, we saved the trained model in each epoch. In this study, when the model has finished training, it is validated on the same test dataset using performance metrics such as Precision, Recall, etc. to reveal the detection accuracy of different models. The computer parameters used for model training and validation and the environment configuration for improving the model are shown in Table 3. In the classification task, four types appear in the prediction results. True Positive (TP) represents a predicted positive sample that is actually positive and correctly determines the type of disease. False positives (FP) represent samples that were predicted to be positive and were actually negative. True Negative (TN) predicts a negative sample and is actually a negative sample. False negative (FN) represents a negative prediction and a positive actual. The following common classification metrics are commonly used to evaluate the model.

Table 3
www.frontiersin.org

Table 3. Computer parameters and improved model environmental resource allocation.

Precision rate P: the ratio of the number of samples predicted to be positive and correct to the total number of samples predicted to be positive. The formula is shown as follows:

P=TPTP+FP

Recall R: The ratio of the number of samples whose predictions are positive and correct to the total number of samples that are actually positive. The formula is shown as follows:

R=TPTP+FN

Specificity S: The ratio of correctly predicted as a negative sample to the actual negative sample. It is an indicator that evaluates the ability to judge negative samples. The specificity was calculated as follows:

S=TNTN+FN

F1-measure: weighted summed mean of precision rate P and recall rate R. The accuracy and coverage of the model can be evaluated in a comprehensive manner. Higher F1 scores indicate that the model strikes a better balance between precision and recall. The formula is shown as follows:

F1=2×P×RP+R

Accuracy (Acc): The ratio of the number of correctly categorized samples to the total sample number. Accuracy is one of the commonly used evaluation measures. The formula is shown as follows:

Acc=TP+TNTP+FP+TN+FN

3.2 Experimental results and analysis

3.2.1 Comparison of data augmentation results

Model training and validation were performed on the original and augmented datasets. The experimental results show that using the original model for training and validation on the original and augmented datasets, the accuracy was improved from 0.946 to 0.951. Using the improved model for training and validation respectively, the accuracy was improved from 0.961 to 0.974, which is an improvement of 1.3%. The experimental data proved the effectiveness of the data augmentation operation. Therefore, the rest of the experiments were conducted on the augmented dataset.

3.2.2 Performance comparison of improved modules

The Swin Transformer model is applied to detect cotton insect pests in complex environments in the field, which may be affected by complex environments. Therefore, this paper introduces a residual module and skip connection to improve the detection accuracy of the model. In order to verify the effectiveness of the improvement scheme proposed in this paper, we take the Swin Transformer model as the original model and add different improvement modules respectively, so as to verify the performance of the improved model. One of the residual module1 is mentioned in 3.2 and is located after the picture input. The skip connection and residual module 2 are the ones mentioned in 3.3 of the text and are located between the Swin Transformer blocks and after the skip connection, respectively. The experimental results are shown in Table 4, where the accuracy is improved from 0.951 to 0.959 by adding the residual module 1 to the original model. The accuracy is improved by 1.3% and 1.2% by adding skip connection and residual module 2, respectively. Finally, the improved model improves the accuracy from 0.951 to 0.974 compared to the original model, an improvement of 2.3%. Comparison results can prove that the residual module and skip connection introduced in this paper contribute to the model performance.

Table 4
www.frontiersin.org

Table 4. Impact of improved modules on model performance.

3.2.3 Analysis of training results

The relationship between the training results and the number of epochs before and after the model improvement is shown in Figure 6, which shows that the original model has better generalization ability. But after training up to 25 epochs, you can see that the accuracy and loss show a straight line and no longer change. It is possible that the gradient disappears or the gradient explodes. The improved model continues the generalization ability of the original model, solves the problems such as gradient vanishing, and increases the classification accuracy from 0.951 to 0.974, and reduces the loss from 0.2019 to 0.1089.

Figure 6
www.frontiersin.org

Figure 6. Accuracy- epoch and Loss-epoch curve of model before and after improvement.

3.2.4 Analysis of test results

The model weights with the highest rate of accuracy during training are saved and the model is tested on the TEST dataset (1620 sheets). In terms of the accuracy of the model, the accuracy rate reaches 97.4%. Meanwhile, in the process of testing the model, the prediction time for the model was also evaluated, and the average time for single prediction was 0.00434s, and the improved model can judge 230 RGB images in 1s.

The precision, recall, specificity, and F1  score is calculated and analyzed for each pest by testing. The performance of different pests before and after model improvement is shown in Table 5. It can be seen that the improved model has improved and more balanced in most indicators, among which the performance of recognizing cotton aphid damage pictures and cotton mirids damage pictures is not much different, and the accuracy rate is 0.978 and 0.983 respectively. The picture will be slightly inferior in identifying cotton leaf mite damage, but it also reaches 0.957. Combined with the collection of cotton pests, we analyzed that due to the occurrence of cotton leaf mites at the late stage of the cotton planting period, which is in the August-October period in the Xinjiang region, the strong light irradiation and overexposure during the filming, and the lack of characteristics of leaf infestation, led to the difficulties encountered by the model in the detection of infestation characteristics.

Table 5
www.frontiersin.org

Table 5. Test results of different insect pests.

3.2.5 Comparison with classical classification networks

Finally, classical classification models were used to train on the same dataset separately. The training results of the classical model and the results of the improved model are shown in Table 6. Among them, the classical classification models VGG11, VGG11-BN, Resnet18, and mobilenetv2 achieve high accuracy, which is as high as 0.969, 0.927, 0.952, and 0.949, respectively. The accuracy of VIT and Swin Transformer small, Swin Transformer base in the Transformer series reached 0.911, 0.895 and 0.894 respectively. The model size of the improved model is not the smallest, but it is smaller than the Transformer series and the VGG network, although in the model size increased, the improved model performs the best with an accuracy of 0.974.

Table 6
www.frontiersin.org

Table 6. Classification results of different algorithms for cotton insect pests.

As shown in Figure 7, the predictions of the model are visualized with the real labels through the confusion matrix. Figures 8 shows the prediction results of the improved model and other classical models on the test dataset, which shows that the number of images classified correctly by the improved Swin Transformer model is relatively balanced, indicating that there is no obvious classification preference of the improved model. Among the results of misjudgments, the number of images of cotton infested with cotton leaf mites that were judged to be infested with cotton aphids was higher than the rest of the misjudgments. Comparison of the leaf images of the two infestations revealed that the appearance of brown spots on the leaves of cotton infested with the cotton leaf mite was more similar to the characteristics of the black cotton aphid on the leaves, resulting in a misjudgment.

Figure 7
www.frontiersin.org

Figure 7. Confusion matrix.

Figure 8
www.frontiersin.org

Figure 8. Visual results.

Compared to images with a single background, images with complex backgrounds increase the difficulty of the model in acquiring leaf features, resulting in some decrease in the accuracy rate. Leaf images with complex backgrounds can contain information such as soil, weeds, and other information, as well as staggered and obscured leaves. The color of the soil in model training is similar to the color of the spots, which can cause learning difficulties; interlocking and partially shaded leaves can cause the model to miss information about the infestation as it learns, leading to less-than optimal learning results.

Figure 8 shows the recognition visualization results of the improved model and each model in the comparison experiment of 3.2.5. The test results of the three pest images were compared by Grad-CAM visualization (Selvaraju et al., 2016; Chattopadhyay et al., 2018), from left to right, the original image and the visualization results of each model, where Swin transformer is the original model. It can be observed that the improved model can capture more detailed information. For example, as shown in the cotton aphid test results, the improved model focuses on the curly lines at the edge of the leaves. According to the visualization results, the improved model has better detection effect, more accurate localization, and the model has strong anti-interference ability in the complex environment.

4 Discussion

In this study, the residual module and skip connection were used to improve the Swin Transformer model, and the improved model could be used for the identification of cotton insect pests in a field environment. Comparing several convolutional neural networks before choosing the benchmark model, it was found that the Swin Transformer model is not limited by the convolutional kernel and can obtain more global information. The Swin Transformer model collects image information through a self-attention mechanism. Effective operation under irregular inputs is achieved by dynamically computing the weight matrix of the self-attention mechanism (Zhu et al., 2022; Zhou et al., 2023). Therefore, Swin Transformer model is chosen as the backbone network in this study.

The inclusion of the residual module helps to enhance the network expressiveness and improve the model classification accuracy. Compared with single background images, cotton pest images in complex backgrounds contain information such as soil, weeds, and so on. There are also problems such as plants occluding each other, which increases the difficulty of obtaining classification features. Feature extraction and fusion of the input information through the residual module preserves the input information and enhances the classification features, thus improving the classification accuracy (He et al., 2024; Wu et al., 2024). Due to the cotton pest early leaf characteristics of small, mostly yellow, red and other small spots on the edge of the leaf, and cotton planting areas are windy and sandy, it is easy to cause the characteristics of the disease spots and complex background confusion. It is easy to lose the insect pest features during the model training process, thus affecting the classification accuracy of the model. In deep networks, skip connections can realize the transfer of information across layers, increase the mobility of information, and solve the problem that pest features are easy to be lost during the training process (Li et al., 2023; V and Kiran, 2022). The results proved that the accuracy of the improved model increased from 94.62% to 97.38%, and the improved model had no obvious classification preference for the three cotton insect pests with balanced classification results.

In recent years, many researchers have used deep learning techniques to study cotton pests and diseases with extensive, high-quality results. As shown in Table 7, many researchers used convolutional neural network as the backbone network, optimized and improved it to make it more suitable for cotton pest and disease detection, and finally realized the identification of cotton pests and diseases. Among them, the model proposed by (Islam et al., 2023) had the highest accuracy of 98.70%, but the model could only determine whether cotton was diseased or not, and could not accurately determine the type of pest or disease. The model proposed in this study had the highest accuracy of 97.38% compared to the remaining three models that can determine the type of diseases and pests.

Table 7
www.frontiersin.org

Table 7. Performance comparison of the proposed model with previous studies.

Deep learning techniques had some success in detecting plant leaf diseases (Liang et al., 2019; Bhujel et al., 2022; Nandhini and Ashokkumar, 2021), with some models achieving classification results almost close to 1 on the Plant Village dataset. However, the photos collected in practical applications contain complex backgrounds such as soil and weeds, and there is also the problem of occlusion, which can lead to a decrease in classification accuracy. Therefore, it is more practical to use complex background image data for model training (Wang et al., 2024; Davidson et al., 2024). With the rapid development of UAVs technology and spectral technology, UAVs have attracted the attention of many scholars by showing convenience and safety in the application of large-scale detection of plant pests (Zhang and Zhang, 2023; Sato et al; Zhang et al., 2024). (Xiao et al., 2021) achieved accurate and timely detection of Fusarium powdery mildew in wheat by combining hyperspectral images captured by utilize unmanned aerial vehicles (UAVs) with ground data and selecting the optimal window size of the gray-level co-occurrence matrix (GLCM) to extract texture features. However, the high cost of hyperspectral equipment makes it difficult to popularize its use on a large scale. (Ma et al., 2024) realized the detection of cotton verticillium wilt disease by combining multispectral image data with ground sampling data using simple linear regression (LR) and multiple linear regression (MLR) methods. However, compared to the above models, deep learning-based models are more advantageous in terms of inter-temporal migration. The improved Swin Transformer model does not require manual image processing to achieve accurate identification of cotton insect pests in the field environment, which reduces the influence of human factors to a certain extent and improves the applicability of the model. The edge device could perform pest detection on the captured images by calling the cotton pest model in the cloud server, and will realizes the cotton pest distribution map finally. The whole process is shown in Figure 9, where the cotton image acquisition work is carried out by means of a UAV, the acquired utilize unmanned aerial vehicles (UAVs) images are cropped into multiple image segments, which are uploaded to a cloud server for cotton pest detection, and the detection results are annotated using equal-sized patches with different colors, and the detection results are spliced together in accordance with the cropping order to ultimately generate a pest distribution map. Based on the pest distribution map, end users can generate application prescription maps according to the medication guide, which provides technical support for precise and variable medication application and promotes the intelligent development of agricultural diseases.

Figure 9
www.frontiersin.org

Figure 9. Pest Distribution Mapping Process.

5 Conclusions

In order to realize the accurate identification of cotton insect pests in complex environments, a cotton insect pest detection model based on improved Swin Transformer was proposed in this study, and the main conclusions are as follows.

Aiming at the characteristics of small and unbalanced samples of cotton pest data, the data set was expanded by digital image processing, and the residual module and skip connection were introduced, which effectively improved the model accuracy from 94.62% to 97.38%. The results of the study proved that the improved model can realize the recognition of cotton insect pests in the field environment.

Through the test of the proposed model, the average single prediction time is 0.00434s, which can complete the judgment of 230 pictures of cotton pests within 1s. It can provide technical reference for the utilization of edge devices such as utilize unmanned aerial vehicles (UAVs).

Based on the results of the current research, the proposed model has great potential for cotton pest detection. However there are several issues that need to be considered in the follow-up study. (1) This study only focuses on the detection of cotton insect pests and ignores the detection of cotton diseases, so the detection of cotton diseases needs to be included in the subsequent study. (2) The size of the improved model is 159.98 MB, which needs to be combined with the computational power of the edge devices in the subsequent research, focusing on the lightweight model, to achieve the application of the equipment in the field, such as the precise application of medicine, to solve the actual production problems in the field.

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Author contributions

TZ: Conceptualization, Formal analysis, Investigation, Methodology, Validation, Visualization, Writing – original draft. JZ: Investigation, Methodology, Software, Validation, Visualization, Writing – review & editing. FZ: Software, Writing – review & editing. SZ: Software, Writing – review & editing. WL: Supervision, Writing – review & editing. RH: Supervision, Writing – review & editing. HD: Supervision, Writing – review & editing. QH: Supervision, Writing – review & editing. CT: Supervision, Writing – review & editing. PL: Supervision, Writing – review & editing.

Funding

The author(s) declare that financial support was received for the research, authorship, and/or publication of this article. This research was funded by Bingtuan Science and Technology Program (No. 2022DB001), the Master Talent Project of the Tarim University Presidents Fund (No. TDZKSS202228) and The University Synergy Innovation Program of Anhui Province (GXXT-2023-101).

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Abbas, A., Jain, S., Gour, M., Vankudothu, S. (2021). Tomato plant disease detection using transfer learning with C-GAN synthetic images. Comput. Electron. Agric. 187, 106279. doi: 10.1016/j.compag.2021.106279

Crossref Full Text | Google Scholar

Alaeddine, H., Jihene, M. (2023). Plant leaf disease classification using wide residual networks. Multimedia Tools Appl. 82, 40953–40655. doi: 10.1007/s11042-023-15226-y

Crossref Full Text | Google Scholar

Arnal Barbedo, J. G. (2013). Digital image processing techniques for detecting, quantifying and classifying plant diseases. SpringerPlus 2, 660. doi: 10.1186/2193-1801-2-660

PubMed Abstract | Crossref Full Text | Google Scholar

Bhujel, A., Kim, N.-E., Arulmozhi, E., Kumar Basak, J., Kim, H.-T. (2022). A lightweight attention-based convolutional neural networks for tomato leaf disease classification. Agriculture 12, 2285. doi: 10.3390/agriculture12020228

Crossref Full Text | Google Scholar

Chattopadhyay, A., Sarkar, A., Howlader, P., Balasubramanian, V. N. (2018). “Grad-CAM++: generalized gradient-based visual explanations for deep convolutional networks,” in IEEE Winter Conference on Applications of Computer Vision (WACV). 839–847.

Google Scholar

Chaudhary, A., Kolhe, S., Kamal, R. (2016). An improved random forest classifier for multi-class classification. Inf. Process. Agric. 3, 215–225. doi: 10.1016/j.inpa.2016.08.002

Crossref Full Text | Google Scholar

Chen, P., Xiao, Q., Zhang, J., Xie, C., Wang, B. (2020). Occurrence prediction of cotton pests and diseases by bidirectional long short-term memory networks with climate and atmosphere circulation. Comput. Electron. Agric. 176, 105612. doi: 10.1016/j.compag.2020.105612

Crossref Full Text | Google Scholar

Davidson, S. J., Saggese, T., Krajňáková, J. (2024). Deep learning for automated segmentation and counting of hypocotyl and cotyledon regions in mature Pinus radiata D. Don. Somatic embryo images. Front. Plant Sci. 15. doi: 10.3389/fpls.2024.1322920

Crossref Full Text | Google Scholar

Ding, Y., Mu, D., Zhang, J., Qin, Z., You, Li., Qin, Z., et al. (2024). A cascaded framework with cross-modality transfer learning for whole heart segmentation. Pattern Recognition 147, 110088. doi: 10.1016/j.patcog.2023.110088

Crossref Full Text | Google Scholar

Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., et al. (2021). An Image Is Worth 16x16 Words: Transformers for Image Recognition at Scale. arXiv. Available at: http://arxiv.org/abs/2010.11929. 1–22

Google Scholar

Ferraz, J. C., Matos, C. H. C., De Oliveira, C. R. F., De Sá, M. D. G. R., Cunha Da Conceição, A. G. (2017). Acaricidal activity of Juazeiro leaf extract against red spider mite in cotton plants. Pesquisa Agropecuária Bras. 52, 493–995. doi: 10.1590/s0100-204x2017000700003

Crossref Full Text | Google Scholar

Guo, Y., Lan, Y., Chen, X. (2022). CST: convolutional Swin transformer for detecting the degree and types of plant diseases. Comput. Electron. Agric. 202, 107407. doi: 10.1016/j.compag.2022.107407

Crossref Full Text | Google Scholar

Haoyu, Y., Haiye, Y., Xu, L., Lei, Z., Yuanyuan, S., et al. (2010). Diagnosis of cucumber diseases and insect pests by fluorescence spectroscopy technology based on PCA-svm. Guang Pu Xue Yu Guang Pu fen xi 30, 30183021. doi: 10.3964/j.issn.1000-0593(2010)11-3018-04

PubMed Abstract | Crossref Full Text | Google Scholar

He, Q., Zhang, J., Chen, W., Zhang, H., Wang, Z., Xu, T. (2024). OENet: an overexposure correction network fused with residual block and transformer. Expert Syst. Appl. 250, 123709. doi: 10.1016/j.eswa.2024.123709

Crossref Full Text | Google Scholar

Irfan, M., Lukman, N., Alfauzi, A. A., Jumadi, J. (2019). Comparison of algorithm support vector machine and C4.5 for identification of pests and diseases in chili plants. J. Physics: Conf. Ser. 1402, 66104. doi: 10.1088/1742-6596/1402/6/066104

Crossref Full Text | Google Scholar

Islam, Md. M., Talukder, Md. A., Sarker, Md. R. A., Uddin, Md A., Akhter, A., Sharmin, S., et al. (2023). A deep learning model for cotton disease prediction using fine-tuning with smart web application in agriculture. Intelligent Syst. Appl. 20, 200278. doi: 10.1016/j.iswa.2023.200278

Crossref Full Text | Google Scholar

Keys, R. (1981). “Cubic convolution interpolation for digital image processing,” in IEEE Transactions on Acoustics, Speech, and Signal Processing, Vol. 29 (6), 1153–1160. doi: 10.1109/TASSP.1981.1163711

Crossref Full Text | Google Scholar

Kiruthika, A., Murugan, M., Jeyarani, S., Sathyamoorthy, N. K., Senguttuvan, K. (2022). Population dynamics of sucking pests in dual season cotton ecosystem and its correlation with weather factors. Int. J. Environ. Climate Change , 586–595. doi: 10.9734/ijecc/2022/v12i1131010

Crossref Full Text | Google Scholar

Kong, J., Yang, C., Xiao, Y., Lin, S., Ma, K., Zhu, Q. (2022). A graph-related high-order neural network architecture via feature aggregation enhancement for identification application of diseases and pests. Comput. Intell. Neurosci. , 2022 (May)1–16. doi: 10.1155/2022/4391491

Crossref Full Text | Google Scholar

Li, P., Liu, J., Yang, L. (2023). Feature pyramid network based on skip connect for vehicle prediction. J. Physics: Conf. Ser. 2456, 0120245. doi: 10.1088/1742-6596/2456/1/012024

Crossref Full Text | Google Scholar

Li, R., He, Y., Li, Y., Qin, W., Abbas, A., Ji, R., et al. (2024). Identification of cotton pest and disease based on CFNet- VoV-GCSP -LSKNet-YOLOv8s: A new era of precision agriculture. Front. Plant Sci. 15. doi: 10.3389/fpls.2024.1348402

Crossref Full Text | Google Scholar

Li, Y., Wang, H., Dang, L.M., Sadeghi-Niaraki, A., Moon, H. (2020). Crop pest recognition in natural scenes using convolutional neural networks. Comput. Electron. Agric. 169, 105174. doi: 10.1016/j.compag.2019.105174

Crossref Full Text | Google Scholar

Li, L., Zhang, S., Wang, B. (2021). Plant disease detection and classification by deep learning—A review. IEEE Access 9, 56683–56698. doi: 10.1109/ACCESS.2021.3069646

Crossref Full Text | Google Scholar

Liang, J., Cao, J., Sun, G., Zhang, K., Van Gool, L., Timofte, R. (2021). “SwinIR: image restoration using swin transformer,” in 2021 IEEE/CVF International Conference on Computer Vision Workshops (ICCVW), Montreal, BC, Canada. 1833–1844, IEEE. doi: 10.1109/ICCVW54120.2021.00210

Crossref Full Text | Google Scholar

Liang, Q., Xiang, S., Hu, Y., Coppola, G., Zhang, D., Sun, W. (2019). PD2SE-net: computer-assisted plant disease diagnosis and severity estimation network. Comput. Electron. Agric. 157, 518–529. doi: 10.1016/j.compag.2019.01.034

Crossref Full Text | Google Scholar

Liu, Z., Lin, Y., Cao, Y., Hu, H., Wei, Y., Zhang, Z., et al. (2021). Swin transformer: hierarchical vision transformer using shifted windows. arXiv. Available at: http://arxiv.org/abs/2103.14030. 1–14

Google Scholar

Luo, Y., Cai, X., Qi, J., Guo, D., Che, W. (2023). FPGA–accelerated CNN for real-time plant disease identification. Comput. Electron. Agric. 207, 107715. doi: 10.1016/j.compag.2023.107715

Crossref Full Text | Google Scholar

M., A., Zekiwos, M., Bruck, A. (2021). Deep learning-based image processing for cotton leaf disease and pest diagnosis. J. Electrical Comput. Eng. 2021, 1–10. doi: 10.1155/2021/9981437

Crossref Full Text | Google Scholar

Ma, R., Zhang, N., Zhang, X., Bai, T., Yuan, X., Bao, H., et al. (2024). Cotton verticillium wilt monitoring based on UAV multispectral-visible multi-source feature fusion. Comput. Electron. Agric. 217, 108628. doi: 10.1016/j.compag.2024.108628

Crossref Full Text | Google Scholar

Manavalan, R. (2022). Towards an intelligent approaches for cotton diseases detection: A review. Comput. Electron. Agric. 200, 107255. doi: 10.1016/j.compag.2022.107255

Crossref Full Text | Google Scholar

Mithal Rind, M., Sayed, S., Ali Sahito, H., Rind, K. H., Rind, N. A., Shar, A. H., et al. (2021). Effects of seasonal variation on the biology and morphology of the dusky cotton bug, oxcarenus laetus (Kirby). Saudi J. Biol. Sci. 28, 3186–3192. doi: 10.1016/j.sjbs.2021.03.065

PubMed Abstract | Crossref Full Text | Google Scholar

Nandhini, S., Ashokkumar, K. (2021). Improved crossover based monarch butterfly optimization for tomato leaf disease classification using convolutional neural network. Multimedia Tools Appl. 80, 18583–18610. doi: 10.1007/s11042-021-10599-4

Crossref Full Text | Google Scholar

Ramcharan, A., Baranowski, K., McCloskey, P., Ahmed, B., Legg, J., Hughes, D. P. (2017). Deep learning for image-based cassava disease detection. Front. Plant Sci. 8. doi: 10.3389/fpls.2017.01852

PubMed Abstract | Crossref Full Text | Google Scholar

Reddy, S. R. G., Varma, G. P. S., Davuluri, R. L. (2023). Resnet-based modified red deer optimization with DLCNN classifier for plant disease identification and classification. Comput. Electrical Eng. 105, 108492. doi: 10.1016/j.compeleceng.2022.108492

Crossref Full Text | Google Scholar

Reder, S., Mund, J.-P., Albert, N., Waßermann, L., Miranda, L. (2021). Detection of windthrown tree stems on UAV-orthomosaics using U-net convolutional networks. Remote Sens. 14, 755. doi: 10.3390/rs14010075

Crossref Full Text | Google Scholar

Rezaei, M., Diepeveen, D., Laga, H., Jones, M. G. K., Sohel, F. (2024). Plant disease recognition in a low data scenario using few-shot learning. Comput. Electron. Agric. 219, 108812. doi: 10.1016/j.compag.2024.108812

Crossref Full Text | Google Scholar

Sato, Y., Tsuji, T., Matsuoka, M. (2024). Estimation of rice plant coverage using sentinel-2 based on UAV-observed data. Remote Sens. 16, 16285. doi: 10.3390/rs16091628

Crossref Full Text | Google Scholar

Selvaraju, R. R., Das, A., Vedantam, R., Cogswell, M., Parikh, D., Batra, D. (2016). Grad-CAM: Why did you say that? Visual Explanations from Deep Networks via Gradient-based Localization. arXiv. doi: 10.48550/arXiv.1610.02391

Crossref Full Text | Google Scholar

Skelsey, P., With, K. A., Garrett, K. A. (2013). Pest and disease management: why we shouldn’t go against the grain. PloS One 8, e758925. doi: 10.1371/journal.pone.0075892

Crossref Full Text | Google Scholar

Steckel, S., Williams, M., Stewart, S. (2021). Efficacy of selected insecticides for managing cotton aphid in cotton 2020. Arthropod Manage. Tests 46, tsab044. doi: 10.1093/amt/tsab044

Crossref Full Text | Google Scholar

Syed-Ab-Rahman, S. F., Hesamian, M. H., Prasad, M. (2022). Citrus disease detection and classification using end-to-end anchor-based deep learning model. Appl. Intell. 52, 927–385. doi: 10.1007/s10489-021-02452-w

Crossref Full Text | Google Scholar

Thivya Lakshmi, R. T., Katiravan, J., Visu, P. (2024). CoDet: A novel deep learning pipeline for cotton plant detection and disease identification. Automatika 65, 662–745. doi: 10.1080/00051144.2024.2317093

Crossref Full Text | Google Scholar

V, A., Kiran, A.G. (2022). SynthNet: A skip connected depthwise separable neural network for novel view synthesis of solid objects. Results Eng. 13, 100383. doi: 10.1016/j.rineng.2022.100383

Crossref Full Text | Google Scholar

Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., et al. (2023). Attention is all you need. arXiv. Available at: http://arxiv.org/abs/1706.03762. 1–15.

Google Scholar

Wang, K., Hu, X., Zheng, H., Lan, M., Liu, C., Liu, Y., et al. (2024). Weed detection and recognition in complex wheat fields based on an improved YOLOv7. Front. Plant Sci. 15. doi: 10.3389/fpls.2024.1372237

Crossref Full Text | Google Scholar

Wu, R., Qin, C., Huang, G., Tao, J., Liu, C. (2024). Precise cutterhead clogging detection for shield tunneling machine based on deep residual networks. Int. J. Control Automation Syst. 22, 1090–11045. doi: 10.1007/s12555-022-0576-8

Crossref Full Text | Google Scholar

Xiao, Y., Dong, Y., Huang, W., Liu, L., Ma, H. (2021). Wheat fusarium head blight detection using UAV-based spectral and texture features in optimal window size. Remote Sens. 13, 24375. doi: 10.3390/rs13132437

Crossref Full Text | Google Scholar

Zhai, Z., Martínez, J. F., Beltran, V., Martínez, N. (2020). Decision support systems for agriculture 4.0: survey and challenges. Comput. Electron. Agric. 170, 105256. doi: 10.1016/j.compag.2020.105256

Crossref Full Text | Google Scholar

Zhang, J., Abdelraheem, A., Zhu, Y., Elkins-Arce, H., Dever, J., Whitelock, D., et al. (2022). Studies of evaluation methods for resistance to fusarium wilt race 4 (Fusarium oxysporum f. Sp. Vasinfectum) in cotton: effects of cultivar, planting date, and inoculum density on disease progression. Front. Plant Sci. 13. doi: 10.3389/fpls.2022.900131

Crossref Full Text | Google Scholar

Zhang, Y., Ma, B., Hu, Y., Li, C., Li, Y. (2022). Accurate cotton diseases and pests detection in complex background based on an improved YOLOX model. Comput. Electron. Agric. 203, 107484. doi: 10.1016/j.compag.2022.107484

Crossref Full Text | Google Scholar

Zhang, W., Peng, X., Bai, T., Wang, H., Takata, D., Guo, W. (2024). A UAV-based single-lens stereoscopic photography method for phenotyping the architecture traits of orchard trees. Remote Sens. 16, 15705. doi: 10.3390/rs16091570

Crossref Full Text | Google Scholar

Zhang, S., Zhang, C. (2023). Modified U-net for plant diseased leaf image segmentation. Comput. Electron. Agric. 204, 107511. doi: 10.1016/j.compag.2022.107511

Crossref Full Text | Google Scholar

Zhang, N., Zhang, X., Shang, P., Ma, R., Yuan, X., Li, L., et al. (2023). Detection of cotton verticillium wilt disease severity based on hyperspectrum and GWO-SVM. Remote Sens. 15, 33735. doi: 10.3390/rs15133373

Crossref Full Text | Google Scholar

Zhang, J., Zheng, Y., Qi, D., Li, R., Yi, X., Li, T. (2018). Predicting citywide crowd flows using deep spatio-temporal residual networks. Artif. Intell. 259, 147–166. doi: 10.1016/j.artint.2018.03.002

Crossref Full Text | Google Scholar

Zhou, X., Zhou, W., Fu, X., Hu, Y., Liu, J. (2023). MDvT: introducing mobile three-dimensional convolution to a vision transformer for hyperspectral image classification. Int. J. Digital Earth 16, 1469–1905. doi: 10.1080/17538947.2023.2202423

Crossref Full Text | Google Scholar

Zhu, Y., Elkins-Arce, H., Wheeler, T. A., Dever, J., Whitelock, D., Hake, K., et al. (2023). Effect of growth stage, cultivar, and root wounding on disease development in cotton caused by fusarium wilt race 4 ( Fusarium oxysporum f. Sp. Vasinfectum ). Crop Sci. 63, 101–145. doi: 10.1002/csc2.20839

Crossref Full Text | Google Scholar

Zhu, Y., Xia, Q., Wen, J. (2022). SRDD: A lightweight end-to-end object detection with transformer. Connection Sci. 34, 2448–2655. doi: 10.1080/09540091.2022.2125499

Crossref Full Text | Google Scholar

Keywords: cotton pests, swin transformer, complex background, deep learning, unmanned aerial vehicles

Citation: Zhang T, Zhu J, Zhang F, Zhao S, Liu W, He R, Dong H, Hong Q, Tan C and Li P (2024) Residual swin transformer for classifying the types of cotton pests in complex background. Front. Plant Sci. 15:1445418. doi: 10.3389/fpls.2024.1445418

Received: 07 June 2024; Accepted: 08 August 2024;
Published: 27 August 2024.

Edited by:

Juan Wang, Hainan University, China

Reviewed by:

Shanwen Zhang, Xijing University, China
Weixiang Yao, Shenyang Agricultural University, China
Tiwei Zeng, East China Jiaotong University, China

Copyright © 2024 Zhang, Zhu, Zhang, Zhao, Liu, He, Dong, Hong, Tan and Li. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Ping Li, 119980015@taru.edu.cn

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.