- 1School of Computer Science and Engineering, Galgotias University, Greater Noida, India
- 2Department of Computer Engineering, Chosun University, Gwangju, Republic of Korea
- 3Department of Computer Application, Dharma Samaj (D.S.) College, Aligarh, India
- 4Department of Information Technology (IT), Institute of Technology and Science, Mohan Nagar Ghaziabad, India
- 5Department of Computer Science and Engineering (CSE), School of Technology, Pandit Deendayal Energy University, Gandhinagar, Gujarat, India
- 6Department of Computer Engineering and Information, College of Engineering in Wadi Alddawasir, Prince Sattam Bin Abdulaziz University, Wadi Alddawasir, Saudi Arabia
- 7School of Information Technology (IT) and Engineering, Melbourne Institute of Technology, Melbourne, VIC, Australia
Introduction: Cotton, being a crucial cash crop globally, faces significant challenges due to multiple diseases that adversely affect its quality and yield. To identify such diseases is very important for the implementation of effective management strategies for sustainable agriculture. Image recognition plays an important role for the timely and accurate identification of diseases in cotton plants as it allows farmers to implement effective interventions and optimize resource allocation. Additionally, deep learning has begun as a powerful technique for to detect diseases in crops using images. Hence, the significance of this work lies in its potential to mitigate the impact of these diseases, which cause significant damage to the cotton and decrease fibre quality and promote sustainable agricultural practices.
Methods: This paper investigates the role of deep transfer learning techniques such as EfficientNet models, Xception, ResNet models, Inception, VGG, DenseNet, MobileNet, and InceptionResNet for cotton plant disease detection. A complete dataset of infected cotton plants having diseases like Bacterial Blight, Target Spot, Powdery Mildew, Aphids, and Army Worm along with the healthy ones is used. After pre-processing the images of the dataset, their region of interest is obtained by applying feature extraction techniques such as the generation of the biggest contour, identification of extreme points, cropping of relevant regions, and segmenting the objects using adaptive thresholding.
Results and Discussion: During experimentation, it is found that the EfficientNetB3 model outperforms in accuracy, loss, as well as root mean square error by obtaining 99.96%, 0.149, and 0.386 respectively. However, other models also show the good performance in terms of precision, recall, and F1 score, with high scores close to 0.98 or 1.00, except for VGG19. The findings of the paper emphasize the prospective of deep transfer learning as a viable technique for cotton plant disease diagnosis by providing a cost-effective and efficient solution for crop disease monitoring and management. This strategy can also help to improve agricultural practices by ensuring sustainable cotton farming and increased crop output.
1 Introduction
Cotton plants are significantly impacted by pests, bacteria or fungal infections which leads to reduced yields and contributes to economic challenges, particularly within the agricultural sector. These diseases pose a serious threat, adversely affecting the development, growth, and overall health of cotton plants. Aphids, Army Worm, Bacterial Blight, Powdery Mildew, and Target Spot are among the most common diseases affecting cotton plants. Aphids are caused by sap-sucking insects that weaken the plant by draining essential nutrients. Army Worm, a pest caused by larvae of Spodoptera moths, results in severe leaf defoliation and damage to buds. Bacterial Blight, triggered by Xanthomonas axonopodis pv. malvacearum, leads to leaf lesions and reduced fiber quality. Powdery Mildew, caused by fungi like Erysiphe, appears as white powdery patches on leaves, hindering photosynthesis. Target Spot, caused by Corynespora cassiicola, creates brown circular lesions, reducing yield. Each disease significantly impacts crop health and productivity (Arthy et al., 2024) (Figure 1 shows the leaf of cotton plant being damaged by bacterial blight). These diseases cause significant damage to the crops and decrease its fibre quality as well as throws farmers who deal with the cotton plants in a financial loss. Hence, it is very important to timely, efficiently, and precisely identify such diseased cotton plants (Bhatti et al., 2020). Various traditional techniques are available such as laboratory testing, visually inspecting, etc but they have certain limitations also. Examining any disease on the cotton plant visually is mostly prone to errors as it is completely based on the skill or expertise of a human being (Caldeira et al., 2021).
Figure 1. Cotton plant leaf damaged by bacterial blight. The red bounding box highlights the area affected by bacterial blight on the leaf.
In fact, to differentiate symptoms caused by these infections from those caused by other factors, such as dietary deficits or environmental stress, can be also difficult. Likewise, laboratory testing is more objective, thereby it is time-consuming and expensive. Moreover, collecting, transporting, and processing under specialized facilities samples can also cause delays in diagnosis and disease management (Zhu et al., 2022).
In these years, deep learning techniques have proven to be as a viable strategy for classifying image-based disease in a multiple fields like plant pathology. By using neural networks, especially convolutional neural networks (CNNs) (Ferentinos, 2018), deep learning models can automatically pull-out hierarchical information from digital images of diseased cotton plant parts like leaves and stems. This image-based approach can learn discriminative features, handle large datasets, and generalize well to previously unseen data. We may accomplish precise and efficient classification of cotton plant diseases using deep learning models trained on labeled datasets of cotton plant pictures, providing farmers, agronomists, and researchers with timely and trustworthy information for disease management methods (Bera and Kumar, 2023; Montalbo, 2022).
Numerous research endeavors have focused on leveraging deep learning methodologies to address the challenge of identifying and classifying diseases affecting cotton plants. Kumar et al. (2022) integrated Convolutional Neural Networks (CNNs) into a mobile application designed to assist farmers to identify and recommend cotton diseases and suitable pesticides respectively. They achieved this by converting the TensorFlow Tflite model (Google AI Edge, 2024) into a Core ML model (Apple Developer, 2024) for seamless integration with iOS apps. Similarly, Memon et al. (2022) proposed a meta-Deep Learning model using a dataset of 2,385 images of both healthy and diseased cotton leaves. They expanded the dataset to enhance the performance of model which resulted in a notable accuracy rate of 98.53%, surpassing other models tested on the Cotton Dataset. Rajasekar et al. (2021) addressed the challenge of detecting cotton plant diseases by developing a hybrid network combining ResNet and Xception models. Their approach outperformed existing techniques, with ResNet-50 achieving a training accuracy (0.95) and a validation accuracy (0.98), along with a training loss (0.33) and a validation loss (0.5). In another study, Naeem et al. (2023) detected, identified, as well as diagnosed cotton leaf diseases using the Root Mean Square Propagation (RMSprop) as well as Adaptive Moment Estimation (ADAM) optimizers. Apart from this, they combined Inception and VGG-16 as feature extractors which resulted in the highest mean accuracy, with the CNN achieving an overall accuracy of 98%. Kalaiselvi and Narmatha (2023) worked on 400 images of cotton plant disease and segmented them using fuzzy rough C-means (FRCM) clustering technique combined with CNN for classification. The researchers demonstrated a remarkable 99% accuracy in diagnosing diseases like Bacterial Blight and Cercospora Leaf Spot. Meanwhile, (Odukoya et al. (2023)) utilized image processing techniques, including watershed segmentation, Edge Detection, Support Vector Machine as well as K-Means Clustering, to detect fungal infections in cotton plants. Arathi and Dulhare (2023) leveraged the DenseNet-121 pre-trained model for enhanced disease classification in cotton leaves, achieving a classification accuracy of 91%. Gülmez (2023) focused on analyzing leaf images to determine plant health, employing the Grey Wolf Optimization technique to identify the most efficient model architecture. Hyder and Talpur (2024) reported a high accuracy rate of 95% using deep transfer learning techniques on a dataset of over 10,000 images of cotton plants. Their model’s effectiveness stemmed from its training on a large-scale dataset. Kukadiya et al. (2024) used a hybrid model combining VGG16 + InceptionV3 to early detect diseases in cotton leaf. By optimizing hyperparameters such as number of epochs and learning rate using stochastic gradient descent (SGD), their ensemble model achieved superior performance with 98% as training accuracy and 95% testing accuracy. Similarly, Mohmmad et al. (2024) worked on the identification of five different cotton crop diseases which includes Aphids, Bacterial Blight, Curly Leaves, Powdery Mildew, and Verticillium Wilt—along with a healthy class. They had dataset of 1,200 images of cotton leaves and were applied on VGG-16, MobileNet, VGG-19, and custom CNN models for classification. These efforts underscore the potential of deep learning in accurately identifying and classifying diseases in cotton leaves, offering effective solutions for agricultural management.
Although the models have demonstrated strong performance, they also face several limitations. One major issue is the use of relatively small or homogenous datasets, which can constrain the generalization and robustness of the models. This limitation is particularly evident when some models achieve high accuracy while others exhibit lower accuracy, as they may be prone to overfitting due to insufficient data diversity. Furthermore, despite some models showing promising results, there is a notable lack of comprehensive comparative analysis between different meta-architectures and optimization strategies. Addressing these gaps could provide valuable insights into model performance and guide the development of more robust and generalizable solutions (Memon et al., 2022; Gülmez, 2023; Naeem et al., 2023; Hyder and Talpur, 2024; Mohmmad et al., 2024).
Hence, the motivation for using proposed models stems from cotton’s pivotal role in global agriculture and the economy. Diseases such as Aphids, Target Spot (a fungal disease), Army Worm, Bacterial Blight, and Powdery Mildew can cause substantial crop losses if not detected early. An automated deep learning-based system offers a more efficient, accurate, and scalable approach to early disease detection. This not only helps safeguard crop yields but also reduces economic losses and promotes sustainability in cotton farming. Apart from this, our work will also try to work on the challenges faced by the existing researchers by conducting comparisons which could provide valuable insights into the most effective approaches for cotton disease detection. Additionally, datasets will also be expanded to encompass a broader range of disease stages, cotton varieties and regularization techniques etc will be used to enhance the execution of model.
In this paper, an automated approach was built that can predict and categorize five types of plant diseases such as Army Worm, Aphids, Bacterial Blight, Target Spot, and Powdery Mildew by evaluating cotton plants alongside healthy ones.
This research has made significant contributions toward achieving this objective:
Comprehensive Image Preprocessing Pipeline: One of the limitations of many existing studies is the use of small or homogeneous datasets, which can hinder model generalization. The paper introduces a robust preprocessing pipeline for handling the COTTON PLANT DISEASE (CPD) dataset, which includes 36,000 images (Dhamodharan, 2023). The preprocessing steps involve removing noise and converting images to grayscale, which enhances the clarity and quality of the data for subsequent analysis. It graphically presents the pre-processed data to reveal patterns and characteristics within the dataset, aiding in a more informed extraction of regions of interest.
Advanced Region Extraction Techniques: The challenge is highlighted regarding the susceptibility of models to overfitting, particularly when there is limited diversity in the data. The paper develops techniques for extracting significant regions of interest from the images by generating the largest contour and identifying extreme points, ensuring that key features are accurately captured. These techniques are employed to further refine the regions of interest, enhancing the precision of the data used for model training and evaluation. Besides this, it will also minimize the risk of overfitting, as the models are trained on well-defined and precise data.
Application and Evaluation of Deep Transfer Learning Models: The study applies a diverse set of advanced transfer learning models - InceptionV3, EfficientNetB3, Xception, ResNet152V2, VGG19, DenseNet169, MobileNetV2, ResNet50V2, EfficientNetB0, as well as InceptionResNetV2, to classify images of healthy leaves and various diseases such as Aphids, Target Spot, Army Worm, Bacterial Blight, and Powdery Mildew. The performance of these models is rigorously evaluated using multiple metrics including F1 score, accuracy, loss, recall, and precision, all computed through the confusion matrix values. This comprehensive evaluation allows for a detailed comparison of model performance.
Selection of Optimal Model: Based on the evaluation metrics, this work identifies and selects the most effective model to classify cotton plant diseases. This selection is guided by an analysis of the models’ performance metrics, ensuring that the chosen model delivers the highest accuracy and reliability. Moreover, the performance of the optimal model has also been compared based on multiple attributes such as dataset, classes, techniques, and accuracy of the existing work which were not reported in some of the previous papers.
The structure of the manuscript is as followed: Section 1 has been already mentioned as an introduction part, which gives us information about the contribution of the researcher’s in the area of detecting cotton plant disease, Section 2 informs us about the methodology that has been used to develop the model for cotton plant disease detection and classification followed by Section 3 where results are analyzed and discussed. Further, the real time implications, improvements as well as the future scope of this research are presented in Section 4 while as the whole paper is at the end summarized and concluded in Section 5.
2 Research methodology
This section focuses on the process of developing a model to identify and classify diseases in cotton plants. The model uses images of the plants and applies several image processing techniques, as well as deep learning classifiers. Apart from this, the novelty of this work lies in the implementation of advanced preprocessing techniques for the Cotton Plant Disease (CPD) dataset. These techniques include comprehensive noise removal, conversion of images to grayscale, and the graphical representation of data to discern underlying patterns. Key preprocessing steps involve extracting Regions of Interest (ROIs) by calculating various image characteristics, identifying the largest contour, locating extreme points, and applying cropping and adaptive thresholding. These refined ROIs are subsequently used as inputs for the deep learning models, thereby enhancing the ability of the model to detect as well as classify plant diseases with improved accuracy. Figure 2 displays the structure of the proposed model.
A pseudo code (Algorithm 1) in disease detection in cotton plants using deep learning techniques has also been mentioned. It outlines the flow of steps involved in training various advanced CNN models for classifying images of healthy and diseased cotton leaves. Let D be the CPD dataset containing images of healthy leaves and diseased leaves labeled as Aphids, Target Spot, Army Worm, Bacterial Blight, Powdery Mildew, and healthy leaves of cotton plant.
Algorithm 1. Algorithmic flow of cotton plant disease detection and classification using Deep Transfer
Learning Models.
In summary, the proposed system demonstrates a well-structured approach to use advanced image preprocessing techniques and deep transfer learning models for the detection of cotton plant diseases. The application of these methods ensures improved accuracy, minimizing the risks of overfitting and maximizing model performance. Thus, the subsequent section will detail utilized in this study, highlighting its significance in the context of cotton plant disease detection.
2.1 Dataset
The data for the detection and classification of diseases in cotton leaves have been taken from the Cotton Plant Disease database (Dhamodharan, 2023). It is a dataset hosted on the Kaggle platform, which is a popular online community for data scientists and machine learning practitioners. The dataset focuses on diseases affecting cotton plants which include Army Worm, Aphids, Target spot, Powdery Mildew, and Bacterial Blight. It also includes healthy leaf dataset for comparison with the diseased plant. Figure 3 represents some samples of disease affected cotton leaves which have been taken from the dataset.
The dataset mainly focuses on the disease which occurs only on leaves, and it does not have any reference images for diseases on stem, buds, flowers and boll. It is intended to aid researchers, scientists, and agricultural experts in studying and understanding different diseases that commonly affect cotton crops. By providing access to this dataset, the aim is to enable the development of machine learning models and data-driven approaches to help diagnose and manage cotton plant diseases effectively.
2.2 Data pre-processing
In computer vision and image analysis tasks, pre-processing of cotton leaves images is a critical step for improving the quality of the images and facilitate subsequent analyses. OpenCV, a widely used library in computer vision, is employed due to its rich collection of functions and tools tailored for such tasks (Kotian et al., 2022). In the preprocessing phase for cotton leaf image analysis, the images are first converted to grayscale to simplify processing and reduce computational complexity (Figure 4). The grayscale conversion is achieved using the following mathematical formula:
where , , and represent the intensity values of the red, green, and blue channels of a pixel, respectively. The coefficients (0.299, 0.587, and 0.114) are derived from the luminance contribution of each color channel in the human visual system. This formula converts the RGB image into a single-channel grayscale image, effectively reducing its dimensionality while retaining the essential visual information necessary for further analysis.
Following the grayscale conversion, Gaussian blurring is applied using a Gaussian kernel to create a smoothed version of the image. This step reduces high-frequency noise and artifacts, enhancing the overall clarity of the image and making it easier to identify relevant features during subsequent analyses. The Gaussian blurring operation is mathematically represented by the convolution of the image with a Gaussian kernel:
where represents the Gaussian kernel function centered at and k is the kernel radius. This convolution operation smooths the image, aiding in noise reduction and improving feature detection.
The combination of grayscale conversion and Gaussian blurring contributes to the overall goal of preparing the images for analysing it further and classifying the tasks in a computationally efficient manner.
2.3 Feature extraction
After examining the dataset and its quality, the next main step is to extract region of interest from each image so that the model can be trained well in future. In this paper, Contour feature extraction has been used which is a fundamental technique in image processing and computer vision used to capture essential shape and boundary information of objects within an image. Contour features are characteristics derived from the contours of regions of interest or object within an image. Contours represent the boundaries of connected components with the same intensity or color. By extracting contour features, we can quantitatively describe the shape and characteristics of these objects, providing valuable information for various image analysis tasks.
In this paper, we have calculated various parameters of an image such as width, area, aspect ratio, height, maximum and minimum value location, mean intensity, extreme right and leftmost, bottom and topmost values etc. in Table 1. The area is calculated from the height and width of the object’s bounding rectangle, with height and width derived from the bounding rectangle function. The perimeter measures the distance around the object, and epsilon represents an approximation accuracy for contour approximation. Aspect ratio is the proportion between the width and height of the object, while extent is the ratio of the object’s area to the area of its bounding rectangle. The equivalent diameter corresponds to the diameter of a circle with the same area as the object. Minimum and maximum values, along with their locations, are obtained using specific OpenCV functions, and the mean color of the object is computed from pixel values. The extreme leftmost point is the farthest left coordinate of the contour. These parameters help in detailed shape analysis and characterization of objects in images. The extreme rightmost point of an object is the contour point with the maximum x-coordinate value, representing the farthest point to the right. Similarly, the extreme topmost point is the contour point with the minimum y-coordinate value, which is the highest point on the object. The extreme bottommost point is the contour point with the maximum y-coordinate, marking the lowest part of the object. These points are useful for understanding the spatial boundaries of an object in image analysis and can be obtained using contour analysis functions in OpenCV.
After computing the characteristics of an image using contour features, the next steps involve several image processing techniques for further analysis and extraction of regions of interest (ROIs) as shown in Figure 5.
Figure 5. Obtaining region of interest from the images using feature extraction techniques (A) Aphids (B) Army Worm (C) Bacterial Blight (D) Powdery Mildew (E) Target Spot (F) Healthy Leaves.
First, the biggest contour, representing the most prominent object in the image, is identified. This can be achieved by finding the contour with the largest area among all detected contours. In the second step, the extreme points of this biggest contour are generated in the form of a continuous curve around the object. Later, the images are cropped within the boundary of those curves to isolate it from the rest of the features.
In the next phase, adaptive thresholding is applied to the cropped image to adjust the threshold value for each pixel in the image based on its local neighbourhood to enhance the quality of an image. In addition to this, it also improves the segmentation of the object from the background, most particularly in those cases where there is a variation in the lighting conditions or uneven illumination. Finally, the last step is obtaining the region of interest for which another contour detection technique is applied.
This contour delineates the boundary of the main object, which can be further utilized for shape analysis, feature extraction, or object recognition tasks. The combination of all these techniques provides a systematic approach to extract and analyse specific regions in the form of features from the original image, for providing more accurate image analysis and computer vision tasks. Later the dataset has been split into training and validation in 80:20 based on which the performance of the applied models has been examined.
2.4 Applied deep learning classifiers
This section provides a concise overview of the many deep learning classifiers that have been used for the detection and classification of cotton plant diseases. Additionally, the section also defines the hyperparameters used to build model during training it with the cotton plant disease dataset, as presented in Table 2.
EfficientNet is a family of CNNs that significantly improve the efficiency of deep learning models through a combination of model scaling, depthwise separable convolutions, and a compound scaling method (Tan, 2019). EfficientNet employs a novel approach to balance network depth, width, and resolution, to achieve performance on various benchmarks during the maintenance of computational efficiency. The key innovation of EfficientNet lies in its compound scaling method, which uniformly scales all dimensions of the network, leading to a better trade-off between accuracy and computational cost. This architecture has been particularly effective in a range of computer vision tasks due to its ability to deliver high performance with less parameter compared to traditional deep learning models.
In paper two versions of EfficientNet model i.e. EfficientNetB0 and EfficientNetB3 have been used (Babu and Jeyakrishnan, 2022; Sun et al., 2022). EfficientNetB0 is a convolutional neural network architecture made up of a stem convolution and a base architecture with repeating building blocks that allows it to successfully extract features in a hierarchical order. Custom dense layers are added to adapt it for cotton plant disease detection, followed by an output layer with a softmax activation function to match the number of classes. Transfer learning allows the use of already trained weights from ImageNet as well as data augmentation techniques and appropriate optimisation approaches to improve the model’s performance during training. By leveraging EfficientNetB0 and fine-tuning hyperparameters, one can achieve robust and accurate cotton plant disease detection and classification (Sun et al., 2022). On the contrary, EfficientNetB3 is a type of CNN that provides promising results to detect as well as classify cotton plant diseases. The architecture has a balancing width, depth, resolution, and compound scaling approach, for optimizing the performance. This model comprises of multiple blocks of depthwise separable convolutions, which split channel-wise and spatial convolutions for reducing the computational complexity. These blocks enable the model to learn hierarchical features at different scales, crucial for identifying disease-related patterns in cotton plant images. For multi-class classification, the architecture includes a global average pooling layer and fully connected layers, along with softmax activation. The model focuses on achieving accurate and efficient detection along with the classification of multiple diseases in cotton plant images in order to empower agricultural applications for identifying diseases at their earliest stages and improving crop management.
DenseNet is a convolutional neural network architecture characterized by its dense connectivity pattern (He et al., 2016). Unlike traditional networks where each layer receives input from the previous layer, DenseNet works in a feed forward fashion by connecting each layer to every other layer. This dense connectivity improves gradient flow, reduces the risk of vanishing gradients, and encourages feature reuse, leading to more efficient training and better performance. DenseNet models are known for their reduced number of parameters compared to other deep networks while maintaining high accuracy, making them effective for a variety of image recognition tasks. In this research, its variant has been used which is DenseNet169 (Akbar et al., 2023; Huang et al., 2017). It presents an effective solution for detecting and classifying cotton plant diseases. The architecture is characterized by promoting strong feature reuse, densely connected layers, and gradient flow across the network, leading to better parameter efficiency and performance. A comprehensive dataset comprising of images of both healthy and diseased cotton plants is gathered and categorized with corresponding disease classifications, specifically for the purpose of detecting and classifying cotton plant diseases. DenseNet-169’s distinctive dense blocks, each with 32 growth rate feature maps, facilitate the extraction of intricate and disease-specific features from the input images. Between dense blocks, transition blocks with convolutional layers and average pooling help manage computational complexity while maintaining relevant information flow.
Xception, short for “Extreme Inception,” is designed to enhance existing convolutional networks by employing depthwise separable convolutions (Rai and Pahuja, 2024). This architecture effectively combines spatial and channel-wise filtering, resulting in improved computational efficiency and performance. The model is based on the idea of Inception modules, but instead of using regular convolutions, it employs depthwise separable convolutions. This technique significantly decreases the number of parameters as well as computational complexity. Xception is made up of several depthwise separable convolution blocks, each containing residual connections for improved gradient flow and information propagation.
MobileNetV2 has an architecture that is lightweight and built for efficient and quick performance on mobile and embedded devices (Verma and Singh, 2022). The usage of inverted residual blocks, which are designed to balance computational efficiency and model correctness, is a significant component of MobileNetV2. These blocks use stride in the depthwise separable convolution followed by a pointwise convolution with expansion and squeeze-and-excitement operations to improve how features are represented and how information flows. MobileNetV2 also uses linear bottlenecks to reduce computing costs while preserving performance. The architecture is designed to strike a balance between model size, speed, as well as accuracy, which makes it well-suited for the applications and environments in real time with limited number of resources.
ResNet (Residual Networks), revolutionized deep learning with its introduction of residual blocks, which help mitigate the vanishing gradient problem in very deep networks (He et al., 2016). ResNet allows for the training of extremely deep networks by introducing shortcut connections that skip one or more layers. These residual connections enable gradients to flow more effectively through the network during backpropagation, which significantly improves training efficiency and model performance. ResNet architectures are known for their remarkable ability to achieve high accuracy on various image recognition tasks and have set new benchmarks in computer vision competitions, making them a fundamental building block in modern deep learning.
The extended version of ResNet is ResNet50V2V2 and ResNet152V2 (Ahad et al., 2023; Kumar et al., 2023). The architecture of ResNet50V2 consists of 50 layers, comprising multiple stacked residual blocks with increasing complexity. The first few layers could perform initial feature extraction from input images of healthy and diseased cotton plants. The subsequent layers would progressively learn more abstract and intricate disease-related features. The network’s final layers would include dense layers for classification, with softmax activation to produce the probabilities for different disease classes. Similarly, another extended version of ResNet152V2. It is an extension of the ResNet architecture with 152 layers, showcasing deeper and more powerful representations. The architecture begins with initial convolutional layers to extract low-level features from input images of healthy and diseased cotton plants. Then, a series of residual blocks with increasing complexity would be stacked on top of each other, enabling the network to learn hierarchical and abstract disease-related features. The residual connections within each block facilitate the smooth flow of information and alleviate the vanishing gradient problem which allows for effective training of very deep networks. Towards the end of the network, global average pooling would be employed to reduce spatial dimensions, and dense layers with softmax activation would perform the final classification into different disease classes.
VGG models are characterized by their simplicity and uniformity (Simonyan and Zisserman, 2014). The VGG network design emphasizes the use of small 3x3 convolutional filters and 2x2 max-pooling layers, which contribute to a deep architecture while keeping the model design straightforward. VGG networks, particularly VGG16 and VGG19, have become widely used due to their high performance in classifying images and their ability to serve as powerful feature extractors for various transfer learning applications. Their straightforward architecture and high accuracy make them a popular choice in the deep learning community.
In this paper, VGG19 is being used (Peyal et al., 2022). The architecture has a deep convolutional neural network that can be adapted for the detection and classification of cotton plant diseases. VGG19 consists of 19 layers, including multiple convolutional and max-pooling layers. The architecture follows a simple and uniform design, where each layer contains 3x3 convolutional filters, followed by ReLU activations to introduce non-linearity. Max-pooling layers are used for downsampling and reducing spatial dimensions. Towards the end of the network, fully connected layers are employed for classification, with softmax activation to predict the probabilities of different disease classes. To adapt VGG19 for cotton plant disease detection, the final dense layers would be modified to match the number of disease classes. By training the model on a diverse dataset of cotton plant images, VGG19 would aim to accurately detect and classify various cotton plant diseases, aiding in early disease identification and effective crop management in agriculture.
Inception networks employ the “Inception module” for feature extraction which considers features at different scales and comprise several convolutional filters of different sizes in the same layer. It also enables model to obtain multiple forms of representation of feature, which is advantageous. Inception networks are incredibly popular due to their high effective and computational ability in image classification as well as other computer vision application. InceptionV3 model is a deep convolutional neural network and it has been found to be useful for the detection and classification of diseases affecting cotton plant (Mary et al., 2022). It is an improvement of the others previously created by Google Research such as Inception and InceptionV2 and aims at achieving a good level of accuracy in addition to ensuring efficient use of computations through its implementation. There are several inception modules, which are convolutional blocks with multiple branches of different kernel size to make it able to capture features of different scales. The same concept exists in InceptionV3 where factorization has been performed to split them into more compact convolutions. For detecting the cotton plant diseases, the last dense layers of InceptionV3 can be trained according to the specific diseases class to help in the agricultural applications & improving crop management.
Hybrid (InceptionResNetV2) incorporates concepts from both the Inception and ResNet models (Sadiq et al., 2023). The architecture incorporates residual connections as well as inception modules via ResNet and the inception family, respectively. InceptionResNetV2 is made up of deep convolutional layers followed by inception blocks with several branches and varying kernel sizes for feature extraction. Furthermore, residual blocks are used to ensure that information flows smoothly across the network, to allow for the effective training of very deep models. InceptionResNetV2 is well-known for its excellent accuracy and is extensively used in classification of images, detecting objects, and segmentation-like applications in computer vision. However, due to its complex nature, it may necessitate significant computer resources for computation.
In addition to this, Table 3 provides a comparison of various applied models by detailing their total number of parameters which includes the breakdown of trainable and non-trainable parameters. This information helps to understand the computational complexity and potential performance of each model during training and validation. Here, VGG19 has the highest number of parameters at 143.67 million, all of which are trainable, indicating a highly complex model but potentially more prone to overfitting. ResNet152V2 follows with over 60 million parameters, emphasizing its depth and strong learning capacity. In contrast, models like MobileNetV2 and EfficientNetB0 have significantly fewer parameters (3.53 million and 5.27 million, respectively), making them more efficient for real-time applications, though possibly sacrificing some accuracy due to reduced complexity. Xception, InceptionV3, and DenseNet169 strike a balance between complexity and performance, with substantial parameter counts but remains computationally feasible.
2.5 Performance metrics
Performance metrics are used to examine the quality and the performance of the models for any data. These metrics provide quantitative measures that assess how well a model performs its intended task, such as classification or regression. The selection of performance measures is based upon the specific scenario and the objectives of the analysis. Here are the metrics that have been used to evaluate the performance of the models (Latif et al., 2021; Elaraby et al., 2022; Kinger et al., 2022; Nalini and Rama, 2022; Kanna et al., 2023):
Accuracy is a broad measure of how well a model performs in correctly identifying positive and negative examples. If the accuracy is having greater value, it suggests that the classification done is very precise while as Loss is a mathematical function that measures the difference between the predicted values of a model as well as the actual values in the training data. Similarly, Root Mean Square Error is defined as the square root of the mean of the squared residuals, which are the differences between the predicted values as well as the true values. Moving to another set of parameters, Precision provides information related to the ability of the model for avoiding false positives. It means that informs the correctly identification of positive instances without generating large number of incorrectly positive predictions. Recall is used for examining the ability of model to identify accurately the positive classes. In short, recall, enables to identify all positive instances and avoids false negatives whereas F1 Score is a performance metric that considers both precision and recall, giving a balanced measure of how accurate a classification model is.
3 Results
The section of this research paper serves as the focal point where empirical findings are presented, unveiling the outcomes of rigorous investigations and data analysis. After framing and executing the objectives and methodology of our study in the preceding sections, we now turn our attention to the fundamental outcomes that emerged from our investigation. In this section, we have presented a concise and informative analysis of the applied deep transfer learning models based on different performance measures as mentioned in section (2.5) for both training as well as validation dataset.
Initially we conducted a complete evaluation of various applied deep learning models for the whole cotton plant disease dataset. To calculate the efficacy of each model, we rigorously measured three critical performance metrics: accuracy, loss, and root mean square error across both the training as well as validation datasets. Employing a ten-epoch training regimen, we obtained valuable insights into the learning dynamics and convergence patterns of each model, as shown in Table 4. Such detailed assessment enables us to make informed decisions about model selection, parameter tuning, and potential avenues for further improvement.
Among the models examined during training phase, DenseNet169 demonstrated remarkable accuracy by achieving 99.86% and similarly, EfficientNetB0, EfficientNetB3, MobileNetV2, Xception, ResNet50V2, InceptionV3, and InceptionResNetV2 also exhibited the high accuracy rates of 99.81%, 99.81%, 99.45%, 99.70%, 99.76%, 99.55%, and 99.50% respectively and also surpasses 99% benchmark. Likewise, for validation phase, the top accuracy has been obtained by EfficientNetB3 with 99.96% closely followed by InceptionResNetV2 having accuracy of 99.95%. Similarly, EfficientNetB0, Xception, MobileNetV2, ResNet50V2, and InceptionV3 maintained accuracy levels within the range of 99% to 100%, which signifies their strong generalization capabilities. However, a notable exception was ResNet152V2 and VGG19, which exhibited displayed relatively lower accuracies where their values dropped to 98.49%, 76.97% during the training phase and 97.94%, 74.13% during validation phases, respectively. These results suggest that the layers of the ResNet152V2 and VGG19 model may not have been effectively trained and are struggling to capture essential patterns in both the training as well as validation datasets.
Moving to the loss values, they represent the discrepancy between predicted and actual outputs, indicating how well the models fit the data. Notably, DenseNet169 showcased the lowest loss during both training (0.189) and EfficientNetB3 during validation phase (0.149), which implies a superior fit to the dataset. On the other hand, VGG19 had significantly higher losses, reaching 0.930 during training and 0.961 during validation, indicating poorer model performance in capturing the data patterns.
Furthermore, the RMSE values which assess the average magnitude of errors in our predictions, has been also computed. On assaying, it has been found that EfficientNetB3 demonstrated the lowest RMSE (0.386) during validation, reflecting its ability to make precise predictions on continuous data. In contrast, VGG19 exhibited the highest RMSE values as compared to the other models, both during training (0.964) and validation (0.980), indicating less precise predictions.
Besides this, the accuracy and loss performance of the models are also examined based on their curves that have been generated during both training as well as validation phases in Figure 6.
From the analysis of the training and validation process depicted in the figure, it is evident that epoch 10 yielded the lowest training and validation loss values across all the models, indicating that the models achieved their good performance in terms of minimizing the difference between actual and predicted values at this epoch. This suggests that after 10 epochs of training, the models converged to a state of optimized loss values. However, when examining the accuracy metric, a more diverse pattern emerges. Different models reached their peak accuracy scores at various epochs during the training process. EfficientNetB0 attained its highest accuracy at epoch 2, DenseNet169 at epoch 3, Xception at epoch 7, ResNet50V2 at epoch 5, ResNet152V2 at epoch 6, InceptionV3 at epoch 6, InceptionResNetV2 at epoch 4, VGG19 at epoch 9, MobileNetV2 at epoch 10, and EfficientNetB3 at epoch 5.
After examining the models for the whole dataset, they have been also evaluated for the different classes of the dataset such as Aphids, Healthy Leaves, ArmyWorm, Powdery Mildew, Bacterial Blight, and Target Spot as mentioned in Table 5 (The table provides a comprehensive overview of the results, showing that for the Aphids class, the top performance in training was achieved by DenseNet169, while EfficientNetB0 delivered the best result during validation).
In Table 5, for Aphids class, EfficientNetB0, DenseNet169, Xception, EfficientNetB3, and MobileNet V2 achieved high accuracies during training, with values ranging from 99.19% to 99.49%. However, during validation, their accuracies varied, with EfficientNetB0, Xception, MobileNetV2, ResNet50V2, InceptionV3, EfficientNetB3, and InceptionResNetV2 maintaining accuracies above 99%, while DenseNet169 lower accuracy at 98.16%. However, VGG19 exhibited significantly lower accuracy values of 76.49% and 74.79% during training and validation, respectively, indicating challenges in capturing important features associated with aphids in the dataset. Regarding the loss and RMSE metrics, models such as DenseNet169 and MobileNetV2 demonstrated lower values for training as well as validation phase respectively, while as VGG19 and ResNet152 V2 exhibited higher values which suggest improvement in their prediction.
For ArmyWorm class, all the models except DenseNet169 and VGG19 have computed the accuracies above 97% for training phase, while as during validation phase these same models excluding ResNet152V2 have obtained the accuracies above 98%. On analysing, it has been observed that IncpetionV3 model has the highest validation accuracy of 99.52% while as VGG 19 once again has the lowest accuracy value of 74.49%. Similarly for the loss and root mean square error value, Xception model computed the lowest value of 0.224 and 0.473 respectively during validation phase, and on the contrary, MobileNetV2 obtained the good loss and RMSE score for validation dataset with 0.119 and 0.344 respectively.
Similarly, the models have been also computed for the other classes of the cotton plant disease dataset such as Powdery Mildew, Bacterial Blight, Healthy leaves, and Target Spot using the same metrics. On analyzing the performance of models for these classes, it has been observed that in case of Bacterial Blight, InceptionV3; Target Spot,EfficientNetB0 and InceptionResNetV2; Powdery Mildew,EfficientNetB0 and (MobileNetV2 and InceptionResNetV2); and Healthy Leaves, InceptionResNetV2 and DenseNet169 computed the good accuracies, loss, and RMSE score values respectively. On the other hand, during validation phase the highest accuracy, root mean square error, and loss scores have been generated by EfficientNetB3 for Bacterial Blight, EfficientNetB0 and ResNet50V2V2 for Target Spot, Xception and InceptionResNetV2 for Powdery Mildew, and EfficientNetB0 and InceptionResNetV2for Healthy leaves.
After examining the models such as EfficientNetB0, EfficientNetB3, ResNet50V2, ResNet152V2, InceptionV3, InceptionResNetV2, VGG19, Xception, MobileNetV2, and DenseNet169 for their accuracies, loss, and RMSE scores during training and validation phase, the goal is to evaluate their performance for different set of metrics i.e. recall, precision, and F1 score. Hence, to assess the effectiveness of these models, Figure 7 shows a widely used evaluation tool known as the confusion matrix.
We have generated the confusion matrix of 6x6 i.e. 6 rows and 6 columns which signifies the 6 difference classes of cotton disease plants i.e. Aphids, Powdery Mildew, Army Worm, Target spot, Bacterial Blight, as well as Healthy leaves. The matrix contains information about the number of samples that are classified as belonging to class i but are predicted as class j by the model. The diagonal elements (i.e., the elements with i=j) of the confusion matrix represent the true positive values for each class. These values show how many samples have been correctly categorized for each disease class. Simply put, these instances demonstrate when the model accurately identified a sample as belonging to a specific disease class.
Conversely, the off-diagonal elements, where i ≠ j, correspond to the false positive values. These elements signify the number of misclassifications made by the model, where it predicted a sample as a particular disease class when it belonged to a different one. On the contrary, in case of true negative and false negative, they are calculated by summing up all the samples that are correctly classified for all classes other than the specific class being considered and the instances where samples that belong to a specific class are incorrectly predicted as other classes respectively (Wikipedia, 2024).
Hence, analyzing the diagonal sequence of the confusion matrix allows us to measure the model’s accuracy for each disease class individually. Higher values along the diagonal indicate that the model is performing well in correctly identifying samples for those disease classes, while lower values may indicate areas where the model needs improvement. High performance was observed for most models, but certain models, such as VGG19, showed weaker performance with lower precision and recall values.
Based on the values of confusion matrix, other parameters except accuracy and loss of the applied deep transfer learning models have been computed for the complete cotton plant disease dataset as shown in Table 6.
Looking at the results, we can observe that most of the models such as EfficientNetB0, DenseNet169, Xception, MobileNetV2, ResNet50 V2, ResNet152 V2, Inception V3, EfficientNetB3, and Hybrid (InceptionResNetV2) have achieved excellent performance, with high precision, recall, and F1 scores close to 0.98 or 1.00. It means that these models have demonstrated outstanding capabilities in accurately classifying images into their respective categories, with minimal misclassifications. However, VGG19 seems to lag the other models, with noticeably lower precision, recall, and F1 scores of around 0.78 and 0.74 respectively. This suggests that VGG19 might struggle with certain classes and is relatively less accurate in its predictions on comparing to the other models.
Likewise, the execution of the models has been also assayed for the different classes of the dataset based on the same performance metrics as discussed earlier and their results are graphically represented in Figure 8.
Upon analyzing the pattern of the graph in the figure, a clear trend emerges, indicating that all the models have generally demonstrated excellent precision, recall, and F1 scores for the six different classes of the dataset, with values ranging between 0.92 to 1.00, except for the VGG19 model. The performance of the VGG19 model varies across different classes, particularly for classes with values below 0.90. Albeit, for the Aphids class, the model exhibits a high precision of 0.94 but struggles with recall at 0.52, leading to an F1 score of 0.67. While achieving a respectable recall of 0.85 for the Army Worm class, the precision is relatively lower at 0.76, resulting in an F1 score of 0.80. Similarly, the Bacterial Blight class shows a precision of 0.51, recall of 0.60, and F1 score of 0.55, indicating moderate performance. The model also did not perform well for the Target Spot class with 0.60 as precision, 0.85 as recall of 0.85, and 0.71 as F1 score. For the Powdery Mildew class, the VGG19 model demonstrates a high precision of 0.96 and an average recall of 0.77, resulting in an F1 score of 0.85. Lastly, the Healthy Leaves class shows good precision at 0.88 and recall at 0.83, with an F1 score of 0.85. Overall, while the model achieves high precision in some cases, it struggles with recall for several classes, impacting the F1 score and indicating room for improvement in its performance for certain categories.
Additionally, the computational time frames of various deep learning models have been also computed (Table 7) on applying to the cotton plant disease dataset. As, the system used for this study consists of a standard desktop architecture with an Intel i7 processor, 32GB of RAM, a 4GB GPU, and Windows 11 as the operating system. Hence, the training duration of each model varies depending on these computational requirements and the system’s hardware configuration. It has been found that EfficientNetB0 obtained the longest time of 10 hours and 56 minutes, closely followed by Inception V3 with 7 hours and 48 minutes. On the other, VGG19 took only 2 hours for the computation of the dataset while as moderate timings have been taken by ResNet50V2 and MobileNetV2 with 3 hours 1 minute and 4 hours 40 minute respectively. Apart from this, DenseNet169, Xception, and Hybrid (InceptionResNetV2) required relatively substantial training times which range from 6 to 8 hours. The models with complex architecture require longer computational times as compared to simpler architectures like VGG19.
Additionally, Table 8 presents a comparative analysis of various studies focusing on the detection and classification of diseases affecting cotton plants using machine learning and deep learning techniques. Each study employs different datasets, classes of diseases, techniques, and their outcomes. Firstly, regarding the datasets used, researchers employ both customized datasets specific to cotton plants and publicly available datasets like CIFAR-10 (Canadian Institute for Advanced Research), cotton plant disease dataset (Rajasekar et al., 2021). The classes of diseases vary across studies, ranging from common issues like boll rot and fungal leafspot diseases to more specific diseases such as Cotton leaf curl virus and fusarium wilt. In terms of techniques, convolutional neural networks (CNN) are the most used, with variants like Tflite model, CoreML model (Kumar et al., 2022), and custom architectures like CottonLeafNet (Singh et al., 2023) and CoNet (Thivya Lakshmi et al., 2024) proposed in some studies. Additionally, models like ResNet50, InceptionVGG16, and DenseNet121 are also employed which showcase the diversity in model architectures. The outcomes, measured primarily in terms of accuracy, vary across studies i.e. from 90% to 99.96%. Notably, the highest accuracy is achieved by the current technique used in this paper i.e. the EfficientNetB3 architecture on a dataset encompassing Army Worm, Aphids, Target spot, Powdery Mildew, and Bacterial Blight.
4 Discussion
The potential of deep learning techniques for detecting and classifying cotton plant diseases presents a promising avenue for addressing agricultural challenges on a global scale. However, realizing this potential requires careful consideration of various feasibility factors. Access to high-quality, labelled datasets is paramount, as it forms the foundation for training effective models, although obtaining such data may prove challenging, particularly for rare or localized diseases (Malar et al., 2021). Moreover, while deep learning models have demonstrated remarkable performance in image classification tasks, hence, to ensure their ability in detecting subtle symptoms and adapt to diverse environmental conditions remains critical. This necessitates robust model design and optimization strategies to enhance performance across different agricultural settings. Additionally, deep learning provides detailed insights into the severity as well as location of diseases within a field. By precisely mapping the distribution of diseases, farmers can adopt site-specific management practices, such as adjusted irrigation, targeted pesticide application, and optimized resource allocation. Moreover, the automation inherent as well as scalability of the deep learning models can also enable to efficiently monitor large agricultural areas using platforms like drones or satellites (Annabel et al., 2019). However, challenges such as the availability of diverse and representative datasets, the generalization of models to different environmental conditions, and the seamless integration of deep learning solutions into existing agricultural workflows require careful consideration and ongoing research to fully harness the potential of this technology in combating cotton plant diseases.
Apart from this, there are also several new improvements that can be done to enhance the current research and its practical applications. These improvements include (Askr et al., 2024; Woźniak and Ijaz, 2024):
● To enhance the performance of deep learning techniques for identifying and classifying cotton plant diseases, advanced optimization techniques such as learning rate scheduling, early stopping, and regularization methods should be fine-tuned.
● Integration of Multi-Modal Data where incorporation of additional data modalities such as spectral imaging, hyperspectral imaging, or thermal imaging alongside visual images can provide complementary information for disease detection. Fusion of multi-modal data can enhance the model’s ability to capture subtle disease symptoms and improve overall diagnostic accuracy.
● Explainable AI Techniques such as attention mechanisms, saliency maps, or feature visualization methods can help to interpret the decisions made by deep learning models. In fact, providing explanations for model predictions can increase trust and transparency in the disease detection system to facilitate better decision-making by end-users.
● Active Learning Strategies can optimize the data labelling process by selecting the most informative samples for annotation. By iteratively training the model on a small set of labelled data and actively acquiring labels for the most uncertain samples, the efficiency of the disease detection system can be enhanced while reducing labelling costs.
● Those models should be developed that are robust to environmental variability, such as changes in lighting conditions, camera angles, or plant growth stages, is crucial for real-world deployment. Adapting the models to diverse environmental conditions through data augmentation and domain adaptation techniques can ensure consistent performance in field settings.
● Design of scalable and deployable solutions that can be easily integrated into existing agricultural systems is essential for widespread adoption. Developing lightweight models optimized for edge devices, cloud-based solutions for centralized monitoring, and user-friendly interfaces for farmers can facilitate the practical implementation of disease detection technologies.
However, in the future work, there is significant scope to expand the current research on disease detection in cotton plants using deep learning techniques. It includes (Naga et al., 2024; Woźniak and Ijaz, 2024):
● Extension of the current applied models to handle a broader range of diseases and abnormalities that affect cotton plants. By incorporating additional classes into the classification system, the AI based system can provide more comprehensive insights for farmers and agronomists.
● To explore advanced transfer learning strategies to leverage pre-trained models effectively. Fine-tuning existing models or combining multiple models through ensemble techniques could enhance the overall performance and robustness of the disease detection system.
● Implementing more advanced data augmentation methods for increasing the size and diversity of the training dataset as it improves the ability of the model to work on unseen data with a good accuracy.
● To incorporate techniques such as model visualization or attention mechanisms which can help users in understanding how the model predicts the output?
● Developing a real-time monitoring system by integrating IoT and cloud technique to provide alerts to the farmers on continuously analysing the images of cotton plants so that they will be able to take decision promptly in saving the cotton plants.
● Building mobile applications for end user interaction where farmers can upload images of diseased plants to analyse and review recommendations for treatment. By this way, the system can adapt to local conditions and improve the accuracy of detecting and classifying cotton diseases on time.
Hence, in a nutshell it can be said that by incorporating these new improvements and to explore these avenues for future work into the research on disease detection in cotton plants, the field can advance towards more accurate, interpretable and user-centric solutions that address the challenges faced in sustainable agriculture and contribute to increased crop productivity and food security.
5 Conclusion
The research has shown the capability of artificial intelligence-based learning techniques to detect diseases in the leaves of cotton plants. The paper highlights the application of deep transfer learning approaches for their effective role in identifying and well as classifying various cotton plant diseases such as Target Spot, Bacterial Blight, Aphids, Army Worm, and Powdery Mildew. During the training of the models, it is recommended to prioritize EfficientNetB3 due to its superior performance as it achieved the highest accuracy of 99.96%. However, depending on specific needs such as computational constraints or model complexity, other models like MobileNetV2 or ResNet50V2 could also be considered, as they offer high accuracy with potentially lower computational overhead.
The paper, while showing promising results in detecting various diseases in cotton plants, also faces several limitations that need to be addressed. One major issue is the accurate generation of the region of interest (ROI) in the images. In plant disease detection, it is crucial to identify the specific areas affected by disease. If the ROI is not accurately detected, the model may analyse irrelevant parts of the image which will lead to misclassification and inaccurate predictions. Another limitation is the tendency of models to modelling error, particularly when training on fewer non-diverse datasets. Additionally, the computational complexity of deep learning models used in the paper is high, requiring substantial computational resources for both training and validation. This can be a bottleneck, especially when working with diverse datasets or deploying the model in resource-constrained environments. To mitigate these challenges, the use of optimization techniques, such as fine-tuning model parameters, is necessary to improve performance and reduce the risk of modeling errors. These limitations, if unaddressed, could impact the generalizability and robustness of the results, making it essential to refine the models and techniques used. Apart from this, looking ahead, the future scope of this research lies in enhancing the interpretability of deep learning models, exploring ensemble learning techniques to boost performance, and integrating real-time monitoring systems for proactive disease management in agriculture. By dealing with these challenges and incorporating new techniques, the field of disease detection in cotton plants can evolve to facilitate more efficient and sustainable agricultural practices. This proactive approach not only benefits crop yield and quality but also contributes to the overall resilience and productivity of agricultural systems.
Data availability statement
Publicly available datasets were analyzed in this study. This data can be found here: https://www.kaggle.com/datasets/dhamur/cotton-plant-disease.
Author contributions
PJ: Writing – original draft, Writing – review & editing. KD: Writing – original draft, Writing – review & editing. PS: Writing – original draft, Writing – review & editing. BK: Writing – original draft, Writing – review & editing. SK: Writing – original draft, Writing – review & editing. YK: Writing – original draft, Writing – review & editing. JS: Writing – original draft, Writing – review & editing. MI: Writing – original draft, Writing – review & editing.
Funding
The author(s) declare that financial support was received for the research, authorship, and/or publication of this article. The NRF grant funded by the Korean government (MSIT) made this work possible. (NRF-2023R1A2C1005950). Funding for this research is provided by Prince Sattam bin Abdulaziz University under project number (PSAU/2024/R/1445).
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
The author(s) declared that they were an editorial board member of Frontiers, at the time of submission. This had no impact on the peer review process and the final decision.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
Ahad, M. T., Li, Y., Song, B., Bhuiyan, T. (2023). Comparison of CNN-based deep learning architectures for rice diseases classification. Artif. Intell. Agric. 9, 22–35. doi: 10.1016/j.aiia.2023.07.001
Akbar, W., Soomro, A., Ullah, M., Haq, M. I. U., Khan, S. U., Shah, T. A. (2023). “Performance evaluation of deep learning models for leaf disease detection: A comparative study,” in 2023 4th International Conference on Computing, Mathematics and Engineering Technologies (iCoMET). (Piscataway, New Jersey, USA: IEEE), 01–05.
Annabel, L. S. P., Annapoorani, T., Deepalakshmi, P. (2019). “Machine learning for plant leaf disease detection and classification–A review,” in 2019 International Conference on Communication and Signal Processing (ICCSP). (Piscataway, New Jersey, USA: IEEE), 0538–0542.
Apple Developer (2024). Core ML updates. Apple Developer Documentation. Version 8, June 2024. Available online at: https://developer.apple.com/documentation/updates/coreml (accessed September 10, 2024).
Arathi, B., Dulhare, U. N. (2023). “Classification of cotton leaf diseases using transfer learning-denseNet-121,” in Proceedings of Third International Conference on Advances in Computer Engineering and Communication Systems: ICACECS 2022, Singapore. (Singapore: Springer Nature Singapore), 393–405.
Arthy, R., Ram, S. S., Yakubshah, S. G. M., Vinodhan, M., Vishal, M. (2024). Cotton guard: revolutionizing agriculture with smart disease management for enhanced productivity and sustainability. Int. Res. J. Advanced Eng. Hub (IRJAEH) 2, 236–241. doi: 10.47392/IRJAEH.2024.0038
Askr, H., El-dosuky, M., Darwish, A., Hassanien, A. E. (2024). Explainable ResNet50 learning model based on copula entropy for cotton plant disease prediction. Appl. Soft Computing 164, 112009. doi: 10.1016/j.asoc.2024.112009
Babu, J. M., Jeyakrishnan, V. (2022). Plant leaves disease detection (Efficient net). Int. J. Modern Developments Eng. Sci. 1, 29–31. Available online at: https://journal.ijmdes.com/ijmdes/article/view/66.
Bera, S., Kumar, P. (2023). “Deep learning approach in cotton plant disease detection: A survey,” in 2023 2nd International Conference for Innovation in Technology (INOCON). (Piscataway, New Jersey, USA: IEEE), 1–7.
Bhatti, M. M. A., Umar, U. U. D., Naqvi, S. A. H. (2020). Cotton diseases and disorders under changing climate. Cotton Production Uses: Agronomy Crop Protection Postharvest Technol., 271–282. doi: 10.1007/978-981-15-1472-2_14
Caldeira, R. F., Santiago, W. E., Teruel, B. (2021). Identification of cotton leaf lesions using deep learning techniques. Sensors 21, 3169. doi: 10.3390/s21093169
Dhamodharan, R. (2023). Cotton plant disease. Available online at: https://www.kaggle.com/datasets/dhamur/cotton-plant-disease (accessed September 4, 2024).
Elaraby, A., Hamdy, W., Alruwaili, M. (2022). Optimization of deep learning model for plant disease detection using particle swarm optimizer. Computers Materials Continua 71 (2), 1–13. doi: 10.32604/cmc.2022.022161
Ferentinos, K. P. (2018). Deep learning models for plant disease detection and diagnosis. Comput. Electron. Agric. 145, 311–318. doi: 10.1016/j.compag.2018.01.009
Google AI Edge (2024). LiteRT overview. Google AI for Developers, Version 2.14, November 20, 2024. Available online at: https://ai.google.dev/edge/litert (accessed December 8, 2024).
Gülmez, B. (2023). A novel deep learning model with the Grey Wolf Optimization algorithm for cotton disease detection. J. Universal Comput. Sci. 29, 595. doi: 10.3897/jucs.94183
He, K., Zhang, X., Ren, S., Sun, J. (2016). “Deep residual learning for image recognition,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (Piscataway, New Jersey, USA: IEEE). 770–778.
Huang, G., Liu, Z., van der Maaten, L., Weinberger, K. Q. (2017). “Densely connected convolutional networks,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (Piscataway, New Jersey, USA: IEEE), 4700–4708.
Hyder, U., Talpur, M. R. H. (2024). Detection of cotton leaf disease with machine learning model. Turkish J. Eng. 8, 380–393. doi: 10.31127/tuje.1406755
Kalaiselvi, T., Narmatha, V. (2023). “Cotton crop disease detection using FRCM segmentation and convolution neural network classifier,” in Computational Vision and Bio-Inspired Computing: Proceedings of ICCVBIC 2022, Singapore. 557–577 (New York, USA: Springer Nature Singapore).
Kanna, G. P., Kumar, S. J., Kumar, Y., Changela, A., Woźniak, M., Shafi, J., et al. (2023). Advanced deep learning techniques for early disease prediction in cauliflower plants. Sci. Rep. 13, 18475. doi: 10.1038/s41598-023-45403-w
Kinger, S., Tagalpallewar, A., George, R. R., Hambarde, K., Sonawane, P. (2022). “Deep learning based cotton leaf disease detection,” in 2022 International Conference on Trends in Quantum Computing and Emerging Business Technologies (TQCEBT). 1–10 (IEEE).
Kotian, S., Ettam, P., Kharche, S., Saravanan, K., Ashokkumar, K. (2022). “Cotton leaf disease detection using machine learning,” in Proceedings of the Advancement in Electronics & Communication Engineering. (Singapore: Springer), 118–124.
Kukadiya, H., Arora, N., Meva, D., Srivastava, S. (2024). An ensemble deep learning model for automatic classification of cotton leaves diseases. Indonesian J. Electrical Eng. Comput. Sci. 33, 1942–1949. doi: 10.11591/ijeecs.v33.i3
Kumar, S., Ratan, R., Desai, J. V. (2022). Cotton disease detection using tensorflow machine learning technique. Adv. Multimedia 2022, 1812025. doi: 10.1155/2022/1812025
Kumar, Y., Singh, R., Moudgil, M. R., Kamini (2023). A systematic review of different categories of plant disease detection using deep learning-based approaches. Arch. Comput. Methods Eng. 30, 1–23. doi: 10.1007/s11831-023-09958-1
Latif, M. R., Khan, M. A., Javed, M. Y., Masood, H., Tariq, U., Nam, Y., et al. (2021). Cotton leaf diseases recognition using deep learning and genetic algorithm. Computers Materials Continua 69 (3), 1–16. doi: 10.32604/cmc.2021.017364
Malar, B. A., Andrushia, A. D., Neebha, T. M. (2021). “Deep learning based disease detection in tomatoes,” in 2021 3rd International Conference on Signal Processing and Communication (ICPSC). (Piscataway, New Jersey, USA: IEEE), 388–392.
Mary, X. A., Raimond, K., Raj, A. P. W., Johnson, I., Popov, V., Vijay, S. J. (2022). “Comparative analysis of deep learning models for cotton leaf disease detection,” in Disruptive Technologies for Big Data and Cloud Applications: Proceedings of ICBDCC 2021, Singapore. 825–834 (Singapore: Springer Nature Singapore).
Memon, M. S., Kumar, P., Iqbal, R. (2022). Meta deep learn leaf disease identification model for cotton crop. Computers 11, 102. doi: 10.3390/computers11070102
Mohmmad, S., Varshitha, G., Yadlaxmi, E., Nikitha, B. S., Anurag, D., Shivakumar, C., et al. (2024). “Detection and Classification of Various Diseases in Cotton Crops using advanced Neural Network Approaches,” in 2024 Fourth International Conference on Advances in Electrical, Computing, Communication and Sustainable Technologies (ICAECT). (Piscataway, New Jersey, USA: IEEE), 1–6.
Montalbo, F. J. P. (2022). Automated diagnosis of diverse coffee leaf images through a stage-wise aggregated triple deep convolutional neural network. Mach. Vision Appl. 33, 19. doi: 10.1007/s00138-022-01277-y
Naeem, A. B., Senapati, B., Chauhan, A. S., Kumar, S., Gavilan, J. C. O., Abdel-Rehim, W. M. (2023). Deep learning models for cotton leaf disease detection with VGG-16. Int. J. Intelligent Syst. Appl. Eng. 11 (2), 550–556. Available online at: https://www.ijisae.org/index.php/IJISAE/article/view/2710.
Naga Srinivasu, P., Ijaz, M. F., Woźniak, M. (2024). XAI-driven model for crop recommender system for use in precision agriculture. Comput. Intell. 40 (1), e12629. doi: 10.1111/coin.12629
Nalini, T., Rama, A. (2022). Impact of temperature condition in crop disease analyzing using machine learning algorithm. Measurement: Sensors 24, 100408. doi: 10.1016/j.measen.2022.100408
Odukoya, O. H., Aina, S., Dégbéssé, F. W. (2023). A model for detecting fungal diseases in cotton cultivation using segmentation and machine learning approaches. Int. J. Advanced Comput. Sci. Appl. 14 (1), 1–10. doi: 10.14569/IJACSA.2023.0140169
Peyal, H. I., Pramanik, M. A. H., Nahiduzzaman, M., Goswami, P., Mahapatra, U., Atusi, J. J. (2022). “Cotton leaf disease detection and classification using lightweight CNN architecture,” in 2022 12th International Conference on Electrical and Computer Engineering (ICECE). (Piscataway, New Jersey, USA: IEEE), 413–416.
Rai, C. K., Pahuja, R. (2024). An ensemble transfer learning-based deep convolution neural network for the detection and classification of diseased cotton leaves and plants. Multimedia Tools Appl. 83, 1–34. doi: 10.1007/s11042-024-18963-w
Rajasekar, V., Venu, K., Jena, S. R., Varthini, R. J., Ishwarya, S. (2021). Detection of cotton plant diseases using deep transfer learning. J. Mobile Multimedia 18, 307–324. doi: 10.13052/jmm1550-4646.1828
Sadiq, S., Malik, K. R., Ali, W., Iqbal, M. M. (2023). Deep learning-based disease identification and classification in potato leaves. J. Computing Biomed. Inf. 5 (01), 13–25. Available online at: https://www.jcbi.org/index.php/Main/article/view/155.
Simonyan, K., Zisserman, A. (2014). Very deep convolutional networks for large-scale image recognition. arXiv preprint 6, 1–14. arXiv:1409.1556. doi: 10.48550/arXiv.1409.1556
Singh, P., Singh, P., Farooq, U., Khurana, S. S., Verma, J. K., Kumar, M. (2023). CottonLeafNet: cotton plant leaf disease detection using deep neural networks. Multimedia Tools Appl. 82, 37151–37176. doi: 10.1007/s11042-023-14954-5
Sun, X., Li, G., Qu, P., Xie, X., Pan, X., Zhang, W. (2022). Research on plant disease identification based on CNN. Cogn. Robotics 2, 155–163. doi: 10.1016/j.cogr.2022.07.001
Tan, M. (2019). Efficientnet: Rethinking model scaling for convolutional neural networks. arXiv preprint 97, 1–10 arXiv:1905.11946. Available online at: https://proceedings.mlr.press/v97/tan19a.html?ref=jina-ai-gmbh.ghost.io.
Thivya Lakshmi, R. T., Katiravan, J., Visu, P. (2024). CoDet: A novel deep learning pipeline for cotton plant detection and disease identification. Automatika 65, 662–674. doi: 10.1080/00051144.2024.2317093
Verma, R., Singh, S. (2022). “Plant leaf disease classification using MobileNetV2,” in Emerging Trends in IoT and Computing Technologies (Abingdon, Oxfordshire, United Kingdom: Routledge), 249–255.
Wikipedia (2024). Confusion matrix.November 5, 2024. Available online at: https://en.wikipedia.org/wiki/Confusion_matrix (accessed December 8, 2024).
Woźniak, M., Ijaz, M. F. (2024). Recent advances in big data, machine, and deep learning for precision agriculture. Front. Plant Sci. 15, 1367538. doi: 10.3389/fpls.2024.1367538
Keywords: cotton disease, agriculture, deep learning, bacterial blight, powdery mildew, contour features
Citation: Johri P, Kim S, Dixit K, Sharma P, Kakkar B, Kumar Y, Shafi J and Ijaz MF (2024) Advanced deep transfer learning techniques for efficient detection of cotton plant diseases. Front. Plant Sci. 15:1441117. doi: 10.3389/fpls.2024.1441117
Received: 30 May 2024; Accepted: 26 November 2024;
Published: 19 December 2024.
Edited by:
Julie A. Dickerson, Iowa State University, United StatesReviewed by:
Lingxian Zhang, China Agricultural University, ChinaMa Khan, HITEC University, Pakistan
Yuriy L. Orlov, I.M.Sechenov First Moscow State Medical University, Russia
Copyright © 2024 Johri, Kim, Dixit, Sharma, Kakkar, Kumar, Shafi and Ijaz. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Yogesh Kumar, eW9nZXNoYXJvcmExMDc0NEBnbWFpbC5jb20=; Muhammad Fazal Ijaz, bWZhemFsQG1pdC5lZHUuYXU=
†These authors have contributed equally to this work and share first authorship