Skip to main content

ORIGINAL RESEARCH article

Front. Bioeng. Biotechnol., 27 April 2022
Sec. Bionics and Biomimetics
This article is part of the Research Topic Bio-Inspired Computation and Its Applications View all 72 articles

Image Classification and Recognition of Rice Diseases: A Hybrid DBN and Particle Swarm Optimization Algorithm

  • 1College of Information and Electrical Engineering, Heilongjiang Bayi Agricultural University, Daqing, China
  • 2School of Physics and Electronic Engineering, Northeast Petroleum University, Daqing, China
  • 3Key Laboratory for Metallurgical Equipment and Control of Ministry of Education, Wuhan University of Science and Technology, Wuhan, China

Rice blast, rice sheath blight, and rice brown spot have become the most popular diseases in the cold areas of northern China. In order to further improve the accuracy and efficiency of rice disease diagnosis, a framework for automatic classification and recognition of rice diseases is proposed in this study. First, we constructed a training and testing data set including 1,500 images of rice blast, 1,500 images of rice sheath blight, and 1,500 images of rice brown spot, and 1,100 healthy images were collected from the rice experimental field. Second, the deep belief network (DBN) model is designed to include 15 hidden restricted Boltzmann machine layers and a support vector machine (SVM) optimized with switching particle swarm (SPSO). It is noted that the developed DBN and SPSO-SVM can simultaneously learn three proposed features including color, texture, and shape to recognize the disease type from the region of interest obtained by preprocessing the disease images. The proposed model leads to a hit rate of 91.37%, accuracy of 94.03%, and a false measurement rate of 8.63%, with the 10-fold cross-validation strategy. The value of the area under the receiver operating characteristic curve (AUC) is 0.97, whose accuracy is much higher than that of the conventional machine learning model. The simulation results show that the DBN and SPSO-SVM models can effectively extract the image features of rice diseases during recognition, and have good anti-interference and robustness.

1 Introduction

Rice is one of the most important food crops for our people. In recent years, the yield of the rice planting industry has increased day by day, but there are still some adverse factors affecting its yield, such as rice diseases. At present, rice blast, rice sheath blight, and rice brown spots have become the most common diseases in the cold areas of northern China. It can occur during the whole growth period of rice, which has a great impact on the yield and quality of rice, and even lead to no yield in serious cases. This will not only affect the income of farmers but also have a bad impact on China’s food security and fiscal revenue. At present, the identification methods of rice diseases mainly rely on farmers’ experience to judge, consult disease books or query on the Internet, consult agricultural technicians, or seek help from plant experts. However, the recognition efficiency of human eyes is low, and the recognition is highly subjective. There may be misjudgment, which makes it difficult to realize the accurate identification of diseases. The update speed of disease books is slow, the amount of information on the Internet is large, and the accuracy cannot be guaranteed. By consulting agricultural technicians and plant experts, a lot of material and financial resources will be consumed, a lot of time and cost will be spent, and it is easy to miss the best opportunity for disease treatment. Therefore, finding a new, efficient, and accurate rice disease recognition method to replace the traditional human eye recognition method is the key to solving the problem of rice disease recognition, which has important research significance.

However, so far, no research has involved the integration of the DBN and SPSO-SVM, because the DBN and SPSO-SVM have great potential application in rice disease identification. Based on the shortcomings and limitations of traditional identification methods, we proposed a method based on the DBN and SPSO-SVM to identify rice blast, sheath blight, and brown spot. Using the images in the rice disease database, the DBN and SPSO-SVM models are trained in two steps. First, a DBN model is constructed using RBMs. Second, the gradient descent algorithm is used to train the DBN and SVM optimized by SPSO as a classifier. The DBN and SPSO-SVM are then evaluated by 10-fold cross-validation.

The main innovations of this study are as follows: a novel framework for automatic diagnosis of rice diseases is given to settle the problems in image segmentation of rice disease spot, where the DBN and SPSO-SVM method are applied to accurately extract the edge of rice disease spot; by learning disease spot features, the proposed DBN and SPSO-SVM can identify the region of interest that is obtained by preprocessing the rice disease images. Experimental results show that this method has high accuracy in feature parameters, segmentation, and peak signal-to-noise ratio. This research replaces the traditional rice disease diagnosis method, enhances the in-depth integration of the rice disease identification field and machine learning technology, combines the DBN algorithm with the SPSO-SVM algorithm for rice disease identification for the first time, effectively improves the accuracy and timeliness of rice disease control, and has practical significance.

The rest of the study is given as follows: Section 2 introduces the relevant research status at home and abroad; Section 3 collects and extracts the characteristics of rice diseases; Section 4 describes the architecture of the DBN and SPSO-SVM framework; Section 5 applies the developed DBNs and SPSO-SVM to the identification of three diseases in rice and demonstrates the experimental results; and conclusions are given in Section 6.

2 Related Work

In recent years, with the rapid development of digital image processing, computer vision, and pattern recognition technology, new ways and methods have been provided for rapid and accurate diagnosing, analyzing, classifying, and nondestructive detection of plant diseases (see Zhang D. et al., 2018; Kaur et al., 2019; Li Y. et al., 2019; Fuentes et al., 2020; Li et al., 2020; Madhulatha and Ramadevi, 2020; Nanehkaran et al., 2020; Sethy et al., 2020; Wspanialy and Moussa, 2020; Li and Chao, 2021, and the references therein). These intelligent diagnostic methods have achieved great success in both theoretical research and practical application. However, due to the limitation of theoretical analysis and the need for specific skills and a lot of experience knowledge in the course of training, the effect of diagnosis is not optimal in many cases.

A deep belief network (DBN) has been proposed by Hinton and Salakhutdinov (2006) at the University of Toronto in 2006, and the relevant results have been reported in the study by Hinton (2012), Dong et al. (2015b), Russakovsky et al. (2015), and Dong et al. (2016b). During this period, the DBN is a probabilistic generation model, which consists of multiple restricted Botlzmann machines (RBMs). It can extract features from the original data by stacking the RBMs layer by layer; then we can obtain some high-level representations of the original data. The core of this deep learning model is to optimize the connection weight of the deep neural network by the layer-by-layer greedy learning algorithm. First, in order to mine the disease features of the rice, unsupervised layer-by-layer training is used. Then, by adding the corresponding classifier and reversing the supervised fine-tuning, the disease diagnosis ability of the DBN is optimized.

The switching PSO algorithm has been proposed by (Tang et al., 2011), in which a mode-dependent velocity updating Eq. 7 is introduced to realize a balance between the local search algorithm and the global one. Lu et al. (2015) have employed the SPSO algorithm to solve the optimization problem with constraints by converting it into an optimization problem without constraints and have gained a better performance.

In the last few years, lots of researchers have applied the DBN in many application fields such as image recognition, speech recognition, face recognition, and plant leaf classification (Cheng et al., 2018; Goudarzi et al., 2018; Liu et al., 2018; Cristin et al., 2020; Vedantham and Reddy, 2020; Veisi and Mani, 2020; Chen and Pan, 2021). It is a multi-layer network structure, which provides a basic model for feature extraction and subsequent image recognition of original images. The DBN has become a hot research topic in the field of image recognition.

As for the identification and classification of plant leaf diseases, a convolution neural network has been introduced by some researchers. For example, Rahman and Arko et al. (Rahman et al., 2020) proposed a two-stage small CNN architecture, which effectively avoids the disadvantages that the large model architecture is not suitable for mobile devices, reduces the model size while ensuring accuracy, and is more widely used in practical production. Zhang and Qiao et al. (Zhang X. et al., 2018) improved GoogLeNet and Cifar10 models based on deep learning, which are proposed to identify corn leaf diseases. The experimental results show that the improved method reduces the number of convergence iterations and effectively improves the accuracy of corn leaf diseases. Hassan and Maji et al. (Hassan et al., 2021) have recognized plant leaf diseases using the CNN and transfer learning methods. Sharma and Berwal et al. (Sharma et al., 2020) have used the S-CNN model trained by segmented images to detect plant diseases. Hu and Yang et al. (Hu et al., 2019) have used an improved deep convolutional neural network model to identify tea leaf diseases. Atila and Uçar et al. (Atila et al., 2021) have used an efficient deep learning model classification of plant leaf diseases. Therefore, we propose to combine the DBN and SPSO-SVM for the identification of rice blast, sheath blight, and brown spot, so as to improve the identification efficiency and diagnostic accuracy.

3 Data Acquisition and Feature Extraction

3.1 Data Collection and Processing

Under natural light conditions, using iPhone seven Plus mobile phone (IOS 10.1 operating system, F/1.8 wide-angle, and F/2.8 telefocal dual-lens camera, 12 million pixels), the rice disease images are captured in the rice experimental field of Heilongjiang Academy of Agricultural Reclamation Sciences, China. The mobile phone is fixed on a bracket, using automatic exposure mode, taking a picture at a moderate distance from the diseased rice leaves. The photo size taken by the phone is 3,024 × 4,032 pixels. The images of three rice diseases were obtained in the early, middle, and late stages, respectively. In each cycle, 500 images were collected for each of the three rice disease images. Therefore, a total of 5,600 sample images were collected, including 1,500 images of rice blast, sheath blight, and brown spot, of which 1,100 were uninfected. Each image is stored in the PNG format and compressed into a uniform size of 400, ×, 300. An example of the original rice disease image sample is shown in Figure 1; 70% of the samples are randomly selected from the data set as a training set, and the remaining 30% of the samples are chosen as a testing set. The data set division is shown in Figure 2.

FIGURE 1
www.frontiersin.org

FIGURE 1. Rice diseases and health image. (A) Rice blast image, (B) sheath blight image, (C) brown spot image, and (D) healthy rice image.

FIGURE 2
www.frontiersin.org

FIGURE 2. Data set partition.

3.2 Feature Extraction

The feature extraction of the disease spot image is a key step in rice disease recognition. According to the spot characteristics of rice sheath blight (taking rice sheath blight as an example), three features of color, texture, and shape are extracted from the spot image.

3.2.1 Color Feature Extraction

Color feature is not affected by the direction, size, and angle of the image. It is a typical visual feature in the image and is widely used in the field of image recognition. At present, some researchers have achieved good results in this field. Peihua Shi used machine learning to estimate rice nitrogen nutrition through RGB images. In this study, the color eigenvalues of rice diseases are extracted from RGB, HSV, and lab color spaces. First, the rice disease image containing disease spots is transformed from RGB color space to HSV color space, and the mean, variance, and energy of the color set in HSV space are obtained as part of the color characteristics. Then, the third-order color moment corresponding to each component in HSV, RGB, and lab space is obtained as another part of the color feature. The color moment is calculated as follows:

1) First-order color moment (mean): reflects the average brightness of the image. It is defined as

μ=1Nj=1Npi,j.(1)

2) Second-order color moment (variance): reflects the discreteness of image gray level distribution. It is defined as

σi=1Nj=1Npi,jμi212.(2)

3) Third-order color moment: defines the skewness of color component and color asymmetry. It is defined as

si=1Nj=1Npi,jμi313(3)

where pi,j represents the ith color component of the jth pixel of the color image, and N represents the number of pixels in the image.

3.2.2 Texture Feature Extraction

Texture feature can reflect the organization and arrangement of the components of things. It is an important visual feature in image description. It can calculate the statistical features of multiple pixel regions and has rotation invariance. As the most widely used texture feature extraction method, the gray level co-occurrence matrix method was proposed by Haralick et al. in 1973. It can effectively describe the change information of image in a certain distance and direction. Therefore, the gray level co-occurrence matrix is used to extract texture features in this study. The texture feature extraction effect is shown in Figure 3. The gray level co-occurrence matrix can be expressed by the following formula:

Pa1,a2=#x1,y1,x2,y2Sfx1,y1=a1&fx2,y2=a2#S(4)

FIGURE 3
www.frontiersin.org

FIGURE 3. Texture feature extraction disease spot.

In the formula, f (x, y) represents the image itself, S represents the set of pixel pairs with some spatial relationship in the image, the expression on the numerator on the right side of the equation represents the number of pixel pairs with some spatial relationship and gray value of a1, a2, and the expression on the denominator represents the number of total pixel pairs (# represents the number of statistical elements).

The common feature parameters extracted according to the gray level co-occurrence matrix are as follows:

1) Energy or angular second-order moment: it is a measure of the uniformity of image gray distribution and the thickness of texture.

ASM=i=1Nj=1NPi,j2(5)

2) Entropy: it reflects the nonuniformity or complexity of image texture. It is a measure of the amount of texture information in the image.

ENT=i=1Nj=1NPi,jlogPi,j(6)

3) Contrast: it reflects the clarity of the image and measures the depth of the image texture groove.

CON=i=1Mj=1Nij2pi,j(7)

4) Inverse differential moment: it reflects the clarity and regularity of image texture and the local change of image.

IDM=i=1Mj=1MPi,j1+ij2(8)

3.2.3 Shape Feature Extraction

In this study, on the basis of selecting the conventional morphological features, such as lesion area, lesion circularity, lesion complexity, and the number of lesions, two new morphological features, namely, the number of lesions and the ratio of lesion area to the number of lesions, are proposed. The calculation formula of lesion area s is

S=x=1ny=1mfx,y(9)

where f (x, y) is a binary image function. Circularity C represents the deviation degree between the shape of the lesion area and the circle, and the calculation formula is

C=dmaxdmin(10)

where dmaxdmin is the maximum/minimum diameter of the cross section of the lesion. Lesion complexity E describes the perimeter per unit area of the lesion area, and the calculation formula is

E=4ΠC(11)

Number of disease spots N represents the number of rice disease spots on a diseased leaf. The ratio of lesion area to lesion number is

R=SN(12)

The effect of rice disease feature extraction is shown in Figure 4.

FIGURE 4
www.frontiersin.org

FIGURE 4. Shape feature extraction of rice disease spot.

4 Architecture of DBN and SPSO-SVM Framework

4.1 Restricted Boltzmann Machine

In Figure 5, an energy model called restricted Boltzmann machine (RBM) is shown. It is a special kind of stochastic neural network model that owns a two-layer architecture. The input layer is also called the visible layer, in which visible units represent observations. The other layer is called the hidden layer, in which hidden units represent features. Visible variables S are linked to hidden variables D through undirected weighted connections W. There are no connections in the same layer. This constitutes a bipartite structure graph. The parameters are introduced as follows:

FIGURE 5
www.frontiersin.org

FIGURE 5. Schematic diagram of the RBM.

S=(s1,s2,,sm)Rm, where m is the neurons’ number in the visible layer.

D=(d1,d2,,dn)Rn, where n is the neurons’ number in the hidden layer.

A=(a1,a2,,am)Rm represents bias in the visible layer.

B=(b1,b2,,bn)Rn represents bias in the hidden layer.

W=(wi,j)Rn×m stands for the weight matrix that connects the visible layer and the hidden layer. wi,j is the weight that connects the jth neuron in the hidden layer and the ith neuron in the visible layer.

ϱ = (W, A, B) represents the unknown model parameter.

RBM is an energy based model. Assuming that the variables of both visible and hidden layers obey Bernoulli distribution, the joint probability distribution of visible and hidden variables is given by energy function. For a given state (s, d), an energy function can be defined as follows (Hinton and Salakhutdinov, 2006):

Eϱs,d=i=1maisij=1nbjdji=1mj1ndjwj,isi.(13)

Then, the joint probability distribution of s, d is given as follows:

Pϱs,d=1ZϱeEϱs,d(14)

where Zϱ=s,deEϱ(s,d) represents the normalization factor. Afterward, the marginal probabilities with which the model assigns to a visible vector s and a hidden vector d are given by

Pϱs=dPϱs,d=1ZϱdeEϱs,d(15)
Pϱd=sPϱs,d=1ZϱseEϱs,d.(16)

Combined with the joint probability distribution, marginal probability distribution, and sigmoid function, we can get the probability of activation of a neuron on the hidden layer when given the visible layer state is given as follows:

Pdτ=1|s=sigmoidbτ+i=1mwτ,isi(17)

The probability of activation of a neuron on the visible layer when given the hidden layer state is given as follows:

Psτ=1|d=sigmoidaτ+i=1nwj,τdj(18)

where sigmoid(x)=(1+ex)1.

Given training set X = s1, s2, … , sl, Si=(s1i,s2i,,smi)T, where l is the number of training samples. Training the RBM is to adjust the parameters ϱ = (w, a, b) so that the probability distribution represented by the RBM is consistent with the training data as far as possible.

The parameters ϱ = (w, a, b) can be solved by maximizing the likelihood function of visual layer data as follows:

Pϱs=1ZϱdexpsTwd+aTd+bTs.(19)

Considering the contrastive divergence (CD) algorithm proposed by Hinton (Hinton, 2002), the conditional probability distribution of the hidden layer and visual layer is given as follows:

Pd|s=jPdj|s,Pdj=1|s=11+expiwijsiaj(20)
Ps|d=iPsi|d,Psi=1|d=11+expjwijdjbi.(21)

The updating rules for the parameters are given as follows:

Δwij=ϵsidjdatasidjrecon(22)
Δai=ϵsidatasirecon(23)
Δbj=ϵdjdatadjrecon(24)

where Δ is the learning rate; Δwij, Δai, and Δbj are the updating values of the weight and biases in the visible layer and hidden layer, respectively; ⟨⋅⟩data means the values of visible variable i multiplied by hidden variable j before reconstruction; and ⟨⋅⟩recon means the values after reconstruction. ⟨⋅⟩data and ⟨⋅⟩recon represent the distribution defined by the reconstruction model.

Reconstructed visible variable s′ and hidden variable d′ are one sampling of the original variable p (s, d); therefore, the sample set obtained by multiple sampling can be regarded as an approximation of the original variables.

4.2 Deep Belief Network

The typical deep belief network (DBN) structure was proposed by Hinton and Salakhutdinov (2006). It is a deep learning neural network which consists of three RBM layers and one back propagation (BP) layer. The advantage of the DBN is its capability of extracting nonlinear features by hidden layer units through unlabeled training data in visible layer units. Training a DBN contains two steps. The first step is pretraining: weights of the generative model are obtained through unsupervised greedy layer-by-layer pretraining. The second step is fine-tuning, which improves the generalization ability of the discriminant model using the gradient descent algorithm.

According to the contrastive divergence (CD) algorithm, in the pretraining step, the weights and the bias of DBN are updated by

w=w+Δpd0=1|s0s0Tpd1=1|s1s1T(25)
b=b+Δpd0=1|s0pd1=1|s1(26)
a=a+Δs0s1(27)

where p (d(0) = 1|s(0)) = σ(wjs(0) + bj) is the probability of the neurons in the hidden layer that will be turned on for Bayesian visible neurons, and Δ is the learning rate.

In the fine-tuning step, for the output BP layer, the minimum mean square error (MSE) criterion is used, and the cost function is as follows:

E=1Ni=1NX̂iwl,blXi2(28)

where E is the MSE of DBN, X̂i and Xi represent the expected output and the actual output, respectively, i is the sample index, and (wl, bl) represent the weights and bias parameters to be learned in the lth layer.

The gradient descent method is used to update the weights and bias parameters of the network according to

wl,bl=wl,bl+ΔEwl,bl.(29)

4.3 Switching PSO Algorithm

To realize a balance between the local and the global search, a velocity update equation that depends on the mode with Markov parameters is introduced in the PSO algorithm. In general, the particles in the early stage should maintain their independence and diversity, which is helpful to expand the search area and avert premature convergence to local optimum. In the end, all groups may converge to the best particles and obtain a more accurate solution. The evolutions of the particles’ velocity and position are shown as

siτ+1=wξτsiτ+c1ξτγ1p1τxiτ+c2ξτγ2pgτxiτxiτ+1=xiτ+siτ+1(30)

where w (ξ(τ)), c1 (ξ(τ)), and c2 (ξ(τ)) are the inertia weight and acceleration coefficients. They all depend on a Markovian chain. The Markov chain ξ(τ) takes values in a finite state space R=1,2,,N with probability transition matrix Πτ=(πij(τ))N×N. πijτ0(i,jR) stands for the transition probability from i to j satisfying j=1NΠijτ=1. According to the current search information, Πτ can be adjusted by

Divk=1RLi=1Rι=1Oxiιτx̄ιτ2(31)

so as to balance the global and local search abilities, where R represents the size of the swarm. The length of the longest diagonal in the search space is denoted by ∣L∣. The dimension of the object problem is O. x is the ιth value of the ith particle, and the ιth value of the average point x̄ is represented by x̄ι, which is shown as

x̄ι=1Ri=1Rxiιτ(32)

In the study, the BP layer of DBN is replaced by the SPSO-SVM classifier (For detail of switching PSO to optimize the parameters of SVM, refer to the study by Lu et al. (2015).

4.4 Image Processing of Rice Diseases Based on DBN and SPSO-SVM Framework

This section introduces rice disease identification based on the DBN and SPSO-SVM framework. The top layer of the traditional DBN model is the BP neural network used to complete the prediction. In this study, the hybrid kernel support vector machine method optimized by the self-adaptive particle swarm optimization (SPSO) algorithm based on particle swarm concentration is used to replace the BP neural network, Combined with the DBN algorithm to detect rice diseases, the improved DBN model is shown in Figure 6.

FIGURE 6
www.frontiersin.org

FIGURE 6. Schematic diagram of the DBN and SPSO-SVM.

This section introduces the process of rice diseases image via the DBN and SPSO-SVM framework. The flowchart of rice diseases image recognition based on the DBN and SPSO-SVM is shown in Figure 7, and our aim is to learn features to identify the rice diseases. To reduce the computational complexity in the process of feature extraction by the DBN, we extract the region of interest (ROI) containing the lesion. The ROI is extracted by combining interactive and automatic extraction methods. First, the experienced plant protection experts mark the outline of the largest section of the lesion in the rice disease images. Then the maximum outer rectangular frame is constructed with the largest section. Finally, the rectangular frame is moved up, down, left, and right by five pixels to obtain the final ROI frame. We adjust the extracted ROI to a uniform scale of 64 × 64 to compare the discriminatory ability of different images against rice disease, and to adjust the network structure and parameter to analyze the impact of classification diagnosis results.

FIGURE 7
www.frontiersin.org

FIGURE 7. Flowchart of the DBN and SPSO-SVM-based rice diseases image recognition.

The selection of the input feature is very important for the DBN because it plays a decisive role in the performance of identification. In this study, the color feature, the shape feature, and the texture feature are selected as the input features of the DBN. The color feature is selected because infected rice tissue usually changes its color from green to brown to yellow. The texture feature is selected to detect the rice disease as there are great differences in thickness and direction between healthy and diseased tissues. Also, the shape feature can better capture the edge and internal pixel features of the lesion target.

In the pretraining stage, after three features are input into the visible layer of the first RBM, each layer of the DBN network is trained independently by an unsupervised algorithm, then we obtain the weight of the generated model. In the fine-tuning stage, the SPSO-SVM is proposed to take the output eigenvector of RBM as its input eigenvector and then train the SPSO-SVM classifier using the supervised algorithm. Moreover, in order to ensure the whole DBN eigenvector mapping is optimal, the backpropagation network will also forecast the error information from back to front to each layer, and then the weights of the DBN are fine-tuned. Moreover, for the sake of avoiding falling into the local optimum of the SVM, the SPSO has been introduced.

5 Rice Disease Identification Experiment

5.1 Test Platform and Training Parameters

The images are processed and the algorithm is implemented by using the Python 3.6 and TensorFlow 2.0 + Keras 2.2.4 open-source framework for deep learning with the Ununtu 18.04 X64 operating system. The computer has 64 GB memory and Intel (R) i7 5960K CPU at3.5 GHz processor. NVIDIA Titan XP graphics card is used to improve the image processing speed.

The identification of rice blast, sheath blight, and brown spot based on the DBN and SPSO-SVM mainly refers to the influence on the accuracy. After experimental comparison, it was finally determined that the number of RBM hidden layers was 15, the number of hidden layer nodes was 128, the batch size was 64, the learning rate was 0.01, the optimizer was SGD, and the momentum was 0.9. The comparison of test results is shown in Figure 8.

FIGURE 8
www.frontiersin.org

FIGURE 8. Model parameter setting.

5.2 Experiment Results

For the sake of verifying the stability and performance of the DBN and SPSO-SVM models, the 10-fold cross-validation method is applied to validate the recognition results. After the model recognizes the rice diseases 30 times, the hit rate, error rate, accuracy, and the area under the receiver operating characteristic (ROC) curve (AUC) value of the model are counted. Finally, the average values of the four performance evaluation indicators are calculated. The results of cross validation are shown in Table1. Table1 shows that the hit rate and accuracy of the model are 91.37 and 94.03%, respectively, and the error recognition rate is 8.63%.The accuracy comparison curves of the four models are shown in Figure 9. It shows that the model achieves high recognition ability and guarantees a low false detection rate. The significance of the AUC value is that the closer the AUC value is to 1, the better is the recognition performance of the model. The AUC value of the recognition model proposed in this study reaches 0.97, which shows that the model has good recognition performance. The closer the ROC curve of the model is to the Y coordinate axis on the left and the closer to the y = 1 line on the top, the better is the performance of the model. The ROC curve of the model is presented in Figure 10. Figure 10 shows the ROC comparison curves of the four models. Through the comparison, it can be seen that the DBN and SPSO-SVM models used in this study have good recognition performance and are better than the other three models. It can effectively learn the implicit feature information in the data set so as to achieve better recognition effect.

TABLE 1
www.frontiersin.org

TABLE 1. Performance parameters of the model.

FIGURE 9
www.frontiersin.org

FIGURE 9. Accuracy comparison curve.

FIGURE 10
www.frontiersin.org

FIGURE 10. ROC curve.

5.3 Experimental Discussions

The proposed DBN and SPSO-SVM models were used to identify rice blast, sheath blight, and brown spot. Under the same experimental conditions and the same data set, support vector machine, the SPSO-SVM and convolutional neural network (CNN) models are used for comparison. The parameters of the support vector machine are set as follows: the recognition accuracy under different combinations of SVM penalty parameter C and kernel function parameter g is compared by the grid search method; when the penalty factor C of the support vector machine is 2.8 and the kernel function parameter g is 50, the recognition accuracy is the highest. The SPSO parameters are set as follows: initial population 80, inertia weight 0.9, the maximum number of iterations 300, and learning factor C1 is 1.6 and C2 is 1.8. The parameters of the CNN are set as follows: the batch size is 64 and the number of iterations is 500. Using the cross-entropy loss function, the Adam optimizer is used to minimize the loss function, and the learning rate is set to 0.001. The comparison of the results of the four models is shown in Table2. It can be seen that the DBN and SPSO-SVM models gain better results.

TABLE 2
www.frontiersin.org

TABLE 2. Simulation results using the proposed method compared with the SVM, SPSO-SVM, and CNN.

6 Conclusion

This study presents a new rice disease identification method based on the DBN and SPSO-SVM. Data preprocessing, the structure design of the identification model, and the implementation technology of the DBN and SPSO-SVM have been elaborated in detail. Three features (including the color, shape, and texture) have been proposed for the DBN and SPSO-SVM method to learn in order to successfully recognize rice disease from the region of interest that is obtained by preprocessing rice disease images. Furthermore, several indices have also been proposed to verify the presented DBN and SPSO-SVM method, and demonstrate that the DBN and SPSO-SVM method achieve better recognition effect and have good anti-interference and robustness. The developed method has exhibited a better training performance, faster convergence rate, and better recognition ability than the SVM, SPSO-SVM, and CNN models.

Although the DBN and SPSO-SVM have achieved better performance in rice diseases recognition fields, there still exist some challenging problems. The first is how to find the optimal weights and bias parameters. The second is how to get high-quality rice disease sample images due to the influence of light, shooting angle, and foresight. The last one is how to improve model training effect, recognition level, and application ability.

In a future study, we plan to do the following work: investigating other optimization algorithms used to find the optimal parameters of the DBN or other machine learning models, enriching positive sample set and reducing negative sample noise interference, and extending the results in this study to some complex systems (Dong et al., 2014; Dong et al., 2015a; Dong et al., 2016a; Dong et al., 2016b; Jiang et al., 2019a; Jiang et al., 2019b; Li G. et al., 2019; Chen et al., 2021a; Chen et al., 2021b; Chen et al., 2021c; Chen et al., 2021d; Chen et al., 2021e; Huang et al., 2021; Jiang et al., 2021a; Jiang et al., 2021b; Liu et al., 2021b; Liu et al., 2021c; Liu X. et al., 2021; Zhao et al., 2021; Chen et al., 2022a; Chen et al., 2022b; Chen et al., 2022c; Sun et al., 2022; Wu et al., 2022).

Data Availability Statement

The raw data supporting the conclusion of this article will be made available by the authors, without undue reservation.

Author Contributions

YL: writing—original draft, investigation, and software. JD: conceptualization, supervision, and funding acquisition. PL: methodology, writing—review and editing, and project administration. YZ: formal analysis and visualization. ZH: validation.

Funding

This study was supported in part by the National Natural Science Foundation of China under Grants 61873058, 61933007 and U21A2019; the Key Projects of Heilongjiang Natural Science Foundation under Grant ZD 2019F001; the Heilongjiang Natural Science Foundation under Grant LH2020F042; the China Postdoctoral Science Foundation under Grant 2016M591560; the Heilongjiang Postdoctoral Financial Assistance under Grant LBH-Z15185; the Scientific Research Starting Foundation for Post Doctor from Heilongjiang under Grant LBH-Q17134; the Heilongjiang Bayi Agricultural University Innovative Research Team Foundation under Grant TDJH201807; the Open Fund of the Key Laboratory for Metallurgical Equipment; and the Control of Ministry of Education in Wuhan University of Science and Technology under Grants 2018A02 and MECOF 2019B02.

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors, and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Atila, Ü., Uçar, M., Akyol, K., and Uçar, E. (2021). Plant Leaf Disease Classification Using Efficientnet Deep Learning Model. Ecol. Inform. 61, 101182. doi:10.1016/j.ecoinf.2020.101182

CrossRef Full Text | Google Scholar

Chen, Q., and Pan, G. (2021). A Structure-Self-Organizing Dbn for Image Recognition. Neural Comput. Appl. 33, 877–886. doi:10.1007/s00521-020-05262-2

CrossRef Full Text | Google Scholar

Chen, T., Jin, Y., Yang, J., and Cong, G. (2022a). Identifying Emergence Process of Group Panic Buying Behavior under the Covid-19 Pandemic. J. Retailing Consumer Serv. 67, 102970. doi:10.1016/j.jretconser.2022.102970

CrossRef Full Text | Google Scholar

Chen, T., Peng, L., Yang, J., and Cong, G. (2021a). Analysis of User Needs on Downloading Behavior of English Vocabulary Apps Based on Data Mining for Online Comments. Mathematics 9, 1341. doi:10.3390/math9121341

CrossRef Full Text | Google Scholar

Chen, T., Peng, L., Yang, J., Cong, G., and Li, G. (2021b). Evolutionary Game of Multi-Subjects in Live Streaming and Governance Strategies Based on Social Preference Theory during the Covid-19 Pandemic. Mathematics 9, 2743. doi:10.3390/math9212743

CrossRef Full Text | Google Scholar

Chen, T., Qiu, Y., Wang, B., and Yang, J. (2022b). Analysis of Effects on the Dual Circulation Promotion Policy for Cross-Border E-Commerce B2b export Trade Based on System Dynamics during Covid-19. Systems 10, 13. doi:10.3390/systems10010013

CrossRef Full Text | Google Scholar

Chen, T., Rong, J., Yang, J., Cong, G., and Li, G. (2021c). Combining Public Opinion Dissemination with Polarization Process Considering Individual Heterogeneity. Healthcare (Basel) 9, 176. doi:10.3390/healthcare9020176

CrossRef Full Text | Google Scholar

Chen, T., Rong, J., Yang, J., and Cong, G. (2022c). Modeling Rumor Diffusion Process with the Consideration of Individual Heterogeneity: Take the Imported Food Safety Issue as an Example during the Covid-19 Pandemic. Front. Public Health 10, 313. doi:10.3389/fpubh.2022.781691

CrossRef Full Text | Google Scholar

Chen, T., Yin, X., Peng, L., Rong, J., Yang, J., and Cong, G. (2021d). Monitoring and Recognizing enterprise Public Opinion from High-Risk Users Based on User Portrait and Random Forest Algorithm. Axioms 10, 106. doi:10.3390/axioms10020106

CrossRef Full Text | Google Scholar

Chen, T., Yin, X., Yang, J., Cong, G., and Li, G. (2021e). Modeling Multi-Dimensional Public Opinion Process Based on Complex Network Dynamics Model in the Context of Derived Topics. Axioms 10, 270. doi:10.3390/axioms10040270

CrossRef Full Text | Google Scholar

Cheng, F., Zhang, H., Fan, W., and Harris, B. (2018). Image Recognition Technology Based on Deep Learning. Wireless Pers Commun. 102, 1917–1933. doi:10.1007/s11277-018-5246-z

CrossRef Full Text | Google Scholar

Cristin, R., Kumar, B. S., Priya, C., and Karthick, K. (2020). Deep Neural Network Based Rider-Cuckoo Search Algorithm for Plant Disease Detection. Artif. Intell. Rev. 53, 4993–5018. doi:10.1007/s10462-020-09813-w

CrossRef Full Text | Google Scholar

Dong, H., Wang, Z., Alsaadi, F. E., and Ahmad, B. (2015a). Event-Triggered Robust Distributed State Estimation for Sensor Networks with State-Dependent Noises. Int. J. Gen. Syst. 44, 254–266. doi:10.1080/03081079.2014.973726

CrossRef Full Text | Google Scholar

Dong, H., Wang, Z., Ding, S. X., and Gao, H. (2014). Finite-Horizon Estimation of Randomly Occurring Faults for a Class of Nonlinear Time-Varying Systems. Automatica 50, 3182–3189. doi:10.1016/j.automatica.2014.10.026

CrossRef Full Text | Google Scholar

Dong, H., Wang, Z., Ding, S. X., and Gao, H. (2015b). Finite-Horizon Reliable Control with Randomly Occurring Uncertainties and Nonlinearities Subject to Output Quantization. Automatica 52, 355–362. doi:10.1016/j.automatica.2014.11.020

CrossRef Full Text | Google Scholar

Dong, H., Wang, Z., Ding, S. X., and Gao, H. (2016a). On H-Infinity Estimation of Randomly Occurring Faults for a Class of Nonlinear Time-Varying Systems with Fading Channels. IEEE Trans. Automat. Contr. 61, 479–484. doi:10.1109/TAC.2015.2437526

CrossRef Full Text | Google Scholar

Dong, H., Wang, Z., Shen, B., and Ding, D. (2016b). Variance-Constrained H Control for a Class of Nonlinear Stochastic Discrete Time-Varying Systems: The Event-Triggered Design. Automatica 72, 28–36. doi:10.1016/j.automatica.2016.05.012

CrossRef Full Text | Google Scholar

Fuentes, A., Yoon, S., and Park, D. S. (2020). “Deep Learning-Based Techniques for Plant Diseases Recognition in Real-Field Scenarios,” in Advanced Concepts for Intelligent Vision Systems. Editors J. Blanc-Talon, P. Delmas, W. Philips, D. Popescu, and P. Scheunders (Cham: Springer International Publishing), 3–14. doi:10.1007/978-3-030-40605-9_1

CrossRef Full Text | Google Scholar

Goudarzi, S., Kama, M., Anisi, M., Soleymani, S., and Doctor, F. (2018). Self-Organizing Traffic Flow Prediction with an Optimized Deep Belief Network for Internet of Vehicles. Sensors 18, 3459. doi:10.3390/s18103459

PubMed Abstract | CrossRef Full Text | Google Scholar

Hassan, S. M., Maji, A. K., Jasiński, M., Leonowicz, Z., and Jasińska, E. (2021). Identification of Plant-Leaf Diseases Using Cnn and Transfer-Learning Approach. Electronics 10, 1388. doi:10.3390/electronics10121388

CrossRef Full Text | Google Scholar

Hinton, G. E. (2012). A Practical Guide to Training Restricted Boltzmann Machines. Berlin, Heidelberg: Springer Berlin Heidelberg, 599–619. doi:10.1007/978-3-642-35289-8_32

CrossRef Full Text | Google Scholar

Hinton, G. E., and Salakhutdinov, R. R. (2006). Reducing the Dimensionality of Data with Neural Networks. Science 313, 504–507. doi:10.1126/science.1127647

PubMed Abstract | CrossRef Full Text | Google Scholar

Hinton, G. E. (2002). Training Products of Experts by Minimizing Contrastive Divergence. Neural Comput. 14, 1771–1800. doi:10.1162/089976602760128018

PubMed Abstract | CrossRef Full Text | Google Scholar

Hu, G., Yang, X., Zhang, Y., and Wan, M. (2019). Identification of Tea Leaf Diseases by Using an Improved Deep Convolutional Neural Network. Sustain. Comput. Inform. Syst. 24, 100353. doi:10.1016/j.suscom.2019.100353

CrossRef Full Text | Google Scholar

Huang, L., Fu, Q., He, M., Jiang, D., and Hao, Z. (2021). Detection Algorithm of Safety Helmet Wearing Based on Deep Learning. Concurrency Computat Pract. Exper 33, e6234. doi:10.1002/cpe.6234

CrossRef Full Text | Google Scholar

Jiang, D., Li, G., Sun, Y., Hu, J., Yun, J., and Liu, Y. (2021a). Manipulator Grabbing Position Detection with Information Fusion of Color Image and Depth Image Using Deep Learning. J. Ambient Intell. Hum. Comput 12, 10809–10822. doi:10.1007/s12652-020-02843-w

CrossRef Full Text | Google Scholar

Jiang, D., Li, G., Sun, Y., Kong, J., Tao, B., and Chen, D. (2019a). Grip Strength Forecast and Rehabilitative Guidance Based on Adaptive Neural Fuzzy Inference System Using Semg. Pers Ubiquit Comput. 2019, 1–10. doi:10.1007/s00779-019-01268-3

CrossRef Full Text | Google Scholar

Jiang, D., Li, G., Tan, C., Huang, L., Sun, Y., and Kong, J. (2021b). Semantic Segmentation for Multiscale Target Based on Object Recognition Using the Improved Faster-Rcnn Model. Future Generation Comput. Syst. 123, 94–104. doi:10.1016/j.future.2021.04.019

CrossRef Full Text | Google Scholar

Jiang, D., Zheng, Z., Li, G., Sun, Y., Kong, J., Jiang, G., et al. (2019b). Gesture Recognition Based on Binocular Vision. Cluster Comput. 22, 13261–13271. doi:10.1007/s10586-018-1844-5

CrossRef Full Text | Google Scholar

Kaur, P., Pannu, H. S., and Malhi, A. K. (2019). Plant Disease Recognition Using Fractional-Order Zernike Moments and Svm Classifier. Neural Comput. Applic 31, 8749–8768. doi:10.1007/s00521-018-3939-6

CrossRef Full Text | Google Scholar

Li, D., Wang, R., Xie, C., Liu, L., Zhang, J., Li, R., et al. (2020). A Recognition Method for rice Plant Diseases and Pests Video Detection Based on Deep Convolutional Neural Network. Sensors 20, 578. doi:10.3390/s20030578

PubMed Abstract | CrossRef Full Text | Google Scholar

Li, G., Li, J., Ju, Z., Sun, Y., and Kong, J. (2019a). A Novel Feature Extraction Method for Machine Learning Based on Surface Electromyography from Healthy Brain. Neural Comput. Appl. 31, 9013–9022. doi:10.1007/s00521-019-04147-3

CrossRef Full Text | Google Scholar

Li, Y., and Chao, X. (2021). Semi-Supervised Few-Shot Learning Approach for Plant Diseases Recognition. Plant Methods 17, 1–10. doi:10.1186/s13007-021-00770-1

PubMed Abstract | CrossRef Full Text | Google Scholar

Li, Y., Qian, M., Liu, P., Cai, Q., Li, X., Guo, J., et al. (2019b). The Recognition of rice Images by Uav Based on Capsule Network. Cluster Comput. 22, 9515–9524. doi:10.1007/s10586-018-2482-7

CrossRef Full Text | Google Scholar

Liu, J., Yang, S., Cheng, Y., and Song, Z. (2018). “Plant Leaf Classification Based on Deep Learning,” in 2018 Chinese Automation Congress (CAC), Xi’an, China, November 30–December 2, 2018, 3165–3169. doi:10.1109/CAC.2018.8623427

CrossRef Full Text | Google Scholar

Liu, X., Jiang, D., Tao, B., Jiang, G., Sun, Y., Kong, J., et al. (2021a). Genetic Algorithm-Based Trajectory Optimization for Digital Twin Robots. Front. Bioeng. Biotechnol. 9, 793782. doi:10.3389/fbioe.2021.793782

PubMed Abstract | CrossRef Full Text | Google Scholar

Liu, Y., Jiang, D., Duan, H., Sun, Y., Li, G., Tao, B., et al. (2021c). Dynamic Gesture Recognition Algorithm Based on 3d Convolutional Neural Network. Comput. Intelligence Neurosci. 2021, 1–12. doi:10.1155/2021/4828102

CrossRef Full Text | Google Scholar

Liu, Y., Jiang, D., Yun, J., Sun, Y., Li, C., Jiang, G., et al. (2021b). Self-Tuning Control of Manipulator Positioning Based on Fuzzy Pid and Pso Algorithm. Front. Bioeng. Biotechnol. 9, 817723. doi:10.3389/fbioe.2021.817723

PubMed Abstract | CrossRef Full Text | Google Scholar

Lu, Y., Zeng, N., Liu, X., and Yi, S. (2015). A New Hybrid Algorithm for Bankruptcy Prediction Using Switching Particle Swarm Optimization and Support Vector Machines. Discrete Dyn. Nat. Soc. 2015, 1–7. doi:10.1155/2015/294930

CrossRef Full Text | Google Scholar

Madhulatha, G., and Ramadevi, O. (2020). “Recognition of Plant Diseases Using Convolutional Neural Network,” in 2020 Fourth International Conference on I-SMAC (IoT in Social, Mobile, Analytics, and Cloud) (I-SMAC), 2020, 738–743. doi:10.1109/I-SMAC49090.2020.9243422

CrossRef Full Text | Google Scholar

Nanehkaran, Y. A., Zhang, D., Chen, J., Tian, Y., and Al-Nabhan, N. (2020). Recognition of Plant Leaf Diseases Based on Computer Vision. J. Ambient Intell. Hum. Comput 2020, 1–18. doi:10.1007/s12652-020-02505-x

CrossRef Full Text | Google Scholar

Rahman, C. R., Arko, P. S., Ali, M. E., Iqbal Khan, M. A., Apon, S. H., Nowrin, F., et al. (2020). Identification and Recognition of rice Diseases and Pests Using Convolutional Neural Networks. Biosyst. Eng. 194, 112–120. doi:10.1016/j.biosystemseng.2020.03.020

CrossRef Full Text | Google Scholar

Russakovsky, O., Deng, J., Su, H., Krause, J., Satheesh, S., Ma, S., et al. (2015). Imagenet Large Scale Visual Recognition Challenge. Int. J. Comput. Vis. 115, 211–252. doi:10.1007/s11263-015-0816-y

CrossRef Full Text | Google Scholar

Sethy, P. K., Barpanda, N. K., Rath, A. K., and Behera, S. K. (2020). Deep Feature-Based rice Leaf Disease Identification Using Support Vector Machine. Comput. Electron. Agric. 175, 105527. doi:10.1016/j.compag.2020.105527

CrossRef Full Text | Google Scholar

Sharma, P., Berwal, Y. P. S., and Ghai, W. (2020). Performance Analysis of Deep Learning Cnn Models for Disease Detection in Plants Using Image Segmentation. Inf. Process. Agric. 7, 566–574. doi:10.1016/j.inpa.2019.11.001

CrossRef Full Text | Google Scholar

Sun, Y., Zhao, Z., Jiang, D., Tong, X., Tao, B., Jiang, G., et al. (2022). Low-Illumination Image Enhancement Algorithm Based on Improved Multi-Scale Retinex and Abc Algorithm Optimization. Front. Bioeng. Biotechnol. 369, 1–16. doi:10.3389/fbioe.2022.865820

CrossRef Full Text | Google Scholar

Vedantham, R., and Reddy, E. S. (2020). A Robust Feature Extraction with Optimized Dbn-Smo for Facial Expression Recognition. Multimed Tools Appl. 79, 21487–21512. doi:10.1007/s11042-020-08901-x

CrossRef Full Text | Google Scholar

Veisi, H., and Haji Mani, A. (2020). Persian Speech Recognition Using Deep Learning. Int. J. Speech Technol. 23, 893–905. doi:10.1007/s10772-020-09768-x

CrossRef Full Text | Google Scholar

Wspanialy, P., and Moussa, M. (2020). A Detection and Severity Estimation System for Generic Diseases of Tomato Greenhouse Plants. Comput. Electron. Agric. 178, 105701. doi:10.1016/j.compag.2020.105701

CrossRef Full Text | Google Scholar

Wu, X., Jiang, D., Yun, J., Liu, X., Sun, Y., Tao, B., et al. (2022). Attitude Stabilization Control of Autonomous Underwater Vehicle Based on Decoupling Algorithm and Pso-Adrc. Front. Bioeng. Biotechnol. 10, 843020. doi:10.3389/fbioe.2022.843020

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhang, D., Zhou, X., Zhang, J., Lan, Y., Xu, C., and Liang, D. (2018a). Detection of rice Sheath Blight Using an Unmanned Aerial System with High-Resolution Color and Multispectral Imaging. PLoS ONE 13, e0187470. doi:10.1371/journal.pone.0187470

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhang, X., Qiao, Y., Meng, F., Fan, C., and Zhang, M. (2018b). Identification of maize Leaf Diseases Using Improved Deep Convolutional Neural Networks. IEEE Access 6, 30370–30377. doi:10.1109/ACCESS.2018.2844405

CrossRef Full Text | Google Scholar

Zhao, G., Jiang, D., Liu, X., Tong, X., Sun, Y., Tao, B., et al. (2021). A Tandem Robotic Arm Inverse Kinematic Solution Based on an Improved Particle Swarm Algorithm. Front. Bioeng. Biotechnol. 2021, 363. doi:10.3389/fbioe.2022.832829

CrossRef Full Text | Google Scholar

Keywords: image classification, image recognition, rice diseases, deep belief networks, switching particle swarm optimization algorithm

Citation: Lu Y, Du J, Liu P, Zhang Y and Hao Z (2022) Image Classification and Recognition of Rice Diseases: A Hybrid DBN and Particle Swarm Optimization Algorithm. Front. Bioeng. Biotechnol. 10:855667. doi: 10.3389/fbioe.2022.855667

Received: 15 January 2022; Accepted: 28 March 2022;
Published: 27 April 2022.

Edited by:

Tinggui Chen, Zhejiang Gongshang University, China

Reviewed by:

Sanjay Goswami, Indian Institute of Sugarcane Research (ICAR), India
Dairi Abdelkader, Oran University of Science and Technology—Mohamed Boudiaf, Algeria
Bo Shen, Donghua University, China
Xihai Zhang, Northeast Agricultural University, China
Shidrokh Goudarzi, National University of Malaysia, Malaysia

Copyright © 2022 Lu, Du, Liu, Zhang and Hao. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Yang Lu, luyanga@sina.com

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.