Skip to main content

ORIGINAL RESEARCH article

Front. Plant Sci., 18 January 2023
Sec. Technical Advances in Plant Science
This article is part of the Research Topic Digital Innovations in Sustainable Agri-Food Systems View all 8 articles

JujubeNet: A high-precision lightweight jujube surface defect classification network with an attention mechanism

Lingjie Jiang,,Lingjie Jiang1,2,3Baoxi Yuan,,*Baoxi Yuan1,2,3*Wenyun MaWenyun Ma4Yuqian WangYuqian Wang5
  • 1School of Electronic Information, Xijing University, Xi’an, China
  • 2Shaanxi Key Laboratory of Integrated and Intelligent Navigation, The 20th Research Institute of China Electronics Technology Group Corporation, Xi’an, China
  • 3Xi’an Key Laboratory of High Precision Industrial Intelligent Vision Measurement Technology, Xijing University, Xi’an, China
  • 4Humanities Teaching Department, Gansu University of Chinese Medicine, Dingxi, China
  • 5Graduate Office, Xijing University, Xi’an, China

Surface Defect Detection (SDD) is a significant research content in Industry 4.0 field. In the real complex industrial environment, SDD is often faced with many challenges, such as small difference between defect imaging and background, low contrast, large variation of defect scale and diverse types, and large amount of noise in defect images. Jujubes are naturally growing plants, and the appearance of the same type of surface defect can vary greatly, so it is more difficult than industrial products produced according to the prescribed process. In this paper, a ConvNeXt-based high-precision lightweight classification network JujubeNet is presented to address the practical needs of Jujube Surface Defect (JSD) classification. In the proposed method, a Multi-branching module using Depthwise separable Convolution (MDC) is designed to extract more feature information through multi-branching and substantially reduces the number of parameters in the model by using depthwise separable convolutions. What’s more, in our proposed method, the Convolutional Block Attention Module (CBAM) is introduced to make the model concentrate on different classes of JSD features. The proposed JujubeNet is compared with other mainstream networks in the actual production environment. The experimental results show that the proposed JujubeNet can achieve 99.1% classification accuracy, which is significantly better than the current mainstream classification models. The FLOPS and parameters are only 30.7% and 30.6% of ConvNeXt-Tiny respectively, indicating that the model can quickly and effectively classify JSD and is of great practical value.

1. Introduction

SDD, also known as Automated Optical Inspection (AOI) or Automated Surface Inspection (ASI), is a significant research content in Industry 4.0 field (Penumuru et al., 2020). It is a technology that can acquire images by using machine vision equipment to determine whether there are defects in the acquired images. At present, surface defect equipment based on artificial vision has been widely used to replace manual visual inspection in various industrial fields, including automobile, household appliances, machinery manufacturing, semiconductor and electronics, chemical industry, medicine, aerospace, light industry, and other industries (Liu Y et al., 2020). For example: Ferguson et al. (2017) applied machine learning technology to the detection of surface defect on automobile parts. The GDXray Casting dataset adopted included 2,727 X-ray images, which were mainly from automobile parts, including aluminum wheels and steering knuckles. Tabernik et al. (2020) applied automated surface inspection technology to the detection of metal surface defects and published Kolektor SDD data set. Mayr et al. (2019) proposed a SDD method for solar panels. Huang et al. (2018) proposed a SDD method for magnetic tile surfaces. He et al. (2020) proposed An end-to-end steel SDD approach via fusing multiple hierarchical features. Gan et al. (2017) proposed hierarchical extractor-based visual rail surface inspection system. Yang et al. (2018) proposed automatic pixel-level crack detection and measurement using full convolutional network. Li et al. (2019) proposed detection algorithm for bridge cracks based on Deep Learning (DL). Tang et al. (2019) proposed an online Printed Circuit Board (PCB) defect detector on a new PCB defect dataset.

Compared with other computer vision tasks, SDD does not have a large and unified data set such as ImageNet (Deng et al., 2009), PASCAL-VOC (Everingham et al., 2010) and COCO (Lin et al., 2014). Defect detection mainly studies specific applications in different detection objects and scenarios. In the real complex industrial environment, SDD is often faced with many challenges, such as small difference between defect imaging and background, low contrast, large variation of defect scale and diverse types, large amount of noise in defect images, and even large amount of interference in the imaging of defects in natural environment. Therefore, it is faced with greater challenges.

The agriculture 4.0 model is an intelligent agricultural development model characterized by being intelligent and unmanned, and it is also a resource-integrated agricultural development model (Da Silveira et al., 2021). In the Agriculture 4.0 mode and the others, the difference between the agricultural management and service system is that, in the former mode, all the intelligent machinery and equipment, including their corresponding elements related to agriculture, such as agricultural production and circulation markets, are interconnected through the Internet of Things network. With the help of new internet technologies such as big data, cloud computing, and artificial intelligence, intelligent decisions on agricultural activities are made to improve the efficiency in resource utilization, labor productivity, and agricultural production. Lack of per capita resources, shortage of labor force, and urgent forms of environmental protection are scientific problems throughout the development of agricultural modernization (Zhai et al., 2020). Agriculture 4.0 is an in-depth development stage of agricultural modernization construction. Precise and intelligent agricultural production can be achieved with a higher level of intensity, precision, and coordination, and the three problems above can be fundamentally solved.

Jujube is a high-quality tonic native to China, rich in various vitamins, with high nutritional, edible, and medicinal values (Rashwan et al., 2020). Jujube has a history of being cultivated for more than 4,000 years. The quality is best for those with red color, thick, plump meat, small kernel, and sweet taste. The rapid increase in planting area of red jujube in China is in sharp contrast to the backward processing technology of post-harvest jujube. The jujube industry has boomed in recent years, with 3.3 million hectares under jujube cultivation and 7.46 million tons of production in China as of 2019. With the continuous improvement of living standards and the increasing popularity of food health knowledge, people’s demand for jujube products is also proliferating, and they have higher requirements for the quality of the fruit (Chen et al., 2017). However, during the natural growth, harvesting, and processing of jujube, defects such as deformation and rotten and cracked jujube fruit are often caused by pests, harsh environments, improper processing methods, and storage methods (Pham and Liou, 2020). If these defective products come onto the market mixed with the choicest jujube, it will seriously affect the jujube’s quality, sales, and prices, causing economic losses to jujube farmers and enterprises. Therefore, effective identification and classification of defective jujube is a primary means to ensure the quality of jujube products and has a significant economic value (GenG et al., 2019). Jujube appearance quality sorting technology is a crucial technology to improve jujube quality and product-added value in the process of jujube industrialization. Jujube sorting quality is also an important factor affecting the prices and markets of jujube. After the appearance and quality sorting, the market value of high-quality jujube can be increased, and the defective jujube can be made into new products through secondary deep processing, which is the direction and trend of the future revenue of the jujube industry. The traditional industrialization process of jujube requires a lot of manpower, and the deep processing of defective jujube can effectively save resources and reduce the pressure of environmental protection. Therefore, the application of artificial vision technology to the industrialization process of jujube production and the improvement in the level of automation can provide a beneficial reference for the combination of Agriculture 4.0 and Industry 4.0 (Araújo et al., 2021).

Machine vision technology has developed rapidly in recent years and has been gradually applied in the quality detection of agricultural products. However, at present, a large number of factories still use manual methods to classify JSD (Bhargava et al., 2022), which have obvious disadvantages such as low efficiency and high costs. Manual quality sorting is subject to significant fluctuations in human factors, and the phenomenon of wrong inspection and omission often occurs, which leads to the uneven overall quality of jujube commodities (Dong et al., 2022). Therefore, it is urgently necessary to introduce advanced technologies to innovate and replace the simple manual sorting methods to improve the quality of jujube products and achieve their automatic sorting.

Traditional SDD methods based on machine vision usually use conventional image processing algorithms or manually designed features and put into the classifier for classification. In general, the corresponding imaging scheme is usually designed according to the defect characteristics of the object surface to be inspected. A reasonable imaging scheme is helpful to obtain the image with uniform illumination and clearly reflect the surface defects of the objects. A common method is to select a light source based on the surface color of the object being inspected. For example, Jing et al. (2016) selected a composite white light source to image the various types of defects on the surface of colored fabrics. Another method is to select different imaging schemes according to the reflective properties of the object surfaces to be detected., mainly including bright field imaging, dark field imaging, and mixed imaging. As an example, Chen et al. (2016) designed two concentrically placed conic annular bright field light sources to illuminate the central and peripheral areas at the bottom of the metal can for SDD at the concave and convex bottom of the can. For the SDD algorithm of red jujube, Wu et al. (2016) used hyperspectral imaging techniques and machine vision algorithms based on Support Vector Machines (SVM) to achieve the quality classification.

In the real production environment, complex SDD often faces many challenges, such as low contrast, large variation of the defect scales, multiple defect types, noise, interference, etc. In this case, classical methods are often helpless, and it is difficult to obtain good detection results. The introduction of classical techniques did solve the automated sorting of red jujube to a certain extent. However, they require a high inspection environment and have problems such as low accuracy and poor real-time performance, and, therefore, are not conducive to large-scale promotion (Li et al., 2022). DL, a significant branch of machine learning, has made breakthroughs in recent years, especially with Convolutional Neural Networks (CNN), which have become widely used in various image recognition scenarios because of their powerful feature extraction and nonlinear representation capabilities, and some defect detection methods based on DL begin to have wide application in various industrial scenes. In 2014, Soukup and Huber-Mörk (2014) innovatively trained a CNN model to accurately classify cavity defects on the track surface by collecting photometric stereo images. The whole network consists of two convolutional layers, two pooling layers and the last fully connected layer. Ultimately, a recognition accuracy of 98.98% was achieved in the rail surface defect dataset. Park et al. (2016) designed a CNN-based SDD system for the automated detection of various defects, such as dirt, scratches, burrs, and wear, on the surface of parts in industrial production. This work shows that the method can achieve 98% classification accuracy on the validation dataset., and its detection speed is 5285 samples/min. Kyeong and Kim (2018) proposed a CNN framework to classify mixed-type defect patterns in Wafer Bin Map (WBM) of semiconductor industry. Deitsch et al. (2019) used the modified VGG19 network to identify defects in 300×300 resolution solar panel images, and the accuracy of the network reached 88.42%, which exceeds a variety of hand-designed features, including KAZE, Scale-Invariant Feature Transform (SIFT), Speeded Up Robust Feature (SURF), and the performance of the classifier exceeded the effect of the SVM method. Liang et al. (2019) proposed a method built on top of ShuffleNetV2 and achieved 99.88% classification accuracy on an in-line code inspection apparatus in the plastic container industry. The methods of directly using original images for SDD were widely used in many fields, such as welding defect classification (Zhang et al., 2019), lithium polymer battery bleb defect classification (Ma et al., 2019), PCB defect classification (Deng et al., 2018), etc. In addition, two-stage Faster R-CNN series and one-stage You Only Look Once (YOLO) series target detection networks are also used for SDD. For example: Tao et al. (2020) improved the two stage Faster R-CNN network for insulator defect location in drone power inspection, and Xue and Li (2018) realized shield tunnel lining defects detection based on the improved Faster R-CNN. Li et al. (2018) proposed a method based on MobileNet-SSD for the detection of sealing surface defects of filling production line containers. Liu et al. (2020) used MobileNet-SSD network to locate the high-speed rail catenary support components. Zhang et al. (2020) applied YOLOv3 to bridge surface defect location.

Several scholars have applied it to fruit sorting tasks and achieved successful results (Yin et al., 2022; Lin et al., 2022). Some researchers have also carried out relevant studies in the field of surface defect identification of jujube, which will be elaborated in the section “Related Work”. A CNN-based JSD classification algorithm named JujubeNet is proposed in this paper. The proposed algorithm is built on the basis of ConvNeXt (Liu et al., 2022) for the JSD detection. To solve the problem of SDD of jujube, the ConvNeXt model was improved as follows: firstly, by introducing a novel MDC module with a multi-branch structure instead of the original ConvNeXt module, the classification accuracy of the model is effectively improved, while the number of model parameters is reduced; Secondly, the CBAM is combined with ConvNeXt to make the model focus on different classes of jujube defects features (Woo et al., 2018). In the experiment of this paper, we compare JujubeNet with other mainstream algorithm models, and the results show that our JujubeNet has higher classification accuracy in the JSD classification scenario. The main contributions of this paper are as follows.

1. In the actual production environment, we collected 12000 images of jujube with 5 categories of defects (2000 images for each type of defective product) and 2000 images of good product, and created a dataset named ‘Jujube2000’, specifically for the classification study of surface defects of jujube. The jujube defect image dataset is released at https://pan.baidu.com/s/1mQZa_aoJ0uCitnSHta0UCg.

2. A MDC module with a multi-branch structure is designed, the CBAM is introduced to improve the ConvNeXt model, and finally, JujubeNet is proposed.

3. The effectiveness of the MDC module and CBAM Attention Mechanism (AM) was verified by ablation experiments, respectively.

4. The performance advantages of the proposed network in this paper are verified by comparing JujubeNet with the other mainstream networks.

The rest of this paper is structured as follows: Section 2 introduces the current DL-based algorithms for JSD. Section 3 introduces the ‘Jujube2000’ dataset and the improvement method proposed in this paper and presents JujubeNet. Section 4 compares and analyzes the experimental results of JujubeNet with the mainstream network on the ‘Jujube2000’ dataset. Section 5 summarizes the main work of this paper and indicates the directions for future research.

2. Related work

In recent years, DL methods represented by CNN have made significant breakthroughs in computer vision. The great success of AlexNet in the 2012 ImageNet competition also marks the beginning of the era of DL. Take image classification as an example, such as VGGNet (Simonyan and Zisserman, 2014), ResNet (He et al., 2016), and DenseNet (Huang et al., 2017), which have achieved satisfactory results in the traditional vision domain (Yang et al., 2022). Along with the development needs of smart agriculture and intelligent manufacturing, DL methods are increasingly introduced into various fields (Zhu et al., 2018). At present, scholars have gradually applied DL methods to fruit defect classification scenarios (Altalak et al., 2022). In this section, the current DL-based JSD classification methods will be discussed.

Geng et al. (2018) first applied a deep CNN with a two-branch structure to the defect classification task of jujube. In the first branch, the authors adopted a migration learning strategy to train the jujube dataset using SqueezeNet; in the second branch, the authors proposed a fusion module, which broadens the structure of the network and improves the classification accuracy of the model by contacting multiple feature maps. Finally, the classification model achieves an average accuracy of 99.3% on the self-built dataset of jujube. Sun et al. (2019) designed a CNN with low computational cost and high classification accuracy specifically for the real-time detection and classification of JSD. The model is based on DenseNet and simulates the visual mechanism of organisms by adding SE attention. In the authors’ self-built dataset of JSD, the model achieves classification accuracy comparable to mainstream networks with an accuracy of 91.9% with real-time availability. Wen et al. (2020) proposed a residual network-based method for classifying surface defects of jujube. First, the authors separated the G component from the RGB color map as the network’s input, improved the residual network’s activation and loss function, and then introduced the Dropout method to avoid overfitting. Finally, with the advantage of the residual module, the model achieved 96.1% classification accuracy on the self-built dataset of jujube. Guo et al. (2020) conducted a study on the impact of the jujube dataset on DL classification algorithms and used generative adversarial networks and rigid transformation to enhance the image data to solve the problem of an uneven sample of defective jujube. Experiments showed that the classification accuracy of the algorithm could be effectively improved, and based on the ResNet18 algorithm, the authors achieved a classification accuracy of 99.2%. Ju et al. (2022) proposed an improved ResNet for JSD classification. The authors first performed simulated data augmentation on small sample data to build a dataset containing five classes of JSD; secondly, the original ResNet was improved by embedding the SE module and applying Triplet loss and Center loss instead of Softmax loss; and finally, transfer learning was used to train the model. Their report shows that the classification accuracy of the algorithm can reach 94.2% in the authors’ self-built dataset of JSD. Yu (2022) proposed a multiple attention blending method for jujube grading. Using DenseNet121 as the backbone network, the authors obtained the final output by designing multiple attention mixing modules, specifically, constructing spatial attention, channel attention, and channel-space attention branches in the module and averaging the outputs of the three branches. The results show that the method can achieve a classification accuracy of 95.7%.

In summary, it is shown that the DL-based image classification algorithm is feasible for the JSD classification task. However, these studies are deficient in several aspects overall: Firstly, the dataset of JSD is not publicly available for academic research, which is detrimental to promoting the research on JSD. Secondly, the above studies are based on classical networks as the research basis of the model, which are not as good as the state-of-the-art networks in model accuracy and structural optimization design. In the Result section (Table 1), this paper demonstrates ConvNeXt’s superiority over other classical networks. In order to achieve more efficient industrial production, cutting-edge and practical techniques in academia should be applied to actual production. Finally, these studies lacked detailed analysis of the misidentified defects. A comprehensive and accurate analysis can be targeted to assist in the optimization of the model to improve the model’s classification accuracy and reduce misidentified. The purpose of this research is to design a lightweight and high-precision network for the classification of JSD. Considering the shortcomings of current research, this paper first collects and discloses the ‘Jujube2000’ dataset for academic research on the classification of JSD. Secondly, a novel MDC module is designed, and CBAM is introduced to improve the advanced ConvNeXt. Next, JujubeNet is proposed, specifically used for JSD classification. Then, the superiority of JujubeNet is demonstrated by extensive experiments. Finally, the direction of the model optimization in the following work was given by a detailed analysis of the misidentified case. In Table 2, this paper further summarizes the JSD classification methods mentioned above in terms of the algorithm used by the model, the composition of the data set, and the classification accuracy.

TABLE 1
www.frontiersin.org

Table 1 Comparison of the performance for each model on the ‘Jujube2000’ test set.

TABLE 2
www.frontiersin.org

Table 2 Experimental results of algorithms in their own datasets in the relevant literature.

Since the dataset used in the relevant literature has not been publicly downloaded, it cannot be directly compared with the experimental results in the relevant papers. Therefore, this paper conducted a comparative experiment with the underlying networks in the related papers on the ‘Jujube2000’ dataset. The results show that the test accuracy of the base algorithms in the relevant papers on the ‘Jujube2000’ dataset is generally lower than that on their own datasets, which also reflects the fact that the ‘Jujube2000’ data set is more challenging and difficult to classify. Based on the ConvNeXt model, this paper achieved an improvement in model accuracy while significantly reducing FLOPS and the number of parameters. The experimental results show that our JujubeNet has a significant advantage in terms of prediction accuracy and parameter computation: the prediction accuracy reaches 99.1% on the ‘Jujube2000’ data set, and the number of parameters is only 8.5M. Please see the Result section of this paper for detailed results.

3. Materials and methods

3.1. Image shooting and processing

Up to now, as the relevant literature has not disclosed the download links of their data sets, this paper has built a jujube image acquisition platform in the actual production scene, which is specially used for the collection of JSD images. The workflow of the acquisition platform is shown in Figure 1. The acquisition equipment mainly consists of LED light source, CCD industrial camera (MER-500-14U3C), roller conveyor, motor switch, and PC (CPU: AMD Ryzen 7 4800H 2.90GHz, RAM:16GB, SSD:500G).

FIGURE 1
www.frontiersin.org

Figure 1 Workflow diagram of ‘Jujube2000’ data set collection platform.

When the motor switch is activated, the roller conveyor drives the jujubes forward at a uniform speed, and at this time, the PC controls the camera to take and save the image. The camera can take a large image containing multiple jujubes at a time. Then, according to the gap between the rollers of the roller conveyor, the large image is divided into sub-images containing only one jujube. These sub-images are saved to the hard disk in png format. In order to obtain sufficient training images and guarantee the balance of the data set, 12000 images were selected from a large number of Jujube images to form the ‘Jujube2000’ data set, which contains 5 kinds of surface defects (2000 images for each defect) and 2000 high quality jujube images. The ‘Jujube2000’ dataset contains six categories: deformed, wrinkled, cracked, moldy, bird-pecked, and normal jujubes. The data set is divided into training, test and verification sets in a ratio of 7:2:1. Some sample images are shown in Figure 2.

FIGURE 2
www.frontiersin.org

Figure 2 Examples of six kinds of defect samples: (A) deformed; (B) wrinkled; (C) cracked; (D) moldy; (E) bird-pecked; (F) normal. (A-E) are jujubes with surface defects, and (F) is one of high quality.

During model training, multiple images need to be combined into a batch to feed the model (the batch size depends on the memory of the GPU), and the images must be of the same size in one batch, which makes the images need to be normalized before the model training. In a large dataset, there will inevitably be images of various sizes. Therefore, setting the height and width of images to be the same is an optimal solution considering the speed and accuracy of the model training (disadvantage: resize may lead to deformation of the object and thus affect the accuracy of the model). Numerous researches (He et al., 2016; Liu et al., 2022) have shown that using an input image size of 224 × 224 is the most balanced choice after considering the model accuracy and computational effort. For example, the default feature map size is 7 × 7 with a downsample multiplicity of 2. If the image size is adjusted to 112 × 112: the image information will be seriously lost. However, resizing the image to 448 × 448 will result in a significant increase in computing load. Therefore, in order to facilitate the model training, the jujube image size is normalized to 224×224 in this paper. Meanwhile, in order to improve the acquisition sample’s quality and increase the model’s generalization ability, this paper performs data enhancement on the images. The following are the approaches taken for the JSD classification scenario.

● Gaussian noise: Add noise to the image in line with the Gaussian distribution to simulate signal interference during image acquisition.

● Random cropping: On the one hand, The cropping of random regions on the image can have the effect of data enhancement; On the other hand, the stability of the model can be enhanced, and the model overfitting can be effectively prevented.

● Hybrid enhancement: Hybrid enhancement refers to the enhancement process of superimposing the contrast, brightness, and color of an image, and this operation contains a variety of image processing methods.

● Horizontal and vertical flip: Randomly flip the image horizontally or vertically without changing the original image information.

Finally, the ‘Jujube2000’ dataset contains 60,000 images after the image enhancement, of which 42,000 images are used for model training, 12,000 for model testing, and 6,000 for model validation.

3.2. Basic network selection

In order to achieve better classification results on the ‘Jujube2000’ dataset, this paper refers to the algorithms used by different authors in related works (Geng et al., 2018; Sun et al., 2019; Wen et al., 2020; Guo et al., 2020; Ju et al., 2021; Yu et al., 2022). However, the relevant literature did not disclose the code in their papers, and the data sets used were not publicly available for download. Therefore, this paper can neither directly identify the algorithms from the relevant literature for the ‘Jujube2000’ dataset for practical production nor directly compare the algorithms proposed in this paper with those in the relevant papers. For this reason, this paper evaluated the classification effect of the base models (SqueezeNet, ResNet18, ResNet34, ResNet50, DenseNet121) used in the related papers and the current mainstream base networks on the ‘Jujube2000’ dataset (detailed results of the experimental tests shown in Section 4.3). The evaluation results show that, firstly, SqueezeNet has the worst classification accuracy (only 75.7% accuracy in test dataset), which can not be directly applied in a real production environment; secondly, DenseNet121 has the slowest inference speed (DenseNet121: 40 images per second inference), which is not conducive to subsequent device deployment; Although VGG16 performed well in the training set, achieving a classification accuracy of 94.5%, it performed poorly in the test set, achieving only 91.8% classification accuracy, which also indicates the poor generalization ability of the network. Also, comparing the ResNet family, although they achieve similar accuracy on the test set, the difference in inference speed is more pronounced (ResNet18: 93.4% test accuracy, 251 images per second inference; ResNet34: 93.6% test accuracy, 168 images per second inference; ResNet50: 93.4% test accuracy, 115 images per second inference). Finally, ConvNeXt-Tiny achieved a classification accuracy of 97.5% on the training set. On the test set, ConvNeXt-Tiny has the best (98.6% test accuracy) generalization performance among all the mainstream networks participating in the comparison. Therefore, this paper further improved the network based on ConvNeXt-Tiny and got a better model named JujubeNet. After our improvements, the JujubeNet is a much lighter model than the ConvNeXt-Tiny, with less than 30.1% parameters and FLOPS, but 0.5% more accuracy, which makes it even more suitable for the processing scenario of industrial production of jujube.

3.3. ConvNeXt

ConvNeXt is a pure CNN proposed by FAIR in 2022 (Liu et al., 2022). It eliminates the need for tedious operations such as window shifting and relative position bias and offers better performance with less computation than the currently popular transformer network. The overall structure of ConvNeXt is based on the design of ResNet using residual blocks. It incorporates many advanced network design approaches to further improve the network’s overall performance. The detailed structure of ConvNeXt-Tiny and ConvNeXt Block are shown in Figure 3.

FIGURE 3
www.frontiersin.org

Figure 3 (A) ConvNeXt-Tiny overall network structure; (B) ConvNeXt Block structure.

3.4. MDC block

Increasing cardinality is a more effective way of gaining accuracy than going deeper or wider. This idea was first proposed in ResNeXt (Xie et al., 2017). Drawing inspiration from it, this paper introduced a two-branch structure in the ConvNeXt Block. In one branch, the paper followed the original modular design; while in the other, it used two consecutive convolutional designs in reference to the Wide Residual Network (Zagoruyko and Komodakis, 2016). Of these two branches, each one’s channel dimension of each branch is half of that on the main branch. Meanwhile, this paper uses depthwise separable convolution in wide residuals instead of ordinary convolution to reduce the number of model parameters. Finally, after the two branches are dimensionally spliced, the information passing on the main branch is randomly discarded by Drop Path, which can effectively prevent the model from overfitting and improve the overall performance. The structure of the MDC block is shown in Figure 4.

FIGURE 4
www.frontiersin.org

Figure 4 The structure of the MDC block.

3.5. CBAM

In general, when CNN extracts features from the images, it will inevitably be disturbed by the background and noise, which can directly affect the classification effect of the networks. Such problems can be effectively overcome by introducing an AM, which enables the network to focus on the helpful feature information and suppress the useless noise and interference. This will improve the model’s classification accuracy. CBAM is a general and efficient AM proposed by Woo et al. in 2018, which perceives feature information in different dimensions through the channel attention module and focuses on location information in the feature map through the spatial attention module (Woo et al., 2018). Not only that but CBAM can also be easily integrated into CNN for end-to-end training (Zhong et al., 2022). The structure of CBAM is shown in Figure 5, which mainly consists of a channel attention module and a spatial attention module.

FIGURE 5
www.frontiersin.org

Figure 5 Convolution Block Attention Module.

As shown in Figure 5, the feature map first passes through the channel attention module, which generates the corresponding channel attention map using the channel relationship between different features. Then, the input feature map is multiplied by the channel attention map. The output is again input to the spatial attention module, which generates the spatial attention map using the spatial position relationship between features. It multiplies the output of the channel attention module with the spatial attention map to obtain the final output feature map. The mathematical expressions of the above operations are shown in (1) and (2), where F (C×H×W) denotes the input feature map, Mc(F) (C×1×1) is the one-dimensional channel attention map, Ms(F’) (1×H×W) is the two-dimensional spatial attention map, ⊗ denotes the element multiplication operation, F’ (C×H×W) is the output after the channel attention module, and F’’ (C×H×W) is the final output of the CBAM.

F=McFF(1)
F=MsFF(2)

This paper further explores the embedding position of CBAM in the model and designed the following three structures (Kim et al., 2021), as shown in Figure 6. Where (a) indicates the use of the CBAM after each ConvNeXt Block operation; (b) indicates that the model uses CBAM before each downsample; and (c) indicates that the network uses CBAM after each downsample, and according to the experiments in Section 4.1, (c) shows more excellent results, so our model will adopt the CBAM of (c) embedding position design.

FIGURE 6
www.frontiersin.org

Figure 6 CBAM embedded position design.

3.6. JujubeNet

In this paper, a novel MDC module with multi-branch structure based on ConvNeXt Block is first designed, then introduced the CBAM AM into ConvNeXt, and finally proposed a high-precision lightweight classification network named JujubeNet, which is specifically designed for JSD classification. The experiments show that JujubeNet can perform the JSD classification excellently, and its overall network structure is shown in Figure 7.

FIGURE 7
www.frontiersin.org

Figure 7 The network structure of the JujubeNet model. It consists of CBAMs (The details of the CBAM structure is shown in Section 3.5), MDC blocks, downsample modules, a stem module, and a classifier.

3.7. Operating environment and parameter setup

All experiments were conducted on the same high-performance DL server (Central processing unit: Intel Xeon Silver 4210 CPU @2.20GHz; Graphics processing unit: NVIDIA GeForce RTX 2080 Ti 11GB; Memory: 128 GB; Deep learning framework: Python 3.8.10, Cuda 10.2, torch 1.8.1, torchvision 0.9.1; Operating system: Windows 10). In the experiments, uniform training parameters were set for all the models in this paper. The training image size is fixed at 224×224, the BatchSize set at 32, and the cross-entropy loss function and AdamW optimizer are applied in the model training. In the initial training phase of the models, this paper first performed a 2-epoch warm-up training, during which one-dimensional linear interpolation was employed to update the learning rate of each iteration. After the warm-up training, the learning rate is decayed using a cosine annealing function, where the initial learning rate is 0.0005, and the minimum learning rate is 0.00005. It is worth noting that we did not train the models using transfer learning in order to make a fairer comparison of the effects of network model modifications. Finally, each model was trained for 300 epochs.

3.8. Evaluate metrics

In DL methods, Accuracy, Recall, Precision, and F1-scores are significant metrics for evaluating the merits of classification models (Khasawneh et al., 2022). The accuracy rate is the percentage of the correct samples predicted by the model to the total number of samples; the recall rate is the percentage of the model correctly predicted as a positive sample to the total number of positive samples; the precision rate is the percentage of the number of positive samples predicted by the model that genuinely belongs to positive samples, and the F1-score is the best balance point that the model measures both the precision rate and the recall rate and achieves, and this value also reflects the overall performance of the model more comprehensively. Their specific formulae are shown below, where TP is the number of true positive samples, FP is the number of false positive samples, FN is the number of false negative samples, and TN is the number of true negative samples.

Accuracy=TP+TNP+N(3)
Recall=TPTP+FN(4)
Precision=TPTP+FP(5)
F1Socre=2×TP2×TP+FP+FN(6)

4. Results

4.1. CBAM embedding position experiment

Table 3 shows the detailed experimental results of the three CBAM embedding position schemes proposed in Section 3.5 of this paper. By observation, it can be found that as the state-of-the-art classification network, ConvNeXt exhibits strong classification performance, and the CBAM embedding position of the scheme (c) can further improve the model’s performance. Therefore, CBAM will be used after each downsample performed by the network.

TABLE 3
www.frontiersin.org

Table 3 Performance comparison of different CBAM embedding position schemes.

4.2. Ablation experiments

To verify the effectiveness of the MDC module and CBAM in JujubeNet, this paper conducted ablation experiments on the test set, and the results are shown in Table 4.

TABLE 4
www.frontiersin.org

Table 4 Comparison of ablation experiments on the test set.

The data in the table shows that CBAM can effectively improve the model’s classification accuracy with almost no increase in the number of the model parameters, and the accuracy is improved by 0.3% compared with the original network. By using the MDC module, the model’s accuracy is also improved by 0.4%, and the FLOPS and the number of parameters are reduced by about 70%, making the model more efficient and lighter. Finally, the classification accuracy of JujubeNet reached 99.1% with the introduction of the two modules. The FLOPS and the number of parameters were only 43.7G and 8.5M, indicating that the MDC module and the CBAM AM can effectively enhance the model’s recognition of JSD and significantly reduce the number of the parameters.

4.3. Model comparison analysis

Under the experimental platform set in this paper, the performance of JujubeNet and the current mainstream classification models are compared on the ‘Jujube2000’ dataset. Figure 8 shows the trend of validation accuracy and training loss with the number of epochs for each model. In Figure 8A, the horizontal axis indicates the number of rounds of model training and the vertical axis indicates the validation accuracy of the model. Similarly, in Figure 8B, the horizontal axis indicates the number of rounds of model training and the vertical axis indicates the training loss of the model.

FIGURE 8
www.frontiersin.org

Figure 8 Training and validation of each model on the ‘Jujube2000’ dataset. (A) Trend of the validation accuracy curve of the model with the number of epochs. (B) Trend of the training loss curve of the model with the number of epochs. Colors denote corresponding models.

Figure 8A shows that all the models start to converge at 30 epochs and slowly level off, reaching full convergence at 300 epochs. Among these models, JujubeNet achieved the highest validation accuracy, up to 96.0%, followed by ConvNeXt with 94.7% accuracy, and the VGG16 model had the worst validation accuracy of 88.3%. Figure 8B shows that the training loss curves of the CNN architecture models are relatively consistent, with the ConvNeXt model performing the best, reflecting its excellent fitting ability, while JujubeNet also shows its excellent ability. Swin-Tiny has the lowest convergence effect, which may be related to the fact that it is not trained using transfer learning because the transformer architecture model is more influenced by the number of training epochs and the dataset size.

All trained models were tested on the test set according to each performance metric, and the specific experimental results are shown in Table 1. The results show that JujubeNet has a significant advantage in terms of prediction accuracy and parameter computation. Based on the ConvNeXt model, this paper achieved a slight improvement in model accuracy while significantly reducing FLOPS and the number of parameters. Finally, JujubeNet achieves a prediction accuracy of 99.1%, and the number of parameters is only 8.5M. In addition, this paper also tested the classification effectiveness of the underlying networks in the related papers on the ‘Jujube2000’ dataset. The results show that the test accuracies of each network model on our dataset are generally lower than the test accuracies on their respective datasets, which also reflects the fact that ‘Jujube2000’ is more challenging and more difficult to classify.

4.4. Confusion matrix

The confusion matrix is a visual tool to evaluate the performance of a classification model, which describes the relationship between the prediction results of the model and the actual sample data. In addition, the confusion matrix can be used to more intuitively discern the strengths and weaknesses of the classification models and analyze their problems. As shown in Figure 9, this paper visualized the confusion matrix of each model in the previous subsection. The horizontal axis represents the true labels of the images, and the vertical axis represents the results predicted by the models. The diagonal part indicates that the predicted results of the models are the same as the true labels, and the remaining part indicates that the predicted results of the models do not match the true labels.

FIGURE 9
www.frontiersin.org

Figure 9 Confusion matrix for each model. (A) VGG16. (B) ResNet34. (C) DenseNet121. (D) Swin-Tiny. (E) ConvNeXt-Tiny. (F) The proposed JujubeNet.

The analysis reveals that, firstly, the deformation defects of jujubes are the easiest to be misjudged because their defective characteristics are the least obvious. Second, crack defects are more likely to be misclassified as moldy, because the black spots of mold are more similar to the cracks of crack defects. Third, due to the variability of moldy defects, making this defect is more easily identified as other defect categories. Among the mainstream networks, Swin-Tiny and ConvNeXt-Tiny achieved better classification performance because they are more novel and have a more reasonable network structure compared with VGG16, ResNet34 and DenseNet121, among which, ConvNeXt-Tiny performs the best. Furthermore, the classification performance of ConvNeXt-Tiny can be further optimized by constructing novel MDC module and introducing the CBAM AM in JujubeNet.

Jujubes are naturally growing plants, and defects of the same type may vary while defects of different types may differ less. Therefore, it is much more difficult than the defect detection of industrial products produced according to the specified process. Though JujubeNet shows state-of-the-art classification performance in many experiments, it has relatively more misidentified extruded, cracked, and moldy JSD in some cases. The specific details are shown in Figure 10. This paper tries to explore and analyze the causes of model discrimination errors for classifying anomalous images. Figure 10A shows that JujubeNet incorrectly identifies a deform defect as normal. The possible reason is that the shooting angle and lighting caused the defective sample to show more features of normal jujube, thus diluting the distortion. As shown in Figure 10B, even the human eye cannot accurately discern the specific defect of this jujube, as there appeared to be moldy in the cracked areas, which made it difficult for JujubeNet to make an accurate judgment. Figure 10C shows that a moldy defect was misidentified as normal. In this image, the moldy part is located on the edge of the jujube, causing the model to mistake it for a shaded area, leading to a classification error. Through a rational analysis of the failure cases, we will carry out targeted optimization on the proposed model to reduce misidentification in future work.

FIGURE 10
www.frontiersin.org

Figure 10 Examples of failure cases. The yellow dashed boxes indicate the defect areas that the model needs to focus. (A) A deformation defect was mistaken for a normal jujube. (B) A crack defect was mistaken for a moldy defect. (C) A moldy defect was mistaken for a normal jujube.

4.5. Visual interpretation of the model

In order to further visualize and analyze the process of classifying surface defects of jujubes, this paper introduce Gradient-weighted Class Activation Mapping (Grad-CAM), which is a method of weighted summation of specific feature maps in the model by weight and outputs a heat map of the specified class (Li and Li, 2022). In the heat map, the higher the weight, the redder the region’s color, indicating that the image information of the region has a more significant influence on the model for category discrimination. Conversely, the smaller the weight, the bluer the region’s color, indicating that the image information of the region has less influence on the model for category discrimination. Figure 11 shows the heat maps generated by each model using Grad-CAM in different categories of JSD.

FIGURE 11
www.frontiersin.org

Figure 11 Comparison of heat maps generated by Grad-CAM for each model.

The heat map gives us a clear view of the defect areas of interest to the model. The VGG16 model focuses on a more scattered region, so its classification is the least effective. ResNet34 and DenseNet121 models have more similar heat maps, which focus on a broader range of regions and have the problem of inaccurate focus on defective regions. Swin-Tiny naturally has the advantage of long-range information dependence because it is a transformer architecture model. Hence, it not only focuses on the most expansive area but also suffers from the problem of imprecise focus on defective regions. It can be found that ConvNeXt-Tiny can focus on the defective regions of jujube more precisely compared with the other networks, which makes the network achieve better classification results. In JujubeNet, by constructing the MDC module and introducing the CBAM AM, the position attention of the model to the defective regions is further optimized. Therefore, the classification performance is improved.

5. Conclusions

SDD is a significant research content in Industry 4.0 field. In the real complex industrial environment, SDD is often faced with many challenges, such as small difference between defect imaging and background, low contrast, large variation of defect scale and diverse types, large amount of noise in defect images. Jujubes are naturally growing plants, so the problem of SDD of jujube is also related to agriculture 4.0 field. Lack of per capita resources, shortage of labor force, and urgent forms of environmental protection are scientific problems throughout the development of agricultural modernization Agriculture 4.0 is an in-depth development stage of agricultural modernization construction. Precise and intelligent agricultural production can be achieved with a higher level of intensity, precision, and coordination, and the three problems above can be fundamentally solved. The rapid increase in planting area of red jujube is in sharp contrast to the backward processing technology of post-harvest jujube. The traditional industrialization process of jujube requires a lot of manpower, and the deep processing of defective jujube can effectively save resources and reduce the pressure of environmental protection. Therefore, the application of artificial vision technology to the industrialization process of jujube production and the improvement in the level of automation can provide a beneficial reference for the combination of Agriculture 4.0 and Industry 4.0.

In the actual production environment, we collected 12000 images and created a dataset named ‘Jujube2000’, which is specifically used for the classification study of surface defects of jujube. In this paper, a ConvNeXt-based high-precision lightweight classification network named JujubeNet is proposed for the defect classification of jujubes. Firstly, a MDC module with a multi-branch structure is designed, then, the CBAM is introduced to improve the ConvNeXt model, and finally, JujubeNet is proposed. In the ablation experiment phase, the effectiveness of the MDC module and CBAM AM was verified, respectively. In this paper, a comparative experiment is carried out on the ‘Jujube2000’ dataset with the underlying network in the relevant papers. By comparison, the performance advantage of JujubeNet is verified. The results show that our model exhibits better recognition accuracy and the FLOPS and number of parameters are much lower than the other models with the same performance, proving the effectiveness of the improved method proposed in this paper. In addition, some cases of classification errors were analyzed by confusion matrix and visualized. Future work will continue to study these difficult samples and further optimize the algorithm. The research results in this paper are not only applicable to the defect classification of jujubes but also can be extended to other defect classification scenarios.

Data availability statement

The original contributions presented in the study are included in the article/Supplementary Materials. Further inquiries can be directed to the corresponding author.

Author contributions

LJ and BY conceptualized of this study, selected the algorithm, designed and performed the experiment, analyzed the data, trained the algorithms, and wrote the original draft. WM collected data and revised the manuscript. YW and WM conceived the study and participated in its design. All authors contributed to this article and approved the submitted version.

Funding

This research was funded by the foundation of Shaanxi Key Laboratory of Integrated and Intelligent Navigation [SKLIIN-20190102]. Natural Science Foundation of Shaanxi Province [2021JM-537, 2019JQ-936, 2021GY-341]

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fpls.2022.1108437/full#supplementary-material

References

Altalak, M., Uddin, M. A., Alajmi, A., Rizg, A. (2022). A hybrid approach for the detection and classification of tomato leaf diseases. Appl. Sci. 12 (16), 8182. doi: 10.3390/app12168182

CrossRef Full Text | Google Scholar

Araújo, S. O., Peres, R. S., Barata, J., Lidon, F., Ramalho, J. C. (2021). Characterising the agriculture 4.0 landscape–emerging trends, challenges and opportunities. Agronomy 11 (4), 667. doi: 10.3390/agronomy11040667

CrossRef Full Text | Google Scholar

Bhargava, A., Bansal, A., Goyal, V. (2022). Machine learning–based detection and sorting of multiple vegetables and fruits. Food Anal. Methods 15 (1), 228–242. doi: 10.1007/s12161-021-02086-1

CrossRef Full Text | Google Scholar

Chen, J., Liu, X., Li, Z., Qi, A., Yao, P., Zhou, Z., et al. (2017). A review of dietary ziziphus jujuba fruit (Jujube): Developing health food supplements for brain protection. Evidence-Based Complement. Altern. Med. 24, 1–10. doi: 10.1155/2017/3019568

CrossRef Full Text | Google Scholar

Chen, T., Wang, Y., Xiao, C., Wu, Q. J. (2016). A machine vision apparatus and method for can-end inspection. IEEE Trans. Instrument. Measurement 65 (9), 2055–2066. doi: 10.1109/TIM.2016.2566442

CrossRef Full Text | Google Scholar

Da Silveira, F., Lermen, F. H., Amaral, F. G. (2021). An overview of agriculture 4.0 development: Systematic review of descriptions, technologies, barriers, advantages, and disadvantages. Comput. Electron. Agric. 189, 106405. doi: 10.1016/j.compag.2021.106405

CrossRef Full Text | Google Scholar

Deitsch, S., Christlein, V., Berger, S., Buerhop-Lutz, C., Maier, A., Gallwitz, F., et al. (2019). Automatic classification of defective photovoltaic module cells in electroluminescence images. Solar Energy 185, 455–468. doi: 10.1016/j.solener.2019.02.067

CrossRef Full Text | Google Scholar

Deng, J., Dong, W., Socher, R., Li, L.-J., Li, K., Fei-Fei, L. (2009). “ImageNet: A large-scale hierarchical image database,” in 2009 IEEE Conference on Computer Vision and Pattern Recognition. (Miami, FL, USA; IEEE), 248–255. doi: 10.1109/CVPR.2009.5206848

CrossRef Full Text | Google Scholar

Deng, Y.-S., Luo, A.-C., Dai, M.-J. (2018). “Building an automatic defect verification system using deep neural network for pcb defect classification,” in 2018 4th International Conference on Frontiers of Signal Processing (ICFSP). (Poitiers, France; IEEE), 145–149 (IEEE). doi: 10.1109/ICFSP.2018.8552045

CrossRef Full Text | Google Scholar

Dong, Y. Y., Huang, Y. S., Xu, B. L., Li, B. C., Guo, B. (2022). Bruise detection and classification in jujube using thermal imaging and DenseNet. J. Food Process Eng. 45 (3), e13981. doi: 10.1111/jfpe.13981

CrossRef Full Text | Google Scholar

Everingham, M., Van Gool, L., Williams, C. K. I., Winn, J., Zisserman, A. (2010). The pascal visual object classes (VOC) challenge. Int. J. Comput. Vis. 88, 303–338. doi: 10.1007/s11263-009-0275-4

CrossRef Full Text | Google Scholar

Ferguson, M., Ak, R., Lee, Y.-T. T., Law, K. H. (2017). “Automatic localization of casting defects with convolutional neural networks,” in 2017 IEEE International Conference on Big Data. (Boston, MA, USA; IEEE), 1726–1735 (Big Data). doi: 10.1109/BigData.2017.8258115

CrossRef Full Text | Google Scholar

Gan, J., Li, Q., Wang, J., Yu, H. (2017). A hierarchical extractor-based visual rail surface inspection system. IEEE Sens. J. 17, 7935–7944. doi: 10.1109/JSEN.2017.2761858

CrossRef Full Text | Google Scholar

GenG, L., Ma, M., Xiao, Z., Liu, Y. (2019). Jujube classification based on a convolution neural network with multi-channel weighting and information aggregation. Food Sci. Technol. Res. 25 (5), 647–656. doi: 10.3136/fstr.25.647

CrossRef Full Text | Google Scholar

Geng, L., Xu, W., Zhang, F., Xiao, Z., Liu, Y. (2018). Dried jujube classification based on a double branch deep fusion convolution neural network. Food Sci. Technol. Res. 24 (6), 1007–1015. doi: 10.3136/fstr.24.1007

CrossRef Full Text | Google Scholar

Guo, Z., Zheng, H., Xu, X., Ju, J., Zheng, Z., You, C., et al. (2021). Quality grading of jujubes using composite convolutional neural networks in combination with RGB color space segmentation and deep convolutional generative adversarial networks. J. Food Process Eng. 44 (2), e13620. doi: 10.1111/jfpe.13620

CrossRef Full Text | Google Scholar

He, Y., Song, K., Meng, Q., Yan, Y. (2020). An end-to-End steel surface defect detection approach via fusing multiple hierarchical features. IEEE Trans. Instrument. Measurement 69, 1493–1504. doi: 10.1109/TIM.2019.2915404

CrossRef Full Text | Google Scholar

He, K., Zhang, X., Ren, S., Sun, J. (2016). “Deep residual learning for image recognition,” in Proceedings of the IEEE conference on computer vision and pattern recognition. (Las Vegas, NV, USA; IEEE), 770–778. doi: 10.1109/CVPR.2016.90

CrossRef Full Text | Google Scholar

Huang, G., Liu, Z., van der Maaten, L., Weinberger, K. Q. (2017). “Densely connected convolutional networks,” in Proceedings of the IEEE conference on computer vision and pattern recognition. (Honolulu, HI, USA; IEEE), 4700–4708. doi: 10.1109/CVPR.2017.243

CrossRef Full Text | Google Scholar

Huang, Y., Qiu, C., Guo, Y., Wang, X., Yuan, K. (2018). “Surface defect saliency of magnetic tile,” in 2018 IEEE 14th International Conference on Automation Science and Engineering (CASE). (Munich, Germany; IEEE), 612–617. doi: 10.1109/COASE.2018.8560423

CrossRef Full Text | Google Scholar

Jing, J., Liu, S., Li, P., Zhang, L. (2016). The fabric defect detection based on CIE l* a* b* color space using 2-d gabor filter. J. Textile Institute 107 (10), 1305–1313. doi: 10.1080/00405000.2015.1102458

CrossRef Full Text | Google Scholar

Ju, J., Zheng, H., Xu, X., Guo, Z., Zheng, Z., Lin, M. (2022). Classification of jujube defects in small data sets based on transfer learning. Neural Computing Appl. 34 (5), 3385–3398. doi: 10.1007/s00521-021-05715-2

CrossRef Full Text | Google Scholar

Khasawneh, N., Faouri, E., Fraiwan, M. (2022). Automatic detection of tomato diseases using deep transfer learning. Appl. Sci. 12 (17), 8467. doi: 10.3390/app12178467

CrossRef Full Text | Google Scholar

Kim, M., Jeong, J., Kim, S. (2021). ECAP-YOLO: Efficient channel attention pyramid YOLO for small object detection in aerial image. Remote Sens. 13 (23), 4851. doi: 10.3390/rs13234851

CrossRef Full Text | Google Scholar

Kyeong, K., Kim, H. (2018). Classification of mixed-type defect patterns in wafer bin maps using convolutional neural networks. IEEE Trans. Semiconductor Manufacturing 31 (3), 395–402. doi: 10.1109/TSM.2018.2841416

CrossRef Full Text | Google Scholar

Liang, Q., Zhu, W., Sun, W., Yu, Z., Wang, Y., Zhang, D. (2019). In-line inspection solution for codes on complex backgrounds for the plastic container industry. Measurement 148, 106965. doi: 10.1016/j.measurement.2019.106965

CrossRef Full Text | Google Scholar

Li, Y., Huang, H., Xie, Q., Yao, L., Chen, Q. (2018). Research on a surface defect detection algorithm based on MobileNet-SSD. Appl. Sci. 8, 1678. doi: 10.3390/app8091678

CrossRef Full Text | Google Scholar

Li, X., Li, S. (2022). Transformer help CNN see better: A lightweight hybrid apple disease identification model based on transformers. Agriculture 12 (6), 884. doi: 10.3390/agriculture12060884

CrossRef Full Text | Google Scholar

Li, Y., Ma, B., Hu, Y., Yu, G., Zhang, Y. (2022). Detecting starch-head and mildewed fruit in dried hami jujubes using Visible/Near-infrared spectroscopy combined with MRSA-SVM and oversampling. Foods 11 (16), 2431. doi: 10.3390/foods11162431

PubMed Abstract | CrossRef Full Text | Google Scholar

Li, L., Ma, W., Li, L., Lu, C. (2019). Research on detection algorithm for bridge cracks based on deep learning. ActaAutom Sin. 45 (9), 1727–1742. doi: 10.16383/j.aas.2018.c170052

CrossRef Full Text | Google Scholar

Lin, J., Chen, X., Pan, R., Cao, T., Cai, J., Chen, Y., et al. (2022). GrapeNet: A lightweight convolutional neural network model for identification of grape leaf diseases. Agriculture 12 (6), 887. doi: 10.3390/agriculture12060887

CrossRef Full Text | Google Scholar

Lin, T.-Y., Maire, M., Belongie, S., Hays, J., Perona, P., Ramanan, D., et al. (2014). “Microsoft COCO: Common objects in context,” in Computer vision – ECCV 2014 lecture notes in computer science. Eds. Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T. (Cham: Springer International Publishing), 740–755. doi: 10.1007/978-3-319-10602-1_48

CrossRef Full Text | Google Scholar

Liu, Z., Liu, K., Zhong, J., Han, Z., Zhang, W. (2020). A high-precision positioning approach for catenary support components with multiscale difference. IEEE Trans. Instrument. Measurement 69, 700–711. doi: 10.1109/TIM.2019.2905905

CrossRef Full Text | Google Scholar

Liu, Z., Mao, H., Wu, C.-Y., Feichtenhofer, C., Darrell, T., Xie, S. (2022). “A convnet for the 2020s,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. (New Orleans, LA, USA; IEEE), 11976–11986. doi: 10.1109/CVPR52688.2022.01167

CrossRef Full Text | Google Scholar

Liu, Y., Ma, X., Shu, L., Hancke, G. P., Abu-Mahfouz, A. M. (2020). From industry 4.0 to agriculture 4.0: Current status, enabling technologies, and research challenges. IEEE Trans. Ind. Inf. 17 (6), 4322–4334. doi: 10.1109/TII.2020.3003910

CrossRef Full Text | Google Scholar

Ma, L., Xie, W., Zhang, Y. (2019). Blister defect detection based on convolutional neural network for polymer lithium-ion battery. Appl. Sci. 9 (6), 1085. doi: 10.3390/app9061085

CrossRef Full Text | Google Scholar

Mayr, M., Hoffmann, M., Maier, A., Christlein, V. (2019). “Weakly supervised segmentation of cracks on solar cells using normalized lp norm,” in 2019 IEEE International Conference on Image Processing (ICIP). (Taipei, Taiwan; IEEE), 1885–1889. doi: 10.1109/ICIP.2019.8803116

CrossRef Full Text | Google Scholar

Park, J.-K., Kwon, B.-K., Park, J.-H., Kang, D.-J. (2016). Machine learning-based imaging system for surface defect inspection. Int. J. Precis. Eng. Manufacturing Green Technol. 3 (3), 303–310. doi: 10.1007/s40684-016-0039-x

CrossRef Full Text | Google Scholar

Penumuru, D. P., Muthuswamy, S., Karumbu, P. (2020). Identification and classification of materials using machine vision and machine learning in the context of industry 4.0. J. Intelligent Manufacturing 31 (5), 1229–1241. doi: 10.1007/s10845-019-01508-6

CrossRef Full Text | Google Scholar

Pham, Q. T., Liou, N.-S. (2020). Hyperspectral imaging system with rotation platform for investigation of jujube skin defects. Appl. Sci. 10 (8), 2851. doi: 10.3390/app10082851

CrossRef Full Text | Google Scholar

Rashwan, A. K., Karim, N., Shishir, M. R. I., Bao, T., Lu, Y., Chen, W. (2020). Jujube fruit: A potential nutritious fruit for the development of functional food products. J. Funct. Foods 75, 104205. doi: 10.1016/j.jff.2020.104205

CrossRef Full Text | Google Scholar

Simonyan, K., Zisserman, A. (2014). Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv 1409, 1556. doi: 10.48550/arXiv.1409.1556

CrossRef Full Text | Google Scholar

Soukup, D., Huber-Mörk, R. (2014). “Convolutional neural networks for steel surface defect detection from photometric stereo images,” in International symposium on visual computing, (Berlin, Germany; Springer), 668–677. doi: 10.1007/978-3-319-14249-4_64

CrossRef Full Text | Google Scholar

Sun, X., Ma, L., Li, G. (2019). Multi-vision attention networks for on-line red jujube grading. Chin. J. Electron. 28 (6), 1108–1117. doi: 10.1049/cje.2019.07.014

CrossRef Full Text | Google Scholar

Tabernik, D., Šela, S., Skvarč, J., Skočaj, D. (2020). Segmentation-based deep-learning approach for surface-defect detection. J. Intell. Manuf 31, 759–776. doi: 10.1007/s10845-019-01476-x

CrossRef Full Text | Google Scholar

Tang, S., He, F., Huang, X., Yang, J. (2019) Online PCB defect detector on a new PCB defect dataset. Available at: http://arxiv.org/abs/1902.06197 (Accessed December 21, 2022).

Google Scholar

Tao, X., Zhang, D., Wang, Z., Liu, X., Zhang, H., Xu, D. (2020). Detection of power line insulator defects using aerial images analyzed with convolutional neural networks. IEEE Trans. Systems Man Cybernetics: Syst. 50, 1486–1498. doi: 10.1109/TSMC.2018.2871750

CrossRef Full Text | Google Scholar

Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., et al. (2017). “Attention is all you need,” in Advances in neural information processing systems, (Long Beach, CA, USA; NeurIPS), vol. 30. .

Google Scholar

Wen, H., Wang, J., Han, F. (2020). Research on defect detection and classification of jujube based on improved residual network. Food Machinery 36 (1), 161–165. doi: 10.13652/j.issn.1003-5788.2020.01.028

CrossRef Full Text | Google Scholar

Woo, S., Park, J., Lee, J.-Y., Kweon, I. S. (2018). “Cbam: Convolutional block attention module,” in Proceedings of the European conference on computer vision (ECCV), (Munich, Germany; Springer), Vol. 3-19. doi: 10.1007/978-3-030-01234-2_1

CrossRef Full Text | Google Scholar

Wu, L., He, J., Liu, G., Wang, S., He, X. (2016). Detection of common defects on jujube using vis-NIR and NIR hyperspectral imaging. Postharvest Biol. Technol. 112, 134–142. doi: 10.1016/j.postharvbio.2015.09.003

CrossRef Full Text | Google Scholar

Xie, S., Girshick, R., Dollár, P., Tu, Z., He, K. (2017). “Aggregated residual transformations for deep neural networks,” in Proceedings of the IEEE conference on computer vision and pattern recognition. (Honolulu, HI, USA; IEEE), 1492–1500. doi: 10.1109/CVPR.2017.634

CrossRef Full Text | Google Scholar

Xue, Y., Li, Y. (2018). A fast detection method via region-based fully convolutional neural networks for shield tunnel lining defects. Computer Aided Civil Infrastructure Eng. 33, 638–654. doi: 10.1111/mice.12367

CrossRef Full Text | Google Scholar

Yang, X., Li, H., Yu, Y., Luo, X., Huang, T., Yang, X. (2018). Automatic pixel-level crack detection and measurement using fully convolutional network. Computer Aided Civil Infrastructure Eng. 33, 1090–1109. doi: 10.1111/mice.12412

CrossRef Full Text | Google Scholar

Yang, S., Wang, Y., Wang, P., Mu, J., Jiao, S., Zhao, X., et al. (2022). Automatic identification of landslides based on deep learning. Appl. Sci. 12 (16), 8153. doi: 10.3390/app12168153

CrossRef Full Text | Google Scholar

Yin, C., Zeng, T., Zhang, H., Fu, W., Wang, L., Yao, S. (2022). Maize small leaf spot classification based on improved deep convolutional neural networks with a multi-scale attention mechanism. Agronomy 12 (4), 906. doi: 10.3390/agronomy12040906

CrossRef Full Text | Google Scholar

Yu, Y. (2022). Study on jujube classification method based on visual attention mechanism (MA thesis; Tarim University; Xinjiang China). doi: 10.27708/d.cnki.gtlmd.2022.000127

CrossRef Full Text | Google Scholar

Zagoruyko, S., Komodakis, N. (2016). Wide residual networks. arXiv preprint arXiv 1605, 07146. doi: 10.5244/C.30.87

CrossRef Full Text | Google Scholar

Zhai, Z., Martínez, J. F., Beltran, V., Martínez, N. L. (2020). Decision support systems for agriculture 4.0: Survey and challenges. Comput. Electron. Agric. 170, 105256. doi: 10.1016/j.compag.2020.105256

CrossRef Full Text | Google Scholar

Zhang, C., Chang, C., Jamshidi, M. (2020). Concrete bridge surface damage detection using a single-stage detector. Computer Aided Civil Infrastructure Eng. 35, 389–409. doi: 10.1111/mice.12500

CrossRef Full Text | Google Scholar

Zhang, Z., Wen, G., Chen, S. (2019). Weld image deep learning-based on-line defects detection using convolutional neural networks for Al alloy in robotic arc welding. J. Manufacturing Processes 45, 208–216. doi: 10.1016/j.jmapro.2019.06.023

CrossRef Full Text | Google Scholar

Zhong, Y., Huang, B., Tang, C. (2022). Classification of cassava leaf disease based on a non-balanced dataset using transformer-embedded ResNet. Agriculture 12 (9), 1360. doi: 10.3390/agriculture12091360

CrossRef Full Text | Google Scholar

Zhu, N., Liu, X., Liu, Z., Hu, K., Wang, Y., Tan, J., et al. (2018). Deep learning for smart agriculture: Concepts, tools, applications, and opportunities. Int. J. Agric. Biol. Eng. 11 (4), 32–44. doi: 10.25165/j.ijabe.20181104.4475

CrossRef Full Text | Google Scholar

Keywords: agriculture 4.0, industry 4.0, artificial vision, CBAM, ConvNeXt

Citation: Jiang L, Yuan B, Ma W and Wang Y (2023) JujubeNet: A high-precision lightweight jujube surface defect classification network with an attention mechanism. Front. Plant Sci. 13:1108437. doi: 10.3389/fpls.2022.1108437

Received: 26 November 2022; Accepted: 29 December 2022;
Published: 18 January 2023.

Edited by:

Roberto Ciccoritti, Council for Agricultural Research and Economics, Italy

Reviewed by:

Nisha Pillai, Mississippi State University, United States
Francesca Antonucci, CREA - Centro di ricerca Ingegneria e Trasformazioni agroalimentari, Italy

Copyright © 2023 Jiang, Yuan, Ma and Wang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Baoxi Yuan, ybxbupt@163.com

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.