- 1School of Information Engineering, Henan Institute of Science and Technology, Xinxiang, China
- 2School of Artificial Intelligence, Beijing University of Posts and Telecommunications, Beijing, China
- 3School of Internet, Anhui University, Hefei, China
Introduction: Efficient and accurate varietal classification of wheat grains is crucial for maintaining varietal purity and reducing susceptibility to pests and diseases, thereby enhancing crop yield. Traditional manual and machine learning methods for wheat grain identification often suffer from inefficiencies and the use of large models. In this study, we propose a novel classification and recognition model called SCGNet, designed for rapid and efficient wheat grain classification.
Methods: Specifically, our proposed model incorporates several modules that enhance information exchange and feature multiplexing between group convolutions. This mechanism enables the network to gather feature information from each subgroup of the previous layer, facilitating effective utilization of upper-layer features. Additionally, we introduce sparsity in channel connections between groups to further reduce computational complexity without compromising accuracy. Furthermore, we design a novel classification output layer based on 3-D convolution, replacing the traditional maximum pooling layer and fully connected layer in conventional convolutional neural networks (CNNs). This modification results in more efficient classification output generation.
Results: We conduct extensive experiments using a curated wheat grain dataset, demonstrating the superior performance of our proposed method. Our approach achieves an impressive accuracy of 99.56%, precision of 99.59%, recall of 99.55%, and an F1-score of 99.57%.
Discussion: Notably, our method also exhibits the lowest number of Floating-Point Operations (FLOPs) and the number of parameters, making it a highly efficient solution for wheat grains classification.
1 Introduction
Wheat, being one of the most extensively cultivated crops globally (Li et al., 2019; Zhou et al., 2021), holds vital genetic and morphological information within its seeds. The distinct characteristics and values exhibited by different wheat grain varieties underscore the importance of selecting high quality varieties. This selection is not only pivotal for augmenting wheat yields and enhancing quality but also crucial for safeguarding crops against pests and diseases (Mefleh et al., 2019; Saeed et al., 2022).
The purity of wheat grain varieties is of paramount importance to breeding specialists, wheat cultivators, and consumers at large (Hussain et al., 2022). Unfortunately, the integrity of seed markets faces challenges from unscrupulous traders who engage in deceptive practices. They market low-quality seed varieties as high-quality ones, posing a threat to consumers and disrupting the seed market. To counteract such issues, accurate classification techniques are imperative (Fanelli et al., 2023).
Historically, professionals relied on traditional methods for varietal identification of wheat grains. However, these methods are slow, labor-intensive, and susceptible to subjective biases. The inherent similarity in the characteristics of various wheat grains further complicates the identification process.
In recent years, the integration of computer vision techniques into wheat grain recognition has witnessed significant advancements (Li et al., 2019). Researchers have explored two primary types of approaches for feature extraction and classification: machine learning-based methods and deep learning-based methods.
Machine learning methods, while effective, require substantial agricultural knowledge, manual feature selection, and classifier design (Lu et al., 2022). This process demands significant human effort and may not match the recognition speed achieved by deep learning approaches.
Deep learning methods offer notable advantages, automating feature extraction and achieving superior classification accuracy. They exhibit strong generalization capabilities, streamlining model training and significantly enhancing recognition speed. However, challenges such as the need for extensive training data and the high number of parameters in deep learning models can impede deployment on resource-constrained devices. For instance, the computational intensity associated with these models can overwhelm devices with limited resources, leading to frequent crashes during usage. The sheer volume of computations required may exceed the processing capacity of these devices, compromising their stability and usability. Moreover, resource-constrained devices may lack the storage capacity necessary to accommodate the extensive parameters of these models, rendering deployment infeasible.
To address these challenges, we propose the Sparsely Connected Group Convolution Network (SCGNet) for efficient and accurate wheat grain classification. Our model is designed to offer a non-destructive, efficient, and rapid classification solution, aligning with the overarching goal of addressing the complexities associated with wheat grain identification and classification. We highlight the key contributions of this paper as follows:
● We introduce a novel approach known as “Group Mixing(GM)”, which involves splitting and rearranging group convolutions based on a strategic criterion. This innovative technique resolves issues related to information exchange among groups, enhances feature multiplexing, and simultaneously reduces the Floating-Point Operations (FLOPs) of the convolutional layers.
● We present a method for connected group convolutions, called “Sparsely Connected(SC)”, facilitating the cascading transfer of feature information between groups without compromising vital details. This advancement further decreases the computational demands of the convolutional layers.
● We incorporate 3-D convolution and revamp the convolutional classification layer within SCGNet. This novel approach replaces traditional layers like pooling and fully connected layers commonly found in conventional convolutional neural networks (CNNs). The result is a reduction in the overall number of model parameters, leading to a more streamlined architecture and faster recognition.
The structure of this paper is organized as follows: Section 2 provides a brief review and summary of various methods employed for wheat grain recognition, along with the challenges they address. In Section 3, we present comprehensive details regarding the proposed SCGNet architecture. Section 4 encompasses our creation of a wheat grains dataset, outlines our experimental procedures, and presents the experimental results. We also perform an in-depth analysis and comparison of classification outcomes generated by various network models using the wheat grain dataset. Finally, in Section 5, we summarize the primary contributions of this paper and engage in a discussion regarding potential directions for future research.
2 Related works
Currently, various identification methods have been gradually applied to wheat grain classification, and in the following, we provide an overview and summary of these studies and summarize the advantages and disadvantages of all methods in Table 1.
2.1 Machine learning-based methods
Machine learning-based methods leverage digital image processing techniques to preprocess data acquired from collected images, followed by manual feature design, feature extraction, and ultimately, classification and recognition employing suitable classifiers like Support Vector Machine (SVM). For instance, Delwiche et al. (Delwiche et al., 2013) employed optical-grade reflectors to capture wheat grain images for assessing surface damage. They parameterized kernel morphology and texture features from both main and reflected views, employing parametric (Linear Discriminant Analysis, LDA) and non-parametric (k-Nearest Neighbors, KNN) classification models, respectively. This approach achieved a recognition accuracy of up to 94%. Güneş et al. (Güneş et al., 2014) proposed a method for recognizing wheat varieties using digital image processing techniques. Their system extracted image features using the Gray Level Covariance Matrix (GLCM) and Linear Binary Pattern (LBP) methods, classifying them with a k-nearest neighbor classifier. Kurtulmuş et al. (Kurtulmuş et al., 2016) introduced a recognition method combining machine vision and neural networks. They calculated features from different color components and constructed a feature database using chili pepper seeds as the study object. Sequential feature selection methods with various criterion functions were employed to select effective features, achieving variety classification of eight pepper seeds with a Multilayer Perceptron (MLP) accuracy of 84.94%. Sabanci et al. (Sabanci et al., 2017) extracted four shape features, three color features, and five texture features, inputting these features into an artificial neural network (ANN) constructed as a multilayer perceptron (MLP), resulting in improved classification results. Ni et al. (Ni et al., 2019)designed an automatic maize surface defect inspection system. Initially, they pre-processed touching kernels using a novel k-means clustering guided curvature method, enhancing the identification of broken kernels and system robustness. Subsequently, they integrated a deep convolutional network into the system for detecting maize surface defects, achieving an accuracy of 98.2%.
2.2 Deep learning-based methods
Deep learning-based methods typically involve the construction of specialized deep learning models for recognition and classification tasks (Huang et al., 2017; Zhang et al., 2022b; Zhang et al., 2023). In these approaches, the deep learning model takes the original image data as input, processes it at the pixel level, and automatically extracts contextual information and global features from the image by employing various combinations of convolution and pooling operations. Finally, the model produces classification and recognition results through specific functions. For instance, Kozłowski et al. (Kozłowski et al., 2019) conducted a comparison of nine different CNNs for wheat grain classification. They used reference performance indicators such as training time, inference speed, and accuracy rates and compared them with traditional machine learning methods. The results showed that traditional methods achieved a relatively low classification accuracy of around 75%, whereas CNN methods achieved an accuracy exceeding 93%. Javanmardi et al. (Javanmardi et al., 2021) proposed a method that utilizes CNNs as generalized feature extractors, combined with artificial neural networks, for feature extraction and classification. They tested this approach on 2250 test samples, achieving a correct classification rate of 98.1% with a total processing time of 26.8 seconds. Deep learning-based methods can achieve satisfactory results in terms of accuracy metrics, but it is worth noting that these high-precision CNNs are accompanied by a high number of parameters and FLOPs, and thus some scholars have focused on how to thin the models.
One noteworthy approach is MobileNet (Howard et al., 2017), which introduces depth-wise separable convolutions, which split standard convolutions into depth-wise convolutions and point-wise convolutions, this reduces the number of parameters and computations significantly. ShuffleNet (Zhang et al., 2018) employs group convolutions and channel shuffling to enhance the fusion of channel information while reducing computational cost, group convolutions split the input channels into separate groups, reducing the complexity of convolutions. MobileNetV2 (Sandler et al., 2018) utilizes the inverted residuals structure, which helps in maintaining a balance between computational efficiency and representational power, it uses linear bottlenecks and shortcut connections to improve information flow. Moreover, BiSeNetV2 (Yu et al., 2021) presented a branching network where the detail branch focused on underlying details using a larger spatial dimension, while the semantic branch captured advanced semantics with large convolutional kernels, these branches were then fused through an aggregation layer, enhancing the model’s capabilities. Inspired by these innovations, Yang et al. (Yang et al., 2021a) devised a branch network by modifying the VGG16 model. By removing the fully connected layer and adjusting the position of the Batch Normalization (BN) layer, they crafted a novel network capable of classifying peanut varieties. This tailored model exhibited remarkable accuracy improvements over the original design while maintaining a reduced parameter count. Zhao et al. (Zhao et al., 2022) employed YOLOv5 for detecting the quality of wheat grains and introduced a lightweight wheat grain detection network, WGNet, based on YOLOv5. WGNet utilized the FPN neck module and hybrid attention module to address performance degradation issues and reduced network parameters through network pruning. This approach significantly improved inference speed while maintaining high detection accuracy.
2.3 Methods based on hyperspectral imaging combined with deep learning
Neural networks have demonstrated outstanding performance, prompting some scholars to explore their combination with hyperspectral imaging techniques. For instance, Weng et al. (Weng et al., 2021) aimed to characterize the degree of Fusarium head blight (FHB) infestation on wheat grains. They extracted reflectance spectra from hyperspectral images of healthy and FHB-infected wheat grains with varying levels of infestation (light, moderate, heavy). Five effective wavelengths (EWs) were randomly selected from the spectra, and different combinations of EWs were used to generate reflectance images (RIs) with LeNet-5. Additionally, a residual attention convolutional neural network (RACNN) was constructed, increasing width and depth, and incorporating channel attention and residual modules to recognize varying degrees of FHB infection in wheat grains. Shen et al. (Shen et al., 2021) proposed a spectral imaging-based method for detecting impurities in wheat. They employed spectral imaging to study the spectral features of the target data and converted them into frequency domain spectra for terahertz pseudo-color imaging of wheat and its impurities. This was combined with a CNN to create a model called Wheat-V2, designed for identifying impurities in wheat images. Zhang et al. (Zhang et al., 2022a) utilized 2D convolution with an attentional mechanism to extract spatial and textural features, while 3-D convolution was used for spatial and inter-spectral information extraction in maize cultivar identification. This combination of 2-D and 3-D convolution showed good feature extraction complementarity. However, classification methods based on hyperspectral imaging often require a substantial amount of data and high-quality equipment for data acquisition. This instrument-dependent nature can be a burden for economically underdeveloped regions where the high cost of multispectral and hyperspectral cameras is prohibitive for widespread adoption. To address this, some scholars (Yang et al., 2021b) designed the Spectrogram Generative Adversarial Network (SPGAN) to expand the wheat grain dataset. SPGAN utilizes a generative adversarial network to generate synthetic datasets based on a small set of real datasets. These synthetic datasets serve as the foundation for the Progressive Neural Structure Search (PNAS) generative network structure, which classifies three types of wheat grains. The SPGAN-PNAS framework achieved an F1-score of 96.2%, outperforming traditional neural networks.
To summarize, machine learning-based methods do not require much data to have better performance although they need to extract features manually, while deep learning-based methods require a large amount of data to support them, at the same time, deep learning-based methods eliminate the need to manually design and extract features, which means that researchers do not need to have a richer background in agricultural knowledge. Hyperspectral-based methods combined with deep learning can extract richer feature information but require sophisticated data acquisition equipment.
3 Proposed SCGNET
Figure 1 provides an exhaustive overview of the proposed innovative SCGNet architecture. The left segment of the figure illustrates the overall structure of SCGNet, while the central part delves into the constituent sub-modules, namely the Downsample layer and SCG block. The right-hand portion zooms in further to unveil the sub-components of the SCG block. SCGNet is a comprehensive network composed of various elements, including a conventional convolutional layer, multiple repetitively stacked Downsample layers, SCG blocks, and a 3-D classification convolutional layer. The primary objective of these components is to capture and process the intricate features of the input image data.
Figure 1 Given a 224×224×3 image of wheat grain, it passes through an initial convolutional layer to extract the coarse-grained features. Following that, we stack four successive SCG blocks to obtain the fine-grained features of the image and reduce the parameters. In addition, we add down sampling modules before each SCG block module to ensure image size consistency in the computation process. Finally, the output classification result is determined by a 3-D convolutional classification layer that we have constructed.
The Conventional Convolutional Layer utilizes a 3×3 convolutional kernel with a stride of 1. Its role is to perform an initial extraction of the coarse-grained features inherent in the image data.
The Downsample Layer is crucial in reducing image size and controlling channel dimensionality. It consists of two convolutional layers: one using a 1×1 kernel with a stride of 1, and the other utilizing a 2×2 kernel with a stride of 2. This combination enables control of image size and channel simultaneously.
The SCG block represents the core of our architecture, consisting of a series of repetitively stacked down sampling layers and SCG block components(GM module and SC module). More details about the important sub-modules of the SCG block: the GM module and the SC module, will be elaborated upon in subsections 3.1.1 and 3.1.2. Additionally, it incorporates two 1×1 ordinary convolutional layers to regulate channel dimensions. The primary purpose of the SCG block is to extract fine grained features from the image and simultaneously reduce the overall parameter count of the entire SCGNet, thereby enhancing efficiency.
The 3-D Convolutional Classification Layer is specifically designed for classification tasks and employs 3-D convolutions to produce the final classification results.
In Table 2, we present a detailed breakdown of each module within SCGNet, offering a comprehensive reference for the configuration and specifications of our network’s components.
3.1 SCG block
The SCG block serves as the main module of the entire network. Within the SCG block, we integrate the Group Mixing module and the Sparsely Connected module, along with two convolutional layers using 1×1 kernel sizes, to constitute the SCG block. These two 1×1 kernel convolutional layers play distinct roles: one for increasing and the other for decreasing the channel dimensionality. The GM module and SC module serve as the core components within the SCG block. Following the Depthwise Convolution (DW Conv), we introduce a BN layer to normalize the data, thus contributing to the potential acceleration of CNN training. The 1×1 Group convolution (G-Conv) primarily serves the purposes of parameter sharing and feature interaction among subgroups.
This aids in reducing the model’s parameter count while facilitating mutual feature learning among subgroups. After the G-Conv layer, we not only apply the BN layer but also opt for the Swish activation function over the traditional ReLU. The Swish activation function has a smoothness that enhances the forward propagation optimization, in addition, the function exhibits a high saturation threshold, which remains unsaturated even when the inputs converge to 0, thus facilitating the flow of gradients during the training process.
In various deep learning architectures, the concept of feature fusion is crucial. Such as ResNet (He et al., 2015) and FPN (Lin et al., 2017), these architectures often employ the element-wise Add operation for feature fusion.
The Add operation directly combines the matrix information from input features and output features without altering the image’s dimensionality. The number of channels remains the same, but the operation increases the amount of information along each dimension. However, in certain cases, such as ShuffleNet (Zhang et al., 2018), a Concat operation is used instead of Add during feature fusion.
The Concat operation, unlike Add, applies different weights to feature maps and then merges them based on the number of channels in the input matrix. This can increase the image’s dimensionality while preserving information along each dimension. The Concat operation aligns input features with the output feature map and leverages the semantic information from feature maps of different scales to achieve superior performance by expanding the number of channels. Therefore, to leverage semantic information from feature maps of different scales and increase dimensionality, we employ the Concat for feature fusion.
3.1.1 Group mixing
Traditional CNNs primarily consist of convolutional layers, activation functions, pooling layers, and fully-connected layers (Krizhevsky et al., 2012). The trainable layers within CNNs typically comprise convolutional layers and fully-connected layers (Gao et al., 2018). Among them, the main role of the convolutional layer is to perform feature extraction on the input image. Each neuron in a convolutional layer is connected to multiple neurons in spatially proximate regions of the preceding layer. The convolution operation involves sequentially applying a convolution kernel to the input features through element-wise matrix multiplication and aggregating the results while incorporating biases.
The presence of a large number of convolution operations in convolutional layers leads to a substantial increase in the number of parameters and FLOPs. To illustrate this, let’s define the input feature map as F ∈ ℝh×w×c, and the convolution kernel as K ∈ ℝh×w×c, with Kn representing the number of convolution kernels. A standard convolutional operation is performed between the feature map and Kn convolutional kernels, with a default stride of 1. The number of parameters for this operation is calculated as Equation 1:
At this point, the FLOPs are determined as Equation 2:
Here, Fh and Fw represent the height and width of the input feature map, respectively. Kh and Kw denote the height and width of the convolution kernel, Fc is the number of channels in the feature map, and Kc is the number of channels in the convolution kernel. in which case, Kc = Fc. To address this computational complexity, MobileNet (Howard et al., 2017) introduced depthwise separable convolution. This technique divides the convolution operation into two steps: depthwise convolution and pointwise convolution. In the depthwise convolution, a single convolution kernel is applied independently to each channel in the depth direction of the feature map. The outputs are then concatenated to generate the same number of output channels, followed by pointwise convolution using a 1×1 unit convolution kernel. With this decomposition, the number of convolutional parameters as Equation 3:
And the FLOPs as Equation 4:
Clearly, there is a significant reduction in both parameters and FLOPs after applying depthwise convolution and pointwise convolution. However, constrained by the computational power of GPUs, there is a need to further reduce parameters and FLOPs. To address this, AlexNet (Krizhevsky et al., 2012) introduced group convolution. This technique involves grouping different feature maps of the input layer and applying different convolution kernels to each group. Group convolution has been successfully employed in various networks, including Xception (Chollet, 2017), MobileNet (Howard et al., 2017), ResNeXt (Xie et al., 2017), and others, demonstrating excellent performance.
Group convolution is a technique used in CNNs to reduce the computational cost of convolutional layers. It divides the input feature map into mutually exclusive groups based on channels, where each group operates with a 1×1 convolution kernel. This division results in each group having a subset of the input channels, with a proportionate reduction in the number of parameters. The number of parameters and FLOPs are reduced to 1/G of the original values, where G is the number of groups.
However, group convolution also brings an Issue of Independence. While group convolution significantly reduces computational requirements, it has a drawback. The feature information in each subgroup is relatively independent, and there is limited interaction between the groups. This can lead to a lack of effective information exchange between channels.
To address the issue of independence and enhance information exchange between the groups, we propose a “Group Mixing” approach. First, group convolution is divided into G primary groups (Gi). Each primary group contains a subset of the channel’s feature information. Then, each primary group (Gi) is further divided into j subgroups (), where j ranges from 1 to i. This secondary division allows for a more fine-grained separation of channel information within each primary group. Finally, the critical step in Group Mixing involves taking one subgroup () from each primary group (Gi) and combining them in an ordered manner. These subgroups are concatenated to create new subgroups () in a way that disrupts and recombines the feature information.
By using the Group Mixing method, feature information from each primary group is mixed in an orderly manner to generate new groups (). For example, G1 contains information from all . This process enhances the interaction and information exchange between different groupings and channels. For the new G1, it comprises each component, and each new mixed group is defined as Equation 5:
Group Mixing is a strategy to balance computational efficiency (achieved through group convolution) with the need for information exchange and interaction between feature channels, particularly in the context of group convolution. This helps in maintaining the representational power of the network while reducing computational complexity. It disrupts and recombines feature information in an ordered manner, allowing for more effective interaction between subgroups, thus addressing the issue of independence observed in group convolution. Figure 2 provides a visual representation of our proposed method.
3.1.2 Sparsely connected
Traditionally, in CNNs, the output of group convolutions is connected to the subsequent layer in a manner that resembles a fully connected layer, as depicted in Figure 3A. This design choice is made to ensure that most of the feature information is preserved since there is typically no information exchange among individual subgroups. As a result, dense connections are used to pass feature information to the next layer. However, with the introduction of “Group Mixing” as discussed in Section 3.1.1, the problem of information exchange between group convolutions has already been addressed to a significant extent, making the dense connections unnecessary. In the essence of convolutional operations, where a feature map is convolved with a kernel, the operation spans both height and width dimensions, constituting a spatial convolution. The 1×1 convolution operation, often utilized in CNNs, is equivalent to a fully connected operation. Building on this understanding, Sparsely Connected offers an alternative perspective on the convolution operation, specifically in the channel dimension.
Figure 3 Difference between Fully Connected and Sparsely Connected. (A) Fully Connected, (B) Sparsely Connected.
Building on our previous discussion, when the input feature map F ∈ ℝh×w×c is convolved with the convolution kernel K ∈ ℝh×w×c, this convolution is equivalent to a fully connected operation in both the spatial dimensions and the channel dimension. With this understanding, we introduce “ Sparsely Connected,” a method primarily aimed at reducing the number of fully connected operations. When the input feature map F ∈ ℝh×w×c is convolved with the convolution kernel K ∈ ℝh×w×c in the channel dimension, we no longer perform a fully connected operation. Instead, we adopt a sparsely connected approach in the channel dimension by employing a certain stride. Simply put, the feature map is convolved with only a part of the convolution kernels. For instance, with stride=3, 2 convolution kernels after one convolution operation are discarded. Regarding the choice of stride, we conducted a series of comparative experiments in the ablation study in Section 4.5, as shown in Table 3, verifying that the best performance is achieved when stride = 3. Also, as the stride gets larger, the model exhibits worse performance.
Table 3 Results of the sparsely connected method when different strides are taken for ablation studies of SC modules.
By implementing Sparsely Connected, we eliminate the necessity for Fc to be multiplied by Kn, thereby significantly reducing the computational burden associated with convolution operations while preserving the essential information required for subsequent processing.
Based on the previous description, we can integrate “Sparsely Connected” with “Group Mixing”.
In the after of Group Mixing, we obtain Kn feature maps following group convolution. Subsequently, we partition these Kn feature maps into g groups and employ g independent convolution kernels K ∈ ℝh×w×c. The step size between feature maps and convolution kernels for each convolution operation is set to g, and Kc ≤ Fc. By performing convolution on the entire input Kn/g times, the number of feature maps is reduced to Kn/g. At this juncture, the parameter count for convolutional computation is calculated as Equation 6:
The FLOPs are determined as Equation 7:
Clearly, by incorporating the “Sparsely Connected” approach, we further reduce the number of parameters in the CNN, enabling it to operate more efficiently. Moreover, this approach facilitates the seamless transfer of feature information from one layer to the next without compromising the effectiveness of the feature information. Figure 3B nicely illustrates our proposed sparsely connected approach.
3.2 3-D convolutional classification layer
The classical structure of a CNN typically includes convolutional layers, pooling layers, and fully connected layers. Traditionally, these networks used average pooling and multiple fully connected layers. Earlier networks like AlexNet (Krizhevsky et al., 2012), VGGNet (Simonyan and Zisserman, 2015), and GoogLeNet (Szegedy et al., 2015), for instance, featured three consecutive fully connected layers. However, these fully connected layers contained a large number of parameters due to their fully connected nature. In fact, in the case of AlexNet (Krizhevsky et al., 2012) on the ImageNet dataset, the three fully connected layers accounted for approximately 96% of the total number of parameters in the entire network, which is nearly the sum of all parameters in the network. Importantly, experimental results demonstrated that this design did not substantially compromise the classification performance of the CNN.
Many studies have revealed that the weight matrix of these fully connected categorization layers is often very sparse, suggesting that only a few features are essential for category prediction. The problem of excessively large fully connected layers has garnered attention from researchers. Consequently, in recent works, (Tan and Le, 2020; Dai et al., 2021; Liu et al., 2022), the last two fully connected layers in the network structure were replaced with a global average pooling layer followed by a single fully connected layer. This modification significantly reduced the total number of parameters in the CNN. For example, in the case of the lightweight network MobileNet (Howard et al., 2017) on the ImageNet dataset, the fully connected layer accounted for approximately 24% of the total network parameters.
In the initial design of SCGNet, we also adopted this approach: replacing the last two layers with a single global pooling layer and a single fully connected layer. However, during the design process, we discovered that even with only one fully connected layer, it still accounted for a significant portion of the CNN’s parameters. This means that the number of parameters in the fully connected layer is still a large percentage of the overall CNNs. To address this, we propose a novel classification layer based on 3-D convolutional operations. This new layer completely replaces the traditional global pooling and fully connected layers, resulting in a reduction in the number of parameters and FLOPs in the network.
In a typical CNN, several convolutional layers are employed to progressively extract features essential for image recognition and classification through convolution operations. These convolutional layers output larger-sized feature maps. Subsequently, these feature maps are passed through a Global Average Pooling layer, which serves the purpose of further downsizing these large feature maps. After traversing multiple pooling layers, these feature maps are then used as inputs for the fully connected layer. The role of the fully connected layer is to connect each node to all nodes in the previous layer and ultimately output a one-dimensional vector. The classification result is obtained by applying the softmax classification function.
Based on our prior description of global pooling and fully connected layers, we can simplify the process when an input feature map F ∈ ℝh×w×c passes through the global pooling layer as follows:
Firstly, we can streamline this step by employing Fc convolution kernels, each with dimensions Fh ×Fw ×1. This implies that Fh and Fw remain consistent with Kh and Kw, respectively, while maintaining uniform weights set at 1/Fh ×Fw.
Secondly, to facilitate the seamless connection between the output of the pooling layer and the input of the fully-connected layer, we approximate the fully-connected layer by substituting it with a convolutional layer. The size of the convolutional kernel in this context becomes 1×1×Kc.
Subsequently, we amalgamate these two convolutional operations, yielding a 3-D convolution operation with a convolutional kernel size of Fh × Fw × Kc. This 3-D convolution layer is predominantly utilized in constructing the entire classification layer. This approach simplifies both the pooling layer and the fully connected layer into a single 3-D convolutional layer, as eloquently depicted in Figure 4.
When this integrated layer performs the classification task, assuming the number of categories to be classified is denoted as i, we adjust Kc to be Fc − i +1. In other words, the prediction for each category necessitates only Fc − i +1 input feature maps. This disambiguation strategy optimally conserves computational resources within the CHANNEL dimension, thereby reducing parameters and computational complexity while preserving the essential category connections for accurate prediction and ensuring efficient categorization output.
Our proposed method combines global pooling and fully connected layers into a unified 3-D convolutional layer, resulting in significant computational efficiency gains without compromising classification accuracy. By implementing this approach, we significantly reduce computational resources in the CHANNEL dimension. This reduction aids in lowering the number of parameters and computations while preserving the connectivity necessary for category prediction and maintaining efficient categorization output.
4 Experimental results and analysis
In this chapter, we begin by introducing the wheat grains dataset that we have utilized, as well as detailing the preprocessing procedures it underwent. Subsequently, we delve into an exploration of the impact of specific parameter settings on the classification capabilities of SCGNet. Following that, we conduct a comprehensive comparison with a series of CNNs commonly employed for image classification. Our objective is to evaluate and highlight the advantages of our proposed SCGNet, with a focus on key metrics such as accuracy, parameters, FLOPs, and other relevant factors.
4.1 Dataset description
The dataset samples we utilized were sourced from the experimental field of the School of Life Science and Technology at Henan Institute of Science and Technology. These samples were generously provided by our colleagues at the School of Life Science and Technology. Following a careful evaluation, we specifically selected the following wheat varieties for inclusion in our dataset: “Bainong 419,” “Bainong 207,” “Bainong 307,” “Luomai 28,” “Xinmai 26,” “Hengshui 6632,” “Nongda 3416-18,” and “Neile 288.” These varieties represent commonly cultivated wheat types in China and serve as a comprehensive representation of wheat diversity.
To capture high-quality images of these wheat samples, we employed a stereo microscope, as depicted in Figure 5.
To minimize any potential external environmental interference, we utilized black light-absorbing flannel as the background for capturing wheat grains images. The image collection process was conducted under natural indoor lighting conditions. For each wheat grain, we captured three images from different angles. It’s important to note that when assembling the dataset, wheat grains from the same variety but with varying angles were categorized together.
To ensure the quality and consistency of our dataset, a meticulous data collection process was employed. Initially, all wheat grains were subjected to a drying procedure in a well-ventilated indoor environment. Subsequently, a total of 8,000 seeds, with 1,000 grains selected from each wheat variety, were carefully handpicked. We chose three specific shooting angles for image capture: Ventral groove downward, Ventral groove toward the front, and Ventral groove upward. These images were saved in PNG format with a pixel resolution of 2688×1520. This comprehensive approach to data collection ensured the richness and completeness of our dataset, contributing to the robustness of our study. Consequently, we amassed a total of 24,000 images. The correspondence between each wheat variety and its corresponding number is detailed in Table 4.
Table 4 Correspondence between number, quantities and species name of wheat grains of different varieties in the dataset.
4.2 Data pre-processing
After the initial dataset collection, we diligently undertook a comprehensive dataset preprocessing pipeline. Our approach encompassed several crucial steps which are briefly described below (Zhuang et al., 2022).
Background Removal: In Figure 6, noticeable artifacts such as small white dots and lines were discernible in the original images. These imperfections arose from the inherent characteristics of the stereo microscope, capturing minute particles like lint and dust during the imaging process. To mitigate the influence of these extraneous elements on the subject matter, we employed sophisticated keying algorithms known as Background Matting and Background Matting V2 (Sengupta et al., 2020; Lin et al., 2021) to effectuate background removal across our dataset. The image with the background removed is shown in Figure 7.
Figure 6 The presentation of the same sample in the dataset with different shooting angles, the image has high clarity and the interference of the background on the subject content is obvious. (A) Ventral groove upward, (B) Ventral groove toward the front, (C) Ventral groove downward.
Figure 7 Using the advanced keying algorithm (Sengupta et al., 2020) (Lin et al., 2021) results. (A–C) correspond to (A–C) in Figure 6, respectively.
Image Resizing: The original images, as captured by the stereo microscope, featured dimensions of 2688×1520 pixels. Following the background removal in the previous step, we uniformly resized the images to 800×800 pixels, specifically focusing on isolating wheat seed grains. Subsequently, we further scaled down the image dimensions to 224×224 pixels, maintaining proportional scaling.
Standardization: To ensure consistency and facilitate convergence during training, we are based on experience standardized each image by setting the mean and standard deviation to (0.485, 0.456, 0.406) and (0.229, 0.224, 0.225), respectively. This standardization process was pivotal in optimizing the numerical properties of the images.
Dataset Split: For proper model evaluation, we randomly partitioned the dataset into the training set, validation set, and testing set with 7:1:2 ratio. This division allowed us to validate the model’s performance on unseen data, adhering to best practices in experimental design.
Transfer Learning: During the training phase, we employed transfer learning techniques by loading weight files pre-trained on the ImageNet dataset into our training model and the comparative models used in our experiments. This practice leveraged knowledge acquired from a large-scale dataset to enhance the performance of our models on the specific task at hand.
4.3 Evaluation criteria
In our evaluation of the network models, we employ several key metrics, including accuracy, precision, recall, and the F1-score, to assess the recognition performance of each model. The mathematical expressions for these metrics are provided in Equations 8–11.
Here, let’s clarify the definitions of the variables used in these equations. True Positive (TP) represents the instances that truly belong to a category and are correctly recognized by the classifier, while False Negative (FN) represents instances that belong to a category but are incorrectly categorized. On the other hand, False Positive (FP) signifies instances that do not belong to a category but are incorrectly recognized as belonging to that category, and True Negative (TN) corresponds to instances that do not belong to a category in reality and are correctly recognized as such. In addition, we show the confusion matrix (Srinivasu et al., 2022)of the experimental results in Figure 8.
Figure 8 Confusion matrix of SCGNet for eight wheat grains classification results, Among them, the correspondence between number and type is detailed in Table 4.
4.4 Comparison experiment
Throughout our experiments, when assessing the performance of different networks, we consider not only recognition accuracy but also other critical metrics, such as the number of parameters, average recognition time, and FLOPs. These metrics hold particular importance in our work, as our primary focus is on reducing these values to enhance the feasibility of deploying these models on mobile devices, thus improving their speed and efficiency in mobile applications. To quantify these metrics, we utilized the open-source project torchstat, which allowed us to calculate parameters, FLOPs, and other relevant statistics for each network.
To assess the efficacy of our proposed SCGNet, we present a comprehensive analysis of nine deep learning models for image classification to assess their effectiveness and suitability for various practical applications. We employ a consistent training, validation and testing dataset to ensure a fair and robust comparison, focusing on evaluating key performance metrics such as accuracy, precision, recall, F1-score, FLOPs, the number of parameters, and average recognition speed.
Our analysis covers a spectrum of network architectures, including traditional models with classical design principles models such as ResNet50, EfficientNet and RegNetX, lightweight models optimized for resource-constrained environments such as MobileNetV3, and ShuffleNetV2, network models with a transformer structure such as Vision Transformer and Swin Transformer, as well as deep models aimed at achieving SOTA accuracy such as RepLKNet and MAGE.
Figure 9 visually presents the accuracy results obtained during training for each of these network models, providing an intuitive overview of their performance. Additionally, Figure 10 reports the training loss, validation loss, training accuracy, and validation accuracy for SCGNet. Tables 5, 6 present a comprehensive summary of the results from the comparative experiments in the testset, encompassing various evaluation metrics.
Figure 10 The SCGNet training loss (A), training accuracy (B), validation loss (C), and validation accuracy (D).
Table 5 Differences between SCGNet and other comparison methods in terms of Accuracy, Precision, Recall and F1-Socre under the same testset.
Table 6 Differences between SCGNet and other comparison methods in terms of FLOPs, Parameters and Average recognition speed.
For comprehensive reference, we have meticulously documented all parameter settings utilized in the training of SCGNet. We use the validation set to evaluate the performance of the model with different parameter settings and finalize all the hyperparameters. These settings are presented in Table 7, allowing for a clear understanding of the experimental setup and facilitating reproducibility. In the comparative experiments involving different networks, since there are different versions of baseline and various improvements, we adhered to officially recommended parameter settings for these models to ensure consistency and fairness in our evaluations.
For Traditional Network Models, while foundational in the field, displayed suboptimal performance compared to more recent innovations. ResNet50, with its increased depth and residual connections, improves recognition accuracy but still falls short due to its relatively shallow architecture. If the depth of the network is increased without restriction, although the model is able to achieve better performance, the number of parameters of the model will also increase dramatically. EfficientNet-B0 incorporated Neural Architecture Search (NAS) principles to amalgamate depth, width, and channel scaling, achieving formidable recognition capabilities but at the expense of increased parameter complexity. RegNetX-200MF refined the NAS approach, achieving substantial parameter reduction while maintaining accuracy, albeit with a slight deficit compared to SCGNet.
For Lightweight Network Models, including MobileNetV3, and ShuffleNetV2, demonstrated a harmonious balance between accuracy and computational efficiency. MobileNetV3, building upon its predecessor, introduced NetAdapt and various NetPruningVersions (NPVs) alongside an algorithm to optimize convolutional kernels and channels, further enhancing its performance. ShuffleNetV2 adopts a split-and-concatenate strategy to reduce overall computational demands.
For the Transfomer Network Models, namely Vision Transformer and Swin Transformer, their performance is already very close to that of SCGNet, and, Swin Transformer is capable of suboptimal performance in the precision rate metric. However, it should not be overlooked that they possess a huge number of parameters.
For the SOTA Network Models, include RepLKNet and MAGE. MAGE exhibits the highest accuracy, recall and F1-Socre, while RepLKNet displays the highest recall. They both exhibit exceptional precision performance. However, they place excessive emphasis on metrics such as accuracy, neglecting the balance between speed and precision, average recognition speed is relatively slow. Moreover, they have a large number of parameters, FLOPs, with RepLKNet and MAGE’s FLOPs being 384 and 614 times higher than that of SCGNet, respectively.
For SCGNet, the proposed SCGNet exhibits commendable performance in resource utilization metrics such as FLOPs, average recognition speed and the number of parameters, surpassing alternative models in these aspects. Despite its sub-optimal performance in accuracy and F1-Socre compared to the MAGE model, the marginal 0.02% difference in accuracy is deemed negligible. We maintain that sacrificing such a small improvement in accuracy for the reduction in FLOPs, average recognition speed and parameter count makes SCGNet highly cost-effective. This is especially favorable for the model’s deployment on mobile devices with limited resources.
This strategic trade-off in favor of resource efficiency positions SCGNet as a compelling candidate for deployment in practical scenarios, where considerations of computational cost are pivotal. Such efficiency gains can contribute significantly to the feasibility and scalability of deploying deep learning models in resource-constrained environments.
4.5 Ablation study
To ascertain the individual contributions of each module within the SCGNet model to its overall performance, we conducted a series of ablation studies. These studies encompass the following scenarios: (1) SCG block without the GM module (-w/o GM), (2) SCG block without the SC module (-w/o SC), (3) Instead of applying a 3-D convolutional classification layer, a traditional classification layer is used instead, and (4) Replacing the Swish activation function with the traditional ReLU activation function(-w/o Swish).
The impact of each ablation study on the model’s performance is summarized in Table 8.
Table 8 Discriminatory results of different modules for the implementation of ablation studies on test samples.
Specifically, (1) -w/o GM exhibits a pronounced effect on recognition performance. This is attributed to the GM module’s role in facilitating the exchange of feature information among group convolutions. The absence of the GM module impedes individual group convolutions from effectively learning additional features from one another.
Contrastingly, (2) -w/o SC demonstrates minimal impact on recognition performance. However, it results in an increase in the number of parameters, along with metrics such as FLOPs, due to the use of fully connected.
Moreover, (3) -w/o 3-D Conv yields a tiny effect on recognition performance. However, the number of parameters and FLOPs are dramatically increased due to the large number of fully connected computations involved in the traditional classification layer.
Finally, (4)-w/o Swish, the Swish activation function has a smoothness that enhances the forward propagation optimization, and replacing ReLU with Swish brings about a lesser drop in FLOPs without any loss of accuracy. In addition, our investigation delves into the impact of varying strides on the model’s performance within the sparsely connected methodology. We systematically evaluate the effects of strides set at 2, 3, 5, 7, and 9, employing three key metrics: accuracy, FLOPs, and the number of parameters. The results, summarized in Table 3, elucidate the influence of each stride value on model performance.
Remarkably, when the stride is set to 3, the model demonstrates optimal recognition performance. As the stride increases, the computational load of the model diminishes. Simultaneously, however, there is a discernible and precipitous decline in the recognition accuracy of the model. This decline is particularly pronounced when the stride is set to 9, resulting in a precipitous drop akin to a cliff.
5 Conclusion
In this research, we introduce a specialized CNN for precise wheat grain classification. We propose “Group Mixing” to address information flow issues in group convolution, and “Sparsely Connected” methodology to reduce parameter redundancy, minimizing FLOPs and parameters. In addition, we have innovatively devised a new classification output layer predicated on 3-D convolution, supplanting the conventional maximum pooling layer and fully connected layer, replacing traditional classification layers without sacrificing accuracy. Drawing from the foregoing advancements, we have conceived an efficient Sparsely Connected Group Convolution Network, custom-tailored for the high-resolution classification of wheat grains.
Numerous rigorous experimental evaluations substantiate the prowess of our proposed SCGNet, which attains an impressive accuracy rate of 99.56%. Moreover, our approach is notably characterized by a parsimonious parameter count and reduced FLOPs, rendering it exceptionally suitable for deployment on mobile devices.
However, we acknowledge limitations in our dataset and SCGNet architecture. The dataset lacks diversity in wheat varieties, necessitating the acquisition of more varied datasets. SCGNet, tested in controlled high-resolution conditions, needs validation for low-resolution images from mobile devices.
The amalgamation of computer vision techniques with the automated non-destructive classification of individual wheat grains portends significant potential across diverse applications. In forthcoming endeavors, our research trajectory will encompass the collection of images representing a broader spectrum of wheat varieties and possibly other crop seeds. Building upon these comprehensive datasets, we endeavor to enhance the efficacy of the SCGNet architecture, with a particular focus on bolstering its robustness, reducing its parameter count and FLOPs, and venturing into deployment on mobile terminals. The latter imposes stringent constraints on model size, an exigent challenge we are poised to tackle.
Data availability statement
The original contributions presented in the study are included in the article/supplementary materials, further inquiries can be directed to the corresponding author.
Author contributions
XS: Writing – original draft, Writing – review & editing. YL: Writing – review & editing. GL: Writing – original draft, Writing – review & editing. SJ: Writing – review & editing. WYZ: Writing – review & editing. ZL: Writing – review & editing. WDZ: Writing – review & editing.
Funding
The author(s) declare financial support was received for the research, authorship, and/or publication of this article. This work was supported in part by the Natural Science Foundation of Henan Province under Grant 232300420428, in part by the Teacher Education Curriculum Reform Research of Henan Province under Grant 2024-JSJYYB-029, in part by the Key Specialized Research and Development Program of Science and Technology of Henan Province under Grants 232102210018, 23212210058, 232102111127.
Acknowledgments
We are very grateful to our colleagues in the School of Life Sciences and Technology at our university for generously providing the datasets used in this study.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
Chollet, F. (2017). “Xception: Deep learning with depthwise separable convolutions,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 1251–1258.
Dai, Z., Liu, H., Le, Q. V., Tan, M. (2021). Coatnet: Marrying convolution and attention for all data sizes. Adv. Neural Inf. Process. Syst. 34, 3965–3977.
Delwiche, S. R., Yang, I.-C., Graybosch, R. A. (2013). Multiple view image analysis of freefalling US wheat grains for damage assessment. Comput. Electron. Agric. 98, 62–73. doi: 10.1016/j.compag.2013.07.002
Ding, X., Zhang, X., Han, J., Ding, G. (2022). “Scaling up your kernels to 31x31: Revisiting large kernel design in cnns,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 11963–11975.
Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., et al. (2020). An image is worth 16x16 words: Transformers for image recognition at scale. arXiv.
Fanelli, V., Dellino, M., Taranto, F., De Giovanni, C., Sabetta, W., De Vita, P., et al. (2023). Varietal identification in pasta through an SSR-based approach: A case study. J. Sci. Food Agric. 103, 5521–5528. doi: 10.1002/jsfa.12631
Gao, H., Wang, Z., Ji, S. (2018). Channelnets: Compact and efficient convolutional neural networks via channel-wise convolutions. Adv. Neural Inf. Process. Syst. 31, 2570–2581.
Güneş, E. O., Aygün, S., Kırcı, M., Kalateh, A., Çakır, Y. (2014). “Determination of the varieties and characteristics of wheat seeds grown in Turkey using image processing techniques,” in 2014 The Third International Conference on Agro-Geoinformatics (IEEE). 1–4.
He, K., Zhang, X., Ren, S., Sun, J. (2015). “Deep residual learning for image recognition,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 770–778.
Howard, A., Sandler, M., Chu, G., Chen, L.-C., Chen, B., Tan, M., et al. (2019). “Searching for mobilenetv3,” in Proceedings of the IEEE/CVF International Conference on Computer Vision. 1314–1324.
Howard, A. G., Zhu, M., Chen, B., Kalenichenko, D., Wang, W., Weyand, T., et al. (2017). MobileNets: efficient convolutional neural networks for mobile vision applications. arXiv.
Huang, G., Liu, Z., van der Maaten, L., Weinberger, K. Q. (2017). “Densely connected convolutional networks,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 4700–4708.
Hussain, M. A., Naeem, A., Sulieman, S., Pitann, B., Mühling, K. H. (2022). Sulfur uptake and distribution, grain yield, and quality of hybrid and inbred winter wheat (Triticum aestivum L.) varieties under early and late waterlogging. J. Plant Nutr. Soil Sci. 185, 622–631. doi: 10.1002/jpln.202200149
Javanmardi, S., Ashtiani, S.-H. M., Verbeek, F. J., Martynenko, A. (2021). Computer-vision classification of corn seed varieties using deep convolutional neural network. J. Stored Products Res. 92, 101800. doi: 10.1016/j.jspr.2021.101800
Kozłowski, M., Górecki, P., Szczypiński, P. M. (2019). Varietal classification of barley by convolutional neural networks. Biosyst. Eng. 184, 155–165. doi: 10.1016/j.biosystemseng.2019.06.012
Krizhevsky, A., Sutskever, I., Hinton, G. E. (2012). Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 60, 84–90. doi: 10.1145/3065386
Kurtulmuş, F., Alibaş, İ., Kavdır, I. (2016). Classification of pepper seeds using machine vision based on neural network. Int. J. Agric. Biol. Eng. 9, 51–62. doi: 10.3965/j.ijabe.20160901.1790
Li, H., Zhou, Y., Xin, W., Wei, Y., Zhang, J., Guo, L. (2019). Wheat breeding in northern China: Achievements and technical advances. Crop J. 7, 718–729. doi: 10.1016/j.cj.2019.09.003
Li, T., Chang, H., Mishra, S., Zhang, H., Katabi, D., Krishnan, D. (2023). “Mage: Masked generative encoder to unify representation learning and image synthesis,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2142–2152.
Lin, S., Ryabtsev, A., Sengupta, S., Curless, B. L., Seitz, S. M., Kemelmacher-Shlizerman, I. (2021). “Real-time high-resolution background matting,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 8762–8771.
Lin, T.-Y., Dollár, P., Girshick, R., He, K., Hariharan, B., Belongie, S. (2017). “Feature pyramid networks for object detection,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2117–2125.
Liu, Z., Lin, Y., Cao, Y., Hu, H., Wei, Y., Zhang, Z., et al. (2021). “Swin transformer: Hierarchical vision transformer using shifted windows,” in Proceedings of the IEEE/CVF International Conference on Computer Vision. 10012–10022.
Liu, Z., Mao, H., Wu, C.-Y., Feichtenhofer, C., Darrell, T., Xie, S. (2022). “A convnet for the 2020s,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 11976–11986.
Lu, Y., Wang, R., Hu, T., He, Q., Chen, Z. S., Wang, J., et al. (2022). Nondestructive 3D phenotyping method of passion fruit based on X-ray micro-computed tomography and deep learning. Front. Plant Sci. 13. doi: 10.3389/fpls.2022.1087904
Ma, N., Zhang, X., Zheng, H.-T., Sun, J. (2018). “Shufflenet v2: Practical guidelines for efficient cnn architecture design,” in Proceedings of the European Conference on Computer Vision (ECCV). 116–131.
Mefleh, M., Conte, P., Fadda, C., Giunta, F., Piga, A., Hassoun, G., et al. (2019). From ancient to old and modern durum wheat varieties: Interaction among cultivar traits, management, and technological quality. J. Sci. Food Agric. 99, 2059–2067. doi: 10.1002/jsfa.9388
Ni, C., Wang, D., Vinson, R., Holmes, M., Tao, Y. (2019). Automatic inspection machine for maize kernels based on deep convolutional neural networks. Biosyst. Eng. 178, 131–144. doi: 10.1016/j.biosystemseng.2018.11.010
Radosavovic, I., Kosaraju, R. P., Girshick, R., He, K., Dollár, P. (2020). “Designing network design spaces,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 10428–10436.
Sabanci, K., Kayabasi, A., Toktas, A. (2017). Computer vision-based method for classification of wheat grains using artificial neural network. J. Sci. Food Agric. 97, 2588–2593. doi: 10.1002/jsfa.8080
Saeed, M., Masood Quraishi, U., Malik, R. N. (2022). Identification of arsenic-tolerant varieties and candidate genes of tolerance in spring wheat (Triticum aestivum L.). Chemosphere 308, 136380. doi: 10.1016/j.chemosphere.2022.136380
Sandler, M., Howard, A., Zhu, M., Zhmoginov, A., Chen, L.-C. (2018). “Mobilenetv2: Inverted residuals and linear bottlenecks,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 4510–4520.
Sengupta, S., Jayaram, V., Curless, B., Seitz, S. M., Kemelmacher-Shlizerman, I. (2020). “Background matting: The world is your green screen,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2291–2300.
Shen, Y., Yin, Y., Li, B., Zhao, C., Li, G. (2021). Detection of impurities in wheat using terahertz spectral imaging and convolutional neural networks. Comput. Electron. Agric. 181, 105931. doi: 10.1016/j.compag.2020.105931
Simonyan, K., Zisserman, A. (2015). Very deep convolutional networks for large-scale image recognition. arXiv.
Srinivasu, P. N., JayaLakshmi, G., Jhaveri, R. H., Praveen, S. P. (2022). Ambient assistive living for monitoring the physical activity of diabetic adults through body area networks. Mobile Inf. Syst. 2022, 1–18. doi: 10.1155/2022/3169927
Szegedy, C., Liu, W., Jia, Y., Sermanet, P., Reed, S., Anguelov, D., et al. (2015). “Going deeper with convolutions,” in 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (Boston, MA, USA: IEEE), 1–9. doi: 10.1109/CVPR.2015.7298594
Tan, M., Le, Q. V. (2020). EfficientNet: rethinking model scaling for convolutional neural networks. arXiv. doi: 10.48550/arXiv.1905.11946
Weng, S., Han, K., Chu, Z., Zhu, G., Liu, C., Zhu, Z., et al. (2021). Reflectance images of effective wavelengths from hyperspectral imaging for identification of Fusarium head blightinfected wheat kernels combined with a residual attention convolution neural network. Comput. Electron. Agric. 190, 106483. doi: 10.1016/j.compag.2021.106483
Xie, S., Girshick, R., Dollár, P., Tu, Z., He, K. (2017). “Aggregated residual transformations for deep neural networks,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 1492–1500.
Yang, X., Guo, M., Lyu, Q., Ma, M. (2021b). Detection and classification of damaged wheat kernels based on progressive neural architecture search. Biosyst. Eng. 208, 176–185. doi: 10.1016/j.biosystemseng.2021.05.016
Yang, H., Ni, J., Gao, J., Han, Z., Luan, T. (2021a). A novel method for peanut variety identification and classification by Improved VGG16. Sci. Rep. 11, 15756. doi: 10.1038/s41598-021-95240-y
Yu, C., Gao, C., Wang, J., Yu, G., Shen, C., Sang, N. (2021). Bisenet v2: Bilateral network with guided aggregation for real-time semantic segmentation. Int. J. Comput. Vision 129, 3051–3068. doi: 10.1007/s11263-021-01515-2
Zhang, W., Li, Z., Sun, H.-H., Zhang, Q., Zhuang, P., Li, C. (2022a). SSTNet: spatial, spectral, and texture aware attention network using hyperspectral image for corn variety identification. IEEE Geosci. Remote Sens. Lett. 19, 1–5. doi: 10.1109/LGRS.2022.3225215
Zhang, W., Sun, X., Zhou, L., Xie, X., Zhao, W., Liang, Z., et al. (2023). Dual-branch collaborative learning network for crop disease identification. Front. Plant Sci. 14. doi: 10.3389/fpls.2023.1117478
Zhang, W., Wang, Y., Li, C. (2022b). Underwater image enhancement by attenuated color channel correction and detail preserved contrast enhancement. IEEE J. Oceanic Eng. 47, 718–735. doi: 10.1109/JOE.2022.3140563
Zhang, X., Zhou, X., Lin, M., Sun, J. (2018). “Shufflenet: An extremely efficient convolutional neural network for mobile devices,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 6848–6856.
Zhao, W., Liu, S., Li, X., Han, X., Yang, H. (2022). Fast and accurate wheat grain quality detection based on improved YOLOv5. Comput. Electron. Agric. 202, 107426. doi: 10.1016/j.compag.2022.107426
Zhou, J., Zhang, C., Du, B., Cui, H., Fan, X., Zhou, D., et al. (2021). Soil and foliar applications of silicon and selenium effects on cadmium accumulation and plant growth by modulation of antioxidant system and Cd translocation: Comparison of soft vs. durum wheat varieties. J. Hazard. Mater. 402, 123546. doi: 10.1016/j.jhazmat.2020.123546
Keywords: wheat grains classification, feature multiplexing, sparsely connected, 3-D convolution, the number of parameters
Citation: Sun X, Li Y, Li G, Jin S, Zhao W, Liang Z and Zhang W (2023) SCGNet: efficient sparsely connected group convolution network for wheat grains classification. Front. Plant Sci. 14:1304962. doi: 10.3389/fpls.2023.1304962
Received: 30 September 2023; Accepted: 01 December 2023;
Published: 22 December 2023.
Edited by:
Yuriy L. Orlov, I.M.Sechenov First Moscow State Medical University, RussiaReviewed by:
Parvathaneni Naga Srinivasu, Prasad V. Potluri Siddhartha Institute of Technology, IndiaZhaoyu Zhai, Nanjing Agricultural University, China
Chu Zhang, Huzhou University, China
Copyright © 2023 Sun, Li, Li, Jin, Zhao, Liang and Zhang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Guohou Li, ligh@hist.edu.cn