Skip to main content

ORIGINAL RESEARCH article

Front. Plant Sci., 29 August 2023
Sec. Technical Advances in Plant Science
This article is part of the Research Topic Rapid and Nondestructive Detection Techniques for Agri-Product Quality View all 17 articles

Non-destructive detection of single-seed viability in maize using hyperspectral imaging technology and multi-scale 3D convolutional neural network

Yaoyao Fan,Yaoyao Fan1,2Ting AnTing An2Qingyan WangQingyan Wang2Guang YangGuang Yang2Wenqian HuangWenqian Huang2Zheli WangZheli Wang2Chunjiang Zhao,*Chunjiang Zhao1,2*Xi Tian*Xi Tian2*
  • 1College of Information and Electrical Engineering, Shenyang Agricultural University, Shenyang, China
  • 2Intelligent Equipment Research Center, Beijing Academy of Agriculture and Forestry Sciences, Beijing, China

The viability of Zea mays seed plays a critical role in determining the yield of corn. Therefore, developing a fast and non-destructive method is essential for rapid and large-scale seed viability detection and is of great significance for agriculture, breeding, and germplasm preservation. In this study, hyperspectral imaging (HSI) technology was used to obtain images and spectral information of maize seeds with different aging stages. To reduce data input and improve model detection speed while obtaining more stable prediction results, successive projections algorithm (SPA) was used to extract key wavelengths that characterize seed viability, then key wavelength images of maize seed were divided into small blocks with 5 pixels ×5 pixels and fed into a multi-scale 3D convolutional neural network (3DCNN) for further optimizing the discrimination possibility of single-seed viability. The final discriminant result of single-seed viability was determined by comprehensively evaluating the result of all small blocks belonging to the same seed with the voting algorithm. The results showed that the multi-scale 3DCNN model achieved an accuracy of 90.67% for the discrimination of single-seed viability on the test set. Furthermore, an effort to reduce labor and avoid the misclassification caused by human subjective factors, a YOLOv7 model and a Mask R-CNN model were constructed respectively for germination judgment and bud length detection in this study, the result showed that mean average precision (mAP) of YOLOv7 model could reach 99.7%, and the determination coefficient of Mask R-CNN model was 0.98. Overall, this study provided a feasible solution for detecting maize seed viability using HSI technology and multi-scale 3DCNN, which was crucial for large-scale screening of viable seeds. This study provided theoretical support for improving planting quality and crop yield.

1 Introduction

Single-seed sowing is a crucial strategy to boost corn production, save seeds, and reduce labor, but it demands high-quality seeds (Li et al., 2017). On October 11th, 2020, a new standard has been released by China, which raises the germination rate index for single-seed sowing from 85% to 93%. The viability is a critical indicator for evaluating the quality and practicality of seed. Assessment of seed viability could ensure each seed has the potential for germination and healthy growth and promotes the popularization of single-seed sowing. This not only facilitates mechanized sowing and reduces the laboriousness of manual interplanting and seedling transplantation, but also significantly reduces the amount of seed used and conserves a considerable amount of seed production area (Liang et al., 2020). Therefore, the determination of seed viability is of utmost importance in reducing the cost and time loss resulting from planting failures and conserving human resources.

Seed viability is a quality characteristic at the individual level rather than a quantitative trait at the population level. Loss of viability among individuals in the same population is not synchronous, making it challenging to detect the viability of single-seed. According to the International Seed Testing Association (ISTA) rules (Association, I.S.T, 1999), common methods for seed viability detection include germination and staining (Cheng et al., 2023). The conventional germination method is the most accurate, but it is time-consuming and requires a lot of material resources. On the other hand, staining is only suitable for a small number of samples. Therefore, it is necessary to develop a rapid-nondestructive technique for single-seed viability detection in large quantities.

In the field of seed quality detection, hyperspectral imaging technology has been widely utilized. However, research on seed viability detection is relatively limited. Jannat Yasmin et al. (2022) presented an online detection system of watermelon seed viability based on longwave near-infrared (LWNIR) HSI, demonstrating its potential application in predicting seed viability. Wang et al. (2021) developed the discrimination models of seed viability using the feature wavelengths and full wavelengths of the visible and shortwave near-infrared (Vis-SWNIR), the result revealed that both models attained an accuracy rate surpassing 95%, suggesting that the seeds with different aging stages exhibited unique spectral features, and the characteristic wavelengths can effectively provide the key information of Zea mays seed quality. Pang et al. (2021) conducted a germination experiment on maize seeds with different aging stages, a 2D convolutional neural network (2DCNN) model was developed by combing deep learning algorithms with hyperspectral technology. The accuracy of this model reached 99.96%, which was significantly higher than machine learning and one-dimensional convolutional neural network (CNN). It was worth pointing out that the model demonstrated a relatively fast convergence speed, which highlighted the feasibility and effectiveness of combining deep learning with hyperspectral technology to determine the viability of single-seed. Ambrose et al. (2016) investigated the feasibility of using HSI technology to differentiate the viability of maize seeds. One group of maize samples was subjected to microwave heat treatment, while the other group served as the control. PLS-DA was employed to classify the heat-treated (aged) and untreated (normal) maize seeds. The results showed that the classification model achieved the highest classification accuracy in the LWNIR region, with calibration set accuracy of 97.6% and prediction set accuracy of 95.6%. These studies achieve high accuracy by predicting the aging level or treatment condition of seeds instead of the actual results of germination experiments. And they mainly rely on overall image information for seed viability classification. However, they overlook the significance of local information within seeds and fail to consider subtle variations and characteristics in different seed regions.

Generally, the evaluation of germination rate of seeds mainly depends on manual labor, which is time-consuming and cumbersome. Zhao et al. (2022) proposed a detection method for the germination rate of rice seeds using deep learning models, which took an average of 0.011 seconds for each image while achieving a mAP of 0.9539, meeting the demands of real-time detection, indicating that the YOLO-r model had great potential for rapidly and precisely determining the germination status of seeds. Bai et al. (2023) developed an improved discriminative approach for the detection of seed germination using YOLOv5. This technique enables the swift evaluation of parameters such as wheat seed germination rate, germination potential, germination index, and average germination days.

The emergence ability of seedlings is crucial for seed growth and crop yield improvement (Cui et al., 2020). In recent studies, significant progress has been made in correlating seed germination ability and seedling growth through various measurement methods. However, traditional manual measurement techniques for assessing parameters like bud length have been found to be inefficient and prone to errors due to the complex and twisted nature of buds. To address this issue, Adegbuyi and Burris (1988) found there was a significant correlation between seed germination ability and seedling growth by measuring comprehensive growth parameters. However, manual measurement method of bud length is inefficient and error-prone due to their curved and twisted nature. Gaikwad et al. (2019) developed a semi-automated tool for measuring leaf length, width, and area. Abdelaziz Triki et al. (2021) used the Mask R-CNN algorithm to effectively segment and measure leaf characteristics and obtained an error rate of around 5%. An enhanced algorithm based on the mask RCNN was introduced by Shen et al. (2023) to recognize defective wheat kernels. The experimental outcomes showed that this refined algorithm facilitated quicker and more precise detection of unsound kernels, effectively tackling issues linked to kernel adhesion. Masood et al. (2021) propose an automated method that utilizes the Mask RCNN model to achieve precise localization and segmentation of brain tumors. Cui et al. (2022) constructed a recognition model using hyperspectral data and feature extraction algorithms to predict maize root length, showing a significant correlation between root length and viability. Therefore, it is of great significance to measure and predict the seed viability using computer technology.

The above study highlighted the significance of seed viability determination and emphasized the need of developing rapid and non-destructive technology for single-seed viability detection. HSI has been established as a useful tool for seed quality detection, and the integration of deep learning and hyperspectral technology can establish an effective seed viability detection model. However, previous studies commonly used relatively simple models, and lacking the prediction model of maize seeds viability developed by 3DCNN and hyperspectral images. This study proposed an improved method for identifying the viability of maize seeds based on germination experiments. The aim of the study is to explore the potential of using hyperspectral images and 3DCNN to identify the viability of maize seeds. Specifically, the objectives are to: (1) select characteristic wavelengths that represent seed viability, (2) combine HSI with 3DCNN to establish the optimal classification model for maize seed viability, (3) evaluate the feasibility of using YOLOv7 model instead of the human eye to determine the seed germination status, (4) evaluate the ability of Mask R-CNN in bud segmentation and bud length prediction.

2 Materials and methods

2.1 Maize sample preparation

2.1.1 Aging experiment

Due to the high quality and the resistance to multiple stressors, “Jingke 968” maize is extensively cultivated in eastern and northern China. Therefore, it was selected as the experiment sample in this study. To ensure the accuracy of the experiment, seeds with uniform size and shape were manually selected, then all seeds were disinfected by soaking them in a 0.5% sodium hypochlorite solution for 5 minutes, followed by rinsing with distilled water five times, and air-dried under natural conditions.

To simulate the natural aging process of seeds, the experiment samples were artificially aged. All seeds were exposed to high temperature and high humidity conditions (45 °C and a relative humidity of 95%) and stirred twice a day to ensure uniform exposure (Zhang et al., 2020). 150 maize seeds were taken out randomly at aging 2, 4, 6, and 8 days, respectively. Additionally, 150 untreated seeds were selected as the control group (CK). Therefore, a total of 750 maize seeds within five aging stages were obtained and used for subsequent experimentation.

2.1.2 Hyperspectral imaging system

Two HSI systems, the Vis-SWNIR and LWNIR, have been built in the Intelligent Detection Laboratory of the China Agricultural Equipment Technology Research Center (Fan et al., 2018). The Vis-SWNIR system is capable of acquiring hyperspectral images within the wavelength range of 327-1098 nm, encompassing 1000 spectral variables, while the LWNIR system can capture images within the range of 930-2548 nm, containing 256 spectral variables. The Vis-SWNIR system includes an imaging spectrometer, an electron-multiplying charge-coupled device camera with a resolution of 502×500, a camera lens, and a spectraCube data acquisition software. Similarly, the LWNIR system includes an imaging spectrometer, a charge-coupled device camera with a resolution of 320×256, a camera lens, and a spectral acquisition software (Tian et al., 2021). And the acquisition software of both systems was developed using LabVIEW (National Instruments Inc., Austin, TX, USA) to facilitate the acquisition of spectral images, as well as to manage the camera and motor operations. Both systems share two 300-watt halogen lamps to provide stable illumination. In addition, an electrically operated moving platform and a computer are available for sample placement (Capable of accommodating up to 96 samples simultaneously) and hyperspectral image acquisition (Figure 1A) (Liu et al., 2022).

FIGURE 1
www.frontiersin.org

Figure 1 Diagram of the 3DCNN for hyperspectral image classification (A) Hyperspectral image acquisition device, (B) Regional voting, (C) Conventional 3DCNN model, (D) Multi-scale 3DCNN model.

To ensure the accuracy and reliability of the hyperspectral images (Eraw), calibration operation is essential to eliminate the effects of uneven illumination of the light source and camera dark current changes (An et al., 2022). The calibration operation involved using a white reflection board (with a reflectance of 99%) (Ew) to acquire a standard white reference image in the same sampling environment as the sample, while turning off the light source and covering the lens to obtain a black reference image (with a reflectance of 0%) (Ed). The calibrated image can be calculated using the following formula:

Ec=ErawEdEwEd(1)

After calibration, in the Vis-SWNIR region, a subset of 347 spectral variables within the 420-1000 nm range was selected for further analysis, considering the abundance of spectral data and the presence of duplicate information in adjacent spectra. On the other hand, in the near-infrared region, due to the limited number of available bands, all spectral variables (256) were directly included in the analysis. To separate maize seeds from the background, a mask was applied to segment the hyperspectral image. The gray-scale images at 801 nm and 1098 nm were selected as the mask images for the Vis-SWNIR and LWNIR bands, respectively. The average spectral curves were obtained by calculating the mean reflectance under the mask. Lastly, in order to eliminate the influence of the instrument, the Savitzky-Golay (SG) and Standard Normal Variate (SNV) methods were utilized to preprocess the spectra.

2.1.3 Standard germination test

A transparent box measuring 25cm×25cm was used as a germination chamber, and 75 seeds were placed in each box. A total of 10 boxes were used in the experiment. Prior to the germination test, the germination boxes were sterilized with 75% ethanol (Suksungworn et al., 2021), and three layers of gauze were placed in each germination box to provide continuous moisture for the seeds. A black gauze was placed on the top layers as the background for photography (Figure 2A). An equal amount of distilled water was added to each box, and the temperature was set to 25°C with 12-hour intervals of light and dark (Figure 2B). Throughout the 7-day germination experiment (Long et al., 2022), the germination progress of maize seeds was monitored daily at specific time intervals. According to the ISTA standard, the germination rate was determined (Wang et al., 2022c).

FIGURE 2
www.frontiersin.org

Figure 2 Diagram of the standard germination experiment (A) Corn seed samples, (B) Germination of seeds in a climate chamber, (C) Sprouted seeds, (D) RGB iamge acquisition device.

2.1.4 RGB image acquisition

RGB images of maize seeds were captured using BASLER industrial cameras (acA1920-25um/uc, BASLER AG, Germany, 2.4 MP,100 fps) during germination test (Figure 2D) (Shen et al., 2023). An adjustable camera platform was built to ensure consistency of the images and prevent camera shake. The position of the germination box relative to the lens was kept fixed during each image capture. Indoor lighting was turned on and curtains were drawn for each capture. After placing the seeds into the boxes (Day 0), images of each box were immediately captured. Subsequently, images were captured every 15 hours for 7 consecutive days (Figure 2C). The dataset used in this study consisted of a total of 3000 maize seeds (All the captured RGB images collectively contain 3000 seeds). Among them, 2250 seeds were designated as training samples, while the remaining seeds were allocated to the test set.

2.2 Data processing

2.2.1 Successive projections algorithm

Hyperspectral data typically consists of numerous bands, and certain bands may exhibit high correlation or contain redundant information (Han et al., 2022).When training 3DCNN with full-band data, it will lead to a significant increase in the number of networks training parameters, resulting in a more complex model. This phenomenon is commonly referred to as the curse of dimensionality. (Köppen, 2000). However, band selection (Sun and Du, 2019) allows retaining spectral bands that are closely related to seed vigor assessment while removing irrelevant bands, thereby enhancing the feature extraction and discriminative capabilities of the model.

Additionally, the use of dimensionality reduction data sets can effectively reduce the complexity of the model, mitigating the risk of overfitting and enhancing the model’s generalization ability and stability (Aloupogianni et al., 2023). Moreover, fewer computing resources are required during model training and inference, leading to a significant improvement in the computational efficiency of the model (XingJia et al., 2022).

Successive projections algorithm is a classical band selection method that can map high-dimensional spectral data to a low-dimensional space through multiple projections(de Almeida et al., 2018). SPA is a forward iterative search method used for selecting spectral information with minimal redundancy to address collinearity issues. The steps of SPA are shown in Table 1. The SPA is widely used in hyperspectral image processing attributed to its advantages of fast computation speed and easy implementation (Chen et al., 2023b). Therefore, the SPA was used in this study to perform feature selection on the processed average spectra of Vis-SWNIR and LWNIR, in order to perform dimensionality reduction on the hyperspectral data.

TABLE 1
www.frontiersin.org

Table 1 Successive projections algorithm.

2.2.2 Machine learning

Support Vector Machine (SVM) (Cortes and Vapnik, 1995) is a powerful algorithm for classification and regression that finds an optimal hyperplane to separate data points of different classes. It handles high-dimensional datasets, avoids overfitting, and can handle non-linear problems using kernel functions. K-Nearest Neighbor (KNN) (Zhang, 2022) is a basic algorithm that selects the K nearest samples based on their feature values and uses their labels as predictions. Subspace Discriminant Analysis (SDA) (Zhao and Phillips, 1999) is a pattern classification method that aims to find a low-dimensional subspace to maximize the separation between different classes. In this study, the aforementioned machine learning methods were used to classify the viability of maize seeds at different aging stages for optimal classification accuracy.

2.2.3 Deep convolutional neural network

The CNN combines the concepts of convolutional filtering and neural networks by utilizing local receptive fields and weight sharing to reduce the number of network parameters and speed up model training (Ghaderizadeh et al., 2021). Compared to the widespread use of two-dimensional convolution, three-dimensional convolution is less commonly used in practice. However, HSI contain rich spectral information, and using two-dimensional convolution may make the interband correlations of HSIs underutilized (Ge et al., 2020). To address this issue, this study introduced a 3DCNN, which can thoroughly extract feature relationships across different feature channels (Figure 1C), thereby enabling it to concurrently extract integrated spectral and spatial features from hyperspectral imagery (Sun et al., 2022).

Before inputting hyperspectral images into the network, standardization is performed to ensure that the data is within the same scale and range, enabling the network to learn weights faster and converge more easily during training. Moreover, data standardization can help avoid the problems of gradient disappearance or explosion, and improve the stability and generalization ability of the network. To obtain multiple convolutional features of HSI, multi-scale convolution is employed in the same convolutional layer, which can acquire both global and local information. Four different convolution kernels of 3×3×3, 3×3×5, 3×5×5, and 5×5×5 were selected to extract feature information and fused on the channel. This method can enhance the classification accuracy of the model. As illustrated in Figure 1D, each convolution kernel in the first convolution module has 16 filters, each kernel in the second convolution module has 32 filters, and each kernel in the third convolution module has 64 filters. The activation function in the three-dimensional convolution module uses Rectified Linear Unit (RELU) and is compressed by the pooling layer to reduce the amount of data and parameters, as well as alleviate the overfitting phenomenon. To ensure that the features extracted by different convolution kernels in the same module can be effectively connected, different parameters need to be set according to different situations, such as stride and padding. Finally, the output is produced through 1 fully connected layer and 1 output layer, and the output layer employs the SoftMax activation function.

To extract features from hyperspectral images of maize seeds at a more microscopic level and increase the amount of data, a window size of 5×5 was selected for segmentation (Figure 1B). To eliminate the influence of background on classification, small blocks containing 0-pixel points were discarded. As the size of maize seeds varies, the number of blocks obtained from different segments of maize seeds is also inconsistent. To address this issue, this study employed a majority principle labeling aggregation method, as Table 2.

TABLE 2
www.frontiersin.org

Table 2 Majority principle labeling aggregation method.

In this study, the germination experiment showed that 404 viable samples and 346 nonviable samples were collected from 750 seeds. Given that the hyperspectral images were collected in a sequential manner based on the aging gradients of the seeds, it was crucial to maintain a balanced distribution of germinated and non-germinated samples in the test set. Therefore, a representative test set was carefully selected, consisting of 75 seeds, including the first seed, the 10th seed, the 20th seed, and so on. The remaining 675 seeds were allocated for the training phase. Through this meticulous approach, it was ensured that the test set encompassed samples from diverse categories, enabling an accurate evaluation of the classification model’s performance.

2.2.4 Establishment of Mask R-CNN model for bud length detection

In order to measure the length of maize seed bud, the Mask R-CNN (He et al., 2017) (With resnet50_fpn as backbone) model was utilized to segment the bud from single-seed image firstly, then a skeleton extraction algorithm was applied to extract the skeleton of the bud (Figure 3A). Next, the bud length detection algorithm was used to remove the branches in the skeleton for obtaining the central skeleton image. Finally, the actual bud length was calculated by converting pixels to actual length (Figure 3B).

FIGURE 3
www.frontiersin.org

Figure 3 Diagram of the maize bud length detection process (A) Process of bud segmentation, (B) Process of bud length detection.

Mask R-CNN is a deep learning model that combines object detection and instance segmentation. It extends Faster R-CNN by generating binary masks for each region of interest (ROI), achieving pixel-level segmentation. The network consists of three main components: a backbone network, a Region Proposal Network (RPN) responsible for generating candidate object regions, and two parallel branches dedicated to object detection and mask prediction. Mask R-CNN excels in instance segmentation, object detection, and keypoint detection, making significant contributions to computer vision advancements (Casado-García et al., 2019). The model employs a multi-task loss function, comprising classification loss (Lcls), bounding box loss (Lbbox), and predicted mask loss (Lmask), as represented by equations (2) to (5) (Cong et al., 2023).

L = Lcls+LLbox+Lmask(2)
Lcls=ilogpipi*+1pi*1pi(3)
Lbbox1Nregipi*R (titi*)(4)
Lmask=1m21 i, j myij*logyij+1yij*log1yij(5)

Lcls measures the deviation between predicted and actual values for overall accuracy assessment. Lbbox quantifies the disparity between predicted and actual position parameters, assessing the model’s accuracy in bud localization. Lmask evaluates the model’s confidence in pixel-level classification using binary cross-entropy. Combining these components into a multi-task loss function allows for comprehensive evaluation across multiple tasks, resulting in enhanced overall performance.

The skeleton extraction algorithm is a technique used to extract the central line or skeleton of an object in a binary image (Fu et al., 2023). By progressively shrinking connected regions within the object contour, the algorithm produces a concise contour that provides valuable information for image processing tasks like recognition and matching. Various algorithms, such as Zhang-Suen, Morphological Thinning, and Medial Axis Transform, can be employed for this purpose. The Medial Axis Transform (MAT) algorithm, specifically, extracts the object’s central line by iteratively dilating boundary pixels and identifying the nearest internal pixels as skeleton pixels. This process continues until the skeleton pixels stabilize, resulting in a stable and versatile representation suitable for subsequent image processing tasks. The MAT algorithm handles different object shapes and can process grayscale information within binary images. Seed germination images exhibit a wide range of shape features, such as bud length, curvature, and angle. However, traditional methods for measuring bud length rely on manual measurements, which are time-consuming and prone to significant subjective biases. The MAT (Medial Axis Transform) skeleton extraction algorithm was chosen to obtain the central line of buds. However, the resulting skeleton may contain branches that need to be eliminated to derive the center skeleton. The process of centerline skeleton extraction is illustrated in the following Figure 3B.

In this study, a transparent box with a side length of 250 mm was used as a reference to convert pixels to actual lengths in millimeters. The calculation formula is:

Ratio =Lbox/1164(6)

Here, Lbox represents the side length of the transparent box, and 1164 is the number of pixels corresponding to the transparent box in the image. According to the calculation formula, it can be derived that one pixel corresponds to 0.215 mm.

2.2.5 Establishment of YOLOv7 model for seed germination detection

The seed quality detection methods such as germination and staining techniques are time-consuming and rely heavily on human intervention, which may lead to inaccurate results due to human error. In order to develop an automated and standardized method for detecting seed germination that is efficient, accurate, and reliable, the YOLOv7 (Wang et al., 2022b) object detection algorithm was selected in this study, which is one of the most widely used algorithms for object detection since its release in 2015 (Dewi et al., 2023). YOLOv7 is a real-time object detection algorithm (Soeb et al., 2023), which has evolved from YOLOv5 and has faster inference speed, improved detection accuracy, and reduced computational complexity. The algorithm consists of three main parts: the input layer, backbone layer, and output layer (Tang et al., 2023), and uses either a loss function with or without an auxiliary training head (Zhou et al., 2023).

The loss function is used to update the gradient loss during the training process (Cai et al., 2023). The YOLOv7 algorithm is evaluated using various metrics such as precision, mAP, recall, and F1 score (Zhao et al., 2023), and curves such as the F1-Confidence curve, precision-confidence curve, recall-confidence curve, and precision-recall curve are used to optimize the algorithm’s performance and achieve the best balance between precision and recall.

This study utilized a self-built dataset of maize seeds, comprising images of seeds from various angles and sizes, each with corresponding labels in YOLO format. The data collection and preprocessing process was conducted using the same method as Mask R-CNN. The dataset used in this study consisted of a total of 7000 maize seeds. Among these, 4200 seeds were designated as training samples, 1400 seeds were allocated for the test sets, and the remaining seeds were assigned to the validation sets. To enhance the accuracy and robustness of the model, the YOLOv7.pt (https://github.com/WongKinYiu/yolov7) pretrained weights provided by the official website were employed for training. These weights were trained on a large-scale dataset, which can significantly reduce the training time while improving the training effect. The Adam optimizer, a widely used optimizer that can optimize at different learning rates, was used to update the model parameters during training. The parameters of the Adam optimizer were adjusted based on the size of the learning rate in the training process to achieve better training results. A batch size of 2 and a training iteration of 300 were used in this study.

3 Results and discussion

3.1 Seed germination result

The experimental results showed that the degree of seed aging was significantly correlated with the germination rate. As shown in the Figure 4, on the seventh day of observation, all seeds that were not aged can germinate, and only a few seeds that aged for 2 days failed to do so. Most seeds that aged for 4 days still retained their viability, with only a few seeds that aged for 6 days able to germinate. Seeds that aged for 8 days experience almost complete mortality. Thus, it can be inferred that seed aging leads to a decline in the germination rate, and the more prolonged the aging process, the more apparent the decline in the germination rate.

FIGURE 4
www.frontiersin.org

Figure 4 Germination levels of seeds at different aging times.

3.2 Average spectrum

By analyzing the spectral curve features (Figures 5A, B), it is easy to observe that the spectral reflectance of both wavelength regions increased with the decreasing of maize seed viability, indicating that the light absorption capacity of maize tissue increases with the aging degree. The spectral curves are monotonic in the Vis-SWNIR region, with the average spectral curve gradually increasing in the 400-800 nm region and then slowly decreasing. However, in the LWNIR region, the spectral curve is more complex, capturing two distinct reflection peaks located around 1100 nm and 1300 nm, respectively. The former could potentially be associated with the presence of C-H bonds in lipids, while the latter could be described as a combination of the first overtone of N-H stretching along with the fundamental N-H in-plane bending and C-N stretching with N-H in-plane bending vibrations (Wang et al., 2022d).The spectral curve characteristics can be used to discriminate maize seeds with different germination potentials. As shown in Figure 6, the spectral data of maize seeds with different viability have similar trends in the Vis-SWNIR and LWNIR regions. However, in the Vis-SWNIR region, these curves are basically mixed together, making it difficult to distinguish clearly. In contrast, there are significant differences in the LWNIR region, which may be related to the breakdown of chemicals during the aging process of organisms. Nevertheless, some mixed situations still exist, indicating that it is difficult to distinguish the seeds with or without viability according to the average spectra of hyperspectral image.

FIGURE 5
www.frontiersin.org

Figure 5 Average spectra and the distribution of optimal bands (A) Near-infrared average spectra, (B) Visible and near-infrared average spectra. (C) Selection of characteristic bands in near-infrared spectra, (D) Selection of characteristic bands in visible and near-infrared spectra.

FIGURE 6
www.frontiersin.org

Figure 6 Classification accuracy curves of maize seed viability based on conventional 3DCNN models using Vis-SWNIR hyperspectral image (A) Classification accuracy curves in test set (400 iterations), (B) Classification accuracy curves in test set (70 iterations), (C) Classification accuracy curves in train set (400 iterations), (D) Classification accuracy curves in train set (70 iterations).

3.3 Key wavelength selection of maize seed viability

During the aging process of maize seeds, a series of changes occurs in the internal chemical substances (Xin et al., 2011), with the extent of these changes depending on the degree of seed vitality. These chemical substances include stored energy and nutrients, such as starch, proteins, and lipids (Xu et al., 2022). Proteins may undergo degradation, leading to the release of amino acids and structural damage to proteins. At the same time, the lipid content in the seed gradually oxidizes, resulting in lipid decomposition and the generation of free radicals, thereby affecting the seed’s metabolism and viability. Additionally, starch gradually degrades into soluble sugars. This difference is the main reason for spectral changes during the aging process. After SG and SNV preprocessing, 18 and 11 characteristic bands were extracted from the Vis-SWNIR region and LWNIR region (Figures 5C, D). These characteristic bands were located at the peaks and valleys of the spectrum, reflecting the changes in water content and protein levels of the seeds.

3.4 Maize seed viability detection based on full-wavelength spectra and machine learning

By analyzing the classification accuracy obtained from SVM and Ensemble analysis, there was no significant difference between Vis-SWNIR and LWNIR regions in predicting seed viability (Table 3). However, KNN exhibited slightly higher accuracy with LWNIR, indicating its greater universality and better performance in detecting seed viability. However, due to the minimal differences between seeds with adjacent aging gradients (Feng et al., 2018), particularly those seeds that aged for 4 days and 6 days, these distinctions may not be immediately discernible, presenting a challenge in accurately determining the germination potential of seeds with similar levels of aging. The germination experiment also showed that the seeds with relatively mild aging did not have inherent germination trends and were easily misclassified by the prediction model. This discrepancy may arise from the fact that maize seeds may not exhibit overt phenotypic changes across different stages of aging (Wang et al., 2022e). However, in actuality, mRNA molecules associated with protein synthesis undergo oxidation through physiological mechanisms. More specifically, research unveiled significantly elevated expression levels of mature enzyme genes and ribosomal protein genes in embryonic roots and shoots as compared to other parts(Wang et al., 2022a). This obstruction hampers protein synthesis, consequently impeding the normal physiological functions of the seeds.

TABLE 3
www.frontiersin.org

Table 3 The classification result of maize seed viability based on full-wavelength spectra and machine learning.

3.5 Maize seed viability detection based on key wavelength and 3DCNN model

After 70 training epochs on the Vis-SWNIR hyperspectral images, the accuracy of the training set has stabilized at a high level of 100% (Figure 6B), and the accuracy of the test set has also reached its peak. In order to further validate the stability and robustness of the model, the number of training epochs was increased to 400. After 400 iterations, the accuracy of the training set remained at around 100%, while the accuracy of the test set remained at around 90% (Figures 6A, C).

By using 3DCNN to process the data, not only the spectral information was considered (Wu et al., 2021), but also the image information was integrated, making the evaluation of maize seed quality more comprehensive and accurate (Collins et al., 2021). Compared with machine learning methods that using all spectral bands as input data, the 3DCNN method only used few representative bands. Traditional machine learning methods tend to lose a lot of information, while the 3DCNN method used in this study can learn more complex features and achieved higher accuracy with fewer bands, with an average accuracy increase of 7 percentage points (Table 4). It was worth noting that 3DCNN performs better on the test set and converges faster, which indicated that 3DCNN was an effective method for seed viability detection and had advantages over machine learning classification method in dealing with such problems.

TABLE 4
www.frontiersin.org

Table 4 The classification performance of the maize seed viability based on 3DCNN.

Conventional 3DCNN and multi-scale 3DCNN exhibit different characteristics. Traditional 3DCNN can achieve high accuracy, but they often exhibit slower convergence compared with multi-scale 3DCNN (Figure 6D). Multi-scale 3DCNN incorporated convolutional layers with different-sized kernels and pooling layers, allowing the network to process features of varying scales simultaneously (Lin et al., 2020). This enhanced the network’s robustness and improved its tolerance to noise, distortion, and artifacts in the data, and ultimately led to a faster convergence. In addition, the stability of conventional 3DCNN may not be satisfactory and may exhibit some fluctuations and instability. In contrast, multi-scale 3DCNN perform better, possibly due to their utilization of multi-scale convolutional kernels, enabling them to extract more abundant feature information (Shi et al., 2021) (Figure 6A). Furthermore, the block-based method effectively increased the amount of data and helped to alleviate overfitting. In the final discrimination, this study adopted a majority principle labeling aggregation method to improve the discrimination accuracy (Table 4). To explore the optimal block size, several experiments were conducted, the input images were segmented into different block sizes, including 5 pixels ×5 pixels, 10 pixels ×10 pixels, and 20 pixels × 20 pixels. As shown in Table 4, the model achieved a relatively high overall accuracy when 5 pixels ×5 pixels was used. This suggested that the small blocks with 5 pixels ×5 pixels size can effectively capture more local features of the seedy and provides more discriminative information. Conversely, larger blocks may result in information blurring and confusion, thereby impacting the classification accuracy. Consequently, the block-based method with 5 pixels ×5 pixels was finally selected to enhances the detection accuracy of seed viability.

3.6 Maize seed germination detection based on YOLOv7 model

Figure 7 shows the detection results of germinated maize seeds using the YOLOv7 model, demonstrating its remarkable precision and recall rates of 99.7% and 99.0%, respectively. Additionally, the model achieves a mAP of 99% when applying an Intersection over Union (IoU) threshold of 0.5. Furthermore, the mAP, calculated across a range of IoU thresholds from 0.5 to 0.95, reaches a value of 71%.

FIGURE 7
www.frontiersin.org

Figure 7 Detection performance of YOLOv7 model for maize seed germination (A) The precision of YOLOv7, (B) The recall of YOLOv7, (C) The mAP@0.5 of YOLOv7, (D) The mAP@0.5:0.95 of YOLOv7.

In Table 5, the YOLOv7 model exhibits an impressive F1 score (The F1 score balances precision and recall, providing a comprehensive evaluation of model accuracy) of 0.99 on all target categories with a confidence threshold set at 0.663, highlighting its exceptional detection performance. Consequently, the YOLOv7 model can achieve both high precisions, accurately identifying true positive predictions, and high recall, effectively capturing all relevant targets during detection. With a confidence threshold set to 0.896, the YOLOv7 model achieves a perfect precision accuracy of 100% for the target categories. This noteworthy precision metric showcases the model’s ability to correctly identify true positive predictions among all the positive predictions made, indicating its reliability and precision in detecting target objects. The model impressively achieves a recall rate (The recall rate quantifies the model’s ability to correctly identify positive targets) of 1.00 with a confidence threshold set to 0.000, indicating that it accurately detects all targets of all categories without any missed detections. This ideal performance underscores the model’s high accuracy and proficiency in target detection tasks Additionally, the model exhibits an mAP (The mAP commonly used to evaluate object detection algorithms’ accuracy and robustness) of 0.991 for all target categories when applying an Intersection over Union (IoU) threshold of 0.5. This further demonstrates the model’s superior detection capabilities across various categories, affirming its exemplary performance.

TABLE 5
www.frontiersin.org

Table 5 The detection result of YOLOv7 model for maize seed germination.

R=TPTP+FN(7)
AP =01PRdR(8)
F1=2*P*RP+R(9)
P =TPTP+FP(10)
mAP =1ni=1nAP(11)

In these formulas, True Positives (TP) represent the number of samples where the predicted label is positive and the actual label is also positive. T represents the total number of samples, and False Negatives (FN) indicate the number of samples where the predicted label is negative, but the actual label is positive. Similarly, False Positives (FP) represent the number of samples where the predicted label is positive, but the actual label is negative. Moreover, the area under the precision-recall (P-R) curve, denoted as AP, provides a measure of the model’s performance.

Figure 8A is the confusion matrix of germinated maize seed based on YOLOv7 model, which provides a visual representation of the classification performance, showing the counts of true positive, true negative, false positive, and false negative predictions. The detection accuracy was 95% for germinated seeds and 99% for ungerminated seeds, respectively. Background FP refers to the number when the background is erroneously predicted as a target, fortunately there was no background area was incorrectly classified as a target in this study. Figure 8B shows the actual detection results of YOLOv7 for discriminating seed germination.

FIGURE 8
www.frontiersin.org

Figure 8 Confusion matrix and detection results of germination maize seed based on YOLOv7 model (A) Confusion matrix, (B) Image of detection results.

All indicators mean that the model can essentially replace manual observation for determining seed germination status. Therefore, although this method required some time and manpower for data annotation and training, the overall cost was much lower than manual operation, and can provide a reference for rapid detection of seed germination in crops. On the other hand, the algorithm suffered from the problem of duplicate detection in practical applications (Chen et al., 2023a), resulting in some seeds may be simultaneously labeled as germinated and non-germinated. This phenomenon may lead to a misclassification and reduce the practicality and reliability of the algorithm. Hence, future work will focus on improving the algorithm to solve the duplicate detection problem.

3.7 Maize seed bud length detection based on Mask R-CNN

The Mask R-CNN model achieved an impressive mAP score of 0.9571, indicating its effectiveness and accuracy in detecting and localizing objects. The mAP is a widely used evaluation metric for object detection models, and a high mAP score indicates that the model performs well in both precision and recall, making it a reliable choice for seed germination analysis. Additionally, the loss value during training decreased significantly, stabilizing around 0.21 from an initial value of 2.61, which is a clear indication of the model’s ability to learn and adapt effectively.

Figure 9A showcases a successful instance of skeleton extraction for maize seed germination, resulting in a clear main skeleton after removing branches, which allows for accurate calculation of the bud length. The detection results of bud length for germinated maize seeds, depicted in Figure 9B, demonstrate Mask R-CNN’s impressive capability to accurately segment the seedlings, even when instances overlap or are occluded. This highlights the superiority of the Mask R-CNN model in instance segmentation tasks, making it a valuable tool for precise analysis of seed germination and growth.

FIGURE 9
www.frontiersin.org

Figure 9 The bud length detection of germinated maize seeds (A) The process of skeleton extraction, (B) Probability map of predicting maize sprouts, (C) Prediction of maize seedling length, (D) Regression analysis of actual and predicted corn sprout length values.

Figures 9C, D shows the detection result of bud length with R-squared value of 0.98 and an RMSE of 1.64, demonstrating that the integration of Mask R-CNN model and skeleton extraction method could detect the bud length during seed germination accurately and rapidly. The R-squared value, also known as the coefficient of determination, is a statistical measure that indicates the proportion of the variance in the dependent variable (Bud length in this case) that can be explained by the independent variable (The predicted bud length). Meanwhile, RMSE quantifies the average magnitude of the differences between the predicted bud lengths and the actual observed bud lengths. It is worth mentioning that the bud length of germinated seeds is closely related to their viability (Adebisi et al., 2014). Therefore, the bud length of seeds can be obtained using this algorithm, and the relationship between bud length and viability can be further explored. This not only has important significance for agricultural production but also provides valuable insights for research in other biological fields.

SSR =i=1nYiYi^2 (12)
SST =i=1nYiY¯2 (13)
RMSE=1ni=1nYiYi^2                             (14)
R2=1SSRSST(15)

In these formulas, SSR (Sum of Squares of Residuals) refers to the regression sum of squares, which represents the sum of squared differences between the predicted values and the true values. On the other hand, SST (Total Sum of Squares) stands for the total sum of squares, representing the sum of squared differences between the true values and their mean.Yi refers to the actual value of the i-th observation, while Yi^ represents the predicted value of the i-th observation from the regression model. And n denotes the sample size.

4 Conclusions

The rapid and successful detection of maize seed viability was achieved by leveraging HSI technology in combination with the multi-scale 3DCNN method. In seed viability detection, the 3DCNN method, which utilizes a limited number of representative spectral bands, was found to learn more complex features and achieve higher accuracy compared to using full-wavelength spectra and machine learning methods. By introducing the multi-scale 3DCNN model, the comprehensive consideration of both spectral and image information enabled a more comprehensive and accurate assessment of maize seed quality. Experimental results demonstrated that the adoption of small block sizes (5 pixels × 5 pixels) significantly improved the accuracy of seed viability detection. Furthermore, the YOLOv7 model and Mask R-CNN model were introduced for germination judgment and bud length detection of maize seeds. Both models exhibited outstanding performance in germination judgment and bud length detection, demonstrating excellent detection capabilities. Based on these exceptional detection results, a novel solution for the rapid detection of maize seed germination and bud length was provided. In brief, this study proposed a reliable and effective method for the evaluation of maize seed viability, providing valuable references for agricultural production and germplasm resource preservation.

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Author contributions

YF: Conceptualization, Data collection, Data analysis, Writing – original draft, Writing – review & editing. TA, GY, QW, WH, ZW: providing language help. CZ: Resources, Supervision. TX: Resources, Review-editing, Supervision. All authors contributed to the article and approved the submitted version.

Funding

This study was supported by China Postdoctoral Science Foundation (2022M720492), Postdoctoral Scientific Research Fund of Beijing Academy of Agricultural and Forestry Sciences (2022-ZZ-018), Beijing Postdoctoral Science Foundation (2023-117).

Conflict of interest

The research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Adebisi, M. A., Kehinde, T. O., Porbeni, J. B. O., Oduwaye, O. A., Biliaminu, K., Akintunde, S. A. (2014). Seed and seedling vigour in tropical maize inbred lines. Plant Breed. Seed Sci. 67 (1), 87–102. doi: 10.2478/v10129-011-0072-4

CrossRef Full Text | Google Scholar

Adegbuyi, E., Burris, J. S. (1988). Field criteria used in determining the vigor of seed corn (Zea mays L.) as influenced by drying injury. J. Agron. Crop Sci. 161 (3), 171–177. doi: 10.1111/j.1439-037X.1988.tb00651.x

CrossRef Full Text | Google Scholar

Aloupogianni, E., Ishikawa, M., Ichimura, T., Hamada, M., Murakami, T., Sasaki, A., et al. (2023). Effects of dimension reduction of hyperspectral images in skin gross pathology. Skin Res. Technol. 29 (2), e13270. doi: 10.1111/srt.13270

PubMed Abstract | CrossRef Full Text | Google Scholar

Ambrose, A., Kandpal, L. M., Kim, M. S., Lee, W.-H., Cho, B.-K. (2016). High speed measurement of corn seed viability using hyperspectral imaging. Infrared Phys. Technol. 75, 173–179. doi: 10.1016/j.infrared.2015.12.008

CrossRef Full Text | Google Scholar

An, T., Huang, W., Tian, X., Fan, S., Duan, D., Dong, C., et al. (2022). Hyperspectral imaging technology coupled with human sensory information to evaluate the fermentation degree of black tea. Sensors Actuators B: Chem. 366, 131994. doi: 10.1016/j.snb.2022.131994

CrossRef Full Text | Google Scholar

Association, I.S.T (1999). International rules for seed testing. Rules 1999. (Zurich, Switzerland: International Seed Testing Association).

Google Scholar

Bai, W., Zhao, X., Luo, B., Zhao, W., Huang, S., Zhang, H. (2023). Research on wheat seed germination detection method utilizing YOLOv5. Acta Agriculturae Zhejiangensis 35 (02), 445–454. doi: 10.3969/j.issn.1004-1524.2023.02.22

CrossRef Full Text | Google Scholar

Cai, L., Liang, J., Xu, X., Duan, J., Yang, Z. (2023). Banana pseudostem visual detection method based on improved YOLOV7 detection algorithm. Agronomy 13 (4), 999. doi: 10.3390/agronomy13040999

CrossRef Full Text | Google Scholar

Casado-García, Á., Domínguez, C., García-Domínguez, M., Heras, J., Inés, A., Mata, E., et al. (2019). CLoDSA: a tool for augmentation in classification, localization, detection, semantic segmentation and instance segmentation tasks. BMC Bioinf. 20 (1), 1–14. doi: 10.1186/s12859-019-2931-1

CrossRef Full Text | Google Scholar

Chen, X., Li, F., Chang, Q. (2023b). Combination of continuous wavelet transform and successive projection algorithm for the estimation of winter wheat plant nitrogen concentration. Remote Sens. 15 (4), 997. doi: 10.3390/rs15040997

CrossRef Full Text | Google Scholar

Chen, J., Ma, B., Ji, C., Zhang, J., Feng, Q., Liu, X., et al. (2023a). Apple inflorescence recognition of phenology stage in complex background based on improved YOLOv7. Comput. Electron. Agric. 211, 108048. doi: 10.1016/j.compag.2023.108048

CrossRef Full Text | Google Scholar

Cheng, T., Chen, G., Wang, Z., Hu, R., She, B., Pan, Z., et al. (2023). Hyperspectral and imagery integrated analysis for vegetable seed vigor detection. Infrared Phys. Technol. 131, 104605. doi: 10.1016/j.infrared.2023.104605

CrossRef Full Text | Google Scholar

Collins, T., Maktabi, M., Barberio, M., Bencteux, V., Jansen-Winkeln, B., Chalopin, C., et al. (2021). Automatic recognition of colon and esophagogastric cancer with machine learning and hyperspectral imaging. Diagnostics 11 (10), 1810. doi: 10.3390/diagnostics11101810

PubMed Abstract | CrossRef Full Text | Google Scholar

Cong, P., Li, S., Zhou, J., Lv, K., Feng, H. (2023). Research on instance segmentation algorithm of greenhouse sweet pepper detection based on improved mask RCNN. Agronomy 13 (1), 196. doi: 10.3390/agronomy13010196

CrossRef Full Text | Google Scholar

Cortes, C., Vapnik, V. (1995). Support-vector networks. Support-vector networks. Machine Learning 20 (3), 273–297. doi: 10.1007/BF00994018

CrossRef Full Text | Google Scholar

Cui, H., Bing, Y., Zhang, X., Wang, Z., Li, L., Miao, A. (2022). Prediction of maize seed vigor based on first-order difference characteristics of hyperspectral data. Agronomy 12 (8), 1899. doi: 10.3390/agronomy12081899

CrossRef Full Text | Google Scholar

Cui, H., Cheng, Z., Li, P., Miao, A. (2020). Prediction of sweet corn seed germination based on hyperspectral image technology and multivariate data regression. Sensors 20 (17), 4744. doi: 10.3390/s20174744

PubMed Abstract | CrossRef Full Text | Google Scholar

de Almeida, V. E., de Araújo Gomes, A., de Sousa Fernandes, D. D., Goicoechea, H. C., Galvão, R. K. H., Araújo, M. C. U. (2018). Vis-NIR spectrometric determination of Brix and sucrose in sugar production samples using kernel partial least squares with interval selection based on the successive projections algorithm. Talanta 181, 38–43. doi: 10.1016/j.talanta.2017.12.064

PubMed Abstract | CrossRef Full Text | Google Scholar

Dewi, C., Chen, A. P. S., Christanto, H. J. (2023). Deep learning for highly accurate hand recognition based on yolov7 model. Big Data Cogn. Computing 7 (1), 53. doi: 10.3390/bdcc7010053

CrossRef Full Text | Google Scholar

Fan, S., Li, C., Huang, W., Chen, L. (2018). Data fusion of two hyperspectral imaging systems with complementary spectral sensing ranges for blueberry bruising detection. Sensors 18 (12), 4463. doi: 10.3390/s18124463

PubMed Abstract | CrossRef Full Text | Google Scholar

Feng, L., Zhu, S., Zhang, C., Bao, Y., Feng, X., He, Y. (2018). Identification of maize kernel vigor under different accelerated aging times using hyperspectral imaging. Molecules 23 (12), 3078. doi: 10.3390/molecules23123078

PubMed Abstract | CrossRef Full Text | Google Scholar

Fu, Y., Xia, Y., Zhang, H., Fu, M., Wang, Y., Fu, W., et al. (2023). Skeleton extraction and pruning point identification of jujube tree for dormant pruning using space colonization algorithm. Front. Plant Sci. 13. doi: 10.3389/fpls.2022.1103794

CrossRef Full Text | Google Scholar

Gaikwad, J., Triki, A., Bouaziz, B. (2019). Measuring morphological functional leaf traits from digitized herbarium specimens using traitEx software. Biodiversity Inf. Sci. Standards 3, e37091. doi: 10.3897/biss.3.37091

CrossRef Full Text | Google Scholar

Ge, Z., Cao, G., Li, X., Fu, P. (2020). Hyperspectral image classification method based on 2D–3D CNN and multibranch feature fusion. IEEE J. Selected Topics Appl. Earth Observations Remote Sens. 13, 5776–5788. doi: 10.1109/jstars.2020.3024841

CrossRef Full Text | Google Scholar

Ghaderizadeh, S., Abbasi-Moghadam, D., Sharifi, A., Zhao, N., Tariq, A. (2021). Hyperspectral image classification using a hybrid 3D-2D convolutional neural networks. IEEE J. Selected Topics Appl. Earth Observations Remote Sens. 14, 7570–7588. doi: 10.1109/jstars.2021.3099118

CrossRef Full Text | Google Scholar

Han, X., Jiang, Z., Liu, Y., Zhao, J., Sun, Q., Li, Y. (2022). A spatial–spectral combination method for hyperspectral band selection. Remote Sens. 14 (13), 3217. doi: 10.3390/rs14133217

CrossRef Full Text | Google Scholar

He, K., Gkioxari, G., Dollár, P., Girshick, R. (2017). Mask r-cnn in Proceedings of the IEEE international conference on computer vision. 2961–2969.

Google Scholar

Köppen, M. (2000). The curse of dimensionality, in 5th online world conference on soft computing in industrial applications. (WSC5). 1, 4–8.

Google Scholar

Li, Y., Zhang, H., Shen, Q. (2017). Spectral–spatial classification of hyperspectral imagery with 3D convolutional neural network. Remote Sens. 9 (1), 67. doi: 10.3390/rs9010067

CrossRef Full Text | Google Scholar

Liang, X.-y., Guo, F., Feng, Y., Zhang, J.-L., Yang, S., Meng, J.-J., et al (2020). Single-seed sowing increased pod yield at a reduced seeding rate by improving root physiological state of Arachis hypogaea. J. Integr. Agric. 19 (4), 1019–1032. doi: 10.1016/S2095-3119(19)62712-7

CrossRef Full Text | Google Scholar

Lin, B., Zhang, S., Bao, F. (2020). “Gait recognition with multiple-temporal-scale 3D convolutional neural network,” in Proceedings of the 28th ACM International Conference on Multimedia. (Seattle, WA, USA: Association for Computing Machinery).

Google Scholar

Liu, C., Chu, Z., Weng, S., Zhu, G., Han, K., Zhang, Z., et al. (2022). Fusion of electronic nose and hyperspectral imaging for mutton freshness detection using input-modified convolution neural network. Food Chem. 385, 132651. doi: 10.1016/j.foodchem.2022.132651

PubMed Abstract | CrossRef Full Text | Google Scholar

Long, Y., Wang, Q., Tang, X., Tian, X., Huang, W., Zhang, B. (2022). Label-free detection of maize kernels aging based on Raman hyperspcectral imaging techinique. Comput. Electron. Agric. 200, 107229. doi: 10.1016/j.compag.2022.107229

CrossRef Full Text | Google Scholar

Masood, M., Nazir, T., Nawaz, M., Javed, A., Iqbal, M., Mehmood, A. (2021). Brain tumor localization and segmentation using mask RCNN. Front. Comput. Sci. 15 (6). doi: 10.1007/s11704-020-0105-y

PubMed Abstract | CrossRef Full Text | Google Scholar

Pang, L., Wang, J., Men, S., Yan, L., Xiao, J. (2021). Hyperspectral imaging coupled with multivariate methods for seed vitality estimation and forecast for Quercus variabilis. Spectrochimica Acta Part A: Mol. Biomolecular Spectrosc. 245, 118888. doi: 10.1016/j.saa.2020.118888

CrossRef Full Text | Google Scholar

Shen, R., Zhen, T., Li, Z. (2023). Segmentation of unsound wheat kernels based on improved mask RCNN. Sensors 23 (7), 3379. doi: 10.3390/s23073379

PubMed Abstract | CrossRef Full Text | Google Scholar

Shi, W., Du, C., Gao, B., Yan, J. (2021). Remote sensing image fusion using multi-scale convolutional neural network. J. Indian Soc. Remote Sens. 49 (7), 1677–1687. doi: 10.1007/s12524-021-01353-2

CrossRef Full Text | Google Scholar

Soeb, M. J. A., Jubayer, M. F., Tarin, T. A., Al Mamun, M. R., Ruhad, F. M., Parven, A., et al (2023). Tea leaf disease detection and identification based on YOLOv7 (YOLO-T). Sci. Rep. 13 (1), 6078. doi: 10.1038/s41598-023-33270-4

PubMed Abstract | CrossRef Full Text | Google Scholar

Suksungworn, R., Sanevas, N., Wongkantrakorn, N., Fangern, N., Vajrodaya, S., Duangsrisai, S. (2021). Phytotoxic effect of Haldina cordifolia on germination, seedling growth and root cell viability of weeds and crop plants. NJAS: Wageningen J. Life Sci. 78 (1), 175–181. doi: 10.1016/j.njas.2016.05.008

CrossRef Full Text | Google Scholar

Sun, W., Du, Q. (2019). Hyperspectral band selection: A review. IEEE Geosci. Remote Sens. Magazine 7 (2), 118–139. doi: 10.1109/mgrs.2019.2911100

CrossRef Full Text | Google Scholar

Sun, K., Wang, A., Sun, X., Zhang, T. (2022). Hyperspectral image classification method based on M-3DCNN-Attention. J. Appl. Remote Sens. 16 (02), 026507. doi: 10.1117/1.Jrs.16.026507

CrossRef Full Text | Google Scholar

Tang, F., Yang, F., Tian, X. (2023). Long-distance person detection based on YOLOv7. Electronics 12 (6), 1502. doi: 10.3390/electronics12061502

CrossRef Full Text | Google Scholar

Tian, X., Zhang, C., Li, J., Fan, S., Yang, Y., Huang, W. (2021). Detection of early decay on citrus using LW-NIR hyperspectral reflectance imaging coupled with two-band ratio and improved watershed segmentation algorithm. Food Chemistry 360, 130077. doi: 10.1016/j.foodchem.2021.130077

PubMed Abstract | CrossRef Full Text | Google Scholar

Triki, A., Bouaziz, B., Gaikwad, J., Mahdi, W. (2021). Deep leaf: Mask R-CNN based leaf detection and segmentation from digitized herbarium specimen images. Pattern Recognition Lett. 150, 76–83. doi: 10.1016/j.patrec.2021.07.003

CrossRef Full Text | Google Scholar

Wang, C. -Y., Bochkovskiy, A., Liao, H. -Y. M. (2022b). “YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors”, in: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition), 7464–7475. doi: 10.48550/arXiv.2207.02696

CrossRef Full Text | Google Scholar

Wang, Z., Huang, W., Tian, X., Long, Y., Li, L., Fan, S. (2022e). Rapid and non-destructive classification of new and aged maize seeds using hyperspectral image and chemometric methods. Front. Plant Sci. 13. doi: 10.3389/fpls.2022.849495

CrossRef Full Text | Google Scholar

Wang, W., Huang, W., Yu, H., Tian, X. (2022d). Identification of maize with different moldy levels based on catalase activity and data fusion of hyperspectral images. Foods 11 (12), 1727. doi: 10.3390/foods11121727

PubMed Abstract | CrossRef Full Text | Google Scholar

Wang, S., Sun, G., Zheng, B., Du, Y. (2021). A crop image segmentation and extraction algorithm based on mask RCNN. Entropy 23 (9), 1160. doi: 10.3390/e23091160

PubMed Abstract | CrossRef Full Text | Google Scholar

Wang, J., Yan, L., Wang, F., Qi, S., Jin, X.-B. (2022c). SVM classification method of waxy corn seeds with different vitality levels based on hyperspectral imaging. J. Sensors 1–13. doi: 10.1155/2022/4379317

CrossRef Full Text | Google Scholar

Wang, B., Yang, R., Ji, Z., Zhang, H., Zheng, W., Zhang, H., et al. (2022a). Evaluation of biochemical and physiological changes in sweet corn seeds under natural aging and artificial accelerated aging. Agronomy 12 (5), 1028. doi: 10.3390/agronomy12051028

CrossRef Full Text | Google Scholar

Wu, H., Li, D., Wang, Y., Li, X., Kong, F., Wang, Q. (2021). Hyperspectral image classification based on two-branch spectral–spatial-feature attention network. Remote Sens. 13 (21), 4262. doi: 10.3390/rs13214262

CrossRef Full Text | Google Scholar

Xin, X., Lin, X.-H., Zhou, Y.-C., Chen, X.-L., Liu, X., Lu, X.-X. (2011). Proteome analysis of maize seeds: the effect of artificial ageing. Physiologia Plantarum 143 (2), 126–138. doi: 10.1111/j.1399-3054.2011.01497.x

PubMed Abstract | CrossRef Full Text | Google Scholar

XingJia, T., PengChang, Z., ZongBen, X., BingLiang, H., Lakshmanna, K. (2022). Calligraphy and painting identification 3D-CNN model based on hyperspectral image MNF dimensionality reduction. Comput. Intell. Neurosci. 2022, 1–19. doi: 10.1155/2022/1418814

CrossRef Full Text | Google Scholar

Xu, Y., Ma, P., Niu, Z., Li, B., Lv, Y., Wei, S., et al. (2022). Effects of artificial aging on physiological quality and cell ultrastructure of maize (Zea mays L.). Cereal Res. Commun. 51, 615–626. doi: 10.1007/s42976-022-00328-4

CrossRef Full Text | Google Scholar

Yasmin, J., Ahmed, M. R., Wakholi, C., Lohumi, S., Mukasa, P., Kim, G., et al. (2022). Near-infrared hyperspectral imaging for online measurement of the viability detection of naturally aged watermelon seeds. Front. Plant Sci. 13. doi: 10.3389/fpls.2022.986754

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhang, S. (2022). Challenges in KNN classification. IEEE Trans. Knowledge Data Eng. 34 (10), 4663–4675. doi: 10.1109/tkde.2021.3049250

CrossRef Full Text | Google Scholar

Zhang, T., Fan, S., Xiang, Y., Zhang, S., Wang, J., Sun, Q. (2020). Non-destructive analysis of germination percentage, germination energy and simple vigour index on wheat seeds during storage by Vis/NIR and SWIR hyperspectral imaging. Spectrochimica Acta Part A: Mol. Biomolecular Spectrosc. 239, 118488. doi: 10.1016/j.saa.2020.118488

CrossRef Full Text | Google Scholar

Zhao, J., Ma, Y., Yong, K., Zhu, M., Wang, Y., Luo, Z., et al. (2022). Deep-learning-based automatic evaluation of rice seed germination rate. J. Sci. Food Agric. 103 (4), 1912–1924. doi: 10.1002/jsfa.12318

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhao, J., Ma, Y., Yong, K., Zhu, M., Wang, Y., Wang, X., et al. (2023). Rice seed size measurement using a rotational perception deep learning model. Comput. Electron. Agric. 205, 107583. doi: 10.1016/j.compag.2022.107583

CrossRef Full Text | Google Scholar

Zhao, W., Chellappa, R., Phillips, P. J. (1999). Subspace linear discriminant analysis for face recognition. University of Maryland at College Park, USA: Citeseer.

Google Scholar

Zhou, J., Zhang, Y., Wang, J. (2023). RDE-YOLOv7: an improved model based on YOLOv7 for better performance in detecting dragon fruits. Agronomy 13 (4), 1042. doi: 10.3390/agronomy13041042

CrossRef Full Text | Google Scholar

Keywords: viability detection, maize seeds, hyperspectral imaging, YOLOv7 model, 3D convolution neural network

Citation: Fan Y, An T, Wang Q, Yang G, Huang W, Wang Z, Zhao C and Tian X (2023) Non-destructive detection of single-seed viability in maize using hyperspectral imaging technology and multi-scale 3D convolutional neural network. Front. Plant Sci. 14:1248598. doi: 10.3389/fpls.2023.1248598

Received: 27 June 2023; Accepted: 11 August 2023;
Published: 29 August 2023.

Edited by:

Jianwei Qin, Agricultural Research Service (USDA), United States

Reviewed by:

Pappu Kumar Yadav, University of Florida, United States
Ebenezer Olaniyi, Mississippi State University, United States
Princess Tiffany D. Mendoza, Kansas State University, United States

Copyright © 2023 Fan, An, Wang, Yang, Huang, Wang, Zhao and Tian. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Chunjiang Zhao, Zhaocj@nercita.org.cn; Xi Tian, tianx2019@sina.com

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.