Skip to main content

ORIGINAL RESEARCH article

Front. Plant Sci., 20 November 2023
Sec. Technical Advances in Plant Science
This article is part of the Research Topic IPPS 2022 - Plant Phenotyping for a Sustainable Future View all 14 articles

Comparing CNNs and PLSr for estimating wheat organs biophysical variables using proximal sensing

  • 1Biosystems Dynamics and Exchanges, TERRA Teaching and Research Center, Gembloux Agro-Bio Tech, University of Liège, Gembloux, Belgium
  • 2Plant Sciences, TERRA Teaching and Research Center, Gembloux Agro-Bio Tech, University of Liège, Gembloux, Belgium

Estimation of biophysical vegetation variables is of interest for diverse applications, such as monitoring of crop growth and health or yield prediction. However, remote estimation of these variables remains challenging due to the inherent complexity of plant architecture, biology and surrounding environment, and the need for features engineering. Recent advancements in deep learning, particularly convolutional neural networks (CNN), offer promising solutions to address this challenge. Unfortunately, the limited availability of labeled data has hindered the exploration of CNNs for regression tasks, especially in the frame of crop phenotyping. In this study, the effectiveness of various CNN models in predicting wheat dry matter, nitrogen uptake, and nitrogen concentration from RGB and multispectral images taken from tillering to maturity was examined. To overcome the scarcity of labeled data, a training pipeline was devised. This pipeline involves transfer learning, pseudo-labeling of unlabeled data and temporal relationship correction. The results demonstrated that CNN models significantly benefit from the pseudolabeling method, while the machine learning approach employing a PLSr did not show comparable performance. Among the models evaluated, EfficientNetB4 achieved the highest accuracy for predicting above-ground biomass, with an R² value of 0.92. In contrast, Resnet50 demonstrated superior performance in predicting LAI, nitrogen uptake, and nitrogen concentration, with R² values of 0.82, 0.73, and 0.80, respectively. Moreover, the study explored multi-output models to predict the distribution of dry matter and nitrogen uptake between stem, inferior leaves, flag leaf, and ear. The findings indicate that CNNs hold promise as accessible and promising tools for phenotyping quantitative biophysical variables of crops. However, further research is required to harness their full potential.

1 Introduction

Biophysical vegetation variables are critical indicators of plant growth and health, providing essential information for understanding complex plant-environment interactions (Hawkesford and Riche, 2020; Lemaire and Ciampitti, 2020). Among these variables, Leaf Area Index (LAI), Aboveground Biomass (AGB), and Nitrogen Uptake (Nupt) stand out as key parameters that aid in crop monitoring and yield prediction. Additionally, they play a pivotal role in unraveling the underlying physiological processes that govern the intricate associations between final yield, genotype, and the surrounding environment. As concerns about climate change and the human food security continue to intensify, the accurate assessment of vegetation variables becomes increasingly crucial (Hickey et al., 2019). Timely and reliable information on crop growth and health can help optimize agricultural practices, enhance resource utilization, and help breeders and researchers improve crops.

Recent developments in phenotyping systems, utilizing multiple remote sensing platforms such as satellites, drones, and ground platforms equipped with various sensors (e.g., RGB, spectral data, thermal, LiDAR, etc…), have led to an improvement in the high-throughput and non-destructive screening of crops (Reynolds et al., 2020; Araus et al., 2022; Sun et al., 2022). These technologies have enabled the collection of large volumes of image data, facilitating the rapid, non-invasive, and detailed acquisition of plant phenotyping traits throughout the entire crop lifecycle (Verrelst et al., 2019). Remote sensing, which has lower spatial resolution, can capture the canopy in its entirety in a fast way. In contrast, proximal sensing provides more precise measurements at the organ level and might better handle the impact of unwanted factors (Deery et al., 2014). Ground-based phenotyping systems equipped with multiple sensors can acquire high-resolution data, facilitating improved identification of plant organs, diseases, or yellow and green plant parts (Carlier et al., 2022; Dandrifosse, 2022; Serouart et al., 2022; Tanner et al., 2022; Xu and Li, 2022). The integration of big data and machine/deep learning techniques further enhances the potential for precision phenotyping, enabling more accurate and efficient analyses of crop characteristics for enhanced agricultural management and breeding practices (Verrelst et al., 2019).

The assessment of such biophysical variables using remote sensing and proximal sensing methods requires a comprehensive understanding of agronomy, image and data analysis, given the inherent complexity of these traits and their susceptibility to various influencing factors. Usual methods for estimating AGB and LAI rely on crop architecture, vegetation indices, radiative transfer models, or a combination of these models (Tilly et al., 2015; Brocks and Bareth, 2018; Yue et al., 2019; Raj et al., 2021; Schiefer et al., 2021; Wan et al., 2021). Such methods are also widely used for assessing crop nitrogen status (Berger et al., 2020).

The algorithm pipeline commonly used in plant phenotyping comprises several stages, which involve feature extraction through image analysis methods, including color information collection, thresholding, edge detection, or/and pattern recognition. While these methods can be effective, their reliance on handcrafted features and hyperparameter tuning often results in a lack of robustness. This limitation becomes particularly evident when dealing with complex environmental conditions, such as the presence of soil, weeds, and biotic and abiotic stresses, as well as variations in plant characteristics like growth stage and canopy architecture. Thus, many phenotyping studies focus solely on local areas or specific agricultural practices, leading to limitations in the broader applicability and generalization of proposed models (Chao et al., 2019).

These challenges can lead to suboptimal performance and reduced accuracy in plant phenotyping tasks (Kamilaris and Prenafeta-Boldú, 2018; Nabwire et al., 2021). Yet, it becomes paramount to design studies that effectively capture the diversity present within crop populations and account for the variability of growing conditions. By doing so, we could unlock valuable insights into the intricate interactions shaping these biophysical variables, fostering more robust and adaptable solutions for the future (Hawkesford and Riche, 2020). To address these issues, researchers have been exploring the potential of deep learning and artificial intelligence techniques also in agricultural applications. These approaches have shown promising results in overcoming the limitations of traditional methods by automatically learning relevant features and adaptively adjusting to diverse conditions.

By leveraging advanced machine learning algorithms, such as deep neural networks and convolutional neural networks (CNNs), plant phenotyping can benefit from improved accuracy and generalization across varying scenarios (Singh et al., 2018; Kattenborn et al., 2021; Arya et al., 2022). These methods excel in handling complex datasets and can effectively capture intricate patterns and relationships in plant-related data. Additionally, they reduce the need for manual feature engineering and parameter tuning, leading to more efficient and reliable analyses. For instance, when predicting wheat biomass during early growth stages, CNNs demonstrated less susceptibility to plant density variations compared to alternative methods (Ma et al., 2019). Moreover, these innovative approaches enhance the ability to accurately estimate traits and unlock the extraction of more advanced parameters, such as crop growth rate, particularly when applied to time-series data (Buxbaum et al., 2022). Furthermore, their remarkable ability to solve highly complex patterns makes them ideal for multi-output purposes, enabling the production of multi-trait outputs using a single model (Pound et al., 2017; Nguyen et al., 2023).

The accessibility of ready-to-use libraries, datasets, and emerging methodologies like transfer learning has enabled the application of sophisticated algorithms to crop characterization. The ever-growing availability of neural networks architectures and hyperparameters can present a challenge when it comes to selecting or designing the most suitable architecture. While some authors have successfully created their own neural architectures that perform comparably to well-known ones in terms of accuracy (Li et al., 2021), it is still highly recommended to use established and widely recognized architectures. Nonetheless, ensuring the accuracy and robustness of these models is crucial, and their training and validation with large ground-truth datasets remain essential. This becomes particularly challenging when dealing with biophysical variables, such as AGB, which require a significant amount of human labor and destructive measurements to construct a dataset (Jiang and Li, 2020). This could explain why regression CNN is not yet widely adopted.

To address the need for data, several methods have been proposed to train robust models with a limited amount of labeled data. One approach is to use pre-trained models with transfer learning, which has been successful in estimating forage biomass (Castro et al., 2020; de Oliveira et al., 2021). However, when dealing with multispectral images, pre-trained models that are generally trained on RGB images may not perform well. Another approach is to use data augmentation to artificially increase the dataset size by applying transformations to the images. Advanced data augmentation methods, such as generative adversarial networks (GANs), have been used to improve wheat yield estimation (Zhang et al., 2022). Yet, phenotyping users often acquire large amounts of unlabeled data that still can be used to train a part of a CNN. Semi-supervised learning methods could be used to pre-trained the convolutional parts of CNN from unlabeled datasets (Zbontar et al., 2021). Additionally, one can predict labels for unlabeled data and subsequently insert them into the training dataset if they meet certain criteria; this technique is known as pseudo-labeling (Lee, 2013).

The use of CNNs in various domains has shown promise, and their potential in agriculture for regression purposes needs more investigation. The current study investigate the use of CNN for estimating biophysical variables such as AGB, LAI, nitrogen concentration, and nitrogen uptake from proximal images of wheat. While some studies have already shown some good examples of the use of CNNs for biomass or LAI prediction (Ma et al., 2019; Li et al., 2021; Sapkota et al., 2022; Schreiber et al., 2022; Zheng et al., 2022), many questions remain unanswered. These unanswered questions encompass identifying the optimal CNN architectures for achieving superior performance in estimating LAI, above-ground biomass, nitrogen uptake, and nitrogen concentration for wheat organs utilizing RGB and multispectral close-range images. Additionally, addressing the challenges related to insufficient training data and devising an effective training pipeline is imperative. Furthermore, there is a need to evaluate the effectiveness of multi-output models in assessing dry matter and nitrogen uptake partitioning, as well as nitrogen concentration partitioning in various wheat organs. Lastly, the best-performing CNN methods will be compared to a traditional machine learning approach, a Partial Least Squares regression (PLSr), using feature engineering.

2 Materials and methods

2.1 Experimental design

Data were acquired on winter wheat trials during four years in the Hesbaye area, Belgium (50 33’50” N and 4 42’00” E). Experimental microplots measuring 1.95 m × 6 m were sown with an inter-row spacing of 0.14 m, on homogeneous deep silt loamy soil in a temperate climate. The microplots were fertilized with 27% ammonium nitrate during the tillering, stem elongation and flag leaf stage corresponding to the BBCH 28, 30 and 39 growth stages, respectively. The trials were of two types: (i) trials testing different fertilization fractioning, noted as F and detailed in Table S1 and in Table S2, (ii) trials composed of different fertilization fractioning combined with different fungicide application programs, noted as FP and detailed in Table S3. These abbreviations, along with the year of experimentation, are used in the trial names presented in Table 1.

TABLE 1
www.frontiersin.org

Table 1 Field trial details.

2.1.1 Reference measurements

Manual measurements were conducted on major phenological growth stages (Table 1), which mainly consisted of tillering, stem elongation, flag leaf, flowering, grain development, and maturity stages. The F trials involved five treatments, while the FP trials involved seven treatments, with three and four replicates conducted, respectively. Fresh AGB was sampled from the three central rows of the microplot over a length of 0.50 m. In the laboratory, the samples were manually separated into ear, stem, flag leaf (L1) and inferior leaves (Linf) groups. Each part was subsequently dried to determine the associated dry matter (DM) expressed in t/ha. The nitrogen concentration (%N) was then measured using the Dumas method, and nitrogen uptake (Nupt) was calculated by multiplying the DM by the corresponding %N, expressed in kgN/ha. Organs DM and Nuptake values of organs were expressed as relative values, representing the partitioning of DM and N uptake among each organ. These relative values indicate the proportion of each organ in relation to the total plant values. Additionally, the Nitrogen Nutrition Index (NNI) was computed using the traditional approach described in (Justes, 1994).

To determine LAI, plants were sampled by taking one row measuring 0.50 m in length. The leaves were separated from the stems, weighed, spread on a white paper using a transparent adhesive sheet, and scanned. An Otsu segmentation method was employed to isolate the leaves from the white background (Otsu, 1979). The leaf surface area was calculated by summing the areas of the scanned paper sheets multiplied by the proportion of pixels segmented as leaf. Since this protocol was time-consuming, only five microplots with contrasting fertilization were selected for manual LAI measurements at each collection date. These LAI values were correlated with the associated fresh masses by means of a linear regression to predict the LAI of the other microplots. Each correlation had a really high correlation above 0.9, thus validate this method as a reference.

2.1.2 Image acquisitions

To capture nadir frames of wheat microplots, a phenotyping platform was designed (Figure 1). In 2019, a single RGB camera was utilized, while a sensor pod combining two types of cameras was employed in 2020, 2021, and 2022. The sensor pod comprised of two close-up RGB cameras dedicated to stereovision. These RGB cameras were GO-5000C-USB cameras from JAI A/S in Copenhagen, Denmark, and featured a 2560 × 2048 CMOS sensor. Additionally, a multispectral camera, the Micro-MCA from Tetracam Inc. in Gainesville, FL, USA, was used. It had six 1280 × 1024 pixel CMOS sensors, each of them equipped with narrow filter centered respectively at 490, 550, 680, 720, 800, and 900 nm. To avoid shadows from the rest of the platform in the images, both cameras were installed on a cantilever beam. The height of the cameras was adjusted at each acquisition date to maintain a consistent distance between the cameras and the top of the canopy. The height was about 1 m in 2019 and 1.6 m for the other years. Two to four images were taken per microplots for both cameras.

FIGURE 1
www.frontiersin.org

Figure 1 Experimental setup. A ground mobile platform (on the left) was equipped with a camera pod (on the right) comprising two high-resolution RGB cameras, a multispectral camera, and an incident light spectrometer, all positioned at a height of 1.6m above the canopy.

The RGB images were recorded using a color depth of 12-bit per pixel in 2019, 2020 and 2021, which were then converted to 8-bit to match the following algorithms. The multispectral grey scale images were converted from 10 to 8-bit, in accordance with the constructor recommendations. In 2019, the RGB camera auto-exposure algorithm was used. Then, a custom exposure algorithm was developed to limit the number of saturated pixels to less than 1%. The multispectral auto-exposure algorithm was based on a master-slaves principle. The 800 nm filter served as the master and its exposure time was determined automatically using the manufacturer algorithm. The exposure time of each slave filters was then defined as a ratio of the master time. These ratios were adjusted across the season to avoid saturated pixels.

The cropping seasons were thoroughly covered, with multiple image acquisitions from tillering to maturity (Figure 2). Nevertheless, some unforeseen events occurred, such as the COVID-19 pandemic and a violent storm in 2021, which disrupted data acquisition.

FIGURE 2
www.frontiersin.org

Figure 2 Overview of the data acquisitions during the cropping seasons. Green diamonds represent the image acquisitions, and the blue crosses the agronomic samples.

The multispectral images underwent two pre-processing steps. The first step involved image registration to correct for shifts between the gray-scale images caused by the proximity to the canopy and the physical lenses gap. The considered method proposed by Dandrifosse et al. (2021) employs a b-spline approach to achieve pixel-wise alignment. The second step involved correcting the multispectral images for different light conditions during acquisition, using the method described by Dandrifosse et al. (2022). A laboratory calibration was performed to convert the digital numbers of the images to Bi-directional Reflectance Factor (BRF), known as reflectance, using an Incident Light Spectrometer, specifically an AvaSpec-ULS2048 from Avantes, Apeldoorn, The Netherlands.

2.2 Partial Least Squares regression approach

A conventional machine learning approach was tested to confront the CNN models presented below. As machine learning algorithms require relevant image features to be extracted, additional processes were applied after performing the pre-processing steps as described in the previous section. Firstly, a stereovision process was used to extract plant height information using the 95th percentile of the height map (Dandrifosse et al., 2020). Secondly, the plant ratio was computed as the proportion of plants in the scene, using a simple threshold method on the 800 nm image as detailed in Dandrifosse et al. (2022). Finally, twelve vegetation indices (see Table S7) were computed using the six BRFs.

A Partial Least Squares regression (PLSr) model was trained and validated using these twenty features for DM, %N and Nupt of the entire plant. It is worth noting that PLSr has previously exhibited good performance in analogous studies (Freitas Moreira et al., 2021). To fine-tune the model, a sequential backward feature selection approach was employed similar to (Song et al., 2022). This method involved generating all possible feature subsets of size n - 1, where n represents the total number of features. Each subset was rigorously assessed using a 5-fold cross-validation technique on the training dataset. The feature to be removed at each step was determined based on the subset’s performance, with the least contributing feature being eliminated. This iterative process continued until the maximum R² value was achieved. It is important to mention that the training data did not encompass the 2019 dataset, primarily due to the limited availability of only one RGB camera during that period. Furthermore, the efficacy of the pseudo-labeling strategy, as described in Section 2.3.3 was also explored for PLSr. This training was performed using the PLSr default parameters from the Scikit-Learn 1.3 Python library (Pedregosa et al., 2011).

2.3 CNN training

2.3.1 Architecture

Three CNN architectures available in the python library Tensorflow 2.4. and Keras 2.4 were tested in this study. They were Resnet50 (He et al., 2015) and EfficientNetB0 and B4 (Tan and Le, 2020). They represent the actual state-of-the-art CNN models with different properties (i.e., architecture and number of parameters) and purposes. Resnet 50 was already used for biomass prediction by (Zheng et al., 2022) and EfficientNet is a cutting-edge neural network architecture with a remarkable ability to seamlessly scale from smaller to larger sizes while maintaining good efficiency.

The CNN architectures were customized to perform two tasks: (i) a single-output model to estimate LAI, DM, %N and Nupt of the whole plant respectively; and (ii) a multi-output model to estimate DM, %N, or Nupt of each wheat organ, also referred as partitioning model in the rest of this paper. Multi-output, also known as multi-task model have already been successfully used in phenotyping by (Nguyen et al., 2023) to predict a set of traits using a single model. Whereas a multivariate model deals with multiple dependent variables and aims to model their relationships, a multi-output model is a machine learning model designed to predict multiple output variables simultaneously. A linear activation function was considered for the last neuron of each single-output model. Regarding the multi-output models, four output neurons were considered, one for each organ. A linear activation function was used for the estimation of %N whereas the softmax activation function were used for the relative values of DM and Nupt, i.e., the proportion, in order to keep the values between 0 and 1. All models were initialized with weights from the ImageNet dataset (Deng et al., 2009).

The CNN architectures were originally designed for three-channel images, but the multispectral images used in this study had six channels. To accommodate this, a 2D convolutional layer with three filters and a kernel size of (1,1) was added at the beginning of each model when using multispectral images. It allowed to provide a three channels input required for the selected CNN models with pre-trained weights.

2.3.2 Dataset configuration

The study used a dataset consisting of 1809 RGB images and 1391 multispectral images with their corresponding reference measurements. These numbers correspond to the multiplication of dates, samples, replicates, and images per microplots. Each image was associated with a specific combination of agronomic variables. From this dataset, two treatments from F trials (Tables S1, S2), and one treatment from FP trials (Table S3) were selected for the validation dataset that included 424 RGB images and 341 multispectral images.

In addition to the images acquired on the same days as the manual sampling, each trial was monitored continuously throughout the season, as illustrated in Figure 2. All those acquisitions yielded a dataset comprising 16 812 RGB images and 14 491 multispectral images. To prepare the data for the CNN models, some pre-processing steps were taken.

The first pre-processing step involved determining the image size, which is a trade-off between retaining as much information as possible and limiting the computing time and resources required. Additionally, when using pre-trained models, it is recommended to set the input image size to match the size used during initial training. Therefore, all images were resized to 224 x 224 for the ResNet50 and EfficientNetB0 models, and to 380 x 380 for the EfficientNetB4 model. It is worth noting that the images were previously cropped into a square to avoid distortion.

In addition to image resizing, the pixel scaling was also adjusted for each model. For the RGB images, pixel scaling was adapted according to the Keras documentation and the requirements of each model. For the multispectral images, Bi-directional Reflectance Factor (BRF) values were first normalized between 0 and 1. Next, the data was standardized based on the mean and standard deviation of the training dataset as advice by Tensorflow. To further enhance the dataset, data augmentation techniques, namely random flip up/down and right/left, were applied. These techniques increase the diversity of the dataset, which can improve the generalization performance of the models.

2.3.3 Training pipeline

In the field of phenotyping, researchers often encounter a substantial amount of unlabeled data. However, these data hold untapped potential for enhancing the performance of machine learning models. In this study, a pseudo-labeling method was employed to leverage the unlabeled data effectively. Pseudo-labeling involves predicting the labels of unlabeled data using a model that demonstrates acceptable performance. These predicted labels, known as pseudo-labels, can then be incorporated into the training dataset, subject to a predefined confidence threshold. For classification tasks, this confidence threshold is based on class probabilities. Nevertheless, regression tasks utilize a linear activation function, leading to the absence of probabilities. To overcome this challenge, this research proposes a novel approach. The predicted biophysical variables from each microplot were plotted against time to generate a crop growth curve. This curve characterizes the growth pattern of the crop over time and can be harnessed to rectify the predicted values.

Based on this idea, a well-defined pipeline was constructed (see Figure 3). The pipeline entailed utilizing CNN models pre-trained on ImageNet through transfer learning. The initial training phase involved training the CNN models for 40 epochs with a learning rate of 1×10−3. During this process, only the last layer, specifically the linear dense layer, was trained, while keeping the remaining layers frozen.

FIGURE 3
www.frontiersin.org

Figure 3 Proposed training pipeline. (1) is the training with transfer learning, and (2) is the training with pseudo-labels. Ytrue corresponds to the reference measurements. Ypred are predicted labels. A curve is fitted to provide the Ypseu which represent the corrected pseudo-labels. n and m correspond to the number of reference measurements and the total number of images respectively.

Following this, a fine-tuning stage was conducted for 10 epochs, with a reduced learning rate of 1×10−5. During fine-tuning, the last convolutional layer block was unfrozen and retrained, leading to the creation of Model 1.

Next, Model 1 was utilized to generate predicted labels (Ypred) for the complete training dataset. These predicted labels were then plotted against the Photo-Vernalo-Thermic Units (°C-days) (Duchene et al., 2021). A cubic B-Spline for LAI and a cubic polynomial function for the other variables was fitted with a high smoothing condition. These curves are traditionally used in biophysical variables modeling (van Eeuwijk et al., 2019). Basic correction conditions were also implemented to help that fitting, such as setting organ values to 0 when they were not present at specific times. The outcome of this process yielded a fitted curve from which “corrected” pseudo-labels (Ypseu) could be extracted.

Last, pre-trained CNNs from ImageNet were trained on the corrected pseudo-labels (Ypseu) for 30 epochs, using a learning rate of 1×10−5. This resulted in the development of Model 2, which was thus trained on a much larger dataset compared to Model 1.

The Mean Square Error (MSE) loss function and Adam optimizer were used in all models. However, in the case of multi-output model for %N, the MSE calculation was limited to true labels above 0. This means that if an organ was not yet visible (e.g., the ear during tillering growth stage), the loss function did not take it into account, which prevented it from interfering with the loss function. Additionally, a weight was applied to the loss calculation when working with relative multi-output models. Specifically, the flag leaf pool weights were multiplied by twenty to ensure consistency with the order of magnitude of the other organ pools. This helped to balance the contributions of different organ pools and prevent one pool from dominating the loss calculation. All models were trained on an NVidia Tesla V100 GPUs.

To evaluate the performance of all models, two metrics were used: the determination coefficient (R²) and the root mean square error (RMSE).

3 Results

3.1 Variations of winter wheat biophysical variables

The descriptive statistics reveal significant variations in the four biophysical variables across the different growth stages: biomass ranging from 0.51 to 27.89 T/ha, LAI from 0.69 to 8.66, nitrogen concentration from 0.61 to 4.76%, and nitrogen uptake from 13.49 to 338.59 kg N/ha (Table 2). This wide variability in the datasets was attributed to diverse factors, including variations in growing stages, repeated measurements over multiple years, and heterogeneous treatments, particularly variations in nitrogen inputs. The analysis, specifically employing ANOVA, indicates that most of these biophysical variables exhibit significant differences among treatments (Table S4). Both the training and validation datasets exhibit similar statistics, which validates the appropriateness of the dataset splitting method. Furthermore, correlations between these variables were examined, and the results show that biomass demonstrated a Pearson correlation of -0.27, -0.71, and 0.87 with LAI, nitrogen concentration, and nitrogen uptake, respectively. The correlation between nitrogen concentration and nitrogen uptake was found to be -0.44.

TABLE 2
www.frontiersin.org

Table 2 Descriptive statistics of dry matter (T/ha), LAI, N concentration (%) and N uptake (kg N/ha).

3.2 Plant biophysical variable modeling

This study evaluated various models for predicting plant biophysical variables. The EfficientNetB4 model trained on pseudo-labels demonstrated the highest performance for DM, achieving an R² of 0.92 and a low RMSE of 1.50 on the validation dataset (Table 3). In contrast, the PLSr model had an R² of 0.77 and a higher RMSE of 2.58, indicating weaker predictive ability.

TABLE 3
www.frontiersin.org

Table 3 Model performances for DM of the plant.

Regarding LAI, the ResNet50 model trained on pseudo-labels yielded the best R² of 0.82 (Table 4). For nitrogen concentration prediction using multispectral images, the ResNet50 model achieved an R² of 0.80 (Table 5) and an R² of 0.73 for Nitrogen uptake (Table 6).

TABLE 4
www.frontiersin.org

Table 4 Model performances for LAI.

TABLE 5
www.frontiersin.org

Table 5 Model performances for %N of the plant.

TABLE 6
www.frontiersin.org

Table 6 Model performances for Nupt of the plant.

The other CNN models investigated in this study exhibited robust and comparable performance levels when subjected to the pseudo-labeling pipeline during training. The utilization of pseudo-labels played a pivotal role in mitigating disparities between the outcomes observed on the validation and training datasets. It is worth noting that the PLSr model did not yield any discernible advantages from the pseudo-labeling technique, consistently falling short of the CNN models in terms of performance. One noteworthy observation is that this pseudo-labeling method appeared to exacerbate the disparities between the performance of the training and validation sets. Furthermore, the results of the backward feature selection analysis, as depicted in Figures S1 to S4, indicated that the augmentation of the dataset via this approach led to an increased requirement for features to achieve optimal performance levels.

Throughout the growing season, the models successfully assessed the variables, as evidenced by Figures 4 and 5. However, there were some outliers that significantly deviated from the ideal 1:1 relationship between predicted and true values. Additionally, a saturation effect was observed, where the models struggled to accurately predict the maximum values of each variable, leading to a lack of detail in certain growing seasons. These observations provide valuable insights for further refining the modeling approach and improving predictive accuracy.

FIGURE 4
www.frontiersin.org

Figure 4 Comparison between observed and predicted values of DM of the whole plant and LAI for both training and validation datasets, using the EfficientNetB4 model for DM and the ResNet50 model for LAI. The dots are color-coded according to the stages in the season, with darker dots indicating later stages. The dark line represents the 1:1 line.

FIGURE 5
www.frontiersin.org

Figure 5 Comparison between observed and predicted values of %N and Nupt of the whole plant for both training and validation datasets, using the ResNet50 model. The dots are color-coded according to the stages in the season, with darker dots indicating later stages. The dark line represents the 1:1 line.

3.3 Organs biophysical variable modeling

The utilization of multi-output models yielded diverse outcomes regarding the proportion of dry matter and nitrogen uptake, as indicated in Table S5. Table 7 displays the performance of the multiplication of both the single output models and the multi-output models for dry matter and nitrogen uptake, and solely the multi-output model for nitrogen concentration.

TABLE 7
www.frontiersin.org

Table 7 R² of the different multi-outputs models to predict nitrogen uptake, dry matter and nitrogen concentration of each organ.

Among the models evaluated, EfficientNetB0 demonstrated superior performance for predicting nitrogen uptake, achieving commendable R² values of 0.7, 0.59, 0.69, and 0.86 for stem, inferior leaves, flag leaf, and ear, respectively. ResNet50 exhibited R² values of 0.87, 0.62, 0.38, and 0.94 for dry matter, and 0.50, 0.76, 0.69, and -1.07 for nitrogen concentration, indicating its effectiveness in certain cases.

Analyzing individual organs, the ear and stems exhibited higher prediction accuracy by the models, while the flag leaf showed comparatively poorer prediction, depending on the specific model employed. Concerning the %N models, the stem and inferior leaf pools were accurately predicted, but the prediction performance for the ear was notably inadequate.

Interestingly, while the pseudo-labeling method led to reduced performance for the multi-output models (Table S5), its combination with the single output models, which significantly benefit from pseudo-labels, did not have a substantial impact on the prediction of DM and %N for each organ. This suggests that the pseudo-labeling approach is effective in enhancing the single output models but may require further optimization for multi-output models.

The Figure 6 presents the predicted partitioning of wheat dry matter and nitrogen uptake over the growing season for a single microplot. It offers a nice alternative to provide valuable insight about the partitioning of the matter within the plant. Moreover, both RGB and multispectral models successfully detected the emergence of new organs, such as the flag leaf and ear. Notably, the dry matter model showed an earlier appearance of ears compared to the nitrogen uptake model in this specific example.

FIGURE 6
www.frontiersin.org

Figure 6 Predicted partitioning of dry matter and nitrogen uptake throughout the season for a microplot from the 22-F trial. This results from the use of the multi-output ResNet50 (Table S5) multiplied by the single output ResNet50 (Table 3).

4 Discussion

4.1 Convolutional neural networks as an effective approach for predicting biophysical variables

This study presents a comprehensive investigation into the potential of recent CNNs in accurately predicting biophysical vegetation variables, such as dry matter, leaf area index, and nitrogen uptake and concentration. The research demonstrate that our CNN-based approach stands as one of the most advanced methods for this task, even though direct performance comparisons with prior studies are hindered by the limited availability of benchmark datasets.

In this study, CNN models outperformed a PLSr approach, consistent with previous research findings (Ma et al., 2019; Castro et al., 2020). Thus, CNN stands out as a potent tool in this context due to its ability to autonomously extract features, eliminating the need for manual feature extraction. It demonstrates remarkable adaptability in handling the evolving features of crops throughout the growing season, including changes in physiology and color. This adaptability negates the necessity for fine-tuning models to specific growth stages or cultivars, as highlighted by previous machine learning research (Yue et al., 2019). These studies commonly adopt a strategy of employing one model for each growth stage, alongside a single overarching model that typically yields less satisfactory results (Wang et al., 2022a). Nevertheless, it would remain intriguing to explore the performance of the presented CNN models on new cultivars, which may exhibit distinct characteristics.

Moreover, the results pertaining to nitrogen concentration are particularly intriguing. One might expect that CNNs would prioritize features related to plant architecture, which would be more closely associated with nitrogen uptake. However, the observed medium correlation (-0.44) between nitrogen concentration and uptake tends to limit this assumption, suggesting that the CNNs might have also identified some kind of vegetation indices contributing to the predictions. Despite these remarkable outcomes, interpreting the specific features extracted by CNNs remains challenging. To enhance our understanding of the underlying mechanisms and improve model interpretability, ongoing research is dedicated to developing techniques for explaining CNN predictions. One such approach, Grad-CAM (Selvaraju et al., 2020), shows promise in providing insights into the regions of the image that significantly influence the model’s decisions.

Among the CNN architectures explored, ResNet50 exhibited high performance, consistent with similar studies (Castro et al., 2020). Notably, EfficientNet also yielded promising results, especially for DM prediction of the entire plant. However, it is worth considering that the advantage of EfficientNetB4 might be attributed to its capacity to capture finer details in larger images. Interestingly, recent research has shown that performance gains may saturate beyond a certain image size (Li et al., 2021). This behavior could be dependent on the architecture, as EfficientNet is explicitly designed for scalable optimization on specific datasets (Tan and Le, 2020).

In contrast, the machine learning approach utilizing PLSr and feature fusion from the multi-sensor system consistently delivered inferior performance when compared to CNN models. Nevertheless, it is important to underscore that this method still achieved commendable results, boasting an R² value exceeding 0.6, which aligns with the findings reported in Yue et al. (2019).

An intriguing aspect arising from the backward feature selection analysis was the observed increase in the number of selected features between Ytrue and Ypseu, signifying the heightened demand for features in constructing models with a larger dataset (See Supplementary Material). Additionally, both sets of features exhibited substantial similarities, affirming their efficacy in modeling agronomical parameters. Among these features, plant height emerged as the most frequently utilized, followed by plant ratio and MCARI index. Furthermore, it is noteworthy that DM and Nuptake shared three out of four features, a logical outcome given that Nuptake was derived from DM. Other features selected for nitrogen-related analysis included well-established indices such as MCARI, mNDB, and GR.

However, it is essential to exercise caution when drawing overarching conclusions solely based on this method. Notably, the selection of these features can be intricate, as they may exhibit seasonal variations, as documented in (Yue et al., 2019; Wang et al., 2022a, b). It is conceivable that more advanced methods may yield superior results, as suggested in (Wang et al., 2021).

4.2 The significance of the amount of ground truth data in deep learning for regression of biophysical variables

Deep learning techniques, especially in regression tasks involving biophysical variables, encounter a substantial challenge due to the scarcity of sufficient training datasets. The limited availability of labeled data necessitates the development of innovative approaches to overcome this issue. A training pipeline was devised in this study, which capitalizes on the abundance of unlabeled data commonly found in highthroughput phenotyping installations. This method represents a practical approach to leverage unlabeled data, leading to optimized performance of CNN models in phenotyping applications.

The pseudo-labeling method emerged as an effective strategy to mitigate overfitting of the model. As a result, the performance gap between the training and validation datasets was reduced, signifying enhanced generalization. To perform such data correction, a polynomial cubic curve was chosen for its simplicity in representing biophysical curves and ease of fitting. Finer curves more related to plant growth pattern, such as P-splines or logistic curves, could have been used, but the fitting process may prove difficult (van Eeuwijk et al., 2019). These finer curves often require more frequent measurements (one to two per week) for accurate fitting (Roth et al., 2020), a frequency that our data did not meet. To address potential bias, correcting conditions were introduced, particularly essential for organ models. For example, when an organ was absent at a specific time (t), the corresponding pseudo-label was set to 0, a correction that, while seemingly straightforward, significantly contributed to the accuracy of representations. The effectiveness of traditional machine learning might also be a good option to generate pseudo-labels in the case of fewer ground truth data.

During the research, we also examined more advanced data augmentation techniques, such as 90° rotation and color space transformations without success. It is crucial to exercise caution when employing such methods, as their indiscriminate application may adversely affect model performance, as observed in certain models in (Castro et al., 2020). Conversely, (Ma et al., 2019) reported clear performance improvements with these methods. The discrepancy in results may be attributed to the risk of the model becoming overly reliant on specific features, such as wheat lines in the case of image rotation. Hence, prudent consideration of data augmentation is warranted based on the specific characteristics of the dataset and model.

4.3 Limitations and perspectives

An effective approach for evaluating model performance is to combine their predictions into a single other variable. In this study, we used DM and %N of the plant, predicted from their respective models, to calculate the Nitrogen Nutrition Index (NNI). The R² values for the training and validation datasets were 0.71 and 0.33, respectively, suggesting the potential utility of this method for measuring NNI as well. Although the dataset contains a substantial amount of heterogeneous ground truth data, the performance of the models may raise questions due to its limited size in terms of crop architecture and color, which only includes a few genotypes. The observed patterns in predicted values in Figures 5 and 4 appeared scattered, resembling a cloud rather than forming a clear line, and some outliers were evident, indicating room for improvement in the models. Overfitting was also observed, particularly with ResNet50, which frequently achieved R² values above 0.95 for the training dataset. To address this, a prudent approach would be to initially use small architectures and acquire more data. Despite the need for improvements concerning trait saturation and accuracy within specific growth stages Figure 4, the models’ potential is significant. They can be employed to compute advanced traits, such as growth rate and spot ideotypes using temporal curves, as demonstrated in a recent study (Roth et al., 2022).

By leveraging diverse and large-scale datasets, CNNs can yield more robust and precise models, reducing the need for heavily relying on study-specific feature engineering. Therefore, the phenotyping community should prioritize the development of extensive and well-annotated datasets for essential phenotyping challenges, such as the Global Wheat Head Detection (GWHD) dataset (David et al., 2021). Additionally, exploring alternative solutions, such as self-supervised learning (Zbontar et al., 2021) or generating synthetic data using Functional-Structural Plant Models (FSPM) (Gao et al., 2023), can further enhance model training and performance.

Research on the allocation of major plant elements, such as sink/source regulation processes and their relationship with grain nitrogen content, heavily relies on dry matter and nitrogen uptake partitioning (Martre et al., 2003; Gaju et al., 2014). The multi-output models proposed in this study have shown promising results (Table 7), with good performance in most cases. However, certain organs exhibited poor performance, such as %N of the ear, which may be attributed to the lack of visible traits that could account for it, like a greener ear. The subpar performance of DM and Nupt for flag leaf could be mainly attributed to the multi-output proportion model’s poor performance for this organ (Table S5), despite assigning it a higher weight in the loss function. Additional images specifically featuring flag leaves might be needed to improve its representation, as the ear rapidly develops behind them.

This multi-output model exemplifies the potential of such approaches for plant phenotyping. While this study employed a simple approach by sharing a common loss function, the benefits of multi-output learning can be substantial. For instance, a single model assessing both dry matter and leaf area index can significantly reduce computational costs and processing time, while maintaining high accuracy for both tasks. In fact, when tasks share complementary information, they can act as regularizers for each other, enhancing prediction performance (Standley et al., 2020). However, combining complex associations between tasks, such as classification and regression tasks, requires careful consideration of model architecture, loss function, and training strategy to achieve optimal performance. Ongoing research in this area is actively being pursued (Vafaeikia et al., 2020; Vandenhende et al., 2020).

The ability of the models to autonomously discover the appearance of new organs, such as ears and flag leaves, is particularly intriguing and opens exciting new research avenues too. This suggests the feasibility of developing growth stage estimation models per RGB image in a similar manner. Such models could be further utilized for various purposes, such as optimizing crop models (Yang et al., 2021).

5 Conclusions

In this study, a robust training pipeline that leverages unlabeled data through the innovative combination of pseudo-labeling and temporal relationship correction were developed and implemented. The results demonstrate the significant advantages of employing CNN models over a PLSr approach, as they achieve superior performance without the need for labor-intensive feature engineering. Notably, EfficientNetB4 was better in predicting above-ground biomass, while ResNet50 exhibited superior performances in predicting LAI, nitrogen uptake, and nitrogen concentration. Additionally, our exploration of multi-output models provided valuable insights into the distribution of dry matter and nitrogen uptake among different plant organs, enriching our understanding of plant biophysical characteristics.

While CNN models show great promise, it is evident that further investigation is required to fully unlock their potential. This research effectively demonstrates the capabilities of CNNs in predicting biophysical vegetation variables and offers valuable insights into addressing limitations and future perspectives in plant phenotyping. Moving forward, data sharing within the phenotyping community will be critical to optimize model performance. Access to large and diverse datasets, such as the Global Wheat Head Detection dataset, is indispensable for advancing phenotyping research and enhancing the performances of CNN models. By fostering data sharing and continued research efforts, CNNs can continue to revolutionize plant phenotyping and make profound contributions to agricultural and environmental sciences.

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Author contributions

AC and SD performed experiments and data collection; AC build the models, performed the statistical analysis, interpreted the results, and prepared the first draft. BD and BM contributed to the result interpretation and supervised the project. All authors revised the manuscript. All authors contributed to the article and approved the submitted version.

Funding

This research was funded by the Agriculture, Natural Resources and Environment Research Direction of the Public Service of Wallonia (Belgium), project D65-1412/S1 PHENWHEAT, and the National Fund of Belgium F.R.S-FNRS (FRIA grant).

Acknowledgments

Computational resources have been provided by the Consortium des Équipements de Calcul Intensif (CÉCI), funded by the Fonds de la Recherche Scientifique de Belgique (F.R.S.-FNRS) under Grant No. 2.5020.11 and by the Walloon Region. The authors thank the research and teaching support units Agriculture Is Life of TERRA Teaching and Research Centre, Liege` University for giving access to the field trials. The authors are grateful to Jerome Heens, Jesse Jap, Franc¸oise Thys and Gauthier Lepage for their help. The authors also thank CRA-W/Agromet.be for the meteorological data.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fpls.2023.1204791/full#supplementary-material

Supplementary Table 1 | Fertilization trial (F) in 2020, 2021 and 2022.

Supplementary Table 2 | Fertilization trial (F) in 2019.

Supplementary Table 3 | Trials composed of different fertilization fractioning combined with different fungicide application programs (FP) in 2020, 2021 and 2022.

Supplementary Table 4 | Dry matter, LAI, nitrogen concentration and nitrogen uptake of the whole plant used in this study. Data are means +/- the standard deviation (n=3 or 4 repetitions). Significance of treatment effects were analyzed using a one-way ANOVA (*: P ≤ 0.05; **: P ≤ 0.01; **: P ≤ 0.001).

Supplementary Table 5 | R² of the different models to predict DM and Nupt proportion of each organ.

Supplementary Table 6 | Pearson correlation table of the agronomic data

Supplementary Table 7 | Vegetation indices selected in this study.

Supplementary Figure 1 | Backward feature selection with PLSr for DM Estimation: On the left, selected features for estimating DM from Ytrue include Plant Ratio, 95th Percentile of Height, MCARI, and BRF 490. On the right, the selected features for estimating DM from Ypseu comprise Plant Ratio, 95th Percentile of Height, MCARI, and BRF 550.

Supplementary Figure 2 | Backward feature selection with PLSr for LAI Estimation: On the left, selected features fo estimating LAI from Ytrue include SR, GNDVI, MCARI, CIgree, BRF 900 and BRF 720. On the right, the selected features for estimating LAI from Ypseu comprise NDRE, SR, GNDVI, CIgreen, CIrede and BRF 550.

Supplementary Figure 3 | Backward feature selection with PLSr for Nuptake Estimation: On the left, selected features for estimating Nuptake from Ytrue include Plant Ratio, 95th percentile of height and MCARI. On the right, the selected features for estimating Nuptake from Ypseu comprise Plant Ratio, 95th percentile of height, GR, MCARI, mNDB and BRF 550.

Supplementary Figure 4 | Backward feature selection with PLSr for Nrate Estimation: On the left, selected features for estimating Nrate from Ytrue include 95th percentile of height and MCARI. On the right, the selected features for estimating Nrate from Ypseu comprise 95th percentile of height, GR and mNDB.

References

Araus, J. L., Buchaillot, M. L., Kefauver, S. C. (2022). “High Throughput Field Phenotyping,” in Wheat Improvement. Eds. Reynolds, M. P., Braun, H.-J. (Cham: Springer International Publishing), 495–512. doi: 10.1007/978-3-030-90673-327

CrossRef Full Text | Google Scholar

Arya, S., Sandhu, K. S., Singh, J., kumar, S. (2022). Deep learning: As the new frontier in high-throughput plant phenotyping. Euphytica 218, 47. doi: 10.1007/s10681-022-02992-3

CrossRef Full Text | Google Scholar

Berger, K., Verrelst, J., Féret, J.-B., Wang, Z., Wocher, M., Strathmann, M., et al. (2020). Crop nitrogen monitoring: Recent progress and principal developments in the context of imaging spectroscopy missions. Remote Sens. Environ. 242, 111758. doi: 10.1016/j.rse.2020.111758

PubMed Abstract | CrossRef Full Text | Google Scholar

Brocks, S., Bareth, G. (2018). Estimating barley biomass with crop surface models from oblique RGB imagery. Remote Sens. 10, 268. doi: 10.3390/rs10020268

CrossRef Full Text | Google Scholar

Buxbaum, N., Lieth, J. H., Earles, M. (2022). Non-destructive plant biomass monitoring with high spatio-temporal resolution via proximal RGB-D imagery and end-to-end deep learning. Front. Plant Sci. 13. doi: 10.3389/fpls.2022.758818

PubMed Abstract | CrossRef Full Text | Google Scholar

Carlier, A., Dandrifosse, S., Dumont, B., Mercatoris, B. (2022). Wheat ear segmentation based on a multisensor system and superpixel classification. Plant Phenomics 2022. doi: 10.34133/2022/9841985

PubMed Abstract | CrossRef Full Text | Google Scholar

Castro, W., Marcato Junior, J., Polidoro, C., Osco, L. P., Gonc¸alves, W., Rodrigues, L., et al. (2020). Deep learning applied to phenotyping of biomass in forages with UAV-based RGB imagery. Sensors 20, 4802. doi: 10.3390/s20174802

PubMed Abstract | CrossRef Full Text | Google Scholar

Chao, Z., Liu, N., Zhang, P., Ying, T., Song, K. (2019). Estimation methods developing with remote sensing information for energy crop biomass: a comparative review. Biomass Bioenergy 122, 414–425. doi: 10.1016/j.biombioe.2019.02.002

CrossRef Full Text | Google Scholar

Dandrifosse, S. (2022). Dynamics of wheat organs by close-range multimodal machine vision. Ph.D. thesis, ULiège. GxABT - Liège Université. Gembloux Agro-Bio Tech Gembloux Belgium.

Google Scholar

Dandrifosse, S., Bouvry, A., Leemans, V., Dumont, B., Mercatoris, B. (2020). Imaging wheat canopy through stereo vision: overcoming the challenges of the laboratory to field transition for morphological features extraction. Front. Plant Sci. 11. doi: 10.3389/fpls.2020.00096

PubMed Abstract | CrossRef Full Text | Google Scholar

Dandrifosse, S., Carlier, A., Dumont, B., Mercatoris, B. (2021). Registration and fusion of closeRange multimodal wheat images in field conditions. Remote Sens. 13, 1380. doi: 10.3390/rs13071380

CrossRef Full Text | Google Scholar

Dandrifosse, S., Carlier, A., Dumont, B., Mercatoris, B. (2022). In-field wheat reflectance: how to reach the organ scale? Sensors 22, 3342. doi: 10.3390/s22093342

PubMed Abstract | CrossRef Full Text | Google Scholar

David, E., Serouart, M., Smith, D., Madec, S., Velumani, K., Liu, S., et al. (2021). Global Wheat Head Detection 2021: an improved dataset for benchmarking wheat head detection methods. Plant Phenomics 2021. doi: 10.34133/2021/9846158

PubMed Abstract | CrossRef Full Text | Google Scholar

Deery, D., Jimenez-Berni, J., Jones, H., Sirault, X., Furbank, R. (2014). Proximal remote sensing buggies and potential applications for field-based phenotyping. Agronomy 4, 349–379. doi: 10.3390/agronomy4030349

CrossRef Full Text | Google Scholar

Deng, J., Dong, W., Socher, R., Li, L.-J., Li, K., Fei-Fei, L. (2009). “ImageNet: A large-scale hierarchical image database,” in 2009 IEEE Conference on Computer Vision and Pattern Recognition. IEEE. 248–255. doi: 10.1109/CVPR.2009.5206848

CrossRef Full Text | Google Scholar

de Oliveira, G. S., Marcato Junior, J., Polidoro, C., Osco, L. P., Siqueira, H., Rodrigues, L., et al. (2021). Convolutional neural networks to estimate dry matter yield in a Guineagrass breeding program using UAV remote sensing. Sensors 21, 3971. doi: 10.3390/s21123971

PubMed Abstract | CrossRef Full Text | Google Scholar

Duchene, O., Dumont, B., Cattani, D. J., Fagnant, L., Schlautman, B., DeHaan, L. R., et al. (2021). Processbased analysis of Thinopyrum intermedium phenological development highlights the importance of dual induction for reproductive growth and agronomic performance. Agric. For. Meteorol, 301–302. doi: 10.1016/j.agrformet.2021.108341

CrossRef Full Text | Google Scholar

Freitas Moreira, F., Rojas de Oliveira, H., Lopez, M. A., Abughali, B. J., Gomes, G., Cherkauer, K. A., et al. (2021). High-throughput phenotyping and random regression models reveal temporal genetic control of soybean biomass production. Front. Plant Sci. 12. doi: 10.3389/fpls.2021.715983

CrossRef Full Text | Google Scholar

Gaju, O., Allard, V., Martre, P., Le Gouis, J., Moreau, D., Bogard, M., et al. (2014). Nitrogen partitioning and remobilization in relation to leaf senescence, grain yield and grain nitrogen concentration in wheat cultivars. Field Crops Res. 155, 213–223. doi: 10.1016/j.fcr.2013.09.003

CrossRef Full Text | Google Scholar

Gao, Y., Li, Y., Jiang, R., Zhan, X., Lu, H., Guo, W., et al. (2023). Enhancing green fraction estimation in rice and wheat crops: A self-supervised deep learning semantic segmentation approach. Plant Phenomics 5, 64. doi: 10.34133/plantphenomics.0064

CrossRef Full Text | Google Scholar

Hawkesford, M., Riche, A. (2020). Impacts of G x E x M on nitrogen use efficiency in wheat and future prospects. Front. Plant Sci. 11. doi: 10.3389/fpls.2020.01157

CrossRef Full Text | Google Scholar

He, K., Zhang, X., Ren, S., Sun, J. (2015). Deep Residual Learning for Image Recognition. doi: 10.48550/arXiv.1512.03385

CrossRef Full Text | Google Scholar

Hickey, L. T., Hafeez, A. N., Robinson, H., Jackson, S. A., Leal-Bertioli, S. C. M., et al. (2019). Breeding crops to feed 10 billion. Nat. Biotechnol. 37, 744–754. doi: 10.1038/s41587-019-0152-9

PubMed Abstract | CrossRef Full Text | Google Scholar

Jiang, Y., Li, C. (2020). Convolutional neural networks for image-based high-throughput plant phenotyping: a review. Plant Phenomics, 1–22. doi: 10.34133/2020/4152816

CrossRef Full Text | Google Scholar

Justes, E. (1994). Determination of a critical nitrogen dilution curve for winter wheat crops. Ann. Bot. 74, 397–407. doi: 10.1006/anbo.1994.1133

CrossRef Full Text | Google Scholar

Kamilaris, A., Prenafeta-Boldú, F. X. (2018). Deep learning in agriculture: A survey. Comput. Electron. Agric. 147, 70–90. doi: 10.1016/j.compag.2018.02.016

CrossRef Full Text | Google Scholar

Kattenborn, T., Leitloff, J., Schiefer, F., Hinz, S. (2021). Review on Convolutional Neural Networks (CNN) in vegetation remote sensing. ISPRS J. Photogrammetry Remote Sens. 173, 24–49. doi: 10.1016/j.isprsjprs.2020.12.010

CrossRef Full Text | Google Scholar

Lee, D.-H. (2013). Pseudo-label : the simple and efficient semi-supervised learning method for deep neural networks. ICML 2013 Workshop Challenges Representation Learn. (WREPL).

Google Scholar

Lemaire, G., Ciampitti, I. (2020). Crop mass and N status as prerequisite covariables for unraveling nitrogen use efficiency across genotype-by-environment-by-management scenarios: a review. Plants 9, 1309. doi: 10.3390/plants9101309

PubMed Abstract | CrossRef Full Text | Google Scholar

Li, Y., Liu, H., Ma, J., Zhang, L. (2021). Estimation of leaf area index for winter wheat at early stages based on convolutional neural networks. Comput. Electron. Agric. 190, 106480. doi: 10.1016/j.compag.2021.106480

CrossRef Full Text | Google Scholar

Ma, J., Li, Y., Chen, Y., Du, K., Zheng, F., Zhang, L., et al. (2019). Estimating above ground biomass of winter wheat at early growth stages using digital images and deep convolutional neural network. Eur. J. Agron. 103, 117–129. doi: 10.1016/j.eja.2018.12.004

CrossRef Full Text | Google Scholar

Martre, P., Porter, J. R., Jamieson, P. D., Tribo¨ı, E. (2003). Modeling grain nitrogen accumulation and protein composition to understand the sink/source regulations of nitrogen remobilization for wheat. Plant Physiol. 133, 1959–1967. doi: 10.1104/pp.103.030585

PubMed Abstract | CrossRef Full Text | Google Scholar

Nabwire, S., Suh, H.-K., Kim, M. S., Baek, I., Cho, B.-K. (2021). Review: application of artificial intelligence in phenomics. Sensors 21, 4363. doi: 10.3390/s21134363

PubMed Abstract | CrossRef Full Text | Google Scholar

Nguyen, C., Sagan, V., Bhadra, S., Moose, S. (2023). UAV multisensory data fusion and multi-task deep learning for high-throughput maize phenotyping. Sensors 23, 1827. doi: 10.3390/s23041827

PubMed Abstract | CrossRef Full Text | Google Scholar

Otsu, N. (1979). A threshold selection method from gray-level histograms. IEEE Trans. Systems Man Cybernetics 9, 62–66. doi: 10.1109/TSMC.1979.4310076

CrossRef Full Text | Google Scholar

Pedregosa, F., Varoquaux, G., Gramfort, A., Michel, V., Thirion, B., Grisel, O., et al. (2011). Scikit-learn: machine learning in python. J. Mach. Learn. Res. 12, 2825–2830. doi: 10.48550/arXiv.1201.0490

CrossRef Full Text | Google Scholar

Pound, M. P., Atkinson, J. A., Wells, D. M., Pridmore, T. P., French, A. P. (2017). “Deep learning for multi-task plant phenotyping,” in 2017 IEEE International Conference on Computer Vision Workshops (ICCVW), 2055–2063. doi: 10.1109/ICCVW.2017.241

CrossRef Full Text | Google Scholar

Raj, R., Walker, J. P., Pingale, R., Nandan, R., Naik, B., Jagarlapudi, A. (2021). Leaf area index estimation using top-of-canopy airborne RGB images. Int. J. Appl. Earth Observation Geoinformation 96, 102282. doi: 10.1016/j.jag.2020.102282

CrossRef Full Text | Google Scholar

Reynolds, M., Chapman, S., Crespo-Herrera, L., Molero, G., Mondal, S., Pequeno, D. N., et al. (2020). Breeder friendly phenotyping. Plant Sci. 295, 110396. doi: 10.1016/j.plantsci.2019.110396

PubMed Abstract | CrossRef Full Text | Google Scholar

Roth, L., Barendregt, C., Bétrix, C.-A., Hund, A., Walter, A. (2022). High-throughput field phenotyping of soybean: spotting an ideotype. Remote Sens. Environ. 269, 112797. doi: 10.1016/j.rse.2021

CrossRef Full Text | Google Scholar

Roth, L., Camenzind, M., Aasen, H., Kronenberg, L., Barendregt, C., Camp, K.-H., et al. (2020). Repeated multiview imaging for estimating seedling tiller counts of wheat genotypes using drones. Plant Phenomics 2020. doi: 10.34133/2020/3729715

PubMed Abstract | CrossRef Full Text | Google Scholar

Sapkota, B., Popescu, S., Rajan, N., Leon, R., Reberg-Horton, C., Mirsky, S., et al. (2022). Use of synthetic images for training a deep learning model for weed detection and biomass estimation in cotton. Sci. Rep. 12. doi: 10.1038/s41598-022-23399-z

PubMed Abstract | CrossRef Full Text | Google Scholar

Schiefer, F., Schmidtlein, S., Kattenborn, T. (2021). The retrieval of plant functional traits from canopy spectra through RTM-inversions and statistical models are both critically affected by plant phenology. Ecol. Indic. 121. doi: 10.1016/j.ecolind.2020.107062

CrossRef Full Text | Google Scholar

Schreiber, L., Atkinson Amorim, J., Guimãraes, L., Motta Matos, D., Maciel da Costa, C., Parraga, A. (2022). Above-ground biomass wheat estimation: deep learning with UAV-based RGB images. Appl. Artif. Intell. doi: 10.1080/08839514.2022.2055392

CrossRef Full Text | Google Scholar

Selvaraju, R. R., Cogswell, M., Das, A., Vedantam, R., Parikh, D., Batra, D. (2020). Grad-CAM: visual explanations from deep networks via gradient-based localization. Int. J. Comput. Vision 128, 336–359. doi: 10.1007/s11263-019-01228-7

CrossRef Full Text | Google Scholar

Serouart, M., Madec, S., David, E., Velumani, K., Lopez Lozano, R., Weiss, M., et al. (2022). SegVeg: segmenting RGB images into green and senescent vegetation by combining deep and shallow methods. Plant Phenomics 2022. doi: 10.34133/2022/9803570

PubMed Abstract | CrossRef Full Text | Google Scholar

Singh, A. K., Ganapathysubramanian, B., Sarkar, S., Singh, A. (2018). Deep learning for plant stress phenotyping: trends and future perspectives. Trends Plant Sci. 23, 883–898. doi: 10.1016/j.tplants.2018.07.004

PubMed Abstract | CrossRef Full Text | Google Scholar

Song, X., Yang, G., Xu, X., Zhang, D., Yang, C., Feng, H. (2022). Winter wheat nitrogen estimation based on ground-level and UAV-mounted sensors. Sensors 22. doi: 10.3390/s22020549

CrossRef Full Text | Google Scholar

Standley, T., Zamir, A. R., Chen, D., Guibas, L., Malik, J., Savarese, S. (2020). Which tasks should be learned together in multi-task learning? doi: 10.48550/arXiv.1905.07553

CrossRef Full Text | Google Scholar

Sun, D., Robbins, K., Morales, N., Shu, Q., Cen, H. (2022). Advances in optical phenotyping of cereal crops. Trends Plant Sci. 27, 191–208. doi: 10.1016/j.tplants.2021.07.015

PubMed Abstract | CrossRef Full Text | Google Scholar

Tan, M., Le, Q. V. (2020). EfficientNet: rethinking model scaling for convolutional neural networks. doi: 10.48550/arXiv.1905.11946

CrossRef Full Text | Google Scholar

Tanner, F., Tonn, S., de Wit, J., Van den Ackerveken, G., Berger, B., Plett, D. (2022). Sensorbased phenotyping of above-ground plant-pathogen interactions. Plant Methods 18, 35. doi: 10.1186/s13007-022-00853-7

PubMed Abstract | CrossRef Full Text | Google Scholar

Tilly, N., Aasen, H., Bareth, G. (2015). Fusion of plant height and vegetation indices for the estimation of barley biomass. Remote Sens. 7, 11449–11480. doi: 10.3390/rs70911449

CrossRef Full Text | Google Scholar

Vafaeikia, P., Namdar, K., Khalvati, F. (2020). A brief review of deep multi-task learning and auxiliary task learning. doi: 10.48550/arXiv.2007.01126

CrossRef Full Text | Google Scholar

Vandenhende, S., Georgoulis, S., Proesmans, M., Dai, D., Gool, L. (2020). Revisiting multi-task learning in the deep learning era. ArXiv. doi: 10.1109/TPAMI.2021.3054719

CrossRef Full Text | Google Scholar

van Eeuwijk, F. A., Bustos-Korts, D., Millet, E. J., Boer, M. P., Kruijer, W., Thompson, A., et al. (2019). Modelling strategies for assessing and increasing the effectiveness of new phenotyping techniques in plant breeding. Plant Sci. 282, 23–39. doi: 10.1016/j.plantsci.2018.06.018

PubMed Abstract | CrossRef Full Text | Google Scholar

Verrelst, J., Malenovsky,´, Z., van der Tol, C., Camps-Valls, G., Gastellu-Etchegorry, J.-P., Lewis, P., et al. (2019). Quantifying vegetation biophysical variables from imaging spectroscopy data: a review on retrieval methods. Surveys Geophysics 40, 589–629. doi: 10.1007/s10712-018-9478-y

CrossRef Full Text | Google Scholar

Wan, L., Zhang, J., Dong, X., Du, X., Zhu, J., Sun, D., et al. (2021). Unmanned aerial vehicle-based field phenotyping of crop biomass using growth traits retrieved from PROSAIL model. Comput. Electron. Agric. 187, 106304. doi: 10.1016/j.compag.2021.106304

CrossRef Full Text | Google Scholar

Wang, Z., Lu, Y., Zhao, G., Sun, C., Zhang, F., He, S. (2022b). Sugarcane biomass prediction with multi-mode remote sensing data using deep archetypal analysis and integrated learning. Remote Sens. 14, 4944. doi: 10.3390/rs14194944

CrossRef Full Text | Google Scholar

Wang, W., Wu, Y., Zhang, Q., Zheng, H., Yao, X., Zhu, Y., et al. (2021). AAVI: A novel approach to estimating leaf nitrogen concentration in rice from unmanned aerial vehicle multispectral imagery at early and middle growth stages. IEEE J. Selected Topics Appl. Earth Observations Remote Sens. 14, 6716–6728. doi: 10.1109/JSTARS.2021.3086580

CrossRef Full Text | Google Scholar

Wang, F., Yang, M., Ma, L., Zhang, T., Qin, W., Li, W., et al. (2022a). Estimation of above-ground biomass of winter wheat based on consumer-grade multi-spectral UAV. Remote Sens. 14, 1251. doi: 10.3390/rs14051251

CrossRef Full Text | Google Scholar

Xu, R., Li, C. (2022). A review of high-throughput field phenotyping systems: focusing on ground robots. Plant Phenomics 2022. doi: 10.34133/2022/9760269

PubMed Abstract | CrossRef Full Text | Google Scholar

Yang, K.-W., Chapman, S., Carpenter, N., Hammer, G., McLean, G., Zheng, B., et al. (2021). Integrating crop growth models with remote sensing for predicting biomass yield of sorghum. silico Plants 3, diab001. doi: 10.1093/insilicoplants/diab001

CrossRef Full Text | Google Scholar

Yue, J., Yang, G., Tian, Q., Feng, H., Xu, K., Zhou, C. (2019). Estimate of winter-wheat above-ground biomass based on UAV ultrahigh-ground-resolution image textures and vegetation indices. ISPRS J. Photogrammetry Remote Sens. 150, 226–244. doi: 10.1016/j.isprsjprs.2019.02.022

CrossRef Full Text | Google Scholar

Zbontar, J., Jing, L., Misra, I., LeCun, Y., Deny, S. (2021). Barlow twins: self-supervised learning via redundancy reduction. doi: 10.48550/arXiv.2103.03230

CrossRef Full Text | Google Scholar

Zhang, J., Tian, H., Wang, P., Tansey, K., Zhang, S., Li, H. (2022). Improving wheat yield estimates using data augmentation models and remotely sensed biophysical indices within deep neural networks in the Guanzhong Plain, PR China. Comput. Electron. Agric. 192, 106616. doi: 10.1016/j.compag.2021.106616

CrossRef Full Text | Google Scholar

Zheng, C., Abd-Elrahman, A., Whitaker, V. M., Dalid, C. (2022). Deep learning for strawberry canopy delineation and biomass prediction from high-resolution images. Plant Phenomics 2022. doi: 10.34133/2022/9850486

CrossRef Full Text | Google Scholar

Keywords: phenotyping, close-range sensing, wheat, CNN, biophysical variables, multi-task, PLSr

Citation: Carlier A, Dandrifosse S, Dumont B and Mercatoris B (2023) Comparing CNNs and PLSr for estimating wheat organs biophysical variables using proximal sensing. Front. Plant Sci. 14:1204791. doi: 10.3389/fpls.2023.1204791

Received: 14 April 2023; Accepted: 30 October 2023;
Published: 20 November 2023.

Edited by:

Jennifer Clarke, University of Nebraska-Lincoln, United States

Reviewed by:

Kang Yu, Technical University of Munich, Germany
Qingfeng Song, Chinese Academy of Sciences (CAS), China
Nisha Pillai, Mississippi State University, United States

Copyright © 2023 Carlier, Dandrifosse, Dumont and Mercatoris. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Benoit Mercatoris, YmVub2l0Lm1lcmNhdG9yaXNAdWxpZWdlLmJl

ORCID: Benjamin Dumont, orcid.org/0000-0001-8411-3990

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.