- 1Facultad de Telemática, Universidad de Colima, Colima, Mexico
- 2International Maize and Wheat Improvement Center (CIMMYT), Texcoco, Edo. de Mexico, Mexico
- 3Facultad de Ciencias, Universidad Nacioanl Autónoma de México (UNAM), México City, Mexico
- 4Independent consultant, Zinacatepec, Puebla, Mexico
- 5Centro de Investigación en Computación (CIC), Instituto Politécnico Nacional (IPN), México City, Mexico
- 6Centro Universitario de Ciencias Exactas e Ingenierías (CUCEI), Universidad de Guadalajara, Guadalajara, Jalisco, Mexico
- 7Department of Statistics and Operations Research, King Saud University, Riyah, Saudi Arabia
- 8Louisiana State University, Baton Rouge, LA, United States
- 9Distinguished Scientist Fellowship Program, King Saud University, Riyah, Saudi Arabia
- 10Instituto de Socieconomia, Estadistica e Informatica, Colegio de Postgraduados, Montecillos, Edo. de México, Texcoco, Mexico
Introduction: Because Genomic selection (GS) is a predictive methodology, it needs to guarantee high-prediction accuracies for practical implementations. However, since many factors affect the prediction performance of this methodology, its practical implementation still needs to be improved in many breeding programs. For this reason, many strategies have been explored to improve the prediction performance of this methodology.
Methods: When environmental covariates are incorporated as inputs in the genomic prediction models, this information only sometimes helps increase prediction performance. For this reason, this investigation explores the use of feature engineering on the environmental covariates to enhance the prediction performance of genomic prediction models.
Results and discussion: We found that across data sets, feature engineering helps reduce prediction error regarding only the inclusion of the environmental covariates without feature engineering by 761.625% across predictors. These results are very promising regarding the potential of feature engineering to enhance prediction accuracy. However, since a significant gain in prediction accuracy was observed in only some data sets, further research is required to guarantee a robust feature engineering strategy to incorporate the environmental covariates.
Introduction
The global population’s rapid growth is increasing food demand, but climate change impacts crop productivity. Plant breeding is essential for high-yield, quality cultivars. Wheat production soared from 200 million tons in 1961 to 775 million tons in 2023 without expanding cultivation, thanks to improved cultivars and agricultural practices (FAO, 2023). Traditional methods used pedigree and observable traits, but DNA sequencing introduced genomic insights. Genomic selection (GS) relies on DNA markers, offering advantages over traditional methods (Crossa et al., 2017).
Numerous studies have investigated the efficacy of GS compared to traditional phenotypic selection across various crops and livestock. Butoto et al. (2022) observed that both GS and phenotypic selection were equally effective in enhancing resistance to Fusarium ear rot and reducing feminizing contamination in maize. Similarly, Sallam and Smith (2016) demonstrated that integrating GS into barley breeding programs targeting yield and Fusarium head blight (FHB) resistance yielded comparable gains in selection response to traditional phenotypic methods. Moreover, GS offered the added benefits of shorter breeding cycles and reduced costs. In contrast, research in maize breeding conducted by Beyene et al. (2015) and Gesteiro et al. (2023) revealed that GS outperformed phenotypic selection, resulting in superior genetic gains. These comparative findings underscore the considerable advantages of GS in optimizing breeding outcomes across diverse agricultural settings.
GS revolutionizes plant and animal breeding by leveraging high-density markers across the genome. It operates on the principle that at least one genetic marker is in linkage disequilibrium with a causative QTL (Quantitative Trait Locus) for the desired trait (Meuwissen et al., 2001). This method transforms breeding in several ways: a) Identifying promising genotypes before planting; b) Improving precision in selecting superior individuals; c) Saving resources by reducing extensive phenotyping; d) Accelerating variety development by shortening breeding cycles; e) Intensifying selection efforts; f) Facilitating the selection of traits difficult to measure; g) Enhancing the accuracy of the selection process (Bernardo and Yu, 2007; Heffner et al., 2009; Desta and Ortiz, 2014; Abed et al., 2018; Budhlakoti et al., 2022).
The GS methodology, embraced widely, expedites genetic improvements in plant breeding programs (Desta and Ortiz, 2014; Bassi et al., 2016; Xu et al., 2020). Utilizing advanced statistical and machine learning models (Montesinos-López et al., 2022), GS efficiently selects individuals within breeding populations. Deep learning, a subset of machine learning, has also shown promise in GS (Montesinos-López et al., 2021; Wang et al., 2023). This selection process relies on data from a training population, encompassing both phenotypic and genotypic information (Crossa et al., 2017).
The Deep Neural Network Genomic Prediction (DNNGP) method of Wang et al. (2023) represents a novel advanced on deep-learning genomic predictive approach. The authors compared the DNNGP with other genomic prediction methods for various traits using genotypic and transcriptomics on maize data. They demonstrated that DNNGP outperformed GBLUP in most datasets. For instance, for maize days to anthesis (DTA) trait, DNNGP showed superiority over GBLUP by 619.840% and 16.420% using gene expression and Single Nucleotide Polymorphism (SNP) data, respectively. When utilizing genotypic data, DNNGP achieved a prediction accuracy of 0.720 for DTA, while GBLUP reached 0.580. However, the study found varied patterns in prediction accuracy for other traits.
Following rigorous training, these models utilize genotypic data to predict breeding or phenotypic values for traits within a target population (Budhlakoti et al., 2022). The GS methodology is versatile, accommodating various scenarios including multi-trait considerations (Calus and Veerkamp, 2011), known major genes and marker-trait associations, Genotype × Environment interaction (GE) (Crossa et al., 2017), and integration of other omics data (Hu et al., 2021; Wu et al., 2022) such as transcriptomics, metabolomics, and proteomics. GE influences phenotypic trait values across diverse environments, underscoring its importance in association and prediction models. Jarquin et al. (2014) introduced a framework significantly improving prediction accuracy in the presence of GE, yet without considering environmental covariates. To enhance accuracy further, recent studies are integrating environmental information into genomic prediction models.
Jarquin et al. (2014) framework lacks consideration of environmental covariates, prompting recent studies to integrate such information to enhance prediction accuracy. For instance, Montesinos-López et al. (2023) and Costa-Neto et al. (2021a, 2021b) demonstrated significant improvements. Conversely, studies by Monteverde et al. (2019); Jarquin et al. (2020), and Rogers et al. (2021) showed modest or negligible enhancements, revealing the ongoing challenge of effectively integrating environmental data into genomic prediction models.
Achieving high prediction accuracy in GS faces significant challenges due to genetic complexities, environmental variations, and data constraints (Juliana et al., 2018). Complex traits involve multiple gene influences, while environmental conditions can alter trait expression (Desta and Ortiz, 2014; Crossa et al., 2017). Phenotyping and marker data quality are critical, and issues like overfitting and population structure can compromise prediction precision (Budhlakoti et al., 2022). Ongoing research focuses on improving models, increasing marker density, and enhancing data quality to refine genomic prediction accuracy (Crossa et al., 2017; Budhlakoti et al., 2022).
Ongoing efforts focus on refining GS accuracy through various optimizations. This includes fine-tuning training and testing sets for improved precision (Rincent et al., 2012; Akdemir et al., 2015). Researchers are also evaluating diverse statistical machine learning methods to develop robust models with minimal fine-tuning yet high accuracy (Montesinos-López et al., 2022). Moreover, integrating additional omics data, such as phenomics and transcriptomics, aims to bolster GS accuracy and identify potent predictors for target traits (Montesinos-López et al., 2017; Krause et al., 2019; Monteverde et al., 2019; Hu et al., 2021; Costa-Neto et al., 2021a, b; Rogers and Holland, 2022; Wu et al., 2022). These endeavors seek to enhance GS predictive capabilities by leveraging diverse information sources.
Feature engineering (FE) is crucial in improving machine learning model performance by selecting, modifying, or creating new features from raw data. It transforms input data into a more representative and informative format, capturing relevant patterns and relationships, and enhancing the model’s generalization ability. FE involves various tasks like selecting optimal features, generating new features, normalization/scaling, handling missing values, and encoding categorical variables. For instance, techniques like Principal Component Analysis (PCA) can transform correlated features into uncorrelated ones (Lam et al., 2017; Dong and Liu, 2018; Khurana et al., 2018). FE’s popularity is rising due to its ability to enhance model performance, extract meaningful information from complex data, improve interpretability, and boost efficiency. Successful implementations include sentiment analysis, image recognition, and predictive maintenance, showcasing FE’s effectiveness across domains (Nargesian et al., 2017; Carrillo-de-Albornoz et al., 2018; Yurek and Birant, 2019). In genomic prediction, FE has also been successful, as demonstrated by Bermingham et al. (2015) and Afshar and Usefi (2020). These examples underscore FE critical role in various domains, leading to more accurate machine learning applications (Dong and Liu, 2018).
The impact of feature engineering (FE) on reducing prediction error varies depending on the dataset, problem, and quality of FE. Well-crafted features can notably minimize prediction error in some cases, but the exact improvement is context-specific and not guaranteed. Effective FE can enhance model performance significantly, albeit its extent varies case by case (Heaton, 2016; Dong and Liu, 2018).
To optimize genomic selection’s predictive accuracy, it’s vital to adopt innovative methodologies that account for its multifaceted influences. FE in genomic prediction offers a promising approach by enhancing prediction quality, uncovering genetic insights, customizing models to specific needs, improving interpretability, and minimizing data noise. In this paper, we investigate FE applied to environmental covariates to assess its potential in enhancing prediction performance within the context of genomic selection.
Materials and methods
Dataset USP
The University of São Paulo (USP) Maize, Zea mays L., dataset is sourced from germplasm developed by the Luiz de Queiroz College of Agriculture at the University of São Paulo, Brazil. An experiment was conducted between 2016 and 2017 involving 49 inbred lines, yielding a total of 906 F1 hybrids, of which 570 were assessed across eight diverse environments for grain yield (GY). These environments were created by combining two locations, two years, and two nitrogen levels. However, we specifically used data from four distinct environments for this research, each containing 100 hybrids. It’s important to note that these environments had varying soil types and climatic conditions, and the study integrated data from 248 covariates related to these environmental factors. The parent lines underwent genotyping through the Affymetrix Axiom Maize Genotyping Array, resulting in a dataset of 54,113 high-quality SNPs after applying stringent quality control procedures. Please refer to Costa-Neto et al. (2021a) for further comprehensive information on this dataset.
Dataset Japonica
The Japonica dataset comprises 320 rice (Oryza sativa L.) genotypes drawn from the Japonica tropical rice population. This dataset underwent evaluations for the same four traits (GY, PHR: percentage of head rice, GC: percentage of chalky grains, PH: plant height) as the Indica population, but in this case, it was conducted across five distinct environments spanning from 2009 to 2013. Covariates were meticulously measured three times a year, covering three developmental stages (maturation, reproductive, and vegetative). This dataset comprises a non-balanced set of 1,051 assessments recorded across these five diverse environments. Additionally, each genotype within this dataset was meticulously evaluated for 16,383 SNP markers that remained after rigorous quality control procedures, with each marker being represented as 0, 1, or 2. For more comprehensive information on this dataset, please refer to Monteverde et al. (2019).
Dataset G2F
These three distinct datasets correspond to the Maize Crop, Zea mays L., for years 2014 (G2F_2014), 2015 (G2F_2015), and 2016 (G2F_2016) from the Genomes to Fields maize project (Lawrence-Dill, 2017), as outlined by Rogers and Holland (2022). These datasets collectively encompass a wealth of phenotypic, genotypic, and environmental information. To narrow the focus, our analysis primarily includes four specific traits: Grain_Moisture_BLUE (GM_BLUE), Grain_Moisture_weight (GM_Weight), Yield_Mg_ha_BLUE (YM_BLUE), and Yield_Mg_ha_weight (YM_Weight), carefully selected from a larger pool of traits detailed by Rogers and Holland (2022). Across these three years, the study involves 18, 12, and 18 distinct environments for the years 2014 (G2F_2014), 2015 (G2F_2015) and 2016 (G2F_2016), respectively. Regarding genotype numbers, the dataset for 2014 consisted of 781 genotypes, the dataset for 2015 featured 1,011 genotypes, and the dataset for 2016 comprised 456 genotypes. The analysis relies on 20,373 SNP markers that have already undergone imputation and filtering, following the methodology outlined by Rogers et al. (2021) and Rogers and Holland (2022). Additive allele calls are documented as minor allele counts, represented as 0, 1, or 2. For more detailed insights into these datasets, we recommend consulting the comprehensive description provided in Lawrence-Dill (2017) and Rogers and Holland (2022).
It is worth noting that each data set presents unique sets of environments. However, concerning traits, the G2F_2014, G2F_2015, and G2F_2016 datasets share identical traits, as do the Japonica dataset.
Statistical models
The four predictors under a genomic best linear unbiased predictor (GBLUP; Habier et al., 2007; VanRaden, 2008) model are described below.
Predictor P1: E+G
This predictor is represented as
where denotes the response variable in environment i and genotype j. denotes the population mean; are the random effects of environments, , denotes the random effects of lines, and denotes the random error components in the model assumed to be independent normal random variables with mean 0 and variance . In the context of this predictor E+G, X, denotes the matrix of markers and the matrix of centered and standardized markers. Then (VanRaden, 2008), where is the number of markers. is the design matrix of genotypes (lines) of order is the genomic relationship-matrix computed using markers (VanRaden, 2008). Therefore, the random effect of lines is distributed as . This model (1) was implemented in the BGLR library of Pérez and de los Campos (2014). Therefore, the linear kernel matrix for the genotype effect was determined by calculating the “covariance” structure of the genotype predictor (g) as .
On the other hand, the linear kernel matrix for the Environment effect was computed using three different techniques: not using environmental covariates (NoEC), with environmental covariates (EC), and with environmental covariates with FE.
∘ NoEC: Under this NoEC technique, the resulting linear kernel of environments was computed as , where denotes the number of environments and the design matrix of environments with zeros and ones, with ones in positions of specific environments.
∘ EC: The EC technique involved selecting and scaling the environmental covariates (EC) that exhibited a relevant Pearson´s correlation with the response variable. Covariates are selected if their Pearson’s correlation with the response variable exceeds 0.5 in each training set per trait. Notably, covariate selection excludes response variables in the testing set, representing the environment to predict. Covariates meeting a correlation of at least 0.5 are used; otherwise, lower thresholds like 0.3 or 0.4 are considered. Correlations below these values indicate training without environmental covariates.
∘ The resulting set of selected EC’s was then used to compute an environmental linear kernel, denoted as of order . After using this kernel, the expanded environmental kernel was computed as , which was used in the Bayesian model. The scaling of each environmental covariate was done by subtracting its respective mean and dividing by its corresponding standard deviation.
∘ FE: The Feature Engineering (FE) technique involved computing various mathematical transformations between all possible pairs of ECs, including addition, difference, product, and ratio, as well as other commonly used transformations such as inverses, square powers, root squares, logarithms, and some Box-Cox transformations for each EC. These transformations were used to generate new variables through FE. The transformation of addition, difference, product and ratio were implemented for each pair of environmental covariates, that is, there were built a total the n_cov choose two new covariates, with n_cov denoting the number of environmental covariates in each data set. While with transformations such as inverses (), square powers (, root squares (, natural logarithms [], and Box-Cox transformations for each environmental covariate was created only one new environmental covariate. Then the original and new environmental covariates were concatenated in a matrix and then were submitted to the selection process explained above. Then under the FE approach these resulting covariates are used to compute the new environmental kernel matrix .
Predictor P2: E+G+GE
The E+G+GE predictor is similar to P1 (Equation 1) but also accounts for the differential response of cultivars in environments, that is GE. This is achieved by taking the product of the kernel matrices of the genotype (G) and environment (E) predictors, that is, they were computed as (for NoEC), (for EC) or (for FE), which serves as the kernel matrix for the GE. In general, adding the GE interaction to the statistical machine learning model increases the genomic prediction accuracy (Jarquin et al., 2014; Crossa et al., 2017). Also, it is important to point out that under this predictor (P2) variance components and heritability of each trait in each data set were obtained under a Bayesian framework using the complete data set (i.e., no missing values allowed). For this computation all the terms were entered as random effects into the model but without taking into account the environmental covariates.
Predictor P3: E+G+BRR
The E+G+BRR predictor is similar to P1 (Equation 1), but incorporating the ECs as fixed effects in a Bayesian Ridge Regression (BRR) framework, that is, regression coefficients are assigned normal independent and identically distributed normal distributions, with mean zero and variance . See details of BRR in Pérez and de los Campos (2014).
Predictor P4: E+G+GE+BRR
The E+G+GE+BRR predictor is similar to P2, but also incorporates ECs as fixed effects in a Bayesian Ridge Regression (BRR) framework (see Appendix for brief details on Bayesian Ridge Regression). The priors used for GBLUP and BRR in BGLR are those default settings which are given with details in Pérez and de los Campos (2014). In this study, we found these default settings to be suitable, as we experimented with various configurations of the prior hyperparameters for the GBLUP and BRR models on the USP and G2F_2014 datasets. Remarkably, all configurations yielded identical predictions. Consequently, for the remaining datasets, we opted to utilize only the default settings.
Evaluation of prediction performance
The cross-validation approach used in this study involved leaving one environment out. In each iteration, the data from a single-environment served as the testing set, while the data from all other families constituted the training set (Montesinos-López et al., 2022). The number of iterations was equal to the number of environments to ensure that each environment was used as the testing set exactly one time. This method was employed to assess the model’s ability to predict information from a complete environment using data from other environments.
To evaluate the predictive performance we used the Mean Square Error (MSE) that quantifies the prediction error by measuring the squared deviation between observed and predicted values on the testing set. The MSE was computed for each scenario evaluated (NoEC, EC and FE) and then for comparing these three scenarios was computed the relative efficiencies as:
compares the prediction performance of EC vs NoEC, compares the prediction performance of FE vs NoEC and compares the prediction performance of FE vs EC. When the best prediction performance was obtained by the EC strategy, while when the strategy NoEC was the best. While when the relative efficiencies are equal to 1 means that both methods had equal prediction performance. The same interpretation applies for the other comparisons in terms of RE.
Results
The results are given in three sections for three datasets (Japonica, USP and G2F_2016). For each section we provided the results for the four predictor models under study (E+G, E+G+GE, E+G+BRR, E+G+GE+BRR) and under each predictor we compared three strategies for the use of the environmental covariates: NoEC, using environmental covariables (EC) and using environmental covariables with FE. Additionally, Appendix A contains comprehensive details of the BRR model utilized in this study. Furthermore, Appendix B offers extensive information on the outcomes for Japonica, USP, and G2F_2016 datasets, which are outlined in Table B1–Table B2, Table B3, Table B4, Table B4–Table B5 respectively. Additionally, Table B7 in this appendix presents the variance components and heritability of each trait within every dataset. For the results pertaining to datasets G2F_2014 and G2F_2015, please refer to the Supplementary Materials section.
Table B1 The prediction performance and the relative efficiency (RE) for Japonica dataset in terms of mean squared error (MSE) for each Environment and for each trait, for the predictors E+G and E+G+GE under three different techniques to compute the Kernel for the effect of the Environment: without Environmental Covariates (NoEC), using Environmental covariates (EC) and using Environmental Covariates with Feature Engineering (FE).
Table B2 The prediction performance and the relative efficiency (RE) for Japonica dataset in terms of mean squared error (MSE) for each Environment and for each trait, for the predictors E+G+BRR and E+G+GE+BRR under three different techniques to compute the Kernel for the effect of the Environment: without Environmental Covariates (NoEC), using Environmental covariates (EC) and using Environmental Covariates with Feature Engineering (FE).
Table B3 The prediction performance and the relative efficiency (RE) for USP dataset in terms of mean squared error (MSE) for each Environment and for each trait, for the predictors E+G and E+G+GE under three different techniques to compute the Kernel for the effect of the Environment: without Environmental Covariates (NoEC), using Environmental covariates (EC) and using Environmental Covariates with Feature Engineering (FE).
Table B4 The prediction performance and the relative efficiency (RE) for USP dataset in terms of mean squared error (MSE) for each Environment and for each trait, for the predictors E+G+BRR and E+G+GE+BRR under three different techniques to compute the Kernel for the effect of the Environment: without Environmental Covariates (NoEC), using Environmental covariates (EC) and using Environmental Covariates with Feature Engineering (FE).
Table B5 The prediction performance and the relative efficiency (RE) for G2F_2016 dataset in terms of mean squared error (MSE) for each Environment and for each trait, for the predictors E+G and E+G+GE under three different techniques to compute the Kernel for the effect of the Environment: without Environmental Covariates (NoEC), using Environmental covariates (EC) and using Environmental Covariates with Feature Engineering (FE).
Table B6 The prediction performance and the relative efficiency (RE) for G2F_2016 dataset in terms of mean squared error (MSE) for each Environment and for each trait, for the predictor E+G+BRR and E+G+GE+BRR under three different techniques to compute the Kernel for the effect of the Environment: without Environmental Covariates (NoEC), using Environmental covariates (EC) and using Environmental Covariates with Feature Engineering (FE).
Table B7 Variance components (Var_Comp) for environment (Env) Line and Genotype by environment (Env:Line) interaction for each data set. CV denotes coefficient of variation and n_Env denotes the average of number of environments in each data set.
Japonica dataset
Predictor: E+G
Figure 1A provides a summary of Table B1 across traits and reveals that FE outperformed EC in most environments with improvements of 20.260% (2010), 38.920% (2011), 1.750% (2012), and 25.470% (2013). This results in an average RE of 1.1567. EC, on the other hand, outperformed NoEC in most environments with improvements of 121.200% (2009), 48.080% (2010), and 8.140% (2012), resulting in an average RE of 1.277. Likewise, FE outperformed NoEC in 101.240% (2009), 59.560% (2010), and 4.710% (2012), with slight losses in other environments, but an average RE of 1.2814. This indicates that using EC and FE surpassed NoEC by 27.730% and 28.140%, respectively. These calculations are derived from the results presented in Table B1.
Figure 1 The three relative efficiencies, considering EC_vs_FE, NoEC_vs_EC, and NoEC_vs_FE, for Japonica dataset, for predictors (A) E+G, (B) E+G+GE, (C) E+G+BRR and (D) E+G+GE+BRR in terms of mean squared error (MSE) for each Environment across traits.
Predictor: E+G+GE
Figure 1B summarizes the findings from Table B1 across traits, illustrating the comparative performance of FE, EC, and NoEC techniques in various environments. The results indicate that FE outperformed EC in the majority of environments, with improvements of 4.280% (2010), 40.050% (2011), and 20.220% (2013), resulting in an average RE of 1.099. On the other hand, EC outperformed NoEC in most environments, with improvements of 78.070% (2009), 16.100% (2012), and 147.980% (2013), yielding an average RE of 1.430. Furthermore, FE surpassed the conventional NoEC technique by 68.990% (2009), 1.780% (2012), and 178.280% (2013), with an average RE of 1.462. These results indicate that using EC and FE techniques outperformed the conventional NoEC technique by 43.040% and 46.150%, respectively. The calculations are derived from the outcomes presented in Table B1.
Predictor: E+G+BRR
Figure 1C provides an overview of Table B2 across traits. It reveals that FE outperformed EC only in environments 2010 (9.630%) and 2011 (25.340%), resulting in an average RE of 0.975. On the other hand, EC outperformed NoEC in all environments, with percentages of improvement of 92.640% (2009), 20.690% (2010), 15.960% (2011), 36.170% (2012), and 9.070% (2013), and an average RE of 1.349. Additionally, FE outperformed the NoEC technique in 80.390% (2009), 34.120% (2010), 13.690% (2011), and 21.950% (2012) of the environments with a slight loss in 2013, but an average RE of 1.269. These findings indicate that using EC and FE techniques surpassed NoEC in 34.910% and 26.940% of the environments, respectively. The calculations are based on the results presented in Table B2.
Predictor: E+G+GE+BRR
Figure 1D summarizes the findings from Table B2 across traits. It reveals that FE displayed a superior performance over EC in environments 2010 (14.770%), 2011 (21.700%), and 2013 (17.870%), resulting in an average RE of 1.064. On the other hand, EC outperformed NoEC in most environments, namely 67.750% (2009), 28.390% (2010), 27.210% (2011), and 183.970% (2013), with an average RE of 1.614. Moreover, FE outperformed NoEC in most environments, specifically 54.260% (2009), 35.520% (2010), 33.140% (2011), and 197.980% (2013), with an average RE of 1.604. These findings indicate that using EC and FE surpassed NoEC in 61.390% and 60.460% of cases, respectively. The computations for these results were based on the findings presented in Table B2.
USP dataset
Predictor: E+G
Figure 2A and Table B3 provide the results of our comparison between the NoEC and FE techniques using the RE metric. FE outperformed the NoEC technique only in Env1 (1.107), displaying an improvement of 10.670%. However, in Env2 (0.910), Env3 (0.8123), and Env4 (0.989), the NoEC technique surpassed FE, resulting in an average RE of 0.955. This average RE indicates a general loss of 4.520% when using FE compared to NoEC (see Table B3).
Figure 2 The three relative efficiencies, considering EC_vs_FE, NoEC_vs_EC, and NoEC_vs_FE, for USP dataset, for predictors (A) E+G, (B) E+G+GE, (C) E+G+BRR and (D) E+G+GE+BRR in terms of mean squared error (MSE) for each Environment.
Predictor: E+G+GE
Figure 2B and Table B3 provide the results of our comparison between the NoEC and FE techniques based on the RE metric, including the fact that the use of FE outperformed the use of NoEC in environments Env1 (1.167), Env2 (1.016), and Env4 (1.064), resulting in respective improvements of 16.670%, 1.550%, and 6.390%. However, in Env3 (0.912), the NoEC technique outperformed FE, resulting in an average RE of 1.040. This average RE indicates a general improvement of 4.000% of the FE technique regarding the NoEC method. For more detailed information, see Table B3.
Predictor: E+G+BRR
Based on Figure 2C and Table B4, our comparison between the NoEC and FE techniques using the RE metric reveals that FE outperformed the NoEC technique in environments Env1 (1.216), Env3 (1.189), and Env4 (1.435), displaying improvements of 21.580%, 18.890%, and 43.500%, respectively. However, in Env2 (0.768), the NoEC technique outperformed using FE. In general, FE outperformed NoEC by 15.200% since an average RE of 1.152 was observed (see Table B4).
Predictor: E+G+GE+BRR
Finally, based on the analysis presented in Figure 2D and Table B4, we compared the NoEC and FE techniques using the RE metric. The results indicate that FE outperformed NoEC in Env1 (1.231), Env3 (1.368), and Env4 (1.491), displaying improvements of 23.090%, 36.760%, and 49.080%, respectively. However, in Env2 (0.901), the NoEC technique outperformed FE, although, FE outperformed the NoEC in general terms, since an average RE of 1.248 was observed (see Table B4).
G2F_2016 dataset
Predictor: E+G
Figure 3A summarizes Table B5 across different environments for each trait. It reveals that FE outperformed EC in all traits, achieving improvements of 87.970% (Grain_Moisture_BLUE), 58.100% (Grain_Moisture_weight), 21.030% (Yield_Mg_ha_BLUE), and 89.600% (Yield_Mg_ha_weight), resulting in an average RE of 1.642. In contrast, EC outperformed NoEC in most traits, with improvements of 63.960% (Grain_Moisture_BLUE), 1682.340% (Grain_Moisture_weight), and 52.860% (Yield_Mg_ha_weight), yielding an average RE of 5.497. Additionally, FE surpassed NoEC in all traits, with enhancements of 119.370% (Grain_Moisture_BLUE), 245.980% (Grain_Moisture_weight), 1.400% (Yield_Mg_ha_BLUE), and 22.630% (Yield_Mg_ha_weight), resulting in an average RE of 1.974. These findings indicate that both EC and FE techniques outperformed NoEC by 449.740% and 97.350%, respectively. The computations are based on the results presented in Table 5B.
Figure 3 The three relative efficiencies, considering EC_vs_FE, NoEC_vs_EC, and NoEC_vs_FE, for G2F_2016 dataset, for predictors (A) E+G, (B) E+G+GE, (C) E+G+BRR and (D) E+G+GE+BRR in terms of mean squared error (MSE) for each trait across environments.
Predictor: E+G+GE
Figure 3B and Table B5 shows that for the Yield_Mg_ha_weight trait, the NoEC technique achieved the best performance in most environments, as shown by the MSE values (DEH1_2016 [0.051], GAH1_2016 [0.026], IAH1_2016 [2.914], IAH2_2016 [0.069], MIH1_2016 [0.055], MNH1_2016 [0.146], NEH1_2016 [0.033], NYH2_2016 [0.449] and OHH1_2016 [1.202]). On average, there were slight losses of 2.210% and 2.570% when comparing EC versus NoEC and FE versus NoEC, respectively. This suggests that EC and FE techniques could have performed more adequately than the conventional NoEC technique. However, comparing EC and FE techniques based on RE showed that FE outperformed EC in most environments under NoEC, resulting in an average RE of 1.339, indicating a superiority of 33.930% for FE (see Table 5B).
Predictor: E+G+BRR
Figure 3C summarizes the findings from Table B6 across environments for each trait. It shows that FE outperformed EC in all characteristics, with improvements of 67.090% (Grain_Moisture_BLUE), 167.270% (Grain_Moisture_weight), 10.650% (Yield_Mg_ha_BLUE), and 3.960% (Yield_Mg_ha_weight), resulting in an average RE of 1.622. Additionally, EC outperformed NoEC in all traits, with improvements of 84.880% (Grain_Moisture_BLUE), 249.510% (Grain_Moisture_weight), 3.780% (Yield_Mg_ha_BLUE), and 51.630% (Grain_Moisture_weight), resulting in an average RE of 1.975. Furthermore, FE outperformed NoEC only in the traits Grain_Moisture_BLUE (129.850%) and Grain_Moisture_weight (25.410%), with an average RE of 1.360. These results indicate that EC and FE techniques outperformed the conventional NoEC technique in 62.240% and 36.020% of cases, respectively. These calculations are derived from the results presented in Table B6.
Predictor: E+G+GE+BRR
Figure 3D summarizes the results from Table B6 across different traits. It shows that FE outperformed EC in the majority of traits, specifically by 29.090% for Grain_Moisture_BLUE, 689.960% for Grain_Moisture_weight, and 38.420% for Yield_Mg_ha_weight. This leads to an average RE of 2.893. On the other hand, EC outperformed NoEC in all traits, with improvements of 65.180% for Grain_Moisture_BLUE, 408.510% for Grain_Moisture_weight, 11.690% for Yield_Mg_ha_BLUE, and 22.200% for Yield_Mg_ha_weight. The average RE for EC compared to NoEC is 2.269. Furthermore, FE outperformed NoEC in all traits, with improvements of 125.150% for Grain_Moisture_BLUE, 240.900% for Grain_Moisture_weight, 9.490% for Yield_Mg_ha_BLUE, and 11.380% for Yield_Mg_ha_weight. The average RE for FE compared to NoEC is 1.967. These results indicate that using EC and FE outperformed NoEC by 126.890% and 96.730%, respectively. These computations are derived from the outcomes of Table B6.
Summary across data sets for each predictor
In Table 1 we can observe that in any of the four predictors using environmental covariates improve prediction accuracy at least 61.400% regarding of not using the environmental covariates (NoEC_vs_EC). Also, we can see in this same table that using FE improves the prediction performance in the four predictors regarding of using the original environmental covariates (EC_vs_FE) in at least 347.300%. Regarding using FE and not using environmental covariates (NoEC_vs_FE) we can observe that also in the four predictors using FE outperform by at least 113.100% not using the environmental covariates. Also, we observed that in many cases adding directly the environmental covariates (EC) not improve (and even reduce) the prediction performance and for this reason, we observe that the gain in terms of prediction performance of NoEC_vs_FE is less pronounced regarding comparing EC_vs_FE.
Discussions
Due to the fact, that still the practical implementation of the GS methodology is challenging since not always is possible to guarantee high genomic-enabled prediction accuracy, many strategies had been developed to improve the machine learning genomic prediction ability (Sallam and Smith, 2016). For this reason, since the GS methodology is still not optimal, this investigation explored FE on the environmental covariates. FE is a crucial step in machine learning and data science that involves creating new features or modifying existing ones to improve the performance of a model. FE is a creative and essential aspect of the machine learning workflow, and it can significantly impact the success of one’s models. It is a skill that improves with experience and a deep understanding of the data and problem. For this reason, FE has been applied successfully in solving natural language processing, computer vision, time series and other issues.
FE is not new in the context of GS, since some studies had been conducted exploring feature engineering techniques from the feature selection point of view. For example, Long et al. (2011) used dimension reduction and variable selection for genomic selection to predict milk yield in Holsteins. Tadist et al. (2019) present a systematic and structured literature review of the feature-selection techniques used in studies related to big genomic data analytics. While Meuwissen et al. (2017) proposed variable selection models for genomic selection using whole-genome sequence data and singular value decomposition. More recently Montesinos-López et al. (2023) proposed feature selection methods for selecting environmental covariables to enhance genomic prediction accuracy. However, these studies are only focused on feature selection and not create new features from the original inputs.
From our results across traits and data sets, we can state that including environmental covariates significantly improves the prediction performance, since comparing no environmental covariates (NoEC) vs adding environmental covariates (EC), the resulting improvement was of 167.900% (RE=2.679 of NoEC_vs_EC), 142.100 (RE=2.242 of NoEC_vs_EC), 56.100% (RE=1.561 of NoEC_vs_EC) and 421.300% (RE=5.213 of NoEC_vs_EC) under predictor E+G, E+G+GE, E+G+BRR and E+G+GE+BRR respectively. However, it is very interesting to point out that the prediction performance can be even improved when the covariates are included but using FE. We found that the improvement of the prediction performance using FE only including only the EC was of 816.600% (RE=9.166 of EC_vs_FE), 372.900% (RE=4.729 of EC_vs_FE), 616.100% (RE=716.100 of EC_vs_FE) and 1240.900% (RE=13.409% of EC_vs_FE) under predictors E+G, E+G+GE, E+G+BRR and E+G+GE+BRR respectively. The larger gain in prediction performance was observed under the most complex predictor (E+G+GE+BRR), while the lowest gain was observed under predictor E+G+GE. Our results show that FE in genomic prediction holds tremendous potential for advancing our understanding of genetics and improving predictions related to various aspects of genomics. For this reason, FE should be considered an important tool to unlock the potential of genomic data for research and practical applications of genomic prediction.
Although our results are very promising for the use of FE, its practical implementation is very challenging, since we observed a significant improvement in some data sets but not in all, and for practical implementations, we need to be able to identify with a high degree of accuracy when the use of FE will be beneficial and when the use of this approach will not be successful. Also, it is important to point out that we have opted against utilizing the Pearson’s correlation coefficient as a performance metric for predicting outcomes. This decision is principally rooted in the lack of substantial improvement linked to this measure we observed. The marginal benefits observed with this metric can be partly ascribed to our exclusive focus on feature selection within the realm of environmental covariates. Additionally, this can be attributed to the assessment of environmental covariates not at the genotype level but rather at the environmental (location) level.
Three reasons why the FE works well for some data but not very well for others are: (1) that those data sets with low efficiency with FE are those in which the environmental covariates are less correlated with the response variable, (2) that we speculate that not for all data sets the type of FE we implemented are efficient and (3) FE capture complex relationships between the inputs and the response variable. These mean that the nature of each data set affects substantially the performance of any FE strategy. For these reasons some challenges for its implementation are: a) Domain Knowledge Requirement: Effective FE often requires a deep understanding of the domain. With domain expertise, it can be easier to identify relevant features or transformations that could enhance model performance; b) Data Quality and Quantity: Obtaining high-quality and sufficient data for FE can be challenging in many practical scenarios. Limited or noisy data can hinder the creation of meaningful features; c) Time and Resource Constraints: Implementing FE can be time-consuming, and in some real-world applications, there might be strict time and resource constraints. This makes exploring and experimenting with a wide range of FE techniques challenging; d) Dynamic Data: Real-world data often changes over time. Features that are effective at one point in time may become less relevant or even obsolete as the data distribution evolves. Maintaining and updating features in dynamic environments can be challenging; e) Overfitting Risks: Aggressive. FE can lead to overfitting, especially when the number of features is large compared to the amount of available data. Overfit models perform well on training data but generalize poorly to new, unseen data; f) Complexity and Interpretability: As the number and complexity of features increase, the resulting models can become difficult to interpret. This lack of interpretability can be challenging, especially in applications where understanding the model’s decisions is crucial; g) Automated Feature Selection: While manual FE can be effective, the process is often subjective and time-consuming. Automated feature selection methods exist, but selecting the right techniques and parameters can be challenging; h) Curse of Dimensionality: As the number of features increases, the curse of dimensionality becomes more pronounced. This can lead to increased computational requirements and decreased model performance, making it challenging to strike the right balance.
The results of this study demonstrate that the feature engineering strategy for incorporating environmental covariates effectively enhances genomic prediction accuracy. However, further research is warranted to refine the methodology for integrating environmental covariates into genomic prediction models, particularly in the context of modeling genotype-environment interactions (GE). For instance, employing the factor analytic (FA) multiplicative operator to describe cultivar effects in different environments has shown promise as a robust and efficient machine learning approach for analyzing multi-environment breeding trials (Piepho, 1998; Smith et al., 2005). Factor analysis offers solutions for modeling GE with heterogeneous variances and covariances, either alongside the numerical relationship matrix (based on pedigree information) (Crossa et al., 2006) or utilizing the genomic similarity matrix to assess GE (Burgueño et al., 2012). Further research is needed to comprehensively explore the application of the FA approach for feature engineering of environmental covariates within the framework of genomic prediction.
Conclusions
This study delved into the impact of feature engineering on environmental covariates to enhance the predictive capabilities of genomic models. Our findings demonstrate a consistent improvement in prediction performance, as measured by MSE, across most datasets when employing feature engineering techniques compared to models without such enhancements. While some datasets showed no significant gains, others exhibited notably substantial improvements. These results underscore the potential of feature engineering to bolster prediction accuracy in genomic studies. However, it’s imperative to acknowledge the inherent complexity and challenges associated with practical implementation, as various factors can influence its efficacy. Therefore, we advocate for further exploration and adoption of feature engineering methodologies within the scientific community to accumulate more empirical evidence and harness its full potential in genomic prediction.
Data availability statement
The original contributions presented in the study are included in the article/Supplementary Material. Further inquiries can be directed to the corresponding authors.
Author contributions
OM: Writing – review & editing, Writing – original draft, Software, Methodology, Investigation, Conceptualization. LC: Writing – review & editing, Conceptualization. CS: Writing – review & editing, Supervision, Project administration, Investigation. BC: Writing – review & editing, Software, Methodology, Formal Analysis, Data curation, Conceptualization. GH: Writing – review & editing, Software, Conceptualization. BA: Writing – review & editing, Software, Methodology, Investigation, Data curation. SR: Writing – review & editing, Software, Methodology, Investigation. GG: Writing – review & editing, Methodology, Investigation, Data curation. KA: Writing – review & editing, Methodology, Investigation. RF: Writing – review & editing, Methodology, Investigation, Conceptualization. AM: Writing – review & editing, Software, Methodology, Investigation, Conceptualization. JC: Writing – review & editing, Writing – original draft, Investigation, Conceptualization.
Funding
The author(s) declare financial support was received for the research, authorship, and/or publication of this article. Open Access fees were received from the Bill & Melinda Gates Foundation. We acknowledge the financial support provided by the Bill & Melinda Gates Foundation (INV-003439 BMGF/FCDO Accelerating Genetic Gains in Maize and Wheat for Improved Livelihoods (AGG)) as well as the USAID projects (Amend. No. 9 MTO 069033, USAID-CIMMYT Wheat/AGGMW, Genes 2023, 14, 927 14 of 18AGG-Maize Supplementary Project, AGG (Stress Tolerant Maize for Africa)) which generated the CIMMYT data analyzed in this study. We are also thankful for the financial support provided by the Foundation for Research Levy on Agricultural Products (FFL) and the Agricultural Agreement Research Fund (JA) through the Research Council of Norway for grants 301835 (Sustainable Management of Rust Diseases in Wheat) and 320090 (Phenotyping for Healthier and more Productive Wheat Crops). We acknowledge the support of the Window 1 and 2 funders to the Accelerated Breeding Initiative (ABI).
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
The handling editor HL declared a past co-authorship with the author JC.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Supplementary material
The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fpls.2024.1349569/full#supplementary-material
References
Abed, A., Pérez-Rodríguez, P., Crossa, J., Belzile, F. (2018). When less can be better: how can we make genomic selection more cost-effective and accurate in Barley? Theor. Appl. Genet. 131, 1873–1890. doi: 10.1007/s00122-018-3120-8
Afshar, M., Usefi, H. (2020). High-dimensional feature selection for genomic datasets. Knowledge-Based Syst. 206, 106370. doi: 10.1016/j.knosys.2020.106370
Akdemir, D., Sanchez, J. I., Jannink, J. L. (2015). Optimization of genomic selection training populations with a genetic algorithm. Genet. Selection Evolution. 47, 1–10. doi: 10.1186/s12711-015-0116-6
Bassi, F. M., Bentley, A. R., Charmet, G., Ortiz, R., Crossa, J. (2016). Breeding schemes for the implementation of genomic selection in wheat (Triticum spp.). Plant Sci. 242, 23–36. doi: 10.1016/j.plantsci.2015.08.021
Bermingham, M. L., Pong-Wong, R., Spiliopoulou, A., Hayward, C., Rudan, I., Campbell, H., et al. (2015). Application of high-dimensional feature selection: evaluation for genomic prediction in man. Sci. Rep. 5, 10312. doi: 10.1038/srep10312
Bernardo, R., Yu, J. (2007). Prospects for genomewide selection for quantitative traits in maize. Crop Sci. 47, 1082–1090. doi: 10.2135/cropsci2006.11.0690
Beyene, Y., Semagn, K., Mugo, S., Tarekegne, A., Babu, R., Meisel, B., et al. (2015). Genetic gains in grain yield through genomic selection in eight bi-parental maize populations under drought stress. Crop Sci. 55, 154–163. doi: 10.2135/cropsci2014.07.0460
Budhlakoti, N., Kushwaha, A. K., Rai, A., Chaturvedi, K. K., Kumar, A., Pradhan, A. K., et al. (2022). Genomic selection: A tool for accelerating the efficiency of molecular breeding for development of climate-resilient crops. Front. Genet. 13. doi: 10.3389/fgene.2022.832153
Burgueño, J., de los Campos, G., Weigel, K., José Crossa, J. (2012). Genomic prediction of breeding values when modeling genotype × environment interaction using pedigree and dense molecular markers. Crop Sci. 52, 707–719. doi: 10.2135/cropsci2011.06.0299
Butoto, E. N., Brewer, J. C., Holland, J. B. (2022). Empirical comparison of genomic and phenotypic selection for resistance to Fusarium ear rot and fumonisin contamination in maize. Theor. Appl. Genet. 135, 2799–2816. doi: 10.1007/s00122-022-04150-8
Calus, M. P. L., Veerkamp, R. F. (2011). Accuracy of multi-trait genomic selection using different methods. Genet. Sel Evol. 43, 1–14. doi: 10.1186/1297-9686-43-26
Carrillo-de-Albornoz, J., Rodriguez Vidal, J., Plaza, L. (2018). Feature engineering for sentiment analysis in e-health forums. PloS One 13, e0207996. doi: 10.1371/journal.pone.0207996
Costa-Neto, G., Crossa, J., Fritsche-Neto, R. (2021b). Enviromic assembly increases accuracy and reduces costs of the genomic prediction for yield plasticity in maize. Front. Plant Sci. 12. doi: 10.3389/fpls.2021.717552
Costa-Neto, G., Fritsche-Neto, R., Crossa, J. (2021a). Nonlinear kernels, dominance, and envirotyping data increase the accuracy of genome-based prediction in multi-environment trials. Heredity 126, 92–106. doi: 10.1038/s41437-020-00353-1
Crossa, J., Burgueño, J., Cornelius, P. L., McLaren, G., Trethowan, R., Krishnamachari, A. (2006). Modeling genotype X environment interaction using additive genetic covariances of relatives for predicting breeding values of wheat genotypes. Crop Sci. 46, 1722–1733. doi: 10.2135/cropsci2005.11-0427
Crossa, J., Pérez-Rodríguez, P., Cuevas, J., Montesinos-López, O., Jarquín, D., de los Campos, G., et al. (2017). Genomic selection in plant breeding: methods, models, and perspectives. Trends Plant Sci. 22, 961–975. doi: 10.1016/j.tplants.2017.08.011
Desta, Z. A., Ortiz, R. (2014). Genomic selection: genome-wide prediction in plant improvement. Trends Plant Sci. 19, 592–601. doi: 10.1016/j.tplants.2014.05.006
Dong, G., Liu, H. (2018). Feature engineering for machine learning and data analytics (California, USA: CRC press).
FAO. (2023). “The state of food security and nutrition in the world 2023,” in Urbanization, agrifood systems transformation and healthy diets across the rural–urban continuum (Rome, Italy: FAOSTAT).
Gesteiro, N., Ordás, B., Butrón, A., de la Fuente, M., Jiménez-Galindo, J. C., Samayoa, L. F., et al. (2023). Genomic versus phenotypic selection to improve corn borer resistance and grain yield in maize. Front. Plant Sci. 14. doi: 10.3389/fpls.2023.1162440
Habier, D., Fernando, R. L., Dekkers, J. C. M. (2007). The impact of genetic relationship information on genome-assisted breeding values. Genetics 177, 2389–2397. doi: 10.1534/genetics.107.081190
Heaton, J. (2016). “An empirical analysis of feature engineering for predictive modeling,” in SoutheastCon 2016. (pp. 1–(pp.6) (Norfolk, Virginia, USA: IEEE). doi: 10.1109/SECON.2016.7506650
Heffner, E. L., Sorrells, M. E., Jannink, J.-L. (2009). Genomic selection for crop improvement. Crop Sci. 49, 1–12. doi: 10.2135/cropsci2008.08.0512
Hu, H., Campbell, M. T., Yeats, T. H., Zheng, X., Runcie, D. E., Covarrubias-Pazaran, G., et al. (2021). Multi-omics prediction of oat agronomic and seed nutritional traits across environments and in distantly related populations. Theor. Appl. Genet. 134 (12), 4043–4054. doi: 10.1007/s00122-021-03946-4
Jarquin, D., Crossa, J., Lacaze, X., Du Cheyron, P., Daucourt, J., Lorgeou, J., et al. (2014). A reaction norm model for genomic selection using high-dimensional genomic and environmental data. Theor. Appl. Genet. 127 (3), 595–607. doi: 10.1007/s00122-013-2243-1
Jarquin, D., de Leon, N., Romay, C., Bohn, M., Buckler, E. S., Ciampitti, I., et al. (2020). Utility of climatic information via combining ability models to improve genomic prediction for yield within the genomes to fields maize project. Front. Genet. 11, 592769. doi: 10.3389/fgene.2020.592769
Juliana, P., Singh, R. P., Poland, J., Mondal, S., Crossa, J., Montesinos-López, O. A., et al. (2018). Prospects and challenges of applied genomic selection-A new paradigm in breeding for grain yield in bread wheat. Plant Genome 11, 1–17. doi: 10.3835/plantgenome2018.03.0017
Khurana, U., Samulowitz, H., Turaga, D. (2018). “Feature engineering for predictive modeling using reinforcement learning,” in Proceedings of the AAAI Conference on Artificial Intelligence (New Orleans, LA, USA), Vol. 32.
Krause, M. R., González-Pérez, L., Crossa, J., Pérez-Rodríguez, P., Montesinos-López, O., Singh, R. P., et al. (2019). Hyperspectral reflectance derived relationship matrices for genomic prediction of grain yield in wheat. G3 Genes Genomes Genet. 9, 1231–1247. doi: 10.1534/g3.118.200856
Lam, H. T., Thiebaut, J. M., Sinn, M., Chen, B., Mai, T., Alkan, O. (2017). One button machine for automating feature engineering in relational databases. arXiv. doi: 10.48550/arXiv.1706.00327
Lawrence-Dill, C. J. (2017) Genomes to fields: GxE Field Experiment. Available online at: https://www.genomes2fields.org/resources/#sop (Accessed 2021 January 11).
Long, N., Gianola, D., Rosa, G. J., Weigel, K. A. (2011). Dimension reduction and variable selection for genomic selection: application to predicting milk yield in Holsteins. J. Anim. Breed Genet. 128, 247–257. doi: 10.1111/jbg.2011.128.issue-4
Meuwissen, T. H. E., Hayes, B. J., Goddard, M. E.. (2001). Prediction of total genetic value using genome-wide dense marker maps. Genetics 157 (12), 1819–1829. doi: 10.1093/GENETICS/157.4.1819
Meuwissen, T. H. E., Indahl, U. G., Ødegård, J. (2017). Variable selection models for genomic selection using whole-genome sequence data and singular value decomposition. Genet. Sel Evol. 49, 94. doi: 10.1186/s12711-017-0369-3
Montesinos-López, O. A., Montesinos-López, A., Pérez-Rodríguez, P., Barrón-López, J. A., Martini, J. W. R., Fajardo-Flores, S. B., et al. (2021). A review of deep learning applications for genomic selection. BMC Genomics 22 (1), 19. doi: 10.1186/s12864-020-07319-x
Montesinos-López, O. A., Montesinos-López, A., Crossa, J. (2022). Multivariate statistical Machine Learning Methods for Genomic Prediction (Cham, Switzerland: Springer International Publishing). doi: 10.1007/978-3-030-89010-0
Montesinos-López, O. A., Crespo-Herrera, A., Saint Pierre, J., Bentley, J., de la Rosa-Santamaria, J., Ascencio-Laguna, J., et al (2023). Do feature selection methods for selecting environmental covariables enhance genomic prediction accuracy?. Front. Genet. 14, 1209275. doi: 10.3389/fgene.2023.1209275
Montesinos-López, A., Montesinos-López, O. A., Cuevas, J., Mata-López, W. A., Burgueño, J., Mondal, S., et al. (2017). Genomic Bayesian functional regression models with interactions for predicting wheat grain yield using hyper-spectral image data. Plant Methods 13, 1–29. doi: 10.1186/s13007-017-0212-4
Monteverde, E., Gutierrez, L., Blanco, P., Pérez de Vida, F., Rosas, J. E., Bonnecarrère, V., et al. (2019). Integrating molecular markers and environmental covariates to interpret genotype by environment interaction in rice (Oryza sativa L.) grown in subtropical areas. G3 (Bethesda). 9, 1519–1531. doi: 10.1534/g3.119.400064
Nargesian, F., Samulowitz, H., Khurana, U., Khalil, E. B., Turaga, D. S. (2017). “Learning feature engineering for classification,” in Ijcai (Melbourne Australia), Vol. 17. 2529–2535.
Pérez, P., de los Campos, G. (2014). Genome-wide regression and prediction with the BGLR statistical package. Genetics 198, 483–495. doi: 10.1534/genetics.114.164442
Piepho, H. P. (1998). Empirical best linear unbiased prediction in cultivar trials using using factor analytic variance covariance structure. Theor. Appl. Genet. 97, 195–201. doi: 10.1007/s001220050885
Rincent, R., Laloë, D., Nicolas, S., Altmann, T., Brunel, D., Revilla, P., et al. (2012). Maximizing the reliability of genomic selection by optimizing the calibration set of reference individuals: comparison of methods in two diverse groups of maize inbreds (Zea mays L.). Genetics 192, 715–728. doi: 10.1534/genetics.112.141473
Rogers, A. R., Dunne, J. C., Romay, C., Bohn, M., Buckler, E. S., Ciampitti, I. A., et al. (2021). The importance of dominance and genotype-by-environment interactions on grain yield variation in a large-scale public cooperative maize experiment. G3 (Bethesda) 11, jkaa050. doi: 10.1093/g3journal/jkaa050
Rogers, A. R., Holland, J. B. (2022). Environment-specific genomic prediction ability in maize using environmental covariates depends on environmental similarity to training data. G3 Genes|Genomes|Genetics 12, jkab440. doi: 10.1093/g3journal/jkab440
Sallam, A. H., Smith, K. P. (2016). Genomic selection performs similarly to phenotypic selection in Barley. Crop Sci. 56, 2871–2881. doi: 10.2135/cropsci2015.09.0557
Smith, A. B., Cullis, B. R., Thompson, R. (2005). The analysis of crop cultivar breeding and evluation trials: An overview of current mixed model approaches. J. Agric. Sci. 143, 1–14. doi: 10.1017/S0021859605005587
Tadist, K., Najah, S., Nikolov, N. S., Mrabti, F., Zahi, A. (2019). Feature selection methods and genomic big data: a systematic review. J. Big Data 6, 79. doi: 10.1186/s40537-019-0241-0
VanRaden, P. M. (2008). Efficient methods to compute genomic predictions. J. Dairy Sci. 91, 4414–4423. doi: 10.3168/jds.2007-0980
Wang, K., Abid, M. A., Rasheed, A., et al. (2023). DNNGP, a deep neural network-based method for genomic prediction using multi-omics data in plants. Mol. Plant 16, 279–293. doi: 10.1016/j.molp.2022.11.004
Wu, P. Y., Stich, B., Weisweiler, M., Shrestha, A., Erban, A., Westhoff, P., et al. (2022). Improvement of prediction ability by integrating multi-omic datasets in barley. BMC Genomics 23, 200. doi: 10.1186/s12864-022-08337-7
Xu, Y., Liu, X., Fu, J., Wang, H., Wang, J., Huang, C., et al. (2020). Enhancing genetic gain through genomic selection: from livestock to plants. Plant Commun. 1, 100005. doi: 10.1016/j.xplc.2019.100005
Yurek, O. E., Birant, D. (2019). “Remaining useful life estimation for predictive maintenance using feature engineering,” in 2019 Innovations in Intelligent Systems and Applications Conference (ASYU). 1–5 (Izmir, Turkeyand: IEEE). doi: 10.1109/ASYU48272.2019.894639
Appendix A
Bayesian ridge regression
Bayesian Ridge Regression (BRR) is a probabilistic approach to linear regression that incorporates Bayesian principles. It is a regularized regression method that extends traditional linear regression by introducing a prior distribution over the regression coefficients. This approach provides a way to express uncertainty in the model parameters and helps prevent overfitting by introducing regularization.
The model assumptions assumes a traditional linear regression, with a linear relationship between the independent variables and the dependent variable. The BRR assumes that the coefficients of the regression model follow a Gaussian (normal) distribution. This introduces a regularization term that penalizes large coefficients, helping to prevent overfitting.
The model formulation assumes that X is an independent variables with and a dependent variable y, such that the BRR can be written as
where y is the dependent variable. X is the matrix of independent variables, β is the vector of regression coefficients and ϵ is the residual (error) term. From a Bayesian perspective, the prior distribution for β is assumed to be Gaussian (normal) with α being a hyperparameter controlling the strength of the regularization and I is the identity matrix. The goal is to estimate the posterior distribution of β given the data. The posterior distribution is proportional to the product of the likelihood and the prior . Once the posterior distribution is obtained, Bayesian inference can be performed with. point estimates (mean or mode) of the posterior distribution can be used as the regression coefficients. additionally, credible intervals can be computed to quantify uncertainty.
Appendix B
Japonica dataset
Predictor: E+G
Table B1 shows an adequate performance for the results under NoEC for the GC trait across all environments. The MSE values for 2009, 2010, 2011, 2012, and 2013 were 0.0035, 0.0110, 0.0019, 0.0281, and 0.0017, respectively. Comparing the NoEC results to the EC and FE techniques using Relative Efficiency (RE), all RE values were below 1. On average, NoEC presented 50.050% better performance compared to EC and 42.230% better performance compared to FE. However, when comparing EC and FE techniques based on RE, FE outperformed EC in 2010, 2011, 2012, and 2013, with RE values of 1.287, 2.686, 1.139, and 1.586, respectively. In 2009, EC had a lower RE value of 0.522. On average, the use of FE outperformed EC by 44.410%. Please refer to Table B1 for more detailed information.
Concerning the GY trait, Table B1 shows that the use of EC led to a superior performance in most environments based on MSE (796,963 [2009], 2,488,872 [2010] and 1,157,280 [2012]). However, the exceptions occurred in 2011 and 2013, when FE achieved the best MSE values of 2,615,758 and 377,719, respectively. By contrast, when comparing NoEC versus EC and NoEC versus FE using RE, most RE values were greater than 1. On average, the EC technique displayed an improvement of 105.610% (NoEC_vs_EC) regarding the NoEC method, and an improvement of 77.570% (NoEC_vs_FE) was observed with the use of FE compared to the conventional NoEC technique. Nonetheless, when assessing the performance of EC and FE techniques based on RE, FE only outperformed EC in 2011 (RE = 1.091) and 2013 (RE = 1.087). EC, on the other hand, outperformed FE in 2009 (RE = 0.777), 2010 (RE = 0.817), and 2012 (RE = 0.806), resulting in an average RE of 0.916. This indicates an overall performance loss of 8.450% when using FE compared to EC. Table B1 provides further details.
In terms MSE for the PH trait, Table B1 shows that the use of FE achieved the best performance in most environments (15.872 [2009], 10.959 [2010], and 164.039 [2012]). However, there were exceptions in 2011 and 2013, where the best MSE values were 28.573 (EC) and 18.363 (NoEC), respectively. On the other hand, when comparing NoEC versus EC and NoEC versus FE techniques using RE, most RE values were greater than 1. On average, the use of EC and FE displayed improvements of 61.570% and 70.210%, respectively, compared to the use of NoEC. Furthermore, when comparing the performance of EC and FE techniques based on RE, FE outperformed EC in all environments, resulting in an average RE of 1.0389. This indicates that using FE surpassed EC by 3.88% (Table B1).
In terms of MSE for the PHR trait, Table B1 indicates that the use of FE yielded the best performance in most environments (0.001 [2009], 0.001 [2010], and 0.001[2013]). However, exceptions were found in 2011 and 2012, when the best MSE values were 0.001 (EC) and 0.006 (NoEC), respectively. On the other hand, when comparing EC versus FE and NoEC versus FE techniques using Relative Efficiency (RE), most RE values were at least 1. On average, the use of FE displayed a general improvement of 22.790%, compared to EC and 7.020% compared to the conventional NoEC technique. However, evaluating the performance of EC versus NoEC techniques based on RE showed that NoEC outperformed EC in most environments, resulting in an average RE of 0.938. This indicates a general accuracy loss of 6.200% when using EC compared to the conventional NoEC technique (Table B1).
Predictor: E+G+GE
Table B1 shows that, in most environments, the conventional NoEC technique yielded the best performance for the GC trait, with MSE values of 0.001 (2009), 0.013 (2010), and 0.002 (2011). The exceptions occurred in 2012 and 2013, with the best MSE values of 0.025 (EC) and 0.0023 (FE). The average RE for the comparison of NoEC versus EC and NoEC versus FE techniques across environments was 0.919 and 0.9023, respectively, indicating general losses of 8.080% and 9.740% for EC and FE compared to the conventional NoEC.
Regarding the GY trait, MSE values from Table B1 reveal that the use of EC achieved the best performance in most environments (1152261.030 [2009], 3653811.510 [2010], and 989127.170 [2012]). However, exceptions were observed in 2011 and 2013, where the best MSE values were 1834248.25 (NoEC) and 30980.32 (FE), respectively. On the other hand, when comparing NoEC versus EC and NoEC versus FE techniques using RE, most RE values were greater than 1. The average RE for NoEC versus EC and NoEC versus FE was 2.219 and 2.075, respectively, indicating general improvements of 121.860% and 107.520% compared to the use of NoEC. However, an evaluation of the performance of EC and FE techniques based on RE showed that FE outperformed EC only in 2011 (1.0267) and 2013 (1.122), while EC outperformed FE in 2009 (0.789), 2010 (0.849), and 2012 (0.7278). Consequently, the average RE for EC versus FE was 0.9029, implying a general loss of 9.710% when using FE compared to EC (Table B1).
Concerning the PH trait, the analysis of MSE values from Table B1 reveals that the use of FE yielded the best performance in most environments (17.631 [2009] and 23.544 [2012]). However, exceptions were observed in 2010, 2011, and 2013, where the best MSE values were 12.954 (EC), 44.689 (NoEC), and 164.891 (NoEC), respectively. On the other hand, comparing NoEC versus EC and NoEC versus FE techniques using RE showed that most RE values were greater than 1. The average RE for NoEC versus EC and NoEC versus FE was 1.618 and 1.700, respectively, indicating general improvements of 61.810% and 70.000% compared to the conventional NoEC technique. Furthermore, when evaluating the performance of EC and FE techniques based on RE, FE consistently outperformed EC in most environments. The average RE for EC versus FE was 1.047, indicating a 4.710% advantage in favor of FE (Table B1).
Moreover, in the case of the PHR trait, the analysis of MSE values from Table B1 shows that the use of FE yielded the best performance in most environments (0.001 [2009], 0.002 [2010], and 0.001 [2013]). However, there were exceptions in 2011 and 2012, where the best MSE values were 0.001 (EC) and 0.005 (NoEC), respectively. Furthermore, when comparing the RE values between NoEC versus EC and NoEC versus FE techniques, the average RE values of 0.966 and 1.168 indicate a slight loss of 3.440% and an improvement of 16.800%, respectively, for the use of EC and FE compared to the conventional NoEC technique. Nevertheless, when evaluating the performance of FE versus EC techniques based on RE, FE consistently outperformed EC in most environments. The average RE for FE versus EC was 1.282, indicating a significant improvement of 28.240% in accuracy for using FE compared to (Table B1).
Predictor: E+G+BRR
According to Table B2, the GC trait displayed superior performances with the conventional NoEC technique in most environments, yielding MSE values of 0.004 (2009), 0.002 (2011), and 0.0012 (2013). However, exceptions were found in 2010 and 2012, where FE achieved the best MSE values of 0.0680 and 0.009, respectively. Comparing the RE values between NoEC versus EC and NoEC versus FE techniques showed that most RE values were below 1. Nonetheless, the average RE of 1.104 (NoEC_vs_EC) and 1.189 (NoEC_vs_FE) indicated that EC and FE outperformed the conventional NoEC technique by 10.360% and 18.930%, respectively. Furthermore, when evaluating the performance of EC and FE techniques based on RE, FE presented the best performance in 2009 (1.151), 2010 (1.353), 2011 (2.044), and 2012 (1.0623), while EC outperformed FE in 2013 (0.529). Overall, the average RE 1.228 indicated that FE outperformed EC by 22.800% (Table B2).
Regarding the GY trait, Table B2 indicates that the conventional NoEC technique displayed superior performances in most environments, with MSE values of 5,683,515.750 (2010), 2,749,626.080 (2012), and 405,886.860 (2013). However, exceptions were observed in 2009 and 2011, where FE achieved the best MSE values of 3,049,246.320 and 4,024,422.450, respectively. When comparing the RE values between NoEC_vs_EC and NoEC_vs_FE techniques, most values were below 1. Nevertheless, the average RE of 1.124 (NoEC_vs_EC) and 0.896 (NoEC_vs_FE) indicated an overall improvement of 12.430% for EC and a general loss of 10.450% for FE compared to the conventional NoEC technique. However, when comparing the performance of EC and FE techniques based on RE, only FE presented a superior performance in 2010 (1.029), resulting in an average RE of 0.797, which indicates a general loss of 20.350% for FE compared to EC (Table B2).
For the PH trait, Table B2 shows that FE yielded the best performance in environments 2009 (15.281) and 2012 (159.312), while EC led to superior performances in environments 2010 (22.962) and 2013 (10.981). Most notably, when comparing the RE values for NoEC_vs_EC and NoEC_vs_FE, values exceeding 1 were observed. The average RE values of 1.634 (NoEC_vs_EC) and 1.5434 (NoEC_vs_FE) indicated substantial improvements of 63.350% and 54.350% respectively for using EC and FE, compared to the conventional NoEC technique. However, in evaluating the performance of EC and FE based on RE, FE exhibited a superior performance in most environments, but still resulting in an average RE of 0.954. This suggests that EC marginally outperformed FE by 4.650%. For further details, see Table B2.
Additionally, for the PHR trait, using FE displayed a superior performance in most environments, as indicated in Table B2. The best MSE values were observed in 2009 (0.001), 2010 (0.001), and 2013 (0.001). However, exceptions were noted in 2011 and 2012, where the use of EC and NoEC resulted in the best MSE values of 8e-04 and 0.0055, respectively. Furthermore, most RE values comparing NoEC_vs_EC and NoEC_vs_FE techniques were greater than 1. The average RE values of 1.535 (NoEC_vs_EC) and 1.449 (NoEC_vs_FE) indicate significant improvements of 53.530% and 44.930% respectively, compared to the conventional NoEC technique. However, when comparing the performance of the EC versus the FE techniques, the RE values were lower than 1 in most environments, resulting in an average RE of 0.9212. This suggests a general accuracy loss of 7.820% in for using FE compared to using the EC technique (Table B2).
Predictor: E+G+GE+BRR
According to Table B2, the GC trait displayed superior performances with the conventional NoEC technique in most environments, yielding MSE values of 0.004 (2009), 0.002 (2011), and 0.0012 (2013). However, exceptions were found in 2010 and 2012, where FE achieved the best MSE values of 0.0680 and 0.009, respectively. Comparing the RE values between NoEC versus EC and NoEC versus FE techniques showed that most RE values were below 1. Nonetheless, the average RE of 1.104 (NoEC_vs_EC) and 1.189 (NoEC_vs_FE) indicated that EC and FE outperformed the conventional NoEC technique by 10.360% and 18.930%, respectively. Furthermore, when evaluating the performance of EC and FE techniques based on RE, FE presented the best performance in 2009 (1.151), 2010 (1.353), 2011 (2.044), and 2012 (1.0623), while EC outperformed FE in 2013 (0.529). Overall, the average RE 1.228 indicated that FE outperformed EC by 22.800% (Table B2).
Regarding the GY trait, the analysis in Table B2 reveals that the use of EC yielded superior results in most environments (2009 [1333530.864], 2012 [1690390.524], and 2013 [584945.854]). However, exceptions were observed in 2010 and 2011, where the NoEC approach resulted in the best MSE values of 4339466.437 and 1834248.259, respectively. Moreover, most RE values for the comparison of NoEC_vs_EC and NoEC_vs_FE techniques were greater than 1. The average RE values of 1.570 (NoEC_vs_EC) and 1.198 (NoEC_vs_FE) indicate general improvements of 57.030% and 19.790% for the use of EC and FE, respectively, compared to the use of NoEC. However, when comparing the performance of EC and FE techniques based on RE, the FE technique did not outperform EC only in 2010, resulting in an average RE of 0.773. This suggests a general loss of 22.670% accuracy for using FE compared to EC.
Regarding the PH trait, Table B2 shows that the use of FE achieved the best performance in environments 2009 (17.332) and 2011 (22.026), while the use of EC achieved the best performance in environments 2010 (14.9561) and 2013 (11.071). Similarly, most of the RE values for the comparison of NoEC_vs_EC and NoEC_vs_FE techniques were greater than 1. The average RE values of 2.5259 (NoEC_vs_EC) and 2.362 (NoEC_vs_FE) indicate general improvements of 152.590% and 136.210% for using EC and FE, respectively, compared to the conventional NoEC technique. However, when comparing the performance of EC and FE techniques based on RE, EC outperformed FE in most environments, resulting in an average RE of 0.909. This indicates that using EC achieved a 9.100% improvement compared to using FE. For more detailed information, refer to Table 2.
Table B2 displays that using EC yielded the best performance for the PHR trait in most environments, as indicated by the MSE. Specifically, the MSE values were as follows: 2009 (0.001), 2010 (0.001), 2011 (0.001), and 2013 (0.001). However, in 2012, the best MSE values were 0.005, achieved using both EC and NoEC. Comparing NoEC_vs_EC and NoEC_vs_FE techniques, most RE values were at least 1, with average improvements of 60.350% and 48.570% when using EC and FE, respectively, compared to NoEC. Conversely, when comparing EC versus FE techniques, most environments resulted in an average RE of 0.877, indicating a 12.260% decrease in accuracy when using FE compared to EC (Table B2).
USP dataset
Predictor: E+G
Upon examining Table B3, it becomes apparent that the conventional NoEC technique achieved the best performance in terms of MSE in environments Env2 (4.073) and Env3 (5.246). However, exceptions were found in Env1 and Env4, where the optimal MSE values were 3.141 (FE) and 7.814 (EC), respectively. For further detail, refer to Table B3.
Table B3 present our comparison results between the NoEC and EC techniques, assessed through the RE metric. The EC technique displayed its best performance in environments Env1 (1.059) and Env4 (1.046), showcasing improvements of 5.920% and 4.610% over the NoEC technique, respectively. However, NoEC outperformed EC in environments Env2 (0.869) and Env3 (0.831), resulting in an average RE of 0.951. This average RE indicates a general loss of 4.890% in accuracy when using EC compared to NoEC (see Table B3).
In terms MSE for the PH trait, Table B1 shows that the use of FE achieved the best performance in most environments (15.872 [2009], 10.959 [2010], and 164.039 [2012]). However, there were exceptions in 2011 and 2013, where the best MSE values were 28.573 (EC) and 18.363 (NoEC), respectively. On the other hand, when comparing NoEC versus EC and NoEC versus FE techniques using RE, most RE values were greater than 1. On average, the use of EC and FE displayed improvements of 61.570% and 70.210%, respectively, compared to the use of NoEC. Furthermore, when comparing the performance of EC and FE techniques based on RE, FE outperformed EC in all environments, resulting in an average RE of 1.0389. This indicates that using FE surpassed EC by 3.88% (Table B1).
The EC and FE techniques were compared, using the RE metric to assess their performance. The findings indicate that the FE technique achieved its best performance in environments Env1 (1.045) and Env2 (1.048), displaying improvements of 4.480% and 4.790% over EC. However, EC exhibited a slightly better performance in environments Env3 (0.979) and Env4 (0.946), resulting in an average RE of 1.004. This average RE suggests a modest improvement of 0.430% when using FE compared to EC (see Table B3).
Predictor: E+G+GE
Table B3 reveals the performance of the FE technique in terms of MSE across different environments. The FE technique achieved its best performance in environments Env1 (2.789) and Env2 (4.636), although exceptions were found in Env3 and Env4, where the optimal MSE values were 5.833 (NoEC) and 7.792 (EC), respectively (see Table 3).
Table B3 present our comparison results between the NoEC and EC techniques, based on the RE metric. The EC technique displayed its best performance in environments Env1 (1.107) and Env4 (1.120), showing improvements of 10.72% and 12.040% over the NoEC technique. However, the NoEC technique outperformed EC in environments Env2 (0.961) and Env3 (0.925), resulting in an average RE of 1.028. This average RE indicates a general improvement of 2.840% of the EC method regarding the NoEC technique (see Table B3).
The EC and FE techniques were compared, using the RE metric to assess their performance. The findings indicate that the FE technique achieved its best performance in environments Env1 (1.054) and Env2 (1.057), displaying improvements of 5.380% and 5.650% over EC. However, using EC exhibited a better performance in environments Env3 (0.986) and Env4 (0.949), resulting in an average RE of 1.012. This average RE indicates a 1.150% improvement of the FE technique over EC (see Table B3).
Predictor: E+G+BRR
Table B4 presents the results of our analysis regarding the MSE about the FE technique. The FE technique performed best in Env1 (2.859) and Env3 (4.413) environments. However, exceptions were observed in Env2 and Env4, where the optimal MSE values were 4.073 (NoEC) and 5.638 (EC), respectively. For further details, see Table B4.
The results of our comparison between the NoEC and EC techniques, based on the RE metric, are presented in Table B4. The EC technique exhibited its best performance in environments Env1 (1.171) and Env4 (1.450), suggesting improvements of 17.1000% and 45.000%, respectively, compared to the NoEC technique. However, the NoEC technique outperformed EC in environments Env2 (0.823) and Env3 (0.836), resulting in an average RE of 1.070. This average RE indicates a general improvement of 7.000% of the EC regarding the NoEC technique (see Table B4).
We compared the EC and FE techniques, evaluating their performance with the RE metric. The findings indicate that the FE technique achieved its best performance in environments Env1 (1.038) and Env3 (1.423), displaying respective improvements of 3.840% and 42.290% over EC. However, EC performed better in environments Env2 (0.934) and Env4 (0.990), resulting in an average RE of 1.096. This average RE indicates a 9.600% better performance of the FE technique over EC (see Table B4).
Predictor: E+G+GE+BRR
Table B4 presents the performance results of the FE technique in terms of MSE. The best performance was observed in environments Env1 (2.644), Env3 (4.265), and Env4 (5.856). The only exception was Env2, where the optimal MSE value was 4.708, achieved using NoEC. For further information, see Table B4.
Based on the RE metric, the results of our comparison between the NoEC and EC techniques are presented in Table B4. EC performed best in environments Env1 (1.175) and Env4 (1.465), with improvements of 17.510% and 46.530%, respectively, compared to the NoEC technique. However, the NoEC technique outperformed EC in environments Env2 (0.958) and Env3 (0.915), resulting in an average RE of 1.128. This average RE indicates a general improvement of 12.830% of EC regarding NoEC. For more specific information, see Table B4.
We compared the EC and FE techniques based on the RE metric. The analysis revealed that the FE technique displayed its best performance in Env1 (1.047), Env3 (1.494), and Env4 (1.017). These results indicate improvements of 4.740%, 49.430%, and 1.740%, respectively, when compared to using EC. However, EC displayed a better performance in Env2 (0.941), but in general, the FE technique outperformed EC by 12.500%, since an average RE of 1.125 was observed (see Table B4).
G2F_2016 dataset
Predictor: E+G
Table B5 illustrates that FE yielded the best performance for the Grain_Moisture_BLUE trait in most environments. MSE values were 4.645 (DEH1_2016), 2.154 (GAH1_2016), 2.703 (IAH1_2016), 0.467 (IAH4_2016), 0.668 (MOH1_2016), 3.598 (NCH1_2016), 2.092 (NYH2_2016), and 1.601 (WIH2_2016). The average RE values showed that FE outperformed EC and NoEC by 87.970% and 119.370%, respectively. Additionally, EC displayed an average RE improvement of 63.960% over NoEC. For further detail, see Table B5.
For the Grain_Moisture_weight trait, EC presented the best performance based on MSE values in several environments listed in Table 5 (ARH1_2016 [24.235], DEH1_2016 [0.207], IAH1_2016 [2.568], ILH1_2016 [2.172], INH1_2016 [0.210], MOH1_2016 [7.450], OHH1_2016 [0.454] and WIH2_2016 [0.194]). The average RE values revealed that EC and FE outperformed the conventional NoEC technique by 1682.340% and 245.980%, respectively. Furthermore, FE displayed a 58.100% improvement over EC (See Table B5).
Regarding the Yield_Mg_ha_BLUE trait, NoEC displayed a superior performance in most environments based on MSE values listed in Table B5 (GAH1_2016 [3.579], IAH4_2016 [2.576], MIH1_2016 [4.045], MNH1_2016 [1.268], NYH2_2016 [16.252], OHH1_2016 [1.830] and WIH1_2016 (3.665]). The average RE values indicated that FE resulted in general improvements of 21.030% and 1.400% over EC and NoEC, respectively. However, a comparison between NoEC and EC showed a slight decrease of 0.190% in average RE for EC (see Table B5).
For the Yield_Mg_ha_weight trait, NoEC showed the best performance based on MSE values in most environments (DEH1_2016 [0.078], IAH4_2016 [0.091], ILH1_2016 [0.351], MIH1_2016 [0.1156], MNH1_2016 [0.391], NYH2_2016 [0.087], WIH1_2016 [0.063] and WIH2_2016 [0.019]). The average RE values indicated general improvements of 52.860% and 22.630% for EC and FE, respectively, compared to NoEC. Moreover, on average, FE outperformed EC by 89.600% (see Table B5).
Predictor: E+G+GE
Table B5 shows that FE yielded the best performance for the Grain_Moisture_BLUE trait in the majority of environments, with MSE values ranging from 0.519 to 5.813 (IAH4_2016, ILH1_2016, MNH1_2016, NEH1_2016, NYH2_2016, OHH1_2016 and WIH1_2016). Comparing RE values, using FE outperformed EC and NoEC techniques by 42.480% and 114.740%, respectively. Additionally, EC outperformed NoEC with an average RE of 1.552, indicating a superiority of 55.210% for EC. For further details, see Table B5.
For the Grain_Moisture_weight trait, Table B5 reveals that FE displayed a better performance in most environments, as indicated by the MSE values (DEH1_2016 [0.132], IAH3_2016 [0.418], IAH4_2016 [139.446], MIH1_2016 [1.668], MNH1_2016 [1.316], NCH1_2016 [6.953], NYH2_2016 [5.565], OHH1_2016 [0.195] and WIH1_2016 [1.508]). Moreover, the average RE values showed that FE outperformed EC and NoEC by 831.910% and 825.260%, respectively. Comparing NoEC and EC techniques, there was a general improvement of 357.000% for EC over NoEC, with an average RE of 4.570 (see Table B5).
Regarding the Yield_Mg_ha_BLUE trait, Table B5 shows that the use of NoEC achieved the best performance in most environments, as indicated by the MSE values (GAH1_2016 [3.379], IAH1_2016 [2.287], IAH2_2016 [7.505], IAH4_2016 [3.565], MIH1_2016 [4.748], NYH2_2016 [17.271], WIH1_2016 [2.210] and WIH2_2016 [4.667]). However, most RE values comparing NoEC_vs_EC and NoEC_vs_FENoEC_vs_FEtechniques were greater than 1. On average, EC displayed a 7.450% improvement and FE showed an 11.690% improvement compared to the conventional NoEC technique. Furthermore, comparing EC and FE techniques, an average RE of 1.227 was observed, indicating that FE outperformed NoEC by 22.700% (see Table B5).
In terms of the Yield_Mg_ha_weight trait, Table B5 shows that the use of NoEC achieved the best performance in most environments, as evident from the MSE values (DEH1_2016 [0.051], GAH1_2016 [0.026], IAH1_2016 [2.914], IAH2_2016 [0.0689], MIH1_2016 [0.055], MNH1_2016 [0.146], NEH1_2016 [0.033], NYH2_2016 [0.449] and OHH1_2016 [1.202]). The average RE values indicated slight losses of 2.210% and 2.570% when comparing EC versus NoEC and FE versus NoEC, respectively. This implies that EC and FE techniques did not perform as adequately as the conventional NoEC technique. However, comparing EC and FE techniques based on RE showed that FE outperformed EC in most environments, resulting in an average RE of 1.339, indicating a 33.930% superiority of FE over EC. For more detailed information, see Table B5.
Predictor: E+G+BRR
In Table B6, it is evident that for the Grain_Moisture_BLUE trait, the use of FE provided the best performance in most environments, as indicated by the MSE values (DEH1_2016 [4.376], GAH1_2016 [2.002], IAH1_2016 [2.036], IAH3_2016 [1.237], IAH4_2016 [0.496], MNH1_2016 [3.685], MOH1_2016 [0.678], NCH1_2016 [3.499], NYH2_2016 [2.213] and WIH2_2016 [1.648]). On average, the RE values indicate that FE outperformed EC and NoEC by 67.090% and 129.850%, respectively. Additionally, comparing NoEC and EC techniques showed that EC outperformed NoEC by an average of 84.880%. For further information, see Table B6.
For the Grain_Moisture_weight trait, Table B6 shows that the use of NoEC provided the best performance in most environments, as indicated by the MSE values (GAH1_2016 [1.272], IAH1_2016 [401.574], IAH3_2016 [0.199], ILH1_2016 [5.447], MIH1_2016 [0.715[, NCH1_2016 [1.174] and NEH1_2016 [42.758]). On average, the RE values indicate that FE outperformed EC and NoEC by 167.270% and 25.410%, respectively. Furthermore, comparing NoEC and EC shows that EC outperformed NoEC with an average RE of 3.495, representing a general improvement of 149.510%. For more detailed information, see Table B6.
Table B6, for the Yield_Mg_ha_BLUE trait, shows that the use of NoEC led to the best performance in most environments, as indicated by the MSE values (ARH1_2016 [3.713], GAH1_2016 [3.579], IAH4_2016 [2.576], INH1_2016 [2016], MIH1_2016 [4.045], MNH1_2016 [1.268], NYH2_2016 [16.252], OHH1_2016 [1.829] and WIH2_2016 [4.629]). On average, the RE values indicate general improvements of 10.650% for FE compared to EC, and 3.780% for EC compared to NoEC. However, when comparing the performance of NoEC and FE techniques, an average RE of 0.986 indicates a slight loss for FE compared to NoEC. For more detailed information, see Table B6.
For the Yield_Mg_ha_weight trait, the use of NoEC achieved the best performance in most environments, as indicated by the MSE values (ARH1_2016 [0.989], GAH1_3016 [0.035], IAH2_2016 [0.175], IAH3_2016 [0.783], MIH1_2016 [0.1201], MHH1_2016 [0.393], MOH1_2016 [0.232], NYH2_2016 [0.402], OHH1_2016 [0.533] and WIH2_2016 [0.055]). On average, the RE values indicate a general improvement of 51.630% for EC compared to NoEC and 3.960% for FE compared to EC. However, when comparing the performance of NoEC and FE based on RE, the best performance was displayed by NoEC in most environments, resulting in an average RE of 0.9012, indicating that NoEC outperformed FE by 9.820%. For more detailed information, see Table B6.
Predictor: E+G+GE+BRR
Table B6 shows that EC yielded the most favorable results for the Grain_Moisture_BLUE trait in various environments. The corresponding MSE values for EC were 4.2801 (DEH1_2016), 0.519 (IAH4_2016), 4.964 (ILH1_2016), 4.762 (MNH1_2016), 6.047 (NEH1_2016), 4.030 (NYH2_2016), 3.072 (OHH1_2016), and 3.495 (WIH1_2016). Additionally, the average RE values indicated that using FE outperformed both EC and NoEC by 29.090% and 125.150%, respectively (1.291 for EC_vs_FE, and 2.252 for NoEC_vs_FE). Furthermore, when comparing the NoEC and EC techniques, an average RE of 1.6512 displays the superior performance of EC over NoEC by 65.180%. For more comprehensive information, see Table B6.
When considering the Grain_Moisture_weight trait, the use of EC presented amor adequate performance in most environments based on the MSE values provided in Table 6 (DEH1_2016 [0.595], IAH1_2016 [360.363], IAH2_2016 [1.219], IAH4_2016 [120.354], ILH1_2016 [10.357], NCH1_2016 [2.0245], NYH2_2016 [2.534], and OHH1_2016 [4.124]). Moreover, the average RE values reveal that EC and FE outperformed the conventional NoEC by 408.510% and 240.900% respectively (5.085 for NoEC_vs_EC and 3.409 for NoEC_vs_FE). Furthermore, a comparison between EC and FE techniques indicates that an average RE of 7.899 suggests that FE outperformed EC by 689.960%. For more detailed information, see Table B6.
When examining the Yield_Mg_ha_BLUE trait, the use of NoEC displayed the best performance in most environments based on the MSE values presented in Table B6 (ARH1_2016 [3.928], GAH1_2016 [3.379], IAH1_2016 [2.287], IAH2_2016 [7.505], IAH4_2016 [2.565], MIH1_2016 [4.748], NYH2_2016 [17.271], WIH1_2016 [2.210] and WIH2_2016 [4.667]). However, it is worth noting that EC and FE outperformed the conventional NoEC by 11.690% and 9.490% in terms of average RE values (1.117 for NoEC_vs_EC and 1.095 for NoEC_vs_FE). Nevertheless, when comparing FE versus EC techniques, a slight loss of 0.400% was observed for using FE compared to EC, as indicated by an average RE of 0.996. For more detailed information, see Table B6.
Regarding the Yield_Mg_ha_weight trait, Table B6 shows that the use of EC yielded the best performance in most environments, as evidenced by the following MSE values: ARH1_2016 (0.719), DEH1_2016 (0.020), IAH1_2016 (2.808), IAH4_2016 (0.058), ILH1_2016 (0.594), NYH2_2016 (0.418), and WIH2_2016 (0.073). The average RE values indicated improvements of 22.200% (NoEC_vs_EC) and 11.380% (NoEC_vs_FE), highlighting the superior performance of EC and FE over the conventional NoEC technique. Conversely, when comparing EC and FE techniques, most environments performed better with an average RE of 1.384, indicating that FE outperformed EC by 38.420%. For additional information, see Table B6.
Keywords: genomic selection, plant breeding, environmental covariates, feature engineering, feature selection
Citation: Montesinos-López OA, Crespo-Herrera L, Pierre CS, Cano-Paez B, Huerta-Prado GI, Mosqueda-González BA, Ramos-Pulido S, Gerard G, Alnowibet K, Fritsche-Neto R, Montesinos-López A and Crossa J (2024) Feature engineering of environmental covariates improves plant genomic-enabled prediction. Front. Plant Sci. 15:1349569. doi: 10.3389/fpls.2024.1349569
Received: 04 December 2023; Accepted: 11 April 2024;
Published: 15 May 2024.
Edited by:
Huihui Li, Chinese Academy of Agricultural Sciences, ChinaReviewed by:
Tingxi Yu, Chinese Academy of Agricultural Sciences (CAAS), ChinaJoão Ricardo Bachega Feijó Rosa, RB Genetics & Statistics Consulting, Brazil
Copyright © 2024 Montesinos-López, Crespo-Herrera, Pierre, Cano-Paez, Huerta-Prado, Mosqueda-González, Ramos-Pulido, Gerard, Alnowibet, Fritsche-Neto, Montesinos-López and Crossa. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Abelardo Montesinos-López, YWJlbGFyZG8ubW9udGVzaW5vczAyMzNAYWNhZGVtaWNvcy51ZGcubXg=; José Crossa, ai5jcm9zc2FAY2dpYXIub3Jn
†ORCID: José Crossa, orcid.org/0000-0001-9429-5855