Skip to main content

BRIEF RESEARCH REPORT article

Front. Genet., 24 November 2020
Sec. Computational Genomics

A Deep Learning Approach to Population Structure Inference in Inbred Lines of Maize

  • 1Department of Computer Sciences and Industries, Catholic University of the Maule, Talca, Chile
  • 2Instituto de Ciencias Agroalimentarias, Animales y Ambientales, Universidad de O’Higgins, San Fernando, Chile
  • 3Institute of Biological Sciences, University of Talca, Talca, Chile
  • 4Departamento de Agronomia, Universidade Estadual de Maringá, Maringá, Brazil

Analysis of population genetic variation and structure is a common practice for genome-wide studies, including association mapping, ecology, and evolution studies in several crop species. In this study, machine learning (ML) clustering methods, K-means (KM), and hierarchical clustering (HC), in combination with non-linear and linear dimensionality reduction techniques, deep autoencoder (DeepAE) and principal component analysis (PCA), were used to infer population structure and individual assignment of maize inbred lines, i.e., dent field corn (n = 97) and popcorn (n = 86). The results revealed that the HC method in combination with DeepAE-based data preprocessing (DeepAE-HC) was the most effective method to assign individuals to clusters (with 96% of correct individual assignments), whereas DeepAE-KM, PCA-HC, and PCA-KM were assigned correctly 92, 89, and 81% of the lines, respectively. These findings were consistent with both Silhouette Coefficient (SC) and Davies–Bouldin validation indexes. Notably, DeepAE-HC also had better accuracy than the Bayesian clustering method implemented in InStruct. The results of this study showed that deep learning (DL)-based dimensional reduction combined with ML clustering methods is a useful tool to determine genetically differentiated groups and to assign individuals into subpopulations in genome-wide studies without having to consider previous genetic assumptions.

Introduction

Analysis of population structure and genetic variation is a common practice in genome-wide studies and is an important guideline to understand and infer the evolutionary processes and the demographic history in ecological and evolutionary studies (Stift et al., 2019). Knowledge of the population genetic structure is very helpful in many applications, which plays an important role for breeding purposes and selection strategies. In this sense, high-throughput DNA sequencing technologies have allowed the generation of large sets of genomic data in diverse populations routinely (Ho et al., 2019), which has been used to study patterns of genetic variation across the genome and to characterize the evolutionary forces in different plant species (Padhukasahasram, 2014; Ho et al., 2019). For instance, markers based on single-nucleotide polymorphisms (SNPs) have provided a rapid way of delineating genetic structure and of understanding the basis of the taxonomic discrimination, providing novel information such as founder effects, bottlenecks, evolutionary relationships, and migration history of natural populations (Padhukasahasram, 2014; Shultz et al., 2016).

Population structure analysis is a major area of interest within the field of genetics and bioinformatics (Alhusain and Hafez, 2018). In this sense, several bioinformatics methods have been developed to examine the population structure in genetically diverse plant germplasm based on high-throughput genomic data. Among the methods currently available, the Bayesian clustering algorithm developed by Pritchard et al. (2000) (i.e., STRUCTURE) is one of the most widely used population analysis tools, which allows researchers to infer population structure patterns in sample sets (Porras-Hurtado et al., 2013). The underlying genetic model of this algorithm assumes that populations are in Hardy–Weinberg equilibrium (Pritchard et al., 2000), which is not met, for instance, in populations with high levels of inbreeding. In this sense, Gao et al. (2007) proposed an extension to the STRUCTURE algorithm denominated InStruct, which eliminates the assumption of Hardy–Weinberg equilibrium within populations and takes inbreeding or selfing into account. This method applies a Bayesian inference to simultaneously assign individuals into subpopulations but can be very time-consuming. Another successful approach to infer population structure has been implemented in the ADMIXTURE software (Alexander et al., 2009; Alexander and Lange, 2011), a maximum-likelihood-based method that updates the log-likelihood as it converges on a solution for the ancestry proportions and allele frequencies that maximize the likelihood function (Alexander and Lange, 2011). Other authors have emphasized the use of non-parametric methods such as K-means (KM) and hierarchical clustering (HC) (Bouaziz et al., 2012; Meirmans, 2012; Alhusain and Hafez, 2018). KM and HC approaches correspond to machine learning (ML) methods that do not require the assumptions of the Hardy–Weinberg principle and use external dimension reduction techniques, such as principal component analysis (PCA) (Kobak and Berens, 2019), commonly used in several data-intensive biological fields. KM is an iterative descent algorithm that minimizes the within-cluster sum of squares (Meirmans, 2012). On the other hand, the HC method allows the formation of genetic groups to be mutually exclusive, in which each cluster is distinct from each other, and the members of each cluster are similar with respect to the input information (Ward, 1963). Stift et al. (2019) found that ADMIXTURE and KM were computationally faster than STRUCTURE; however, ADMIXTURE had less power to detect structure compared to STRUCTURE and KM clustering.

The analytical Bayesian inference-based methods (STRUCTURE and InStruct) and the most traditional ML algorithms require that the data provided need to be of numerical type (Pritchard et al., 2000; Yokota and Wu, 2018). Label encoder (LE) is a useful method to help normalize labels so that they can transform non-numerical values into numerical values (Joshi et al., 2016). In genomic data, for instance, Agajanian et al. (2019) used LE to assign to each nucleotide a unique numeric data value. Other ML methods [e.g., deep autoencoder (DeepAE), a likelihood-free inference framework] consider a framework in which the information of the input variables is compressed and subsequently reconstruct the input data, minimizing the loss function. In this sense, the deep learning (DL) approach is a class of neural networks and has been an active area of ML research, emerging as a powerful tool in genetics and genomics studies, e.g., schizophrenia classification through datasets of SNP and functional magnetic resonance imaging (Li et al., 2020), gene expression prediction from SNP genotypes (Xie et al., 2017), MADS-box gene classification system for angiosperms (Chen et al., 2019), and RNA secondary structure prediction (Zhang et al., 2019) and to predict quantitative phenotypes from SNPs (Liu et al., 2019). Unlike the traditional artificial neural network, DL algorithms consider many hidden layers during the network training (Xie et al., 2016). The advantages of the DL approach have been well described by Qu et al. (2019) and can be summarized as the capacity of (1) learning from data without prediction features, (2) learning from increasingly large and high-dimensional datasets, and (3) capturing non-linear dependencies in genetic sequences. Therefore, in this study, a genome-wide data assessment of maize inbred lines was performed using the DL (DeepAE) approach, combined with ML methods (HC and KM) and a Bayesian clustering approach (InStruct), to infer population genetic structure and assign individuals into each subpopulation. A better understanding of the use of these novel methods could provide recommendations for genetic diversity and differentiation studies.

Materials and Methods

The step-by-step description of the proposed methodology for the clustering of populations, through the use of DL and ML, is illustrated in Figure 1. The respective codes are available in Supplementary Data Sheet S1. The first step is to preprocess SNP dataset and apply DeepAE with two layers in both the encoder and decoder, without considering the input and output. The second step is based on applying ML clustering algorithms in an unsupervised way based on the data obtained from the DeepAE in order to group and identify subpopulations.

FIGURE 1
www.frontiersin.org

Figure 1. Overview of the deep autoencoder (DeepAE) system and population clustering analysis from a large single-nucleotide polymorphism (SNP) dataset. (A) Schematic representation of the entity system. Input data are composed of a matrix of n samples (rows) and k SNPs (columns). Through autoencoder, categorical variables, i.e., the SNPs and samples, were represented as an m-dimensional vector. (B) Samples are projected into m dimensional sample entity vector space. The DeepAE learned the feature of the sample solely from input matrix, such that similar samples are clustered in close proximity. (C) The t-distributed stochastic neighbor embedding (t-SNE) representation of sample entity matrix from the SNP dataset transformed by DeepAE. The 183 inbred lines are labeled with a different color according to subpopulations of maize (dent corn and popcorn). (D) Population clustering analysis through unsupervised clustering techniques.

Genotyping and Data Processing

These inbred lines correspond to a panel of 183 maize genotypes from the Department of Agronomy of the State University of Maringá, which consist of 97 dent field corn and 86 popcorn genotypes (for more details, see Supplementary Data Sheet S2). Seedlings were grown in a growth chamber at 27/20°C day/night temperatures and a 12-h photoperiod. The youngest leaves of five plants were sampled from each genotype ∼30 days after germination. The DNA samples were sent to the Genomic Diversity Institute of Cornell University for SNP discovery via genotyping-by-sequencing (GBS), which is described in Elshire et al. (2011). The TASSEL 5.2 software (Bradbury et al., 2007) was used to align the raw data of GBS with the Zea mays version AGPV3 reference genome (B73 RefGen v3), resulting in a total of 1,014,070 SNPs. Subsequently, these SNPs were filtered through TASSEL considering a minor allele frequency > 0.15 and the absence of missing data, yielding a final subset of 4,812 SNPs (distributed on all chromosomes).

Dimensionality Reduction Methods

Unsupervised Learning Using Deep Autoencoder

DeepAE was applied to find a mathematical representation of SNPs and to reduce the dimensionality of the dataset. This architecture contains multiple encoding and decoding stages made up of a sequence of encoding layers followed by a stack of decoding layers. First, the SNPs were encoded to a numerical representation through one hot encoding process, as follows: A: [1,0,0,0], T: [0,1,0,0], G: [0,0,1,0], C: [0,0,0,1]. The depth of the network was varied from one to four hidden layers in order to minimize the loss function (cross-entropy function; Supplementary Data Sheet S3), in which the best results were obtained considering two hidden layers. Therefore, DeepAE was performed considering the following parameters: an entrance of 4,812 features corresponding to SNP markers (represented by one hot encode), two hidden layers with 2,000 and 700 neurons, respectively, in both the encoder and decoder, a bottleneck hidden layer of 40 neurons, and a learning rate of 0.001. Details about DeepAE are shown in Supplementary Data Sheet S3. The Adam optimizer was used to minimize the loss function (cross-entropy function). The rectified linear unit (ReLU) was used as the activation function. DeepAE was implemented in python 3.7 language using the libraries Keras 2.2.4 and TensorFlow 1.14.0.

Principal Component Analysis

The PCA describes the variation of a dataset in terms of a set of uncorrelated variables, where each of these is a linear combination of the original variables. These new variables are sorted in descending order of importance, where the first variable (or first principal component) accounts for a majority of the variation in the original data, and the following variables account for a large amount of the remaining variation of the data that is not correlated with the previous variables. The PCA was performed in TASSEL 5.2 (Bradbury et al., 2007).

Visualization of Reduced Data by Deep Autoencoder

The t-distributed stochastic neighbor embedding (t-SNE) is a technique that allows visualization of high-dimension data giving each data point a location in a low dimension (Maaten and Hinton, 2008). This method maps the different high-dimension instances into new low-dimension instances keeping up the similarities found in the original data. Encoded SNPs with DeepAE were visualized by two-dimensional t-SNE implemented with perplexity = 30, iterations equal to 1,000, and a learning rate of 200.

Clustering Analysis

Three types of unsupervised clustering algorithms were applied: KM (Macqueen, 1967), HC (Abbas, 2008), and InStruct (Gao et al., 2007). Details about these three clustering algorithms are shown in Supplementary Data Sheet S3. The entrance for these methods corresponds, on one hand, to the SNP genomic data represented with LE and, on the other hand, to the encoded SNP data with dimensionality reduction techniques: DeepAE and PCA. Specifically, LE is a numerical representation to transform non-numerical labels to numerical labels (Joshi et al., 2016); in this case, the SNP markers were processed as follows: A:[0], T:[1], G:[2], C:[3]. The genomic data represented by LE and the dimension reduction techniques were used as inputs in the ML clustering methods, while the Bayesian method only used the dataset codified with LE (Supplementary Figure S4). The optimal number of clusters was determined by two validation metrics: Silhouette coefficient (SC; Rousseeuw, 1987) and Davies–Bouldin index (DBI; Davies and Bouldin, 1979) for the ML-based clustering algorithms (details about evaluation metrics are shown in Supplementary Data Sheet S3). On the other hand, the optimal number of clusters (K) in Bayesian-based clustering algorithm (InStruct) was determined with the highest ΔK method, as proposed by Evanno et al. (2005), and the lowest value of deviance information criterion (DIC) (Gao et al., 2007).

Results

Population Clustering Analysis With K-Means, Hierarchical Clustering, and InStruct

The results of clustering analysis with KM and HC varied depending on the data preprocessing algorithms being studied (i.e., LE, DeepAE, and PCA). The results of KM and HC methods showed that LE was less accurate than DeepAE and PCA according to SC and DBI measures (Table 1). In fact, these validation indexes (SC and DBI) were ∼10 and ∼8 times higher for DeepAE and PCA, respectively, than LE, when K = 2. The SC values showed that the reliability of clusters generated by the three data preprocessing algorithms decreases as the amount of K clusters increases (Table 1), achieving the best accuracy measures for K = 2. Consistently, in the three preprocessing algorithms, the DBI was higher when the number of clusters increased. Moreover, KM and HC in combination with DeepAE and PCA showed the best results in terms of accuracy when K = 2. The high values of SC obtained for PCA and DeepAE in combination with both clustering methods (close to 1; Table 1) indicate that an inbred line is well matched to its own genetic cluster and poorly matched to the neighboring group or subpopulation. On the other hand, the SC value for LE in combination with HC (LE-HC) was close to zero and achieved the same value for K = 2, 4, and 5, while DBI for LE-HC revealed that the optimal number of clusters was K = 4. According to these results, through the classical representation (LE), it was not possible to achieve a consistent clustering performance with neither ML clustering method (KM and HC), thus this representation was discarded from the posterior cross-tab analysis. In the case of DeepAE or PCA, it was possible to achieve the optimal number of clusters.

TABLE 1
www.frontiersin.org

Table 1. Validation indexes for the optimal number of clusters (K) according to Silhouette coefficient (SC) and Davies–Bouldin index (DBI).

The Bayesian clustering analysis with InStruct indicated that the 183 inbred lines were grouped into two clusters (K = 2) according to the lowest DIC and the highest second-order change rate of the probability function with respect to K (ΔK). This result was expected, since the inbred lines come from two well-defined maize subpopulations (i.e., dent corn and popcorn), which was confirmed by DIC and ΔK values obtained from InStruct and both SC and DBI validation measures in the clusters formed by both ML clustering methods in combination with DeepAE and PCA (Table 1). In this study, the majority of the dent corn lines were grouped in cluster 1, whereas the majority of popcorn lines were assigned to cluster 2.

A simple cross-tab analysis was performed to evaluate the ability of clustering and preprocessing methods to assign individuals to their putative subpopulation (i.e., dent corn and popcorn). The results of this analysis are shown in Table 2 for KM and HC ML clustering algorithms in combination with DeepAE and PCA dimension reduction algorithms and the Bayesian clustering method implemented in InStruct. DeepAE combined with both HC (DeepAE-HC) and KM (DeepAE-KM) methods grouped the smallest amount of popcorn lines within cluster 1 (which should be composed of only dent corn lines). The Bayesian approach implemented in InStruct grouped 17 popcorn lines within cluster 1, while PCA combined with HC (PCA-HC) and KM (PCA-KM) grouped the greatest amount of popcorn lines together with dent corn lines (cluster 1) (Table 2). Interestingly, the SC validation index was higher in DeepAE than PCA (for both clustering methods), which implies that average within-cluster distances were low (high compactness in clusters), whereas between-cluster distances were high (high separation between clusters). It should be noted that the Bayesian clustering method (InStruct) and ML clustering algorithms (HC and KM) grouped coincidently eight popcorn lines into cluster 1, i.e., the cluster containing dent corn lines. Overall, DeepAE-HC was the most effective method to assign individuals to the clusters (96% of correct individual assignments), whereas DeepAE-KM, PCA-HC, PCA-KM, and InStruct assigned correctly 92, 89, 81, and 91%, respectively, of the lines into their respective clusters (Table 2).

TABLE 2
www.frontiersin.org

Table 2. Cross-tab analysis among subpopulations of maize (popcorn and dent corn) and clusters predicted by machine learning (ML) clustering methods in combination with dimensionality reduction techniques.

Visualization of the Genetic Structure With t-Distributed Stochastic Neighbor Embedding and Principal Component Analysis

Figure 2 shows the visualization with t-SNE (for DeepAE) and the PCA representation for SNP data. t-SNE and PCA clearly separated the inbred lines into two clusters, which correspond to the subpopulations of popcorn (blue) and dent corn (orange) (Figure 2). The t-SNE method allowed to define the clusters with any of its dimensions (1 and 2), while the PCA only separated the clusters with dimension 1. Moreover, t-SNE clustered the individuals of each putative subpopulation closer together than the PCA method. In this sense, t-SNE preserved the local structure (more than the larger-scale structure) of data by matching pairwise similarity distributions in the higher-dimensional space (original data) in a lower-dimensional projected space (Chan et al., 2018), and thus, as opposed to PCA, t-SNE grouped the samples in a low-dimensional space, while keeping the distributions of original data space.

FIGURE 2
www.frontiersin.org

Figure 2. Visualization of the genetic structure of inbred lines of maize with t-distributed stochastic neighbor embedding (t-SNE) for the DeepAE (A) and representation of principal component analysis (PCA) (B) for single-nucleotide polymorphism (SNP) data. The colored dots orange and blue represent different inbred lines from the two subpopulations of maize (dent corn and popcorn, respectively).

Discussion

Analysis of population genetic structure is a major area of interest within the field of genetics and bioinformatics, which is a common practice in genome-wide studies, including association mapping, ecology, and evolution studies in crop species such as maize (Li et al., 2019; Mafra et al., 2019; Maldonado et al., 2019; Wang et al., 2019). The present study proposed an ML-based analysis of population structure and individual assignment usually performed in several data-intensive biological fields. According to the results, HC in combination with the two data preprocessing algorithms (DeepAE and PCA) presented higher accuracy in assigning maize lines to their respective clusters as compared to KM. These findings agree with a previous study by Kaur and Kaur (2013), who reported that the hierarchical algorithm provides better results and higher quality than KM. Additionally, the results of this study were consistent with the findings of previous research, indicating that dent corn and popcorn lines from Brazilian germplasms are grouped into two genetically differentiated clusters (Coan et al., 2018; Camacho et al., 2019; Maldonado et al., 2019; Senhorinho et al., 2019). In this regard, the results of this study showed that the DeepAE-based data preprocessing had better accuracy values than those achieved by PCA. In this sense, PCA-HC and PCA-KM had a high number of lines incorrectly assigned to cluster 1 (20 and 34, respectively). On the other hand, InStruct showed an accuracy value lower than DeepAE-HC and DeepAE-KM when assigning maize lines. In this sense, Stift et al. (2019) found that InStruct revealed more inconsistency than KM in the clustering results, which was derived from a lack of convergence across replicate runs of the algorithm.

The conventional clustering methods, e.g., self-organizing map algorithm (Kohonen and Somervuo, 1998), Gaussian mixture models (Reynolds, 2015), KM (Arthur and Vassilvitskii, 2007), and HC (Bouaziz et al., 2012), usually have poor clustering performance on high-dimensional data due to high computational complexity (Min et al., 2018). For this reason, dimensionality reduction methods have been widely studied to represent the raw data into a low-dimensional space to ensure that the data are easier to separate when using clustering methods. The most popular methods for dimensionality reduction include linear transformation with PCA and non-linear transformation with autoencoder (Vincent et al., 2008; Chazan et al., 2019; Kobak and Berens, 2019). However, the non-linear nature of an autoencoder has been shown to reconstruct complex data more accurately than PCA (Xie et al., 2016). Sakaue et al. (2020) pointed out that the main linear technique for dimensionality reduction, PCA, was not sufficient to fully capture the fine and subtle genomic structure within a Japanese population (n = 169,719), while non-linear dimensionality reduction methods (t-SNE and uniform manifold approximation and projection) could detect a fine and discrete population structure with a high resolution. Tan et al. (2014) showed that the use of denoising autoencoders was efficient to identify and extract complex patterns from a large collection of breast cancer gene expression data, which allowed for successfully constructed features that contain both clinical and molecular information. In this sense, Yue and Wang (2018) pointed out denoising autoencoders are effective in extracting biological insights, since the reconstruction loss of autoencoder ensures that the network learns a feasible representation and avoids obtaining trivial solutions. On the other hand, Almotiri et al. (2017) showed that the non-linear autoencoder achieved better accuracy than the linear PCA method in the classification of handwritten numerals. In accordance with the present study, Manning-Dahan (2018) observed that autoencoder had an accuracy 68% higher than PCA, with much less false positives found in the classification of images. This author also pointed out that PCA creates linear maps and, thus, is limited to learn linear relationships between variables, whereas autoencoders can be used for encoding and decoding large datasets with the flexibility of learning both non-linear and linear mappings. Xie et al. (2016) also pointed out that KM has a better performance when it is employed on a set of data preprocessed by autoencoder than when they have not been preprocessed. Therefore, a key aspect of the methodology proposed in this study is the correct mathematical representation of the SNP dataset, which is not achieved with a classical technique such as LE or PCA but is achieved through the implementation of more complex techniques, such as DeepAE.

Artificial neural network models have been used before in order to genetically evaluate crop germplasm collections, such as maize (Ferreira et al., 2018; Kulka et al., 2018) and grapevine (Costa et al., 2020), in which the clustering analysis was based upon competitive learning-based neural networks. This alternative method was able to analyze population structure based on not only bi-allelic but also multi-allelic data (Peña-Malavera et al., 2014; Ferreira et al., 2018) and has been demonstrated to be computationally faster than MCMC methods (Nikolic et al., 2009) and does not consider the assumption of Hardy–Weinberg equilibrium in the population being studied (Ferreira et al., 2018). In this sense, the ML algorithms required less time for its run as compared to InStruct. ML clustering algorithms in combination with DeepAE or PCA dimension reduction algorithms require approximately 3 s for execution, while with InStruct, the time required was about 3 weeks. InStruct is based on the Markov chain method for parameter estimation, which is computationally time-consuming with respect to other unsupervised clustering methods (Gao et al., 2007; Stift et al., 2019). It should be noted that the artificial neural networks have the advantage of being a non-parametric method, which does not require detailed information about the physical processes being modeled and is able to analyze data containing missing data (Azevedo et al., 2015; Costa et al., 2020). Interestingly, our results confirm that DeepAE neural networks provide precise results in the identification of genetically differentiated groups and the assignment of lines into subpopulations (Table 2).

On the other hand, the t-SNE algorithm, in combination with DeepAE data, was able to visually identify both subpopulations of maize (Figure 2). This is of great relevance since it is the starting point in the unsupervised clustering algorithms and identifying clusters. Kobak and Berens (2019) pointed out that when applied to high-dimensional but well-clustered data, t-SNE tends to produce a visualization with distinctly isolated clusters, which often are in agreement with the clusters found by a dedicated clustering algorithm. Also, these authors mentioned that the combination of t-SNE with a variational autoencoder better preserves the global structure of the data and produces more interpretable visualizations than standard t-SNE. In this sense, this study found that t-SNE was better than PCA in preserving the local structure by grouping genotypes of each putative subpopulation closer together. Moreover, t-SNE could separate the subpopulations with any dimension (1 and 2), while a PCA separated the subpopulations only with the first dimension.

Finally, the use of the novel dimensionality reduction method, DL, combined with ML clustering methods allowed to assign popcorn and dent corn lines into their respective maize subpopulations. This analytical methodology can be applied to uncover the genetic structure in diverse populations worldwide, without having to consider previous genetic assumptions such as Hardy–Weinberg and linkage disequilibrium.

Data Availability Statement

Publicly available datasets were analyzed in this study. These data can be found here: https://doi.org/10.6084/m9.figshare.12934913.

Author Contributions

FM-P, CM, CS, and XL-C conducted and designed this study. CM, XL-C, and FM implemented the database and web application. FM-P, CM, and CS performed the data curation. FM-P, CM, and XL-C wrote the manuscript. All authors reviewed and approved the manuscript for publication.

Funding

The authors are grateful to the Chilean National Fund for Scientific and Technological Development (FONDECYT) grant number 1201973 and the Program of International Cooperation (PCI-CONICYT) grant number REDI170172.

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Supplementary Material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fgene.2020.543459/full#supplementary-material

References

Abbas, O. (2008). Comparisons between data clustering algorithms. Int. Arab J. Inf. Technol. 5, 320–325.

Google Scholar

Agajanian, S., Oluyemi, O., and Verkhivker, G. M. (2019). Integration of random forest classifiers and deep convolutional neural networks for classification and biomolecular modeling of cancer driver mutations. Front. Mol. Biosci. 6:44. doi: 10.3389/fmolb.2019.00044

PubMed Abstract | CrossRef Full Text | Google Scholar

Alexander, D. H., and Lange, K. (2011). Enhancements to the ADMIXTURE algorithm for individual ancestry estimation. BMC Bioinform. 12:246. doi: 10.1186/1471-2105-12-246

PubMed Abstract | CrossRef Full Text | Google Scholar

Alexander, D. H., Novembre, J., and Lange, K. (2009). Fast model-based estimation of ancestry in unrelated individuals. Genome Res. 19, 1655–1664. doi: 10.1101/gr.094052.109

PubMed Abstract | CrossRef Full Text | Google Scholar

Alhusain, L., and Hafez, A. M. (2018). Nonparametric approaches for population structure analysis. Hum. Genet. 12:25.

Google Scholar

Almotiri, J., Elleithy, K., and Elleithy, A. (2017). “Comparison of autoencoder and principal component analysis followed by neural network for e-learning using handwritten recognition,” in Proceedings of the 2017 IEEE Long Island Systems, Applications and Technology Conference (LISAT), Farmingdale, NY.

Google Scholar

Arthur, D., and Vassilvitskii, S. (2007). “K-Means++: the advantages of careful seeding,” in Proceedings of the Eighteenth Annual ACM-SIAM Symposium on Discrete Algorithms, SODA 2007, New Orleans, LA.

Google Scholar

Azevedo, A. M., Andrade Júnior, V. C. D., Pedrosa, C. E., Oliveira, C. M. D., Dornas, M. F. S., Cruz, C. D., et al. (2015). Application of artificial neural networks in indirect selection: a case study on the breeding of lettuce. Bragantia 74, 387–393. doi: 10.1590/1678-4499.0088

CrossRef Full Text | Google Scholar

Bouaziz, M., Paccard, C., Guedj, M., and Ambroise, C. (2012). SHIPS: spectral hierarchical clustering for the inference of population structure in genetic studies. PLoS One 7:e45685. doi: 10.1371/journal.pone.0045685

PubMed Abstract | CrossRef Full Text | Google Scholar

Bradbury, P. J., Zhang, Z., Kroon, D. E., Casstevens, T. M., Ramdoss, Y., and Buckler, E. S. (2007). TASSEL: software for association mapping of complex traits in diverse samples. Bioinformatics 23, 2633–2635. doi: 10.1093/bioinformatics/btm308

PubMed Abstract | CrossRef Full Text | Google Scholar

Camacho, L. F. S., Coan, M. M. D., Scapim, C. A., Barth Pinto, R. J., Tessmann, D. J., and Contreras-Soto, R. I (2019). A genome-wide association study for partial resistance to southern corn rust in tropical maize. Plant Breed. 138, 770–780. doi: 10.1111/pbr.12718

CrossRef Full Text | Google Scholar

Chan, D. M., Rao, R., Huang, F., and Canny, J. F. (2018). “t-SNE-CUDA: GPU-accelerated t-SNE and its applications to modern data,” in Proceedings of the International Symposium on Computer Architecture and High Performance Computing, Lyon.

Google Scholar

Chazan, S. E., Gannot, S., and Goldberger, J. (2019). “Deep clustering based on a mixture of autoencoders,” in Proceedings of the IEEE 29th International Workshop on Machine Learning for Signal Processing (MLSP), Pittsburgh, PA.

Google Scholar

Chen, Y. T., Chang, C. C., Chen, C. W., Chen, K. C., and Chu, Y. W. (2019). MADS-box gene classification in Angiosperms by clustering and machine learning approaches. Front. Genet. 9:707. doi: 10.3389/fgene.2018.00707

PubMed Abstract | CrossRef Full Text | Google Scholar

Coan, M., Senhorinho, H. J., Pinto, R. J., Scapim, C. A., Tessmann, D. J., Williams, W. P., et al. (2018). Genome-wide association study of resistance to ear rot by Fusarium verticillioides in a tropical field maize and popcorn core collection. Crop Sci. 58, 564–578. doi: 10.2135/cropsci2017.05.0322

CrossRef Full Text | Google Scholar

Costa, M. O., Capel, L. S., Maldonado, C., Mora, F., Mangolin, C. A., and Machado, M. D. F. P. D. (2020). High genetic differentiation of grapevine rootstock varieties determined by molecular markers and artificial neural networks. Acta Sci. Agron. 42:e43475. doi: 10.4025/actasciagron.v42i1.43475

CrossRef Full Text | Google Scholar

Davies, D. L., and Bouldin, D. W. (1979). “A cluster separation measure,” in Proceedings of the IEEE Transactions on Pattern Analysis and Machine Intelligence, Piscataway, NJ.

Google Scholar

Elshire, R. J., Glaubitz, J. C., Sun, Q., Poland, J. A., Kawamoto, K., Buckler, E. S., et al. (2011). A robust, simple genotyping-by-sequencing (GBS) approach for high diversity species. PLoS One 6:e19379. doi: 10.1371/journal.pone.019379

CrossRef Full Text | Google Scholar

Evanno, G., Regnaut, S., and Goudet, J. (2005). Detecting the number of clusters of individuals using the software STRUCTURE: a simulation study. Mol. Ecol. 14, 2611–2620. doi: 10.1111/j.1365-294x.2005.02553.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Ferreira, F., Scapim, C. A., Maldonado, C., and Mora, F. (2018). SSR-based genetic analysis of sweet corn inbred lines using artificial neural networks. Crop Breed. Appl. Biot. 18, 309–313. doi: 10.1590/1984-70332018v18n3n45

CrossRef Full Text | Google Scholar

Gao, H., Williamson, S., and Bustamante, C. D. (2007). A Markov chain Monte Carlo approach for joint inference of population structure and inbreeding rates from multilocus genotype data. Genetics 176, 1635–1651. doi: 10.1534/genetics.107.072371

PubMed Abstract | CrossRef Full Text | Google Scholar

Ho, S. S., Urban, A. E., and Mills, R. E. (2019). Structural variation in the sequencing era. Nat. Rev. Genet. 21, 171–189. doi: 10.1038/s41576-019-0180-9

PubMed Abstract | CrossRef Full Text | Google Scholar

Joshi, P., Hearty, J., Sjardin, B., Massaron, L., and Boschetti, A. (2016). Python: Real World Machine Learning. Birmingham: Packt Publishing Ltd.

Google Scholar

Kaur, M., and Kaur, U. (2013). Comparison between k-means and hierarchical algorithm using query redirection. Int. J. Adv. Res. Comput. Sci. Softw. Eng. 3, 1454–1459.

Google Scholar

Kobak, D., and Berens, P. (2019). The art of using t-SNE for single-cell transcriptomics. Nat. Commun. 10, 1–14. doi: 10.1002/9783527678679.dg11959

CrossRef Full Text | Google Scholar

Kohonen, T., and Somervuo, P. (1998). Self-organizing maps of symbol strings. Neurocomputing 21, 19–30. doi: 10.1016/s0925-2312(98)00031-9

CrossRef Full Text | Google Scholar

Kulka, V. P., Silva, T. A. D., Contreras-Soto, R. I., Maldonado, C., Mora, F., and Scapim, C. A. (2018). Diallel analysis and genetic differentiation of tropical and temperate maize inbred lines. Crop Breed. Appl. Biot. 18, 31–38. doi: 10.1590/1984-70332018v18n1a5

CrossRef Full Text | Google Scholar

Li, G., Han, D., Wang, C., Hu, W., Calhoun, V. D., and Wang, Y. P. (2020). Application of deep canonically correlated sparse autoencoder for the classification of schizophrenia. Comput. Meth. Prog. Biol. 183:105073. doi: 10.1016/j.cmpb.2019.105073

PubMed Abstract | CrossRef Full Text | Google Scholar

Li, J., Chen, G. B., Rasheed, A., Li, D., Sonder, K., Zavala Espinosa, C., et al. (2019). Identifying loci with breeding potential across temperate and tropical adaptation via EigenGWAS and EnvGWAS. Mol. Ecol. 28, 3544–3560. doi: 10.1111/mec.15169

PubMed Abstract | CrossRef Full Text | Google Scholar

Liu, Y., Wang, D., He, F., Wang, J., Joshi, T., and Xu, D. (2019). Phenotype prediction and genome-wide association study using deep convolutional neural network of soybean. Front. Genet. 10:1091. doi: 10.3389/fgene.2018.001091

CrossRef Full Text | Google Scholar

Maaten, L. V. D., and Hinton, G. (2008). Visualizing data using t-SNE. J. Mach. Learn. Res. 9, 2579–2605.

Google Scholar

Macqueen, J. (1967). “Some methods for classification and analysis of multivariate observations,” in Proceedings of the Fifth Berkeley Symposium on Mathematical Statistics and Probability, Berkeley, CA.

Google Scholar

Mafra, G. S., Do Amaral Júnior, A. T., Almeida, F. J. E. D., Vivas, M., Araújo Diniz-Santos, P. H., Saltires-Santos, J., et al. (2019). SNP-based mixed model association of growth-and yield-related traits in popcorn. PLoS One 14:e0218552. doi: 10.1371/journal.pone.0218552

PubMed Abstract | CrossRef Full Text | Google Scholar

Maldonado, C., Mora, F., Bertagna, F. A. B., Kuki, M. C., and Scapim, C. A. (2019). SNP-and haplotype-based GWAS of flowering-related traits in maize with network-assisted gene prioritization. Agronomy 9:725. doi: 10.3390/agronomy9110725

CrossRef Full Text | Google Scholar

Manning-Dahan, T. (2018). PCA and Autoencoders. Montreal: Concordia University, INSE 6220.

Google Scholar

Meirmans, P. G. (2012). AMOVA-based clustering of population genetic data. J. Hered. 103, 744–750. doi: 10.1093/jhered/ess047

PubMed Abstract | CrossRef Full Text | Google Scholar

Min, E., Guo, X., Liu, Q., Zhang, G., Cui, J., and Long, J. (2018). A survey of clustering with deep learning: from the perspective of network architecture. IEEE Access. 6, 39501–39512. doi: 10.1109/access.2018.2855437

CrossRef Full Text | Google Scholar

Nikolic, N., Park, Y. S., Sancristobal, M., Lek, S., and Chevalet, C. (2009). What do artificial neural networks tell us about the genetic structure of populations? The example of European pig populations. Genet. Res. 91, 121–132. doi: 10.1017/s0016672309000093

PubMed Abstract | CrossRef Full Text | Google Scholar

Padhukasahasram, B. (2014). Inferring ancestry from population genomic data and its applications. Front. Genet. 5:204. doi: 10.3389/fgene.2014.00204

PubMed Abstract | CrossRef Full Text | Google Scholar

Peña-Malavera, A., Bruno, C., Fernandez, E., and Balzarini, M. (2014). Comparison of algorithms to infer genetic population structure from unlinked molecular markers. Stat. Appl. Genet. Mol. 13, 391–402.

Google Scholar

Porras-Hurtado, L., Ruiz, Y., Santos, C., Phillips, C., Carracedo, Á, and Lareu, M. (2013). An overview of STRUCTURE: applications, parameter settings, and supporting software. Front. Genet. 4:98. doi: 10.3389/fgene.2013.00098

PubMed Abstract | CrossRef Full Text | Google Scholar

Pritchard, J. K., Stephens, M., and Donnelly, P. (2000). Inference of population structure using multilocus genotype data. Genetics 155, 945–959.

Google Scholar

Qu, Y., Tran, D., and Ma, W. (2019). Deep Learning approach to biogeographical ancestry inference. Proc. Comput. Sci. 159, 552–561. doi: 10.1016/j.procs.2019.09.210

CrossRef Full Text | Google Scholar

Reynolds, D. (2015). “Gaussian mixture models,” in Encyclopedia of Biometrics, eds S. Z. Li and A. K. Jain (Boston, MA: Springer).

Google Scholar

Rousseeuw, P. J. (1987). Silhouettes: a graphical aid to the interpretation and validation of cluster analysis. J. Comput. Appl. Math. 20, 53–65. doi: 10.1016/0377-0427(87)90125-7

CrossRef Full Text | Google Scholar

Sakaue, S., Hirata, J., Kanai, M., Suzuki, K., Akiyama, M., Too, C. L., et al. (2020). Dimensionality reduction reveals fine-scale structure in the Japanese population with consequences for polygenic risk prediction. Nat. Commun. 11, 1–11.

Google Scholar

Senhorinho, H. J. C., Coan, M. M. D., Marino, T. P., Kuki, M. C., Pinto, R. J. B., Scapim, C. A., et al. (2019). Genomic-wide association study of popping expansion in tropical popcorn and field corn germplasm. Crop Sci. 59, 2007–2019. doi: 10.2135/cropsci2019.02.0101

CrossRef Full Text | Google Scholar

Shultz, A. J., Baker, A. J., Hill, G. E., Nolan, P. M., and Edwards, S. V. (2016). SNPs across time and space: population genomic signatures of founder events and epizootics in the house finch (Haemorhous mexicanus). Ecol. Evol. 6, 7475–7489. doi: 10.1002/ece3.2444

PubMed Abstract | CrossRef Full Text | Google Scholar

Stift, M., Koláø, F., and Meirmans, P. G. (2019). STRUCTURE is more robust than other clustering methods in simulated mixed-ploidy populations. Heredity 123, 429–441. doi: 10.1038/s41437-019-0247-6

PubMed Abstract | CrossRef Full Text | Google Scholar

Tan, J., Ung, M., Cheng, C., and Greene, C. S. (2014). Unsupervised feature construction and knowledge extraction from genome-wide assays of breast cancer with denoising autoencoders. Pac. Symp. Biocomput. 20, 132–143.

Google Scholar

Vincent., P., Larochelle, H., Bengio, Y., and Manzagol, P. A. (2008). “Extracting and composing robust features with denoising autoencoders,” in Proceedings of the 25th International Conference on Machine Learning, Helsinki

Google Scholar

Wang, J., Li, X., Do Kim, K., Scanlon, M. J., Jackson, S. A., Springer, N. M., et al. (2019). Genome-wide nucleotide patterns and potential mechanisms of genome divergence following domestication in maize and soybean. Genome Biol. 20, 1–16.

Google Scholar

Ward, J. H. Jr. (1963). Hierarchical grouping to optimize an objective function. J. Am. Stat. Assoc. 58, 236–244. doi: 10.1080/01621459.1963.10500845

CrossRef Full Text | Google Scholar

Xie, J., Girshick, R., and Farhadi, A. (2016). “Unsupervised deep embedding for clustering analysis,” in Proceedings of the 33rd International Conference on International Conference on Machine Learning, Washington, DC.

Google Scholar

Xie, R., Wen, J., Quitadamo, A., Cheng, J., and Shi, X. (2017). A deep auto-encoder model for gene expression prediction. BMC Genom. 18:845. doi: 10.1186/s12864-017-4226-0

PubMed Abstract | CrossRef Full Text | Google Scholar

Yokota, R., and Wu, W. (2018). “Supercomputing frontiers,” in Proceedings of the 4th Asian Conference, SCFA 2018, Singapore.

Google Scholar

Yue, T., and Wang, H. (2018). Deep learning for genomics: a concise overview. arXiv [Preprint]. Available online at: https://arxiv.org/abs/1802.00810 (accessed January 28, 2020).

Google Scholar

Zhang, H., Zhang, C., Li, Z., Li, C., Wei, X., Zhang, B., et al. (2019). A new method of RNA secondary structure prediction based on convolutional neural network and dynamic programming. Front. Genet. 10:467. doi: 10.3389/fgene.2019.00467

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: deep learning, genome-wide studies, machine learning, single-nucleotide polymorphisms, dimensionality reduction

Citation: López-Cortés XA, Matamala F, Maldonado C, Mora-Poblete F and Scapim CA (2020) A Deep Learning Approach to Population Structure Inference in Inbred Lines of Maize. Front. Genet. 11:543459. doi: 10.3389/fgene.2020.543459

Received: 17 March 2020; Accepted: 19 October 2020;
Published: 24 November 2020.

Edited by:

Jie Chen, Augusta University, United States

Reviewed by:

Margaret Woodhouse, United States Department of Agriculture, United States
Francesca Taranto, Italian National Research Council, Italy

Copyright © 2020 López-Cortés, Matamala, Maldonado, Mora-Poblete and Scapim. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Freddy Mora-Poblete, morapoblete@gmail.com

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.