Skip to main content

ORIGINAL RESEARCH article

Front. Psychol., 11 April 2022
Sec. Quantitative Psychology and Measurement

Evaluation of Prediction-Oriented Model Selection Metrics for Extended Redundancy Analysis

  • 1Department of Psychology, University of Manitoba, Winnipeg, MB, Canada
  • 2Department of Psychology, McGill University, Montreal, QC, Canada

Extended redundancy analysis (ERA) is a statistical method that relates multiple sets of predictors to response variables. In ERA, the conventional approach of model evaluation tends to overestimate the performance of a model since the performance is assessed using the same sample used for model development. To avoid the overly optimistic assessment, we introduce a new model evaluation approach for ERA, which utilizes computer-intensive resampling methods to assess how well a model performs on unseen data. Specifically, we suggest several new model evaluation metrics for ERA that compute a model’s performance on out-of-sample data, i.e., data not used for model development. Although considerable work has been done in machine learning and statistics to examine the utility of cross-validation and bootstrap variants for assessing such out-of-sample predictive performance, to date, no research has been carried out in the context of ERA. We use simulated and real data examples to compare the proposed model evaluation approach with the conventional one. Results show the conventional approach always favor more complex ERA models, thereby failing to prevent the problem of overfitting in model selection. Conversely, the proposed approach can select the true ERA model among many mis-specified (i.e., underfitted and overfitted) models.

Introduction

Extended redundancy analysis (ERA; Takane and Hwang, 2005) is a statistical modeling framework that relates multiple sets of predictors to response variables. In ERA, a component is extracted from each set of predictors in such a way that it accounts for the maximum variation of response variables. ERA provides several benefits in terms of interpretability and predictability, especially when many predictors need to be studied in a regression problem. For example, ERA provides a parsimonious interpretation of directional relationships using the extracted components whose number is much smaller than that of the original predictors. Also, using domain-specific knowledge concerning which predictors are to be put together can improve the interpretability of the extracted components. Moreover, ERA allows researchers to generate predictive models because it searches for components that maximize predictive accuracy, without having to eliminate any predictors of interest to avoid multicollinearity. Indeed, the practical usefulness of ERA lies in its predictive nature; ERA has been well blended with many statistical techniques for prediction problems, e.g., regularizations (Hwang et al., 2015; Lee et al., 2019; Kim et al., 2020) and a tree-based supervised learning method (Kim and Hwang, 2021).

The present study concerns model evaluation in ERA. How to assess a model’s performance is critical for model comparison and model selection. One existing method is to calculate FIT (Takane and Hwang, 2005), an R2-type overall goodness-of-fit (GOF) metric for ERA. Several other GOF metrics for parametric ERA, such as AICERA and BICERA (DeSarbo et al., 2015), are also available, which are computed based on penalized-likelihood criteria taking model complexity into account. A model’s GOF based on these metrics provides diagnostic information that helps explain model performance by quantifying the degree of similarity between observed and predicted values. However, all of these conventional metrics represent “in-sample” model evaluation, i.e., the same dataset is used both to construct the model and to assess its performance. This naturally leads to an overly optimistic view of a model’s performance: the more closely we fit the model to the training sample—a set of data used to estimate parameters, the better it will perform when being evaluated on the same sample. This is a well-known statistical phenomenon called “optimism” (Efron, 1983; Efron and Tibshirani, 1997; Hastie et al., 2009, Chapter 7).

When researchers are interested in predicting important health or behavioral outcomes to the benefit of the broader population, relying on such an “optimistic” view based on in-sample model evaluation is not ideal because it provides little information about the model’s performance on “out-of-sample”—data that are not used for model building. For example, in studies on cognitive impairment in older adults (Choi and Jin, 2018; Na, 2019), patient responses to treatments for depression (Cuijpers et al., 2013), or user response patterns in online advertising (Zhang et al., 2016), the goal of model evaluation is to select a prediction model that can best assist practitioners with their decision making in unseen cases (e.g., treatment recommendation for new patients). To develop such prediction models that can generalize beyond the current sample, it would be better to assess a model’s performance based on out-of-samples.

Thus, in this paper, we introduce several new model evaluation metrics for ERA, each of which aims to quantify how well a model performs in out-of-sample data. But before discussing the out-of-sample metrics, we also investigate the degree of optimism in the existing in-sample model evaluation based on a simulation study. Hastie et al. (2009, Chapter 7) discussed, in a general regression problem, the optimism of in-sample model evaluation decreases linearly as the training sample size increases but increases with model complexity. Thus, in the simulation study, we examine how the degree of optimism is affected by different training sample sizes and the number of ERA parameters under several model mis-specification conditions.

We then illustrate a new model evaluation framework for ERA which aims to avoid the overly optimistic assessment in the conventional in-sample metrics. Specifically, we apply various sample-reuse or resampling methods, such as cross-validation (CV; Stone, 1974; Geisser, 1975) and the bootstrap (Efron, 1979, 1983), to assess out-of-sample model performance. The basic idea behind this framework is to avoid the optimism by using non-overlapping data for model development and evaluation, which in turn provides a more accurate estimate of model performance in a new sample (Hastie et al., 2009, Chapter 7; Shmueli, 2010). Although considerable work has been done on the use of CV and bootstrap variants for out-of-sample model assessment in various prediction models (e.g., Steyerberg et al., 2001; Molinaro et al., 2005; Smith et al., 2014; Austin and Steyerberg, 2017), to date, no research has applied these general tools to ERA. Thus, we formulate various out-of-sample prediction error estimators for ERA based on (1) c-fold CV (c = 3, 5, and 10), (2) leave-one-out CV (LOOCV), (3) out-of-bag (OOB) bootstrap, (4) 0.632 bootstrap, and (5) 0.632+ bootstrap, and carry out a simulation to examine their relative behaviors under different simulation conditions.

The remainder of this paper is structured as follows. First, we briefly recapitulate the ERA model and provide a formulation of a prediction error estimator for ERA. We then describe the use of the abovementioned resampling methods for out-of-sample model evaluation in ERA. Next, we present the results of a series of simulation studies comparing the conventional in-sample metrics and new out-of-sample prediction error estimators. In all simulations, mis-specified (i.e., underfitted and overfitted) models are considered to examine the behavior of each estimator. The empirical usefulness of the proposed out-of-sample model evaluation is illustrated using a real-world dataset. Finally, we summarize the findings, provide a guideline for practitioners on which resampling approach should be favored under which conditions, and discuss the limitation of the study.

Materials and Methods

Extended Redundancy Analysis

Let yi denote the ith value of the response variable (i = 1,  ..., N). Assume that there are K different sets of predictors, each of which consists of Pk predictors (k = 1,  ..., K). Let xikp denote the ith value of the pth variable in the kth predictor set (p = 1,  ..., Pk) and xi = (xi11,  ..., xikp) denote a 1 by P vector of predictors for the ith observation, where P=k=1KPk. Let wkp denote a component weight assigned to xikp and wk =wk1,...,wkPkdenote a Pk by 1 vector of component weights in the kth predictor set. Let fik=p=1Pkxikpwkp denote the ith component score of the kth component, which is the sum of weighted predictors for the ith observation in the kth predictor set. Let bk denote the regression coefficient relating the kth component to the response variable. Let ei denote an error term for yi. We assume that all predictors are standardized with zero means and unit variances (Takane and Hwang, 2005).

The ERA model (Takane and Hwang, 2005) is then expressed as,

yi=k=1K[p=1Pkxikpwkp]bk+ei=xiWb+ei=k=1Kfikbk+ei=fib+ei,    (1)

where W=w100wk, fi=fi1,,fiK, and b=b1,,bK.

This can also be expressed as,

y=XWb+e=Fb+e    (2)

where y is an N by 1 vector of response variables, X is an N by P matrix of predictors, W is a P by K matrix of component weights, F is a N by K matrix of regression coefficients, and e is an N by 1 vector of errors. For identification, a standardization constraint is imposed on F such that diag(FF) = NI, where diag(∙) denotes a diagonal matrix with the elements of FF on its diagonal. As shown in (1), each set of predictors reduces to a single component, which in turn influences the response variable. The component weight wkp shows the contribution of each predictor to obtaining its component as in data reduction methods such as principal component analysis or canonical correlation analysis, whereas the regression coefficient bk signifies the effect of each component on the response variable as in linear regression. In this regard, ERA carries out data reduction and linear regression simultaneously.

Figure 1 displays an example of the ERA model, where each component is associated with two predictors (P1 = P2 = 2) and a response variable is influenced by two components (K = 2). For this example, the W and b are given as,

FIGURE 1
www.frontiersin.org

Figure 1. A prototypical ERA model. Square boxes indicate observed predictor and response variables. Circles represent components and an error term.

W=w110w1200w210w22andb=b1b2

The ERA model contains two sets of parameters—component weights (W) and regression coefficients (b). These unknown parameters are estimated by minimizing the following objective function:

ϕ=SSyXWb,    (3)

with respect to W and b, subject to the constraint diag(FF) = NI, where SS(a) = aa, i.e., the sum of squares for all the elements of a vector a. An alternating least-squares (ALS) algorithm (de Leeuw et al., 1976) was developed to minimize the objective function (Takane and Hwang, 2005). Refer to Takane and Hwang (2005) for a more detailed description of the parameter estimation algorithm.

To assess the performance of ERA models, an overall GOF metric, called FIT (Takane and Hwang, 2005), can be calculated to evaluate how well a given ERA model explains the variance of the response variable:

FIT=1ϕSSy    (4)

The values of FIT range from 0 to 1, and larger values indicate more explained variance of the response variable. The limitation of using FIT in conventional ERA will be discussed later but note that it only informs how well a model fits to the training set, thus being of little utility when assessing a model’s prediction capability.

Assessment of Predictive Performance

Consider a continuous response variable Y that is related to a predictor matrix X by a prediction model h: X → Y. Then, ĥX denotes the predicted responses estimated from the observed training set T = {x1,y1, ⋯,xi,yi, ⋯,xN,yN}. We assume that the observations in T are random samples from a distribution H. Let denote LY,ĥX a loss function for measuring errors between Y and ĥX. For example, in many regression-based models, a common choice for a continuous Y is the squared-error loss, i.e., LY,ĥX=YĥX2. The term “loss” in mathematical optimization is used to describe how much a model is losing compared to having made perfect predictions. Thus, a loss function is a measure of how good a model does in terms of being able to predict the expected responses. Note that, as shown in (3), ERA defines its objective function in terms of the mean squared-error loss and seeks to minimize it over all possible parameter values, which shows its predictive nature.

To quantify the overall predictive performance of an ERA model, root mean square error (RMSE),

RMSE=1Ni=1Nyiĥxi2    (5)

is a reasonable choice because ERA parameters are estimated to minimize the mean squared-error loss, as discussed above. This metric can range from 0 to ∞ and is negatively oriented scores, which means lower values are better. Also, it is indifferent to the direction of errors since the errors are squared before they are averaged.

Estimating Generalization Error Using Resampling Methods

In predictive modeling, we wish to obtain a model that not only performs well on the training data, but also on independent unseen data. Thus, it is important to understand how to estimate the true error rate of a model when it is used to predict the future responses. Let x0,y0 is a new independent test sample randomly drawn from H. The true test error or generalization error (Efron, 1983; Efron and Tibshirani, 1997; Hastie et al., 2009, Chapter 7) is the prediction error for x0,y0,

ErrT=Ex0,y0Ly0,ĥx0|Τ    (6)

Note that, in (6), only x0,y0 is random with T being fixed, meaning that the true test error refers to the conditional error for the particular training set T. In practice, it is more amenable to estimate a model’s prediction error as the expectation of ErrT (Efron, 1983; Efron and Tibshirani, 1997; Hastie et al., 2009, Chapter 7),

Err=EErrT=ELY,ĥX    (7)

where everything random is averaged over. In many machine learning applications, where a large independent test set is available, the goal of model selection is to find a model that gives the minimum expected test error in (7).

In the absence of a large independent test set, the simplest way to estimate ErrT is to use the training error or apparent error, defined by the average loss over the training data,

ErrTrain=1Ni=1NLyi,ĥxi    (8)

As discussed previously, ErrTrain is typically smaller than ErrT, i.e., ErrTrain tends to be biased downward as an estimate of ErrT because the same observations are used twice, both for fitting ĥX and for evaluating the prediction error of ĥX. Note that, in conventional ERA, model evaluation is based only on this apparent performance.

A straightforward approach for correcting the inherent optimism in ErrTrain is to randomly split the observed data T in two parts: one for developing the model (learning set) and the other for measuring its predictive performance (validation set). With this split-sample approach, model performance can be assessed on independent data not used for model development. There are, however, two criticisms of this procedure. First, it is inefficient when the size of T is small, owing to its reduction of the size of both learning and validation sets. Second, high variability in the estimated predictive performance can be introduced because of its reliance on a single split of T.

Various computer-intensive resampling methods can be employed to avoid these limitations of the split-sample approach. Let Err¯(Resampling Method) denote the estimated generalization error (7) of an ERA model, which is formulated as the averaged predictive performance on validation data, where the predictive performance is computed as described in (5) and validation data are generated based on a resampling method. Using this out-of-sample prediction error estimator, we can now define the optimal model as the one gives the lowest Err¯(Resampling Method) (i.e., the smallest prediction error on validation data) in model selection. Key references on the use of different resampling methods (for validation data creation) include Efron (1983), Breiman and Spector (1992), and Efron and Tibshirani (1997).

c-fold CV is one of the most preferred methods used to generate validation data. This method randomly assigns N observations to one of c partitions such that the partitions are of nearly equal size. Subsequently, the learning set contains all but one of the partitions which is labeled the validation set. After fitting a model to the learning set, the prediction error (RMSE) of the fitted model is computed using the validation set. We then repeat this procedure for all c folds and average the prediction errors, resulting in the c-fold CV estimate of prediction error: Err¯(cv,c). LOOCV is the most extreme case of c-fold CV, where the number of folds equals the number of observations (i.e., c = N) and each observation is individually assigned to the validation set.

Another popular approach is based on the bootstrap (Efron, 1979). Efron (1983) proposed and compared several bootstrap variants for the assessment of a model’s predictive performance, which are generally referred to as out-of-bag (OOB) estimators in the statistics and machine learning literature. Calculating the OOB prediction error begins with bootstrap sampling. Let B be the number of bootstrap replications. For each draw, a bootstrap sample contains only 63.2% of the original data on average (referred to as in-bag sample) due to the sampling with replacement. The prediction error is assessed on the remaining 37% of the data (OOB data) for each bootstrap draw and subsequently averaged over the B iterations, resulting in the OOB estimate of prediction error: Err¯(OOB).

There are two more variations of the OOB estimator: 0.632 estimator (Efron, 1983) and 0.632+ estimators (Efron and Tibshirani, 1997). Both aim to correct the underestimated ErrTrain as a weighted combination of ErrTrain and Err¯(OOB) such that ω·ErrTrain+ (1-ωErr¯(OOB). The value of ω is fixed as 0.632 for the 0.632 estimator, whereas ω is determined based on the so-called “no-information error rate” for the 0.632+ estimator (Efron and Tibshirani, 1997). In brief, the weight is dependent on the relative amount of overfitting coefficient R: ω = 0.632/(1–0.368·R). The relative overfitting R is large when the difference between ErrTrain and Err¯(OOB) is relatively large. In this case, R and ω approach 1, indicating that the estimated prediction error is largely based on Err¯(OOB). When the overfitting is small, R approaches 0 and ω 0.632, resulting in similarity between the 0.632 and 0.632+ estimators. In this paper, we also consider these two bootstrap variants for ERA: Err¯(.632) and Err¯(.632+).

Simulation Study

We conducted a simulation study to examine the behaviors of the abovementioned in-sample and out-of-sample prediction error estimators for ERA. Specifically, under different model mis-specification conditions, we investigated the degree of optimism in ErrTrain in terms of (4) and (5) and compared the differences between the resampling methods in the estimation of ErrT: Err¯(cv), Err¯(LOOCV), Err¯(OOB), Err¯(.632), and Err¯(.632+). We particularly focused on the influence of sample size and model complexity. For this, we considered five different sample sizes (N = 50, 100, 200, 500, 1,000) for training data (T) and varied the number of predictors per component (Np = 2, 4, 6, 8).

For data generation, we specified an ERA model that was composed of two components (K = 2) and a response variable (see Figure 1). We fixed the first regression coefficient b1 to 0.3, and the variance explained by two components R2 to 0.4, which in turn resulted in the value of the second regression coefficient, b2 = 0.56. We assumed no correlation between the components. Each component was linked to Np predictors, where each component weight, wkp, was randomly sampled from a uniform distribution between 0.8 and 1. Following the data generation approach of Becker et al. (2013), the variance-covariance matrix of the predictor and response variables, Σ, was obtained based on the ERA parameters described above. We generated 1,000 datasets from a multivariate normal distribution with zero means and Σ for each combination of sample sizes (N) and the number of predictors per component (Np).

For each simulation scenario, we considered three different conditions of model mis-specification and investigated if the true (data generating) model was chosen using the various out-of-sample prediction error estimators. The three mis-specified models and the true model were as follow: (1) under-specified model, f0:yi=b1fi1+ei, (2) correctly specified model (i.e., data generating model), f1:yi=b1fi1+b2fi2+ei, (3) over-specified model with a component interaction term, f2:yi=b1fi1+b2fi2+b3(fi1fi2)+ei, and (4) over-specified model with interaction and quadratic terms, f3:yi=b1fi1+b2fi2+b3(fi1fi2)+b4(fi12)+b5(fi22)+ei.

All data generation and computations were carried out using the R system for statistical computing (Version 3.4). We wrote an R code to implement ERA and to compute the various in-sample and out-of-sample prediction error estimators for ERA. We used the R package “mlr” (Bischl et al., 2016) to create random samples generated by c-fold CV, LOOCV, OOB, 0.632, and 0.632+ bootstrap. See Data and Code Availability for details regarding the code used for the simulation.

Optimism in In-Sample Metrics

Figure 2 shows the training set or apparent FIT values (i.e., in-sample model performance measured by FIT) for four different ERA models (f0, …, f3), including the true and three mis-specified ones, in relation to the sample size (N) and the number of indicators per component (Np). Boxplots were constructed to show the distribution of estimated FIT values over 1,000 repetitions. For each simulation condition, we also estimated the Err estimate in (7)—which was computed using 1,000 independent test sets (of size N each)—and added it as a dotted line, so that we can easily view the bias and variability of FIT, as a model performance estimator for ErrT.

FIGURE 2
www.frontiersin.org

Figure 2. Behavior of in-sample FIT values as the training set sample size (N = 50, 100, 200, 500, 1,000) and model complexity (Np = 2, 4, 6, 8) are varied for different model mis-specification conditions (f0, f1, f2, and f3). f1 is the true model, f0 is an under-specified model, and f2 and f3 are over-specified models.

As the boxplots show, when sample size was large and the number of predictors per component was small, the median apparent performance of FIT (centers of boxplots) approached the true test error (dotted lines). FIT’s variability (each boxplot’s length) also decreased in such conditions. However, the true model (f1) was never selected by the FIT, because the overfitted models (f2 and f3) had higher FIT values under all simulation conditions. In addition, the median apparent performance for f1 was always above the dotted line, showing FIT’s optimism as an estimate of ErrT, as was theoretically expected. For the overfitted models, the apparent performance of FIT tended to reach its maximum value (i.e., 1) rapidly even when sample size was small and the number of predictors per component was large, indicating that the optimism in the apparent FIT metric can be of particular problem when overfitting happens.

In Figure 3, boxplots show the in-sample model performance measured by RMSE in (5). The differences between Figure 2 and Figure 3 are minimal in terms of the optimism of in-sample model assessment: the apparent RMSE values were clearly estimated below the dotted lines, indicating overly underestimated prediction error. However, RMSE showed less variability than FIT: the difference in the variabilities of apparent RMSE values for f0, f1, f2, and f3 became very minimal when the sample size was larger than 50 (N ≥ 100).

FIGURE 3
www.frontiersin.org

Figure 3. Behavior of in-sample RMSE values as the training set sample size (N = 50, 100, 200, 500, 1,000) and model complexity (Np = 2, 4, 6, 8) are varied for different model mis-specification conditions (f0, f1, f2, and f3). f1 is the true model, f0 is an under-specified model, and f2 and f3 are over-specified models.

Behavior of Out-of-Sample Estimators

In this section, all simulation results are discussed but only a limited number of figures are displayed because the differences in the error estimators across the resampling methods were negligible as the sample size increased, e.g., N > 100. The full compilation of simulation results is archived on the first author’s GitHub, see Data and Code Availability for more details.

Figure 4 displays the behavior of various out-of-sample prediction error estimators based on different resampling methods for the correctly and incorrectly specified ERA models (f0, f1, f2, and f3) when N = 100. The figures in the top panels show the results of Np = 2, where those in the bottom portray the results obtained when Np = 4. We obtained the results for (1) c-fold CV estimators with c = 3, 5, and 10 (i.e., Err¯(cv,3), Err¯(cv,5), Err¯(cv,10)), (2) LOOCV estimator (Err¯(LOOCV)), (3) the regular OOB bootstrap estimators with B = 20, 50, and 100 (Err¯(OOB,B=20), Err¯(OOB,B=50), Err¯(OOB,B=100)), (4) the 0.632 estimators with B = 20, 50, and 100 (Err¯(.632,B=20), Err¯(.632,B=50), Err¯(.632,B=100)), and (5) the 0.632+ estimators with B = 20, 50, and 100 (Err¯(.632+,B=20), Err¯(.632+,B=50), Err¯(.632+,B=100)). We also computed the apparent RMSE (i.e., in-sample RMSE) and displayed it with other prediction error estimators to see if the optimism remained. This is denoted by “(1) App” in the figure. As there was no noticeable difference between the 0.632 and 0.632+ estimators in all simulation conditions, the results of the 0.632+ estimators are not included in Figure 4. The results obtained from the OOB bootstrap and 0.632 estimators with B = 100 are not displayed as well because there was minimal improvement over those with B = 50. In the figure, each error bar represents one standard deviation of the expected value of the estimated errors over 1,000 repetitions. Each dotted line indicates the Err estimate in (7) which is estimated over 1,000 simulated independent test sets (of size N each).

FIGURE 4
www.frontiersin.org

Figure 4. Behavior of out-of-sample prediction error estimators based on various resampling methods, as the model complexity (Np = 2 and 4) are varied different conditions of model mis-specifications (f0, f1, f2, and f3). The training sample size is N = 100.

Most noticeably, all out-of-sample prediction error estimators resulted in the smallest error for f1, on average; thus, the true ERA model was always selected across all simulation conditions. In this true model condition, all of the resampling estimators—except the LOOCV—successfully corrected the optimism in the apparent error estimator: They reported larger values than the underestimated in-sample RSME [“(1) App” in the figure], and at the same time, located closer to the dotted lines. They also exhibited smaller variabilities, compared to those in the over-specified conditions (i.e., f2 and f3). In all sample size and model complexity conditions, the LOOCV estimator for f1 had the lowest variability but showed noticeable downward bias. This is possible because each learning set in the LOOCV procedure is very similar to the full observed data T (Efron and Tibshirani, 1997). The c-fold CV estimators had smaller biases than other bootstrap variants, regardless of the number of folds, but the difference among the resampling methods was marginal.

When overfitting occurred (f2 and f3), however, the CV estimators overestimated the true prediction error with a high variability when the number of folds was small (c = 3 and 5). Similarly, the OOB bootstrap estimators showed large upward biases, but such bias was substantially reduced in the 0.632 and 0.632+ estimators (The advantage of increasing B from 20 to 50 was minimal in all bootstrap variant estimators). The variability of both 0.632 and 0.632+ estimators were smaller than that of 10-fold CV. Unlike simulation results in which various regression-based models were considered (e.g., Efron and Tibshirani, 1997; Steyerberg et al., 2001), 0.632+ did not outperform LOOCV in the context of ERA, particularly in small samples and with many predictors.

Empirical Application

We examined the utility of the proposed out-of-sample model evaluation metrics with a real data application. We used the same ERA model considered in Kim and Hwang (2021) for the analysis of nicotine dependence in the United States. Their model considered the effects of three components of 12 predictors—early exposure to substances (f1), mental health (f2), and socioeconomic status (SES, f3)—on nicotine dependence among American adults. Table 1 shows which component is associated with which predictors. The authors applied the ERA model to the National Survey on Drug Use and Health (NSDUH) data collected in 2012, but for this empirical study, we utilized new data from more recent survey years of NSDUH, 2016–2019 (U.S. Department of Health and Human Services et al., 2017, 2018, 2019, 2020). Table 1 presents a description of the variables included, as well as their summary statistics. As shown in the table, the number of respondents (N) varied across survey years, from 6,697 to 7,477. Particularly, the data collected in 2019 (N = 6,697) were used as an independent test set for assessing the generalization performance of a model.

TABLE 1
www.frontiersin.org

Table 1. Description of variables and summary statistics for the NSDUH data (2016 ~ 2019).

Based on the ERA model in Kim and Hwang (2021), we constructed possible candidate models with different levels of model complexity. See Table 2 for the candidate models. As shown in the table, M4 represents the original ERA model. Our particular interest lied in selecting the best-performing model among the various candidate models. Thus, we estimated their performances using the model evaluation metrics discussed in the previous section in order to rank them against each other. More specifically, we computed the following model evaluation metrics using the NSDUH data from 2016 to 2018: (1) apparent (in-sample) RMSE, (2) Err¯(cv,10), (3) Err¯(OOB,B=50), and (4) Err¯(.632,B=50). In addition, we computed the models’ RMSE using the 2019 NSDUH data. The objectives of this investigation were 2-fold. First, it aimed to estimate the generalization performance of each model using the test set. Second, by plotting it, one could visually detect when overfitting started to occur. The results are summarized in Table 2, which are also graphically displayed in Figures 5, 6.

TABLE 2
www.frontiersin.org

Table 2. Out-of-sample model performance metrics for the nicotine dependence ERA model using NSDUH data.

FIGURE 5
www.frontiersin.org

Figure 5. In-sample and out-of-sample RMSE for 12 candidate models using the NSDUH data from 2016 to 2018 (Total N = 21,553). The RMSE values are consistent with those in Table 2.

FIGURE 6
www.frontiersin.org

Figure 6. Test set RMSE for 12 models using the 2019 NSDUH data (N = 6,696). The RMSE values are consistent with those in Table 2.

In Figures 5, 6, we plotted the 12 candidate models on the x-axis, from simple to complex models, against the RMSE values on the y-axis. In Figure 5, the apparent RMSE shows a typical learning curve: The model performance on the training set consistently decreases with model complexity. By looking at the test set RMSE in Figure 6, where the generalization error goes down only until a turning point M4, we can clearly see M5 to M12 overfit the training data. Interestingly, all the out-of-sample model evaluation metrics did not show the same pattern as the apparent RMSE: after M4, they almost stopped decreasing with increasing model complexity. Moreover, their error trend even changed from descending to ascending after M10. In addition, all of them alleviated the optimism in the apparent RMSE. These findings support the simulation results that the out-of-sample metrics are more likely to avoid overfitted models than the traditional in-sample metrics for ERA. To build a model that can generalize to unseen samples, especially when overfitting is of concern, the use of the out-of-sample model assessment metrics is crucial in model selection.

Concluding Remarks

In many psychology studies, it is often assumed that a sample at hand (i.e., a training set of data) is a good reflection of what will be encountered in the future; thus, the final model is selected as the one optimized for the training data (Shmueli, 2010; Bryant, 2016). It is not ideal to compare candidate models based on such in-sample GOF assessment because GOF model evaluation metrics always favor more complex models (which fit the training data too tightly), thereby limiting the generalizability of the selected model. But estimating a model’s future performance only using information contained in the current sample is also a hard problem. To respond to this, we discussed the use of computer-intensive resampling methods, including variants of CV and the bootstrap, in order to provide new ways of assessing generalizability of ERA models. There has been no discussion in the ERA literature as to out-of-sample error estimation for model selection, and no comparison of widely used resampling methods has been conducted to date. Thus, we discussed the predictive modeling of ERA in the context of model selection, particularly the derivation and evaluation of new model selection criteria.

Our simulation study showed that the optimism of conventional in-sample model evaluation metrics was negligible in large samples (e.g., N ≥ 500) but never disappeared. This implies that comparing candidate models based only on conventional FIT or in-sample RMSE is not recommended because these model evaluation metrics always favor more complex models, thereby being unable to select a model that results in a reproducible conclusion for future data. The simulation study also highlighted the advantage of adopting CV and bootstrap methods to avoid overly optimistic assessment of a model’s predictive performance. Over a wide range of different sample sizes and model complexities under correctly and incorrectly specified model conditions, all of the out-of-sample prediction error estimators favored the true model with the highest frequency. In fact, owing to improvements in statistical computing over the past years, it has become substantially easier to execute various resampling methods on modern laptops without much computational burden.

The study’s general conclusions may be summarized as follows. Firstly, the our-of-sample model evaluation metrics based on 10-fold CV, 0.632, and 0.632+ methods outperformed other resampling strategies, but the differences between them became unnoticeable in large samples, e.g., N ≥ 200, even for mis-specified models. Secondly, when highly complex models are considered, the c-fold CV with a small number of folds and the regular OOB bootstrap methods may perform poorly compared to other resampling methods. Thirdly, for largely over-specified models, the LOOCV estimator was a reasonable choice as it resulted in the lowest bias and variability. Fourthly, B = 20 bootstrap replications would be sufficient for the regular OOB bootstrap estimator and its variants. The advantage of increasing B from 20 to 100 was minimal in terms of the variability of estimators. Lastly, overall, the 0.632 and 0.632+ estimators outperformed other estimators, but the 10-fold CV prediction error estimate approximated those of 0.632 and 0.632+ in almost all settings. Thus, for computationally burdensome analyses, 10-fold CV may be preferable over the OOB bootstrap estimators.

As the first initiative in investigating out-of-sample model performance of ERA, the broader goal of this study is to bring this discussion into the field of psychology so that such predictive model assessment metrics can be effectively utilized for investigation of reproducibility and generalizability of psychological and behavioral data science. Thus, we also demonstrated how the proposed model evaluation metrics could be utilized in practice, using a well-known national survey data set in the US. The out-of-sample model evaluation metrics did not show the continuously decreasing trend of the conventional in-sample metric. This suggests that the out-of-sample metrics are reliable in avoiding overfitted models and can be utilized as a tool to estimate the generalization error in the absence of a large independent test set. Using out-of-sample methodological practices to access a model’s predictive performance received little attention in the psychology literature, however, it comes to be applied more frequently in recent years, especially with the reliance on CV (e.g., Yarkoni and Westfall, 2017; Koul et al., 2018; de Rooij and Weeda, 2020).

In this paper, we discussed the assessment of prediction performance in terms of RMSE. Thus, the simulation results can be widely applicable for other ERA models for continuous responses fit by expected squared-error loss. However, when response variables are discrete—which is often termed classification problems in statistics and machine learning, the best choices of resampling methods may differ substantially. Simulation studies on prediction error estimation in classification problems (e.g., Efron, 1979; Hastie et al., 2009, Chapter 7.3) demonstrate that the bias and variance of expected test error in (7) behave considerably differently for classification loss functions (e.g., 0–1 loss, the negative binomial log-likelihood known as deviance or cross-entropy) than they do for squared-error loss. Thus, optimistic estimation of model performance in classification problems should be further examined. A variety of loss functions for prediction error need to be considered. Especially, understanding the effect of the number of events per variable (EPV), instead of simple sample sizes, is important when binary or multinomial response variables are considered.

Data and Code Availability

Original NSDUH data are openly available in the Substance Abuse and Mental Health Data Archive (SAMHDA) at https://datafiles.samhsa.gov/. The aggregated data used in Empirical Application are available at https://github.com/QuantMM/PredictiveERAmodeling. Simulation source code is also available at the same GitHub page.

Data Availability Statement

Publicly available datasets were analyzed in this study. This data can be found at: https://www.datafiles.samhsa.gov/.

Author Contributions

SK contributed to conception and methodology, organized the data, performed the simulation and statistical analysis, and wrote and edit the original draft. HH read the draft and contributed to manuscript revision. All authors contributed to the article and approved the submitted version.

Funding

This work was supported in part by University of Manitoba’s University Research Grants Program (project no. 54366) awarded to SK.

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Austin, P. C., and Steyerberg, E. W. (2017). Events per variable (EPV) and the relative performance of different strategies for estimating the out-of-sample validity of logistic regression models. Stat. Methods Med. Res. 26, 796–808. doi: 10.1177/0962280214558972

PubMed Abstract | CrossRef Full Text | Google Scholar

Becker, J.-M., Rai, A., and Rigdon, E. (2013). “Predictive validity and formative measurement in structural equation modeling: embracing practical relevance.” in the International Conference on Information Systems (ICIS). Available at: https://scholarworks.gsu.edu/marketing_facpub

Google Scholar

Bischl, B., Lang, M., Kotthoff, L., Schiffner, J., Richter, J., Studerus, E., et al. (2016). Mlr: machine learning in R. J. Mach. Learn. Res. 17, 1–5.

Google Scholar

Breiman, L., and Spector, P. (1992). Submodel selection and evaluation in regression. The X-random case. Int. Stat. Rev. 60:291. doi: 10.2307/1403680

CrossRef Full Text | Google Scholar

Bryant, F. B. (2016). Enhancing predictive accuracy and reproducibility in clinical evaluation research. J. Eval. Clin. Pract. 22, 829–834. doi: 10.1111/JEP.12669

PubMed Abstract | CrossRef Full Text | Google Scholar

Choi, H., and Jin, K. H. (2018). Predicting cognitive decline with deep learning of brain metabolism and amyloid imaging. Behav. Brain Res. 344, 103–109. doi: 10.1016/j.bbr.2018.02.017

PubMed Abstract | CrossRef Full Text | Google Scholar

Cuijpers, P., Berking, M., Andersson, G., Quigley, L., Kleiboer, A., and Dobson, K. S. (2013). A meta-analysis of cognitive-behavioural therapy for adult depression, alone and in comparison with other treatments. Can. J. Psychiatr. 58, 376–385. doi: 10.1177/070674371305800702

PubMed Abstract | CrossRef Full Text | Google Scholar

de Leeuw, J., Young, F. W., and Takane, Y. (1976). Additive structure in qualitative data: an alternating least squares method with optimal scaling features. Psychometrika 41, 471–503. doi: 10.1007/BF02296971

CrossRef Full Text | Google Scholar

de Rooij, M., and Weeda, W. (2020). Cross-validation: a method Every psychologist should know. Adv. Methods Pract. Psychol. Sci. 3, 248–263. doi: 10.1177/2515245919898466

CrossRef Full Text | Google Scholar

DeSarbo, W. S., Hwang, H., Blank, A., and Kappe, E. (2015). Constrained stochastic extended redundancy analysis. Psychometrika 80, 516–534. doi: 10.1007/s11336-013-9385-6

PubMed Abstract | CrossRef Full Text | Google Scholar

Efron, B. (1979). Bootstrap methods: another look at the jackknife. Ann. Stat. 7, 1–26. doi: 10.1214/AOS/1176344552

CrossRef Full Text | Google Scholar

Efron, B. (1983). Estimating the error rate of a prediction rule: improvement on cross-validation. J. Am. Stat. Assoc. 78:316. doi: 10.2307/2288636

CrossRef Full Text | Google Scholar

Efron, B., and Tibshirani, R. (1997). Improvements on cross-validation: the 632+ bootstrap method. J. Am. Stat. Assoc. 92, 548–560. doi: 10.1080/01621459.1997.10474007

CrossRef Full Text | Google Scholar

Geisser, S. (1975). The predictive sample reuse method with applications. J. Am. Stat. Assoc. 70:320. doi: 10.2307/2285815

CrossRef Full Text | Google Scholar

Hastie, T., Tibshirani, R., and Friedman, J. H. (2009). The Elements of Statistical Learning: Data Mining, Inference, and Prediction. 2nd Edn. New York: Springer.

Google Scholar

Hwang, H., Suk, H. W., Takane, Y., Lee, J., and Lim, J. (2015). Generalized functional extended redundancy analysis. Psychometrika 80, 101–125. doi: 10.1007/S11336-013-9373-X

PubMed Abstract | CrossRef Full Text | Google Scholar

Kim, S., and Hwang, H. (2021). Model-based recursive partitioning of extended redundancy analysis with an application to nicotine dependence among US adults. Br. J. Math. Stat. Psychol. 74, 567–590. doi: 10.1111/bmsp.12240

PubMed Abstract | CrossRef Full Text | Google Scholar

Kim, S., Lee, S., Cardwell, R. L., Kim, Y., Park, T., and Hwang, H. (2020). “An application of regularized extended redundancy analysis via generalized estimating equations to the study of co-occurring substance use Among US adults”, in Quantitative Psychology. IMPS 2019. Springer Proceedings in Mathematics & Statistics. Vol. 322. eds. M. Wiberg, D. Molenaar, J. González, U. Böckenholt, and J. S. Kim (Springer, Cham).

Google Scholar

Koul, A., Becchio, C., and Cavallo, A. (2018). Cross-validation approaches for replicability in psychology. Front. Psychol. 9:1117. doi: 10.3389/fpsyg.2018.01117

PubMed Abstract | CrossRef Full Text | Google Scholar

Lee, S., Kim, S., Kim, Y., Oh, B., Hwang, H., and Park, T. (2019). Pathway analysis of rare variants for the clustered phenotypes by using hierarchical structured components analysis. BMC Med. Genet. 12:100. doi: 10.1186/s12920-019-0517-4

PubMed Abstract | CrossRef Full Text | Google Scholar

Molinaro, A. M., Simon, R., and Pfeiffer, R. M. (2005). Prediction error estimation: a comparison of resampling methods. Bioinformatics 21, 3301–3307. doi: 10.1093/bioinformatics/bti499

PubMed Abstract | CrossRef Full Text | Google Scholar

Na, K. S. (2019). Prediction of future cognitive impairment among the community elderly: a machine-learning based approach. Sci. Rep. 9, 3335–3339. doi: 10.1038/s41598-019-39478-7

PubMed Abstract | CrossRef Full Text | Google Scholar

Shmueli, G. (2010). To explain or to predict? Stat. Sci. 25, 289–310. doi: 10.1214/10-STS330

CrossRef Full Text | Google Scholar

Smith, G. C. S., Seaman, S. R., Wood, A. M., Royston, P., and White, I. R. (2014). Correcting for optimistic prediction in small data sets. Am. J. Epidemiol. 180, 318–324. doi: 10.1093/aje/kwu140

PubMed Abstract | CrossRef Full Text | Google Scholar

Steyerberg, E. W., Harrell, F. E., Borsboom, G. J. J. M., Eijkemans, M. J. C., Vergouwe, Y., and Habbema, J. D. F. (2001). Internal validation of predictive models: efficiency of some procedures for logistic regression analysis. J. Clin. Epidemiol. 54, 774–781. doi: 10.1016/S0895-4356(01)00341-9

CrossRef Full Text | Google Scholar

Stone, M. (1974). Cross-Validatory choice and assessment of statistical predictions. J. R. Stat. Soc. 36, 111–147. doi: 10.2307/2984809

CrossRef Full Text | Google Scholar

Takane, Y., and Hwang, H. (2005). An extended redundancy analysis and its applications to two practical examples. Comput. Stat. Data Analy. 49, 785–808. doi: 10.1016/j.csda.2004.06.004

CrossRef Full Text | Google Scholar

U.S. Department of Health and Human Services, Substance Abuse and Mental Health Services Administration, and Center for Behavioral Health Statistics and Quality (2017). National Survey on Drug Use and Health 2016 (NSDUH-2016-DS0001). Available at: https://www.datafiles.samhsa.gov/study-dataset/national-survey-drug-use-and-health-2016-nsduh-2016-ds0001-nid17185

Google Scholar

U.S. Department of Health and Human Services, Substance Abuse and Mental Health Services Administration, and Center for Behavioral Health Statistics and Quality (2018). National Survey on Drug Use and Health 2017 (NSDUH-2017-DS0001). Available at: https://www.datafiles.samhsa.gov/study-dataset/national-survey-drug-use-and-health-2017-nsduh-2017-ds0001-nid17939

Google Scholar

U.S. Department of Health and Human Services, Substance Abuse and Mental Health Services Administration, and Center for Behavioral Health Statistics and Quality (2019). National Survey on Drug Use and Health 2018 (NSDUH-2018-DS0001). Available at: https://www.datafiles.samhsa.gov/study-dataset/national-survey-drug-use-and-health-2018-nsduh-2018-ds0001-nid18758

Google Scholar

U.S. Department of Health and Human Services, Substance Abuse and Mental Health Services Administration, and Center for Behavioral Health Statistics and Quality (2020). National Survey on Drug Use and Health 2019 (NSDUH-2019-DS0001). Available at: https://www.datafiles.samhsa.gov/study-dataset/national-survey-drug-use-and-health-2019-nsduh-2019-ds0001-nid19016

Google Scholar

Yarkoni, T., and Westfall, J. (2017). Choosing prediction over explanation in psychology: lessons from machine learning. Perspect. Psychol. Sci. 12, 1100–1122. doi: 10.1177/1745691617693393

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhang, W., Du, T., and Wang, J. (2016). “Deep Learning over Multi-Field Categorical Data,” in Advances in Information Retrieval. ECIR 2016. Lecture Notes in Computer Science. Vol. 9626. eds. N. Ferro, C. Fabio, M. Marie-Francine, M. F. C. Josiane, and M. D. N. Giorgio (Cham: Springer), 45–57.

Google Scholar

Keywords: out-of-sample model evaluation, predictive performance analysis, cross-validation, bootstrap, extended redundancy analysis

Citation: Kim S and Hwang H (2022) Evaluation of Prediction-Oriented Model Selection Metrics for Extended Redundancy Analysis. Front. Psychol. 13:821897. doi: 10.3389/fpsyg.2022.821897

Received: 25 November 2021; Accepted: 18 February 2022;
Published: 11 April 2022.

Edited by:

Holmes Finch, Ball State University, United States

Reviewed by:

Dong Seo, Hallym University, South Korea
Chester Chun Seng Kam, University of Macau, China

Copyright © 2022 Kim and Hwang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Sunmee Kim, c3VubWVlLmtpbUB1bWFuaXRvYmEuY2E=

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.