- Psychology Department, Fordham University, New York, NY, United States
Many currently available effect size measures for mediation have limitations when the predictor is nominal with three or more categories. The mediation effect size measure υ was adopted for this situation. A simulation study was conducted to investigate the performance of its estimators. We manipulated several factors in data generation (number of groups, sample size per group, and effect sizes of paths) and effect size estimation [different R-squared (R2) shrinkage estimators]. Results showed that the Olkin–Pratt extended adjusted R2 estimator had the least bias and the smallest MSE in estimating υ across conditions. We also applied different estimators of υ in a real data example. Recommendations and guidelines were provided about the use of this estimator.
Introduction
Mediation analysis has enabled behavioral researchers to better understand the mechanistic relationships between variables. Many researchers are specifically interested in the role of mediators (M) that account for the relation between a predictor variable (X; independent variable) and an outcome variable (Y; dependent variable). A variety of effect size measures have been developed for mediation analysis. However, most of these effect size measures have limitations, including non-monotonicity and spurious inflation. Lachowicz et al. (2018) developed a new effect size measure, upsilon (υ), which has overcome these two limitations. The current study extends its work by applying this effect size metric to mediation models with a multicategorical predictor.
A simple mediation model shown in Figure 1A represents a mediation process in which a predictor X indirectly influences an outcome Y through a mediator M. This causal sequence suggests that X exerts a direct effect on M (a path), which, in turn, causally affects outcome Y (b path). X can have a direct effect on and Y (c′ path), in which X directly influences Y.
Figure 1. Path diagram of a simple mediation model and the total effect. (A) Shows the simple mediation model; (B) shows the total effect.
This model can be estimated by a set of regression models or by structural equation modeling when the effects are linear and M and Y are treated as continuous. Equations 1, 2 are required to estimate the effects of the a path and b path, respectively. The indirect effect of X on Y is the product of a and b. The direct effect is c′ (Eq 2). The total effect is c in Figure 1B and Eq 3, which equals the sum of X's direct and indirect effects on Y (Eq 4).
In this simple mediation model, predictor X can be dichotomous, or it can be treated as continuous. However, mediation analysis with a multicategorical predictor is common (Kalyanaraman and Sundar, 2006). When X is multicategorical, it can be expressed by applying coding strategies in regression analysis (Hayes and Preacher, 2014). When there are k groups comprising a multicategorical predictor X, (k − 1) coded variables are computed to represent each group. Different coding strategies are available, and the choice of strategy depends on the research question. Hayes and Preacher (2014) suggested using a dummy or contrast coding for a multicategorical predictor X in mediation analysis. In dummy coding, (k − 1) dummy variables (Di, i = 1, …, k − 1) are constructed, where Di is set to 1 to represent the cases in group i; otherwise, it is set to 0. The kth group is the reference group and is coded 0 in all Dis (see Table 1 for a three-group example). The effect of Di is the difference between the ith group and the reference group. In contrast coding, Di is constructed such that its effect represents the difference between the ith group and the average of the (i + 1)th group to the kth group (see Table 1 for a three-group example). A mediation model with a multicategorical predictor X is expressed in Figure 2A and Equations 5, 6.
Figure 2. Path diagram of a simple mediation model with a k-Group predictor and the total effect. (A) shows the simple mediation model; (B) shows the total effect.
Hayes termed aib the relative indirect effect and the relative direct effect. Equation 7 and Figure 2B show the relative total effect of X on Y, ci, which represents the sum of the corresponding relative direct and indirect effects (Equation 8). The effects are “relative” because they are based on comparisons between the group i and the other group (s), depending on the coding strategy of Di.
For statistical inference of the multicategorical mediation analysis, we can use the non-parametric bootstrapping method, Monte Carlo, or product of moments methods to calculate the confidence intervals of relative indirect effects (Hayes and Preacher, 2014). Creedon et al. (2016) developed an omnibus test of the indirect effect of X on Y through M. They suggest using R-squared (R2) to capture the overall effect of X on M from Equation 5 and construct the confidence interval for the product of this R2 and b to provide an omnibus test of the indirect effect. They tested the performance of this method using Smith, Wherry-1, Wherry-2, Olkin–Pratt, Pratt, and Claudy R2 shrinkage estimators for R2 (Creedon et al., 2016), which are supposed to produce less biased estimates for R2, and the non-parametric bootstrapping method for confidence intervals (R2 × b). While there is no mathematical proof that the product of R2 and b quantifies the overall mediation effect properly, this method has shown satisfactory results in maintaining the type I error rate at a set level (Creedon et al., 2016).
In addition to statistical inferences, reporting on effect sizes is highly encouraged or mandated by many peer-reviewed journals and organizations, including the American American Psychological Association (2010, p. 33) and the International Committee of Medical Journal Editors via the Consolidated Standard of Reporting Trials (CONSORT; Moher et al., 2010). Reporting on the effect size of an indirect effect is challenging because mediation analysis is a non-standard regression-based analysis, such that standardized mean differences, correlation coefficients, and proportions of variance explained are effect size metrics that are not sufficient to capture the entirety of an indirect effect (Lachowicz et al., 2018). Preacher (2011) have listed the desiderata for a good effect size measure: (a) An effect size should have an interpretable scale, (b) its confidence interval can be calculated, and (c) its sample estimation of the population parameter should be unbiased, consistent, and efficient. Wen (2015) have since added a desideratum that (d) the function of the effect size measure should be a monotonic representation, either in raw or absolute form, of the quantified effect.
In this study, we focus on a mediation effect size measure υ developed by Lachowicz et al. (2018), which is a modification of MacKinnon's (2008) effect size measure. MacKinnon (2008) recommends three proportions of variance-explained measures termed , , and . The subscripts (4.5), (4.6), and (4.7) are based on the original equation numbers in MacKinnon (2008), and these notations have continued to be referenced in subsequent literature (e.g., Lachowicz et al., 2018; Preacher, 2011). In the simple mediation model in Figure 1A, these effect sizes are calculated as follows (Equations 9–11):
is the squared correlation of Y and M. is the squared correlation of Y and X. is the squared multiple correlations of Y jointly explained by M and X. is the squared correlation between M and X. is the squared partial correlation of Y and M controlling for X. is the explained variance in Y jointly by M and X. is the proportion of variance in Y accounted for solely by M, weighted by the proportion of explained variance in M by X. is difficult to interpret on an R2 metric since it is the product of two proportions of variance from different variables (Preacher, 2011). is divided by , which is interpreted as the proportion of explained variance in Y jointly explained by M and X.
Among the three effect size measures mentioned previously, Lachowicz et al. (2018) modified and proposed a new measure υ to address the problem of spurious inflation. To illustrate the issue of spurious inflation, one needs to consider the elements of the simple mediation model in Equations 1–3. It is assumed that Y is dependent on M, and Y and M are mutually dependent on X, given that all other assumptions hold (temporal precedence, covariation of the cause and effect, etc.,). The zero-order correlation between M and Y has two components: (a) the conditional correlation between M and Y independent of X, and (b) the correlation due to the mutual dependence of Y and M on X. The first component is often referred to as true correlation, and the latter component is considered spuriously inflated. A true correlation between M and Y is needed to create a mediation effect size without spurious inflation (Lachowicz et al., 2018). Lachowicz et al. decomposed the correlation between Y and M (rYM):
where βa is the standardized a path. is the standardized c′ path. βb is the standardized b path, which captures the true correlation between M and Y in the mediation model. Lachowicz et al. pointed out that is the component that is spuriously inflated, indicating this is a key limitation of . If there is no indirect effect, either βb is zero (i.e., ), or the inflated term, , is zero (i.e., βa = 0; rYM = βb). If there is no direct effect, the spuriously inflated term is zero (i.e., rYM = βb). As a result, rYM cannot distinguish whether an indirect effect is present. Lachowicz et al. (2018) developed the effect size υ by removing the spurious inflation from ; υ measures the variance in Y explained jointly by M and X, correcting for the spurious inflation between M and Y on X (Equation 13). The term is the squared true correlation between M and Y. is the difference between the total variance in Y explained by M and X () and the total variance in Y explained solely by X (). Equation 14 is equal to Equation 13, which replaces with the squared from Equation 12. Equation 15 is also equivalent to Equations 13 and 14 (Lachowicz et al., 2018).
According to Lachowicz et al., υ has numerous desirable properties as an effect size measure of the indirect effect: (a) It is standardized (scale-free), (b) it is independent of sample size, (c) its function in the absolute value of the indirect effect is monotonic, and (d) its confidence interval can be constructed. Using Equation 15, Lachowicz et al. proposed sample estimators of υ for the simple mediation model (Figure 1A). The first estimator is where is the sample estimator. Their simulation study found that it was upwardly biased. Based on the relationship that , where E(•) is the expectation function, B is the unstandardized regression coefficient, and is the variance of , they propose the second estimator:
where and are the variances of â and , respectively. and are the sample variances of X and Y. The confidence interval of υ can be calculated via non-parametric bootstrapping. Their simulation study found that this estimator had an acceptable bias with only four experimental conditions resulting in percent relative biases > 5% at N = 100. The confidence interval of υ can be constructed using this estimator via non-parametric bootstrapping; their simulation showed that the bootstrapped confidence interval had an acceptable coverage rate (i.e., between 92.5 and 97.5%) at N = 250.
Current study
Because of the desirable properties of υ, we applied it in a simple mediation model with a multicategorical predictor (Figure 2A). We conducted a simulation study to investigate the performance of a sample estimator of υ. Using Equation 15, the υ for each relative indirect effect aib equals , where βai is the standardized ai path and βai = ai(σDi/σM). In addition, we believe that researchers would be interested in calculating υ for the overall indirect effect of X on Y through M. Equation 15 and its sample estimators are difficult to apply in this situation because of multiple ai paths; thus, we chose Equation 14 and proposed an estimator of υ:
Following Equation 16, we chose in the hope of getting a less biased estimate of . and are the shrinkage estimators of and , respectively. Following Creedon et al. (2016), we hoped that the shrinkage estimators would provide less biased estimates of and . Therefore, all the components in Equation 14 (i.e., population υ) were adjusted for small-sample biases in Equation 17.
Table 2 summarizes the formulas of the shrinkage estimators to adjust R2 and r2 (i.e., R2 in single-predictor regression). Based on the results from previous simulation studies on the performance of the shrinkage estimators (Raju et al., 1999; Yin, 2001; Walker, 2007; Wang and Thompson, 2007; Shieh, 2008; Creedon et al., 2016), Pratt, Ezekiel, Smith, Wherry-2, Walker, and Olkin–Pratt extended formulas performed well in at least one study. Nevertheless, none of these studies directly tested the mediation effect size measures. It is difficult to determine which formula is the best option for the sample adjustments of and when estimating υ. We conducted a simulation study to examine the following shrinkage estimators: Claudy, Ezekiel, Olkin–Pratt, Olkin–Pratt extended, Pratt, Smith, Walker, and Wherry.
Methods
A simulation study was conducted to evaluate the performance of sample estimator υ in Equation 17 under finite samples. The simulation was based on the mediation model in Figure 2A. Five factors were manipulated:
(1) A number of groups in X: k = 3, 4, 5. The conditions followed those in Creedon et al. (2016).
(2) Sample size per group, n = 10, 20, 25, 50, 100. The range of n covered small-to-large sample sizes across groups.
(3) Effect size of ai paths were manipulated as the mean difference between adjacent groups on M. The mean difference was set in terms of Cohen's d = 0, 0.2, 0.5, 0.8, representing null, small, medium, and large effects, respectively (Cohen, 1992).
(4) Size of b path, b = 0, 0.15, 0.39, 0.59. The conditions followed those in Lachowicz et al. (2018).
(5) Effect size of paths were manipulated as the mean difference between adjacent groups on Y. The mean difference was set in terms of Cohen's d = 0.1, 0.2, 0.3.
In all conditions, the residual of the mediator M, eM, was a normal variable with variance determined by , and the residual of the outcome Y, eY, was a standard normal variable. The simulation used a full factorial design with a total of 720 conditions (3 × 5 × 4 × 4 × 3). For each condition, 10,000 replications were created. For each condition and replication, nine and estimators were used to estimate υ: unadjusted, Claudy, Ezekiel, Olkin–Pratt, Olkin–Pratt extended, Pratt, Smith, Walker, and Wherry. The unadjusted sample estimates of υ were calculated using Equation 14, in which βb, , and were not adjusted. The rest of the estimators were calculated using Equation 17. The 95% confidence interval of υ was constructed using non-parametric bootstrapping with 1,000 bootstrap samples. The simulation study was conducted in R (Version 3.5.3; Windows system), and the packages boot 1.3–20 (Canty, 2017), dummies 1.5.6 (Brown, 2012), effsize 0.7.4 (Torchiano, 2018), and lm.beta 1.5–1 (Behrendt, 2014) were utilized.
To evaluate the performance of sample estimators of υ, the following outcomes were used: bias, standardized bias, mean squared error (MSE), and coverage rate. For any parameter θ, bias is the difference between the expectation of the sample estimates and the parameter (Equation 18).
In each condition, the population value of υ was calculated using Equation 14. The online Supplementary material present the population value of υ in each simulation condition.
Standardized bias is the bias divided by the standard deviation of the sample estimates (Equation 19). Standardized bias within ±0.4 can be regarded as acceptable (Collins et al., 2001).
Mean squared error is the expected squared difference between the sample estimates and the parameter. It is equal to the sum of the variance of sample estimates and squared bias (Equation 20). An unbiased sample estimator would produce an MSE equal to the variance of the estimator.
Coverage rate is the proportion of the sample in which the population parameter is contained within the 95% confidence interval across replications in a condition. An acceptable coverage rate is between 92.5 and 97.5% (Bradley, 1978).
We expected that the unadjusted would produce upwardly biased estimates, particularly for small sample size (N) and small effect size conditions. The bias of unadjusted would decrease with increasing N and effect size. For the shrinkage adjusted , it was expected that the estimates would have acceptable bias and that the changes in N would not change the bias. We expected that the MSE of the unadjusted would decrease as N increased since larger sample sizes decrease both sampling error and bias. For the same reason, we expected that the MSE of the shrinkage adjusted would also decrease as N increased. Finally, for both unadjusted and shrinkage adjusted , we expected that the coverage rate would approach the acceptable range as N increased.
We conducted one-way analyses of variance (ANOVAs) to study the effects of each manipulated factor (number of groups, sample size per group, effect size of ai paths, size of b path, and effect size of paths) on the bias, standardized bias, MSE, and coverage rate. For each significant factor, we conducted post-hoc pairwise comparisons with Tukey's HSD tests and produced boxplots by different conditions within the factor.
Results
Table 3 shows the bias, standardized bias, MSE, and coverage rate of using different R2 shrinkage method across a number of groups, group sizes, and sizes of ai, b, and paths. ANOVAs and post hoc pairwise comparison results of each estimator are provided in the online supplements. All the R2 shrinkage methods performed similarly. The Olkin–Pratt extended method had the least bias and the lowest MSE. The Smith method yielded the least standardized bias. However, the unadjusted method and all the R2 shrinkage methods produced very large standardized biases and were beyond the acceptable range (>0.4). None of the estimators produced satisfactory coverage rates (>0.9) across conditions. The unadjusted had the highest coverage rate among all the sample estimators. This finding was consistent with Lachowicz et al. (2018). Based on the results, we decided to focus on the Olkin–Pratt extended shrinkage method () hereafter. The online supplements provide the results of ANOVAs and post hoc pairwise comparisons of each manipulated factor on the bias, standardized bias, MSE, and coverage rate of .
Bias
The size of the b path had the strongest effect on the bias of , F(3, 716) = 448.22, p < 0.0001; η2 = 0.65. Figure 3A shows the boxplots of bias of by different b path conditions. The bias of at b = 0.59 was significantly higher than all the other b path conditions. The bias at b = 0.39 was also significantly higher than the b = 0.15 and b = 0 conditions. A higher effect of b path was associated with a more positive bias .
Figure 3. Boxplots of bias of grouped by (A) size of b path, (B) effect size of ai path, (C) effect size of path, and (D) number of groups.
The effect size of ai paths had the second strongest effect on the bias of , F(3, 716) = 39.13, p < 0.0001, η2 = 0.14. Figure 3B shows the boxplots of bias of by different effect sizes of ai paths. All pairwise comparisons were significant (ps < 0.05) except (d = 0.2 vs. d = 0). The bias of at d = 0.8 was significantly lower than d = 0.5, 0.2, and 0. Bias of was the lowest when d = 0.5. The effect size of paths were also affected by the bias of with F(2, 717) = 8.26, p < 0.001 η2 = 0.02. Figure 3C shows that the larger effect sizes of paths had a larger bias of . The number of groups had a small effect on the bias of , F(2, 717) = 3.20, p < 0.05, η2 = 0.009 (Figure 3D). The bias of at groups of k = 5 was significantly smaller than that at k = 3, and the bias was the lowest at k = 5. Group size had no significant effect on the bias of , F(4, 715) = 0.57, p = 0.69, η2 = 0.003.
Standardized bias
The size of the b path had an effect on the standardized bias of , F(3, 716) = 25.73, p < 0.0001, η2 = 0.10. had the most positive standardized bias when b = 0.59 with (Figure 4A). When b = 0.39, had more than 50% of cases with positive standardized bias > 0.4. When b = 0.15 and 0, the median standardized bias was within the ±0.4 criterion, yet some cases were beyond this range. The effect size of ai paths had an effect on the standardized bias of , F(3, 716) = 19.98, p < 0.0001, η2 = 0.08. Figure 4B shows that the standardized bias of decreased when the effect size of ai paths increased. When d = 0 and 0.2, the standardized bias of was > 0.4. When d = 0.5, the median standardized bias was < 0.4. When d = 0.8, the standardized bias of was the lowest and the median standardized bias was close to 0. The effect size of paths did not have a significant effect on the standardized bias of , F(2, 717) = 0.25, p = 0.78, η2 = 0.001.
Figure 4. Boxplots of standardized bias of grouped by (A) size of b path, (B) effect size of ai path, and (C) sample size per group. Due to the limit of y-axis, 6.53% of the observations are not shown in each box plot; the acceptable range of standardized bias (i.e., between −0.4 and 0.4) is indicated by the solid horizontal lines.
The number of groups had no significant effect on the standardized bias of , F(2, 717) = 2.84, p = 0.06, η2 = 0.008. Group size had a significant effect on the standardized bias of , F(4, 715) = 7.66, p < 0.0001, η2 = 0.04. Figure 4C shows that the standardized bias decreased when group size increased. The standardized bias of was the most positive when n= 10 (median > 0.4), and its standardized bias was significantly more positive than those in other conditions (ps < 0.05). The median standardized bias was >0.4 when group size = 25 and 100. When group size = 100 and 200, the median standardized bias was < 0.4.
Mean squared error
The size of the b path had a significant and large effect on the MSE of , F(3, 716) = 301.85, p < 0.0001, η2 = 0.56. MSE of increased when b increased (Figure 5A). The MSE at b = 0.59 was significantly higher than those in other conditions (b = 0.39, 0.15, 0; ps < 0.05), and the MSE at b = 0.39 was significantly higher than those at b = 0.15 and 0 (ps < 0.05). When b = 0 and 0.15, the MSE were close to zero. The effect sizes of ai and paths did not have significant effects on the MSE of [ai: F(3, 716) = 0.90, p = 0.44, η2 = 0.004; : F(2, 717) = 2.29, p = 0.10, η2 = 0.006].
Figure 5. Boxplots of MSE of grouped by experimental conditions. Due to limit of y-axis, 0.97% and 15.83% of the observations are not shown in (A) and (B), respectively.
The number of groups had no significant effect on the MSE of , F(2, 717) = 1.11, p = 0.33, η2 = 0.003. Group size had a significant effect on the MSE of , F(4, 715) = 23.18, p < 0.0001, η2 = 0.12. Figure 5B supports the hypothesis that the MSE of would decrease as group size increased in general. MSE of was the highest when group size = 10, and its MSE was significantly higher than those in other conditions (ps < 0.05). When group size = 25, 50, and 100, the median MSEs were close to zero.
Coverage rate
The size of the b path had a significant effect on the coverage rate of , F(3, 716) = 99.18, p < 0.0001, η2 = 0.29. All pairwise comparisons were significant (ps < 0.05). Figure 6A shows that the coverage rate decreased when b increased, and only when b = 0 did reach satisfactory coverage rate (>0.9). The effect size of ai paths had a significant effect on the coverage rate of , F(3, 716) = 2.24, p < 0.0001, η2 = 0.08. Figure 6B shows that coverage rate at d = 0 had a significantly lower coverage rate than all other conditions (d = 0.2, 0.5, 0.8; ps < 0.05). For d = 0.2, 0.5, and 0.8 conditions, did not have satisfactory coverage rates (< 0.9). The effect size of paths did not have a significant effect on the coverage rate of , F(2, 717) = 0.01, p = 0.99, η2 < 0.0001. None of the conditions for paths reached a majority of satisfactory coverage rates.
Figure 6. Boxplots of coverage rate of grouped by (A) size of b path, (B) effect size of ai path, and (C) sample size per group. The cutoff of satisfactory coverage rate (coverage rate > 0.9) is indicated by the solid horizontal line.
The number of groups had no significant effect on the coverage rate of , F(2, 717) = 1.83, p = 0.16, η2 = 0.005. Group size had a significant effect on the coverage rate of , F(4, 715) = 36.26, p < 0.0001, η2 = 0.17. Figure 6C shows that the coverage rate decreased when the group size increased. The coverage rate at group size = 10 was the highest and was significantly higher than those in other conditions (ps < 0.05). Group size = 10 reached the satisfactory cutoff with a mean coverage rate of 0.91. The coverage was the lowest at group size = 200.
Estimating υ without finite sample adjustment to βb
Results suggest that the performance of was influenced by the size of the b path. According to Equation 16, the finite sample adjustment of has two parts: (1) Adjusting βb and (2) adjusting and . and were unlikely to be influenced by the size of b path. Therefore, we conducted additional analyses which estimated υ using Equation 14 without finite sample adjustment to βb and with the Olkin–Pratt extended shrinkage method for and . The results are provided in the online supplements. Similar to the previous results of the with adjusted βb (Equation 16), the size of the b path had significant effects on the bias, standardized bias, and MSE of this estimator. We conclude that this estimator did not further improve .
Empirical example
We present an empirical example to help facilitate the interpretation of the υ effect size measure with a multicategorical predictor. The data were acquired through the NIMH-funded (PI: Rivera Mindt; K23MH079718) Medication Adherence Study at the Icahn School of Medicine at Mount Sinai (ISMMS) in New York City. We were interested in the mediation process between race/ethnicity and antiretroviral medication adherence via perceived racial and ethnic discrimination. There were 90 participants with completed measures. Race/ethnicity was a multicategorical variable with four groups: non-Hispanic white (group 1; n1 = 26), Hispanic (group 2; n2 = 37), Afro-Hispanic (group 3; n3 = 12), and others (group 4: n4 = 15). Dummy coding was used and non-Hispanic white was the reference group. The mediator, perceived discrimination (M1 = 5.04, SD1 = 1.34; M2 = 5.27, SD2 = 0.93; M3 = 5.25, SD3 = 0.97; M4 = 5.40, SD4 = 0.91), was measured by the Perceived Ethnic Discrimination Questionnaire—Community Version Brief (PEDQ-CVB, scaled from 1 to 5; Brondolo et al., 2005). The outcome variable, medication adherence (M1 = 77.44, SD1 = 19.26; M2 = 61.88, SD2 = 35.57; M3 = 61.90, SD3 = 32.44; M4 = 52.35, SD4 = 32.57), was calculated as the percentage (0 to 100%) of doses taken on schedule as assessed by the Medication Event Monitoring System (MEMS; Group AARDEX, 2022).
Table 4 presents the results for unadjusted s and finite sample adjusted s by correcting and different R2 shrinkage methods. Similar to simulation results, all R2 shrinkage methods were performed equally, with ranging from 0.146 to 0.147, meaning the variance of medication adherence explained by race/ethnicity via perceived discrimination was equal to 0.15. All adjusted s had similar 95% bootstrapped confidence intervals [−0.23, 1.55].
Discussion
The goal of this study was to extend the mediation effect size υ developed by Lachowicz et al. (2018) to mediation models involving a multicategorical predictor. Theoretically, υ has many desirable properties as a mediation effect size, particularly for its monotonicity, and bootstrapping can be used to construct its confidence interval. Their simulation showed that the unadjusted and finite sample adjusted sample estimators were consistent effect size measures. Furthermore, this effect size measure is standardized with an invariance of a linear transformation, so it is independent of the predictor scales, the mediator, and the outcome variable. We applied υ to a mediation model with a multicategorical predictor. In this scenario, our simulation results showed that υ did not retain some of the desiderata asserted by Lachowicz et al. (2018). Based on our results, the size of b path was the most important factor that negatively influenced the accuracy of the sample estimator of υ. In our data generation process, there were no large differences between the effect sizes of the a path and those of the b path. On further scrutiny, the performance of the sample estimator with uncorrected was similar to the sample estimator with the correction. Therefore, the large effect of b path remained undiscovered. R2 shrinkage methods produced slightly less biased υ estimates. The R2 shrinkage methods performed similarly on adjusting and . R2 shrinkage methods might not be the source of high values of bias, standardized bias, and MSE.
Since achieved higher performance among all the sample estimators of υ across all the experimental conditions, it is recommended that researchers use for simple mediation models involving a multicategorical predictor, as illustrated earlier. However, researchers should be cautious about the scenarios mentioned below:
(1) When the size of b path reaches 0.39–0.59, could become an upwardly biased effect size measure. A large b path (e.g., b = 0.59) could also result in a large value of MSE in .
(2) does not perform well and will be positively biased when ai paths have small effects (i.e., Cohen's d close or equal to 0).
(3) Researchers should pay attention to small group sizes. could have high standardized bias and high MSE when n = 10.
There are several limitations to this study. First, further analyses should be conducted on the relationship between the effect of the size of b path and the bias of the sample estimator of υ. Second, since this effect size measure is appropriate for the mediation model specified in this study, future studies should further develop this effect size measure to be suitable for more complex mediation models, such as models with multicategorical mediators or outcomes, models with moderations or latent variables, and models with multilevel or longitudinal data structures. Finally, since there are more rigorous assumptions that need to be made to justify an indirect effect as a causal effect, it is possible for researchers to introduce unknown bias into the estimation of υ. Future studies should investigate the performance of the estimator when such moderate violations of assumptions occur.
Data availability statement
The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.
Ethics statement
The studies involving human participants were reviewed and approved by NIMH-funded study (Grant# K23MH079718; Principal Investigator: M. Rivera Mindt) at the Icahn School of Medicine at Mount Sinai (ISMMS) https://reporter.nih.gov/search/XK3XEZYfbEuZQESg5r-YNA/project-details/7681039. The patients/participants provided their written informed consent to participate in this study.
Author contributions
All authors listed have made a substantial, direct, and intellectual contribution to the work and approved it for publication.
Funding
This research was supported by a K23 grant from NIMH (K23MH079718) and an R56 grant from the NIMHD (R56AG075744).
Acknowledgments
This research was supported by a K23 grant from the National Institutes of Mental Health (NIMH) (K23MH079718) and an R56 grant from the National Institute on Aging (NIA) (R56AG075744). All participants provided written informed consent, and the ISMMS Institutional Review Board approved the study protocol. The content is solely the responsibility of the authors and does not necessarily represent the official views of the National Institutes of Mental Health or the National Institute on Aging.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher's note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Supplementary material
The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fpsyg.2023.1101440/full#supplementary-material
References
American Psychological Association (2010). Publication manual of the American Psychological Association (6th ed.). Washington, DC: American Psychological Association.
Behrendt, S. (2014). lm.beta: Add standardized regression coefficients to lm-objects (R package version 1, 5–1). [Computer software]. The Comprehensive R Archive Network. Available online at: https://CRAN.R-project.org/package=lm.beta
Bradley, J. V. (1978). Robustness? Br. J. Math. Statistic. Psychol. 31, 144–152. doi: 10.1111/j.2044-8317.1978.tb00581.x
Brondolo, E., Kelly, K. P., Coakley, V., Gordon, T., Thompson, S., Levy, E., et al. (2005). The perceived ethnic discrimination questionnaire: development and preliminary validation of a community version. J. Appl. Soc. Psychol., 35, 335–365. doi: 10.1111/j.1559-1816.2005.tb02124.x
Brown, C. (2012). dummies: Create dummy/indicator variables flexibly and efficiently (R package version 1, 5.6) [Computer software]. The Comprehensive R Archive Network. Available online at: https://CRAN.R-project.org/package=dummies
Canty, A. (2017). boot: Bootstrap r (s-plus) functions (R package version 1, 3–20.) [Computer software]. The Comprehensive R Archive Network. Available online at: https://cran.r-project.org/web/packages/boot/index.html
Collins, L. M., Schafer, J. L., and Kam, C. (2001). A comparison of inclusive and restrictive strategies in modern missing data procedures. Psychol. Method. 6, 330–351. doi: 10.1037/1082-989X.6.4.330
Creedon, P. S., Hayes, A. F., and Preacher, K. J. (2016). Omnibus Tests of the Indirect Effect in Statistical Mediation Analysis With a Multicategorical Independent Variable. Available online at: http://www.afhayes.com/public/chp2016.pdf
Hayes, A. F., and Preacher, K. J. (2014). Statistical mediation analysis with a multicategorical independent variable. Br. J. Math. Statist. Psychol. 67, 451–470. doi: 10.1111/bmsp.12028
Kalyanaraman, S., and Sundar, S. S. (2006). The psychological appeal of personalized content in web portals: Does customization affect attitudes and behavior? J. Commun. 56, 110–132. doi: 10.1111/j.1460-2466.2006.00006.x
Lachowicz, M. J., Preacher, K. J., and Kelley, K. (2018). A novel measure of effect size for mediation analysis. Psychol. Method. 23, 244–261. doi: 10.1037/met0000165
MacKinnon, D. P. (2008). Introduction to Statistical Mediation Analysis. Routledge: Taylor and Francis.
Moher, D., Hopewell, S., Schulz, K., Montori, V., Gøtzsche, P. C., Devereaux, P. J., et al. (2010). Consort 2010 explanation and elaboration: updated guidelines for reporting parallel group randomized trials. Br. Med. J. 340, 698–702. doi: 10.1136/bmj.c869
Preacher, K. J. (2011). Effect size measures for mediation models: quantitative strategies for communicating indirect effects. Psychol. Methods 16, 93–115. doi: 10.1037/a0022658
Raju, N. S., Bilgic, R., and Edwards, J. E. (1999). Accuracy of population validity and cross-validity estimation: an empirical comparison of formula-based, traditional empirical, and equal weights procedures. Appl. Psychol. Meas. 23, 99–115. doi: 10.1177/01466219922031220
Shieh, G. (2008). Improved shrinkage estimation of squared multiple correlation coefficient and squared cross-validity coefficient. Organ. Res. Methods 11, 387–407. doi: 10.1177/1094428106292901
Torchiano, M. (2018). Effsize: efficient effect size computation (R package version 0, 7.4) [Computer software]. The Comprehensive R Archive Network. Available online at: https://CRAN.R-project.org/package=effsize
Walker, D. A. (2007). A comparison of eight shrinkage formulas under extreme conditions. J. Modern Appl. Stat. Methods 6, 162–172. doi: 10.22237/jmasm/1177992900
Wang, Z., and Thompson, B. (2007). Is the Pearson r2 Biased, and if So, What Is the Best Correction Formula? J. Exper. Educ. 75, 109–125. doi: 10.3200/JEXE.75.2.109-125
Wen, Z. (2015). Monotonicity of effect sizes: questioning kappa-squared as mediation effect size measure. Psychol. Methods, 20, 193–203. doi: 10.1037/met0000029
Keywords: effect size, mediation analysis, categorical predictor, simulation studies, R-squared
Citation: Cao Z, Cham H, Stiver J and Rivera Mindt M (2023) Effect size measure for mediation analysis with a multicategorical predictor. Front. Psychol. 14:1101440. doi: 10.3389/fpsyg.2023.1101440
Received: 17 November 2022; Accepted: 13 February 2023;
Published: 10 March 2023.
Edited by:
Jakob Pietschnig, University of Vienna, AustriaReviewed by:
Ulrich S. Tran, University of Vienna, AustriaHolly Patricia O'Rourke, Arizona State University, United States
Copyright © 2023 Cao, Cham, Stiver and Rivera Mindt. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Zihuan Cao, emNhbzIyJiN4MDAwNDA7Zm9yZGhhbS5lZHU=