
94% of researchers rate our articles as excellent or good
Learn more about the work of our research integrity team to safeguard the quality of each article we publish.
Find out more
ORIGINAL RESEARCH article
Front. Child Adolesc. Psychiatry, 09 April 2025
Sec. Developmental Psychopathology and Mental Health
Volume 4 - 2025 | https://doi.org/10.3389/frcha.2025.1528196
In the context of parent-child interaction, the Emotional Availability Scales have been developed to capture a dyad's emotional connection in an observational setting by four parental and two child-related scales. This study aims to test the psychometric foundation of the EAS, including basic descriptive preconditions on the item level and structural validity on the scale level, for a preschool-aged sample; as such, it complements analyses by Aran for a sample of infants. The sample of parents and their preschool-aged children is a mixed clinically referred and non-clinical sample from a midsize city in Germany. Interactions were observed in a free-play setting and rated with the EAS by two blind and certificated raters. Several model tests indicate violations for the structural model as well as all six measurement models. An additional post hoc exploratory factor analysis with parallel analysis suggests a non-interpretable two-factor structure. Psychometric analyses did not validate the EAS's postulated structure and measurement model. A post hoc literature review showed that ceiling effects on the item and scale levels are not unique to our study. However, traditionally important concepts or terms of parent-child relationships covered by the EAS are not rejected by our study and can be assessed with alternative measures, but these also need psychometric evaluation in the future.
The Emotional Availability Scales (1) represent a widely known instrument (2) that promises to cover central aspects of a parent-child relationship by four parental scales (sensitivity, structuring, non-intrusiveness, and non-hostility) and two child scales [responsiveness and involvement (1)]. The EAS origin is based on terms coined by Ainsworth (3), such as maternal sensitivity/insensitivity, interference/cooperation, acceptance/rejection (4), and accessibility/inaccessibility. Assessing several scales is reasonable if each scale is informative in the sense of providing additional information not assessed by the remaining scales. This implies that the scales have to be sufficiently independent. In contrast to this assumption, studies have consistently reported relatively high to extremely high intercorrelations for the EAS (5). The sizes of theses correlations introduce doubt about whether the six scales are empirically discriminable. Because a psychometric analysis including exploratory or confirmatory factor analysis has not been reported yet, profound interpretation of the six-scale scores is not supported (6). Aran et al. (7) examined the postulated factor structure introduced by Biringen (8) and concluded that one latent factor underlies all six scales, which is in line with other findings (9).
In general, observational instruments are—in contrast to questionnaire-based assessments (10)—seldom examined with respect to their psychometric foundation (2, 11). One reason might be that such studies require greater time and personnel investments, which frequently result in smaller sample sizes, usually below N = 100 (2). According to older rules of thumb, sample sizes for factor analyses were generally recommended to exceed the number of variables multiplied by ten [12 cited in 13]; yet, these guidelines were not derived by theoretical or empirical knowledge and may prevent researchers from creating favorable statistical models. Meanwhile, several simulation studies (13, 14) have suggested that a minimum sample size of N = 60 is sufficient for a one-factor model assessed by seven items, like the single EAS scales, with low communalities (i.e., explained item variance by the factor ranging between .2 and .4). Similarly, testing the single scales of the structural model with the six scales used to measure an overall emotional availability score allows for an even smaller sample size. A greater sample size is need if one aims to test both the structural model along with all the single scales and accompanied items. However, simulation studies suggest that a multidimensional analyses for six independent factors measured by seven items each with high communalities (ranging between .6 and .8) require a sample size of only N = 110. Finally, because a smaller sample size would lower the power for rejecting a measurement or structural model, using a larger sample size is therefore a progressive approach to hypothesis testing, a larger sample size makes it easier to detect model violations.
Other psychometric characteristics like interrater reliability do not seem to be a core issue of the EAS (2, 15, 16). High internal consistency among the dimensions has also been repeatedly attested (17). However, high interrater reliability and high internal consistency are in agreement with a one-factor model, or, conversely, are results of a one-dimensional instrument.
Our intention of this psychometric analysis was to investigate the assumed underlying factor structure of the EAS based on emotional availability theory (8). We assumed that a child's age is a very important aspect in this matter, as parental behavior may depend on the child's level of development. For example, acting out structuring may be less important in a sample of infants and more important in a sample of preschool children. The results presented by Aran (7) may therefore be limited to infant samples. We therefore examined the factor structure of the parent and child scales, including the hypothesis of scale homogeneity on each of the six scales. This is preceded by an examination of item- and scale-based descriptive statistics and interrater reliabilities at item and scale level. If the factor structure is not supported by the data, we ask which structure is supported by an exploratory factor analysis. All planned tests are described in the Methods section below.
A summary of the structural and psychometric hypothesis is given in Figure 1. First, we tested the hypothesis about model fit of the overall model presented in Figure 1 (left side) and, therefore, included the structural hypothesis as well as the six measurement models of the EAS in one global testing approach. Because we expected model violation, we examined a number of submodels and started our second hypothesis with a proof of the structural hypothesis in Figure 1 (mid), which was based solely on the scale scores, and we assumed the measurement models were psychometrically sound. The measurement models in Figure 1 (mid), were tested subsequently to investigate unidimensionality for each scale.
Figure 1. Left side: Complete model including structural and measurement models of the emotional availability scales. Mid: Structural model only without measurement models. Right side: Separated measurement model of all single scales.
The study was approved by the Ethics Committee of the Medical Association of the Physicians Chamber Westfalen-Lippe (AZ: 2013-620-f). Informed Consent was obtained for all participants in this study. Approximately 80% of the contacted parents agreed to study participation. The EAS scales 4th edition is generally applicable without exclusion criteria related to mental health problems in the version for preschool-aged children (1). However sufficient skills in the German language were necessary. Note that our sample consisted of a clinical and a non-clinical sample within a cross-sectional study design. The clinical sample was recruited during their treatment in the Family Day Hospital for preschool children, which is part of the Clinic for Child and Adolescent Psychiatry, Psychosomatics and Psychotherapy of the University Hospital Muenster between December 17, 2015 and October 5, 2016. To be treated in the Family Day Hospital, a child has to have at least one diagnosis of a behavioral or emotional disorder according to the International Statistical Classification of Diseases and Related Health Problems 10th Revision (ICD-10), while exclusion from treatment in the Family Day Hospital occurs when a child has a pervasive developmental disorder (F84; World Health Organization, 1992) or a parent has an acute mental illness requiring inpatient treatment. A description of the clinical treatment approach and the intake population sample of the Family Day Hospital has been described previously (18, 19).
The non-clinical sample was approached in the Family Day Hospital (20) in January 18, 2016 and March 31, 2018. Mothers of the non-clinical group were recruited from several day care centers in Münster. After receiving the study information and giving informed consent, the mother completed several questionnaires, including the Child Behaviour Checklist [CBCL (21)]. Eligibility required a score below the clinical cut-off on the CBCL, and two dyads were excluded. The remaining mother and child attended the Family Day Hospital, and participating dyads were videotaped interacting in a playroom with age-appropriate toys, e.g., a wooden train set (tracks, trains, barriers), knights, animals, small dolls to play with together.
Overall, data from N = 116 parent-child dyads were analyzed in the current study, with the clinical sample accounting for n = 86. Of participating adults, 115 (99%) were female; parents' ages ranged between 18 and 56 years. The mean age of participating children was M = 4.61 years (SD = 1.61; 61.21% boys). It was a Western, educated, industrialized, rich, and democratic (WEIRD) sample. The most common types of main diagnoses at the time of admission were emotional disorders with onset specific to childhood (F93; 39%), other behavioral and emotional disorders with onset usually occurring in childhood and adolescence (F98; 24%), and conduct disorders (F91; 19%).
The participating dyads were recorded as they interacted in a free-play setting in a playroom, where age-appropriate toys, e.g., a wooden train set, were presented. To ensure observation of a natural and valid free-play interaction, dyads were instructed to play and enjoy their time together. Ten minutes of video material were then rated with regard to the observed emotional availability by two independent and, with respect to the clinical status of the children, blind raters. Raters completed a three-day online training as a requirement for application. The fee-based training costs $603 per person, and the obtained certificate is valid for two years. Both raters were certificated in terms of reliable application. The videos were rated in an individually randomized order by each rater between October 9, 2020 and August 18, 2022.
The EAS instrument is not based on counting specific behaviors but on a global judgement additionally informed by the observers' clinical opinion, according to the guidelines in the manual (8). Four parental scales aim to assess the following aspects: sensitivity (authentic warmth, responsiveness to the child's cues) structuring (providing security, support, and guidance), non-intrusiveness (following the child's lead, absence of invasive actions), and non-hostility (absence of overt and covert negativity). Two child-related scales aim to assess responsiveness (authentic positive affect, responsiveness to the adult) and involvement (positive initiation of interaction, balanced autonomy and inclusion). Each of the six scales is represented by seven items: The first two items are scored on a seven-point rating scale, while five complementary items are scored on a three-point rating scale. Therefore, the score of each scale, as the sum score of all seven items in a scale, ranges from seven to 29.
We calculated the basic descriptives of mean and standard deviation based on the original ratings. On this basis, interrater reliability on the item level was computed to examine the statistical precondition of any bi- and multivariate analysis. To weight each rater equally, the ratings were standardized prior to computing aggregated scores for all subsequent analyses. This standardization excludes the measurement error of rater differences in order to avoid the possibility that the model test fails not because of a deviation in the measurement structure, but because of rater influences, which could confound the test of structural validity. Note that the investigation of rater influences—which is beyond the scope of this analysis—is rather complex and can best be investigated after factorial validity has been confirmed. Nevertheless, the results should be interpreted bearing in mind that we have excluded rater differences that become relevant in clinical practice. The basic description of aggregated item means and standard deviations was repeated on the scale level consulting the aggregated scores. We additionally reported Cronbach's alpha.
We have not excluded any data and refer to the simulation studies mentioned in the introduction to ensure a sufficient minimum sample size. A preliminary analysis with a smaller sample was performed by the second author on February 27, 2021, and the results of this analysis along with the model specification in (7) led to a detailed conceptualization of the EAS structure and related model testing. Notably, our analysis includes model testing on the item level, which gives additional insight into the measurement model for each scale. The EAS coding material is copyright protected and is only assessable via Zeynep Biringen. Note that because of possible violation of copyright, Aran (7) did not report detailed results on the item level.
To examine the model's structural validity, we first tested the complete model (Figure 1, left side), after which we tested the path model (Figure 1, mid) of the structural model, assuming that the six measurement models (Figure 1, right side) are psychometrically sound. Subsequently, each of the six measurement models (Figure 1, right side) was tested by a confirmatory factor analysis to locate potential sources of model violations. The model testing was conducted with SAS PROC CALIS (SAS 9.4; Statistical Analysis Systems; CALIS = covariance analysis of linear structural equations) to evaluate model fit according to Schermelleh-Engel (22). The parameter estimation was conducted by maximum likelihood (ML) but repeated with SAS robust and OLS (ordinary least squares) estimation. Because the results from these two approaches essentially did not differ, only the results from the ML method are presented.
After model testing, we aimed to explore the cause of psychometric inconsistencies identified earlier. We, therefore, carried out a post hoc analysis to extract the number of factors determined by the 42 EAS items. We compared the observed eigenvalue of each factor with eigenvalues estimated based on random numbers by applying the SAS program for determining the number of components using parallel analysis by Velicer's minimum average partial test, which focuses on the common variance in a correlation matrix test, as done by O'Connor (23). In a second step, we described the pattern of item loadings according to the factors determined in step one, and we evaluated whether the new factors could be interpreted according to emotional availability theory (8).
Additionally, and with knowledge of ceiling effects, we conducted a further literature search to examine whether ceiling effects were unique to our data or were also reported in other publications. Finally, we computed the descriptive test score PDTS (probability of distinct test scores) (24, 25), which reports the average probability of distinguishing between two randomly selected test scores using the critical differences for each EAS dimension separately, as this probability can be lowered when ceiling effects are observed. In these cases, the PDTS constitutes a superior parameter compared to Cronbach's alpha, which can be high even if the test score distribution shows ceiling effects. As ceiling effects, in general, mean that a number of test scores in a sample are located in a small range, the PDTS performs a complete comparison on any two test scores in a sample and returns the number of significant comparisons divided by the number of all comparisons. Therefore, the PDTS score can vary between 0% and 100%, and it is lower when the test score distribution shows ceiling effects. We interpreted the PDTS index according to (26) as “very poor” < 30%; “poor” = 30%–45%; “moderate” = 45%–60%; “good” = 60%–75%; “very good” = 75%–90%, and “excellent” > 90%.
Table 1 shows the rating format, means, and standard deviations on the item level for rater A and rater B based on the original unstandardized scale scores and the Pearson correlation between rater A and B along with the part-whole corrected item-to-scale correlation. In Table 2 we present means and standard deviations of the original unstandardized scale scores on the scale level along with means and standard deviations for the clinical and non-clinical sample. These values reveal ceiling effects on the item level, especially for items with the three-point rating format, as well as on the sum score level. Table 2 additionally includes Cronbach's alpha, the PDTS score, and the correlation between the scale scores. Note that the reported correlations between scale scores are not attenuated by the scales' reliability, which may lead to an underestimation of these correlations. The attenuation is an inherent adjustment within an SEM approach.
Table 1. Emotional availability scales (EAS) item descriptives for 42 items, showing the range of answer formats, means, standard deviations, part-whole corrected item-to-scale correlation based on aggregated ratings from rater A and rater B, and correlation between raters A and B.
Table 2. Emotional availability scales (EAS) scale descriptives with range of scales from a minimum of 7 to a maximum of 29, sum scale scores for raters A and B, mean differences between A and B by t-test, correlation between sum scale scores for raters A and B, and correlation between the aggregate sum scale scores of raters A and B for the complete, clinical and community sample of N = 116.
In Table 3 we report model fit indices based on maximum likelihood estimates for the structural model as well as all tested measurement models depicted in Figure 1. Note that we replicated all statistical tests with robust ordinary least square parameter estimates in order to address the non-normally distributed data, but we did not observe differing results. Moreover, the estimation of path coefficients in the structural model exceeded, in general, a level of .80, indicating a solution with high communalities. The path coefficients in the measurement models varied considerably. For the sensitivity model, the path coefficients ranged from .54 to .96, for structuring from .37 to 1.00, for non-intrusiveness from .71 to .96, for non-hostility from −.06 to .89, for responsiveness from .42 to .92, and for involvement from .26 to .96; most path coefficients fell in the range of .75–.95. This, in turn, indicates, as mentioned in the simulation studies (13, 14), that our study had a sufficient number of observations for the intended model testing. Overall, all conducted tests revealed misfits and, therefore, suggest rejection of the models.
Table 3. Model fit indices of emotional availability models according Figure 1, including the complete, the structural and the six measurement models with N = 116 estimated by SAS PROC CALIS.
In order to evaluate post hoc whether the number of factors in the EAS model was mainly responsible for the model rejections, we conducted a parallel analysis that estimated eigenvalues based on random answers. Based on simulation studies, we computed Velicer's minimum average partial (MAP), which is the upper confidence interval (95th percentile) of randomly generated eigenvalues for a given sample size in combination with the number of items in a real data set, with SAS-Macro software (23). The comparison of eigenvalues based on random vs. real data is displayed in Figure 2, and it suggests that the EAS item covariation can be best described by two factors. To get an impression of the item loading pattern in a two-factorial solution, oblique (not orthogonal) rotation was used, which allows the factors to be correlated. This is justified if the first factor is considered to represent parental characteristics while the second factor contains the child characteristics, with both factors being expected to be correlated. However, the new first dimension consisted of 18 parental and two child items, while the new second dimension consisted of 12 parental and 12 child items (see Table 4). Given the theoretical superordinate distinction of parental and child items by Biringen, we were not able to interpret this pattern.
Figure 2. Exploratory factor analysis (EFA) scree plot on all 42 items of the emotional availability scales (solid line) in comparison with the scree plot of 42 items with random numbers according Velicer's minimum average partial (MAP), which is the upper confidence interval (95th percentile) of randomly generated eigenvalues for the same sample size.
This study examined the psychometric foundation of the 4th edition EAS (1). Our findings for a sample of preschool-aged children and their mothers were not in agreement with the underlying theoretical expectations. In its current version, none of the six scales—neither the superordinate parental, child-related, nor a global emotional availability factor—showed an acceptable model fit. Our interpretation that EAS scores lack a psychometric foundation is in agreement with Aran (7) and their results reported for a sample of infants. The main reason for the rejection of the model is revealed by the post hoc analysis using exploratory factor analysis, which identified only two factors instead of six. In addition, these two factors each included a mixture of both parental and child-related items. Thus, as the six scales of the EAS contain items that were assigned to two different factors, this led to the rejection of one-dimensional models. Note that the majorities of our observed commonalities were mainly between 0.6 and 0.8, which is considered a “wide” commonality condition [see Mundfrom (13)], and with 7 indicators per factor this requires a minimum sample size of N = 110 for 6 factors to be confirmed at an excellent level criterion (0.98)—which is the case for the EAS, and for a good level criterion (0.92) only N = 55 observations are required.
Our findings additionally explain the high correlations among the scales, which have previously been observed in many studies applying the EAS (27). The results should not be overinterpreted and do not generally call into question the terms related to parental sensitivity, structuring, intrusiveness, or hostility or to child responsiveness and involvement, nor do they question their testability or distinction. The shortcomings of the EAS are more likely to be attributed to the limitations of test development, in particular the generation of a large item pool, a repeated item selection process guided by factor and structural models to locate items on the intended content or, finally, scales of facets of emotional availability. The review by Lotzin (2) provides an overview of a number of alternative observation tools. However, Lotzin mentioned that the majority of observational measures have also not been developed with the help of psychometric methods, probably because old rules of thumb suggest large sample sizes before conducting a factor analysis; such sample sizes can hardly be achieved for observational studies given the effort needed to recruit clinical samples, which requires an enormous investment in personal and room facilities plus double-rated, blinded, and trained observers. A positive exception is the work of Wilson and Durbin (28), which was based on results of a factor analysis, did not require training, and assessed five parenting scales (involvement, positivity, hostility, intrusiveness, discipline). However, given the empirical and conceptual overlap with previous terms introduced from Baumrind (29), like “parental control,” as well as newer developments [see Grolnick, (30)] and additional constructs mentioned in Pritchett (10), experts in the field are still faced with a multitude of terms and construct and misses an integrated and consistent thesaurus. Therefore, many more studies on observational instruments in the field are needed to provide evidence regarding their internal structure (Standard 1.13, p. 26), their relationships with conceptually related constructs (Standard 1.16), and their relationships with criteria [Standard 1.17; all Standards defined by the American Educational Research Association, American Psychological Association, & National Council on Measurement in Education, (31)]. Once a researcher can use a structurally valid observation tool, many subsequent issues related to the application can be investigated, such as participant characteristics, settings, instructions, and the assessment procedure, including live or video-based observation, duration of observation, or training (32).
Our mean scale scores appeared around the upper quarter of the scale range, which is not a specific effect of the non-clinical participant. Such ceiling effects have been detected in other studies across different conditions, e.g., for shorter [5 min (33)] and longer observation times [25 min (7, 15)], across stressful and non-stressful situations (33), and with infants (7, 34) as well as toddlers (16, 35) and preschool-aged children (36). The identified effects are, therefore, not unique to our sample, setting, or duration time, and such effects also depend on the changing rating format across and within different versions of the EAS. Note that there is a certain robustness for test statistics to non-normality like ceiling effects in confirmatory factor analysis (37, 38), and all model tests were repeated with robust estimates in SAS PROC CALIS, which did not lead to differing results. In general, the EAS appears to be more appropriate for clinical samples and therefore we do not expect the relatively small number of non-clinical participants to limit the main conclusions of the study.
Do the ceiling effects limits the applicability of the EAS? A first impression of the internal consistency (Cronbach's alpha) seems to be sufficient and in a comparable range as reported in other publications (17). However, Cronbach's alpha does not reflect limitations related to ceiling effects, which became more visible when computing the alternative test score PDTS. In our sample, the average probability of distinguishing between two test scores ranged from “very poor” (22.2% for non-hostility) to “moderate” (54.54% for non-intrusiveness), which leads to a practical limitation in using the EAS scores in a clinical context.
Interrater reliability is another issue that affects the applicability of an observational tool while our main intention was the examination of the structural validity. Therefore our analysis were based on the raters aggregate and exclude this source of measurement error. However a short note is given to our results before we aggregated. We observed mean score differences between both raters for scores on sensitivity, non-intrusiveness, and responsiveness, along with correlations ranging from r = .62 (structuring) up to r = .71 (non-hostility). Such rater influences have been reported previously (39). The analysis of rater differences is rather complex and was handled by building an aggregate score to exclude this source of error from our intended analysis, which focused on item and scale properties.
The results presented in our study are limited by the clinical sample taken at admission from the Family Day Hospital for preschool-aged children in a child and adolescent psychiatry unit, and the non-clinical sample from a midsize city in Germany. Further, the recruiting procedure excluded severe developmental disorders. The observed scores were based on observation in a setting of free-play interaction for a limited duration of observation.
The Emotional Availability Scales (EAS) is a multidimensional observational instrument that is used to understand the emotional connection between parents and young children. Our analysis found statistical problems with the psychometric foundation scale scores, indicating that the suggested six scales are not supported by the data, corresponding to findings by Aran et al. (7). We found that a two-factor model, not a six-factor model as represented by the six scales of the EAS, fit best to the data, but the two-factor model was difficult to interpret. However, some important concepts covered by the EAS still seem valuable in clinical practice and can potentially be assessed using alternative measures.
The datasets presented in this study can be found in online repositories. The names of the repository/repositories and accession number(s) can be found below: Müller, J. M. (2024). EAS data from two raters for a sample of preschool aged children. https://doi.org/10.23668/psycharchives.14072; Müller, J. M. (2024). SAS EAS psychometric analysis, two raters, SEM model. https://doi.org/10.23668/psycharchives.14044.
The studies involving humans were approved by Ethics Committee of the Medical Association of the Physicians Chamber Westfalen-Lippe (AZ: 2013-620-f). The studies were conducted in accordance with the local legislation and institutional requirements. Written informed consent for participation in this study was provided by the participants' legal guardians/next of kin.
JM: Conceptualization, Data curation, Formal Analysis, Funding acquisition, Investigation, Methodology, Project administration, Resources, Software, Supervision, Validation, Visualization, Writing – original draft, Writing – review & editing. CE: Data curation, Formal Analysis, Investigation, Methodology, Software, Validation, Writing – original draft, Writing – review & editing, Resources.
The author(s) declare that no financial support was received for the research and/or publication of this article.
We would like to thank Celeste Brennecka for proofreading and English correction. An earlier version of this paper was presented as conference poster presentations at the DGKJP-Forschungstagung. Universität Duisburg-Essen, March 16, 2023–March 17, 2023.
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
The author(s) declare that no Generative AI was used in the creation of this manuscript.
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/frcha.2025.1528196/full#supplementary-material
1. Biringen Z. Emotional Availability (EA) Scales Manual, 4th Edition. Part 1: Infancy Early Childhood Version (child age: 0–5 Years). Boulder, Colorado (2008). Available at: emotionalavailability.com
2. Lotzin A, Lu X, Kriston L, Schiborr J, Musal T, Romer G, et al. Observational tools for measuring parent-infant interaction: a systematic review. Clin Child Fam Psychol Rev. (2015) 18(2):99–132. doi: 10.1007/s10567-015-0180-z
3. Ainsworth MDS, Blehar MC, Waters E, Wall S. Patterns of Attachment: A Psychological Study of the Strange Situation. Oxford: Lawrence Erlbaum (1978).
4. Rothenberg WA, Ali S, Rohner RP, Lansford JE, Britner PA, Di Giunta L, et al. Effects of parental acceptance-rejection on children’s internalizing and externalizing behaviors: a longitudinal, multicultural study. J Child Fam Stud. (2022) 31(1):29–47. doi: 10.1007/s10826-021-02072-5
5. Oppenheim D. Emotional availability: research advances and theoretical questions. Dev Psychopathol. (2012) 24(1):131–6. doi: 10.1017/S0954579411000721
6. Gridley N, Blower S, Dunn A, Bywater T, Whittaker K, Bryant M. Psychometric properties of parent-child (0–5 years) interaction outcome measures as used in randomized controlled trials of parent programs: a systematic review. Clin Child Fam Psychol Rev. (2019) 22(2):253–71. doi: 10.1007/s10567-019-00275-3
7. Aran P, Lewis AJ, Watson SJ, MacMillan KK, Power J, Galbally M. A psychometric study of the emotional availability scales: construct validity and measurement invariance between depressed and nondepressed mother–infant dyads. Psychol Assess. (2022) 34(1):70–81. doi: 10.1037/pas0001067
8. Biringen Z, Derscheid D, Vliegen N, Closson L, Easterbrooks MA. Emotional availability (EA): theoretical background, empirical research using the EA scales, and clinical applications. Dev Rev. (2014) 34(2):114–67. doi: 10.1016/j.dr.2014.01.002
9. Rodrigo MJ, León I, Góngora D, Hernández-Cabrera JA, Byrne S, Bobes MA. Inferior fronto-temporo-occipital connectivity: a missing link between maltreated girls and neglectful mothers. Soc Cogn Affect Neurosci. (2016) 11(10):1658–65. doi: 10.1093/scan/nsw080
10. Pritchett R, Kemp J, Wilson P, Minnis H, Bryce G, Gillberg C. Quick, simple measures of family relationships for use in clinical practice and research a systematic review. Fam Pract. (2011) 28(2):172–87. doi: 10.1093/fampra/cmq080
11. Prasad AH, Keevers Y, Kaouar S, Kimonis ER. Conception and development of the warmth/affection coding system (WACS): a novel hybrid behavioral observational tool for assessing parent-to-child warmth. Res Child Adolesc Psychopathol. (2023) 51(9):1357–69. doi: 10.1007/s10802-023-01055-y
13. Mundfrom DJ, Shaw DG, Ke TL. Minimum sample size recommendations for conducting factor analyses. Int J Test. (2005) 5(2):159–68. doi: 10.1207/s15327574ijt0502_4
14. Wolf EJ, Harrington KM, Clark SL, Miller MW. Sample size requirements for structural equation models: an evaluation of power, bias, and solution propriety. In: Boyle GJ, O’Gorman JG, Fogarty GJ, editors. Work and Organisational Psychology: Research Methodology; Assessment and Selection; Organisational Change and Development; Human Resource and Performance Management; Emerging Trends: Innovation/Globalisation/Technology, Vols 1–5. Thousand Oaks, CA: Sage Publications, Inc (2016). p. 177–97.
15. Kohlhoff J, Morgan S, Briggs N, Egan R, Niec L. Parent–child interaction therapy with toddlers in a community-based setting: improvements in parenting behavior, emotional availability, child behavior, and attachment. Infant Ment Health J. (2020) 41(4):543–62. doi: 10.1002/imhj.21864
16. Licata M, Zietlow A-L, Träuble B, Sodian B, Reck C. Maternal emotional availability and its association with maternal psychopathology, attachment style insecurity and theory of mind. Psychopathology. (2016) 49(5):334–40. doi: 10.1159/000447781
17. Derscheid DJ, Fogg LF, Julion W, Johnson ME, Tucker S, Delaney KR. Emotional availability scale among three US race/ethnic groups. West J Nurs Res. (2019) 41(3):409–30. doi: 10.1177/0193945918776617
18. Furniss T, Müller JM, Achtergarde S, Wessing I, Averbeck-Holocher M, Postert C. Implementing psychiatric day treatment for infants, toddlers, preschoolers and their families: a study from a clinical and organizational perspective. Int J Ment Health Syst. (2013) 7:1–12. doi: 10.1186/1752-4458-7-12
19. Müller JM, Averbeck-Holocher M, Romer G, Fürniss T, Achtergarde S, Postert C. Psychiatric treatment outcomes of preschool children in a family day hospital. Child Psychiatry Hum Dev. (2015) 46(2):257–69. doi: 10.1007/s10578-014-0465-3
20. Janßen M, Holodynski M, Müller JM, Reinersmann A, Romer G. Impaired predictability: enhanced fluctuations in the parenting behaviour of mothers of pre-school children with clinical diagnoses across three different play tasks. Eur Child Adolesc Psychiatry. (2019) 28(12):1645–58. doi: 10.1007/s00787-019-01330-8
21. Achenbach TM, Rescorla LA. Manual for the ASEBA Preschool Forms and Profiles. Burlington, VT: Research Center for Children, Youth, and Families, University of Vermont (2000).
22. Schermelleh-Engel K, Moosbrugger H, Müller H. Evaluating the fit of structural equation models: tests of significance and descriptive goodness-of-fit measures. MPR-online. (2003) 8(2):23–74. doi: 10.23668/psycharchives.12784
23. O’Connor BP. SPSS and SAS programs for determining the number of components using parallel analysis and velicer’s MAP test. Behav Res Methods Instr Comput. (2000) 32(3):396–402. doi: 10.3758/BF03200807
24. Müller JM. The probability of obtaining two statistically different test scores as a test index. Educ Psychol Meas. (2006) 66(4):601–11. doi: 10.1177/0013164405284034
25. Müller JM. SAS macros to compute the descriptive test index PDTS. Appl Psychol Meas. (2006) 30(4):345–6. doi: 10.1177/0146621606286207
26. Müller JM, Postert C, Beyer T, Furniss T, Achtergarde S. Comparison of eleven short versions of the symptom checklist 90-revised (SCL-90-R) for use in the assessment of general psychopathology. J Psychopathol Behav Assess. (2010) 32(2):246–54. doi: 10.1007/s10862-009-9141-5
27. Bornstein MH, Gini M, Putnick DL, Haynes OM, Painter KM, Suwalsky JTD. Short-term reliability and continuity of emotional availability in mother-child dyads across contexts of observation. Infancy. (2006) 10(1):1–16. doi: 10.1207/s15327078in1001_1
28. Wilson S, Durbin CE. The laboratory parenting assessment battery: development and preliminary validation of an observational parenting rating system. Psychol Assess. (2012) 24(4):823–32. doi: 10.1037/a0028352
29. Baumrind D. Current patterns of parental authority. Dev Psychol. (1971) 4(1 Pt. 2):1–103. doi: 10.1037/h0030372
30. Grolnick WS. The relations among parental power assertion, control, and structure. Hum Dev. (2012) 55(2):57–64. doi: 10.1159/000338533
31. Eignor DR. The standards for educational and psychological testing. In: Geisinger KF, Bracken BA, Carlson JF, Hansen J-IC, Kuncel NR, Reise SP, et al., editors. APA handbook of Testing and Assessment in Psychology, Vol 1: Test Theory and Testing and Assessment in Industrial and Organizational Psychology. Washington, DC: American Psychological Association (2013). p. 245–50.
32. Müller JM, Hoppermann S, Elvert C, Janssen M. Are emotionally competent adults emotionally available parents? Examination of theoretical and empirical relationships in a video-recorded parent-child interaction clinical preschool-aged sample. J Clin Psychol. (2023):1–16. doi: 10.1002/jclp.23510
33. Shakiba N, Doron G, Gordon HA, Egotubov A, Wagner NJ, Gueron SN. Mother-infant emotional availability through the COVID-19 pandemic: examining continuity, stability, and bidirectional associations. Infancy. (2023) 28(1):34–55. doi: 10.1111/infa.12517
34. Endendijk JJ, Groeneveld MG, Deković M, van den Boomen C. Short-term test–retest reliability and continuity of emotional availability in parent–child dyads. Int J Behav Dev. (2019) 43(3):271–7. doi: 10.1177/0165025419830256
35. Susman-Stillman A, Pleuss J, Englund MM. Attitudes and beliefs of family- and center-based child care providers predict differences in caregiving behavior over time. Early Child Res Q. (2013) 28(4):905–17. doi: 10.1016/j.ecresq.2013.04.003
36. Harkoma SM, Sajaniemi NK, Suhonen E, Saha M. Impact of pedagogical intervention on early childhood professionals’ emotional availability to children with different temperament characteristics. Eur Early Child Educ Res J. (2021) 29(2):183–205. doi: 10.1080/1350293X.2021.1895264
37. Curran PJ, West SG, Finch JF. The robustness of test statistics to nonnormality and specification error in confirmatory factor analysis. Psychol Methods. (1996) 1(1):16–29. doi: 10.1037/1082-989X.1.1.16
38. Micceri T. The unicorn, the normal curve, and other improbable creatures. Psychol Bull. (1989) 105(1):156–66. doi: 10.1037/0033-2909.105.1.156
Keywords: emotional availability scales, parent-child interaction, preschool age, psychometric analysis, assessment
Citation: Müller JM and Elvert C (2025) Psychometric analysis of the emotional availability scales for video-recorded interactions between parents and their preschool-aged children. Front. Child Adolesc. Psychiatry 4:1528196. doi: 10.3389/frcha.2025.1528196
Received: 14 November 2024; Accepted: 21 March 2025;
Published: 9 April 2025.
Edited by:
Niclas Schaper, University of Paderborn, GermanyCopyright: © 2025 Müller and Elvert. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Jörg Michael Müller, am9lcmdtaWNoYWVsLm11ZWxsZXJAdWttdWVuc3Rlci5kZQ==
Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.
Research integrity at Frontiers
Learn more about the work of our research integrity team to safeguard the quality of each article we publish.