Skip to main content

ORIGINAL RESEARCH article

Front. Psychol., 23 November 2023
Sec. Quantitative Psychology and Measurement

A new look at an old well-being construct: evaluating the psychometric properties of 9, 5, and 1-item versions of emotional exhaustion metrics

\r\nCaitlin L. PennyCaitlin L. Penny1Kathryn C. Adair,Kathryn C. Adair2,3Allan S. FrankelAllan S. Frankel4Michael W. LeonardMichael W. Leonard4Joshua ProulxJoshua Proulx4Paul J. Mosca,Paul J. Mosca5,6J. Bryan Sexton,*J. Bryan Sexton2,3*
  • 1Duke University School of Medicine, Duke University Health System, Durham, NC, United States
  • 2Duke Center for the Advancement of Well-being Science, Duke University Health System, Durham, NC, United States
  • 3Department of Psychiatry, Duke University School of Medicine, Duke University Health System, Durham, NC, United States
  • 4Vizient Safe and Reliable Healthcare, Evergreen, CO, United States
  • 5Duke Network Services, Duke University Health System, Durham, NC, United States
  • 6Department of Surgery, Duke University School of Medicine, Duke University Health System, Durham, NC, United States

Objective: To compare the relative strengths (psychometric and convergent validity) of four emotional exhaustion (EE) measures: 9- and 5-item scales and two 1-item metrics.

Patients and methods: This was a national cross-sectional survey study of 1409 US physicians in 2013. Psychometric properties were compared using Cronbach’s alpha, Confirmatory Factor Analysis (CFA), Exploratory Factor Analysis (EFA), and Spearman’s Correlations. Convergent validity with subjective happiness (SHS), depression (CES-D10), work-life integration (WLI), and intention to leave current position (ITL) was assessed using Spearman’s Correlations and Fisher’s R-to-Z.

Results: The 5-item EE scale correlated highly with the 9-item scale (Spearman’s rho = 0.828), demonstrated excellent internal reliability (alpha = 0.87), and relative to the 9-item, exhibited superior CFA model fit (RMSEA = 0.082, CFI = 0.986, TLI = 0.972). The 5-item EE scale correlated as highly as the 9-item scale with SHS, CES-D10, and WLI, and significantly stronger than the 9-item scale to ITL. Both 1-item EE metrics had significantly weaker correlation with SHS, CES-D10, WLI, and ITL (Fisher’s R-to-Z; p < 0.05) than the 5- and 9-item EE scales.

Conclusion: The 5-item EE scale was repeatedly found equivalent or superior to the 9-item version across analyses, particularly with respect to the CFA results. As there is no cost to using the briefer 5-item EE scale, the burden on respondents is smaller, and widespread access to administering and interpreting an excellent wellbeing metric is enhanced at a critical time in global wellbeing research. The single item EE metrics exhibited lower convergent validity than the 5- and 9-item scales, but are acceptable for detecting a signal of EE when using a validated EE scale is not feasible. Replication of psychometrics and open-access benchmarking results for use of the 5-tem EE scale further enhance access and utility of this metric.

Introduction

Healthcare worker (HCW) burnout is a global problem that negatively impacts wellbeing and quality of patient care, as highlighted by the COVID-19 pandemic (Shanafelt et al., 2010, 2019; Cimiotti et al., 2012; Kang et al., 2013; Welp et al., 2014; Wurm et al., 2016; Baier et al., 2018; Tawfik et al., 2018, 2019; Menon et al., 2020; Sexton et al., 2022b). Burnout is a complex construct with many contributing factors, thought to develop from an imbalance of high work demands and stressors (e.g., heavy workloads, difficult patient encounters, documentation requirements) with a lack of rewards and resources (e.g., helping others, recognition and income) (Demerouti et al., 2001; Lee and Mylod, 2019).

Healthcare worker burnout is associated with risks to patients through increased adverse events and patient mortality (Shanafelt et al., 2010; Cimiotti et al., 2012; Kang et al., 2013; Welp et al., 2014; Baier et al., 2018; Tawfik et al., 2018, 2019), and is associated with risks to the HCWs themselves through increased depression and suicidal ideation (Wurm et al., 2016; Menon et al., 2020). Burnout also increases financial costs for healthcare systems, physician turnover, and reduced productivity (Dewa et al., 2014; Shanafelt et al., 2016; Han et al., 2019; Adair et al., 2020b). With escalating demands on HCWs during the COVID-19 pandemic, concerns over the impact of burnout continue to increase (Bruyneel et al., 2021; Douglas et al., 2021; Haidari et al., 2021; Lasalvia et al., 2021; Macía-Rodríguez et al., 2021; Nguyen et al., 2021). As HCW wellbeing draws more attention and interventions are developed (Sexton and Adair, 2019; Adair et al., 2020b; Profit et al., 2021; Sexton et al., 2021b,2022a), the field suffers from a lack of consensus regarding rigorous assessment of “burnout,” which is evaluated using a variety of metrics and cutoff scores (Rotenstein et al., 2018).

After almost 50 years, the Maslach Burnout Inventory (MBI) is still considered the gold standard for assessing burnout among human services professionals, including HCWs (Maslach et al., 1997). It includes 22 questions and 3 subscales: emotional exhaustion (EE), depersonalization (DP), and personal accomplishment (PA). To date, EE is the most widely studied and reported component of the MBI and refers to being emotionally overextended and exhausted by the demands of work (Maslach et al., 2001). Compared to DP and PA, the 9-question EE subscale produces the largest and most consistent internal reliability estimates (Wheeler et al., 2011; Kleijweg et al., 2013; Loera et al., 2014), the highest test–retest reliability (Maslach et al., 1997), and has been demonstrated to be the only subscale adequately precise for individual-level measurement (Brady et al., 2020). EE also has the best ability to distinguish outpatients with an ICD-10 diagnosis of work-related neurasthenia or DSM-IV diagnosis of undifferentiated somatoform disorder, the closest clinical proxies to burnout (Schaufeli et al., 2001; Kleijweg et al., 2013).

Unfortunately, the length and cost of the MBI, which must be licensed and charges per use, limit its inclusion in health system surveys which often concurrently assess other important parameters (e.g., safety culture, engagement, and work-life integration) (Sexton et al., 2018). Assessing burnout rates and the impact of burnout interventions requires a valid, brief, and affordable metric that can be administered easily and quickly to busy HCWs while minimizing respondent burden. Longer surveys can cause fatigue among respondents and inaccurate results (Galesic and Bosnjak, 2009; Böckenholt and Lehmann, 2015; Liu and Wronski, 2018). As health systems increasingly focus on measuring and tracking burnout among HCWs, it can be a challenge to decide among the various metrics to administer, with several versions of the EE scale in particular, and a spectrum of cutoff scores used (Rotenstein et al., 2018). In order to facilitate efficient and accurate assessment of EE among populations of HCWs, abbreviated scales including 1-item (West et al., 2009; Li-Sauerwine et al., 2020) and 5-item (Sexton et al., 2018; Sexton and Adair, 2019; Profit et al., 2021) metrics have been developed. Notably, there are benefits to having scales with more than 1 item, including allowing for the assessment of internal reliability, a prerequisite of psychometrically valid assessments (Robinson, 2018).

A 5-item abbreviated version of the original 9-item EE scale has been carefully developed using items that are face valid and the scale has performed well psychometrically in prior research (Cronbach’s alphas ranging 0.84 to 0.92) (Adair et al., 2018, 2020a,2020c; Sexton et al., 2018; Schwartz et al., 2019; Sexton and Adair, 2019; Profit et al., 2021). It includes 4 items from the 9-item EE scale, in addition to a new item “events at work affect my life in an emotionally unhealthy way” (see Supplementary material for all items and response options). The response scale was also changed from a seven-point scale of frequency throughout the year ranging from “never” to “every day” (Maslach et al., 1997) to a five-point Likert scale ranging from “strongly disagree” to “strongly agree.” The transition from a seven-point to five-point scale was done to minimize respondent burden by maintaining consistency with other wellbeing and HCW-specific metrics that commonly use five-point scales (Sexton et al., 2006, 2018; Singer et al., 2007). Scales using responses on the agree-disagree spectrum can be applied to a wide array of constructs and are quickly and easily administered (Revilla et al., 2014). There is no significant difference in standard variation, skewness, or kurtosis between five-point and seven-point scales (Dawes, 2008), nor is there a difference in three key measures of response bias (Weijters et al., 2010). Despite the potential to increase reliability, validity, and discriminating power using seven-point response scales, the benefit of validated five-point scales is that they help to prevent respondents from being frustrated or demotivated, therefore improving response quality in settings of increased time constraint or longer surveys (Preston and Colman, 2000).

Given the importance of accurate and consistent EE assessment in contemporary healthcare, and the need for more clarity amidst various scales, items, response options and cutoffs, the current study aims to: (1) compare the psychometric properties of the 5-item, 9-item, and two 1-item EE metrics, and (2) assess the convergent construct validity of each EE scale with other important metrics used to assess wellbeing, including happiness, depression, work-life integration, and intention to leave current position. It is hypothesized that given the excellent psychometrics (Adair et al., 2018, 2020a,c; Sexton et al., 2018; Schwartz et al., 2019; Sexton and Adair, 2019; Profit et al., 2021) of the 5-item EE metric, it will have equivalent psychometric properties and higher convergent validity with the other metrics of interest relative to the 9-item, followed by the two 1-item metrics.

Materials and methods

Design and patient population

This study used cross-sectional survey data collected on a national sample of physicians, employed in a variety of hospital settings across the United States in 2013. Participants were members of a large healthcare system, contacted by email. They voluntarily responded to an anonymous electronic survey prior to initiating a continuing medical education (CME) activity to enhance wellbeing in clinical practice. In addition to emotional exhaustion, this wellbeing activity included assessments of happiness, depression, work-life integration, and intention to leave current position.

Demographic information

The survey captured gender, years in in current position, years of work experience, shift type, and name of facility where the HCW was employed.

Emotional exhaustion

Four measures of EE were captured in the survey: a 9-item, a 5-item, and two single item metrics. The 9-item EE scale from the MBI using a five-point Likert response scale (disagree strongly to agree strongly) was included (EE9item). The 5-item EE derivative of the 9-item version includes four items from the 9-item scale [(i.e., I feel…) (1) burned out, (2) fatigued, (3) frustrated, and (4) working too hard], and the additional item, “events at work affect my life in an emotionally unhealthy way” (EE5item). The EE5item has been tested in numerous large samples, has demonstrated good psychometric properties, and responsiveness to interventions (Adair et al., 2018, 2020a,c; Sexton et al., 2018; Schwartz et al., 2019; Sexton and Adair, 2019; Profit et al., 2021). Two single items of burnout were assessed. The first single item is “I feel burned out” using a five-point agree-disagree scale (EE1item5pt), and the second single item is “how often I feel burned out from my work” using a 7-point frequency response scale (EE1item7pt) (West et al., 2009). Table 2 lists the items included in each EE metric. For detailed information on scales, scoring, and benchmarking see the Supplementary material.

Responses to EE questions were averaged and rescaled from 0 to 100, with higher scores indicating more severe EE. Consistent with prior research, for five-point agree-disagree scales (EE9item, EE5item, EE1item5pt) a score of <50 indicates no EE (on average disagreeing slightly or strongly to all questions), 50–74 indicates mild EE (on average being neutral or agreeing slightly), 75–95 indicates moderate EE (on average agreeing slightly or strongly), and >95 indicates severe EE (agreeing strongly to all items). Scores ≥50 indicate concerning levels of EE, because the respondent is not disagreeing (i.e., neutral or higher) with the EE statements. To date, these cutoff scores have not been considered diagnostic, but rather to gauge the severity of EE present. The cutoffs provide an anchor for interpretation and a tool for communicating trends in the data. For the EE1item7pt metric, a cutoff of “once a week or more” is often used to indicate concerning levels (West et al., 2012; Brady et al., 2020; Li-Sauerwine et al., 2020). Responses to this item were also averaged and converted to a 0–100 scale to allow comparison to the five-point scales, with 66.67 on this scale indicating “once a week or more.”

Depression

The Center for Epidemiological Studies Depression Scale 10-item (CES-D10) was used to measure depressive symptoms (Andresen et al., 1994). It uses a 4-point frequency scale with the prompt “during the past week, how often did this occur?” followed by phrases such as “I could not get going.” Responses were summed for a scale ranging from 0 to 30, with a score ≥10 considered a concerning frequency of symptoms (Björgvinsson et al., 2013).

Subjective happiness

The Subjective Happiness Scale (SHS) is a validated, psychometrically sound, internationally used scale including 4 items on a seven-point scale, for example “In general I consider myself (1 = not a very happy person; 7 = a very happy person)” (Lyubomirsky and Lepper, 1999; Lin et al., 2010). Responses were averaged, with higher scores indicating higher subjective happiness. Prior studies have consistently demonstrated a mean SHS score of about 5 (Lyubomirsky and Lepper, 1999), so scores <5 were used to indicate concerning levels.

Work-life integration

Items included in the measurement of work-life integration (WLI) are from the work-life climate scale (Sexton et al., 2017) using the prompt “during the past week, how often did this occur?” with a 4-point frequency scale ranging from “less than 1 day” to “5–7 days a week.” Examples include: “skipped a meal,” “arrived home late from work,” and “slept less than 5 hours.” Responses were averaged with higher scores indicating greater work-life imbalance. Consistent with prior research, mean scores >2, indicating average work-life imbalance >2 days a week, were used to indicate concerning WLB (Sexton et al., 2017; Schwartz et al., 2019; Tawfik et al., 2021).

Intention to leave

To measure respondents’ intention to leave (ITL) their current position, 3 items using a five-point Likert scale ranging from “disagree strongly” to “agree strongly” were used (Sexton et al., 2019). Items included “I have plans to leave this job within the next year.” Higher intent to stay has been consistently shown to reduce job turnover (Brewer et al., 2012). Responses were averaged and converted to a scale ranging from 0 to 100, with higher scores indicating greater ITL. A threshold of ≥50 was used to indicate concerning ITL, as this is equivalent to responding “neutral” or higher, and not disagreeing with the questions.

Statistical analysis

Descriptive analyses were performed for demographic variables. For all metrics, “not applicable” responses were treated as missing, Respondents who left 2 or more EE questions blank were excluded from all analyses. If 1 or more question was left blank on one of the other metrics (CES-D10, SHS, WLI, ITL), then a score for that metric was not included in the analyses. Cronbach’s alpha was used to measure internal reliability, assessing how closely related the set of questions are as a group. Alphas range from 0 to 1, with a reliability coefficient of at least 0.70 being acceptable for early stage research, 0.80 for implementing cutoff scores, and 0.90 for individual assessment or if clinically important decisions are being made (Nunnally, 1978; Nunnally et al., 1994). A unidimensional Confirmatory factor analysis (CFA) with maximum likelihood estimation was used to test how well the data fit the hypothesized underlying model construct of EE. The following fit indices were used: root mean square error approximation (RMSEA) with adequate fit <0.08, Tucker-Lewis fit index (TLI) with adequate fit >0.95, confirmatory fit index (CFI) with adequate fit >0.95, and standardized root mean square residual (SRMR) with <0.08 considered adequate fit (Browne and Cudeck, 1992; Hu and Bentler, 1999).

Analyses similar to prior research evaluating the EE1item7pt measure were also used (West et al., 2009). Spearman’s rank correlation coefficients were calculated to assess the convergent relationships among the different EE metrics and the SHS, CES-D10, WLI, and ITL metrics. This evaluates how well the relationship between two scales can be described using a monotonic function. Additionally, Spearman’s coefficients were calculated comparing the single item EE metrics to the EE scales while excluding shared items. To evaluate concordance of responses to the various EE scales, respondents were grouped based on their response to the EE1item5pt and EE1item7pt metrics, and mean EE9item and EE5item scores were calculated for those groups.

Convergent construct validity for each EE metric was assessed by comparing correlations to SHS, CES-D10, WLI, and ITL metrics. The correlations between each EE metric and the other metrics of interest were compared by converting correlation coefficients into a z-score using Fisher’s R-to-Z transformation then testing the difference between the two dependent correlations (Lee and Preacher, 2013). Additionally, the percent of respondents scoring above the concerning threshold for SHS (<5), CES-D10 (≥10), WLI (>2), and ITL (≥50) at various thresholds for severity of EE was calculated for all EE metrics.

Analyses were performed using JMP version 15.1 (SAS Institute Inc, 2019) and Mplus version 8.5 (Muthén and Muthén, 2010). The study was approved by the Duke University Health System Institutional Review Board (Pro00063703).

Results

Of the 1836 physicians invited to participate, 1627 completed this survey (88.6% response rate). Of those, 120 (7.4%) did not consent to being in the research study and 98 (6.0%) left 2 or more EE questions blank and were excluded from analysis. Remaining results from 1409 participants were included in analyses. Participants reported working in over 200 different healthcare facilities and physician groups located in over 150 cities. Overall, 3.5% of participants did not report their clinical setting. The majority were male (64.4%). The most common responses for years in current position were 5–10 years (22.6%), 3–4 years (18.9%), and 1–2 years (15.6%). The most common responses for total years of professional experience were 21 + years (34.7%), 11–20 years (29.2%), and 5–10 years (21.8%). The majority worked day shifts (82.5%) with few working evening, night, or variable shifts (Table 1).

TABLE 1
www.frontiersin.org

Table 1. Respondent demographics (N = 1409).

Aim 1: compare the psychometric properties of the EE9item, EE5item, EE1item5pt, and EE1item7pt

After converting to a 0–100 scale, the mean scores and 95% CIs for the EE metrics in descending order are: EE5item 36.6 (35.2–38.0), EE1item5pt 34.3 (32.6–36.0), EE9item 31.6 (30.3–32.8), EE1item7pt 29.7 (28.4–31.0) (Table 2).

TABLE 2
www.frontiersin.org

Table 2. Spearman’s correlation coefficients between EE, SHS, CES-D10, WLI, and ITL scales with correlations of unshared items in parentheses.

Both the EE9item and EE5item demonstrated excellent internal reliability with Cronbach’s alpha of 0.91 and 0.87, respectively. The CFA for the EE9item revealed poor fit with RMSEA = 0.219 (90% CI 0.211–0.228), CFI = 0.790, TLI = 0.720, and SRMR = 0.088. The CFA for the EE5item demonstrated improved fit with RMSEA = 0.082 (90% CI 0.063–0.103), CFI = 0.986, TLI = 0.972, and SRMR = 0.018.

Spearman’s correlation coefficient for the full EE5item and EE9item was rho = 0.956, (p < 0.0001). Spearman’s correlation coefficient for the EE5item and EE9item with the 4 shared questions excluded (I feel fatigued, burned out, frustrated, working too hard), was rho = 0.828 (p < 0.0001). Spearman’s correlation coefficient for the EE1item5pt and the EE9item with that question excluded (I feel burned out), was rho = 0.812 (p < 0.0001), and compared to the EE5item with that question excluded, was rho = 0.802 (p < 0.0001). When comparing the EE1item7pt the Spearman’s correlation coefficient for the EE9item excluding “I feel burned out” was rho = 0.671 (p < 0.0001), and compared to the EE5item excluding the burnout question rho = 0.663 (p < 0.0001) (Table 2).

To determine how the EE severity for a respondent would vary based on the EE metric used, all respondents who answered the same level of severity on the EE1item5pt and EE1item7pt (for example all respondents who select “agree strongly” to feeling burned out) were grouped. Then, the average EE scores on the EE5item and EE9item were calculated for each of those groups. This demonstrated that severe EE as assessed through single item EE metrics corresponded to lower EE using the EE9item relative to the EE5item. For example, using EE1item5pt, the average respondents who “agrees strongly” that they were burned out from work (score = 100/100), had an average response of only “agree slightly” for the EE9item (score = 74.4/100), and fell between slight and strong agreement for the EE5item (score = 86.1/100). For the EE1item7pt, there was a similar effect, with respondents who reported feeling burned out “every day” (score = 100/100) on the EE1item7pt only responding on average “once a week” to “a few times a week” (score = 76.1/100) on the EE9item (Table 3). In comparison, those who reported feeling burned out “every day” (score = 100/100) on the EE1item7pt, reported an average score of 88.6/100 on the EE5item.

TABLE 3
www.frontiersin.org

Table 3. Translating single item EE scores to their corresponding 5 and 7 item scale scores.

Aim 2: assess the convergent construct validity of each EE metric for other important metrics assessing wellbeing including happiness, depression, work-life integration, and intention to leave current position

For completeness, Cronbach’s alphas were calculated for the other wellbeing metrics and were as follows: SHS = 0.83; CES-D10 = 0.80; WLI = 0.83; ITL = 0.87. Spearman’s correlation coefficients were calculated (Table 2) and used for the test of difference between two dependent correlations using Fisher’s transformation (Table 4). Results demonstrate that the EE5item is more strongly correlated with SHS, WLI, and ITL relative to the other EE metrics, but only significantly stronger than the EE9item for ITL. The EE9item and EE5item have significantly stronger correlations with all other metrics than the EE1item5pt and EE1item7pt. The EE1item5pt is significantly stronger correlated with SHS and ITL than EE1item7pt.

TABLE 4
www.frontiersin.org

Table 4. Z-scores and p-values for two-tailed test of the difference between two dependent correlations for each EE metric and scores for SHS, CES-D10, WLI, and ITL using Fisher’s transformation.

The ability of various EE cutoff values to identify concerning levels on the SHS, CES-D10, WLI, and ITL metrics is presented in Table 5. Results indicate that there is a large increase in CES-D10, WLI, and ITL scores when comparing those who endorse moderate vs. severe EE for both the EE9item and EE5item.1

TABLE 5
www.frontiersin.org

Table 5. Proportion of respondents reporting concerning levels on the SHS, CES-D10, WLI, and ITL scales at various thresholds on the EE metrics.

Discussion

Concern about HCW emotional exhaustion due to the fast paced and high intensity nature of working in healthcare has been elevated by the COVID-19 pandemic (Sexton et al., 2022b; Shanafelt et al., 2022). Healthcare systems are increasingly motivated to assess and reduce the level of EE in the workforce. Having brief, no-cost, reliable and validated metrics of EE is essential to minimize respondent burden and facilitate inclusion in larger health system surveys evaluating a variety of safety culture, engagement, and wellbeing domains. This evaluation of the EE9item, EE5item, EE1item5pt, and EE1item7pt metrics in a large sample of HCW across the country revealed that the EE5item has strong psychometric reliability and construct validity. The EE1item5pt and EE1item7pt do not exhibit as strong convergent construct validity as the EE5item (i.e., they exhibited significantly weaker correlations to all other outcomes), however, in situations where a quick screening assessment is needed for a signal of EE, they can be used with relative confidence (with a slight advantage for EE1item5pt over EE1item7pt). A guide to using, scoring and interpreting the EE5item, including verbatim items, instructions and benchmarking results from a sample of over 30,000 healthcare workers can be found in the Supplementary material.

The EE5item is equivalent or superior to the EE9item in several ways. The internal reliability (alpha = 0.87) was only slightly lower than the EE9item (alpha = 0.91), despite the fact that higher numbers of items artificially inflate the alpha. The EE5item alpha was consistent with previous alphas reported (0.84 to 0.92) (Adair et al., 2018, 2020a,c; Sexton et al., 2018, 2022b; Schwartz et al., 2019; Sexton and Adair, 2019; Profit et al., 2021). These values rest solidly in the >0.80 range acceptable for implementing cutoff scores and border the >0.90 threshold for clinically important decisions (Nunnally, 1978; Nunnally et al., 1994). Cronbach’s alpha can be artificially inflated by either increasing the number of items so that the quantity overpowers the quality of items, or reducing the number of items in a scale to select only the most highly correlated items (Kopalle and Lehmann, 1997; Cho and Kim, 2015). Considering this, the CFA adds more granularity to how the two scales model the underlying construct of EE. The CFA fit indices for the EE5item indicated much better fit overall than for the EE9item, in which none of the indices indicated acceptable model fit (Browne and Cudeck, 1992; Hu and Bentler, 1999). The exploratory factor analysis (EFA) results demonstrate that EE emerges as a distinct construct from depression when using both the EE9item and EE5item (see Supplementary material). The EFA also demonstrated that the two items “working with people directly puts too much stress on me” and “working with people all day is really a strain for me,” loaded on their own factor. These items are part of the EE9item, but not included in the EE5item.

Previous EFAs and CFAs of the EE9item have shown the two “stress” and “strain” items to be particularly problematic, because they are very similar and locally dependent (Worley et al., 2008; Mukherjee et al., 2020). Both of these items also strongly load on the DP factor of burnout rather than EE (Poghosyan et al., 2009). Item response theory analyses also revealed that higher EE symptom severity is required to endorse the “stress” or “strain” items (a z-score > 1.57 standard deviations above the mean), or “I feel like I am at the end of my rope” (z-score > 1.0 standard deviations above the mean) (Brady et al., 2020). These “stress,” “strain,” and “rope” questions are not included in the EE5item. The high symptom severity required to endorse those EE9item questions also helps explain why the mean score is lower for the EE9item (31.6) than the EE5item (36.6). Analyses further demonstrated that the EE5item correlates as well as the EE9item with SHS, WLI, and CES-D10 and significantly stronger with ITL. It is overall more sensitive and less specific than the EE9item for detecting concerning levels on these other metrics, making it a useful screening tool.

The EE9item and EE5item predicted the 4 other wellbeing metrics better than the single item EE measures, and the EE5item was slightly better than the EE9item. Correlations among both single item measures and every other metric (SHS, CES-D10, WLI, and ITL) are significantly weaker than the EE9item and EE5item. This demonstrates the value of using an EE metric with multiple items to capture the full range of the underlying dimension of EE (Robinson, 2018). Although the 1-item questions can be administered to quickly gauge a signal of EE in a population, they do not detect these other important wellbeing outcomes as effectively.

The EE1item7pt had the lowest mean EE score (29.7), likely because most respondents (43.6%) endorsed feeling burned out “a few times a year or less,” the second lowest frequency, and those who selected lower frequencies had on average a higher agreement with EE on the EE5item and EE9item. The seven-point response scale inherently does not assess current burnout, but rather its frequency across a whole year. It is possible for a respondent to report burnout frequency as “a few times a year or less” yet be severely burned out when they complete the survey. This is an important consideration when selecting which metric to administer. The EE1item7pt also correlates weaker than the EE1item5pt to the EE5item (0.66 vs. 0.80, respectively) and the EE9item (0.67 vs. 0.81, respectively).

A variety of thresholds have been used for measurements of burnout and EE (Rotenstein et al., 2018), and they can be useful to identify a subset of respondents who may require additional support or intervention. Even without a consensus clinical definition for burnout syndrome, it is important to detect which HCWs may be dealing with more exhaustion and emotional burden at work, provide resources and track trends over time. Thresholds such as “% emotionally exhausted” are useful in debriefings and for summarizing results for those less accustomed to dealing with survey data. Individuals may find it easier to grasp that, for instance, 59% of their unit is experiencing emotional exhaustion, compared to hearing that the mean EE score for the group is 3.1. An alternative to thresholds is to use measures of central tendency, such as means, to track trends in data over time. This can also be valuable to assess how a population is changing, because respondents who score just two points away from each other might be identified as clinically different even though in reality their EE level is virtually the same. We recommend reporting the mean and the percent for clarity. According to the thresholds for EE used here, a majority of participants reporting severe EE on the EE9item and EE5item have concerning levels of depression, work-life imbalance, subjective happiness and ITL their current position.

Ultimately when deciding which metric for EE to administer, the EE5item has optimal psychometrics and convergent construct validity. Compared to the 9-item version, the EE5item is free to administer, less burdensome for respondents, has improved model fit for EE, and has either equivalent or superior correlation with other important metrics: SHS, CES-D10, WLI, and ITL. Particularly with a move among health systems to administer shorter surveys more frequently in “pulses” rather than longer surveys with more time between them, identifying valid and abbreviated EE metrics to gauge burnout is crucial. The questions retained in the EE5item (I feel burned out, fatigued, frustrated, working too hard) as well as the newly developed item (events at work) were chosen because they are more representative of the overall EE construct, as reflected in the improved CFA model fit compared to the EE9item. One thing to keep in mind is that respondents on average score slightly higher EE on the EE5item, likely because several items from the EE9item that require severe EE to endorse have been removed. For a screening tool, the results from the EE5item are the sought after criteria. The EE1item5pt and EE1item7pt do not have as strong convergent construct validity for other important metrics, but can be used to quickly assess or track trends in the EE level of a population.

This study is limited by the age of the dataset, which was collected in 2013. For this reason, the level of burnout in the sample is not timely (EE is worse now) (Sexton et al., 2022b). Nevertheless, the psychometric analyses and convergent validity correlations set forth as aims should not be affected by the year data were collected. Recent EE5item results from over 30,000 HCWs in 2021/2022 show the reliability of the scale was above 0.90 (Supplementary material). There is little reason to believe that relationships between items within scales or associations among the four EE metrics and other wellbeing outcomes would be significantly different based on the age of the data. Another limitation is that five-point response options were used for the EE9item, whereas the original 9-Item EE scale from the MBI uses seven-point response options. Nonetheless, all survey items were the same. The study’s respondents were from diverse healthcare settings across the United States, therefore the results may be fairly generalizable. However, given these participants were physicians, it is possible that the pattern of results could differ in other HCW roles, therefore the generalizability to other groups is not yet clear. The study is also limited by its use of self-report data which poses a risk of response, selection, and social desirability biases. The high response rate and use of psychometrically validated scales helped to minimize these biases. Finally, because this is a cross-sectional study, test–retest reliability could not be examined. Nevertheless, the responsiveness of the EE5item to interventions (Sexton and Adair, 2019; Adair et al., 2020a,c, 2022; Profit et al., 2021; Sexton et al., 2021a,2022a) and changing work stresses (Sexton et al., 2022b), strongly suggests that changes in scores very consistently move in the predicted direction and magnitude.

Conclusion

There is considerable interest in assessing EE, a key component of HCW wellbeing, using brief and valid metrics. The current study evaluated the psychometric properties of four measures of EE (EE9item, EE5item, EE1item5pt, and EE1item7pt). The EE5item was equivalent or superior to the EE9item across analyses. EE5item has no cost, is easy to administer and interpret, and is a reliable metric for assessing EE. The decreased length reduces respondent burden and is of great value when assessing healthcare workers to make clinical and operational decisions. The lack of cost associated with this metric opens researchers to include a brief, reliable scale to assess emotional exhaustion during a time that wellbeing research is increasingly important. EE1item5pt and EE1item7pt did not correlate as well with happiness, depression, WLI, and ITL, however, they can serve as a proxy for the longer scales to detecting a signal of EE, when administering validated scales is not feasible.

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Ethics statement

The studies involving humans were approved by the Duke University Health System Institutional Review Board. The studies were conducted in accordance with the local legislation and institutional requirements. The Ethics Committee/Institutional Review Board waived the requirement of written informed consent for participation from the participants or the participants’ legal guardians/next of kin because of the low-risk level of the data and its deidentification.

Author contributions

CP: Formal analysis, Investigation, Methodology, Writing – original draft, Writing – review and editing. KA: Formal analysis, Investigation, Methodology, Project administration, Supervision, Writing – review and editing. AF: Conceptualization, Data curation, Writing – review and editing. ML: Conceptualization, Writing – review and editing. JP: Data curation, Writing – review and editing. PM: Resources, Supervision, Writing – review and editing. JS: Conceptualization, Data curation, Investigation, Methodology, Supervision, Writing – review and editing.

Funding

The author(s) declare financial support was received for the research, authorship, and/or publication of this article. The authors declare that this study received funding from Vizient Safe and Reliable Healthcare. The funder was not involved in the study design, collection, analysis, interpretation of data, the writing of this article, or the decision to submit it for publication.

Conflict of interest

AF, ML, and JP are employed by Vizient Safe and Reliable Healthcare.

The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fpsyg.2023.1267660/full#supplementary-material

Footnotes

  1. ^ Divergent validity was assessed using Exploratory Factor Analyses (EFAs). Since depression is a similar but distinct construct from EE, we conducted two EFAs to test whether the depression items (CES-D10 items) loaded onto different factors than the items from the EE5item and EE9item. Indeed, both EFAs revealed that none of the depression items loaded on the same factor as any of the EE items, indicating discriminant validity. These results are described in detail in the Supplementary file (section VI).

References

Adair, K., Heath, A., Frye, M., Frankel, A., Proulx, J., Rehder, K., et al. (2022). The psychological safety scale of the safety, communication, operational, reliability, and engagement (SCORE) Survey: A brief, diagnostic, and actionable metric for the ability to speak up in healthcare settings. J. Patient Saf. 18, 513–520. doi: 10.1097/PTS.0000000000001048

PubMed Abstract | CrossRef Full Text | Google Scholar

Adair, K., Kennedy, L., and Sexton, J. (2020a). Three good tools: Positively reflecting backwards and forwards is associated with robust improvements in well-being across three distinct interventions. J. Posit. Psychol. 15, 613–622. doi: 10.1080/17439760.2020.1789707

PubMed Abstract | CrossRef Full Text | Google Scholar

Adair, K., Rehder, K., and Sexton, J. (2020b). “How Healthcare Worker Well-Being Intersects with Safety Culture, Workforce Engagement, and Operational Outcomes,” in Connecting Healthcare Worker Well-Being, Patient Safety and Organisational Change: The Triple Challenge, eds A. Montgomery, M. van der Doef, E. Panagopoulou, and M. Leiter (Cham: Springer International Publishing).

Google Scholar

Adair, K., Rodriguez-Homs, L., Masoud, S., Mosca, P., and Sexton, J. (2020c). Gratitude at Work: Prospective cohort study of a web-based, single-exposure well-being intervention for health care workers. J. Med. Internet Res. 22:e15562. doi: 10.2196/15562

PubMed Abstract | CrossRef Full Text | Google Scholar

Adair, K., Quow, K., Frankel, A., Mosca, P., Profit, J., Hadley, A., et al. (2018). The Improvement Readiness scale of the SCORE survey: a metric to assess capacity for quality improvement in healthcare. BMC Health Serv. Res. 18:975. doi: 10.1186/s12913-018-3743-0

PubMed Abstract | CrossRef Full Text | Google Scholar

Andresen, E., Malmgren, J., Carter, W., and Patrick, D. (1994). Screening for depression in well older adults: Evaluation of a short form of the CES-D (Center for Epidemiologic Studies Depression Scale). Am. J. Prev. Med. 10, 77–84.

Google Scholar

Baier, N., Roth, K., Felgner, S., and Henschke, C. (2018). Burnout and safety outcomes - a cross-sectional nationwide survey of EMS-workers in Germany. BMC Emerg. Med. 18:24. doi: 10.1186/s12873-018-0177-2

PubMed Abstract | CrossRef Full Text | Google Scholar

Björgvinsson, T., Kertz, S., Bigda-Peyton, J., McCoy, K., and Aderka, I. (2013). Psychometric properties of the CES-D-10 in a psychiatric sample. Assessment 20, 429–436. doi: 10.1177/1073191113481998

PubMed Abstract | CrossRef Full Text | Google Scholar

Böckenholt, U., and Lehmann, D. (2015). On the limits of research rigidity: the number of items in a scale. Mark Lett. 26, 257–260.

Google Scholar

Brady, K., Ni, P., Sheldrick, R., Trockel, M., Shanafelt, T., Rowe, S., et al. (2020). Describing the emotional exhaustion, depersonalization, and low personal accomplishment symptoms associated with Maslach Burnout Inventory subscale scores in US physicians: an item response theory analysis. J. Patient-Rep Outcomes 4:42. doi: 10.1186/s41687-020-00204-x

PubMed Abstract | CrossRef Full Text | Google Scholar

Brewer, C., Kovner, C., Greene, W., Tukov-Shuser, M., and Djukic, M. (2012). Predictors of actual turnover in a national sample of newly licensed registered nurses employed in hospitals. J. Adv. Nurs. 68, 521–538. doi: 10.1111/j.1365-2648.2011.05753.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Browne, M., and Cudeck, R. (1992). Alternative ways of assessing model fit. Sociol. Methods Res. 21, 230–258.

Google Scholar

Bruyneel, A., Smith, P., Tack, J., and Pirson, M. (2021). Prevalence of burnout risk and factors associated with burnout risk among ICU nurses during the COVID-19 outbreak in French speaking Belgium. Intensive Crit. Care Nurs. 16:103059. doi: 10.1016/j.iccn.2021.103059

PubMed Abstract | CrossRef Full Text | Google Scholar

Cho, E., and Kim, S. (2015). Cronbach’s Coefficient Alpha: Well Known but Poorly Understood. Organ. Res. Methods 18, 207–230.

Google Scholar

Cimiotti, J., Aiken, L., Sloane, D., and Wu, E. (2012). Nurse staffing, burnout, and health care-associated infection. Am. J. Infect. Control 40, 486–490.

Google Scholar

Dawes, J. (2008). Do data characteristics change according to the number of scale points used? An experiment using 5-point, 7-point and 10-point scales. Int. J. Mark Res. 50:613.

Google Scholar

Demerouti, E., Bakker, A., Nachreiner, F., and Schaufeli, W. (2001). The job demands-resources model of burnout. J. Appl. Psychol. 86, 499–512.

Google Scholar

Dewa, C., Loong, D., Bonato, S., Thanh, N., and Jacobs, P. (2014). How does burnout affect physician productivity? A systematic literature review. BMC Health Serv. Res. 14:325. doi: 10.1186/1472-6963-14-325

PubMed Abstract | CrossRef Full Text | Google Scholar

Douglas, D., Choi, D., Marcus, H., Muirhead, W., Reddy, U., Stewart, T., et al. (2021). Wellbeing of Frontline Health Care Workers After the First SARS-CoV-2 Pandemic Surge at a Neuroscience Centre: A Cross-sectional Survey. J. Neurosurg. Anesthesiol. 34, 333–338. doi: 10.1097/ANA.0000000000000767

PubMed Abstract | CrossRef Full Text | Google Scholar

Galesic, M., and Bosnjak, M. (2009). Effects of Questionnaire Length on Participation and Indicators of Response Quality in a Web Survey. Public Opin. Q. 73, 349–360.

Google Scholar

Haidari, E., Main, E., Cui, X., Cape, V., Tawfik, D., Adair, K., et al. (2021). Maternal and neonatal health care worker well-being and patient safety climate amid the COVID-19 pandemic. J. Perinatol. 41, 961–969. doi: 10.1038/s41372-021-01014-9

PubMed Abstract | CrossRef Full Text | Google Scholar

Han, S., Shanafelt, T., Sinsky, C., Awad, K., Dyrbye, L., Fiscus, L., et al. (2019). Estimating the Attributable Cost of Physician Burnout in the United States. Ann. Intern. Med. 170, 784–790.

Google Scholar

Hu, L., and Bentler, P. (1999). Cutoff criteria for fit indexes in covariance structure analysis: Conventional criteria versus new alternatives. Struct. Equ. Model Multidiscip. J. 6, 1–55. doi: 10.1080/00273171.2022.2163476

PubMed Abstract | CrossRef Full Text | Google Scholar

Kang, E., Lihm, H., and Kong, E. (2013). Association of intern and resident burnout with self-reported medical errors. Korean J. Fam. Med. 34, 36–42.

Google Scholar

Kleijweg, J., Verbraak, M., and Van Dijk, M. (2013). The clinical utility of the Maslach Burnout Inventory in a clinical population. Psychol. Assess. 25, 435–441. doi: 10.1037/a0031334

PubMed Abstract | CrossRef Full Text | Google Scholar

Kopalle, P., and Lehmann, D. (1997). Alpha Inflation? The Impact of Eliminating Scale Items on Cronbach’s Alpha. Organ. Behav. Hum. Decis. Process. 70, 189–197.

Google Scholar

Lasalvia, A., Amaddeo, F., Porru, S., Carta, A., Tardivo, S., Bovo, C., et al. (2021). Levels of burn-out among healthcare workers during the COVID-19 pandemic and their associated factors: a cross-sectional study in a tertiary hospital of a highly burdened area of north-east Italy. BMJ Open. 11:e045127. doi: 10.1136/bmjopen-2020-045127

PubMed Abstract | CrossRef Full Text | Google Scholar

Lee, I., and Preacher, K. (2013). Calculation for the test of the difference between two dependent correlations with one variable in common [Computer software]. Geneva: Quantpsy.

Google Scholar

Lee, T., and Mylod, D. (2019). Deconstructing burnout to define a positive path forward. JAMA Intern. Med. 179, 429–430. doi: 10.1001/jamainternmed.2018.8247

PubMed Abstract | CrossRef Full Text | Google Scholar

Lin, J., Lin, P., and Wu, C. (2010). Wellbeing perception of institutional caregivers working for people with disabilities: use of Subjective Happiness Scale and Satisfaction with Life Scale analyses. Res. Dev. Disabil. 31, 1083–1090. doi: 10.1016/j.ridd.2010.03.009

PubMed Abstract | CrossRef Full Text | Google Scholar

Li-Sauerwine, S., Rebillot, K., Melamed, M., Addo, N., and Lin, M. (2020). A 2-Question Summative Score Correlates with the Maslach Burnout Inventory. West J. Emerg. Med. 21, 610–617. doi: 10.5811/westjem.2020.2.45139

PubMed Abstract | CrossRef Full Text | Google Scholar

Liu, M., and Wronski, L. (2018). Examining completion rates in web surveys via over 25,000 real-world surveys. Soc. Sci. Comput. Rev. 36, 116–124.

Google Scholar

Loera, B., Converso, D., and Viotti, S. (2014). Evaluating the psychometric properties of the Maslach Burnout Inventory-Human Services Survey (MBI-HSS) among Italian nurses: How many factors must a researcher consider? PLoS One 9:e114987. doi: 10.1371/journal.pone.0114987

PubMed Abstract | CrossRef Full Text | Google Scholar

Lyubomirsky, S., and Lepper, H. S. A. (1999). Measure of subjective happiness: Preliminary reliability and construct validation. Soc. Indic. Res. 46, 137–155. doi: 10.1017/sjp.2017.8

PubMed Abstract | CrossRef Full Text | Google Scholar

Macía-Rodríguez, C., Alejandre de Oña, Á, Martín-Iglesias, D., Barrera-López, L., Pérez-Sanz, M., and Moreno-Diaz, J. (2021). Burn-out syndrome in Spanish internists during the COVID-19 outbreak and associated factors: a cross-sectional survey. BMJ Open 11:e042966. doi: 10.1136/bmjopen-2020-042966

PubMed Abstract | CrossRef Full Text | Google Scholar

Maslach, C., Jackson, S., and Leiter, M. (1997). “The Maslach Burnout Inventory: Third Edition,” in Evaluating Stress: A Book of Resources, eds C. P. Zalaquett and R. J. Wood (Lanham, MD: Scarecrow Education). doi: 10.3390/brainsci13020251

PubMed Abstract | CrossRef Full Text | Google Scholar

Maslach, C., Schaufeli, W., and Leiter, M. (2001). Job burnout. Annu. Rev. Psychol. 52, 397–422.

Google Scholar

Menon, N., Shanafelt, T., Sinsky, C., Linzer, M., Carlasare, L., Brady, K., et al. (2020). Association of physician burnout with suicidal ideation and medical errors. JAMA Netw. Open 3:e2028780.

Google Scholar

Mukherjee, S., Tennant, A., and Beresford, B. (2020). Measuring burnout in pediatric oncology staff: Should we be using the maslach burnout inventory? J. Pediatr. Oncol. Nurs. 37, 55–64. doi: 10.1177/1043454219873638

PubMed Abstract | CrossRef Full Text | Google Scholar

Muthén, L., and Muthén, B. (2010). Mplus. Los Angeles, CA: Muthén & Muthén.

Google Scholar

Nguyen, J., Liu, A., McKenney, M., Liu, H., Ang, D., and Elkbuli, A. (2021). Impacts and challenges of the COVID-19 pandemic on emergency medicine physicians in the United States. Am. J. Emerg. Med. 48, 38–47.

Google Scholar

Nunnally, J. (1978). Psychometric Theory. New York, NY: McGraw-Hill.

Google Scholar

Nunnally, J., Jum, N., Bernstein, I., and Bernstein, I. (1994). Psychometric Theory. New York, NY: McGraw-Hill Companies.

Google Scholar

Poghosyan, L., Aiken, L., and Sloane, D. (2009). Factor structure of the Maslach burnout inventory: an analysis of data from large scale cross-sectional surveys of nurses from eight countries. Int. J. Nurs. Stud. 46, 894–902. doi: 10.1016/j.ijnurstu.2009.03.004

PubMed Abstract | CrossRef Full Text | Google Scholar

Preston, C., and Colman, A. (2000). Optimal number of response categories in rating scales: reliability, validity, discriminating power, and respondent preferences. Acta Psychol. 104, 1–15. doi: 10.1016/s0001-6918(99)00050-5

PubMed Abstract | CrossRef Full Text | Google Scholar

Profit, J., Adair, K., Cui, X., Mitchell, B., Brandon, D., Tawfik, D., et al. (2021). Randomized controlled trial of the “WISER” intervention to reduce healthcare worker burnout. J. Perinatol. 41, 2225–2234. doi: 10.1038/s41372-021-01100-y

PubMed Abstract | CrossRef Full Text | Google Scholar

Revilla, M., Saris, W., and Krosnick, J. (2014). Choosing the number of categories in agree–disagree scales. Sociol. Methods Res. 43, 73–97.

Google Scholar

Robinson, M. (2018). Using multi-item psychometric scales for research and practice in human resource management. Hum. Resour. Manage. 57, 739–750.

Google Scholar

Rotenstein, L., Torre, M., Ramos, M., Rosales, R., Guille, C., Sen, S., et al. (2018). Prevalence of Burnout Among Physicians: A Systematic Review. JAMA 320, 1131–1150.

Google Scholar

SAS Institute Inc (2019). JMP§Pro, Version 15.1. Cary, NC: SAS Institute Inc.

Google Scholar

Schaufeli, W., Bakker, A., Hoogduin, K., Schaap, C., and Kladler, A. (2001). On the clinical validity of the maslach burnout inventory and the burnout measure. Psychol. Health 16, 565–582.

Google Scholar

Schwartz, S., Adair, K., Bae, J., Rehder, K., Shanafelt, T., Profit, J., et al. (2019). Work-life balance behaviours cluster in work settings and relate to burnout and safety culture: a cross-sectional survey analysis. BMJ Qual. Saf. 28, 142–150. doi: 10.1136/bmjqs-2018-007933

PubMed Abstract | CrossRef Full Text | Google Scholar

Sexton, J., and Adair, K. (2019). Forty-five good things: a prospective pilot study of the Three Good Things well-being intervention in the USA for healthcare worker emotional exhaustion, depression, work-life balance and happiness. BMJ Open 9:e022695. doi: 10.1136/bmjopen-2018-022695

PubMed Abstract | CrossRef Full Text | Google Scholar

Sexton, J., Adair, K., Cui, X., Tawfik, D., and Profit, J. (2022a). Effectiveness of a bite-sized web-based intervention to improve healthcare worker wellbeing: A randomized clinical trial of WISER. Front. Public Health 10:1016407. doi: 10.3389/fpubh.2022.1016407

PubMed Abstract | CrossRef Full Text | Google Scholar

Sexton, J., Adair, K., Proulx, J., Profit, J., Cui, X., Bae, J., et al. (2022b). Emotional Exhaustion Among US Health Care Workers Before and During the COVID-19 Pandemic, 2019-2021. JAMA Netw. Open 5:e2232748. doi: 10.1001/jamanetworkopen.2022.32748

PubMed Abstract | CrossRef Full Text | Google Scholar

Sexton, J., Adair, K., Leonard, M., Frankel, T., Proulx, J., Watson, S., et al. (2018). Providing feedback following Leadership WalkRounds is associated with better patient safety culture, higher employee engagement and lower burnout. BMJ Qual. Saf. 27, 261–270. doi: 10.1136/bmjqs-2016-006399

PubMed Abstract | CrossRef Full Text | Google Scholar

Sexton, J., Adair, K., Profit, J., Bae, J., Rehder, K., Gosselin, T., et al. (2021a). Safety culture and workforce well-being associations with positive leadership walk rounds. J. Comm. J. Qual. Patient Saf. 47, 403–411. doi: 10.1016/j.jcjq.2021.04.001

PubMed Abstract | CrossRef Full Text | Google Scholar

Sexton, J., Adair, K., and Rehder, K. (2021b). The science of healthcare worker burnout: Assessing and improving healthcare worker well-being. Arch. Pathol. Lab. Med. 145, 1095–1109.

Google Scholar

Sexton, J., Frankel, A., Leonard, M., and Adair, K. (2019). SCORE: Assessment of your work setting Safety, Communication, Operational Reliability, and Engagement. Durham: Duke University.

Google Scholar

Sexton, J., Helmreich, R., Neilands, T., Rowan, K., Vella, K., Boyden, J., et al. (2006). The Safety Attitudes Questionnaire: psychometric properties, benchmarking data, and emerging research. BMC Health Serv. Res. 6:44. doi: 10.1186/1472-6963-6-44

PubMed Abstract | CrossRef Full Text | Google Scholar

Sexton, J., Schwartz, S., Chadwick, W., Rehder, K., Bae, J., Bokovoy, J., et al. (2017). The associations between work-life balance behaviours, teamwork climate and safety climate: cross-sectional survey introducing the work-life climate scale, psychometric properties, benchmarking data and future directions. BMJ Qual. Saf. 26, 632–640. doi: 10.1136/bmjqs-2016-006032

PubMed Abstract | CrossRef Full Text | Google Scholar

Shanafelt, T., Balch, C., Bechamps, G., Russell, T., Dyrbye, L., Satele, D., et al. (2010). Burnout and medical errors among American surgeons. Ann. Surg. 251, 995–1000.

Google Scholar

Shanafelt, T., Mungo, M., Schmitgen, J., Storz, K., Reeves, D., Hayes, S., et al. (2016). Longitudinal study evaluating the association between physician burnout and changes in professional work effort. Mayo Clin. Proc. 91, 422–431. doi: 10.1016/j.mayocp.2016.02.001

PubMed Abstract | CrossRef Full Text | Google Scholar

Shanafelt, T., West, C., Sinsky, C., Trockel, M., Tutty, M., Satele, D., et al. (2019). Changes in Burnout and Satisfaction With Work-Life Integration in Physicians and the General US Working Population Between 2011 and 2017. Mayo Clin. Proc. 94, 1681–1694.

Google Scholar

Shanafelt, T., West, C., Sinsky, C., Trockel, M., Tutty, M., Wang, H., et al. (2022). Changes in Burnout and Satisfaction With Work-Life Integration in Physicians and the General US Working Population Between 2011 and 2020. Mayo Clin. Proc. 97, 491–506. doi: 10.1016/j.mayocp.2021.11.021

PubMed Abstract | CrossRef Full Text | Google Scholar

Singer, S., Meterko, M., Baker, L., Gaba, D., Falwell, A., and Rosen, A. (2007). Workforce perceptions of hospital safety culture: Development and validation of the patient safety climate in healthcare organizations survey. Health Serv. Res. 42, 1999–2021. doi: 10.1111/j.1475-6773.2007.00706.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Tawfik, D., Profit, J., Morgenthaler, T., Satele, D., Sinsky, C., Dyrbye, L., et al. (2018). Physician Burnout, Well-being, and Work Unit Safety Grades in Relationship to Reported Medical Errors. Mayo Clin. Proc. 93, 1571–1580. doi: 10.1016/j.mayocp.2018.05.014

PubMed Abstract | CrossRef Full Text | Google Scholar

Tawfik, D., Scheid, A., Profit, J., Shanafelt, T., Trockel, M., Adair, K., et al. (2019). Evidence relating health care provider burnout and quality of care: A systematic review and meta-analysis. Ann. Intern. Med. 171, 555–567.

Google Scholar

Tawfik, D., Sinha, A., Bayati, M., Adair, K., Shanafelt, T., Sexton, J., et al. (2021). Frustration with technology and its relation to emotional exhaustion among health care workers: Cross-sectional observational study. J. Med. Internet Res. 23:e26817. doi: 10.2196/26817

PubMed Abstract | CrossRef Full Text | Google Scholar

Weijters, B., Cabooter, E., and Schillewaert, N. (2010). The effect of rating scale format on response styles: The number of response categories and response category labels. Int. J. Res. Mark. 27, 236–247.

Google Scholar

Welp, A., Meier, L., and Manser, T. (2014). Emotional exhaustion and workload predict clinician-rated and objective patient safety. Front. Psychol. 5:1573. doi: 10.3389/fpsyg.2014.01573

PubMed Abstract | CrossRef Full Text | Google Scholar

West, C., Dyrbye, L., Satele, D., Sloan, J., and Shanafelt, T. (2012). Concurrent validity of single-item measures of emotional exhaustion and depersonalization in burnout assessment. J. Gen. Intern. Med. 27, 1445–1452. doi: 10.1007/s11606-012-2015-7

PubMed Abstract | CrossRef Full Text | Google Scholar

West, C., Dyrbye, L., Sloan, J., and Shanafelt, T. (2009). Single item measures of emotional exhaustion and depersonalization are useful for assessing burnout in medical professionals. J. Gen. Intern. Med. 24, 1318–1321. doi: 10.1007/s11606-009-1129-z

PubMed Abstract | CrossRef Full Text | Google Scholar

Wheeler, D., Vassar, M., Worley, J., and Barnes, L. (2011). A Reliability Generalization Meta-Analysis of Coefficient Alpha for the Maslach Burnout Inventory. Educ. Psychol. Meas. 71, 231–244.

Google Scholar

Worley, J., Vassar, M., Wheeler, D., and Barnes, L. (2008). Factor structure of scores from the Maslach Burnout Inventory: A review and meta-analysis of 45 exploratory and confirmatory factor-analytic studies. Educ. Psychol. Meas. 68, 797–823.

Google Scholar

Wurm, W., Vogel, K., Holl, A., Ebner, C., Bayer, D., Mörkl, S., et al. (2016). Depression-burnout overlap in physicians. PLoS One 11:e0149913. doi: 10.1371/journal.pone.0149913

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: burnout, emotional exhaustion, Maslach burnout inventory, wellbeing, physician wellbeing assessment, burnout metrics, emotional exhaustion metrics

Citation: Penny CL, Adair KC, Frankel AS, Leonard MW, Proulx J, Mosca PJ and Sexton JB (2023) A new look at an old well-being construct: evaluating the psychometric properties of 9, 5, and 1-item versions of emotional exhaustion metrics. Front. Psychol. 14:1267660. doi: 10.3389/fpsyg.2023.1267660

Received: 26 July 2023; Accepted: 18 October 2023;
Published: 23 November 2023.

Edited by:

Fahmida Laghari, Xijing University, China

Reviewed by:

Kotaro Shoji, University of Human Environments, Japan
Chaorong Wu, The University of Iowa, United States

Copyright © 2023 Penny, Adair, Frankel, Leonard, Proulx, Mosca and Sexton. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: J. Bryan Sexton, bryan.sexton@duke.edu

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.