Skip to main content

ORIGINAL RESEARCH article

Front. Artif. Intell., 19 June 2024
Sec. Medicine and Public Health

A machine learning framework to classify musculoskeletal injury risk groups in military service members

\r\nMatthew B. Bird,
Matthew B. Bird1,2*Megan H. Roach,,Megan H. Roach1,2,3Roberts G. NelsonRoberts G. Nelson4Matthew S. HeltonMatthew S. Helton5Timothy C. Mauntel,,Timothy C. Mauntel1,2,3
  • 1Extremity Trauma and Amputation Center of Excellence, Defense Health Agency, Falls Church, VA, United States
  • 2Department of Clinical Investigations, Womack Army Medical Center, Fort Liberty, NC, United States
  • 3Department of Surgery, Uniformed Services University of the Health Sciences, Bethesda, MD, United States
  • 4Artificial Intelligence Integration Center, Army Futures Command, Pittsburgh, PA, United States
  • 5U.S. Army, Tripler Army Medical Center, Honolulu, HI, United States

Background: Musculoskeletal injuries (MSKIs) are endemic in military populations. Thus, it is essential to identify and mitigate MSKI risks. Time-to-event machine learning models utilizing self-reported questionnaires or existing data (e.g., electronic health records) may aid in creating efficient risk screening tools.

Methods: A total of 4,222 U.S. Army Service members completed a self-report MSKI risk screen as part of their unit's standard in-processing. Additionally, participants' MSKI and demographic data were abstracted from electronic health record data. Survival machine learning models (Cox proportional hazard regression (COX), COX with splines, conditional inference trees, and random forest) were deployed to develop a predictive model on the training data (75%; n = 2,963) for MSKI risk over varying time horizons (30, 90, 180, and 365 days) and were evaluated on the testing data (25%; n = 987). Probability of predicted risk (0.00–1.00) from the final model stratified Service members into quartiles based on MSKI risk.

Results: The COX model demonstrated the best model performance over the time horizons. The time-dependent area under the curve ranged from 0.73 to 0.70 at 30 and 180 days. The index prediction accuracy (IPA) was 12% better at 180 days than the IPA of the null model (0 variables). Within the COX model, “other” race, more self-reported pain items during the movement screens, female gender, and prior MSKI demonstrated the largest hazard ratios. When predicted probability was binned into quartiles, at 180 days, the highest risk bin had an MSKI incidence rate of 2,130.82 ± 171.15 per 1,000 person-years and incidence rate ratio of 4.74 (95% confidence interval: 3.44, 6.54) compared to the lowest risk bin.

Conclusion: Self-reported questionnaires and existing data can be used to create a machine learning algorithm to identify Service members' MSKI risk profiles. Further research should develop more granular Service member-specific MSKI screening tools and create MSKI risk mitigation strategies based on these screenings.

Introduction

Musculoskeletal injuries (MSKIs) impose a burden on the United States (U.S.) Military Healthcare System (MHS). MSKIs account for more than 2 million medical encounters annually across the U.S. military (Teyhen et al., 2014) and cost the MHS more than $500 million dollars per year (Teyhen et al., 2018). More importantly, MSKIs limit military Service members' abilities to train and perform required work duties (Ruscio et al., 2010; Teyhen et al., 2018). Despite the significant negative impact MSKIs have on operational readiness and the financial burden on the MHS, there are few actionable and deployable MSKI risk classification models that can help inform clinical practice to reduce MSKI risks among Service members (Rhon et al., 2022b).

Multi-factorial MSKI risk modeling has been the focus of MSKI risk research for the past decade (Bahr, 2016; Bittencourt et al., 2016). Commonly, previous MSKI risk modeling research has discretized variables before modeling [e.g., body mass index (BMI): underweight, healthy, overweight, and obese] to simplify a model's clinical interpretability (Teyhen et al., 2020); however, discretization is not recommended to optimize model performance (Carey et al., 2018). MSKI risk assessment models that are overly simplistic, demonstrate poor model performance, are not yet validated in a new cohort, or are too complex for interpretation, which may misguide clinicians (Bullock et al., 2021, 2022). However, validated machine learning models that provide clinically relevant results are attainable with the correct tools and proper framework.

Machine learning models, built-in frameworks designed for modularity (e.g., flexibility in the prediction model and/or variables included in the model) and scalability (e.g., handling large amounts of data), may have the greatest opportunity for deployment in clinical or operational settings. In addition, machine learning models that are flexible (e.g., random forests) may further allow for modularity and account for non-linearity in the data when compared to more traditional statistical approaches (e.g., regressions). Machine learning models are typically built across a specific time horizon (e.g., within 1 year) and do not account for varying time horizons (i.e., exposure to time), which can limit their utility in clinical settings (Van Eetvelde et al., 2021). Thus, other models must be employed to address the complexity of MSKI risk data.

Time-to-event or survival models estimate the probability of an event occurring before a specified time horizon (Nielsen et al., 2019a). Unlike binary classification models (e.g., logistic regression), which make predictions at a one-time horizon, survival models predict MSKI risk across a continuous range of time horizons accounting for censored individuals. The most common of these survival models is the Cox proportional hazard model. The major limitation of the Cox model is that it makes a strong assumption that the hazard functions between individuals are proportional (Andersen, 2022). Thus, to account for censoring and to utilize flexible survival models (e.g., decision trees), it is necessary to look to other fields deploying survival machine learning techniques for potential applications for MSKI risk stratification (Wang et al., 2019). These existing techniques deployed in MSKI risk stratification may better enable clinicians to estimate survival probability outcomes (i.e., time-to-MSKI) in a methodologically valid and flexible framework.

The purpose of this study was to determine if self-reported questionnaires and existing data sources (e.g., demographics) from military Service members could be used to create a modular machine learning model for assessing MSKI risk. The use of existing self-report data is essential to make the models scalable to a large number of U.S. military Service members [~1.3 million active-duty Service members (U.S. Department of Defense, 2022)]. Second, we will assess varying statistical modeling approaches to provide guidance for appropriate tools to classify MSKI risk in a machine learning framework. We hypothesize that a high-throughput self-report MSKI risk screen, combined with existing data sources, will provide the data necessary to build a machine learning model with acceptable model performance in identifying Service members' MSKI risk levels.

Materials and methods

We conducted a retrospective review of existing self-report MSKI screen data and electronic medical records to develop MSKI risk screening models. A total of 4,222 Service members completed a self-report MSKI risk screen as part of their standard in-processing to a U.S. Army Airborne Division. Complete datasets were accessible for 3,950 Service members (female gender: =12.2%, age: 24.1 ± 5.6 years, height: 175.4 ± 8.8 cm, mass: 78.71 ± 12.0 kg) and were used for analyses. This retrospective cohort study was deemed exempt from the Institutional Review Board review by the Womack Army Medical Center Human Research Protections Program Office.

Data types and reduction

Military Service members in-processing to a U.S. Army Airborne Division completed a self-report MSKI risk screen between December 2020 and March 2022. The self-reported questionnaire consisted of general health, physical fitness, and demographic-type questions (Table 1) (Roach et al., 2023). Additionally, Service members self-reported if they experienced pain during three movement screens: (1) shoulder clearing (bilateral), (2) spinal extension clearing, and (3) squat-jump-land. Pain with either shoulder clearing screen (i.e., left or right) was documented as one, regardless of whether the pain was experienced unilaterally or bilaterally. The three movement screen pain items were summed for a total movement screen pain score (score range = 0–3) for each Service member. The movement screen assessment methods have been previously described (Roach et al., 2023).

Table 1
www.frontiersin.org

Table 1. Model features.

Occupation codes, age, race, gender, and marital status were collected via the Defense Enrollment Eligibility Reporting System (DEERS) within the MHS Management Analysis and Reporting Tool (MHS MART [M2]) closest to the Service member's in-processing date (Table 1). The occupation index codes were categorized into three groups: (1) combat arms (e.g., infantry and field artillery); (2) combat support (e.g., artillery repair and counterintelligence); and (3) combat service support (e.g., medical, instructors, and transportation) (Teyhen et al., 2018).

The MSKI data [International Classification of Diseases - Tenth Revision (ICD-10 Codes)] were collected via the Comprehensive Ambulatory Provider Encounter Record (CAPER) within M2 from the time of the Service member's in-processing screening date up to 1 year post-screening. The CAPER provides direct care outpatient encounter records. Thus, any MSKI treated and diagnosed by an MHS healthcare provider is documented in the CAPER. MSKI encounters were labeled via an MSKI classification matrix (Hando et al., 2023) using the ICD-10 Codes extracted from CAPER. Days until MSKI were calculated as the date of the first MSKI encounter minus the in-processing date. For analysis, subjects were right-censored if they did not sustain an MSKI (noMSKI) within 1 year from their in-processing date, and no duplicate Service members were present in the analysis.

Data analysis: model performance measurements

The area under the curve (AUC) is defined as the area under the receiver operating characteristic (ROC) curve, which quantifies the discriminative power of a variable or multivariable model to correctly classify outcome occurrences (e.g., MSKI) across all probability thresholds. Because MSKI occurrence changes over time, the time-dependent AUC (t-AUC) is used to evaluate model discrimination as a function of time (Heagerty et al., 2000). The t-AUC values range from 0.50 to 1, with 1 being perfect discrimination and 0.5 being a random guess.

Calibration is a measurement of the agreement between the predicted probability of the model and the actual risk. A perfectly calibrated model will have prediction probabilities that exactly match the actual risk outcomes in the data. When plotting actual risk (y-axis) against predicted risk (x-axis), perfect calibration is marked by the line y = x. Any deviation from this line indicates an over- or under-optimistic model. In survival data with censoring, the actual risk is not directly observed, so we estimate it using the Kaplan–Meier estimate (Austin et al., 2020).

The Brier score measures both model discrimination and calibration in one metric. In the absence of censoring, the Brier score is the mean-squared error for binary outcomes. In the presence of right-censoring, the Brier score is weighted with an inverse probability of censoring. A Brier score of 0 indicates a perfectly accurate model, a score of 1 indicates a perfectly inaccurate model, and a score of 0.25 indicates a randomly guessing model if the incidence was 50%. Because the incidence changes over time in survival models, the Brier score for a random guess will also change. To account for this change, the scaled Brier score, or index prediction accuracy (IPA), is defined as the percent improvement over random guess, where 0% is an uninformative model and 100% is a perfectly accurate model (Kattan and Gerds, 2018).

Data analysis: survival models

The missing data were evaluated before analysis and were determined to be low (pay grade: 5.5%, occupation: 1.0%, race: 0.6%, height: 0.1%). Thus, complete cases (n = 3,950) were used for analyses. The data were randomly split into training (75%) and test (25%) sets with an equal proportion of Service members with MSKI in each set. The Cox proportional hazard regression (COX), COX with splines (COX-S), conditional inference trees (CTREE), and random forest (RF) were used to train multiple models on the training data set. All models accounted for right-censored data utilizing a survival function, and the models were trained and tuned individually.

COX generates multivariable survival models that allow multiple variables (continuous or categorical) to be simultaneously assessed via survival probability. Hazard ratios were calculated to provide event rates in one group compared to the other (Deo et al., 2021). Two COX models were developed; COX included all features in their raw format without transformation, while COX-S applied a restricted cubic spline basis expansion to continuous features where the number of knots on each predictor was determined by selecting the model with the lowest Akaike's information criterion (AIC) (Rutherford et al., 2015). The number of knots for each predictor was allowed to take on integer values between 1 and 5.

The CTREE algorithm generates a single binary decision tree where the split at each node is based on the p-value of a statistical test. In the case of survival outcomes, the splits are made using the log-rank test (Hothorn et al., 2015). The significance criterion hyperparameter for the CTREE model was selected using 10-fold cross-validation to maximize the t-AUC at 180 days. The significance criterion hyperparameter was evaluated at the following values: 0.8, 0.85, 0.9, 0.95, and 0.99.

The RF algorithm creates an ensemble of decision trees that fit the bootstraps of the data, so predictions are made based on majority voting across the ensemble (Biau and Scornet, 2016). RF is extended to survival data by approximating the cumulative hazard function in each leaf node with the Nelson–Aalen estimator and splitting nodes based on maximizing the concordance index of the tree (Ishwaran et al., 2008). The hyperparameters for the RF model were tuned by maximizing the out-of-bag concordance index over a grid search for the minimum number of observations per node (3, 5, 7, and 9) and the number of features to consider at each split (2, 4, 6, 8, and 10). The final model was trained with these chosen hyperparameters and an ensemble size of 500 trees.

After tuning model hyperparameters, all models were trained on the full training set and evaluated on the testing data. Model performances on the test set were evaluated using the t-AUC and scaled Brier score for time horizons between 30 and 365 days. The final model selection was performed by minimizing the Brier score at 180 days. Based on this criterion, the COX model was selected for risk stratification.

Data analysis: musculoskeletal injury risk stratification

The COX model's calibration curve was binned by quartiles to create four equal “risk bins” to compare predicted risk to estimated risk in each bin across different time horizons (30, 90, 180, and 365 days). Calibration was assessed qualitatively by comparing the mean predicted risk in each group with their Kaplan–Meier estimate. Incidence rates (IRs) and incidence rate ratios (IRRs) were used to compare MSKI incidences across risk bins and time horizons (risk bin 1 served as the reference). Comparisons were made across the risk bins using Fisher's exact test for categorical features and the one-way ANOVA for continuous features. Additional post-hoc comparisons were performed by comparing risk bin 1 to risk bin 2, risk bin 3, and risk bin 4 with a Holm–Bonferroni adjustment on the p-values. See Figure 1 (data analysis: machine learning pipeline) for our data analysis framework. R version 4.2 was used for all statistical analysis, namely packages caret, riskRegression, tidymodels, survival, party, ranger, dplyr, readr, ggfortify, and gridExtra (R Core Team, 2019). R scripts are provided in Supplementary material.

Figure 1
www.frontiersin.org

Figure 1. MSKI risk assessment framework in the military; (A) application of MSKI risk assessment (green): collect relevant existing data on service members; (B) data analysis: machine learning pipeline (blue): data analysis pipeline utilized for the creation of machine learning models in the manuscript. In application, the Service member's data would be inputted into the MSKI risk machine learning model for categorization of the risk bin; (C) future model: in-depth screening and evaluation in the high-performance model (orange): In application, if the service member is flagged in risk bin four, then further screening assessments are performed to determine if any modifiable risk factors can be addressed by the human performance staff for further intervention.

Results

The MSKI incidence rates were 780.9 per 1,000 person-years, 785.1 per 1,000 person-years, and 768.3 per 1,000 person-years for the entire data set (100%), the training data set (75%), and the test data set (25%), respectively, during the 1 year surveillance period. The t-AUC performance for COX, COX-S, and RF were similar at each time horizon, while the CTREE t-AUC was ~0.05 less at each time horizon. There was a 0.06 t-AUC decline from 30 to 365 days averaged across all models (Figure 2A). The absolute Brier score demonstrated similar performance across each model at each time horizon, with no models greater than the null model (0 variables; Figure 2B). The scaled Brier score (i.e., IPA, Figure 2C) was similar across all models up to 90 days (90 days = 10–12%), while the CTREE declined over time, and the COX IPA performance was greater at all time horizons past 90 days compared to the other models. Overall, the COX model had the best performance compared to the COX-S, RF, and CTREE when evaluating the t-AUC and IPA. Based on the criterion at 180 days, the COX model was chosen as the final model for evaluation (t-AUC = 0.70 and IPA = 12.4%) and further analysis. The COX model demonstrated the top five MSKI risk factors were {reported as hazard ratio and 95% confidence interval [HR (95% CI)]}: “other” race [3.3 (1.4–8.2)], three movement pain items [2.2 (1.7–2.9)], two movement pain items [1.9 (1.6–2.4)], female gender [1.6 (1.3–1.9)], and prior MSKI [1.5 (1.3–1.8)] (Figure 3).

Figure 2
www.frontiersin.org

Figure 2. (A) Time-dependent area under the curve receiver operator characteristic curve (t-AUC); (B) absolute Brier score, dotted black line denotes null model; (C) Brier score scaled to the null model (0 variables) or index prediction accuracy (IPA).

Figure 3
www.frontiersin.org

Figure 3. Cox proportional hazards regression (COX) model (final model) coefficients. Red = significant (p < 0.05), Black = non-significant; dots are hazard ratios with bars signifying a 95% confidence interval.

The COX model calibration curve binned into quartiles demonstrated four groups of predicted and estimated risk across 30, 90, 180, and 365 days (Figure 4). The estimated risk fell within the mean predicted risk point estimate and standard error, indicating good calibration of the COX final model (Figure 4). The median survival time (95% confidence interval) in risk bin 4 was 119 (85–142) days, in risk bin 3 was 331 (268-NA) days, and for risk bins 1 and 2 could not be calculated within 1 year (Figure 5).

Figure 4
www.frontiersin.org

Figure 4. Cox proportional hazards regression (COX) model (final model) calibration curves for each time horizon. Dots represent mean calibration within each risk group bin, with standard error of the counts of subjects across the bin. Histograms represent the distribution of Service member counts in each risk bin across the predicted risk. Histograms shading represent risk bins: light gray = risk bin 1 … black = risk bin 4.

Figure 5
www.frontiersin.org

Figure 5. Kaplan–Meier curve, stratified by four risk bins and number of subjects at risk for each time horizon.

Risk bin 4 had the highest IR and IRR at each time horizon. The overall IR was lowest at the time horizon of 365 days when compared to 30, 90, and 180 days, respective to each risk bin (Table 2). In risk bin 4, compared to risk bin 1, there were a greater proportion of female members, enlisted Service members, Service members with prior MSKIs, Service members with pain during the Army Combat Fitness Test (ACFT), those in combat service support occupations, married Service members, non-white Service members, Service members with an MSKI-related limited duty profile, those with prior stress fracture, those with prior surgery, and those with >1 pain across movement screens (Table 3). In addition, risk bin 4 Service members were shorter in height (bin 4: 173.2 ± 9.4 cm, bin 1: 176.8 ± 7.1 cm; p = < 0.001), reported less sleep (bin 4: 6.0 ± 1.1 h, bin 1: 6.6 ± 0.9 h; p < 0.001), and were older (bin 4: 25.6 ± 6.4 y, bin 1: 22.9 ± 4.5 y; p < 0.001) than risk bin 1 (Table 3). Risk bin 3 had similar findings as risk bin 4 when compared to risk bin 1, while risk bin 2 was more comparable to risk bin 1.

Table 2
www.frontiersin.org

Table 2. Risk bins across time horizons incidence rates.

Table 3
www.frontiersin.org

Table 3. Risk bin descriptive statistics.

Discussion

Our overall objective was to evaluate if data elements extracted from self-reported questionnaires and existing data sources in military Service members could be used to create a predictive model for assessing MSKI risk. We (1) created a valid and modular MSKI risk predictive model, (2) determined what key features are important for MSKI risk modeling, and (3) provided a framework for MSKI risk stratification through self-reported and existing data sources.

Overall, the models we evaluated were discriminative (t-AUC COX and RF, 30- to 180-day time horizons = 0.73–0.70), and all models performed better than the null model (0 variables) across all time horizons (Figure 2). The COX model demonstrated the best overall performance across all time horizons compared to the other models. However, our models performed worse than previously described military MSKI risk models (Rhon et al., 2022b; Shaw et al., 2023). In an Army Ranger cohort, the logistic regression and cross-validation techniques produced a discriminative (cross-validation AUC = 0.90) and accurate (Brier score = 0.12) MSKI risk model with demographic, biomechanical, and Army physical fitness scores (Rhon et al., 2022b). Similarly, an RF model for medial tibial stress syndrome in military cadets, with an externally validated dataset of Australian Navy recruits, demonstrated discriminative (test set AUC = 0.92) capabilities (Shaw et al., 2023). In general, the worse performance of our models may relate to the granularity of variables being assessed. In our study, dichotomized variables (e.g., yes/no response) were used and may not have provided enough information for the algorithms to “learn” or the variables were not as closely related to MSKI when compared to more granular data (e.g., biomechanical data). Furthermore, previous research did not apply survival-type models; thus, it is difficult to make model performance comparisons (Rhon et al., 2022b; Shaw et al., 2023).

Survival analyses are rarely utilized in MSKI risk research unless for inference and association rather than prediction (e.g., epidemiologic research) (Sharma et al., 2017; Hando et al., 2023). For example (Nielsen et al., 2019a), out of 31 original research studies, only two used “time-to-event” or survival analyses (Drew and Finch, 2016) when associating training load to MSKI outcomes. In MSKI risk research, it may be more advantageous to analyze a rate (e.g., survival–hazard ratios) compared to a static risk (e.g., odds ratios), especially when analyzed over an extended exposure time (e.g., 1 year). By calculating a rate, exposure time is included in the modeling, which could be helpful in determining how useful a screening tool/measurement is up to a specified time horizon, as determined by t-AUC or expected survival time among particular groups (Nielsen et al., 2019a). To a clinician, time-to-event may assist in allocating resources, such as staff or additional screening, to mitigate MSKI risk among a group of Service members before the MSKI occurrence at a specified time horizon (Nielsen et al., 2019b).

Our baseline COX model outperformed all models when evaluating the varying performance metrics. The RF model performed similarly to COX, but the COX model was chosen as the final model to aid clinical interpretation due to the slightly better performance at 180 days (t-AUC 180 days: COX = 0.70, RF = 0.69; IPA 180 days: COX = 12.4%, RF = 10.2%). The COX model may have outperformed the RF model due to the large number of dichotomized variables (RF is robust to continuous type variables but may lead to overfitting) or a few variables explaining the majority of model performance. While the variables most important for modeling (i.e., prior MSKI, female gender, and movement pain; Figure 3) are in agreement with decades of previous research (Rhon et al., 2022a; Hando et al., 2023), many of these are non-modifiable and subject to minimal change states over an extended exposure time (e.g., occupation changes infrequently; Table 1).

Models such as RF are typically deemed “black box” algorithms as there are ensembles of decision trees that are utilized for model building, and extrapolating a single decision tree does not reflect the overall ensemble of trees (Ishwaran et al., 2008; Price, 2018). On the other hand, the CTREE is a single decision tree that is considered a “white box” algorithm and may be easier for a clinician to interpret (Loyola-Gonzalez, 2019). However, the CTREE model had the worst overall model performance (t-AUC 180 days = 0.65). We included the CTREE decision tree (Supplementary Figure 1) and the COX model hazard ratios (Figure 3) to demonstrate the interoperability of these “white box” algorithms for immediate MSKI risk decision support tools.

To empirically evaluate the COX model, Service members in the test data were stratified into quartiles based on predicted risk at four time horizons. In the COX model, when compared to risk bin 1, risk bin 4 had a 9.41 × (30 days), 5.31 × (90 days), 4.74 × (180 days), and 3.58 × (365 days) significantly greater MSKI risk (Table 2). Thus, we successfully stratified groups into low- to high-MSKI risk bins. Our approach may be utilized as a clinical decision support tool to identify Service members who require more in-depth screenings to assess their MSKI risk more thoroughly. Our recommended utilization for the MSKI risk prediction model is to stratify Service members into quartiles or use the top percentage of participants (e.g., top 10%) predicted probability risk to create an “at risk” bin for further screening and potential intervention (Figure 1B) (Roach et al., 2023). An example may be to follow up on the “at-risk” participants with biomechanical assessments that provide feedback to mitigate potential dysfunctional movement patterns (Bird et al., 2022) or in-depth clinical assessments to identify (i.e., Figure 1C: additional screening), properly diagnose (i.e., Figure 1C: Human Performance Staff), and treat present MSKIs (i.e., Figure 1C: Intervention). We stratified the Service members into quartiles, but, dependent on the staffing and time allotted for additional screening and/or interventions, other stratification methods may be used (e.g., quintiles) to increase/decrease the number of Service members who screen into each quantile. Regardless of the number of MSKI risk categories, identifying a “high-risk group” with routinely collected variables (i.e., self-reported MSKI risk screen data = ~30min time to assess 150 Service members) or already collected data (i.e., health records), this may be a method to drastically reduce the number of Service members requiring more in-depth MSKI risk screening (Roach et al., 2023).

Congruent with previous research, we found that female gender, prior MSKI, and pain with movement assessments were MSKI risk factors (Rhon et al., 2022a; Roach et al., 2023). Interestingly, the selection of “other race” had the highest hazard ratio [3.3 (1.4–8.2)] (Figure 3), and Service members who identified as “other” were all distributed to the high-risk bin after stratification (Table 3). All race data were extracted from DEERS and included the following categories: White, Asian or Pacific Islander, Black, American Indian or Alaskan, Other, and Unknown. While several studies have identified an association between race and MSKI risk, the findings are contradictory, and there is no clear association as to which race is at the highest MSKI risk (Sammito et al., 2021). Furthermore, MHS data systems do not fully capture race data, as Service members are unable to view or edit their DEERS data (Michael-Anne Browne, 2023). Thus, we are unable to determine whether disparities exist, and further investigation into race and MSKI risk is needed. On the other hand, modifiable risk factors (e.g., pain with three and two movement assessments) were highly associated with MSKI (Figure 3). Thus, pain with movement assessments may provide a guided in-depth screening to understand the underlying mechanism of pain, allowing a human performance staff team member to potentially intervene (Figure 1).

The primary strength of our study was developing a training model and testing the model on hold-out test data, as many original research articles conclude their findings with the training model (Van Eetvelde et al., 2021). Additionally, we provided potential cut-points to assess MSKI risk for clinicians (Table 3) and a framework for MSKI risk modeling (Figure 1) using existing data sources. However, our study is not without limitations. Our primary limitation is that our analyses include MSKI diagnostic codes, as MSKIs are only identified if the Service member sought care for the MSKI. Service members often do not seek medical care, and MSKIs may go unreported (Sauers et al., 2016). Additionally, ~300 observations were removed due to data being incomplete, which is common when utilizing secondary data sources. Furthermore, our framework only designates high-risk Service members (n = 247) for in-depth screening, which inherently leaves 740 Service members without more in-depth MSKI risk screening data. Since our model is based on MSKI risk probability, we decided to structure our framework to further screen Service members at the greatest estimated MSKI risk probability to reduce the time burden on clinicians due to the large number of U.S. military Service members.

Conclusion

Service members' MSKI risk levels can be determined from self-reported MSKI risk screens and existing military data. This high-throughput approach will improve the ability to complete initial MSKI risk assessment en masse and reduce the number of Service members requiring “in-depth” screenings, as healthcare providers and strength and conditioning professionals can focus their efforts on the highest risk individuals. Decreasing additional screenings across all Service members reduces the time burden on the Service member and clinical staff and increases resource allocation to other necessities across the MHS. Thus, the next step is to deploy this type of algorithm directly into an electronic health record or within an operational military unit. This original research is a continuation of the modernization of the MHS with 21st-century capabilities by delivering a framework that may be deployable in the MHS infrastructure.

Data availability statement

The datasets presented in this article are not readily available because the datasets generated and analyzed during the current study are not publicly available because of data sharing restrictions on data generated within the United States Department of Defense; however, data may be available from the corresponding author on reasonable request, following approval from all required regulatory bodies. Requests to access the datasets should be directed to MB, matthew.b.bird.civ@health.mil.

Ethics statement

This study received an exempt determination as non-human subject research by the Womack Army Medical Center Human Research Protections Office (Protocol #22-14804; Approved 14 June 2022).

Author contributions

MB: Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Validation, Visualization, Writing – original draft, Writing – review & editing. MR: Conceptualization, Data curation, Investigation, Methodology, Writing – original draft, Writing – review & editing. RN: Formal analysis, Methodology, Validation, Visualization, Writing – original draft, Writing – review & editing. MH: Conceptualization, Investigation, Writing – original draft, Writing – review & editing. TM: Conceptualization, Data curation, Investigation, Methodology, Project administration, Supervision, Writing – original draft, Writing – review & editing.

Funding

The author(s) declare that no financial support was received for the research, authorship, and/or publication of this article.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Author disclaimer

The views expressed in this manuscript are of those of the authors, and do not necessarily reflect the views, opinions, or policies of the Uniformed Services University of the Health Sciences, the US Departments of Army/Navy/Air Force, Department of Defense, nor the U.S. Government.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/frai.2024.1420210/full#supplementary-material

Supplementary Figure 1. Random forest (RF) variable importance.

Supplementary Figure 2. Conditional inference trees (CTREE).

References

Andersen, P. K. (2022). Fifty years with the Cox proportional hazards regression model. J. Ind. Inst. Sci. 102, 1135–1144. doi: 10.1007/s41745-021-00283-9

Crossref Full Text | Google Scholar

Austin, P. C., Harrell, F. E. Jr, and Van Klaveren, D. (2020). Graphical calibration curves and the integrated calibration index (ICI) for survival models. Stat. Med. 39, 2714–2742. doi: 10.1002/sim.8570

PubMed Abstract | Crossref Full Text | Google Scholar

Bahr, R. (2016). Why screening tests to predict injury do not work—and probably never will…: a critical review. Br. J. Sports Med. 50, 776–780. doi: 10.1136/bjsports-2016-096256

PubMed Abstract | Crossref Full Text | Google Scholar

Biau, G., and Scornet, E. (2016). A random forest guided tour. Test 25, 197–227. doi: 10.1007/s11749-016-0481-7

Crossref Full Text | Google Scholar

Bird, M. B., Mi, Q., Koltun, K. J., Lovalekar, M., Martin, B. J., Fain, A., et al. (2022). Unsupervised clustering techniques identify movement strategies in the countermovement jump associated with musculoskeletal injury risk during US Marine Corps Officer Candidates School. Front. Physiol. 787:868002. doi: 10.3389/fphys.2022.868002

PubMed Abstract | Crossref Full Text | Google Scholar

Bittencourt, N. F., Meeuwisse, W., Mendonça, L., Nettel-Aguirre, A., Ocarino, J., and Fonseca, S. (2016). Complex systems approach for sports injuries: moving from risk factor identification to injury pattern recognition—narrative review and new concept. Br. J. Sports Med. 50, 1309–1314. doi: 10.1136/bjsports-2015-095850

PubMed Abstract | Crossref Full Text | Google Scholar

Bullock, G. S., Hughes, T., Arundale, A. H., Ward, P., Collins, G. S., and Kluzek, S. (2022). Black box prediction methods in sports medicine deserve a red card for reckless practice: a change of tactics is needed to advance athlete care. Sports Med. 1–7. doi: 10.1007/s40279-022-01771-3

Crossref Full Text | Google Scholar

Bullock, G. S., Hughes, T., Sergeant, J. C., Callaghan, M. J., Riley, R. D., and Collins, G. S. (2021). Clinical prediction models in sports medicine: a guide for clinicians and researchers. J. Orthopaed. Sports Phys. Therapy 51, 517–525. doi: 10.2519/jospt.2021.10697

PubMed Abstract | Crossref Full Text | Google Scholar

Carey, D. L., Crossley, K. M., Whiteley, R., Mosler, A., Ong, K.-L., Crow, J., et al. (2018). Modeling training loads and injuries: the dangers of discretization. Med. Sci. Sports Exerc. 50, 2267–2276. doi: 10.1249/MSS.0000000000001685

PubMed Abstract | Crossref Full Text | Google Scholar

Deo, S. V., Deo, V., and Sundaram, V. (2021). Survival analysis—part 2: Cox proportional hazards model. Ind. J. Thorac. Cardiovasc. Surg. 37, 229–233. doi: 10.1007/s12055-020-01108-7

PubMed Abstract | Crossref Full Text | Google Scholar

Drew, M. K., and Finch, C. F. (2016). The relationship between training load and injury, illness and soreness: a systematic and literature review. Sports Med. 46, 861–883. doi: 10.1007/s40279-015-0459-8

PubMed Abstract | Crossref Full Text | Google Scholar

Hando, B. R., Bryant, J., Pav, V., Haydu, L., Hogan, K., Mata, J., et al. (2023). Musculoskeletal injuries in US Air Force Tactical Air Control Party trainees: an 11-year longitudinal retrospective cohort study and presentation of a musculoskeletal injury classification matrix. BMJ Mil Health e002417. doi: 10.1136/military-2023-002417

PubMed Abstract | Crossref Full Text | Google Scholar

Heagerty, P. J., Lumley, T., and Pepe, M. S. (2000). Time-dependent ROC curves for censored survival data and a diagnostic marker. Biometrics 56, 337–344. doi: 10.1111/j.0006-341X.2000.00337.x

PubMed Abstract | Crossref Full Text | Google Scholar

Hothorn, T., Hornik, K., and Zeileis, A. (2015). ctree: Conditional Inference Trees. The Comprehensive R Archive Network, 8. Available online at: https://cran.r-project.org/web/packages/partykit/vignettes/ctree.pdf

Google Scholar

Ishwaran, H., Kogalur, U. B., Blackstone, E. H., and Lauer, M. S. (2008). Random survival forests. Ann. Appl. Stat. 2, 841–860. doi: 10.1214/08-AOAS169

Crossref Full Text | Google Scholar

Kattan, M. W., and Gerds, T. A. (2018). The index of prediction accuracy: an intuitive measure useful for evaluating risk prediction models. Diagn. Progn. Res. 2, 1–7. doi: 10.1186/s41512-018-0029-2

PubMed Abstract | Crossref Full Text | Google Scholar

Loyola-Gonzalez, O. (2019). Black-box vs. white-box: understanding their advantages and weaknesses from a practical point of view. IEEE Access 7, 154096–154113. doi: 10.1109/ACCESS.2019.2949286

Crossref Full Text | Google Scholar

Michael-Anne Browne, M. (2023). Decision Brief: Eliminating Racial and Ethnic Health Disparities in the Military Health System. Available online at: https://www.health.mil/Reference-Center/Presentations/2023/06/28/Eliminating-Racial-and-Ethnic-Health-Disparities-in-the-MHS

Google Scholar

Nielsen, R. O., Bertelsen, M. L., Ramskov, D., Møller, M., Hulme, A., Theisen, D., et al. (2019a). Time-to-event analysis for sports injury research part 1: time-varying exposures. Br. J. Sports Med. 53, 61–68. doi: 10.1136/bjsports-2018-099408

PubMed Abstract | Crossref Full Text | Google Scholar

Nielsen, R. O., Bertelsen, M. L., Ramskov, D., Møller, M., Hulme, A., Theisen, D., et al. (2019b). Time-to-event analysis for sports injury research part 2: time-varying outcomes. Br. J. Sports Med. 53, 70–78. doi: 10.1136/bjsports-2018-100000

PubMed Abstract | Crossref Full Text | Google Scholar

Price, W. N. (2018). Big data and black-box medical algorithms. Sci. Transl. Med. 10:eaao5333. doi: 10.1126/scitranslmed.aao5333

PubMed Abstract | Crossref Full Text | Google Scholar

R Core Team (2019). R: A Language and Environment for Statistical Computing, 3.6.1 Edn. R Foundation for Statistical Computing. Available online at: https://www.R-project.org/ (accessed January 2, 2024).

Google Scholar

Rhon, D. I., Molloy, J. M., Monnier, A., Hando, B. R., and Newman, P. M. (2022a). Much work remains to reach consensus on musculoskeletal injury risk in military service members: a systematic review with meta-analysis. Eur. J. Sport Sci. 22, 16–34. doi: 10.1080/17461391.2021.1931464

PubMed Abstract | Crossref Full Text | Google Scholar

Rhon, D. I., Teyhen, D. S., Collins, G. S., and Bullock, G. S. (2022b). Predictive models for musculoskeletal injury risk: why statistical approach makes all the difference. BMJ Open Sport Exerc. Med. 8:e001388. doi: 10.1136/bmjsem-2022-001388

PubMed Abstract | Crossref Full Text | Google Scholar

Roach, M. H., Bird, M. B., Helton, M. S., and Mauntel, T. C. (2023). “Musculoskeletal injury risk stratification: a traffic light system for military service members”, in Healthcare (MDPI), 1675.

PubMed Abstract | Google Scholar

Ruscio, B. A., Jones, B. H., Bullock, S. H., Burnham, B. R., Canham-Chervak, M., Rennix, C. P., et al. (2010). A process to identify military injury prevention priorities based on injury type and limited duty days. Am. J. Prev. Med. 38, S19–S33. doi: 10.1016/j.amepre.2009.10.004

PubMed Abstract | Crossref Full Text | Google Scholar

Rutherford, M. J., Crowther, M. J., and Lambert, P. C. (2015). The use of restricted cubic splines to approximate complex hazard functions in the analysis of time-to-event data: a simulation study. J. Stat. Comput. Simul. 85, 777–793. doi: 10.1080/00949655.2013.845890

Crossref Full Text | Google Scholar

Sammito, S., Hadzic, V., Karakolis, T., Kelly, K. R., Proctor, S. P., Stepens, A., et al. (2021). Risk factors for musculoskeletal injuries in the military: a qualitative systematic review of the literature from the past two decades and a new prioritizing injury model. Milit. Med. Res. 8, 1–40. doi: 10.1186/s40779-021-00357-w

PubMed Abstract | Crossref Full Text | Google Scholar

Sauers, S. E., Smith, L. B., Scofield, D. E., Cooper, A., and Warr, B. J. (2016). Self-management of unreported musculoskeletal injuries in a US Army brigade. Milit. Med. 181, 1075–1080. doi: 10.7205/MILMED-D-15-00233

PubMed Abstract | Crossref Full Text | Google Scholar

Sharma, J., Dixon, J., Dalal, S., Heagerty, R., and Spears, I. (2017). Musculoskeletal injuries in British Army recruits: a prospective study of incidence in different Infantry Regiments. BMJ Milit. Health 163, 406–411. doi: 10.1136/jramc-2016-000657

PubMed Abstract | Crossref Full Text | Google Scholar

Shaw, A., Newman, P., Witchalls, J., and Hedger, T. (2023). Externally validated machine learning algorithm accurately predicts medial tibial stress syndrome in military trainees: a multicohort study. BMJ Open Sport Exerc. Med. 9:e001566. doi: 10.1136/bmjsem-2023-001566

PubMed Abstract | Crossref Full Text | Google Scholar

Teyhen, D., Bergeron, M. F., Deuster, P., Baumgartner, N., Beutler, A. I., De La Motte, S. J., et al. (2014). Consortium for health and military performance and American College of Sports Medicine Summit: utility of functional movement assessment in identifying musculoskeletal injury risk. Curr. Sports Med. Rep. 13, 52–63. doi: 10.1249/JSR.0000000000000023

PubMed Abstract | Crossref Full Text | Google Scholar

Teyhen, D. S., Goffar, S. L., Shaffer, S. W., Kiesel, K., Butler, R. J., Tedaldi, A.-M., et al. (2018). Incidence of musculoskeletal injury in US Army unit types: a prospective cohort study. J. Orthopaed. Sports Phys. Therapy 48, 749–757. doi: 10.2519/jospt.2018.7979

PubMed Abstract | Crossref Full Text | Google Scholar

Teyhen, D. S., Shaffer, S. W., Goffar, S. L., Kiesel, K., Butler, R. J., Rhon, D. I., et al. (2020). Identification of risk factors prospectively associated with musculoskeletal injury in a warrior athlete population. Sports Health 12, 564–572. doi: 10.1177/1941738120902991

PubMed Abstract | Crossref Full Text | Google Scholar

U.S. Department of Defense (2022). 2022 Demographics Profile of the Military Community. Department of Defense, Under Contract with Icf. Available online at: https://demographics.militaryonesource.mil/ (accessed January 2, 2024).

Google Scholar

Van Eetvelde, H., Mendonça, L. D., Ley, C., Seil, R., and Tischer, T. (2021). Machine learning methods in sport injury prediction and prevention: a systematic review. J. Exp. Orthopaed. 8, 1–15. doi: 10.1186/s40634-021-00346-x

PubMed Abstract | Crossref Full Text | Google Scholar

Wang, P., Li, Y., and Reddy, C. K. (2019). Machine learning for survival analysis: a survey. ACM Comp. Surv. 51, 1–36. doi: 10.1145/3214306

Crossref Full Text | Google Scholar

Keywords: survival, Cox proportional hazard regression, decision trees, screening, secondary data

Citation: Bird MB, Roach MH, Nelson RG, Helton MS and Mauntel TC (2024) A machine learning framework to classify musculoskeletal injury risk groups in military service members. Front. Artif. Intell. 7:1420210. doi: 10.3389/frai.2024.1420210

Received: 19 April 2024; Accepted: 27 May 2024;
Published: 19 June 2024.

Edited by:

Tse-Yen Yang, China Medical University Hospital, Taiwan

Reviewed by:

TaChen Chen, Chia Nan University of Pharmacy and Science, Taiwan
Antonio Sarasa-Cabezuelo, Complutense University of Madrid, Spain

Copyright © 2024 Bird, Roach, Nelson, Helton and Mauntel. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Matthew B. Bird, matthew.b.bird.civ@health.mil

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.