Skip to main content

ORIGINAL RESEARCH article

Front. Digit. Health , 14 February 2025

Sec. Health Informatics

Volume 7 - 2025 | https://doi.org/10.3389/fdgth.2025.1538793

Equitable hospital length of stay prediction for patients with learning disabilities and multiple long-term conditions using machine learning

\r\nEmeka AbakasangaEmeka Abakasanga1Rania KousovistaRania Kousovista1Georgina Cosma
Georgina Cosma1*Ashley AkbariAshley Akbari2Francesco ZaccardiFrancesco Zaccardi3Navjot Kaur,Navjot Kaur3,4Danielle FittDanielle Fitt2Gyuchan Thomas JunGyuchan Thomas Jun4Reza KianiReza Kiani5Satheesh Gangadharan\r\nSatheesh Gangadharan5
  • 1Computer Science Department, School of Science, Loughborough University, Loughborough, United Kingdom
  • 2Faculty of Medicine, Health and Life Science, Swansea University, Swansea, United Kingdom
  • 3Leicester Real World Evidence Unit, Diabetes Research Centre, University of Leicester, Leicester, United Kingdom
  • 4School of Design and Creative Arts, Loughborough University, Loughborough, United Kingdom
  • 5Learning Disability Service (Agnes Unit), Leicestershire Partnership NHS Trust, Leicester, United Kingdom

Purpose: Individuals with learning disabilities (LD) often face higher rates of premature mortality and prolonged hospital stays compared to the general population. Predicting the length of stay (LOS) for patients with LD and multiple long-term conditions (MLTCs) is critical for improving patient care and optimising medical resource allocation. However, there is limited research on the application of machine learning (ML) models to this population. Furthermore, approaches designed for the general population often lack generalisability and fairness, particularly when applied across sensitive groups within their cohort.

Method: This study analyses hospitalisations of 9,618 patients with LD in Wales using electronic health records (EHR) from the SAIL Databank. A Random Forest (RF) ML model was developed to predict hospital LOS, incorporating demographics, medication history, lifestyle factors, and 39 long-term conditions. To address fairness concerns, two bias mitigation techniques were applied: a post-processing threshold optimiser and an in-processing reductions method using an exponentiated gradient. These methods aimed to minimise performance discrepancies across ethnic groups while ensuring robust model performance.

Results: The RF model outperformed other state-of-the-art models, achieving an area under the curve of 0.759 for males and 0.756 for females, a false negative rate of 0.224 for males and 0.229 for females, and a balanced accuracy of 0.690 for males and 0.689 for females. Bias mitigation algorithms reduced disparities in prediction performance across ethnic groups, with the threshold optimiser yielding the most notable improvements. Performance metrics, including false positive rate and balanced accuracy, showed significant enhancements in fairness for the male cohort.

Conclusion: This study demonstrates the feasibility of applying ML models to predict LOS for patients with LD and MLTCs, while addressing fairness through bias mitigation techniques. The findings highlight the potential for equitable healthcare predictions using EHR data, paving the way for improved clinical decision-making and resource management.

Introduction

Learning disability (LD), also referred to as intellectual disabilities in some contexts, have been defined by the Learning Disabilities Observatory (1) as the presence of: “a significantly reduced ability to understand new or complex information, to learn new skills (impaired intelligence), with a reduced ability to cope independently (impaired social functioning); which started before adulthood, with a lasting effect on development.” There are approximately 1.1 million adults aged 18 years and older living with an LD in the UK, including over 54,000 individuals from Wales (2). Existing sources show that individuals with LD often experience poorer physical and mental health, as well as higher rates of multiple long-term conditions (MLTCs) and avoidable mortality compared to those without LD (310). This demographic presents unique needs and challenges that impact their hospitalisations (11).

Effectively managing healthcare resources while ensuring optimal patient outcomes poses particular challenges for patients with LD. An important outcome of interest for patients with LD is a reliable prediction of the length of hospital stay (LOS) of their admission and the underlying factors that could influence their LOS (12). Predicting the LOS can lead to enhanced healthcare services and further initiate proactive measures to prevent prolonged LOS. A recent study conducted in the UK discovered that at any given time, approximately 2,000 patients with LD and/or autism in long-stay hospitals have been hospitalised, with over half having spent over 2 years in hospital care. This includes 350 LD patients who have been in long-stay hospitals for more than a decade (13). The extended LOS in their study was attributed to either the patient’s personal characteristics or limitations of the system supporting them. Another study found that other general factors contributing to prolonged hospital stay for patients who are medically fit for discharge included hospital-acquired infections, falls, and other medical errors (14). Conversely, there is also a downside to patients being discharged prematurely, as it may result in readmissions or, in severe cases, preventable deaths (15, 16). Although these studies were carried out on the general population, the conclusion and outcome may still apply to people with LD with MLTCs.

The above-mentioned studies highlight the importance of proactively managing patient discharges as early as possible during their hospitalisation to optimise the LOS.

Several studies have utilised machine learning (ML) models to predict the LOS of individual patients in the general population by analysing large datasets of hospital admission records (1719). These models typically assess the influence of various factors—such as sex, age, diagnosis, admission method, and illness severity—on the likelihood of hospital stays exceeding a predefined threshold. However, there is limited research focusing on the prediction of hospitalisations for adults with learning disabilities using ML. This cohort faces unique challenges, underscoring the importance of tracking their clinical pathways to better inform medical interventions and reduce average hospital durations.

Most existing LOS prediction studies using ML focus on specific long-term conditions, medical units, or single hospitals or geographic regions. As a result, many of these models may not be applicable to patients with varied MLTCs [e.g., those with multiple co-occurring long-term conditions (LTCs)] or to individuals from diverse socio-geographic backgrounds. Furthermore, a vast majority of these studies fail to provide adequate explanations regarding the accuracy of ML models across distinct patient subgroups, such as ethnicity, age, and sex (2022). This lack of transparency raises concerns about the applicability of these models to sensitive groups. Such concerns are particularly important from an ethical standpoint, as the models may be biased towards underrepresented communities. These limitations underscore the urgent need for fairness in healthcare ML models, ensuring that predictions are explainable, and that model performance is thoroughly evaluated and optimised across diverse patient groups (23).

This study aims to predict the LOS of patients with LD and MLTCs in Wales using ML models. The model is built on a dataset of 62,243 hospital admission records from 9,618 patients with LD. It incorporates 39 long-term conditions (S1), identified through a literature review and expert consensus (24, 25), as well as demographics, medication history, prior hospitalisations, and lifestyle factors to account for potential confounding variables.

Bias in ML classification models in healthcare often stems from imbalanced datasets, systemic inequities, or suboptimal feature selection, which can result in unfair outcomes for sensitive patient groups (e.g., ethnic minorities, sex, socioeconomic status, or specific medical conditions). To address these issues, bias mitigation techniques are typically applied at three stages of the ML pipeline: preprocessing (before model training), in-processing (during training), and post-processing (after training) (26, 27). This study focuses on fairness in predicting outcomes across sensitive ethnic groups by employing two bias mitigation techniques. The first is the threshold optimiser, a post-processing method applied after model training. It adjusts the classifier’s output scores based on a fairness constraint to reduce bias (27, 28). The second is the reductions method with exponentiated gradient (EG), an in-processing approach that modifies the classifier’s weights during training to ensure fairness in real time (22, 27). Both techniques were rigorously evaluated to assess their impact on model performance across different ethnic groups.

The structured electronic healthcare datasets used in this study presented inherent complexities due to various inconsistencies in patient records over time. These inconsistencies can stem from factors such as data entry errors, biased data collection, missing documentation, issues with patient compliance, and changes in patient status that are not captured in the electronic health record (EHR) (29). To address these complexities, we employed statistical preprocessing techniques, systematically handled missing data, and applied algorithms to mitigate biases. The resulting ML model demonstrated strong performance, validating our approach to managing complex healthcare data. The contributions of this paper are as follows:

• Analysis of hospitalised Welsh LD patient population—provided statistics on demographics, prevalence of 39 LTCs, previous hospitalisation data including prior admissions, episodes, days, and condition prevalence.

• Identification of primary conditions treated during hospitalisation and the prevalent LTCs for hospitalised patients with LD, along with admission rates per patient.

• Identification of prevalent LTCs linked to prolonged hospital stays (129 days).

• Statistical analysis to identify factors associated with hospital stays 4 days using non-parametric tests.

• Development and evaluation of machine learning models to predict whether a patient’s LOS would be <4 days or 4 days, using patient data available up to the first 24 h of admission.

• Assessment of model performance differences across ethnic groups. Application and comparison of two bias mitigation algorithms: threshold optimisation and reductions algorithm using an exponentiated gradient.

• Demonstrated potential of applying ML with effective bias mitigation on electronic health records data to promote equitable prediction across groups when predicting LOS.

Methods

The LOS is defined as the number of days an inpatient is hospitalised during a single admission (12). For each admission, the LOS value was obtained by calculating the difference between the admission and discharge dates. An LOS threshold ψ was obtained by taking the ceiling value of the mean LOS across all extracted admission records, excluding the “outliers.” With the mean LOS of 3.015 days (standard deviation of 4.064 days), ψ was set to ψ=3.015=4 days. Therefore, the LOS threshold (ψ) was set to 4 days. Throughout this study, a “long stay” refers to hospitalisation with LOS 4 days, while a “short stay” denotes LOS <4 days. The “long-stay rate,” is defined as the percentage of admissions lasting at least 4 days, calculated using

Long-stay rate=Number of admissions with LOS4Total number of admissions(1)

This study addresses three key research questions: (1) How can ML models be used to accurately predict the LOS for patients with LD and MLTCs using EHR data? (2) What are the key sources of bias in ML models for LOS prediction, and how do they impact accuracy and fairness across sensitive groups, such as ethnic minorities? (3) Can bias mitigation techniques effectively improve fairness across ethnic groups while maintaining or enhancing predictive performance? To tackle these challenges, a structured methodology was developed as illustrated in Figure 1, integrating data extraction, preprocessing, ML model development, and applying bias mitigation techniques to ensure robust and equitable predictions for this population.

Figure 1
www.frontiersin.org

Figure 1. Machine learning framework diagram.

Stage I: data extraction

SAIL database

This study utilised EHR data sources of hospital admissions for patients with LD, contained in the Secure Anonymised Information Linkage (SAIL) Databank, the national Trusted Research Environment (TRE) for Wales, enabling the use of anonymised individual-level, population-scale, and linked data sources (3032). SAIL partners with the National Health Service (NHS) and Welsh government to organise routinely collected longitudinal health and administrative data for approximately 5 million Welsh residents, accessed securely under strict conditions compliant with the General Data Protection Regulation (GDPR). Specifically for this study the Welsh Longitudinal General Practice (WLGP) data containing information on primary care General Practice (GP) records, the Patient Episode Database for Wales (PEDW) containing information on secondary care inpatient hospitalisation admissions, the Welsh Demographic Service Dataset (WDSD) containing patient demographic, residency, and registration history, and the Annual District Death Extract (ADDE) containing mortality records from the Office for National Statistics (ONS) were accessed. An Anonymised Linking Field (ALF) facilitates longitudinal anonymised linkage through all data sources in SAIL (33). Data captured within primary care is via read codes. The International Classification of Diseases version 10 (ICD-10) codes capture diagnosis, and Office of Population Censuses and Surveys (OPCS) Classification of Interventions and Procedures version 4 (OPCS-4) captures operations in hospital admissions.

Inclusion and exclusion criteria

The study focused on Welsh residents aged 18 years or older identified with LD during the period from 1 January 2000 to 31 December 2021, as depicted in Figure 2: Exclusions were implemented for individuals younger than 18 years, those not residing in Wales, individuals not registered with a SAIL GP at the study start date, and those without LD. This process led to a total of 14,323 unique patients during the data extraction phase. Subsequently, patients without any hospital admission records between the study start and end dates were omitted. The resulting admissions data were then associated with 39 LTCs (Supplementary Table S1), documented primarily through Read codes in primary care and ICD-10 codes in secondary care (34, 35). The dataset comprised 62,523 unique admissions from 9,630 LD patients, spanning the period from January 2000 to December 2021. Exclusions were applied to records with missing discharge dates and negative LOS, resulting in a cohort of 62,521 unique admissions from 9,630 LD patients. Finally, admissions with no data within the first 24 h were excluded, yielding the final cohort of 9,618 unique patients (4,929 males, 4,689 females) with 62,243 unique admissions (32,275 males, 29,968 females). Refer to Supplementary Table S2 for further details on demographic distribution. Additionally, Supplementary Table S3 provides a comprehensive list of all variables extracted for each patient across their admissions.

Figure 2
www.frontiersin.org

Figure 2. Consort flow diagram.

Dataset demographic description

Age group. Age was categorised into seven groups, as shown in Supplementary Table S2. Patient and admission counts generally followed a normal distribution across age groups for both sexes, with fewer admissions in ages above 60–69 and below 30. Possible factors for lower admission rates in ages above 60–69 include home care reducing hospital need and mortality (see Supplementary Table S4 for mortality statistics on this study cohort). The statistics on mortality obtained in this study are consistent with the life expectancy statistics (66 and 67 years for LD males and females, respectively) (36). Notably, ages above 60–69 exhibited higher rates of long stays, similar for both sexes.

Welsh Index of Multiple Deprivation. Patients’ socioeconomic status was described using the Welsh Index of Multiple Deprivation (WIMD) version 2019 for Wales (37), an area-level weighted index across seven deprivation domains assigned based on the individual’s residence using their Lower-layer Super Output Areas (LSOA) version 2011, with each LSOA representing an area of around 1,500 people. The seven deprivation domains include income, employment, health and disability, education skills and training, barriers to housing and services, living environment, and crime. WIMD quintiles were used, categorising the area of residence from 1 (most deprived) to 5 (least deprived) as shown in Supplementary Table S2. Patients without an LSOA or associated WIMD quintile were grouped as “Unknown.” The highest representation of hospital admissions was from the most deprived quintile (quintile 1), comprising 28% of admissions for males and 27% of admissions for females, indicating heightened hospital demand with increasing deprivation. This aligns with research done on the general population, attributing higher admission rates in impoverished areas to factors such as inadequate allied healthcare and local resources and potential underuse of community medical resources (38, 39). Supplementary Table S4 also shows higher mortality rates with increasing deprivation for the study cohort. This finding is consistent with a recently published study on the impact of deprivation on mortality (7).

Ethnic group. Ethnic groups were classified using the ONS, UK categorisation. The methodology by Akbari et al. (40) was used to extract and harmonise the ethnic group details from the various data sources available to the project. The cohort was not uniformly represented (i.e., unbalanced) in terms of ethnic groups: 73% of patients (79% of admissions) and 74% of patients (80% of admissions) were from the “White” group for the male and the female sex, respectively (see Supplementary Table S2). Within the cohort, 25% of males and 24% of females had no ethnic group records, classified as “Unknown” (19.24% and 17.53% of male and female admissions, respectively). The remaining ethnic groups (Black, Asian, Mixed, Other) each represented <3% of patients and admissions for both sexes. Long-stay rates varied widely across the ethnic groups, with the “Black” group having the highest rate (M: 52% and F: 65.9%; M denotes male sex and F denotes female sex) and the “Other” group the lowest (M: 34.2% and F: 30.4%). A similar finding was observed in a study on inpatient discharges for the general patient population in the United States, which revealed that Black patients had significantly longer LOSs compared to other groups (41).

Feature extraction

Extracting the inputs. The ML models in this study utilise a dataset of 9,618 patients with LD (see Supplementary Table S2) and a total of 62,243 hospital admission records to predict each patient’s LOS (i.e., target variable). For each admission, data relating to the patient’s health up until the first 24 h of admission were extracted to be applied as inputs to the ML models. As detailed in Supplementary Table S3, this data includes variables describing: the patient’s lifestyle and history [body mass index (BMI), smoking, alcohol consumption, physical exercise, autism]; prior 1-year and 3-year hospitalisation data (previous admissions and hospital episodes, cumulative hospital days from past admissions, and the number of MLTCs from previous admissions); prescribed antipsychotic, antidepressant, and anti-manic/anti-epileptic medications (Supplementary Table S5 provides the medications list); and other variables from first 24 h of current admission, indicating the prevalence of the 39 LTCs (see Supplementary Table S1 for the full list of LTCs). Age group was also included as an input into the model (i.e., as a predictor) because MLTC counts increase with age (5, 42) and consequently have an impact on the length of stay. There is a deficit of studies that apply ML to predict the LOS for adults with LD. Therefore, the variables used as inputs in this study were selected through a review by a professional advisory board, supplemented by insights from existing studies utilising ML for hospital predictions in the general population (18, 19, 43, 44).

Extracting the target variable. The LOS variable indicating the number of hospital days was replaced with a binary target variable (LOSClass) for machine learning classification purposes, as follows: values with an LOS 4 days where replaced with 1 indicating long stay, and values with LOS < 4 days where replaced with 0 indicating short stay, as shown in Supplementary Table S3.

Stage II: preprocessing the inputs for ML

Step 1: data cleaning

Supplementary Table S3 describes the ML model input and target variables, and how they were preprocessed for ML training and analysis. Supplementary Table S3 also shows the demographic variables and these were not applied as inputs to the ML model (except age group) but were utilised for describing the cohort and for bias analysis of ML model performance. To clean the data two steps were carried out: coding the longitudinal variables and handling missing data

Preprocessing longitudinal risk factors. BMI, alcohol consumption, smoking, physical activity, and medications are longitudinal variables, gathered and coded by GPs or practice nurses. These risk factors change over time across patients. Hence, the variable “BMI” was coded categorically using the patient’s BMI value documented closest to the admission date. This enables the model to use the most recent patient BMI description for each unique admission. Alcohol and smoking records, which are self-reported risk factors, were found to be noisy and inconsistent across patients. For example, some chronological records indicated declarations such as “ex-smoker,” followed by “smoker,” and then “never smoked,” suggesting potential dishonesty or imputation errors. To address this, these risk factors were recoded based on historical data rather than current status, assuming that past behaviours could have a lasting impact on the patient’s physiology (45, 46). The variables “ALCOHOL_HISTORY” and “SMOKING_HISTORY” were categorised based on intake history up to the admission date (see Supplementary Figures S1, S2). “PHYSICAL” was coded categorically using the patient’s physical activity status (i.e., if the patient engages in light or regular exercise) documented nearest admission. The “MEDICATIONS” variable was coded as binary for intake history at each admission and assumed lifetime use from the first prescription.

Missing values. Missing values for all categorical variables were classified into “unknown” categories for their respective variables This allows models to utilise the partial information from observations with missing data rather than discarding or imputing them during preprocessing, consequently introducing more bias into the model. There were no missing data for the numerical variables.

Step 2: feature selection

A correlation test explored relationships between the numerical features (columns) in the dataset. Correlation analysis aids ML model building by detecting redundant inputs, simplifying interpretations, and improving target prediction performance. Supplementary Table S3 describes all features for predicting LOS. The Kolmogorov–Smirnov test assessed feature normality before selecting an appropriate correlation analysis. As shown in Supplementary Table S6, all features had non-normal distributions. Consequently, the non-parametric Spearman rank correlation coefficient was evaluated. Supplementary Figures S3, S4 illustrate no statistically significant associations between any input and the outcome (LOSClass) for male and female cohorts. However, several input variable pairs exhibited collinearity, with correlation coefficients exceeding ±0.5 for both groups. This study found that the interaction of the highly correlated variable pairs provided useful information to the model for the cohort examined. “While their correlation did not imply causation, this observation, combined with feedback from clinicians on the professional advisory panel, supported their inclusion in the analysis.”

Step 3: convert categorical variables to one-hot encoded variables

ML models require numerical inputs/target variables. To enable this, the categorical input variables were converted into numeric representations via one-hot encoding. The steps for one-hot encoding a categorical variable using the variable “PHYSICAL” as an example, are as follows:

1. Identify unique categories and count per variable, e.g., “PHYSICAL” has three categories: Yes, No, Unknown (Supplementary Table S3).

2. Create a new binary variable for each category. For each admission, only one of the new binary variables is “hot” (1), indicating its category. This is illustrated by example in Supplementary Tables S7, S8 for the variable “PHYSICAL” recorded over three admissions.

3. Concatenate these new binary columns to the dataset. Hence, from the example, instead of one three-value variable, “PHYSICAL” is now encoded into three separate binaries that ML models can easily use.

Step 4: data normalisation

Z-score normalisation was applied to all numerical input variables in Supplementary Table S3. Through this, each numerical variable was centred to have zero mean and unit variance. The standardised data retains the skewness and kurtosis shape properties of the original dataset. Data normalisation ensures that variables with different scales are rescaled to a common range, preventing larger features from dominating ML models. This improves model performance, training stability, and convergence while reducing the impact of outliers. In this study, normalisation ensured equal contributions from all features, enhancing accuracy, fairness, and generalisability in predicting the LOS.

Stage III: model training and evaluation methodology

Step 1: initial train and test split

The preprocessed dataset comprising the model inputs was split into two separate sets: a training set and a test set, utilising a stratified 50–50 partition. Hence, the training and test set each had 16,121 samples for the male dataset and 15,094 samples for the female group.

Step 2: downsampling of the training set

To ensure a balanced representation across the two classes of the target variable (long stay: LOSClass = 1 and short stay: LOSClass = 0), the majority class (short stay) in the training set was downsampled. Specifically, the number of training samples belonging to the short-stay class was reduced to match the number of examples from the smaller, long-stay class. Applying this balanced downsampling serves to mitigate potential modelling inefficiencies caused by class imbalance, where models may ignore or not properly learn underrepresented classes. Also, by harmonising class distributions, downsampling can improve model evaluation metrics related to average performance across classes, such as balanced accuracy, as both classes are weighted and assessed equally. Supplementary Tables S9, S10 show the demographic distribution of the training and test data applied in classifying the LOS for males and females, respectively.

Step 3: model training and testing

To identify an optimal ML model for predicting hospitalisation duration (LOSClass), eight classifiers were evaluated including logistic regression (LR) (47, 48), support vector machine (SVM) (47, 49), Random Forest (RF) (47, 49), eXtreme Gradient Boosting (XGBoost) (47), histogram-based gradient boosting (HISTGBoost) (48), XGBoost (48), k-nearest neighbor (KNN) (47), and a sequential neural network (NN) model (49). Supplementary Table S11 indicates the parameter configurations set for each classifier. Supplementary Tables S12, S13 detail the parameter configurations for the NN.

Step 4: performance evaluation

Several evaluation metrics were employed to evaluate each classifier’s performance on the test set. Let |TP| denote the number of unique admissions for which the long stay (i.e., LOSClass = 1) was correctly classified; |TN| be the number of short stays (LOSClass = 0) correctly classified; |FP| be the number of short stays incorrectly classified as long stays; |FN| be the number of long stays incorrectly classified as short stays; |P| be the total number of long-stay admissions, where |P|=|TP|+|FN|; and |N| represents the total number of short-stay admissions, where |N|=|TN|+|FP|. The following metrics defined in Equations 26 were utilised to evaluate the performance of the ML models.

True positive rate ( TPR) =|TP||TP|+|FN|(0,1)(2)
True negative rate ( TNR) =|TN||TN|+|FP|(0,1)(3)
Balanced accuracy=TPR + TNR2(0,1).(4)

The closer the values of the abovementioned metrics are to 1, the better the performance of the model. The FNR and FPR are given by the expression

FNR=1TPR=|FN||TP|+|FN|,(0,1),(5)
FPR=1TNR=|FP||TN|+|FP|,(0,1).(6)

The closer the values of the FNR and the FPR are to 0, the better the performance of the model. Another important evaluative measure is the receiver operating characteristic (ROC) curve, which plots the true positive rate (TPR) against the FPR at different threshold values. This creates a curve from (0,0) to (1,1). The area under the ROC curve (AUC) measures the two-dimensional area beneath this curve. A greater AUC shows better performance in predicting the long (LOS 4 days) and short (LOS < 4 days) hospital stays.

Assessing model generisability. After the optimal model (RF) was selected by comparing the metrics, it was further evaluated using repeated random train/test splits to scrutinise the model’s generalisability. Specifically, the dataset was randomly split into dedicated 50% sized training sets and 50% sized testing sets a total of 10 times. This generated 10 distinct train/test set combinations, allowing for evaluation across different data partitions. For each of the 10 train/test splits, the corresponding training set was downsampled as described in step 2 of Stage III and used to train the model. The model performance was evaluated at each iteration using its corresponding test set. Finally, the evaluation performance across the 10 iterations with different dataset splits was averaged to assess generalisability.

Stage IV: bias analysis and mitigation

To check for fairness across ethnic groups, the optimal random forest model was analysed by ethnic groups. Particularly, this analysis refers to the performance range that is obtained by taking the difference between the maximum and minimum values for each metric across ethnic groups. For instance, if the FNR for an ML model has a maximum of 0.17 for the “White” ethnic group and a minimum of 0.51 for the “Asian” ethnic group, the performance range is calculated as follows:

PerformancerangeFNR=max{FNR}min{FNR},

which equals 0.34 or 34%. The closer the performance range is to 0 the better the performance of the model. In this study, two bias mitigation techniques (reductions with EG and threshold optimiser) were evaluated to balance performance metrics across ethnic groups.

Reductions with exponentiated gradient

The reductions algorithm (22) was applied during training to limit performance ranges across ethnic groups. A fairness constraint was defined on FNR parity, requiring the FNR range between ethnic groups to be at most 0.2. The model was trained and evaluated to measure unfairness based on this constraint. The EG algorithm assigned weights to training instances that reduce the overall violation of the fairness constraint, with higher weights to instances contributing more to unfairness. The weighted training data was used to retrain the model to focus more on instances contributing to unfairness. These steps were repeated 10 times, updating instance weights iteratively to reduce bias while maintaining accuracy across ethnic groups. At each iteration, the algorithm updates the model parameters by considering the gradient of an objective function incorporating both predictive balanced accuracy and fairness constraints. The model parameters are then updated to effectively adjust the model to reduce bias while maintaining predictive balanced accuracy.

Threshold optimiser

To check for performance discrepancies across demographic groups, the post-processing threshold-optimiser approach (28) was utilised. Specifically, the threshold optimiser tuned the decision boundary of the random forest classifier to achieve parity in the balanced accuracy metric between ethnic groups subject to enforcing constraints on the FNRs per group. This optimised the fairness-balanced accuracy trade-off solution without needing to modify the underlying ML model or training procedure. The threshold optimiser takes an already trained model and fits a transformation function to the model’s outputs to satisfy certain fairness constraints. This approach allows for mitigating unfairness when developers have no control over the training process of the model, which may occur due to practical limitations or considerations around security or privacy.

Results

This section presents the findings from the analysis of hospital admission records for patients with LD and MLTCs. The results are structured to address key research questions, including the primary and prevalent conditions for hospital admissions, the distribution and factors influencing LOS, and the application of ML models to predict LOS while addressing bias across ethnic groups. Insights from the study are further supported by performance evaluations of ML models and the application of bias mitigation techniques to improve fairness and generalisability.

What were the primary conditions for hospital admissions of patients with LD?

The PEDW data includes a variable “diag_num”—a number used to identify the position of diagnosis assigned to a patient during a unique admission. Value “1” relates to the primary ICD-10 Diagnostic Code, which is the main condition treated or investigated during the relevant episode of healthcare. Values > 1 relate to secondary ICD-10 diagnostic codes. Analysis of 18,541 admissions of men and 17,587 admissions of women with LD from the last 10 years of study duration (January 2011–December 2021) showed cancer as the primary condition for admission, with 1,703 (9.2%) male and 2,149 (12.2%) female admissions. The subsequent top primary conditions differed by sex, respectively, as shown in Figure 3: epilepsy, chronic pneumonia, chronic airway diseases, and mental illness for males; vs. chronic kidney disease, epilepsy, chronic pneumonia, and chronic airway diseases for females. Supplementary Tables S14, S15 detail the top 10 primary conditions for both sexes including admission counts, patient numbers, and admission rates per patient. The high standard deviation (SD) values in the admission rate for some conditions indicate the variation in admission rates per individual, as some patients may be admitted more times than others. The admission rates are strongly influenced by the number of MLTCs across individuals.

Figure 3
www.frontiersin.org

Figure 3. The top five primary conditions and top five prevalent (common) conditions treated or investigated during hospitalisations for the (a) males and (b) females with LD and MLTCs. The primary conditions indicate the main condition treated or investigated during the admission, and the prevalent conditions indicate the most frequently treated or investigated conditions (includes both primary and secondary diagnostic codes) across all unique hospitalisations of patients with LD and MLTCs.

What were the prevalent conditions in hospital admissions for patients with LD?

Further 2011–2021 admission analysis was made on the PEDW data for the prevalent conditions treated or investigated during unique hospital admissions. This includes both primary and secondary diagnostic codes. The analysis revealed epilepsy as the most commonly treated LTCs during admissions, present in 29.4% of male and 24.1% of female admissions (Figure 3). The next most prevalent conditions in the male group were diabetes, chronic airway diseases, mental illness, and cancer, while the female group had chronic airway diseases, diabetes, thyroid disorders, and mental illness as the next most prevalent conditions after epilepsy. Figure 3 also indicates slightly higher rates of epilepsy, diabetes, and mental illness admissions in males compared to females, and higher rates of chronic airway diseases in females compared to males. Supplementary Table S16 provides the ranking of common LTCs treated during hospital admissions for the stated period.

What is the distribution of the LOS across patients?

The LOS for all extracted patients’ hospitalisation records from birth ranged from a minimum of 0 day (indicating no overnight stay during admission) to over 5,000 days. The combined male and female group comprised 67,377 admissions. The median hospitalisation was Q2=2 days with first and third quartiles of 0 day (Q1) and 7 days (Q3), respectively, giving an interquartile range (IQR) of Q3Q1. This is illustrated using the box plot in Supplementary Figure S5, which includes lower and upper whiskers. All admissions with LOS days below the lower whiskers or above the upper whiskers are described as outliers. The lower whisker was the smallest LOS value in days greater than Q11.5IQR, equal to Q1 (0 day). The upper whisker was obtained as the largest LOS value in days less than Q3+1.5IQR, obtained at 17 days.

Outliers

For the entire hospitalisation records of patients from birth, outliers (admissions with LOS >17 days) were further analysed to understand patterns among admissions with the most extended stays. Hence, quartile values from boxplots of outlier records were obtained: Q1=24 days, Q2=36 days, Q3=66 days, and an IQR of 42 days. The upper whisker was the largest hospital stay under Q3+1.5IQR, obtained at 129 days. Consequently, admissions with LOS 129 days were numerically evaluated. This amounted to 934 unique admissions. The majority of these admissions are related to mental illness. Generally, hospitalisations with very long LOS are common for mental health admissions, especially for those with challenging behaviours and autism/personality disorders posing safety risks (50, 51). Figure 4a further illustrates these findings, depicting the common conditions for stays 129 days. For these admissions with LOS 129, the most common condition was mental illness and epilepsy, followed by diabetes, dementia, and cerebral palsy. Supplementary Table S17 provides a full breakdown of related LTCs for admissions with LOS 129 days. Figure 4b further depicts, for all admissions with LOS 129 days, the age distribution for admissions involving mental illness and without mental illness. For the latter, age was normally distributed and skewed towards the older age groups. Given the study cohort includes patients with MLTCs, most patients admitted for a primary condition also experienced hospital episodes associated with other secondary conditions.

Figure 4
www.frontiersin.org

Figure 4. Distribution of patients with LOS 129 days. (a) The 4 most common conditions and (b) age distribution of patients with and without mental illness.

What factors were prevalent in hospital stays 4 days?

Analysis was conducted prior to the ML predictions to identify common trends among admissions with LOS 4 days within the study duration. The extracted data was analysed by sex (male and female) across all admissions in the study duration. The one-sample chi-square and binomial tests were applied to all categorical features to test the null hypothesis that the variable categories occur with equal probability. Results stated in Supplementary Table S18 rejected this hypothesis for all categorical variables.

Excluding “unknown” groups, Supplementary Table S19 revealed 4 days stays were predominantly in patients aged 50 years, from more deprived socioeconomic quintiles, obese, and less physically active, compared to patients with short stays (LOS < 4 days). This observation was similar for both sexes. In addition, females with prescribed antipsychotic, antidepressant, or anti-manic/anti-epileptic medications were also seen to have more long stay admissions with LOS 4 days, compared to female admissions with LOS <4 days. Other factors influencing 4 days stays include the primary long-term condition for admission and MLTCs counts, which can increase hospital episodes in a single admission. Patients in this study cohort had between 1 and 21 comorbid conditions per person. Figures 5a,b illustrate the average MLTCs counts by age group for combined and separate sexes, showing a linear rise with age. Further analysis was also conducted on previous hospitalisation data that included the number of previous admissions and hospital episodes, cumulative hospital days from past admissions, and the number of MLTCs from previous admissions (see Supplementary Table S3). Patients with a higher number of MLTCs, cumulative hospital days from past admissions, and long-term conditions treated in previous admissions were more likely to be hospitalised for 4 days compared to patients with stays less than 4 days. Patients with more frequent prior admissions and higher counts of previous hospital episodes tended to have short LOS in their current admission. This is illustrated in Supplementary Figures S6–S9. Those patients with LOS <1 day were most likely to be attending routine appointments rather than emergency visits.

Figure 5
www.frontiersin.org

Figure 5. Distribution of mean MLTC count across age groups for (a) combined sexes and (b) individual sexes.

Can machine learning predict the LOS across hospital admissions?

Several classification models were developed for the prediction of the LOS using the selected features described in Supplementary Table S3. The models were designed to predict whether a patient’s admission would have LOS <4 days or 4 days. Predictions were made using available patient data up until the first 24 h of admission. The classification performance of each model is presented in Supplementary Table S20. Additionally, Figures 6a,b depict the ROC curve for each classifier, along with the AUC and optimal ROC point values. An optimal ROC point refers to the point on an ROC curve that provides the best balance between the TPR and the FPR for a given classification model.

Figure 6
www.frontiersin.org

Figure 6. ROC curves across models indicating their optimal points (FPR and TPR) for (a) male and (b) female cohorts.

Classification performance

In this study, the HISTGBoost and RF classifiers showed the best performance compared to the other models as shown in Supplementary Table S20. The HISTGBoost classifier achieved higher AUC (M: 0.771, F: 0.773) and balanced accuracy (M: 0.701, F: 0.705). This was followed by the XGBoost classifier [AUC = 0.763 (M), 0.761 (F); balanced accuracy = 0.695 (M), 0.692 (F)] and the RF classifier [AUC = 0.759 (M), 0.756 (F); balanced accuracy = 0.690 (M), 0.689 (F)]. Regarding the FNR (indicating patients predicted to be discharged early when a longer (i.e., 4 days) hospital stay is required), the RF classifier returned lower values compared to the other models [FNR = 0.224 (M), 0.229 (F)]. The FNR value was approximately 7% lower for the RF classifier than the HISTGBoost model for the male group and 6% lower than the XGBoost model for the female group. The RF demonstrated optimal performance for both the male and female groups as shown by its low FNRs and high balanced accuracy. The low FNR indicates fewer missed cases where patients require longer hospitalisation 4 days, while the balanced accuracy shows high predictive performance.

Is the performance of the best model consistent across different train/test combinations?

The selected RF model was further evaluated using 10 randomly selected distinct train/test combinations to assess its performance across different samples. The mean performance (with SD) across all 10 iterations is provided in Supplementary Table S21 for the male and female cohorts. The mean value obtained for each parameter across sexes is similar to the performance detailed in Supplementary Table S20 for the male and female cohorts, and the SD is less than 0.007 for all parameters as shown in Supplementary Table S21. This suggests the RF model achieves consistent performance across different data samples, indicating it has good generalisability and is not overfitting on the given training set. The low SD also shows the model yields reliable and stable predictions.

How did the best model perform across ethnic groups?

The RF model that showed optimal performance was further evaluated to address potential performance bias associated with the ethnic groups. Supplementary Tables S22, S23 provide an overview of the RF model’s performance segmented by ethnic groups for the male and female groups, respectively.

Males

For males, the “Other” ethnic group exhibited the highest FNR at 33.3% followed by the “Black” group with FNR of 27.3%. The “Black” group had the lowest balanced accuracy at 59.7%. In contrast, the “Asian” group demonstrated the lowest FNR of 19.5%, and the “White” group had the highest balanced accuracy of 69.2%. Overall, there was an approximate 13.8% range in the FNR and a 9.5% range in balanced accuracy across the considered ethnic groups. The model’s performance for the “White” ethnic group closely resembled the overall model performance (detailed in Supplementary Table S20 for the RF model), most likely because the majority of the training data (79%) originated from the “White” group. The “White” group constituted 79% of the entire male cohort.

Females

For females, the model’s performance for the “White” group closely matched the overall classifier performance, attributable to the high representation of the “White” group (80.6% of training, 80% of total data). The model underperformed in predicting LOS for the “Black” group with the lowest balanced accuracy (66.7%), and highest FPR (50%). Both the “Black” and “Other” groups had the lowest representation in the extracted data (0.15% and 0.19%, respectively), suggesting insufficient data for optimal modelling. The “Other” group, despite having a low representation of admission records, demonstrated improved performance with the lowest FNR (11.1%) and the highest balanced accuracy (77.8%). This suggests the classifier was able to effectively model outcomes for the “Other” ethnic group given the available training samples. Overall, for females, there was a performance range in the FNR, and balanced accuracy of approximately 12%, and 11.1%, respectively across ethnic groups.

Can consistency in ML performance be achieved across ethnic groups?

In the context of this study, an ideally fair model would exhibit consistent performance in predicting the LOS across ethnic groups. To improve the fairness of the LOS prediction models, two bias mitigation algorithms: one post-processing (threshold optimiser) and one in-processing (reductions approach with exponentiated gradient) were empirically investigated. Both approaches aimed to minimise the range of each performance metric across the ethnic groups. Each bias mitigation algorithm was assessed and compared to the unmitigated model to determine its effectiveness.

In-processing (reductions with EG)

Supplementary Tables S24, S25 overview the EG reductions performance. This approach worked best in reducing the performance range of the FNR for the males by 9% compared to the unmitigated model. However, the performance range was not optimised for the FPRs and balanced accuracies for both sexes (see Supplementary Tables S26, S27).

Post-processing (threshold optimiser)

Supplementary Tables S24, S25 depict the performance of the threshold-optimised model. For females, compared to the unmitigated model, the threshold optimiser reduced the FNR range across ethnic groups by 3.6% (see Supplementary Table S27). However, the range for the FPR and balanced accuracy increased by 9.8% and 5.6%, respectively across ethnic groups compared to the unmitigated model. Specifically, the optimiser did not substantially improve the model fairness across the female group. The threshold optimiser yielded better performance in males, reducing the range for the FPR and the balanced accuracy values across ethnic groups by 7.4% and 1.8%, respectively. However, the range for FNR increased slightly by 1.7% for the threshold optimiser compared to the unmitigated model. Particularly, the unmitigated classifier had a lower FNR range across the ethnic groups than the optimiser for males (see Supplementary Table S26).

In summary, although the fairness goal of equal performance across ethnic groups was not fully met, the post-processing threshold optimiser approach was more effective at improving performance uniformity across ethnic groups compared to the reduction with exponentiated gradient.

Discussion

Prolonged hospital stays pose significant patient risks, including increased susceptibility to infections, falls, sleep deprivation, and physical and mental decline. To address these issues, NHS England’s Reducing Length of Stay (RLoS) program (52) aims to improve patient care and optimise resource use by minimising unnecessary delays in hospital discharges. The program’s national goal to reduce hospital stays of 21+ days by 25%–40% highlights the importance of timely discharge planning in improving patient outcomes, increasing capacity in urgent and emergency care, and freeing up hospital beds. However, individuals with LD experience poorer health outcomes compared to the general population, which often leads to prolonged hospital stays (10). Accurately predicting the LOS for this population is therefore crucial for optimising resource allocation and preventing unnecessarily prolonged or premature discharges (12).

This study contributes to these goals by developing ML models to predict hospital LOS for patients with LD and MLTCs while addressing fairness concerns across ethnic groups. An overview of findings from this study is detailed in Box 1. As ML increasingly guides healthcare decision-making, ensuring algorithmic fairness is critical to avoid exacerbating existing health disparities. However, the dataset used in this study exhibited imbalanced representation across ethnic groups, with underrepresentation of minority patients and missing ethnic information for a significant proportion of records. These imbalances resulted in performance disparities in the Random Forest (RF) model, particularly in FNRs for minority females, indicating poorer predictions of long stays for this subgroup.

www.frontiersin.org

Box 1. Overview of findings

To address these challenges, two bias mitigation techniques were applied: the threshold optimizer (post-processing) and reductions method with exponentiated gradient (in-processing). These techniques successfully reduced performance discrepancies across ethnic groups while maintaining strong overall predictive accuracy. However, this study underscores the critical importance of improving the completeness and consistency of data recording. Enhancing the quality of such data would provide more representative training examples, improving both the fairness and reliability of ML models and ultimately enabling more equitable healthcare outcomes.

In summary, this study advances the use of ML for LOS prediction by integrating fairness-focused methodologies, thereby contributing to equitable healthcare delivery for disadvantaged populations. These findings demonstrate the potential of ML models to enhance care planning, optimise resource allocation, and align with broader policy initiatives, such as the RLoS program. Future research should explore the scalability of these methods across different healthcare systems and address data gaps to further improve fairness and generalisability.

Limitations

While the techniques employed in this study show promise, challenges remain in achieving a balance between fairness and accuracy, as highlighted by the results, addressing intersectional biases, and ensuring generalisability across diverse patient populations. Healthcare datasets are often constrained by privacy concerns and small sample sizes, further complicating the implementation of effective bias mitigation strategies. The dataset of Welsh patients used in this study exhibited significant imbalances in ethnic group representation, limiting the model’s ability to generalise and contributing to performance disparities across groups. Additionally, missing and incomplete data, categorised as “unknown” during preprocessing, may have obscured important nuances or introduced biases. Future research should focus on acquiring complete and consistent datasets, particularly for sensitive attributes such as ethnicity and socioeconomic status, to enhance model fairness and accuracy. Additionally, consistent annual health checks for the LD cohort could enhance data integrity and the accuracy of time-varying predictors such as BMI, smoking status, alcohol intake, and medication history (53).

Future research directions

Future research should focus on expanding datasets to include larger and more diverse cohorts, enhancing applicability across regions, healthcare systems, and population groups. Incorporating longitudinal data and social determinants of health, such as income, education, living composition (e.g., whether a patient with LD lives alone or with a carer), and communication or mobility difficulties, could further refine predictions and improve utility for this cohort. Addressing intersectional biases based on overlapping attributes (e.g., age, ethnicity, and socioeconomic status) and developing advanced fairness algorithms to balance equity and accuracy remain essential. For real-world implementation, integrating the models into clinical decision support systems for adults with LD and conducting pilot studies in hospitals are key steps to evaluating usability and effectiveness. Training clinicians and administrators to interpret AI-driven predictions will be crucial for fostering trust and adoption. Lastly, extending the binary classification of hospital stays (short vs. long) into more granular ordinal or continuous prediction models could provide deeper insights into healthcare needs and outcomes for the study cohort.

Data availability statement

The datasets presented in this article are not readily available as all proposals to use SAIL data are subject to review by the independent IGRP. The anonymised individual-level data sources used in this study are available in the SAIL Databank at Swansea University, Swansea, UK, Before any data can be accessed, approval must be given by the IGRP. The IGRP gives careful consideration to each project to ensure proper and appropriate use of SAIL data. When access has been granted, it is gained through a privacy-protecting safe haven and remote access system referred to as the SAIL Gateway. SAIL has established an application process to be followed by anyone who would like to access data via SAIL at: https://www.saildatabank.com/application-process.

Ethics statement

The studies involving humans were approved by SAIL Databank’s independent Information Governance Review Panel (IGRP Project: 1375). The studies were conducted in accordance with the local legislation and institutional requirements. Written informed consent for participation was not required from the participants or the participants’ legal guardians/next of kin in accordance with the national legislation and institutional requirements.

Author contributions

EA: Conceptualization, Data curation, Formal Analysis, Investigation, Methodology, Validation, Visualization, Writing – original draft, Writing – review & editing. RK: Data curation, Writing – review & editing. GC: Conceptualization, Data curation, Formal Analysis, Funding acquisition, Investigation, Methodology, Project administration, Supervision, Validation, Visualization, Writing – review & editing. AA: Data curation, Funding acquisition, Software, Writing – review & editing. FZ: Data curation, Funding acquisition, Writing – review & editing. NK: Data curation, Writing – review & editing. DF: Data curation, Writing – review & editing. GJ: Funding acquisition, Supervision, Writing – review & editing. RK: Writing – review & editing. SG: Data curation, Formal Analysis, Funding acquisition, Supervision, Writing – review & editing.

Funding

The authors declare financial support was received for the research, authorship, and/or publication of this article. Data-driven machinE-learning aided stratification and management of multiple long-term COnditions in adults with intellectual disabilitiEs (DECODE) project (NIHR203981) is funded by the NIHR AI for Multiple Long-term Conditions (AIM) Programme.

Acknowledgments

This work uses data provided by patients and collected by the NHS as part of their care and support. We would also like to acknowledge all data providers who make anonymised data available for research.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declare that no Generative AI was used in the creation of this manuscript.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fdgth.2025.1538793/full#supplementary-material

References

1. Emerson E, Heslop P. A working definition of learning disabilities. In: Durham: Improving Health & Lives: Learning Disabilities Observatory. Durham: Improving Health & Lives: Learning Disabilities Observatory (2010).

Google Scholar

2. Mencap. Data from: How common is learning disability? Available online at: https://www.mencap.org.uk/learning-disability-explained/research-and-statistics/how-common-learning-disability (accessed January 2024).

Google Scholar

3. Mencap. Data from: Health inequalities. Available online at: https://www.mencap.org.uk/learning-disability-explained/research-and-statistics/health/health-inequalities (accessed January 2024).

Google Scholar

4. White A, Sheehan R, Ding J, Roberts C, Magill N, Keagan-Bull R, et al. Learning from lives and deaths – people with a learning disability and autistic people (LeDeR) report for 2021 (Tech. rep.). King’s College London (2022).

Google Scholar

5. Tyrer F, Morriss R, Kiani R, Gangadharan SK, Kundaje H, Rutherford MJ. Comparing the number and length of primary care consultations in people with and without intellectual disabilities and health needs: observational cohort study using electronic health records. Fam Pract. (2022) 41: cmac135. doi: 10.1093/fampra/cmac135

Crossref Full Text | Google Scholar

6. Tyrer F, Morriss R, Kiani R, Gangadharan SK, Kundaje H, Rutherford MJ. Health needs and their relationship with life expectancy in people with and without intellectual disabilities in England. Int J Environ Res Public Health. (2022) 19:6602. doi: 10.3390/ijerph19116602

PubMed Abstract | Crossref Full Text | Google Scholar

7. Tyrer F, Morriss R, Kiani R, Gangadharan SK, Rutherford MJ. Mortality disparities and deprivation among people with intellectual disabilities in England: 2000–2019. J Epidemiol Community Health. (2022) 76:168–74. doi: 10.1136/jech-2021-216798

PubMed Abstract | Crossref Full Text | Google Scholar

8. Tyrer F, Kiani R, Rutherford MJ. Mortality, predictors and causes among people with intellectual disabilities: a systematic narrative review supplemented by machine learning. J Intellect Dev Disabil. (2021) 46:102–14. doi: 10.3109/13668250.2020.1834946

PubMed Abstract | Crossref Full Text | Google Scholar

9. Heslop P, Blair P, Fleming P, Hoghton M, Marriott A, Russ L, et al. Confidential inquiry into premature deaths of people with learning disabilities (CIPOLD) (Tech. rep.). Bristol: Norah Fry Research Centre (2013).

Google Scholar

10. Carey IM, Shah SM, Hosking FJ, DeWilde S, Harris T, Beighton C, et al. Health characteristics and consultation patterns of people with intellectual disability: a cross-sectional database study in English general practice. Br J Gen Pract. (2016) 66:e264–70. doi: 10.3399/bjgp16X684301

PubMed Abstract | Crossref Full Text | Google Scholar

11. Hatton C, Glover G, Emerson E, Brown I. Data from: Learning disabilities observatory. People with learning disabilities in England in 2015: main report (2016). Available online at: https://www.gov.uk/government/uploads/system/uploads/attachment_data/file/613182/PWLDIE_2015_main_report_NB090517.pdf (Accessed February 06, 2025).

Google Scholar

12. Stone K, Zwiggelaar R, Jones P, Mac Parthaláin N. A systematic review of the prediction of hospital length of stay: towards a unified framework. PLOS Digit Health. (2022) 1:e0000017. doi: 10.1371/journal.pdig.0000017

PubMed Abstract | Crossref Full Text | Google Scholar

13. Ince R, Glasby J, Miller R, Glasby AM. “Why are we stuck in hospital?” Understanding delayed hospital discharges for people with learning disabilities and/or autistic people in long-stay hospitals in the UK. Health Soc Care Community. (2022) 30:e3477–92. doi: 10.1111/hsc.13964

PubMed Abstract | Crossref Full Text | Google Scholar

14. Rodziewicz TL, Hipskind JE. Medical error reduction and prevention. In: StatPearls. StatPearls Publishing (2023). Available online at: https://www.ncbi.nlm.nih.gov/books/NBK499956/ (accessed January 2024).

Google Scholar

15. Alper E, O’Malley TA, Greenwald J, Aronson M, Park L. Hospital discharge and readmission. In: Auerbach AD, Han Li, MD, editors. UpToDate Waltham: UpToDate. Alphen aan den Rijn, Netherlands: Wolters Kluwer (2023). Available online at: https://www.uptodate.com/contents/hospital-discharge-and-readmission? (accessed January 2024).

Google Scholar

16. Loch AA. Discharged from a mental health admission ward: is it safe to go home? a review on the negative outcomes of psychiatric hospitalization. Psychol Res Behav Manag. (2014) 7:137–45. doi: 10.2147/PRBM.S35061

PubMed Abstract | Crossref Full Text | Google Scholar

17. Yang CC, Bamodu OA, Chan L, Chen JH, Hong CT, Huang YT, et al. Risk factor identification and prediction models for prolonged length of stay in hospital after acute ischemic stroke using artificial neural networks. Front Neurol. (2023) 14:1085178. doi: 10.3389/fneur.2023.1085178

PubMed Abstract | Crossref Full Text | Google Scholar

18. Ricciardi C, Marino MR, Trunfio TA, Majolo M, Romano M, Amato F, et al. Evaluation of different machine learning algorithms for predicting the length of stay in the emergency departments: a single-centre study. Front Digit Health. (2024) 5:1323849. doi: 10.3389/fdgth.2023.1323849

PubMed Abstract | Crossref Full Text | Google Scholar

19. Ippoliti R, Falavigna G, Zanelli C, Bellini R, Numico G. Neural networks and hospital length of stay: an application to support healthcare management with national benchmarks and thresholds. Cost Eff Resour Alloc. (2021) 19:1–20. doi: 10.1186/s12962-021-00322-3

PubMed Abstract | Crossref Full Text | Google Scholar

20. Agarwal C, Krishna S, Saxena E, Pawelczyk M, Johnson N, Puri I, et al. Openxai: towards a transparent evaluation of model explanations. Adv Neural Inf Process Syst. (2022) 35:15784–99. doi: 10.48550/arXiv.2206.11104

Crossref Full Text | Google Scholar

21. Chen RJ, Wang JJ, Williamson DF, Chen TY, Lipkova J, Lu MY, et al. Algorithmic fairness in artificial intelligence for medicine and healthcare. Nat Biomed Eng. (2023) 7:719–42. doi: 10.1038/s41551-023-01056-8

PubMed Abstract | Crossref Full Text | Google Scholar

22. Agarwal A, Beygelzimer A, Dudik M, Langford J, Wallach H. A reductions approach to fair classification. In: Dy J, Krause A, editors. Proceedings of the 35th International Conference on Machine Learning (2018), vol. 80. p. 60–9.

Google Scholar

23. Mittermaier M, Raza MM, Kvedar JC. Bias in AI-based models for medical applications: challenges and mitigation strategies. NPJ Digit Med. (2023) 6:113. doi: 10.1038/s41746-023-00858-z

PubMed Abstract | Crossref Full Text | Google Scholar

24. Mann C, Jun GT, Tyrer F, Kiani R, Lewin G, Gangadharan SK. A scoping review of clusters of multiple long-term conditions in people with intellectual disabilities and factors impacting on outcomes for this patient group. J Intellect Disabil. (2023) 27:1045–61. doi: 10.1177/17446295221107275

PubMed Abstract | Crossref Full Text | Google Scholar

25. Abakasanga E, Kousovista R, Cosma G, Jun GT, Kiani R, Gangadharan S. Cluster and trajectory analysis of multiple long-term conditions in adults with learning disabilities. In: International Conference on AI in Healthcare. Springer (2024). p. 3–16.

Google Scholar

26. Caton S, Haas C. Fairness in machine learning: a survey. ACM Comput Surv. (2024) 56:1–38. doi: 10.1145/3616865

Crossref Full Text | Google Scholar

27. Pessach D, Shmueli E. A review on fairness in machine learning. ACM Comput Surv. (2022) 55:51. doi: 10.1145/3494672

Crossref Full Text | Google Scholar

28. Hardt M, Price E, Price E, Srebro N. Equality of opportunity in supervised learning. In: Lee D, Sugiyama M, Luxburg U, Guyon I, Garnett R editors. Advances in Neural Information Processing Systems. Curran Associates, Inc. (2016), vol. 29.

Google Scholar

29. Kim MK, Rouphael C, McMichael J, Welch N, Dasarathy S. Challenges in and opportunities for electronic health record-based data analysis and interpretation. Gut Liver. (2024) 18:201. doi: 10.5009/gnl230272

PubMed Abstract | Crossref Full Text | Google Scholar

30. Lyons RA, Jones KH, John G, Brooks CJ, Verplancke JP, Ford DV, et al. The SAIL databank: linking multiple health and social care datasets. BMC Med Inform Decis Mak. (2009) 9:1–8. doi: 10.1186/1472-6947-9-3

PubMed Abstract | Crossref Full Text | Google Scholar

31. Ford DV, Jones KH, Verplancke JP, Lyons RA, John G, Brown G, et al. The SAIL databank: building a national architecture for e-health research and evaluation. BMC Health Serv Res. (2009) 9:1–12. doi: 10.1186/1472-6963-9-157

PubMed Abstract | Crossref Full Text | Google Scholar

32. Rodgers SE, Demmler JC, Dsilva R, Lyons RA. Protecting health data privacy while using residence-based environment and demographic data. Health Place. (2012) 18:209–17. doi: 10.1016/j.healthplace.2011.09.006

PubMed Abstract | Crossref Full Text | Google Scholar

33. Jones KH, Ford DV, Jones C, Dsilva R, Thompson S, Brooks CJ, et al. A case study of the secure anonymous information linkage (SAIL) gateway: a privacy-protecting remote access system for health-related research and evaluation. J Biomed Inform. (2014) 50:196–204. doi: 10.1016/j.jbi.2014.01.003

PubMed Abstract | Crossref Full Text | Google Scholar

34. Benson T. The history of the read codes: the inaugural james read memorial lecture 2011. Inform Prim Care. (2011) 19:173–82. doi: 10.14236/jhi.v19i3.811

PubMed Abstract | Crossref Full Text | Google Scholar

35. Cartwright DJ. ICD-9-CM to ICD-10-CM codes: what? why? how?. Adv Wound Care. (2013) 2:588–92. doi: 10.1089/wound.2013.0478

PubMed Abstract | Crossref Full Text | Google Scholar

36. NHS Digital. Health and care of people with learning disabilities, experimental statistics: 2018 to 2019 [pas] (Tech. rep.). NHS (2020). Available online at: https://digital.nhs.uk/data-and-information/publications/statistical/health-and-care-of-people-with-learning-disabilities/experimental-statistics-2018-to-2019/condition-prevalence (accessed January 2024).

Google Scholar

37. StatsWales. WIMD indicator data from 2019 (Tech. rep.). StatsWales (2019). Available online at: https://statswales.gov.wales/Catalogue/Community-Safety-and-Social-Inclusion/Welsh-Index-of-Multiple-Deprivation/WIMD-2019 (accessed January 2024).

Google Scholar

38. Cournane S, Byrne D, Conway R, O’Riordan D, Coveney S, Silke B. Social deprivation and hospital admission rates, length of stay and readmissions in emergency medical admissions. Eur J Intern Med. (2015) 26:766–71. doi: 10.1016/j.ejim.2015.09.019

PubMed Abstract | Crossref Full Text | Google Scholar

39. AlJuburi G, Laverty AA, Green SA, Phekoo KJ, Bell D, Majeed A. Socio-economic deprivation and risk of emergency readmission and inpatient mortality in people with sickle cell disease in england: observational study. J Public Health. (2013) 35:510–7. doi: 10.1093/pubmed/fdt100

PubMed Abstract | Crossref Full Text | Google Scholar

40. Akbari A, Torabi F, Bedston S, Lowthian E, Abbasizanjani H, Fry R, et al. Developing a research ready population-scale linked data ethnicity-spine in wales. medRxiv [Preprint]. (2022). doi: 10.1101/2022.11.28.22282810

Crossref Full Text | Google Scholar

41. Ghosh AK, Geisler BP, Ibrahim S. Racial/ethnic and socioeconomic variations in hospital length of stay: a state-based analysis. Medicine. (2021) 100:e25976. doi: 10.1097/MD.0000000000025976

PubMed Abstract | Crossref Full Text | Google Scholar

42. Salive ME. Multimorbidity in older adults. Epidemiol Rev. (2013) 35:75–83. doi: 10.1093/epirev/mxs009

PubMed Abstract | Crossref Full Text | Google Scholar

43. Rahimian F, Salimi-Khorshidi G, Payberah AH, Tran J, Ayala Solares R, Raimondi F, et al. Predicting the risk of emergency admission with machine learning: development and validation using linked electronic health records. PLoS Med. (2018) 15:e1002695. doi: 10.1371/journal.pmed.1002695

PubMed Abstract | Crossref Full Text | Google Scholar

44. Mistry C, Palin V, Li Y, Martin GP, Jenkins D, Welfare W, et al. Development and validation of a multivariable prediction model for infection-related complications in patients with common infections in UK primary care and the extent of risk-based prescribing of antibiotics. BMC Med. (2020) 18:1–13. doi: 10.1186/s12916-020-01581-2

PubMed Abstract | Crossref Full Text | Google Scholar

45. Liu J, Feng Z, Gao R, Liu P, Meng F, Fan L, et al. Analysis of risk factors for autoimmune thyroid disease based on blood indicators and urinary iodine concentrations. Front Endocrinol. (2024) 15:1453828. doi: 10.3389/fendo.2024.1453828

PubMed Abstract | Crossref Full Text | Google Scholar

46. Nagao T, Nogawa K, Sakata K, Morimoto H, Morita K, Watanabe Y, et al. Effects of alcohol consumption and smoking on the onset of hypertension in a long-term longitudinal study in a male workers’ cohort. Int J Environ Res Public Health. (2021) 18:11781. doi: 10.3390/ijerph182211781

PubMed Abstract | Crossref Full Text | Google Scholar

47. Zeleke AJ, Palumbo P, Tubertini P, Miglio R, Chiari L. Machine learning-based prediction of hospital prolonged length of stay admission at emergency department: a gradient boosting algorithm analysis. Front Artif Intell. (2023) 6:1179226. doi: 10.3389/frai.2023.1179226

PubMed Abstract | Crossref Full Text | Google Scholar

48. Zeng X. Length of stay prediction model of indoor patients based on light gradient boosting machine. Comput Intell Neurosci. (2022) 2022:9517029. doi: 10.1155/2022/9517029

PubMed Abstract | Crossref Full Text | Google Scholar

49. Daghistani TA, Elshawi R, Sakr S, Ahmed AM, Al-Thwayee A, Al-Mallah MH. Predictors of in-hospital length of stay among cardiac patients: a machine learning approach. Int J Cardiol. (2019) 288:140–7. doi: 10.1016/j.ijcard.2019.01.046

PubMed Abstract | Crossref Full Text | Google Scholar

50. Dorning H, Davies A, Blunt I. Focus on: people with mental ill health and hospital use (Tech. rep.). QualityWatch (2015).

Google Scholar

51. Siddiqui N, Dwyer M, Stankovich J, Peterson G, Greenfield D, Si L, et al. Hospital length of stay variation and comorbidity of mental illness: a retrospective study of five common chronic medical conditions. BMC Health Serv Res. (2018) 18:498. doi: 10.1186/s12913-018-3316-2

PubMed Abstract | Crossref Full Text | Google Scholar

52. NS England. Data from: Reducing length of stay. Available online at: https://www.england.nhs.uk/urgent-emergency-care/reducing-length-of-stay/ (accessed March 2024).

Google Scholar

53. Carey IM, Hosking F, Harris T, De Wilde S, Beighton C, Cook DG. An evaluation of the effectiveness of annual health checks and quality of health care for adults with intellectual disability: an observational study using a primary care database. Health Serv Deliv Res. (2017) 5:1–170. doi: 10.3310/hsdr05250

Crossref Full Text | Google Scholar

Keywords: learning disabilities, length of stay, bias mitigation, threshold optimiser, exponentiated gradient

Citation: Abakasanga E, Kousovista R, Cosma G, Akbari A, Zaccardi F, Kaur N, Fitt D, Jun GT, Kiani R and Gangadharan S (2025) Equitable hospital length of stay prediction for patients with learning disabilities and multiple long-term conditions using machine learning. Front. Digit. Health 7:1538793. doi: 10.3389/fdgth.2025.1538793

Received: 3 December 2024; Accepted: 27 January 2025;
Published: 14 February 2025.

Edited by:

Xiaofeng Li, Heilongjiang International University, China

Reviewed by:

Kumar P., Rajalakshmi Engineering College, India
Prosper Yeng, Norwegian University of Science and Technology Gjøvik, Norway

Copyright: © 2025 Abakasanga, Kousovista, Cosma, Akbari, Zaccardi, Kaur, Fitt, Jun, Kiani and Gangadharan. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Georgina Cosma, Zy5jb3NtYUBsYm9yby5hYy51aw==

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.

Research integrity at Frontiers

Man ultramarathon runner in the mountains he trains at sunset

94% of researchers rate our articles as excellent or good

Learn more about the work of our research integrity team to safeguard the quality of each article we publish.


Find out more