- School of Electronics Engineering, VIT-AP University, Amaravati, Andhra Pradesh, India
Coronary artery disease poses a significant challenge in decision-making when predicting the length of stay for a hospitalized patient. This study presents a predictive model—a Particle Swarm Optimized-Enhanced NeuroBoost—that combines the deep autoencoder with an eXtreme gradient boosting model optimized using particle swarm optimization. The model uses a fuzzy set of rules to categorize the length of stay into four distinct classes, followed by data preparation and preprocessing. In this study, the dimensionality of the data is reduced using deep neural autoencoders. The reconstructed data obtained from autoencoders is given as input to an eXtreme gradient boosting model. Finally, the model is tuned with particle swarm optimization to obtain optimal hyperparameters. With the proposed technique, the model achieved superior performance with an overall accuracy of 98.8% compared to traditional ensemble models and past research works. The model also scored highest in other metrics such as precision, recall, and particularly F1 scores for all categories of hospital stay. These scores validate the suitability of our proposed model in medical healthcare applications.
1 Introduction
Nearly 31% of all deaths around the globe are due to cardiovascular diseases, among which coronary artery disease (CAD) is the most common. By 2030, an estimated 22 million people will likely be affected by CAD (Workina et al., 2022). The risk factors causing CAD are inactivity, smoking, excessive alcohol consumption, poor diet, and a sedentary lifestyle (Balen et al., 2024). To determine the patient mortality risk in the health sector, clinicians often consider the patient's length of stay (LOS) as a critical indicator (Awad et al., 2017). The statistics show that approximately 33% of older adult patients admitted to intensive care units (ICUs) due to prolonged LOS do not survive (Li et al., 2024). Most statistical analysis and current research solely focus on predicting the overall target LOS or singular LOS class, i.e., short or long stay. This limits the ability to determine the important prediction insights of each hospital stay duration, showing a significant research gap. Therefore, to address this gap, our study focuses on four distinct LOS classes: short, medium, long, and extended stays (Abdurrab et al., 2024; Junior et al., 2024; Momo et al., 2024). A short stay typically signifies quick recoveries, though it may sometimes indicate early mortality. Whereas long stay often indicates the presence of severe health issues or long-term illnesses. On the other hand, extended stay refers to prolonged hospital stay due to severe illness (Heyland et al., 2015). Meanwhile, a medium stay indicates ongoing treatment or monitoring of moderate medical issues.
In the medical sector, the LOS plays a crucial role in managing various health conditions such as diabetes (Ata et al., 2023), cancer (Jung et al., 2023), tumors (Alzubi et al., 2019; Muhlestein et al., 2019), chronic kidney disease (Neyra et al., 2023), inflammatory conditions (Mangalesh et al., 2023), and infectious diseases (Saadatmand et al., 2023). The accurate prediction of LOS at the preliminary stage can aid clinicians and patients in decision-making about treatment and recovery planning, resource, and budget allocation and help patients reduce mortality rate (Asadi-Lari et al., 2004). Therefore, it highlights the need for a single, powerful machine-learning model to predict the LOS. Moreover, researchers use various artificial intelligence techniques to classify the LOS with automated decision-making systems (Masood et al., 2022). However, the challenge lies in dealing with the variability of data and the inherent uncertainties in predicting LOS. Therefore, with advanced machine learning (ML), deep learning (DL), comprehensive medical data, and fuzzy logic principles, our model aims to develop a robust classifier model capable of predicting LOS for CAD patients. This work proposes an enhanced modeling strategy with Particle Swarm Optimization (PSO) to improve accuracy. The main contributions of our research are:
• Developing a Particle Swarm Optimized-Enhanced NeuroBoost (PSO-XnB) model combining PSO with eXtreme Gradient Boosting (XGBoost). In this technique, the enhancement is done using deep autoencoder techniques for feature selection and dimensionality reduction.
• Developing Tuning rules of PSO to maintain convergence of optimal solutions and balance exploration and exploitation of the model hyperparameters.
• For multi-class LOS classification, the PSO-XnB model is evaluated on performance metrics, i.e., accuracy, F1 score, precision, recall, and area under the curve (AUC) score. Then, it is compared with other traditional ensemble models.
This study is structured as follows: Section 2 presents the related works on the importance of feature selection and classifiers for predictions. Section 3 elaborates on the proposed PSO-XnB framework, including the methodology and model design. Section 4 presents experimental observations and performance results comparing the PSO-XnB model with traditional ensemble methods. Finally, Section 5 concludes the findings and discusses potential future directions.
2 Related works
This section presents a literature survey highlighting previous research works limited to the past 4 to 5 years based on the importance of feature selection and the role of classifiers for predictions. This section will also present various feature selection methods used with classifiers to improve the medical diagnostic process. Shah et al. (2020) focused on improving the feature selection by combining the mean fisher-based feature selection algorithm (MFFSA) and accuracy-based feature selection algorithm (AFSA). The authors used a hybrid combination of feature selection methods to train and using principal component analysis (PCA), they selected the best features. These selected features were further trained with an RBF-based support vector machine (SVM); the overall accuracy scored was 92.1% using the Switzerland heart disease dataset on 123 instances. The Fisher score algorithm (FSA), F_score algorithm (FA), and extra trees classifier algorithms (ETCA) were used by Nasarian et al. (2020) to select features with the highest scores, specifically targeting the top 30%, 40%, and 50%. Subsequently, the selected features were trained with four baseline classifiers, namely, Decision Tree (DT), Gaussian Naive Bayes (GNB), Random Forest (RF), and XGBoost classifiers. The average accuracy scored from the four models was 83.94%, 81.58%, and 92.58%, using Hungarian, Long-beach-va, and Z-Alizadeh Sani datasets, respectively. Another work by Amarbayasgalan et al. (2021) used a two-step approach where PCA first divides initial data into commonly distributed and highly biased groups. In the next step, two variational autoencoders (VAE) were used to add samples in highly biased groups. Finally, highly biased group samples were trained with two Deep Neural Networks (DNN) models, and the model scored 89.2% accuracy and 91.5% F1 score. Mienye and Sun (2021) proposed a DL approach to improve heart disease prediction, utilizing a stacked sparse autoencoder network (SSAE) for feature selection. The selected features were trained with the soft-max classifier optimized by PSO; with this multilayer architecture, the authors scored 97.3% accuracy using the Framingham dataset. Another work used a sparse autoencoder (SAE) and softmax classifier; this combination scored 91% accuracy, as reported by Ebiaredoh-Mienye et al. (2020). This work did not use any optimization step; therefore, the model's parameters were not fully optimized for heart disease prediction. Ali et al. (2019) focused on eliminating irrelevant features followed by preprocessing with the statistical model chi-square (χ2). The over-fitting issue was minimized by configuring the χ2-DNN model through an exhaustive search strategy, because of which the model achieved 93.3% accuracy in heart disease prediction.
In discussing the use of ML models, Miriyala et al. (2021) used a single LGBM to diagnose CAD in 1,190 instances, reporting 93.3% accuracy; the model has an overfit issue, which can be solved through optimization. Barfungpa et al. (2023) optimized features using an enhanced sparrow search algorithm (E-SSA). Their focus was to minimize the dimensionality of data using deep-dense residual attention Aquila convolutional network (Deep-DenseAquilaNet). The model's approach was to update the weights using the Aquila optimization algorithm; this resulted in 99.57% accuracy using the comprehensive heart dataset with <2,000 instances. A heart disease prediction model employing a Chaos Game Optimization Recurrent Neural Network (CGO-RNN) model was reported by Alam and Muqeem (2023). The authors used integrated Kernel Principal Component Analysis (KPCA) as a strategic choice for handling the data's dimensionality and computational intensity. The model novelty helped the authors to achieve an accuracy of 98 ± 0.9%. However, the works reported by Barfungpa et al. (2023) and Alam and Muqeem (2023) could not converge during optimization.
As the research advanced, the LOS performance was improved using various ML models. In this context, Harerimana et al. (2021) propose three classes of LOS, where a hierarchical attention network (HAN) was trained with the MIMIC-III text-oriented dataset, achieving an AUC-ROC score of 0.82. Zou et al. (2023) used MIMIC-III and divided it into three datasets with different feature sizes, and the LOS labels from the datasets were extracted from the PostgreSQL DBMS. The authors proposed a generative adversarial network (GAN) called Wasserstein-GAN to predict the LOS class with a gradient penalty. For three datasets and various LOS ranges, the highest class scored 96.6% accuracy among other classes. Another work by Hempel et al. (2023) performed prediction using the MIMIC-IV dataset; in this work, the LOS was categorized into two classes: short and long stay. The authors used the default parameters of the model in optimization and classifier training using Logistic Regression (LR), RF, SVM, and XGBoost; the result shows that RF scored the highest accuracy of 81%. A framework for LOS prediction using RF with three over-sampling and three under-sampling techniques was proposed by Alsinglawi et al. (2022). The oversampling technique using RF scored 98% accuracy with a 95% confidence interval. González-Nóvoa et al. (2023) developed the XGBoost model optimized with Bayesian techniques to predict early ICU readmissions. The authors utilized SHAP to determine important features, scoring 0.92 ± 0.03 AUC-ROC.
In discussing LOS prediction using the MIMIC dataset for other diseases, El-Rashidy et al. (2022) developed a predictive model for detecting sepsis in patients admitted to the post-ICU during the initial 6 h of their stay. This innovative approach combined the strengths of the non-dominated sorting genetic algorithm-II (NSGA-II) and neural networks to extract optimal features. This hybrid model was followed by a classifier-stacked deep ensemble learning model to train the dataset. This model scored an accuracy of 91.3% and AUC score of 0.906. The effect of lung cancer on ICU patients was studied by Qian et al. (2023) using the MIMIC-III dataset. The study analyzed 1,242 ICU admissions, scaling them on illness severity AUC score; the predicted short-term and long-term mortality scores were 0.714 and 0.717, respectively. Bozkurt and Aşuroglu (2023) targeted their model in LOS prediction for patients admitted with breast, lung, prostate, and stomach cancer. The authors used the MIMIC-IV dataset and LR-based feature selection. The selected features were trained on five ML models: LR, DT, RF, SVM, and Multilayer Perceptron. The performance given by RF was the highest among others; the F1 scores ranged from 0.73 to 0.82, and AUC-ROC scores ranged from 0.88 to 0.96. The research discussed in this section 2 is compared in Table 1 according to the problem statement, the models used, and the limitations observed.
The recent pandemic has forced researchers to use nature-inspired metaheuristic optimization algorithms to target complex datasets. Among these, algorithms such as the capuchin search algorithm (Braik et al., 2023a), particle swarm optimization sine cosine algorithm (Somgiat and Chansamorn, 2022), whale optimizer (Nadimi-Shahraki et al., 2022), and snake optimizer (Braik et al., 2023b) have shown promising results in ML and DL applications. However, our study uses PSO optimization due to its robustness in determining complex, multidimensional search spaces from clinical data and its convergence toward optimal solutions (Kennedy and Eberhart, 1995). The existing works primarily focused on the binary classification of the LOS. In contrast, our proposed model PSO-XnB focuses on showing the best performance across multiple categories where the model is implemented with deep autoencoders for feature selection and dimensionality reduction, XGBoost for predictive modeling, and PSO for fine-tuning the model's hyperparameters. With the development of specific tuning rules, our model also achieves a balance between exploration and exploitation, leading to enhanced model convergence and accuracy. Section 3 discusses the strategic choices for developing PSO-XnB in detail.
3 Materials and methods
Our methodology of the PSO-XnB framework is depicted in Figure 1. The work starts with data preparation, followed by preprocessing, discussed in Section 3.1.
The framework of PSO-XnB model development to predict LOS for CAD patients is discussed in Section 3.2.
3.1 Dataset preparation and preprocessing
This study uses the MIMIC-III clinical dataset (Goldberger et al., 2000; Johnson et al., 2016). From the large dataset, our selection process exclusively focuses on medical cases relevant to CAD disease. Therefore, the keyword extraction method is performed from the “admit diagnosis” feature. Keywords such as CAD, heart, coronary, chest pain, myocardial infarction, angina, cardiac aortic, and STEMI were utilized for extraction. After this extraction, our final dataset included 9,989 patient instances with 23 features. These features cover admission details, clinical and care metrics, and medical outcomes, including diagnoses, procedures, inputs/outputs, and length of stay. The continuous patient records were gathered from the dataset. Data handling practices were ensured to maintain the study's integrity, reliability, and reproducibility. The “LOSdays” (length of stay in days) is selected as a target variable to predict hospital stay durations. This work categorizes the LOS classes with fuzzy followed by three steps. In the first step, the fuzzy sets for each LOS category are defined as short, medium, long, and extended stays. The second step is to define fuzzy rules based on data summary statistical analysis to understand the inherent variability and patterns. The rules given are as follows (Jena et al., 2022):
• IF LOSDays are ≤ 4.33 days, THEN the stay is likely a “short stay.”
• IF LOSDays are >4.33 days AND are ≤ 6.75 days, THEN the stay is likely “medium stay.”
• IF LOSDays are >6.75 days AND are ≤ 10.21 days, THEN the stay is likely a “long stay.”
• IF LOSDays are >10.21 days, THEN the stay is likely an “extended stay.”
This observation defines fuzzy sets with boundaries reflecting the data's natural distribution, not arbitrary intervals. These rules can capture overlaps and gradations of the LOS class. In the final step through the inference mechanism, fuzzy rules are applied with the trapezoidal membership to identify the complex patterns where a patient's LOS can simultaneously exhibit characteristics of multiple categories. These rules help to make predictions more accurate and adaptable because they are considered with different possibilities and variations in LOS. Further, label encoding was used as a preprocessing step to process the categorical variables into numerical format (Boulif et al., 2023), and this preprocess data was split into training and testing sets. The LOS classes obtained from a fuzzy set of rules are distributed across training and testing data, which is visualized in Figure 2, revealing the data imbalance. Therefore, the Synthetic Minority Over Sampling Technique (SMOTE), a standard technique in ML, is applied to balance the data. This technique effectively addresses the imbalance issue and augments the training and testing datasets from 7,991 and 1,998 instances to 18,188 and 4,548 instances. The data augmentation does not fabricate new data randomly. Instead, it analyses the existing data to identify underlying patterns specifically within the underrepresented classes and creates additional synthetic data to enhance those minority class representations.
3.2 Proposed PSO-XnB model
After data preprocessing, the MIMIC III dataset shows high dimensionality due to the multiplicity of features and instances. Therefore, our work uses a deep autoencoder, an artificial neural network developed with TensorFlow (Abadi et al., 2016) and Keras (Chollet, 2015). The deep autoencoder can find patterns within the data that are not easily visible due to the complex relationXships between features in medical datasets (Bank et al., 2023; Shinde et al., 2023). Let us denote the original data with the dimensions of m×n, where m and n, denote the number of instances and features. The autoencoder is a combination of encoder and decoder, and it is initialized with an encoder function E(X, θE), where θE determines the network parameters of the encoder E. The encoder network starts with the input layer of the original data Xin that matches the feature space of the entire data where , and ℝn determines the n-dimensional space of real numbers. In this encoder, the two mappings are typically used to reduce the dimensionality of the input data. The first and second mappings E1 and E2 are used as functions to transform data from ℝn → ℝ128 and ℝ128 → ℝ64. The function E1 transforms input data from n-dimensions to 128 dimensions by adding features and E2 then reduces this data from 128 dimensions to 64 dimensions, focusing on obtaining the most relevant features. Thus, and use the rectified linear activation (ReLU) function to introduce non-linearity forming latent encoded representation Z = E2(ReLU(E1(Xin))).
The subsequent decoder function R, mirrors the two encoder mappings as , and . These mappings used in the decoder transform the encoded low-dimensional data back to the reconstructed original data, so the output of the decoder is Xout = R2(ReLU(R1(Z))). The complete operation of the autoencoder A functions as A(X; θE) = R(E(X)), where X, parameterized by θE, determines the process initially transformed by the encoder E into lower dimensional representations, and then the decoder reconstructs the input data from this encoder representation. The optimization of encoder parameters θE continues iterations with the mean square error until the lowest possible difference between the original data and the reconstructed data is achieved. The reconstruction data obtained from the autoencoder at minimal loss shows reduced dimensionality and a focus on important features.
This proposed approach uses the reconstructed data as input for the XGBoost classifier to be optimized with PSO (Farahnakian and Heikkonen, 2018). Le et al. (2019) and Jiang et al. (2020) previously developed the PSO-XGBoost model, but their studies faced challenges in predicting minority classes (Jiang et al., 2020) and handling a high number of input variables (Le et al., 2019). In contrast, our model, PSO-XnB, addresses these issues using autoencoders, an effective approach to predict outcomes across all LOS classes. The flowchart of the PSO-XnB model is shown in Figure 3. Our work uses the XGBoost model, enhanced by a gradient-boosting technique, for efficient performance in real-world applications (Chen and Guestrin, 2016). The model XGBoost trained without regularization could overfit the training data, capturing noise from the underlying patterns (Kigo et al., 2023). Therefore, fine-tuning the hyperparameters is crucial to balance the model's performance, ensuring it generalizes well without overfitting (Yewale et al., 2023). In this study, Particle Swarm Optimization (PSO) is used for hyperparameter tuning, a method inspired by natural patterns in bird flocking and fish schooling introduced by Kennedy and Eberhart (1995). This optimization, combined with deep autoencoder reconstructed data, aims to improve the predictive capacity of the XGBoost model, yielding a highly accurate outcome. The entire model development is structured into seven distinct steps:
1. Encoding Train and Test Input Data:
The reconstructed data Xout and the target variable Y is divided into training and testing sets defined in Equation (1):
Where Xout, train, Ytrain and Xout, test, Ytest represents training and testing input and output data.
2. Defining Hyperparameters for XGBoost:
The hyperparameters of the XGBoost model utilized in this study are represented by a vector, θX = [lr, ne, md, mcw, ss, cbt], and the variables in the vector are detailed in Table 2.
3. Define Search Space and PSO Initialization:
Initially, search space is defined for PSO as a multidimensional space, aiming to optimize different hyperparameters of the XGBoost model within each dimension. The range for each XGBoost hyperparameter used in this process determines the boundaries of the search space (El Dor et al., 2012). Before starting the PSO process, the configuration parameters and their values play an important role in directing the particle's movement throughout the hyperparameter space. Our configurations start with ten particles, each representing a possible solution within the XGBoost hyperparameter search space. This specific number of particles is selected to maintain computational efficiency while providing a diverse set of initial candidate solutions. Six dimensions are categorized for each particle. The hyperparameters of XGBoost are outlined in Table 2. For the learning factors, the local learning factor (C1) and global learning factor (C2) are set at 0.5 and 0.3. The setting value of C1 guides each particle toward its best optimal position by fine-tuning solutions based on past performance. The lower C2 value, on the other hand, encourages particles to seek out new potential solutions in new regions of the search space. Our model using PSO is limited to 50 iterations because this value achieved efficient convergence at initial tests. Finally, ten independent runs were conducted to ensure the reliability of the optimization process. These runs help to validate the stability and replicability of our hyperparameter tuning results. After preliminary testing, the best hyperparameters were ultimately chosen based on observing the optimal balance between computational efficiency and the depth of exploration within the hyperparameter space, ensuring the robustness of our findings.
Each particle's i initial position and velocity in PSO are described by vectors and that correspond to the values of hyperparameters, i.e., learning rate, number of estimators, max depth, min child weight, subsample size, and colsample by tree. The position vector represents the initial point in the search space, which is defined as . The velocity vector sets the initial speed as zero or a value close to zero, allowing particles to explore the search space without bias, defined as.
4. PSO Iterative Process:
In the iteration process t, the particle positions and velocities are updated using the predefined PSO rules. Therefore, the velocity for each dimension d of i is adjusted and updated by Equation (2):
From the above equation, w determines the inertia weight that balances exploration and exploitation, impacting the particle's current velocity, c1 determines cognitive coefficients, and c2 determines social acceleration coefficients. r1 and r2 are random numbers ranging between 0 to 1 providing stochasticity to the search, which is the best individual position found in the d and is the best position obtained in the swarm by any particle found in d. The term represents a particle's memory of its best position, marked as a cognitive component, and represents the particle's cooperation with the swarm as a social component. Therefore, with the iterations, each particle's position in the search space updates according to the velocity . This updated position becomes the candidate solution for the next iteration , defined in Equation (3).
Through these equations, the PSO algorithm allows each particle to explore the search space, adjusting its trajectory based on its own experience and the swarm's collective knowledge. By updating each particle, the swarm converges to the best solution of hyperparameters that results in the best model performance according to the chosen fitness metric.
5. Fitness Evaluation and Model Training with Dimensionality:
The XGBoost model is trained using the hyperparameters represented by the particle's position (Qin et al., 2021), and the F1 score is chosen as the fitness function because it balances precision and recall and can predict all classes accurately. This fitness metric is especially suitable for imbalanced datasets. The fitness function used in this work is defined by Equation (4) as:
The Equation (4) represents f(xi) a fitness function of the particle i, and calculates the F1 score of the XGBoost model trained with the reconstructed data and target, using the hyperparameters of position vectors across all dimensions D for the particle i. However, depending on the specific requirements of the task, other metrics such as accuracy or precision could also serve as the fitness score.
6. PSO Convergence to Best Hyperparameters with Dimensionality:
To obtain the best hyperparameters across all dimensions D, the global best positions defined across each dimension are defined in Equation (5) as:
The PSO algorithm fine-tune the hyperparameters iteratively to minimize a convex loss function. Simultaneously, the F1 score is also used as a fitness measure. This dual-focus strategy ensures the best hyperparameter solutions , which enhances the model's accuracy and reliability. To ensure the reach of the convergence of the PSO, 10 independent runs with 50 iterations are executed. The optimal set of hyperparameters is then selected based on the lowest convex loss observed from the XGBoost model evaluations during these runs, as presented in Table 3.
7. Training and Prediction with the Enhanced NeuroBoost Model with Dimensionality:
Finally, the NeuroBoost model is trained with , obtained from the PSO process defined by Equation (6) as:
The final model is expected to demonstrate enhanced performance due to being fine-tuned with the dimensionally optimized hyperparameters. Further, the model is employed across various classes of LOS as C = {cshort, cmedium, clong, cextended}, where each class in the set belongs to a specific LOS class. The model predictions on the reconstructed test data are defined by the Equation (7) as:
Therefore, the final objective function of the PSO-XnB model, at iterations t, sums over all classes k, and computes with the best parameters given by Equation (8)as:
Where l is a differentiable convex loss function that measures the difference between actual ym, k and predicted test observations of instances m with k, is the regularization term that penalizes the complexity of the model to prevent overfitting. is the predictive function of the model using the best hyperparameters for input features of instances xm across all target classes k. Due to the internal mechanism of XGBoost, the L1 regularization terms get explicitly modified as best based on cost per leaf γbest and the L2 regularization term based on leaf weights λbest, both optimized via hyperparameters defined in the Equation (9):
Here, T is the number of leaves in the tree, and is the vector of scores on the leaves. Finally, the performance metrics are very important to evaluate the performance of a multi-class classification problem, as discussed in Miriyala et al. (2021). In this work, the overall model's performance is observed using accuracy as a primary metric, and the F1 score acts as an alternative metric to understand the positive class predictions from precision and recall across each LOS class. The Precision-Recall (PR) curve and Receiver Operating Characteristic (ROC) curve are also examined to gain insights into the model distinguishing the target classes. These visual representations offer an understanding of its capabilities. The Average Precision (AP) is also derived from the PR curve and effectively summarizes the model's performance for each category (Le et al., 2019).
4 Experimental results and performance observations
In this section, the simulation setup of the PSO-XnB model is performed on a computational environment with NVIDIA® GeForce RTX™ 3050 Ti graphics card, Windows operating system, and an 11th generation Intel Core™ i7-11390H CPU for simulation. PyCharm is the simulation environment used with the Python 3.6.3 programming language (Van Rossum and Drake Jr, 1995). The libraries that were utilized were NumPy, Pandas, Matplotlib, SciPy, scikit-learn (Fandango, 2017), PySwarms (James and Miranda, 2018), Bayesian optimization (Nogueira, 2014), imbalanced-learn (LemaÃŽtre et al., 2017), TensorFlow (Abadi et al., 2016), SHAP (Lundberg and Lee, 2017), and a few other utilities for preprocessing data and evaluating models from scikit-learn. Our primary data source was the MIMIC III dataset, and the Pandas library was used mostly in data preparation. Initially, the reconstructed data is obtained from deep autoencoder data with the minimum MSE loss shown in Figure 4, where the training and validation loss starts at 1.7450 and 0.8830 at Epoch 1. However, over 50 iterations, the training loss consistently reduced to 0.3567, and the fluctuations in the validation loss were identified in Epochs 30 and 50, indicating the challenges in generalization. Therefore, to improve the model's performance toward the data, for further optimization is required.
4.1 PSO-XnB model results
In our study, to predict the LOS, the best XGBoost hyperparameters from PSO are obtained to minimize convex loss, showing the characteristic exploratory nature of PSO. These values across ten independent runs are tabulated in Table 3. Each row corresponds to a separate optimization run, showcasing the unique set of hyperparameters. Figure 5A visualizes the PSO algorithm's efficacy, showcasing the convergence of loss for ten independent runs to a mean of 0.01275. Figure 5B, showing a boxplot, highlights a median loss of 0.0128, indicating a robust optimization process with narrow variability. From Table 3, the best and worst losses are observed as 0.013 and 0.0123, respectively, indicating tight clustering of outcomes and suggesting stability in the optimization process across ten independent runs. The standard deviation is 0.000217, and there is an extremely low variance. This visual and statistical confirmation of the PSO's performance underscores the reliability of our model for healthcare predictive analytics.
Figure 5. Model tuning results of PSO-XnB for ten independent runs. (A) Convergence graph. (B) Loss dispersion graph.
With the selected hyperparameters highlighted in Table 3, the PSO-XnB is trained on the test dataset. The performance metrics of PSO-XnB are calculated and shown in Table 4. The PSO-XnB on extended stay class scored high in precision, recall, and F1 scores metrics, suggesting that the model predicted accurately. However, the perfect recall score for extended stay could indicate a need for more diversity in the training set, potentially limiting the model's ability to handle a range of cases. On the other hand, there is a lower recall score for the medium stay class, indicating that there may be fewer true positive predictions for this category, which could be an area focusing on improving the model. The overall accuracy of the PSO-XnB was 98.8%, proving successful hyperparameter tuning using PSO-XnB. From these performance results, it is clear and promising that the deep autoencoder reconstructed data showed a considerable scope for future research aiming for refinement with the tools for broader applications in healthcare analytics.
4.2 Comparison with other models
In this subsection, the proposed model PSO-XnB is compared with other ensemble models such as Random Forest (RF) (Hempel et al., 2023), Gradient Boosting (GB) (Naemi et al., 2021), Bagging (Tully et al., 2023), and XGBoost (Hempel et al., 2023), to observe the effectiveness of predicting LOS.
The overall accuracy of the other traditional ensemble models, such as RF, GB, Bagging, and XGBoost, scored 85.74%, 69.04%, 86.98%, and 94%. The F1 scores obtained from each ensemble model are compared with PSO-XnB, as shown in Figure 6, and the comparison of performance metrics is shown in Table 5. PSO-XnB showed the highest F1 scores of 0.99 for short stays, 0.98 for medium stays, 0.99 for long stays, and 1.00 for extended stays. The XGBoost scored second highest for three classes, i.e., short stay, medium stay, and long stay, and the RF model performed highest only for extended stay. Compared to the F1 score, gradient-boosting scores are the lowest in all four classes. Along with the F1 score, the precision and recall scores are also shown in Table 5, where the PSO-XnB model scored higher for each LOS class than other ensemble models.
For observing the significant validation, the consistent performance of PSO-XnB over the baseline ensemble models is performed with ten test splits, as shown in Supplementary Table S1. These evaluations are further statistically tested with the paired t-test to show the significance of the four LOS classes, i.e., short stay (SS), medium stay (MS), long stay (LS), and extended stay (ES). The results from the statistical analysis are summarized in Table 6. As reported in Table 6, the PSO-XnB model shows significance over ensemble models across all categories of LOS. For each class of short, medium, and long stay, the p-values are <0.001, showing that our model PSO-XnB is statistically significant compared to other baseline traditional ensemble models. A similar trend of p-values close to 0.001 is observed for an extended LOS stay. These lowest p-values determine a high degree of statistical confidence. Observing the statistical analysis, the PSO-XnB model demonstrates superiority in predicting the LOS categories compared to other baseline ensemble models.
Figure 7A compares ROC-AUC scores as percentages; the perfect score of 1.00 achieved by PSO-XnB represents a rate of 100% with no false positives across all thresholds. In contrast, the XGBoost and Bagging models scored AUC with 98% and 97%, covering 2% and 3% less ROC space than the PSO-XnB model. Similarly, both GB and RF models have an AUC of 96%, covering 4% of the ROC area compared to PSO-XnB. These percentages represent the varying capacities of each model to differentiate between categories. While examining the Precision-Recall curves in Figure 7B, it becomes evident that the PSO-XnB model establishes a standard with an Average Precision (AP) score of 99.65%. Relative to this, the XGBoost, GB, Bagging, and RF model shows a 4%, 5%, 10%, and 11% decrease in an area with an AP with 95.36%, 94.54%, 90.23%, and 88.92%, respectively. In practical terms, these percentages of AP reflect the degree to which each model's performance on precision and recall falls short of the near-perfect score achieved by the PSO-XnB. A lower AP score means the model is either less precise, less sensitive, or both, especially in predicting positive instances within imbalanced datasets (Jiang et al., 2020).
Figure 7. Performance evaluation graphs of PSO-XnB with other ensemble models. (A) ROC-AUC curve. (B) Precision recall curve.
The impact of the PSO-XnB model on features over four LOS classes is determined by SHAP analysis, as shown in Figure 8; the main two features impacting each class are discussed. In Figure 8A, the short stay shows the significance of NumNotes and NumTransfers features, indicating the number of medical notes and transfers between departments on patient records. These are important to predict short stays. Similarly, medium stay in Figure 8B shows NumCallouts determining the number of times staff were called to attend a patient as an impactful feature, along with NumTransfers. The features such as NumDiagnosis and NumNotes highly influence the long stay class in Figure 8C. These features show that the number of documented diagnoses and notes is important for long-stay predictions. Lastly, in Figure 8D, NumDiagnosis and NumProcs are influential for extended stays, where NumProcs determines the medical procedure that the patient has undergone. These high-impactful features on predictions are represented in red, whereas low features are depicted in blue. The insights into features impacting LOS classes can help clinicians optimize resource allocation. Also, hospitals can reduce complication risks and improve capabilities, enhancing patient outcomes.
Figure 8. Feature impact model analysis using SHAP (A) Short stay. (B) Medium stay. (C) Long stay. (D) Extended stay.
From the literature review, some of the limitations observed were resolved with our proposed model PSO-XnB, such as overfitting, sensitivity of feature space dimensionality, and class imbalance. Although the PSO-XnB performs well in interpreting clinical data, its specific design is a limitation when considering its utility for various data analysis applications. Another limitation is that the optimization heavily depends on complex computational processes (Cao et al., 2023). This high computational demand could restrict the model's practical use, especially in healthcare environments. Addressing these challenges will broaden the model's applicability in decision-making and real-time processing, ensuring its utility in diverse healthcare scenarios. The comparison of our proposed model with other works from past literature is shown in Table 7.
5 Conclusion
In this research, the Particle Swarm Optimized-Enhanced NeuroBoost (PSO-XnB) model showed significant achievement in predicting Hospital Length of Stay for CAD patients. This model combines particle swarm optimization with an eXtreme gradient-boosting model that utilizes a deep autoencoder technique for dimensionality reduction. Our proposed model PSO-XnB model showed outstanding performance efficiency with 98.8% accuracy. The model also scored the highest F1 score of all four LOS categories, compared with other ensemble models, with 0.99 for short stays, 0.98 for medium stays, 0.99 for long stays, and 1.00 for extended stays. This proposed model also demonstrated the high scores for other metrics such as precision, recall, and AUC. With accurate LOS predictions, healthcare services and patients can use our PSO-XnB model to take immediate initiatives. In the future, distributed and federated learning will be focused on observing the effects on classifier performance to reduce computational demands.
Data availability statement
The original contributions presented in the study are included in the article/Supplementary material, further inquiries can be directed to the corresponding author.
Author contributions
GM: Writing – review & editing, Writing – original draft, Visualization, Validation, Software, Resources, Project administration, Methodology, Investigation, Formal analysis, Data curation, Conceptualization. AS: Writing – review & editing, Supervision.
Funding
The author(s) declare that no financial support was received for the research, authorship, and/or publication of this article.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher's note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Supplementary material
The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/frai.2024.1381430/full#supplementary-material
Supplementary Table S1. Performance metrics of PSO-XnB and Baseline traditional Ensemble models on ten-test splits.
References
Abadi, M., Barham, P., Chen, J., Chen, Z., Davis, A., Dean, J., et al. (2016). “TensorFlow: a system for large-scale machine learning,” in 12th USENIX Symposium on Operating Systems Design and Implementation (OSDI 16) (Savannah, USA), 265–283.
Abdurrab, I., Mahmood, T., Sheikh, S., Aijaz, S., Kashif, M., Memon, A., et al. (2024). “Predicting the length of stay of cardiac patients based on pre-operative variables—bayesian models vs. machine learning models,” in Healthcare (MDPI), 12. doi: 10.3390/healthcare12020249
Alam, A., and Muqeem, M. (2023). An optimal heart disease prediction using chaos game optimization-based recurrent neural model. Int. J. Inf. Technol. 1:8. doi: 10.1007/s41870-023-01597-w
Ali, L., Rahman, A., Khan, A., Zhou, M., Javeed, A., and Khan, J. A. (2019). An automated diagnostic system for heart disease prediction based on χ2 statistical model and optimally configured deep neural network. IEEE Access 7, 34938–34945. doi: 10.1109/ACCESS.2019.2904800
Alsinglawi, B., Alshari, O., Alorjani, M., Mubin, O., Alnajjar, F., Novoa, M., et al. (2022). An explainable machine learning framework for lung cancer hospital length of stay prediction. Sci. Rep. 12:607. doi: 10.1038/s41598-021-04608-7
Alzubi, J. A., Kumar, A., Alzubi, O.mar,., and Manikandan, R. (2019). Efficient approaches for prediction of brain tumor using machine learning techniques. Indian J. Public Health Res. Dev. 10:267. doi: 10.5958/0976-5506.2019.00298.5
Amarbayasgalan, T., Pham, V.-H., Theera-Umpon, N., Piao, Y., and Ryu, K. H. (2021). An efficient prediction method for coronary heart disease risk based on two deep neural networks trained on well-ordered training datasets. IEEE Access 9, 135210–135223. doi: 10.1109/ACCESS.2021.3116974
Asadi-Lari, M., Tamburini, M., and Gray, D. (2004). Patients' needs, satisfaction, and health related quality of life: towards a comprehensive model. Health Qual. Life Outc. 2, 1–15. doi: 10.1186/1477-7525-2-1
Ata, F., Khan, A. A., Khamees, I., Iqbal, P., Yousaf, Z., Mohammed, B. Z. M., et al. (2023). Clinical and biochemical determinants of length of stay, readmission and recurrence in patients admitted with diabetic ketoacidosis. Ann. Med. 55, 533–542. doi: 10.1080/07853890.2023.2175031
Awad, A., Bader–El–Den, M., and McNicholas, J. (2017). Patient length of stay and mortality prediction: a survey. Health Serv. Manage. Res. 30, 105–120. doi: 10.1177/0951484817696212
Balen, F., Routoulp, S., Charpentier, S., Azema, O., Houze-Cerfon, C.-H., Dubucs, X., et al. (2024). Impact of emergency department length of stay on in-hospital mortality: a retrospective cohort study. Eur. J. Emer. Med. 31, 39–45. doi: 10.1097/MEJ.0000000000001079
Bank, D., Koenigstein, N., and Giryes, R. (2023). “Autoencoders,” in Machine Learning for Data Science Handbook, ed. L. Rokach, (Switzerland, AG: Springer Cham), 353–374. doi: 10.1007/978-3-031-24628-9_16
Barfungpa, S. P., Deva Sarma, H. K., and Samantaray, L. (2023). An intelligent heart disease prediction system using hybrid deep dense Aquila network. Biomed. Signal Proc. Control 84:104742. doi: 10.1016/j.bspc.2023.104742
Boulif, A., Ananou, B., Ouladsine, M., and Delliaux, S. (2023). A literature review: ecg-based models for arrhythmia diagnosis using artificial intelligence techniques. Bioinform. Biol. Insights 17:11779322221149600. doi: 10.1177/11779322221149600
Bozkurt, C., and Aşuroglu, T. (2023). Mortality prediction of various cancer patients via relevant feature analysis and machine learning. SN Comput. Sci. 4:264. doi: 10.1007/s42979-023-01720-5
Braik, M., Awadallah, M. A., Al-Betar, M. A., Hammouri, A. I., and Alzubi, O. A. (2023b). Cognitively enhanced versions of capuchin search algorithm for feature selection in medical diagnosis: a COVID-19 case study. Cognit. Comput. 15, 1884–1921. doi: 10.1007/s12559-023-10149-0
Braik, M., Sh. Hammouri, A. I., Awadallah, M. A., Al-Betar, M. A., and Alzubi, O. A. (2023a). Improved versions of snake optimizer for feature selection in medical diagnosis: a real case COVID-19. Soft. Comput. 27, 17833–17865. doi: 10.1007/s00500-023-09062-3
Cao, X., Başar, T., Diggavi, S., Eldar, Y. C., Letaief, K. B., Poor, H. V., et al. (2023). Communication-efficient distributed learning: an overview. IEEE J. Select. Areas Commun. 41, 851–873. doi: 10.1109/JSAC.2023.3242710
Chen, T., and Guestrin, C. (2016). “XGBoost,” in Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (New York, USA), 785–794. doi: 10.1145/2939672.2939785
Chollet, F. (2015). Keras. GitHub 12. Available online at: https://github.com/fchollet/keras (accessed April 5, 2024).
Ebiaredoh-Mienye, S. A., Esenogho, E., and Swart, T. G. (2020). Integrating enhanced sparse autoencoder-based artificial neural network technique and softmax regression for medical diagnosis. Electronics 9:1963. doi: 10.3390/electronics9111963
El Dor, A., Clerc, M., and Siarry, P. (2012). A multi-swarm PSO using charged particles in a partitioned search space for continuous optimization. Comput. Optim. Appl. 53, 271–295. doi: 10.1007/s10589-011-9449-4
El-Rashidy, N., Abuhmed, T., Alarabi, L., El-Bakry, H. M., Abdelrazek, S., Ali, F., et al. (2022). Sepsis prediction in intensive care unit based on genetic feature optimization and stacked deep ensemble learning. Neural Comput. Appl. 34, 3603–3632. doi: 10.1007/s00521-021-06631-1
Farahnakian, F., and Heikkonen, J. (2018). “A deep auto-encoder based approach for intrusion detection system,” in 2018 20th International conference on advanced communication technology (ICACT), Chuncheon, South Korea, 178–183, doi: 10.23919/ICACT.2018.8323687
Gentimis, T., Alnaser, A. J., Durante, A., Cook, K., and Steele, R. (2017). “Predicting hospital length of stay using neural networks on MIMIC III data., in 2017 IEEE 15th intl conf on dependable, autonomic and secure computing,” in 15th intl conf on pervasive intelligence and computing, 3rd intl conf on big data intelligence and computing and cyber science and technology congress (DASC/PiCom/DataCom/CyberSciTech), Orlando, USA, 1194–1201. doi: 10.1109/DASC-PICom-DataCom-CyberSciTec.2017.191
Goldberger, A. L., Amaral, L. A. N., Glass, L., Hausdorff, J. M., Ivanov, P. Ch., et al. (2000). PhysioBank, physiotoolkit, and physionet. Circulation 101, e215–e220. doi: 10.1161/01.CIR.101.23.e215
González-Nóvoa, J. A., Campanioni, S., Busto, L., Fariña, J., Rodríguez-Andina, J. J., Vila, D., et al. (2023). Improving intensive care unit early readmission prediction using optimized and explainable machine learning. Int. J. Environ. Res. Public Health 20:3455. doi: 10.3390/ijerph20043455
Harerimana, G., Kim, J. W., and Jang, B. (2021). A deep attention model to forecast the length of stay and the in-hospital mortality right on admission from ICD codes and demographic data. J. Biomed. Inform. 118:103778. doi: 10.1016/j.jbi.2021.103778
Hempel, L., Sadeghi, S., and Kirsten, T. (2023). Prediction of intensive care unit length of stay in the MIMIC-IV dataset. Appl. Sci. 13:6930. doi: 10.3390/app13126930
Heyland, D., Cook, D., Bagshaw, S. M., Garland, A., Stelfox, H. T., Mehta, S., et al. (2015). The very elderly admitted to ICU. Crit. Care Med. 43, 1352–1360. doi: 10.1097/CCM.0000000000001024
James, V., and Miranda, L. (2018). PySwarms: a research toolkit for particle swarm optimization in python. J. Open Sour. Softw. 3:433. doi: 10.21105/joss.00433
Jena, K. K., Bhoi, S. K., Prasad, M., and Puthal, D. (2022). A fuzzy rule-based efficient hospital bed management approach for coronavirus disease-19 infected patients. Neural Comput. Appl. 34, 11361–11382. doi: 10.1007/s00521-021-05719-y
Jiang, H., He, Z., Ye, G., and Zhang, H. (2020). Network intrusion detection based on PSO-XGBoost model. IEEE Access 8, 58392–58401. doi: 10.1109/ACCESS.2020.2982418
Johnson, A. E. W., Pollard, T. J., Shen, L., Lehman, L. H., Feng, M., Ghassemi, M., et al. (2016). MIMIC-III, a freely accessible critical care database. Sci. Data 3:160035. doi: 10.1038/sdata.2016.35
Jung, H., Park, H. W., Kim, Y., and Hwangbo, Y. (2023). “Machine learning-based prediction for 30-day unplanned readmission in all-types cancer patients,” in 2023 IEEE International Conference on Big Data and Smart Computing (BigComp), 132–135. doi: 10.1109/BigComp57234.2023.00029
Junior, J. C., Caneo, L. F., Turquetto, A. L. R., Amato, L. P., Arita, E. C. T. C., da Silva Fernandes, A. M., et al. (2024). Predictors of in-ICU length of stay among congenital heart defect patients using artificial intelligence model: a pilot study. Heliyon 10:e25406. doi: 10.1016/j.heliyon.2024.e25406
Kennedy, J., and Eberhart, R. (1995). “Particle swarm optimization,” in Proceedings of ICNN'95 - International Conference on Neural Networks, Perth, Australia, 1942–1948.
Kigo, S. N., Omondi, E. O., and Omolo, B. O. (2023). Assessing predictive performance of supervised machine learning algorithms for a diamond pricing model. Sci. Rep. 13:17315. doi: 10.1038/s41598-023-44326-w
Le, L. T., Nguyen, H., Zhou, J., Dou, J., and Moayedi, H. (2019). Estimating the heating load of buildings for smart city planning using a novel artificial intelligence technique PSO-XGBoost. Appl. Sci. 9:2714. doi: 10.3390/app9132714
LemaÃŽtre, G., Nogueira, F., and Aridas, C. K. (2017). Imbalanced-learn: a python toolbox to tackle the curse of imbalanced datasets in machine learning. J. Mach. Lear. Res. 18, 1–5. doi: 10.48550/arXiv.1609.06570
Li, W., Zhang, Y., Zhou, X., Quan, X., Chen, B., Hou, X., et al. (2024). Ensemble learning-assisted prediction of prolonged hospital length of stay after spine correction surgery: a multi-center cohort study. J. Orthop. Surg. Res. 19:112. doi: 10.1186/s13018-024-04576-4
Lundberg, S. M., and Lee, S.-I. (2017). “A unified approach to interpreting model predictions,” in Proceedings of the 31st International Conference on Neural Information Processing Systems (NIPS'17), California, USA, 4768–4777.
Mangalesh, S., Dudani, S., and Malik, A. (2023). The systemic immune-inflammation index in predicting sepsis mortality. Postgrad. Med. 135, 345–351. doi: 10.1080/00325481.2022.2140535
Masood, F., Masood, J., Zahir, H., Driss, K., Mehmood, N., and Farooq, H. (2022). Novel approach to evaluate classification algorithms and feature selection filter algorithms using medical data. J. Comput. Cogn. Eng. 2, 57–67. doi: 10.47852/bonviewJCCE2202238
Mienye, I. D., and Sun, Y. (2021). Improved heart disease prediction using particle swarm optimization based stacked sparse autoencoder. Electronics 10:2347. doi: 10.3390/electronics10192347
Miriyala, G. P., Sinha, A. K., Jayakody, D. N. K., and Sharma, A. (2021). “A review on recent machine learning algorithms used in CAD diagnosis,” in 2021 10th International Conference on Information and Automation for Sustainability (ICIAfS), 269–274. doi: 10.1109/ICIAfS52090.2021.9605854
Momo, L. N. W., Moorosi, N., Nsoesie, E. O., Rademakers, F., and De Moor, B. (2024). Length of stay prediction for hospital management using domain adaptation. Eng. Appl. Artif. Intell. 133:108088. doi: 10.1016/j.engappai.2024.108088
Muhlestein, W. E., Akagi, D. S., Davies, J. M., and Chambless, L. B. (2019). Predicting inpatient length of stay after brain tumor surgery: developing machine learning ensembles to improve predictive performance. Neurosurgery 85, 384–393. doi: 10.1093/neuros/nyy343
Nadimi-Shahraki, M. H., Zamani, H., and Mirjalili, S. (2022). Enhanced whale optimization algorithm for medical feature selection: a COVID-19 case study. Comput. Biol. Med. 148:105858. doi: 10.1016/j.compbiomed.2022.105858
Naemi, A., Schmidt, T., Mansourvar, M., Ebrahimi, A., and Wiil, U. K. (2021). Quantifying the impact of addressing data challenges in prediction of length of stay. BMC Med. Inform. Decis. Mak. 21:298. doi: 10.1186/s12911-021-01660-1
Nasarian, E., Abdar, M., Fahami, M. A., Alizadehsani, R., Hussain, S., Basiri, M. E., et al. (2020). Association between work-related features and coronary artery disease: a heterogeneous hybrid feature selection integrated with balancing approach. Patt. Recognit. Lett. 133, 33–40. doi: 10.1016/j.patrec.2020.02.010
Neyra, J. A., Ortiz-Soriano, V., Liu, L. J., Smith, T. D., Li, X., Xie, D., et al. (2023). Prediction of mortality and major adverse kidney events in critically ill patients with acute kidney injury. Am. J. Kidney Dis. 81, 36–47. doi: 10.1053/j.ajkd.2022.06.004
Nogueira, F. (2014). Bayesian Optimization: Open source constrained global optimization tool for Python. Available online at: https://github.com/bayesian-optimization/BayesianOptimization (accessed February 5, 2024).
Qian, J., Qin, R., Hong, L., Shi, Y., Yuan, H., Zhang, B., et al. (2023). Characteristics and clinical outcomes of patients with lung cancer requiring ICU admission: a retrospective analysis based on the MIMIC-III database. Emer. Cancer Care 2:1. doi: 10.1186/s44201-022-00017-2
Qin, C., Zhang, Y., Bao, F., Zhang, C., Liu, P., and Liu, P. (2021). XGBoost optimized by adaptive particle swarm optimization for credit scoring. Math. Probl. Eng. 2021, 1–18. doi: 10.1155/2021/6655510
Saadatmand, S., Salimifard, K., Mohammadi, R., Kuiper, A., Marzban, M., and Farhadi, A. (2023). Using machine learning in prediction of ICU admission, mortality, and length of stay in the early stage of admission of COVID-19 patients. Ann. Oper. Res. 328, 1043–1071. doi: 10.1007/s10479-022-04984-x
Shah, S. M. S., Shah, F. A., Hussain, S. A., and Batool, S. (2020). Support vector machines-based heart disease diagnosis using feature subset, wrapping selection and extraction methods. Comput. Electr. Eng. 84:106628. doi: 10.1016/j.compeleceng.2020.106628
Shinde, K., Itier, V., Mennesson, J., Vasiukov, D., and Shakoor, M. (2023). Dimensionality reduction through convolutional autoencoders for fracture patterns prediction. Appl. Math. Model. 114, 94–113. doi: 10.1016/j.apm.2022.09.034
Somgiat, W., and Chansamorn, S. (2022). “A new hybrid PSO-SCA using horse optimization algorithm's group behavior update,” in 2022 19th International joint conference on computer science and software engineering (JCSSE), Bangkok, Thailand, 1–6. doi: 10.1109/JCSSE54890.2022.9836292
Tavakolian, A., Rezaee, A., Hajati, F., and Uddin, S. (2023). Hospital readmission and length-of-stay prediction using an optimized hybrid deep model. Fut. Internet 15:304. doi: 10.3390/fi15090304
Tully, J. L., Zhong, W., Simpson, S., Curran, B. P., Macias, A. A., Waterman, R. S., et al. (2023). Machine learning prediction models to reduce length of stay at ambulatory surgery centers through case resequencing. J. Med. Syst. 47:71. doi: 10.1007/s10916-023-01966-9
Van Rossum, G., and Drake, F. L. Jr (1995). Python reference manual. Centrum voor Wiskunde en Informatica Amsterdam.
Wang, S., McDermott, M. B. A., Chauhan, G., Ghassemi, M., Hughes, M. C., and Naumann, T. (2020). “Mimic-extract: A data extraction, preprocessing, and representation pipeline for MIMIC-III,” in Proceedings of the ACM conference on health, inference, and learning, New York, USA. 222–235. doi: 10.1145/3368555.3384469
Workina, A., Habtamu, A., Diribsa, T., and Abebe, F. (2022). Knowledge of modifiable cardiovascular diseases risk factors and its primary prevention practices among diabetic patients at jimma university medical centre: a cross-sectional study. PLoS Global Public Health 2:e0000575. doi: 10.1371/journal.pgph.0000575
Yewale, D., Vijayaragavan, S. P., and Munot, M. (2023). “An optimized XGBoost based classification model for effective analysis of heart disease prediction,” in AIP Conference Proceedings (AIP Publishing). doi: 10.1063/5.0148268
Keywords: coronary artery disease, length of stay prediction, uncertainty, deep autoencoder, boosting algorithms, particle swarm optimization
Citation: Miriyala GP and Sinha AK (2024) PSO-XnB: a proposed model for predicting hospital stay of CAD patients. Front. Artif. Intell. 7:1381430. doi: 10.3389/frai.2024.1381430
Received: 13 February 2024; Accepted: 11 April 2024;
Published: 03 May 2024.
Edited by:
Kezhi Li, University College London, United KingdomReviewed by:
Omar A. Alzubi, Al-Balqa Applied University, JordanMiodrag Zivkovic, Singidunum University, Serbia
Copyright © 2024 Miriyala and Sinha. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Arun Kumar Sinha, YXJ1bmt1bWFyLnMmI3gwMDA0MDt2aXRhcC5hYy5pbg==