Next Article in Journal
New Biomarkers and Treatment Advances in Triple-Negative Breast Cancer
Next Article in Special Issue
Synthetic Attenuation Correction Maps for SPECT Imaging Using Deep Learning: A Study on Myocardial Perfusion Imaging
Previous Article in Journal
Generalist Vision Foundation Models for Medical Imaging: A Case Study of Segment Anything Model on Zero-Shot Medical Segmentation
Previous Article in Special Issue
Artificial Intelligence, Wearables and Remote Monitoring for Heart Failure: Current and Future Applications
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Heart Failure Emergency Readmission Prediction Using Stacking Machine Learning Model

by
Md. Sohanur Rahman
1,
Hasib Ryan Rahman
1,
Johayra Prithula
1,
Muhammad E. H. Chowdhury
2,*,
Mosabber Uddin Ahmed
1,*,
Jaya Kumar
3,
M. Murugappan
4 and
Muhammad Salman Khan
2
1
Department of Electrical and Electronics Engineering, University of Dhaka, Dhaka 1000, Bangladesh
2
Department of Electrical Engineering, Qatar University, Doha 2713, Qatar
3
Department of Physiology, Faculty of Medicine, University Kebangsaan Malaysia, Kuala Lumpur 56000, Malaysia
4
Intelligent Signal Processing (ISP) Research Lab, Department of Electronics and Communication Engineering, Kuwait College of Science and Technology, Block 4, Doha 13133, Kuwait
*
Authors to whom correspondence should be addressed.
Submission received: 5 April 2023 / Revised: 16 May 2023 / Accepted: 26 May 2023 / Published: 2 June 2023
(This article belongs to the Special Issue Artificial Intelligence in Cardiology Diagnosis )

Abstract

:
Heart failure is a devastating disease that has high mortality rates and a negative impact on quality of life. Heart failure patients often experience emergency readmission after an initial episode, often due to inadequate management. A timely diagnosis and treatment of underlying issues can significantly reduce the risk of emergency readmissions. The purpose of this project was to predict emergency readmissions of discharged heart failure patients using classical machine learning (ML) models based on Electronic Health Record (EHR) data. The dataset used for this study consisted of 166 clinical biomarkers from 2008 patient records. Three feature selection techniques were studied along with 13 classical ML models using five-fold cross-validation. A stacking ML model was trained using the predictions of the three best-performing models for final classification. The stacking ML model provided an accuracy, precision, recall, specificity, F1-score, and area under the curve (AUC) of 89.41%, 90.10%, 89.41%, 87.83%, 89.28%, and 0.881, respectively. This indicates the effectiveness of the proposed model in predicting emergency readmissions. The healthcare providers can intervene pro-actively to reduce emergency hospital readmission risk and improve patient outcomes and decrease healthcare costs using the proposed model.

1. Introduction

A significant proportion of hospitalizations and deaths among elderly individuals is associated with heart failure (HF). A physiological condition of HF is either not enough blood flow from the heart to meet the body’s metabolic demands or a compensatory neurohormonal response resulting in increased pressure in the left ventricle [1]. HF has been further divided into three subtypes based on ejection fraction, natriuretic peptide levels, structural heart disease, and diastolic dysfunction in recent years. The three subtypes of heart failure are heart failure with a reduced ejection fraction (HFrEF), heart failure with a preserved ejection fraction (HFpEF), and heart failure with a mid-range ejection fraction (HFmrEF) [2].

1.1. Background

In the United States, more than a million patients are hospitalized each year because of HF. Patients with HF often require hospital readmissions, which cost on average USD 23,000 per hospitalization. There is an estimated increase of USD 70 billion in costs for these admissions by 2030 [3,4,5,6]. The number of people suffering from heart failure in the US is currently over 6.2 million [7]. The prevalence of this disease is predicted to increase by 46% by 2030, reaching over 8 million [8]. Despite advances in medical treatment, hospital readmission rates remain high, with over half of patients readmitted within six months of discharge [9,10,11]. It is imperative to understand why some hospitals perform well while others struggle in order to deal with this issue. Previous studies have examined the first readmission, but not all the patients experience re-hospitalizations after the initial hospitalization for HF. Several interventions for patients with HF have been shown to reduce the risk of readmission by providing higher-quality care [12,13].
Predicting hospital readmission of heart failure patients is crucial for several reasons, including improving the quality of care, reducing healthcare costs, optimizing resource allocation, and enhancing patient outcomes [14]. High rates of readmission may indicate inadequate care during the initial hospitalization and discharge planning, leading to negative impacts on patient health and well-being. Predictive models can help healthcare providers identify patients at the highest risk of readmission and allocate resources effectively to improve outcomes. Additionally, predicting readmission risk can help healthcare organizations meet performance metrics and identify areas for improvement in care delivery. It has been observed that the elderly population with a previous HF diagnosis and a history of cardiac revascularization have a higher chance of unplanned readmission [15].

1.2. Related Works

A review by Ouwerker et al. of 117 models found that blood urea nitrogen and sodium were the strongest predictors for HF hospitalization [16]. Four subgroups of models were identified, with prospective registry-type studies using a large number of clinical predictor variables being the most accurate in predicting mortality. Blood sugar levels are also found to be crucial for HF patients [17]. The Center for Medicare and Medicaid Services (CMS) has emphasized reducing readmission rates to reduce costs and improve care quality [18]. When examining claims data from 2003–2004, it was discovered that a significant number of Medicare beneficiaries were readmitted to the hospital within 30 days (19.6%) and 90 days (34.0%) after being discharged. Both medical and surgical patients were affected, but medical patients had a higher readmission rate, accounting for 77.1% of the rehospitalizations. The top conditions associated with high 30-day readmission rates were heart failure (26.9%), psychoses (24.6%), recent vascular surgery (23.9%), chronic obstructive pulmonary disease (22.6%), and pneumonia (20.1%) [19]. Despite efforts to reduce readmissions, risk-adjusted 30-day readmission rates among Medicare beneficiaries have remained fairly stable over the past decade. According to a study conducted by Artetxe et al., it has been found that machine learning methods can enhance the accuracy of predictions as compared to conventional statistical techniques [20]. The researchers noted that logistic regression (LR) and survivability analysis are the frequently utilized techniques for developing models to anticipate readmission rates. A machine learning approach to HF treatment, including diagnosis, classification, readmissions, and medication, has the potential to improve treatment quality and cut costs [21].
Data extracted from routinely collected electronic healthcare records (EHRs) have been used in a few studies. Based on clinical and physiological data available in EHR systems, Beecy et al. developed a comprehensive approach to predicting 30-day unplanned readmissions and all-cause mortality (ACM) [22]. They designed three different predictive models, using an extreme gradient boosting (XGBoost) classifier, which were based on index admission, index discharge, and feature-aggregation. Due to the lack of previous data on heart failure rates in China, Zhang Z. et al. studied the symptoms of HF patients admitted to Zigong Fourth People’s Hospital between December 2016 and June 2019. Based on their work, baseline clinical characteristics (including body temperature, pulse, respiration rate, body mass index (BMI), etc.) and primary drug categories were recorded on admission. The majority of other studies focus only on 30-day emergency readmission rates, which may not represent the majority of patients. This study used follow-up periods of 28 days, 3 months, and 6 months for better statistical analysis and interpretation [23]. Data mining techniques and significant features have been found to boost the accuracy of cardiovascular survivor prediction by Ishaq et al. By using synthetic minority oversampling technique (SMOTE) to offset the imbalance in the dataset, they achieved a remarkable accuracy of 0.926 using the ExtraTree (ET) classifier algorithm [24]. A deep-learning-based approach was performed by Golas et al., as they developed a risk prediction model for hospital readmission including structured and un-structured data, constructed using deep unified networks (DUNs) to avoid over-fitting [25]. The DUNs model had the best performance after 10-fold cross-validation, reaching an AUC of 0.705  ±  0.015 and 76.4% accuracy. Another deep-learning-based study by Liu et al. proposed a CNN model with an F1 score of 0.756 and 0.733 in general readmission prediction and 30-day readmission prediction, respectively [26].

1.3. Research Objectives and Contribution

Our aim was to determine whether patients with heart failure who had previously undergone treatment and been discharged would be likely to return to the hospital in the event of an emergency. In order to accomplish this goal, an EHR dataset of HF patients was used. The dataset included comprehensive information about the patient’s physical and medical conditions, including symptoms and signs of HF, and their discharge status. We used 13 variants of classical machine learning (ML) models and a stacking machine learning (or meta-learner)-based model. In this study, the following novel contributions are made:
  • This study examines a large dataset to predict emergency readmissions for heart failure patients.
  • Thirteen variants of classical machine learning models were examined using different feature selection techniques along with a stacking machine learning model.
  • Based on nine biomarkers, this study produced very good predictions for emergency department readmission for heart failure patients.
  • LACE-related feature incorporation to the feature set provided a significant performance boost in the study to predict hospital readmission due to heart failure emergency.
The rest of the paper is organized as follows: Section 2 discusses the methodology of the study by describing the datasets used in this paper and the details of data pre-processing stages for machine learning classifiers and the stacking machine learning algorithm. Section 3 discusses the result of the classification models and the stacking machine learning model. Finally, the article is concluded in Section 4.

2. Methodology

This study attempts to predict the emergency readmission of HF patients after discharge from a medical facility using a dataset that is made publicly available in Physionet [27]. The dataset has three different forms of data: the electronic health record, medication, and dictionary. An extensive analysis of the data led to the selection of the most important features for the model. Data from electronic health records and drug databases were combined, and feature engineering was performed. An important parameter, a LACE score, was added to the feature list for identifying readmissions cases reliably. Thirteen variants of classical ML models were trained with a reduced feature set selected by a feature-ranking technique. A stacking ML model was trained using the top three classical ML models, which improved the overall performance. A solid and reliable model for forecasting HF patients’ emergency readmissions was developed as a result of the approach. The flowchart in Figure 1 provides a general picture of how the study was handled.

2.1. Dataset Description

We collected this EHR dataset from a public source available from Physionet [27]. The EHR dataset was collected by Zhang Z. et al. to study the symptoms of HF patients admitted to Zigong Fourth People’s Hospital between December 2016 and June 2019. The study was approved by the ethics committee of Zigong Fourth People’s Hospital (Approval Number: 2020-010). As there was a scarcity of datasets available on HF, the researchers targeted Zigong Fourth People’s Hospital in Sichuan, China. They extracted data from a routinely collected EHR. In addition, follow-up data were collected on hospital admissions, mortality rates, and required follow-up visits at 28 days, 3 months, and 6 months after discharge. In the event that a patient had a medical emergency and was unable to travel to the clinical facility, their information was collected by phone. Participants’ data were collected between December 2016 and June 2019, providing researchers with an in-depth look at their health conditions and outcomes over time. Researchers were able to create more effective medications and interventions after understanding HF and the variables that affect patient outcomes.
The clinical data were collected on the day of the subject patients’ hospital admission. Clinical data included information from a variety of medical tests, including body temperature, pulse, respiration rate, systolic and diastolic blood pressure, mean arterial blood pressure, weight, height, body mass index (BMI), type of HF, New York Heart Association (NYHA) cardiac function, and Killip Grade (Class 1: no rales, no third heart sound; Class 2: rales in 0.5 lung fields or the presence of a third heart sound). This sub-dataset contained 2008 records and 166 features excluding a redundant index column. Information related to admission, discharge, and medical test reports was included in these features.
A separate dataset included information on the medications consumed by the subject individuals while in the hospital. There was no mention of the time that the drugs were taken; only the medications were mentioned. In the dataset, there were multiple instances of the same medicine because values were inputted every time a medication was taken, indicating more than one dose was administered. This sub-dataset contains 15,362 records and 2 feature columns, including inpatient number and medication name. We were able to combine the two datasets into one by combining the inpatient number in each of the sub-datasets.

2.2. Statistical Analysis

The variables of this study were statistically analyzed using Python 3.8. Ages were classified according to decades. There were eight age categories from 19 to 110. Three categories were available for the readmission period: 28 days, three months, and six months. By applying Pearson correlation coefficient (PCC) for each feature with other features, we were able to exclude highly correlated features with a score greater than 0.80 that were highly similar to the label data and could lead to overfitting our model. PCC formula is given as below [28]:
r = i = 1 n x i x ¯ y i y ¯ i = 1 n x i x ¯ 2 i = 1 n y i y ¯ 2
where r, x i , x ¯ , y i , y ¯ indicate the Pearson correlation coefficient, feature value, mean feature value, other feature value, other mean feature value. Here, features x and y have a length of n.
Table 1 shows the characteristics of the patients’ return to the Emergency Department within six months, the NYHA cardiac function classification, the use of Digoxin tablets, the presence of a magnesium serum, and the presence of potassium ions. The 1951 patients had a return home rate of 67.5% at 1318, a hospitalization rate of 22.5% at 436, an unknown rate of 9.8% at 192, and a mortality rate of 0.2% at 5. Due to the small number of patients who passed away in hospitals, they were excluded from the dataset to avoid low-quality features. The number of male patients was 845 (42.09%) and the number of female patients was 1163 (57.91%). Approximately 7.657 medications were consumed by each patient, on average, with a standard deviation of 2.4204. Among 2008 patients, 92.8% (1865) had myocardial infarction and 93.2% (1872) had congestive HF. The most common disease was connective tissue disease (99.7%).

2.3. Preprocessing

A physical examination was conducted on each patient to verify the accuracy and authenticity of the dataset. This included height, weight, body mass index (BMI), type of heart failure, and heart condition. However, certain anomalies were discovered when the data were examined. The height, weight, and BMI measurements of some patients were atypical, and some of these measurements were recorded as 0. The anomalous values were not consistent with the known range of human measurements. As an example, the shortest individual ever listed in the Guinness Book of Records, who was about 0.54 m tall, did not fit with certain height numbers that were especially unusual, such as 0.35 m [29]. The ages of the patients were reviewed to confirm the accuracy of these data. Analysis revealed that 0.697% of discharged patients died in the hospital after they were discharged. These records were excluded from the dataset.
Height and weight outliers were missing data that were filled in with 0 and NaNs (Not a Number) were substituted with 0. Additionally, the ages of the patients were listed in decade form instead of as exact numbers. As a result, the age range was converted from an object type to an integer by replacing it with the average of the range. A total of 15 columns with NaN values exceeding 65% were removed from the dataset in order to avoid overfitting and skewed results. We removed columns for blood temperature, blood gas, and leukemia because they would not be relevant to the model’s training and would interfere with standardizing and normalizing the dataset.
The objective of this project was to develop a model for discharged patients to be readmitted in the event of an emergency. Therefore, we needed to select the right predicting label file. A total of two columns matched the target. In one column, patients were listed who got readmitted after days of prior admission; in the other, the number of days after the initial admission that the patients had to be in the emergency department within 6 months was listed. The correlation between these two columns was calculated using Equation (1), and we found that it was 0.99. Therefore, any patient who needed to be admitted again would have to go through the emergency department. In this regard, the ‘re.admission.time.days.from.admission.’ column is set as the label file. All values greater than 0 in that column were transformed into 1, indicating the patient had been readmitted, and NaN values were transformed into 0, meaning the patient had not been readmitted. We dropped the ‘time.to.emergency.department.within.6.months’ variable to prevent overfitting.

2.3.1. Encoding

Despite having a plethora of categorical columns, the EHR dataset would be difficult to use in its current form without encoding. An encoding process was carried out on the entire dataset in order to prepare it in a way that can be easily fed into the classification models in order to have greater accuracy. Each distinct categorical item was represented by an ordinal number. Drug datasets used an additional encoding method, where a new vector was integrated for each distinct categorical value. It is now possible to concatenate the EHR datasets. The medication dataset’s form was altered due to encoding, which also created 25 new columns with the parent ID column.

2.3.2. Dataset Mapping

After encoding the dataset in two different encoding methods, it was necessary to merge the datasets into a single dataset, which was achieved through dataset mapping. Each EHR dataset and medication dataset had the same Patient ID to identify each patient individually. There were multiple entries under the same Patient ID in the medication dataset. Each patient taking the same medicine multiple times was summed under a single patient in a column, so the recurrence of Patient ID was eliminated. The data had been merged between two datasets based on Patient ID. Therefore, all drug name columns were concatenated with the EHR dataset, and the shape of the dataset was the same as that of the EHR dataset.

2.3.3. Missing Data Imputation

Due to the practical difficulty of collecting all information from a subject, missing data were likely in the dataset. The missing values were represented by NaN. Due to the fact that machine learning models are mathematical in nature, they are not capable of processing data of any other kind. It is therefore imperative to adopt essential safeguards. The removal of NaN values is one way to deal with the missing values, but this approach would result in substantial data loss and rather poor ML performance. Alternatively, NaN values can be replaced with appropriate values based on the records’ pattern. By using the entire feature data, the data imputation approach fills in the missing values [30]. The inclusion of missing indicator variables is preferred in some studies [31,32], while it is also controversial [33] in medical research. Several other studies totally excluded missing data in order to prevent high bias [34,35,36].
In our study, we carefully examined the EHR dataset and found that many key predictors had missing parameters. In order to prevent data loss, an important data imputation method was used. In this study, MICE, or Multivariate Imputation by Chained Equations, was employed as the imputation technique [37,38]. As part of this method, both mean and regression methods are applied successively within a feature column to make the missing records comparable to other records that have similar values. Using this imputation method, the analyzed data were clinical data, which made the imputation method successful.

2.3.4. LACE Feature Extraction

The LACE index, also known as Length of stay, Acuity of admission, Comorbidities, and Emergency department use, is widely used in the medical field in the US and Canada [39,40,41]. This mathematical model was developed by van Walraven et al. [42] to predict the risk of unplanned readmission or death within 30 days after discharge from the hospital for medical and surgical patients. Four main factors can help healthcare providers make informed decisions about patient care [43]. They can identify patients who may require extra support after discharge, prevent unplanned readmissions, and improve patient outcomes.
After connecting the one-hot encoded data with the EHR information, we identified the medications taken by the subject patients. Because HF groups are usually associated with a high number of prescription drugs, we introduced the total number of medications patients take to represent their risk factors. More medications taken by a patient increases the risk of their health deteriorating [44]. A pharmaceutical column was removed after the previous feature had been extracted. Using the LACE score, we were able to identify patients with high hospitalization risks. As shown in Table 2, the LACE score [42] was designed according to its fully developed phrase using the relevant elements.
Six hundred and seventy-five patients, or nearly 33.87% of all records, had LACE scores indicating that they were at high risk. In total, 63.67% of the study’s participants were in the moderate-risk zone, while 2.46% were in the low-risk zone. LACE alone can be a good predictor of hospital readmission for HF patients and in this study, we have incorporated this LACE feature with other features, which made our model robust.

2.3.5. Feature Scaling

Occasionally, varying ranges and different scaling of data cause the model to run slowly and increase the cost function. Furthermore, all forms of variations and scaling are unlikely to fit the label data effectively [45]. It is also possible that some observations may yield false predictions due to differences in scaling. An ML model should yield a favorable outcome if the quality of the data is guaranteed. The training feature data need to be scaled properly in order to achieve quality data in comparison to the target data. Our study assessed the effectiveness of two normalization methods, min-max normalization, and Z-score normalization to normalize the data [46].
As a result of the min-max normalization method, each feature is scaled with its minimum and maximum value, allowing the data to remain in the 0–1 range. The following formula is used to scale the data (Equation (2)). If X is considered to be a feature value of a certain feature, X s c a l e d , X m a x , and X m i n are scaled output, maximum feature value, and the minimum feature value of that certain feature.
X s c a l e d = X X m i n X m a x X m i n
Using Z-score normalization, the dataset is transformed into a gaussian distribution. Therefore, it is possible to comprehend the probability of an event occurring within the boundary of a gaussian distribution. This phrase is used in this way (Equation (3)).
Z = X μ σ
Here, Z, X, μ , and σ are, respectively, the output, feature value, mean, and standard deviation of the variable values.
This approach dramatically improved computation speed by scaling feature values to have a mean of 0 and a standard deviation of 1. Despite this, this strategy may produce outlier values since some observations showed negative instances. In our particular case, min-max scaling performed better than Z-score scaling method while training the model because our dataset contained some outlier values.

2.3.6. Feature Selection

Three different feature selection models were used in this work: XGBoost, Random Forest, and ExtraTrees. Additionally, the feature importance values for each of these selection models have been computed. XGBoost method provides feature importance based on average gain or reduction in the objective function, which provides insights of how a specific feature can contribute in reducing the loss of the model [47]. For Random Forest and ExtraTrees, feature importance is extracted based on mean decrease accuracy, which reduces the impurity in the decision tree when a particular feature is split [48]. In order to achieve the highest accuracy, these feature models were listed in descending order and trained using the logistic regression method’s single feature incrementation, which draws out an optimum number of features needed for the model to score a best result. We can formulate logistic regression as follows:
P = e a + b X 1 + e a + b X
where P is the logistic regression output, a denotes the expected value of log odds, b stands for change in log odds, and X is the input feature.
In order to draw out the number of features necessary to generate the highest accuracy from the model, the model was trained on the features in increments of 1 feature at time. This method provides valuable insights of how the model performance changes based on the features it is trained on. The outcome of such a method can be observed in Figure 2, which dictates training on 9 features that would present us the best possible performance. Supplementary Table S1 shows the list of features obtained in Figure 2.
  • XGBoost
We conducted a univariate logistic regression analysis to identify the independent variables associated with death. We identified the top 1, top 2, and top 9 features. Gradient-based ensemble tree methods such as XGBoost and GB (Gradient Boosting) are both gradient-descent-based methods that utilize weak learners [47,49,50]. A predictive model was developed by Xu et al. [51] based on XGBoost for ischemic stroke patients readmitted within 90 days of hospital discharge. As a result, they were able to achieve an AUC score of 0.782. XGBoost classifier is trained with a 0.001 learning rate and Logloss as the loss function.
  • Random Forest (RF)
RF is a powerful machine learning algorithm that excels at handling non-linear datasets and outperforms other tree-based algorithms. It utilizes multiple trees with randomized inputs and splits and effectively addresses the problem of overfitting found in decision trees. The RF algorithm solves this issue by averaging the output of each tree and comparing it with the original output. According to Yeshvendra et al. [52], they developed an RF model that achieved 85.81% accuracy in predicting heart disease. This classifier is trained with the following hyperparameters: (bootstrap = True, class_weight = None, criterion = ‘gini’, max_depth = 10, max_features = ‘auto’, max_leaf_nodes = None, min_impurity_decrease = 0.0, min_samples_leaf = 1, min_samples_split = 2, min_weight_fraction_leaf = 0.0, n_estimators = 100, oob_score = False).
  • ExtraTrees Classifier (ET)
An ExtraTrees classifier uses the traditional top-down method to create an unpruned decision tree [53]. It involves strong randomization both in attribute selection and in cut-point selection when splitting a node. It differs from other tree-based ensemble methods in that it splits nodes fully randomly and grows trees using the whole training sample instead of bootstrap replicas. The final prediction is made by a majority vote among all three predictions. ExtraTrees reduce variance and bias and are computationally efficient [54]. According to Shafique et al. [55], ExtraTrees were found to be the most accurate models for predicting cardiovascular disease, with an accuracy rate of 96%. ET was trained with the following hyperparameters: (max_depth = 8, min_samples_split =10).

2.3.7. Stacking Machine Learning Model

We have investigated 13 variants of classical ML models for initial prediction performance for hospital readmission of HF patients. Then, we proposed using a stacking ML model that uses 3 classical best-performing ML models. Multiple learners are integrated into stacking-based models, which improve prediction capabilities. This was motivated by a three-layer stacking model demonstrated by Zhang et al. to predict hospital readmission risk [56], where they achieved an AUC of 0.720 outperforming other individual models.
We converted the data using fivefold cross-validation and divided it into training, testing, and validation sets with 80% train and 20% test splits, where 20% of training data used for validation. The training and stacking procedure is graphically shown in Figure 3. The training set had a shape of p × q, where p represents the number of observations and q represents the top features selected using the top feature approach. A stacking classification model was created based on the predictions of the three best-performing models, allowing for greater precision in categories where the model had previously been unclear. To construct the stacking classification model, we used the best classifier predictions as input to the new meta classifier. A significant improvement in accuracy and performance was achieved with this approach.

2.4. Evaluation Metrics

There were a number of performance evaluation metrics considered for this work, as accuracy alone is not good metric [57,58,59]. As the scores were acknowledged to be the weighted average of all classes, accuracy as the sole metric for evaluation could have skewed the model towards the best-performing class [60]. Furthermore, there are other metrics necessary to explain the AUROC curve in addition to accuracy.
The test set was used to measure the performance metrics of the ML model with the largest area under the receiver operating characteristic curve (AUROC). There were several measures included to compute the performance metrics, including positive likelihood ratios, true positives (TP) and false positives (FP), true negatives (TN) and false negatives (FN), as well as positive predictive value (PPV) [61]. According to AUROC curves, c-statistics are a measure of model discrimination performance for binary classification studies. By comparing patients who actually had the outcome with those who did not, we are able to determine whether a model predicts a higher probability of the outcome. These metrics can be represented mathematically as follows:
A c c u r a c y = T P + T N T P + T N + F P + F N
P r e c i s i o n = T P T P + F P
R e c a l l   S e n s i t i v i t y = T P T P + F N
F 1   s c o r e =   2     Precision   Recall Precision + Recall
S p e c i f i c i t y   T r u e   N e g a t i v e   R a t e = T N T N + F P

3. Results and Discussions

3.1. Statistical Analysis

The analysis of this dataset provided significant insights into the characteristics and data that should be focused on when admitting patients at risk. An emergency room admission decision for an HF patient should be made with greater care based on prior diagnostic factors and patient information. It is important to consider the patient’s previous hospitalization data and parameters when prompting them to contact a hospital to prevent exacerbation of the illness. Table 3 summarizes the important feature statistics for the dataset.
In the dataset, there were 900 (45.16%) cases that needed emergency readmission, and 1093 (54.84%) cases were not readmitted. Average discharge day for the subject for readmitted and not readmitted were 10 ± 9.01 (days) and 9 ± 7.06 (days). The cholesterol level for not readmitted cases was 3.77 ± 1.07 (mmol/L) and for the readmitted case 3.66 ± 1.11 (mmol/L). Not readmitted patients had a mean weight of 52 ± 10.7 (kg) and for the other class, the mean was counted as 52.5 ± 11.1 (kg). Additionally, this study indicates the mean reduced Hemoglobin of 4.3 ± 6.22 (%) for the not readmitted class and 4.1 ± 6.0 (%) for the readmitted cases. For both non-readmitted and readmitted cases, both Systolic and Diastolic Blood pressure had a statistically observed level: 133 ± 25.0 (mmHg) and 129 ± 24.2 (mmHg), 77 ± 14.6 (mmHg), and 76 ± 14.2 (mmHg), respectively. Readmitted patients were having 2 ± 3.64 (mg/L) D-dimer in their blood and not readmitted patients were showing 3 ± 6.34 (mg/L), which clearly acted as an indicator by drawing a substantial difference in statistical metrics for both cases. Mitral Valve EMS also proved to be a proper indicator for both the classes, as it showed 6 ± 42.9 (m/s) for not readmitted patients and 3 ± 29.9 for the readmitted patients, although it showed a considerably low correlation value. Readmitted patients were observed to have comparably smaller statistical value range regarding High-Sensitivity Protein having 24 ± 33.5 (mg/L), whereas the patients who were not accounted for emergency readmission showed a value of 26 ± 35.5 (mg/L). We also observed a Glomerular Filtration Rate of 71 ± 36.5 (mL/min/1.73 m2) for the patients who did not necessarily need an emergency readmission, where readmitted patients had a filtration rate of 66 ± 36.5 (mL/min/1.73 m2). p-value < 0.05 was used to check the significance of the features for the classes.

3.2. Feature Ranking

According to the XGBoost ranking algorithm, the top features to identify patients who will require readmission were admission.way, LACE.score, reduced.hemoglobin, type.of.heart.failure, occupation, mitral valve EMS, CCI.score, and white.globulin.ratio. After initial investigation, it was observed that the features identified by XGBoost model outperformed other feature selection-based features. The feature rankings by XGBoost model is shown in ascending order in Figure 4. Therefore, these features were used for the classification investigations. Feature ranking using Random Forest and ExtraTree algorithms are reported in Supplementary Figures S1 and S2.

3.3. Classification Performance

Individual Model Performance: We implemented 13 variants of classical machine learning models including multilayer perceptron (MLP), Linear Discriminant Analysis (LDA), XGBoost, Random Forest, logistic regression (LR), Support Vector Machine (SVM), ExtraTrees (ET), AdaBoost, K-nearest neighbor (KNN), CatBoost, Gradient Boosting, Light GB (LGB), and ElasticNet. Table 4 summarizes the accuracy, precision, recall, specificity, and F1-score for all the models. Catboost, Adaboost, and Gradient Boosting algorithms are the three top-performing algorithms, which were used to develop the stacking ML model.
Stacking Model Performance: The top three performing models were used to create a meta-classifier, and the stacking model was trained based on those results. This last stage of classification provided us with the best results. The stacking ML model provides an overall accuracy, and weighted average of precision, recall, specificity, and F1-score of 89.41%, 90.10%, 89.41%, 87.83%, and 89.28%, respectively. Table 5 summarizes the performance of the stacking ML model, where it is seen that the XGB model as a meta-learner performs the best.
To assess the quality of our model’s predictive performance, we showed the confusion matrix [62], which summarizes the overall performance of the model based on the performance evaluation metrics introduced earlier. Figure 5 shows the confusion matrix, which shows the correct and incorrect predictions. It can be seen that the model can predict very well the patients who do not need readmission. However, the model suffers for patients who need readmission compared to the other class. The reason of slightly less accurate prediction for admission class might be due to the LACE feature. The LACE scoring method requires the number of emergency department visits in the last 6 months, which is not provided in our dataset. Therefore, we utilized the scoring table to use only the first two scores and leave the rest. There was also a small information gap between the number of days between readmissions and the time to the emergency department in the last six months.
The ROC curve represents the degree of separability between binary classes. The larger the area covered by the ROC curve, the more effectively the model utilizes the training dataset. Figure 6 shows the ROC curves for the three best-performing models and the stacking ML model. It can be observed that the AUC scores are very close. XGBoost achieved an excellent AUC score of 0.881, which shows a very good discrimination capability of the model.

3.4. Discussion and Comparison with Similar Works

As per the XGBoost ranking algorithm, the foremost attributes for identifying patients who may require readmission comprise admission.way, LACE.score, reduced.hemoglobin, type.of.heart.failure, occupation, mitral valve EMS, CCI.score, and white.globulin.ratio. Upon preliminary examination, it was noted that the features identified by the XGBoost model exhibited greater efficacy than other feature selection-based attributes. Thirteen machine learning models were tested to predict heart failure readmission, and Catboost, Adaboost, and Gradient Boosting were found to be the top-performing models. These three models were used to create a stacking model, which achieved an overall accuracy of 89.41% and weighted average precision, recall, specificity, and F1-score of 90.10%, 89.41%, 87.83%, and 89.28%, respectively. The model performed well in predicting patients who did not require readmission but moderately performed with those who did. XGBoost had the best AUC score of 0.881, indicating excellent discrimination capabilities.
Assessing different models using the same criteria presents a considerable challenge due to the variance in methodologies and variables used to build them. Moreover, the majority of machine-learning-based heart failure readmission prediction models have yet to be verified through an independent prospective cohort of heart failure patients. For our stacking model, it can be seen that the XGB model as a meta-learner performs the best.
The study conducted an in-depth analysis of the experimental results and found that the proposed stacking ensemble model can effectively make use of the predictive abilities of various models to understand the data space and structure. By leveraging this understanding, different models are able to learn from one another, leading to a mutual enhancement in their accuracy of predictions. This suggests that the stacking ensemble model can be beneficial for improving the performance of prediction models, especially when dealing with complex data structures. This claim is further proved by a similar study by Chiu et al. that predicts a model by stacking six classifiers, namely: RF, SVC, KNN, LGBM, Bagging, and Adaboost [63]. According to the stacking model results provided by them, it has achieved an accuracy level of 95.25% alongside an AUROC score of 82.55%. This is consistent with the results of this study. Jing suggested an ML model aimed at precise predictions of one-year, all-cause mortality for a significant number of HF patients. The best prediction performance was achieved by the nonlinear XGBoost model with an AUC score of 0.77 [64]. Chen et al. achieved an accuracy of 0.652 and AUC of 0.634 (95%CI: 0.599–0.646) using LR [65]. Overall, our study highlights the importance of considering the interplay between multiple prediction models in the development of effective machine learning algorithms. It also provides a novel study on predicting emergency readmission of HF patients using only nine biomarkers.

3.5. Limitations of Our Work

In terms of dataset, the limitation is that it is obtained from only one center; therefore, the generalizability of the model cannot be guaranteed. The data in the dataset have been merged into a solitary row per hospitalization. Therefore, only superficial portrayals of patient stays are accessible, and time series data for the entire hospitalization are not present. Thus, critical information like medication distribution timings is not provided in the dataset [27]. The study relies on retrospective data where the researchers typically have less control of the dataset variables compared to prospectively collected data.
One of the main limitations of this study is that the features selected by our pipeline did not include any of the important biomarkers of HF: for example, B-type natriuretic peptide (BNP), N-terminal pro-B-type natriuretic peptide (NT-proBNP), Troponin, etc. This requires further investigation and opens up a new area for our future research goal. The reliability of our model is limited to the nine features it currently incorporates, and the absence of any of these features may compromise the accuracy of its predictions. Furthermore, the dataset under scrutiny is geographically confined to a single region, which raises concerns regarding the applicability of our findings to other regions around the world. Given the modest size of our dataset, which comprises approximately two thousand patient records, it is not very suitable to train deep learning models. Consequently, we intend to augment our dataset by aggregating additional datasets pertaining to heart failure, thereby enabling us to develop a comprehensive, one-for-all machine learning pipeline.

4. Conclusions

This study aimed to use classical machine learning models based on EHR data to predict emergency readmissions for discharged heart failure patients. Early identification of heart failure patients is crucial to prevent complications and mortality. The proposed method uses the LACE score as a key component to predict subsequent readmissions, enabling healthcare providers to proactively intervene and decrease emergency hospital readmissions while simultaneously improving patient outcomes and reducing healthcare costs. The proposed design accurately predicts emergency readmissions for discharged HF patients with an AUC score of 0.883 using stacking ML model. While the dataset used in this study is limited to one center, our proposed model could still help healthcare providers improve the standard of care for HF patients by fine-tuning the model on a specific population data. Further research is needed to investigate important biomarkers of heart failure not included in the selected features, and augmenting the dataset could enable the development of a more comprehensive framework. The study can be expanded by accumulating multiple datasets from different regions to increase the generalizability of the model. This will also help construct a one-for-all model regardless of race and geographical position. The research demonstrates that using a stacking model can enhance the accuracy of prediction models, particularly when dealing with intricate data structures. Overall, this study is expected to contribute to filling a knowledge gap in multi-feature-based ML models and advance the field of emergency readmission prediction for HF patients, thereby enhancing the quality of heart failure care.

Supplementary Materials

The following supporting information can be downloaded at: https://0-www-mdpi-com.brum.beds.ac.uk/article/10.3390/diagnostics13111948/s1, Figure S1: Feature ranking using Random Forest; Figure S2: Feature ranking using Random Forest; Table S1: Ranked 15 features using XGB model.

Author Contributions

Conceptualization M.S.R., H.R.R., M.E.H.C., J.K. and M.U.A.; data curation M.S.R., H.R.R. and J.P.; formal analysis M.S.R., H.R.R. and J.P., funding acquisition M.E.H.C., M.S.K. and M.M.; investigation M.S.R., H.R.R. and J.P.; methodology M.E.H.C., J.K., M.U.A., M.M. and M.S.K.; project administration M.E.H.C., J.K., M.U.A., M.M. and M.S.K.; resources M.E.H.C., M.U.A.; Supervision M.E.H.C., J.K., M.U.A., M.M. and M.S.K.; validation M.E.H.C., J.K. and M.U.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by and Qatar University student grant QUST-1-CENG-2023-933. The statements made herein are solely the responsibility of the authors. The open-access publication cost is covered by Qatar National Library.

Institutional Review Board Statement

The authors of this paper have used the publicly available dataset. The study was conducted in accordance with the Declaration of Helsinki, and approved by the Ethics Committee of Zigong Fourth People’s Hospital the Ethics Committee of Zigong Fourth People’s Hospital (Approval Number: 2020-010).

Informed Consent Statement

Informed consent was waived due to the retrospective study of the dataset. It should be noted that the dataset was collected from a publicly available dataset on Physionet.

Data Availability Statement

This work used publicly available datasets collected from Physionet. The dataset can be found here https://0-doi-org.brum.beds.ac.uk/10.13026/5m60-vs44 [27].

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Tan, L.-B.; Williams, S.G.; Tan, D.; Cohen-Solal, A. So many definitions of heart failure: Are they all universally valid? A critical appraisal. Expert Rev. Cardiovasc. Ther. 2010, 8, 217–228. [Google Scholar] [CrossRef]
  2. Ponikowski, P.; Voors, A.A.; Anker, S.D.; Bueno, H.; Cleland, J.G.F.; Coats, A.J.S.; Falk, V.; González-Juanatey, J.R.; Harjola, V.-P.; Jankowska, E.A.; et al. 2016 ESC Guidelines for the diagnosis and treatment of acute and chronic heart failure: The Task Force for the diagnosis and treatment of acute and chronic heart failure of the European Society of Cardiology (ESC) Developed with the special contribution of the Heart Failure Association (HFA) of the ESC. Eur. Heart J. 2016, 37, 2129–2200. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Chamberlain, A.M.; Dunlay, S.M.; Gerber, Y.; Manemann, S.M.; Jiang, R.; Weston, S.A.; Roger, V.L. Burden and Timing of Hospitalizations in Heart Failure: A Community Study. Mayo Clin. Proc. 2017, 92, 184–192. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. O’Connor, C.M.; Miller, A.B.; Blair, J.E.; Konstam, M.A.; Wedge, P.; Bahit, M.C.; Carson, P.; Haass, M.; Hauptman, P.J.; Metra, M.; et al. Causes of death and rehospitalization in patients hospitalized with worsening heart failure and reduced left ventricular ejection fraction: Results from efficacy of vasopressin antagonism in heart failure outcome study with tolvaptan (EVEREST) program. Am. Hear. J. 2010, 159, 841–849.e1. [Google Scholar] [CrossRef]
  5. Dharmarajan, K.; Hsieh, A.F.; Lin, Z.; Bueno, H.; Ross, J.; Horwitz, L.; Barreto-Filho, J.A.; Kim, N.; Bernheim, S.M.; Suter, L.G.; et al. Diagnoses and Timing of 30-Day Readmissions After Hospitalization for Heart Failure, Acute Myocardial Infarction, or Pneumonia. JAMA 2013, 309, 355–363. [Google Scholar] [CrossRef] [PubMed]
  6. Arora, S.; Lahewala, S.; Virk, H.U.H.; Setareh-Shenas, S.; Patel, P.; Kumar, V.; Tripathi, B.; Shah, H.; Patel, V.; Gidwani, U.; et al. Etiologies, Trends, and Predictors of 30-Day Readmissions in Patients With Diastolic Heart Failure. Am. J. Cardiol. 2017, 120, 616–624. [Google Scholar] [CrossRef]
  7. Virani, S.S.; Alonso, A.; Benjamin, E.J.; Bittencourt, M.S.; Callaway, C.W.; Carson, A.P.; Chamberlain, A.M.; Chang, A.R.; Cheng, S.; Delling, F.N.; et al. Heart Disease and Stroke Statistics—2020 Update: A Report from the American Heart Association. Circulation 2020, 141, e139–e596. [Google Scholar] [CrossRef]
  8. Mozaffarian, D.; Benjamin, E.J.; Go, A.S.; Arnett, D.K.; Blaha, M.J.; Cushman, M.; Das, S.R.; de Ferranti, S.; Després, J.-P.; Fullerton, H.J.; et al. American Heart Association Statistics Committee; Stroke Statistics Subcommittee. Heart Disease and Stroke Statistics–2016 Update: A report from the American Heart Association. Circulation 2016, 133, e38–e360. [Google Scholar] [CrossRef]
  9. Chun, S.; Tu, J.V.; Wijeysundera, H.C.; Austin, P.C.; Wang, X.; Levy, D.; Lee, D.S. Lifetime Analysis of Hospitalizations and Survival of Patients Newly Admitted With Heart Failure. Circ. Heart Fail. 2012, 5, 414–421. [Google Scholar] [CrossRef] [Green Version]
  10. Krumholz, H.M.; Merrill, A.R.; Schone, E.M.; Schreiner, G.C.; Chen, J.; Bradley, E.H.; Wang, Y.; Wang, Y.; Lin, Z.; Straube, B.M.; et al. Patterns of Hospital Performance in Acute Myocardial Infarction and Heart Failure 30-Day Mortality and Readmission. Circ. Cardiovasc. Qual. Outcomes 2009, 2, 407–413. [Google Scholar] [CrossRef] [Green Version]
  11. Joynt, K.E.; Jha, A.K. Who has higher readmission rates for heart failure, and why? Implications for efforts to improve care using financial incentives. Circ. Cardiovasc. Qual. Outcomes 2011, 4, 53–59. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  12. Krumholz, H.M.; Amatruda, J.; Smith, G.L.; Mattera, J.A.; Roumanis, S.A.; Radford, M.J.; Crombie, P.; Vaccarino, V. Randomized trial of an education and support intervention to preventreadmission of patients with heart failure. J. Am. Coll. Cardiol. 2002, 39, 83–89. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Phillips, C.O.; Wright, S.M.; Kern, D.E.; Singa, R.M.; Shepperd, S.; Rubin, H.R. Comprehensive discharge planning with postdischarge support for older patients with congestive heart failure: A meta-analysis. JAMA 2004, 291, 1358–1367. [Google Scholar] [CrossRef]
  14. Giamouzis, G.; Kalogeropoulos, A.; Georgiopoulou, V.; Laskar, S.; Smith, A.L.; Dunbar, S.; Triposkiadis, F.; Butler, J. Hospitalization Epidemic in Patients With Heart Failure: Risk Factors, Risk Prediction, Knowledge Gaps, and Future Directions. J. Card. Fail. 2011, 17, 54–75. [Google Scholar] [CrossRef] [PubMed]
  15. Kossovsky, M.P.; Sarasin, F.P.; Perneger, T.V.; Chopard, P.; Sigaud, P.; Gaspoz, J.-M. Unplanned readmissions of patients with congestive heart failure: Do they reflect in-hospital quality of care or patient characteristics? Am. J. Med. 2000, 109, 386–390. [Google Scholar] [CrossRef]
  16. Ouwerkerk, W.; Voors, A.A.; Zwinderman, A.H. Factors Influencing the Predictive Power of Models for Predicting Mortality and/or Heart Failure Hospitalization in Patients With Heart Failure. JACC Hear. Fail. 2014, 2, 429–436. [Google Scholar]
  17. Iribarren, C.; Karter, A.J.; Go, A.S.; Ferrara, A.; Liu, J.Y.; Sidney, S.; Selby, J.V. Glycemic Control and Heart Failure Among Adult Patients With Diabetes. Circulation 2001, 103, 2668–2673. [Google Scholar] [CrossRef] [Green Version]
  18. Upadhyay, S.; Stephenson, A.; Smith, D.G. Readmission Rates and Their Impact on Hospital Financial Performance: A Study of Washington Hospitals. Inq. J. Health Care Organ. Provis. Financ. 2019, 56, 0046958019860386. [Google Scholar] [CrossRef]
  19. Theobald, C.N.; Anctil, B.; Vasilevskis, E.E. Reducing Hospital Readmission Rates: Current Strategies and Future Directions. Annu. Rev. Med. 2014, 65, 471–485. [Google Scholar]
  20. Artetxe, A.; Beristain, A.; Graña, M. Predictive models for hospital readmission risk: A systematic review of methods. Comput. Methods Programs Biomed. 2018, 164, 49–64. [Google Scholar] [CrossRef]
  21. Awan, S.E.; Sohel, F.; Sanfilippo, F.; Bennamoun, M.; Dwivedi, G. Machine learning in heart failure: Ready for prime time. Curr. Opin. Cardiol. 2018, 33, 190–195. [Google Scholar] [CrossRef]
  22. Beecy, A.N.; Gummalla, M.; Sholle, E.; Xu, Z.; Zhang, Y.; Michalak, K.; Dolan, K.; Hussain, Y.; Lee, B.C.; Zhang, Y.; et al. Utilizing electronic health data and machine learning for the prediction of 30-day unplanned readmission or all-cause mortality in heart failure. Cardiovasc. Digit. Health J. 2020, 1, 71–79. [Google Scholar] [CrossRef] [PubMed]
  23. Zhang, Z.; Cao, L.; Chen, R.; Zhao, Y.; Lv, L.; Xu, Z.; Xu, P. Electronic healthcare records and external outcome data for hospitalized patients with heart failure. Sci. Data 2021, 8, 46. [Google Scholar] [CrossRef] [PubMed]
  24. Ishaq, A.; Sadiq, S.; Umer, M.; Ullah, S.; Mirjalili, S.; Rupapara, V.; Nappi, M. Improving the Prediction of Heart Failure Patients’ Survival Using SMOTE and Effective Data Mining Techniques. IEEE Access 2021, 9, 39707–39716. [Google Scholar] [CrossRef]
  25. Golas, S.B.; Shibahara, T.; Agboola, S.; Otaki, H.; Sato, J.; Nakae, T.; Hisamitsu, T.; Kojima, G.; Felsted, J.; Kakarmath, S.; et al. A machine learning model to predict the risk of 30-day readmissions in patients with heart failure: A retrospective analysis of electronic medical records data. BMC Med. Inform. Decis. Mak. 2018, 18, 44. [Google Scholar] [CrossRef] [Green Version]
  26. Liu, X.; Chen, Y.; Bae, J.; Li, H.; Johnston, J.; Sanger, T. Predicting Heart Failure Readmission from Clinical Notes Using Deep Learning. In Proceedings of the 2019 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), San Diego, CA, USA, 18–21 November 2019; pp. 2642–2648. [Google Scholar] [CrossRef] [Green Version]
  27. ZZhang Cao, L.; Zhao, Y.; Xu, Z.; Chen, R.; Lv, L.; Xu, P. Hospitalized patients with heart failure: Integrating electronic healthcare records and external outcome data (version 1.3). In PhysioNet; PhysioNet: Cambridge, MA, USA, 2022. [Google Scholar]
  28. Freedman, D.; Robert, P.; Purves, R. Statistics, 4th ed.; W.W. Norton & Company: New York, NY, USA, 2007. [Google Scholar]
  29. Ibrahim, H. 20-Year-Old Iranian Confirmed as World’s Shortest Man. Guinness World Records. 2022. Available online: https://www.guinnessworldrecords.com/news/2022/12/20-year-old-iranian-confirmed-as-worlds-shortest-man-730170 (accessed on 5 January 2023).
  30. Quintero, M.; LeBoulluec, A. Missing Data Imputation for Ordinal Data. Int. J. Comput. Appl. 2018, 181, 10–16. [Google Scholar] [CrossRef]
  31. Hu, Z.; Melton, G.B.; Arsoniadis, E.G.; Wang, Y.; Kwaan, M.R.; Simon, G.J. Strategies for handling missing clinical data for automated surgical site infection detection from the electronic health record. J. Biomed. Inform. 2017, 68, 112–120. [Google Scholar] [CrossRef]
  32. Sharafoddini, A.; Dubin, J.A.; Maslove, D.M.; Lee, J. A New Insight Into Missing Data in Intensive Care Unit Patient Profiles: Observational Study. JMIR Public Health Surveill. 2019, 7, e11605. [Google Scholar] [CrossRef] [Green Version]
  33. Van Smeden, M.; Groenwold, R.H.; Moons, K.G. A cautionary note on the use of the missing indicator method for handling missing data in prediction research. J. Clin. Epidemiol. 2020, 125, 188–190. [Google Scholar]
  34. Shin, S.; Austin, P.C.; Ross, H.J.; Abdel-Qadir, H.; Freitas, C.; Tomlinson, G.; Chicco, D.; Mahendiran, M.; Lawler, P.R.; Billia, F.; et al. Machine learning vs. conventional statistical models for predicting heart failure readmission and mortality. ESC Heart Fail. 2020, 8, 106–115. [Google Scholar] [CrossRef]
  35. Frizzell, J.D.; Liang, L.; Schulte, P.J.; Yancy, C.W.; Heidenreich, P.A.; Hernandez, A.F.; Bhatt, D.L.; Fonarow, G.C.; Laskey, W.K. Prediction of 30-Day All-Cause Readmissions in Patients Hospitalized for Heart Failure. JAMA Cardiol. 2017, 2, 204–209. [Google Scholar] [CrossRef] [Green Version]
  36. Jiang, W.; Siddiqui, S.; Barnes, S.; Barouch, L.A.; Korley, F.; Martinez, D.A.; Toerper, M.; Cabral, S.; Hamrock, E.; Levin, S. Readmission Risk Trajectories for Patients With Heart Failure Using a Dynamic Prediction Approach: Retrospective Study. JMIR Public Health Surveill. 2019, 7, e14756. [Google Scholar] [CrossRef] [PubMed]
  37. Van Buuren, S.; Groothuis-Oudshoorn, K. Mice: Multivariate Imputation by Chained Equations in R. J. Stat. Softw. 2011, 45, 1–67. [Google Scholar] [CrossRef] [Green Version]
  38. Jazayeri, A.; Liang, O.S.; Yang, C.C. Imputation of Missing Data in Electronic Health Records Based on Patients’ Similarities. J. Health Inform. Res. 2020, 4, 295–307. [Google Scholar] [CrossRef] [PubMed]
  39. Damery, S.; Combes, G. Evaluating the predictive strength of the LACE index in identifying patients at high risk of hospital readmission following an inpatient episode: A retrospective cohort study. BMJ Open 2017, 7, e016921. [Google Scholar] [CrossRef] [Green Version]
  40. Hakim, M.; Garden, F.L.; Jennings, M.D.; Dobler, C.C. Performance of the LACE index to predict 30-day hospital readmissions in patients with chronic obstructive pulmonary disease. Clin. Epidemiol. 2017, 10, 51–59. [Google Scholar] [CrossRef] [Green Version]
  41. Robinson, R.; Bhattarai, M.; Hudali, T.; Vogler, C. Predictors of 30-day hospital readmission: The direct comparison of number of discharge medications to the HOSPITAL score and LACE index. Future Health J. 2019, 6, 209–214. [Google Scholar] [CrossRef] [Green Version]
  42. Van Walraven, C.; Dhalla, I.A.; Bell, C.; Etchells, E.; Stiell, I.G.; Zarnke, K.; Austin, P.C.; Forster, A.J. Derivation and validation of an index to predict early death or unplanned readmission after discharge from hospital to the community. Can. Med. Assoc. J. 2010, 182, 551–557. [Google Scholar] [CrossRef] [Green Version]
  43. Saluk, J.L.; Blackwell, R.H.; Gange, W.; Zapf, M.A.C.; Kothari, A.N.; Kuo, P.C.; Quek, M.L.; Flanigan, R.C.; Gupta, G.N. The LACE Score as a Tool to Identify Radical Cystectomy Patients at Increased Risk of 90-Day Readmission and Mortality. Curr. Urol. 2018, 12, 20–26. [Google Scholar] [CrossRef]
  44. Wu, J.-R.; Moser, D.K. Medication Adherence Mediates the Relationship Between Heart Failure Symptoms and Cardiac Event-Free Survival in Patients With Heart Failure. J. Cardiovasc. Nurs. 2018, 33, 40–46. [Google Scholar] [CrossRef]
  45. Singh, D.; Singh, B. Investigating the impact of data normalization on classification performance. Appl. Soft Comput. 2020, 97, 105524. [Google Scholar] [CrossRef]
  46. Kasim, S.; Malek, S.; Song, C.; Ahmad, W.A.W.; Fong, A.; Ibrahim, K.S.; Safiruz, M.S.; Aziz, F.; Hiew, J.H.; Ibrahim, N. In-hospital mortality risk stratification of Asian ACS patients with artificial intelligence algorithm. PLoS ONE 2022, 17, e0278944. [Google Scholar] [CrossRef]
  47. Athanasiou, M.; Sfrintzeri, K.; Zarkogianni, K.; Thanopoulou, A.C.; Nikita, K.S. An explainable XGBoost–based approach towards assessing the risk of cardiovascular disease in patients with Type 2 Diabetes Mellitus. In Proceedings of the 2020 IEEE 20th International Conference on Bioinformatics and Bioengineering (BIBE), Cincinnati, OH, USA, 26–28 October 2020; pp. 859–864. [Google Scholar] [CrossRef]
  48. Nguyen, T.T.; Huang, J.Z.; Nguyen, T.T. Unbiased feature selection in learning random forests for high-dimensional data. Sci. World J. 2015, 2015, 471371. [Google Scholar] [CrossRef] [Green Version]
  49. Chen, T.; Guestrin, C. XGBoost: A Scalable Tree Boosting System. In Proceedings of the KDD’16: 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016. [Google Scholar] [CrossRef] [Green Version]
  50. Friedman, J.H. Greedy function approximation: A gradient boosting machine. Ann. Stat. 2001, 29, 1189–1232. [Google Scholar] [CrossRef]
  51. Xu, Y.; Yang, X.; Huang, H.; Peng, C.; Ge, Y.; Wu, H.; Wang, J.; Xiong, G.; Yi, Y. Extreme Gradient Boosting Model Has a Better Performance in Predicting the Risk of 90-Day Readmissions in Patients with Ischaemic Stroke. J. Stroke Cerebrovasc. Dis. 2019, 28, 104441. [Google Scholar] [CrossRef]
  52. Singh, Y.K.; Sinha, N.; Singh, S.K. Heart Disease Prediction System Using Random Forest. In Communications in Computer and Information Science; Singh, M., Gupta, P.K., Tyagi, V., Sharma, A., Ören, T., Grosky, W., Eds.; Springer: Singapore, 2017; pp. 613–623. [Google Scholar] [CrossRef]
  53. Geurts, P.; Ernst, D.; Wehenkel, L. Extremely randomized trees. Mach. Learn. 2006, 63, 3–42. [Google Scholar] [CrossRef] [Green Version]
  54. Alfian, G.; Syafrudin, M.; Fahrurrozi, I.; Fitriyani, N.L.; Atmaji, F.T.D.; Widodo, T.; Bahiyah, N.; Benes, F.; Rhee, J. Predicting Breast Cancer from Risk Factors Using SVM and Extra-Trees-Based Feature Selection Method. Computers 2022, 11, 136. [Google Scholar] [CrossRef]
  55. Shafique, R.; Mehmood, A.; Ullah, S.; Choi, G.S. Cardiovascular Disease Prediction System Using Extra Trees Classifie. Res. Sq. 2019, 4, 1–7. [Google Scholar] [CrossRef]
  56. Zhang, Z.; Qiu, H.; Li, W.; Chen, Y. A stacking-based model for predicting 30-day all-cause hospital readmissions of patients with acute myocardial infarction. BMC Med. Inform. Decis. Mak. 2020, 20, 335. [Google Scholar] [CrossRef]
  57. Yusuf, M.; Atal, I.; Li, J.; Smith, P.; Ravaud, P.; Fergie, M.; Callaghan, M.; Selfe, J. Reporting quality of studies using machine learning models for medical diagnosis: A systematic review. BMJ Open 2020, 10, e034568. [Google Scholar] [CrossRef] [Green Version]
  58. Alba, A.C.; Agoritsas, T.; Walsh, M.; Hanna, S.; Iorio, A.; Devereaux, P.J.; McGinn, T.; Guyatt, G. Discrimination and Calibration of Clinical Prediction Models: Users’ Guides to the Medical Literature. JAMA 2017, 318, 1377–1384. [Google Scholar] [CrossRef] [PubMed]
  59. Liu, Y.; Chen, P.-H.C.; Krause, J.; Peng, L. How to Read Articles That Use Machine Learning. JAMA 2019, 322, 1806–1816. [Google Scholar] [CrossRef]
  60. Valverde-Albacete, F.J.; Peláez-Moreno, C. 100% Classification Accuracy Considered Harmful: The Normalized Information Transfer Factor Explains the Accuracy Paradox. PLoS ONE 2014, 9, e84217. [Google Scholar] [CrossRef] [Green Version]
  61. Sharma, V.; Kulkarni, V.; Mcalister, F.; Eurich, D.; Keshwani, S.; Simpson, S.H.; Voaklander, D.; Samanani, S. Predicting 30-Day Readmissions in Patients with Heart Failure Using Administrative Data: A Machine Learning Approach. J. Card. Fail. 2021, 28, 710–722. [Google Scholar] [CrossRef] [PubMed]
  62. Bat-Erdene, B.-I.; Zheng, H.; Son, S.H.; Lee, J.Y. Deep learning-based prediction of heart failure rehospitalization during 6, 12, 24-month follow-ups in patients with acute myocardial infarction. Health Inform. J. 2022, 28, 14604582221101529. [Google Scholar] [CrossRef] [PubMed]
  63. Chiu, C.-C.; Wu, C.-M.; Chien, T.-N.; Kao, L.-J.; Li, C.; Jiang, H.-L. Applying an Improved Stacking Ensemble Model to Predict the Mortality of ICU Patients with Heart Failure. J. Clin. Med. 2022, 11, 6460. [Google Scholar] [CrossRef]
  64. Jing, L.; Cerna, A.E.U.; Good, C.W.; Sauers, N.M.; Schneider, G.; Hartzel, D.N.; Leader, J.B.; Kirchner, H.L.; Hu, Y.; Riviello, D.M.; et al. A Machine Learning Approach to Management of Heart Failure Populations. JACC Heart Fail. 2020, 8, 578–587. [Google Scholar]
  65. Chen, S.; Hu, W.; Yang, Y.; Cai, J.; Luo, Y.; Gong, L.; Li, Y.; Si, A.; Zhang, Y.; Liu, S.; et al. Predicting Six-Month Re-Admission Risk in Heart Failure Patients Using Multiple Machine Learning Methods: A Study Based on the Chinese Heart Failure Population Database. J. Clin. Med. 2023, 12, 870. [Google Scholar] [CrossRef]
Figure 1. General overview of the methodology.
Figure 1. General overview of the methodology.
Diagnostics 13 01948 g001
Figure 2. Feature incrementation vs. Accuracy (in %). The number of features is in increments of 1 step each time, and an increase in accuracy of logistic regression classifier is observed. Highest accuracy is observed at 9 features.
Figure 2. Feature incrementation vs. Accuracy (in %). The number of features is in increments of 1 step each time, and an increase in accuracy of logistic regression classifier is observed. Highest accuracy is observed at 9 features.
Diagnostics 13 01948 g002
Figure 3. Classical and stacking ML model training framework.
Figure 3. Classical and stacking ML model training framework.
Diagnostics 13 01948 g003
Figure 4. Feature ranking using XGBoost model.
Figure 4. Feature ranking using XGBoost model.
Diagnostics 13 01948 g004
Figure 5. Confusion matrix for the best-performing XGBoost stacking ML classifier.
Figure 5. Confusion matrix for the best-performing XGBoost stacking ML classifier.
Diagnostics 13 01948 g005
Figure 6. ROC curve for the three best-performing models and the stacked classifier.
Figure 6. ROC curve for the three best-performing models and the stacked classifier.
Diagnostics 13 01948 g006
Table 1. All types of diseases including HF that are present in the dataset [23].
Table 1. All types of diseases including HF that are present in the dataset [23].
Disease NameYesNo
Myocardial Infarction1865143
Congestive Heart Failure1872136
Peripheral Vascular Disease1907101
Cerebrovascular Disease1858150
Dementia1893115
Chronic Obstructive Pulmonary Disease (COPD)1775233
Connective Tissue Disease20044
Peptic Ulcer Disease196145
Chronic Kidney Disease1532474
Hemiplegia199612
Solid Tumor196939
Liver Disease192384
Table 2. LACE scoring scheme.
Table 2. LACE scoring scheme.
LACE Score ComponentsScore
L: Length of stay (days)
Associated features in the dataset:
‘dischargeDay’
11
22
33
4–64
7–145
≥147
A: Acuity of admission
Associated feature in the dataset:
‘admission.way’
Non-emergency0
Emergency3
C: Charlson comorbidity score
Associated feature in the dataset: ‘CCI.score’
00
11
22
33
≥44
E: Emergency department visits in the preceding 6 months
Associated feature in the dataset:
‘return.to.emergency.department.within.6.months’
00
11
22
3
≥4
3
4
Table 3. Statistics of the different important features for readmitted and not readmitted cases.
Table 3. Statistics of the different important features for readmitted and not readmitted cases.
FeaturesUnitMissing Rate (%)Not Readmitted (0)Readmitted (1)Overallp-Value
[Mean ± std][Mean ± std][Mean ± std]
Neutrophil RatioN/A10.756 ± 0.1070.744 ± 0.100.751 ± 0.104<0.05
Discharge DayDays09 ± 7.0610 ± 9.019 ± 8.030.10
Cholesterolmmol/L103.77 ± 1.073.66 ± 1.113.72 ± 1.090.06
Sodiummmol/L<1138.5 ± 4.91137.9 ± 4.87138.2 ± 4.90<0.05
Partial Pressure of CO2mmHg5136 ± 10.435.1 ± 8.1336 ± 9.560.10
Direct Bilirubinμmol/L59 ± 9.769 ± 9.289 ± 9.55<0.05
Albuming/L536.3 ± 5.0536.9 ± 4.8736.5 ± 4.98<0.05
Globuling/L528.7 ± 6.3328.4 ± 5.7028.6 ± 6.06<0.05
FiO2(%)033.0 ± 5.1832.33 ± 4.1532.7 ± 4.760.07
Systolic Blood PressuremmHg0133 ± 25.0129 ± 24.2131 ± 24.70.07
Cystatinmg/L21.8 ± 0.9711.89 ± 0.9251.8 ± 0.9510.05
Potassium Ionmmol/L513.84 ± 0.6523.98 ± 0.6853.89 ± 0.6690.07
White Globulin RatioN/A51.32 ± 0.3161.35 ± 0.301.33 ± 0.309<0.05
D-dimermg/L83 ± 6.342.0 ± 3.642 ± 5.330.08
Indirect Bilirubinμmol/L513 ± 9.5714.0 ± 8.8314 ± 9.25<0.05
FucosidaseU/L2619.3 ± 6.3119.5 ± 5.7819.4 ± 6.08<0.05
Sodium Ionmmol/L51136.4 ± 5.14135.9 ± 4.71136.2 ± 4.970.09
Left Ventricular End Diastolic Diameter, LVcm3552.4 ± 10.654.2 ± 11.353.1 ± 10.90.12
Eosinophil RatioN/A10.017 ± 0.0310.02 ± 0.0310.019 ± 0.0310.05
High Sensitivity Proteinmg/L5326 ± 35.524 ± 33.525 ± 34.7<0.05
Measured Residual Basemmol/L51−1.8 ± 4.90−2.1 ± 4.44−1.9 ± 4.72<0.05
Glomerular Filtration RatemL/min/1.73 m2371 ± 36.566 ± 36.569 ± 36.60.05
Mean Hemoglobin Volumepg129.9 ± 3.2830.0 ± 3.5929.9 ± 3.42<0.05
Mitral Valve EMSm/s516 ± 42.93 ± 29.95 ± 38.6<0.05
Diastolic Blood PressuremmHg077 ± 14.676 ± 14.277 ± 14.5<0.05
Weightkg052 ± 10.752.5 ± 11.152 ± 10.9<0.05
Basophil Count×109/L10.030 ± 0.0310.033 ± 0.0280.031 ± 0.029<0.05
Platelet Hematocrit(%)50.176 ± 0.0680.17 ± 0.0670.174 ± 0.068<0.05
Prothrombin Activity(%)267 ± 18.365 ± 18.666 ± 18.40.05
Reduced Hemoglobin(%)514.3 ± 6.224.1 ± 6.04.2 ± 6.130.05
Ureammol/L19.4 ± 5.849.8 ± 5.169.6 ± 5.55<0.05
Table 4. Different evaluation metrics result from 5 folded cross-validated data for different classical machine learning models. Best performances marked in bold.
Table 4. Different evaluation metrics result from 5 folded cross-validated data for different classical machine learning models. Best performances marked in bold.
Model NameAccuracyPrecisionRecallSpecificityF1-Score
MLP Classifier82.1982.282.1981.2782.12
Linear Discriminant Analysis78.0278.1378.0377.9478.06
XGBClassifier75.5175.4675.5174.4475.41
Random Forest Classifier76.1276.0676.1175.2776.05
Logistic Regression78.9879.0178.9878.7378.99
Support Vector Machine47.1758.3247.1655.8334.84
ExtraTree Classifier70.770.6670.769.9970.67
AdaBoost Classifier8787.1187.0186.0486.94
K-nearest neighbor Classifier70.5570.4770.5569.6270.48
Gradient Boosting Classifier84.984.9284.984.0784.84
CatBoost88.3688.8988.3686.8688.23
LGB Classifier80.7380.7880.7379.5880.62
ElasticNet78.9879.0278.9878.7478.99
Table 5. Performance of the stacking model with different evaluation metrics.
Table 5. Performance of the stacking model with different evaluation metrics.
ModelAccuracyPrecisionRecallSpecificityF1-Score
MLP Classifier89.1190.189.1187.2788.94
Linear Discriminant Analysis88.2188.6988.2186.7688.09
XGBClassifier89.4190.189.4187.8389.28
Random Forest Classifier89.1690.0789.1687.3989.01
Logistic Regression87.3687.6187.3686.1687.27
Support Vector Machine (SVM)29.429.8929.4130.3329.25
ExtraTrees Classifier89.0690.0189.0687.2588.9
AdaBoost Classifier89.0689.8689.0687.3688.91
K-nearest Neighbours Classifier 86.486.6386.485.286.31
Gradient Boosting Classifier85.8586.0285.8584.7285.76
CatBoost89.1189.9289.1187.488.96
LGB classifier88.9189.7488.9187.1888.75
ElasticNet87.3187.5687.3186.0987.21
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Rahman, M.S.; Rahman, H.R.; Prithula, J.; Chowdhury, M.E.H.; Ahmed, M.U.; Kumar, J.; Murugappan, M.; Khan, M.S. Heart Failure Emergency Readmission Prediction Using Stacking Machine Learning Model. Diagnostics 2023, 13, 1948. https://0-doi-org.brum.beds.ac.uk/10.3390/diagnostics13111948

AMA Style

Rahman MS, Rahman HR, Prithula J, Chowdhury MEH, Ahmed MU, Kumar J, Murugappan M, Khan MS. Heart Failure Emergency Readmission Prediction Using Stacking Machine Learning Model. Diagnostics. 2023; 13(11):1948. https://0-doi-org.brum.beds.ac.uk/10.3390/diagnostics13111948

Chicago/Turabian Style

Rahman, Md. Sohanur, Hasib Ryan Rahman, Johayra Prithula, Muhammad E. H. Chowdhury, Mosabber Uddin Ahmed, Jaya Kumar, M. Murugappan, and Muhammad Salman Khan. 2023. "Heart Failure Emergency Readmission Prediction Using Stacking Machine Learning Model" Diagnostics 13, no. 11: 1948. https://0-doi-org.brum.beds.ac.uk/10.3390/diagnostics13111948

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop