This document summarizes a research paper that proposed using an ensemble learning approach to predict early hospital mortality for intensive care unit patients. It discusses previous work on mortality prediction that had limitations in handling missing data, imbalanced data, and lack of early prediction. The proposed method uses data mining techniques like random forest, decision trees, naive Bayes and PART on clinical variables from the first 6 hours of ICU patients. It conducted experiments with 10-fold cross validation to evaluate and compare the performance of these techniques as well as existing scoring models. The results showed the ensemble approach achieved higher accuracy than previous methods in early hospital mortality prediction.
How to do quick user assign in kanban in Odoo 17 ERP
Early Hospital Mortality Prediction Using Ensemble Learning
1. A review on
Early Hospital Mortality Prediction of Intensive Care Unit
Patients Using an Ensemble Learning Approach
Authors: Awad, A., Bader-El-Den, M., McNicholas, J., & Briggs, J.
By
Reza Sadeghi
Sadeghi.2@ wright.edu
Instructor
Dr. Romine
Nov 2017
2. Outline
• The importance of early mortality prediction
• Previous work pros and cons
• The proposed methods
• Experiments
• Conclusion
3. Early mortality prediction usage
• A decision support systems for intensivists
• The more accurate and quick decision provides the more benefit for
patients and health care resources
• The quality and cost of treatment
• The ICU patient is highly monitored using electronic equipment
4. Previous work pros and cons
Customized models
Score based models
Data mining models
- perform better than traditional scores [1]
- patients may suffering from heterogeneous diseases then selecting the
proper model is hard job
- rely on panels of experts or statistical models
- refined for use within specified geographical areas
- extracting features from Electronic Medical Records
- only few models are designed for early mortality
- low discrimination power
- many of required attributes are not always available at ICU admission
- higher performance rather than the previous methods
- descriptive modelling as it explains hidden clinical implications
- the results of different algorithm on different data set are not constant.
- it depends on the population of interest, the variables measured and the
outcome being tested
5. Missing Data Type:
1. missing completely at random (MCAR)
2. missing at random (MAR)
3. Missing not at random (MNAR) [2]
Missing Data Handling:
1. interpretation
2. ignoring those records from the dataset that
are not complete
3. substitute the missing value by the mean or
mode value of each attribute.
("ReplaceMissing-Values“ filter in weka)
4. predicted by using a learning algorithm, such
as Multiple Imputation or EMImputation [3]
Imbalance data Handling:
1. re-sampling [4,5]: under sampling and
oversampling.
2. making the classier 'cost sensitive' [46]
3. hybrid method
Data mining models issues
6. The structure of experiments
Data set:
MIMIC II
Age of patients:
patients at the age of 16 or above with a single ICU stay
Place of patients:
in either the Medical ICU (MICU), Surgical ICU (SICU) or Cardiac Surgery
Recovery Unit (CSRU)
Early Hospital Mortality Prediction :
define as predicting death inside the hospital via both the chart and lab-test
variables collected in the first 6 hours
Handling missing data :
ignored attributes with coverage below 10%
7. Selected Attributes
Random Forest, Decision
Trees, Naive Bayes and PART
ANN
Selected DM techniques
10 times 10-fold cross-validated
Experiment strategy
The Photo originated from
https://articles.mercola.com/sites/articles/archive/2010/03/13/hospitals-now-kill-48000-in-us-per-year-up-nearly-500-percent.aspx
The Photo originated fromhttps://www.statnews.com/2016/09/07/hospital-icu-modernize/
[1]-> [11]
1.interpretation: if an individual patients record has multiple entries missing, it may be explained that this is because they
were regarded as being less sick than others, so they were not prioritized. Equally, the patient may have
been regarded as being extremely sick, so they died before much can be done. Distinguishing these cases is
not simple in the absence of other information.
[2] -> [42]
[3] -> [44]
[4,5] -> [42, 45]
6 hour: This interval was reached after consulting several intensivists and considering gaps in literature
They selected 33 chart attributes and 25 lab-tests from the initially identified attributes with high coverage.
Attributes with higher coverage were considered, resulting in a total of 20 unique variables; 29 if we count
maximums and minimums.
In experiment A, the dataset contained all 20 unique variables (29 if counting maximums and minimums)
In experiment B, the dataset contained only the eight unique vital signs variables (10 if counting maximums and minimums)
the mean (rep1)
EMImputation algorithm (rep2)
We conducted five different experiments. In experiments A and B, we evaluated each of the four data
mining algorithms on each of six versions of the dataset (second column of Table II):
1. original dataset (original),
2. dataset after modified by applying the Synthetic Minority Oversampling Technique (SMOTE) [48],
an oversampling technique that involves increasing the size of the minority class with the insertion of
synthetic data (original+smote),
3. dataset after replacing missing values with the mean (rep1) to handle the issue of missing values
4. dataset after replacing missing values with mean and then applying SMOTE (rep1+smote),
5. datasets after replacing missing values using the EMImputation algorithm (rep2),
6. dataset after replacing missing values using EMImputation algorithm and applying SMOTE (rep2+smote).
1. dataset with eliminated records and the 20 unique variables (original),
2. dataset with eliminated records and the 20 unique variables and then applying smote (original+smote),
3. dataset with eliminated records and the top filtered ranked variables only (filtered top, 5, 10 and 15),
and
4. dataset with eliminated records and the top filtered ranked variables only and then applying smote
(filter+smote).
Graphical model of the roles of different accuracy parameters
In general in the top 5, 10 and 15 experiment categories, the models developed with the original attributes
(without any ltering) performed better than those with ltering. In the experiments without ltering, top
10 (original) and (original+smote) performed best (AUROC = 0.89 0.02), followed by top 5 (original)
(AUROC = 0.86 0.02). As for the ltered experiments, top 10 (lter) and (lter+smote) also performed
best (AUROC = 0.87 0.03), followed by top 15 (lter) and (lter+smote) (AUROC = 0.82 0.04).