Machine Learning
Mentors: Professor Nikola Banovic and Anindya Das Antar
The Challenge
Developing methods for patient specific predictions of in-hospital mortality
The timely and accurate detection of people at risk can save lives!
2
Abbreviations
3
Metric definitions
4
TN
FP
FN
TP
Predicted
Dead
Alive
Actual
Alive
Dead
Fraction of predicted positives that were actually positive
Fraction of actual positives that were correctly predicted
Fraction of predictions that were correctly classified
Group 1: MaSH
Markov model, SVM Hybrid
Anvit Garg, Alejandra Solis Sala, Ian Maywar, Rhea Verma
Group 1: Predicting ICU Mortality using HMMs and SVM
Anvit Garg, Alejandra Solis Sala, Ian Maywar, Rhea Verma
Model Illustration
Model Architecture
7
Feature selection and preprocessing
8
Models (background)
9
Hidden
Observed
Models
10
Cross Validation
11
SVM
HMM_dead
HMM_alive
time window transformation
Full Model Evaluation
12
1187
535
62
216
Predicted
Dead
Alive
Actual
Alive
Dead
Conclusions and Future Work
With more time, further developments can be made on the model:
While there were limitations in the dataset due to missing values, results may still inform future models that can allow for augmented resource allocation and confirmation of clinical decisions.
13
Thank You
Any questions?
Contact Information:
14
References
15
MACHINE LEARNING �GROUP 2
Felicia Zhang
Madeline J Peterson
Maya Nitsche Taylor
16
Mentors: Professor Nikola Banovic and Anindya Das Antar
The Model
17
Canva. (n.d.). Donut Decision Maker. Design a Superb Decision Tree Online with Canva. https://www.canva.com/graphs/decision-trees/ .
Yiu, T. (2019, August 14). Understanding random forest. Medium. https://towardsdatascience.com/understanding-random-forest-58381e0602d2.
Why Random Forests?
18
The goal: provide clinicians and hospitals with a bigger picture of patient survival in the ICU that may help with large-scale vision
The challenge: applying random forest to time-series data when the algorithm is not meant for time-series data
Methods
19
10,000 patients
development
2500 samples
training
5000 samples
testing
2500 samples
48 hours
8 hours
window and feature selection
summarize over each window
For a given time-series variable:
Understanding Time Windows
20
To use random forest, the time-series data must be collapsed over time windows to create 2D feature vectors:
Development - Choosing Window Size
Ran cross-validation on the development set using various window sizes, optimizing on the f1 score produced by the random forest.
Based on this analysis we decided to move forward with six windows.
21
In parallel with window selection, we also chose a set of summary statistics for best performance (using the development data set).
Development - Choosing Summary Statistics
Set 1: mean, min, max
Set 2: mean, min, max, median, std deviation
Set 3: mean, min, max, median, std deviation, 25th and 75th quantiles, count
22
Methods
23
Max Features
Final: 100
N Estimators
Final: 1000
Others:
Bootstrap: True
Class Weight: Balanced
Max Depth: 4
Methods
24
Comparison to other ML Models
F1 score
25
Final Metrics on the Testing Set
Note:
-1 = survival
1 = death
Important takeaways:
Final Values on test data
Precision: 0.377
F1 Score: 0.501
Recall: 0.749
Confusion Matrix
26
Receiver Operating Characteristic (ROC) Curves
27
Understanding the Model - Example Decision Tree
Top 10 important features:
GCSmedian5, GCSquant755, GCSmax5, GCSmean5, GCSquant255, GCSmin5, GCSmedian4, GCSmean4, mean_Urine, quant25_BUN
Note: the number at the end denotes the window. GCS (Glasgow coma scale) quantifies degree of consciousness.
28
Takeaways and Future Work
Takeaways
Future Work
29
References
Canva. (n.d.). Donut Decision Maker. Design a Superb Decision Tree Online with Canva. https://www.canva.com/graphs/decision-trees/ .
Goldberger, A., Amaral, L., Glass, L., Hausdorff, J., Ivanov, P. C., Mark, R., ... & Stanley, H. E. (2000). PhysioBank, PhysioToolkit, and PhysioNet: Components of a new research resource for complex physiologic signals. Circulation [Online]. 101 (23), pp. E215–e220.
Pedregosa, F., Varoquaux, G., Gramfort, A., Michel, V., Thirion, B., Grisel, O., Blondel, M., Prettenhofer, P., Weiss, R., Dubourg, V., Vanderplas, J., Passos, A., Cournapeau, D., Brucher, M., Perrot, M., & Duchesnay, É. (1970, January 1). Scikit-learn: Machine learning in Python. Journal of Machine Learning Research. https://jmlr.csail.mit.edu/papers/v12/pedregosa11a.html.
Yiu, T. (2019, August 14). Understanding random forest. Medium. https://towardsdatascience.com/understanding-random-forest-58381e0602d2.
Thank you!
Felicia Zhang, University of Michigan - fyzhang@umich.edu
Madeline Peterson, Albion College - mjp12@albion.edu
Maya Taylor, Brown University - maya_taylor@brown.edu
30
Long Short-Term Memory (LSTM) neural network to predict ICU mortality
By: Rami Shams, Sabir Meah, Esther Adegoke, Brian Lin
The Promise of LSTMs
Neural Network
RNN
LSTM
Raw Data
t = 1 Hour
L = 48 Hours
Cubic Spline Interpolation
≥ 3 measurements per feature
Forward and back filling
≥ 1 measurements per feature
Mean Imputation
0 measurements per feature
Z-Score standardization
Data Loader
Batch Size = 64
Time Variant(35)
Time Invariant(8)
Data Preprocessing and Missing Data approaches
10000 Patients
__________________
8000 Train
1000 Validation
1000 Test
Mean aggregation by time
Stratify by outcome label
Time Variant data(35)
Time Invariant data(8)
LSTM(40)
Hidden layer ∈ R48
Hidden layer ∈ R40
Internal state
Sigmoid output
t = 48
LSTM Architecture
ReLU
Methods - Data cleaning
ICU | CCU | CSRU | MI | SI |
Mortality | 0.13 | 0.05 | 0.20 | 0.15 |
Num of Patients (N = 10000) | 1476 | 2076 | 3609 | 2839 |
Methods - Hyperparameters
Loss with Adam Optimizer
Loss with Gradient Descent Optimizer
Results - Final Model Performance
Loss in Final Model
F1 Score in Final Model
Discussion
References
Contacts
Brian Lin - Carnegie Mellon University
Sabir Meah - University of Michigan
Rami Shams - University of Michigan
Esther Adegoke - Tufts University