Chandan Singh
Useful interpretations for machine learning
in science & medicine
Advisor: Bin Yu
Input
Prediction
Interpretable ml extracts useful, relevant knowledge
Problem, data, audience
Model
Post hoc analysis
Iterate
Predictive accuracy
Descriptive accuracy
Relevancy
Existing methods don’t capture interactions well (e.g. LIME, SHAP)....
Interpreting + improving neural networks
Hierarchical interactions: ICLR ‘19
CS*, Murdoch*, & Yu
Interaction regularization: ICML ‘20
Rieger, CS, Murdoch, & Yu
Transformation importances: ICLR Workshop ‘21
CS*, Ha*, Lanusse, Boehm, Liu, & Yu
Negative
“not very good”
Prediction
very good
not very good
not
very
good
Positive
Negative
Interpretation
not | very | good |
Pred
| |
| |
| |
| |
| very | good |
| |
| |
| |
| |
not | | |
| |
| |
| |
| |
=
+
Original input
Relevant part
Irrelevant part
CD importance of very good
=
+
=
+
=
CD importance
Irrelevant
+
Relevant (at layer i)
Linear / conv
Irrelevant (at layer i)
Relu
Pooling
Importances work for
transformations via a reparameterization trick
Evaluation 1: human accuracy at identifying a corrupted model
Human accuracy
Evaluation 2: penalizing importances improves performance
Before penalty
0.67 F1
After penalty
0.73 F1
Molecular-partner prediction
with interpretable neural networks
CS*, Li*, Ruan, Song, Dang, He, Upadhyayula, & Yu
Clathrin-mediated endocytosis
Cremona 2001
Tracking molecular partners is a central challenge
Hey
Hey partner
...but it’s experimentally difficult
given one partner, we predict the other
X
Fitting a network for spike prediction
Time
Amplitude
Auxilin
Aguet et al
‘13, ‘16
Binary classification: spike?
Clathrin
Spike!
Interpretation
Positive
Negative
Prediction
Spike!
Interpreting individual predictions with CD
Positive
Negative
DALLE-2
Ha, CS, Lanusse, Upadhyayula, & Yu NeurIPS `21
Prediction
Our interpretation is incomplete
Thousands of parameters….
Solution: distillation
Prediction
<20 parameters!
Wavelet transform
z
x
Wavelet function
Wavelet function can vary
x
z
small
Orthonormal basis under these conditions (mallat, 1998):
Haar
Mexican hat
db4
sym4
Adaptive wavelet
x
z
Adaptive wavelet + distillation
AWD is accurate and interpretable
x
z
Input
Wavelet reconstruction
Prediction
AWD | Neural network | Baseline |
0.263 | 0.237 | 0.197 |
R2
AWD works on a completely different problem
AWD | Neural network | Roberts cross* |
1.029 | 1.156 | 1.259 |
x 10-4
*Ribli et al (2019) Nature Astronomy
(RMSE)
Clinical-decision rule modeling
Kornblith*, CS*, Devlin, Addo, Streck, Holmes,
Kuppermann, Grupp-Phelan, Fineman, Butte, & Yu
203/12044
6/5034
112/1963
38/826
36/2532
6/955
2/305
1/34
2/395
Abdo trauma / SBS
GCS 3-13
Abdo tenderness
Thoracic wall trauma
Abdo pain
↓ or no BS
Vomiting
FIGS: Fast interpretable greedy-tree sums
Tan*, Singh*, Nasseri, Agarwal, & Yu
Number of rules
FIGS
Trees compete with each other to predict the outcome
+
+
i
models
Singh*, Nasseri*, Tan, Tang, & Yu
JOSS, ‘21
: a python library for interpretability
⭐ 700+
Acknowledgements
Acknowledgements
Thesis committee
Internship advisors
Undergrad advisors
Yu Group
Berkeley friends
Old friends
Family
Thank you
Interpreting individual predictions with CD
Positive
Negative
Positive
Negative
Unsure
Time
Random stuff
w/ Alejo Rico-Guevara, Xin Cheng
w/ Gang-Yu Liu, Jiali Zhang
w/ Mike Eickenberg, Reza Abbasi-Asl, Mike Oliver
w/ Summer Devlin, Jamie Murdoch
w/ Raaz Dwivedi, Martin Wainwright
w/ Yu-group
w/ James Duncan, Rush Kapoor, Sahil Saxena
Hierarchical shrinkage for trees
Agarwal*, Tan*, Ronen, Singh, & Yu
arXiv, submitted to ICML ‘22
Covid county-level
forecasting / data curation
Yu Group + Response4Life, HDSR 2021
Evaluating interpretation: simulations
X
Y
Training
Importance
CD | Deeplift | Shap | IG |
0.4 | 3.6 | 4.0 | 4.2 |
Error (%)
FIGS can potentially improve clinical decision rules
IAI
CSI
TBI
Ha, Singh, Lanusse, Upadhyayula, & Yu, NeurIPS `21
DALLE-2
bin yu
jamie murdoch
karl kumbier
reza
abbasi-asl
aaron kornblith
françois lanusse
wooseok ha
vanessa boehm
gokul
upadhyayula
xiongtao
ruan
xiao
li
yu group
yan shuo tan
Raaz Dwivedi
laura rieger
Definitions
Interpretability is the “ability to explain or to present in understandable terms to a human” (doshi-velez & kim, 2017)
Related to trust, causality, transferability, informativeness, fairness (lipton 2017)
“Explanations are... the currency in which we exchanged beliefs” (lombrozo 2006)
Interpretations should be useful
Improve predictive accuracy
Uncertainty
Make causal recommendations
not | very | good |
| |
| |
| |
| |
| very | good |
| |
| |
| |
| |
not | | |
| |
| |
| |
| |
Original input
Baseline importance of very good
pred
Relevant part
pred2
pred3
Irrelevant part
Molecular partners underlie biological processes
Images by DALLE 2
Predicting via peak counts
Predict using nearest-neighbor
Filtered values
at the local maxima
Binning into histogram
Ex: Pneumonia-death prediction
Caruana et al. 2015
Spurious correlation: asthma correlated with lower risk in the training data
cell biology
science &
medicine
cosmological inference
neuroscience
interaction summarization
interpreting neural nets
interpretable distillation
explanation penalization
building
interpretable models
rule-based algorithms
software
Back to biology: interpreting many predictions
Back to biology: interpreting a DNN prediction
transformation importance
singh, ha, lanusse, boehm, & yu
ICLR 2020 workshop
evaluating (hierarchical) CD importance
qualitative
human experiments
improving models
recovering groundtruth
and more...
Evaluating interpretation: penalizing scores improves performance
Unpenalized
0.67 (f1)
Penalized
0.73 (f1)
Patch becomes less important