interpretability workshop
(with notebooks)
interpretability depends on context
data
audience
overview
How can we build a
simple model?
example 1: sparse integer linear model
struck et al. 2017
example 2: optimal classification tree
bertsimas & dunn 2017
example 3: bayesian rule list
letham et al. 2015
example 4: rulefit
molnar et al. 2019
example 5: (causal) structural equation model
bouttou et al. 2013
example 6: prototypical neural networks
chen et al. 2018
Which features are globally important?
global linear feature importances
sobol’s indices (sobol 1993)
finding important variables: no interactions
screening unimportant variables: use interactions
only permutation importance requires a model
permutation importance (breiman 2001)
delta index (borgonovo 2007)
How does the model use different features?
PDP ICE Plot
ALE Plot
molnar 2019
friedman 2001
apley 2016
SHAP-Interact
lundberg et al. 2019
SHAP-Interact
lundberg et al. 2019
xdxd
How can we understand
one prediction?
LIME (ribeiro et al. 2016)
SHAP (lundberg & lee, 2017)
C
Sampling
the problem with local explanations
Sampling 2: conditional sampling can introduce spurious attribution for interactions
Let f(x) = X1
henin & metayer 2019
easy, effective uncertainty
What else is out there?