LIME
Agnieszka Grala, Zuzanna Kurnicka
LIME - Local Interpretable Model-agnostic Explanations
Key idea
Locally approximate a black-box model by a simpler glass-box model, which is easier to interpret.
Intuition
Goal: understand the factors that influence a black-box model around a single instance of interest
Method
Data spaces
Note that models f and g may operate on different data spaces.
f : X → R
is defined on a large, p-dimensional space X corresponding to the p explanatory variables used in the model.
g : X’ → R
is defined on a q-dimensional space X’ with q<<p, often called the “space for interpretable representation”.
Algorithm
Data types
Data space transformation
Text data:
Image data:
Tabular data:
Leads to various implementations and potentially different results.
Image data space transformation example
How do you get the variations of the data?
Perform perturbations on the selected instance:
choose random numer of variables and switch (negate) their values
Challenge: Generating realistic perturbations can be difficult due to complex feature correlations.
Weights and Local Neighborhood
Purpose of Weights: To make the interpretable model a local approximation. Prioritize samples closer to the instance being explained.
Weight=Kernel(Distance)
Fitting the glass-box model
LIME for text data
Example: Spam classification
Creating variations of comments
Explanations
The word “channel” indicates a high probability of spam.
LIME for tabular data
Neighbourhood in tabular data
Choosing kernel width
Example: Titanic data
Define interpretable space
Steps
Black-box model: random forest model
Strengths
Limitations
Summary
The most useful applications of LIME are limited to high-dimensional data for which one can define a low-dimensional interpretable data representation, as in image analysis, text analysis, or genomics.
References
Thank you
<imagine funny meme here>