1 of 22

Local Interpretable Model Agnostic Explanations (LIME)

and Discussion about other Heatmap Methods

2 of 22

Features vs Interpretable Representations

Interpretable explanations need to use a representation that is understable to humans

Ex - In text that representation would be binary vector indicating presence or absence of a word vs word embeddings

Ex - In image it can be a binary vector indicating presence/absence of a super-pixel (vs actual pixel values)

3 of 22

Some Notations

x - original representation of an instance being explained

x’ - binary vector for its interpretable representation

Explanation as a model g coming from a class G of interpretable models

Domain of g is binary (presence/absence of interpretable components)

Ω(g) - measure of complexity of model

f - model being explained

Πx - proximity measure between z and x (local point around x)

L - measure of unfaithfulness of g in approximating f in the local neighbourhood

4 of 22

General Equation of their formulation

Our task is to find a g which is interpretable and faithfully represents f in local neighbourhood

explanation(x)=argmin(g∈G) L(f,g,πx)+Ω(g)

A

5 of 22

6 of 22

7 of 22

8 of 22

9 of 22

10 of 22

Few Results

11 of 22

Global Reliability of models

Explanation of a single prediction provides some understanding into the reliability of the classifier

Still not sufficient to evaluate and assess trust in the model as a whole

Global understanding/reliability of the model by explaining a set of individual instances (carefully selected by their provided method)

12 of 22

Problems

Complexity of the explanation has to be defined in advance (compromise between fidelity and sparsity)

Instability of explanations (Source) - Stability of two very close points varied a lot in simulated settings

If you repeat the sampling (local points), you get different explanations (Bad) (Source)

13 of 22

Ref on LIME

Paper

Talks

Blog by Author and Code

Other interesting resources - [1], [2]

14 of 22

Some of the other heamap methods

15 of 22

Occlusion

16 of 22

Gradient based Heatmaps

17 of 22

What we are doing

We have a model which would classify the image based on the centre pixel regardless of the content of the image

18 of 22

Ideal Heatmap

How would the ideal model explanation look like?

It would be a binary image with 1 at centre and 0 everywhere representing that my centre pixel is the only important pixel for the classification

19 of 22

Applying occlusion for our model

Perfect result when applying occlusion with a 1x1 patch

20 of 22

Applying other heatmap methods?

21 of 22

Applying other heatmap methods?

22 of 22

Our model

f(x) = Softmax(Wx)

where x is your flattened input image.