1
XClusters: Explainability-first Clustering
Hyunseung Hwang and Steven Whang
Data Intelligence Lab, KAIST
Explaining Clusters
2
…..
…..
C1
C2
Explaining Clusters
3
…..
…..
C1
C2
Online
True
C2
Car Sales
Shopping
False
C1
C1
C2
Decision trees provide succinct summaries
Explainability-first Clustering
4
Explainable Clustering
Explainability-first Clustering
5
Explainable Clustering
Root
Online
20s
30s
30s
40s
Offline
Explainability-first Clustering
6
Root
Online
Offline
Explainable Clustering
Explainability-first Clustering
Root
Online
20s
30s
30s
40s
Offline
Two Types of Features
7
Extend Clustering
8
Holistic Optimization
9
Monotonic Properties
10
c1
c2
c3
c1
c1
c2
c3
c1
c2
c3
c1
c2
c3
c1
k increases → N increases, D decreases
α increases → N decreases, D increases
Monotonic Optimization
11
Branch-Reduce-and-Bound
12
B1
1
8
k
0
1
α
B2
B3
B3
B4
B5
Computing Upper, Lower Bounds
≥ D(k2,α) + λN(k1,α)
≥ D(k2,α1) + λN(k1,α2)
13
(k1,α2)
(k2,α1)
B
k1
k2
k
α1
α2
α
Example
14
Box | Lower bound | Upper Bound |
B2 | 0.3 | 0.4 |
B3 | 0.2 | 0.5 |
B2
B3
Example
15
B3 lower
< B2 lower
Box | Lower bound | Upper Bound |
B2 | 0.3 | 0.4 |
B3 | 0.2 | 0.5 |
Box | Lower bound | Upper Bound |
B2 | 0.3 | 0.4 |
B4 | 0.41 | 0.6 |
B5 | 0.5 | 0.7 |
B2
B3
B2
B4
B5
Example
16
B3 lower
< B2 lower
Discard B4, B5.
B2 = best block!
Return B2’s Upper bound
Box | Lower bound | Upper Bound |
B2 | 0.3 | 0.4 |
B3 | 0.2 | 0.5 |
Box | Lower bound | Upper Bound |
B2 | 0.3 | 0.4 |
B4 | 0.41 | 0.6 |
B5 | 0.5 | 0.7 |
0.41+ε > 0.4
0.5+ε > 0.4
B2
B3
B2
B4
B5
Experiments
We demonstrate:
17
Experiment Settings
18
Monotonicity Observations
19
k versus D
k versus N
α versus D
α versus N
Explainability and Distortion Results
Our algorithm achieves accurate and explainable results with a small runtime
20
Dataset | Method | D+λN | D | N | Runtime (s) |
Credit Card | 2-Step | 1.005 | 0.166 | 0.840 | 0.426 |
GS | 0.625 | 0.391 | 0.234 | 64.455 | |
BO | 0.657 | 0.364 | 0.293 | 16.595 | |
XClusters | 0.656 | 0.385 | 0.271 | 6.336 |
Explainability and Distortion Results
Our algorithm achieves accurate and explainable results with a small runtime
21
High N, Low D
Dataset | Method | D+λN | D | N | Runtime (s) |
Credit Card | 2-Step | 1.005 | 0.166 | 0.840 | 0.426 |
GS | 0.625 | 0.391 | 0.234 | 64.455 | |
BO | 0.657 | 0.364 | 0.293 | 16.595 | |
XClusters | 0.656 | 0.385 | 0.271 | 6.336 |
Explainability and Distortion Results
XClusters achieves a lowest objective value with a small runtime
22
High Runtime
Dataset | Method | D+λN | D | N | Runtime (s) |
Credit Card | 2-Step | 1.005 | 0.166 | 0.840 | 0.426 |
GS | 0.625 | 0.391 | 0.234 | 64.455 | |
BO | 0.657 | 0.364 | 0.293 | 16.595 | |
XClusters | 0.656 | 0.385 | 0.271 | 6.336 |
Comparison with Explainable Clustering
XClusters performs better due to its holistic optimization
23
Dataset | Method | D+λN | D | N |
Credit Card | Explainable k-Means [1] | 1.081 | 0.253 | 0.828 |
XClusters with Explainable k-Means | 0.708 | 0.565 | 0.143 |
[1] Moshkovitz et al. “Explainable k-Means and k-Medians Clustering,” 2020 ICML
Conclusion
24
Conclusion
25
Thank you!