Variational Autoencoders pursue PCA directions (by accident)
July 2, 2019
Michal Rolínek
Max-Planck Institute for Intelligent Systems
1.
A LITTLE BIT ABOUT ME
BRIEF TRAJECTORY
MY WORK BEFORE...
… AND AFTER
[Rolínek, Martius]
[Rolínek*, Zietlow*, Martius]
2.
WE NEED TO TALK ABOUT DISENTANGLEMENT
[Higgins et al., Beta-VAE: Learning Visual Concepts with a Constrained Variational Framework, ICLR 2017]
“Learning meaningful and compact representations with structurally disentangled semantics.”
WHAT IS DISENTANGLEMENT?
Also formal approaches to the definition exist
[Higgins et al., Towards a Definition of Disentangled Representations, 2018]
BRIEF HISTORY
We were here (and failing)
From [Locatello et al.] ICML 2019
Rotations matter
3.
MAIN RESULT
FORMALIZED...
USUAL REACTIONS
WHY NOT OBVIOUS?
4.
FROM UNDER THE RUG
THE CLASSICAL VAE STORY
CANONICAL IMPLEMENTATION
β
What doesn’t explain choice of alignment...
PROOF STRATEGY
PROOF STRATEGY II
WHERE THE MATH IS FRAGILE
5.
THE HAPPY EXPERIMENTS
ORTHOGONALITY
VS.
DISENTANGLEMENT
COMMON DEGENERATE CASE WITH PCA
What are the two
principle components?
Ambiguous!
COMMON DEGENERATE CASE WITH PCA
Four restarts of linear Beta-VAE
In particular, (Beta-)VAE does not optimize for statistical independence.
DISCUSSION
6.
FINAL WORD ABOUT AN ONGOING PROJECT
BACK TO BASICS - WHAT IS THE POWER OF DEEP LEARNING?
EXACT ALGORITHMS ARE COOL!
PRELIMINARY RESULTS
(FOR NOW) SYNTHETIC EXAMPLES
THANK YOU!