Data Geometry and DL - Lecture 9
Hyperbolic Neural Networks
Recent survey: https://arxiv.org/abs/2101.04562
Embeddings: what and why
Word2Vec – learn the context-dependent probab.
We get a dictionary-sized vector for each word
The result works remarkably like euclidean space !!
Hierarchical relationships
Typically, a tree structure is approximates well the graph.
Needed:
low distortion
metric trees embeddings
in NN’s latent space.
“Spaces that are almost like trees”
(here is an infinitely large triangle → )
Gromov hyperbolicity
(Other definitions are available, without need of geodesics)
Difference between Euclidean and δ-Hyperbolic spaces
(probabilistic proof, uses Johnson-Lindenstrauss to project to dimension m=logN and compares covering numbers set of all graphs on N vertices, with the one in R^m)
Therefore, Euclidean space embeddings of general graphs, and high-branching trees, is inefficient, and Hyperbolic space embeddings of high-branching trees is efficient.
These occur in social networks (paper) and in data with strong hierarchical content, such as semantic dependencies.
Hyperbolic Neural Networks
Main idea: use Hyperbolic space as a replacement for Euclidean space for embeddings
Consequence: we introduce hyperbolic analogues of vector space operations, and work on manifolds.
standard distance on Poincare disk.
Gyrovector spaces approach
(relativity) introduces several intuitive operations
NN operations lifted from tangent space
Non-commutativity: gyrations
Hyperbolic Neural Networks – more tools
Hyperbolic Neural Networks – more tools
Start in R^d
Hyperbolic Graph Neural Networks
Hyperbolic Graph Neural Networks
Hyperbolic Neural Networks – another viewpoint
Hyperbolic Neural Networks – Poincare latent space VAE
Example: learn random diffusion process outcome, via a Hyperbolic VAE (paper)
For the VAE, use Riemannian Normals or project normal from tangent space, loss according to Hyp. vol.
Hyperbolic Neural Network modules