1 of 41

The Six Families of�Deep Generative Models

Yannis Pantazis

Journal Club on AI @ FORTH

Friday, October 18th

2 of 41

Logistics

  • Biweekly meetings – Friday @ 14:00
  • Location: Vassilis Dougalis Room
  • JC list: jc-list@iacm.forth.gr
  • Email Panos Evaggelidakis for subscribing το the JC list
    • --> panosevangelidakis@gmail.com

3 of 41

November’s Presentations

Ammar Qammaz

Accessible AI,Easily locally available State-of-the-Art Methods for VLMs, LLMs, Image, Voice and Music Synthesis

1/11/2024

Gregory Tsagkatakis

Deep Learning for Inverse Imaging Problem

15/11/2024

Yiannis Kamarianakis

Statistical Models for Spatial, Spatiotemporal and 4D data

29/11/2024

4 of 41

Introduction – What is a Generative Model?

What is required:

    • Families of Generative Models
    • Algorithms to train these GMs
    • Neural network architectures
    • Loss functions & distances between probability density functions

5 of 41

Families of Generative Models – Taxonomy based on Likelihood Function

Planar

Coupling

MAFs/IAFs

(R)NADE

WaveNet

WaveRNN

GPT

Vanilla

β-VAE

VQ-VAE

diffusion

denoising

score

Belief nets

Boltzmann

machines

 

GMs

Exact

ARMs

NFs

VAEs

EBMs

DPMs

Approximate

Implicit

GANs

6 of 41

AutoRegressive Models (ARMs)

7 of 41

AutoRegressive Models (ARMs)

 

 

  • Softmax when discrete
  • Gaussian when continuous

Empirical observation:

  • Neural networks perform better when the output is discrete.
  • Thus, even when data are continuous, they are often quantized and treated as discrete variables.
  • Current trend: Tokenize everything

8 of 41

AutoRegressive Models (ARMs)

 

  • Dilated convolutions
  • Recurrent Architectures
  • Transformers (decoder only)

 

9 of 41

AutoRegressive Models (ARMs)

1542M

762M

345M

117M parameters

GPT released June 2018

GPT-2 released Nov. 2019 with 1.5B parameters

GPT-3: 175B parameters trained on 45TB texts

10 of 41

Normalizing Flows (NFs)

Many small steps adds up

to big results.

11 of 41

Normalizing Flows (NFs)

 

12 of 41

Normalizing Flows (NFs)

13 of 41

Normalizing Flows (NFs) – RealNVP 2016

14 of 41

Variational Autoencoders (VAEs) - Motivation

15 of 41

Variational Autoencoders (VAEs) - Motivation

observed

latent/hidden

16 of 41

Variational Autoencoders (VAEs)

z

17 of 41

Variational Autoencoders (VAEs)

18 of 41

Variational Autoencoders (VAEs)

19 of 41

Variational Autoencoders (VAEs)

20 of 41

Variational Autoencoders (VAEs)

21 of 41

Variational Autoencoders (VAEs)

  • Training VAEs requires:
    • Approximate the model evidence with a lower bound called ELBO (from Evidence Lower BOund) and maximize ELBO instead of the evidence.
    • Reparametrization trick for efficient gradient estimation.

  • Typically, the latent variable is continuous.�However, there are extensions to discrete latent variables (Vector Quantized VAE – VQ-VAE).

  • Latent variables can be disentangled (β-VAE and InfoVAE)

22 of 41

Energy-based Models (EBMs)

23 of 41

Energy-based Models (EBMs)

  • The first family of Deep Generative Models!
    • Inspired by Statistical Physics and Boltzmann distribution

  • Interesting algorithms for training have been proposed

    • Contrastive divergence

    • Score function

    • Noise contrastive estimation

24 of 41

Energy-based Models (EBMs) – Product of Experts

25 of 41

Generative Adversarial Networks (GANs)

  • Main Idea: Instead of using KLD minimization (ie, log-likelihood maximization) use a different “distance” to minimize.

26 of 41

Generative Adversarial Networks (GANs)

27 of 41

Generative Adversarial Networks (GANs)

28 of 41

Generative Adversarial Networks (GANs)

29 of 41

Generative Adversarial Networks (GANs)

30 of 41

Generative Adversarial Networks (GANs) – Cycle GAN 2017 – Unpaired Matching

31 of 41

Generative Adversarial Networks (GANs) – Cycle GAN 2017 – Unpaired Matching

32 of 41

Diffusion Probabilistic Models (DPMs)

33 of 41

Diffusion Probabilistic Models (DPMs)

34 of 41

Diffusion Probabilistic Models (DPMs)

35 of 41

Diffusion Probabilistic Models (DPMs)

36 of 41

Diffusion Probabilistic Models (DPMs)

Simplified training objective:

37 of 41

Diffusion Probabilistic Models (DPMs)

38 of 41

Diffusion Probabilistic Models (DPMs)

Palette: Image-to-Image Translation

39 of 41

Challenges in Generative Models

40 of 41

Challenges in Generative Models

  • How to evaluate their performance?
    • Likelihood-based
    • Inception score for images
    • But nothing conclusive and rigorous

  • Avoid “spreading mis-information”
    • On the usage of generative models - Ethics
    • On their reliability

41 of 41

Reading on Generative Models

Introduction to Deep GMs Course

https://www.csd.uoc.gr/~hy673/