1 of 27

Special Course in Machine Learning:

CS236 Deep Generative Models

Mikhail Papkov

11.02.2020

Slides available here: eid.ee/4yo

2 of 27

Agenda

  • Course organization
  • Introducing ourselves
  • Short discussion of Lecture 1 (Introduction and Background)
  • Kahoot quiz

3 of 27

Logistics

Delta-1022, Tuesday, 16:15 - 17:45 (we can stay longer or leave earlier)

Do we need to discuss changing time and place?

(Wednesday or Thursday evening are also fine)

4 of 27

Prerequisites

  • Probability theory, linear algebra, calculus
  • Python (we’ll most probably use NumPy and PyTorch)
  • MTAT.03.227 Machine Learning (lectures)

5 of 27

Resources

  1. courses.cs.ut.ee/2020/scml/spring
  2. piazza.com/ut.ee/spring2020/mtat03317 (invitations were sent)
  3. deepgenerativemodels.github.io
  4. deeplearningbook.org
    1. Chapter 3. Probability and Information Theory
    2. Chapter 20. Deep Generative Models
    3. Other chapters that you find useful
  5. Course videos (I hope so, sent an email to TA; meanwhile YouTube)

6 of 27

Organization

This is a reverse classroom seminar! I am also a student and present only today

  1. Presentation based on lecture material (see schedule) — 45 minutes
    • Preferably not only using published slides
    • You are welcome to prepare a lecture based on related external source (YouTube lecture)
    • I can help you to prepare (or at least try to)
  2. Solve and discuss HW task — 30 minutes
    • Solving every HW individually is not required (simply does not fit in 3 ECTS) but encouraged
  3. Quiz — 10 minutes
    • We’ll try Kahoot today and see how it goes
    • You can prepare your own homework instead (possibly based on CS236 HW) with deadline next week
  4. Discussion — some more minutes

7 of 27

How to pass?

  1. Prepare and conduct one seminar: lecture, HW, quiz (sign up here)
  2. Attend at least 9 times = solve 9 quizzes (60%)
  3. Read/watch material at home, participate in the discussion

Do you need more motivation?

8 of 27

Questions?

9 of 27

About me

  • Mikhail, 1st year PhD student in Health Image Analysis
  • Research in image quality enhancement (denoising, super-resolution)
  • Want to try applying GANs for cell microscopy images
    • Data generation
    • Domain adaptation

10 of 27

About you

(this would probably help us to define directions + I’ll try to remember you)

  • What do you do?
  • How could you apply generative models in your work?

(it’s fine if you do not yet have a clear perspective or take this course for fun!)

11 of 27

Background

12 of 27

Summary

  1. Tasks for generative models
  2. What is a distribution?
  3. Conditional independence
  4. Chain rule and Bayes rule
  5. Bayes network
  6. Naive Bayes (feature independence)
  7. Logistic regression
  8. Neural networks (more parameters)

13 of 27

Data

Hereinafter, CS236 materials were used (by Stefano Ermon and Aditya Grover, MIT License)

14 of 27

Generation

15 of 27

Tasks

16 of 27

Distribution

You kick the black box, it gives you some value according to certain rules:

  • Bernoulli (coin flip), Gaussian, uniform etc

Can be joint (RGB)

Are all the pixels (parameters) independent? We can assume so, but most likely not

17 of 27

Conditional independence

18 of 27

Rules

19 of 27

Bayes rule terminology

P(A) — prior

P(A|B) — posterior

P(B|A) — likelihood

20 of 27

Bayes network

We can assume that some events are conditionally independent

Figure from Meelis Kull: Fall 2019 MTAT.03.227 Machine Learning Lecture 11, slide 32

21 of 27

Naive Bayes

Assume features to be conditionally independent given label

22 of 27

Generative vs Discriminative

Naive Bayes is a generative model (although we use it for classification)

23 of 27

Logistic regression (discriminative)

Parametrized with α

24 of 27

Logistic regression (discriminative)

Parametrized with α

25 of 27

From logistic regression to neural network

26 of 27

Why using generative models?

27 of 27

Kahoot!