1 of 47

Deep Learning (DEEP-0001)�

Week 1 - Introduction

2 of 47

Book

http://udlbook.com

Slides based on the material from the book and the CM20315 course (University of Bath)

3 of 47

4 of 47

5 of 47

6 of 47

7 of 47

8 of 47

Supervised learning

  • Define a mapping from input to output
  • Learn this mapping from paired input/output data examples

9 of 47

What is a supervised learning model?

  • An equation relating input (age) to output (height)
  • Search through family of possible equations to find one that fits training data well

10 of 47

What is a supervised learning model?

  • Deep neural networks are just a very flexible family of equations
  • Fitting deep neural networks = “Deep Learning”

11 of 47

Regression

  • Univariate regression problem (one output, real value)
  • Fully connected network

12 of 47

Graph regression

  • Multivariate regression problem (>1 output, real value)
  • Graph neural network

13 of 47

Text classification

  • Binary classification problem (two discrete classes)
  • Transformer network

14 of 47

Music genre classification

  • Multiclass classification problem (discrete classes, >2 possible values)
  • Recurrent neural network (RNN)

15 of 47

Image classification

  • Multiclass classification problem (discrete classes, >2 possible classes)
  • Convolutional network

16 of 47

Image segmentation

  • Multivariate binary classification problem (many outputs, two discrete classes)
  • Convolutional encoder-decoder network

17 of 47

Depth estimation

  • Multivariate regression problem (many outputs, continuous)
  • Convolutional encoder-decoder network

18 of 47

Pose estimation

  • Multivariate regression problem (many outputs, continuous)
  • Convolutional encoder-decoder network

19 of 47

20 of 47

Translation

21 of 47

Image captioning

22 of 47

Image generation from text

23 of 47

What do these last three examples have in common?

  • Very complex relationship between input and output
  • Sometimes may be many possible valid answers
  • But outputs (and sometimes inputs) obey rules: can be learned without the need for output labels

Language obeys grammatical rules

Natural images also have “rules”

24 of 47

25 of 47

Unsupervised Learning

  • Learning about a dataset without labels
    • Clustering
    • Generating new examples
      • Filling in missing data

26 of 47

DeepCluster: Deep Clustering for Unsupervised Learning of Visual Features (Caron et al., 2018)

27 of 47

DeepCluster: Deep Clustering for Unsupervised Learning of Visual Features (Caron et al., 2018)

28 of 47

Unsupervised Learning

  • Learning about a dataset without labels
    • e.g., clustering
  • Generative models can create examples
    • e.g., generative adversarial networks

29 of 47

Unsupervised Learning

  • Learning about a dataset without labels
    • e.g., clustering
  • Generative models can create examples
    • e.g., generative adversarial networks
  • PGMs learn distribution over data
    • e.g., variational autoencoders,
    • e.g., normalizing flows,
    • e.g., diffusion models

30 of 47

Generative models

31 of 47

Latent variables

32 of 47

Interpolation

33 of 47

Conditional synthesis

34 of 47

35 of 47

36 of 47

Reinforcement learning

  • A set of states
  • A set of actions
  • A set of rewards

  • Goal: take actions to change the state so that you receive rewards

  • You don’t receive any data – you have to explore the environment yourself to gather data as you go

37 of 47

Example: chess

  • States are valid states of the chess board
  • Actions at a given time are valid possible moves
  • Positive rewards for taking pieces, negative rewards for losing them

38 of 47

Example: chess

  • States are valid states of the chess board
  • Actions at a given time are valid possible moves
  • Positive rewards for taking pieces, negative rewards for losing them

39 of 47

Why is this difficult?

  • Stochastic
    • Make the same move twice, the opponent might not do the same thing
    • Rewards also stochastic (opponent does or doesn’t take your piece)
  • Temporal credit assignment problem
    • Did we get the reward because of this move? Or because we made good tactical decisions somewhere in the past?
  • Exploration-exploitation trade-off
    • If we found a good opening, should we use this?
    • Or should we try other things, hoping for something better?

40 of 47

Landmarks in Deep Learning

  • 1958 Perceptron (Simple `neural’ model)
  • 1986 Backpropagation (Practical Deep Neural networks)
  • 1989 Convolutional networks (Supervised learning)
  • 2012 AlexNet Image classification (Supervised learning)
  • 2014 Generative adversarial networks (Unsupervised learning)
  • 2014 Deep Q-Learning -- Atari games (Reinforcement learning)
  • 2016 AlphaGo (Reinforcement learning)
  • 2017 Machine translation (Supervised learning)
  • 2019 Language models ((Un)supervised learning)
  • 2022 Dall-E2 Image synthesis from text prompts ((Un)supervised learning)
  • 2022 ChatGPT ((Un)supervised learning)
  • 2023 GPT4 Multimodal model ((Un)supervised learning)

41 of 47

2018 Turing award winners

42 of 47

43 of 47

This course

Deep neural networks

How to train them

How to measure their performance

How to make that performance better

44 of 47

This course

Networks specialized in images

Image classification, object detection

Image segmentation

Pose estimation

45 of 47

This course

Networks specialized to text

Text generation

Automatic translation

ChatGPT

46 of 47

This course

Generative learning (unsupervised)

Generating random cats!

47 of 47

This course

Play Atari games, AlphaGo

Robots that learn how to walk