Join at slido.com�#3715895
The Slido app must be installed on every computer you’re presenting from
Do not edit�How to change the design
3715895
Course Evaluation
The final course survey is intended to help us improve future offerings of the course.
�If you complete our internal anonymous survey AND the official Course Evaluation (http://course-evaluations.berkeley.edu) you will get 0.5% extra credit points to your grade. If 90% of the class also submits both, then everyone who submitted both will receive an additional 0.5% percentage point on their final grade.
�Internal course survey link: https://bit.ly/cs189-eval
3715895
Self-Supervised Learning
Lecture 24
Data understanding using self-supervised representation learning
EECS 189/289, Fall 2025 @ UC Berkeley
Joseph E. Gonzalez and Narges Norouzi
EECS 189/289, Fall 2025 @ UC Berkeley
Joseph E. Gonzalez and Narges Norouzi
Roadmap
3715895
Self-Supervised Learning
3715895
Supervised vs. Unsupervised Learning
Let's build methods that learn from "raw" data with no annotations required.
None of these represent how humans learn.
3715895
Self-Supervised Learning
3715895
Self-Supervised Learning Example
3715895
How?�Transfer Learning
3715895
Transfer Learning
3715895
Transfer Learning with CNN
Cat
Trained feature extractor
Linear classifier
3715895
Transfer Learning with CNN
Trained feature extractor
Linear classifier
3715895
Given a small labeled dataset for a new classification task, what is the most effective initial approach to leverage a pretrained CNN?
The Slido app must be installed on every computer you’re presenting from
Do not edit�How to change the design
3715895
Transfer Learning With CNN
3715895
Fine-tuning
Cat
3715895
Fine-tuning
Bakery
Initialize with pre-trained, then train with low learning rate
3715895
Why Pretext Learning Is Important
17
Goodfellow
3715895
Self-Supervised Learning Challenges
3715895
Self-Supervised Learning Approaches
3715895
Approaches
Generative: Predict part of the input signal
Discriminative: Predict something about the input signal
Multimodal: Use some additional signal in addition to RGB images.
3715895
Generative: Autoencoders
3715895
Autoencoders
Encoder
Decoder
Code/Feature
3715895
Autoencoder Idea
3715895
Convolutional Autoencoder
Information Bottleneck
|
|
|
|
|
|
|
|
3715895
Denoising Autoencoders
Encoder
Decoder
Latent space representation
Denoised Input
Noisy Input
3715895
Denoising Autoencoders
3715895
What are valid training objectives for a denoising auto encoder?
The Slido app must be installed on every computer you’re presenting from
Do not edit�How to change the design
3715895
0
0
3715895
Denoising Autoencoders - Process
Apply Noise
3715895
Denoising Autoencoders - Process
Encode And Decode
DAE
DAE
3715895
Denoising Autoencoders - Process
DAE
DAE
3715895
Denoising Autoencoders - Process
Compare
3715895
Demo
3715895
Generative: Colorization
3715895
Image Colorization
Sky is blue
Cloud is white
Mountain is green
3715895
Image Colorization
3715895
Image Colorization
3715895
Generative: Cross-Channel Prediction
3715895
Cross-Channel Prediction
3715895
Cross-Channel Prediction
3715895
Cross-Channel Prediction
3715895
Generative: Context-Encoders (Inpaiting)
3715895
Context Encoders
3715895
Context Encoders
3715895
Context Encoders
3715895
Context Encoders
Input image
3715895
Context Encoders
Input image
Joint loss
3715895
Context Encoders
Central region
Random block
Random region
3715895
Context Encoders
3715895
Generative: Image Super Resolution
3715895
Image Super-Resolution
3715895
Image Super-Resolution
3715895
Discriminative: Rotation
3715895
Image Rotation
3715895
Image Rotation
3715895
Image Rotation - Generalizability
56
3715895
ImageNet Top-1 Classification Results
57
With non-linear layers
3715895
Task Generalization
58
With linear layers
3715895
Relative Patch Position
3715895
Relative Patch Position
3715895
Relative Patch Position
3715895
Relative Patch Position
3715895
Discriminative: Jigsaw
3715895
Image Jigsaw Puzzle
3715895
Image Jigsaw Puzzle
3715895
Discriminative: Clustering
3715895
Deep Clustering
3715895
Deep Clustering
3715895
Contrastive Representation Learning
3715895
Contrastive Representation Learning
Same object
Different object
3715895
Formulation for Contrastive Learning
3715895
Contrastive Loss
This is similar to cross-entropy loss for a N-way softmax classifier.
3715895
SimCLR: A Simple Framework for Contrastive Learning
3715895
Momentum Contrastive Learning (MoCo)
Decouples negative sample size from minibatch size; allows large batch training without TPUs.
3715895
Contrastive Language Image Pre-Training (CLIP)
3715895
Self-Supervised Learning
Lecture 24
Reading: Partially covered in Chapter 19 of Bishop Deep Learning Textbook