Machine Learning for English Analysis
Week 7 – Unsupervised Learning
Prof. Seungtaek Choi
Today
Plan for Remaining Semester
| Week | |||||||
| W9 | W10 | W11 | W12 | W13 | W14 | W15 | W16 |
Assign#3 (project proposal) | start | | end | | | | | |
Assign#4 (data collection and analysis) | | start | | end | | | | |
Assign#5 (model training and evaluation) | | | start | | end | | | |
Assign#6 (real usage & final report) | | | | start | | end | | |
Assign#7 (presentation) | | | | | | lecture | | |
Lecture Topic | Unsup. Learning | Unsup.�Learning | Deep Learning | Deep Learning | Advanced�Topics | Advanced�Topics | X | Final Exam |
Announcement: Assignment #3
Midterm Explanation
Stats about Midterm
Midterm: OX
Midterm: OX
Midterm: OX
Midterm: Multiple-Choice
Midterm: Multiple-Choice
Midterm: Multiple-Choice
Midterm: Multiple-Choice
test data leakage
optimistic but …
counter: 85 vs. 15
cannot be sure without “cost”
Midterm: Multiple-Choice
covered in OX
covered in assignment2 (random init vs. word2vec init)
approximation with random sampling
Midterm: Multiple-Choice
dropout is for regularization (better generalization)
too strong regularization = low representation power
for p = 0, validation loss will increase (overfitting)
not guaranteed
Midterm: Multiple-Choice
n classifiers will multiply computing n-times
[0.1, 0.3, 0.4, 0.1, 0.1]
+ [0.2, 0.2, 0.3, 0.2, 0.1]
= [0.3, 0.5, 0.7, 0.3, 0.2]
🡪 class C
= 3 * e^2 * (1 – e) + e^3 = 3 * 0.2^2 * 0.8 + 0.2^3 = 0.104 (error) 🡪 0.896 (accuracy)
Midterm: Multiple-Choice
Scores
- Complete confusion matrix 🡪 4 points
- Correct accuracy 🡪 4 points
| Pred | |||
A | B | C | ||
True | A | 5 | 1 | 0 |
B | 0 | 4 | 2 | |
C | 2 | 0 | 4 | |
Accuracy = (# correct) / (# total) = 13 / 18
Midterm: Multiple-Choice
Scores
- Correct order 🡪 7 points
- Correct format 🡪 1 points (uppercase/lowercase, numbers, etc.)
A > C > B
#1: A
#2: C
#3: B
Model A | Pred | ||
P | N | ||
True | P | 160 | 40 |
N | 160 | 640 | |
Model B | Pred | ||
P | N | ||
True | P | 100 | 100 |
N | 25 | 775 | |
Model C | Pred | ||
P | N | ||
True | P | 120 | 80 |
N | 60 | 740 | |
Cost(A)
= 1 * 160 + 5 * 40
= 360
Cost(B)
= 1 * 25 + 5 * 100
= 525
Cost(C)
= 1 * 60 + 5 * 80
= 460
Midterm: Multiple-Choice
(i) [<PAD>, in, paris]
(ii) [in, paris, spring]
(iii) [paris, spring, <PAD>]
For more interactive understanding, please see this URL
https://claude.ai/public/artifacts/33cc6421-a2f0-46c8-b429-94a9af7fed00
Midterm: Multiple-Choice
Midterm: Multiple-Choice
Midterm: Multiple-Choice
ReLU = max(0, z) cannot return negative numbers
Midterm: Multiple-Choice
Scores
- Correct (a) = dog, dog 🡪 4 points
- Correct (b) = park, park 🡪 4 points
Unsupervised Learning
Clustering
Recap: Supervised vs. Unsupervised Learning
Recap: Supervised vs. Unsupervised Learning
Recap: Supervised vs. Unsupervised Learning
Supervised Learning | Unsupervised Learning |
Building a model from labeled data | Clustering from unlabeled data |
| |
| |
Data Clustering
Data Clustering: Similarity (~ Distance)
Basic Intuition
Basic Intuition
Basic Intuition
Basic Intuition
Basic Intuition
Basic Intuition
Basic Intuition
Basic Intuition
K-Means (Iterative) Algorithm
K-Means (Iterative) Algorithm
K-Means (Iterative) Algorithm
K-Means (Iterative) Algorithm
K-Means (Iterative) Algorithm
K-Means with Python
K-Means with Python
K-Means with Python
Example: K-Means on Word2Vec
Example: K-Means on Word2Vec
Some Issues of K-Means
Some Issues of K-Means
Some Issues of K-Means
Some Issues of K-Means
Limitations of K-Means (1/4)
Limitations of K-Means (2/4)
Limitations of K-Means (3/4)
Limitations of K-Means (4/4)