CS60050: Machine Learning
Sourangshu Bhattacharya
CSE, IIT Kharagpur
Course organization
Classroom: NR - 412
Slots:
Monday (8:00 – 9:55)
Tuesday (12:00 – 12:55)
Website:�https://sourangshu.github.io/IITKGP-ML-Spring-2026/ �
Moodle (for assignment submission):�https://moodlecse.iitkgp.ac.in/moodle/
Teaching Assistants
Saptarshi Mondal
Suman Kumar Bera
Vaishnovi Arun
Bibhudatta Bhanja
Dhyana Vardhan
Evaluation
Grades:
Midsem, Endsem: 50 - 60
Class Tests: 20 - 30
Term Project: 20 - 30
Class tests will be surprise ! If you miss it, you loose marks.
List of Term Projects will be share with you this week. You need to form groups of size 3 – 4 and select one project.
Both Term Project and assignment will require you to write code.
Assignments will not be graded.
Tentative Schedule (Changeable)
Week | Date-1 | Date-2 | Topic |
1 | 5/1/26 | 6/1/26 | Introduction classification |
2 | 12/1/26 | 13/1/26 | Linear models regression |
3 | 19/1/26 | 20/1/26 | SVM - kernel |
4 | 26/1/26 | 27/1/26 | Probabilistic ML Naive Bayes, Bayesian regression |
5 | 2/2/26 | 3/2/26 | DT, Bagging, Random forests |
6 | 9/2/26 | 10/2/26 | Boosting, Xgboost |
| 16/2/26 | 17/2/26 | Mid-sem |
| 23/2/26 | 24/2/26 | Mid-sem |
7 | 2/3/26 | 3/3/26 | Clustering GMM |
8 | 9/3/26 | 10/3/26 | Graphical Models |
9 | 16/3/26 | 17/3/26 | Neural Networks |
10 | 23/3/26 | 24/3/26 | Learning Theory and Metrics for problems |
11 | 30/3/26 | 31/3/26 | Id-ul-fitr |
12 | 6/4/26 | 7/4/26 | Active, Transfer, multi-task learning |
13 | 13/4/26 | 14/4/26 | Explainability and Trustworthiness |
COURSE BACKGROUND
Turing Test
image from http://en.wikipedia.org/wiki/Turing_test
Turing Test on Unsuspecting Judges
Turing Test
Turing Test
What is Artificial Intelligence
“[The automation of] activities that we associate with human thinking, activities such as decision making, problem solving, learning” (Bellman 1978)
“The study of mental faculties through the use of computational models” (Charniak & McDermott, 1985)
Good Old AI Days
Representing Knowledge
A Few Statements
Predicates
Rule Based Inference Example
(R1) if gas_in_engine and does not start, then problem(spark_plugs).
(R2) if not (does not start) and not (lights_on), then problem(battery).
(R3) if not(turns_over) and light_on, then problem(starter). (R4) if gas_in_tank and gas_in_carb, then gas_in_engine
Semantic Nets
Elephant
Africa
head
Nellie
Animal
Is a
Is a
Lives in
has
Hit the Wall
“The spirit is willing but the flesh is weak” becomes “The vodka is good but the meat is rotten”
AI Winter
Machine Learning
Data
Data
(10-15,15-20,20-25,...).
Data
Machine Learning
Emergence of Supervised Learning
Supervised Learning Human Perspective
Cats Dogs
Supervised Learning Human Perspective
Supervised Learning Human Perspective
Cats Dogs
How did a child know?
Supervised Learning Human Perspective
Cats Dogs
Eyes
Facial Features
Supervised Learning Model Perspective
Cats
Dogs
Train
Machine Learning Model
Supervised Learning Model Perspective
Training Dataset
Testing Dataset
Supervised Learning Model Perspective
Machine
Learning Model
Feature Extractor
Cats
Dogs
Feed
Predict
Problems with Data Annotation
Unsupervised Learning
When data is unlabelled
Unsupervised Learning
Type of learning where data is unlabelled/unknown and models learn these type of data for hidden patterns or data groupings.
Unsupervised Learning
Unsupervised Learning
Unsupervised Learning
Recommender Systems
Example of Unsupervised Learning
Cluster of Detective Novels
Those who bought this
also bought these
Semi Supervised Learning
Semi Supervised Learning
Books are present, know labels of some and not of others
Semi Supervised Learning
History Chemistry
We know the labels of these 2 books among other books
How can we get to know about the class of other books?
Semi Supervised Learning
History
Chemistry
We use Clustering Algorithm (Unsupervised Learning) to cluster similar books (based on content) and label them!
Semi Supervised Learning
History
Chemistry
We use Clustering Algorithm (Unsupervised Learning) to cluster similar books (based on content) and label them!
Semi Supervised Learning
History Chemistry
Model
Learned Parameters
We then use these labelled data and train our Model in a Supervised Learning Manner!
Types of Learning
Supervised Learning Unsupervised Learning Semi-Supervised Learning.
Ensemble Learning [combining models together] Self Supervised Learning [how we learn language] Reinforcement Learning [how robot learns]
Discriminative And Generative Tasks
Generative Task is a Discriminative Task!
Consider the well known sentence containing all the English Alphabets -
“The quick brown fox jumps over a lazy dog”
Suppose Given
“The quick brown
The model needs to generate the entire sentence
Generative Task is a Discriminative Task!
Task given “The quick brown” classify which of the following words will be the next word.
[fox, ox, tiger, ant, duck] — the model l classify “fox” as 1 and the rest as 0.
Recursively Given “The quick brown fox”, classify which word will be the next word — the model classify “jumps”
Given .“The quick brown fox jumps”, classify which word will be the next word — the model will classify “over” and so on….
We observe that if a series of such Prediction (or) Classification task (or) Discriminative Task are done, and each word predicted is appended with the phrase and run again for Prediction, we get a sentence.
Generative Task is a Discriminative Task!
Et Voila! We find that Generative Task is a Sequence of Discriminative Task!
Concept Learning Example
Version Spaces
Concept learning
Example / Instance: an atomic (real life) situation / object over which we want to learn.�
Instance space: Set of all possible instances.�
Attributes: observable quantities which describe a situation.�
Concept: a Boolean valued function over set of examples.�
Hypothesis space: subset of all Boolean valued functions over instance space.
Concept Learning - example
Attributes: Sky, Air temp, Humidity, Wind, Weather, Forecast.�
Instance space X. What is the size ?�
Hypothesis space: conjunction of literals ( which are conditions over attributes).
Conditions are of the form: (attr=val) or (attr=?) or (attr=φ)�
What is the size of hypothesis space ?
Concept Learning - example
Inductive learning problem
Training examples: D={ (x1,c(x1) , … , (xn,(c(xn)) }�
Problem: Given D, learn hϵH, such that for all xϵX , h(x)=c(x).�
Inductive learning assumption:
Any hypothesis found to approximate target concept well over sufficiently large training set, will also approximate it well over unseen examples.
General to specific ordering
Example x is said to be positive if c(x) = 1, else negative.�
Hypothesis h “satisfies” x, if h(x) = 1 .�
Hypothesis h2 is said to be “more general or equal to” h1 if�for all x: h1(x) = 1 implies h2(x) = 1
General to specific ordering
Find - S
Finding maximally specific hypothesis
Find – S Example
Find – S Problems
Can’t tell whether it has learned the concept�
Can’t tell whether the data is inconsistent�
Picks maximally specific hypothesis�
There might be several maximally specific hypothesis.
Version Space
Version space representation
Version space
Candidate Elimination
Candidate Elimination
If d is a negative example:
Example Problem
Example
Workout …
Convergence
Candidate elimination will converge to the target concept if:
Training data doesn’t have errors.
Target concept lies in the hypothesis space.
Otherwise
G and S sets become null.
Partially learned concept
What next training example ?
<Sunny, Warm, Normal, Light, Warm, Same>
Observations
The hypothesis space is biased.
Example: XOR concept cannot be expressed.
Unbiased learner – disjunction of conjunctions.
Learned Version space:
S set: all positive examples
G set: compliment of all negative examples
Can we use the partially learned concept from above ?
There is perfect ambiguity for all examples not in training set.
Unbiased learning
Learning in an unbiased hypothesis space is futile as it cannot generalize to examples other than training examples.
End of Slides