CS365�Foundations of Data Science
Lecture 8 (2/15)
Charalampos E. Tsourakakis �ctsourak@bu.edu
Reminder
The Thumbtack problem
Heads
The Thumbtack problem
Assumptions
The Thumbtack problem
Maximum Likelihood Estimate (MLE)
Maximum Likelihood Estimate (MLE)
p* value
Maximum Likelihood Estimate (MLE)
Optimize to find p*
Almost done: ��- We need to verify that we get a maximum for pMLE��(whiteboard)�
Maximum Likelihood Estimate (MLE)
Confidence intervals (reminder)
Sampling theorem: Given n independent 0-1 RVs Xi such that Pr(Xi=1)=p (i=1…n), ��where then the following holds: ��
Two important properties of the MLE
Billionaire with prior beliefs
p now becomes a random variable
Inference using Bayes’ rule
���������
Does not depend on p.
Notice the notation Pr(D|p) instead of Pr(D;p)
Bayesian inference - Summary
Maximum a posteriori (MAP) estimate is the mode of the posterior distribution (as we did in this lecture).
Important observation
Conjugate priors
Definition ��“ If the posterior distribution p(θ | x) is in the same probability distribution family as the prior probability distribution p(θ), the prior and posterior are then called conjugate distributions, and the prior is called a conjugate prior for the likelihood function p(x | θ).”��- For Binomial, conjugate prior is Beta distribution.
Bayesian inference for the thumbtack problem
Where is the gamma function �� �
Bayesian inference for the thumbtack problem
��
Posterior has same form
as prior!
Method of moments
Suppose our model has parameters θ=(θ1,..,θk).
Method of moments: example 1
Let X1,...,Xn~Bernoulli(p).
Remark: Here, MoM is same as MLE, but this is not always the case!
Method of moments: example 2
MLE for Gaussian
Readings