Probabilistic Foundations of Machine Learning
I am: Yaniv / Professor Yacoby
I use: he/they
Announcements
Today: Optimization
So far, we:
Next, we want to perform the MLE: but
Today:
Our generative process
in code.
Performing the MLE:
Checking for Convergence
Our goal: to minimize our “loss”
How can we minimize a function? Let’s get some intuition:
Idea: look for places where the derivative of the loss is 0.
Global vs. Local Optima
Looking at figure,
Answer: Yes. How?
Initial Algorithm
…shall we try it?
Analytic MLE, Step 1: What’s our model?
Analytic MLE, Step 2: What’s our joint data likelihood?
Analytic MLE, Step 3: What’s our loss function?
Analytic MLE, Step 4: Minimize our loss
Pros and Cons of Analytic MLE
Pros:
Cons:
Alternative: Use Numeric Optimization Algorithm
Gradient Descent
Simulations (see chapter)
Gradient Descent in Jax
Note: you should not implement/use this. Use the code we provided.
Challenges with Numeric Optimization
Constraining Parameters to Valid Ranges
In pairs: continue implementing the IHH-ER model in NumPyro!
That’s all for today!
Questions?