ML Systems Fail, Part II:
How to Manage Mistakes at Model Training
June 18th, 2023
Me
You
Recap
https://bit.ly/MLFailsOne
Interacting with ML Systems
Accepting the Perfect Imperfections
Feedback
Disclaimer!!!
Two Tasks:
Find the wrong text span
Suggest an appropriate replacement
A Trivial Dataset
Model-centric Approach
Setting Expectations
0.67
0.67
0.55
0.82
Training Many Models
Cross Validation
Training Many Models
Cross Validation
Training Many Models
Cross Validation
Training Many Models
Cross Validation
Training Many Models
Cross Validation
Training Many Models
Hyperparameter Tuning
Regularization
Preventing overfitting and improving generalizability.
Model Versioning
Tracking and managing changes to your model, hyperparameters, its evaluation results, and other related artifacts.
A Gold Standard
Today, Large Language Models seem to fit the role of a gold standard.
Summary
Data-centric Approach
Splits
Splits: Train, Validate, Test
Splits: Random Splitting… or Not
It works, but suffers when inter-record dependencies exist.
Splits: Random Splitting… or Not
Avoid:
Splits: Training on Validation Data
Splits: Summary
Slices
Slices: Not all inputs are equal.
Context is important. Some predictions have far greater consequences.
Slices: Subpopulations
Slices: Capabilities
Slices
Slices: Generating Slices
Slices: Summary
Questions??
Thank You🎈🎈🎈
Up Next
ML Systems Fail, Part III: How to Manage Mistakes while Planning Requirements