How accurate are Open Phil's predictions?
Javier Prieto
GPI Workshop on Forecasting Existential Risks and the Long-term Future
(June 2022)
Outline
User interface (a table in a Google doc)
Predictions | Scoring (you can leave this blank until you're able to score) | |||
With X% confidence, | I predict that…(yes/no or confidence interval prediction) | …by Y time (ideally a date, not e.g. "in 1 year"): | Score (please stick to True / False / Not Assessed) | Comments or caveats about your score |
80% | GPI will hire at least 2 postdocs | 2022/7/1 | | |
60% | GPI will publish at least 3 papers about population ethics in peer-reviewed journals | 2022/7/1 | | |
40% | GPI will run a workshop on forecasting x-risk | 2022/7/1 | True | I’m here! |
Why make predictions?
Calibration
Accuracy
Brier Score = E(P[Y] - Y)2
= E(P[Y|p] - p)2 - E(P[Y|p] - P[Y])2 + P[Y](1 - P[Y])
= Miscalibration - Resolution + Entropy
= 0.004 - 0.037 + 0.250
= 0.217
Caveats and sources of bias
Closing thoughts
Thank you!
Questions?
☝️EA Forum post☝️
Appendix: Prediction-level Brier scores
¼ of our predictions are worse than chance
Half of our predictions have an edge of ≥10%
¼ of our predictions have an edge of ≥25%
Q1 = .252 = .0625
Q2 = .42 = .16
Q3 = .52 = .25
We don’t seem to be getting more accurate over time
Predictions cluster around “round numbers” + 25% & 75%
Appendix: Range distribution (all predictions)
Appendix: Accuracy vs Range (scored predictions only)