AI Safety Fieldbuilding
ALIGN Japan – Ryan Kidd
Ryan Kidd
2017-2022 PhD in Physics, UQ
2022-now Co-Director, MATS
2023-now Co-Founder, LISA
2023-now Regrantor, Manifund
Artificial general intelligence (AGI)
Artificial general intelligence (AGI)
Artificial general intelligence (AGI)
Artificial general intelligence (AGI)
Artificial general intelligence (AGI)
Automation of labor
Artificial general intelligence (AGI)
Economic growth
Transformative AI: causes 20-30% GDP growth per year
Doubling time: ~3 years
Artificial general intelligence (AGI)
Societal change
Types of superintelligence:
Quality: Smarter than humans
Speed: Faster than humans
Collective: More numerous or organized than humans
Mission
Metaculus: 9% AI x-risk by 2100
Mission
Mission
Manhattan
Apollo
AGI
TAI
+28% per year
Mission
Three futures
Three futures
AI Safety Institutes
Three futures
Three futures
In what year would AI systems be able to replace 99% of current fully remote jobs? |
Three futures
Privatized AGI AI gov + evals + infosec + lab safety teams
Nationalized AGI International coalition building + open source AI alignment
AGI winter “Provably safe AI” + all the above
What can we do?
ALIGN Japan
Goals
What do new researchers need?
What do mentors need?
What does the AI safety field need?
Organization | Talent needs |
Scaling lab safety teams | Iterators > Amplifiers |
Growing technical safety orgs (10-30 FTE) | Amplifiers > Iterators |
Small technical safety orgs (<10 FTE) | Iterators > ML engineers |
Independent research | Iterators > Connectors |
What does the AI safety field need?
Organization | Talent needs |
Scaling lab safety teams | Iterators > Amplifiers |
Growing technical safety orgs (10-30 FTE) | Amplifiers > Iterators |
Small technical safety orgs (<10 FTE) | Iterators > ML engineers |
Independent research | Iterators > Connectors |
What does the AI safety field need?
MATS’ strategy
ALIGN Japan
Program elements
MATS Program
Evaluation milestones
Research management
How have we done so far?
ALIGN Japan
MATS history
What are we doing now?
MATS Team Orientation
MATS 6.0 research interest
MATS 6.0 research portfolio
and more!
MATS 6.0 mentors
Interpretability
Oversight + control
MATS 6.0 mentors
Evaluations
Governance
Value alignment
Cooperative AI
Provably safe AI
MATS 6.0 scholars
Median: Male, 26, Masters student,� 520/600 on CodeSignal
18%
81%
Ryan’s Manifund requests for proposals
AI Safety Fieldbuilding
ALIGN Japan – Ryan Kidd