Potential Existential Risks from Artificial Intelligence
Ben Garfinkel�Future of Humanity Institute (University of Oxford)
Plan for the talk
Background
What is AI?
History of the field
Some present-day capabilities
Some present-day capabilities
Some present-day capabilities
Some present-day capabilities
Some present-day capabilities
Future developments
Where is the field headed?
Human-substitute AI (or “Human-level AI”)
Studying a world with HSAI
Superintelligence
Major uncertainties: Specialization
Major uncertainties: Continuity
Potential risks
Potential risks
Diversity in existential risk concerns
Potential risks: Instability
Instability
Potential risks: Bad attractor state
Bad attractor state
Technological determinism
| Forager societies | Farmer societies | Industrialized societies |
Political institutions | Consensus-based | Hierarchical, rarely democratic | Hierarchical, often democratic |
Living standards | Low, stagnant | Low, stagnant | High, rising |
Forced labor | Rare | Common | Rare |
Inequality | Low/medium | High | High |
Gender roles | Somewhat distinct (probably) | Highly distinct | Somewhat distinct |
Interpersonal violence | High (probably) | Moderate | Low |
Assessing previous major economic revolutions
Assessing the AI Revolution: Concerns about humans
Assessing the AI Revolution: Concerns about AI systems
Conditional technological determinism
Potential risks: Lock-in of contingencies
Lock-in concerns
The “choosing our successors” argument
Potential risks: Technical failure
Technical failure
Two kinds of failure: Incompetence
Two kinds of failure: Misalignment
Explaining misalignment: one story
Existential misalignment risk
Diversity in alignment risk concerns
Conclusions
Summing up
Summing up