Class 8: The risks of generative AI
DPI-681M
The Science and Implications of Generative AI
Profs. Goel, Levy, and Svoronos
Harvard Kennedy School
A taxonomy of risks
There are four broad categories of risks from generative AI:
Limitations and misuse
Large language models learn to produce language by digesting everything on the internet and beyond. That doesn’t always lead to good outcomes!
Misinformation
Misinformation
Using PingPong, try to get GPT-4 to return incorrect information.
🚴 Activity 🚴
Misuse
Misuse
Fine-tuning LLMs
To mitigate misinformation and misuse, large language models are fine-tuned to align with desired behavior. �[ We align using demonstration and comparison data. ]
Society-wide disruptions
Society-wide disruptions
Existential risks
Though there is much disagreement on the topic, many researchers and policymakers are concerned with “existential risks” from AI [ i.e., significant loss of life ].
Such risk may stem from malicious individuals using powerful AI to cause damage, or from AI that acts in ways that are not “aligned” with human interests.
“Suppose we have an AI whose only goal is to make as many paper clips as possible.”�[ Nick Bostrom ]
What could go wrong?
“Suppose we have an AI whose only goal is to make as many paper clips as possible.
The AI will realize quickly that it would be much better if there were no humans because humans might decide to switch it off. Because if humans do so, there would be fewer paper clips.
Also, human bodies contain a lot of atoms that could be made into paper clips.
The future that the AI would be trying to gear towards would be one in which there were a lot of paper clips but no humans.”�[ Nick Bostrom ]
Existential risks
AGI is artificial general intelligence
AGI
No AGI
Harm
No harm
Harm
No harm
Generative AI and elections
Thinking back to our discussion of elections in the last class, what are some risks in each of these categories?�
II. Mitigation strategies
Mitigating misinformation
II. Mitigation strategies
The creation and dissemination of misinformation was cited by a majority of people in the class as an immediate or medium-term risk of generative AI.
Mitigating misinformation
“Comprehensive legislation and attendant regulations that not only prohibits misinformation but requires disclosures and reporting on the use of generative AI in any political campaign or respecting any political content.”
Mitigating misinformation
“Enforce watermark or watermark-like requirements (embedded patterns, metadata) for any LLM provider.”
Mitigating misinformation
“A government-sponsored deep fake detector.”
Mitigating misinformation
“Implementing Robust Fact-Checking and Content Verification Systems. This strategy aims to combat misinformation by developing and integrating advanced verification systems to authenticate AI-generated content before it's widely shared.”
Mitigating misinformation
“Implementing comprehensive digital literacy programs that include education on generative AI capabilities and how to critically assess digital content.”
Regulatory levers
Guha et al. identify four commonly proposed regulatory approaches to mitigating AI risks:
Regulatory levers
Guha et al. identify four commonly proposed regulatory approaches to mitigating AI risks:
Regulatory levers
In assessing proposed policies, we must consider:
Mitigating misinformation
How effective, feasible, and supported do you think these proposed strategies are? [ Discuss with your neighbor. ]
III. Case study: AI for healthcare
Potential Benefits of GenAI in healthcare
III. Case study: AI for healthcare
Source: https://www.economist.com/technology-quarterly/2024-03-30
Potential Benefits of GenAI in healthcare
III. Case study: AI for healthcare
Source: https://www.economist.com/technology-quarterly/2024-03-30
Potential Benefits of GenAI in healthcare
III. Case study: AI for healthcare
Source: https://www.economist.com/technology-quarterly/2024-03-30
Potential Benefits of GenAI in healthcare
III. Case study: AI for healthcare
Source: https://www.economist.com/technology-quarterly/2024-03-30
Potential Benefits of GenAI in healthcare
III. Case study: AI for healthcare
Source: https://www.economist.com/technology-quarterly/2024-03-30
Potential risks of GenAI in healthcare
Known limitations: | Misuse: |
Society-wide disruptions: | Existential risks: |
Potential risks of GenAI in healthcare
Known limitations:
| Misuse:
|
Society-wide disruptions:
| Existential risks:
|
In your group, discuss possible mitigation strategies for one of these two risks:
Come up 2-3 concrete, specific proposals to mitigate these potential risks.
Keep in mind the regulatory alignment problem when discussing!
🚴 Activity 🚴
Mitigation strategies for Inaccuracy and bias
Mitigation strategies for Overreliance
Key Takeaways
Takeaways
Key Takeaways