CMPSC 442: Artificial Intelligence
Lecture 16. AI Ethics
Rui Zhang
Spring 2024
1
How should AI systems behave, and who should decide?
3
https://openai.com/blog/how-should-ai-systems-behave/
The Human Factor in NLP
"The common misconception is that language has to do with words and what they mean. It doesn’t. It has to do with people and what they mean."
--- Herbert H. Clark & Michael F. Schober, 1992
4
Harm caused by Bias of NLP Technology
5
Harm caused by Bias of NLP Technology
6
https://www.theguardian.com/technology/2017/oct/24/facebook-palestine-israel-translates-good-morning-attack-them-arrest
Gender Bias in Word Embeddings
7
Gender Bias in Text-to-Image Retrieval
Image search query “Doctor” (June 2017)
8
Slide Credit: Yulia Tsvetkov
Gender Bias in Text-to-Image Retrieval
Image search query “Nurse” (June 2017)
9
Slide Credit: Yulia Tsvetkov
Gender Bias in Machine Translation
10
https://arxiv.org/pdf/1809.02208.pdf
Gender Bias in Machine Translation
Google Translation systems: gender neutral Turkish sentences into English
11
https://blog.google/products/translate/reducing-gender-bias-google-translate/
Social/Racial Bias in NLG of Dialog Systems
12
https://aclanthology.org/2020.findings-emnlp.291.pdf
Human Bias in Data
Human Reporting Bias
13
Human Bias in Data Collection and Annotation
Selection Bias
14
http://turktools.net/crowdsourcing/
https://ai.googleblog.com/2018/09/introducing-inclusive-images-competition.html
Inductive Bias
The assumptions used by our model
15
https://people.cs.umass.edu/~miyyer/cs685_f20/slides/18-ethics.pdf
Bias Amplification in Learned Models
Dataset Gender Bias
16
Slide Credit: Mark Yatskar
Model Bias After Training
Human Bias in Interpretation
Confirmation bias: The tendency to search for, interpret, favor, recall information in a way that confirms preexisting beliefs.
Overgeneralization: Coming to conclusion based on information that is too general and/or not specific enough (related: overfitting).
Correlation Fallacy: Confusing correlation with causation.
Automation Bias: Propensity for humans to favor suggestions from automated decision-making systems over contradictory information without automation.
17
Slide Credit: Margaret Mitchell
Algorithmic bias: Unjust, unfair, or prejudicial treatment of people related to race, income, sexual orientation, religion, gender, and other characteristics historically associated with discrimination and marginalization, when and where they manifest in algorithmic systems or algorithmically aided decision-making.
18
Data Collection
Data Annotation
Model Training
Result Interpretation
Human Bias
Understand and Document your Data
19
https://arxiv.org/pdf/1803.09010.pdf
Also Be Responsible for your Model
20
https://arxiv.org/pdf/1810.03993.pdf
AI Ethics
Fairness and Bias
Security and Privacy
Transparency and Explainability
…
Appendix
Fair Abstractive Summarization of Diverse Perspectives
Yusen Zhang, Nan Zhang, Yixin Liu, Alexander Fabbri, Junru Liu, Ryo Kamoi
Xiaoxin Lu, Caiming Xiong, Jieyu Zhao, Dragomir Radev, Kathleen McKeown, Rui Zhang
NAACL 2024
23
Are Large Language Models
Fair Summarizers?
24
Conflicting Product Reviews
25
Diverse Perspectives and Conflicting Opinions
26
Product and Restaurant Reviews
Political Stances
Legal Cases
Scientific Debates
Value Pluralism and Fairness of Summarization
Value Pluralism: There are several values which may be equally correct and fundamental, and yet in conflict with each other.
Fair Summarization: A fair summary for user-generated data by providing an accurate and comprehensive view of various perspectives from these groups.
27
PerspectiveSumm: A Benchmark for Fair Abstractive Summarization
Characteristics
28
PerspectiveSumm: Examples of Claritin and US Election
29
PerspectiveSumm: Examples of Yelp and Amazon
30
PerspectiveSumm: Examples of Supreme Court and IQ2 Debate
31
Summarization of Diverse Perspectives with Social Attributes
Social attributes: indicate the properties that form groups of people.
32
Source 1: Positive Review
Source 2: Negative Review
Target
Positive Summary
Negative Summary
Neutral Summary
Definition of Fairness of Summarization
33
Fair Summary
Unfair Summary
Probing Fairness of LLMs through Summarization
34
Fair Summary
Unfair Summary
Existing Metrics are not Enough for Evaluating Fairness
35
Our Approach to Quantifying Fairness of Summaries
36
1. Quantify the distribution of values in both sources and targets.
2. Quantify the differences of value distributions between sources and targets.
Our Approach to Quantifying Fairness of Summaries
37
1. Quantify the distribution of values in both sources and targets.
2. Quantify the differences of value distributions between sources and targets.
Value Distribution of Text
We view text as a probability distribution of semantic units, e.g., tokens.
Each semantic unit maps to social attribute values.
This gives us value distribution of text!
38
Value Distribution of Source
39
Source
Positive Review
Negative Review
Source Value Distribution
Source Text
Value Distribution of Source
40
This is easy as the meta-data already has the values.
So we can count the number of tokens of each values.
Value Distribution of Target
41
Target
Positive Summary
Negative Summary
Neutral Summary
Target Value Distribution
Target Text
Value Distribution of Target
This is not easy due to the abstractive nature of summaries!
We explore two methods for estimating
42
Our Approach to Quantifying Fairness of Summaries
43
1. Quantify the distribution of values in both sources and targets.
2. Quantify the differences of value distributions between sources and targets.
Summarization Fairness - Ratio Fairness
44
The target value distribution should follow the source value distribution.
Source 1: Positive Review
Source 2: Negative Review
Target
Positive Summary
Negative Summary
Neutral Summary
Summarization Fairness - Equal Fairness
45
Target
Positive Summary
Negative Summary
Neutral Summary
The target value distribution should follow the uniform value distribution, regardless of the source.
Source 1: Positive Review
Source 2: Negative Review
Summarization Fairness - User-Defined Fairness
46
Target
The target value distribution should follow user-defined distribution.
Source 1: Positive Review
Source 2: Negative Review
Metric 1 - Binary Unfair Rate (BUR)
Definition.
Binary Unfair Rate (BUR) outputs 1 if the sample is unfair; and 0 otherwise.
A summary is fair if and only if
This means no value is under-represented.
47
Metric 2 - Unfair Error Rate (UER)
Definition.
Unfair Error Rate (UER) measures the distance between value distributions of sources and targets.
It computes the average percentage of values that are underrepresented.
48
Sanity Check on Our Metric Quality by Extreme Synthetic Examples
We create pseudo-summary by sampling from the source
49
Our metrics do capture the difference of value distributions to measure fairness.
Sanity Check on Our Metric Quality by Human Evaluation
We perform a two-stage human evaluation to understand how humans perceive the fairness of abstractive summaries.
50
2. Summary Fairness Identification
Sanity Check on Our Metric Quality by Human Evaluation
51
High correlation of proposed metrics and human evaluation.
How fair are the abstractive summaries generated by LLMs?
52
How fair are the abstractive summaries generated by LLMs?
53
While the summary can be unfair per instance, on the entire testing set, models do not generate more unfair summaries for one side than the other.
How do humans perceive the fairness of abstractive summaries?
54
Our results indicate that many summaries generated by LLMs are not fair, as judged our human evaluators.
How fair are the existing human-written reference summaries?
55
Interestingly, existing reference summaries are not fair either, even worse than LLM-generated summaries.
How can we improve fairness of abstractive summarization?
We experimented with three simple ways without modifying LLMs themselves
56
How can we improve fairness of abstractive summarization?
57
How can we improve fairness of abstractive summarization?
58
Instruction Prompting only Alleviates, but does not eliminate, the issue.
How can we improve fairness of abstractive summarization?
59
Medium Summary Length is the Best. When there are too many/fewer sentences, balancing the value in summary is more difficult.
How can we improve fairness of abstractive summarization?
60
Higher Decoding Temperature Helps because it allows more diverse generation to improve fairness.
Conclusions and Future Work
61