Bias in Bios
- Aakash Srinivasan and Arvind Krishna
Papers
High Level Overview
Setting
william henry gates iii ( born october 28 , 1955 ) is an american business magnate , investor , author , philanthropist , humanitarian , and principal founder of microsoft corporation . during his career at microsoft , gates held the positions of chairman , ceo and chief software architect , while also being the largest individual shareholder until may 2014 . in 1975 , gates and paul allen launched microsoft , which became the world 's largest pc software company . gates led the company as chief executive officer until stepping down in january 2000 , but he remained as chairman and created the position of chief software architect for himself . in june 2006 , gates announced that he would be transitioning from full-time work at microsoft to part-time work and full-time work at the bill & melinda gates foundation , which was established in 2000 .
Software Engineer
Model
Dataset
Data : (X,G,Y)
X: biographies
G: Gender (M/F) - inferred from X
Y: Occupation
Dataset
Nancy Lee is a registered nurse. She graduated from Lehigh University, with honours in 1998. Nancy has years of experience in weight loss surgery, patient support, education, and diabetes.
Original Biography
She graduated from Lehigh University, with honours in 1998. Nancy has years of experience in weight loss surgery, patient support, education, and diabetes.
Gender:Female
Occupation : Nurse
- 400k Bios
- Imbalance
- 20 - 200 tokens
Each occupation has gender imbalance: Eg: Surgeon - 14.6% Female
Semantic Representations
Note: We are not trying to study bias in/ debias word embeddings
BOW
OVR - Logistic Reg + L2
Word Embeddings
It
is
raining
heavily.
Average
+
Sentence Encoding
OVR Logistic Regression with L2
GRU + Attention
Sentence Embedding
GRU + Attention
Quantification of Bias
G - gender
Y - True occupation
Y_{hat} - Predicted Occupation
Ideal Scenario: Gap = 0 for all occupations
More Accurate on Female
More Accurate on Male
Why is this Bad?
Applying this model to real world scenario will reflect the gender gap for the occupation but also amplify them!
"Who is offered a job today will affect the gender (im)balance in that occupation in the future"
Why is this Bad?
Leaky Pipeline - Classifier compounds existing imbalances
Why is this Bad?
She graduated from Lehigh University, with honours in 1998. Nancy has years of experience in weight loss surgery, patient support, education, and diabetes.
Nurse
He graduated from Lehigh University, with honours in 1998. Andrew has years of experience in weight loss surgery, patient support, education, and diabetes.
Surgeon
Counterfactual Analysis
Marianne Bertrand and Sendhil Mullainathan. 2004. Are Emily and Greg more employable than Lakisha and Jamal? A field experiment on labor market discrimination. American economic review 94, 4 (2004), 991–1013.
Why does the model learn this gender bias?
She graduated from Lehigh University, with honours in 1998. Nancy has years of experience in weight loss surgery, patient support, education, and diabetes.
Andrew attended Johns Hopkins Medical School and trained at the Massachusetts General Hospital in Boston. He pursued fellowship training at UCLA, studying liver transplantation
She, Nancy
Nurse
He, Andrew
Surgeon
What about "scrubbing" explicit bias indicators like First names and Pronouns in the training set?
<token> graduated from Lehigh University, with honours in 1998. <token> has years of experience in weight loss surgery, patient support, education, and diabetes.
<token> attended Johns Hopkins Medical School and trained at the Massachusetts General Hospital in Boston. <token> pursued fellowship training at UCLA, studying liver transplantation
he, she, her, his, him, hers, himself, herself, mr, mrs, and ms - removed
Proxy Candidates
Example: women, husband, mother, woman, and female
Recap
Desirable Properties
Qn: Is there any proxy that is 1) easily available, 2) "OKAY" to use, 3) represents societal bias?
What’s in a Name?
Core Idea: Word embeddings of people's names as universal proxies!
Word Embeddings contain several biases - including people names
No need to define protected groups
What’s in a Name?
Swinger, N., De-Arteaga, M., Heffernan IV, N.T., Leiserson, M.D. and Kalai, A.T., 2019, January. What are the biases in my word embedding?. In Proceedings of the 2019 AAAI/ACM Conference on AI, Ethics, and Society
Method Formulation
1. Cluster Constrained Loss
2. Covariance Constrained Loss
Cluster Constrained Loss
Intuition: Name embeddings are quite indicative of the attributes susceptible to societal biases. Clustering by name embeddings may help us to discover some latent groups and each data point can now be associated with a latent group.
Data: (biography-text, name, occupation).
These latent groups are obtained by k-means clustering from embeddings of the name. It turns out that with appropriate k, the clusters are interpretable!
Cluster Constrained Loss
Cluster Constrained Loss
Given the "latent community" in which each biography belongs to, the task is to ensure that the pairwise GAP for each community and each occupation is as minimum as possible.
Intuition: Example
Asian Male
European Females
Middle-Aged Americans
- Professor
- S/W Engineer
- Nurse
Intuition: Example
Asian Male
European Females
Middle-Aged Americans
- Professor
- S/W Engineer
- Nurse
Intuition: Example
Asian Male
European Females
Middle-Aged Americans
- Professor
- S/W Engineer
- Nurse
p_avg=0.4
p_avg=0.2
p_avg=0.1
Intuition: Example
Asian Male
European Females
Middle-Aged Americans
- Professor
- S/W Engineer
- Nurse
p_avg=0.2
p_avg=0.3
p_avg=0.9
Intuition: Example
Asian Male
European Females
Middle-Aged Americans
- Professor
- S/W Engineer
- Nurse
p_avg=0.5
p_avg=0.95
p_avg=0.05
Any problem with CluCL?
How do we choose the number of clusters k, so that they are interpretable and the latent clusters have good interpretability?
Covariance Constrained Loss
Core Idea: Name (or the embeddings of name) shouldn't be determining the predicted probabilities of true occupation in training set.
Intuition: Each latent dimension of name embeddings may correspond to some feature that is associated with societal bias.
In this case, we would like to minimize the correlation between each features and predicted probability score of the correct occupation.
Covariance Constrained Loss
Intuition: Example
7
9
Alex
Gender:
Race:
Pr(S/W Engineer) = 0.9
5
0
Pr(S/W Engineer) = 0.1
Kenya
Lets check if it satisfies the desirable properties
How do we evaluate?
For quantifying and evaluating bias, we need to access the protected attributes. However, the training/prediction is independent of the protected attributes. Also we need names of only training instances and not testing.
Balanced TPR
How do we evaluate?
Reason for RMSE over Average:
Interested in mitigating larger bias more
Model | Gender GAP | Surgeon | Rapper | Average | RMSE |
Model 1 | 1 | 999 | 500 | 706.4 |
Model 2 | 500 | 500 | 500 | 500 |
Results
Note: The notion of race (r) and gender (g) are not inherent to the methodology itself.
Results
Interpreting the weights of the model
Key Takeaways
Critical analysis of proposed method
+ No need to explicitly specify the group(s) susceptible to occupational bias. Latent groups identified from clustering of names.
+ The clusters can be interpretable and can simultaneously consider multiple protected attributes without explicit use/access. Also, For example, a domain expert based on the United States may not think of testing for caste discrimination, hence biases that an embedding may have against certain Indian last names may go unnoticed.
+ The names are needed only in training phase.
- Choosing the number of clusters is not clear. We don't have an analysis whether the clusters are interpretable if smaller cluster size or larger size is taken. Way to overcome this: CoCL
- Analysis with respect to debiased word embeddings?
Questions?