New methods, old problems
Ethics and Bias in Natural Language Processing
Ben Batorsky, PhD
Presented as part of the NLP Summit in 2021
About me
Part of Ciox Health, health information management company representing 3/5ths of US hospitals
Focused on providing actionable health information and insights in the hands of researchers using AI + NLP technology
The “old days” of machine learning
1950s: Extensive rules engine-based MT
Into the era of neural models
https://towardsdatascience.com/a-logistic-regression-from-scratch-3824468b1f88
Lots of hand engineering (work)
2013: The advent of “generalized” word embeddings
Can instead “swap-in” pre-trained embedding
Circa 2000s Language Model
A Neural Probabilistic Language Model (https://papers.nips.cc/paper/1839-a-neural-probabilistic-language-model.pdf)
Word embeddings used across settings and languages
med2vec - Embeddings based on medical records
Large neural Language Models are the SOTA...and they keep getting larger
Vaswani, Ashish, et al. "Attention is all you need." https://arxiv.org/pdf/1706.03762.pdf
On the Dangers of Stochastic Parrots (Bender 2021)
Neural model performance on Machine Translation
But remember: Garbage in, garbage out
The Internet
Where does the data for these methods come from?
Whose perspectives are represented in Wikipedia and web text?
What are the impacts of flawed AI systems?
GPT-2 producing problematic passages from race/gender/orientation seeds
Ethics and bias (what we’ll be talking about)
(Some) Types of bias in ML
Gender/racial bias in word embeddings
Word vector similarity between gender words and occupations
Word embeddings quantify 100 years of gender and ethnic stereotypes (https://www.pnas.org/content/pnas/115/16/E3635.full.pdf)
Racial/ethnic bias in the “most similar” occupations
Word embeddings encode history
Word embeddings quantify 100 years of gender and ethnic stereotypes (https://www.pnas.org/content/pnas/115/16/E3635.full.pdf)
Try it yourself
(Some) Definitions of fairness
Group-level fairness
Statistical parity
Individual-level fairness
Similar individual = similar outcome
Cynthia Dwork - Finding Fairness (https://www.youtube.com/watch?v=i_avLd49f8I&feature=youtu.be&t=1548)
Language under-representation: ~90% of languages have almost no labelled data available
Availability of data by language class (see table)
Which perspectives are represented in labels?
Locations of Mechanical Turk labellers
http://turktools.net/crowdsourcing/
Upshot: Variable performance across languages, multi-language models generally worse
Fairness: Representation that better matches the speaker distribution
“Technology cannot be accessible if it is only available for English speakers with a standard accent.” (Sebastian Ruder https://ruder.io/nlp-beyond-english/)
“The handful of languages on which NLP systems are trained and tested are often related...leading to a typological echo-chamber. As a result, a vast majority of typologically diverse linguistic phenomena are never seen by our NLP systems” (Joshi 2021 [2004.09095] The State and Fate of Linguistic Diversity and Inclusion in the NLP World)
General strategies for addressing bias in prediction
Some methods for monitoring/addressing bias
De-biasing word embeddings
Equalize
Equalize pairs
We fixed it!
“To measure racial discrimination by people, we must create controlled circumstances in the real world where only race differs. For an algorithm, we can create equally controlled just by feeding it the right data and observing its behavior.”
We...fixed it?
Some questions worth asking in development and usage
Thank you!
More resources
Get in touch!
Website: https://benbatorsky.com/
Blog: https://bpben.github.io/
Twitter: https://twitter.com/bpben2
Is it worth it?