Discovering and Mitigating Biases in CLIP-based Text-to-Image Generation
�Md Mehrab Tanjim1, Krishna Kumar Singh2, Kushal Kafle2, Ritwik Sinha2, Garrison W. Cottrell1
1UC San Diego, 2Adobe Research
University of California San Diego
Adobe Research
Motivation
Contributions
Our Debiasing Framework
References
Discovering Biases
CLIP + ID
CLIP + LPIPS
Output of StyleCLIP
Original
CLIP + Gender + ID
CLIP + Gender + LPIPS
CLIP + LPIPS + ID
CLIP + Gender + LPIPS + ID
Text-based Debiasing
CLIP + Gender
[1] lec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, et al. Learning transferable visual models from natural language super vision. In PMLR, 2021
[2]Or Patashnik, Zongze Wu, Eli Shechtman, Daniel Cohen-Or, and Dani Lischinski. Styleclip: Text-driven manipulation of stylegan imagery. In Proceedings CVPR, 2021
[3] Sandhini Agarwal, Gretchen Krueger, Jack Clark, Alec Radford, Jong Wook Kim, and Miles Brundage. Evaluating CLIP: towards characterization of broader capabilities and downstream implications. arXiv preprint arXiv:2108.02818, 2021.
[4] Md Mehrab Tanjim, Ritwik Sinha, Krishna Kumar Singh, Sridhar Mahadevan, David Arbour, Moumita Sinha, and Garrison W Cottrell. Generating and controlling diversity in image�search. In Proceedings WACV, 2022.
[5] Jiankang Deng, Jia Guo, Niannan Xue, and Stefanos Zafeiriou. Arcface: Additive angular margin loss for deep face recognition. In Proceedings of the IEEE/CVF conference on CVPR.
[6] Richard Zhang, Phillip Isola, Alexei A Efros, Eli Shechtman, and Oliver Wang. The unreasonable effectiveness of deep features as a perceptual metric. In CVPR, 2018
‘Plumber’, ‘Nurse’, ‘Administrative Assistant’, ‘Farmer’, ‘Security Guard’, ‘Executive Manager’, ‘Military Person’, ‘Maids & Housekeepers’.
Related Work
Original
Original
Face of a carpenter
Face of a nurse
After Debiasing
Before Debiasing
After Debiasing
Before Debiasing
Original
After Debiasing
Before Debiasing
Original
After Debiasing
Before Debiasing
Face of a software engineer
Face of an administrative assistant
ROC curve for gender subset
ROC curve for race subset
Misrank comparison for race subset
Misrank comparison for gender subset
Original Image
“An image of a plumber”
GradCAM
Original Image
GradCAM
“An image of a farmer”
Original Image
GradCAM
Original Image
GradCAM
Some examples of biases in CLIP-based generative model, StyleCLIP [2], and results from our debiasing framework are shown above.