Evidence for Hypodescent in Visual Semantic AI

05/22/2022
by   Robert Wolfe, et al.
0

We examine the state-of-the-art multimodal "visual semantic" model CLIP ("Contrastive Language Image Pretraining") for the rule of hypodescent, or one-drop rule, whereby multiracial people are more likely to be assigned a racial or ethnic label corresponding to a minority or disadvantaged racial or ethnic group than to the equivalent majority or advantaged group. A face morphing experiment grounded in psychological research demonstrating hypodescent indicates that, at the midway point of 1,000 series of morphed images, CLIP associates 69.7 label over a White text label, and similarly prefers Latina (75.8 (89.1 female morphs, reflecting hypodescent. Additionally, assessment of the underlying cosine similarities in the model reveals that association with White is correlated with association with "person," with Pearson's rho as high as 0.82 over a 21,000-image morph series, indicating that a White person corresponds to the default representation of a person in CLIP. Finally, we show that the stereotype-congruent pleasantness association of an image correlates with association with the Black text label in CLIP, with Pearson's rho = 0.48 for 21,000 Black-White multiracial male images, and rho = 0.41 for Black-White multiracial female images. CLIP is trained on English-language text gathered using data collected from an American website (Wikipedia), and our findings demonstrate that CLIP embeds the values of American racial hierarchy, reflecting the implicit and explicit beliefs that are present in human minds. We contextualize these findings within the history and psychology of hypodescent. Overall, the data suggests that AI supervised using natural language will, unless checked, learn biases that reflect racial hierarchies.

READ FULL TEXT

page 14

page 15

page 16

research
07/01/2022

American == White in Multimodal Language-and-Image AI

Three state-of-the-art language-and-image AI models, CLIP, SLIP, and BLI...
research
12/21/2022

Contrastive Language-Vision AI Models Pretrained on Web-Scraped Multimodal Data Exhibit Sexual Objectification Bias

Nine language-vision AI models trained on web scrapes with the Contrasti...
research
05/23/2022

Markedness in Visual Semantic AI

We evaluate the state-of-the-art multimodal "visual semantic" model CLIP...
research
08/05/2018

Improving Deep Visual Representation for Person Re-identification by Global and Local Image-language Association

Person re-identification is an important task that requires learning dis...
research
08/25/2016

Semantics derived automatically from language corpora contain human-like biases

Artificial intelligence and machine learning are in a period of astoundi...
research
05/29/2023

Marked Personas: Using Natural Language Prompts to Measure Stereotypes in Language Models

To recognize and mitigate harms from large language models (LLMs), we ne...
research
02/16/2021

Could you become more credible by being White? Assessing Impact of Race on Credibility with Deepfakes

Computer mediated conversations (e.g., videoconferencing) is now the new...

Please sign up or login with your details

Forgot password? Click here to reset