TY - GEN
T1 - Adultification Bias in LLMs and Text-To-Image Models
AU - Castleman, Jane
AU - Korolova, Aleksandra
N1 - Publisher Copyright:
© 2025 Copyright held by the owner/author(s).
PY - 2025/6/23
Y1 - 2025/6/23
N2 - The rapid adoption of generative AI models in domains such as education, policing, and social media raises significant concerns about potential bias and safety issues, particularly along protected attributes, such as race and gender, and when interacting with minors. Given the urgency of facilitating safe interactions with AI systems, we study bias along axes of race and gender in young girls. More specifically, we focus on "adultification bias,"a phenomenon in which Black girls are presumed to be more defiant, sexually intimate, and culpable than their White peers.Advances in alignment techniques show promise towards mitigating biases but vary in their coverage and effectiveness across models and bias types. Therefore, we measure explicit and implicit adultification bias in widely used LLMs and text-To-image (T2I) models, such as OpenAI, Meta, and Stability AI models. We find that LLMs exhibit explicit and implicit adultification bias against Black girls, assigning them harsher, more sexualized consequences in comparison to their White peers. Additionally, we find that T2I models depict Black girls as older and wearing more revealing clothing than their White counterparts, illustrating how adultification bias persists across modalities.We make three key contributions: (1) we measure a new form of bias in generative AI models, (2) we systematically study adultification bias across modalities, and (3) our findings emphasize that current alignment methods are insufficient for comprehensively addressing bias. Therefore, new alignment methods that address biases such as adultification are needed to ensure safe and equitable AI deployment.
AB - The rapid adoption of generative AI models in domains such as education, policing, and social media raises significant concerns about potential bias and safety issues, particularly along protected attributes, such as race and gender, and when interacting with minors. Given the urgency of facilitating safe interactions with AI systems, we study bias along axes of race and gender in young girls. More specifically, we focus on "adultification bias,"a phenomenon in which Black girls are presumed to be more defiant, sexually intimate, and culpable than their White peers.Advances in alignment techniques show promise towards mitigating biases but vary in their coverage and effectiveness across models and bias types. Therefore, we measure explicit and implicit adultification bias in widely used LLMs and text-To-image (T2I) models, such as OpenAI, Meta, and Stability AI models. We find that LLMs exhibit explicit and implicit adultification bias against Black girls, assigning them harsher, more sexualized consequences in comparison to their White peers. Additionally, we find that T2I models depict Black girls as older and wearing more revealing clothing than their White counterparts, illustrating how adultification bias persists across modalities.We make three key contributions: (1) we measure a new form of bias in generative AI models, (2) we systematically study adultification bias across modalities, and (3) our findings emphasize that current alignment methods are insufficient for comprehensively addressing bias. Therefore, new alignment methods that address biases such as adultification are needed to ensure safe and equitable AI deployment.
UR - https://www.scopus.com/pages/publications/105010815724
UR - https://www.scopus.com/pages/publications/105010815724#tab=citedBy
U2 - 10.1145/3715275.3732178
DO - 10.1145/3715275.3732178
M3 - Conference contribution
AN - SCOPUS:105010815724
T3 - ACMF AccT 2025 - Proceedings of the 2025 ACM Conference on Fairness, Accountability,and Transparency
SP - 2751
EP - 2767
BT - ACMF AccT 2025 - Proceedings of the 2025 ACM Conference on Fairness, Accountability,and Transparency
PB - Association for Computing Machinery, Inc
T2 - 8th Annual ACM Conference on Fairness, Accountability, and Transparency, FAccT 2025
Y2 - 23 June 2025 through 26 June 2025
ER -