Intersectional Stereotypes in Large Language Models: Dataset and Analysis

Weicheng Ma, Brian Chiang, Tong Wu, Lili Wang, Soroush Vosoughi


Abstract
Despite many stereotypes targeting intersectional demographic groups, prior studies on stereotypes within Large Language Models (LLMs) primarily focus on broader, individual categories. This research bridges this gap by introducing a novel dataset of intersectional stereotypes, curated with the assistance of the ChatGPT model and manually validated. Moreover, this paper offers a comprehensive analysis of intersectional stereotype propagation in three contemporary LLMs by leveraging this dataset. The findings underscore the urgency of focusing on intersectional biases in ongoing efforts to reduce stereotype prevalence in LLMs.
Anthology ID:
2023.findings-emnlp.575
Volume:
Findings of the Association for Computational Linguistics: EMNLP 2023
Month:
December
Year:
2023
Address:
Singapore
Editors:
Houda Bouamor, Juan Pino, Kalika Bali
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
8589–8597
Language:
URL:
https://aclanthology.org/2023.findings-emnlp.575
DOI:
10.18653/v1/2023.findings-emnlp.575
Bibkey:
Cite (ACL):
Weicheng Ma, Brian Chiang, Tong Wu, Lili Wang, and Soroush Vosoughi. 2023. Intersectional Stereotypes in Large Language Models: Dataset and Analysis. In Findings of the Association for Computational Linguistics: EMNLP 2023, pages 8589–8597, Singapore. Association for Computational Linguistics.
Cite (Informal):
Intersectional Stereotypes in Large Language Models: Dataset and Analysis (Ma et al., Findings 2023)
Copy Citation:
PDF:
https://aclanthology.org/2023.findings-emnlp.575.pdf