Musical Ethnocentrism in Large Language Models

Anna Kruspe


Abstract
Large Language Models (LLMs) reflect the biases in their training data and, by extension, those of the people who created this training data. Detecting, analyzing, and mitigating such biases is becoming a focus of research. One type of bias that has been understudied so far are geocultural biases. Those can be caused by an imbalance in the representation of different geographic regions and cultures in the training data, but also by value judgments contained therein. In this paper, we make a first step towards analyzing musical biases in LLMs, particularly ChatGPT and Mixtral. We conduct two experiments. In the first, we prompt LLMs to provide lists of the “Top 100” musical contributors of various categories and analyze their countries of origin. In the second experiment, we ask the LLMs to numerically rate various aspects of the musical cultures of different countries. Our results indicate a strong preference of the LLMs for Western music cultures in both experiments.
Anthology ID:
2024.nlp4musa-1.11
Volume:
Proceedings of the 3rd Workshop on NLP for Music and Audio (NLP4MusA)
Month:
November
Year:
2024
Address:
Oakland, USA
Editors:
Anna Kruspe, Sergio Oramas, Elena V. Epure, Mohamed Sordo, Benno Weck, SeungHeon Doh, Minz Won, Ilaria Manco, Gabriel Meseguer-Brocal
Venues:
NLP4MusA | WS
SIG:
Publisher:
Association for Computational Lingustics
Note:
Pages:
62–68
Language:
URL:
https://aclanthology.org/2024.nlp4musa-1.11/
DOI:
Bibkey:
Cite (ACL):
Anna Kruspe. 2024. Musical Ethnocentrism in Large Language Models. In Proceedings of the 3rd Workshop on NLP for Music and Audio (NLP4MusA), pages 62–68, Oakland, USA. Association for Computational Lingustics.
Cite (Informal):
Musical Ethnocentrism in Large Language Models (Kruspe, NLP4MusA 2024)
Copy Citation:
PDF:
https://aclanthology.org/2024.nlp4musa-1.11.pdf