Unmasking the Hidden Meaning: Bridging Implicit and Explicit Hate Speech Embedding Representations

Nicolás Benjamín Ocampo, Elena Cabrio, Serena Villata


Abstract
Research on automatic hate speech (HS) detection has mainly focused on identifying explicit forms of hateful expressions on user-generated content. Recently, a few works have started to investigate methods to address more implicit and subtle abusive content. However, despite these efforts, automated systems still struggle to correctly recognize implicit and more veiled forms of HS. As these systems heavily rely on proper textual representations for classification, it is crucial to investigate the differences in embedding implicit and explicit messages. Our contribution to address this challenging task is fourfold. First, we present a comparative analysis of transformer-based models, evaluating their performance across five datasets containing implicit HS messages. Second, we examine the embedding representations of implicit messages across different targets, gaining insight into how veiled cases are encoded. Third, we compare and link explicit and implicit hateful messages across these datasets through their targets, enforcing the relation between explicitness and implicitness and obtaining more meaningful embedding representations. Lastly, we show how these newer representation maintains high performance on HS labels, while improving classification in borderline cases.
Anthology ID:
2023.findings-emnlp.441
Original:
2023.findings-emnlp.441v1
Version 2:
2023.findings-emnlp.441v2
Volume:
Findings of the Association for Computational Linguistics: EMNLP 2023
Month:
December
Year:
2023
Address:
Singapore
Editors:
Houda Bouamor, Juan Pino, Kalika Bali
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
6626–6637
Language:
URL:
https://aclanthology.org/2023.findings-emnlp.441
DOI:
10.18653/v1/2023.findings-emnlp.441
Bibkey:
Cite (ACL):
Nicolás Benjamín Ocampo, Elena Cabrio, and Serena Villata. 2023. Unmasking the Hidden Meaning: Bridging Implicit and Explicit Hate Speech Embedding Representations. In Findings of the Association for Computational Linguistics: EMNLP 2023, pages 6626–6637, Singapore. Association for Computational Linguistics.
Cite (Informal):
Unmasking the Hidden Meaning: Bridging Implicit and Explicit Hate Speech Embedding Representations (Ocampo et al., Findings 2023)
Copy Citation:
PDF:
https://aclanthology.org/2023.findings-emnlp.441.pdf