Subversive Toxicity Detection using Sentiment Information

Eloi Brassard-Gourdeau, Richard Khoury


Abstract
The presence of toxic content has become a major problem for many online communities. Moderators try to limit this problem by implementing more and more refined comment filters, but toxic users are constantly finding new ways to circumvent them. Our hypothesis is that while modifying toxic content and keywords to fool filters can be easy, hiding sentiment is harder. In this paper, we explore various aspects of sentiment detection and their correlation to toxicity, and use our results to implement a toxicity detection tool. We then test how adding the sentiment information helps detect toxicity in three different real-world datasets, and incorporate subversion to these datasets to simulate a user trying to circumvent the system. Our results show sentiment information has a positive impact on toxicity detection.
Anthology ID:
W19-3501
Volume:
Proceedings of the Third Workshop on Abusive Language Online
Month:
August
Year:
2019
Address:
Florence, Italy
Editors:
Sarah T. Roberts, Joel Tetreault, Vinodkumar Prabhakaran, Zeerak Waseem
Venue:
ALW
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
1–10
Language:
URL:
https://aclanthology.org/W19-3501
DOI:
10.18653/v1/W19-3501
Bibkey:
Cite (ACL):
Eloi Brassard-Gourdeau and Richard Khoury. 2019. Subversive Toxicity Detection using Sentiment Information. In Proceedings of the Third Workshop on Abusive Language Online, pages 1–10, Florence, Italy. Association for Computational Linguistics.
Cite (Informal):
Subversive Toxicity Detection using Sentiment Information (Brassard-Gourdeau & Khoury, ALW 2019)
Copy Citation:
PDF:
https://aclanthology.org/W19-3501.pdf