Neural Word Decomposition Models for Abusive Language Detection

Sravan Bodapati, Spandana Gella, Kasturi Bhattacharjee, Yaser Al-Onaizan


Abstract
The text we see in social media suffers from lots of undesired characterstics like hatespeech, abusive language, insults etc. The nature of this text is also very different compared to the traditional text we see in news with lots of obfuscated words, intended typos. This poses several robustness challenges to many natural language processing (NLP) techniques developed for traditional text. Many techniques proposed in the recent times such as charecter encoding models, subword models, byte pair encoding to extract subwords can aid in dealing with few of these nuances. In our work, we analyze the effectiveness of each of the above techniques, compare and contrast various word decomposition techniques when used in combination with others. We experiment with recent advances of finetuning pretrained language models, and demonstrate their robustness to domain shift. We also show our approaches achieve state of the art performance on Wikipedia attack, toxicity datasets, and Twitter hatespeech dataset.
Anthology ID:
W19-3515
Original:
W19-3515v1
Version 2:
W19-3515v2
Volume:
Proceedings of the Third Workshop on Abusive Language Online
Month:
August
Year:
2019
Address:
Florence, Italy
Venues:
ACL | ALW | WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
135–145
Language:
URL:
https://aclanthology.org/W19-3515
DOI:
10.18653/v1/W19-3515
Bibkey:
Cite (ACL):
Sravan Bodapati, Spandana Gella, Kasturi Bhattacharjee, and Yaser Al-Onaizan. 2019. Neural Word Decomposition Models for Abusive Language Detection. In Proceedings of the Third Workshop on Abusive Language Online, pages 135–145, Florence, Italy. Association for Computational Linguistics.
Cite (Informal):
Neural Word Decomposition Models for Abusive Language Detection (Bodapati et al., 2019)
Copy Citation:
PDF:
https://aclanthology.org/W19-3515.pdf