@inproceedings{park-fung-2017-one,
title = "One-step and Two-step Classification for Abusive Language Detection on {T}witter",
author = "Park, Ji Ho and
Fung, Pascale",
editor = "Waseem, Zeerak and
Chung, Wendy Hui Kyong and
Hovy, Dirk and
Tetreault, Joel",
booktitle = "Proceedings of the First Workshop on Abusive Language Online",
month = aug,
year = "2017",
address = "Vancouver, BC, Canada",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/W17-3006",
doi = "10.18653/v1/W17-3006",
pages = "41--45",
abstract = "Automatic abusive language detection is a difficult but important task for online social media. Our research explores a two-step approach of performing classification on abusive language and then classifying into specific types and compares it with one-step approach of doing one multi-class classification for detecting sexist and racist languages. With a public English Twitter corpus of 20 thousand tweets in the type of sexism and racism, our approach shows a promising performance of 0.827 F-measure by using HybridCNN in one-step and 0.824 F-measure by using logistic regression in two-steps.",
}
<?xml version="1.0" encoding="UTF-8"?>
<modsCollection xmlns="http://www.loc.gov/mods/v3">
<mods ID="park-fung-2017-one">
<titleInfo>
<title>One-step and Two-step Classification for Abusive Language Detection on Twitter</title>
</titleInfo>
<name type="personal">
<namePart type="given">Ji</namePart>
<namePart type="given">Ho</namePart>
<namePart type="family">Park</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Pascale</namePart>
<namePart type="family">Fung</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<originInfo>
<dateIssued>2017-08</dateIssued>
</originInfo>
<typeOfResource>text</typeOfResource>
<relatedItem type="host">
<titleInfo>
<title>Proceedings of the First Workshop on Abusive Language Online</title>
</titleInfo>
<name type="personal">
<namePart type="given">Zeerak</namePart>
<namePart type="family">Waseem</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Wendy</namePart>
<namePart type="given">Hui</namePart>
<namePart type="given">Kyong</namePart>
<namePart type="family">Chung</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Dirk</namePart>
<namePart type="family">Hovy</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Joel</namePart>
<namePart type="family">Tetreault</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<originInfo>
<publisher>Association for Computational Linguistics</publisher>
<place>
<placeTerm type="text">Vancouver, BC, Canada</placeTerm>
</place>
</originInfo>
<genre authority="marcgt">conference publication</genre>
</relatedItem>
<abstract>Automatic abusive language detection is a difficult but important task for online social media. Our research explores a two-step approach of performing classification on abusive language and then classifying into specific types and compares it with one-step approach of doing one multi-class classification for detecting sexist and racist languages. With a public English Twitter corpus of 20 thousand tweets in the type of sexism and racism, our approach shows a promising performance of 0.827 F-measure by using HybridCNN in one-step and 0.824 F-measure by using logistic regression in two-steps.</abstract>
<identifier type="citekey">park-fung-2017-one</identifier>
<identifier type="doi">10.18653/v1/W17-3006</identifier>
<location>
<url>https://aclanthology.org/W17-3006</url>
</location>
<part>
<date>2017-08</date>
<extent unit="page">
<start>41</start>
<end>45</end>
</extent>
</part>
</mods>
</modsCollection>
%0 Conference Proceedings
%T One-step and Two-step Classification for Abusive Language Detection on Twitter
%A Park, Ji Ho
%A Fung, Pascale
%Y Waseem, Zeerak
%Y Chung, Wendy Hui Kyong
%Y Hovy, Dirk
%Y Tetreault, Joel
%S Proceedings of the First Workshop on Abusive Language Online
%D 2017
%8 August
%I Association for Computational Linguistics
%C Vancouver, BC, Canada
%F park-fung-2017-one
%X Automatic abusive language detection is a difficult but important task for online social media. Our research explores a two-step approach of performing classification on abusive language and then classifying into specific types and compares it with one-step approach of doing one multi-class classification for detecting sexist and racist languages. With a public English Twitter corpus of 20 thousand tweets in the type of sexism and racism, our approach shows a promising performance of 0.827 F-measure by using HybridCNN in one-step and 0.824 F-measure by using logistic regression in two-steps.
%R 10.18653/v1/W17-3006
%U https://aclanthology.org/W17-3006
%U https://doi.org/10.18653/v1/W17-3006
%P 41-45
Markdown (Informal)
[One-step and Two-step Classification for Abusive Language Detection on Twitter](https://aclanthology.org/W17-3006) (Park & Fung, ALW 2017)
ACL