@inproceedings{chen-etal-2019-unraveling,
    title = "Unraveling the Search Space of Abusive Language in {W}ikipedia with Dynamic Lexicon Acquisition",
    author = "Chen, Wei-Fan  and
      Al Khatib, Khalid  and
      Hagen, Matthias  and
      Wachsmuth, Henning  and
      Stein, Benno",
    editor = "Feldman, Anna  and
      Da San Martino, Giovanni  and
      Barr{\'o}n-Cede{\~n}o, Alberto  and
      Brew, Chris  and
      Leberknight, Chris  and
      Nakov, Preslav",
    booktitle = "Proceedings of the Second Workshop on Natural Language Processing for Internet Freedom: Censorship, Disinformation, and Propaganda",
    month = nov,
    year = "2019",
    address = "Hong Kong, China",
    publisher = "Association for Computational Linguistics",
    url = "https://aclanthology.org/D19-5009/",
    doi = "10.18653/v1/D19-5009",
    pages = "76--82",
    abstract = "Many discussions on online platforms suffer from users offending others by using abusive terminology, threatening each other, or being sarcastic. Since an automatic detection of abusive language can support human moderators of online discussion platforms, detecting abusiveness has recently received increased attention. However, the existing approaches simply train one classifier for the whole variety of abusiveness. In contrast, our approach is to distinguish explicitly abusive cases from the more ``shadowed'' ones. By dynamically extending a lexicon of abusive terms (e.g., including new obfuscations of abusive terms), our approach can support a moderator with explicit unraveled explanations for why something was flagged as abusive: due to known explicitly abusive terms, due to newly detected (obfuscated) terms, or due to shadowed cases."
}<?xml version="1.0" encoding="UTF-8"?>
<modsCollection xmlns="http://www.loc.gov/mods/v3">
<mods ID="chen-etal-2019-unraveling">
    <titleInfo>
        <title>Unraveling the Search Space of Abusive Language in Wikipedia with Dynamic Lexicon Acquisition</title>
    </titleInfo>
    <name type="personal">
        <namePart type="given">Wei-Fan</namePart>
        <namePart type="family">Chen</namePart>
        <role>
            <roleTerm authority="marcrelator" type="text">author</roleTerm>
        </role>
    </name>
    <name type="personal">
        <namePart type="given">Khalid</namePart>
        <namePart type="family">Al Khatib</namePart>
        <role>
            <roleTerm authority="marcrelator" type="text">author</roleTerm>
        </role>
    </name>
    <name type="personal">
        <namePart type="given">Matthias</namePart>
        <namePart type="family">Hagen</namePart>
        <role>
            <roleTerm authority="marcrelator" type="text">author</roleTerm>
        </role>
    </name>
    <name type="personal">
        <namePart type="given">Henning</namePart>
        <namePart type="family">Wachsmuth</namePart>
        <role>
            <roleTerm authority="marcrelator" type="text">author</roleTerm>
        </role>
    </name>
    <name type="personal">
        <namePart type="given">Benno</namePart>
        <namePart type="family">Stein</namePart>
        <role>
            <roleTerm authority="marcrelator" type="text">author</roleTerm>
        </role>
    </name>
    <originInfo>
        <dateIssued>2019-11</dateIssued>
    </originInfo>
    <typeOfResource>text</typeOfResource>
    <relatedItem type="host">
        <titleInfo>
            <title>Proceedings of the Second Workshop on Natural Language Processing for Internet Freedom: Censorship, Disinformation, and Propaganda</title>
        </titleInfo>
        <name type="personal">
            <namePart type="given">Anna</namePart>
            <namePart type="family">Feldman</namePart>
            <role>
                <roleTerm authority="marcrelator" type="text">editor</roleTerm>
            </role>
        </name>
        <name type="personal">
            <namePart type="given">Giovanni</namePart>
            <namePart type="family">Da San Martino</namePart>
            <role>
                <roleTerm authority="marcrelator" type="text">editor</roleTerm>
            </role>
        </name>
        <name type="personal">
            <namePart type="given">Alberto</namePart>
            <namePart type="family">Barrón-Cedeño</namePart>
            <role>
                <roleTerm authority="marcrelator" type="text">editor</roleTerm>
            </role>
        </name>
        <name type="personal">
            <namePart type="given">Chris</namePart>
            <namePart type="family">Brew</namePart>
            <role>
                <roleTerm authority="marcrelator" type="text">editor</roleTerm>
            </role>
        </name>
        <name type="personal">
            <namePart type="given">Chris</namePart>
            <namePart type="family">Leberknight</namePart>
            <role>
                <roleTerm authority="marcrelator" type="text">editor</roleTerm>
            </role>
        </name>
        <name type="personal">
            <namePart type="given">Preslav</namePart>
            <namePart type="family">Nakov</namePart>
            <role>
                <roleTerm authority="marcrelator" type="text">editor</roleTerm>
            </role>
        </name>
        <originInfo>
            <publisher>Association for Computational Linguistics</publisher>
            <place>
                <placeTerm type="text">Hong Kong, China</placeTerm>
            </place>
        </originInfo>
        <genre authority="marcgt">conference publication</genre>
    </relatedItem>
    <abstract>Many discussions on online platforms suffer from users offending others by using abusive terminology, threatening each other, or being sarcastic. Since an automatic detection of abusive language can support human moderators of online discussion platforms, detecting abusiveness has recently received increased attention. However, the existing approaches simply train one classifier for the whole variety of abusiveness. In contrast, our approach is to distinguish explicitly abusive cases from the more “shadowed” ones. By dynamically extending a lexicon of abusive terms (e.g., including new obfuscations of abusive terms), our approach can support a moderator with explicit unraveled explanations for why something was flagged as abusive: due to known explicitly abusive terms, due to newly detected (obfuscated) terms, or due to shadowed cases.</abstract>
    <identifier type="citekey">chen-etal-2019-unraveling</identifier>
    <identifier type="doi">10.18653/v1/D19-5009</identifier>
    <location>
        <url>https://aclanthology.org/D19-5009/</url>
    </location>
    <part>
        <date>2019-11</date>
        <extent unit="page">
            <start>76</start>
            <end>82</end>
        </extent>
    </part>
</mods>
</modsCollection>
%0 Conference Proceedings
%T Unraveling the Search Space of Abusive Language in Wikipedia with Dynamic Lexicon Acquisition
%A Chen, Wei-Fan
%A Al Khatib, Khalid
%A Hagen, Matthias
%A Wachsmuth, Henning
%A Stein, Benno
%Y Feldman, Anna
%Y Da San Martino, Giovanni
%Y Barrón-Cedeño, Alberto
%Y Brew, Chris
%Y Leberknight, Chris
%Y Nakov, Preslav
%S Proceedings of the Second Workshop on Natural Language Processing for Internet Freedom: Censorship, Disinformation, and Propaganda
%D 2019
%8 November
%I Association for Computational Linguistics
%C Hong Kong, China
%F chen-etal-2019-unraveling
%X Many discussions on online platforms suffer from users offending others by using abusive terminology, threatening each other, or being sarcastic. Since an automatic detection of abusive language can support human moderators of online discussion platforms, detecting abusiveness has recently received increased attention. However, the existing approaches simply train one classifier for the whole variety of abusiveness. In contrast, our approach is to distinguish explicitly abusive cases from the more “shadowed” ones. By dynamically extending a lexicon of abusive terms (e.g., including new obfuscations of abusive terms), our approach can support a moderator with explicit unraveled explanations for why something was flagged as abusive: due to known explicitly abusive terms, due to newly detected (obfuscated) terms, or due to shadowed cases.
%R 10.18653/v1/D19-5009
%U https://aclanthology.org/D19-5009/
%U https://doi.org/10.18653/v1/D19-5009
%P 76-82
Markdown (Informal)
[Unraveling the Search Space of Abusive Language in Wikipedia with Dynamic Lexicon Acquisition](https://aclanthology.org/D19-5009/) (Chen et al., NLP4IF 2019)
ACL