2024
pdf
bib
Proceedings of the Fourth Workshop on Threat, Aggression & Cyberbullying @ LREC-COLING-2024
Ritesh Kumar
|
Atul Kr. Ojha
|
Shervin Malmasi
|
Bharathi Raja Chakravarthi
|
Bornini Lahiri
|
Siddharth Singh
|
Shyam Ratan
Proceedings of the Fourth Workshop on Threat, Aggression & Cyberbullying @ LREC-COLING-2024
2022
pdf
bib
abs
The ComMA Dataset V0.2: Annotating Aggression and Bias in Multilingual Social Media Discourse
Ritesh Kumar
|
Shyam Ratan
|
Siddharth Singh
|
Enakshi Nandi
|
Laishram Niranjana Devi
|
Akash Bhagat
|
Yogesh Dawer
|
Bornini Lahiri
|
Akanksha Bansal
|
Atul Kr. Ojha
Proceedings of the Thirteenth Language Resources and Evaluation Conference
In this paper, we discuss the development of a multilingual dataset annotated with a hierarchical, fine-grained tagset marking different types of aggression and the “context” in which they occur. The context, here, is defined by the conversational thread in which a specific comment occurs and also the “type” of discursive role that the comment is performing with respect to the previous comment. The initial dataset, being discussed here consists of a total 59,152 annotated comments in four languages - Meitei, Bangla, Hindi, and Indian English - collected from various social media platforms such as YouTube, Facebook, Twitter and Telegram. As is usual on social media websites, a large number of these comments are multilingual, mostly code-mixed with English. The paper gives a detailed description of the tagset being used for annotation and also the process of developing a multi-label, fine-grained tagset that has been used for marking comments with aggression and bias of various kinds including sexism (called gender bias in the tagset), religious intolerance (called communal bias in the tagset), class/caste bias and ethnic/racial bias. We also define and discuss the tags that have been used for marking the different discursive role being performed through the comments, such as attack, defend, etc. Finally, we present a basic statistical analysis of the dataset. The dataset is being incrementally made publicly available on the project website.
2021
pdf
bib
Proceedings of the 18th International Conference on Natural Language Processing: Shared Task on Multilingual Gender Biased and Communal Language Identification
Ritesh Kumar
|
Siddharth Singh
|
Enakshi Nandi
|
Shyam Ratan
|
Laishram Niranjana Devi
|
Bornini Lahiri
|
Akanksha Bansal
|
Akash Bhagat
|
Yogesh Dawer
Proceedings of the 18th International Conference on Natural Language Processing: Shared Task on Multilingual Gender Biased and Communal Language Identification
pdf
bib
abs
ComMA@ICON: Multilingual Gender Biased and Communal Language Identification Task at ICON-2021
Ritesh Kumar
|
Shyam Ratan
|
Siddharth Singh
|
Enakshi Nandi
|
Laishram Niranjana Devi
|
Akash Bhagat
|
Yogesh Dawer
|
Bornini Lahiri
|
Akanksha Bansal
Proceedings of the 18th International Conference on Natural Language Processing: Shared Task on Multilingual Gender Biased and Communal Language Identification
This paper presents the findings of the ICON-2021 shared task on Multilingual Gender Biased and Communal Language Identification, which aims to identify aggression, gender bias, and communal bias in data presented in four languages: Meitei, Bangla, Hindi and English. The participants were presented the option of approaching the task as three separate classification tasks or a multi-label classification task or a structured classification task. If approached as three separate classification tasks, the task includes three sub-tasks: aggression identification (sub-task A), gender bias identification (sub-task B), and communal bias identification (sub-task C). For this task, the participating teams were provided with a total dataset of approximately 12,000, with 3,000 comments across each of the four languages, sourced from popular social media sites such as YouTube, Twitter, Facebook and Telegram and the the three labels presented as a single tuple. For the test systems, approximately 1,000 comments were provided in each language for every sub-task. We attracted a total of 54 registrations in the task, out of which 11 teams submitted their test runs. The best system obtained an overall instance-F1 of 0.371 in the multilingual test set (it was simply a combined test set of the instances in each individual language). In the individual sub-tasks, the best micro f1 scores are 0.539, 0.767 and 0.834 respectively for each of the sub-task A, B and C. The best overall, averaged micro f1 is 0.713. The results show that while systems have managed to perform reasonably well in individual sub-tasks, especially gender bias and communal bias tasks, it is substantially more difficult to do a 3-class classification of aggression level and even more difficult to build a system that correctly classifies everything right. It is only in slightly over 1/3 of the instances that most of the systems predicted the correct class across the board, despite the fact that there was a significant overlap across the three sub-tasks.
2020
pdf
bib
abs
KMI-Panlingua-IITKGP @SIGTYP2020: Exploring rules and hybrid systems for automatic prediction of typological features
Ritesh Kumar
|
Deepak Alok
|
Akanksha Bansal
|
Bornini Lahiri
|
Atul Kr. Ojha
Proceedings of the Second Workshop on Computational Research in Linguistic Typology
This paper enumerates SigTyP 2020 Shared Task on the prediction of typological features as performed by the KMI-Panlingua-IITKGP team. The task entailed the prediction of missing values in a particular language, provided, the name of the language family, its genus, location (in terms of latitude and longitude coordinates and name of the country where it is spoken) and a set of feature-value pair are available. As part of fulfillment of the aforementioned task, the team submitted 3 kinds of system - 2 rule-based and one hybrid system. Of these 3, one rule-based system generated the best performance on the test set. All the systems were ‘constrained’ in the sense that no additional dataset or information, other than those provided by the organisers, was used for developing the systems.
pdf
bib
Proceedings of the Second Workshop on Trolling, Aggression and Cyberbullying
Ritesh Kumar
|
Atul Kr. Ojha
|
Bornini Lahiri
|
Marcos Zampieri
|
Shervin Malmasi
|
Vanessa Murdock
|
Daniel Kadar
Proceedings of the Second Workshop on Trolling, Aggression and Cyberbullying
pdf
bib
abs
Developing a Multilingual Annotated Corpus of Misogyny and Aggression
Shiladitya Bhattacharya
|
Siddharth Singh
|
Ritesh Kumar
|
Akanksha Bansal
|
Akash Bhagat
|
Yogesh Dawer
|
Bornini Lahiri
|
Atul Kr. Ojha
Proceedings of the Second Workshop on Trolling, Aggression and Cyberbullying
In this paper, we discuss the development of a multilingual annotated corpus of misogyny and aggression in Indian English, Hindi, and Indian Bangla as part of a project on studying and automatically identifying misogyny and communalism on social media (the ComMA Project). The dataset is collected from comments on YouTube videos and currently contains a total of over 20,000 comments. The comments are annotated at two levels - aggression (overtly aggressive, covertly aggressive, and non-aggressive) and misogyny (gendered and non-gendered). We describe the process of data collection, the tagset used for annotation, and issues and challenges faced during the process of annotation. Finally, we discuss the results of the baseline experiments conducted to develop a classifier for misogyny in the three languages.
2018
pdf
bib
abs
Language Identification and Morphosyntactic Tagging: The Second VarDial Evaluation Campaign
Marcos Zampieri
|
Shervin Malmasi
|
Preslav Nakov
|
Ahmed Ali
|
Suwon Shon
|
James Glass
|
Yves Scherrer
|
Tanja Samardžić
|
Nikola Ljubešić
|
Jörg Tiedemann
|
Chris van der Lee
|
Stefan Grondelaers
|
Nelleke Oostdijk
|
Dirk Speelman
|
Antal van den Bosch
|
Ritesh Kumar
|
Bornini Lahiri
|
Mayank Jain
Proceedings of the Fifth Workshop on NLP for Similar Languages, Varieties and Dialects (VarDial 2018)
We present the results and the findings of the Second VarDial Evaluation Campaign on Natural Language Processing (NLP) for Similar Languages, Varieties and Dialects. The campaign was organized as part of the fifth edition of the VarDial workshop, collocated with COLING’2018. This year, the campaign included five shared tasks, including two task re-runs – Arabic Dialect Identification (ADI) and German Dialect Identification (GDI) –, and three new tasks – Morphosyntactic Tagging of Tweets (MTT), Discriminating between Dutch and Flemish in Subtitles (DFS), and Indo-Aryan Language Identification (ILI). A total of 24 teams submitted runs across the five shared tasks, and contributed 22 system description papers, which were included in the VarDial workshop proceedings and are referred to in this report.
2012
pdf
bib
Developing a POS tagger for Magahi: A Comparative Study
Ritesh Kumar
|
Bornini Lahiri
|
Deepak Alok
Proceedings of the 10th Workshop on Asian Language Resources