Michał Satława


pdf bib
SRPOL DIALOGUE SYSTEMS at SemEval-2021 Task 5: Automatic Generation of Training Data for Toxic Spans Detection
Michał Satława | Katarzyna Zamłyńska | Jarosław Piersa | Joanna Kolis | Klaudia Firląg | Katarzyna Beksa | Zuzanna Bordzicka | Christian Goltz | Paweł Bujnowski | Piotr Andruszkiewicz
Proceedings of the 15th International Workshop on Semantic Evaluation (SemEval-2021)

This paper presents a system used for SemEval-2021 Task 5: Toxic Spans Detection. Our system is an ensemble of BERT-based models for binary word classification, trained on a dataset extended by toxic comments modified and generated by two language models. For the toxic word classification, the prediction threshold value was optimized separately for every comment, in order to maximize the expected F1 value.

pdf bib
Multilingual Entity and Relation Extraction Dataset and Model
Alessandro Seganti | Klaudia Firląg | Helena Skowronska | Michał Satława | Piotr Andruszkiewicz
Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume

We present a novel dataset and model for a multilingual setting to approach the task of Joint Entity and Relation Extraction. The SMiLER dataset consists of 1.1 M annotated sentences, representing 36 relations, and 14 languages. To the best of our knowledge, this is currently both the largest and the most comprehensive dataset of this type. We introduce HERBERTa, a pipeline that combines two independent BERT models: one for sequence classification, and the other for entity tagging. The model achieves micro F1 81.49 for English on this dataset, which is close to the current SOTA on CoNLL, SpERT.