BERTweetFR : Domain Adaptation of Pre-Trained Language Models for French Tweets

Yanzhu Guo, Virgile Rennard, Christos Xypolopoulos, Michalis Vazirgiannis


Abstract
We introduce BERTweetFR, the first large-scale pre-trained language model for French tweets. Our model is initialised using a general-domain French language model CamemBERT which follows the base architecture of BERT. Experiments show that BERTweetFR outperforms all previous general-domain French language models on two downstream Twitter NLP tasks of offensiveness identification and named entity recognition. The dataset used in the offensiveness detection task is first created and annotated by our team, filling in the gap of such analytic datasets in French. We make our model publicly available in the transformers library with the aim of promoting future research in analytic tasks for French tweets.
Anthology ID:
2021.wnut-1.49
Volume:
Proceedings of the Seventh Workshop on Noisy User-generated Text (W-NUT 2021)
Month:
November
Year:
2021
Address:
Online
Editors:
Wei Xu, Alan Ritter, Tim Baldwin, Afshin Rahimi
Venue:
WNUT
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
445–450
Language:
URL:
https://aclanthology.org/2021.wnut-1.49
DOI:
10.18653/v1/2021.wnut-1.49
Bibkey:
Cite (ACL):
Yanzhu Guo, Virgile Rennard, Christos Xypolopoulos, and Michalis Vazirgiannis. 2021. BERTweetFR : Domain Adaptation of Pre-Trained Language Models for French Tweets. In Proceedings of the Seventh Workshop on Noisy User-generated Text (W-NUT 2021), pages 445–450, Online. Association for Computational Linguistics.
Cite (Informal):
BERTweetFR : Domain Adaptation of Pre-Trained Language Models for French Tweets (Guo et al., WNUT 2021)
Copy Citation:
PDF:
https://aclanthology.org/2021.wnut-1.49.pdf