@inproceedings{adebara-etal-2020-translating,
title = "Translating Similar Languages: Role of Mutual Intelligibility in Multilingual Transformers",
author = "Adebara, Ife and
Nagoudi, El Moatez Billah and
Abdul Mageed, Muhammad",
booktitle = "Proceedings of the Fifth Conference on Machine Translation",
month = nov,
year = "2020",
address = "Online",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2020.wmt-1.42",
pages = "381--386",
abstract = "In this work we investigate different approaches to translate between similar languages despite low resource limitations. This work is done as the participation of the UBC NLP research group in the WMT 2019 Similar Languages Translation Shared Task. We participated in all language pairs and performed various experiments. We used a transformer architecture for all the models and used back-translation for one of the language pairs. We explore both bilingual and multi-lingual approaches. We describe the pre-processing, training, translation and results for each model. We also investigate the role of mutual intelligibility in model performance.",
}
<?xml version="1.0" encoding="UTF-8"?>
<modsCollection xmlns="http://www.loc.gov/mods/v3">
<mods ID="adebara-etal-2020-translating">
<titleInfo>
<title>Translating Similar Languages: Role of Mutual Intelligibility in Multilingual Transformers</title>
</titleInfo>
<name type="personal">
<namePart type="given">Ife</namePart>
<namePart type="family">Adebara</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">El</namePart>
<namePart type="given">Moatez</namePart>
<namePart type="given">Billah</namePart>
<namePart type="family">Nagoudi</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Muhammad</namePart>
<namePart type="family">Abdul Mageed</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<originInfo>
<dateIssued>2020-11</dateIssued>
</originInfo>
<typeOfResource>text</typeOfResource>
<relatedItem type="host">
<titleInfo>
<title>Proceedings of the Fifth Conference on Machine Translation</title>
</titleInfo>
<originInfo>
<publisher>Association for Computational Linguistics</publisher>
<place>
<placeTerm type="text">Online</placeTerm>
</place>
</originInfo>
<genre authority="marcgt">conference publication</genre>
</relatedItem>
<abstract>In this work we investigate different approaches to translate between similar languages despite low resource limitations. This work is done as the participation of the UBC NLP research group in the WMT 2019 Similar Languages Translation Shared Task. We participated in all language pairs and performed various experiments. We used a transformer architecture for all the models and used back-translation for one of the language pairs. We explore both bilingual and multi-lingual approaches. We describe the pre-processing, training, translation and results for each model. We also investigate the role of mutual intelligibility in model performance.</abstract>
<identifier type="citekey">adebara-etal-2020-translating</identifier>
<location>
<url>https://aclanthology.org/2020.wmt-1.42</url>
</location>
<part>
<date>2020-11</date>
<extent unit="page">
<start>381</start>
<end>386</end>
</extent>
</part>
</mods>
</modsCollection>
%0 Conference Proceedings
%T Translating Similar Languages: Role of Mutual Intelligibility in Multilingual Transformers
%A Adebara, Ife
%A Nagoudi, El Moatez Billah
%A Abdul Mageed, Muhammad
%S Proceedings of the Fifth Conference on Machine Translation
%D 2020
%8 November
%I Association for Computational Linguistics
%C Online
%F adebara-etal-2020-translating
%X In this work we investigate different approaches to translate between similar languages despite low resource limitations. This work is done as the participation of the UBC NLP research group in the WMT 2019 Similar Languages Translation Shared Task. We participated in all language pairs and performed various experiments. We used a transformer architecture for all the models and used back-translation for one of the language pairs. We explore both bilingual and multi-lingual approaches. We describe the pre-processing, training, translation and results for each model. We also investigate the role of mutual intelligibility in model performance.
%U https://aclanthology.org/2020.wmt-1.42
%P 381-386
Markdown (Informal)
[Translating Similar Languages: Role of Mutual Intelligibility in Multilingual Transformers](https://aclanthology.org/2020.wmt-1.42) (Adebara et al., WMT 2020)
ACL