@inproceedings{gowda-may-2020-finding,
title = "Finding the Optimal Vocabulary Size for Neural Machine Translation",
author = "Gowda, Thamme and
May, Jonathan",
booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2020",
month = nov,
year = "2020",
address = "Online",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2020.findings-emnlp.352",
doi = "10.18653/v1/2020.findings-emnlp.352",
pages = "3955--3964",
abstract = "We cast neural machine translation (NMT) as a classification task in an autoregressive setting and analyze the limitations of both classification and autoregression components. Classifiers are known to perform better with balanced class distributions during training. Since the Zipfian nature of languages causes imbalanced classes, we explore its effect on NMT. We analyze the effect of various vocabulary sizes on NMT performance on multiple languages with many data sizes, and reveal an explanation for why certain vocabulary sizes are better than others.",
}
<?xml version="1.0" encoding="UTF-8"?>
<modsCollection xmlns="http://www.loc.gov/mods/v3">
<mods ID="gowda-may-2020-finding">
<titleInfo>
<title>Finding the Optimal Vocabulary Size for Neural Machine Translation</title>
</titleInfo>
<name type="personal">
<namePart type="given">Thamme</namePart>
<namePart type="family">Gowda</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Jonathan</namePart>
<namePart type="family">May</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<originInfo>
<dateIssued>2020-11</dateIssued>
</originInfo>
<typeOfResource>text</typeOfResource>
<relatedItem type="host">
<titleInfo>
<title>Findings of the Association for Computational Linguistics: EMNLP 2020</title>
</titleInfo>
<originInfo>
<publisher>Association for Computational Linguistics</publisher>
<place>
<placeTerm type="text">Online</placeTerm>
</place>
</originInfo>
<genre authority="marcgt">conference publication</genre>
</relatedItem>
<abstract>We cast neural machine translation (NMT) as a classification task in an autoregressive setting and analyze the limitations of both classification and autoregression components. Classifiers are known to perform better with balanced class distributions during training. Since the Zipfian nature of languages causes imbalanced classes, we explore its effect on NMT. We analyze the effect of various vocabulary sizes on NMT performance on multiple languages with many data sizes, and reveal an explanation for why certain vocabulary sizes are better than others.</abstract>
<identifier type="citekey">gowda-may-2020-finding</identifier>
<identifier type="doi">10.18653/v1/2020.findings-emnlp.352</identifier>
<location>
<url>https://aclanthology.org/2020.findings-emnlp.352</url>
</location>
<part>
<date>2020-11</date>
<extent unit="page">
<start>3955</start>
<end>3964</end>
</extent>
</part>
</mods>
</modsCollection>
%0 Conference Proceedings
%T Finding the Optimal Vocabulary Size for Neural Machine Translation
%A Gowda, Thamme
%A May, Jonathan
%S Findings of the Association for Computational Linguistics: EMNLP 2020
%D 2020
%8 November
%I Association for Computational Linguistics
%C Online
%F gowda-may-2020-finding
%X We cast neural machine translation (NMT) as a classification task in an autoregressive setting and analyze the limitations of both classification and autoregression components. Classifiers are known to perform better with balanced class distributions during training. Since the Zipfian nature of languages causes imbalanced classes, we explore its effect on NMT. We analyze the effect of various vocabulary sizes on NMT performance on multiple languages with many data sizes, and reveal an explanation for why certain vocabulary sizes are better than others.
%R 10.18653/v1/2020.findings-emnlp.352
%U https://aclanthology.org/2020.findings-emnlp.352
%U https://doi.org/10.18653/v1/2020.findings-emnlp.352
%P 3955-3964
Markdown (Informal)
[Finding the Optimal Vocabulary Size for Neural Machine Translation](https://aclanthology.org/2020.findings-emnlp.352) (Gowda & May, Findings 2020)
ACL