@inproceedings{scherrer-etal-2019-analysing,
title = "Analysing concatenation approaches to document-level {NMT} in two different domains",
author = {Scherrer, Yves and
Tiedemann, J{\"o}rg and
Lo{\'a}iciga, Sharid},
editor = "Popescu-Belis, Andrei and
Lo{\'a}iciga, Sharid and
Hardmeier, Christian and
Xiong, Deyi",
booktitle = "Proceedings of the Fourth Workshop on Discourse in Machine Translation (DiscoMT 2019)",
month = nov,
year = "2019",
address = "Hong Kong, China",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/D19-6506",
doi = "10.18653/v1/D19-6506",
pages = "51--61",
abstract = "In this paper, we investigate how different aspects of discourse context affect the performance of recent neural MT systems. We describe two popular datasets covering news and movie subtitles and we provide a thorough analysis of the distribution of various document-level features in their domains. Furthermore, we train a set of context-aware MT models on both datasets and propose a comparative evaluation scheme that contrasts coherent context with artificially scrambled documents and absent context, arguing that the impact of discourse-aware MT models will become visible in this way. Our results show that the models are indeed affected by the manipulation of the test data, providing a different view on document-level translation quality than absolute sentence-level scores.",
}
<?xml version="1.0" encoding="UTF-8"?>
<modsCollection xmlns="http://www.loc.gov/mods/v3">
<mods ID="scherrer-etal-2019-analysing">
<titleInfo>
<title>Analysing concatenation approaches to document-level NMT in two different domains</title>
</titleInfo>
<name type="personal">
<namePart type="given">Yves</namePart>
<namePart type="family">Scherrer</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Jörg</namePart>
<namePart type="family">Tiedemann</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Sharid</namePart>
<namePart type="family">Loáiciga</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<originInfo>
<dateIssued>2019-11</dateIssued>
</originInfo>
<typeOfResource>text</typeOfResource>
<relatedItem type="host">
<titleInfo>
<title>Proceedings of the Fourth Workshop on Discourse in Machine Translation (DiscoMT 2019)</title>
</titleInfo>
<name type="personal">
<namePart type="given">Andrei</namePart>
<namePart type="family">Popescu-Belis</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Sharid</namePart>
<namePart type="family">Loáiciga</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Christian</namePart>
<namePart type="family">Hardmeier</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Deyi</namePart>
<namePart type="family">Xiong</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<originInfo>
<publisher>Association for Computational Linguistics</publisher>
<place>
<placeTerm type="text">Hong Kong, China</placeTerm>
</place>
</originInfo>
<genre authority="marcgt">conference publication</genre>
</relatedItem>
<abstract>In this paper, we investigate how different aspects of discourse context affect the performance of recent neural MT systems. We describe two popular datasets covering news and movie subtitles and we provide a thorough analysis of the distribution of various document-level features in their domains. Furthermore, we train a set of context-aware MT models on both datasets and propose a comparative evaluation scheme that contrasts coherent context with artificially scrambled documents and absent context, arguing that the impact of discourse-aware MT models will become visible in this way. Our results show that the models are indeed affected by the manipulation of the test data, providing a different view on document-level translation quality than absolute sentence-level scores.</abstract>
<identifier type="citekey">scherrer-etal-2019-analysing</identifier>
<identifier type="doi">10.18653/v1/D19-6506</identifier>
<location>
<url>https://aclanthology.org/D19-6506</url>
</location>
<part>
<date>2019-11</date>
<extent unit="page">
<start>51</start>
<end>61</end>
</extent>
</part>
</mods>
</modsCollection>
%0 Conference Proceedings
%T Analysing concatenation approaches to document-level NMT in two different domains
%A Scherrer, Yves
%A Tiedemann, Jörg
%A Loáiciga, Sharid
%Y Popescu-Belis, Andrei
%Y Loáiciga, Sharid
%Y Hardmeier, Christian
%Y Xiong, Deyi
%S Proceedings of the Fourth Workshop on Discourse in Machine Translation (DiscoMT 2019)
%D 2019
%8 November
%I Association for Computational Linguistics
%C Hong Kong, China
%F scherrer-etal-2019-analysing
%X In this paper, we investigate how different aspects of discourse context affect the performance of recent neural MT systems. We describe two popular datasets covering news and movie subtitles and we provide a thorough analysis of the distribution of various document-level features in their domains. Furthermore, we train a set of context-aware MT models on both datasets and propose a comparative evaluation scheme that contrasts coherent context with artificially scrambled documents and absent context, arguing that the impact of discourse-aware MT models will become visible in this way. Our results show that the models are indeed affected by the manipulation of the test data, providing a different view on document-level translation quality than absolute sentence-level scores.
%R 10.18653/v1/D19-6506
%U https://aclanthology.org/D19-6506
%U https://doi.org/10.18653/v1/D19-6506
%P 51-61
Markdown (Informal)
[Analysing concatenation approaches to document-level NMT in two different domains](https://aclanthology.org/D19-6506) (Scherrer et al., DiscoMT 2019)
ACL