Towards Efficient Machine Translation Evaluation by Modelling Annotators

Nitika Mathur, Timothy Baldwin, Trevor Cohn


Abstract
Accurate evaluation of translation has long been a difficult, yet important problem. Current evaluations use direct assessment (DA), based on crowd sourcing judgements from a large pool of workers, along with quality control checks, and a robust method for combining redundant judgements. In this paper we show that the quality control mechanism is overly conservative, which increases the time and expense of the evaluation. We propose a model that does not rely on a pre-processing step to filter workers and takes into account varying annotator reliabilities. Our model effectively weights each worker's scores based on the inferred precision of the worker, and is much more reliable than the mean of either the raw scores or the standardised scores. We also show that DA does not deliver on the promise of longitudinal evaluation, and propose redesigning the structure of the annotation tasks that can solve this problem.
Anthology ID:
U18-1010
Volume:
Proceedings of the Australasian Language Technology Association Workshop 2018
Month:
December
Year:
2018
Address:
Dunedin, New Zealand
Editors:
Sunghwan Mac Kim, Xiuzhen (Jenny) Zhang
Venue:
ALTA
SIG:
Publisher:
Note:
Pages:
77–82
Language:
URL:
https://aclanthology.org/U18-1010
DOI:
Bibkey:
Cite (ACL):
Nitika Mathur, Timothy Baldwin, and Trevor Cohn. 2018. Towards Efficient Machine Translation Evaluation by Modelling Annotators. In Proceedings of the Australasian Language Technology Association Workshop 2018, pages 77–82, Dunedin, New Zealand.
Cite (Informal):
Towards Efficient Machine Translation Evaluation by Modelling Annotators (Mathur et al., ALTA 2018)
Copy Citation:
PDF:
https://aclanthology.org/U18-1010.pdf
Data
WMT 2016