Proceedings of the First International Workshop on Knowledge-Enhanced Machine Translation

Arda Tezcan, Víctor M. Sánchez-Cartagena, Miquel Esplà-Gomis (Editors)


Anthology ID:
2024.kemt-1
Month:
June
Year:
2024
Address:
Sheffield, United Kingdom
Venues:
KEMT | WS
SIG:
Publisher:
European Association for Machine Translation (EAMT)
URL:
https://aclanthology.org/2024.kemt-1
DOI:
Bib Export formats:
BibTeX MODS XML EndNote
PDF:
https://aclanthology.org/2024.kemt-1.pdf

pdf bib
Proceedings of the First International Workshop on Knowledge-Enhanced Machine Translation
Arda Tezcan | Víctor M. Sánchez-Cartagena | Miquel Esplà-Gomis

pdf bib
Incorporating Hypernym Features for Improving Low-resource Neural Machine Translation
Abhisek Chakrabarty | Haiyue Song | Raj Dabre | Hideki Tanaka | Masao Utiyama

Parallel data is difficult to obtain for low-resource languages in machine translation tasks, making it crucial to leverage monolingual linguistic features as auxiliary information. This article introduces a novel integration of hypernym features into the model by combining learnable hypernym embeddings with word embeddings, providing semantic information. Experimental results based on bilingual and multilingual models showed that: (1) incorporating hypernyms improves translation quality in low-resource settings, yielding +1.7 BLEU scores for bilingual models, (2) the hypernym feature demonstrates efficacy both in isolation and in conjunction with syntactic features, and (3) the performance is influenced by the choice of feature combination operators and hypernym-path hyperparameters.

pdf bib
Exploring Inline Lexicon Injection for Cross-Domain Transfer in Neural Machine Translation
Jesujoba O. Alabi | Rachel Bawden

Domain transfer remains a challenge in machine translation (MT), particularly concerning rare or unseen words. Amongst the strategies proposed to address the issue, one of the simplest and most promising in terms of generalisation capacity is coupling the MT system with external resources such as bilingual lexicons and appending inline annotations within source sentences. This method has been shown to work well for controlled language settings, but its usability for general language (and ambiguous) MT is less certain. In this article we explore this question further, testing the strategy in a multi-domain transfer setting for German-to-English MT, using the mT5 language model fine-tuned on parallel data. We analyse the MT outputs and design evaluation strategies to understand the behaviour of such models. Our analysis using distractor annotations suggests that although improvements are not systematic according to automatic metrics, the model does learn to select appropriate translation candidates and ignore irrelevant ones, thereby exhibiting more than a systematic copying behaviour. However, we also find that the method is less successful in a higher-resource setting with a larger lexicon, suggesting that it is not a magic solution, especially when the baseline model is already exposed to a wide range of vocabulary.

pdf bib
Adding soft terminology constraints to pre-trained generic MT models by means of continued training
Tommi Nieminen

This article describes an efficient method of adding terminology support to existing machine translation models. The training of the pre-trained models is continued with parallel data where strings identified as terms in the source language data have been annotated with the lemmas of the corresponding target terms. Evaluation using standard test sets and methods confirms that continued training from generic base models can produce term models that are competitive with models specifically trained as term models.

pdf bib
Leveraging Synthetic Monolingual Data for Fuzzy-Match Augmentation in Neural Machine Translation: A Preliminary Study
Thomas Moerman | Arda Tezcan

pdf bib
Can True Zero-shot Methods with Large Language Models be Adopted for Sign Language Machine Translation?
Euan McGill | Horacio Saggion