Trainable Greedy Decoding for Neural Machine Translation

Jiatao Gu, Kyunghyun Cho, Victor O.K. Li


Abstract
Recent research in neural machine translation has largely focused on two aspects; neural network architectures and end-to-end learning algorithms. The problem of decoding, however, has received relatively little attention from the research community. In this paper, we solely focus on the problem of decoding given a trained neural machine translation model. Instead of trying to build a new decoding algorithm for any specific decoding objective, we propose the idea of trainable decoding algorithm in which we train a decoding algorithm to find a translation that maximizes an arbitrary decoding objective. More specifically, we design an actor that observes and manipulates the hidden state of the neural machine translation decoder and propose to train it using a variant of deterministic policy gradient. We extensively evaluate the proposed algorithm using four language pairs and two decoding objectives and show that we can indeed train a trainable greedy decoder that generates a better translation (in terms of a target decoding objective) with minimal computational overhead.
Anthology ID:
D17-1210
Volume:
Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing
Month:
September
Year:
2017
Address:
Copenhagen, Denmark
Editors:
Martha Palmer, Rebecca Hwa, Sebastian Riedel
Venue:
EMNLP
SIG:
SIGDAT
Publisher:
Association for Computational Linguistics
Note:
Pages:
1968–1978
Language:
URL:
https://aclanthology.org/D17-1210
DOI:
10.18653/v1/D17-1210
Bibkey:
Cite (ACL):
Jiatao Gu, Kyunghyun Cho, and Victor O.K. Li. 2017. Trainable Greedy Decoding for Neural Machine Translation. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, pages 1968–1978, Copenhagen, Denmark. Association for Computational Linguistics.
Cite (Informal):
Trainable Greedy Decoding for Neural Machine Translation (Gu et al., EMNLP 2017)
Copy Citation:
PDF:
https://aclanthology.org/D17-1210.pdf
Video:
 https://aclanthology.org/D17-1210.mp4