Boosting Neural Machine Translation with Similar Translations

Jitao XU, Josep Crego, Jean Senellart

Abstract Paper Share

Machine Translation Long Paper

Session 2B: Jul 6 (09:00-10:00 GMT)
Session 3A: Jul 6 (12:00-13:00 GMT)
Abstract: This paper explores data augmentation methods for training Neural Machine Translation to make use of similar translations, in a comparable way a human translator employs fuzzy matches. In particular, we show how we can simply present the neural model with information of both source and target sides of the fuzzy matches, we also extend the similarity to include semantically related translations retrieved using sentence distributed representations. We show that translations based on fuzzy matching provide the model with ``copy'' information while translations based on embedding similarities tend to extend the translation ``context''. Results indicate that the effect from both similar sentences are adding up to further boost accuracy, combine naturally with model fine-tuning and are providing dynamic adaptation for unseen translation pairs. Tests on multiple data sets and domains show consistent accuracy improvements. To foster research around these techniques, we also release an Open-Source toolkit with efficient and flexible fuzzy-match implementation.
You can open the pre-recorded video in a separate window.
NOTE: The SlidesLive video may display a random order of the authors. The correct author list is shown at the top of this webpage.

Similar Papers