Auto-Encoding Variational Neural Machine Translation

Open Access
Authors
Publication date 2019
Host editors
  • I. Augenstein
  • S. Gella
  • S. Ruder
  • K. Kann
  • J. Welbl
  • A. Conneau
  • X. Ren
  • M. Rei
Book title The 4th Workshop on Representation Learning for NLP (RepL4NLP-2019)
Book subtitle ACL 2019 : proceedings of the workshop : August 2, 2019, Florence, Italy
ISBN (electronic)
  • 9781950737352
Event 4th Workshop on Representation Learning for NLP (RepL4NLP)
Pages (from-to) 124–141
Number of pages 18
Publisher Stroudsburg, PA: The Association for Computational Linguistics
Organisations
  • Interfacultary Research - Institute for Logic, Language and Computation (ILLC)
Abstract
We present a deep generative model of bilingual sentence pairs for machine translation. The model generates source and target sentences jointly from a shared latent representation and is parameterised by neural networks. We perform efficient training using amortised variational inference and reparameterised gradients. Additionally, we discuss the statistical implications of joint modelling and propose an efficient approximation to maximum a posteriori decoding for fast test-time predictions. We demonstrate the effectiveness of our model in three machine translation scenarios: in-domain training, mixed-domain training, and learning from a mix of gold-standard and synthetic data. Our experiments show consistently that our joint formulation outperforms conditional modelling (i.e. standard neural machine translation) in all such scenarios.
Document type Conference contribution
Note This project has received funding from the Dutch Organization for Scientific Research VICI Grant No 277-89-002 and from the European Union’s Horizon 2020 research and innovation programme under grant agreement No 825299 (GoURMET).
Language English
Published at https://doi.org/10.18653/v1/W19-4315
Downloads
W19-4315 (Final published version)
Permalink to this page
Back