On Target Representation in Continuous-output Neural Machine Translation

Open Access
Authors
Publication date 2022
Host editors
  • S. Gella
  • H. He
  • B.P. Majumder
  • B. Can
  • E. Giunchiglia
  • S. Cahyawijaya
  • S. Min
  • M. Mozes
  • X.L. Li
  • I. Augenstein
  • A. Rogers
  • K. Cho
  • E. Grefenstette
  • L. Rimell
  • C. Dyer
Book title The 7th Workshop on Representation Learning for NLP (RepL4NLP 2022)
Book subtitle proceedings of the workshop : ACL : May 26, 2022
ISBN (electronic)
  • 9781955917483
Event 7th Workshop on Representation Learning for NLP
Pages (from-to) 227–235
Publisher Stroudsburg, PA: Association for Computational Linguistics
Organisations
  • Faculty of Science (FNWI) - Informatics Institute (IVI)
Abstract
Continuous generative models proved their usefulness in high-dimensional data, such as image and audio generation. However, continuous models for text generation have received limited attention from the community. In this work, we study continuous text generation using Transformers for neural machine translation (NMT). We argue that the choice of embeddings is crucial for such models, so we aim to focus on one particular aspect”:" target representation via embeddings. We explore pretrained embeddings and also introduce knowledge transfer from the discrete Transformer model using embeddings in Euclidean and non-Euclidean spaces. Our results on the WMT Romanian-English and English-Turkish benchmarks show such transfer leads to the best-performing continuous model.
Document type Conference contribution
Note With supplementary video
Language English
Published at https://doi.org/10.18653/v1/2022.repl4nlp-1.24
Downloads
2022.repl4nlp-1.24 (Final published version)
Supplementary materials
Permalink to this page
Back