Graph Convolutional Encoders for Syntax-aware Neural Machine Translation

Open Access
Authors
Publication date 2017
Host editors
  • M. Palmer
  • R. Hwa
  • S. Riedel
Book title The Conference on Empirical Methods in Natural Language Processing
Book subtitle proceedings of the conference : EMNLP 2017 : September 9-11, 2017, Copenhagen, Denmark
ISBN (electronic)
  • 9781945626838
Event 2017 Conference on Empirical Methods in Natural Language Processing, EMNLP 2017
Pages (from-to) 1957-1967
Publisher Stroudsburg, PA: Association for Computational Linguistics
Organisations
  • Interfacultary Research - Institute for Logic, Language and Computation (ILLC)
Abstract
We present a simple and effective approach to incorporating syntactic structure into neural attention-based encoder-decoder models for machine translation. We rely on graph-convolutional networks (GCNs), a recent class of neural networks developed for modeling graph-structured data. Our GCNs use predicted syntactic dependency trees of source sentences to produce representations of words (i.e. hidden states of the encoder) that are sensitive to their syntactic neighborhoods. GCNs take word representations as input and produce word representations as output, so they can easily be incorporated as layers into standard encoders (e.g., on top of bidirectional RNNs or convolutional neural networks). We evaluate their effectiveness with English-German and English-Czech translation experiments for different types of encoders and observe substantial improvements over their syntax-agnostic versions in all the considered setups.
Document type Conference contribution
Note Later version also available.
Language English
Published at https://doi.org/10.18653/v1/D17-1209
Downloads
D17-1209v2 (Other version)
Permalink to this page
Back