Structured Reordering for Modeling Latent Alignments in Sequence Transduction

Open Access
Authors
Publication date 2022
Host editors
  • M. Ranzato
  • A. Beygelzimer
  • Y. Dauphin
  • P.S. Liang
  • J. Wortman Vaughan
Book title 35th Conference on Neural Information Processing Systems (NeurIPS 2021)
Book subtitle online, 6-14 December 2021
ISBN
  • 9781713845393
Series Advances in Neural Information Processing Systems
Event 35th Conference on Neural Information Processing Systems, NeurIPS 2021
Volume | Issue number 16
Pages (from-to) 13378-13391
Number of pages 14
Publisher San Diego, CA: Neural Information Processing Systems Foundation
Organisations
  • Interfacultary Research - Institute for Logic, Language and Computation (ILLC)
Abstract

Despite success in many domains, neural models struggle in settings where train and test examples are drawn from different distributions. In particular, in contrast to humans, conventional sequence-to-sequence (seq2seq) models fail to generalize systematically, i.e., interpret sentences representing novel combinations of concepts (e.g., text segments) seen in training. Traditional grammar formalisms excel in such settings by implicitly encoding alignments between input and output segments, but are hard to scale and maintain. Instead of engineering a grammar, we directly model segment-to-segment alignments as discrete structured latent variables within a neural seq2seq model. To efficiently explore the large space of alignments, we introduce a reorder-first align-later framework whose central component is a neural reordering module producing separable permutations. We present an efficient dynamic programming algorithm performing exact marginal and MAP inference of separable permutations, and, thus, enabling end-to-end differentiable training of our model. The resulting seq2seq model exhibits better systematic generalization than standard models on synthetic problems and NLP tasks (i.e., semantic parsing and machine translation).

Document type Conference contribution
Note With supplementary file.
Language English
Published at https://papers.nips.cc/paper_files/paper/2021/hash/6f46dd176364ccec308c2760189a4605-Abstract.html
Other links https://www.proceedings.com/63069.html https://www.scopus.com/pages/publications/85121628090
Downloads
Supplementary materials
Permalink to this page
Back