Can Transformer be Too Compositional? Analysing Idiom Processing in Neural Machine Translation

Open Access
Authors
Publication date 2022
Host editors
  • S. Muresan
  • P. Nakov
  • A. Villavicencio
Book title The 60th Annual Meeting of the Association for Computational Linguistics
Book subtitle ACL 2022 : proceedings of the conference : May 22-27, 2022
ISBN (electronic)
  • 9781955917216
Event 60th Annual Meeting of the Association for Computational Linguistics, ACL 2022
Volume | Issue number 1
Pages (from-to) 3608-3626
Number of pages 19
Publisher Stroudsburg, PA: Association for Computational Linguistics
Organisations
  • Interfacultary Research - Institute for Logic, Language and Computation (ILLC)
Abstract

Unlike literal expressions, idioms' meanings do not directly follow from their parts, posing a challenge for neural machine translation (NMT). NMT models are often unable to translate idioms accurately and over-generate compositional, literal translations. In this work, we investigate whether the non-compositionality of idioms is reflected in the mechanics of the dominant NMT model, Transformer, by analysing the hidden states and attention patterns for models with English as source language and one of seven European languages as target language. When Transformer emits a non-literal translation - i.e. identifies the expression as idiomatic - the encoder processes idioms more strongly as single lexical units compared to literal expressions. This manifests in idioms' parts being grouped through attention and in reduced interaction between idioms and their context. In the decoder's cross-attention, figurative inputs result in reduced attention on source-side tokens. These results suggest that Transformer's tendency to process idioms as compositional expressions contributes to literal translations of idioms.

Document type Conference contribution
Note With software and video
Language English
Published at https://doi.org/10.18653/v1/2022.acl-long.252
Other links https://paperswithcode.com/paper/can-transformer-be-too-compositional-1 https://www.scopus.com/pages/publications/85139891673
Downloads
2022.acl-long.252 (Final published version)
Supplementary materials
Permalink to this page
Back