A comparison of architectures and pretraining methods for contextualized multilingual word embeddings

Authors
Publication date 2020
Book title AAAI-20, IAAI-20, EAAI-20 proceedings
Book subtitle Thirty-Fourth AAAI Conference on Artificial Intelligence, Thirty-Second Conference on Innovative Applications of Artificial Intelligence, The Tenth Symposium on Educational Advances in Artificial Intelligence : February 7–12th, 2020, New York Hilton Midtown, New York, New York, USA
ISBN
  • 9781577358350
Series Proceedings of the AAAI Conference on Artificial Intelligence
Event 34th AAAI Conference on Artificial Intelligence, AAAI 2020
Volume | Issue number 5
Pages (from-to) 9090-9097
Publisher Palo Alto, California: AAAI Press
Organisations
  • Interfacultary Research - Institute for Logic, Language and Computation (ILLC)
Abstract
The lack of annotated data in many languages is a well-known challenge within the field of multilingual natural language processing (NLP). Therefore, many recent studies focus on zero-shot transfer learning and joint training across languages to overcome data scarcity for low-resource languages. In this work we (i) perform a comprehensive comparison of state-of-the-art multilingual word and sentence encoders on the tasks of named entity recognition (NER) and part of speech (POS) tagging; and (ii) propose a new method for creating multilingual contextualized word embeddings, compare it to multiple baselines and show that it performs at or above state-of-the-art level in zero-shot transfer settings. Finally, we show that our method allows for better knowledge sharing across languages in a joint training setting.
Document type Conference contribution
Language English
Published at https://doi.org/10.1609/aaai.v34i05.6443
Permalink to this page
Back