A comparison of architectures and pretraining methods for contextualized multilingual word embeddings
| Authors | |
|---|---|
| Publication date | 2020 |
| Book title | AAAI-20, IAAI-20, EAAI-20 proceedings |
| Book subtitle | Thirty-Fourth AAAI Conference on Artificial Intelligence, Thirty-Second Conference on Innovative Applications of Artificial Intelligence, The Tenth Symposium on Educational Advances in Artificial Intelligence : February 7–12th, 2020, New York Hilton Midtown, New York, New York, USA |
| ISBN |
|
| Series | Proceedings of the AAAI Conference on Artificial Intelligence |
| Event | 34th AAAI Conference on Artificial Intelligence, AAAI 2020 |
| Volume | Issue number | 5 |
| Pages (from-to) | 9090-9097 |
| Publisher | Palo Alto, California: AAAI Press |
| Organisations |
|
| Abstract |
The lack of annotated data in many languages is a well-known challenge within the field of multilingual natural language processing (NLP). Therefore, many recent studies focus on zero-shot transfer learning and joint training across languages to overcome data scarcity for low-resource languages. In this work we (i) perform a comprehensive comparison of state-of-the-art multilingual word and sentence encoders on the tasks of named entity recognition (NER) and part of speech (POS) tagging; and (ii) propose a new method for creating multilingual contextualized word embeddings, compare it to multiple baselines and show that it performs at or above state-of-the-art level in zero-shot transfer settings. Finally, we show that our method allows for better knowledge sharing across languages in a joint training setting.
|
| Document type | Conference contribution |
| Language | English |
| Published at | https://doi.org/10.1609/aaai.v34i05.6443 |
| Permalink to this page | |