Modelling Lexical Ambiguity with Density Matrices

Open Access
Authors
Publication date 2020
Host editors
  • R, Fernández
  • T. Linzen
Book title The 24th Conference on Computational Natural Language Learning (CoNNL)
Book subtitle CoNNL 2020 : proceedings of the conference : November 19-20, 2020, Online
ISBN (electronic)
  • 9781952148637
Event 24th Conference on Computational Natural Language Learning, CoNLL 2020
Pages (from-to) 276-290
Number of pages 15
Publisher Stroudsburg, PA: The Association for Computational Linguistics
Organisations
  • Interfacultary Research - Institute for Logic, Language and Computation (ILLC)
Abstract

Words can have multiple senses. Compositional distributional models of meaning have been argued to deal well with finer shades of meaning variation known as polysemy, but are not so well equipped to handle word senses that are etymologically unrelated, or homonymy. Moving from vectors to density matrices allows us to encode a probability distribution over different senses of a word, and can also be accommodated within a compositional distributional model of meaning. In this paper we present three new neural models for learning density matrices from a corpus, and test their ability to discriminate between word senses on a range of compositional datasets. When paired with a particular composition method, our best model outperforms existing vector-based compositional models as well as strong sentence encoders.

Document type Conference contribution
Note With supplementary material.
Language English
Published at https://doi.org/10.18653/v1/2020.conll-1.21
Other links https://www.scopus.com/pages/publications/85173612807
Downloads
2020.conll-1.21 (Final published version)
Supplementary materials
Permalink to this page
Back