Improving entity linking by modeling latent relations between mentions

Open Access
Authors
Publication date 2018
Host editors
  • I. Gurevych
  • Y. Miyao
Book title ACL 2018 : The 56th Annual Meeting of the Association for Computational Linguistics
Book subtitle proceedings of the conference : July 15-20, 2018, Melbourne, Australia
ISBN (electronic)
  • 9781948087322
Event 56th Annual Meeting of the Association for Computational Linguistics, ACL 2018
Volume | Issue number 1
Pages (from-to) 1595-1604
Number of pages 10
Publisher Stroudsburg, PA: The Association for Computational Linguistics
Organisations
  • Interfacultary Research - Institute for Logic, Language and Computation (ILLC)
Abstract

Entity linking involves aligning textual mentions of named entities to their corresponding entries in a knowledge base. Entity linking systems often exploit relations between textual mentions in a document (e.g., coreference) to decide if the linking decisions are compatible. Unlike previous approaches, which relied on supervised systems or heuristics to predict these relations, we treat relations as latent variables in our neural entity-linking model. We induce the relations without any supervision while optimizing the entity-linking system in an end-to-end fashion. Our multi-relational model achieves the best reported scores on the standard benchmark (AIDA-CoNLL) and substantially outperforms its relation-agnostic version. Its training also converges much faster, suggesting that the injected structural bias helps to explain regularities in the training data.

Document type Conference contribution
Note With supplementary poster
Language English
Published at https://doi.org/10.18653/v1/p18-1148
Other links https://github.com/lephong/mulrel-nel https://www.scopus.com/pages/publications/85063077508
Downloads
P18-1148 (Final published version)
Supplementary materials
Permalink to this page
Back