SlotGAN: Detecting Mentions in Text via Adversarial Distant Learning

Open Access
Authors
Publication date 2022
Host editors
  • A. Vlachos
  • P. Agrawal
  • A. Martins
  • G. Lampouras
  • C. Lyu
Book title Sixth Workshop on Structured Prediction for NLP
Book subtitle Proceedings of the Workshop : SPNLP 2022 : May 27, 2022
ISBN (electronic)
  • 9781955917513
Event 6th Workshop on Structured Prediction for NLP, SPNLP 2022
Pages (from-to) 32-39
Number of pages 8
Publisher Stroudsburg, PA: Association for Computational Linguistics
Organisations
  • Faculty of Science (FNWI) - Informatics Institute (IVI)
Abstract

We present SlotGAN, a framework for training a mention detection model that only requires unlabeled text and a gazetteer. It consists of a generator trained to extract spans from an input sentence, and a discriminator trained to determine whether a span comes from the generator, or from the gazetteer. We evaluate the method on English newswire data and compare it against supervised, weakly-supervised, and unsupervised methods. We find that the performance of the method is lower than these baselines, because it tends to generate more and longer spans, and in some cases it relies only on capitalization. In other cases, it generates spans that are valid but differ from the benchmark. When evaluated with metrics based on overlap, we find that SlotGAN performs within 95% of the precision of a supervised method, and 84% of its recall. Our results suggest that the model can generate spans that overlap well, but an additional filtering mechanism is required.

Document type Conference contribution
Language English
Published at https://doi.org/10.18653/v1/2022.spnlp-1.4
Other links https://www.scopus.com/pages/publications/85137434170
Downloads
2022.spnlp-1.4 (Final published version)
Permalink to this page
Back