Big Generalizations with Small Data: Exploring the Role of Training Samples in Learning Adjectives of Size

Open Access
Authors
Publication date 2019
Host editors
  • A. Mogadala
  • D. Klakow
  • S. Pezzelle
  • M.-F. Moens
Book title Beyond Vision and LANguage: inTEgrating Real-world kNowledge (LANTERN)
Book subtitle EMNLP-IJCNLP 2019 : proceedings of the workshop : November 3, 2019, Hong Kong, China
ISBN (electronic)
  • 9781950737758
Event 1st Workshop on Beyond Vision and LANguage: inTEgrating Real-World kNowledge, LANTERN@EMNLP-IJCNLP 2019
Pages (from-to) 18-23
Publisher Stroudsburg, PA: The Association for Computational Linguistics
Organisations
  • Interfacultary Research - Institute for Logic, Language and Computation (ILLC)
Abstract
In this paper, we experiment with a recently proposed visual reasoning task dealing with quantities – modeling the multimodal, contextually-dependent meaning of size adjectives (‘big’, ‘small’) – and explore the impact of varying the training data on the learning behavior of a state-of-art system. In previous work, models have been shown to fail in generalizing to unseen adjective-noun combinations. Here, we investigate whether, and to what extent, seeing some of these cases during training helps a model understand the rule subtending the task, i.e., that being big implies being not small, and vice versa. We show that relatively few examples are enough to understand this relationship, and that developing a specific, mutually exclusive representation of size adjectives is beneficial to the task.
Document type Conference contribution
Language English
Related dataset MALeViC
Published at https://doi.org/10.18653/v1/D19-6403
Downloads
D19-6403 (Final published version)
Permalink to this page
Back