Analysing Neural Language Models: Contextual Decomposition Reveals Default Reasoning in Number and Gender Assignment
| Authors | |
|---|---|
| Publication date | 2019 |
| Host editors |
|
| Book title | The 23rd Conference on Computational Natural Language Learning |
| Book subtitle | CoNLL 2019 : proceedings of the conference : November 3-4, 2019, Hong Kong, China |
| ISBN (electronic) |
|
| Event | 23rd Conference on Computational Natural Language Learning |
| Pages (from-to) | 1-11 |
| Publisher | Stroudsburg, PA: The Association for Computational Linguistics |
| Organisations |
|
| Abstract |
Extensive research has recently shown that recurrent neural language models are able to process a wide range of grammatical phenomena. How these models are able to perform these remarkable feats so well, however, is still an open question. To gain more insight into what information LSTMs base their decisions on, we propose a generalisation of Contextual Decomposition (GCD). In particular, this setup enables us to accurately distil which part of a prediction stems from semantic heuristics, which part truly emanates from syntactic cues and which part arise from the model biases themselves instead. We investigate this technique on tasks pertaining to syntactic agreement and co-reference resolution and discover that the model strongly relies on a default reasoning effect to perform these tasks.
|
| Document type | Conference contribution |
| Note | With attachment. |
| Language | English |
| Published at | https://doi.org/10.18653/v1/K19-1001 |
| Downloads |
K19-1001
(Final published version)
|
| Supplementary materials | |
| Permalink to this page | |