Analysing the potential of seq-to-seq models for incremental interpretation in task-oriented dialogue

Open Access
Authors
Publication date 2018
Host editors
  • T. Linzen
  • G. Chrupała
  • A. Alishahi
Book title The 2018 EMNLP Workshop BlackboxNLP: Analyzing and Interpreting Neural Networks for NLP
Book subtitle EMNLP 2018 : proceedings of the First Workshop : November 1, 2018, Brussels, Belgium
ISBN (electronic)
  • 9781948087711
Event 2018 EMNLP Workshop BlackboxNLP
Pages (from-to) 165–174
Publisher Stroudsburg, PA: The Association for Computational Linguistics
Organisations
  • Interfacultary Research - Institute for Logic, Language and Computation (ILLC)
Abstract
We investigate how encoder-decoder models trained on a synthetic dataset of task-oriented dialogues process disfluencies, such as hesitations and self-corrections. We find that, contrary to earlier results, disfluencies have very little impact on the task success of seq-to-seq models with attention. Using visualisations and diagnostic classifiers, we analyse the representations that are incrementally built by the model, and discover that models develop little to no awareness of the structure of disfluencies. However, adding disfluencies to the data appears to help the model create clearer representations overall, as evidenced by the attention patterns the different models exhibit.
Document type Conference contribution
Language English
Published at https://doi.org/10.18653/v1/W18-5419
Published at https://arxiv.org/abs/1808.09178
Downloads
W18-5419 (Final published version)
Permalink to this page
Back