Why are Sequence-to-Sequence Models So Dull? Understanding the Low-Diversity Problem of Chatbots

Open Access
Authors
Publication date 2018
Host editors
  • A. Chuklin
  • J. Dalton
  • J. Kiseleva
  • A. Borisov
  • M. Burtsev
Book title Search-Oriented Conversational AI (SCAI)
Book subtitle EMNLP 2018 : Proceedings of the 2018 EMNLP Workshop SCAI: The 2nd International Workshop on Search-Oriented Conversational AI : October 31, 2018, Brussels, Belgium
ISBN (electronic)
  • 9781948087759
Event 2nd International Workshop on Search-Oriented Conversational AI
Pages (from-to) 81-86
Number of pages 6
Publisher Association for Computational Linguistics
Organisations
  • Faculty of Science (FNWI) - Informatics Institute (IVI)
Abstract
Diversity is a long-studied topic in information retrieval that usually refers to the requirement that retrieved results should be non-repetitive and cover different aspects. In a conversational setting, an additional dimension of diversity matters: an engaging response generation system should be able to output responses that are diverse and interesting. Sequence-to-sequence (Seq2Seq) models have been shown to be very effective for response generation. However, dialogue responses generated by Seq2Seq models tend to have low diversity. In this paper, we review known sources and existing approaches to this low-diversity problem. We also identify a source of low diversity that has been little studied so far, namely model over-confidence. We sketch several directions for tackling model over-confidence and, hence, the low-diversity problem, including confidence penalties and label smoothing.
Document type Conference contribution
Language English
Published at https://doi.org/10.48550/arXiv.1809.01941 https://doi.org/10.18653/v1/W18-5712
Published at https://arxiv.org/abs/1809.01941
Downloads
W18-5712 (Final published version)
Permalink to this page
Back