Effective Estimation of Deep Generative Language Models

Open Access
Authors
Publication date 2020
Host editors
  • D. Jurafsky
  • J. Chai
  • N. Schluter
  • J. Tetreault
Book title The 58th Annual Meeting of the Association for Computational Linguistics
Book subtitle ACL 2020 : Proceedings of the Conference : July 5-10, 2020
ISBN (electronic)
  • 9781952148255
Event Association for Computational Linguistics 2020
Pages (from-to) 7220-7236
Publisher Stroudsburg, PA: The Association for Computational Linguistics
Organisations
  • Interfacultary Research - Institute for Logic, Language and Computation (ILLC)
Abstract
Advances in variational inference enable parameterisation of probabilistic models by deep neural networks. This combines the statistical transparency of the probabilistic modelling framework with the representational power of deep learning. Yet, due to a problem known as posterior collapse, it is difficult to estimate such models in the context of language modelling effectively. We concentrate on one such model, the variational auto-encoder, which we argue is an important building block in hierarchical probabilistic models of language. This paper contributes a sober view of the problem, a survey of techniques to address it, novel techniques, and extensions to the model. To establish a ranking of techniques, we perform a systematic comparison using Bayesian optimisation and find that many techniques perform reasonably similar, given enough resources. Still, a favourite can be named based on convenience. We also make several empirical observations and recommendations of best practices that should help researchers interested in this exciting field.
Document type Conference contribution
Language English
Published at http://10.18653/v1/2020.acl-main.646
Downloads
2020.acl-main.646 (Final published version)
Permalink to this page
Back