Efficient Gradient-Based Inference through Transformations between Bayes Nets and Neural Nets

Open Access
Authors
Publication date 2014
Journal JMLR Workshop and Conference Proceedings
Event International Conference on Machine Learning (ICML 2014)
Volume | Issue number 32
Pages (from-to) 1782-1790
Organisations
  • Faculty of Science (FNWI) - Informatics Institute (IVI)
Abstract
Hierarchical Bayesian networks and neural networks with stochastic hidden units are commonly perceived as two separate types of models. We show that either of these types of models can often be transformed into an instance of the other, by switching between centered and differentiable non-centered parameterizations of the latent variables. The choice of parameterization greatly influences the efficiency of gradient-based posterior inference; we show that they are often complementary to eachother, we clarify when each parameterization is preferred and show how inference can be made robust. In the non-centered form, a simple Monte Carlo estimator of the marginal likelihood can be used for learning the parameters. Theoretical results are supported by experiments.
Document type Article
Note International Conference on Machine Learning, 22-24 June 2014, Bejing, China. Editors: Eric P. Xing, Tony Jebara.
Language English
Published at http://jmlr.org/proceedings/papers/v32/kingma14.html
Downloads
kingma14 (Final published version)
Permalink to this page
Back