Transport Score Climbing: Variational Inference Using Forward KL and Adaptive Neural Transport

Open Access
Authors
Publication date 08-2023
Journal Transactions on Machine Learning Research
Article number 1118
Volume | Issue number 2023
Number of pages 17
Organisations
  • Faculty of Science (FNWI) - Informatics Institute (IVI)
Abstract
Variational inference often minimizes the “reverse” Kullbeck-Leibler (KL) divergence, DKL(q||p) from the approximate distribution q to the posterior p. Recent work instead studies the “forward” KL divergence, DKL(p||q), which unlike reverse KL does not lead to variational approximations that underestimate uncertainty. To optimize the forward KL, these methods leveraged Markov chain Monte Carlo (MCMC) methods to evaluate the intractable expectation with respect to the posterior p. This paper introduces Transport Score Climbing (TSC), a method that optimizes DKL(p||q) by using Hamiltonian Monte Carlo (HMC). For improved performance the HMC chain is run on a transformed, or warped, space. A function called the transport map performs the transformation by acting as a change-of-variable from the latent variable space. TSC uses HMC samples to dynamically train the transport map while optimizing DKL(p||q). TSC leverages synergies, where better transport maps lead to better HMC sampling, which then leads to better transport maps. We demonstrate TSC on synthetic and real data, including using TSC to train variational auto-encoders. We find that TSC achieves competitive performance on the experiments.
Document type Article
Note With supplementary material
Language English
Published at https://openreview.net/forum?id=7KW7zvKd7J
Other links http://jmlr.org/tmlr/papers/
Downloads
Transport Score Climbing (Final published version)
Supplementary materials
Permalink to this page
Back