Self Normalizing Flows

Open Access
Authors
Publication date 2021
Journal Proceedings of Machine Learning Research
Event 38th International Conference on Machine Learning
Volume | Issue number 139
Pages (from-to) 5378-5387
Organisations
  • Faculty of Science (FNWI) - Informatics Institute (IVI)
Abstract
Efficient gradient computation of the Jacobian determinant term is a core problem in many machine learning settings, and especially so in the normalizing flow framework. Most proposed flow models therefore either restrict to a function class with easy evaluation of the Jacobian determinant, or an efficient estimator thereof. However, these restrictions limit the performance of such density models, frequently requiring significant depth to reach desired performance levels. In this work, we propose Self Normalizing Flows, a flexible framework for training normalizing flows by replacing expensive terms in the gradient by learned approximate inverses at each layer. This reduces the computational complexity of each layer’s exact update from O(D3) to O(D2), allowing for the training of flow architectures which were otherwise computationally infeasible, while also providing efficient sampling. We show experimentally that such models are remarkably stable and optimize to similar data likelihood values as their exact gradient counterparts, while training more quickly and surpassing the performance of functionally constrained counterparts.
Document type Article
Note International Conference on Machine Learning, 18-24 July 2021, Virtual. - With supplementary file.
Language English
Published at https://arxiv.org/abs/2011.07248 https://proceedings.mlr.press/v139/keller21a.html
Other links https://youtu.be/6Q3b3MergqI
Downloads
keller21a (Final published version)
Supplementary materials
Permalink to this page
Back