Bayesian Compression for Deep Learning

Open Access
Authors
Publication date 2018
Host editors
  • U. von Luxburg
  • I. Guyon
  • S. Bengio
  • H. Wallach
  • R. Fergus
  • S.V.N. Vishwanathan
  • R. Garnett
Book title 31st Conference on Advances in Neural Information Processing Systems (NIPS 2017)
Book subtitle Long Beach, California, USA, 4-9 December 2017
ISBN
  • 9781510860964
Series Advances in Neural Information Processing Systems
Event 31st Conference on Advances in Neural Information Processing Systems
Volume | Issue number 5
Pages (from-to) 3289-3299
Publisher La Jolla, CA: Neural Information Processing Systems
Organisations
  • Faculty of Science (FNWI) - Informatics Institute (IVI)
Abstract
Compression and computational efficiency in deep learning have become a problem of great significance. In this work, we argue that the most principled and effective way to attack this problem is by adopting a Bayesian point of view, where through sparsity inducing priors we prune large parts of the network. We introduce two novelties in this paper: 1) we use hierarchical priors to prune nodes instead of individual weights, and 2) we use the posterior uncertainties to determine the optimal fixed point precision to encode the weights. Both factors significantly contribute to achieving the state of the art in terms of compression rates, while still staying competitive with methods designed to optimize for speed or energy efficiency.
Document type Conference contribution
Note With supplemental files
Language English
Published at https://papers.nips.cc/paper/6921-bayesian-compression-for-deep-learning
Other links http://www.proceedings.com/39083.html
Downloads
6921-bayesian-compression-for-deep-learning (Accepted author manuscript)
Supplementary materials
Permalink to this page
Back