FlipOut: Uncovering Redundant Weights via Sign Flipping
| Authors |
|
|---|---|
| Publication date | 2021 |
| Host editors |
|
| Book title | Artificial Intelligence and Machine Learning |
| Book subtitle | 32nd Benelux Conference, BNAIC/Benelearn 2020, Leiden, The Netherlands, November 19–20, 2020 : revised selected papers |
| ISBN |
|
| ISBN (electronic) |
|
| Series | Communications in Computer and Information Science |
| Event | 32nd Benelux Conference on Artificial Intelligence and Belgian-Dutch Conference on Machine Learning, BNAIC/Benelearn 2020 |
| Pages (from-to) | 15-29 |
| Publisher | Cham: Springer |
| Organisations |
|
| Abstract |
We propose a novel pruning method which uses the oscillations around 0, i.e. sign flips, that a weight has undergone during training in order to determine its saliency. Our method can perform pruning before the network has converged, requires little tuning effort due to having good default values for its hyperparameters, and can directly target the level of sparsity desired by the user. Our experiments, performed on a variety of object classification architectures, show that it is competitive with existing methods and achieves state-of-the-art performance for levels of sparsity of 99.6% and above for 2 out of 3 of the architectures tested. For reproducibility, we release our code at https://github.com/AndreiXYZ/flipout.
|
| Document type | Conference contribution |
| Language | English |
| Related publication | FlipOut: Uncovering Redundant Weights via Sign Flipping |
| Published at | https://doi.org/10.1007/978-3-030-76640-5_2 |
| Published at | http://bnaic.liacs.leidenuniv.nl/wordpress/wp-content/uploads/papers/BNAICBENELEARN_2020_Final_paper_25.pdf https://arxiv.org/abs/2009.02594 http://bnaic.liacs.leidenuniv.nl/bnaic2020proceedings.pdf |
| Other links | https://github.com/AndreiXYZ/flipout |
| Downloads |
FlipOut
(Other version)
|
| Permalink to this page | |