Yandex at ICML 2017: Variational Dropout Sparsifies Deep Neural Networks
Last week, Sydney, Australia hosted 34th International Conference on Machine Learning (ICML 2017), one of a few major machine learning conferences. Yandex team presented the result of the joint research project between Yandex, Skolkovo Institute of Science and Technology and National Research University Higher School of Economics. In our paper "Variational Dropout Sparsifies Deep Neural Networks" (https://arxiv.org/abs/1701.05369), we introduced Sparse Variational Dropout, a simple yet effective technique for sparsification of neural networks.
Modern deep learning models are extremely complex and may contain hundreds of millions of parameters. Just to store such models one would need hundreds of megabytes of disk space. The size of these models limits their applicability on devices with low memory and computational resources, e.g. embedded or mobile devices. It is also hard to deploy such models, as they cannot be sent to mobile devices via low-speed internet connection. Such overparameterization also leads to overfitting.
There are different methods of sparsification of deep neural networks. Pruning has been one of the most popular sparsification technique during past several years. However, this technique is purely heuristical, makes the training process more complicated and slow, and the result is not that impressive - we can do much better!
In our Sparse Variational Dropout technique, presented at ICML 2017, we minimize the KL-distance between the true posterior distribution and a fully-factorized (e.g. green or red plot) approximation with a sparsity inducing Log-Uniform prior distribution (blue plot). To address this problem we maximize the special objective function Stochastic Gradient Variational Lower Bound.
In practice Sparse VD leads to extremely high compression levels both in fully-connected and convolutional layers. It and allows us to reduce the number of parameters up to 280 times on LeNet architectures and up to 68 times on VGG-like networks with almost no decrease in accuracy.
Recently, it was shown that the CNNs are capable of memorizing the data even with random labeling. The standard dropout as well as other regularization techniques did not prevent this behaviour. Following that work, we also experimented with the random labeling of data. Most existing regularization techniques do not help in this case. However, Sparse Variational Dropout decides to remove all weights of the neural network and yield a constant prediction. This way Sparse VD provides the simplest possible architecture that performs the best on the testing set.
Sparse Variational Dropout is a novel state-of-the art method for sparsification of Deep Neural Networks. It can be used for compression of existing neural networks, for optimal architecture learning and for effective regularization.
The source code is available at our Github page https://github.com/ars-ashuha/variational-dropout-sparsifies-dnn.