10
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Article: found
      Is Open Access

      Reducing the Model Order of Deep Neural Networks Using Information Theory

      Preprint
      , , ,

      Read this article at

      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          Deep neural networks are typically represented by a much larger number of parameters than shallow models, making them prohibitive for small footprint devices. Recent research shows that there is considerable redundancy in the parameter space of deep neural networks. In this paper, we propose a method to compress deep neural networks by using the Fisher Information metric, which we estimate through a stochastic optimization method that keeps track of second-order information in the network. We first remove unimportant parameters and then use non-uniform fixed point quantization to assign more bits to parameters with higher Fisher Information estimates. We evaluate our method on a classification task with a convolutional neural network trained on the MNIST data set. Experimental results show that our method outperforms existing methods for both network pruning and quantization.

          Related collections

          Author and article information

          Journal
          2016-05-16
          Article
          1605.04859
          65cf6caf-968c-4b2d-b7ef-8b35632cae80

          http://arxiv.org/licenses/nonexclusive-distrib/1.0/

          History
          Custom metadata
          To appear in ISVLSI 2016 special session
          cs.LG cs.NE

          Neural & Evolutionary computing,Artificial intelligence
          Neural & Evolutionary computing, Artificial intelligence

          Comments

          Comment on this article