110
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Article: found
      Is Open Access

      Net2Net: Accelerating Learning via Knowledge Transfer

      Preprint
      , ,

      Read this article at

      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          We introduce techniques for rapidly transferring the information stored in one neural net into another neural net. The main purpose is to accelerate the training of a significantly larger neural net. During real-world workflows, one often trains very many different neural networks during the experimentation and design process. This is a wasteful process in which each new model is trained from scratch. Our Net2Net technique accelerates the experimentation process by instantaneously transferring the knowledge from a previous network to each new deeper or wider network. Our techniques are based on the concept of function-preserving transformations between neural network specifications. This differs from previous approaches to pre-training that altered the function represented by a neural net when adding layers to it. Using our knowledge transfer mechanism to add depth to Inception modules, we demonstrate a new state of the art accuracy rating on the ImageNet dataset.

          Related collections

          Author and article information

          Journal
          2015-11-17
          2016-04-23
          Article
          1511.05641
          5770b9eb-213a-49f7-aba9-52fc9f519269

          http://arxiv.org/licenses/nonexclusive-distrib/1.0/

          History
          Custom metadata
          ICLR 2016 submission
          cs.LG

          Artificial intelligence
          Artificial intelligence

          Comments

          Comment on this article