1
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Article: found
      Is Open Access

      Self-Net: Lifelong Learning via Continual Self-Modeling

      research-article

      Read this article at

      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          Learning a set of tasks over time, also known as continual learning (CL), is one of the most challenging problems in artificial intelligence. While recent approaches achieve some degree of CL in deep neural networks, they either (1) store a new network (or an equivalent number of parameters) for each new task, (2) store training data from previous tasks, or (3) restrict the network's ability to learn new tasks. To address these issues, we propose a novel framework, Self-Net, that uses an autoencoder to learn a set of low-dimensional representations of the weights learned for different tasks. We demonstrate that these low-dimensional vectors can then be used to generate high-fidelity recollections of the original weights. Self-Net can incorporate new tasks over time with little retraining, minimal loss in performance for older tasks, and without storing prior training data. We show that our technique achieves over 10X storage compression in a continual fashion, and that it outperforms state-of-the-art approaches on numerous datasets, including continual versions of MNIST, CIFAR10, CIFAR100, Atari, and task-incremental CORe50. To the best of our knowledge, we are the first to use autoencoders to sequentially encode sets of network weights to enable continual learning.

          Related collections

          Most cited references40

          • Record: found
          • Abstract: not found
          • Article: not found

          Gradient-based learning applied to document recognition

            Bookmark
            • Record: found
            • Abstract: not found
            • Article: not found

            Learning representations by back-propagating errors

              Bookmark
              • Record: found
              • Abstract: found
              • Article: not found

              Interplay of hippocampus and prefrontal cortex in memory.

              Recent studies on the hippocampus and the prefrontal cortex have considerably advanced our understanding of the distinct roles of these brain areas in the encoding and retrieval of memories, and of how they interact in the prolonged process by which new memories are consolidated into our permanent storehouse of knowledge. These studies have led to a new model of how the hippocampus forms and replays memories and how the prefrontal cortex engages representations of the meaningful contexts in which related memories occur, as well as how these areas interact during memory retrieval. Furthermore, they have provided new insights into how interactions between the hippocampus and prefrontal cortex support the assimilation of new memories into pre-existing networks of knowledge, called schemas, and how schemas are modified in this process as the foundation of memory consolidation. Copyright © 2013 Elsevier Ltd. All rights reserved.
                Bookmark

                Author and article information

                Contributors
                Journal
                Front Artif Intell
                Front Artif Intell
                Front. Artif. Intell.
                Frontiers in Artificial Intelligence
                Frontiers Media S.A.
                2624-8212
                09 April 2020
                2020
                : 3
                : 19
                Affiliations
                Department of Computer Science, Georgia State University , Atlanta, GA, United States
                Author notes

                Edited by: Balaraman Ravindran, Indian Institute of Technology Madras, India

                Reviewed by: German I. Parisi, University of Hamburg, Germany; Srinivasa Chakravarthy, Indian Institute of Technology Madras, India

                *Correspondence: Rolando Estrada restrada1@ 123456gsu.edu

                This article was submitted to Machine Learning and Artificial Intelligence, a section of the journal Frontiers in Artificial Intelligence

                †These authors have contributed equally to this work

                Article
                10.3389/frai.2020.00019
                7861283
                d8fa17a9-3be7-418b-aadd-decc73843396
                Copyright © 2020 Mandivarapu, Camp and Estrada.

                This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

                History
                : 02 October 2019
                : 17 March 2020
                Page count
                Figures: 9, Tables: 0, Equations: 5, References: 41, Pages: 14, Words: 8019
                Funding
                Funded by: Directorate for Computer and Information Science and Engineering 10.13039/100000083
                Award ID: 1849946
                Categories
                Artificial Intelligence
                Original Research

                deep learning,continual learning,autoencoders,manifold learning,catastrophic forgetting

                Comments

                Comment on this article