6
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Article: found
      Is Open Access

      A solution to the learning dilemma for recurrent networks of spiking neurons

      research-article

      Read this article at

      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          Recurrently connected networks of spiking neurons underlie the astounding information processing capabilities of the brain. Yet in spite of extensive research, how they can learn through synaptic plasticity to carry out complex network computations remains unclear. We argue that two pieces of this puzzle were provided by experimental data from neuroscience. A mathematical result tells us how these pieces need to be combined to enable biologically plausible online network learning through gradient descent, in particular deep reinforcement learning. This learning method–called e-prop–approaches the performance of backpropagation through time (BPTT), the best-known method for training recurrent neural networks in machine learning. In addition, it suggests a method for powerful on-chip learning in energy-efficient spike-based hardware for artificial intelligence.

          Abstract

          Bellec et al. present a mathematically founded approximation for gradient descent training of recurrent neural networks without backwards propagation in time. This enables biologically plausible training of spike-based neural network models with working memory and supports on-chip training of neuromorphic hardware.

          Related collections

          Most cited references28

          • Record: found
          • Abstract: found
          • Article: not found

          LSTM: A Search Space Odyssey

          Several variants of the long short-term memory (LSTM) architecture for recurrent neural networks have been proposed since its inception in 1995. In recent years, these networks have become the state-of-the-art models for a variety of machine learning problems. This has led to a renewed interest in understanding the role and utility of various computational components of typical LSTM variants. In this paper, we present the first large-scale analysis of eight LSTM variants on three representative tasks: speech recognition, handwriting recognition, and polyphonic music modeling. The hyperparameters of all LSTM variants for each task were optimized separately using random search, and their importance was assessed using the powerful functional ANalysis Of VAriance framework. In total, we summarize the results of 5400 experimental runs ( ≈ 15 years of CPU time), which makes our study the largest of its kind on LSTM networks. Our results show that none of the variants can improve upon the standard LSTM architecture significantly, and demonstrate the forget gate and the output activation function to be its most critical components. We further observe that the studied hyperparameters are virtually independent and derive guidelines for their efficient adjustment.
            Bookmark
            • Record: found
            • Abstract: not found
            • Article: not found

            Backpropagation through time: what it does and how to do it

              Bookmark
              • Record: found
              • Abstract: found
              • Article: not found

              Generating coherent patterns of activity from chaotic neural networks.

              Neural circuits display complex activity patterns both spontaneously and when responding to a stimulus or generating a motor output. How are these two forms of activity related? We develop a procedure called FORCE learning for modifying synaptic strengths either external to or within a model neural network to change chaotic spontaneous activity into a wide variety of desired activity patterns. FORCE learning works even though the networks we train are spontaneously chaotic and we leave feedback loops intact and unclamped during learning. Using this approach, we construct networks that produce a wide variety of complex output patterns, input-output transformations that require memory, multiple outputs that can be switched by control inputs, and motor patterns matching human motion capture data. Our results reproduce data on premovement activity in motor and premotor cortex, and suggest that synaptic plasticity may be a more rapid and powerful modulator of network activity than generally appreciated.
                Bookmark

                Author and article information

                Contributors
                maass@igi.tugraz.at
                Journal
                Nat Commun
                Nat Commun
                Nature Communications
                Nature Publishing Group UK (London )
                2041-1723
                17 July 2020
                17 July 2020
                2020
                : 11
                : 3625
                Affiliations
                ISNI 0000 0001 2294 748X, GRID grid.410413.3, Institute of Theoretical Computer Science, , Graz University of Technology, ; Inffeldgasse 16b, Graz, Austria
                Author information
                http://orcid.org/0000-0002-4278-9527
                http://orcid.org/0000-0002-7333-9860
                http://orcid.org/0000-0001-9183-5852
                http://orcid.org/0000-0002-8724-5507
                http://orcid.org/0000-0002-1178-087X
                Article
                17236
                10.1038/s41467-020-17236-y
                7367848
                32681001
                6648bdba-7b63-4c2d-b918-cdba49a4616d
                © The Author(s) 2020

                Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.

                History
                : 9 December 2019
                : 16 June 2020
                Funding
                Funded by: Human Brain Project (Grand Agreement number 785907) and the SYNCH project (Grand Agreement number 824162) of the European Union.
                Funded by: Human Brain Project (Grand Agreement number 785907), the SYNCH project (Grand Agreement number 824162) of the European Union and a grant from Intel.
                Categories
                Article
                Custom metadata
                © The Author(s) 2020

                Uncategorized
                network models,neuroscience,learning algorithms,synaptic plasticity,electrical and electronic engineering

                Comments

                Comment on this article