6
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Article: found
      Is Open Access

      Extracting automata from neural networks using active learning

      research-article

      Read this article at

      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          Deep learning is one of the most advanced forms of machine learning. Most modern deep learning models are based on an artificial neural network, and benchmarking studies reveal that neural networks have produced results comparable to and in some cases superior to human experts. However, the generated neural networks are typically regarded as incomprehensible black-box models, which not only limits their applications, but also hinders testing and verifying. In this paper, we present an active learning framework to extract automata from neural network classifiers, which can help users to understand the classifiers. In more detail, we use Angluin’s L* algorithm as a learner and the neural network under learning as an oracle, employing abstraction interpretation of the neural network for answering membership and equivalence queries. Our abstraction consists of value, symbol and word abstractions. The factors that may affect the abstraction are also discussed in the paper. We have implemented our approach in a prototype. To evaluate it, we have performed the prototype on a MNIST classifier and have identified that the abstraction with interval number 2 and block size 1 × 28 offers the best performance in terms of F1 score. We also have compared our extracted DFA against the DFAs learned via the passive learning algorithms provided in LearnLib and the experimental results show that our DFA gives a better performance on the MNIST dataset.

          Related collections

          Most cited references30

          • Record: found
          • Abstract: found
          • Article: not found

          Deep learning.

          Deep learning allows computational models that are composed of multiple processing layers to learn representations of data with multiple levels of abstraction. These methods have dramatically improved the state-of-the-art in speech recognition, visual object recognition, object detection and many other domains such as drug discovery and genomics. Deep learning discovers intricate structure in large data sets by using the backpropagation algorithm to indicate how a machine should change its internal parameters that are used to compute the representation in each layer from the representation in the previous layer. Deep convolutional nets have brought about breakthroughs in processing images, video, speech and audio, whereas recurrent nets have shone light on sequential data such as text and speech.
            Bookmark
            • Record: found
            • Abstract: found
            • Article: not found

            Deep learning in neural networks: An overview

            In recent years, deep artificial neural networks (including recurrent ones) have won numerous contests in pattern recognition and machine learning. This historical survey compactly summarizes relevant work, much of it from the previous millennium. Shallow and Deep Learners are distinguished by the depth of their credit assignment paths, which are chains of possibly learnable, causal links between actions and effects. I review deep supervised learning (also recapitulating the history of backpropagation), unsupervised learning, reinforcement learning & evolutionary computation, and indirect search for short programs encoding deep and large networks.
              Bookmark
              • Record: found
              • Abstract: not found
              • Article: not found

              Test selection based on finite state models

                Bookmark

                Author and article information

                Contributors
                Journal
                PeerJ Comput Sci
                PeerJ Comput Sci
                peerj-cs
                peerj-cs
                PeerJ Computer Science
                PeerJ Inc. (San Diego, USA )
                2376-5992
                19 April 2021
                2021
                : 7
                : e436
                Affiliations
                [1 ]College of Computer Science and Software Engineering, Shenzhen University , Shenzhen, China
                [2 ]School of Computing, Engineering and Digital Technologies, Teesside University , Middlesbrough, United Kingdom
                Author information
                http://orcid.org/0000-0003-1826-6213
                http://orcid.org/0000-0003-3028-8191
                Article
                cs-436
                10.7717/peerj-cs.436
                8064235
                589a5642-c793-4681-801c-569b28b81372
                © 2021 Xu et al.

                This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, reproduction and adaptation in any medium and for any purpose provided that it is properly attributed. For attribution, the original author(s), title, publication source (PeerJ Computer Science) and either DOI or URL of the article must be cited.

                History
                : 10 November 2020
                : 17 February 2021
                Funding
                Funded by: National Natural Science Foundation of China
                Award ID: 61972260, 61772347 and 61836005
                Funded by: Guangdong Basic and Applied Basic Research Foundation
                Award ID: 2019A1515011577
                This work was supported by the National Natural Science Foundation of China (Nos. 61972260, 61772347, 61836005) and the Guangdong Basic and Applied Basic Research Foundation (No. 2019A1515011577). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
                Categories
                Artificial Intelligence
                Theory and Formal Methods

                automata learning,neural network,active learning
                automata learning, neural network, active learning

                Comments

                Comment on this article