29
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Article: not found

      Definitions, methods, and applications in interpretable machine learning

      , , , ,
      Proceedings of the National Academy of Sciences
      Proceedings of the National Academy of Sciences

      Read this article at

      ScienceOpenPublisherPMC
      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          Machine-learning models have demonstrated great success in learning complex patterns that enable them to make predictions about unobserved data. In addition to using models for prediction, the ability to interpret what a model has learned is receiving an increasing amount of attention. However, this increased focus has led to considerable confusion about the notion of interpretability. In particular, it is unclear how the wide array of proposed interpretation methods are related and what common concepts can be used to evaluate them. We aim to address these concerns by defining interpretability in the context of machine learning and introducing the predictive, descriptive, relevant (PDR) framework for discussing interpretations. The PDR framework provides 3 overarching desiderata for evaluation: predictive accuracy, descriptive accuracy, and relevancy, with relevancy judged relative to a human audience. Moreover, to help manage the deluge of interpretation methods, we introduce a categorization of existing techniques into model-based and post hoc categories, with subgroups including sparsity, modularity, and simulatability. To demonstrate how practitioners can use the PDR framework to evaluate and understand interpretations, we provide numerous real-world examples. These examples highlight the often underappreciated role played by human audiences in discussions of interpretability. Finally, based on our framework, we discuss limitations of existing methods and directions for future work. We hope that this work will provide a common vocabulary that will make it easier for both practitioners and researchers to discuss and choose from the full range of interpretation methods.

          Related collections

          Most cited references32

          • Record: found
          • Abstract: not found
          • Book: not found

          Principal Component Analysis

            Bookmark
            • Record: found
            • Abstract: not found
            • Article: not found

            Sparse coding with an overcomplete basis set: A strategy employed by V1?

              Bookmark
              • Record: found
              • Abstract: not found
              • Book: not found

              Causal Inference for Statistics, Social, and Biomedical Sciences

                Bookmark

                Author and article information

                Journal
                Proceedings of the National Academy of Sciences
                Proc Natl Acad Sci USA
                Proceedings of the National Academy of Sciences
                0027-8424
                1091-6490
                October 29 2019
                October 29 2019
                October 29 2019
                October 16 2019
                : 116
                : 44
                : 22071-22080
                Article
                10.1073/pnas.1900654116
                6825274
                31619572
                9a156fd4-e48c-4a6b-a9b0-5dece03a375c
                © 2019

                Free to read

                https://www.pnas.org/site/aboutpnas/licenses.xhtml

                History

                Comments

                Comment on this article