12
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Article: found
      Is Open Access

      High-throughput adaptive sampling for whole-slide histopathology image analysis (HASHI) via convolutional neural networks: Application to invasive breast cancer detection

      research-article

      Read this article at

      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          Precise detection of invasive cancer on whole-slide images (WSI) is a critical first step in digital pathology tasks of diagnosis and grading. Convolutional neural network (CNN) is the most popular representation learning method for computer vision tasks, which have been successfully applied in digital pathology, including tumor and mitosis detection. However, CNNs are typically only tenable with relatively small image sizes (200 × 200 pixels). Only recently, Fully convolutional networks (FCN) are able to deal with larger image sizes (500 × 500 pixels) for semantic segmentation. Hence, the direct application of CNNs to WSI is not computationally feasible because for a WSI, a CNN would require billions or trillions of parameters. To alleviate this issue, this paper presents a novel method, High-throughput Adaptive Sampling for whole-slide Histopathology Image analysis (HASHI), which involves: i) a new efficient adaptive sampling method based on probability gradient and quasi-Monte Carlo sampling, and, ii) a powerful representation learning classifier based on CNNs. We applied HASHI to automated detection of invasive breast cancer on WSI. HASHI was trained and validated using three different data cohorts involving near 500 cases and then independently tested on 195 studies from The Cancer Genome Atlas. The results show that (1) the adaptive sampling method is an effective strategy to deal with WSI without compromising prediction accuracy by obtaining comparative results of a dense sampling (∼6 million of samples in 24 hours) with far fewer samples (∼2,000 samples in 1 minute), and (2) on an independent test dataset, HASHI is effective and robust to data from multiple sites, scanners, and platforms, achieving an average Dice coefficient of 76%.

          Related collections

          Most cited references28

          • Record: found
          • Abstract: found
          • Article: found
          Is Open Access

          Deep Learning in Neural Networks: An Overview

          (2014)
          In recent years, deep artificial neural networks (including recurrent ones) have won numerous contests in pattern recognition and machine learning. This historical survey compactly summarises relevant work, much of it from the previous millennium. Shallow and deep learners are distinguished by the depth of their credit assignment paths, which are chains of possibly learnable, causal links between actions and effects. I review deep supervised learning (also recapitulating the history of backpropagation), unsupervised learning, reinforcement learning & evolutionary computation, and indirect search for short programs encoding deep and large networks.
            Bookmark
            • Record: found
            • Abstract: found
            • Article: not found

            Breast cancer histopathology image analysis: a review.

            This paper presents an overview of methods that have been proposed for the analysis of breast cancer histopathology images. This research area has become particularly relevant with the advent of whole slide imaging (WSI) scanners, which can perform cost-effective and high-throughput histopathology slide digitization, and which aim at replacing the optical microscope as the primary tool used by pathologist. Breast cancer is the most prevalent form of cancers among women, and image analysis methods that target this disease have a huge potential to reduce the workload in a typical pathology lab and to improve the quality of the interpretation. This paper is meant as an introduction for nonexperts. It starts with an overview of the tissue preparation, staining and slide digitization processes followed by a discussion of the different image processing techniques and applications, ranging from analysis of tissue staining to computer-aided diagnosis, and prognosis of breast cancer patients.
              Bookmark
              • Record: found
              • Abstract: found
              • Article: not found

              A Deep Convolutional Neural Network for segmenting and classifying epithelial and stromal regions in histopathological images.

              Epithelial (EP) and stromal (ST) are two types of tissues in histological images. Automated segmentation or classification of EP and ST tissues is important when developing computerized system for analyzing the tumor microenvironment. In this paper, a Deep Convolutional Neural Networks (DCNN) based feature learning is presented to automatically segment or classify EP and ST regions from digitized tumor tissue microarrays (TMAs). Current approaches are based on handcraft feature representation, such as color, texture, and Local Binary Patterns (LBP) in classifying two regions. Compared to handcrafted feature based approaches, which involve task dependent representation, DCNN is an end-to-end feature extractor that may be directly learned from the raw pixel intensity value of EP and ST tissues in a data driven fashion. These high-level features contribute to the construction of a supervised classifier for discriminating the two types of tissues. In this work we compare DCNN based models with three handcraft feature extraction based approaches on two different datasets which consist of 157 Hematoxylin and Eosin (H&E) stained images of breast cancer and 1376 immunohistological (IHC) stained images of colorectal cancer, respectively. The DCNN based feature learning approach was shown to have a F1 classification score of 85%, 89%, and 100%, accuracy (ACC) of 84%, 88%, and 100%, and Matthews Correlation Coefficient (MCC) of 86%, 77%, and 100% on two H&E stained (NKI and VGH) and IHC stained data, respectively. Our DNN based approach was shown to outperform three handcraft feature extraction based approaches in terms of the classification of EP and ST regions.
                Bookmark

                Author and article information

                Contributors
                Role: ConceptualizationRole: Data curationRole: Formal analysisRole: Funding acquisitionRole: InvestigationRole: MethodologyRole: ResourcesRole: SoftwareRole: VisualizationRole: Writing – original draftRole: Writing – review & editing
                Role: Data curationRole: ValidationRole: Writing – review & editing
                Role: Data curationRole: ValidationRole: Writing – review & editing
                Role: Data curationRole: ValidationRole: Writing – review & editing
                Role: Data curationRole: ValidationRole: Writing – review & editing
                Role: Data curationRole: ValidationRole: Writing – review & editing
                Role: Data curationRole: Writing – review & editing
                Role: ConceptualizationRole: Funding acquisitionRole: ResourcesRole: SupervisionRole: ValidationRole: Writing – review & editing
                Role: ConceptualizationRole: Formal analysisRole: Funding acquisitionRole: MethodologyRole: ResourcesRole: SoftwareRole: SupervisionRole: ValidationRole: Writing – original draftRole: Writing – review & editing
                Role: Editor
                Journal
                PLoS One
                PLoS ONE
                plos
                plosone
                PLoS ONE
                Public Library of Science (San Francisco, CA USA )
                1932-6203
                2018
                24 May 2018
                : 13
                : 5
                : e0196828
                Affiliations
                [1 ] School of Engineering, Universidad de los Llanos, Villavicencio, Meta, Colombia
                [2 ] Dept. of Computing Systems and Industrial Engineering, Universidad Nacional de Colombia, Bogotá, Cundinamarca, Colombia
                [3 ] University Hospitals Case Medical Center, Cleveland, OH, United States of America
                [4 ] Inspirata Inc., Tampa, FL, United States of America
                [5 ] Hospital of the University of Pennsylvania, Philadelphia, PA, United States of America
                [6 ] Cancer Institute of New Jersey, New Brunswick, NJ, United States of America
                [7 ] University at Buffalo, The State University of New York, Buffalo, NY, United States of America
                [8 ] Case Western Reserve University, Cleveland, OH, United States of America
                Beijing University of Technology, CHINA
                Author notes

                Competing Interests: Drs Madabhushi, Feldman, Ganesan, and Tomaszewski are scientific consultants for the digital pathology company Inspirata Inc. Drs Madabhushi, Feldman, Ganesan, and Tomaszewski also serve on the scientific advisory board for the digital pathology company Inspirata Inc. Dr. Madabhushi also has an equity stake in Inspirata Inc. and Elucid Bioimaging Inc.

                Author information
                http://orcid.org/0000-0003-3389-8913
                Article
                PONE-D-17-28266
                10.1371/journal.pone.0196828
                5967747
                29795581
                8182f31c-1f12-4fa1-8c46-bc0586ea52b7
                © 2018 Cruz-Roa et al

                This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.

                History
                : 31 July 2017
                : 22 April 2018
                Page count
                Figures: 7, Tables: 5, Pages: 23
                Funding
                Funded by: funder-id http://dx.doi.org/10.13039/100007637, Departamento Administrativo de Ciencia, Tecnología e Innovación;
                Award ID: 528/2011
                Award Recipient :
                Funded by: Universidad de los Llanos (CO)
                Award ID: C03-F02-35-2015
                Award Recipient :
                Funded by: funder-id http://dx.doi.org/10.13039/100000054, National Cancer Institute;
                Award ID: 1U24CA199374-01, R01CA202752-01A1 R01CA208236-01A1 R21CA179327-01; R21CA195152-01
                Funded by: funder-id http://dx.doi.org/10.13039/100000062, National Institute of Diabetes and Digestive and Kidney Diseases;
                Award ID: R01DK098503-02
                Funded by: funder-id http://dx.doi.org/10.13039/100000097, National Center for Research Resources;
                Award ID: 1 C06 RR12463-01
                Funded by: DOD Prostate Cancer Synergistic Idea Development (US)
                Award ID: PC120857
                Funded by: DOD Lung Cancer Idea Development New Investigator (US)
                Award ID: LC130463
                Funded by: DOD Prostate Cancer Idea Development (US)
                Funded by: DOD Peer Reviewed Cancer Research Program (US)
                Award ID: W81XWH-16-1-0329
                Funded by: Case Comprehensive Cancer Center Pilot Grant (US)
                Funded by: VelaSano Grant from the Cleveland Clinic (US)
                Funded by: Wallace H. Coulter Foundation Program in the Department of Biomedical Engineering at Case Western Reserve University (US)
                Funded by: funder-id http://dx.doi.org/10.13039/501100002753, Universidad Nacional de Colombia;
                Funded by: Universidad de los Llanos (CO)
                Award ID: C05-F02-039-2016
                Award Recipient :
                Research reported in this publication was funded by doctoral fellowship grant from the Administrative Department of Science, Technology and Innovation - Colciencias (528/2011), Universidad Nacional de Colombia, projects C03-F02-35-2015 and C05-F02-039-2016 from Universidad de los Llanos, the National Cancer Institute of the National Institutes of Health under award numbers 1U24CA199374-01; R01CA202752-01A1; R01CA208236-01A1; R21CA179327-01; R21CA195152-01 the National Institute of Diabetes and Digestive and Kidney Diseases under award number R01DK098503-02, National Center for Research Resources under award number 1 C06 RR12463-01; the United States Department of Defense Prostate Cancer Synergistic Idea Development Award (PC120857); the United States Department of Defense Lung Cancer Idea Development New Investigator Award (LC130463); the United States Department of Defense Prostate Cancer Idea Development Award; the United States Department of Defense Peer Reviewed Cancer Research Program W81XWH-16-1-0329; the Case Comprehensive Cancer Center Pilot Grant; VelaSano Grant from the Cleveland Clinic; the Wallace H. Coulter Foundation Program in the Department of Biomedical Engineering at Case Western Reserve University. The content is solely the responsibility of the authors and does not necessarily represent the official views of the National Institutes of Health.
                Categories
                Research Article
                Medicine and Health Sciences
                Oncology
                Cancers and Neoplasms
                Breast Tumors
                Breast Cancer
                Medicine and Health Sciences
                Pathology and Laboratory Medicine
                Anatomical Pathology
                Histopathology
                Research and Analysis Methods
                Imaging Techniques
                Image Analysis
                Research and Analysis Methods
                Imaging Techniques
                Biology and Life Sciences
                Cell Biology
                Cellular Types
                Animal Cells
                Neurons
                Biology and Life Sciences
                Neuroscience
                Cellular Neuroscience
                Neurons
                Computer and Information Sciences
                Neural Networks
                Biology and Life Sciences
                Neuroscience
                Neural Networks
                Medicine and Health Sciences
                Oncology
                Cancers and Neoplasms
                Invasive Tumors
                Biology and Life Sciences
                Neuroscience
                Cognitive Science
                Cognitive Psychology
                Learning
                Biology and Life Sciences
                Psychology
                Cognitive Psychology
                Learning
                Social Sciences
                Psychology
                Cognitive Psychology
                Learning
                Biology and Life Sciences
                Neuroscience
                Learning and Memory
                Learning
                Custom metadata
                TCGA whole-slide histopathology images of breast cancer are now available from the The Cancer Genome Atlas in http://cancergenome.nih.gov/ and Genomic Data Commons (GDC) Data Portal in https://portal.gdc.cancer.gov/. The ground-truth region annotations from whole-slide histopathology images, CNN used in the paper (CS256-FC256) trained with Torch7 from HUP and UHCMC/CWRU data sets, scaled images whole-slide histopathology images from the different data cohorts (HUP, UHCMC/CWRU, CINJ, TCGA) used for training, validation and testing and their corresponding binary masks of invasive breast cancer regions annotated by pathologists are publicly available in the Dryad database: https://doi.org/10.5061/dryad.1g2nt41.

                Uncategorized
                Uncategorized

                Comments

                Comment on this article