11
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: not found
      • Article: not found

      Deep learning ensembles for melanoma recognition in dermoscopy images

      Read this article at

      ScienceOpenPublisher
      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Related collections

          Most cited references21

          • Record: found
          • Abstract: not found
          • Conference Proceedings: not found

          DeepFace: Closing the Gap to Human-Level Performance in Face Verification

            Bookmark
            • Record: found
            • Abstract: found
            • Article: found
            Is Open Access

            FaceNet: A Unified Embedding for Face Recognition and Clustering

            , , (2015)
            Despite significant recent advances in the field of face recognition, implementing face verification and recognition efficiently at scale presents serious challenges to current approaches. In this paper we present a system, called FaceNet, that directly learns a mapping from face images to a compact Euclidean space where distances directly correspond to a measure of face similarity. Once this space has been produced, tasks such as face recognition, verification and clustering can be easily implemented using standard techniques with FaceNet embeddings as feature vectors. Our method uses a deep convolutional network trained to directly optimize the embedding itself, rather than an intermediate bottleneck layer as in previous deep learning approaches. To train, we use triplets of roughly aligned matching / non-matching face patches generated using a novel online triplet mining method. The benefit of our approach is much greater representational efficiency: we achieve state-of-the-art face recognition performance using only 128-bytes per face. On the widely used Labeled Faces in the Wild (LFW) dataset, our system achieves a new record accuracy of 99.63%. On YouTube Faces DB it achieves 95.12%. Our system cuts the error rate in comparison to the best published result by 30% on both datasets. We also introduce the concept of harmonic embeddings, and a harmonic triplet loss, which describe different versions of face embeddings (produced by different networks) that are compatible to each other and allow for direct comparison between each other.
              Bookmark
              • Record: found
              • Abstract: found
              • Article: not found

              A methodological approach to the classification of dermoscopy images.

              In this paper a methodological approach to the classification of pigmented skin lesions in dermoscopy images is presented. First, automatic border detection is performed to separate the lesion from the background skin. Shape features are then extracted from this border. For the extraction of color and texture related features, the image is divided into various clinically significant regions using the Euclidean distance transform. This feature data is fed into an optimization framework, which ranks the features using various feature selection algorithms and determines the optimal feature subset size according to the area under the ROC curve measure obtained from support vector machine classification. The issue of class imbalance is addressed using various sampling strategies, and the classifier generalization error is estimated using Monte Carlo cross validation. Experiments on a set of 564 images yielded a specificity of 92.34% and a sensitivity of 93.33%.
                Bookmark

                Author and article information

                Journal
                IBM Journal of Research and Development
                IBM J. Res. & Dev.
                IBM
                0018-8646
                0018-8646
                July 2017
                July 1 2017
                : 61
                : 4/5
                : 5:1-5:15
                Article
                10.1147/JRD.2017.2708299
                d0c6d431-efd6-4696-8a70-89f651f02352
                © 2017
                History

                Comments

                Comment on this article