47
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: not found
      • Article: not found

      Identifying facial phenotypes of genetic disorders using deep learning

      Read this article at

      ScienceOpenPublisherPubMed
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Related collections

          Most cited references21

          • Record: found
          • Abstract: not found
          • Conference Proceedings: not found

          DeepFace: Closing the Gap to Human-Level Performance in Face Verification

            • Record: found
            • Abstract: found
            • Article: found
            Is Open Access

            FaceNet: A Unified Embedding for Face Recognition and Clustering

            , , (2015)
            Despite significant recent advances in the field of face recognition, implementing face verification and recognition efficiently at scale presents serious challenges to current approaches. In this paper we present a system, called FaceNet, that directly learns a mapping from face images to a compact Euclidean space where distances directly correspond to a measure of face similarity. Once this space has been produced, tasks such as face recognition, verification and clustering can be easily implemented using standard techniques with FaceNet embeddings as feature vectors. Our method uses a deep convolutional network trained to directly optimize the embedding itself, rather than an intermediate bottleneck layer as in previous deep learning approaches. To train, we use triplets of roughly aligned matching / non-matching face patches generated using a novel online triplet mining method. The benefit of our approach is much greater representational efficiency: we achieve state-of-the-art face recognition performance using only 128-bytes per face. On the widely used Labeled Faces in the Wild (LFW) dataset, our system achieves a new record accuracy of 99.63%. On YouTube Faces DB it achieves 95.12%. Our system cuts the error rate in comparison to the best published result by 30% on both datasets. We also introduce the concept of harmonic embeddings, and a harmonic triplet loss, which describe different versions of face embeddings (produced by different networks) that are compatible to each other and allow for direct comparison between each other.
              • Record: found
              • Abstract: found
              • Article: found
              Is Open Access

              Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification

              Rectified activation units (rectifiers) are essential for state-of-the-art neural networks. In this work, we study rectifier neural networks for image classification from two aspects. First, we propose a Parametric Rectified Linear Unit (PReLU) that generalizes the traditional rectified unit. PReLU improves model fitting with nearly zero extra computational cost and little overfitting risk. Second, we derive a robust initialization method that particularly considers the rectifier nonlinearities. This method enables us to train extremely deep rectified models directly from scratch and to investigate deeper or wider network architectures. Based on our PReLU networks (PReLU-nets), we achieve 4.94% top-5 test error on the ImageNet 2012 classification dataset. This is a 26% relative improvement over the ILSVRC 2014 winner (GoogLeNet, 6.66%). To our knowledge, our result is the first to surpass human-level performance (5.1%, Russakovsky et al.) on this visual recognition challenge.

                Author and article information

                Journal
                Nature Medicine
                Nat Med
                Springer Nature
                1078-8956
                1546-170X
                January 2019
                January 7 2019
                January 2019
                : 25
                : 1
                : 60-64
                Article
                10.1038/s41591-018-0279-0
                30617323
                3b526ab9-52b2-49fc-8160-2872c799fe91
                © 2019

                http://www.springer.com/tdm

                History

                Comments

                Comment on this article

                Related Documents Log