0
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Article: found
      Is Open Access

      A hybrid neural network approach for classifying diabetic retinopathy subtypes

      research-article

      Read this article at

      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          Objective

          Diabetic retinopathy is a prevalent complication among diabetic patients that, if not predicted and treated promptly, can lead to blindness. This paper proposes a method for accurately and swiftly predicting the degree of diabetic retinopathy using a hybrid neural network model. Timely prediction of diabetic retinopathy is crucial in preventing blindness associated with this condition.

          Methods

          This study aims to enhance the prediction accuracy of diabetic retinopathy by utilizing the hybrid neural network model EfficientNet and Swin Transformer. The specific methodology includes: (1) combining local and global features to accurately capture lesion characteristics by leveraging the strengths of both Swin Transformer and EfficientNet models; (2) improving prediction accuracy through a comprehensive analysis of the model’s training details and applying data augmentation techniques such as Gaussian blur to enhance the hybrid model’s performance; (3) validating the effectiveness and utility of the proposed hybrid model for diabetic retinopathy detection through extensive experimental evaluations and comparisons with other deep learning models.

          Results

          The hybrid model was trained and tested on the large-scale real-world diabetic retinopathy detection dataset APTOS 2019 Blindness Detection. The experimental results show that the hybrid model in this paper achieves the best results in all metrics, including sensitivity of 0.95, specificity of 0.98, accuracy of 0.97, and AUC of 0.97. The performance of the model is significantly improved compared to the mainstream methods currently employed. In addition, the model provides interpretable neural network details through class activation maps, which enables the visualization of diabetic retinopathy. This feature helps physicians to make more accurate diagnosis and treatment decisions. The model proposed in this paper shows higher accuracy in detecting and diagnosing diabetic retinopathy, which is crucial for the treatment and rehabilitation of diabetic patients.

          Conclusion

          The hybrid neural network model based on EfficientNet and Swin Transformer significantly contributes to the prediction of diabetic retinopathy. By combining local and global features, the model achieves improved prediction accuracy. The validity and utility of the model are verified through experimental evaluations. This research provides robust support for the early diagnosis and treatment of diabetic patients.

          Related collections

          Most cited references23

          • Record: found
          • Abstract: found
          • Article: not found

          Attention Is All You Need

          The dominant sequence transduction models are based on complex recurrent or convolutional neural networks in an encoder-decoder configuration. The best performing models also connect the encoder and decoder through an attention mechanism. We propose a new simple network architecture, the Transformer, based solely on attention mechanisms, dispensing with recurrence and convolutions entirely. Experiments on two machine translation tasks show these models to be superior in quality while being more parallelizable and requiring significantly less time to train. Our model achieves 28.4 BLEU on the WMT 2014 English-to-German translation task, improving over the existing best results, including ensembles by over 2 BLEU. On the WMT 2014 English-to-French translation task, our model establishes a new single-model state-of-the-art BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small fraction of the training costs of the best models from the literature. We show that the Transformer generalizes well to other tasks by applying it successfully to English constituency parsing both with large and limited training data. 15 pages, 5 figures
            Bookmark
            • Record: found
            • Abstract: not found
            • Conference Proceedings: not found

            Swin Transformer: Hierarchical Vision Transformer using Shifted Windows

              Bookmark
              • Record: found
              • Abstract: found
              • Article: not found

              EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks

              Convolutional Neural Networks (ConvNets) are commonly developed at a fixed resource budget, and then scaled up for better accuracy if more resources are available. In this paper, we systematically study model scaling and identify that carefully balancing network depth, width, and resolution can lead to better performance. Based on this observation, we propose a new scaling method that uniformly scales all dimensions of depth/width/resolution using a simple yet highly effective compound coefficient. We demonstrate the effectiveness of this method on scaling up MobileNets and ResNet. To go even further, we use neural architecture search to design a new baseline network and scale it up to obtain a family of models, called EfficientNets, which achieve much better accuracy and efficiency than previous ConvNets. In particular, our EfficientNet-B7 achieves state-of-the-art 84.3% top-1 accuracy on ImageNet, while being 8.4x smaller and 6.1x faster on inference than the best existing ConvNet. Our EfficientNets also transfer well and achieve state-of-the-art accuracy on CIFAR-100 (91.7%), Flowers (98.8%), and 3 other transfer learning datasets, with an order of magnitude fewer parameters. Source code is at https://github.com/tensorflow/tpu/tree/master/models/official/efficientnet. ICML 2019
                Bookmark

                Author and article information

                Contributors
                URI : https://loop.frontiersin.org/people/2309702/overviewRole: Role: Role: Role: Role: Role: Role: Role: Role: Role:
                Role:
                URI : https://loop.frontiersin.org/people/2503952/overviewRole:
                URI : https://loop.frontiersin.org/people/1884717/overviewRole: Role: Role: Role: Role:
                Journal
                Front Med (Lausanne)
                Front Med (Lausanne)
                Front. Med.
                Frontiers in Medicine
                Frontiers Media S.A.
                2296-858X
                04 January 2024
                2023
                : 10
                : 1293019
                Affiliations
                [1] 1The School of Medical Information Engineering, Anhui University of Chinese Medicine , Hefei, China
                [2] 2NHC Key Laboratory of Hormones and Development, Chu Hsien-I Memorial Hospital and Tianjin Institute of Endocrinology, Tianjin Medical University , Tianjin, China
                [3] 3The School of Medical Information Engineering, Guangzhou University of Chinese Medicine , Guangzhou, China
                Author notes

                Edited by: Binh P. Nguyen, Victoria University of Wellington, New Zealand

                Reviewed by: Liang Zhou, Shanghai Jiao Tong University, China

                Jin Zhong, Hefei Normal University, China

                *Correspondence: Fangliang Huang, hfl@ 123456ahtcm.edu.cn
                Article
                10.3389/fmed.2023.1293019
                10794511
                38239623
                9bc30670-4467-4d97-a220-019fe041b74c
                Copyright © 2024 Xu, Shao, Fang and Huang.

                This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

                History
                : 12 September 2023
                : 07 December 2023
                Page count
                Figures: 10, Tables: 1, Equations: 15, References: 23, Pages: 13, Words: 7263
                Funding
                Funded by: Key Project of Scientific Research in Anhui Higher Education Institutions of China
                Award ID: KJ2021A0587
                Award ID: SK2020A0244
                Funded by: Provincial Quality Engineering Project in Anhui Higher Education Institutions of China
                Award ID: 2020jyxm1029
                Funded by: Anhui Provincial Quality Engineering Project
                Award ID: 2022jyxm857
                The author(s) declare financial support was received for the research, authorship, and/or publication of this article. The research is supported by the Key Project of Scientific Research in Anhui Higher Education Institutions of China under Grant Nos. KJ2021A0587, 2023AH050770, 2023AH050780 and SK2020A0244, the Provincial Quality Engineering Project in Anhui Higher Education Institutions of China under Grant No. 2020jyxm1029 and Anhui Provincial Quality Engineering Project No. 2022jyxm857.
                Categories
                Medicine
                Original Research
                Custom metadata
                Ophthalmology

                diabetic retinopathy classifications,hybrid neural network,efficientnet,swin transformer,cad systems

                Comments

                Comment on this article