20
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Article: found
      Is Open Access

      Wireless capsule endoscopy multiclass classification using three-dimensional deep convolutional neural network model

      research-article

      Read this article at

      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          Background

          Wireless capsule endoscopy (WCE) is a patient-friendly and non-invasive technology that scans the whole of the gastrointestinal tract, including difficult-to-access regions like the small bowel. Major drawback of this technology is that the visual inspection of a large number of video frames produced during each examination makes the physician diagnosis process tedious and prone to error. Several computer-aided diagnosis (CAD) systems, such as deep network models, have been developed for the automatic recognition of abnormalities in WCE frames. Nevertheless, most of these studies have only focused on spatial information within individual WCE frames, missing the crucial temporal data within consecutive frames.

          Methods

          In this article, an automatic multiclass classification system based on a three-dimensional deep convolutional neural network (3D-CNN) is proposed, which utilizes the spatiotemporal information to facilitate the WCE diagnosis process. The 3D-CNN model fed with a series of sequential WCE frames in contrast to the two-dimensional (2D) model, which exploits frames as independent ones. Moreover, the proposed 3D deep model is compared with some pre-trained networks. The proposed models are trained and evaluated with 29 subject WCE videos (14,691 frames before augmentation). The performance advantages of 3D-CNN over 2D-CNN and pre-trained networks are verified in terms of sensitivity, specificity, and accuracy.

          Results

          3D-CNN outperforms the 2D technique in all evaluation metrics (sensitivity: 98.92 vs. 98.05, specificity: 99.50 vs. 86.94, accuracy: 99.20 vs. 92.60). In conclusion, a novel 3D-CNN model for lesion detection in WCE frames is proposed in this study.

          Conclusion

          The results indicate the performance of 3D-CNN over 2D-CNN and some well-known pre-trained classifier networks. The proposed 3D-CNN model uses the rich temporal information in adjacent frames as well as spatial data to develop an accurate and efficient model.

          Supplementary Information

          The online version contains supplementary material available at 10.1186/s12938-023-01186-9.

          Related collections

          Most cited references49

          • Record: found
          • Abstract: not found
          • Article: not found

          ImageNet classification with deep convolutional neural networks

            Bookmark
            • Record: found
            • Abstract: found
            • Article: not found

            3D convolutional neural networks for human action recognition.

            We consider the automated recognition of human actions in surveillance videos. Most current methods build classifiers based on complex handcrafted features computed from the raw inputs. Convolutional neural networks (CNNs) are a type of deep model that can act directly on the raw inputs. However, such models are currently limited to handling 2D inputs. In this paper, we develop a novel 3D CNN model for action recognition. This model extracts features from both the spatial and the temporal dimensions by performing 3D convolutions, thereby capturing the motion information encoded in multiple adjacent frames. The developed model generates multiple channels of information from the input frames, and the final feature representation combines information from all channels. To further boost the performance, we propose regularizing the outputs with high-level features and combining the predictions of a variety of different models. We apply the developed models to recognize human actions in the real-world environment of airport surveillance videos, and they achieve superior performance in comparison to baseline methods.
              Bookmark
              • Record: found
              • Abstract: not found
              • Conference Proceedings: not found

              Imagenet classification with deep convolutional neural networks

                Bookmark

                Author and article information

                Contributors
                ms_helfroush@sutech.ac.ir
                Journal
                Biomed Eng Online
                Biomed Eng Online
                BioMedical Engineering OnLine
                BioMed Central (London )
                1475-925X
                15 December 2023
                15 December 2023
                2023
                : 22
                : 124
                Affiliations
                [1 ]Department of Electrical Engineering, Shiraz University of Technology, ( https://ror.org/04bxa3v83) Shiraz, Iran
                [2 ]Department of Internal Medicine, Gastroenterohepatology Research Center, School of Medicine, Shiraz University of Medical Sciences, ( https://ror.org/01n3s4692) Shiraz, Iran
                Article
                1186
                10.1186/s12938-023-01186-9
                10722702
                38098015
                babeb320-8b66-4704-b507-0517560c8dc7
                © The Author(s) 2023

                Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver ( http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

                History
                : 10 August 2023
                : 29 November 2023
                Categories
                Research
                Custom metadata
                © BioMed Central Ltd., part of Springer Nature 2023

                Biomedical engineering
                wireless capsule endoscopy,image classification,deep learning,3d convolutional neural network

                Comments

                Comment on this article