11
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Article: found
      Is Open Access

      CheXaid: deep learning assistance for physician diagnosis of tuberculosis using chest x-rays in patients with HIV

      research-article

      Read this article at

      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          Tuberculosis (TB) is the leading cause of preventable death in HIV-positive patients, and yet often remains undiagnosed and untreated. Chest x-ray is often used to assist in diagnosis, yet this presents additional challenges due to atypical radiographic presentation and radiologist shortages in regions where co-infection is most common. We developed a deep learning algorithm to diagnose TB using clinical information and chest x-ray images from 677 HIV-positive patients with suspected TB from two hospitals in South Africa. We then sought to determine whether the algorithm could assist clinicians in the diagnosis of TB in HIV-positive patients as a web-based diagnostic assistant. Use of the algorithm resulted in a modest but statistically significant improvement in clinician accuracy ( p = 0.002), increasing the mean clinician accuracy from 0.60 (95% CI 0.57, 0.63) without assistance to 0.65 (95% CI 0.60, 0.70) with assistance. However, the accuracy of assisted clinicians was significantly lower ( p < 0.001) than that of the stand-alone algorithm, which had an accuracy of 0.79 (95% CI 0.77, 0.82) on the same unseen test cases. These results suggest that deep learning assistance may improve clinician accuracy in TB diagnosis using chest x-rays, which would be valuable in settings with a high burden of HIV/TB co-infection. Moreover, the high accuracy of the stand-alone algorithm suggests a potential value particularly in settings with a scarcity of radiological expertise.

          Related collections

          Most cited references24

          • Record: found
          • Abstract: found
          • Article: found
          Is Open Access

          Deep-learning-assisted diagnosis for knee magnetic resonance imaging: Development and retrospective validation of MRNet

          Background Magnetic resonance imaging (MRI) of the knee is the preferred method for diagnosing knee injuries. However, interpretation of knee MRI is time-intensive and subject to diagnostic error and variability. An automated system for interpreting knee MRI could prioritize high-risk patients and assist clinicians in making diagnoses. Deep learning methods, in being able to automatically learn layers of features, are well suited for modeling the complex relationships between medical images and their interpretations. In this study we developed a deep learning model for detecting general abnormalities and specific diagnoses (anterior cruciate ligament [ACL] tears and meniscal tears) on knee MRI exams. We then measured the effect of providing the model’s predictions to clinical experts during interpretation. Methods and findings Our dataset consisted of 1,370 knee MRI exams performed at Stanford University Medical Center between January 1, 2001, and December 31, 2012 (mean age 38.0 years; 569 [41.5%] female patients). The majority vote of 3 musculoskeletal radiologists established reference standard labels on an internal validation set of 120 exams. We developed MRNet, a convolutional neural network for classifying MRI series and combined predictions from 3 series per exam using logistic regression. In detecting abnormalities, ACL tears, and meniscal tears, this model achieved area under the receiver operating characteristic curve (AUC) values of 0.937 (95% CI 0.895, 0.980), 0.965 (95% CI 0.938, 0.993), and 0.847 (95% CI 0.780, 0.914), respectively, on the internal validation set. We also obtained a public dataset of 917 exams with sagittal T1-weighted series and labels for ACL injury from Clinical Hospital Centre Rijeka, Croatia. On the external validation set of 183 exams, the MRNet trained on Stanford sagittal T2-weighted series achieved an AUC of 0.824 (95% CI 0.757, 0.892) in the detection of ACL injuries with no additional training, while an MRNet trained on the rest of the external data achieved an AUC of 0.911 (95% CI 0.864, 0.958). We additionally measured the specificity, sensitivity, and accuracy of 9 clinical experts (7 board-certified general radiologists and 2 orthopedic surgeons) on the internal validation set both with and without model assistance. Using a 2-sided Pearson’s chi-squared test with adjustment for multiple comparisons, we found no significant differences between the performance of the model and that of unassisted general radiologists in detecting abnormalities. General radiologists achieved significantly higher sensitivity in detecting ACL tears (p-value = 0.002; q-value = 0.019) and significantly higher specificity in detecting meniscal tears (p-value = 0.003; q-value = 0.019). Using a 1-tailed t test on the change in performance metrics, we found that providing model predictions significantly increased clinical experts’ specificity in identifying ACL tears (p-value < 0.001; q-value = 0.006). The primary limitations of our study include lack of surgical ground truth and the small size of the panel of clinical experts. Conclusions Our deep learning model can rapidly generate accurate clinical pathology classifications of knee MRI exams from both internal and external datasets. Moreover, our results support the assertion that deep learning models can improve the performance of clinical experts during medical imaging interpretation. Further research is needed to validate the model prospectively and to determine its utility in the clinical setting.
            Bookmark
            • Record: found
            • Abstract: found
            • Article: not found

            A Deep Learning Mammography-based Model for Improved Breast Cancer Risk Prediction

            Background Mammographic density improves the accuracy of breast cancer risk models. However, the use of breast density is limited by subjective assessment, variation across radiologists, and restricted data. A mammography-based deep learning (DL) model may provide more accurate risk prediction. Purpose To develop a mammography-based DL breast cancer risk model that is more accurate than established clinical breast cancer risk models. Materials and Methods This retrospective study included 88 994 consecutive screening mammograms in 39 571 women between January 1, 2009, and December 31, 2012. For each patient, all examinations were assigned to either training, validation, or test sets, resulting in 71 689, 8554, and 8751 examinations, respectively. Cancer outcomes were obtained through linkage to a regional tumor registry. By using risk factor information from patient questionnaires and electronic medical records review, three models were developed to assess breast cancer risk within 5 years: a risk-factor-based logistic regression model (RF-LR) that used traditional risk factors, a DL model (image-only DL) that used mammograms alone, and a hybrid DL model that used both traditional risk factors and mammograms. Comparisons were made to an established breast cancer risk model that included breast density (Tyrer-Cuzick model, version 8 [TC]). Model performance was compared by using areas under the receiver operating characteristic curve (AUCs) with DeLong test (P < .05). Results The test set included 3937 women, aged 56.20 years ± 10.04. Hybrid DL and image-only DL showed AUCs of 0.70 (95% confidence interval [CI]: 0.66, 0.75) and 0.68 (95% CI: 0.64, 0.73), respectively. RF-LR and TC showed AUCs of 0.67 (95% CI: 0.62, 0.72) and 0.62 (95% CI: 0.57, 0.66), respectively. Hybrid DL showed a significantly higher AUC (0.70) than TC (0.62; P < .001) and RF-LR (0.67; P = .01). Conclusion Deep learning models that use full-field mammograms yield substantially improved risk discrimination compared with the Tyrer-Cuzick (version 8) model. © RSNA, 2019 Online supplemental material is available for this article. See also the editorial by Sitek and Wolfe in this issue.
              Bookmark
              • Record: found
              • Abstract: found
              • Article: found
              Is Open Access

              Randomized controlled trial of a 12-week digital care program in improving low back pain

              Low back pain (LBP) is the leading cause of disability throughout the world and is economically burdensome. The recommended first line treatment for non-specific LBP is non-invasive care. A digital care program (DCP) delivering evidence-based non-invasive treatment for LBP can aid self-management by engaging patients and scales personalized therapy for patient-specific needs. We assessed the efficacy of a 12-week DCP for LBP in a two-armed, pre-registered, randomized, controlled trial (RCT). Participants were included based on self-reported duration of LBP, but those with surgery or injury to the lower back in the previous three months were excluded. The treatment group (DCP) received the 12-week DCP, consisting of sensor-guided exercise therapy, education, cognitive behavioral therapy, team and individual behavioral coaching, activity tracking, and symptom tracking – all administered remotely via an app. The control group received three digital education articles only. All participants maintained access to treatment-as-usual. At 12 weeks, an intention-to-treat analysis showed each primary outcome—Oswestry Disability Index (p < 0.001), Korff Pain (p < 0.001) and Korff Disability (p < 0.001)—as well as each secondary outcome improved more for participants in the DCP group compared to control group. For participants who completed the DCP (per protocol), average improvement in pain outcomes ranged 52-64% (Korff: 48.8–23.4, VAS: 43.6–16.5, VAS impact on daily life: 37.3–13.4; p < 0.01 for all) and average improvement in disability outcomes ranged 31–55% (Korff: 33.1–15, ODI: 19.7–13.5; p < 0.01 for both). Surgical interest significantly reduced in the DCP group. Participants that completed the DCP had an average engagement, each week, of 90%. Future studies will further explore the effectiveness of the DCP for long-term outcomes beyond 12 weeks and for a LBP patient population with possibly greater baseline pain and disability. In conclusion, the DCP resulted in improved LBP outcomes compared to treatment-as-usual and has potential to scale personalized evidence-based non-invasive treatment for LBP patients.
                Bookmark

                Author and article information

                Contributors
                pranavsr@cs.stanford.edu
                Journal
                NPJ Digit Med
                NPJ Digit Med
                NPJ Digital Medicine
                Nature Publishing Group UK (London )
                2398-6352
                9 September 2020
                9 September 2020
                2020
                : 3
                : 115
                Affiliations
                [1 ]GRID grid.168010.e, ISNI 0000000419368956, Stanford University Department of Computer Science, ; Stanford, CA USA
                [2 ]GRID grid.32224.35, ISNI 0000 0004 0386 9924, Massachusetts General Hospital Department of Anesthesia, ; Boston, MA USA
                [3 ]GRID grid.168010.e, ISNI 0000000419368956, Stanford University AIMI Center, ; Stanford, CA USA
                [4 ]GRID grid.7836.a, ISNI 0000 0004 1937 1151, Department of Medicine, , University of Cape Town, ; Cape Town, South Africa
                Author information
                http://orcid.org/0000-0002-8030-3727
                http://orcid.org/0000-0003-4471-477X
                http://orcid.org/0000-0002-7335-3339
                http://orcid.org/0000-0002-4300-0372
                Article
                322
                10.1038/s41746-020-00322-2
                7481246
                31934645
                81af6cb8-233e-42df-92c9-aa5c69b9d507
                © The Author(s) 2020

                Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.

                History
                : 8 January 2020
                : 14 August 2020
                Categories
                Article
                Custom metadata
                © The Author(s) 2020

                diagnosis,machine learning
                diagnosis, machine learning

                Comments

                Comment on this article