8
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Article: found
      Is Open Access

      Data Science Approaches for Effective Use of Mobile Device–Based Collection of Real‐World Data

      article-commentary

      Read this article at

      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          The use of mobile health for monitoring disease outside of the clinic has opened new opportunities for drug testing and monitoring. In particular, these tools are providing new experimental designs for collection of real‐world data. These technologies and queries, although promising, require the application of analytical methods that can accommodate the uncontrolled, unmonitored, individualized, and, often, near continuous data streams. Here, we discuss opportunities and ramifications on analytical considerations. Mobile health, that is, the evaluation of health outside of the clinic using wearables and smartphones, and, more broadly, the collection of real‐world evidence,1 provide opportunities to advance multiple goals for monitoring drug response, including the monitoring of efficacy through digital biomarkers that can be used as primary end points for drug efficacy, monitoring of patient‐reported outcomes and/or quality of life measures, and of toxicities and/or response to long‐term exposures. Although digital end points are of interest to regulatory agencies such as the US Food and Drug Administration (FDA) and the European Medicines Agency (EMA) and are starting to be integrated as primary end points into clinical trials,2 adoption is slow. In part, this is due to difficulties in quantifying the accuracy of measures when they are collected in an unmonitored manner and in an uncontrolled setting. Indeed, the ability to develop robust measures that are reliably accurate requires both an expanded validation plan designed to pressure‐test the measure across a range of conditions and a good understanding of the impact that variations in daily living can have on data collection. Because interpretation of mobile health data involves the processing and analysis of high dimensional, longitudinal sensor data collected in continuous or near continuous data streams, it requires the use of statistical approaches that account for repeat measures as well as extensive use of signal processing and/or machine‐learning techniques. These approaches provide opportunity for sensitive, individualized monitoring of drug responses. Here, we provide a short introduction to the importance of appropriate usage of analytic and machine‐learning techniques for the interpretation of mobile health data (see also refs. 3, 4, 5). This includes a description of the types of experiments and data that can be collected using mobile health and some examples from the literature that highlight important analytical considerations. Although these observations are relevant to any device that is collecting sensor data in a continuous or near continuous manner, we exemplify these issues using our own experience with the development and analysis of smartphone‐based measures. From Sensors to Machine Learning and Some Consequences Modern smartphones have opened a wealth of possibilities to extend electronic health monitoring for two reasons: (i) The always connected nature and computational power of smartphones allows for rapid data collection and (ii) the large number of embedded sensors allows for multimodal data collection.6 A typical phone has sensors that can measure acceleration, rotation rates, magnetic fields, sound levels, record audio and video, and record time and touch through the screen, among other capabilities. Sensor‐based data collection performed in the context of protocols, or tasks designed to capture disease relevant behavior, can be used to generate hundreds of phenotypic measurements, including those that mimic evaluations typically performed in the clinic (e.g., sit to stand test for mobility or blood pressure measurement). They can also be used to passively collect measurements during daily activities (e.g., mobility analysis performed while an individual is walking). In either case, high dimensional data streams are generated that require extensive processing and analysis to be converted into phenotypic measures. For gait analysis, inertial measuring units embedded in wearables and phones collect time series data consisting of 100 Hz recordings on 6 axes (three from the accelerometer and three from the gyroscope). Data such as these can be analyzed in three ways. First, features with established clinical relevance (e.g., gait speed) can be extracted through signal processing. For this approach, algorithms are manually evaluated and tuned to maximally approximate the desired phenotypic measure. Although this first approach provides measures with clear clinical interpretation, it can limit use of the full spectrum of information provided in the collected data. To address this, one could opt to use traditional machine learning in conjunction with signal processing to select a subset of promising features from a larger set of exploratory features generated by signal processing. This data‐driven approach might be better able to distinguish disease state across heterogeneous populations, as it works by optimizing on the outcome of interest. In the case of gait, this second approach is suited to identifying a broader set of gait disturbances in addition to gait speed. Finally, machine‐learning methods based on deep‐learning models have also been used to generate features in an automatic and data‐driven way, bypassing the need for signal processing. Although machine‐learning approaches provide the opportunity to develop more comprehensive digital measures, the use of machine learning must be done appropriately in order to avoid subtle errors. Because they are data driven, machine‐learning approaches will leverage any source of variation in a dataset, including variability due to biology, technical artifacts, and even random noise (especially in small datasets). Identification of biologically relevant measures requires disciplined analysis. This is typically addressed by using two datasets—training data is used to train models and select potential features, whereas a separate validation dataset, assumed to contain similar biological but different technical variation, is used to evaluate the predictive performance of the trained model and confirm the biological relevance of the new features (e.g., by comparing then to existing clinically validated outcomes or severity measures). Because two datasets are not always available in mobile health studies, a single dataset is often split to support both training and validation functions. This can be problematic for small datasets. There are many papers reporting positive validation results of digital measures. Many of these results are developed using machine learning in small sample size studies, which can promote exaggerated results that will not replicate in other datasets. This is best addressed by reporting the uncertainty in measure performance. As an example, a study reporting diagnostic accuracy using the area under the receiver operating curve should be expected to report error bars as a means to help readers understand the uncertainty in the reported performance. In addition, the choice of performance metric is dependent on the nature of the data. Reporting on the incorrect metric (e.g., area under the receiver operating curve in extremely imbalanced datasets) can lead to inflated interpretation of accuracy.7 Application and Consequences of Longitudinal Sampling A major benefit of mobile health is the opportunity to tailor health monitoring to each individual. This is of particular benefit for conditions and treatments that present in a highly heterogeneous manner across individuals or change dynamically over time. Because mobile health provides longitudinal data collection with frequent sampling, it can be used to capture individualized changes over time by using personalized models or n‐of‐1 analysis.3 Analysis of frequently sampled longitudinal data requires an analytical approach that is distinct from those used for sparsely sampled data. Although repeated measures collected from an individual are autocorrelated, a common mistake observed in the literature is to assume that these repeated measures are independent. If not taken into account, autocorrelation can lead to an inaccurate estimate of the number of false‐positive discoveries in an analysis. Notably, this can result in either an underestimate or an overestimate depending on whether the autocorrelation is positive or negative.3, 4 Furthermore, the incorrect use of the repeated measurements in population level analysis, such as classification studies can lead to identity confounding artifacts, where the classifier is mostly distinguishing differences across individuals instead of differences across conditions or disease states. A recent literature review of mobile health classification studies demonstrated that 47% had artificially inflated the performance of their measures through failure to account for the identity of individual data points.8 Our own quantification of this effect across three studies showed that identity confounding can be many times larger than the effect of the condition that was being studied.9 As with the analytical issues described above, proper interpretation of analyses using mobile health studies for classification requires reporting of how repeat measures were handled. Postmarket Monitoring, Open Enrollment, and The Effects of Confounders Fully remote mobile health studies can support low cost enrollment of large swaths of the population as compared to in‐clinic studies. Many studies relying solely on mobile health measures have enrolled in the tens of thousands from across distributed geographic regions, providing the opportunity for broad sampling across diverse populations in the real‐world setting. This approach can be a good option for postmarket monitoring studies, including to evaluate real‐world drug efficacy and toxicity as well as market fit. It can also be used to prescreen for enrollees into clinical trials. In these contexts, data are often collected using open enrollment techniques. Because these can lead to biased sampling of the population, they must be carefully evaluated and interpreted. For example, we recently recruited 17,000 individuals into a Parkinson’s disease study using an open enrollment approach. The control population tended to be significantly younger than the Parkinson’s disease population (average age 38 vs. 61). Because age was correlated with disease status, machine‐learning methods could trivially distinguish between cases and controls by selecting features related to age rather than those related to disease state. With careful consideration these issues can be both assessed and accounted for.10 In this case, we did so by rebalancing the populations according to clinical covariates and by measuring performance of the classifier both before and after correction for known covariates.10 Conclusions The use of mobile health to collect frequent measures in a real‐world setting provides a promising tool to aid in drug development and monitoring. Appropriate use and interpretation of these approaches, which also provide great opportunity to monitor lived experience, require careful attention to analytical techniques. Much of the success of mHealth will be dependent on comprehensive validation of developed measure and objective benchmarking of analytical techniques used in their interpretation. With appropriate application, these approaches stand to greatly advance our ability to objectively assess the impact of treatments on individuals’ lives. Funding No funding was received for this work. Conflict of Interest The authors declared no competing interests for this work.

          Related collections

          Most cited references9

          • Record: found
          • Abstract: found
          • Article: found
          Is Open Access

          The need to approximate the use-case in clinical machine learning

          Abstract The availability of smartphone and wearable sensor technology is leading to a rapid accumulation of human subject data, and machine learning is emerging as a technique to map those data into clinical predictions. As machine learning algorithms are increasingly used to support clinical decision making, it is vital to reliably quantify their prediction accuracy. Cross-validation (CV) is the standard approach where the accuracy of such algorithms is evaluated on part of the data the algorithm has not seen during training. However, for this procedure to be meaningful, the relationship between the training and the validation set should mimic the relationship between the training set and the dataset expected for the clinical use. Here we compared two popular CV methods: record-wise and subject-wise. While the subject-wise method mirrors the clinically relevant use-case scenario of diagnosis in newly recruited subjects, the record-wise strategy has no such interpretation. Using both a publicly available dataset and a simulation, we found that record-wise CV often massively overestimates the prediction accuracy of the algorithms. We also conducted a systematic review of the relevant literature, and found that this overly optimistic method was used by almost half of the retrieved studies that used accelerometers, wearable sensors, or smartphones to predict clinical outcomes. As we move towards an era of machine learning-based diagnosis and treatment, using proper methods to evaluate their accuracy is crucial, as inaccurate results can mislead both clinicians and data scientists.
            Bookmark
            • Record: found
            • Abstract: found
            • Article: found
            Is Open Access

            Best practices for analyzing large-scale health data from wearables and smartphone apps

            Smartphone apps and wearable devices for tracking physical activity and other health behaviors have become popular in recent years and provide a largely untapped source of data about health behaviors in the free-living environment. The data are large in scale, collected at low cost in the “wild”, and often recorded in an automatic fashion, providing a powerful complement to traditional surveillance studies and controlled trials. These data are helping to reveal, for example, new insights about environmental and social influences on physical activity. The observational nature of the datasets and collection via commercial devices and apps pose challenges, however, including the potential for measurement, population, and/or selection bias, as well as missing data. In this article, we review insights gleaned from these datasets and propose best practices for addressing the limitations of large-scale data from apps and wearables. Our goal is to enable researchers to effectively harness the data from smartphone apps and wearable devices to better understand what drives physical activity and other health behaviors.
              Bookmark
              • Record: found
              • Abstract: found
              • Article: found
              Is Open Access

              Use of Mobile Devices to Measure Outcomes in Clinical Research, 2010–2016: A Systematic Literature Review

              Background: The use of mobile devices in clinical research has advanced substantially in recent years due to the rapid pace of technology development. With an overall aim of informing the future use of mobile devices in interventional clinical research to measure primary outcomes, we conducted a systematic review of the use of and clinical outcomes measured by mobile devices (mobile outcomes) in observational and interventional clinical research. Method: We conducted a PubMed search using a range of search terms to retrieve peer-reviewed articles on clinical research published between January 2010 and May 2016 in which mobile devices were used to measure study outcomes. We screened each publication for specific inclusion and exclusion criteria. We then identified and qualitatively summarized the use of mobile outcome assessments in clinical research, including the type and design of the study, therapeutic focus, type of mobile device(s) used, and specific mobile outcomes reported. Results: The search retrieved 2,530 potential articles of interest. After screening, 88 publications remained. Twenty-five percent of the publications ( n = 22) described mobile outcomes used in interventional research, and the rest ( n = 66) described observational clinical research. Thirteen therapeutic areas were represented. Five categories of mobile devices were identified: (1) inertial sensors, (2) biosensors, (3) pressure sensors and walkways, (4) medication adherence monitors, and (5) location monitors; inertial sensors/accelerometers were most common (reported in 86% of the publications). Among the variety of mobile outcomes, various assessments of physical activity were most common (reported in 74% of the publications). Other mobile outcomes included assessments of sleep, mobility, and pill adherence, as well as biomarkers assessed using a mobile device, including cardiac measures, glucose, gastric reflux, respiratory measures, and intensity of head-related injury. Conclusion: Mobile devices are being widely used in clinical research to assess outcomes, although their use in interventional research to assess therapeutic effectiveness is limited. For mobile devices to be used more frequently in pivotal interventional research – such as trials informing regulatory decision-making – more focus should be placed on: (1) consolidating the evidence supporting the clinical meaningfulness of specific mobile outcomes, and (2) standardizing the use of mobile devices in clinical research to measure specific mobile outcomes (e.g., data capture frequencies, placement of device). To that aim, this manuscript offers a broad overview of the various mobile outcome assessments currently used in observational and interventional research, and categorizes and consolidates this information for researchers interested in using mobile devices to assess outcomes in interventional research.
                Bookmark

                Author and article information

                Contributors
                larsson.omberg@sagebionetworks.org
                elias.chaibub.neto@sagebionetworks.org
                lara.mangravite@sagebionetworks.org
                Journal
                Clin Pharmacol Ther
                Clin. Pharmacol. Ther
                10.1002/(ISSN)1532-6535
                CPT
                Clinical Pharmacology and Therapeutics
                John Wiley and Sons Inc. (Hoboken )
                0009-9236
                1532-6535
                09 February 2020
                April 2020
                09 February 2020
                : 107
                : 4 , Data Science ( doiID: 10.1002/cpt.v107.4 )
                : 719-721
                Affiliations
                [ 1 ] Sage Bionetworks Seattle Washington USA
                Author notes
                [*] [* ] Correspondence: Larsson Omberg ( larsson.omberg@ 123456sagebionetworks.org ), Elias Chaibub Neto ( elias.chaibub.neto@ 123456sagebionetworks.org ), and Lara M. Mangravite ( lara.mangravite@ 123456sagebionetworks.org )

                Article
                CPT1781
                10.1002/cpt.1781
                7158202
                32036612
                5257f370-d384-4fb8-b10a-be5bfa714e13
                © 2020 The Authors. Clinical Pharmacology & Therapeutics published by Wiley Periodicals, Inc. on behalf of American Society for Clinical Pharmacology and Therapeutics.

                This is an open access article under the terms of the http://creativecommons.org/licenses/by-nc/4.0/ License, which permits use, distribution and reproduction in any medium, provided the original work is properly cited and is not used for commercial purposes.

                History
                : 18 November 2019
                : 13 January 2020
                Page count
                Figures: 0, Tables: 0, Pages: 4, Words: 2015
                Categories
                Perspective
                Perspectives
                Perspectives
                Custom metadata
                2.0
                April 2020
                Converter:WILEY_ML3GV2_TO_JATSPMC version:5.8.0 mode:remove_FC converted:14.04.2020

                Pharmacology & Pharmaceutical medicine
                Pharmacology & Pharmaceutical medicine

                Comments

                Comment on this article