100
views
0
recommends
+1 Recommend
1 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Article: found
      Is Open Access

      Developing an implementation strategy for a digital health intervention: an example in routine healthcare

      Read this article at

      ScienceOpenPublisher
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          Background Evidence on how to implement new interventions into complex healthcare environments is often poorly reported and indexed, reducing its potential to inform initiatives to improve healthcare services. Using the implementation of a digital intervention within routine National Health Service (NHS) practice, we provide an example of how to develop a theoretically based implementation plan and how to report it transparently. In doing so we also highlight some of the challenges to implementation in routine healthcare. Methods The implemented intervention was HeLP-Diabetes, a digital self-management programme for people with Type 2 Diabetes, which was effective in improving diabetes control. The target setting for the implementation was an inner city London Clinical Commissioning Group in the NHS comprised of 34 general practices. HeLP-Diabetes was designed to be offered to patients as part of routine diabetes care across England. Evidence synthesis, engagement of local stakeholders, a theory of implementation (Normalization Process Theory), feedback, qualitative interviews and usage data were used to develop an implementation plan. Results A new implementation plan was developed to implement HeLP-Diabetes within routine practice. Individual component strategies were selected and developed informed by Normalization Process Theory. These strategies included: engagement of local opinion leaders, provision of educational materials, educational visits, educational meetings, audit and feedback and reminders. Additional strategies were introduced iteratively to address barriers that arose during the implementation. Barriers largely related to difficulties in allocating resources to implement the intervention within routine care. Conclusion This paper provides a worked example of implementing a digital health intervention. The learning from this work can inform others undertaking the work of planning and executing implementation activities in routine healthcare. Of particular importance is: the selection of appropriate theory to guide the implementation process and selection of strategies; ensuring that enough attention is paid to planning implementation; and a flexible approach that allows response to emerging barriers. Electronic supplementary material The online version of this article (10.1186/s12913-018-3615-7) contains supplementary material, which is available to authorized users.

          Related collections

          Most cited references44

          • Record: found
          • Abstract: found
          • Article: found
          Is Open Access

          Fostering implementation of health services research findings into practice: a consolidated framework for advancing implementation science

          Background Many interventions found to be effective in health services research studies fail to translate into meaningful patient care outcomes across multiple contexts. Health services researchers recognize the need to evaluate not only summative outcomes but also formative outcomes to assess the extent to which implementation is effective in a specific setting, prolongs sustainability, and promotes dissemination into other settings. Many implementation theories have been published to help promote effective implementation. However, they overlap considerably in the constructs included in individual theories, and a comparison of theories reveals that each is missing important constructs included in other theories. In addition, terminology and definitions are not consistent across theories. We describe the Consolidated Framework For Implementation Research (CFIR) that offers an overarching typology to promote implementation theory development and verification about what works where and why across multiple contexts. Methods We used a snowball sampling approach to identify published theories that were evaluated to identify constructs based on strength of conceptual or empirical support for influence on implementation, consistency in definitions, alignment with our own findings, and potential for measurement. We combined constructs across published theories that had different labels but were redundant or overlapping in definition, and we parsed apart constructs that conflated underlying concepts. Results The CFIR is composed of five major domains: intervention characteristics, outer setting, inner setting, characteristics of the individuals involved, and the process of implementation. Eight constructs were identified related to the intervention (e.g., evidence strength and quality), four constructs were identified related to outer setting (e.g., patient needs and resources), 12 constructs were identified related to inner setting (e.g., culture, leadership engagement), five constructs were identified related to individual characteristics, and eight constructs were identified related to process (e.g., plan, evaluate, and reflect). We present explicit definitions for each construct. Conclusion The CFIR provides a pragmatic structure for approaching complex, interacting, multi-level, and transient states of constructs in the real world by embracing, consolidating, and unifying key constructs from published implementation theories. It can be used to guide formative evaluations and build the implementation knowledge base across multiple studies and settings.
            Bookmark
            • Record: found
            • Abstract: found
            • Article: found

            Better reporting of interventions: template for intervention description and replication (TIDieR) checklist and guide

            Without a complete published description of interventions, clinicians and patients cannot reliably implement interventions that are shown to be useful, and other researchers cannot replicate or build on research findings. The quality of description of interventions in publications, however, is remarkably poor. To improve the completeness of reporting, and ultimately the replicability, of interventions, an international group of experts and stakeholders developed the Template for Intervention Description and Replication (TIDieR) checklist and guide. The process involved a literature review for relevant checklists and research, a Delphi survey of an international panel of experts to guide item selection, and a face to face panel meeting. The resultant 12 item TIDieR checklist (brief name, why, what (materials), what (procedure), who provided, how, where, when and how much, tailoring, modifications, how well (planned), how well (actual)) is an extension of the CONSORT 2010 statement (item 5) and the SPIRIT 2013 statement (item 11). While the emphasis of the checklist is on trials, the guidance is intended to apply across all evaluative study designs. This paper presents the TIDieR checklist and guide, with an explanation and elaboration for each item, and examples of good reporting. The TIDieR checklist and guide should improve the reporting of interventions and make it easier for authors to structure accounts of their interventions, reviewers and editors to assess the descriptions, and readers to use the information.
              Bookmark
              • Record: found
              • Abstract: found
              • Article: found
              Is Open Access

              Implementation strategies: recommendations for specifying and reporting

              Implementation strategies have unparalleled importance in implementation science, as they constitute the ‘how to’ component of changing healthcare practice. Yet, implementation researchers and other stakeholders are not able to fully utilize the findings of studies focusing on implementation strategies because they are often inconsistently labelled and poorly described, are rarely justified theoretically, lack operational definitions or manuals to guide their use, and are part of ‘packaged’ approaches whose specific elements are poorly understood. We address the challenges of specifying and reporting implementation strategies encountered by researchers who design, conduct, and report research on implementation strategies. Specifically, we propose guidelines for naming, defining, and operationalizing implementation strategies in terms of seven dimensions: actor, the action, action targets, temporality, dose, implementation outcomes addressed, and theoretical justification. Ultimately, implementation strategies cannot be used in practice or tested in research without a full description of their components and how they should be used. As with all intervention research, their descriptions must be precise enough to enable measurement and ‘reproducibility.’ We propose these recommendations to improve the reporting of implementation strategies in research studies and to stimulate further identification of elements pertinent to implementation strategies that should be included in reporting guidelines for implementation strategies.
                Bookmark

                Author and article information

                Contributors
                Journal
                BMC Health Services Research
                BMC Health Serv Res
                Springer Science and Business Media LLC
                1472-6963
                December 2018
                October 19 2018
                December 2018
                : 18
                : 1
                Article
                10.1186/s12913-018-3615-7
                8f3b18ee-961b-418f-bb8e-75d5f63d465a
                © 2018

                http://creativecommons.org/licenses/by/4.0/

                History

                Comments

                Comment on this article