Skip to main content

BRIEF RESEARCH REPORT article

Front. Public Health, 02 September 2022
Sec. Public Health Education and Promotion
This article is part of the Research Topic Implementation Mapping for Selecting, Adapting and Developing Implementation Strategies View all 16 articles

Implementation mapping for tobacco cessation in a federally qualified health center

\nAriel M. Domlyn
Ariel M. Domlyn1*Carolyn CrowderCarolyn Crowder2Howard Eisenson,Howard Eisenson2,3Kathryn I. Pollak,,Kathryn I. Pollak3,4,5James M. Davis,James M. Davis5,6Patrick S. Calhoun,,Patrick S. Calhoun1,5,7Sarah M. Wilson,,,Sarah M. Wilson1,4,5,7
  • 1VA Health Services Research and Development Center of Innovation to Accelerate Discovery and Practice Transformation, Durham, NC, United States
  • 2Lincoln Community Health Center, Durham, NC, United States
  • 3Department of Family Medicine and Community Health, Duke University School of Medicine, Durham, NC, United States
  • 4Department of Population Health Sciences, Duke University School of Medicine, Durham, NC, United States
  • 5Duke Cancer Institute, Duke University Health System, Durham, NC, United States
  • 6Department of Medicine, Duke University School of Medicine, Durham, NC, United States
  • 7Department of Psychiatry and Behavioral Sciences, Duke University School of Medicine, Durham, NC, United States

Background: Implementation mapping (IM) is a promising five-step method for guiding planning, execution, and maintenance of an innovation. Case examples are valuable for implementation practitioners to understand considerations for applying IM. This pilot study aimed to determine the feasibility of using IM within a federally qualified health center (FQHC) with limited funds and a 1-year timeline.

Methods: An urban FQHC partnered with an academic team to employ IM for implementing a computerized strategy of tobacco cessation: the 5A's (Ask, Advise, Assess, Assist, Arrange). Each step of IM was supplemented with theory-driven methods and frameworks. Data collection included surveys and interviews with clinic staff, analyzed via rapid data analysis.

Results: Medical assistants and clinicians were identified as primary implementers of the 5A's intervention. Salient determinants of change included the perceived compatibility and relative priority of 5A's. Performance objectives and change objectives were derived to address these determinants, along with a suite of implementation strategies. Despite indicators of adoptability and acceptability of the 5A's, reductions in willingness to adopt the implementation package occurred over time and the intervention was not adopted by the FQHC within the study timeframe. This is likely due to the strain of the COVID-19 pandemic altering health clinic priorities.

Conclusions: Administratively, the five IM steps are feasible to conduct with FQHC staff within 1 year. However, this study did not obtain its intended outcomes. Lessons learned include the importance of re-assessing barriers over time and ensuring a longer timeframe to observe implementation outcomes.

Introduction

Community-engaged dissemination and implementation (CEDI) research is a process of collaboration and shared decision-making between academics and community-based healthcare providers and recipients (1). CEDI is presumed to mitigate health inequities by incorporating the perspectives of individuals typically marginalized from traditional research paradigms (1, 2). Implementation Mapping (IM) (3) is a CEDI method with growing popularity (2, 3). IM hybridizes implementation science principles with a process for multi-level health promotion called intervention mapping. IM defines five change management steps (3). Despite being touted as a promising strategy (4, 5) and multiple examples of planned use via study protocols (4, 6, 7) there are few publicly accessible descriptions of applying the IM process (810) and among these only one reported use through all steps (8). This complete example effectively illustrates IM as a feasible and effective method, however, it was also bolstered by significant resources (a 4-year timeline and five funding sources). We offer an example of using IM on a smaller scale within a busy, urban federally qualified health center (FQHC). The lessons learned from this pilot study offer perspective on the feasibility (11) of conducting IM in resource-limited settings.

Materials and methods

Tobacco cessation is an important public health effort (12). Despite declining rates of tobacco use in recent years, tobacco rates among low-income individuals remain unchanged (13). Community clinics and primary care providers are front line forces for the prevention and treatment of harmful health behaviors, including tobacco use. This project sought to use IM to implement an evidence-based tobacco cessation strategy within a community healthcare center. Table 1 provides the definitions of terms used throughout this text.

TABLE 1
www.frontiersin.org

Table 1. Definitions of key terms.

Setting: Federally qualified health center

Nationally, tobacco use rates are highest among those at or below 200% of the federal poverty level (13). In Durham, North Carolina tobacco use remains a leading cause of death in the area (14). At one local FQHC, 97% of patients have income at or below 200% of the federal poverty level (15). This FQHC serves over 34,000 adult and pediatric patients per year. In 2015, the FQHC attempted to implement an evidence-based specialty tobacco cessation clinic with trained tobacco treatment specialists. Despite early successes, the program was not sustained due to staff turnover. To address this concern, the FQHC's director of behavioral health (CC) partnered with an academic with expertise in implementation science and clinical psychology (SW) to address patient tobacco use and design a sustainable program.

The present project sought to create a package of implementation strategies designed to facilitate uptake and sustainment of an evidence-based, technology-assisted tobacco cessation tool at the FQHC. In consultation with clinician and researcher colleagues, the CEDI leadership team selected computer-facilitated delivery of evidence-based 5A's due to its known impact increasing delivery of tobacco cessation treatment in medical settings (16, 17).

Intervention: 5A's intervention model for tobacco cessation

The 5A's intervention model (Ask, Advise, Assess, Assist, Arrange) was developed as a guide to help clinicians treat tobacco use (12). One method proposed to facilitate clinician use of the 5A's is to use a computerized process with handheld digital devices (1618). While this strategy has proven effective, there are some implementation issues with introducing handheld devices into clinical encounters where they are not normally used (16). The present study sought to overcome implementation barriers to computerized 5A's by implementing this evidence-based intervention into the electronic health record (EHR) system at an FQHC. This would enable the 5A's to be completed with fidelity directly through the EHR rather than using any outside devices or manuals. However, it is recognized that technology-assisted smoking cessation tools may suffer significant challenges in implementation including limited staff knowledge of resources, limited familiarity with tobacco cessation practices, and lack of organizational support (16, 19). Numerous factors (20) may affect uptake and sustainment, including disruption of clinic workflow, as well as perceptions that technology is burdensome and ineffective (21, 22). Systematic implementation planning and support may improve uptake and sustainment of technology-dependent tobacco cessation interventions. In selecting an implementation method, the implementers prioritized equity-focused options that accounted for situations unique to community-academic collaborations.

CEDI method

Use of CEDI methods are critical in FQHC settings, as patients served by these clinics are often among the most disenfranchised (23). Derived from literatures on health promotion and implementation science, implementation mapping (IM) is a CEDI process that includes five steps for assisting organizations in planning and enacting change strategies. The steps detail (1) conducting a needs and assets assessment within the setting, (2) identifying implementation outcomes and performance objectives based on identified change determinants, (3) selecting a theory-based method and strategies to affect these determinants, (4) developing implementation protocols and materials, (5) evaluating implementation outcomes (3). Standardized measures or tools are not yet available for enacting each step, but guidelines exist to inform the process. Key among these is the use of theory to inform each step (3).

The current project used the Consolidated Framework for Implementation Research (CFIR) (24) to design needs assessment materials and identify determinants of change. Determinants are the barriers and facilitators affecting whether the innovation is adopted, scaled, and maintained; these are classified into discrete constructs related to the implementation process or the innovation itself (24). CFIR is a comprehensive model composed of determinants with empirical and theoretical support for implementation relevance, such as characteristics of the intervention, inner setting, and outer setting. This includes knowledge (staff familiarity with the innovation), compatibility (perceived fit between the innovation and organization), relative priority (perceived importance of the innovation), and the implementation climate (staff receptivity to the innovation) (24).

For designing implementation strategies (methods or techniques used to enhance the adoption, implementation, and sustainability of a clinical program or practice) (25, 26), a systems science method (27) was used to assess the variable impact and effort of each potential strategy and adapted for developing implementation strategies within this FQHC (28). Consistent with community-engaged practices (2), this process enabled power-sharing by identifying staff-driven strategies, later mapped onto a taxonomy of expert-identified strategies (26) for consistency in reporting.

A well established conceptualization of implementation outcomes (29) assisted in reporting progress for the final step. Unlike service or population outcomes, implementation outcomes include both expected and observed indicators of successful innovation adoption, usage, and maintenance (30). These types of outcomes identify markers of interim progress in the implementation efforts and may be assessed temporally early, mid, or late in the project (29). Debate about conceptualizing implementation outcomes (30) unfolded in the literature within the timeframe this pilot study was conceptualized and executed. Reporting for step 5 considers the anticipated implementation outcomes as perceived acceptability [degree of satisfaction or palatability of the innovation (29)] and adoptability [the likelihood key decision-makers will decide to put the innovation into place (30)]. Actual implementation outcomes are adoption [the extent key decision makers decide to put the innovation into place (30)] and implementation [the extent the innovation is in place (30)]. Of note: throughout this paper the word “feasibility” refers to the common term for a preparatory study (11) rather than the “feasibility” as defined in implementation outcomes (29).

Data collection and analysis

Participants included multiple groups of FQHC clinical and administrative staff: physicians, advanced practice providers, behavioral health specialists, nurses, medical assistants, patient educators, and administrative and clinical leaders. Inclusion of different clinic roles aimed for diversity of opinions to generate staff-driven solutions. We used quota sampling (31) to ensure representation across clinic roles (physicians, advanced practice providers, behavioral health specialists, nurses, medical assistants, and patient educators) and settings (internal medicine, family medicine, and pediatrics). All clinic staff members who contacted the study coordinator for participation were included in the study. Participants (N = 12) were interviewed using open-ended prompts for the needs assessment and determinant identification. This sample size was selected given its high likelihood of reaching data saturation (51). These interviews were recorded, transcribed, and analyzed using a rapid analytic method in which data reduction occurred prior to coding (32). Concept codes were determined a priori with the goal to rapidly inform process (33). Interview results informed Step 4, the development of implementation protocols and materials. Surveys were then conducted with FQHC clinic and administrative leaders (N = 7), and descriptive statistics reported to identify performance objectives and gather early-stage implementation outcomes. Informed consent was obtained for all participants. Only non-FQHC study staff had access to identifying information of staff participants. All FQHC investigators saw only de-identified, aggregated data.

Figure 1 shows the project timeline by key activity, IM step, and implementation stage. Key activities here are data collection, analysis, and development of materials. Implementation stages are discerned from a common stage framework that determines the stages of change that occur within an organization: Exploration, Preparation/Adoption, Implementation, and Sustainment (EPIS) (47). Both IM and EPIS are heuristics for describing the process, but both can be iterative rather than linear processes. Therefore, key activities do not always occur sequentially according to these steps and stages.

FIGURE 1
www.frontiersin.org

Figure 1. Project timeline.

Results

Over the course of 12 months, all five IM steps were planned and executed iteratively, as described in the original IM process (3). However, not all objectives and outcomes were achieved. Results below illustrate the process as it was planned and unfolded, but readers are advised to note that the implementation package described in Step 4 was not adopted, and therefore the planned activities and objectives in Step 3 were not undertaken. In Step 5, anticipated outcomes showed promising indicators of eventual actual implementation outcomes; this is further described in that step.

IM step 1: Needs assessment

Implementation adopters were identified by soliciting opinions about the most appropriate staff member to oversee the process of rolling out the 5A's tool in the FQHC's EHR. In IM, implementation adopters are the decision makers, such as leaders, who hold power to decide whether an innovation is adopted (3). There was not clear consensus on this question, with most staff endorsing multiple possible adopters, generally among those who already held clinic leadership positions. Most staff endorsed either the clinical chief for each department or the head of behavioral health. Three interviewees (25%) suggested that providers be the ones to decide to adopt the intervention.

Once implementation adopters are identified, IM indicates adopters be involved in the subsequent planning process. Throughout planning, preparation, and implementation there was shared decision-making and collaboration between adopters (CC, HE) and academic partners (SW, JD, PC), in activities including brainstorming sessions, planning meetings, and addressing issues around participant selection, qualitative methodology, selection of implementation framework, and identifying staff engagement strategies.

Choices of appropriate implementers varied across the 5A's steps. Implementers, per IM, are the individuals responsible for regularly executing the innovation and ensuring it becomes routine practice (3). Clinic staff interviewees (N = 12) were given options to endorse one or more roles for each step, thus number of endorsements exceeded the number of interviewees. Clinic staff unanimously identified medical assistants as having the knowledge and skills to conduct the first 5A's step of asking patients about tobacco use. Per the second 5A's step of advising on tobacco cessation, most (N = 11) staff stated that medical providers were the most appropriate implementers. Concerning the third 5A's step (assessing patient willingness to quit) several staff identified multiple potential implementers. Most staff endorsed that medical providers should conduct the motivational interviewing (34) necessary for this step, yet there were also four mentions each of behavioral health providers and medical assistants being capable of conducting this portion of the intervention. The fourth 5A's step (assist the patients to quit) was seen as a joint effort between medical and behavioral health providers, with nine endorsements of medical providers conducting this step and prescribing nicotine replacement therapies, and six endorsements of behavioral health support as necessary for counseling or consultation. No staff suggested medical assistants as implementers for this step. For the fifth 5A's step (following up with patients), staff were divided on the optimal implementers. There was equal endorsement for the medical provider, behavioral health provider, and whomever was conducting the primary intervention (e.g., pharmacotherapy, counseling, etc.). Responses at this step were contingent upon who the interviewee had identified as the primary responsible party in the previous step. Collectively, identified implementers of the 5A's intervention were clinical staff members, particularly medical assistants, nurses, primary care providers, and behavioral health providers.

IM step 2: Identify adoption and implementation outcomes, performance objectives, determinants, and change objectives

Performance objectives and implementation outcomes

Performance objectives were derived from surveys administered to FQHC leadership (N = 7) and from follow-up meetings with the CEDI leadership team (CC, SW, HE, PC, JD). These objectives intend to define the specific steps or behaviors needed to obtain implementation outcomes. Performance objectives gleaned from leadership surveys are displayed in Table 2.

TABLE 2
www.frontiersin.org

Table 2. Performance objectives and implementation outcomes.

Determinants

Interviewers asked FQHC staff open-ended questions pertaining to potential determinants [i.e., barriers and facilitators of successfully implementing the intervention (24)]. These determinants included staff knowledge of the 5A's intervention, compatibility of 5A's with current clinic practices, implementation climate (i.e., staff receptivity to 5A's), and relative priority of implementing 5A's. Most (N = 9) staff reported no familiarity with the 5A's intervention, while the remaining staff (N = 3) stated they had a vague recollection of having learned this previously, such as in graduate training. All reported percieving the 5A's as compatible with current practices, and several staff said they routinely one or more of the steps as part of usual care. One respondent clarified that they would be opposed to the process if it were mandated, preferring it to be optional and limited to patients who were known tobacco users. Two respondents indicated more information would improve perceived compatibility, such as further education on the 5A's or seeing evidence of the innovation's efficacy in other clinics. Five respondents noted that one barrier to compatibility is the perceived burden of time and effort to conducting the 5A's, which could be mitigated by streamlining the documentation process. Regarding implementation climate, the majority (N = 8) of respondents were in favor of integrating the 5A's into the EHR, some expressing strong optimism about its potential. Three interviewees expressed neutral or ambivalent sentiment, such as: “Adding [this to the] chart is both great and challenging. [There are already] so many other things to [the EHR].” One staff member at the pediatric clinic was opposed to the innovation, stating that they already had a template for asking teenagers about smoking and thought the yield would be low in this population. Per the perceived priority, three participants indicated the tobacco cessation intervention was a priority while four staff members indicated it was a low priority. Six respondents did not state whether it seemed like a priority: five of those indicated the intervention seemed feasible and one stated it would depend on the clinic flow.

Change objectives

Change objectives were developed by cross-walking previously identified performance objectives with determinants of change. Change objectives are the behaviors necessary for FQHC staff to exhibit in order to successfully implement the 5A's. In Table 3, a sample of change objectives is shown with columns corresponding to the necessary change in attitudes, knowledge, and skills for various FQHC roles. Each cell lists an observable behavior that would be indicative of a change in attitude, knowledge, or skills. These are marked by the expectation that each change objective would affect the perceived compatibility of 5A's or the perceived relative priority of 5A's.

TABLE 3
www.frontiersin.org

Table 3. Change objectives by implementation role.

Step 3: Select theoretical methods and design implementation strategies

The selection of implementation strategies requires identifying techniques to influence determinants gleaned from the previous step. There is much debate in the literature about best methods for selecting strategies, with general consensus that a systematic and constituent-influenced approach is optimal, with the entire IM process often cited as an option (35, 36). Here, a three-component approach was adapted from the effort-vs-impact assessment method of operations planning, fully described elsewhere (28). In brief, this approach charted strategies according to effort (low/high) and impact (low/high). The first component assessed the potential effort to make the technological strategy usable according to availability (i.e., how accessible the technological infrastructure is to clinic staff) and familiarity (i.e., how much training would be required for staff). The second component assessed potential impact of the strategy (i.e., improving monitoring, communication, or data collection). The third component assessed whether to use or abandon the strategy by cross-referencing results from the previous two components. Rapid analysis of staff interviews and leader surveys were coded according to a spectrum of perceived effort and impact (28). Results identified seven priority strategies, primarily enacted by the CEDI support system (37) (SW, CC) to target behavioral change in the delivery system (medical assistants and clinicians). See Table 4 for a breakdown of the proposed strategies [described with best-practice language from a common taxonomy of implementation strategies (26)], with corresponding change objectives, and specification per best practice guidelines for describing implementation strategies (25).

TABLE 4
www.frontiersin.org

Table 4. Implementation strategies generated by implementation mapping.

Step 4: Produce implementation protocols and materials

This step aims to enact the implementation strategies through content development. For the strategy of incorporating elements of the 5A's intervention into the EHR, the implementation support team enlisted assistance from EHR analysts (from the FQHC and academic affiliate) and tobacco cessation experts from the academic affiliate. This team created pharmacy order sets within the EHR to speed clinician access to different prescription options for the Arrange step while completing the patient visit, and a sample after visit summary page—including cessation tips and guidance on how to use medication therapies—to be provided to patients. The team also created a data analytic strategy for pulling summaries of tobacco users and completion of 5A's steps (i.e., advice to quit, pharmacotherapy prescriptions, printing patient after-visit summaries, and referrals for behavioral treatment).

Educational materials for adopters included workflow diagrams and detailed flowcharts of decision points and documentation requirements for each of the 5A's steps. Different flowcharts were created for the adult and pediatric clinics to account for differing algorithms. These specified when in the patient visit the innovation was to be enacted (i.e., after taking vital signs) and suggested prompts to start the conversation with patients (e.g., during the Advise step a clinician could state “Can I share with you why I think it is important to your health for you to stop using tobacco products, and how I can help you?”). To enhance usability, these flowcharts were limited to one page with clear font and large text.

Faculty from the local academic affiliate (which runs a tobacco treatment specialist training program) provided sample lecture slides and quick-reference handouts for the development of clinician and staff educational materials.

Step 5: Evaluate implementation outcomes

Since this study was an implementation pilot, outcomes focus on the broad feasibility of the IM process (11). The IM process took approximately 12 months. True to the iterative nature of IM, feedback from FQHC staff and leadership informed revisions of the implementation package. The process of conducting the needs assessment and defining determinants, objectives, and strategies was feasible with a small, collaborative team.

Per anticipated implementation outcomes, in the early IM stages staff interviews indicated a majority were either in favor (67%) or neutral toward (17%) implementing the computerized 5A's process, indicative of good acceptability. Similarly, the majority (86%) of clinic leaders were in favor of proceeding with the plan to implement the computerized 5A's, indicative of adoptability. However, during review of the final implementation package, FQHC executive leadership expressed reductions in willingness to integrate the 5A's intervention package as shown in the protocols and materials.

In meetings following the development of the implementation package, leadership and clinicians involved in the CEDI team reported that the COVID-19 pandemic had caused significant strain on the FQHC, as well as its staff and clinicians. Specific barriers to proceeding with implementation were consistent with those originally voiced by both staff and clinic leadership during IM steps 1 and 2. Although in the early steps of IM limited clinical appointment time and risk of staff burnout were perceived as manageable barriers to implementation, they later became salient to organizational leadership as barriers—and perhaps insurmountable due to the pandemic.

Given the limited time frame of the study funding period and competing priorities of FQHC staff and leadership, further work on revising the implementation package has not been possible. Changes to the EHR have not yet been made, trainings have not been completed, and requests have not been made to the medical staff to change care or documentation of tobacco cessation. Despite promising early indicators of acceptability and adoptability, at time of publication the actual implementation outcomes for implementers (medical assistants and clinicians) were unfortunately not achieved.

Discussion

The five steps of implementation mapping were conducted with an FQHC for implementing a computerized tobacco cessation intervention. Despite following IM recommendations and achieving early implementation outcomes of acceptability and adoptability, the intervention was not adopted. While not yet successful in its intended efforts, this project offers important lessons for future use and improvement of IM application in community clinics.

Lessons learned

The most significant barrier to achieving intended outcomes is not accounted for by standard implementation methods: a global pandemic. CFIR and other implementation models recognize the vast effect of broad external factors on implementation success (24, 37). Changes in outer context (local, national, and global) affect the inner context (individual, team, organizational). Without data to investigate the salient factors after the onset of the COVID-19 pandemic, the authors surmise that the substantial effect of the pandemic on healthcare organizations altered FQHC staff perceptions about 5A's priority and compatibility. This is likely due to rapid rollouts of new disease mitigation processes and a sudden increase in telehealth technology needs. Additionally, the determinants may have been affected by the changing financial, temporal, and logistical resources of the FQHC. The initial needs assessment was instrumental for understanding the performance objectives and change objectives and developing the initial implementation package, but a repeat assessment of determinants could have assisted in understanding evolving barriers to uptake and optimal strategies to address them.

Identifying implementation strategies requires assessing and addressing both individual and organizational-level components, a point reinforced by successful IM examples (8). While this project developed multi-tiered strategies by involving multiple stakeholders and conducting IM as an iterative process, logistical barriers preventing this project from including the intended recipients of this innovation. Patients were unavailable for participation in the project during the early phases of the COVID crisis. This omission highlights that patient perspective may be a critical component for IM success.

Additionally, this IM process took 12 months. Compared to other examples that unfolded over several years (8) the time elapsed may have been too brief to achieve practice utilization. In interviews, FQHC staff noted the need for time and resources to adopt and scale this innovation. This highlights the stressors of using limited external funding, which follows grant cycles and stipulations, and may require much greater funds to follow the full implementation process through to the maintenance phase. Here, the external support was limited to one year. It is well documented that successful implementation requires long-term support and strategies (38), which requires funders' long-term investment of implementation projects (39, 40). While external funding sources provide critical supports for knowledge transfer, there remains a lag between the significant resources needed for successful implementation (38) and the structure of funding mechanisms (39, 41).

Implementation mapping

IM remains a promising and feasible method for effectively planning and strategizing implementation efforts (810). The method is continuing to be tested and improved. Several large studies using IM are planned or underway (4, 6, 7), which will further describe and refine the process. While the evidence base grows, the practice-based evidence supplied here bridges implementation practice to implementation science.

Based on this project's findings, IM does not sufficiently guide how to manage contextual changes that occur over time. It is well documented that determinants display variable salience across implementation stages (4244). In our example, during early IM steps the FQHC adopters and implementers reported enthusiasm for the innovation. Yet the relative priority may have changed, likely due to the COVID-19 pandemic shifting organizational needs. Accounting for external disruptions to the implementation process is necessary for both building organizational resilience and enhancing implementation success (45). Relatedly, regardless of changing priorities, preliminary work identified that some strategies are relevant in earlier versus later stages of change (46). Although IM is proposed as an iterative process, it is unclear when and how often users should re-assess determinants and revise strategies. This is likely to vary by context, however—as originally suggested by the IM developers (3)—implementation practitioners would benefit from expanding the literature on how IM can be synchronized with frameworks that account for other influences on the implementation process. Frameworks of implementation stages (4749) may be critical supplements for IM. Timely re-assessment of determinants and strategy selection—with appropriate resources for doing so—could have assisted in effectively adapting implementation protocols for the rapidly changing FQHC context.

Similarly, given resource constraints in certain care settings, prioritization of change objectives is an essential element that should be added to the IM process. Translating determinants into change objectives is a critical step. This effectively decides, across roles and systems, which key elements are needed to affect change. Here, relative priority and compatibility of the 5A's were identified as highly important in the data. However, among the actions prescribed by Table 3, which are most influential? Ideally, implementation protocols would enact strategies to address all the change objectives, yet this is not feasible in practice. IM developers suggested one determinant framework of organizational readiness may aid in the second and third steps (3). Since the initial IM publication, guidelines (5) and tools (45) for systematically prioritizing determinants have been developed for this readiness framework along with proposals to validate readiness measures in FQHCs (50). Relatedly, models accounting for the behaviors, capabilities, opportunities, and motivations (52) of staff could sharpen assessment of determinants and match them appropriately to change objectives. Use of these instruments by implementation practitioners are consistent with IM recommendations to be both integrative and iterative.

Conclusion

Although this pilot did not result in adoption of the computerized 5A's intervention, IM was feasible to conduct in an FQHC with limited resources. Future IM use should allocate more than one year for reaching intended outcomes and re-assess determinants and change objectives at regular intervals. IM users would benefit from explicit instructions for when to re-assess determinants and how to merge IM with other implementation frameworks. These considerations may improve ability to reach sustainment in future projects.

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Ethics statement

The study protocol was reviewed and approved by Duke University Health System Institutional Review Board. The need to obtain written informed consent was waived by the review board. Prior to beginning study procedures, the participants provided either verbal consent or reviewed a consent script in digital format.

Author contributions

This work was conceived, designed, and executed by SW and CC with assistance from HE, KP, JD, and PC. AD analyzed data and drafted the majority of the manuscript. HE, KP, JD, CC, and SW provided substantial revisions. Additionally, PC provided design oversight, consultation, and critical comments. All authors contributed to final read and provided approval for publication of the content.

Funding

This work was supported by resources provided by the National Center for Advancing Translational Sciences (UL1TR002553) and by a Duke School of Medicine Department of Psychiatry & Behavioral Sciences Health Inequities Research Pilot Grant.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

1. Holt CL, Chambers DA. Opportunities and challenges in conducting community-engaged dissemination/implementation research. Transl Behav Med. (2017) 7:389–92. doi: 10.1007/s13142-017-0520-2

PubMed Abstract | CrossRef Full Text | Google Scholar

2. Schlechter CR, Del Fiol G, Lam CY, Fernandez ME, Greene T, Yack M, et al. Application of community–engaged dissemination and implementation science to improve health equity. Prevent Med Rep. (2021) 24:101620. doi: 10.1016/j.pmedr.2021.101620

PubMed Abstract | CrossRef Full Text | Google Scholar

3. Fernandez ME, Ten Hoor GA, Van Lieshout S, Rodriguez SA, Beidas RS, Parcel G, et al. Implementation mapping: using intervention mapping to develop implementation strategies. Front Public Health. (2019) 7:158. doi: 10.3389/fpubh.2019.00158

PubMed Abstract | CrossRef Full Text | Google Scholar

4. Fernandez ME, Schlechter CR, Del Fiol G, Gibson B, Kawamoto K, Siaperas T, et al. QuitSMART Utah: an implementation study protocol for a cluster-randomized, multi-level sequential multiple assignment randomized trial to increase reach and impact of tobacco cessation treatment in community health centers. Implement Sci. (2020) 15:1–3. doi: 10.1186/s13012-020-0967-2

PubMed Abstract | CrossRef Full Text | Google Scholar

5. Domlyn AM, Scott V, Livet M, Lamont A, Watson A, Kenworthy T, et al. R = MC2 readiness building process: a practical approach to support implementation in local, state, and national settings. J Community Psychol. (2021) 49:1228–48. doi: 10.1002/jcop.22531

PubMed Abstract | CrossRef Full Text | Google Scholar

6. Kurlander JE, Helminski D, Lanham M, Henstock JL, Kidwell KM, Krein SL, et al. Development of a multicomponent implementation strategy to reduce upper gastrointestinal bleeding risk in patients using warfarin and antiplatelet therapy, and protocol for a pragmatic multilevel randomized factorial pilot implementation trial. Implement Sci Commun. (2022) 3:1–7. doi: 10.1186/s43058-022-00256-8

PubMed Abstract | CrossRef Full Text | Google Scholar

7. Walker TJ, Kohl HW, Bartholomew JB, Green C, Fernández ME. Using Implementation Mapping to develop and test an implementation strategy for active learning to promote physical activity in children: a feasibility study using a hybrid type 2 design. Implement Sci Commun. (2022) 3:1–9. doi: 10.1186/s43058-022-00271-9

PubMed Abstract | CrossRef Full Text | Google Scholar

8. Ibekwe LN, Walker TJ, Ebunlomo E, Ricks KB, Prasad S, Savas LS, et al. Using implementation mapping to develop implementation strategies for the delivery of a cancer prevention and control phone navigation program: a collaboration with 2-1-1. Health Promot Pract. (2020) 9:1524839920957979. doi: 10.1177/1524839920957979

PubMed Abstract | CrossRef Full Text | Google Scholar

9. Kennedy MA, Bayes S, Newton RU, Zissiadis Y, Spry NA, Taaffe DR, et al. We have the program, what now? development of an implementation plan to bridge the research-practice gap prevalent in exercise oncology. Int J Behav Nutri Physic Activ. (2020) 17:1. doi: 10.1186/s12966-020-01032-4

PubMed Abstract | CrossRef Full Text | Google Scholar

10. Roth IJ, Tiedt MK, Barnhill JL, Karvelas KR, Faurot KR, Gaylord S, et al. Feasibility of implementation mapping for integrative medical group visits. J Alternat Complement Med. (2021) 27:S71. doi: 10.1089/acm.2020.0393

PubMed Abstract | CrossRef Full Text | Google Scholar

11. Pearson N, Naylor PJ, Ashe MC, Fernandez M, Yoong SL, Wolfenden L. Guidance for conducting feasibility and pilot studies for implementation trials. Pilot Feasibil Stud. (2020) 6:1–2. doi: 10.1186/s40814-020-00634-w

PubMed Abstract | CrossRef Full Text | Google Scholar

12. Fiore MC. A clinical practice guideline for treating tobacco use and dependence: a US Public Health Service report. JAMA: J Am Med Assoc. (2000) 20:5.

PubMed Abstract | Google Scholar

13. Drope J, Liber AC, Cahn Z, et al. Who's still smoking? Disparities in adult cigarette smoking prevalence in the united states CA. Cancer J Clinic. (2018) 68:106–15. doi: 10.3322/caac.21444

PubMed Abstract | CrossRef Full Text | Google Scholar

14. Durham County Department of Public Health & Partnership for a Healthy Durham. Durham County Community Health Assessment. Durham County Department of Public Health & Partnership for a Healthy Durham (2018).

Google Scholar

15. Health Resources Services Administration. Lincoln Community Health Center, Inc.Health Center Profile. (2018). Available online at: https://bphc.hrsa.gov/uds/datacenter.aspx?q=d&bid=040910&state=NC&year=2017 (accessed July 3, 2019).

Google Scholar

16. Satterfield JM, Gregorich SE, Kalkhoran S, et al. Computer-facilitated 5a's for smoking cessation: a randomized trial of technology to promote provider adherence. Am J Prev Med. (2018) 55:35–43. doi: 10.1016/j.amepre.2018.04.009

PubMed Abstract | CrossRef Full Text | Google Scholar

17. Strayer SM, Heim SW, Rollins LK, et al. Improving smoking cessation counseling using a point-of-care health intervention tool (IT): From the Virginia Practice Support and Research Network (VAPSRN). J Am Board Fam Med. (2013) 26:116–25. doi: 10.3122/jabfm.2013.02.110078

PubMed Abstract | CrossRef Full Text | Google Scholar

18. Bentz CJ, Bayley KB, Bonin KE, Fleming L, Hollis JF, Hunt JS, et al. Provider feedback to improve 5A's tobacco cessation in primary care: a cluster randomized clinical trial. Nicotine & Tobacco Research. (2007) 9:341–9. doi: 10.1080/14622200701188828

PubMed Abstract | CrossRef Full Text | Google Scholar

19. Martínez C, Castellano Y, Andrés A, Fu M, Antón L, Ballbè MU, et al. Factors associated with implementation of the 5A's smoking cessation model. Tob Induc Dis. (2017) 15:1. doi: 10.1186/s12971-017-0146-7

PubMed Abstract | CrossRef Full Text | Google Scholar

20. Holden RJ, Karsh B-T. The technology acceptance model: its past and its future in health care. J Biomed Inform. (2010) 43:159–72. doi: 10.1016/j.jbi.2009.07.002

PubMed Abstract | CrossRef Full Text | Google Scholar

21. Nápoles AM, Appelle N, Kalkhoran S. Perceptions of clinicians and staff about the use of digital technology in primary care: qualitative interviews prior to implementation of a computer-facilitated 5As intervention. BMC Med Inform Decis Making. (2016) 16:44. doi: 10.1186/s12911-016-0284-5

PubMed Abstract | CrossRef Full Text | Google Scholar

22. Yarbrough AK, Smith TB. Technology acceptance among physicians: a new take on TAM. Med Care Res Rev. (2007) 64:650–72. doi: 10.1177/1077558707305942

PubMed Abstract | CrossRef Full Text | Google Scholar

23. Nath JB, Costigan S, Hsia RY. Changes in demographics of patients seen at federally qualified health centers, 2005-2014. JAMA Intern Med. (2016) 176:712–4. doi: 10.1001/jamainternmed.2016.0705

PubMed Abstract | CrossRef Full Text | Google Scholar

24. Damschroder LJ, Aron DC, Keith RE, Kirsh SR, Alexander JA, Lowery JC. Fostering implementation of health services research findings into practice: a consolidated framework for advancing implementation science. Implement Sci. (2009) 4:1–5. doi: 10.1186/1748-5908-4-50

PubMed Abstract | CrossRef Full Text | Google Scholar

25. Proctor EK, Powell BJ, McMillen JC. Implementation strategies: recommendations for specifying and reporting. Implement Sci. (2013) 8:1. doi: 10.1186/1748-5908-8-139

PubMed Abstract | CrossRef Full Text | Google Scholar

26. Powell BJ, Waltz TJ, Chinman MJ, Damschroder LJ, Smith JL, Matthieu MM, et al. A refined compilation of implementation strategies: results from the Expert Recommendations for Implementing Change (ERIC) project. Implement Sci. (2015) 10:1–4. doi: 10.1186/s13012-015-0209-1

PubMed Abstract | CrossRef Full Text | Google Scholar

27. American Society for Quality,. Impact Effort Matrix. (2022). Available online at: https://asq.org/quality-resources/impact-effort-matrix (accessed March 4, 2022).

Google Scholar

28. Kim B, Wilson SM, Mosher TM, Breland JY. Systematic decision-making for using technological strategies to implement evidence-based interventions: an illustrated case study. Front Psychiatr. (2021) 12:64240. doi: 10.3389/fpsyt.2021.640240

PubMed Abstract | CrossRef Full Text | Google Scholar

29. Proctor E, Silmere H, Raghavan R, Hovmand P, Aarons G, Bunger A, et al. Outcomes for implementation research: conceptual distinctions, measurement challenges, and research agenda. Administr Policy Mental Health Ment Health Serv Res. (2011) 38:65–76. doi: 10.1007/s10488-010-0319-7

PubMed Abstract | CrossRef Full Text | Google Scholar

30. Damschroder LJ, Reardon CM, Opra Widerquist MA, Lowery J. Conceptualizing outcomes for use with the consolidated framework for implementation research (CFIR): the CFIR outcomes. Addendum. (2022) 17:7. doi: 10.1186/s13012-021-01181-5

PubMed Abstract | CrossRef Full Text | Google Scholar

31. Robinson OC. Sampling in interview-based qualitative research: A theoretical and practical guide. Qual Res Psychol Implement Sci. (2014) 11:25–41. doi: 10.1080/14780887.2013.801543

CrossRef Full Text | Google Scholar

32. Watkins DC. Rapid and rigorous qualitative data analysis: The “RADaR” technique for applied research. Int J Qual Methods. (2017) 16:1609406917712131. doi: 10.1177/1609406917712131

PubMed Abstract | CrossRef Full Text | Google Scholar

33. Hamilton AB, Finley EP. Qualitative methods in implementation research: an introduction. Psychiatry Res. (2019) 280:112516. doi: 10.1016/j.psychres.2019.112516

PubMed Abstract | CrossRef Full Text | Google Scholar

34. Lindson-Hawley N, Thompson TP, Begh R. Motivational interviewing for smoking cessation. Cochr Database Systemat Rev. (2015) 3:15. doi: 10.1002/14651858.CD006936.pub3

PubMed Abstract | CrossRef Full Text | Google Scholar

35. Leeman J, Birken SA, Powell BJ, Rohweder C, Shea CM. Beyond “implementation strategies”: classifying the full range of strategies used in implementation science and practice. Implement Sci. (2017) 12:1–9. doi: 10.1186/s13012-017-0657-x

PubMed Abstract | CrossRef Full Text | Google Scholar

36. Waltz TJ, Powell BJ, Fernández ME, Abadie B, Damschroder LJ. Choosing implementation strategies to address contextual barriers: diversity in recommendations and future directions. Implement Sci. (2019) 14:1–5. doi: 10.1186/s13012-019-0892-4

PubMed Abstract | CrossRef Full Text | Google Scholar

37. Wandersman A, Duffy J, Flaspohler P, Noonan R, Lubell K, Stillman L, et al. Bridging the gap between prevention research and practice: the interactive systems framework for dissemination and implementation. Am J Commun Psychol. (2008) 41:171–81. doi: 10.1007/s10464-008-9174-z

PubMed Abstract | CrossRef Full Text | Google Scholar

38. Colditz GA, Emmons KM. The promise and challenges of dissemination and implementation research. Disseminat Implement Res Health: Translat Sci Pract. (2012) 24:3–22. doi: 10.1093/acprof:oso/9780199751877.003.0001

CrossRef Full Text | Google Scholar

39. Smits PA, Denis JL. How research funding agencies support science integration into policy and practice: an international overview. Implement Sci. (2014) 9:1–2. doi: 10.1186/1748-5908-9-28

PubMed Abstract | CrossRef Full Text | Google Scholar

40. Wiltsey Stirman S, Kimberly J, Cook N, Calloway A, Castro F, Charns M. The sustainability of new programs and innovations: a review of the empirical literature and recommendations for future research. Implement Sci. (2012) 7:1–9. doi: 10.1186/1748-5908-7-17

PubMed Abstract | CrossRef Full Text | Google Scholar

41. McLean RK, Graham ID, Tetroe JM, Volmink JA. Translating research into action: an international study of the role of research funders. Health Res Policy Syst. (2018) 16:1–5. doi: 10.1186/s12961-018-0316-y

PubMed Abstract | CrossRef Full Text | Google Scholar

42. Domlyn AM, Wandersman A. Community coalition readiness for implementing something new: using a Delphi methodology. J Community Psychol. (2019) 47:882–97. doi: 10.1002/jcop.22161

PubMed Abstract | CrossRef Full Text | Google Scholar

43. Dunbar J, Hernan A, Janus E, Davis-Lameloise N, Asproloupos D, O'Reilly S, et al. Implementation salvage experiences from the Melbourne diabetes prevention study. BMC Public Health. (2012) 12:806. doi: 10.1186/1471-2458-12-806

PubMed Abstract | CrossRef Full Text | Google Scholar

44. Hoagwood KE, Chaffin M, Chamberlain P, Bickman L, Mittman B. “Implementation salvage strategies: maximizing methodological flexibility in children's mental health research,” In: 4th Annual NIH Conference on the Science of Dissemination and Implementation. Washington, DC. (2011).

Google Scholar

45. Kolodny-Goetz J, Hamm DW, Cook BS, Wandersman A. The readiness, resilience and recovery tool: an emerging approach to enhance readiness amidst disruption. Global Implement Res Applicat. (2021) 1:135–46. doi: 10.1007/s43477-021-00011-6

PubMed Abstract | CrossRef Full Text | Google Scholar

46. Vax S, Farkas M, Mueser KT, Drainoni ML, Russinova Z. Building organizational readiness: initial field testing of an expert-informed typology of implementation strategies. Implement Sci Commun. (2022) 3:1. doi: 10.1186/s43058-022-00257-7

PubMed Abstract | CrossRef Full Text | Google Scholar

47. Aarons GA, Hurlburt M, Horwitz SM. Advancing a conceptual model of evidence-based practice implementation in public service sectors. Adm Policy Ment Health Ment Health Serv Res. (2011) 38:4–23. doi: 10.1007/s10488-010-0327-7

PubMed Abstract | CrossRef Full Text | Google Scholar

48. Metz A, Bartley L. Active implementation frameworks for program success. Zero Three. (2012) 32:11–8.

Google Scholar

49. Chamberlain P, Brown CH, Saldana L. Observational measure of implementation progress in community based settings: the stages of implementation completion (SIC). Implement Sci. (2011) 6:1–8. doi: 10.1186/1748-5908-6-116

PubMed Abstract | CrossRef Full Text | Google Scholar

50. Walker TJ, Brandt HM, Wandersman A, Scaccia J, Lamont A, Workman L, et al. Development of a comprehensive measure of organizational readiness (motivation × capacity) for implementation: a study protocol. Implement Sci Commun. (2020) 1:1. doi: 10.1186/s43058-020-00088-4

PubMed Abstract | CrossRef Full Text | Google Scholar

51. Guest G, Bunce A, Johnson L. How many interviews are enough? an experiment with data saturation and variability. Field Methods. (2006) 18:59–82. doi: 10.1177/1525822X05279903

CrossRef Full Text | Google Scholar

52. Michie S, Atkins L, West R. The Behavior Change Wheel: A Guide to Designing Interventions. London, England (2014).

PubMed Abstract | Google Scholar

Keywords: implementation science (MeSH), implementation mapping, tobacco cessation, 5A's smoking cessation guidelines, community health, community-engaged dissemination and implementation

Citation: Domlyn AM, Crowder C, Eisenson H, Pollak KI, Davis JM, Calhoun PS and Wilson SM (2022) Implementation mapping for tobacco cessation in a federally qualified health center. Front. Public Health 10:908646. doi: 10.3389/fpubh.2022.908646

Received: 30 March 2022; Accepted: 04 August 2022;
Published: 02 September 2022.

Edited by:

Gill Ten Hoor, Maastricht University, Netherlands

Reviewed by:

Allie Slemon, University of British Columbia, Canada
Gerjo Kok, Maastricht University, Netherlands

Copyright © 2022 Domlyn, Crowder, Eisenson, Pollak, Davis, Calhoun and Wilson. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Ariel M. Domlyn, ariel.domlyn@va.gov

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.