Skip to main content

METHODS article

Front. Health Serv., 15 July 2024
Sec. Implementation Science
This article is part of the Research Topic Learning for Action in Policy Implementation View all 12 articles

Design considerations for developing measures of policy implementation in quantitative evaluations of public health policy

  • 1Department of Social and Behavioral Sciences, Harvard TH Chan School of Public Health, Boston, MA, United States
  • 2Mongan Institute Health Policy Research Center, Massachusetts General Hospital, Boston, MA, United States
  • 3Harvard Medical School, Boston, MA, United States
  • 4Human Development and Family Studies Program, Department of Human Ecology, University of California, Davis, CA, United States
  • 5Department of Public Health Policy & Management, Global Center for Implementation Science, New York University School of Global Public Health, New York, NY, United States
  • 6Institute for Health Research and Policy, University of Illinois Chicago, Chicago, IL, United States
  • 7Department of Health Policy and Administration, School of Public Health, University of Illinois Chicago, Chicago, IL, United States

Typical quantitative evaluations of public policies treat policies as a binary condition, without further attention to how policies are implemented. However, policy implementation plays an important role in how the policy impacts behavioral and health outcomes. The field of policy-focused implementation science is beginning to consider how policy implementation may be conceptualized in quantitative analyses (e.g., as a mediator or moderator), but less work has considered how to measure policy implementation for inclusion in quantitative work. To help address this gap, we discuss four design considerations for researchers interested in developing or identifying measures of policy implementation using three independent NIH-funded research projects studying e-cigarette, food, and mental health policies. Mini case studies of these considerations were developed via group discussions; we used the implementation research logic model to structure our discussions. Design considerations include (1) clearly specifying the implementation logic of the policy under study, (2) developing an interdisciplinary team consisting of policy practitioners and researchers with expertise in quantitative methods, public policy and law, implementation science, and subject matter knowledge, (3) using mixed methods to identify, measure, and analyze relevant policy implementation determinants and processes, and (4) building flexibility into project timelines to manage delays and challenges due to the real-world nature of policy. By applying these considerations in their own work, researchers can better identify or develop measures of policy implementation that fit their needs. The experiences of the three projects highlighted in this paper reinforce the need for high-quality and transferrable measures of policy implementation, an area where collaboration between implementation scientists and policy experts could be particularly fruitful. These measurement practices provide a foundation for the field to build on as attention to incorporating measures of policy implementation into quantitative evaluations grows and will help ensure that researchers are developing a more complete understanding of how policies impact health outcomes.

1 Introduction

Public policy plays a major role in improving public health, and most of the greatest public health achievements have resulted from policy action (1, 2). Disciplines such as economics, public administration, political science, and health services research have a long history of evaluating policies and quantifying their effects on health and related outcomes. Methodologically, this is typically operationalized by comparing outcomes in jurisdictions with and without a policy, treating the presence/absence of policies as binary conditions, and using difference-in-differences analyses or related quasi-experimental causal inference methods (3).

However, this approach masks potentially important variability in components of policy implementation, which in turn can affect outcomes and researchers' ability to draw clear conclusions about a policy's effects (36). For example, similar policies can have different provisions that affect implementation. One state with retail restrictions on electronic cigarettes may empower state public health officials to enforce the restrictions, while others may rely on local public health or law enforcement officials to ensure compliance. Likewise, variation in funding for policy enforcement and government capacity for monitoring compliance can result in heterogeneity in implementation and subsequent outcomes. In addition, even when jurisdictions (e.g., states) have the exact same provisions in their policy, how the policy is interpreted and implemented may vary based on the entity responsible for implementation. In one published example, McGinty et al. found that state opioid prescribing laws did not significantly change opioid prescriptions or nonopioid pain treatments (7). Their team's parallel qualitative work describing suboptimal implementation and limited penalties for nonadherence among these laws helped to put the quantitative findings into greater context and provided explanation as to why their findings did not change clinical practice (7, 8).

These qualitative studies demonstrate the importance of conducting research to understand policy implementation, but no measures of policy implementation were included in the quantitative components. Inclusion of quantitative measures of policy implementation, combined with qualitative findings, can generate a more holistic understanding of the mechanisms or processes that contribute to policy impact (3, 6, 9). Recent work has discussed that policy implementation can be conceptualized as an effect modifier or a mediator (6); such analyses are promising analytic approaches that are compatible with standard regression modeling methods.

Using such analytic approaches ultimately depends on being able to measure policy implementation in a rigorous way, and there is a major lack of valid and reliable measures of health policy implementation determinants, processes, and outcomes (1015). As a starting point to address the policy implementation measurement gap, provide guidance for researchers designing studies, and ultimately improve how policy implementation is analytically incorporated into quantitative policy evaluation research, we discuss four considerations for developing or identifying existing measures of policy implementation.

2 Methods

We identified our considerations by drawing on three currently funded NIH research grants that focus on policy implementation across areas of health at the state or local levels, as well as our team's expertise and subject matter knowledge. While these three grants all focus on studying policy implementation and include a quantitative component, their designs are diverse and offer potentially informative comparisons. A key thread linking all of these studies is that all focus on better understanding the “black box” of implementation after a policy has been adopted (16).

Our team met periodically to discuss the development of our considerations and design decisions. Each project lead first completed an implementation research logic model [IRLM, (17)] for their research grant. Group meetings were structured around discussing different components of the IRLM (e.g., determinants, implementation strategies, or implementation outcomes). We additionally used existing implementation science theories, models, or frameworks to guide these discussions. Specifically, we relied on the Consolidated Framework for Implementation Research (18, 19), the Expert Recommendations for Implementing Change compilation of implementation strategies (20), the Bullock et al. policy determinants and process model (21), and the Proctor Implementation Outcomes Framework (22, 23). These helped guide our group discussions and allowed the team to extract key information related to each study in a consistent format. Through group discussions, we iteratively developed a list of key considerations for developing or identifying measures based on commonalities across the three funded studies as well as our combined expertise in quantitative policy evaluation, public policy, and dissemination and implementation science.

2.1 Descriptions of included studies

The VAping POlicy Research (VAPOR) study seeks to (1) characterize the implementation of e-cigarette policies, (2) estimate the impact of these policies – accounting for strength of implementation – on e-cigarette, combustible cigarette, and cannabis use, and (3) project the future impact of alternate policy configurations using simulation modeling.

The Berkeley Choices and Health Environments at ChecKOUT (CHECKOUT) study focuses on the world's first healthy checkout policy, which prohibits the placement of high-added-sugar and high-sodium products and encourages healthy foods and beverages at store checkout areas. This work aims to (1) assess how the policy impacts store food environments, (2) how the policy impacts food purchasing, and (3) examine implementation factors that influence the effectiveness of the policy.

The 988 Lifeline financing study is focused on how states are supporting the implementation of a new three-digit dialing code for the national 988 Suicide & Crisis Lifeline, which was created by a federal law. The study aims to (1) characterize how states are financing 988 implementation, (2) explore perceptions of the financing determinants of 988 implementation success and understand the acceptability and feasibility of different financing strategies, and (3) examine how financing strategies affect policy implementation, mental health crisis, and suicide outcomes.

3 Results

We show key information from each study's design and IRLM in Table 1 and each completed IRLM is included in the Supplementary Appendix. Through our discussions of the IRLMs, we identified four key considerations to developing and identifying measures that researchers may consider as they plan and execute quantitative policy implementation studies. For each consideration discussed below, we provide illustrative examples from each study and provide a summary of the considerations in Table 2.

Table 1
www.frontiersin.org

Table 1 Key policy implementation study components.

Table 2
www.frontiersin.org

Table 2 Design considerations for measuring policy implementation.

3.1 Clearly specify the implementation logic of the policy under study

Differences between the three projects underscored that policy implementation cannot be measured with a one-size-fits-all approach. Each policy area is unique, with different policy actors, contexts, and goals - the who, what, how, and why - similar to recommendations for specifying implementation strategies (24). To appropriately identify or develop measures of policy implementation, it is important to clearly define the policy under study, its hypothesized mechanisms of action, and which implementation determinants and outcomes are relevant. Therefore, a key consideration we identified was for investigators to clearly specify the implementation logic of their policy, using existing tools and frameworks from implementation science (including the IRLM) and related fields (e.g., public administration research, political science). Having this implementation logic specified can help ensure that there is conceptual alignment between the policy exposure, implementation outcomes, and behavioral or health outcomes (25). Beyond conceptual alignment, clear specification can also help define the statistical role different elements of the study may play (e.g., mediator, moderator, confounder). Specifying these elements is essential to understand what variables are needed to sufficiently specify statistical models and identify what needs to be measured to statistically identify an effect of interest with a reasonable degree of precision. This in turn enables researchers to identify what type of measurement tools or approaches are needed, identify existing measures in the literature, and understand if new measures are needed. Measures can be derived from routinely collected administrative data or primary data collection.

3.1.1 VAPOR

The VAPOR study is specifically interested in understanding the variety of implementation strategies for different e-cigarette policies and identifying simple ways to measure the strength of policy implementation. Studying the simultaneous implementation of more than one policy means that the measurement of policy implementation cannot be policy specific. For that reason, the team developed two short questions that will be fielded to the relevant individuals within state governments: (a) the degree to which policies are implemented as written, and (b) whether they have adequate resources to implement the policies. The team has also had to make choices about classifying policies for the purposes of analyses. There is much heterogeneity between policy details, separate from implementation heterogeneity (e.g., some states with flavor restrictions prohibit sales of all flavored e-cigarettes while others allow sales of mint or menthol products). In studies with large sample sizes, this could be managed by including a variety of analytic variables capturing such heterogeneity, but in evaluations of state policies, sample sizes are limited. Thus, the VAPOR team has had to wrestle with how to collapse similar policies across states into meaningful categories while maintaining sample sizes that are needed for analyses.

As an example of how clearly specifying implementation logic can assist with conceptual clarity and model specification, the VAPOR team used the IRLM process to interrogate what mechanisms (a specific component of the IRLM) would operate as mediators or moderators. Through this process, consensus emerged that moderators are typically contextual elements (inner/outer setting) that affect the relationship between the policy and outcomes, while mediators are typically factors that lie along the causal pathway between the policy and outcomes (generally institutional changes that are directly caused by the policy and its implementation strategy). Specificity is crucial to determine what is conceptually a mediator vs. a moderator (6). We provide a general illustration of this conceptualization as well as a specific example from VAPOR in Figure 1.

Figure 1
www.frontiersin.org

Figure 1 Example of policy implementation conceptualized as a mediator or moderator.

3.1.2 CHECKOUT

In comparison to VAPOR, the CHECKOUT study focuses on the implementation of a single policy (a healthy checkout ordinance), implemented in a single jurisdiction (the city of Berkeley), and the first implementation study of this type of policy. As such, their study delves much deeper into the specifics of how the healthy checkout policy is being implemented. In-depth interviews with key stakeholders, combined with a brief quantitative survey, will generate rich data that can be used to develop and refine a broader set of quantitative measures to examine implementation heterogeneity across jurisdictions once healthy checkout policies are more widely adopted.

Although this is the first implementation study of a municipal healthy checkout policy, there are parallels between this policy and others (e.g., SSB excise taxes and restrictions on tobacco placement) that are helpful in developing an intuitive logic model. First, there is evidence from field experiments and voluntary policies that improving the healthfulness of checkouts also improves the healthfulness of consumer purchases (26). Second, prior evaluations of policy implementation have identified the importance of the following for improving health behaviors and outcomes: effective communications with retailers (e.g., definitions and lists of compliant and non-compliant products) (27), retailer compliance (e.g., the extent to which they stock only compliant products at checkout) (28), and enforcement and fines (29). Although this study is assessing implementation in a single city, the researchers expect to observe variability in how store managers and owners understand, interpret, and buy into the policy, and hence their store's compliance. Variability in compliance may also be observed over time based on the timing and robustness of the city's inspections, communications, and fines. The researchers' annual in-store assessments of products at checkout will provide objective quantitative measures of compliance, while the store interviews will indicate reasons for variability in compliance across stores and time. These data will not only inform constructs to assess in future quantitative measures of healthy retail policy implementations, but may also inform how to improve the next healthy checkout ordinance and its implementation.

3.1.3 988 Lifeline

The 988 study focuses specifically on one state financing strategy to support the implementation of the 988 Suicide & Crisis Lifeline: 988 telecom fees. Telecom fees, which are adopted by state legislatures, were identified in the federal law that created 988 as the recommended financing strategy that states should use (though there is no requirement for them to use it). These telecom fees–which are flat monthly fees on cell phone bills (e.g., 30 cents a month)–are consistent with how the 911 system in the United States is in part financed. As of March 2024, eight states had adopted 988 telecom fees. The study conceptualizes the state laws that create the fees as implementation strategies to support federal policy implementation and increase the reach of services provided by the 988 Suicide & Crisis Lifeline. The policies are operationalized as a dichotomous variable (988 telecom fee passed in the state, yes/no) as well as a continuous variable (dollar amount of revenue the 988 telecom fee generated annually per state resident).

3.2 Develop an interdisciplinary team consisting of policy practitioners and researchers with expertise in quantitative methods, public policy and law, implementation science, and subject matter knowledge

The projects illustrate the importance of involving an interdisciplinary team when measuring policy implementation. All study teams included researchers with expertise in methods and models used to evaluate health policies (e.g., difference-in-differences analyses, epidemiological and econometric methods), legal and policy expertise, implementation science, and health subject matter expertise. Implementation science expertise is critical to the clear conceptualization of different components of policy implementation, including the distinction between determinants, implementation strategies, mechanisms, processes, and outcomes – something that the team found difficult throughout our group discussions for this paper. The importance of mixed methods to policy implementation measurement (see next section “Use mixed methods…”) also necessitates team expertise with qualitative, quantitative, and mixed methods approaches. Practice partners can help guide the recruitment of individuals who are best able to provide information on policy implementation components, and support other aspects of data collection or access. Including practice partners also ensures that there is a built-in feedback loop to communicate findings to other policy practitioners who may be considering or implementing similar policies.

3.2.1 VAPOR

The VAPOR team is led by a health policy and health services researcher with expertise in tobacco control and implementation science. Additional investigators and consultants bring expertise in tobacco and e-cigarette policy, addiction medicine, youth vaping, qualitative methods, implementation science, statistics (difference-in-differences methods), and simulation modeling. The team's tobacco and e-cigarette control experts come not just from the study's research center, but from major national organizations, including the Public Health Law Center, the Truth Initiative, and the Tobacco Control Network (TCN) of the Association of State and Territorial Health Officials. These team members help guide the execution of the research, including the recruitment of key implementation actors in different states. Representatives from these organizations constitute an important advisory board that is also called on to help guide analyses - for example, using their practice-based knowledge to help identify what policy synergies are useful to probe for in analyses, both independent and dependent on implementation.

3.2.2 CHECKOUT

The CHECKOUT study is led by a nutritional epidemiologist with experience evaluating food policies and their implementation processes using mixed methods. Other investigators include health economists (difference-in-differences and synthetic control methods) and behavioral scientists, and investigators are working with the director of a food policy NGO and with a community leader and public health advocate who has deep community ties and on-the-ground experience developing, advocating for, and implementing public health policies. These practitioners are particularly helpful in understanding mechanisms of policy action and important contextual factors as well as in advising on the recruitment of participants most knowledgeable about policy implementation.

3.2.3 988 Lifeline

The project team is led by an experienced implementation scientist who focuses on policy and mental health. A statistician with substantive expertise in health policy impact analysis is integral to the team and brings deep expertise in methods related to causal inference and quasi-experimental policy impact evaluations. Given the project's focus on understanding variation across state 988 financing approaches–many of which are codified in statutes–a public health lawyer is a key member of the project team and integral to accurately specifying the different implementation financing approaches used by states. Heterogeneity in the effects of 988 financing strategies across demographic groups and equity considerations are core to the project, so a team member has expertise in racial and ethnic disparities in suicide and mental health crises. The team is also working with a public finance/accounting expert in a school of public administration to help measure and quantify financing. Finally, practice partners have been critical in helping the team stay abreast of the rapidly changing 988 financing and policy implementation environment.

3.3 Use mixed methods to identify, measure, and analyze relevant policy implementation determinants and processes

All three studies use mixed methods, though applied in different ways. Often, qualitative methods are used to understand implementation determinants, strategies, and variability across jurisdictions and implementing actors (30, 31). In all three projects discussed here, qualitative findings drive how policy implementation is measured and incorporated into quantitative policy evaluations and provide important context for quantitative findings.

3.3.1 VAPOR

The VAPOR team is conducting interviews with staff in state agencies to understand how they are taking e-cigarette laws and translating them into action on the ground. The team's interview guides are based on Bullock's policy implementation framework, with significant focus on specific questions about determinants of implementation, for example, the clarity of policy language, the degree of vertical integration within state agencies, and the existence of communication and collaboration between state agencies and outside stakeholders (e.g., businesses). Along with the interviews, the team is asking each state to evaluate (a) the degree to which policies are implemented as written, and (b) whether they have adequate resources to implement the policies. The question is asked with respect to the initial implementation period and the current period, with responses taking values from 1 to 5. These numerical assessments will form the basis for the quantitative policy evaluation, allowing the team to move beyond “0/1” policy coding and determine whether policy impacts on study outcomes are different for states reporting “well-implemented” policies vs. “poorly implemented” policies. Interview data will provide further context on why policies do/do not show evidence of effectiveness, and will help the team decide how to approach the question of additive/multiplicative policies.

3.3.2 CHECKOUT

The CHECKOUT investigators are conducting interviews, accompanied by brief quantitative surveys, with city staff, leaders, and community organizations involved in implementing the healthy checkout policy and with managers and owners of stores that are subject to the policy. The measures will characterize the implementation process and strategies, such as the overall implementation framework and timeline, the teams involved and their degree of coordination, training of inspectors, and communication of policy requirements to stores. Measures will also assess implementation outcomes, such as perceived acceptability of the policy, fidelity of enforcement, and costs of implementation, as well as other determinants of implementation, such as the complexity of policy requirements, presence of champions, prioritization of the policy, resources, and other barriers and facilitators (10, 18). A key implementation outcome–the extent to which stores comply with the policy–is being assessed quantitatively using repeated observations of products at store checkouts in Berkeley and comparison cities (32) and analyzed using difference-in-differences models. These observations of checkouts will be used to identify variability in compliance across stores and over time. Although at the time the evaluation was planned, there was only one city with a healthy checkout policy, another city, Perris, CA has since enacted a similar policy, and there are other jurisdictions also considering such policies. The quantitative measures used in this single-city study have the potential to be used in evaluations of subsequent healthy checkout policies, and the qualitative data will inform the expansion and refinement of quantitative measures.

3.3.3 988 Lifeline

Because the 988 study is largely focused on one specific policy implementation financing strategy (i.e., state telecom fee legislation), the quantitative component uses difference-in-differences analyses to understand the impact of telecom user fee legislation on key implementation outcomes (i.e., measures of 988 implementation fidelity and reach–using call volume and routing data from Vibrant Emotional Health) and effectiveness outcomes (suicide death, using Centers for Disease Control and Prevention mortality data; suicide attempts, using self-report data from the National Survey on Drug Use and Health; and emergency department use for mental health crises and self-harm, using data from the Agency for Healthcare Research and Quality's State Emergency Department Databases). Moderation analyses will assess whether the state per capita amount of telecom fee revenue affects the relationship between user fee legislation and outcomes. Prior to the difference-in-difference analyses, surveys and interviews with “policy implementers” (e.g., 988 Lifeline call center leaders, state suicide prevention coordinators) will be conducted to unpack implementation processes and mechanisms of financing strategies. The information gained from these surveys and interviews may inform decisions in the difference-in-difference analyses (e.g., inform the selection and integration of new covariates) and will aid the interpretation of results. The surveys and interviews will draw from psychometrically validated instruments and assess stakeholders' perceptions of the financing determinants of 988 implementation and the acceptability and feasibility of state legislative financing strategies to improve implementation.

3.4 Build flexibility into project timelines to manage delays and challenges due to the real-world nature of policy

These projects' operationalization of implementation measurement also illustrates that policy implementation studies must wrestle with the real-world nature of policy implementation, which is constantly changing. This consideration is especially important when projects are evaluating policies as they are being implemented. The real-world dynamics of policy implementation work mean that investigators may need to consider backup plans in case data are delayed, unavailable, or change over time. Delays in policy implementation can run up against grant timelines, requiring no-cost extensions and even additional funding to sustain repeated measures longer than the initially anticipated need - though the three studies discussed herein are in early stages and have not yet faced these challenges.

3.4.1 VAPOR

Policies affecting e-cigarettes differ across the US. VAPOR evaluations are happening over a five year grant period, and depending on the state, a policy might be recently enacted or amended, or long-standing. Repeated measures are thus critically important to fully understand implementation processes. One anticipated challenge that VAPOR has addressed is data acquisition. Some data on e-cigarette use is held by state departments of public health. Acquiring these data often requires obtaining state-specific data use agreements, which can be time-consuming to complete, file, and execute.

3.4.2 CHECKOUT

Due to staffing shortages and strains on local governments caused by the COVID-19 pandemic, there has been a delay in some key aspects of the healthy checkout policy implementation, including the rollout of inspections and subsequent technical assistance to stores and fines. Such delays are not uncommon with municipal policy, and as such, policy implementation researchers need to be flexible and prepared to shift timelines for data collection and measurement (e.g., conducting interviews and surveys). Additionally, if government staff become too busy to participate in research, it may become necessary to rely on alternative data sources, such as public records and meeting minutes, to complement interview data. Potential delays in policy implementation also highlight the importance of collecting repeated measures of implementation outcomes. The researchers' multi-year assessments of products at checkout have the potential to detect increases in in-store compliance that may be expected immediately following policy communication, inspections, and fines (33).

3.4.3 988 Lifeline

There has been more federal funding for 988 implementation than originally anticipated, and more states have been substantively supporting implementation through budget appropriations than projected. The research team has needed to modify their data collection approach to ensure that these funds are being adequately tracked and measured - including how much is being distributed to each state and how those dollars are being allocated. The 988 Lifeline has also expanded texting capacity, and thus, the team has had to revisit their initial analytic planning to make sure their variables and data appropriately capture text volume in addition to call volume. There is also a major upcoming change in how calls and texts are routed to local Lifeline centers and thus counted at the state level (34). Routing has been based on area code (high potential for measurement error or misclassification bias), and soon it will be based on geolocation (much lower potential for measurement error or misclassification bias). This is a great advancement for the real-world implementation and impact of the policy (i.e., callers will be routed based on their actual location, rather than their area code which does not necessarily reflect their current location) but poses a significant measurement challenge for the study because pre- and post-policy measures reflect different routing methods.

4 Discussion

It is critical that we incorporate policy implementation into quantitative evaluations of health policies. However, measuring policy implementation is a key gap in the literature. Indeed, while recent work has discussed how policy implementation is conceptualized in evaluations, less work has discussed how to operationalize and measure policy implementation, a prerequisite for including it in any analyses. The subfield of policy (or policy-focused) implementation science is well-poised to address this methodological gap in the literature (3, 16). Through group discussion and comparing the approaches and methods of three NIH-funded research projects, we identified four key design considerations for researchers to use to develop or identify measures of policy implementation for inclusion in quantitative analyses: (1) clearly specify the implementation logic of the policy under study, (2) develop an interdisciplinary team consisting of policy practitioners and researchers with expertise in quantitative methods, public policy and law, implementation science, and subject matter knowledge, (3) use mixed methods to identify, measure, and analyze relevant policy implementation determinants and processes, and (4) build flexibility into project timelines to manage delays and challenges due to the real-world nature of policy.

Our study reinforces the need for more work developing and validating transferrable measures of policy implementation determinants and outcomes (10). This represents a key area where implementation scientists with expertise in measure development and evaluation could greatly enhance policy-focused implementation science. Ideally, determinant and outcome measures would be transferrable across levels of policy (e.g., local, state, national), consistent within content areas, and include a focus on health equity (1315, 35). Transferrable measurements will greatly enhance our ability to derive broadly generalizable knowledge from policy studies like those discussed here. Measures that are too study-specific will have limited generalizability (though potentially higher internal validity) and limit the ability of broader learning for the field of policy-focused implementation science. Including attention to health equity in measure development will help provide comprehensive understanding of how marginalized populations are impacted by policies (15). As measures are developed, improving the coordination and use of common measures through publicly available measure repositories is crucial to improving the efficiency, reproducibility, and learning potential of policy-focused implementation science research (15). Existing repository and field-building efforts can provide guidance for how to build and disseminate such repositories (3638).

Prior systematic reviews have focused primarily on measures of policy implementation determinants and outcomes (10, 13, 14). Another area of research that needs to be expanded is understanding the process by which policy implementation occurs and how it unfolds over time. As one example, a number of studies have examined the process of implementing sugar-sweetened beverage taxes across multiple jurisdictions in the US. The collective impact has been to illustrate how the implementation of these tax policies varies by jurisdiction, including what implementation strategies were deployed across contexts (27, 39, 40). Another example investigated how three states chose to implement new substance use disorder care services under a Medicaid waiver policy and identified key implementation strategies deployed (31). As more work in the field studies policy implementation processes and identifies policy implementation strategies, ensuring that implementation strategies are clearly reported (24) and understanding the mechanisms by which implementation strategies affect implementation outcomes will be a critical next step, similar to work being undertaken in the broader field of implementation science (41).

Beyond reporting on successful implementation processes in the scientific literature, the timely and accessible sharing of this learning with policy- makers, writers, practitioners, implementers, and consumer protection organizations is key for disseminating best practices and informing future policy implementation efforts. Engaging policy practitioners as part of the research team is one avenue for timely dissemination. Also, in the process of recruiting policy implementers to participate in surveys and interviews, researchers can establish a preferred mechanism and format for the timely sharing of findings. This is crucial, as prior research has established that a “one-size-fits-all” approach to dissemination will likely not be successful (4247). Another feedback loop through which the research can strengthen future policy implementations is by presenting to coalitions of public health policy practitioners. For instance, there is a national coalition of healthy retail policy practitioners that invites researchers to present findings that could inform their future policy work.

We urge researchers to be specific about the role that policy plays in their study, particularly when outlining the implementation logic that will drive project decisions. Policy can be conceptualized in many ways in implementation science, including considering policy as the “thing” of interest, policy as an implementation strategy to put an intervention into place, and policy as a “vessel” for interventions (3, 16, 21, 48, 49). Here, all three projects have a common goal of understanding strategies or processes by which policies were (or are being) put into place; the VAPOR and CHECKOUT studies conceptualized policies as the “thing” of interest, while the 988 Lifeline study conceptualized state policies as the implementation financing strategy deployed to support implementation of the federal 988 policy and increase the reach of Lifeline services. All are crucial lines of research that are needed to improve population health, but advancements in the field and collective learning will be impeded without conceptual clarity of the role of policy in individual studies.

A limitation of this work is that due to project constraints, we were limited to three case studies, all of which were in early stages during the development of this manuscript. Our intent was to provide illustrative considerations to measuring policy implementation but are not intended to be inclusive of all possible considerations for measurement, and we do not have evidence on the success of these considerations. However, we also note that our considerations overlap with related work in the policy and implementation science fields (6, 10, 13, 48, 50). For example, Crable et al. discuss the importance of clearly specifying a policy's form and function (48), similar to our suggested practice of clearly specifying the implementation logic of the policy under study using tools such as the IRLM. Second, our practice of considering mixed methods is consistent with considerations outlined in protocol papers from other teams involved in this area (51, 52), and reflections from authors in the field of public administration (53). Third, each of the studies is currently working to handle challenges in measurement because of the real-world nature of policy, consistent with findings from a report on other funded policy implementation studies (54).

5 Conclusions

Quantitative policy evaluations provide critical knowledge of how policies impact behavioral and health outcomes, building the evidence base for further adoption elsewhere. To appropriately evaluate policy impacts on health, we must adequately measure how the policy is implemented, rather than assuming a policy is implemented just because it is “on the books” (3). In turn, this can help researchers better understand the full picture of why policies do or do not affect health outcomes, and their impact on health disparities (3, 10, 15). Policy-focused implementation science research focuses on understanding just this, but the measurement of policy implementation is lacking. Here, we describe four design considerations for policy implementation measurement, particularly when researchers are seeking to include policy implementation quantitative evaluations of health policies. These considerations provide a foundation for the field to build on as attention to measuring policy implementation grows.

Author's note

The content is solely the responsibility of the authors and does not necessarily represent the official views of the National Institutes of Health.

Data availability statement

The original contributions presented in the study are included in the article/Supplementary Material, further inquiries can be directed to the corresponding author.

Author contributions

NS: Conceptualization, Writing – original draft, Writing – review & editing. DL: Conceptualization, Supervision, Writing – original draft, Writing – review & editing. JF: Conceptualization, Supervision, Writing – original draft, Writing – review & editing. JP: Conceptualization, Supervision, Writing – original draft, Writing – review & editing. JC: Conceptualization, Supervision, Writing – original draft, Writing – review & editing.

Funding

The author(s) declare financial support was received for the research, authorship, and/or publication of this article.

This study was supported by K99CA277135 (NS), R01DA054935 (DL), R01DK135099 (JF), and R01MH131649 (JP).

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/frhs.2024.1322702/full#supplementary-material

References

1. Pollack Porter KM, Rutkow L, McGinty EE. The importance of policy change for addressing public health problems. Public Health Rep. (2018) 133(1_suppl):9S–14. doi: 10.1177/0033354918788880

PubMed Abstract | Crossref Full Text | Google Scholar

2. Domestic Public Health Achievements Team. Ten Great Public Health Achievements—United States, 2001–2010. Available online at: https://www.cdc.gov/mmwr/preview/mmwrhtml/mm6019a5.htm (cited June 1, 2021).

3. Chriqui JF, Asada Y, Smith NR, Kroll-Desrosiers A, Lemon SC. Advancing the science of policy implementation: a call to action for the implementation science field. Transl Behav Med. (2023) 13(11):820–5. doi: 10.1093/tbm/ibad034

PubMed Abstract | Crossref Full Text | Google Scholar

4. Tormohlen KN, Bicket MC, White S, Barry CL, Stuart EA, Rutkow L, et al. The state of the evidence on the association between state cannabis laws and opioid-related outcomes: a review. Curr Addict Rep. (2021) 8(4):538–45. doi: 10.1007/s40429-021-00397-1

PubMed Abstract | Crossref Full Text | Google Scholar

5. Ritter A, Livingston M, Chalmers J, Berends L, Reuter P. Comparative policy analysis for alcohol and drugs: current state of the field. Int J Drug Policy. (2016) 31:39–50. doi: 10.1016/j.drugpo.2016.02.004

PubMed Abstract | Crossref Full Text | Google Scholar

6. McGinty EE, Seewald NJ, Bandara S, Cerdá M, Daumit GL, Eisenberg MD, et al. Scaling interventions to manage chronic disease: innovative methods at the intersection of health policy research and implementation science. Prev Sci. (2024) 25(Suppl 1):96–108. doi: 10.1007/s11121-022-01427-8

PubMed Abstract | Crossref Full Text | Google Scholar

7. McGinty EE, Bicket MC, Seewald NJ, Stuart EA, Alexander GC, Barry CL, et al. Effects of state opioid prescribing laws on use of opioid and other pain treatments among commercially insured U.S. adults. Ann Intern Med. (2022) 175(5):617–27. doi: 10.7326/M21-4363

PubMed Abstract | Crossref Full Text | Google Scholar

8. Stone EM, Rutkow L, Bicket MC, Barry CL, Alexander GC, McGinty EE. Implementation and enforcement of state opioid prescribing laws. Drug Alcohol Depend. (2020) 213:108107. doi: 10.1016/j.drugalcdep.2020.108107

PubMed Abstract | Crossref Full Text | Google Scholar

9. Angrist N, Meager R. Implementation Matters: Generalizing Treatment Effects in Education (2023). Available online at: https://osf.io/https://osf.io/qtr83 (cited June 26, 2023).

10. Allen P, Pilar M, Walsh-Bailey C, Hooley C, Mazzucca S, Lewis CC, et al. Quantitative measures of health policy implementation determinants and outcomes: a systematic review. Implement Sci. (2020) 15(1):47. doi: 10.1186/s13012-020-01007-w

PubMed Abstract | Crossref Full Text | Google Scholar

11. Howlett M. Moving policy implementation theory forward: a multiple streams/critical juncture approach. Public Policy Admin. (2019) 34(4):405–30. doi: 10.1177/0952076718775791

Crossref Full Text | Google Scholar

12. Asada Y, Kroll-Desrosiers A, Chriqui JF, Curran GM, Emmons KM, Haire-Joshu D, et al. Applying hybrid effectiveness-implementation studies in equity-centered policy implementation science. Front Health Serv. (2023) 3. doi: 10.3389/frhs.2023.1220629

PubMed Abstract | Crossref Full Text | Google Scholar

13. Pilar M, Jost E, Walsh-Bailey C, Powell BJ, Mazzucca S, Eyler A, et al. Quantitative measures used in empirical evaluations of mental health policy implementation: a systematic review. Implement Res Pract. (2022) 3:26334895221141116. doi: 10.1177/26334895221141116

PubMed Abstract | Crossref Full Text | Google Scholar

14. McLoughlin GM, Allen P, Walsh-Bailey C, Brownson RC. A systematic review of school health policy measurement tools: implementation determinants and outcomes. Implement Sci Commun. (2021) 2(1):67. doi: 10.1186/s43058-021-00169-y

PubMed Abstract | Crossref Full Text | Google Scholar

15. McLoughlin GM, Kumanyika S, Su Y, Brownson RC, Fisher JO, Emmons KM. Mending the gap: measurement needs to address policy implementation through a health equity lens. Transl Behav Med. (2024) 14(4):207–14. doi: 10.1093/tbm/ibae004

PubMed Abstract | Crossref Full Text | Google Scholar

16. Purtle J, Moucheraud C, Yang LH, Shelley D. Four very basic ways to think about policy in implementation science. Implement Sci Commun. (2023) 4(1):111. doi: 10.1186/s43058-023-00497-1

PubMed Abstract | Crossref Full Text | Google Scholar

17. Smith JD, Li DH, Rafferty MR. The implementation research logic model: a method for planning, executing, reporting, and synthesizing implementation projects. Implement Sci. (2020) 15(1):84. doi: 10.1186/s13012-020-01041-8

PubMed Abstract | Crossref Full Text | Google Scholar

18. Damschroder LJ, Reardon CM, Widerquist MAO, Lowery J. The updated consolidated framework for implementation research based on user feedback. Implement Sci. (2022) 17(1):75. doi: 10.1186/s13012-022-01245-0

PubMed Abstract | Crossref Full Text | Google Scholar

19. Damschroder LJ, Aron DC, Keith RE, Kirsh SR, Alexander JA, Lowery JC. Fostering implementation of health services research findings into practice: a consolidated framework for advancing implementation science. Implement Sci. (2009) 4(1):50. doi: 10.1186/1748-5908-4-50

PubMed Abstract | Crossref Full Text | Google Scholar

20. Powell BJ, Waltz TJ, Chinman MJ, Damschroder LJ, Smith JL, Matthieu MM, et al. A refined compilation of implementation strategies: results from the expert recommendations for implementing change (ERIC) project. Implement Sci. (2015) 10(1):21. doi: 10.1186/s13012-015-0209-1

PubMed Abstract | Crossref Full Text | Google Scholar

21. Bullock HL, Lavis JN, Wilson MG, Mulvale G, Miatello A. Understanding the implementation of evidence-informed policies and practices from a policy perspective: a critical interpretive synthesis. Implement Sci. (2021) 16(1):18. doi: 10.1186/s13012-021-01082-7

PubMed Abstract | Crossref Full Text | Google Scholar

22. Proctor E, Silmere H, Raghavan R, Hovmand P, Aarons G, Bunger A, et al. Outcomes for implementation research: conceptual distinctions, measurement challenges, and research agenda. Adm Policy Ment Health. (2011) 38(2):65–76. doi: 10.1007/s10488-010-0319-7

PubMed Abstract | Crossref Full Text | Google Scholar

23. Proctor EK, Bunger AC, Lengnick-Hall R, Gerke DR, Martin JK, Phillips RJ, et al. Ten years of implementation outcomes research: a scoping review. Implement Sci. (2023) 18(1):31. doi: 10.1186/s13012-023-01286-z

PubMed Abstract | Crossref Full Text | Google Scholar

24. Proctor EK, Powell BJ, McMillen JC. Implementation strategies: recommendations for specifying and reporting. Implement Sci. (2013) 8(1):139. doi: 10.1186/1748-5908-8-139

PubMed Abstract | Crossref Full Text | Google Scholar

25. Ding D, Gebel K. Built environment, physical activity, and obesity: what have we learned from reviewing the literature? Health Place. (2012) 18(1):100–5. doi: 10.1016/j.healthplace.2011.08.021

PubMed Abstract | Crossref Full Text | Google Scholar

26. Ejlerskov KT, Sharp SJ, Stead M, Adamson AJ, White M, Adams J. Supermarket policies on less-healthy food at checkouts: natural experimental evaluation using interrupted time series analyses of purchases. PLoS Med. (2018) 15(12):e1002712. doi: 10.1371/journal.pmed.1002712

PubMed Abstract | Crossref Full Text | Google Scholar

27. Falbe J, Grummon AH, Rojas N, Ryan-Ibarra S, Silver LD, Madsen KA. Implementation of the first US sugar-sweetened beverage tax in Berkeley, CA, 2015–2019. Am J Public Health. (2020) 110(9):1429–37. doi: 10.2105/AJPH.2020.305795

PubMed Abstract | Crossref Full Text | Google Scholar

28. Zacher M, Germain D, Durkin S, Hayes L, Scollo M, Wakefield M. A store cohort study of compliance with a point-of-sale cigarette display ban in Melbourne, Australia. Nicotine Tob Res. (2013) 15(2):444–9. doi: 10.1093/ntr/nts150

PubMed Abstract | Crossref Full Text | Google Scholar

29. Jacobson PD, Wasserman J. The implementation and enforcement of tobacco control laws: policy implications for activists and the industry. J Health Polit Policy Law. (1999) 24(3):567–98. doi: 10.1215/03616878-24-3-567

PubMed Abstract | Crossref Full Text | Google Scholar

30. Economou MA, Kaiser BN, Yoeun SW, Crable EL, McMenamin SB. Applying the EPIS framework to policy-level considerations: tobacco cessation policy implementation among California medicaid managed care plans. Implement Res Pract. (2022) 3:26334895221096289. doi: 10.1177/26334895221096289

PubMed Abstract | Crossref Full Text | Google Scholar

31. Crable EL, Benintendi A, Jones DK, Walley AY, Hicks JM, Drainoni ML. Translating medicaid policy into practice: policy implementation strategies from three US states’ experiences enhancing substance use disorder treatment. Implement Sci. (2022) 17(1):3. doi: 10.1186/s13012-021-01182-4

PubMed Abstract | Crossref Full Text | Google Scholar

32. Falbe J, Marinello S, Wolf EC, Solar SE, Schermbeck RM, Pipito AA, et al. Food and beverage environments at store checkouts in California: mostly unhealthy products. Curr Dev Nutr. (2023) 7(6):100075. doi: 10.1016/j.cdnut.2023.100075

PubMed Abstract | Crossref Full Text | Google Scholar

33. Powell LM, Li Y, Solar SE, Pipito AA, Wolf EC, Falbe J. (2022). Report No.: 125. Development and Reliability Testing of the Store CheckOUt Tool (SCOUT) for Use in Healthy Checkout Evaluations. Chicago, Illinois: Policy, Practice, and Prevention Research Center, University of Illinois Chicago. Available online at: https://p3rc.uic.edu/research-evaluation/evaluation-of-food-policies/retail-food-and-beverages/

34. Purtle J, Goldman ML, Stuart EA. Interpreting between-state variation in 988 suicide and crisis lifeline call volume rates. Psychiatr Serv. (2023) 74(9):901. doi: 10.1176/appi.ps.23074015

PubMed Abstract | Crossref Full Text | Google Scholar

35. Phulkerd S, Lawrence M, Vandevijvere S, Sacks G, Worsley A, Tangcharoensathien V. A review of methods and tools to assess the implementation of government policies to create healthy food environments for preventing obesity and diet-related non-communicable diseases. Implement Sci. (2016) 11(1):15. doi: 10.1186/s13012-016-0379-5

PubMed Abstract | Crossref Full Text | Google Scholar

36. McKinnon RA, Reedy J, Berrigan D, Krebs-Smith SM. The national collaborative on childhood obesity research catalogue of surveillance systems and measures registry: new tools to spur innovation and increase productivity in childhood obesity research. Am J Prev Med. (2012) 42(4):433–5. doi: 10.1016/j.amepre.2012.01.004

PubMed Abstract | Crossref Full Text | Google Scholar

37. Rabin BA, Lewis CC, Norton WE, Neta G, Chambers D, Tobin JN, et al. Measurement resources for dissemination and implementation research in health. Implement Sci. (2016) 11(1):42. doi: 10.1186/s13012-016-0401-y

PubMed Abstract | Crossref Full Text | Google Scholar

38. Lewis CC, Mettert KD, Dorsey CN, Martinez RG, Weiner BJ, Nolen E, et al. An updated protocol for a systematic review of implementation-related measures. Syst Rev. (2018) 7(1):66. doi: 10.1186/s13643-018-0728-3

PubMed Abstract | Crossref Full Text | Google Scholar

39. Chriqui JF, Sansone CN, Powell LM. The sweetened beverage tax in cook county, Illinois: lessons from a failed effort. Am J Public Health. (2020) 110(7):1009–16. doi: 10.2105/AJPH.2020.305640

PubMed Abstract | Crossref Full Text | Google Scholar

40. Asada Y, Pipito AA, Chriqui JF, Taher S, Powell LM. Oakland’s sugar-sweetened beverage tax: honoring the “spirit” of the ordinance toward equitable implementation. Health Equity. (2021) 5(1):35–41. doi: 10.1089/heq.2020.0079

PubMed Abstract | Crossref Full Text | Google Scholar

41. Lewis CC, Klasnja P, Lyon AR, Powell BJ, Lengnick-Hall R, Buchanan G, et al. The mechanics of implementation strategies and measures: advancing the study of implementation mechanisms. Implement Sci Commun. (2022) 3(1):114. doi: 10.1186/s43058-022-00358-3

PubMed Abstract | Crossref Full Text | Google Scholar

42. Ashcraft LE, Quinn DA, Brownson RC. Strategies for effective dissemination of research to United States policymakers: a systematic review. Implement Sci. (2020) 15(1):89. doi: 10.1186/s13012-020-01046-3

PubMed Abstract | Crossref Full Text | Google Scholar

43. Purtle J. Disseminating evidence to policymakers: accounting for audience heterogeneity. In: Weber MS, Yanovitzky I, editors. Networks, Knowledge Brokers, and the Public Policymaking Process. Cham: Springer International Publishing (2021). p. 27–48. doi: 10.1007/978-3-030-78755-4_2

Crossref Full Text | Google Scholar

44. Smith NR, Mazzucca S, Hall MG, Hassmiller Lich K, Brownson RC, Frerichs L. Opportunities to improve policy dissemination by tailoring communication materials to the research priorities of legislators. Implement Sci Commun. (2022) 3:24. doi: 10.1186/s43058-022-00274-6

PubMed Abstract | Crossref Full Text | Google Scholar

45. Purtle J, Lê-Scherban F, Wang X, Shattuck PT, Proctor EK, Brownson RC. Audience segmentation to disseminate behavioral health evidence to legislators: an empirical clustering analysis. Implement Sci. (2018) 13(1):121. doi: 10.1186/s13012-018-0816-8

PubMed Abstract | Crossref Full Text | Google Scholar

46. Crable EL, Grogan CM, Purtle J, Roesch SC, Aarons GA. Tailoring dissemination strategies to increase evidence-informed policymaking for opioid use disorder treatment: study protocol. Implement Sci Commun. (2023) 4(1):16. doi: 10.1186/s43058-023-00396-5

PubMed Abstract | Crossref Full Text | Google Scholar

47. Purtle J, Lê-Scherban F, Nelson KL, Shattuck PT, Proctor EK, Brownson RC. State mental health agency officials’ preferences for and sources of behavioral health research. Psychol Serv. (2020) 17(S1):93–7. doi: 10.1037/ser0000364

PubMed Abstract | Crossref Full Text | Google Scholar

48. Crable EL, Lengnick-Hall R, Stadnick NA, Moullin JC, Aarons GA. Where is “policy” in dissemination and implementation science? Recommendations to advance theories, models, and frameworks: EPIS as a case example. Implement Sci. (2022) 17(1):80. doi: 10.1186/s13012-022-01256-x

PubMed Abstract | Crossref Full Text | Google Scholar

49. Oh A, Abazeed A, Chambers DA. Policy implementation science to advance population health: the potential for learning health policy systems. Front Public Health. (2021) 9. doi: 10.3389/fpubh.2021.681602

Crossref Full Text | Google Scholar

50. Hoagwood KE, Purtle J, Spandorfer J, Peth-Pierce R, Horwitz SM. Aligning dissemination and implementation science with health policies to improve children’s mental health. Am Psychol. (2020) 75(8):1130–45. doi: 10.1037/amp0000706

PubMed Abstract | Crossref Full Text | Google Scholar

51. McGinty EE, Tormohlen KN, Barry CL, Bicket MC, Rutkow L, Stuart EA. Protocol: mixed-methods study of how implementation of US state medical cannabis laws affects treatment of chronic non-cancer pain and adverse opioid outcomes. Implement Sci. (2021) 16(1):2. doi: 10.1186/s13012-020-01071-2

PubMed Abstract | Crossref Full Text | Google Scholar

52. McGinty EE, Stuart EA, Caleb Alexander G, Barry CL, Bicket MC, Rutkow L. Protocol: mixed-methods study to evaluate implementation, enforcement, and outcomes of U.S. state laws intended to curb high-risk opioid prescribing. Implement Sci. (2018) 13(1):37. doi: 10.1186/s13012-018-0719-8

PubMed Abstract | Crossref Full Text | Google Scholar

53. Roll S, Moulton S, Sandfort J. A comparative analysis of two streams of implementation research. J Public Nonprofit Affairs. (2017) 3(1):3–22. doi: 10.20899/jpna.3.1.3-22

Crossref Full Text | Google Scholar

54. Asada Y, Smith NR, Kroll-Desrosiers A, Chriqui JF. Policy implementation science: NIH-funded case examples and lessons learned. Consortium for Cancer Implementation Science (2022).

Keywords: health policy, implementation science, policy research, policy implementation science, policy evaluation

Citation: Smith NR, Levy DE, Falbe J, Purtle J and Chriqui JF (2024) Design considerations for developing measures of policy implementation in quantitative evaluations of public health policy. Front. Health Serv. 4:1322702. doi: 10.3389/frhs.2024.1322702

Received: 16 October 2023; Accepted: 20 June 2024;
Published: 15 July 2024.

Edited by:

Yanfang Su, University of Washington, United States

Reviewed by:

Heather Brown, Lancaster University, United Kingdom
Andrea Titus, New York University, United States

© 2024 Smith, Levy, Falbe, Purtle and Chriqui. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Natalie Riva Smith, nataliesmith@hsph.harvard.edu

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.