- 1Behavioural Science Group, Warwick Business School, University of Warwick, Coventry, United Kingdom
- 2Division of Health Sciences, Warwick Medical School, University of Warwick, Coventry, United Kingdom
- 3Liberal Arts, University of Health Sciences and Pharmacy, St Louis, MO, United States
- 4Alliance Manchester Business School, University of Manchester, Manchester, United Kingdom
- 5Institute for Applied Health Research, University of Birmingham, Birmingham, United Kingdom
- 6Heartlands Hospital, University Hospitals Birmingham NHS Foundation Trust, Birmingham, United Kingdom
Background: Hospital decision-makers have limited resources to implement quality improvement projects. To decide which interventions to take forward, trade-offs must be considered that inevitably turn on stakeholder preferences. The multi-criteria decision analysis (MCDA) approach could make this decision process more transparent.
Method: An MCDA was conducted to rank-order four types of interventions that could optimise medication use in England's National Healthcare System (NHS) hospitals, including Computerised Interface, Built Environment, Written Communication, and Face-to-Face Interactions. Initially, a core group of quality improvers (N = 10) was convened to determine criteria that could influence which interventions are taken forward according to the Consolidated Framework for Implementation Research. Next, to determine preference weightings, a preference survey was conducted with a diverse group of quality improvers (N = 356) according to the Potentially All Pairwise Ranking of All Possible Alternatives method. Then, rank orders of four intervention types were calculated according to models with criteria unweighted and weighted according to participant preferences using an additive function. Uncertainty was estimated by probabilistic sensitivity analysis using 1,000 Monte Carlo Simulation iterations.
Results: The most important criteria influencing what interventions were preferred was whether they addressed “patient needs” (17.6%)' and their financial “cost (11.5%)”. The interventions' total scores (unweighted score out of 30 | weighted out of 100%) were: Computerised Interface (25 | 83.8%), Built Environment (24 | 79.6%), Written Communication (22 | 71.6%), and Face-to-Face (22 | 67.8%). The probabilistic sensitivity analysis revealed that the Computerised Interface would be the most preferred intervention over various degrees of uncertainty.
Conclusions: An MCDA was conducted to rank order intervention types that stand to increase medication optimisation across hospitals in England. The top-ranked intervention type was the Computerised Interface. This finding does not imply Computerised Interface interventions are the most effective interventions but suggests that successfully implementing lower-ranked interventions may require more conversations that acknowledge stakeholder concerns.
Background
Hospital quality improvement teams, including operational and transformation management teams, are charged with improving patient outcomes and workforce efficacy (1). Often these teams are presented with multiple options for new interventions but have limited resources to implement them all. To determine which interventions to implement trade-offs must be considered that inevitably depend on stakeholder preferences. While many teams already use frameworks such as Six Sigma Lean and The Model for Improvement to organise a scientific process of change (1–5), these frameworks do not help determine which interventions to implement. The multi-criteria decision analysis (MCDA) approach could make this decision-making process more transparent.
MCDA is an umbrella term that describes structured decision-making approaches to prioritise, rank, or choose options based on multiple criteria (6–8). Any MCDA involves three key steps: (a) defining a decision problem that requires choosing among options, (b) selecting multiple criteria upon which those options can be judged, and (c) constructing a performance matrix to rank the options that can serve as a framework for future decisions (9). The MCDA approach is commonly used to help committees, e.g., select job candidates or prioritise new business opportunities (10–12). For narrower decision problems, the MCDA criteria may be equally weighted (i.e., unweighted), as often little information exists to determine the relative importance of each criterion, e.g., to say whether communication or teamwork skills are more important for a particular job. In contrast, for broader decision problems, determining the relative importance (i.e., weights) of each criterion can be a major concern. Here larger samples of diverse stakeholder perspectives can be considered.
MCDA has been used to rank national treatment and reimbursement priorities, and to inform local hospital formularies (8, 9, 13, 14). Formularies are an important component of medication optimisation with cost implications. In England's National Health Service (NHS), the total expenditure on medicines increased by 4.6%, from £16.4 billion in 2019/20 to £17.1 billion in 2020/21. Roughly half (44.4%) of this cost was issued by hospitals (15). NHS hospital trusts already negotiate for the most cost-effective medications and much of this increase is due to the rising patient demand. Interventions dictating which medications clinicians can prescribe could prove effective, but they may be difficult to implement as they appear to question clinician authority or ability to offer patients the best care. Nudge interventions offer an opportunity to optimise medication use without curtailing clinicians' ability to choose the medications they believe best serve patient needs (16).
A recent literature review found 20 nudge interventions issued to optimise medication use in hospital settings of which 16 were successful (9). These nudge interventions are often inexpensive to deliver, e.g., changing message wordings on paper or computerised instructions; however, criteria beyond cost may influence which nudge interventions are the most appealing. The Consolidated Framework for Implementation Research describes 41 empirically and theoretically support criteria that influence the implementation of evidence-based findings (17). Given the large array of nudge intervention options (at least 20) and implementation criteria (at least 41), quality improvement teams could be well served by a framework rank ordering types of nudge interventions they could take forward.
In the current study, an MCDA approach is used to develop a framework that quality improvement teams can use to make more transparent decisions between nudge interventions to optimise medication use in England's NHS hospitals. Additionally, the current study demonstrates how a large number of perspectives can be synthesised to inform a national framework.
Methods
We adapted the International Society for Pharmacoeconomics and Outcomes Research (ISPOR) eight-step guidelines for conducting and reporting an MCDA (9). Each step is presented in Figure 1 and detailed further below. The initial four steps describe setting up the problem and the available choice options along with the criteria upon which those options can be judged. In the current project, the first four steps were considered pre-protocol work that shaped our cross-sectional survey. The fifth step involves conducting a cross-sectional survey of national stakeholder preferences for changes within and between criteria. The methods for the first to fifth steps are detailed in the methods section, followed by the sixth and seventh steps in the analysis and the results sections. The final step is to report the findings to the funder and stakeholders.
Step 1—defining the problem
During early consultations with hospital teams, we presented up to 20 nudge intervention options to optimise medication use previously identified in a systematic review (16). The hospital teams appeared overwhelmed and often did not choose any option, plausibly due to “choice overload” (18). The research team determined that a multi-criteria decision analysis (MCDA) could provide a robust and transparent framework to facilitate these decisions. To further simplify alternative intervention options, the research team categorise the 20 intervention options into a smaller number of nudge intervention types. Our categorisations were cross-checked and approved by representatives at Public Health England.
Step 2—selecting criteria
To select the criteria upon which the intervention types would be judged, a core group of 10 quality improvers was formed including 4 doctors, 2 pharmacists, and 4 managers. The core group was asked to review the 41 constructs from the Consolidated Framework for Implementation Research (CFIR) (17). Within this framework, constructs are organised across five domains: characteristics of the innovation in question (e.g., two constructs include adaptability and trialability), the individuals involved (e.g., knowledge and self-efficacy), the inner setting (e.g., compatibility and available resources), the outer setting (e.g., incentives and patient needs), and the process encouraging uptake (e.g., planning and opinion leaders). Each improver was asked to rate each construct based on their perceived importance, from 1 (least important) to 5 (most important). Then, each construct was ranked based on the sum of those scores. The top 10 constructs were selected as the criteria considered in the present MCDA.
Step 3—measuring performance
To measure levels of quality for each criterion, three-point Likert scales were created to indicate high, medium, and low performance levels.
Step 4—scoring alternatives
Two researchers (SK and UT) used consensus discussions to assess the expected performance of each type of intervention against each criterion. Next, two additional researchers (KAS and IV), each of whom had at least 10 years of experience implementing nudge interventions in healthcare settings, cross-checked these decisions. Next, a performance matrix was assembled with the agreed rankings. A RAG system was used to indicate the quality criteria rankings, such that low-quality rankings would be designated in red, medium in amber, and high in green.
Step 5—weighting criteria
A cross-sectional survey was conducted between October 2021 and December 2021. The survey was designed using 1,000minds Decision-making software.
Participants
To capture more diverse views across a national context, our participants included staff involved in making suggestions or decisions about the implementation of the quality improvement projects for NHS hospital organisations in England such as doctors, nurses, pharmacists, and quality improvement managers (19). As we planned to recruit our participants via email, we anticipated a low rate of those emails being opened, 20%–25%, and that only 15%–30% of those opened would be completed (20, 21). We worked with the Health Foundation (described below) to identify at least 4,000 potential participants and anticipated that at least 100 would complete the survey.
Identification and recruitment of participants
The opportunity to take part in the survey was initially advertised via email to 4,439 members of the Health Foundation's Q-Community. The Health Foundation is a charitable foundation charged with improving healthcare quality. The Q-Community encompasses individuals who self-identify as healthcare quality improvers in England. At the end of the survey, participants were asked to identify other potential participants, i.e., snowball recruitment. The survey was also advertised by researchers on Twitter and LinkedIn. Participants were asked to complete the survey within two weeks. One reminder was sent to encourage completion.
Survey
The survey was designed according to the Potentially All Pairwise Ranking of All Possible Alternatives (PAPRIKA) method (22). PAPRIKA is a method for scoring additive multi-attribute values using pairwise rankings of alternatives (23). Each question asked participants to choose between two alternative intervention types differing according to two criteria with differing performance levels. An example question appears in Figure 2. Future questions were adaptively selected based on participant responses to previous questions.
Step 6—calculating aggregate scores
Our research team calculated the aggregated scores for each intervention type when the criteria were unweighted and weighted. Unweighted scores were calculated by adding the intervention scores against each criterion based on the researcher team's assigned performance levels. Weighted scores were calculated using an additive model that combines scores and weights in a way that is consistent with stakeholders' expressed preferences. The form of an additive function was given according to the formula
where Vj is the overall value of intervention j, Sij is the score for intervention j on criterion i, and Wi is the weight attached to criterion i. Then, each intervention type was rank ordered according to its total score.
Step 7—dealing with uncertainty
To assess the consistency of participant responses we developed a decision-analytic model comprising criteria, weights, and scores in TreeAge® Pro R1. A probabilistic sensitivity analysis was carried out using a Monte Carlo simulation with 1,000 iterations for the following three scenarios: (a) incorporating uncertainty in scoring the alternatives, (b) incorporating uncertainty in the weighting preference, and (c) incorporating uncertainty for both the scoring alternatives and weighing preference. A gamma distribution was used for scores. Because preference weights are positive proportions that totalled one, we selected a Dirichlet distribution to preference weights in the simulation (24). The means of observed data were used as parameters for this distribution.
Step 8—reporting and examining findings
After the final report is shared with the Health Foundation, a brief synopsis of the results will be made available to members of the Q-Community along with a link to the publication for further information.
Results
Step 1—agreed decision problem
The research team determined that a multi-criteria decision analysis (MCDA) could help quality improvement teams select nudge interventions to optimise medication use. To promote a more generalisable framework, the 20 nudge interventions located in the systematic review were grouped into 4 intervention types, including Computerised Interfaces; Built Environment; Written Communications; and Face-to-Face Interactions. Computerised Interface interventions could include reminder prompts or default choice orders presented on the computer. Built Environment interventions could involve changing the physical space within which care occurs or the materials available. Written Communication interventions could include letters or emails to staff or adjustments to existing paper forms. Face-to-Face Interaction interventions could involve meetings within and across departments.
Step 2—selected criteria
The 10 selected decision criteria are provided in Table 1 in the order the working group advised. All of the Consolidated Framework for Intervention Research's (CFIR's) domains were represented among the selected criteria, including four intervention criteria, two outer setting criteria, two individual characteristics criteria, one inner setting criterion, and one implementation process criterion. Our working group advised that the original terms used in the CFIR were too technical for our intended participants. Informed by each construct's definition, KAS and IV simplified the terms by adding implementation context. In the second column of Table 1, added words are given in italics.
Step 3—measured performance
The measures of performance assigned are stated in the final column of Table 1. For example, the CFIR criterion “adaptability to meet local needs” was measured by using each of the following three values: 1 = Highly adaptable, 2 = Moderately adaptable, or 3 = Less adaptable.
Step 4—alternatives scored
The scores for each intervention type by criteria were assembled in a performance matrix (Table 2). Only the Written Communications and Face-to-Face Interactions intervention types had low-quality ranking aspects.
Step 5—weighted criteria
Participants
The survey was completed by 356 participants, which was greater than anticipated. The average participant completed the survey in 8.6 (SD = 2.3) minutes. Most participants identified as male (57%), were between 31 and 50 years old (76.4%), and had 6 to 10 years of professional work experience (40%). All nine geographic regions of England were represented. Further details are presented in Table 3.
Preference weight for criteria
The weights given by the participants to each criterion are represented in Figure 3. Based on the mean preference values, the most highly weighted criterion appears at the top of the chart, followed, in a clockwise direction, by the second-most highly weighted criterion, and so on. Based on the preference of stakeholders, whether a particular intervention addresses patient needs was the most important (17.6% weight out of 100% was attributed to this criteria) followed by its likely financial cost (11.5%), and evidence supporting its effectiveness (10.5%). Planning required to facilitate the intervention implementation (7.8%) was the least important criterion. In addition to the criteria weights, the weights for each level within the criteria were also calculated, see Table 4.
Step 6 –Aggregate scores
Table 5 displays the ranking of different intervention types based on their total scores in two scenarios: unweighted and weighted. In the unweighted scenario, the Computerised Interface intervention type garnered the highest total score of 25, closely followed by the Built Environment with a score of 24. Both the Written Communication and Face-to-Face intervention types received a total score of 22, jointly ranking third. In the weighted scenario, the top two intervention types remain the same as in the unweighted scenario, with Computerised Interface receiving a total score of 83.8 and the Built Environment receiving 79.6. However, the ranking of Written Communication and Face-to-Face interventions change. Written Communication receives a total score of 71.6, ranking third, while Face-to-Face interaction receives a total score of 67.8, ranking fourth.
These scores represent the preferences given by quality improvers in the NHS to different criteria, including patient needs and resources, cost to implement, adaptability to meet local needs, relative priority and advantage of implementing the intervention, available resources for implementation, evidence about the intervention's impact, knowledge and beliefs about the intervention, identification with the organisation, and planning to facilitate intervention implementation. The weights for these criteria were obtained from the perceived importance or preference assigned by the quality improvers in the NHS. The weighted scores are given as percentages, reflecting the importance assigned to each criterion.
Step 7—capturing uncertainty
Table 6 and Figure 4 illustrate the outcomes of a Monte Carlo simulation (probabilistic sensitivity analysis), presenting the anticipated scores for distinct intervention types across various scenarios. The scenarios comprise the base model (expected scores), uncertainty in scoring, uncertainty in weighting preferences, and uncertainty in both scoring and weighting. In all scenarios, the Computerised Interface intervention type remains the top-ranked intervention, receiving the highest expected scores. For instance, in scenario 1 (uncertainty in scoring), the anticipated values for Computerised Interface interventions indicate that they are likely to be chosen 78 times out of 100. Conversely, the Face-to-Face intervention type does not receive the highest expected scores in any of the scenarios, indicating that it is less probable to be prioritised over the other three types.
The anticipated values for each intervention type are presented in percentage, with their mean scores and 95% confidence intervals. In the base model, the expected scores for Computerised Interface, Built Environment, Written Communication, and Face-to-Face interventions are 83.8, 79.6, 71.6, and 67.8, respectively. In scenario 1, the anticipated scores decrease slightly for all intervention types, but the rank order stays the same.
In scenario 2 (uncertainty in weighting preferences), the anticipated scores for all intervention types are comparable to those in the base model, signifying that the ranking of the intervention types is not substantially affected by uncertainty in weighting preferences. In scenario 3 (uncertainty in scoring and weighting preferences), the expected scores decrease further for all intervention types, but the rank order remains the same as in the base model. Overall, the outcomes suggest that the Computerised Interface intervention type is the most highly prioritised intervention for quality improvement in healthcare settings, with the Built Environment intervention type ranking second.
Discussion
A multicriteria decision analysis (MCDA) was conducted to describe quality improvers' preference for implementing nudge-type interventions to optimise medication use for NHS hospitals in England. Our findings suggest that quality improvers' intervention preferences are most influenced by patient needs followed by the interventions' financial costs. While the top-ranked intervention types, Computerised Interface interventions and Built Environment interventions, were the same for weighted and unweighted scores, preferences for the bottom-ranked interventions were different. Specifically, while the research team's unweighted rankings predicted that the Written Communications and Face-to-Face interventions would not differ, quality improvers preferred Written Communication interventions over Face-to-Face interventions. Our resultant framework could serve as a decision aid to assist emerging quality improvement initiatives across NHS hospitals in England.
MCDA can be a useful tool for decision-making in quality improvement programmes. Based on the authors' knowledge, this is the first paper in which MCDA has been performed for a quality improvement programme. But the MCDA approach is not entirely new. For example, MCDA has been used to prioritise non-pharmacological treatment options for patients with abdominal cancer suffering from chronic pain (25). In that study, a smaller number of health workers than in the current study, just five clinicians, rated each treatment across five criteria including convenience, pain, risk, duration, and cost. Convenience was weighted higher than costs, and similar to our findings their findings differed depending on whether the criteria were weighted. In another study (26), MCDA was used to prioritise different groups of people who would be offered COVID-19 vaccinations when vaccines were in short supply. Here, 60 physicians took part. Note that in both cases, participants were limited to health workers.
Outside of treatment-focused decisions, MCDA has been used to prioritise funding for a national healthcare service in New Zealand (27). Here, focus group discussions took place involving diverse stakeholder groups: 4 members of the public, 10 front-line workers, 7 retirees, 5 public health workers, and 13 members of a health provider group focused on an indigenous population. Stakeholder discussions revolved around not just how to score alternatives, but also around what the criteria meant. How and when to bring different stakeholder groups into the discussion making processes is an important consideration. While considering divergent perspectives may increase buy-in across stakeholder groups, it will also take considerable time and resources to foster a shared understanding of the criteria assessed and the quantitative processes underlying the resultant scores (14).
The present study focuses on the preferences of quality improvers, as this stakeholder group is most directly charged with building and sustaining the momentum necessary for successful implementation. Quality improvers' preferences were assessed across ten of the Consolidated Framework for Implementation Researcher (CFIR) constructs, as this is a widely recognized framework to support implementation. The CFIR constructs include information that quality improvers can meaningfully react to in a self-completion online survey; these constructs could prove challenging for other stakeholder groups to meaningfully react to in this setting. Of course, to build and sustain momentum, quality improvers must understand the preferences of other stakeholders, including patients whose voices should be heard as the intervention takes shape. The MCDA process could be used to foster shared decision-making and understanding across stakeholder groups and determine how a quality improvement programme is implemented, e.g., when, what materials, how often, etc.”
Overall, MCDA can be a versatile tool in healthcare decision-making, from evaluating treatment options, to prioritising project funding and quality improvement programmes.
Limitations
A limitation of the MCDA approach is that the ultimate framework will depend on the participants who take part (28). While our participant sample is large, it is a convenience sample, and we did not collect sufficient demographics to ensure historically underrepresented groups were represented. Further, it is uncertain whether the preferences captured in this survey will remain stable as interventions themselves evolve. That said, as the capabilities of computerised software itself and the people who use that software expand, we suspect that preferences for computerised interventions are likely to remain stable or increase (29, 30).
Another limitation of the MCDA approach is that it only reveals which intervention options people prefer; it does not tell us which intervention options are most likely to be effective. The authors acknowledge that any decision to initiate or implement quality improvement programmes in the NHS would be a multifactorial decision, however, understanding quality improvers' preferences is a key issue. For example, many interventions that have the potential to be effective in optimal conditions fail to be effective where workers do not support their use. Thus, where a less preferred intervention type is rolled out, extra attention in the implementation process should be devoted to increasing people's acceptance (31).
Implications
One implication of our finding is that while the top preferred intervention options may be clear, the ordering of lower-ranked interventions may be less certain. When the top-ranked option is not feasible, researchers should work closely with stakeholders to select and co-design the implementation process for less preferred intervention options. This recommendation likely extends to complex interventions, which may include combinations of high-ranked and lower-ranked intervention types (32). For instance, a computerised decision aid integrated into the hospital prescribing software may include prompts facilitating shared decision-making with pharmacists, patients, and carers. Where these prompts minimise perceived threats to professional autonomy and preserve a natural doctor-patient conversation, they are more likely to be accepted (33). Implementing such an intervention may require face-to-face meetings in advance to co-design an acceptable prompt.
The current study demonstrates how many perspectives can be synthesised to inform a national framework. Implementing such a project required access to quality improvers across England along with statistical expertise and dedicated software. As new decisions arise within local settings, a narrower MCDA may suffice to transparently capture stakeholder perspectives and to consider a broader range of criteria (7). Where possible, patients could be part of this process (34). Wider implementation of the MCDA methodology may require further top-down support and bottom-up capacity (35). For instance, demand issues may stem from board members being unaware that these decisions could be more transparent. Supply issues may stem from quality improvers not knowing how to gather or present this type of support. Multi-tiered initiatives could be used to simultaneously address these supply and demand issues (36).
Conclusion
A multicriteria decision analysis framework revealed that quality improvers prefer medication-optimising interventions that address patient needs and that are less financially costly. Thus, where the costs of two nudge interventions are similar, preference will largely turn on the perceived ability of the intervention to meet patient needs. At the time of this study, the Computerised Interface intervention type was most preferred. This finding aligns with expanding technological capabilities in healthcare and may promote future innovative low-cost opportunities to overcome practical problems. The current research team cautions that stakeholders do not always prefer what is likely to be the most effective intervention. Interventions that have the potential to be effective but are not immediately preferred must be implemented in a manner that respects stakeholders' preferences.
Data availability statement
The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding authors.
Ethics statement
The studies involving human participants were reviewed and approved by Humanities and Social Sciences Research Ethics Committee (HSSREC) at the University of Warwick, United Kingdom (Ref: HSSREC 185/20-21). The patients/participants provided their written informed consent to participate in this study.
Author contributions
Conceptualization, SK, KS, UT, AT, and IV; data curation, SK and SK; formal analysis, SK and KS; funding acquisition, KS, IV, and AT.; investigation, SK, KS, UT, AT, and IV; methodology, SK and KS; project administration, IV; resources, IV; supervision, KS and IV; validation, KS; writing—original draft, SK; writing—review and editing, KS, UT, AT and IV. All authors contributed to the article and approved the submitted version.
Funding
This work received support from the Health Foundation's Behavioural Insights Research programme (Award 807263) and the NIHR ARC West Midlands (NIHR200165). Ivo Vlaev was further supported by the National Institute for Health Research (NIHR) through the Policy Research Unit in Behavioural Science (PR-PRU-1217-20501). The views expressed are those of the author(s) and not necessarily the funders. The funders had no role in the design of the study.
Acknowledgments
The authors would like to acknowledge all the participants who acted as a member of the working group. The authors would also like to thank the quality improvers who participated in the preference survey part of this research.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher's note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Abbreviations
CFIR: Consolidated Framework for Implementation Research; MCDA: Multi-criteria Decision Analysis; NHS: National Health Service; PAPRIKA: Potentially All Pairwise RanKings of all possible Alternatives; PSA: Probabilistic Sensitivity Analysis.
References
1. Amaratunga T, Dobranowski J. Systematic review of the application of lean and six sigma quality improvement methodologies in radiology. J Am Coll Radiol. (2016) 13(9):1088–95.e7. doi: 10.1016/j.jacr.2016.02.033
2. Clark DM, Silvester K, Knowles S. Lean management systems: creating a culture of continuous quality improvement. J Clin Pathol. (2013) 66(8):638–43. doi: 10.1136/jclinpath-2013-201553
3. Courtlandt CD, Noonan L, Feld LG. Model for improvement-part 1: a framework for health care quality. Pediatr Clin N Am. (2009) 56(4):757–78. doi: 10.1016/j.pcl.2009.06.002
4. Crowl A, Sharma A, Sorge L, Sorensen T. Accelerating quality improvement within your organization: applying the model for improvement. J Am Pharm Assoc (2003). (2015) 55(4):e364–76. doi: 10.1331/JAPhA.2015.15533
5. Niñerola A, Sánchez-Rebull M-V, Hernández-Lara A-B. Quality improvement in healthcare: six sigma systematic review. Health Policy. (2020) 124(4):438–45. doi: 10.1016/j.healthpol.2020.01.002
6. Baran-Kooiker A, Czech M, Kooiker C. Multi-criteria decision analysis (MCDA) models in health technology assessment of orphan drugs—a systematic literature review. Next steps in methodology development? Front Public Health. (2018) 6:287. doi: 10.3389/fpubh.2018.00287
7. Marsh K, IJzerman M, Thokala P, Baltussen R, Boysen M, Kaló Z, et al. Multiple criteria decision analysis for health care decision making—emerging good practices: report 2 of the ISPOR MCDA emerging good practices task force. Value Health. (2016) 19(2):125–37. doi: 10.1016/j.jval.2015.12.016
8. Thokala P, Devlin N, Marsh K, Baltussen R, Boysen M, Kalo Z, et al. Multiple criteria decision analysis for health care decision making—an introduction: report 1 of the ISPOR MCDA emerging good practices task force. Value Health. (2016) 19(1):1–13. doi: 10.1016/j.jval.2015.12.003
9. Baltussen R, Marsh K, Thokala P, Diaby V, Castro H, Cleemput I, et al. Multicriteria decision analysis to support health technology assessment agencies: benefits, limitations, and the way forward. Value Health. (2019) 22(11):1283–8. doi: 10.1016/j.jval.2019.06.014
10. Carver SJ. Integrating multi-criteria evaluation with geographical information systems. Int J Geogr Inf Syst. (1991) 5(3):321–39. doi: 10.1080/02693799108927858
11. Hallerbach WG, Spronk J. The relevance of MCDM for financial decisions. J Multi-Criteria Decis Anal. (2002) 11(4-5):187–95. doi: 10.1002/mcda.328
12. Linkov I, Moberg E. Multi-criteria decision analysis: Environmental applications and case studies. Boca Raton: CRC Press (2011).
13. Angelis A, Kanavos P. Multiple criteria decision analysis (MCDA) for evaluating new medicines in health technology assessment and beyond: the advance value framework. Soc Sci Med. (2017) 188:137–56. doi: 10.1016/j.socscimed.2017.06.024
14. Gasol M, Paco N, Guarga L, Bosch JÀ, Pontes C, Obach M. Early access to medicines: use of multicriteria decision analysis (MCDA) as a decision tool in catalonia (Spain). J Clin Med. (2022) 11(5):1353. doi: 10.3390/jcm11051353
15. NHS Business Service Authority. Prescribing Costs in Hospitals and the Community—England 2020/21 2022. Available at: https://www.nhsbsa.nhs.uk/statistical-collections/prescribing-costs-hospitals-and-community-england/prescribing-costs-hospitals-and-community-england-202021
16. Talat U, Schmidtke KA, Khanal S, Chan A, Turner A, Horne R, et al. A systematic review of nudge interventions to optimize medication prescribing. Front Pharmacol. (2022) 13:1–17. doi: 10.3389/fphar.2022.798916
17. Damschroder L, Hall C, Gillon L, Reardon C, Kelley C, Sparks J, et al., The consolidated framework for implementation research (CFIR): progress to date, tools and resources, and plans for the future. Implementation Sci. (2015) 10(Suppl 1):A12. doi: 10.1186/1748-5908-10-S1-A12
18. Chernev A, Böckenholt U, Goodman J. Choice overload: a conceptual review and meta-analysis. J Consum Psychol. (2015) 25(2):333–58. doi: 10.1016/j.jcps.2014.08.002
19. Khan I, Pintelon L, Martin H. The application of multicriteria decision analysis methods in health care: a literature review. Med Decis Making. (2022) 42(2):262–74. doi: 10.1177/0272989X211019040
20. GetResponse. 2022 Email Marketing Benchmarks by GetResponse Arkonska, Poland: GetResponse. (2022). Available at: https://www.getresponse.com/resources/reports/email-marketing-benchmarks
21. PeoplePulse. Survey response rate New South Wales, Australia: PeoplePlus (2019). Available at: https://peoplepulse.com/resources/useful-articles/survey-response-rates/
22. Hansen P, Devlin N. Multi-criteria decision analysis (MCDA) in healthcare decision-making. Oxford: Oxford Research Encyclopedia of Economics and Finance (2019).
23. Hansen P, Ombler F. A new method for scoring additive multi-attribute value models using pairwise rankings of alternatives. J Multi-Criteria Decis Anal. (2008) 15(3-4):87–107. doi: 10.1002/mcda.428
24. Briggs AH, Ades A, Price MJ. Probabilistic sensitivity analysis for decision trees with multiple branches: use of the Dirichlet distribution in a Bayesian framework. Med Decis Making. (2003) 23(4):341–50. doi: 10.1177/0272989X03255922
25. Sir E, Batur Sir GD. Evaluating treatment modalities in chronic pain treatment by the multi-criteria decision making procedure. BMC Med Inform Decis Mak. (2019) 19:1–9. doi: 10.1186/s12911-018-0723-6
26. Chaker Masmoudi H, Rhili A, Zamali I, Ben Hmid A, Ben Ahmed M, Khrouf MR. Multi-criteria decision analysis to prioritize people for COVID-19 vaccination when vaccines are in short supply. Front Health Serv. (2022) 2:18. doi: 10.3389/frhs.2022.760626
27. Sullivan T. Using MCDA (multi-criteria decision analysis) to prioritise publicly-funded health care. Dunedin: University of Otago (2012).
28. Thokala P, Madhavan G. Stakeholder involvement in multi-criteria decision analysis. Cost Eff Resour Alloc. (2018) 16(1):1–3. doi: 10.1186/s12962-018-0086-y
29. Patel B, Usherwood T, Harris M, Patel A, Panaretto K, Zwar N, et al. What drives adoption of a computerised, multifaceted quality improvement intervention for cardiovascular disease management in primary healthcare settings? A mixed methods analysis using normalisation process theory. Implement Sci. (2018) 13(1):1–15. doi: 10.1186/s13012-018-0830-x
30. Sutton RT, Pincock D, Baumgart DC, Sadowski DC, Fedorak RN, Kroeker KI. An overview of clinical decision support systems: benefits, risks, and strategies for success. NPJ Digit Med. (2020) 3(1):1–10. doi: 10.1038/s41746-020-0221-y
31. Herrera-Sánchez IM, León-Pérez JM, León-Rubio JM. Steps to ensure a successful implementation of occupational health and safety interventions at an organizational level. Front Psychol. (2017) 8:2135. doi: 10.3389/fpsyg.2017.02135
32. Skivington K, Matthews L, Simpson SA, Craig P, Baird J, Blazeby JM, et al.. A new framework for developing and evaluating complex interventions: update of medical research council guidance. Br Med J. (2021) 374:n2061. doi: 10.1136/bmj.n2061
33. Moxey A, Robertson J, Newby D, Hains I, Williamson M, Pearson S-A. Computerized clinical decision support for prescribing: provision does not guarantee uptake. J Am Med Inform Assoc. (2010) 17(1):25–33. doi: 10.1197/jamia.M3170
34. Malfait S, Van Hecke A, De Bodt G, Palsterman N, Eeckloo K. Patient and public involvement in hospital policy-making: identifying key elements for effective participation. Health Policy. (2018) 122(4):380–8. doi: 10.1016/j.healthpol.2018.02.007
35. Michie S, Van Stralen MM, West R. The behaviour change wheel: a new method for characterising and designing behaviour change interventions. Implement Sci. (2011) 6(1):1–12. doi: 10.1186/1748-5908-6-1
Keywords: multi criteria decision analysis (MCDA), quality improvment, national health service (England), optimising medicine use, decision aid
Citation: Khanal S, Schmidtke KA, Talat U, Turner AM and Vlaev I (2023) Using multi-criteria decision analysis to describe stakeholder preferences for new quality improvement initiatives that could optimise prescribing in England. Front. Health Serv. 3:1155523. doi: 10.3389/frhs.2023.1155523
Received: 31 January 2023; Accepted: 26 May 2023;
Published: 20 June 2023.
Edited by:
Richard Scheffler, University of California, Berkeley, United StatesReviewed by:
António Teixeira Rodrigues, University of Minho, PortugalAltacílio Aparecido Nunes, University of São Paulo, Brazil
© 2023 Khanal, Schmidtke, Talat, Turner and Vlaev. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Saval Khanal a2hhbmFsc2F2YWxAZ21haWwuY29t Ivo Vlaev aXZvLnZsYWV2QHdicy5hYy51aw==