ORIGINAL RESEARCH article

Front. Public Health, 31 March 2025

Sec. Public Health Policy

Volume 13 - 2025 | https://doi.org/10.3389/fpubh.2025.1562191

This article is part of the Research TopicImpact Evaluation using the Translational Science Benefits Model Framework in the National Center for Advancing Translational Science Clinical and Translational Science Award ProgramView all 5 articles

Navigating the road ahead: using concept mapping to assess Clinical and Translational Science Award (CTSA) program goals


Cathleen Kane
Cathleen Kane1*William TrochimWilliam Trochim2Haim BarHaim Bar3Andie VaughtAndie Vaught4Heather BakerHeather Baker4Munziba KhanMunziba Khan4Robin WagnerRobin Wagner4Kristi HolmesKristi Holmes5Keith HerzogKeith Herzog5Jamie Mihoko DoyleJamie Mihoko Doyle4
  • 1New York University Langone Health, Clinical and Translational Science Institute (CTSI), New York, NY, United States
  • 2The Brooks School of Public Policy, Cornell University, Ithaca, NY, United States
  • 3Department of Statistics, University of Connecticut, Storrs, CT, United States
  • 4Division of Clinical Innovation, National Center for Advancing Translational Sciences, National Institutes of Health, Bethesda, MD, United States
  • 5Northwestern University Feinberg School of Medicine, Northwestern University Clinical and Translational Sciences (NUCATS) Institute, Chicago, IL, United States

Evaluating large-scale programs designed to transform public health demands innovative approaches for navigating their complexity and scope. The Clinical and Translational Science Awards (CTSA) Program, supported by the NIH's National Center for Advancing Translational Sciences (NCATS), represents a significant national investment with over 60 sites or “hubs” spread across the country. Assessing an initiative of this size and complexity requires measures that balance local flexibility with national coherence. To that end, this study used concept mapping, a mixed-methods approach integrating qualitative brainstorming and sorting with quantitative multidimensional scaling and cluster analysis. Participation across the CTSA was unprecedented. Over 100 evaluation stakeholders were engaged across the network of hubs, leading to the identification of more than 80 measures, which were then organized into thematic clusters that reflect a logical progression from CTSA activities to outcomes and impacts, as well as critical foundational factors such as collaboration and education. The results also revealed a pattern where long-term impacts were ranked among the highest in importance but among the lowest in feasibility, particularly for measures tied to the Translational Science Benefits Model (TSBM), a new evaluation framework gaining popularity across the CTSA. The findings of this study underscore the efficacy of concept mapping in incorporating wide-ranging perspectives, identifying areas of consensus, and informing leadership in the development of unified, data-driven evaluation frameworks —such as TSBM and/or a CTSA logic model— critical to maximizing the CTSA's transformative potential for public health.

1 Introduction

It has been over 17 years since the National Institutes of Health (NIH) launched the Clinical and Translational Science Awards (CTSA) Program, an ambitious set of bold initiatives (1) and national investments aimed at improving the process of transforming laboratory, clinical, and community-based discoveries into effective public health interventions (2). The CTSA program is a nationwide network of medical research institutions, referred to as “hubs”, designed to synergize infrastructure and interdisciplinary, scientific expertise to advance clinical and translational science (CTS) research. CTSA hubs facilitate translational research through targeted pilot awards, research support services, community engagement, and multidisciplinary training. In Fiscal Year 2024, the National Center for Advancing Translational Sciences (NCATS) invested more than $629 million (3) to support more than 60 hubs.

Large biomedical research investments, such as the CTSA program, require rigorous process and outcome evaluations to determine whether the program is meeting its goals and if systematic modifications are needed over time. The Foundations for Evidence-based Policymaking Act of 2018 (also known as the Evidence Act) further underscores the need for federal agencies to build evidence in support of programs and decision-making, including the CTSA program and NCATS (4). However, the CTSA program's expansive goals, diverse institutional activities, and decentralized structure create a complex evaluation environment requiring an approach that balances local flexibility with consortium-wide coherence (5). Assessments of programs with this complexity present both practical and theoretical challenges. One practical challenge, for instance, centers on the number of CTSA institutions supported (>60) that are geographically disparate. Another challenge is building consensus among multiple evaluation stakeholders from different hubs that have a wide array of local contexts (rural vs. urban), varying financial resources, and differing roles at their CTSA (ex. CTSA Evaluators vs. Administrators).

Concept mapping is one approach to addressing these challenges by enabling stakeholders to define evaluation measures collaboratively through an asynchronous participatory platform, thereby fostering a quantifiable consensus and shared vision while respecting individual hub and institutional contexts (69). By design, this methodology is an example of participatory evaluation. This study utilized concept mapping to identify a comprehensive set of specific measures for evaluating the CTSA Program's success in meeting its goals. Input was gathered from a diverse range of perspectives and locations, spanning multiple hubs nationwide and involving over 100 key stakeholders, including CTSA Administrators, CTSA Evaluators, and NCATS staff. By engaging CTSA participants from a set of different but complimentary roles, the study sought to uncover areas of consensus or disagreement around key themes while ensuring differing perspectives were represented.

2 Materials and methods

Concept mapping is a mixed-methods approach that applies quantitative analysis to qualitative inputs. This methodology was chosen for this project as opposed to other analogous approaches such as the Delphi Method (10, 11) or Nominal Group Technique (NGT) (10, 12) because: multiple non geo-located stakeholders needed to asynchronously and collaboratively define and organize ideas; both qualitative and quantitative analysis was preferable for structuring a variety of concepts; visualization of conceptual relationships would be more useful than simple ranking; and group consensus-building was a key goal for the process overall. As Trochim describes it, concept mapping is “…a unique integration of qualitative (group process brainstorming unstructured sorting interpretation) and quantitative (multidimensional scaling hierarchical cluster analysis) methods designed to enable a group of people to articulate and depict graphically a coherent conceptual framework or model of any topic or issue of interest” [(13), p. 166]. This method has been used extensively in planning and evaluation since the 1980s (6, 9, 13, 14), and involves four essential components detailed below: Participant Selection, Data Collection, Analysis, and Interpretation.

Implementing this methodology required three waves of primary data collection and participant engagement to interpret findings. All waves of data collection involved soliciting volunteers at regularly scheduled CTSA Administrator and Evaluators meetings, and internal meetings of NCATS staff as well as sending emails directly to these targeted audiences for participation. These three groups of stakeholders—CTSA Administrators, CTSA Evaluators, and NCATS staff—were non-randomly sampled for heterogeneity. More specifically, they were also selected because of their direct and often complementary roles in designing, implementing, and utilizing evaluation data to monitor and convey the value-add and impact of CTSA-funded activities. Participation was voluntary and each participant did not need to participate in all three waves of data collection.

The first wave of data collection involved the brainstorming of measures where participants were asked to respond to a focus prompt with a data collection instrument that was created in REDCap. After providing the CTSA program goals, the data collection instrument included the following focus prompt to guide participants: “Please brainstorm as many measures as you can in response to the following prompt: ‘One specific measure I think should be used in an evaluation of the CTSA program is….” Data collection opened on February 8, 2022 and closed on March 8, 2022. A total of 320 statements were collected from participants. While the focus prompt specifically solicited measures, some participants gave statements about measures instead. Therefore, we refer to the raw data that was collected as “statements.” It is interesting to note that select non-NCATS/NIH staff were invited to participate (N = 3), but ultimately did not participate in any waves of data collection.

The next step involves unitizing the statements, a content analysis methodology (15) that is part of the concept mapping process. Trochim et al. (16) describe “unitizing” as “…the process of dividing a continuous text into smaller units that can then be analyzed” (P. 67). For example, there were instances where responses were double-barreled (e.g., “Describe the impact of CTSA funding on community health or translation into clinical practice”). These responses were then parsed out by two of the authors (CK and JD) into single idea statements (e.g., “Describe the impact of CTSA funding on community health” and “Describe the impact of CTSA funding on translation into clinical practice”). Of the 320 statements originally submitted by participants, statements were then parsed into 499 single idea statements. Two authors (CK and JD) then iteratively and inductively combined these single idea statements into 81 final statements that were used for the remainder of the process and constitute the detailed content of the mapping exercise. More formally, the authors use an inductive and independent blind coding process where similarities between statements arose from the data itself (induction) rather than having a pre-determined list of categories, bins, or statements for which each of the 499 single statements would need to be combined (deduction). The process was also “iterative” in that the 499 single idea statements were exchanged iteratively with two authors until the final list of statements was obtained. This approach, which combines iterative and inductive processes, can be described as inductive content analysis (17). A flow diagram of the brainstorming data collection and arrival of the final statement set for sorting is available in Supplementary Figure S1.

The second wave of data collection involved soliciting the same three groups—CTSA Evaluators, Administrators, and NCATS staff—to sort the 81 measures. For the sorting, participants were given a macro-enabled spreadsheet (18) and asked to assign labels of their choosing next to each statement. The only restrictions that were given to the participant were as follows: (1) spreadsheets could not be reformatted in any way, (2) each statement could only be labeled exclusively in one group, and (3) all statements could not be put into a single group. It is important to mention that while the approach used involves having participants “label” each statement, this is functionally the same as having them physically sort similar statements into piles that are then labeled (ex. “Collaborations”, “Translation Measure”, “Success Stories”, etc.). All submitted labels were later used in a subsequent wave of qualitative analysis to assign representative titles to the clusters in the concept map. Data collection for the sorting part of this activity opened on February 20, 2023, and closed on March 3, 2023. Three participants submitted their sorted statements after the due date (two NCATS staff and a participant from a CTSA hub), with the last submitted received on April 10, 2023. These late submissions were included in the analysis.

Shortly following the sorting activity, the groups were then asked to rate the 81 measures by their feasibility and importance using a REDCap form. More specifically, participants were asked to rate measures on a five-point scale according to their relative feasibility of collection and relative importance for assessing the extent to which the CTSA program is meeting its goals, where: 1 = Not Feasible/Relatively Unimportant; 2 = Somewhat Feasible/Important; 3 = Moderately Feasible/Important; 4 = Very Feasible/Important; and 5 = Extremely Feasible/Important. Participants were asked to spread out their ratings and try to use each of the five rating values at least several times.

The RCMap package in R was used to perform all analyses (19). The analysis begins with the construction from the pile-sorting information of an NxN binary symmetric matrix of similarities, Xij, k, for each sorter. For a single participant (indexed by k) Xij, k = 1 for any two statements i and j, if the two items were placed together in the same pile (category label) by the participant, otherwise a 0 is entered. The total similarity matrix is obtained by summing across all individual participants' matrices (20). Therefore, each cell in this total matrix indicates how many participants sorted the two statements together (regardless of what other statements they may have been sorted with). This total similarity matrix is the input for nonmetric multidimensional scaling (MDS) with a two-dimensional solution, which yields a two-dimensional (x,y) configuration/plot of the statements such that statements that were piled together more frequently are located closer to each other in this space while statements piled together less frequently are further apart (the “point map”). This x,y configuration is the input for hierarchical cluster analysis using Ward's method (21) which effectively partitions the x,y configuration of statements into non-overlapping clusters, called the “cluster map”. The importance and feasibility rating data are averaged across persons for each statement and cluster in a second stage of analysis described below.

Once the basic map structure is determined it is possible to construct any number of pattern match graphs [also called “ladder graphs”, and known in the field of data visualization as a parallel coordinates graph (22)] that either compare two ratings (for all participants or any subgroups) or two groups (for any rating). Groups were determined from the demographic data that was collected. A pattern match or ladder graph is a useful visual device for showing relationships and especially for highlighting the degree of relationship between the entities being displayed. A Bonferroni correction was applied to the differences in means tests reported in the ladder graphs due to the multiple hypothesis tests performed. Finally, a “Go-Zone” plot (13) was generated to visually summarize feasibility and importance measures across all raters. Quadrants for each Go-Zone were generated using overall mean ratings for feasibility and importance, respectively.

The final step of the Concept Mapping process requires engagement with representative stakeholders, which we refer to as the Interpretation Group, to respond to the general layout of the concept map and associated visualizations. This group is tasked with providing final high-level feedback on the graphic representations based on the analysis described above, as well as a review of the concept map cluster titles based on qualitative coding of label names aggregated across all sorting participants Given the hierarchical nature of the relationship between funders and grant recipients, we prioritized capturing this final wave of targeted feedback strictly from the perspective of the hubs (Administrators and Evaluators). A stratified random sample of 12 participants was taken from the list of raters, with role (CTSA Administrator vs. Evaluator) and CTSA hub size (Small, Medium, and Large according to budgeted direct costs of the hub award) as strata (6 evaluators and 6 administrators, with two from each hub size within each sub-strata). Of the 12 participants, 4 did not respond or declined to participate in the interpretation step (2 evaluators and 2 administrators from medium and large hubs). Hubs that already had a participant committed to interpreting the findings were removed from the rating list for resampling, and additional potential participants were then selected. These participants had characteristics that were the same as those who did not respond or declined. All 4 of the newly sampled participants agreed to attend the interpretation session, and only one did not attend the actual meeting (an evaluator from a large hub). NCATS staff did not participate in the interpretation session.

3 Results

Table 1 shows the descriptive statistics of participant characteristics, as well as participating hubs. During the study period, there were 61 active CTSA hubs during the first wave of data collection (i.e., Brainstorming) and 65 active hubs during the second and third waves (i.e., Sorting and Rating). For the Brainstorming stage of Concept Mapping, we had participants from 20 out of 61 hubs (33%) and 33 participants, which included 8 NCATS staff. Participation increased with subsequent waves of data collection, with 44 out of 65 hubs participating in the Sorting stage (68%), and 47 out of 65 hubs (72%) participating in the Rating stage. The number of individual participants also increased with each stage, with a 124% increase (from N = 33 to N = 74) in the number of participants from Brainstorming to Sorting and a 36% increase (from N = 74 to N = 101) from Sorting to Rating. Across all waves of data collection, nearly half of all participating hubs were small in size, which roughly reflects the proportion of total hubs in the portfolio of that size (23). The largest group of participants in the sample across all phases of data collection were CTSA Evaluators.

Table 1
www.frontiersin.org

Table 1. Descriptive statistics of participants.

Figure 1 shows the concept map (specifically, the “cluster map”), a graphic depiction of the composite thinking of all participants based on the cluster analysis, and Table 2 lists all 81 measures represented in the plot. The number of clusters (K = 8) was chosen by examining an “elbow plot” of the within sum of squares by the number of clusters (24). All eight statements in Cluster 8, located in the middle of the map as a “weak center”, were recoded to adjacent clusters based on manual review of the statements by the authors (CK and JD). Statements 27, 56, and 62 were recoded to Cluster 5 (Hub Processes and Operations); Statements 47 and 58 were recoded to Cluster 6 [Diversity, Equity, Inclusion, and Accessibility (DEIA)/Underrepresented in Research (URiR)]; Statements 17 and 55 were recoded to Cluster 4 (Mid-Term Outcomes), and Statement #33 to Cluster 1 (Long-Term Impacts).

Figure 1
www.frontiersin.org

Figure 1. A concept map of CTSA measures.

Table 2
www.frontiersin.org

Table 2. Concept mapping measures.

Given the content of the focus prompt, this map suggests a structured, comprehensive framework to discuss and assess various potential measures to be used to evaluate whether the CTSA program is meeting its stated goals. Reading from left to right and moving clockwise on the map, the clusters can be described as follows:

Hub Process and Operations: This cluster focuses on the operational aspects of CTSA hubs, including measures like time (in days) to complete consultations, the number and types of multi-center trials supported, and the quality of services.

Collaboration: This cluster examines the collaborative efforts within and between CTSA hubs, including partnerships with community members, state and local public health entities, and other CTSA hubs. It also includes measures such as tracking qualitative data on community perceptions and experiences with research.

Diversity, Equity, Inclusion, and Accessibility (DEIA)/Underrepresented in Research (URiR): This cluster includes measures around the integration of diversity and inclusion in clinical research, with examples such as the relative availability of research materials in different languages and the participation of underrepresented populations in clinical trials.

Clinical Translational Research (CTR) Training & Careers: These measures focus on training and career development. This cluster includes measures such as the number of training programs, overall mentorship satisfaction, specific career outcomes, and the retention of trainees in research.

Mid-Term Outcomes: This cluster aims to track mid-term progress such as the more proximal impact of hub-supported projects on subsequent publications or grants, bibliometric indicators or changes in promotion and tenure policies.

Long-Term Impacts: This cluster includes measures for the broad, long-term effects of CTSA activities, such as the number of Food and Drug Administration (FDA) approvals, economic and community health benefits, and/or policy or legislative impacts. It also contains measures around systematic reviews and high-level success stories. Notably, all Translational Science Benefits Model (TSBM) measures (numbers 68, 69, 70, and 78) were sorted thematically into this cluster.

Translational Process: This cluster concentrates on tracking the progression of research from the early stages of discovery to clinical application. Measures include the use of retrospective case studies, or the tracking of research across the translational spectrum (e.g., from bench to clinical trials).

Figure 2 shows a logical progression with the clusters from our map in Figure 1 “flattened” and listed in temporal order. Working clockwise from the bottom of the map in Figure 1, we begin the process in Figure 2 by first listing CTSA Activities such as Hub Processes and Operations (5) (how hubs carry out their work). We can compress the next two clusters in the map: Diversity, Equity, Inclusion, and Accessibility (DEIA)/Underrepresented in Research (URiR; 6) and CTR Training & Careers (2) into a single common phase in Figure 2 focused on key Participants, with Collaboration (3) as a critical component for all relevant participants in the Activities phases. Continuing clockwise around Figure 1 to inform the process phases in Figure 2, we can next list clusters for Outcomes Mid-term Outcomes (4) (immediately observable intermediate results) followed by Long-term Impacts (1) (final consequences or effects) and finally Translational Processes (7) (the ultimate mission of the CTSAs, to move research from discovery to application). Thus, whether viewed clockwise on the concept map (Figure 1) or as a simplified linear progression (Figure 2), our project participants logically grouped the measures related to CTSA Activities (formative evaluation) in specific logical relation to those corresponding to Outcomes and Impacts (summative evaluation). In reviewing both graphic depictions, we can see that project participants showed agreement not only in a comprehensive list of measures in general, but in a rational thematic framework for where these measures belonged relative to the evaluation of the CTSA program.

Figure 2
www.frontiersin.org

Figure 2. CTSA evaluation concept map clusters arranged as a logical process model.

Figure 3 shows the parallel coordinates plots, or “ladder graphs”, which describe the relationship between feasibility and importance for the 81 measures. Bonferroni corrected p-values indicate that the mean feasibility and mean importance for all seven clusters were statistically different except for CTR Training and Careers (p = 0.009); but the magnitude of the difference varied by cluster. On average, more than half of the seven clusters of measures were rated as appreciably more important than feasible, with lower averages overall on the feasibility side of the figure. Translational Process (7) and Long-Term Impacts measures (1) stood out as the least feasible relative to their importance. Notable exceptions were Mid-Term Outcomes (4) and Hub Processes and Operations (5), which were the only clusters rated as less important than feasible on average. Interestingly, these clusters are more closely tied to the daily activities of CTSA hubs and are frequently utilized for external reporting as well as for internal review by hub leadership.

Figure 3
www.frontiersin.org

Figure 3. Parallel coordinates plots (or “ladder graphs”), mean feasibility and mean importance.

Subgroup heterogeneity by participant type is evident when the data were disaggregated, as seen in Figure 4. Figure 4 (“ladder graphs”) stratifies average ratings for Mean Importance (left) and Mean Feasibility (right) by the three primary participant roles: CTSA Administrators, CTSA Evaluators, and NCATS staff. Here the pattern of high importance and low feasibility regarding Long-Term Impacts (1) can be seen in more relief when broken out by participant type. On the left of Figure 4, in regards to relative average importance, the steep angle of Long-Term Impacts (1) illustrates Administrators ranking this cluster in the bottom third, whereas Evaluators and NCATS staff ranked the same cluster in the top third. Mid-term Outcomes (4) and Long-Term Impact (1) represent the most pronounced, statistically different discrepancies in average importance ratings between CTSA Evaluators and Administrators. Moreover, both NCATS staff and CTSA Administrators had average importance ratings that were not statistically different from CTSA Evaluators on these two clusters. In contrast, NCATS staff rated almost all measures as more important on average than their peers in Administration and Evaluation in terms of the magnitude of their means, but were only statistically different from CTSA Evaluators on Collaboration (3). On the right of Figure 4, regarding relative average feasibility, there are more steep lines illustrating differences observed by cluster between CTSA Administrators and NCATS staff, and a pattern of more agreement between the feasibility ratings of CTSA Evaluators and NCATS staff. The mean feasibility ratings between CTSA Evaluators and NCATS staff are not statistically different across nearly all clusters of measures except for Translational Process (7) shown at the bottom of the graph. Overall, these trends in Figure 4, along with low Bonferroni-corrected p-values—show the most measurable disagreement in the feasibility ratings for CTSA Administrators vs. the other two subgroups (CTSA Evaluators and NCATS staff). The Significance Tests for Ladder Graphs by Group are shown in Table 3.

Figure 4
www.frontiersin.org

Figure 4. Feasibility vs. importance “ladder graphs” by primary participant type: CTSA administrators, CTSA evaluators, and NCATS staff.

Table 3
www.frontiersin.org

Table 3. Significance tests for ladder graphs by group.

Figure 5 visually represents all measures as single points based on their average Feasibility and Importance (“GoZone plot”). Measures with highly rated importance and feasibility can be seen as points in the upper right side of the Figure, whereas measures with relatively low importance and feasibility can be seen as points on the lower left. High Feasibility-Low Importance and Low Feasibility-High Importance measures appear in the upper left and lower right, respectively. The lower right quadrant—Low Feasibility-High Importance measures—is of particular interest with respect to the differing rating levels between Administrators, Evaluators and NCATS staff especially in regard to Long-Term Impacts (1) and TSBM. For instance, measures 68 (Number and type of Clinical and Medical Benefits), 69 (Number and type of Economic Benefits), 70 (Number and type of Community and Public Health Benefits), and 78 (Number and type of Policy and Legislative Benefits) directly reference the full scope of the TSBM, a framework gaining popularity across the CTSA consortium (25, 26). Translational Processes (7) is also well-represented in the lower-right quadrant, with measures such as data quality (Statement 60), CTSA hub-level listing of scientific and operational innovations developed, demonstrated and disseminated (Statement 37), tracking hub-supported research from one step to the next on the translational spectrum (Statement 41), the number of outcomes and innovations of CTSA supported and/or funded clinical research (Statement 74) rated, on average, as having high importance but low feasibility.

Figure 5
www.frontiersin.org

Figure 5. Feasibility vs. importance “go-zone” plot for all concept mapping measures.

4 Discussion

The purpose of this study was to conduct a comprehensive analysis of the range of measures for assessing the CTSA program's goals and to identify areas of consensus and differing perspectives. This effort resulted in three main findings. First, the concept mapping activity yielded a broad range of measures (N = 81). In terms of the overall volume of statements per themed cluster, Process and Operations had the greatest number of measures (>50% larger than the median). However, it stands to reason that in any large and complex program such as the CTSA, it is likely that causality will operate through many multiple pathways (the intentional use of multiple processes) toward a common goal (a smaller focused set of desired outcomes and impacts). Second, the clusters in our concept map corresponded with the components of a traditional logic model, illustrating the expected progression from actions to outcomes. Measures focused on CTSA activities and processes are included in the clusters represented on the left side of the map, and progress to measures associated with outcomes and impacts on right side of the map. A related finding was the TSBM measures were arrayed in a tight configuration on the far-right side of the map. This spatial placement and consolidation suggest that many participants classified the TSBM measures similarly. Third, the analyses stratified by role (in Figure 4: “ladder graphs”) showed diverging views on importance and feasibility by participant role (CTSA Evaluator vs. CTSA Administrator vs. NCATS staff), particularly regarding Long-term Impact measures, which included the TSBM (four out of nine measures in the cluster). There was also a striking and widespread consensus on the overall importance of the long-term impact measures. Evaluators and NCATS staff in particular showed marked consensus on the importance of long-term measures, However, this agreement sharply contrasted with the pronounced disagreement regarding the feasibility of implementing these measures in practice, revealing a substantial divide among key stakeholders. Interestingly, the vast majority of long-term impact measures ranked with both highly importance and low feasibility centered almost exclusively on the TSBM (in Figure 5: GoZone plot).

The patterns in these findings could be due to several factors. The discrepancies in perceived feasibility and the heavy representation of processes and operations measures could reflect functional differences in day-to-day responsibilities and scopes of work across roles. For example, Hub Processes and Operations was the cluster with the largest number of measures. Many of these measures are linked to narratives reported in annual progress reports (Measure #45: “Number and type of CTSA hub supported service consultations and services”) and continuous quality improvement activities routinely conducted at most hubs [Measure #64: Time to activation of new clinical trials supported by the CTSA hub (in days)]. These measures lie at the intersection of work that CTSA Evaluators and Administrators perform each year. However, Administrators, specifically, must prioritize proximal process measures aligned with hub operations, whereas Evaluators often find themselves balancing short-term programmatic reporting and deliverables with broader, hypothesis-driven questions NCATS staff are required to monitor program performance and across the consortium. Another consideration is that the contrasting views on feasibility by roles may reflect overall familiarity with measures and their implementation. For instance, in the concept mapping interpretation group, one CTSA Administrator expressed concern that measuring “Number and Type of Economic or Public Health Benefits” in their catchment area would be challenging. They wondered how to access economic data and grappled with the complexities of attribution vs. contribution. Meanwhile, in the same meeting, an Evaluator shared how their team already used several TSBM survey questions, collected through trainee applications, exit interviews and alumni surveys to gather high-level self-reported data. This highlighted a contrast: one side believed they needed sensitive, detailed financial measures in order to operate within the TSBM framework, while the other had already integrated straightforward self-reported surveys to capture essential data. In this short meeting, the Administrators learned that both types of data could be used within TSBM; broad economic indicators and individual success stories both fit in this flexible framework for measuring impact.

It should be noted that this concept mapping study had several limitations that must be considered when interpreting its findings. First, the concept mapping process relied on voluntary participation across the CTSA consortium, which may have introduced self-selection bias, as participants with stronger opinions or familiarity with evaluation practices may have been more likely to contribute. Second, the majority of participants were drawing on the perspective of a single hub (Evaluators and Administrators), and the total number of NCATS staff was relatively low. This means that the greater part of the feedback stemmed from a hub-specific rather than a consortium level perceptions and experiences. Third, all participants were part of the CTSA program in some manner, which may have introduced additional bias based on the preponderance of internal perspective. Fourth, while the RCMap package provided robust analytical tools for clustering and visualizing participant input, the manual reassignment of certain measures to specific clusters introduces a degree of subjectivity, potentially influencing the final cluster configurations. Fifth, given the rapidly evolving nature of translational science and the specific goals outlined in the NCATS Strategic Plan (27), the measures identified here may require regular updates to remain aligned with emerging priorities, technological advancements, and evolving program goals. Finally, in the specific context for this discussion, it is also important to note that concept mapping is a tool for illustrating the composite thinking of a diverse group at a single point in time, rather than a means for providing incontrovertible or static answers. For example, the feasibility and importance ratings illustrated in the Go-Zone charts and ladder graphs are based on subjective assessments at a single point in time, which may be influenced by respondents' individual experiences, familiarity with particular measures, and role-specific priorities.

Nevertheless, these findings suggest that there is no shortage of available measures to assess the CTSA programmatic goals, but there may be a lack of consensus on how to implement them effectively and efficiently. This opens opportunities for future work. These concept mapping results could support multiple complimentary frameworks such as a consortium-wide logic model and the TSBM. While individual CTSAs may have developed logic models to address local needs and individual grant aims (28), a logic model for the consortium has not yet been developed. Using the results of this concept map as a foundation for this model would have the benefit of being a “bottom up” and data-driven exercise representing the thinking of the wide range of participants as opposed to a “top down” exercise with authorship and buy-in limited to a minority of stakeholders. Simultaneously, these same findings highlight that while TSBM measures are currently recognized as highly important, there are significant challenges around perceptions on feasibility. This provides a focused starting point for developing strategies to address and overcome these barriers to evaluation implementation.

This project also revealed practical opportunities for NCATS to provide strategic leadership by integrating the interdisciplinary insights of Evaluators and Administrators. The concept mapping process and results from this analysis create a starting point for future collaborative evaluation activities centered on assessing the CTSA program and its progress toward achieving its goals. Just as there are many roads to Rome, there are many ways to support translation in clinical research. As reflected in the concept map, on the activities side of the logical progression we have numerous interventions and collaborations to support clinical translational research. By the time we get to the outcomes side of the logical progression we are essentially listing impact measures that revisit the central mission of the CTSA program: To increase the pace of development and availability of treatments; to enable more individuals and communities to contribute to and benefit from translational science; and to identify and address inefficiencies in translation that slow and even stop research efforts (27). To fully leverage the strengths of CTSA Evaluators, Administrators, and NCATS staff, it is essential to embrace their distinct roles and responsibilities. Administrators focus on monitoring their own hub's operations, NCATS oversee consortium-wide outcomes and impacts, and Evaluators bridge these perspectives, balancing program-level reporting with broader questions of long-term effects. These differences are not limitations, but integral features of the system's structure, presenting opportunities for collaboration to enhance the full breadth of evaluation of the CTSA program.

Ultimately, the most difficult and pressing work will not lie in the selection of measures, but in driving coordinated CTSA evaluation across the consortium. Frameworks like concept mapping, logic modeling, and TSBM offer concrete signposts on the “many roads to Rome”; but their utility in this navigation depends on coordinated direction. Of all three roles represented in this study, only NCATS has the unique perspective and operational authority to endorse a unified CTSA logic model associated with a specific set of impact measures. They are also the only contributors with the level of access and critical resources necessary to collect and analyze aggregated data for a program of this complexity, scale and importance. By using data from these findings to guide their ongoing efforts, NCATS can strengthen its ability to assess whether the CTSA is meeting its goals and demonstrate the program's broader value. If we are to overcome the roadblocks on the path to evaluation, there is an opportunity ahead to harness and align the unique perspectives and strengths of CTSA Evaluators, Administrators and NCATS consortium leadership. By setting a course centered around a shared vision for the way forward, these frameworks can guide us in the effective evaluation of the long-term impact of the CTSA Program.

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Ethics statement

Ethical review and approval were not required for the study on human participants in accordance with the local legislation and institutional requirements. Written informed consent from the (patients/participants or patients/participants legal guardian/next of kin) was not required to participate in this study in accordance with the national legislation and the institutional requirements. The research was conducted in accordance with ethical principles for integrity, transparency, and responsible scholarship.

Author contributions

CK: Conceptualization, Formal analysis, Methodology, Project administration, Visualization, Writing – original draft, Investigation, Supervision, Resources. WT: Writing – review & editing, Methodology. HBar: Data curation, Formal analysis, Methodology, Resources, Software, Validation, Visualization, Writing – review & editing. AV: Writing – review & editing. HBak: Writing – review & editing. MK: Writing – review & editing. RW: Formal analysis, Writing – review & editing. KHo: Data curation, Methodology, Resources, Writing – review & editing. KHe: Data curation, Methodology, Resources, Writing – review & editing. JD: Conceptualization, Data curation, Formal analysis, Methodology, Project administration, Resources, Software, Supervision, Validation, Visualization, Writing – original draft.

Funding

The author(s) declare that financial support was received for the research and/or publication of this article. The authors were provided partial salary support through the following National Institutes of Health (NIH) grants: UL1TR001445 (Cathleen Kane) and UM1TR005121 (Keith Herzog and Kristi Holmes). The content is solely the responsibility of the authors and does not necessarily represent the official views of the contributors' institutions, NCATS or NIH.

Acknowledgments

The authors wish to extend a special thank you for the stakeholder feedback provided by: Zainab Abedin (Irving Institute for Clinical and Translational Research); Jennifer Croker (UAB Center for Clinical and Translational Science); Rosalina Das (Miami Clinical and Translational Science Institute); Lauren Herget (NYU Langone Clinical and Translational Science Institute); Matt Honore (Oregon Clinical and Translational Research Institute); Verónica Hoyo (Northwestern University Clinical and Translational Sciences); Ashley Kapron (Utah Clinical and Translational Science Institute); Maggie Padek Kalman (Frontiers Clinical and Translational Science Institute); Gerald Moose Stacy (Institute for Translational Medicine); Cinthia Sanchez (Duke Clinical and Translational Science Institute); Shannon Swiatkowski (Clinical and Translational Science Collaborative of Northern Ohio).

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declare that no Gen AI was used in the creation of this manuscript.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fpubh.2025.1562191/full#supplementary-material

References

1. Zerhouni E. Medicine. NIH Roadmap Sci. (2003) 302:63–72. doi: 10.1126/science.1091867

PubMed Abstract | Crossref Full Text | Google Scholar

2. Zerhouni EA. Clinical research at a crossroads: the NIH roadmap. J Investig Med. (2006) 54:171–3. doi: 10.2310/6650.2006.X0016

PubMed Abstract | Crossref Full Text | Google Scholar

3. National Center for Advancing Translational Sciences (2024). Budget Overview. Available online at: https://ncats.nih.gov/about/budget (accessed January 16, 2025).

Google Scholar

4. T HCongress (2019). Foundations for Evidence-Based Policymaking Act of 2018. Available online at: https://www.congress.gov/115/plaws/publ435/PLAW-115publ435.pdf (accessed October 10, 2024).

Google Scholar

5. Trochim WM, Marcus SE, Mâsse LC, Moser RP, Weld PC. The evaluation of large research initiatives: a participatory integrative mixed-methods approach. Am J Eval. (2008) 29:8–28. doi: 10.1177/1098214007309280

Crossref Full Text | Google Scholar

6. Trochim WMK. An introduction to concept mapping for planning and evaluation. Eval Program Plann. (1989) 12:1–16. doi: 10.1016/0149-7189(89)90016-5

Crossref Full Text | Google Scholar

7. Trochim WM, Cabrera DA, Milstein B, Gallagher RS, Leischow SJ. Practical challenges of systems thinking and modeling in public health. Am J Public Health. (2006) 96:538–46. doi: 10.2105/AJPH.2005.066001

PubMed Abstract | Crossref Full Text | Google Scholar

8. Trochim WM, Kane M. Using Concept Mapping in Evaluation. Concept Mapping for Planning and Evaluation. Thousand Oaks, CA: Sage Publications (2011).

Google Scholar

9. Kane M, Trochim WM. Concept Mapping for Planning and Evaluation. Thousand Oaks, CA: Sage Publications, Inc. (2007).

Google Scholar

10. Mcmillan SS, King M, Tully MP. How to use the nominal group and Delphi techniques. Int J Clin Pharm. (2016) 38:655–62. doi: 10.1007/s11096-016-0257-x

PubMed Abstract | Crossref Full Text | Google Scholar

11. Sablatzky T. The Delphi method. Hypoth Res J Health Inform Prof. (2022) 34:26224. doi: 10.18060/26224

Crossref Full Text | Google Scholar

12. Gaskin S. A guide to nominal group technique (NGT) in focus-group research. J Geogr High Educ. (2003) 27:342–7. doi: 10.1080/03098265.2003.12288745

Crossref Full Text | Google Scholar

13. Trochim WM, Mclinden D. Introduction to a special issue on concept mapping. Eval Program Plann. (2017) 60:166–75. doi: 10.1016/j.evalprogplan.2016.10.006

PubMed Abstract | Crossref Full Text | Google Scholar

14. Kagan JM, Kane M, Quinlan KM, Rosas S, Trochim WM. Developing a conceptual framework for an evaluation system for the NIAID HIV/AIDS clinical trials networks. Health Res Policy Syst. (2009) 7:12. doi: 10.1186/1478-4505-7-12

PubMed Abstract | Crossref Full Text | Google Scholar

15. Krippendorf K. Content Analysis: An Introduction to its Methodology. Thousand Oaks: CA, Sage (2013).

Google Scholar

16. Trochim WM, Donnelly JP, Arora K. Research Methods: The Essential Knowledge Base. Boston, MA: Cengage Learning (2016).

Google Scholar

17. Vears D, Gillam L. Inductive content analysis: a guide for beginning qualitative researchers. Focus Health Prof Educ. (2022) 23:111–27. doi: 10.11157/fohpe.v23i1.544

Crossref Full Text | Google Scholar

18. Bar H. RCMap. (2022). Available online at: https://github.com/haimbar/RCMap/blob/main/pilesortingExample.xlsm (accessed December 26, 2024).

Google Scholar

19. Bar H. Group Concept Mapping in R. (2024). Available online at: https://haimbar.github.io/RCMap/ (accessed December 10, 2024).

Google Scholar

20. Weller SC, Romney AK. Systematic Data Collection. Thousand Oaks, CA: Sage Publications (1988).

Google Scholar

21. Ward JH. Hierarchical grouping to optimize an objective function. J Am Stat Assoc. (1963) 58:236–44. doi: 10.1080/01621459.1963.10500845

Crossref Full Text | Google Scholar

22. Inselberg A. Parallel Coordinates: Visual Multidimensional Geometry and Its Applications. New York, NY: Springer (2009).

Google Scholar

23. National Institutes of Health. NIH RePORTER. (2024). Available online at: https://reporter.nih.gov/ (accessed December 20, 2024).

Google Scholar

24. Everitt BS, Landau S, Leese M. Cluster Analysis. Hodder Education Publishers (2001).

Google Scholar

25. Luke DA, Sarli CC, Suiter AM, Carothers BJ, Combs TB, Allen JL, et al. The translational science benefits model: a new framework for assessing the health and societal benefits of clinical and translational sciences. Clin Transl Sci. (2018) 11:77–84. doi: 10.1111/cts.12495

PubMed Abstract | Crossref Full Text | Google Scholar

26. Andersen S, Wilson A, Combs T, Brossart L, Heidbreder J, Mccrary S, et al. The Translational Science Benefits Model, a new training tool for demonstrating implementation science impact: a pilot study. J Clin Transl Sci. (2024) 8:e142. doi: 10.1017/cts.2024.643

PubMed Abstract | Crossref Full Text | Google Scholar

27. National Center for Advancing Translational Sciences. Strategic Plan 2025–2030. (2024). Available online at: https://ncats.nih.gov/about/ncats-overview/strategic-plan (accessed December 19, 2024).

Google Scholar

28. Fishman AY, Lounsbury DW, Lechuga C, Patena J, Marantz P, Kim M, et al. Moving from prove to improve: a collaborative continuous quality improvement process for advancing Clinical and Translational Science. J Clin Transl Sci. (2024) 8:e103. doi: 10.1017/cts.2024.555

PubMed Abstract | Crossref Full Text | Google Scholar

Keywords: Clinical and Translational Science Awards (CTSA), concept mapping, evaluation study, stakeholder participation, mixed-methods research, Translational Science Benefits Model (TSBM), longitudinal impact, National Institutes of Health (NIH)

Citation: Kane C, Trochim W, Bar H, Vaught A, Baker H, Khan M, Wagner R, Holmes K, Herzog K and Doyle JM (2025) Navigating the road ahead: using concept mapping to assess Clinical and Translational Science Award (CTSA) program goals. Front. Public Health 13:1562191. doi: 10.3389/fpubh.2025.1562191

Received: 17 January 2025; Accepted: 10 March 2025;
Published: 31 March 2025.

Edited by:

Douglas Luke, Washington University in St. Louis, United States

Reviewed by:

Nasser Sharareh, The University of Utah, United States
Emmanuel Tetteh, Washington University in St. Louis, United States

Copyright © 2025 Kane, Trochim, Bar, Vaught, Baker, Khan, Wagner, Holmes, Herzog and Doyle. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Cathleen Kane, Y2F0aGxlZW4ua2FuZUBueXVsYW5nb25lLm9yZw==

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.