Skip to main content

ORIGINAL RESEARCH article

Front. Vet. Sci., 24 October 2022
Sec. Veterinary Humanities and Social Sciences

Longitudinal assessment of competency development at The Ohio State University using the competency-based veterinary education (CBVE) model

\nEmma K. Read
Emma K. Read1*Connor MaxeyConnor Maxey2Kent G. Hecker,Kent G. Hecker2,3
  • 1College of Veterinary Medicine, The Ohio State University, Columbus, OH, United States
  • 2Faculty of Veterinary Medicine, University of Calgary, Calgary, AB, Canada
  • 3International Council for Veterinary Assessment, Bismarck, ND, United States

With the development of the American Association of Veterinary Medical Colleges' Competency-Based Veterinary Education (CBVE) model, veterinary schools are reorganizing curricula and assessment guidelines, especially within the clinical rotation training elements. Specifically, programs are utilizing both competencies and entrustable professional activities (EPAs) as opportunities for gathering information about student development within and across clinical rotations. However, what evidence exists that use of the central tenets of the CBVE model (competency framework, milestones and EPAs) improves our assessment practices and captures reliable and valid data to track competency development of students as they progress through their clinical year? Here, we report on validity evidence to support the use of scores from in-training evaluation report forms (ITERs) and workplace-based assessments of EPAs to evaluate competency progression within and across domains described in the CBVE, during the final year clinical training period of The Ohio State University's College of Veterinary Medicine (OSU-CVM) program. The ITER, used at the conclusion of each rotation, was modified to include the CBVE competencies that were assessed by identifying the stage of student development on a series of descriptive milestones (from pre-novice to competent). Workplace based assessments containing entrustment scales were used to assess EPAs from the CBVE model within each clinical rotation. Competency progression and entrustment scores were evaluated on each of the 31 rotations offered and high-stakes decisions regarding student performance were determined by a collective review of all the ITERs and EPAs recorded for each learner across each semester and the entire year. Results from the class of 2021, collected on approximately 190 students from 31 rotations, are reported with more than 55 299 total competency assessments combined with milestone placement and 2799 complete EPAs. Approximately 10% of the class was identified for remediation and received additional coaching support. Data collected longitudinally through the ITER on milestones provides initial validity evidence to support using the scores in higher stakes contexts such as identifying students for remediation and for determining whether students have met the necessary requirements to successfully complete the program. Data collected on entrustment scores did not, however, support such decision making. Implications are discussed.

Introduction

Veterinary medicine recently began adoption of competency-based education (CBE) and this has been formalized by introduction of three components of the competency-based veterinary education (CBVE) model: CBVE Competency Framework, Entrustable Professional Activities and Milestones (14). CBE is widely employed across health sciences education with emphasis on a learner-centered approach and recurring assessments (5). CBE has moved beyond evaluation of learning for the knowledge domain and is inclusive of psychomotor skills, attitudes and values (5). Since single assessments do not adequately capture all these dimensions at once, longitudinal assessment using multiple assessment methods has become a necessity to gather the complete picture of learning or competency across the domains (6).

Longitudinal assessment utilizes a data continuum to provide information about an individual's learning and includes a mixture of low-stakes opportunities and high-stakes decisions (7). What is presented in this study are the preliminary building blocks for a programmatic approach (8) where assessment data are collected longitudinally and periodically reviewed by an oversight clinical educators' committee that renders high-stakes progress decisions regarding learners (5). Here, we report on the use of two assessment methods that incorporate scales which have been reported to result in more reliable and valid scores over time (9). The intent of using these data are to provide students and stakeholders with feedback regarding competency progression through clinical rotations. Before fully committing to an oversight committee, sometimes referred to as a competence committee (5) and a more detailed programmatic assessment approach, here we first wanted to ensure that data can be collected according to a competency profile and entrustable activities, where data regarding milestones and entrustment scales result in reliable scores to assist with identifying students earlier for remediation and demonstrate clear evidence of learner growth over time.

So why the push to adopt CBE or CBVE at all? Increasingly educators are finding value in the demonstration of “assessment as learning” with frequent specific feedback providing more meaning to learners and helping to better guide their development (10).

Convincing evidence for student learning in all domains of competence is being increasingly demanded by educators, accrediting bodies, and stakeholders alike (10). In medical education, a recent study compared competencies measured using Likert type scales to those using entrustment scales (11) and national studies are being conducted to compare milestone ratings in resident training programs with national licensing examination scoring (12). Veterinary specific examples are more limited to date however there has been demonstration of strong validity evidence for a programmatic assessment approach (13). This veterinary study used VetPro (14) as the competency framework with the item scales on the workplace-based assessment methods ranging from 1 (novice) to 5 (expert).

EPAs are described as the daily activities of a practicing veterinarian that can be used as opportunities to observe performance and offer feedback using workplace-based assessment methods (15). To clarify further, competencies describe the veterinarian, while EPAs describe what the veterinarian does (16). Pilot work at the University of Calgary identified an entrustment scale that proved useful with assessing students in general practice veterinary settings in a distributed veterinary teaching hospital and this was used as the basis of establishing a scale for use at OSU-CVM which featured 5 anchor points: Don't trust to perform any aspect of the task, Trust but needs lots of help, Trust but needs some help, Trust but needs on-demand guidance, Trust to perform on own (17). OSU-CVM used EPAs from the CBVE model with EPA 1 separated into 3 individual EPAs, so 10 EPAs were made available for learners to select from.

Here, we report both the use of the CBVE milestone scale within an In-training Evaluation Report (ITER), and the use of scores from entrustment scales for evaluation of CBVE EPA performances. The work here with its underlining assumptions builds our validity argument (18, 19) that scores from these methods will provide reliable and valid evidence to support progression and remediation decisions for final year veterinary students. To affirm this, scores would need to be assessed for their reliability (consistency of scores within and across method overtime) and their validity (changes in scores over time to demonstrate learning progression), and scores would also need to demonstrate a consistent positive correlation between milestone scores and entrustment scores. Such findings could then be used to produce preliminary evidence for scoring and generalization within Kane's argument-based validity framework (18, 19).

With this as background, the purpose of this research is to report on the validity evidence of scores from both in the moment assessments (of EPAs) and clinical rotation milestone scores on competency development, within and across domains, within and across time, over the clinical component of a veterinary program.

Methods

Context

This study was provided exemption from IRB review by The Ohio State University Office of Responsible Research Practices. OSU-CVM has a final year program that includes 31 possible rotations offered across five hospitals in the Veterinary Health System. Each final year student is enrolled in three identical courses that are graded “S” (satisfactory) or “U” (unsatisfactory). Each course spans a single semester and the final year program is comprised of three semesters, that include 8 rotations each. The rotations selected for each student will vary according to the career area of emphasis they select: small animal, farm animal, mixed animal, equine, spectrum of care, and individualized (research, public health). Each student has a total of 24 rotations chosen from a total of 31 available. Beginning with the class of 2021, changes were made to the final year assessments to move from a single summative judgement at the conclusion of each rotation to a suite of workplace-based assessment methods that emphasize student development over time. OSU-CVM introduced a new ITER based on milestone progression, as well as EPAs and workplace-based assessment with corresponding entrustment scales for timely direct observation with formative feedback. The college also required student reflection on their longitudinal competency development across rotations at the end of each semester and created opportunities for more faculty coaching input (not reported here). The 2020–21 final year program was 25 blocks in length (including 2 weeks vacation) with the first 5 blocks being conducted on-line because of the global COVID-19 pandemic. In the first 5 blocks all students were required to take an orientation, followed by rotations in diagnostic imaging, applied pathology, clinical pathology, and preventive medicine. The remaining 20 in-person 2-week long rotations were offered across the veterinary health system. Data are shown from Block 6 onwards.

ITER based on milestone progression within competencies

A modified ITER was created with adoption of the CBVE's 32 competencies and associated four milestones (novice, advanced beginner, competent and proficient). A fifth milestone (pre-novice) was added following consultation of OSU-CVM faculty and was defined for all competencies as “not yet meeting novice level”. This was introduced because faculty members had expressed concerns that some students may arrive into the clinical training portion of the program without yet being able to meet the novice milestone in all competencies. Faculty did not wish to cause students to lose confidence in their training to date should this occur, and they also wanted to encourage learners to keep progressing toward higher level milestones. Brief video training materials were created to orient learners and raters to the assessment process and to encourage standardized use of the tools across rotations. As part of the development, each of the rotation coordinators were asked to identify the competencies from the CBVE framework that their respective rotation could assess in each learner that attended their rotation. The number of observations of each competency are reported in the results (Table 1). The premise at the time was that the sum of the assessment pieces from each rotation would provide a complete picture of overall development as the year progressed and as the learner navigated the 20 clinical rotation blocks that made up the remaining final year program.

TABLE 1
www.frontiersin.org

Table 1. Summary of the observational data: number of observations made per competency, number of students assessed per competency, number of raters providing feedback, mean number of assessments provided per student for each competency, and mean number of ratings provided per rater for each competency.

Any clinician or preceptor within a rotation who worked with a student was able to complete an ITER form and for each rotation all the forms were averaged together to create a ‘total ITER form’ that assigned each student their mean final milestone rating for each competency assessed. Placement of student performance on the CBVE milestone anchors involved the use of a ‘bubble bar’ visual analog scale that incorporated a 25-point scale with equal intervals between milestones. A description of each milestone anchor was provided for each different competency as per the CBVE model (4). This provided the rater flexibility in reporting their observation of student's ability within competency. See Figure 1— The ITER form. Clinicians were asked to provide qualitative comments specific to what went well and what areas needed to be addressed by the learner for further development in the competency area. Clinicians also had the opportunity to “flag” to the Associate Dean for Professional Programs if the student's performance prompted concerns as it related to: clinical skills, communications, mental wellbeing, medical knowledge, clinical reasoning and critical thinking, interprofessional skills, professionalism, time management, and honor code violation. Free text comments were also encouraged to confidentially report the clinicians' observations to the Associate Dean so that a remediation program could be developed to directly address the deficiencies if required.

FIGURE 1
www.frontiersin.org

Figure 1. The Ohio State University College of Veterinary Medicine final year rotation ITER (in-training evaluation report) form showing the use of a “bubble bar” for continuous scoring across milestone anchors.

The “total ITER form” and qualitative comments was ultimately reviewed by each rotation leader and a “global total ITER” score of “S”—satisfactory, “B”—borderline, or “U”—unsatisfactory was recorded. Any discrepancies between milestone scoring, qualitative comments or “flags” were checked by the Associate Dean in consultation with the rotation leader to ensure appropriate interpretation.

A final rotation grade of “S” (satisfactory) or “U” (unsatisfactory) for the rotation was assigned later following use of a borderline regression analysis that plotted mean overall score (from across all domains and competencies scored on the 25-point scale) vs. the “global total ITER” score (S, B or U). The final rotation grade was therefore not determined by the observing clinical faculty but was instead based on review of the borderline regression analysis results, the qualitative comments, and EPAs. The “global total ITER” score assigned by rotation leaders assisted in the analysis but was not the final score for the rotation.

Data analyses

Data from the 2020–21 academic year were analyzed using SPSS v.28 (IBM, Chicago, IL, USA). Generalizability theory, where student was crossed with competency nested within clinical rotation was used to calculate reliability coefficients for the ITER. To assess progression of performance across clinical rotations, composite learning curves were built to determine if milestone scores were different across students, if scores changed over time, and if we can differentiate between competencies. Given that repeated milestone scores (level 1) were nested within CBVE competencies (level 2) nested within student (level 3), a multilevel random coefficient model was used to model variance components for each of these levels, similar to Bok et al. (13). Within student (level 3), clinical rotation block was included to assess whether students begin at the same level (intercept variance) and progressed at the same rate (slope variance) over time. Finally, as reported by Bok et al. (13), these data may change due to modeling linear and non-linear time, therefore data were further analyzed with linear, quadratic and cubic functions and significant changes were assessed using the−2log likelihood chi-square analyses.

Results

Every OSU-CVM rotation (except for preventive medicine) identified the same 5 competencies in common −1.1, 1.2, 1.3, 5.1, and 7.2, however not all the CBVE competencies were assessed on every rotation. All of the competencies were assessed for each student multiple times over the year as they progressed through their rotations. See Table 1 for further details on the number of observations of each competency. See Table 1 and three for Generalizability and multilevel model analyses report for the 5 competencies mentioned above.

During 2020-21, across the final year clinical program there were 55,299 total competency assessments using ITERs and 2,799 complete EPAs collected on 190 students.

Figure 2 shows the visual representation of the average competency growth for the whole cohort across all of the domains and competencies (collapsed together) for the VME IV clinical program in 2020–21. Figure 3 shows the visual representation of the average competency growth for the whole cohort across each of the 9 CBVE domains and 32 competencies for the final year clinical program in 2020–21 according to data collected from these ITERs.

FIGURE 2
www.frontiersin.org

Figure 2. Mean competency scores over time for entire cohort collapsed across all CBVE domains.

FIGURE 3
www.frontiersin.org

Figure 3. Mean competency scores over time for entire cohort for each individual CBVE domain.

Table 1 shows the observational data collected from the ITERs: number of observations per competency, number of students assessed per competency, number of raters contributing to each competency assessment, as well as the mean number of assessments per student for a particular competency, and the mean number of ratings provided by each clinician for each domain.

Table 2 reports the reliability (G-coef) as 0.85. This suggests milestone ratings of students were consistent across competencies and blocks. Greater variance was accounted for student within block (43.16%; indicating scores discriminate high and low performing students consistently) compared to competency within block (1.67%; little discrimination by competency score) and by week (17.50%; scores change over weeks).

TABLE 2
www.frontiersin.org

Table 2. Generalizability Analysis for rotation block (n = 24), student nested within block (n = 191) and competency nested within block (n = 5 competencies).

The random coefficients model (Table 3) nests repeated scores within competency domain within student with time fixed (Model 1) and time modeled as a linear function (Model 2) as suggested in Bok et al. (13). In the first model (Model 1) students' intercepts account for the greatest amount of variance (79.66%) indicating students start at different milestone ratings. The slope indicates that rate of student learning differs negligibly between student (student progress at the same milestone rating over time). The covariance term in level 3 of the model was negative indicating higher initial milestone scores have a less positive change over time. Interestingly there is no variance between competency, suggesting milestone ratings among competencies vary little within rotation. Finally, the residual includes the variance due to repeated measures nested within competency and student (20.06%).

TABLE 3
www.frontiersin.org

Table 3. Multilevel random coefficients models, where repeated measures are nested within competency and students.

Model 2 assumes changes in milestone ratings are constant (linear) over time. There is an improved model fit (χ2(1) = 1,20,065.83–1,19,752.72 = 331.10, p < 0.01). The value of 1.53 is the expected change in student milestone ratings over one block. Given that there are 19 total blocks (clinical rotations) this represents an average student increase of 29.07 indicating a significant change in milestone rating. There is a difference in student variance between Model 1 and Model 2 due to change over time (794.05–164.66 = 629.39). The general trends that were present in Model 1 are reflected in Model 2, the majority of the variance was due to different student intercepts (different milestone ratings; 45.03%), the rate of change does not change between students (slope variance in minimal) and there is again no variance between milestone competency assessment. When we modeled time as a quadratic function (one inflection point) there were no differences in the amount of variance that the model accounts for (χ2(1) = 1,19,752.72–1,19,755.19 =-2.44, p = n.s.), therefore we did not pursue further data analysis modeling time as non-linear.

Table 4 shows the number of students choosing to complete various EPAs. Using entrustment scales as the assessment tool, performance within the various EPAs was shown not to vary over time. Figure 4 shows entrustment scores for EPAs across rotation blocks and specifically, for the most assessed EPA “Performs a Physical Exam.”

TABLE 4
www.frontiersin.org

Table 4. Number of students selecting to complete each EPA.

FIGURE 4
www.frontiersin.org

Figure 4. Box plots presenting categorical entrustment scores on EPAs across clinical rotation blocks. (A) Combined entrustment scores across all EPAs by rotation block. (B) Entrustment scores for the “Performs a Physical Exam” EPA, the EPA most assessed within and across rotation blocks.

Discussion

The main findings of this research are: Milestone scores from ITERs demonstrate adequate reliability (G coef = 0.85). The multilevel model indicates that students begin the final year clinical program ranked at different milestone markers, their performance increases at relatively the same rate as they progress through their clinical rotations, yet faculty assessment between core competencies was not distinguishable. Interestingly there was little variability in entrustment scale scores for the EPAs assessed across rotations.

The unique contributions to veterinary medical education assessment here are two-fold. First, there is an uptake and use of items such as milestones where preceptors/faculty appear to assess student performance consistently within and across rotations and use the scores for student feedback and development. Also there is the implementation and use of EPAs and entrustment scores. However, there was little to no differentiation in entrustment scores of students. There are a number of potential factors why faculty and students may not have used the EPA framework and entrustment scales as intended which include: assessor understanding of the EPAs and their use, assessor understanding of the entrustment scales and their use, and the broad range of assessors completing entrustment scales (including faculty, technical staff, residents and interns). The latter population is different to those completing ITERs (faculty and residents only) who represent a group with more education and assessment experience collectively and by comparison. In addition to differences in how the tools are completed there may be differences in how they are being interpreted by students and used by administration. Entrustment scales were used here as low stakes assessment tools which may have diminished their perceived value for assessors and resulted in less discriminating selection between anchor points with a focus on being just good enough. In other veterinary programs entrustment scale tools and EPAs have been reported for use in a high-stakes capacity (13, 20).

The second unique contribution of this work is demonstration of learner progression over time, supported by our analysis. These findings advance use of the CBVE model components for effective scoring of learners in veterinary medicine.

ITERs: Competencies and milestones

Despite individual students who occasionally rated as pre-novice, there was no rating of the entire cohort below novice for any individual competency or domain, and the cohort achieved competent for all competencies by the end of the training period (Figures 2, 3). Figure 3, showing milestone by domain, could be reflective of OSU's pre-clinical training and how this contributes to their entry-level performance upon arrival into the clinical workplace (21). For example, in domain 1 (individual animal care and management) the initial assessments (as represented near the y-intercept) falls at the novice milestone for performance, yet in domain 5 (communication) the initial assessments (as represented near the y-intercept) falls at the advanced beginner milestone. These milestone ratings may reflect the robust communication training program that has been in place for almost a decade, while the clinical skills program is a more recent addition to the program. These examples highlight that while the multi-level model did not show variance due to competency there are some interesting findings when the Domain scores are considered. Domains 3 (animal population care and management) and 8 (financial and practice management) were more difficult to interpret due to fewer specific data points and because they demonstrated less readily identifiable learning curves compared to the other domains. These particular domains were less frequently identified by rotation leaders as ones that could be reliably assessed in every student on every rotation resulting in fewer data collected. This reinforces that thorough assessment of student progression in all areas requires evaluation of multiple competencies across the entire program and should likely span more than just the clinical training period (6). Fortunately, domain 3 (animal population care and management) is heavily assessed in pre-clinical core courses and competency 8.1 (economic factors in personal and business decision making) is a significant component of the professional development courses.

The ITERs allowed identification of approximately 10% of students per semester that could benefit from remediation. Students participating in remediation activities are required to have formal reporting from supervising instructors and they have all shown improvement in order to progress. Any student required to repeat a rotation is assessed using the same ITER form and to date, all students have shown improvement in their scoring or qualitative comments on the repeated experience. Repeat rotations are typically undertaken with different rotation instructors wherever possible to avoid conflict of interest with grading.

A recent paper described that validity evidence for an assessment system can be supported with demonstration that growth in performance over time follows a theoretically predictable pattern (such as learning curves) (9). Figure 3 shows such patterns for all domains of competence that were assessed in learners at OSU-CVM using the CBVE.

Entrustment scales and EPAs

Studies in medicine have shown the importance of using methods that focus on both the granular and the holistic in programs of assessment, with emphasis being placed on EPAs to provide narrative value (22). EPAs were introduced at OSU-CVM to create opportunities to provide direct observation of performance and facilitate provision of formative feedback using an entrustment scale tool. All students must complete one EPA per rotation block as a program requirement and the student initiates the process by selecting an EPA and asking to be observed performing it. The most commonly selected EPA was 1b “perform a physical exam” (35%), followed by “develop a diagnostic plan and interpret results” (13%), “develop and implement a management/treatment plan” (13%), and “gather a history” (11%). It is likely that these were the most commonly selected EPAs because students already had some comfort and familiarity with these activities and thought that they may have greater success in performance “under pressure.” It may also be that these activities are simply those that a student performs multiple times per day repetitively on each block and it was instead a matter of convenience.

The least commonly selected EPAs were recognizing a patient needing urgent or emergent care (2%), formulate relevant questions and retrieve evidence to advance care (1%), or formulate recommendations for preventive healthcare (2%). The avoidance of selecting the EPA for urgent or emergent care is likely because emergency rotations tend to be fast paced, high-stress learning environments leaving little time to ask questions (23). This makes it difficult to ask a clinical assessor to interrupt an urgent situation to allow a student to take on the lead role. Alternatively, it may also be an avoidance on the part of the student to avoid making higher stakes decisions when time is tight and the situation is urgent. Further investigation is required to determine why student preferences occur as they do. Recall, OSU students are allowed to freely select any EPA they wish to be evaluated on for their single mandated EPA per block. A recent discussion amongst the final year teachers affirmed that the most frequently selected EPAs tend to be amongst some of the most important to a new graduate and therefore rotation leaders were content with students having multiple measures of assessment on the same activities repeatedly and consider this valuable repeated practice over time.

Figure 4 shows combined mean entrustment scores assigned for EPAs across all rotations and the entrustment scores assigned for the “performs a physical exam” EPA. The box plots show that there is relatively little variation in scoring in either the overall cohort or individual student consideration. As mentioned above, the reasons for this lack of variability could be the focus of further research and will warrant further consideration of how best to incorporate entrustment scales into veterinary medical assessment. Other health profession fields report that frontline faculty are struggling with entrustment-anchored scales, and while that may not be the case in veterinary medicine, it's clear that we have much to learn about their implementation (24).

Conclusion

The development of ITERs based on milestones and competences and Entrustment Scale forms at OSU-CVM was modeled on the CBVE model and a previously validated entrustment scale used at the University of Calgary (17). There were changes over time from novice to competent for each domain and competency on ITER scores. Analyses provide foundational support for the reliability of the scores and some supportive validity evidence on which to make important progress decisions. However, entrustment scores provided little differentiation between and within students, on EPAs, over the clinical rotation period. To incorporate and use scores from both methods (in conjunction with other clinical rotation assessment methods) to make competency decisions in a programmatic assessment framework requires further training of all stakeholders in order to maximize their intended use. Gathering of multiple data points per learner over time, using various assessment tools, is key to a more holistic assessment of student ability and is consistent with providing the foundation for more evidence-based practice of programmatic assessment within veterinary medicine.

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Ethics statement

The studies involving human participants were reviewed and approved by The Ohio State University Office of Responsible Research Practices. Written informed consent for participation was not required for this study in accordance with the national legislation and the institutional requirements.

Author contributions

ER conceived, led the design of the assessments, and conducted the study. CM organized the database and performed the statistical analysis. KH advised on the assessment design and statistical analysis. All authors contributed to the manuscript revisions and approved the final version.

Acknowledgments

The authors thank the final year teaching faculty, residents, interns and clinical staff, as well as the professional programs staff of The Ohio State University College of Veterinary Medicine for their roles in data generation and collection respectively. The authors also acknowledge the class of 2021 learners themselves for their contributions and engagement in this process. Finally, the authors wish to thank the Clinical Teachers' Committee for having the vision to implement the CBVE model into their assessment redesign in the final year program. Their collective wisdom and effort has helped produce the system described in this manuscript.

Conflict of interest

Author KH is the Chief Assessment Officer for the International Council for Veterinary Assessment.

The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

1. Danielson JA. Key assumptions underlying a competency-based approach to medical sciences education, and their applicability to veterinary medical education. Front Vet Sci. (2021) 595. doi: 10.3389/fvets.2021.688457

PubMed Abstract | CrossRef Full Text | Google Scholar

2. Molgaard et al. AAVMC working group on competency-based veterinary education. In:Molgaard LK, Hodgson JL, Bok HGJ, Chaney KP, Ilkiw JE, Matthew SM, May SA, Read EK, Rush BR, Salisbury SK, , editors. Competency-Based Veterinary Education: Part 1—CBVE Framework. Washington, DC: Association of American Veterinary Medical Colleges (2018).

Google Scholar

3. Molgaard et al. AAVMC working group on competency-based veterinary education. In:Molgaard LK, Hodgson JL, Bok HGJ, Chaney KP, Ilkiw JE, Matthew SM, May SA, Read EK, Rush BR, Salisbury SK, , editors. Competency-Based Veterinary Education: Part 2—Entrustable Professional Activities. Washington, DC: Association of American Veterinary Medical Colleges (2018).

Google Scholar

4. Salisbury et al. AAVMC working group on competency-based veterinary education. In:Salisbury SK, Chaney KP, Ilkiw JE, Matthew SM, Read EK, Bok HG, Hodgson JL, May SA, Molgaard LK, Rush BR, , editors. Competency-Based Veterinary Education: Part 3—Milestones. Washington, DC: Association of American Veterinary Medical Colleges (2019).

Google Scholar

5. Bok HG, van der Vleuten CPM, de Jong LH. “Prevention is better than cure”: a plea to emphasize the learning function of competence committees in programmatic assessment. Front Vet Sci. (2021) 8:638455. doi: 10.3389/fvets.2021.638455

PubMed Abstract | CrossRef Full Text | Google Scholar

6. Norcini J, Anderson MB, Bollela V, Burch V, Costa MJ, Duvivier R, et al. Consensus framework for good assessment. Med Teach. (2018) 33:206–14. doi: 10.1080/0142159X.2018.1500016

PubMed Abstract | CrossRef Full Text | Google Scholar

7. Schut S, Maggio LA, Heeneman S, van Tartwijk J, van der Vleuten C, Driessen E. Where the rubber meets the road—an integrative review of programmatic assessment in health care professions education. Perspect Med Educ. (2021) 10:6–13. doi: 10.1007/s40037-020-00625-w

PubMed Abstract | CrossRef Full Text | Google Scholar

8. van der Vleuten CPM, Schuwirth LWT. Assessment in the context of problem-based learning. Adv Health Sci Educ. (2019) 24:903–14. doi: 10.1007/s10459-019-09909-1

PubMed Abstract | CrossRef Full Text | Google Scholar

9. Violato C, Cullen MJ, Englander R, Murray KE, Hobday PM, Borman-Shoap E, et al. Validity evidence for assessing entrustable professional activities during undergraduate medical education. Acad Med. (2021) 96:S70–5. doi: 10.1097/ACM.0000000000004090

PubMed Abstract | CrossRef Full Text | Google Scholar

10. Boursicot K, Kemp S, Wilkinson T, Findyartini A, Canning C, Cilliers F, Fuller R. Performance assessment: consensus statement and recommendations from the 2020 Ottawa conference. Med Teach. (2021) 58–67. doi: 10.1080/0142159X.2020.1830052

PubMed Abstract | CrossRef Full Text | Google Scholar

11. Senders ZJ, Brady JT, Ladhani HA, Marks J, Ammori JB. Factors influencing the entrustment of resident operative autonomy: comparing perceptions of general surgery residents and attending surgeons. J Grad Med Educ. (2021) 13:675–81. doi: 10.4300/JGME-D-20-01259.1

PubMed Abstract | CrossRef Full Text | Google Scholar

12. Hamstra S J, Cuddy MM, Jurich D, Yamazaki K, Burkhardt J, Holmboe ES, et al. Exploring the association between USMLE scores and ACGMA milestone ratings: a validity study using national data from emergency medicine. Acad Med. (2021) 96:1324. doi: 10.1097/ACM.0000000000004207

PubMed Abstract | CrossRef Full Text | Google Scholar

13. Bok H. deJong L, O'Neill T, Maxey C, Hecker KG. Validity evidence for programmatic assessment in competency-based education. Perspect Med Educ. (2018) 7:362–72. doi: 10.1007/s40037-018-0481-2

PubMed Abstract | CrossRef Full Text | Google Scholar

14. Bok HG, Jaarsma DADC, Teunissen PW, van der Vleuten CPM, van Beukelen P. Development and validation of a competency framework for veterinarians. J Vet Med Educ. (2011) 38:262–9. doi: 10.3138/jvme.38.3.262

PubMed Abstract | CrossRef Full Text | Google Scholar

15. Salisbury SK, Rush BR, Ilkiw JE, Matthew SM, Chaney KP, Molgaard LK, et al. Collaborative development of core entrustable professional activities for veterinary education. J Vet Med Educ. (2020) 47:607–18. doi: 10.3138/jvme.2019-0090

PubMed Abstract | CrossRef Full Text | Google Scholar

16. Ten Cate O. Competency-based postgraduate medical education: past, present and future. GMS J Med Educ. (2017) 34.

PubMed Abstract | Google Scholar

17. Read EK, Brown A, Maxey C, Hecker KG. Comparing entrustment and competence: an exploratory look at performance-relevant information in the final year of a veterinary program. J Vet Med Educ. (2021) 48:562–72. doi: 10.3138/jvme-2019-0128

PubMed Abstract | CrossRef Full Text | Google Scholar

18. Kane MT. Validating the interpretations and uses of test scores. J Educ Meas. (2013) 50:1–73. doi: 10.1111/jedm.12000

PubMed Abstract | CrossRef Full Text | Google Scholar

19. Kane M. Validation. In:Brennan RL, , editor. Educational Measurement, 4th Edn. Westport, CT. American Council on Education (2006). pp. 17–64.

Google Scholar

20. Duijn CCMA., ten Cate O, Kremer WDJ, Bok HGJ. The development of entrustable professional activities for competency-based veterinary education in farm animal health. J Vet Med Educ. (2019) 46:218–24. doi: 10.3138/jvme.0617-073r

PubMed Abstract | CrossRef Full Text | Google Scholar

21. Pusic MV, Boutis K, Hatala R, Cook DA. Learning curves in health professions education. Acad Med. (2015) 90:1034–42. doi: 10.1097/ACM.0000000000000681

PubMed Abstract | CrossRef Full Text | Google Scholar

22. Hall AK, Rich J, Dagnone JD, Weersink K, Caudle J, Sherbino J, et al. It's a marathon, not a sprint: rapid evaluation of competency-based medical education program implementation. Acad Med. (2020) 95:786–93. doi: 10.1097/ACM.0000000000003040

PubMed Abstract | CrossRef Full Text | Google Scholar

23. Booth M, Rishniw M, Kogan LR. The shortage of veterinarians in emergency practice: a survey and analysis. J Vet Emerg Crit Care. (2021) 31:295–305. doi: 10.1111/vec.13039

PubMed Abstract | CrossRef Full Text | Google Scholar

24. Robinson TJG, Wagner N, Szulewski A, Dudek N, Cheung WJ, Hall AK. Exploring the use of rating scales with entrustment anchors in workplace-based assessment. Med Educ. (2021) 55:1047–55. doi: 10.1111/medu.14573

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: longitudinal assessment, ITER, in-training evaluation report, EPA, entrustable professional activity, entrustment, milestones, CBVE

Citation: Read EK, Maxey C and Hecker KG (2022) Longitudinal assessment of competency development at The Ohio State University using the competency-based veterinary education (CBVE) model. Front. Vet. Sci. 9:1019305. doi: 10.3389/fvets.2022.1019305

Received: 14 August 2022; Accepted: 29 September 2022;
Published: 24 October 2022.

Edited by:

Catherine Elizabeth Stalin, University of Glasgow, United Kingdom

Reviewed by:

Arno H. Werners, St. George's University, Grenada
Virginia Fajt, Texas A&M University, United States

Copyright © 2022 Read, Maxey and Hecker. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Emma K. Read, cmVhZC42NSYjeDAwMDQwO29zdS5lZHU=

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.