- 1Abu Dhabi Healthcare Services, Ambulatory Healthcare Services, Al Ain, United Arab Emirates
- 2Community Medicine Department, UAEU, Al Ain, United Arab Emirates
Background: Competency-Based Medical Education (CBME) is now mandated by many graduate and undergraduate accreditation standards. Evaluating CBME is essential for quantifying its impact, finding supporting evidence for the efforts invested in accreditation processes, and determining future steps. The Ambulatory Healthcare Services (AHS) family medicine residency program has been accredited by the Accreditation Council of Graduate Medical Education-International (ACGME-I) since 2013. This study aims to report the Abu Dhabi program’s experience in implementing CBME and accreditation.
Objectives: 1. Compare the two residents’ cohorts’ performance pre-and post-ACGME-I accreditation.
2. Study the bi-annually reported milestones as a graduating residents’ performance prognostic tool.
Methods: All residents in the program from 2008 to 2019 were included. They are called Cohort one—the intake from 2008 to 2012, before the ACGME accreditation, and Cohort two—the intake from 2013 to 2019, after the ACGME accreditation, with the milestones used. The mandatory annual in-training exam was used as an indication of the change in competency between the two cohorts. Among Cohort two ACGME-I, the biannual milestones data were studied to find the correlation between residents’ early and graduating milestones.
Results: A total of 112 residents were included: 36 in Cohort one and 76 in Cohort two. In Cohort one, before the ACGME accreditation, no significant associations were identified between residents’ graduation in-training exam and their early performance indicators, while in Cohort two, there were significant correlations between almost all performance metrics. Early milestones are correlated with the graduation in-training exam score. Linear regression confirmed this relationship after controlling the residents’ undergraduate Grade Point Average (GPA). Competency development continues to improve even after residents complete training at Post Graduate Year, PGY4, as residents’ achievement in PGY5 continues to improve.
Conclusion: Improved achievement of residents after the introduction of the ACGME-I accreditation is evident. Additionally, the correlation between the graduation in-training exam and graduation milestones, with earlier milestones, suggests a possible use of early milestones in predicting outcomes.
Introduction
Accreditation in medical education was introduced as an evidence-based process to improve outcomes. Accreditation is defined as the formal evaluation of an educational program, institution, or system, against defined standards, by an external body for quality assurance and enhancement (1). Many reports support that accreditation-related activities steer medical education toward establishing processes that are likely to improve the quality of medical education (2–4). Accrediting bodies increasingly view practice outcome measures as an essential step toward better continuous educational quality improvement (5). Thus, the development of competency-based medical education became an obligatory standard for many accrediting bodies, such as the Accrediting Council of Graduate Medical Education and ACGME-International. The recent introduction of milestones and Entrustable Professional Activities (EPA) is an important development in this area, which needs more research and innovation in its implementation and validation with graduates’ assessment of performance in patient care.
Nevertheless, such an assessment is challenging. While milestones and EPA are works in progress, the milestones completed twice annually have helped programs identify individual residents struggling globally or in a specific area earlier in residency. The process has also helped programs identify areas where the curriculum of the residency program may need improvements (5). On the other hand, competency-based medical education (CBME) implementation creates challenges, such as synchronizing the training of residents, as not all trainees progress at the same speed. Another challenge is that the implementation can be labor-intensive for supervisors and educators.
Implementing CBME may be associated with different challenges in different countries, health systems, or even between PGME and UME in the same country. Therefore, experiences from different programs may not be generalizable, but still, the literature needs evidence in this area. In Abu Dhabi, the government mandated, in 2011, that all residency training should be ACGME-I accredited, and most programs achieved this by 2013. The ACGME-I accreditation signed an agreement with Abu Dhabi Healthcare Services to conduct site visits and accredit programs that meet the ACGME-I standards. The first group of residents for the family medicine residency program of the Ambulatory Healthcare Services (AHS), joined on October 1st, 1994. Since then, more than 160 family physicians have graduated. Its curriculum and process were modified to meet the ACGME-I, which accredited the program in 2013 and renewed the accreditation to date. The milestones were introduced in 2014. There is consistency in the program leadership as the program director, and all core faculty are program graduates who have gone through all or most of the accreditation site visits. In addition, they have attended many ACGME-I faculty development courses conducted over the year, particularly on curriculum and CBME assessments and milestones.
The Family Medicine Program of the Ambulatory Healthcare Services (AHS) was among the first accredited and implemented by all required standards. A vital element of the Accreditation System is the measuring and reporting of educational milestones since 2014.
This study aims to retrospectively reflect on the Abu Dhabi residency program experience in implementing CBME and accreditation. Two cohorts of the AHS Family Medicine program residents were compared pre-and post-ACGME-I accreditation regarding their achievements in different assessments, among them the in-training exam (ITE) scores. Also, the second cohort’s bi-annually reported milestones were used as a prognostic tool for graduating residents’ ITE and milestones.
Methods
The study is a sectional observational study comparing two cohorts before the ACGME standards implementation and after. Residents included in the study are called Cohort one, whose intake was from 2008 to 2012 before the ACGME accreditation, and Cohort two, whose intake was from 2013 to 2019, after the ACGME accreditation, with the milestones used. Regarding outcome measures, usually, the quality of graduates is best judged by the Board Certification, and in this program, it is the Arab Board examination. The pass rate for the program is above 97% overall, and the results were pass or fail. Therefore, it is not an appropriate outcome for this small cohort. Therefore, it was decided that the in-training exam of the American Board was a better alternative to judge competency achievement. It has the same structure and similar content focus as the Arab Board exam. It is a Multiple-Choice Written exam and covers all current family medicine curricula. Additional competency indicators from the internal program assessments were clinical long-case exams, Objective Structured Clinical Examination (OSCE), oral exams, the yearly in-training exam, and multiple evaluations required by the accrediting body, ACGME-I, which were used as variables in the study of the residents’ competency. Clinical Competency Committees (CCC) review the many different sources of information for each resident at least four times annually and evaluate each resident’s progress across the milestones twice annually. It is worth noting that other than the clinical exams and oral exams, all other assessments were formative and not summative.
All residents in both cohorts have these assessments: yearly clinical exams, oral exams, and yearly OSCE’s and in-training exams. The two cohorts of residents, who were all residents in the program from 2008 to 2019, were compared in all these assessments. The only added assessment in Cohort two, which was not part of Cohort one assessments, was the milestones, which were introduced as part of the required reporting of performance by the accrediting body, the ACGME-I.
Summative assessments included yearly clinical exams and oral exams, while yearly OSCE and in-training exams were obligatory formative assessments. The in-training exam, which is a requirement in the program to be taken by all residents annually, was used as one indication of the change in competency between the two cohorts. In addition, among Cohort two, ACGME-I milestones’ data, which is collected biannually, was studied to find the correlation between the residents’ early and graduating milestones. The program is for four years, but the fifth year in-training exam is available due to extensions for a few or most residents to remain in the program until their certification exam, which is usually during mid-year 5.
Milestones data from residents included 6 or 8 semi-annual reporting periods from July 2016 to June 2019, using the milestones assessment system reported to the ACGME-I. Family medicine is a 3-year ACGME-I accredited training program. However, some residents need to extend training and will have milestones for an extra year. in the first version of the milestones that the program implanted, there were 22 sub-competencies across the 6 ACGME Core Competencies, and now in version 2.0 there are 18: there are five patient care (PC) sub-competencies, two medical knowledge (MK) sub-competencies, two practice-based learning and improvement (PBLI) sub-competencies, three system-based practice (SBP) sub-competencies, two professionalism sub-competencies, and three interpersonal communication skills (ICS) sub-competencies (Table 1).
The milestones data are scored between level 1 and level 5 in 0.50-unit intervals (and a preceding level 0 to indicate that the learner has not achieved level 1). Level 4 is specified as the recommended graduation target, which indicates readiness for unsupervised practice. Milestones are designed mainly for formative purposes, including providing feedback on progress made and remaining deficiencies, and are not used for accrediting individual residency programs or eligibility determinations for board certification (6). The value of milestones, and other in-training assessments, depends on the extent to which they measure progress and predict final assessments. A surrogate outcome for board certification was used, namely, the graduates’ in-training exam (at year 4 or 5), which reflects their Board Certification preparedness at the point of completion of their training.
To investigate possible influential factors on outcome in this study, we utilized all other available residents’ data and assessments, including age, gender, medical school GPA, and entry interview score.
Statistical methods used include frequency tables and t-tests. As all outcome variables were continuous variables, correlation coefficients and linear regression analysis were used. A two-tailed significance level of 0.05 was used in statistical tests. For comparing Pearson’s correlation coefficients between the two cohorts, a value of 0.35 was considered significant as it reached the 0.05 significance level in the smaller cohort.
Abu Dhabi Healthcare Services, SEHA, and IRB approved this study. The approval number is SEHA-IRB-035. No consent was needed as the data were retrospectively gathered and anonymized.
Results
A total of 112 family medicine residents, 36 in Cohort one and 76 in Cohort two, were included. The majority were females (n = 34, 94.4% in Cohort one and n = 59, 77.6% in Cohort two). The mean ages of graduation for the program were 32.6 years (SD: 3.16) in Cohort one and 28.6 years (SD 2.76) in Cohort two. The GPA average was 2.77 (SD:0.55) in Cohort one and 3.00 (SD: 0.41) in Cohort two. The mean score for the family medicine entry interview was 81.21 (SD: 9.26) in Cohort one and 90.93 (SD:8.86) in Cohort two. All were UAE nationals (Table 2).
The mean ITE scores for PGY1 were similar in Cohorts one and two—237.8 in Cohort one, and 242.80 in Cohort two. For PGY2, the mean ITE score was 260 in Cohort one, while it was 307.12 in Cohort two. The mean ITE for PGY3 in Cohort one was 285.04, while in Cohort two, it was 328.54. For PGY4, the mean ITE score was 286.25 in Cohort one, while in Cohort two, it was 343.49. However, there were no significant differences between Cohorts one and two in these ITE scores and other performance, clinical, and oral exams (Table 2; Figure 1).
Figure 1. The difference in in-training average score between the two cohorts among PGY1, PGY2, PGY3, PGY4, PGY5.
Studying the correlation between different competency outcomes, using Pearson correlation analysis, found a less significant association between the Cohort one residents’ graduation in-training exam and their early performance in residency years, such as clinical exams. The year one oral exam, year two OSCE, and the in-training exams in years 1 and 2 had significant correlations (r = 0.66, r = 0.37, r = 0.45, r = 0.64, respectively) (Table 3). In the newer cohort, Cohort two (2013–2019), there was a more significant correlation between the graduation in-training exam with almost all performance metrics (GPA, OSCE in PGY1 and PGY2, ITE in PGY1 and PGY2, clinical exam in PGY1 and PGY2, oral exam in PGY1 only, and mid and end milestone from in PGY1 to PGY4). More importantly, the graduation in-training exam score was correlated with both ends of the PGY1 milestone (r = 0.47), mid-year 2 (r = 0.65), and the end of the PGY2 milestone (r = 0.71). GPA was not significantly correlated with graduation ITE in both cohorts (Table 3). As shown in Table 4, early, mid, and end year 1 and 2 milestones correlated well with graduation ITE and milestones.
Table 4. Correlation between the milestones assessed at different training years for Cohort two which had the milestones implemented.
This relation between early and graduating milestones was confirmed using univariate linear regression except for the very early milestones (mid-first year) (Table 5). With regards to the milestones sub-competencies in patient care, medical knowledge, professionalism, interprofessional and personal skills, system-based practice, and practice-based learning and improvement, linear regression showed a significant prediction of almost all sub-competencies with the ITE in year four after completion of the three ACGME-I programs’ requirements (Table 5). This was for all years’ sub-competencies from PGY1 to PGY3.
Table 5. Linear regression of outcome with other early performance variables in Cohort one and Cohort two.
It is of value to note that competency development continues with time spent in residency as we can see a continuous development of the residents in the in-training performance and milestones in years 4 and 5. There was no plateau, but the in-training performance and milestones showed continuous improvement (Table 6; Figure 2). A positive correlation exists between graduation ITE score and the early milestone in PGY1 and PGY2. Also, the positive correlation found between GPA and graduation ITE was noted. However, the correlation between graduation GPA and entry interview score was not statistically significant.
Table 6. ACGME milestones with the average score of individual competencies and all competencies average distributed by the time in residency training.
Discussion
Abu Dhabi’s investment in medical education, through accreditation, paved the way for better achievement in medical education, and hopefully, this will reflect in a more capable medical workforce. Accreditation system implementation requires restructuring institutions and training programs involved as it implies incorporating quality assurance and outcome tracking systems. The better mean ITE (In Training Examination) score, through academic years PGY1 to PGY4, after the implementation of ACGME accreditation and milestones and the association between the graduating ITE and earlier years’ ITE scores are in line with a systematic review by McCrary et al., which found a moderate-strong relationship between ITE and board examination performance. ITE scores significantly predict board examination scores for most studies. Performing well on an ITE predicts a passing outcome for the board examination, but there is little evidence that performing poorly on an ITE will result in failing in the associated specialty board examination (7).
What explains the improved ITE scores is mainly the implementation of the accreditation standards, as nothing else has changed, and the implementation change is very large. Such standards demanded structure with responsibilities and task distribution within the program. New processes of generating, communicating, and managing performance data were introduced. In addition to adding new curriculum, new teaching initiatives, and the integration of supervised patient care with competency assessments at the workplace. A major change as well is the shift toward competency-based education. The six ACGME competencies were targeted in building curriculum, evaluations, assessments, and target outcomes.
Improving learners’ performance was preplanned based on data fed to the CCC with the generation of milestones bi-annually and the feedback after the data review and actions recommended by the CCC regarding possible residents’ remediation, development opportunities, or promotion decisions. This is a collective decision from multiple sources and expertise and based on gathered observations and data. That was not available before in the case of Cohort one when the Program Director (PD) had to make such performance improvement initiatives and make decisions on learners’ progress. The feedback system to residents through individualized meetings biannually, after milestones assessments, and promotions are required by accreditation standards. Such meetings end with an individualized action plan followed by PD and supervisors for implementation. Weak residents were identified, and an action plan was discussed with them and followed by the PD and CCC committee.
With regards to educational programs, planning of the sessions utilized feedback from the CCC and the PEC and necessitated coverage of all areas of competencies, sometimes, special workshops were added. Curriculum change was a must with the accreditation. For example, there was no structured sports medicine or geriatric services in the city in the time of Cohort one, and no specialist was there in the whole area. When the program was cited for that, a curriculum was built to cover this, and residents had to monitor the volume of elderly patients seen by them, more time was given to both areas with special workshops built to develop the competencies of residents in these areas as sports medicine workshops and in-depth coverage of geriatrics topics.
Additionally, ITE results were utilized to inform the program of better and worse global performance of the residents, which was aimed at in more board review courses and topics in the educational program.
Finally, the milestone is based on ACGME competencies, and among these competencies, as an example, is medical knowledge, which has multiple sources of data. And on a quarterly basis, the CCC will assess progress, and this may influence the In-training performance of residents as it is targeted among other competencies. Additionally, beyond the effect on the ITE score, we provided evidence that these milestones added value to the implemented accreditation system and represented a more comprehensive approach to competency assessment and predicting residency outcomes. Therefore, milestones could be a key player in the difference noticed between Cohorts one and two. It required mandatory tracking and reporting of changes in competency for individual residents of Cohort two. That could not be made without the faculty’s efforts and data collected and analyzed and generating conclusions that were acted upon. Most importantly, the accreditation standards required ensuring structure and processes and a demand for time and resources for the program that supported the residents and faculty, which was not there for Cohort one.
Quantitative and qualitative data for both committees, at different data points from different sources, need to be processed by committee members to be acted upon by Program Director (PD), faculty, and residents (8–10). Assessment tools, developed to assist with milestone assignments, also captured the growth of residents over time and demonstrated quantifiable differences in achievements between PGY classes, which aid timely intervention in supporting residents’ competency development (11). In summary, the difference brought about by the accreditation is a standardized educational structure and processes, with measurements aiming to improve faculty’s accountability in assessment and enhance transparency, and it requires more adherence to administrative tasks and peer assessments from residents and faculty. In a review by Andolsek, starting similar structures and processes as early as in medical schools was recommended to facilitate better resident selection and development (12).
It is worth noting that both GPA and milestones predicted in-training results at graduation. However, while both utilize classic assessment tools like multiple choice questions, MCQs and OSCE, which measure basic medical knowledge, milestones focus on applying acquired knowledge and skills in the workplace assessment and EPA. According to Holmboe and others, milestones data offers a valid and reliable predictive tool that guides the programs to make formal decisions according to the resident’s progress and needs. This might even extend to provide some prediction on the level of the resident’s readiness to face clinical practice (6, 13). Whether this makes milestones data a better predictor of the quality of care that residents provide to patients deserves further empirical study.
Our study showed a positive correlation between the milestone scores and residents’ performance in the graduating written exam; similar findings were noted in other residency programs, such as the Obstetrics and Gynecology and Surgery residency programs, especially in the medical knowledge sub-competency (14–17). The significant prediction of graduation ITE score by the early sub-competencies is an important observation as the ITE is very similar to the exit certifying exam. Therefore, early signs of poor performance, as judged by milestones, can guide targeted remedial actions and provide better support for residents. Although the ACGME-I accredited program is three years long, Arab Board Certification requires four years, and residents can have extensions of their training until mid-year 5, until the Arab Board Certification exam. This provides an opportunity to explore continuous development and performance, especially in years 4 and 5. Residents are given more responsibilities within the practice, and training is more practice-based, which resembles future practice.
Entry interview score in both cohorts, was not significantly linked to any important outcome similar to other reports internationally; a study by Burkhardt et al. did not find consistent, meaningful correlations between the most common selection factors and milestones at any point in the training (18). In this program, the weightage for interview is 75% of the final interview score, and a weight of 25% is given to the GPA. Therefore, we cannot compare our data to other studies as we know that many programs use different entry composite scores that usually include numerical ratings for Dean’s letters and a weighted score for on-site interviews and the GPA. A study by Prystowsky et al. found that higher scores in their program were associated more with the graduate being in an academic faculty position. Therefore, an entry interview alone is difficult to interpret (19). Not surprisingly, the process of selection is challenged and there are increasing calls to reform the US resident selection process as highlighted by a recent scoping review (20). With the use of the milestones initiative, there is a shift from what abilities physicians, in general, should possess to what medical specialists should possess. An increasing number of postgraduate programs have now started to link competencies to Entrustable Professional Activities (EPAs) (21).
Caution in interpreting this study’s findings is essential as there may be alternative explanations for the better performance of Cohort two. For example, more supervisors were program graduates in Cohort two than in Cohort one. Still, as accreditation standards stress investment in faculty development, we can attribute success to accreditation. Finally, these results highlight the value of the competency-based curriculum and assessment metrics during the internship year to better prepare residents for their first residency year and build a handover system between the two graduate educational levels. Ultimately, the ideas underlying the accreditation process should affect the whole medical curriculum.
Limitation
The outcome, being an exam, does not provide actual work capability in practice. However, the milestones incorporate increasing workplace assessments, which can be a possible strength for milestone assessments. Future research linking milestones to post-graduation performance is crucial to answering this question. And this significant correlation is a start to assessing future performance as graduates. Only one residency program, in one site with two cohorts, was reported in the study, which needs to include a larger number of residency programs.
Our family medicine program is implementing the Entrustable Professional Activities (EPAs) as a further step in competency assessment and development, which will provide future research opportunities to study its impact on the outcome.
Conclusion
The introduction of the ACGME-I accreditation was associated with increasing the residents’ achievements. Similar to other international studies, milestones proved to be a promising instrument for competency acquisition in the Abu Dhabi AHS family medicine program. The correlation between the graduation in-training exam and graduation milestones with earlier milestones suggests a possible use of early milestones in predicting outcomes.
Data availability statement
Data is available on request sent to the corresponding author.
Ethics statement
The studies involving humans were approved by Abu Dhabi Healthcare Services, SEHA IRB. The studies were conducted in accordance with the local legislation and institutional requirements. The ethics committee/institutional review board waived the requirement of written informed consent for participation from the participants or the participants’ legal guardians/next of kin because informed consent was waived by Abu Dhabi Healthcare Services, SEHA, and IRB as the study was designed for retrospective data gathered as part of residency program training and anonymized at analysis.
Author contributions
LB: Conceptualization, Data curation, Formal analysis, Methodology, Project administration, Writing – original draft. NN: Formal analysis, Writing – review & editing. AAA: Writing – review & editing, MK: Writing – review & editing. RA: Writing – review & editing. AMA: Writing – review & editing. SA: Writing – review & editing. MK: Writing – review & editing.
Funding
The author(s) declare that no financial support was received for the research, authorship, and/or publication of this article.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
1. Frank, JR, Taber, S, van Zanten, M, Scheele, F, and Blouin, D, International HPAOC. The role of accreditation in 21st century health professions education: report of an international consensus group. BMC Med Educ. (2020) 20:305. doi: 10.1186/s12909-020-02121-5
2. Barzansky, B, Hunt, D, Moineau, G, Ahn, D, Lai, CW, Humphrey, H, et al. Continuous quality improvement in an accreditation system for undergraduate medical education: benefits and challenges. Med Teach. (2015) 37:1032–8. doi: 10.3109/0142159X.2015.1031735
3. Blouin, D, Tekian, A, Kamin, C, and Harris, IB. The impact of accreditation on medical schools’ processes. Med Educ. (2018) 52:182–91. doi: 10.1111/medu.13461
4. Holt, KD, Miller, RS, Byrne, LM, and Day, SH. The positive effects of accreditation on graduate medical education programs in Singapore. J Grad Med Educ. (2019) 11:213–7. doi: 10.4300/JGME-D-19-00429
5. Bandiera, G, Frank, J, Scheele, F, Karpinski, J, and Philibert, I. Effective accreditation in postgraduate medical education: from process to outcomes and back. BMC Med Educ. (2020) 20:307. doi: 10.1186/s12909-020-02123-3
6. Park, YS, Hamstra, SJ, Yamazaki, K, and Holmboe, E. Longitudinal reliability of milestones-based learning trajectories in family medicine residents. JAMA Netw Open. (2021) 4:e2137179. doi: 10.1001/jamanetworkopen.2021.37179
7. McCrary, HC, Colbert-Getz, JM, Poss, WB, and Smith, BK. A systematic review of the relationship between in-training examination scores and specialty board examination scores. J Grad Med Educ. (2021) 13:43–57. doi: 10.4300/JGME-D-20-00111.1
8. Friedman, KA, Raimo, J, Spielmann, K, and Chaudhry, S. Resident dashboards: helping your clinical competency committee visualize trainees’ key performance indicators. Med Educ Online. (2016) 21:29838. doi: 10.3402/meo.v21.29838
9. Kou, M, Baghdassarian, A, Rose, JA, Levasseur, K, Roskind, CG, Vu, T, et al. Milestone achievements in a national sample of pediatric emergency medicine fellows: impact of primary residency training. AEM Educ Train. (2021) 5:e10575. doi: 10.1002/aet2.10575
10. Meier, AH, Gruessner, A, and Cooney, RN. Using the ACGME milestones for resident self-evaluation and faculty engagement. J Surg Educ. (2016) 73:e150–7. doi: 10.1016/j.jsurg.2016.09.001
11. Goldman, RH, Tuomala, RE, Bengtson, JM, and Stagg, AR. How effective are new milestones assessments at demonstrating resident growth? 1 year of data. J Surg Educ. (2017) 74:68–73. doi: 10.1016/j.jsurg.2016.06.009
12. Andolsek, KM. Improving the medical student performance evaluation to facilitate resident selection. Acad Med. (2016) 91:1475–9. doi: 10.1097/ACM.0000000000001386
13. Holmboe, ES, Yamazaki, K, Nasca, TJ, and Hamstra, SJ. Using longitudinal milestones data and learning analytics to facilitate the professional development of residents: early lessons from three specialties. Acad Med. (2020) 95:97–103. doi: 10.1097/ACM.0000000000002899
14. Bienstock, JL, Shivraj, P, Yamazaki, K, Connolly, AM, Wendel, G, Hamstra, SJ, et al. Correlations between accreditation Council for Graduate Medical Education Obstetrics and Gynecology Milestones and American Board of Obstetrics and Gynecology qualifying examination scores: an initial validity study. Am J Obstet Gynecol. (2021) 224:308.e1–308.e25. doi: 10.1016/j.ajog.2020.10.029
15. Cassaro, S, Jarman, BT, Joshi, ART, Goldman-Mellor, S, Hope, WW, Johna, S, et al. Mid-year medical knowledge milestones and ABSITE scores in first-year surgery residents. J Surg Educ. (2020) 77:273–80. doi: 10.1016/j.jsurg.2019.09.012
16. Kimbrough, MK, Thrush, CR, Barrett, E, Bentley, FR, and Sexton, KW. Are surgical milestone assessments predictive of in-training examination scores. J Surg Educ. (2018) 75:29–32. doi: 10.1016/j.jsurg.2017.06.021
17. Beeson, MS, Holmboe, ES, Korte, RC, Nasca, TJ, Brigham, T, Russ, CM, et al. Initial validity analysis of the emergency medicine milestones. Acad Emerg Med. (2015) 22:838–44. doi: 10.1111/acem.12697
18. Burkhardt, JC, Parekh, KP, Gallahue, FE, London, KS, Edens, MA, Humbert, AJ, et al. A critical disconnect: residency selection factors lack correlation with intern performance. J Grad Med Educ. (2020) 12:696–704. doi: 10.4300/JGME-D-20-00013.1
19. Prystowsky, MB, Cadoff, E, Lo, Y, Hebert, TM, and Steinberg, JJ. Prioritizing the interview in selecting resident applicants: behavioral interviews to determine goodness of fit. Acad Pathol. (2021) 8:23742895211052885. doi: 10.1177/23742895211052885
20. Zastrow, RK, Burk-Rafel, J, and London, DA. Systems-level reforms to the US resident selection process: a scoping review. J Grad Med Educ. (2021) 13:355–70. doi: 10.4300/JGME-D-20-01381.1
Keywords: residency training, milestones, ACGME, in-training exams, competency residency training, performance, competency
Citation: Baynouna AlKetbi L, Nagelkerke N, AlZarouni AA, AlKuwaiti MM, AlDhaheri R, AlNeyadi AM, AlAlawi SS and AlKuwaiti MH (2024) Assessing the impact of adopting a competency-based medical education framework and ACGME-I accreditation on educational outcomes in a family medicine residency program in Abu Dhabi Emirate, United Arab Emirates. Front. Med. 10:1257213. doi: 10.3389/fmed.2023.1257213
Edited by:
Yajnavalka Banerjee, Mohammed Bin Rashid University of Medicine and Health Sciences, United Arab EmiratesReviewed by:
Liesl Visser, Mohammed Bin Rashid University of Medicine and Health Sciences, United Arab EmiratesSamuel Ho, Mohammed Bin Rashid University of Medicine and Health Sciences, United Arab Emirates
Copyright © 2024 Baynouna AlKetbi, Nagelkerke, AlZarouni, AlKuwaiti, AlDhaheri, AlNeyadi, AlAlawi and AlKuwaiti. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Latifa Baynouna AlKetbi, bGF0aWZhLm1vaGFtbWFkQGdtYWlsLmNvbQ==