Skip to main content

ORIGINAL RESEARCH article

Front. Physiol., 10 July 2023
Sec. Integrative Physiology
This article is part of the Research Topic Physiological Education: Preparing for the Future View all 9 articles

Learning and assessment strategies to develop specific and transversal competencies for a humanized medical education

Antonio S. TutorAntonio S. TutorEsther EscuderoEsther EscuderoMaría del Nogal vilaMaría del Nogal ÁvilaJuan Francisco ArandaJuan Francisco ArandaHortensia TorresHortensia TorresJosu G. YagueJosué G. YagueMaría Jos BorregoMaría José BorregoÚrsula MuozÚrsula MuñozMaría C. SdabaMaría C. SádabaIsabel Snchez-Vera
Isabel Sánchez-Vera*
  • Departamento de Ciencias Médicas Básicas, Facultad de Medicina, Universidad San Pablo-CEU, CEU Universities, Madrid, Spain

Introduction: Medical education should promote the development of skills and abilities that can be applied to real-world work performance. The aim of this study is to evaluate technical and methodological knowledge, as well as physician-patient communication skills, as one of the most important transversal competencies that a good physician should acquire; all this in a reliable, accurate and objective way.

Methods: We present a rubric specifically designed and implemented for the evaluation of specific and transversal competencies in the physiology practical sessions, during the second year of the medical degree. The assessment consists in two evaluation tests: 1) a theoretical test that consists of multiple-choice questions. Students must demonstrate that they have acquired adequate theoretical knowledge (specific competency “to know”); 2) a practical test, in which students are evaluated by the rubric through the simulation of a medical consultation. Thus, demonstrating their ability to execute/apply what they have learned in class (specific competency “to know how to do”). They are also evaluated on the transversal competencies that we call “communication with the patient” (transversal competency “to know how to be there”) and “dealing with the patient” (transversal competency “to know how to be”).

Results: We evaluated whether there were differences in the grades obtained by students when the transversal competencies were not assessed (academic years 2017-2018 and 2018-2019; n = 289), and when the transversal competencies were assessed by applying the rubric in the academic years 2019-2020, 2021-2022, and 2022-2023 (n = 526). Furthermore, we present a student perception that supports the use of clinical simulation and our rubric as a good method within the competency learning process.

Discussion: The acquisition of these competencies, starting from the first courses of undergraduate education, helps to raise the students’ awareness in the development of a more humanized medicine, allowing a better response to the patients’ needs. Our rubric, which clearly indicate the performance criteria, have become an excellent method to carry out the assessment of competencies, both for students and teachers, since they allow to obtain clear evidence of the level of acquisition and application of knowledge.

1 Introduction

It has been known for quite some time that academic records (acquired knowledge) do not always provide sufficient information to reliably predict people’s suitability for different jobs or successful careers (Barrett and Depinet, 1991; González Lorente and Martínez Clares, 2015). In this sense, the European Higher Education Area (EHEA), in its Bologna Declaration, established, as one of its fundamental objectives, the transformation of a teaching-centered education into a learning-centered education, in which the student must acquire a more active role in the whole process (Hannay, 2010; Goodman et al., 2018). A learning-centered education involves the development in the acquisition of competencies by students, expanding, without excluding, the traditional content-based approach (Candela et al., 2006; Christianson et al., 2007; Trullàs et al., 2022). Thus, competency-based learning is based on students learning how to transfer knowledge to reality, so that they know how to use it safely and effectively (Chacko, 2014; Lee and Chiu, 2022). Thus, current curricula are designed so that students acquire various competencies before completing the corresponding undergraduate program.

Martínez-Clares and Echeverría Samanes. (2009) listed the 4 types of competencies, specific and transversal, that each student must acquire during his or her program to perform satisfactorily in his or her professional career (Table 1).

TABLE 1
www.frontiersin.org

TABLE 1. Definition of specific and transversal competencies (adapted from Clares, 2009).

The specific competencies are related to the acquisition of knowledge that the student must achieve by taking a specific subject. These competencies provide them with the technical skills or abilities they need to know for their future professional development, acquiring both the appropriate theoretical knowledge, “to know”, and the ability to know how to apply and put it into practice, “to know how to do”. But, in addition, as important as the acquisition of technical and methodological knowledge, is the acquisition of the so-called transversal competencies (Sá and Serpa, 2018), which are those that enable them to work in a team, to be flexible and adapt to different situations, to be reflective, analytical, respectful, empathetic, supportive, and to develop a commitment to others and promote ethics and values aimed at the pursuit of the common good, in short, “to know how to be there” and “to know how to be”.

This common framework for all undergraduate programs is particularly necessary in the health sciences degrees and more specifically in medicine, where the acquisition of transversal competencies is key, since they are extremely important for their practice, and they will apply throughout their professional career. So much so that medical students are assessed for these competencies before graduation through the Objective and Structured Clinical Evaluation (OSCE), which aims to evaluate skills and aptitudes in specific clinical situations (Carraccio and Englander, 2000). The OSCE is currently implemented in all medical schools in Spain, although its teaching is limited to the last years of the program. We strongly believe that it is important that students work on and acquire these skills from the beginning of their academic career. First, because this competency-based approach, applied from the first years of study when basic subjects predominate, will help students to become aware of the importance of this knowledge in their training process (Finnerty et al., 2010). Secondly, because students must finish their studies having acquired other skills that will be useful for them in the development of their professional performance, and for this, they must not only have the necessary theoretical knowledge, but also train the doctor-patient relationship as soon as possible (Belasen, 2018; Vogel et al., 2018). Good communication with patients is essential in medicine, since the patient’s cooperation and thus the outcome of the treatment depends to a large extent on it (Haskard et al., 2008). However, in addition to being able to convey information truthfully, they must do so with empathy, respect, and cordiality (Bakić-Mirić, 2008). It is increasingly necessary to treat the patient as a human being and not as a disease. There is a growing interest in improving the humanistic aspects of healthcare, which requires collaboration in the training of future professionals (Thibault, 2019).

Therefore, the implementation of competency-based learning in medical education is necessary, but it is also a great challenge. It may seem obvious, but it is important to emphasize, that competency-based education must be well designed, planned, and regulated to be effective (Hawkins et al., 2015; Gruppen et al., 2016; Caverzagie et al., 2017). Establishing a competency-based curriculum involves not only revising the definition of objectives and content, but also developing new strategies and methodological tools that enable its implementation and evaluation (Nigaard et al., 2008). It has been proposed to use a type of teaching based on active methods. These include role-playing, problem-based learning, the case method, and gamification, among others, which allow the development of skills such as critical thinking, analytical, argumentative, and reflective skills, as well as the promotion of oral communication skills (Goodman et al., 2018; Torralba and Doo, 2020). Therefore, one of the challenges that teachers face in current curricula is how to assess the competencies that students are supposed to acquire, for which it is necessary to plan the assessment systems appropriately and that they are consistent with the objectives pursued (Holmboe et al., 2010; Lockyer et al., 2017). Assessment by means of a rubric is one of the valid tools for this process, since the use of specific descriptors makes it possible to define levels of acquisition and, in addition, makes students aware of how their knowledge develops and provides them with an integrated view of the attitudes and actions for improvement that they can apply to new situations in different environments (Boateng et al., 2009; Martiáñez-Ramírez et al., 2022).

For all the above reasons, the aim of this study is to design a competency-based learning plan and an evaluation system that will allow us to adequately measure the level of competency-based progress acquired during practical physiology classes. In this way, the following objectives will be achieved:

- Encourage medical students to acquire specific and transversal competencies through role-plays that reflect application of theoretical concepts in practice and work on the doctor-patient relationship.

- To design an effective rubric that, independent of the evaluating teacher and the clinical case evaluated, allows an unbiased evaluation method that accurately reflects the acquisition of both transversal and specific competencies by the students in medical practice.

2 Methods

2.1 Learning and assessment plan

This experience was developed in the practical classes of the Physiology I subject during the second year of the Medical Degree. This subject is compulsory and consists of 7.5 ECTS credits. The theoretical part of the course covers general physiological phenomena, skeletal muscle and blood physiology, and the physiology of the cardiovascular and respiratory systems are studied. The practical classes are directly related to the theoretical content and are developed in 3-h sessions for 7 days, according to the following program (Table 2). The students were informed about the program of the practical classes and the evaluation system.

TABLE 2
www.frontiersin.org

TABLE 2. Physiology practical classes syllabus.

Throughout the physiology practices, students learn to: 1) measure blood pressure and interpret the obtained measurements, 2) to auscultate and take the pulse, checking whether the results are within physiological parameters, 3) to perform an electrocardiogram and analyze basic data from it, 4) to decipher the main parameters of a complete blood count and perform a blood type determination test, and finally 5) to conduct a spirometry and describe the obtained results. The practical activities are reinforced by the content of the theoretical classes. Each practical session has a maximum of 14 students, lasts for 3 h, and is designed as follows:

- Introduction: The professor provides the necessary theoretical knowledge to review the concepts associated with the clinical test that will be performed, as well as the basic notions to carry out the clinical procedure.

- Learning by doing: Students perform the clinical test in pairs, alternating between the roles of doctor and patient. The importance of doctor-patient communication is always emphasized. The professor directly supervises and advises the students, correcting their procedural approach and their interaction in the different roles. It was suggested that during the practical sessions they should simulate clinical care with their colleagues and put transversal competencies into practice. Thus, they were encouraged to be polite and cordial with the patient, to show interest in the patient´s health and the reason for the consultation, to explain the tests to be performed and finally to explain the results obtained.

- Learning by thinking: Students record the obtained data in their practice notebook and answer a series of questions related to the performance of the clinical procedure and the interpretation of the results obtained.

- Feedback: The practice concludes with a joint session between the professor and the students, reviewing the steps of the clinical procedure and the interpretation of the results. Emphasis is placed on how to convey information to patients.

This protocol provides a structure for students to acquire practical skills in the clinical field while reinforcing the theoretical concepts learned in class. The practical learning approach in pairs and constant feedback from the professor promote the development of technical skills and the improvement of doctor-patient communication.

The assessment consists of performing two evaluation tests: a theoretical test and a practical test. The final practical qualification is obtained based on 75% of the score of the theoretical test and 25% of the score of the practical test. The grading scale was always from zero to ten. The theoretical test consists of multiple-choice questions and students must demonstrate that they have acquired adequate theoretical knowledge (specific competency “to know”). The second test is a practical one, through the simulation of a medical consultation in which students alternate the roles of doctor or patient, thus having to demonstrate their ability to execute/apply what they have learned in class (specific competency “to know how to do”). In this second test, the students are also evaluated on the transversal competencies that we call “communication with the patient” (transversal competency “to know how to be there”) and “dealing with the patient” (transversal competency “to know how to be”). For the evaluation of this second practical test, a rubric was designed to evaluate the students more objectively.

2.2 Assessment of competencies. Design of the rubric

As an evaluation instrument for the practical test, a grading rubric (Figure 1) was designed to evaluate the theoretical knowledge and technical skills of the students, as well as the way they treated, behaved, and showed communication skills with a classmate who played the role of the patient. The rubric was designed according to Bloom’s scale (Bloom, 1956) and taken as a model some previous published rubrics (Moni et al., 2005; Ayhan and Türkyılmaz, 2015).

FIGURE 1
www.frontiersin.org

FIGURE 1. Competency assessment rubric. Tool that we have created to assess the theoretical knowledge, technical skills, and doctor-patient communication during Physiology practical sessions. Specific descriptors allow us to define levels of acquisition in each of the competencies we evaluate. SP, Systolic Pressure; DP, Diastolic Pressure; PP, Pulse Pressure; MP, Medium Pressure.

To assess the different competencies, the rubric was divided into several sections:

- First, the specific competency “to know-how to do” must be evaluated. The theoretical knowledge that the students had learned and their ability to put it into practice were evaluated, since they had to demonstrate that they could perform a procedure and know how to interpret the corresponding results. For this purpose, the evaluator, with the help of the rubric, selected and graded the students, according to their performance, in one of the following practical situations: determination of blood type, interpretation of a hemogram and clotting tests; cardiac auscultation and determination of heart rate; determination of blood pressure; interpretation of an electrocardiogram; interpretation of the results of spirometry and vitalography. The “to know how to do” competency assessment accounted for 50% of the final practical examination score.

- Secondly, during the practical examination, the evaluator had to evaluate the students’ ability to communicate with their partner, who played the role of the patient. This aspect therefore assessed the “to know how to be there” competency, as it considered the student’s ability to explain what test had been or would be performed, for what purpose, what the results obtained were and what the implications of these results were. This section was worth 30% of the final practical examination score.

- Finally, the last section of the rubric was used by the evaluator to assess the competency “to know how to be”, since during the practical classes the importance of this competency had been explained to the students and they had been encouraged to always be polite and cordial with the patient, to be empathetic, to be interested in the patient’s state of health and the reason for the consultation. This last part of the rubric accounted for the remaining 20% of the final grade.

All faculty involved in the teaching of the subject reviewed the content of the rubric and made suggestions about the dimensions and descriptors of the rubric. After incorporating the teachers’ comments, the rubric was considered valid for use. The assessment using the rubric was carried out during the academic years 2019-2020, 2021-2022 and 2022-2023.

Students are given advance information about the percentage of the grade that will be assigned to each part of the rubric. Although the weight of each competency in the final grade varies, students are made aware that they are all important to their overall education. They are explained that the primary objective of physiology practices is to learn how to perform and interpret various clinical tests and relate them to physiological processes, but they are also reminded of the significance we place on developing competencies that enable them to communicate and interact better with patients. Therefore, the “ to know how to do” competency carries more weight in the grading than the “to know how to be” and “to know-how to be there” competencies.”

2.3 Data collection

Competency assessment scores obtained by applying the rubric were collected from three different selected cohorts, these corresponded to students from the 2019-2020 academic year (n = 143), the 2021-2022 academic year (n = 179), and the 2022-2023 academic year (n = 204). Likewise, to evaluate whether there were differences in the grades obtained by the students when only specific competencies were evaluated, but not transversal competencies, the data were also collected for the two previous academic years, the 2017-2018 (n = 148) and the 2018-2019 (n = 141) academic years.

For the analysis of the data, the grades were grouped according to several variables such as: specific evaluations of the sections “to know” and “to know how to do”, “to know how to be there” and “to know how to be” sections, respectively, as well as the evaluating lecturer and the practical case to be evaluated.

2.4 Student perception. Opinion survey

The students’ perceptions of the usefulness and validity of the competency assessment were determined using an opinion survey. The survey consisted of 12 statements with which the students had to indicate their level of agreement (Likert scale with five levels: strongly disagree, disagree, neither agree nor disagree, agree, or strongly agree) and an open-ended question (Table 3; Figure 2). The survey was sent to all students in the 2022-2023 cohort. Of these, 137 students responded voluntarily and anonymously. For data analysis, the percentage of students who selected each level of the Likert scale was calculated.

TABLE 3
www.frontiersin.org

TABLE 3. Questions for the perception survey on the assessment of competencies by students.

FIGURE 2
www.frontiersin.org

FIGURE 2. Results of the student opinion survey. Graph showing the percentages of the different response options (strongly disagree, disagree, neither agree nor disagree, agree, or strongly agree) to the different survey questions by the students.

This study was approved by the Ethics Committee of CEU San Pablo University (678/23/68). An informed consent was obtained from all students who wished to participate in the survey. Participation was voluntary and participants were assured of confidentiality.

2.5 Statistical analysis

Statistical analysis was preformed using the SPSS 27 program (SPSS Institute, France), with a p-value < 0.05 being considered statistically significant differences. The results were expressed as the mean ± standard deviation. The comparative study between the different courses analyzed was carried out using ANOVA for independent samples with subsequent Student-Newman-Keuls post hoc analysis for quantitative variables.

3 Results

3.1 Overall evaluation of the practical lessons of the subject: theoretical and practical knowledge

The average of the final practical qualification of the subject and the breakdown into the tests evaluating the theoretical and practical knowledge, since the academic years 2017-2018 to 2022-2023, are presented in Table 4.

TABLE 4
www.frontiersin.org

TABLE 4. Mean scores obtained in the tests that evaluate the practical block of the subject: theoretical knowledge test, global practical test and final practical examination score in the different academic years. Data are presented as mean ± standard deviation. Data in the same column is compared using ANOVA with subsequent Student-Newman-Keuls post hoc analysis. (*) denotes statistically significant differences for the theorical knowledge test: * between 2017 and 2018 academic year with the others and ** between 2021 and 2022 academic year with the others; (#) denotes statistically significant differences for the global practical test between 2017-2018 and 2018-2019 academic years with the others, but there is no difference between each other. (§) denotes statistically significant differences for the final practical examination score: § between 2017 and 2018 academic year with the others and §§ between 2021 and 2022 academic year with the others. In all cases p < 0.05.

As shown in Table 4, there is a variability in both the grades of the theoretical and practical tests and, therefore, in the final grade among the five academic courses analyzed.

The test that assesses theoretical knowledge and measures the level of mastery of the specific competency “to know” consists of multiple-choice questions drawn from a common bank of more than 100 questions of similar difficulty. This bank was developed by the teachers themselves and is accessible to all evaluators, but not to the students. The evaluators can choose the 30 questions they consider most appropriate, while maintaining a certain number of questions for each subject. Since this test is prepared every year using the same question bank, it is reasonable to assume that any statistically significant differences in scores between academic years are not due to differences in the actual difficulty of the test or to students’ prior knowledge of the test, but to the unique composition of students in each year considered.

However, when analyzing the results of the global practical knowledge test, we observed that there is a repeated pattern, being the results similar between the academic years 2017-2018 and 2018-2019, and between the academic years 2019-2020, 2021-2022 and 2022-2023. For the first two academic years analyzed, students obtained significantly higher scores (p < 0.001) than the three subsequent academic years. This fact is very relevant because in the first two courses analyzed, the rubric was not available as an evaluation method, but it was in the following three courses, when this assessment method was introduced. The fact that, until the 2018-2019 course, the grades in the practical test were higher may be due to the lack of objectivity that existed in the evaluation of the way the students performed the test. Since the rubric evaluation was established, we have observed that the grades are more unbiased and aligned to the development of the students in the subject.

3.2 Results of the evaluation of specific and transversal competencies in the global practical knowledge test

The use of the rubric to evaluate the doctor-patient simulation has allowed us to establish a global practical knowledge test that evaluates specific and transversal competencies in the same test. Thus, we can analyze both the skills developed by the students to perform a given clinical procedure and their ability to interact with the patient.

As is shown in Table 5, in the evaluation of the performance of the clinical procedure (specific competency “to know how to do”), there are no significant differences in the scores along the courses analyzed. In the analysis of the scores for the transversal competency “to know how to be” (“dealing with the patient”), there were also no statistically significant differences between the three last academic years, when this competency was evaluated using the rubric. Only in the analysis of the transversal competency “to know how to be there” (“communication with the patient”), we observe a statistically significant increase in the average grade obtained in the last course analyzed (2022-2023) when compared to previous academic years. This could be due to the consolidation of the rubric method over the years and the teacher’s greater insistence on the importance of developing this competency. This could be a determining factor for the students to become aware of its importance and to try to achieve it in the best possible way. In this sense, we can suggest that in recent years, we have noticed a significant increase in the interest and motivation of students to study careers related to health sciences. This increase can be attributed to the impact that the pandemic has had on society. The health crisis has highlighted the importance of having trained professionals who are committed to the health of people. In addition, the experience of the pandemic has raised a collective awareness of the need for more humane and empathetic treatment of patients. As a result, this increased motivation among students may have a positive impact on the quality of medical care in the future. Future healthcare professionals will be able to focus not only on the technical and scientific aspects, but also on the importance of establishing meaningful human connections with patients. It is crucial to encourage and support this increased interest in health science careers and to promote a comprehensive education that includes social and communication skills.

TABLE 5
www.frontiersin.org

TABLE 5. Mean scores obtained on the rubric assessing specific and transversal competences in the different academic years. Data are presented as mean ± standard deviation. Data in the same column is compared using ANOVA with subsequent Student-Newman-Keuls post hoc analysis * denotes statistically significant differences between the 2022-2023 academic year and the others academic years analyzed for the competence “To know how to be there” (p < 0.05).

The average scores obtained by the students of the Faculty of Medicine over the last three academic years (n = 526) were 7.6 ± 1.91 for the competency “to know how to do”, 7.82 ± 1.72 for the competency “communication with the patient” and 7.8 ± 1.83 for the competency “dealing with the patient”. These passing grades indicate that the student has satisfactorily met the established evaluation criteria. This means that he/she has demonstrated an adequate level of knowledge and skills in relation to the objectives of the physiology practices. It is important to keep in mind that the rubric is not designed to produce high grades, but rather to fairly measure the student’s level of competency. The grades received indicate that the evaluation was balanced and provided an objective measure of the student’s level of competency.

3.3 Results of the evaluation of specific and transversal competencies according to the evaluator

One of the main concerns that we had was whether there was variability in the students’ grades in the practical test depending on the teacher who evaluated them. The differences in the grades obtained could be due either to the variability in the students’ learning or to the subjectivity of the evaluator. Therefore, one of the principal reasons to design the rubric was to have a simple grading system that would help different evaluators to objectively evaluate the students during their practical test.

Table 6 shows the results of the grades of specific and transversal competencies analyzed according to different evaluators in the academic years in which the rubric was used.

TABLE 6
www.frontiersin.org

TABLE 6. Mean scores obtained by the rubric assessing specific and transversal competences according to the evaluators. Data are presented as mean ± standard deviation. Data in the same column is compared using ANOVA with subsequent Student-Newman-Keuls post hoc analysis *denotes statistically significant differences between evaluator 9 with the others evaluators for the competency “To know how to be there”. # denotes statistically significant differences between evaluators 7 and 10 for the competency “To know how to be there”. § denotes statistically significant differences between evaluators 7 and 8 for the competency “To know how to be”. In all cases p < 0.05.

It can be observed that there are no statistically significant differences in the grades of “to know how to do” competency of the students depending on the teacher who evaluated them.

However, in communication and patient deal, there are statistically significant differences in some mean scores among some evaluators. This could be due to a different acquisition of competencies by the students, since there is a positive correlation between the score of the practical case evaluated and the score obtained in communication with the patient (r = 0.666; p < 0.001; n = 526), as well as in dealing with the patient (r = 0.409; p < 0.001; n = 526). The results show a certain tendency among the different scores, such that students who have obtained better results in the tests that evaluate the specific competency “to know how to do” would obtain better results in the scores of the transversal competencies “to know how to be” and “know how to be there”, independently of the teacher who evaluated them. Even so, we cannot rule out the possibility of a student being able to satisfactorily perform a clinical procedure, but then not knowing how to communicate and deal with his or her patient adequately, or vice versa. Nor can we rule out that there is still some subjectivity on the part of certain evaluators who have not strictly adhered to the use of the rubric.

Considering all the above, and bearing in mind certain limitations, we dare to assert that the rubric designed is useful to minimize the variability that may exist in the objectivity of the evaluators and is an effective method that helps to evaluate both specific and transversal competencies.

3.4 Results of the evaluation of specific and transversal competencies according to the practical case evaluated

Table 7 shows the results obtained by analyzing the average grades for “to know how to do”, “to know how to be” and “to know how to be there” competencies according to the different practical cases evaluated, during the academic years in which the rubric was used (2019-2020, 2021-2022, and 2022-2023).

TABLE 7
www.frontiersin.org

TABLE 7. Mean scores obtained by the rubric that evaluates specific and transversal competences according to the different practical cases. Data are presented as mean ± standard deviation. The number of students analyzed is shown in brackets. Data in the same column is compared using ANOVA with subsequent Student-Newman-Keuls post hoc analysis * denotes statistically significant differences between the 2019-2020 academic year and the others academic years analyzed for the competency “To know how to do” in the Blood pressure test (p < 0.05). ECG: Electrocardiogram. Different letters indicate statistically significant differences between the different courses analyzed (p < 0.001). ECG: Electrocardiogram.

The data collected in the table shows that there are no statistically significant differences in the grades obtained for any of the competencies analyzed depending on the clinical case evaluated within the same course or between different courses. Exceptionally, we observe statistically significant differences only in the specific case of the competency “to know how to do” for blood pressure measurement in the 2019-2020 academic year. From these data we can conclude that in general there are no clinical cases that are more difficult than others.

3.5 Students’ perceptions of competency-based assessment

For a better analysis of the students’ responses (Figure 2), we have grouped the 12 statements of the survey into two different categories: those related to the way a competency was assessed (questions 2, 5, 7, 8, 9, 11, and 12) and those related to the usefulness of conducting this type of competency assessment (questions 1, 3, 4, 6, and 10).

Virtually all respondents agreed or strongly agreed with the statements that competency-based assessment was useful in their applied clinical learning (90.5%), has allowed them to integrate theoretical knowledge and clinical skills (89%), and strengthened their ability to perform a good physical examination of the patient (82.2%). In addition, a clear majority also agreed or strongly agreed with the statements that the competency-based assessment has helped them to use information in a rational and applied way (79.2%) and helps them to identify and improve their weaknesses related to clinical practice (75%).

Students agree or strongly agree that competency-based assessment is a good complement to the method based on multiple-choice assessment (86.7%) and that it focuses on content relevant to professional practice (86.7%). Most students agree or strongly agree that competency-based assessment is a fair way of evaluation (63.4%) and has clear rules (70.4%). On the other hand, very few (10.4%) agree or strongly agree that this type of competency-based assessment should be eliminated in the next course. In addition, almost half of the students (48.6%) claim to have enjoyed taking the test.

The survey included an open-ended question so that students could add any comments related to competency assessment. Some of the opinions expressed by the students were the following: “I think it is a very good idea because we do not have any subject or any practice in which we can put our knowledge into action as doctors”, “Very useful for life and our future work”, “I think the oral practice exam is even more important than the theoretical one”, “It seems to me a very good way of evaluation because it makes you face a situation that we will have to face in the future and in which we apply the knowledge we have learned”.

4 Discussion

The fact of accumulating knowledge does not necessarily imply being competent. Competence, in professional terms, is measured by the degree of resolution aptitude that people show in the development of their professional activity (Kukkonen et al., 2020). The relevance of all of this is visible in our daily work. It is common to associate good work performance with the ability to solve a problem; on the contrary, if they are not able to solve it, they can be labeled as incompetent and not knowing how to do their job well (García et al., 2020). In neither case are people’s intelligence, knowledge or preparation being judged, but rather how they managed to use it and apply all their training in the performance of their task. Education in competencies is education in knowledge, there is no other way to educate, but it implies a redefinition of curricula, orienting them towards the education of people in a broader sense (Timmerberg et al., 2022). In this way, we will guarantee our students a personal and intellectual development that will lead them to develop a complete education in line with the demands of the labor market (French et al., 2020). Comprehensive education is articulated based on conceptual, procedural, and attitudinal knowledge (Oandasan et al., 2020).

In the specific case of medical education, it is increasingly necessary that the efforts of the professors be focused on offering the students a complete, qualified, and competent training, where they are not only taught what they must know and be able to do, but also encouraged to develop more humanistic attitudes and to train the doctor-patient relationship (Weiss et al., 2019; Forsey et al., 2021). To be able to transmit information truthfully, but also with empathy, respect, and cordiality.

All of this leads us to promote strategies aimed at the acquisition of competencies by students, expanding but not excluding the traditional content-based path. However, although nobody doubts the above, it is necessary to emphasize the importance of a good development of the methods of learning these competencies, as well as the procedures to evaluate their level of acquisition. This requires evaluation systems that are in line with the objectives and that are known to the students, so that they know what is expected from them and are aware of their learning process, development, and personal and professional growth (Chimea, 2020).

Our main purpose in this work has been to instruct our students not only in the development of specific competencies, but also in the development of transversal competencies, specifically the doctor-patient relationship, from the beginning of their medical training and not to postpone it to higher courses (Daniel et al., 2021; Krishnasamy et al., 2022). To this end, we have used active methodologies based on simulation environments and role playing. In this way, students put into practice the knowledge acquired in different clinical situations and worked on the communication and treatment of the patient by informing him/her about the test procedure to be performed and/or transmitting the results of the test itself. The development of competency-based learning strategies must necessarily be accompanied by the design of specific and appropriate evaluation methods to assess the degree of acquisition of these competencies (Min Simpkins et al., 2019). The results further suggest that we have developed a rubric that allows us to appropriately measure the acquisition of the knowledge necessary to perform a given clinical test, as well as the acquisition of the ability to relate to the patient, regardless of the teacher performing the assessment or the clinical test in question.

The rubric provides students with measurable evidence of their learning process, achievement, and development of competencies, and is considered as a valid and a real alternative to traditional evaluation methods (Virk et al., 2020). Moreover, the use of the rubric increases transparency and reduces subjectivity in the evaluation process among the different evaluator, and therefore we can affirm that it makes the process fairer. Its inclusion in the evaluation of communication skills and patient care promotes the acquisition of essential competencies for the professional development of future physicians from the first courses.

Thus, we can consider competency assessment using the rubric, both teachers and students, as a valid complement to written exams (“know how”) and allows making students aware of the importance of striving to develop competencies (“to know how to do”, “to know how to be there” and “to know how to be”) to achieve a comprehensive training that allows them to develop as better professionals.

Data availability statement

The raw data supporting the conclusion of this article will be made available by the authors, without undue reservation.

Ethics statement

The studies involving human participants were reviewed and approved by the Comité de Ética de Investigación de la Universidad San Pablo-CEU. The patients/participants provided their written informed consent to participate in this study.

Author contributions

IS-V conceive and designed the competency-based learning plan and its evaluation system. IS-V analyzed the data. AT and IS-V wrote the main manuscript and prepared figures and table. All authors contributed to the article and approved the submitted version.

Acknowledgments

The authors would like to thank Universidad San Pablo CEU and the Cátedra de Educación Médica y humanización “Nemesio Diez” for funding the open access. We would also like to thank our students. Without them, none of this would make sense.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Ayhan, Ü., and Türkyılmaz, M. U. (2015). Key of language assessment: Rubrics and rubric design. Int. J. Lang. Linguistics 2 (2), 82–92.

Google Scholar

Bakić-Mirić, N. M. (2008). Successful doctor-patient communication and rapport building as the key skills of medical practice. Facta Univers 15 (2), 74–79.

Google Scholar

Barrett, G. V., and Depinet, R. L. (1991). A reconsideration of testing for competence rather than for intelligence. Am. Psychol. 46 (10), 1012–1024. doi:10.1037/0003-066X.46.10.1012

PubMed Abstract | CrossRef Full Text | Google Scholar

Belasen, A. (2018). Doctor-patient communication: A review and a rationale for using an assessment framework. J. health Organ. Manag. 32 (7), 891–907. doi:10.1108/JHOM-10-2017-0262

PubMed Abstract | CrossRef Full Text | Google Scholar

Bloom, B. S. (1956). in Taxonomy of educational objectives: Handbook I. Editor D. McKay (New York: Cognitive domain).

Google Scholar

Boateng, B. A., Bass, L. D., Blaszak, R. T., and Farrar, H. C. (2009). The development of a competency-based assessment rubric to measure resident milestones. J. graduate Med. Educ. 1 (1), 45–48. doi:10.4300/01.01.0008

CrossRef Full Text | Google Scholar

Candela, L. D. L., Dalley, K., and Benzel-Lindley, J. (2006). A case for learning-centered curricula. J. Nurs. Educ. 45 (2), 59–66. doi:10.3928/01484834-20060201-04

PubMed Abstract | CrossRef Full Text | Google Scholar

Carraccio, C., and Englander, R. (2000). The objective structured clinical examination: A step in the direction of competency-based evaluation. Archives Pediatr. Adolesc. Med. 154 (7), 736–741. doi:10.1001/archpedi.154.7.736

CrossRef Full Text | Google Scholar

Caverzagie, K. J., Nousiainen, M. T., Ferguson, P. C., Ten Cate, O., Ross, S., Harris, K. A., et al. (2017). Overarching challenges to the implementation of competency-based medical education. Med. Teach. 39 (6), 588–593. doi:10.1080/0142159X.2017.1315075

PubMed Abstract | CrossRef Full Text | Google Scholar

Chacko, T. V. (2014). Moving toward competency-based education: Challenges and the way forward. Archives Med. Health Sci. 2 (2), 247–253. doi:10.4103/2321-4848.144365

CrossRef Full Text | Google Scholar

Chimea, T. K. (2020). Assessment of clinical competence in competency-based education. Can. J. Dent. Hyg. 54 (2), 83–91.

PubMed Abstract | Google Scholar

Christianson, C. E., McBride, R. B., Vari, R. C., Olson, L. E., and Wilson, H. D. (2007). From traditional to patient-centered learning: Curriculum change as an intervention for changing institutional culture and promoting professionalism in undergraduate medical education. Acad. Med. 82 (11), 1079–1088. doi:10.1097/ACM.0b013e3181574a62

PubMed Abstract | CrossRef Full Text | Google Scholar

Clares, P. M. (2009). Formación basada en competencias. Rev. Investig. Educ. 27 (1), 125–147.

Google Scholar

Daniel, M. M., Morrison, G., Hauer, K. E., Pock, A., Seibert, C., Amiel, J., et al. (2021). Strategies from 11 U.S. Medical schools for integrating basic science into core clerkships. Acad. Med. J. Assoc. Am. Med. Coll. 96 (8), 1125–1130. doi:10.1097/ACM.0000000000003908

CrossRef Full Text | Google Scholar

Finnerty, E. P., Chauvin, S., Bonaminio, G., Andrews, M., Carroll, R. G., and Pangaro, L. N. (2010). Flexner revisited: The role and value of the basic sciences in medical education. Acad. Med. 85 (2), 349–355. doi:10.1097/ACM.0b013e3181c88b09

PubMed Abstract | CrossRef Full Text | Google Scholar

Forsey, J. N., Ng, S., Rowland, P., Freeman, R., Li, C., and Woods, N. N. (2021). The basic science of patient-physician communication: A critical scoping review. Acad. Med. J. Assoc. Am. Med. Coll. 96 (11S), S109–S118. doi:10.1097/ACM.0000000000004323

CrossRef Full Text | Google Scholar

French, M. B., Choate, J. K., Zubek, J., Bryner, R. W., Johnson, K. M. S., and Luttrell, M. J. (2020). Professional skills for physiology majors: Defining and refining. Adv. physiology Educ. 44 (4), 653–657. doi:10.1152/advan.00178.2019

CrossRef Full Text | Google Scholar

García, P. C. A. S., Pinto-Carral, A., Villorejo, J. S., and Marqués-Sánchez, P. (2020). Nurse manager core competencies: A proposal in the Spanish health system. Int. J. Environ. Res. Public Health 17 (9), 3173. doi:10.3390/ijerph17093173

PubMed Abstract | CrossRef Full Text | Google Scholar

González Lorente, C., and Martínez Clares, P. (2015). Expectativas de futuro laboral del universitario de hoy: Un estudio internacional. Rev. Investig. Educ. 34 (1), 167–183. doi:10.6018/rie.34.1.232071

CrossRef Full Text | Google Scholar

Goodman, B. M., Barker, M. K., and Cooke, J. E. (2018). Best practices in active and student-centered learning in physiology classes. Adv. Physiol. Educ. 42 (3), 417–423. doi:10.1152/advan.00064.2018

PubMed Abstract | CrossRef Full Text | Google Scholar

Gruppen, L. D., Burkhardt, J. C., Fitzgerald, J. T., Funnell, M., Haftel, H. M., Lypson, M. L., et al. (2016). Competency-based education: Programme design and challenges to implementation. Med. Educ. 50 (5), 532–539. doi:10.1111/medu.12977

PubMed Abstract | CrossRef Full Text | Google Scholar

Hannay, M. K. (2010). Student-focused strategies for the modern classroom. J. Instr. Pedagogies 2, 1–16.

Google Scholar

Haskard, K. B., Williams, S. L., DiMatteo, M. R., Rosenthal, R., White, M. K., and Goldstein, M. G. (2008). Physician and patient communication training in primary care: Effects on participation and satisfaction. Health Psychol. 27 (5), 513–522. doi:10.1037/0278-6133.27.5.513

PubMed Abstract | CrossRef Full Text | Google Scholar

Hawkins, R. E., Welcher, C. M., Holmboe, E. S., Kirk, L. M., Norcini, J. J., Simons, K. B., et al. (2015). Implementation of competency-based medical education: Are we addressing the concerns and challenges? Med. Educ. 49 (11), 1086–1102. doi:10.1111/medu.12831

PubMed Abstract | CrossRef Full Text | Google Scholar

Holmboe, E. S., Sherbino, J., Long, D. M., Swing, S. R., and Frank, J. R. (2010). The role of assessment in competency-based medical education. Med. Teach. 32 (8), 676–682. doi:10.3109/0142159X.2010.500704

PubMed Abstract | CrossRef Full Text | Google Scholar

Krishnasamy, N. H., Hasamnis, A., and Patil, S. (2022). Developing professional identity among undergraduate medical students in a competency-based curriculum: Educators' perspective. J. Educ. health Promot. 11, 361. doi:10.4103/jehp.jehp_329_22

PubMed Abstract | CrossRef Full Text | Google Scholar

Kukkonen, P. L. K. L., Leino-Kilpi, H., Koskinen, S., Salminen, L., and Strandell-Laine, C. (2020). Nurse managers' perceptions of the competence of newly graduated nurses: A scoping review. J. Nurs. Manag. 28 (1), 4–16. doi:10.1111/jonm.12891

PubMed Abstract | CrossRef Full Text | Google Scholar

Lee, G. B., and Chiu, A. M. (2022). Assessment and feedback methods in competency-based medical education. Ann. allergy, asthma & Immunol. 128 (3), 256–262. doi:10.1016/j.anai.2021.12.010

CrossRef Full Text | Google Scholar

Lockyer, J. C., Carraccio, C., Chan, M. K., Hart, D., Smee, S., Touchie, C., et al. (2017). Core principles of assessment in competency-based medical education. Med. Teach. 39 (6), 609–616. doi:10.1080/0142159X.2017.1315082

PubMed Abstract | CrossRef Full Text | Google Scholar

Martiáñez-Ramírez, N. L. G. B. G., Pineda-Galán, C., Rodríguez-Bailón, M., and Romero-Galisteo, R. P. (2022). Competence assessment rubric in the physiotherapy practicum. PloS one 17 (2), 0264120. doi:10.1371/journal.pone.0264120

CrossRef Full Text | Google Scholar

Martínez Clares, P., and Echevarría Samanes, B. (2009). Formación basada en competencias. Rev. Investig. Educ. 27 (1), 125–147.

Google Scholar

Min Simpkins, A. A. E., Koch, B., Spear-Ellinwood, K., and St John, P. (2019). A developmental assessment of clinical reasoning in preclinical medical education. Med. Educ. online 24 (1), 1591257. doi:10.1080/10872981.2019.1591257

PubMed Abstract | CrossRef Full Text | Google Scholar

Moni, R. W., Beswick, E., and Moni, K. B. (2005). Using student feedback to construct an assessment rubric for a concept map in physiology. Adv. physiology Educ. 29 (4), 197–203. doi:10.1152/advan.00066.2004

CrossRef Full Text | Google Scholar

Nigaard, C. H., Højlt, T., and Hermansen, M. (2008). Learning-based curriculum development. High. Educ. 55, 33–50. doi:10.1007/s10734-006-9036-2

CrossRef Full Text | Google Scholar

Oandasan, I. M., Martin, L., McGuire, M., and Zorzi, R. (2020). Twelve tips for improvement-oriented evaluation of competency-based medical education. Med. Teach. 42 (3), 272–277. doi:10.1080/0142159X.2018.1552783

PubMed Abstract | CrossRef Full Text | Google Scholar

Sá, M. J., and Serpa, S. (2018). Transversal competences: Their importance and learning processes by higher education students. Educ. Sci. 8 (3), 126–138. doi:10.3390/educsci8030126

CrossRef Full Text | Google Scholar

Thibault, G. E. (2019). Humanism in medicine: What does it mean and why is it more important than ever? Acad. Med. 94 (8), 1074–1077. doi:10.1097/ACM.0000000000002796

PubMed Abstract | CrossRef Full Text | Google Scholar

Timmerberg, J. F., Chesbro, S. B., Jensen, G. M., Dole, R. L., and Jette, D. U. (2022). Competency-based education and practice in physical therapy: It's time to act. Phys. Ther. 102 (5), 018. doi:10.1093/ptj/pzac018

CrossRef Full Text | Google Scholar

Torralba, K. D., and Doo, L. (2020). Active learning strategies to improve progression from knowledge to action. Rheum. Dis. Clin. 46 (1), 1–19. doi:10.1016/j.rdc.2019.09.001

CrossRef Full Text | Google Scholar

Trullàs, B. C., Blay, C., Sarri, E., and Pujol, R. (2022). Effectiveness of problem-based learning methodology in undergraduate medical education: A scoping review. BMC Med. Educ. 22 (1), 104. doi:10.1186/s12909-022-03154-8

PubMed Abstract | CrossRef Full Text | Google Scholar

Virk, A. J., Joshi, A., Mahajan, R., and Singh, T. (2020). The power of subjectivity in competency-based assessment. J. Postgrad. Med. 66 (4), 200–205. doi:10.4103/jpgm.JPGM_591_20

PubMed Abstract | CrossRef Full Text | Google Scholar

Vogel, D. M., Meyer, M., and Harendza, S. (2018). Verbal and non-verbal communication skills including empathy during history taking of undergraduate medical students. BMC Med. Educ. 18 (1), 157–164. doi:10.1186/s12909-018-1260-9

PubMed Abstract | CrossRef Full Text | Google Scholar

Weiss, T., Weiss, T., and Swede, M. (2019). Transforming preprofessional health education through relationship-centered care and narrative medicine. J. Teach. Learn. Med. 31 (2), 222–233. doi:10.1080/10401334.2016.1159566

CrossRef Full Text | Google Scholar

Keywords: competency-based learning, humanization, learning by doing, doctor-patient communication, role playing, rubric, student experiences, transversal and specific competencies

Citation: Tutor AS, Escudero E, Nogal Ávila Md, Aranda JF, Torres H, Yague JG, Borrego MJ, Muñoz Ú, Sádaba MC and Sánchez-Vera I (2023) Learning and assessment strategies to develop specific and transversal competencies for a humanized medical education. Front. Physiol. 14:1212031. doi: 10.3389/fphys.2023.1212031

Received: 25 April 2023; Accepted: 27 June 2023;
Published: 10 July 2023.

Edited by:

Luis Monteiro Rodrigues, Universidade Lusófona, Portugal

Reviewed by:

Yasser Mohamed El-Wazir, Suez Canal University, Egypt
Patsie Polly, University of New South Wales, Australia
Jorge M. A. Oliveira, University of Porto, Portugal

Copyright © 2023 Tutor, Escudero, Nogal Ávila, Aranda, Torres, Yague, Borrego, Muñoz, Sádaba and Sánchez-Vera. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Isabel Sánchez-Vera, aXNhbnZlckBjZXUuZXM=

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.