Skip to main content

SYSTEMATIC REVIEW article

Front. Med., 02 October 2024
Sec. Healthcare Professions Education
This article is part of the Research Topic Opportunities and Challenges of Interprofessional Collaboration and Education - Volume II View all 7 articles

Tools for self- or peer-assessment of interprofessional competencies of healthcare students: a scoping review

Sharon Brownie\n,\n\n
\nSharon Brownie1 ,2 * Jia Rong Yap\nJia Rong Yap3 Denise Blanchard\nDenise Blanchard4 Issac Amankwaa\n,\nIssac Amankwaa5 ,6 Amy Pearce\n,\nAmy Pearce3 ,7 Kesava Kovanur Sampath\n,\nKesava Kovanur Sampath3 ,6 Ann-Rong Yan\nAnn-Rong Yan6 Patrea Andersen\n,\nPatrea Andersen3 ,8 Patrick Broman\n,\n,\nPatrick Broman1 ,2 ,3
  • 1School of Health Sciences, Swinburne University of Technology, Hawthorn, VIC, Australia
  • 2School of Medicine and Dentistry, Griffith Health, Griffith University, Southport, QLD, Australia
  • 3Centre for Health and Social Practice and Centre for Sports Science and Human Performance, Waikato Institute of Technology, Hamilton, Waikato, New Zealand
  • 4College of Health, Medicine and Wellbeing, The University of Newcastle, Callaghan, NSW, Australia
  • 5Faculty of Health and Environmental Sciences, Auckland University of Technology, Auckland, Auckland, New Zealand
  • 6School of Health Sciences, Faculty of Health, University of Canberra, Canberra, ACT, Australia
  • 7Division of Health, University of Waikato, Hamilton, Waikato, New Zealand
  • 8School of Nursing, Midwifery and Paramedicine, University of the Sunshine Coast, Maroochydore, QLD, Australia

Introduction: Healthcare professionals are expected to demonstrate competence in the effective management of chronic disease and long-term health and rehabilitation needs. Care provided by groups of collaborating professionals is currently well recognized as a more effective way to support people living with these conditions than routine, single-profession clinical encounters. Clinical learning contexts provide hands-on opportunities to develop the interprofessional competencies essential for health professional students in training; however, suitable assessment tools are needed to support student attainment of interprofessional competencies with self-assessment espoused as an important component of learning.

Method: A structured approach was taken to locate and review existing tools used for the self-assessment and peer assessment of students’ competencies relevant to interprofessional practice.

Results: A range of self- and/or peer assessment approaches are available, including formally structured tools and less structured processes inclusive of focus groups and reflection.

Discussion: The identified tools will usefully inform discussion regarding interprofessional competency self- and peer assessment options by healthcare students participating in a broad range of clinical learning contexts.

Conclusion: Self- and/or peer assessment is a useful approach for those seeking to effectively enhance interprofessional learning and measure the attainment of related competencies.

1 Introduction

An increasing focus on interprofessional education is needed as student health professionals prepare for a context of increasing health complexity, non-communicable disease, co-morbid conditions, and aging populations (1, 2). Programs of care provided by professionals working together are currently well recognized as a more effective way to support people living with these conditions than routine, single-profession clinical encounters. Patients increasingly expect a broader and more coordinated approach to their care (2, 3). Recognition of the need for team-based collaborative care and interprofessional education is not new as shown by documents such as the World Health Organization (2010) Framework for Action on Interprofessional Education and Collaborative Practice (4). Subsequently, curriculum content focused on the development of interprofessional competencies is an increasingly expected component of health professional education (5, 6). Interprofessional competencies form the basis of safety, quality, and patient-centeredness in team collaboration contexts (7, 8). These competencies include team participation, leadership, and communication (9). Interprofessional competence also includes soft skills such as attitudes, values, ethics, and teamwork, facilitating difficult conversations, multi-party communication, and trust building (8, 10, 11). Interprofessional competence is required for effective modern healthcare practice but all too often, various barriers get in the way of teaching interprofessional (IP) competencies, given that training usually involves professionals working in isolation using their own discipline knowledge base (1, 12).

Clinical learning environments, or contexts in which health professional programs are taught and practice placements occur, provide hands-on opportunities to support student attainment of IP competencies. Best practices in clinical education involve continuous feedback as a critical link between teaching and assessment and essential in supporting the educational process (13, 14) with self- and peer assessment espoused and regularly used as an important component of the learning sequence (15, 16). Twenty years ago, Ward et al. (2002) reported that reflection on practice using self- and peer assessment is not without difficulties, raising concerns such as issues in objectivity and reliability of students assessing their own performance (17), and debates have persisted since that time (18). Despite these concerns, self-assessment is widely implemented as an educational learning process (16). In the face of concerns, suitably validated self- and peer assessment tools are needed to guide best practices, complement faculty assessment processes, and effectively maximize learning (19).

Previous reviews identifying interprofessional assessment tools for use with prelicensure students have focused on post-placement or post-intervention assessment of IP competency (20) or the identification of tools for use by faculty in the assessment of student IP development (21). This inquiry aimed to locate assessment tools and assessment processes used by prelicensure healthcare students for the self- and peer assessment of IP competency attainment in clinical learning contexts, including the potential for use within an interprofessional student-led clinic. Student-led clinics (SLCs) are a unique option for the provision of practice placements in health professional programs (22). They are used with increasing frequency to enhance the opportunity and experience for prelicensure students in hands-on practice, especially in primary healthcare settings, while also providing benefits to service users and communities (2224). SLCs May involve students from single professions (22) or May be interprofessional in nature (25, 26). Within both general clinical learning contexts and SLCs, tools May be used to assess either individuals or whole teams in interprofessional competencies.

This study sought to understand what assessment tools and self/peer assessment processes have been used by prelicensure healthcare students during interprofessional self-assessment and peer assessment processes in clinical learning environments with two or more health professionals working together. In developing this search, we noted that “tools,” “techniques,” instruments,” and “scales” are frequent terms used interchangeably in the literature (2729). Definitions are closely aligned and often contradictory (30, 31). For this review, the term ‘tool’ is reported for consistency. Consistent with our research question, we also report processes that did not include the utilization of formally developed ‘tools’ but also other means such as self- or peer reflection and focus group discussions to measure, assess, or reflect on interprofessional competency development.

The inquiry focused on student self- and peer assessment versus assessment undertaken by teaching faculty and on the self-assessment of interprofessional competencies versus profession-specific competencies. The review aimed to answer the following questions:

• What tools and self−/peer assessment processes have been used by prelicensure healthcare students to undertake self- and/or peer assessment of interprofessional competencies in an interprofessional clinical learning context (contexts in which health professional programs are taught and practice placements occur) with two or more health professions working together?

2 Method

2.1 Reporting guideline

A scoping review was considered most appropriate for investigating the research question as this topic has not yet been comprehensively reviewed. In such instances, scoping reviews are suitable to provide a general overview of available evidence (assessment tools) as a precursor to more detailed inquiry (32). A scholarly approach was undertaken in conducting the review using the Preferred Reporting Items for Systematic Reviews and Meta-Analyses Extension for Scoping Reviews (PRISMA-ScR) statement (33).

2.2 Eligibility criteria

This review sought primary studies using qualitative, quantitative, or mixed methods for assessing interprofessional competencies. Specifically, we searched for studies involving healthcare students (from two or more professions working together) at any level of study, participating in interprofessional education activities, and utilizing tools to self-assess interprofessional competence or peer assess other students. The search focused on prelicensure students. Publications in which participants included registered health professionals and those with initiatives to maintain registration or undertake continuing professional development were excluded. Studies that assessed IPE programs more broadly and in which tools were used for the primary purpose of program evaluation, as opposed to specifically assessing student IPE competencies as a result of such programs, were excluded. The selection criteria are summarized in Table 1.

Table 1
www.frontiersin.org

Table 1. Inclusion and exclusion criteria.

2.3 Information sources

The literature search was completed in May 2023 and updated in November 2023 followed by analysis and write-up. Four electronic databases, ProQuest, ERIC, Medline, and Embase, were searched for literature published in the 25 years preceding the search date. By focusing on the last 25 years, the review aligns with the transfer in various nations of hospital-based education to university-based education and captures the most relevant and impactful developments in the field of Interprofessional Education and Collaborative Practice (IPECP). This approach allowed us to concentrate on the period during which these concepts gained significant traction, thereby providing a more focused and pertinent analysis.

2.4 Search process

The search strategy was guided by the research question and the inclusion/exclusion criteria, focusing on three broad concepts: healthcare student, peer- and self-assessment, and interprofessional competence, with refinement through MeSH headings in Medline. The initial search in ERIC used the following keywords: [(Pre-registration OR Pre-licensure) AND (Healthcare student OR Healthcare student) AND (postgraduate OR undergraduate) AND (Evaluate OR Assessment OR assessing OR assess OR outcome OR outcomes OR examin* OR evaluate) OR (measurement OR measure OR measuring) AND (Competenc* OR Competent) AND (interprofession*) AND tools]. The search strategy was then tailored to each database accordingly. Google Scholar was specifically used to search for gray and narrative literature that might have been missed in the focused search as well as to explore reference lists of relevant primary papers in the database search.

2.5 Study selection

Search results were imported into Covidence® (34), an online software for review data management and screening, which automatically removed duplicates. Initial screening of the titles and abstracts was conducted by two sets of independent reviewers. Disagreements regarding paper inclusion were resolved by discussion between a third and fourth reviewer.

Full texts of included studies were then reviewed by two sets of independent reviewers. Discrepancies and conflicts were resolved by a third reviewer.

2.6 Data extraction

Two reviewers independently extracted relevant data from the included studies via Covidence for review and discussion by all authors. This information encompassed the following parts: the characteristics of studies (publication year, country, study design, sample population, and size), participant features (student professional field and level of study), and characteristics relating to the intervention, control, and outcome measures (IPE, interprofessional competency, and self/peer assessment tools). Any conflicts that arose between the reviewers were resolved by consensus.

Focused effort with significant rereading and team discussion was needed to locate studies directly relevant to the research questions. This was because significant literature was identified where students undertook self-assessment activities using published scales; however, on close examination, the student self-assessment data were used to inform tutor evaluation of the effectiveness of the IPE program or intervention rather than for the students’ personal assessment, discussion, and reflection. Examples include (3537). Articles that used student self-assessment data purely to inform program evaluations were excluded in the review process because this review directly related to the question ‘what tools and self/peer assessment processes have been used by prelicensure healthcare students to undertake self- and/or peer assessment of interprofessional competencies in an interprofessional clinical learning context with two or more health professions working together?’ Some studies had a dual-purpose use of the student self-evaluation data—to inform both student self-evaluation and program evaluation. If data were available for student self-assessment and/or reflection, the study was identified as relevant to this review.

3 Results

Twenty studies were identified of direct relevance to the review question (see Figure 1).

Figure 1
www.frontiersin.org

Figure 1. Literature search and PRISMA diagram.

3.1 Characteristics of included studies

Table 2 provides a summary of the characteristics of each study selected for inclusion in this review. Studies were identified across a 25-year timeframe from 2009 and involved quantitative, qualitative, and mixed-methods approaches. A wide range of health professions were reported in the selected studies with nursing and pharmacy the most frequently noted. Studies originated from the United States, Canada, Australia, and the United Kingdom, with the highest number (14 or 70%) having been published in the United States.

Table 2
www.frontiersin.org

Table 2. Characteristics of included studies.

3.2 Analysis of included studies

For the purposes of this review, assessment tools identified May have been used for either self- or peer assessment, with results having been provided to students for the purposes of learning assessment, rather than being used by educators or researchers for program or course evaluation. Table 3 lists each study and provides information about the number of participants, the intervention (IPE learning activity), the participating student population, and the specific assessment tools used. Note that where assessment was undertaken by instructors or faculty in conjunction with self-assessment or peer assessment in a given study, these tools are not listed. For example, Begley et al. (2019) also used the Creighton Interprofessional Collaborative Evaluation (C-ICE) instrument, a “25-point dichotomous tool in which the evaluator awards one point if the interprofessional team demonstrates competency in a 2 specific area, or no point for failure to do so” (p. 477). Because this tool involves evaluator 3 (not self- or peer-) assessment, this tool is not listed or considered further here (38).

Table 3
www.frontiersin.org

Table 3. Interprofessional education self/peer assessment tools.

The 12 specific tools in Table 3 have been used for self-and/or peer assessment of interprofessional competencies across the twenty included studies are shown in Table 4.

Table 4
www.frontiersin.org

Table 4. Tools used for self or peer assessment of interprofessional competencies.

The origins of the frequently used tool can be found in USA which has a strong history of formally established interprofessional learning collaboratives. For example, the ISVS Scale was developed by the Minnesota-based National Centre for Interprofessional Education and Practice (39), the IPEC scale was developed by the Washington-DC-based Interprofessional Education Collaborative. It appears that US educators have the autonomy to choose and utilize various tools or to construct their own approaches. The UK hosts CAIPE—the Centre for the Advancement of Interprofessional Practice and Education established in 1987 to drive interprofessional practice in health (40). However, UK educational providers appear to have less autonomy as UK-based regulators mandate the actual competencies, which must be addressed by each profession. We can only speculate that this May be why only one UK-based manuscript appeared in this search.

The 12 assessment tools vary in different ways, although 11 of the 12 tools are quantitative, Likert-scale measures, with the exception being the Description of a Meaningful Interprofessional Learning Situation Tool developed by Dubouloz et al. (2010) to capture students’ perspectives qualitatively, via open-ended questions. This is the only specific qualitative tool used (41); however, other studies also adopted less structured approaches to self and peer assessment, such as the use of focus groups or written reflection tasks. Table 4 only includes the 14 studies in which formal tools were utilized. Utilization of tools was most frequently reported in mixed-methods studies, in conjunction with a more structured, quantitative approach utilizing a scaled tool (38, 4244). Two studies adopted a solely qualitative approach, with students undertaking self-assessment via reflective written questionnaire/open-ended survey (post-test-only and pre-and-post, respectively) (45, 46).

Among quantitative approaches, the most frequently used tool was the IPEC (43, 44, 47, 48). This is a 5-point Likert scale tool based on the well-known core competency statements developed by the Interprofessional Education Collaborative (IPEC, 2011), a U.S. collaboration involving peak bodies from six health disciplines. An early 42-item scale includes 8 to 11 items for each of the four key domains in the statement (values and ethics, roles, and responsibilities, interprofessional communication, and teams and teamwork), although Nieuwoudt et al. (2021) used a shortened 16-item scale and Porter et al. (2020) modified the scale to use ‘the team’, instead of ‘I’ for each of the competencies (43, 49). Note that an updated version 3 of the IPEC competency standards was released in late 2023 shortly after the conclusion of the search process associated with this manuscript (50). The updated version is available as a resource to inform future studies.

Three of the included studies used ISVS for students to self-assess attitudes, values, and beliefs about the value of interprofessional socialization (44, 51, 52). The second most used tool was the Readiness for Interprofessional Learning Scale (RIPLS), used by Ref. (41, 5355). However, it is important to note that in each of these cases, this tool was used alongside one or more other tools for assessing competencies. In each study including RIPLS the decision to include it is not explained. While a valid and reliable tool, RIPLS (56), was not designed to be an outcome or impact measure. It is designed to measure attitudes toward IPE before starting an IPE intervention. However, it is appropriate to include studies, which have utilized the RIPLS scale on the basis that this scale measures attitudes and values as regards interprofessional educational activities. Collaborative attitudes and values, including the attitude and openness to follow leaders within a team, are important interprofessional competencies (10, 11, 57), and students’ awareness of their own situation is an important part of interprofessional learning. In considering the decision to include studies using the RPILS scale in the findings of this search, it is important to reflect on the variance among the 12 tools highlighted in Table 4. Assessment is a multivariate process. One size does not fit all. Thus, a selection of different types of tools and processes for differing settings is both valid and useful.

Considering differences is also important to differentiate between the most common type of tool, which measures individual competencies (whether for oneself or one’s peers), and those which measure competencies overall, for a team. Most located studies used individual and personal scales, but there were some examples of scales or tools which measured overall team functioning, skills, or approaches. These include the CATME used in Ref. (58) originated in engineering, and which involves individuals assessing self- and team-member contributions to a team, and the Teamwork Assessment Scale used in Ref. (53), which assesses team functioning in a given situation (items include ‘the team roles were distinct without ambiguity’, for example). The CASCD scale used by Ref. (59) measures perceptions of team interaction and satisfaction with decision-making and is thus also more situational and team functioning focused than a scale of individual skills, knowledge, or experience.

Other studies used (either solely or alongside named tools) in-house constructed Likert-scale instruments not listed in Table 3 (51, 57, 60). Validation for these tools, particularly a detailed description of their psychometric properties, was typically lacking (61). Likert-scale ranked approaches were typically used in a pre- and post-design before and after the intervention, but there are also examples of retrospective, post-then-predesign where participants recalled prior knowledge after the fact (48) and ICCAS used by Vyas et al. (2021) is designed to be completed only once, rating abilities after training and also as recalled previously (62). Overall, there was significant variability in the approaches to self- and peer- assessment undertaken by students in these contexts and in the tools and processes used.

4 Discussion

Effective assessment should be designed in a multifaceted manner and include a variety of formative and summative assessment activities and continuous learner feedback with each assessment activity designed to build, test, and affirm learner capability and expand understanding. Using more than one assessment type helps give students a range of ways to demonstrate what they have learned and what still needs to be learned (63). Among the wide milieu to be learned by student health professionals, interprofessional insight and practice capabilities are increasingly important as populations age and levels of chronic and complex care priorities increase (2).

A recent review reported results of a search designed as a resource of interprofessional assessment tools used by faculty (21). This search was designed to complement this study by locating and providing a pointer to tools and processes available for student self-assessment and peer assessment of interprofessional understanding and capability. As highlighted, the search identified studies utilizing and reporting formally developed IP self- and peer assessment tools along with other studies reporting processes such as focus groups and reflection—the benefit being an identification of a broad range of resources that can used to engage with students and their peers and enhance IP related learning among health professional students during their learning experiences. A significant benefit of self- and peer assessment is the extent to which these processes increase student understanding of their current capabilities and learning needs (58, 64).

4.1 Self- and peer assessment tools

Despite concerns in the literature about the objectivity and reliability of students assessing their own performance and/or that of their classmates (17, 18), peer and self-assessments have been shown to significantly contribute to the expansion of student capability and positive learning outcomes (6466). The argument that self-assessment May be unreliable, inflated, and/or biased can be mitigated by including others (for example, peers, colleagues, and clients) in the assessment of self (67). Thus, the review has searched for examples using both self- and peer assessment tools and processes with both noted as being complementary to each other (65).

Documented benefits of self-assessment include the growth that occurs when students learn how to assess their own competencies and/or those of their peers. This includes increased ‘deep-level’ learning, critical thinking, and problem-solving skills (66). Reported benefits also include growth in self-awareness and the transition from tutor-directed learning, to self-directed learning, and ultimately, autonomous, reflective practice (66).

Timing of assessment and the benefits of repeating assessments are important considerations. Students May rate themselves inappropriately high before their learning experience and score lower in terms of comfort or ability after the placement, once they have greater insight into their capabilities and have been provided with an opportunity to reflect (51, 68). Self-assessment has also been reported as more likely to be inflated among first-year students with further instruction and reflection recommended to moderate over-confidence and self-bias among novice learners (69).

Aside from the use of formalized tools to facilitate self and/or peer assessment, verbal or written reflection and engagement in focus groups provide the opportunity for students to safely contemplate and recognize their own strengths and weaknesses and, as such, is a valuable aid to learning. Benefits also include reported increases in empathy, comfort in dealing with complexity, and engagement in the learning process (45, 46, 70, 71).

4.2 Assessment as a comprehensive concept

“Effective assessment is more like a scrapbook of mementos and pictures rather than a single snapshot.”

Wiggins and McTighe, 2005, p 152

Multiple methods are needed to best capture the major aspects of knowledge and competency acquisition among student health professionals (72, 73). While the search has successfully identified self and/or peer assessment options for educators and their learners, it is important to position these within a broader suite of assessment options to maximize the development of a self-reflective health professional. Figure 2 illustrates the comprehensive nature of student assessment and the multivariate approach outlined by Wiggins and McTighe (2005), which is needed to support the development of critically thinking, self-reflective practitioners (63).

Figure 2
www.frontiersin.org

Figure 2. A comprehensive approach to assessment.

Blue et al. (2015) have noted that the lack of progress relating to the assessment of interprofessional competencies continues to create challenges for educators. Various studies conducting assessments have focused on learner attitudes toward IPE as opposed to learner IP knowledge or skill (61, 74, 75). Moreover, existing tools lack sufficient theoretical and psychometric development (61, 75). The Readiness for Interprofessional Learning Scale (56) and the Interdisciplinary Education Perception Scale (IEPS) (76) have been widely used, for both faculty (21) and student self and/or peer assessment (55), and other tools or scales have been locally developed to meet specific institutional goals and objectives (43, 49, 74). Blue et al. (2015) and Nieuwoudt et al. (2021) found that few programs reported systematic processes for evaluating individual student’s skills and behaviors related to interprofessional collaboration. It is clear that rigorous assessment and evaluation methods, standardized and widely used tools, and longitudinal assessment from diverse contexts are needed if the field of IPE is to advance and align with the demands of changing clinical care systems.

4.3 Need for further research

More study is needed to investigate the strengths and merits of qualitative scales versus more qualitative approaches in the assessment of interprofessional competency and within the suite of currently available self and/or peer assessment options. Reflective, deep dive approaches—both have been used in the literature, but little seems to have been done to reconcile them, test the value or otherwise of one over another or within mixed approaches (74) What is clear is that for the field to progress, there needs to be some consensus agreement on which measures to use to most effectively support learning.

5 Limitations

This study focuses on studies reporting self-assessment and peer assessment processes. Findings identify considerable variance among the identified tools and processes and the ways in which they were utilized. Several studies undertook a case study approach or included small cohorts only, so results May not be comprehensive or generalizable. As a detailed description of learning outcomes and psychometric properties of the results was typically lacking, it is not possible to make evidence-based comparative comments about which of the individual tools and/or processes as the most effective aids for learning. Thus, readers are encouraged to consider the recommendations in conjunction with the combination of assessment and feedback processes available to assess interprofessional readiness, capability, and competence and aid student learning.

6 Conclusion/recommendations

This review has identified a range of self- and peer assessment tools and processes to usefully contribute to the assessment of interprofessional competencies. Findings highlight the option of using a range of self and/or peer assessment approaches including formally structured tools and less structured processes, inclusive of focus groups and reflection. Discussion recommends that results identified within this search be used to complement tools, which can be used by faculty and others within a broader mosaic of assessments designed to support learning and the development of competent, self-reflective beginner practitioners. As such, the research provides a useful resource for seeking to effectively enhance interprofessional learning and competencies attainment. Of note is the conclusion that there is still more study to be undertaken in this area including the need for greater clarity and consensus agreement about definitions, tools, and the most appropriate measurement approach.

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Author contributions

SB: Conceptualization, Data curation, Formal analysis, Funding acquisition, Investigation, Methodology, Project administration, Resources, Software, Supervision, Validation, Visualization, Writing – original draft, Writing – review & editing. JY: Data curation, Formal analysis, Investigation, Writing – review & editing. DB: Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Writing – review & editing. IA: Data curation, Formal analysis, Investigation, Writing – review & editing. AP: Data curation, Formal analysis, Investigation, Writing – review & editing. KS: Data curation, Formal analysis, Investigation, Writing – review & editing. A-RY: Data curation, Formal analysis, Investigation, Methodology, Writing – review & editing. PA: Formal analysis, Investigation, Methodology, Writing – review & editing. PB: Conceptualization, Data curation, Formal analysis, Funding acquisition, Investigation, Project administration, Supervision, Writing – review & editing.

Funding

The author(s) declare that financial support was received for the research, authorship, and/or publication of this article. This research is supported by a Trust Waikato Community Impact Grant, Hamilton, New Zealand.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

1. Rawlinson, C, Carron, T, Cohidon, C, Arditi, C, Hong, QN, Pluye, P, et al. An overview of reviews on Interprofessional collaboration in primary care: barriers and facilitators. Int J Integr Care. (2021) 21:32. doi: 10.5334/ijic.5589

PubMed Abstract | Crossref Full Text | Google Scholar

2. Koppel, R, Wilson, C, and Humowiecki, M. Core competencies for frontline complex care providers. National Center for Complex Health and Social Needs, Camden Coalition of Healthcare Providers (2020)

Google Scholar

3. Vachon, B, Huynh, A-T, Breton, M, Quesnel, L, Camirand, M, Leblanc, J, et al. Patients’ expectations and solutions for improving primary diabetes care. Int J Health Care Qual Assur. (2017) 30:554–67. doi: 10.1108/IJHCQA-07-2016-0106

PubMed Abstract | Crossref Full Text | Google Scholar

4. World Health Organization ed. Framework for action on Interprofessional Education & Collaborative Practice. Geneva, Switzerland: World Health Organization (2010).

Google Scholar

5. Thistlethwaite, JE, Dunston, R, and Yassine, T. The times are changing: workforce planning, new health-care models and the need for interprofessional education in Australia. J Interprof Care. (2019) 33:361–8. doi: 10.1080/13561820.2019.1612333

PubMed Abstract | Crossref Full Text | Google Scholar

6. World Health Organization . Transforming and scaling up health Professionals' education and training: World Health Organization guidelines. Geneva: Switzerland World Health Organization (2013).

Google Scholar

7. Ansa, BE, Zechariah, S, Gates, AM, Johnson, SW, Heboyan, V, and De Leo, G. Attitudes and behavior towards Interprofessional collaboration among healthcare professionals in a large Academic Medical Center. Healthcare (Basel). (2020) 8:323–337. doi: 10.3390/healthcare8030323

PubMed Abstract | Crossref Full Text | Google Scholar

8. Andersen, P, Broman, P, Tokolahi, E, Yap, JR, and Brownie, S. Determining a common understanding of interprofessional competencies for pre-registration health professionals in Aotearoa New Zealand: a Delphi study. Front Med. (2023):10. doi: 10.3389/fmed.2023.1119556

Crossref Full Text | Google Scholar

9. Canadian Interprofessional Health Collaborative . A national Interprofessional competency framework. (2010). Available at: https://phabc.org/wp-content/uploads/2015/07/CIHC-National-InterprofessionalCompetency-Framework.pdf (Accessed March 01, 2024).

Google Scholar

10. McLaney, E, Morassaei, S, Hughes, L, Davies, R, Campbell, M, and Di Prospero, L. A framework for interprofessional team collaboration in a hospital setting: advancing team competencies and behaviours. Healthc Manage Forum. (2022) 35:112–7. doi: 10.1177/08404704211063584

PubMed Abstract | Crossref Full Text | Google Scholar

11. Chitsulo, CG, Chirwa, EM, and Wilson, L. Faculty knowledge and skills needs in interprofessional education among faculty at the College of Medicine and Kamuzu College of Nursing, University of Malawi. Malawi Med J. (2021) 33:30–4. doi: 10.4314/mmj.v33is.6

Crossref Full Text | Google Scholar

12. Hero, L-M, and Lindfors, E. Students’ learning experience in a multidisciplinary innovation project. Educ Train. (2019) 61:500–22. doi: 10.1108/ET-06-2018-0138

Crossref Full Text | Google Scholar

13. Kuhlmann Lüdeke, A, and Guillén Olaya, JF. Effective feedback, an essential component of all stages in medical education. Universitas Medica. (2020) 61:32–46. doi: 10.11144/Javeriana.umed61-3.feed

Crossref Full Text | Google Scholar

14. Burgess, A, van Diggele, C, Roberts, C, and Mellis, C. Feedback in the clinical setting. BMC Med Educ. (2020) 20:460. doi: 10.1186/s12909-020-02280-5

Crossref Full Text | Google Scholar

15. Archer, JC . State of the science in health professional education: effective feedback. Med Educ. (2010) 44:101–8. doi: 10.1111/j.1365-2923.2009.03546.x

PubMed Abstract | Crossref Full Text | Google Scholar

16. Lu, F-I, Takahashi, SG, and Kerr, C. Myth or reality: self-assessment is central to effective curriculum in anatomical pathology graduate medical education. Acad Pathol. (2021) 8:23742895211013528. doi: 10.1177/23742895211013528

PubMed Abstract | Crossref Full Text | Google Scholar

17. Mn, W, Gruppen, L, and Regehr, G. Measuring self-assessment: current state of the art. Adv Health Sci Educ. (2002) 7:63–80. doi: 10.1023/a:1014585522084

Crossref Full Text | Google Scholar

18. Yates, N, Gough, S, and Brazil, V. Self-assessment: with all its limitations, why are we still measuring and teaching it? Lessons from a scoping review. Med Teach. (2022) 44:1296–302. doi: 10.1080/0142159X.2022.2093704

PubMed Abstract | Crossref Full Text | Google Scholar

19. Andrade, HL . A critical review of research on student self-assessment. Front Educ. (2019):4. doi: 10.3389/feduc.2019.00087

Crossref Full Text | Google Scholar

20. Almoghirah, H, Nazar, H, and Illing, J. Assessment tools in pre-licensure interprofessional education: a systematic review, quality appraisal and narrative synthesis. Med Educ. (2021) 55:795807. doi: 10.1111/medu.14453

PubMed Abstract | Crossref Full Text | Google Scholar

21. Brownie, S, Blanchard, D, Amankwaa, I, Broman, P, Haggie, M, Logan, C, et al. Tools for faculty assessment of interdisciplinary competencies of healthcare students: an integrative review. Front Med. (2023):10. doi: 10.3389/fmed.2023.1124264

Crossref Full Text | Google Scholar

22. Kavannagh, J, Kearns, A, and McGarry, T. The benefits and challenges of student-led clinics within an Irish context. J Pract Teach Learn. (2015) 13:58–72. doi: 10.1921/jpts.v13i2-3.858

Crossref Full Text | Google Scholar

23. Stuhlmiller, CM, and Tolchard, B. Developing a student-led health and wellbeing clinic in an underserved community: collaborative learning, health outcomes and cost savings. BMC Nurs. (2015) 14:32. doi: 10.1186/s12912-015-0083-9

Crossref Full Text | Google Scholar

24. Tokolahi, E, Broman, P, Longhurst, G, Pearce, A, Cook, C, Andersen, P, et al. Student-led clinics in Aotearoa New Zealand: a scoping review with stakeholder consultation. J Multidiscip Healthc. (2021) 2053-2066. doi: 10.2147/JMDH.S308032

Crossref Full Text | Google Scholar

25. Frakes, K-A, Brownie, S, Davies, L, Thomas, J, Miller, M-E, and Tyack, Z. Experiences from an interprofessional student-assisted chronic disease clinic. J Interprof Care. (2014) 28:573–5. doi: 10.3109/13561820.2014.917404

PubMed Abstract | Crossref Full Text | Google Scholar

26. Frakes, K-a, Brownie, S, Davies, L, Thomas, JB, Miller, M-E, and Tyack, Z. Capricornia allied health partnership (CAHP): a case study of an innovative model of care addressing chronic disease through a regional student-assisted clinic. Australian Health Rev. (2014) 38:483–6. doi: 10.1071/AH13177

PubMed Abstract | Crossref Full Text | Google Scholar

27. Dennis, V, Craft, M, Bratzler, D, Yozzo, M, Bender, D, Barbee, C, et al. Evaluation of student perceptions with 2 interprofessional assessment tools-the collaborative healthcare interdisciplinary relationship planning instrument and the Interprofessional attitudes scale-following didactic and clinical learning experiences in the United States. J Educ Eval Health Profess. (2019) 16:35. doi: 10.3352/jeehp.2019.16.35

Crossref Full Text | Google Scholar

28. Iverson, L, Todd, M, Ryan Haddad, A, Packard, K, Begley, K, Doll, J, et al. The development of an instrument to evaluate interprofessional student team competency. J Interprof Care. (2018) 32:531–8. doi: 10.1080/13561820.2018.1447552

PubMed Abstract | Crossref Full Text | Google Scholar

29. Peltonen, J, Leino-Kilpi, H, Heikkilä, H, Rautava, P, Tuomela, K, Siekkinen, M, et al. Instruments measuring interprofessional collaboration in healthcare - a scoping review. J Interprof Care. (2020) 34:147–61. doi: 10.1080/13561820.2019.1637336

PubMed Abstract | Crossref Full Text | Google Scholar

30. Australian Skills Quality Authority . What is the difference between an assessment tool and an assessment instrument? (Clause 1.8). Available at: https://www.asqa.gov.au/faqs/what-difference-betweenassessment-tool-and-assessment-instrument-clause-18 (Accessed March 01, 2024).

Google Scholar

31. Industry Network Training & Assessment Resources . Difference between ‘tools’ and ‘instruments’ - chapter 3 (2017) Available at: https://www.intar.com.au/resources/training_and_assessing/section_3/chapter3_developing_assessment_tools/lesson2_tools_and_instruments.htm

Google Scholar

32. Munn, Z, Peters, MDJ, Stern, C, Tufanaru, C, McArthur, A, and Aromataris, E. Systematic review or scoping review? Guidance for authors when choosing between a systematic or scoping review approach. BMC Med Res Methodol. (2018) 18:143. doi: 10.1186/s12874-018-0611-x

Crossref Full Text | Google Scholar

33. Liberati, A, Altman, DG, Tetzlaff, J, Mulrow, C, Gøtzsche, PC, Ioannidis, JPA, et al. The PRISMA statement for reporting systematic reviews and Meta-analyses of studies that evaluate health care interventions: explanation and elaboration. PLoS Med. (2009) 6:e1000100. doi: 10.1371/journal.pmed.1000100

PubMed Abstract | Crossref Full Text | Google Scholar

34. Babineau, J . Product review: Covidence (systematic review software). J Can Health Lib Assoc. (2014) 35:68–71. doi: 10.5596/c14-016

Crossref Full Text | Google Scholar

35. Mink, J, Mitzkat, A, Mihaljevic, AL, Trierweiler-Hauke, B, Götsch, B, Schmidt, J, et al. The impact of an interprofessional training ward on the development of interprofessional competencies: study protocol of a longitudinal mixed-methods study. BMC Med Educ. (2019) 19:48. doi: 10.1186/s12909-019-1478-1

Crossref Full Text | Google Scholar

36. MacKenzie, D, Creaser, G, Sponagle, K, Gubitz, G, MacDougall, P, Blacquiere, D, et al. Best practice interprofessional stroke care collaboration and simulation: the student perspective. J Interprof Care. (2017) 31:793–6. doi: 10.1080/13561820.2017.1356272

PubMed Abstract | Crossref Full Text | Google Scholar

37. Mahmood, LS, Mohammed, CA, and Gilbert, JHV. Interprofessional simulation education to enhance teamwork and communication skills among medical and nursing undergraduates using the TeamSTEPPS® framework. Med J Armed Forces India. (2021) 77:S42–s8. doi: 10.1016/j.mjafi.2020.10.026

Crossref Full Text | Google Scholar

38. Begley, K, O'Brien, K, Packard, K, Castillo, S, Haddad, AR, Johnson, K, et al. Impact of Interprofessional telehealth case activities on Students' perceptions of their collaborative care abilities. Am J Pharm Educ. (2019) 83:6880. doi: 10.5688/ajpe6880

PubMed Abstract | Crossref Full Text | Google Scholar

39. National Centre for Interprofessional Practice and Education . Interprofessional socialization and valuing scale (ISVS-21) (2016) Available at: https://nexusipe.org/advancing/assessmentevaluation/interprofessional-socialization-and-valuing-scale-isvs-21 (Accessed March 01, 2024).

Google Scholar

40. The Centre for the Advancement of Interprofessional Education . The Centre for the Advancement of Interprofessional education: collaborative practice through learning together to work together (2024) Available at: https://www.caipe.org/ (Accessed March 01, 2024).

Google Scholar

41. Dubouloz, C-J, Savard, J, Burnett, D, and Guitard, P. An Interprofessional rehabilitation University Clinic in Primary Health Care: a collaborative learning model for physical therapist students in a clinical placement. J Physical Therapy Educ. (2010) 24:19–24.

Google Scholar

42. Pawłowska, M, Del Rossi, L, Kientz, M, McGinnis, P, and Padden-Denmead, M. Immersing students in family-centered interprofessional collaborative practice. J Interprof Care. (2020) 34:50–8. doi: 10.1080/13561820.2019.1608165

Crossref Full Text | Google Scholar

43. Nieuwoudt, L, Hutchinson, A, and Nicholson, P. Pre-registration nursing and occupational therapy students' experience of interprofessional simulation training designed to develop communication and team-work skills: a mixed methods study. Nurse Educ Pract. (2021) 53:103073. doi: 10.1016/j.nepr.2021.103073

PubMed Abstract | Crossref Full Text | Google Scholar

44. Timm, JR, and Schnepper, LL. A mixed-methods evaluation of an interprofessional clinical education model serving students, faculty, and the community. J Interprof Care. (2021) 35:92–100. doi: 10.1080/13561820.2019.1710117

PubMed Abstract | Crossref Full Text | Google Scholar

45. Guitard, P, Dubouloz, C-J, Metthé, L, and Brasset-Latulippe, A. Assessing Interprofessional learning during a student placement in an Interprofessional rehabilitation University Clinic in Primary Healthcare in a Canadian francophone minority context. J Res Interprofessional Pract Educ. (2010) 1:231–246.

Google Scholar

46. Johnson, AM, Woltenberg, LN, Heinss, S, Carper, R, Taylor, S, and Kuperstein, J. Whole person health: using experiential learning and the ICF model as a tool for introductory Interprofessional collaborative practice. J Allied Health (2020) 49:86–91,A.

Google Scholar

47. Roberts, SD, Lindsey, P, and Limon, J. Assessing students’ and health professionals’ competency learning from interprofessional education collaborative workshops. J Interprof Care. (2019) 33:38–46. doi: 10.1080/13561820.2018.1513915

PubMed Abstract | Crossref Full Text | Google Scholar

48. Sevin, AM, Hale, KM, Brown, NV, and McAuley, JW. Assessing Interprofessional education collaborative competencies in service-learning course. Am J Pharm Educ. (2016) 80:32. doi: 10.5688/ajpe80232

PubMed Abstract | Crossref Full Text | Google Scholar

49. Porter, KJ, Nandan, M, Varagona, L, and Maguire, MB. Effect of experiential competency-based Interprofessional education on pre-professional undergraduate students: a pilot study. J Allied Health. (2020) 49:79–85.

PubMed Abstract | Google Scholar

50. Interprofessional Education Collaborative . Core competencies for Interprofessional collaborative practice. Washington, D.C: (2023).

Google Scholar

51. Seaman, K, Saunders, R, Dugmore, H, Tobin, C, Singer, R, and Lake, F. Shifts in nursing and medical students’ attitudes, beliefs and behaviours about interprofessional work: an interprofessional placement in ambulatory care. J Clin Nurs. (2018) 27:3123–30. doi: 10.1111/jocn.14506

PubMed Abstract | Crossref Full Text | Google Scholar

52. August, B, Gortney, J, and Mendez, J. Evaluating interprofessional socialization: matched student self-assessments surrounding underserved clinic participation. Curr Pharm Teach Learn. (2020):12. doi: 10.1016/j.cptl.2020.04.006

Crossref Full Text | Google Scholar

53. Leithead, J 3rd, Garbee, DD, Yu, Q, Rusnak, VV, Kiselov, VJ, Zhu, L, et al. Examining interprofessional learning perceptions among students in a simulation-based operating room team training experience. J Interprof Care. (2019) 33:26–31. doi: 10.1080/13561820.2018.1513464

PubMed Abstract | Crossref Full Text | Google Scholar

54. Nash, WA, Hall, LA, Lee Ridner, S, Hayden, D, Mayfield, T, Firriolo, J, et al. Evaluation of an interprofessional education program for advanced practice nursing and dental students: the oralsystemic health connection. Nurse Educ Today. (2018) 66:25–32. doi: 10.1016/j.nedt.2018.03.021

PubMed Abstract | Crossref Full Text | Google Scholar

55. Seif, G, Coker-Bolt, P, Kraft, S, Gonsalves, W, Simpson, K, and Johnson, E. The development of clinical reasoning and interprofessional behaviors: service-learning at a student-run free clinic. J Interprof Care. (2014) 28:559–64. doi: 10.3109/13561820.2014.921899

PubMed Abstract | Crossref Full Text | Google Scholar

56. Parsell, G, and Bligh, J. The development of a questionnaire to assess the readiness of health care students for interprofessional learning (RIPLS). Med Educ. (1999) 33:95–100.

Google Scholar

57. Anderson, E, Manek, N, and Davidson, A. Evaluation of a model for maximizing interprofessional education in an acute hospital. J Interprof Care. (2006) 20:182–94. doi: 10.1080/13561820600625300

Crossref Full Text | Google Scholar

58. Earnest, M, Madigosky, WS, Yamashita, T, and Hanson, JL. Validity evidence for using an online peer-assessment tool (CATME) to assess individual contributions to interprofessional student teamwork in a longitudinal team-based learning course. J Interprof Care. (2022) 36:923–31. doi: 10.1080/13561820.2022.2040962

PubMed Abstract | Crossref Full Text | Google Scholar

59. Simko, LC, Rhodes, DC, McGinnis, KA, and Fiedor, J. Students' perspectives on Interprofessional teamwork before and after an Interprofessional pain education course. Am J Pharm Educ. (2017) 81:104. doi: 10.5688/ajpe816104

PubMed Abstract | Crossref Full Text | Google Scholar

60. Dobson, RT, Stevenson, K, Busch, A, Scott, DJ, Henry, C, and Wall, PA. A quality improvement activity to promote interprofessional collaboration among health professions students. Am J Pharm Educ. (2009) 73:64. doi: 10.1016/S0002-9459(24)00559-X

PubMed Abstract | Crossref Full Text | Google Scholar

61. Thannhauser, J, Russell-Mayhew, S, and Scott, C. Measures of interprofessional education and collaboration. J Interprof Care. (2010) 24:336–49. doi: 10.3109/13561820903442903

PubMed Abstract | Crossref Full Text | Google Scholar

62. Vyas, D, Ziegler, L, and Galal, SM. A telehealth-based interprofessional education module focused on social determinants of health. Curr Pharm Teach Learn. (2021) 13:1067–72. doi: 10.1016/j.cptl.2021.06.012

PubMed Abstract | Crossref Full Text | Google Scholar

63. Wiggins, G, and McTighe, J. Understanding by design. 2nd Upper Saddle River, NJ: Pearson; (2005). 1–371 p.

Google Scholar

64. New South Wales Government . Peer and self-assessment for students (2023). New South Wales Government. Available at: https://education.nsw.gov.au/teaching-and-learning/professional-learning/teacher-quality-andaccreditation/strong-start-great-teachers/refining-practice/peer-and-self-assessment-for-students (Accessed March 01, 2024).

Google Scholar

65. Cheong, CM, Luo, N, Zhu, X, Lu, Q, and Wei, W. Self-assessment complements peer assessment for undergraduate students in an academic writing task. Assess Eval High Educ. (2023) 48:135–48. doi: 10.1080/02602938.2022.2069225

Crossref Full Text | Google Scholar

66. Kajander-Unkuri, S, Leino-Kilpi, H, Katajisto, J, Meretoja, R, Räisänen, A, Saarikoski, M, et al. Congruence between graduating nursing students’ self-assessments and mentors’ assessments of students’ nurse competence. Collegian. (2016) 23:303–12. doi: 10.1016/j.colegn.2015.06.002

Crossref Full Text | Google Scholar

67. Taylor, SN . Student self-assessment and multisource feedback assessment: exploring benefits, limitations, and remedies. J Manag Educ. (2014) 38:359–83. doi: 10.1177/1052562913488111

Crossref Full Text | Google Scholar

68. Ma, K, Cavanagh, MS, and McMaugh, A. Preservice teachers’ reflections on their teaching SelfEfficacy changes for the first professional experience placement. Austrl J Teach Educ. (2021) 46:62–76. doi: 10.14221/ajte.2021v46n10.4

Crossref Full Text | Google Scholar

69. Pawluk, SA, Zolezzi, M, and Rainkie, D. Comparing student self-assessments of global communication with trained faculty and standardized patient assessments. Curr Pharm Teach Learn. (2018) 10:779–84. doi: 10.1016/j.cptl.2018.03.012

PubMed Abstract | Crossref Full Text | Google Scholar

70. Winkel, AF, Yingling, S, Jones, AA, and Nicholson, J. Reflection as a learning tool in graduate medical education: a systematic review. J Grad Med Educ. (2017) 9:430–9. doi: 10.4300/JGME-D-16-00500.1

PubMed Abstract | Crossref Full Text | Google Scholar

71. Artioli, G, Deiana, L, De Vincenzo, F, Raucci, M, Amaducci, G, Bassi, MC, et al. Health professionals and students’ experiences of reflective writing in learning: a qualitative metasynthesis. BMC Med Educ. (2021) 21:394. doi: 10.1186/s12909-021-02831-4

Crossref Full Text | Google Scholar

72. Tabish, SA . Assessment methods in medical education. Int J Health Sci (Qassim). (2008) 2:3–7.

Google Scholar

73. ClinEd Australia . The role of assessment in clinical practice (2023) Available at: https://www.clinedaus.org.au/topics-category/the-role-of-assessment-in-clinical-placement-174

Google Scholar

74. Blue, A, Chesluk, B, Conforti, L, and Holmboe, E. Assessment and evaluation in Interprofessional education: exploring the field. J Allied Health. (2015):44.

Google Scholar

75. Kenaszchuk, C, Reeves, S, Nicholas, D, and Zwarenstein, M. Validity and reliability of a multiplegroup measurement scale for interprofessional collaboration. BMC Health Serv Res. (2010) 10:83. doi: 10.1186/1472-6963-10-83

Crossref Full Text | Google Scholar

76. Luecht, RM, Madsen, MK, Taugher, MP, and Petterson, BJ. Assessing professional perceptions: design and validation of an interdisciplinary education perception scale. J Allied Health. (1990) 19:181–91.

PubMed Abstract | Google Scholar

Keywords: interdisciplinary education, interdisciplinary communication, interprofessional collaboration, self-assessment, peer assessment, healthcare student

Citation: Brownie S, Yap JR, Blanchard D, Amankwaa I, Pearce A, Sampath KK, Yan A-R, Andersen P and Broman P (2024) Tools for self- or peer-assessment of interprofessional competencies of healthcare students: a scoping review. Front. Med. 11:1449715. doi: 10.3389/fmed.2024.1449715

Received: 15 June 2024; Accepted: 05 September 2024;
Published: 02 October 2024.

Edited by:

Jill Thistlethwaite, University of Technology Sydney, Australia

Reviewed by:

Lindsay Iverson, Creighton University, United States
Vikki Park, Northumbria University, United Kingdom

Copyright © 2024 Brownie, Yap, Blanchard, Amankwaa, Pearce, Sampath, Yan, Andersen and Broman. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Sharon Brownie, c2Jyb3duaWVAc3dpbi5lZHUuYXU=

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.