Skip to main content

ORIGINAL RESEARCH article

Front. Med., 26 July 2024
Sec. Intensive Care Medicine and Anesthesiology
This article is part of the Research Topic Opportunities and Challenges of Interprofessional Collaboration and Education - Volume II View all 3 articles

Effective debriefings in the clinical setting: a pilot study to test the impact of an evidence based debriefing app on anesthesia care providers’ performance

Julia C Seelandt&#x;Julia C Seelandt1Jeannine Schneider&#x;Jeannine Schneider2Michaela Kolbe,Michaela Kolbe1,3Bastian Grande,,
Bastian Grande1,2,3*
  • 1Simulation Center, University Hospital Zurich, Zurich, Switzerland
  • 2Institute of Anesthesiology, University Hospital Zurich, Zurich, Switzerland
  • 3Department of Health Sciences and Technology, ETH Zürich, Zürich, Switzerland

Background: Debriefing enhances team learning, performance, and patient safety. Despite its benefits, it’s underused. To address this, we developed an evidence-based debriefing app.

Methods: This pilot study, conducted at a Swiss hospital, evaluated team performance during two anesthesia inductions using the Team Performance Scale (TPS). Following the first induction, teams engaged with the Zurich Debriefing App, with debriefing sessions meticulously recorded for subsequent evaluation. To mitigate bias, raters underwent comprehensive TPS training. The debriefings were analyzed through the DE-CODE framework. We utilized paired t-tests to examine performance improvements and linear regressions to assess the impact of reflective statements on performance, moderated by psychological safety.

Results: Team performance significantly improved from the first to the second induction (t (9) = −2.512, p = 0.033). Senior physicians’ (n = 8) reflective statements predicted post-assessment TPS scores (R2 = 0.732, p = 0.061), while consultants (n = 7) and nurse anesthetists (n = 10) did not. Interaction analysis revealed no moderation effects, but a main effect indicated the significance of senior physicians’ reflective statements.

Conclusion: This pilot study confirms the efficacy of the evidence-based debriefing app in enhancing anesthesia team performance. Senior physicians’ reflective statements positively influenced performance; however, no moderation effects were observed. The study highlights the potential of debriefing apps to streamline and enhance team debriefing processes, with significant implications for improving clinical practice and patient safety. Further research is needed to validate these findings on a larger scale and optimize the integration of debriefing into routine clinical practice.

Background

Healthcare debriefings have the potential to enhance team learning and team performance in ad hoc teams. They reduce errors and improve patient safety (1, 2). It is a guided conversation among clinicians that aims to explore and understand the relationships among events, actions, thought and feeling processes, and performance outcomes of a clinical situation (37). A core element of debriefings is promoting experiential learning and thus reflecting/shared reflection which in turn may allow the development of strategies that can be applied in future performance episodes (812). Kolb’s Experiential Learning Theory posits that learning is a process where knowledge is created through the transformation of experience, following a cyclical model comprising four stages: Concrete Experience, Reflective Observation, Abstract Conceptualization, and Active Experimentation. This theory emphasizes that effective learning involves actively engaging in experiences, reflecting on them, conceptualizing the insights gained, and then applying these insights in practice (13). Debriefings are likely to be a suitable learning infra-structure (14, 15), particularly for ad hoc teams in healthcare with their temporal instability (16). While the potential of debriefings is increasingly recognized (17, 18) and empirical studies have demonstrated their benefits (1921), they are still underutilized (15, 22, 23). Research has demonstrated that debriefings are only seemingly easy to conduct. In fact, they require a number of challenging conversational skills (24) and knowledge about team functioning (3, 21) which may be discouraging and requires the exploration of ways to help start and conduct debriefings (17). Also, research on organizational behavior suggests that many assumptions exist that may prevent healthcare personnel from engaging in debriefings. The so called “debriefing myths” include debriefing only when disaster strikes, debriefing is a luxury, senior clinicians should determine debriefing content, and debriefers must be neutral and nonjudgmental (1). These myths offer valuable insights into why current debriefing practices are ad hoc and not embedded into daily unit practices (1).

Different tools for conducting debriefings in the clinical setting exist for either hot [immediately after an event (4)] or cold [delayed hours to weeks after an event (4)] debriefings. These tools have in common that they have a similar framework and structure (5) but all of them lack a systematic analysis of the interaction between debriefers and participants namely how actions of debriefers relate to actions of participants; they mostly do not illuminate the debriefing process nor do they focus on specific questions to trigger participants’ double loop learning.

We therefore aimed to develop an evidence-based dynamic debriefing tool that contains evidence for the immediate effectiveness of selected debriefing and participant communications (6).

We have also tried to address the assumptions about debriefings mentioned above and why debriefings are rarely performed in the clinical setting. For example, the moderator of the debriefing is recommended, and participants are given a selection of topics to talk about in the debriefing (e.g., leadership, team coordination, speak up, team communication). In addition, participants are guided through the debriefing while using the debriefing app; for each phase of the debriefing, participants receive suggestions for effective question and they can also access current research results on selected crisis resource principles.

The objective of this observational pilot study was to test the impact of an evidence based debriefing app on anesthesia care providers’ performance. Based on team science and debriefing literature, we hypothesized that using the debriefing app in between two complex induction of anesthesia will enable team members to reflect and thus improve the performance of the second induction. Specifically, we tested the following hypotheses: (1) Team performance during anesthesia induction assessed by the Team Performance Scale (TPS) will increase after the debriefing and (2) the more reflective statements are verbalized during debriefings, the better the team performance is during the second induction for senior consultants, consultants, and registered anesthesia nurses, respectively. This relationship is moderated by psychological safety. Reflective statements were assessed via behavior observation and in situ behavior coding rather than relying on self-reports (79).

Methods

The respective ethics committee determined this study to be exempt KEK-ZH-Nr. 2013-0592.

Study design and inclusion/exclusion criteria

Data collection for this study took place at a central care-providing hospital in Switzerland. The participants included 10 male and 12 female anesthesia care providers. We observed participants performing complex inductions of general anesthesia in teams of 2 or 3. After the first induction, the participants used an app to debrief themselves, followed by a second complex induction of general anesthesia, which we observed again. Inductions were performed in the anesthesia induction room adjacent to the theater. Debriefings were conducted in a separate room immediately after the induction, facilitated by another anesthesia team that relieved the original team for this purpose.

Participants were recruited over 5 months for anesthesia in thoracic, visceral, vascular, or neurosurgery. Inclusion criteria included patients with an ASA classification of two or higher, requiring a central venous catheter, arterial catheter, thoracic epidural catheter, or double lumen tube, and complex patient positioning (e.g., prone or side position) (10). The exclusion criteria were anesthesia inductions in patients with an ASA classification 1 and 2, without extended monitoring or complex positioning. The anesthesia inductions included general anesthesia with and without thoracic epidural anesthesia, and all cases were elective surgical procedures. The teams consisted of one anesthesia consultant, one registrar, and one registered anesthesia nurse.

The anesthesia inductions took place in a designated induction room. After the placement of a thoracic epidural catheter, the usual steps such as preoxygenation, pharmacological induction, and pharmacological stabilization of blood pressure (within the usual range) were carried out.

Data collection

Data were collected anonymously. Participants were informed about the study both verbally and through written documents, and written informed consent was obtained. Patient characteristics (age, physical status, ASA classification), type of surgical procedure, monitoring, duration of anesthesia induction, intubation method, and patient positioning were extracted from the patient file and anesthesia protocol. The debriefings were videotaped.

During inductions of anesthesia, team members were observed and assessed using the Team Performance Scale (TPS). The TPS analyzes the roles and responsibilities of team members and focuses on effective communication (11).

The TPS has been used as surrogate for the quality of the anesthesia induction. Raters were consultant anesthesiologists and anesthesia nurses with years of professional experience. All raters participated a two-hour rater training. The training included general information about the study purpose, a structured introduction into the rating systems and the observation method and rating of one videotaped induction of anesthesia using TPS under the direct guidance. To assess interrater reliability, two additional videotaped anesthesia inductions were evaluated. Training was considered complete if agreement between trainees and expert coders (Intraclass Correlation Coefficient) was.70 for both instruments indicating good interrater reliability (12). During anesthesia induction, the raters were placed closely to the anesthesia team and used TPS in real-time with direct observation.

The observation started with administering the first drug and ended with the handover to the surgical staff (25).

After the first anesthesia induction was finished, the participants used the Zurich Debriefing App for a videotaped debriefing. Afterwards participants performed another induction of general anesthesia and underwent the same procedure. Both anesthesia inductions have been rated by different raters to avoid any biases (Figures 1, 2).

Figure 1
www.frontiersin.org

Figure 1. Screenshot of the application, showing the process of questioning.

Figure 2
www.frontiersin.org

Figure 2. Study process.

Measurements

Participants completed a questionnaire after each debriefing. Psychological safety was measured using a validated German translation (26, 27) of the Team Psychological Safety scale.

Data analysis

The debriefings have been observed remotely by the study team (7). In particular, they applied four codes of the DE-CODE, a valid and reliable coding scheme for assessing debriefers’ and learners’ communication in debriefings (28, 29). The authors focused on learners reflective statements/marker including learners analyses why something happened (DE-CODE: description), mentioning mental models (DE-CODE: mental models), learners conclusions about lessons learned or other actions that s/he could have done (DE-CODE: conclusion) as well as future-oriented action plans (DE-CODE: action plan). The anesthesia teams have been observed from the beginning of the debriefing until the end of the debriefing and reflective statements/marker have been recorded.

Behavioral coding was conducted using a standard personal computer and Excel sheet. To ensure interrater reliability, two coders independently coded 20% (36 out of 180) of the videotaped debriefings.

Statistical analysis

Interrater reliability was assessed using the Intraclass Correlation Coefficient (ICC), suitable for ordinal, interval, and ratio scales. ICC values below 0.40 indicate poor reliability, between 0.40 and 0.59 are considered fair, 0.60–0.74 are good, and above 0.75 are excellent (30). Statistical analyses were performed using IBM SPSS V.26 software.

To evaluate the hypothesis that teams perform better during the second anesthesia induction, paired sample t-tests were conducted.

For the hypothesis concerning the relationship between the verbalization of reflective statements during debriefings and team performance during the second induction, linear regressions were performed. Additionally, a moderation analysis was conducted to examine whether this relationship is moderated by psychological safety.

Results

Interrater reliability

The ICC between two independent coders assessing 20% of the debriefings was 0.73, indicating good interrater reliability.

Participants and descriptive data

Debriefings involved a minimum of two and a maximum of three participants, including attending physicians, resident physicians, and nurses with varying levels of experience in anesthesia. The mean duration of debriefings was 12.5 min, with a range of reflective statements made by participants. The average anesthesia experience was 9.17 years, on average; the team size for induction was 2.5 people. One person had never had simulation training with debriefing until then; all other participants were familiar with debriefing through simulation training.

Descriptive data for inductions

Patients undergoing anesthesia inductions had an average ASA score of 3.05, with procedures primarily neurosurgical or thoracic in nature.

Hypothesis testing

Results from paired sample t-tests revealed a significant increase in team performance from the first to the second anesthesia induction (p = 0.033), confirming the first hypothesis.

Regarding the second hypothesis, linear regression analyses showed that senior consultants’ reflective statements predicted post-assessment team performance scores (R2 = 0.732, p = 0.061), while consultants’ and registered anesthesia nurses’ statements did not significantly predict team performance. Moderation analysis did not reveal significant interactions between reflective statements and psychological safety for any group of anesthesia care providers. Therefore, the first hypothesis was confirmed, while the second hypothesis was partially supported, and the moderation hypothesis was not confirmed (Table 1).

Table 1
www.frontiersin.org

Table 1. ANTS and TPS score.

Discussion

Aim of this pilot study was to test the impact of evidence based, guided debriefing app on anesthesia care providers’ team performance. Based on team science and debriefing literature, we hypothesized that using the debriefing app in between two complex induction of anesthesia will enable team members to reflect and thus improve the performance of the second induction. Specifically, we hypothesized that team performance will increase from first to second induction of anesthesia and that the more reflective statements are verbalized during debriefings, the better the team performance is during the second induction. In addition, we hypothesized that this relationship is moderated by psychological safety. We assessed reflective statements via behavior observation and team performance was assessed by using TPS. Results showed that our first hypothesis is confirmed.

Interpreting effect sizes is a critical aspect of research methodology. Cohen’s benchmarks (1988) classify effect sizes as small (d = 0.2), medium (d = 0.5), and large (d = 0.8), but their application should not be overly rigid. Despite these benchmarks, small effect sizes can hold significant practical implications, as seen in instances like interventions leading to a substantial reduction in suicide rates with an effect size of d = 0.1. While Cohen’s d for between-subject designs can be interpreted as a fraction of the standard deviation, offering a tangible measure, the most meaningful interpretation involves contextualizing the effect within existing literature and elucidating its practical implications. However, there is a lack of clear guidelines on how to undertake this process. Therefore, researchers must exercise discretion in interpreting effect sizes, considering both statistical benchmarks and the broader context of the research field (31).

Teamwork and thus patient safety can be improved by reflexivity, through reflexivity in debriefing, but also in a briefing or during action (32, 33). Based on this information, reflexivity in debriefing should be promoted.

In our study, the second hypothesis was that increased reflexivity in debriefing would lead to an improvement in TPS in the second induction, this was shown to be only partially significant. This was only shown in relation to the reflexivity of the senior doctors’ statements. However, this was probably also due to the small sample in the pilot study. This would have to be analysed again in a larger study and especially the participants’ share of conversation in the debriefing as well as the reflection markers would have to be considered further.

The second hypothesis is only confirmed for senior consultants, a main effect is shown in the reflective statements of the senior consultants and an increased performance post, otherwise no moderation effects were shown. The results show that our second hypothesis is not confirmed.

The strengths of the study are certainly demonstrated by the ease of conducting the debriefing using an app on a smartphone or pad, as this can be done in a resource-efficient and simple way. After all, the use of smartphones in everyday clinical practice is now well accepted by most doctors and nurses (34). Through the app, the team can be guided neutrally through the debriefing and the participants are tempted to reflect on their actions in the team. The limitations of this study are that it is a single center study and has only a small number of cases. Furthermore, organizing the same team for two consecutive complex anesthesia inductions proved to be a challenge.

It is noteworthy to highlight our adherence to recommendations put forth, as evidenced by the alignment of our approach with the findings elucidated in the systematic review on clinical debriefing tools: attributes and evidence for use. Additionally, our reference to authoritative documents such as Healthcare Simulation Standards of Best Practicetm, The Debriefing Process, Reflective debrief and the social space: offload, refuel, and stay on course, and Clinical debriefing: TALK© to learn and improve together in healthcare environments, underscores the robust methodology.

The incorporation of reflexivity during debriefing sessions has been shown in contemporary literature to be conducive to enhancing teamwork dynamics and bolstering patient safety measures (35). This is particularly pertinent given the complexities inherent in healthcare environments. Furthermore, our findings pertaining to the second hypothesis, while partially significant, warrant nuanced interpretation. The observed partial significance could be attributed, in part, to the relatively modest sample size utilized in our study. Moving forward, it may be prudent to delve deeper into the conversational dynamics within debriefings, potentially shedding light on the need to ensure equitable participation beyond senior consultants. It is plausible that other anesthesia providers may have contributed disproportionately to the overall discourse. Consequently, future analyses should prioritize assessing the balance of reflective markers rather than focusing solely on individual contributors.

The accessibility and dynamic nature of our debriefing application are notable, serving as an effective tool in guiding users through the debriefing process. By reducing barriers, such as complexity and time constraints, our application streamlines the debriefing experience, making it more accessible and resource-efficient in clinical settings.

Moreover, our findings underscore the versatility of our approach, as it is suitable for both hot and cold debriefings, as advocated by Sugarman (5). However, it is imperative to acknowledge the limitations inherent in our study design. As a single-center study with a modest sample size, our findings may not be generalizable to broader contexts. Furthermore, the pilot nature of our study posed challenges in ensuring stable team compositions for two sequential inductions, potentially impacting the robustness of our findings. Additionally, our study focused exclusively on a single discipline within healthcare, further limiting the generalizability of our findings.

Finally, despite concerns surrounding the integration of smartphone applications in clinical practice, our findings indicate a prevailing positive attitude among healthcare professionals toward their use. This trend is supported by the burgeoning adoption of smartphones among healthcare professionals over the past decade, with approximately 80% of doctors and 85% of medical trainees utilizing smartphones in their professional capacities (34).

For clinicians, these findings present significant advantages. They allow for systematic and structured debriefings to be conducted without a loss of time. Additionally, they document the learning effect. Furthermore, team members are trained to independently conduct effective debriefings. Based on the findings, the use of the application can be recommended; however, the effect of the subject of debriefing should not be overlooked (36). In this study, only non-critical situations were discussed, aligning with the Safety II concept by Hollnagel et al. (37). Whether this structure yields similarly positive effects in situations involving incidents remains to be seen.

We hope that this pilot study will help to confirm our hypotheses in a larger study and create a tool through this app that can better integrate debriefing into everyday clinical practice and thus improve team performance and patient safety.

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Ethics statement

Ethical review and approval was not required for the study on human participants in accordance with the local legislation and institutional requirements. Written informed consent from the patients/participants was not required to participate in this study in accordance with the national legislation and the institutional requirements.

Author contributions

CS: Conceptualization, Data curation, Formal analysis, Funding acquisition, Investigation, Methodology, Project administration, Resources, Software, Supervision, Validation, Visualization, Writing – original draft. JS: Data curation, Project administration, Visualization, Writing – original draft. MK: Conceptualization, Data curation, Formal analysis, Funding acquisition, Methodology, Project administration, Resources, Supervision, Writing – original draft. BG: Conceptualization, Funding acquisition, Methodology, Resources, Supervision, Validation, Writing – original draft, Writing – review & editing.

Funding

The author(s) declare that financial support was received for the research, authorship, and/or publication of this article. The research was supported by a grant from the Swiss National Science Foundation (Grant No. 100014_152822).

Acknowledgments

The authors thank Michael Ganter, Julio Moral, Reto Thomasin, Eva Streuli, Thomas Rieder and Roland Gsell for their great cooperation.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

The author(s) declared that they were an editorial board member of Frontiers, at the time of submission. This had no impact on the peer review process and the final decision.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

1. Seelandt, JC, Walker, K, and Kolbe, M. “A debriefer must be neutral” and other debriefing myths: a systemic inquiry-based qualitative study of taken-for-granted beliefs about clinical post-event debriefing. Adv Simul. (2021) 6:1–15. doi: 10.1186/s41077-021-00161-5

Crossref Full Text | Google Scholar

2. Keiser, NL, and Arthur, W Jr. A meta-analysis of the effectiveness of the after-action review (or debrief) and factors that influence its effectiveness. J Appl Psychol. (2021) 106:1007–32. doi: 10.1037/apl0000821

PubMed Abstract | Crossref Full Text | Google Scholar

3. Kolbe, M, Grande, B, and Spahn, DR. Briefing and debriefing during simulation-based training and beyond: content, structure, attitude and setting. Best Pract Res Clin Anaesthesiol. (2015) 29:87–96. doi: 10.1016/j.bpa.2015.01.002

PubMed Abstract | Crossref Full Text | Google Scholar

4. Kessler, DO, Cheng, A, and Mullan, PC. Debriefing in the emergency department after clinical events: a practical guide. Ann Emerg Med. (2015) 65:690–8. doi: 10.1016/j.annemergmed.2014.10.019

PubMed Abstract | Crossref Full Text | Google Scholar

5. Sugarman, M, Graham, B, Langston, S, Nelmes, P, and Matthews, J. Implementation of the ‘TAKE STOCK’Hot debrief tool in the ED: a quality improvement project. Emerg Med J. (2021) 38:579–84. doi: 10.1136/emermed-2019-208830

PubMed Abstract | Crossref Full Text | Google Scholar

6. Kolbe, M, Grande, B, Lehmann-Willenbrock, N, and Seelandt, JC. Helping healthcare teams to debrief effectively: associations of debriefers’ actions and participants’ reflections during team debriefings. BMJ Qual Saf. (2022) 32:160–72. doi: 10.1136/bmjqs-2021-014393

Crossref Full Text | Google Scholar

7. Brauner, E, Boos, M, and Kolbe, M. The Cambridge handbook of group interaction analysis. Cambridge: Cambridge University Press (2018).

Google Scholar

8. Kolbe, M, and Boos, M. Laborious but elaborate: the benefits of really studying team dynamics. Front Psychol. (2019) 10:1478. doi: 10.3389/fpsyg.2019.01478

PubMed Abstract | Crossref Full Text | Google Scholar

9. Gerpott, FH, Lehmann-Willenbrock, N, and Scheibe, S. Is work and aging research a science of questionnaires? Moving the field forward by considering perceived versus actual behaviors. Work Aging Retire. (2020) 6:65–70. doi: 10.1093/workar/waaa002

Crossref Full Text | Google Scholar

10. Doyle, DJ, Goyal, A, Bansal, P, and Garmon, EH. American society of anesthesiologists classification. Treasure Island (FL): Stat Pearls (2021).

Google Scholar

11. Sigalet, E, Donnon, T, Cheng, A, Cooke, S, Robinson, T, Bissett, W, et al. Development of a team performance scale to assess undergraduate health professionals. Acad Med. (2013) 88:989–96. doi: 10.1097/ACM.0b013e318294fd45

PubMed Abstract | Crossref Full Text | Google Scholar

12. Seelandt, JC. Quality control: assessing reliability and validity In: E Brauner, M Boos, and M Kolbe, editors. The Cambridge handbook of group interaction analysis. Cambridge: University Press (2018). 227–44.

Google Scholar

13. Yardley, S, Teunissen, PW, and Dornan, T. Experiential learning: transforming theory into practice. Med Teach. (2012) 34:161–4. doi: 10.3109/0142159X.2012.643264

PubMed Abstract | Crossref Full Text | Google Scholar

14. Tannenbaum, SI, and Cerasoli, CP. Do team and individual debriefs enhance performance? A meta-analysis. Hum Factors. (2013) 55:231–45. doi: 10.1177/0018720812448394

PubMed Abstract | Crossref Full Text | Google Scholar

15. Tannenbaum, SI, and Goldhaber-Fiebert, SN. Medical team debriefs: Simple, powerful, underutilized In: E Salas and K Frush, editors. Improving patient safety through teamwork and team training. New York: Oxford University Press (2013). 249–56.

Google Scholar

16. Vashdi, DR, Bamberger, PA, and Erez, M. Can surgical teams ever learn? The role of coordination, complexity, and transitivity in action team learning. Acad Manag J. (2013) 56:945–71. doi: 10.5465/amj.2010.0501

Crossref Full Text | Google Scholar

17. Mullan, PC, Kessler, DO, and Cheng, A. Educational opportunities with postevent debriefing. JAMA. (2014) 312:2333–4. doi: 10.1001/jama.2014.15741

PubMed Abstract | Crossref Full Text | Google Scholar

18. Eppich, WJ, Mullan, PC, Brett-Fleegler, M, and Cheng, A. “Let's talk about it”: translating lessons from health care simulation to clinical event debriefings and coaching conversations. Clinic Pediatr Emerg Med. (2016) 17:200–11. doi: 10.1016/j.cpem.2016.07.001

Crossref Full Text | Google Scholar

19. Weiss, M, Kolbe, M, Grote, G, Spahn, DR, and Grande, B. Why didn’t you say something? Effects of after-event reviews on voice behaviour and hierarchy beliefs in multi-professional action teams. Eur J Work Organ Psy. (2017) 26:66–80. doi: 10.1080/1359432X.2016.1208652

Crossref Full Text | Google Scholar

20. Eddy, ER, Tannenbaum, SI, and Mathieu, JE. Helping teams to help themselves: comparing two team-led debriefing methods. Pers Psychol. (2013) 66:975–1008. doi: 10.1111/peps.12041

Crossref Full Text | Google Scholar

21. Smith-Jentsch, KA, Cannon-Bowers, JA, Tannenbaum, SI, and Salas, E. Guided team self-correction - impacts on team mental models, processes, and effectiveness. Small Group Res. (2008) 39:303–27. doi: 10.1177/1046496408317794

Crossref Full Text | Google Scholar

22. Ahmed, M, Sevdalis, N, Paige, J, Paragi-Gururaja, R, Nestel, D, and Arora, S. Identifying best practice guidelines for debriefing in surgery: a tri-continental study. Am J Surg. (2012) 203:523–9. doi: 10.1016/j.amjsurg.2011.09.024

PubMed Abstract | Crossref Full Text | Google Scholar

23. Gougoulis, A, Trawber, R, Hird, K, and Sweetman, G. ‘Take 10 to talk about it’: use of a scripted, post-event debriefing tool in a neonatal intensive care unit. J Paediatr Child Health. (2020) 56:1134–9. doi: 10.1111/jpc.14856

PubMed Abstract | Crossref Full Text | Google Scholar

24. Rudolph, JW, Foldy, EG, Robinson, T, Kendall, S, Taylor, SS, and Simon, R. Helping without harming: the instructor’s feedback dilemma in debriefing—a case study. Simul Healthc. (2013) 8:304–16. doi: 10.1097/SIH.0b013e318294854e

Crossref Full Text | Google Scholar

25. Künzle, B, Zala-Mezö, E, Wacker, J, Kolbe, M, and Grote, G. Leadership in anaesthesia teams: the most effective leadership is shared. Qual Saf Health Care. (2010) 19:e46–6. doi: 10.1136/qshc.2008.030262

Crossref Full Text | Google Scholar

26. Baer, M, and Frese, M. Innovation is not enough: climates for initiative and psychological safety, process innovations, and firm performance. J Organ Behav. (2003) 24:45–68. doi: 10.1002/job.179

Crossref Full Text | Google Scholar

27. Edmondson, A. Psychological safety and learning behavior in work teams. Adm Sci Q. (1999) 44:350–83. doi: 10.2307/2666999

Crossref Full Text | Google Scholar

28. Seelandt, JC, Grande, B, Kriech, S, and Kolbe, M. DE-CODE: a coding scheme for assessing debriefing interactions. BMJ STEL. (2018) 4:51–8. doi: 10.1136/bmjstel-2017-000233

PubMed Abstract | Crossref Full Text | Google Scholar

29. Brauner, E. Coding interaction In: E Brauner, M Boos, and M Kolbe, editors. The Cambridge handbook of group interaction analysis. Cambridge handbooks in psychology. Cambridge: Cambridge University Press (2018). 165–90.

Google Scholar

30. Evans, JD. Straightforward statistics for the behavioral sciences: Pacific grove. United States: CA Brooks/Cole Pub. Co. (1996).

Google Scholar

31. Lakens, D. Calculating and reporting effect sizes to facilitate cumulative science: a practical primer for t-tests and ANOVAs. Front Psychol. (2013) 4:863. doi: 10.3389/fpsyg.2013.00863

Crossref Full Text | Google Scholar

32. McHugh, SK, Lawton, R, O'Hara, JK, and Sheard, L. Does team reflexivity impact teamwork and communication in interprofessional hospital-based healthcare teams? A systematic review and narrative synthesis. BMJ Qual Saf. (2020) 29:672–83. doi: 10.1136/bmjqs-2019-009921

PubMed Abstract | Crossref Full Text | Google Scholar

33. Schmutz, JB, and Eppich, WJ. Promoting learning and patient care through shared reflection: a conceptual framework for team reflexivity in health care. Acad Med. (2017) 92:1555–63. doi: 10.1097/ACM.0000000000001688

PubMed Abstract | Crossref Full Text | Google Scholar

34. Kerst, A, Zielasek, J, and Gaebel, W. Smartphone applications for depression: a systematic literature review and a survey of health care professionals' attitudes towards their use in clinical practice. Eur Arch Psychiatry Clin Neurosci. (2020) 270:139–52. doi: 10.1007/s00406-018-0974-3

PubMed Abstract | Crossref Full Text | Google Scholar

35. Shrader, SZ, and Zaudke, J. Top ten best practices for interprofessional precepting. J Interprof Edu Prac. (2018) 10:56–60. doi: 10.1016/j.xjep.2017.12.004

Crossref Full Text | Google Scholar

36. Bergerod, IJ, Braut, GS, and Wiig, S. Resilience from a stakeholder perspective: the role of next of kin in Cancer care. J Patient Saf. (2020) 16:e205–10. doi: 10.1097/PTS.0000000000000532

PubMed Abstract | Crossref Full Text | Google Scholar

37. Hollnagel, E, Wears, RL, and Braithwaite, J. From safety-I to safety-II: A white paper The Resilient Health Care Net: University of Southern Denmark, University of Florida, USA, and Macquarie University, Australia (2015).

Google Scholar

Keywords: self-debriefing, reflective statements, team learning, team performance, debriefing application

Citation: Seelandt JC, Schneider J, Kolbe M and Grande B (2024) Effective debriefings in the clinical setting: a pilot study to test the impact of an evidence based debriefing app on anesthesia care providers’ performance. Front. Med. 11:1427061. doi: 10.3389/fmed.2024.1427061

Received: 02 May 2024; Accepted: 03 July 2024;
Published: 26 July 2024.

Edited by:

Heather A. Wolfe, Children’s Hospital of Philadelphia, United States

Reviewed by:

Anthony Paul Breitbach, Saint Louis University, United States
Leon Marković, University Hospital Sveti Duh, Croatia

Copyright © 2024 Seelandt, Schneider, Kolbe and Grande. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Bastian Grande, bastian.grande@usz.ch

These authors have contributed equally to this work and share first authorship

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.