- 1Department of Obstetrics and Gynecology, Aarhus University Hospital, Aarhus, Denmark
- 2Department of Clinical Medicine, Aarhus University, Aarhus, Denmark
- 3Department of Obstetrics and Gynecology, Sunderland Royal Hospital, Sunderland, United Kingdom
- 4Department of Obstetrics and Gynecology, Horsens Regional Hospital, Horsens, Denmark
- 5Fachhochschule Nordwestschweiz (FHNW) School of Applied Psychology, University of Applied Sciences and Arts Northwestern Switzerland, Olten, Switzerland
Introduction: In Northern Europe, vacuum-assisted delivery (VAD) accounts for 6–15% of all deliveries; VAD is considered safe when conducted by adequately trained personnel. However, failed vacuum extraction can be harmful to both the mother and child. Therefore, the clinical performance in VAD must be assessed to guide learning, determine a performance benchmark, and evaluate the quality to achieve an overall high performance. We were unable to identify a pre-existing tool for evaluating the clinical performance in real-life vacuum-assisted births.
Objective: We aimed to develop and validate a checklist for assessing the clinical performance in VAD.
Methods: We conducted a Delphi process, described as an interactive process where experts answer questions until answers converge toward a “joint opinion” (consensus). We invited international experts as Delphi panelists and reached a consensus after four Delphi rounds, described as follows: (1) the panelists were asked to add, remove, or suggest corrections to the preliminary list of items essential for evaluating clinical performance in VAD; (2) the panelists applied weights of clinical importance on a Likert scale of 1–5 for each item; (3) each panelist revised their original scores after reviewing a summary of the other panelists’ scores and arguments; and (4) the TeamOBS-VAD was tested using videos of real-life VADs, and the Delphi panel made final adjustments and approved the checklist.
Results: Twelve Delphi panelists from the UK (n = 3), Norway (n = 2), Sweden (n = 3), Denmark (n = 3), and Iceland (n = 1) were included. After four Delphi rounds, the Delphi panel reached a consensus on the checklist items and scores. The TeamOBS-VAD checklist was tested using 60 videos of real-life vacuum extractions. The inter-rater agreement had an intraclass correlation coefficient (ICC) of 0.73; 95% confidence interval (95% CI) of [0.58, 0.83], and that for the average of two raters was ICC 0.84 95% CI [0.73, 0.91]. The TeamOBS-VAD score was not associated with difficulties in delivery, such as the number of contractions during vacuum extraction delivery, cephalic level, rotation, and position. Failed vacuum extraction occurred in 6% of the video deliveries, but none were associated with the teams with low clinical performance scores.
Conclusion: The TeamOBS-VAD checklist provides a valid and reliable evaluation of the clinical performance of vaginal-assisted vacuum extraction.
1 Introduction
In Northern Europe, 6–15% of women have had a vacuum-assisted delivery (VAD) (1). Notably, most births using a vacuum have good outcomes, and VAD is generally accepted as safe when performed by appropriately trained healthcare providers (2). When delivery is indicated in the second stage of labor, obstetricians must balance the differing risks of instrumental-assisted births with those of second-stage cesarean births. Women delivered successfully using a vacuum have a higher chance of uncomplicated births in subsequent pregnancies than those delivered through a cesarean section. Furthermore, cesarean birth in the second stage can be challenging because of increased maternal and perinatal risks (3). Therefore, instrumental delivery remains a core obstetrical competence, and systematic clinical performance evaluation is crucial to ensure ongoing quality assessment and research to improve clinical training (4).
Available checklists for VAD have been designed to support procedural task execution, as cognitive aids or by evaluating by item-by-item feedback. These existing checklists in vacuum extraction use simple dichotomous items (not done/done) or Likert scales (5–7). However, a growing body of empirical evidence suggests that performance assessment should include weighted checklist items to differentiate between essential and less important actions. Furthermore, the inclusion of time frames helps to create a more refined assessment of performance (8, 9). The main advantage of these performance assessment tools is the production of an objective summative score that is valuable in quality assessment, benchmarking performance and research (10).
We could not identify an existing performance assessment tool for real-life vacuum-assisted births that fulfilled the abovementioned requirements; therefore, this study aimed to develop and validate a checklist for assessing the clinical performance of VAD.
2 Materials and methods
2.1 Delphi method
We developed the TeamOBS-VAD checklist using a Delphi process to evaluate clinical performance (Figure 1). The Delphi method is an interactive process in which experts answer questions in four rounds until the answers reach a consensus. This is an internationally recognized method for solving research questions with different clinical approaches in practice and limited evidence (11–13). Our research aimed to develop a list of core items/tasks that a team should perform when conducting a VAD. The Delphi process was conducted online using Google survey tool forms.
Eighteen international obstetricians were invited to participate in the Delphi panel; 12 accepted and completed the Delphi process. The Delphi panelists were obstetric consultants from the UK (n = 3), Norway (n = 2), Sweden (n = 3), Denmark (n = 3), and Iceland (n = 1). The Delphi process was anonymous to ensure equal weight for all participants’ arguments and suggestions. The Delphi steering committee (NU, LB, OK, and LH) drafted a preliminary list of items identified from the literature and international guidelines, each representing a core task in the VAD.
In the first round, the panel reviewed the preliminary list of items and was asked to add, remove, or suggest the wording of items and argue why. In the second round, panelists weighted each item for clinical importance on a 5-point Likert scale, where “5” was highly important and “1” was least important. In round three, the panelists reassessed the weights after reviewing a summary of the other panelists’ scores and arguments. Consensus was defined when 90% of the panelists’ scores fell within three neighboring categories using the Likert scale range of 1–5.
The TeamOBS-VAD checklist was designed using a predefined blueprint (10). The checklist resulted in a total score calculated using a “weighted score” of 0–100% as a percentage of the highest possible points and the assessors’ subjective global rating, the “patient safety score,” ranging from 0–100% (100% served as a goal for others). The “patient safety score” represented the assessors’ subjective global rating of all treatment actions and offered the opportunity to evaluate aspects of performance that were not captured by the 18 items. The TeamOBS-VAD score was calculated as follows: (weighted score + patient safety score)/2.
2.2 Validity and reliability testing
We used the conceptual definitions and arguments for validity described by Cook et al. (14, 15) to test validity and reliability. We used video recordings of real-life VADs, collected with informed consent from all individuals present (patients and staff) in the videos from two Danish hospitals: Aarhus University and Horsens Regional Hospitals. Aarhus University Hospital, with approximately 5,000 deliveries per year, provides level III maternal care (16) and Horsens Regional Hospital, with approximately 2,000 deliveries per year, provides level II maternal care. All 17 birthing suites at the two hospitals were equipped with two or three high-definition minidome surveillance cameras and a microphone attached to the ceiling, allowing for a comprehensive view of the room. As previously described, the recordings were automatically activated using Bluetooth (17). The instruments used for vacuum extraction were a Malmström or Bird metal size 5–6 cm cup or a soft silicone cup. Videos were obtained with informed consent over 15 months between January 2015 and March 2016 and analyzed between 2018 and 2019.
Consultants LH and LA conducted tests for validity and reliability. They were experienced video raters from previous TeamOBS studies. In a session, they were trained as “raters” to develop familiarity with the checklist, followed by detailed discussions about the high and low score definitions for each item in the tool. The TeamOBS-VAD checklist was then independently applied to 60 real-life VAD video recordings. One month later, they reassessed 20% of the recordings (randomly selected) to evaluate both inter-rater and intra-rater agreements. Notably, all 60 videos were ranked based on their clinical performance scores. NU, LB, and LH reviewed the videos chronologically to determine where to set the low, acceptable, and high team performance levels.
2.3 Ethics
This study was approved by the Central Denmark Region’s legal department, Danish Data Protection Agency (2012-58-006), and Central Denmark Region’s Research Foundation (Case No. 1-16-02-257-14). All the participants (patients and staff) volunteered to participate and provided informed consent.
2.4 Statistical analysis
The clinical performance scores were analyzed on a logit-transformed scale to meet the criteria for normality and back-transformed using the inverse logit function (18). Rater agreement was described as summative using the intraclass correlation (19), Bland–Altman plots, and limits of agreement (20). STATA 17 (StataCorp LP, College Station, TX, USA) was used for statistical analysis.
3 Results
The Delphi panel achieved consensus in round one after adjusting for five items and adding one. Notably, all panelists weighted items according to importance on a Likert scale of 1–5; they reached a consensus on 14/18 items in the second round and a consensus on all 18 items in the next round. The TeamOBS-VAD checklist was developed and tested for usability in a simulation based on the Delphi method, and the raters found it easy to understand and use (Figure 2, Supplementary material 1).
Figure 2. In graphs (A,B) the inter-rater agreement are visualized as Bland Altman plots and limits of agreement. Clinical performance data was analyzed on the log-it scale to meet the assumptions of constant mean, SD, and normality.
The validity and reliability were tested by applying the TeamOBS-VAD checklist to 60 videos of real-life VADs. The inter-rater agreement for an individual rater had an intraclass correlation coefficient (ICC) of 0.73; 95% confidence interval (95% CI) of [0.58, 0.83], and that for the average of two raters ICC 0.84; 95% CI [0.73, 0.91]. The intra-rater agreement was tested as raters re-evaluated 11 videos, and the agreements had an ICC 0.74; 95% CI [0.24, 0.92] for rater one and ICC 0.90; 95% CI [0.67, 0.97] for rater 2. Agreement was described using the Bland–Altman plot. The limits used to indicate low, acceptable, and high clinical performance were a score of <60%, 60–84%, and 85–100%, respectively; Figure 3. The arguments for validity and reliability are presented in Table 1.
4 Discussion
4.1 Main findings
We used an international Delphi process to develop the TeamOBS-VAD checklist to evaluate the clinical performance of VAD. The checklist allowed the calculation of the total performance score. The validity and reliability of the TeamOBS-VAD checklist were high when applied by two raters. However, they were still acceptable when used by one rater.
4.2 Strengths and limitations
A crucial strength of the study is the international, diverse Delphi panel as this ensured a sensible construct along with clinical applicability and increased the possibilities for international adoption and acceptability (21). The panel’s feedback and commentaries rounds guaranteed substantiated adaptations in the described tasks as well of the framework of the checklist (13). We recognize that the inclusion of a larger number of experts and inclusion of other professional groups such as midwives in expert panels could have been of additional benefit. Furthermore, as all experts in the Delphi are based Northern Europe countries, we recognize that the checklist will apply primarily in these countries. However, with minor modifications the checklist may be useful in other countries as well (22).
The use of videos of real-life vacuum deliveries was also a significant strength to the validity and reliability (15). Informed consent was obtained for all included videos, fulfilling all Danish ethical and legal requirements. However, informed consent may introduce potential selection bias, as we cannot exclude the possibility that low-performing teams were less willing to provide consent (23). It was nevertheless reassuring that 95% of the obstetricians consented to include all their videos, and only two videos were deleted when the staff withdrew consent (24).
4.3 Interpretation
The Delphi process was valuable in ensuring the inclusion of different international perspectives in managing vacuum extraction. The respective panelists’ national guidelines had similarities; however, some elements differed (25). Discussions in the Delphi panel included the acceptable number of pulls, when to abandon the attempt, checking the position with ultrasound (and it’s weighting according to importance), and examining and suturing the perineum after delivery. The weighted score of pain relief was discussed, as some panelists thought the idea of delivery using a vacuum without considering further medical pain relief was inappropriate. However, others did not consider the need for additional analgesia. These different views may reflect the expectations of the panelists and the availability or use of epidural and spinal analgesia (26). Two Delphi panelists preferred to include non-technical skills such as communication, leadership, and teamwork in the TeamOBS-VAD.
The steering committee had a priori decided not to include non-technical skills in the checklist, as validated obstetric teamwork assessment tools already exist for this purpose (27). From a methodological point of view tools for rating non-technical skills should be independent of the actual clinical problem while the clinical performance ratings need to address these specifics. Thus, the Delphi process and framework of the checklist did focus on the clinical performance (24, 28).
We did not identify any published checklists which are specifically designed to produce a summative score for VAD performance. Previously published checklists have been designed as cognitive aids, for supporting procedural task execution, or evaluating using item-by-item feedback (5–7). These classic evaluation checklists use simple dichotomous items (i.e., not done/done); however, dichotomous items are often not sufficient for the assessment of more complex tasks. Therefore, tools for performance evaluation include more categories (e.g., not indicated/incorrectly performed/performed late/timely and correctly performed) (10). Other requirements of a performance checklist include weighted checklist items to differentiate between essential and less important actions and time frames to help create a more refined assessment of performance (8, 9). Therefore, we included both weighted items and more categories and these should be considered when developing future evaluation checklists.
The summative clinical performance score is useful in (a) assessing adherence to accepted guidelines, (b) supporting individual learning by mapping the learning curve, and (c) quality assessment within a labor and delivery ward (10). We developed the TeamOBS-VAD as a tool to produce a summative clinical performance score based on items weighted for importance. We ensured that difficult deliveries and deliveries in which the teams abandoned the vaginal delivery attempt did not automatically result in a low score. Conversely, simple outlet deliveries did not automatically result in high scores. The TeamOBS-VAD clinical performance score was not associated with the number of contractions if the number was acceptable, cephalic level, rotation, or position. Failed vacuum extraction was observed in 6% of the videos; none were teams with low clinical performance scores.
Educators and trainees have experienced difficulties developing and maintaining clinical competence in VAD because of reduced working hours and instrumental delivery rates (29). Therefore, we must rethink our learning path for vacuum extraction to improve and speed up trainees’ learning. The first step could be to systematically assess performance, as improving current methods is difficult if we do not measure them objectively (30). Filming vaginal deliveries could be a second step in rethinking our learning path because a systematic assessment of the performance of trainees or departments allows us to investigate our team’s performance and offer targeted training (31). Studies evaluating video use for educational purposes have reported high patient and staff acceptability and compliance (23). Notably, 30% of the staff found that filming provoked mild anxiety; however, they confirmed that the educational value outweighed it (32). Solving the ethical and legal issues associated with video recordings in emergency care may improve our knowledge and serve as a foundation for providing better patient care (33, 34).
External validity must be considered before applying the TeamOBS-VAD checklist in other settings. Our checklist reflects adherence to guidelines and accepted practices in Northern Europe. Thus, before applying the TeamOBS-VAD checklist in different settings, it may be necessary to agree with the present statements in the tool for “done correctly.” Significant differences in opinion, the checklist, and item weights of importance will be imprecise. In addition, scoring an abandoned attempted vaginal delivery is meaningless if an urgent cesarean section is unavailable. Validity testing was conducted in two Danish hospitals, and it may be necessary to re-evaluate the validity if the delivery guidance differs substantially.
5 Conclusion
The TeamOBS-VAD checklist we developed is valid, reliable, and easy to use in assessing clinical vacuum deliveries. It may help train individuals and evaluate team performance in a department.
Data availability statement
The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.
Ethics statement
This study was approved by the Central Denmark Region’s legal department, Danish Data Protection Agency (2012-58-006), and Central Denmark Region’s Research Foundation (Case No. 1-16-02-257-14). All the participants (patients and staff) volunteered to participate and provided informed consent. The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their written informed consent to participate in this study.
Author contributions
LB: Conceptualization, Data curation, Formal analysis, Funding acquisition, Investigation, Methodology, Project administration, Software, Validation, Writing – original draft. KH: Formal analysis, Methodology, Supervision, Writing – review & editing. OK: Conceptualization, Data curation, Funding acquisition, Resources, Supervision, Validation, Writing – review & editing. TM: Conceptualization, Methodology, Supervision, Validation, Writing – review & editing. NU: Conceptualization, Methodology, Supervision, Writing – review & editing. LH: Conceptualization, Data curation, Formal analysis, Investigation, Supervision, Validation, Visualization, Writing – review & editing.
Funding
The author(s) declare financial support was received for the research, authorship, and/or publication of this article. We thank the following organizations and departments for their financial support: Tryg Foundation (Trygfonden) (Grant ID no. 109507), Regional Hospital in Horsens, Department of Obstetrics and Gynecology, and Regional Postgraduate Medical Education Administration Office, North.
Acknowledgments
We thank the Delphi panel and all the patients who volunteered to participate in the study. We thank our colleagues who participated in this study and continued to strive for excellence. A special thanks to Consultant Lars Høj and Midwives Astrid Johansen and Lene Yding for assisting in testing the checklist. Our analysis was approved by Erik Parner, Professor of Statistics at Aarhus University.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Supplementary material
The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fmed.2024.1330443/full#supplementary-material
Abbreviations
Abbreviations: VAD, vacuum-assisted delivery; ICC, intraclass correlation coefficient; CI, confidence interval.
References
1. Muraca G, Sabr Y, Lisonkova S, Skoll A, Brant R, Cundiff G, et al. Morbidity and mortality associated with forceps and vacuum delivery at outlet, low, and midpelvic station. J Obstet Gynaecol Can. (2019) 41:327–37. doi: 10.1016/j.jogc.2018.06.018
2. Murphy D, Strachan B, Bahl R. Assisted vaginal birth: green-top guideline no. 26. BJOG. (2020) 127:e70–112. doi: 10.1111/1471-0528.16092
3. Gei A. Prevention of the first cesarean delivery: the role of operative vaginal delivery. Semin Perinatol. (2012) 36:365–73. doi: 10.1053/j.semperi.2012.04.021
4. Merriam A, Ananth CV, Wright J, Siddiq Z, D’Alton M, Friedman A. Trends in operative vaginal delivery, 2005–2013: a population-based study. BJOG. (2017) 124:1365–72. doi: 10.1111/1471-0528.14553
5. Mannella P, Giordano M, Guevara M, Giannini A, Russo E, Pancetti F, et al. Simulation training program for vacuum application to improve technical skills in vacuum-assisted vaginal delivery. BMC Preg Childbirth. (2021) 21:338. doi: 10.1186/s12884-021-03829-y
6. Maagaard M, Oestergaard J, Johansen M, Andersen L, Ringsted C, Ottesen B, et al. Vacuum extraction: Development and test of a procedure-specific rating scale. Acta Obstet Gynecol Scand. (2012) 91:1453–9. doi: 10.1111/j.1600-0412.2012.01526.x
7. Staat B, Combs C. SMFM Special Statement: Operative vaginal delivery: checklists for performance and documentation. Am J Obstet Gynecol. (2020) 222:B15–21. doi: 10.1016/j.ajog.2020.02.011
8. Donoghue A, Nishisaki A, Sutton R, Hales R, Boulet J. Reliability and validity of a scoring instrument for clinical performance during Pediatric Advanced Life Support simulation scenarios. Resuscitation. (2010) 81:331–6. doi: 10.1016/j.resuscitation.2009.11.011
9. Donoghue A, Durbin D, Nadel F, Stryjewski G, Kost S, Nadkarni V. Effect of high-fidelity simulation on pediatric advanced life support training in pediatric house staff a randomized trial. Pediatr Emerg Care. (2009) 25:139–44.
10. Schmutz J, Eppich W, Hoffmann F, Heimberg E, Manser T. Five steps to develop checklists for evaluating clinical performance: An integrative approach. Acad Med. (2014) 89:996–1005. doi: 10.1097/ACM.0000000000000289
11. Wolf H, Schaap A, Bruinse H, Smolders-de Haas H, van Ertbruggen I, Treffers P. Vaginal delivery compared with caesarean section in early preterm breech delivery: a comparison of long term outcome. Br J Obstet Gynaecol. (1999) 106:486–91.
12. Morgan P, Lam-McCulloch J, Herold-McIlroy J, Tarshis J. Simulation performance checklist generation using the Delphi technique. Can J Anaesth. (2007) 54:992–7. doi: 10.1007/BF03016633
13. Hsu C, Sandford B. The Delphi technique: making sense of consensus. Pract Assess Res Eval. (2007) 12:1–8. doi: 10.1016/S0169-2070(99)00018-7
14. Cook D, Brydges R, Ginsburg S, Hatala R. A contemporary approach to validity arguments: a practical guide to Kane’s framework. Med Educ. (2015) 49:560–75. doi: 10.1111/medu.12678
15. Cook D, Zendejas B, Hamstra S, Hatala R, Brydges R. What counts as validity evidence? Examples and prevalence in a systematic review of simulation-based assessment. Adv Health Sci Educ Theory Pract. (2014) 19:233–50. doi: 10.1007/s10459-013-9458-4
16. Menard M, Kilpatrick S, Saade G, Hollier L, Joseph G, Barfield W, et al. Levels of maternal care This document was developed jointly by the. Am J Obstet Gynecol. (2015) 212:259–71. doi: 10.1016/j.ajog.2014.12.030
17. Brogaard L, Hvidman L, Hinshaw K, Kierkegaard O, Manser T, Musaeus P, et al. Development of the TeamOBS-PPH – targeting clinical performance in postpartum hemorrhage. Acta Obstet Gynecol Scand. (2018) 97:13336. doi: 10.1111/aogs.13336
18. Carstensen B. Comparing Clinical Measurement Methods: A Practical Guide. Hoboken, NJ: Wiley (2010). doi: 10.1136/pgmj.59.687.72-a
19. Koo T, Li M. A guideline of selecting and reporting intraclass correlation coefficients for reliability research. J Chiropr Med. (2016) 15:155–63. doi: 10.1016/j.jcm.2016.02.012
20. Bland J, Altman D. Applying the right statistics: analyses of measurement studies. Ultrasound Obstet Gynecol. (2003) 22:85–93. doi: 10.1002/uog.122
21. Barrett D, Heale R. What are Delphi studies? Evid Based Nurs. (2020) 23:68–9. doi: 10.1136/ebnurs-2020-103303
22. Schaap T, Bloemenkamp K, Deneux-Tharaux C, Knight M, Langhoff-Roos J, Sullivan E, et al. Defining definitions: a Delphi study to develop a core outcome set for conditions of severe maternal morbidity. BJOG. (2019) 126:394–401. doi: 10.1111/1471-0528.14833
23. Brogaard L, Uldbjerg N. Filming for auditing of real-life emergency teams: a systematic review. BMJ Open Qual. (2019) 8:e000588. doi: 10.1136/bmjoq-2018-000588
24. Brogaard L, Kierkegaard O, Hvidman L, Jensen K, Musaeus P, Uldbjerg N, et al. The importance of non-technical performance for teams managing postpartum haemorrhage: video review of 99 obstetric teams. BJOG. (2019) 126:15655. doi: 10.1111/1471-0528.15655
25. Tsakiridis I, Giouleka S, Mamopoulos A, Athanasiadis A, Daniilidis A, Dagklis T. Operative vaginal delivery: A review of four national guidelines. J Perinat Med. (2020) 48:189–98. doi: 10.1515/jpm-2019-0433
26. King T. Epidural anesthesia in labor. Benefits versus risks. J Nurse Midwifery. (1997) 42:377–88.
27. Onwochei D, Halpern S, Balki M. Teamwork assessment tools in obstetric emergencies: A systematic review. Simul Healthc. (2017) 12:165–76. doi: 10.1097/SIH.0000000000000210
28. Marks M, Mathieu J, Zaccaro S. A temporally based framework and taxonomy of team processes. Acad Manage Rev. (2001) 26:356–76.
29. Baskett T. Operative vaginal delivery – An historical perspective. Best Pract Res Clin Obstet Gynaecol. (2019) 56:3–10. doi: 10.1016/j.bpobgyn.2018.08.002
30. Manser T, Brösterhaus M, Hammer A. You can’t improve what you don’t measure: Safety climate measures available in the German-speaking countries to support safety culture development in healthcare. Z Evid Fortbild Qual Gesundhwes. (2016) 114:58–71. doi: 10.1016/j.zefq.2016.07.003
31. Bahl R, Murphy D, Strachan B. Non-technical skills for obstetricians conducting forceps and vacuum deliveries: qualitative analysis by interviews and video recordings. Eur J Obstet Gynecol Reprod Biol. (2010) 150:147–51. doi: 10.1016/j.ejogrb.2010.03.004
32. Davis L, Johnson L, Allen S, Kim P, Sims C, Pascual J, et al. Practitioner perceptions of trauma video review. J Trauma Nurs. (2013) 20:150–4. doi: 10.1097/JTN.0b013e3182a172b6
33. Townsend R, Clark R, Ramenofsky M, Diamond DL. ATLS-based videotape trauma resuscitation review: education and outcome. J Trauma. (1993) 34:133–8. doi: 10.1097/00005373-199301000-00025
Keywords: performance, emergency, obstetric, vacuum extraction, team, video, checklist
Citation: Brogaard L, Hinshaw K, Kierkegaard O, Manser T, Uldbjerg N and Hvidman L (2024) Developing the TeamOBS-vacuum-assisted delivery checklist to assess clinical performance in a vacuum-assisted delivery: a Delphi study with initial validation. Front. Med. 11:1330443. doi: 10.3389/fmed.2024.1330443
Received: 30 October 2023; Accepted: 08 January 2024;
Published: 02 February 2024.
Edited by:
Fedde Scheele, VU Amsterdam, NetherlandsReviewed by:
Florian Recker, University of Bonn, GermanyJessica Van Der Aa, Academic Medical Center, Netherlands
Copyright © 2024 Brogaard, Hinshaw, Kierkegaard, Manser, Uldbjerg and Hvidman. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Lise Brogaard, bGJyakBjbGluLmF1LmRr