- 1Victor Horsley Department of Neurosurgery, National Hospital for Neurology and Neurosurgery, London, United Kingdom
- 2Wellcome/EPSRC Centre for Interventional and Surgical Sciences, University College London, London, United Kingdom
Background: An exoscope heralds a new era of optics in surgery. However, there is limited quantitative evidence describing and comparing the learning curve.
Objectives: This study aimed to investigate the learning curve, plateau, and rate of novice surgeons using an Olympus ORBEYE exoscope compared to an operating microscope (Carl Zeiss OPMI PENTERO or KINEVO 900).
Methods: A preclinical, randomized, crossover, noninferiority trial assessed the performance of seventeen novice and seven expert surgeons completing the microsurgical grape dissection task “Star’s the limit.” A standardized star was drawn on a grape using a stencil with a 5 mm edge length. Participants cut the star and peeled the star-shaped skin off the grape with microscissors and forceps while minimizing damage to the grape flesh. Participants repeated the task 20 times consecutively for each optical device. Learning was assessed using model functions such as the Weibull function, and the cognitive workload was assessed with the NASA Task Load Index (NASA-TLX).
Results: Seventeen novice (male:female 12:5; median years of training 0.4 [0–2.8 years]) and six expert (male:female 4:2; median years of training 10 [8.9–24 years]) surgeons were recruited. “Star’s the limit” was validated using a performance score that gave a threshold of expert performance of 70 (0–100). The learning rate (ORBEYE −0.94 ± 0.37; microscope −1.30 ± 0.46) and learning plateau (ORBEYE 64.89 ± 8.81; microscope 65.93 ± 9.44) of the ORBEYE were significantly noninferior compared to those of the microscope group (p = 0.009; p = 0.027, respectively). The cognitive workload on NASA-TLX was higher for the ORBEYE. Novices preferred the freedom of movement and ergonomics of the ORBEYE but preferred the visualization of the microscope.
Conclusions: This is the first study to quantify the ORBEYE learning curve and the first randomized controlled trial to compare the ORBEYE learning curve to that of the microscope. The plateau performance and learning rate of the ORBEYE are significantly noninferior to those of the microscope in a preclinical grape dissection task. This study also supports the ergonomics of the ORBEYE as reported in preliminary observational studies and highlights visualization as a focus for further development.
Introduction
The operating microscope pioneered in the 1950s by Yasagil (1) remains the gold standard for microneurosurgery. More recently, an “exoscope” system has been introduced as a potential alternative to the microscope (2). Suggested benefits of an exoscope include improved ergonomics and being a valuable educational tool (2–5). A newly developed exoscope is an ORBEYE (Olympus, Tokyo, Japan, 2017), equipped with 3D optics, 4 K imaging quality, and comparable field of view and depth of field to those of the microscope.
The safe introduction of novel technology into clinical practice is central to reducing patient harm (6, 7). Surgeons gain procedural competence as their experience increases with a device (8, 9). This relationship between learning effort and the outcome can be represented using learning curves (10, 11). Factors that affect the learning curve are the initial skill level, the learning rate, and the final skill level achieved—known as the learning plateau (10, 12, 13). Understanding learning curves, both at individual and system levels, is crucial for assessing a new surgical technique or technology, informing surgical training, and evaluating procedures in practice (14, 15). Previous comparative studies suggest the presence of a learning curve for experienced surgeons with the ORBEYE, but there has been no attempt at quantification of the learning curve nor a direct comparison of the learning curve for both the microscope and ORBEYE in relation to novice surgeons (5, 16–18).
We explored the learning rate, learning plateau, and cognitive load of novice surgeons performing a validated microsurgical grape dissection task.
We performed a microsurgical grape dissection task to explore the learning rate, learning plateau, and cognitive load of novice surgeons with limited experience of both the microscope and OREYE; the learning curve of the ORBEYE is not inferior to that of the traditional microscope.
Methods
Protocol and Ethics
The protocol was registered with the local Clinical Governance Committee and was approved by the Institutional Review Board. The Consolidated Standards of Reporting Trials Statement (19) (CONSORT) with noninferiority extension was used.
Participants
Novice and expert surgeons were recruited from a university hospital. Novice surgeons had not performed any operative cases on either the microscope or the ORBEYE. Expert surgeons had completed their neurosurgical training (20, 21). Informed written consent was obtained.
Sample Size
A target sample size of 12 novices was set. Owing to pragmatic constraints and the lack of applicable pilot data, no power calculation was undertaken, but such a number was deemed appropriate based on previous similar studies (20–23).
Randomization
Novice surgeons were randomly allocated to start on either optical device before crossing over. Permuted blocked randomization (block size 2 and 4) using a computer-generated sequence. One author (ZM) performed sequence generation and implementation. Blinding was not possible due to the nature of optical devices.
Interventions
Microsurgical Grape Dissection Task: “Star’s the Limit”
Participants performed a validated microsurgical task “Star’s the limit” (24). A standardized star is drawn on a grape using a stencil with a 5 mm edge length. Participants cut the star and peeled the star-shaped skin off the grape with microscissors and forceps while minimizing damage to the grape flesh (Figure 1). Each novice repeated the task 20 times consecutively before changing the device and repeating the task a further 20 times consecutively. The microsurgical task was validated by experts who repeated the task 20 times consecutively using the microscope only. No feedback or teaching was provided to the participants during the task. If participants were not able to finish within 5 min, they were told to stop, and the next repetition would begin. The microscopes used were an OPMI PENTERO or a KINEVO 900 (Carl Zeiss Co, Oberkochen, Germany).
Figure 1. Microsurgical grape dissection task “Star’s the limit” setup. (A) Microscope trial. (B) “Star’s the limit.” Note the grape has a homogeneous shape, drawn on by stencil, and secured using needles to ensure constant position across the trials.
Outcomes
Primary Outcomes: Learning Plateau, Learning Rate, and Learning Curves
The performance of participants in the task was assessed using a five-item rubric: (1) time to completion; (2) completeness of dissection; (3) degree of grape flesh attached to the star; (4) number of edges incised within the drawn lines; and (5) perforation of grape flesh with the instruments (Table 1). The same assessor (ZM) prepared all 820 grapes and assessed all 17 novice and 7 expert surgeons to reduce scoring intravariability. The raw scores were combined into a composite performance score for each repetition. There was a modification to the performance score algorithm from the original protocol (OpenEd@UCL repository; https://open-education-repository.ucl.ac.uk//620/). The performance score was calculated as follows:
Numerous different learning curve models were tested (Appendix 1). The best-fitting curve was selected using log-likelihood. All curves tested are characterized by the rate of improvement decreasing over time (Appendix 1). The curves were fitted with nonlinear mixed-effects models. Fixed effects were used to
1. investigate the difference between the microscope and ORBEYE and
2. establish the coefficients of the curves.
Random effects were used to account for
1. nonindependence of the data from the same subject and
2. random variations in coefficients between subjects.
The fixed effects of the model output gave the curves averaged across each group. These fixed-effects outputs were used to test the noninferiority of the asymptote and the learning rates of the ORBEYE group compared to thosed of the microscope group. The inferiority threshold was set a priori at 20%, and a one-tailed t-test was used against this inferiority margin to give the plateau performance.
Finally, crossover analysis was performed to evaluate whether the starting optical device (ORBEYE or microscope) had an impact on novices’ performance. The final five trials before and after crossover were considered. The performance score was investigated before and after crossover (fixed-effects), and the difference between each group (fixed-effects) was taken into account by the subject (random-effects). The analysis was performed using linear mixed-effects regression.
Secondary Outcomes: Subjective Impression of Optical Devices
The perceived workload was assessed using the NASA Raw Task Load Index (NASA R-TLX) (25, 26) (Appendix 2). Within the NASA R-TLX are six domains: mental, physical, and temporal demands, performance, effort, and frustration, and these are rated using a 20-point scale. Participants completed the NASA R-TLX immediately after finishing the task. The domain score and total score were used for the secondary outcome. Novice surgeons also reported their subjective impression of the microscope and ORBEYE (Appendix 3).
Statistical Methods
Curve fitting was performed using R v4.1.2 (27); linear mixed-effects regression analysis was made using packages lme4 v1.1 (28), lmerTest v3.1 (29), and nlme v3.1 (30). Noninferiority testing of the learning curves was conducted using the outputs for the estimates, standard errors, and the degrees of freedom of the growth curve coefficients by utilizing the base R functions for Student’s t distributions. Subjective impression analysis was conducted using JASP v0.14.1 and GraphPad Prism v9.2.0. Data are expressed as mean ± 95% confidence intervals or median ± IQR. The threshold for statistical significance was set at α < 0.05. Adjustments for multiple comparisons were made using the Benjamini–Hochberg method for false discovery rates.
Results
Participants
Seventeen novice and seven expert surgeons were recruited (Appendix 4). The novice surgeons (male:female 12:5) had completed a median of 0.4 years of training (0–2.8 years). No novice surgeon had not performed any microsurgical cases using either the ORBEYE or a microscope. The expert surgeons (male:female 4:2) had a median 10 years of training (8.9–24 years) and completed the “Star’s the limit” task using the microscope only to validate the grape dissection model.
Validation of the Composite Performance Score
The time to task completion was the slowest during the first attempts and plateaued by the 16th attempt (Appendix 5). To compare “absolute novice” and “absolute expert” for validation purposes, the first five attempts for novices and the final five attempts for experts were considered. The time taken for task completion by novices and experts demonstrates that participant 2 (expert) and 10 (novice) are outliers and was excluded from the validation analysis.
The Akaike information criterion (AIC) was used to select the best predictive model for discrimination between novice and expert (Table 2). The best performing model was time taken to task completion alone (AIC: 13.1) and the second-best performing model was time taken + edge score (AIC: 15.0). We elected to utilize the latter model despite the higher AIC value to ensure penalization for a fast performance if performed poorly. A validated performance score was created with a threshold of expert performance of 70 (0–100) (Appendix 6).
Table 2. Summary of best selective model function for discrimination between novice and expert performances.
Investigating the Learning Curve
We evaluated the seven different models, and the best fit was the modified Weibull function (AIC 2389.013; log-likelihood −1181.506) (Figure 2; Appendix 7). The primary outcome of the learning rate (ORBEYE −0.94 ± 0.37; microscope −1.30 ± 0.46) and learning plateau (ORBEYE 64.89 ± 8.81; microscope 65.93 ± 9.44) of the ORBEYE was significantly noninferior compared to that of the microscope group (p = 0.009; p = 0.027, respectively) (Figure 3). If considering participants crossing over to the ORBEYE from the microscope, the plateau is not (but nearly significant) noninferior (p = 0.055). ANOVA analysis demonstrates that performance significantly improved after crossover in both groups (microscope to ORBEYE and ORBEYE to microscope).
Figure 2. Composite performance score of novice surgeons completing the microsurgical grape dissection task with a threshold for expert performance (gray dashed line 70; 0–100). Each graph represents a separate novice. In total, participants completed 20 repetitions of the task on each device consecutively. (A) Novice surgeons’ performance scores plotted against the number of trials performed, with the group starting with the ORBEYE and microscope. The first colored graph represents the first device, and the second colored graph represents the crossover to the second device. (B) Modeled learning curves using the modified Weibull function.
Figure 3. (A) Plateau performance between the novices, starting the “Star’s the limit” task on the exoscope or the microscope. The performance score was 0–100, generated through methods outlined in section Methods, with an expert threshold score of 70. (B) Significant noninferiority of the microscope and exoscope based on novice performance compared against the expert performance. (C) Learning rate of novices on the exoscope or the microscope. (D) Learning rate and statistical noninferiority of the exoscope compared to the microscope.
Workload Assessments
The NASA R-TLX demonstrates that mental demand, performance, effort, and frustration scores were not statistically different between the ORBEYE and the microscope (Figure 4A). Analysis of variance (repeated-measure two-way ANOVA) showed no significant main effect of device and group assignment on the workload scores, and a significant interaction was only found in the temporal demand score (p < .05). However, post-hoc comparisons with FDR adjustments did not find significant results.
Figure 4. Subjective impression of the optical device. (A) NASA Raw Task Load Index (NASA R-TLX) score for each dimension and total workload compared between novices with different instruments. (B) Subjective questionnaire (Supplementary Material Table 2): Q1: better visualization; Q2: greater freedom of movement; Q3 more comfortable; Q4: easier to perform a task with; and Q5: prefer to use in the future.
Subjective Impression
Novice surgeons answered the subjective questionnaire following task completion on each device (Figure 4B; Appendix 8). For Q1, 75% (12/16) of novices strongly preferred or preferred the microscope compared to the ORBEYE regarding visualization. Further, 55% (9/16) strongly preferred or preferred the microscope compared to the ORBEYE for task completion. Novices strongly preferred or preferred the greater movement (55%; 9/16) and comfortable ergonomics (63%; 10/16) of the ORBEYE compared to the microscope. Seven novices compared to five novices would prefer to use the microscope rather than the ORBEYE in the future (Figure 4B).
Discussion
Interpretation
This preclinical, randomized, crossover noninferiority trial provides robust data that quantifies for the first time that the learning curve of the ORBEYE is statistically noninferior regarding the learning rate and learning plateau compared to that of the traditional operating microscope in novice surgeons (Figure 3).
The modified Weibull function was the best fit to model learning curves using the composite performance scores. The microsurgical task was satisfactory at discriminating between novice and expert surgeons using the composite performance score. This adds validity to the modeled learning curves as representatives of real-world learning curves.
There was no significant difference in the cognitive workload of the optical devices using the NASA R-TLX. The total workload score was 61 (IQR: 44.75–72.75) for the ORBEYEE and 53 (IQR: 38.75–63) for the microscope. Analysis of variance demonstrated no significant difference in the scores depending on the optical device the novice surgeon started on.
Subjective assessment of the optic devices found that novices preferred the visualization, ease of task completion, and preferential future use of the microscope, while they preferred the ergonomics and greater freedom of movement of the ORBEYE.
Taken together, and considering the microscope has been standard practice for decades, this study does not intend to definitively state that the ORBEYE should replace the microscope. Instead, the ORBEYE has a similar learning plateau and learning rate to the microscope in novice surgeons on a preclinical task; further work must be undertaken to facilitate safe, comparative clinical studies.
Comparison with the Literature
The “learning curve” is frequently used in surgical education literature and represents the relationship between learning effort and the outcome (11, 12, 15, 31). Understanding the learning curve, rate, and plateau provides a mechanism for understanding the development of procedural competency (7, 15). The learning curve is vital when introducing novel technology to surgical practice and should be established before any definitive comparative clinical trials (6, 32, 33). The current ORBEYE literature describes the learning curve subjectively or is inferred from a sample of experienced surgeons (16, 17, 34, 35). The present study provides quantitative data that models the learning curve for novices for both the microscope and the ORBEYE. We demonstrate no significant inferiority for either optical device. This should encourage the international community to ensure trainees develop skills for both optical devices.
The subjective impression from the novice surgeons in this study prefers or strongly prefers the ORBEYE’s freedom of movement and comfortable position. This supports existing literature and descriptive studies (35, 35–37). Regarding visualization, participants strongly preferred or preferred the microscope compared to the ORBEYE (Figure 4). Microscope preference might have been influenced by obstruction of the line of sight or increased “noise” in peripheral vision while completing novel tasks. The subjective feedback supports future ORBEYE development to enhance their visualization. Future advances to the ORBEYE may add further functionality, including the possibility of augmenting data flow due to the digital nature of the ORBEYE, permitting interoperability with other technological innovations such as augmented reality or computer vision.
Strengths and Limitations
The aim was to compare the learning curves of the ORBEYE and microscope in novices. Previous studies ask experts to perform simulations using the ORBEYE or microscope or ask for subjective feedback after performing surgery with the exoscope. This introduces bias based on their previous experience and likely established preference. It was therefore considered more robust to use a single, large, homogeneous group of novice surgeons with limited surgical experience, rather than a small group of experts with varying experience of optical device and technical expertise. We did validate the task with expert surgeons to characterize the learning curve for novices. To ensure we avoided any learning curve with the visualization device, we felt it best for expert surgeons to complete the task on the device they were most familiar with, in this case, the microscope. The validated low-fidelity task was also appropriate to characterize the learning curve, as it was not too easy that novices could perform perfectly but not too challenging that a plateau was not achieved at the end of 20 consecutive repetitions. Our methods were also published a priori; participants were randomized to reduce the risk of bias and modeled over 60 variants of the performance score.
A limitation was using a low-fidelity microsurgical grape dissection task. Although this has precedence within the literature, our findings require further validation with higher fidelity models when evaluating an expert’s learning curve, such as suturing or anastomoses. The novice sample size is small, although again concordant with the literature. Finally, no participant did not crossover; therefore, we cannot control for the effect of switching optical devices.
Conclusion
This is the first study to quantify the ORBEYE learning curve and the first randomized controlled trial to compare the ORBEYE learning curve to the microscope. The plateau performance and learning rate of the ORBEYE are significantly noninferior to those of the microscope in a preclinical grape dissection task. This study also supports the ergonomics of the ORBEYE as reported in preliminary observational studies and highlights visualization as a focus for further future development.
Data Availability Statement
The raw data supporting the conclusions of this article/Supplementary Material will be made available by the authors without undue reservation.
Ethics Statement
The studies involving human participants were reviewed and approved by the National Hospital for Neurology and Neurosurgery, London, UK. The patients/participants provided their written informed consent to participate in this study.
Author Contributions
HLH, ZM, CHK, DZK, WM, DS, and HJM conceived, created, and developed the idea and protocol. HLH and ZM led the data collection with CHK, DZK, WM, and HJM assistance. HLH, ZM, CHK, and DZK wrote the first draft of the manuscript. CHK undertook the statistical analysis with support from HLH and ZM. DS and HJM critically reviewed and edited the manuscript. All authors contributed to the article and approved the submitted version.
Funding
HLH, HJM, CHK, and WM are supported by the Wellcome/EPSRC Centre for Interventional and Surgical Sciences, University College London. DZK is supported by an NIHR Academic Clinical Fellowship. HJM is also funded by the NIHR Biomedical Research Centre at University College London.
Acknowledgments
The authors thank the Theatre Staff at the National Hospital for Neurology and Neurosurgery for facilitating this trial. The Olympus ORBEYE was loaned to the National Hospital for Neurology and Neurosurgery as part of a reference center agreement.
Supplementary Material
The Supplementary Material for this article can be found online at: https://journal.frontiersin.org/article/10.3389/fsurg.2022.920252/full#supplementary-material
Conflict of Interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher's Note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
1. Yaşargi G. A legacy of microneurosurgery: memoirs, lessons, and axioms. Neurosurgery. (1999) 45(5):1025–92. doi: 10.1097/00006123-199911000-00014
2. Nishiyama K. From exoscope into the next generation. J Korean Neurosurg Soc. (2017) 60(3):289–93. doi: 10.3340/jkns.2017.0202.003
3. Mamelak AN, Danielpour M, Black KL, Hagike M, Berci G. A high-definition exoscope system for neurosurgery and other microsurgical disciplines: preliminary report. Surg Innov. (2008) 15(1):38–46. doi: 10.1177/1553350608315954
4. Mamelak AN, Nobuto T, Berci G. Initial clinical experience with a high-definition exoscope system for microneurosurgery. Neurosurgery. (2010) 67(2):476–83. doi: 10.1227/01.NEU.0000372204.85227.BF
5. Takahashi S, Toda M, Nishimoto M, Ishihara E, Miwa T, Akiyama T. Pros and cons of using ORBEYE™ for microneurosurgery. Clin Neurol Neurosurg. (2018) 174:57–62. doi: 10.1016/j.clineuro.2018.09.010
6. Sachdeva AK, Russell TR. Safe introduction of new procedures and emerging technologies in surgery: education, credentialing, and privileging. Surg Clin North Am. (2007) 87(4):853–66. doi: 10.1016/j.suc.2007.06.006
7. Aggarwal R, Darzi A. Innovation in surgical education – a driver for change. The Surgeon. (2011) 9:S30–1. doi: 10.1016/j.surge.2010.11.021
8. Healey P, Samanta J. When does the “Learning Curve” of innovative interventions become questionable practice? Eur J Vasc Endovasc Surg. (2008) 36(3):253–7. doi: 10.1016/j.ejvs.2008.05.006
9. Valsamis EM, Chouari T, O’Dowd-Booth C, Rogers B, Ricketts D. Learning curves in surgery: variables, analysis and applications. Postgrad Med J. (2018) 94(1115):525–30. doi: 10.1136/postgradmedj-2018-135880
10. Valsamis EM, Golubic R, Glover TE, Husband H, Hussain A, Jenabzadeh AR. Modeling learning in surgical practice. J Surg Educ. (2018) 75(1):78–87. doi: 10.1016/j.jsurg.2017.06.015
11. Hopper AN, Jamison MH, Lewis WG. Learning curves in surgical practice. Postgrad Med J. (2007) 83(986):777–9. doi: 10.1136/pgmj.2007.057190
12. Cook JA, Ramsay CR, Fayers P. Using the literature to quantify the learning curve: a case study. Int J Technol Assess Health Care. (2007) 23(2):255–60. doi: 10.1017/S0266462307070341
13. Papachristofi O, Jenkins D, Sharples LD. Assessment of learning curves in complex surgical interventions: a consecutive case-series study. Trials. (2016) 17(1):266. doi: 10.1186/s13063-016-1383-4
14. Harrysson IJ, Cook J, Sirimanna P, Feldman LS, Darzi A, Aggarwal R. Systematic review of learning curves for minimally invasive abdominal surgery: a review of the methodology of data collection, depiction of outcomes, and statistical analysis. Ann Surg. (2014) 260(1):37–45. doi: 10.1097/SLA.0000000000000596
15. Simpson AHRW, Howie CR, Norrie J. Surgical trial design – learning curve and surgeon volume. Bone Jt Res. (2017) 6(4):194–5. doi: 10.1302/2046-3758.64.BJR-2017-0051
16. Siller S, Zoellner C, Fuetsch M, Trabold R, Tonn JC, Zausinger S. A high-definition 3D exoscope as an alternative to the operating microscope in spinal microsurgery. J Neurosurg Spine. (2020) 33(5):705–14. doi: 10.3171/2020.4.SPINE20374
17. Kwan K, Schneider JR, Du V, Falting L, Boockvar JA, Oren J. Lessons learned using a high-definition 3-Dimensional exoscope for spinal surgery. Oper Neurosurg. (2019) 16(5):619–25. doi: 10.1093/ons/opy196
18. Vetrano IG, Acerbi F, Falco J, D’Ammando A, Devigili G, Nazzi V. High-Definition 4 K 3D exoscope (ORBEYETM) in peripheral nerve sheath tumor surgery: a preliminary, explorative, pilot study. Oper Neurosurg. (2020) 19(4):480–8. doi: 10.1093/ons/opaa090
19. Piaggio G, Elbourne DR, Pocock SJ, Evans SJW, Altman DG, CONSORT Group. Reporting of noninferiority and equivalence randomized trials: extension of the CONSORT 2010 statement. JAMA. (2012) 308(24):2594–604. doi: 10.1001/jama.2012.87802
20. Marcus HJ, Hughes-Hallett A, Pratt P, Yang GZ, Darzi A, Nandi D. Validation of martyn to simulate the keyhole supraorbital subfrontal approach. Bull R Coll Surg Engl. (2014) 96(4):120–1. doi: 10.1308/rcsbull.2014.96.4.120
21. Marcus HJ, Payne CJ, Kailaya-Vasa A, Griffiths S, Clark J, Yang G-Z. A “smart” force-limiting instrument for microsurgery: laboratory and in vivo validation. Plos One. (2016) 11(9):e0162232. doi: 10.1371/journal.pone.0162232
22. Pafitanis G, Hadjiandreou M, Alamri A, Uff C, Walsh D, Myers S. The exoscope versus operating microscope in microvascular surgery: a simulation noninferiority trial. Arch Plast Surg. (2020) 47(3):242–9. doi: 10.5999/aps.2019.01473
23. Volpe A, Ahmed K, Dasgupta P, Ficarra V, Novara G, van der Poe H. Pilot validation study of the european association of urology robotic training curriculum. Eur Urol. (2015) 68(2):292–9. doi: 10.1016/j.eururo.2014.10.025
24. Masud D, Haram N, Moustaki M, Chow W, Saour S, Mohanna PN. Microsurgery simulation training system and set up: an essential system to complement every training programme. J Plast Reconstr Aesthet Surg. (2017) 70(7):893–900. doi: 10.1016/j.bjps.2017.03.009
25. Hart SG, Staveland LE. Development of NASA-TLX (task load index): results of empirical and theoretical research. In: Hancock PA, Meshkati N, editors. Advances in psychology. Vol 52. North-Holland: Human Mental Workload (1988). p. 139–83. doi: 10.1016/S0166-4115(08)62386-9
26. Hart SG. Nasa-Task load index (NASA-TLX). 20 Years later. Proc Hum Factors Ergon Soc Annu Meet. (2006) 50(9):904–8. doi: 10.1177/154193120605000909
27. R Core Team. R: A language and environment for statistical computing. R Foundation for Statistical Computing, Vienna, Austria (2021). https://www.R-project.org/
28. Bates D, Mächler M, Bolker B, Walker S. Fitting linear mixed-effects models using lme4. J Stat Softw. (2015) 67:1–48. doi: 10.18637/jss.v067.i01
29. Kuznetsova A, Brockhoff PB, Christensen RHB. lmerTest package: tests in linear mixed effects models. J Stat Softw. (2017) 82:1–26. doi: 10.18637/jss.v082.i13
30. Pinheiro J, Bates D, DebRoy S, Sarkar D. Nlme: Linear and Nonlinear Mixed Effects Models.; 2021. Accessed December 18, 2021. https://CRAN.R-project.org/package=nlme).
31. Feldman LS, Cao J, Andalib A, Fraser S, Fried GM. A method to characterize the learning curve for performance of a fundamental laparoscopic simulator task: defining “learning plateau” and “learning rate.” Surgery. (2009) 146(2):381–6. doi: 10.1016/j.surg.2009.02.021
32. Marcus HJ, Bennett A, Chari A, Day T, Hirst A, Hughes-Hallett A. IDEAL-D framework for device innovation: a consensus statement on the preclinical stage. Ann Surg. (2022) 275(1):73–9. doi: 10.1097/SLA.0000000000004907
33. McCulloch P, Altman DG, Campbell WB, Flum DR, Glasziou P, Marshall JC. No surgical innovation without evaluation: the IDEAL recommendations. The Lancet. (2009) 374(9695):1105–12. doi: 10.1016/S0140-6736(09)61116-8
34. Sack J, Steinberg JA, Rennert RC, Hatefi D, Pannell JS, Levy M. Initial experience using a high-definition 3-dimensional exoscope system for microneurosurgery. Oper Neurosurg. (2018) 14(4):395–401. doi: 10.1093/ons/opx145
35. Rösler J, Georgiev S, Roethe AL, Chakkalakal D, Acker G, Dengler NF. Clinical implementation of a 3D4K-exoscope (Orbeye) in microneurosurgery. Neurosurg Rev. (2022) 45(1):627–35. doi: 10.1007/s10143-021-01577-3
36. Ricciardi L, Chaichana KL, Cardia A, Stifano V, Rossini Z, Olivi A. The exoscope in neurosurgery: an innovative “point of view”. a systematic review of the technical, surgical, and educational aspects. World Neurosurg. (2019) 124:136–44. doi: 10.1016/j.wneu.2018.12.202
Keywords: microscope, exoscope, neurosurgery, learning curve, innovation, education, surgery
Citation: Layard Horsfall H, Mao Z, Koh CH, Khan DZ, Muirhead W, Stoyanov D and Marcus HJ (2022) Comparative Learning Curves of Microscope Versus Exoscope: A Preclinical Randomized Crossover Noninferiority Study. Front. Surg. 9:920252. doi: 10.3389/fsurg.2022.920252
Received: 14 April 2022; Accepted: 12 May 2022;
Published: 6 June 2022.
Edited by:
Cesare Zoia, San Matteo Hospital Foundation (IRCCS), ItalyReviewed by:
Carlo Giussani, University of Milano-Bicocca, ItalyPierlorenzo Veiceschi, University of Insubria, Italy
Copyright © 2022 Layard Horsfall, Mao, Koh, Khan, Muirhead, Stoyanov and Marcus. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Hugo Layard Horsfall hugo.layardhorsfall@ucl.ac.uk
Specialty section: This article was submitted to Neurosurgery, a section of the journal Frontiers in Surgery