Skip to main content

ORIGINAL RESEARCH article

Front. Neurosci., 20 January 2022
Sec. Brain Imaging Methods
This article is part of the Research Topic Advanced Computational Intelligence Methods for Processing Brain Imaging Data View all 62 articles

Continuous Monitoring of Mental Load During Virtual Simulator Training for Laparoscopic Surgery Reflects Laparoscopic Dexterity: A Comparative Study Using a Novel Wireless Device

\r\nNeta B. Maimon,*&#x;Neta B. Maimon1,2*†Maxim Bez&#x;Maxim Bez3†Denis Drobot,Denis Drobot4,5Lior MolchoLior Molcho2Nathan Intrator,Nathan Intrator2,6Eli KakiashvilliEli Kakiashvilli5Amitai Bickel,\r\nAmitai Bickel4,5
  • 1The School of Psychological Sciences, Tel Aviv University, Tel Aviv, Israel
  • 2Neurosteer Ltd, Herzliya, Israel
  • 3Medical Corps, Israel Defense Forces, Ramat Gan, Israel
  • 4Faculty of Medicine in the Galilee, Bar-Ilan University, Safed, Israel
  • 5Department of Surgery A, Galilee Medical Center, Nahariyya, Israel
  • 6Blavatnik School of Computer Science, Tel Aviv University, Tel Aviv, Israel

Introduction: Cognitive Load Theory (CLT) relates to the efficiency with which individuals manipulate the limited capacity of working memory load. Repeated training generally results in individual performance increase and cognitive load decrease, as measured by both behavioral and neuroimaging methods. One of the known biomarkers for cognitive load is frontal theta band, measured by an EEG. Simulation-based training is an effective tool for acquiring practical skills, specifically to train new surgeons in a controlled and hazard-free environment. Measuring the cognitive load of young surgeons undergoing such training can help to determine whether they are ready to take part in a real surgery. In this study, we measured the performance of medical students and interns in a surgery simulator, while their brain activity was monitored by a single-channel EEG.

Methods: A total of 38 medical students and interns were divided into three groups and underwent three experiments examining their behavioral performances. The participants were performing a task while being monitored by the Simbionix LAP MENTOR. Their brain activity was simultaneously measured using a single-channel EEG with novel signal processing (Aurora by Neurosteer®). Each experiment included three trials of a simulator task performed with laparoscopic hands. The time retention between the tasks was different in each experiment, in order to examine changes in performance and cognitive load biomarkers that occurred during the task or as a result of nighttime sleep consolidation.

Results: The participants’ behavioral performance improved with trial repetition in all three experiments. In Experiments 1 and 2, delta band and the novel VC9 biomarker (previously shown to correlate with cognitive load) exhibited a significant decrease in activity with trial repetition. Additionally, delta, VC9, and, to some extent, theta activity decreased with better individual performance.

Discussion: In correspondence with previous research, EEG markers delta, VC9, and theta (partially) decreased with lower cognitive load and higher performance; the novel biomarker, VC9, showed higher sensitivity to lower cognitive load levels. Together, these measurements may be used for the neuroimaging assessment of cognitive load while performing simulator laparoscopic tasks. This can potentially be expanded to evaluate the efficacy of different medical simulations to provide more efficient training to medical staff and measure cognitive and mental loads in real laparoscopic surgeries.

Introduction

Medical training of surgical interns in recent years utilizes digital simulators as an effective tool for acquiring practical skills in a controlled and hazard-free environment. These simulators help train interns and provide their performance outputs for each task, thus supporting the learning process. The surgery simulators offer practice of various tasks and present an evaluation of the performance via behavioral measurements (i.e., performance accuracy and performance length, etc.). Nevertheless, the simulator output measurements have two main drawbacks. First, althoughthe behavioral performance serves as a good indication of laparoscopic dexterity, it does not give an indication regarding the student’s cognitive state and, specifically, his or her cognitive load level during the task performance. Assessing cognitive load and stress levels directly during the surgery simulation is essential to evaluate the intern’s ability to engage in real-life operations. Second, the behavioral performance is measured while completing a task under the surgery simulator. Therefore, it does not allow extraction of similar measurements during laparoscopic procedures inside the operating room. To address these points, we designed three experiments in which medical students and interns perform laparoscopic tasks under a surgery simulator, while recording their brain activity with a wearable single-channel EEG headset. The experiments measured behavioral performance, EEG bands that are known to correlate to cognitive load, and a novel machine-learning-based cognitive load biomarker. In the following paragraphs, we will review the literature regarding cognitive load, EEG biomarkers of cognitive load, medical simulators, and previous measurements of cognitive load during simulator tasks.

Cognitive Load

Any acquirement of a new skill or knowledge must be passed through working memory (WM) before being transferred into long-term memory (LTM). According to Cognitive Load Theory (CLT, Miller, 1956), WM resources, unlike LTM resources, are limited in their capacity for processing or holding new information. However, when performing a complex task, the new information elements are being processed simultaneously and not iterated, so their interactions cause a much higher WM load. According to CLT, optimizing learning processes may be achieved by constructing automated schemas in which the WM load is reduced (Jolles et al., 2013). The amount of cognitive load used is a good predictor of the learning process during a task performance, and continuous measurement is of particular importance (Van Merriënboer and Sweller, 2005; Constantinidis and Klingberg, 2016). For instance, practicing the same task repeatedly will cause some of the interacting information elements to be stored together in a schematic form in LTM, so that they can be extracted and manipulated in WM as single-information elements (Schneider and Shiffrin, 1977).

Cognitive Load Biomarkers

Several methods were previously described and validated as measures of WM and cognitive load. Traditionally, subjective self-rating scales were proven as reliable assessment tools across several studies (Paas et al., 2003). However, as this tool can only be recorded crudely and retrospectively, objective assessment methods with a real-time indication of WM are in demand. By using such an approach, the activity can be broken down into different components that reflect different stages of complex simulations and be used to evaluate the efficacy of each training element. Several biological methods have been reported to successfully assess WM, such as pupil size (Van Gerven et al., 2004), tracked eye movement (Van Gog and Scheiter, 2010), salivary cortisol levels (Bong et al., 2010), and functional magnetic resonance imaging (fMRI) (Van Dillen et al., 2009).

Using electroencephalography (EEG), studies repeatedly show that the theta band (4–7 Hz) measured by frontal electrodes increases with higher cognitive load and task difficulty (Scheeringa et al., 2009; Antonenko et al., 2010). Multiple studies confirmed this correspondence using a variety of cognitive tasks, including the n-back (Gevins et al., 1997), operation span task (Scharinger et al., 2017), visuospatial WM tasks (Bastiaansen et al., 2002), and simulated real-life experiences such as in driving (Di Flumeri et al., 2018) or flight simulators (Dussault et al., 2005).

In addition to theta, delta oscillation (0.5–4 Hz) was also found to increase with higher cognitive load and/or task difficulty such as during reading tasks (Zarjam et al., 2011) and multiple-choice reaction tasks (Schapkin et al., 2020), and to increase with higher vigilance during low-cognitive-load attention tasks (Kim et al., 2017). It has been suggested that delta activity is an indicator of internal attention, and therefore increases while undergoing mental tasks (Harmony et al., 1996). Delta activity was also associated with interference inhibition processes, which occur in order to modulate sensory afferences, which in turn increase internal concentration (Harmony, 2013).

Medical Simulations

Medical simulations are a common and widespread tool for medical education (Gurusamy et al., 2009). Medical simulations can emulate common scenarios in clinical practice, and through interactive interplay and hands-on teaching, they can improve the effectiveness and quality of teaching of healthcare professionals (Kunkler, 2006). These simulations can be particularly beneficial for surgical staff, as they allow residents to practice and perfect complex procedures, ensuring they have enough experience and practice before real patient contact (Thompson et al., 2014). As laparoscopic surgeries demand unique eye–hand coordination and are performed while the surgeon indirectly observes the intra-abdominal contents without tactile sensation but through a camera view, these surgeons are ideal candidates for virtual reality (VR) simulators. Indeed, these simulators have been shown to greatly improve surgical operating skills and reduce operating time (Gallagher et al., 2005; Larsen et al., 2009; Gauger et al., 2010; Yiannakopoulou et al., 2015; Tucker et al., 2017).

Recently, a few studies made the connection between CLT and medical training using surgery simulators (Andersen et al., 2016, 2018; Frederiksen et al., 2020). In these studies, participants completed laparoscopic tasks via VR simulators, while their WM loads were estimated using a secondary behavioral task (e.g., reaction times in response to auditory stimuli). They found that different interventions with medical VR training may modulate WM load. For example, Frederiksen et al. (2020) showed that immersive VR results show higher cognitive load than the conventional VR training, and Andersen et al. (2018) demonstrated that additional VR training sessions may reduce cognitive load and increase the performance of dissection training. However, these studies evaluated cognitive load to assess the effect of an external manipulation, rather than to understand the neurophysiological mechanism underlying the simulation process itself. None of the studies recorded an objective physiological WM load biomarker such as pre-frontal cortex activity or frontal theta oscillation during the training itself. Such measurements would not only exhibit a continuous level of cognitive load, but, if found reliable, could later be measured in real scenarios. Lower levels of WM load, along with efficient performance, may indicate the readiness levels of surgeons. Additionally, a wearable, portable device may enable real-time monitoring of WM load levels inside the operating room, reflecting a surgeon’s abilities and laparoscopic dexterity.

The Present Study

The aim of this study was to explore the neural mechanism that underlies surgery simulation training. Specifically, our aim was to test the relationship between cognitive load, skill acquisition, and the activity levels of different brain oscillations. We aimed to track cognitive load neuro-markers using an EEG device that will enable hand mobility while performing surgical simulator tasks. Importantly, we wanted to test medical students and interns who, on the one hand, have great motivation and motor/cognitive abilities to perform such laparoscopic tasks and, on the other, have no previous hands-on experience, neither in surgery simulators nor with real-life patients. Finally, we intended to compare “online gains”—improvements in performance that occur while undergoing the task—to “offline gains,” the improvements in skill acquisition preceding a consolidation during nighttime sleep between task trainings, without further practice (Issenberg et al., 2011; Fraser et al., 2015; Lugassy et al., 2018). Notwithstanding, frontal theta and other known cognitive load biomarkers were mostly used to measure WM load during task completion itself and not to evaluate the load differences between tasks’ sessions.

To meet these goals, we divided 38 medical students and interns into three experiments. Each experiment included three trials of the same short laparoscopic task administered by a surgical simulator (Simbionix LAP MENTOR). While performing the tasks, participants’ brain activity was recorded using a wearable single-channel EEG device (Aurora by Neurosteer® Inc), from which we continuously measured frontal brain oscillations (i.e., the theta and delta bands) throughout the simulation task. In each experiment, the participants repeated the same task three times, thus allowing them to obtain the new skill and decrease their cognitive load levels.

In addition to known brain oscillations, Neurosteer® also provided us with a machine-learning-based cognitive load biomarker (i.e., VC9). VC9 activity was previously shown to increase with cognitive load in the standard and most common WM tasks (i.e., the n-back task, Maimon et al., 2020), auditory recognition and classification tasks (Molcho et al., 2021), and interruptions paradigm (Bolton et al., 2021). Specifically, Maimon et al. (2021) showed that the VC9 biomarker is more sensitive to the finer differences between cognitive loads. During the n-back task, while both VC9 activity and theta and delta oscillations increased with higher cognitive loads, the VC9 biomarker exhibited higher sensitivity than the theta and delta oscillations to the lower levels of cognitive load. These subtler differences are particularly crucial for the present study’s purposes and experimental design. With each task repetition, participants’ cognitive load levels descended, eventually reaching relatively low levels of cognitive load. Therefore, the ability to detect finer differences between the low cognitive loads is the most critical for estimating participants’ readiness to go into the operating room.

Thus, our hypotheses were as follows: (1) Behavioral performance will improve with repetition of the surgery simulator task. (2) This improvement of behavioral performance with task repetition will result in the decrease of cognitive load that will modulate the EEG features (i.e., theta, delta, and VC9 will decrease with task repetition). (3) EEG features will correlate to some extent with higher individual performance (i.e., decrease with better individual performance), reflecting the reduced need for cognitive resources together with improving laparoscopic dexterity. (4) “Offline gains” will be present following procedural memory consolidation during nighttime sleep. Probing these hypotheses will help reveal new and objective information regarding the efficacy of simulation-based training. For a graphical representation of the data analysis applied in the present study, see Figure 1.

FIGURE 1
www.frontiersin.org

Figure 1. A graphical representation of the study. (A) EEG data are recorded via a single channel using a three-electrode forehead patch by an Aurora EEG system while participants perform the simulator task. Data are pre-processed by a predefined set of wavelet packet atoms to produce 121 Brain Activity Features (BAFs). The BAFs are passed through linear models, which were pretrained on external data, to form a higher-level biomarker, VC9. Then, VC9 activity is averaged per trial and participant to find differences between simulator trials. (B) From the raw EEG data, power spectral density analysis is applied to find energy representation (in dB) produced for 1–50 Hz. The power density per Hz is averaged over frequency bands. Theta and delta activity are averaged per trial and participant to find differences between simulator trials. (C) Behavioral data are obtained from the Simbionix LAP MENTOR (i.e., time, accuracy, and economy of movement). These are averaged per trial and participant to find differences between simulator trials. Finally, correlations are calculated between EEG features (i.e., VC9, theta, and delta bands) and behavioral performance (time, accuracy, and economy of movement).

Experiment 1

Materials and Methods

Sample Size Selection

To estimate the sample size required to observe a significant effect of simulator trial repetition in the present study, we conducted an a priori sample size analysis with G*Power software (Erdfelder et al., 1996). We based this analysis on the study by Andersen et al. (2016). Their study examined 18 novice medical students who underwent VR simulation training while their WM loads were examined using a secondary response time (RT) task. The experimental design included two groups: a control group (n = 9), which received standard instructions, and an intervention group (n = 9), which received CLT-based instructions. Similar to our present study, both groups in the Andersen et al. (2016) study underwent two post-training virtual procedures. Therefore, we based our power analysis on the within-group comparison of RTs between the first and second simulator trials of the control group. Moreover, Andersen et al. (2016) inserted these multiple within-participants fix effects into mixed linear models design, which was applied in the current study, as well; they are therefore most suitable for comparison. We conducted this analysis with a desired alpha of 0.05, and a power of 0.80, and calculated effect size from the control group’s average RTs and standard deviation (SD) difference between the two simulator trials. The output of the analysis software was that the minimum required sample size was nine participants.

Participants

Nineteen participants (68% females, mean age 28, age range 25–36) were enrolled in the first experiment. All participants were healthy medical interns who had completed 6 years of medical studies, never participated in laparoscopic surgery, and had no prior experience using a surgical simulator. Ethical approval for this study was granted by the Galilee Medical Center Institutional Review Board.

Electroencephalography Device

The EEG signal acquisition system included a three-electrode patch attached to the subject’s forehead (Aurora by Neurosteer®., Herzliya, Israel). The medical-grade electrode patch included dry gel for optimal signal transduction. The electrodes were located at Fp1 and Fp2, with a reference electrode at Fpz. The EEG signal was amplified by a factor of 100 and sampled at 500 Hz. Signal processing was performed automatically by Neurosteer® in their cloud (see section “Signal Processing” below and Supplementary Appendix A). We were therefore provided with a sample per second of activity of the brain oscillations (i.e., delta, theta, alpha, beta, and gamma) and the VC9 biomarker.

Signal Processing

Full technical specifications regarding the signal analysis are provided in Supplementary Appendix A. In brief, the signal-processing algorithm interprets the EEG data using a time/frequency wavelet-packet analysis, instead of the commonly used spectral analysis. The Neurosteer® signal-processing algorithm interprets the EEG data using a time/frequency wavelet-packet analysis, creating a presentation of 121 brain activity features (BAFs) composed of the fundamental frequencies and their high harmonics.

To demonstrate this process, let g and h be a set of biorthogonal quadrature filters created from the filters G and H, respectively. These are convolution-decimation operators, where, in a simple Haar wavelet, g is a set of averages and h is a set of differences.

Let ψ1 be the mother wavelet associated with the filters sH, and dG. Then, the collection of wavelet packets ψn, is given by:

ψ2n=Hψn;ψ2n(t)=2jZs(j)ψn(2t-j),(1)
ψ2n+1=Gψn;ψ2n+1(t)=2jZd(j)ψn(2t-j).(2)

The recursive form provides a natural arrangement in the form of a binary tree. The functions ψn have a fixed scale. A library of wavelet packets of any scale s, frequency f, and position p is given by:

ψsfp(t)=2-s/2ψf(2-st-p).(3)

The wavelet packets {ψsfp : pZ} include a large collection of potential orthonormal bases. An optimal basis can be chosen by the best-basis algorithm (Coifman and Wickerhauser, 1992). Furthermore, an optimal mother wavelet can be chosen by Neretti and Intrator (2002). Following robust statistics methods to prune some of the basis functions, one gets 121 basis functions, which we term brain activity features (BAFs). Based on a given labeled-BAFs dataset, various models can be created for different discriminations of these labels. In the linear case, these models are of the form:

Vk(w,x)=Ψ(iwixi),(4)

where, w is a vector of weights and Ψ is a transfer function that can either be linear, e.g., Ψ (y) = y, or sigmoidal for logistic regression Ψ (y) = 1/ (1 + ey).

From these BAFs, several linear and non-linear combinations were obtained using machine learning techniques on previously collected, labeled datasets. These datasets included data about participants undergoing different cognitive and emotional tasks. Specifically, VC9 was found with linear discriminant analysis technique (LDA, Ye and Li, 2005) on the 121 BAFs to be the best separator between an auditory detection task (higher cognitive load) and an auditory classification task (lower cognitive load, Molcho et al., 2021). VC9 was also recently validated as a cognitive load biomarker using the n-back task, auditory detection task, and interruption task (Maimon et al., 2020; Bolton et al., 2021; Molcho et al., 2021). It was shown that VC9 activity increased with increasing levels of cognitive load within cognitively healthy participants (Maimon et al., 2020; Bolton et al., 2021), and showed higher sensitivity than the theta band (Maimon et al., 2021). Neurosteer® provided a sample per second activity of EEG bands (i.e., delta, theta, alpha, beta, and gamma) and VC9.

Procedure

The Aurora Electrode strip with three frontal electrodes was attached to each subject’s forehead and connected to the device for brain activity recording. The participants were then asked to complete a standard beginner’s task under the surgical simulator Simbionix LAP MENTOR (Simbionix, Airport City, Israel), which involved grasping and clamping blood vessels using two different laparoscopic arms. The same task was performed by all subjects. At the end of the task, the participants were rated by the surgical simulator based on three main parameters: accuracy in the execution of the required task, economy of movement (to use as few movements as necessary to produce the technical goal), and time required to accomplish the task. All the behavioral parameters were processed by the virtual simulator Simbionix and were presented in graphs and values at the end of the technical goal. All participants were monitored by the EEG device and were given three attempts to perform the same task. Participants performed three consecutive trials in the same session, with a 5-min break between the trials. The performance of each participant was scored by the surgical simulator’s algorithm based on accuracy, economy of movement, and time to complete the task. The concurrent activity of the theta, delta, and VC9 biomarker was extracted by the Aurora system.

Statistical Analysis

Behavioral performance was extracted from the surgical simulator after each trial of each participant, including accuracy (in percentages), economy of movement (in percentages), and time to exceed the trial (in seconds). An initial analysis of all bands revealed that the theta band (averaged dBm at 4–7 Hz), the delta band (averaged dBm at 0.5–4 Hz), and VC9 biomarker (normalized between 1 and 100) were the only three EEG features that were both significant in at least one effect of the LMM analysis and one correlation with the repeated correlation analysis, and therefore were included in the analysis (see Supplementary Appendix B for the results of the other oscillations). All features were averaged throughout the task completion. Activity of all the dependent variables was averaged per trial per participant and reported as a mean with a SD. Two main analyses were performed on the behavioral and EEG data: On each experiment, seven mixed linear models (LMM) (Boisgontier and Cheval, 2016) were designed (one for each dependent variable) to measure the differences between the three simulator trials for each dependent variable. The model used trial numbers (1, 2, or 3) as a within-participants variable, and participants were inserted into the random slope. Since each trial number variable included three levels, indicator variables (aka dummy variables) were computed, with Trial 1 as the reference group. Accordingly, two effects resulted from the model: one from the second trial, which represents the significance of difference between the first and second trials, and one from the third trial, which represents the significance of difference between the first and third trials. These effects were extracted straight from the LMM models, with no need for multiple comparisons corrections. For significant fixed effects, we calculated effect sizes according to Westfall et al. (2014) using an analog of Cohen’s d (i.e., the expected mean difference divided by the expected variation of an individual observation). Prior to analyses, each dataset for comparison was revised with Shapiro’s test of normal distribution and Levene’s test of equal variance (for all test results, see Supplementary Appendix B). If any were significant, a Wilcoxon nonparametric test was applied instead of LMM. Two-tailed p < 0.05 was considered statistically significant. These analyses were performed using Python Statsmodels (Seabold and Perktold, 2010).

The second analysis was designed to evaluate the correlation between individual performance and the EEG feature. We took into this analysis each of the EEG features’ activity and behavioral performance per each participant’s trial. Since during each experiment, each participant underwent three task trials, the data points are not independent. To consider both results from repeated participants’ trials and between participants, we used the repeated measures correlation (rmcorr, Bakdash and Marusich, 2017). Rmcorr estimates the common regression slope, the association shared among individuals. This analysis was underscored by the rmcorr R package (Bakdash and Marusich, 2017), using 1.4.1717 R studio (RStudio Team, 2021). An initial analysis of all bands revealed that the theta band (averaged dBm at 4–7 Hz), the delta band (averaged dBm at 0.5–4 Hz), and VC9 biomarker (normalized between 1 and 100), were the only three EEG features that were both significant in at least one effect of the LMM analysis and one correlation with the repeated correlation analysis (see Supplementary Appendix B). The statistical analyses took into consideration the averaged VC9, delta, and theta activity during each task repetition (an overall average across experiments and trials of 144.17 s).

Results

Behavioral Measurements

A full description of the mixed linear models’ (LMM) parameters (coefficients, standard errors, Z-scores, p-values, and confidence intervals) for all models used in this study are presented in Table 1. Averages and standard errors of behavioral data in Experiment 1 are presented in Figure 2.

TABLE 1
www.frontiersin.org

Table 1. Coefficients, standard errors, Z-scores, p-values, and confidence intervals for all effects depicted in the three LMM analyses applied in the three experiments of the present study.

FIGURE 2
www.frontiersin.org

Figure 2. The means of accuracy (in percentage), economy of movement (in percentage), and time to exceed the task (in seconds) obtained in the three trials of Experiment 1, as a function of task repetitions: first trial (blue), second trial (red), and third trial (green), for all participants (n = 19). Error bars represent standard errors. *p < 0.05, **p < 0.01, and ***p < 0.001.

A significant increase of the participants’ accuracy was observed between the first and second trials and between the first and third trials (W = 16, p = 0.004; and W = 10, p = 0.004, respectively), as well as a significant increase in the economy of movement (p < 0.001, d = −0.799; p < 0.001, d = −1.054, respectively). The average time required for the completion of the task was also significantly reduced between the first and second trials, and first and third trials (p = 0.005, d = 0.854; p < 0.001, d = 1.562, respectively).

Electroencephalography Features

Distributions of VC9, theta, and delta data in Experiment 1 are presented in Figure 3. VC9 activity was significantly reduced between the first and second attempts and between the first and third attempts (p = 0.011, d = 0.474; p = 0.021, d = 0.503, respectively). Delta activity significantly decreased between the first and second attempts (p = 0.003, d = 0.625). Theta did not exhibit any significant differences between the trials (all p’s > 0.05).

FIGURE 3
www.frontiersin.org

Figure 3. (A) The distribution of participant VC9 activity (normalized between 1 and 100); (B) the distribution of theta (averaged power of 4–7 Hz in dBm); and (C) the distribution of delta (averaged power of 0.5–4 Hz in dBm), obtained in the three trials of Experiment 1, as a function of task repetitions: first trial (blue), second trial (red), and third trial (green), for all participants (n = 19). Dashed lines represent means and SDs. *p < 0.05, **p < 0.01, and ***p < 0.001.

Correlations Between Electroencephalography Features Activity and Behavioral Performance

VC9 activity significantly correlated with all three behavioral measurements and was found to decrease with better performance (i.e., negatively correlated with accuracy and economy of movement and positively correlated with time): rrm = −0.46, p = 0.003; rrm = −0.51, p = 0.001; and rrm = 0.43, p = 0.006 for accuracy, economy of movement, and time, respectively; see Figure 4. Delta also significantly correlated with all three measurements: rrm = −0.57, p < 0.001; rrm = −0.6, p < 0.001; and rrm = 0.5, p = 0.001 for accuracy, economy of movement, and time, respectively. Theta decreased with higher accuracy and economy of movement and increased with time: rrm = −0.41, p = 0.01; rrm = −0.4, p = 0.012; and rrm = 0.35, p = 0.029 for accuracy, economy of movement, and time, respectively.

FIGURE 4
www.frontiersin.org

Figure 4. Mean activity of VC9 as a function of accuracy (A), economy (B), and time (C); theta as a function of accuracy (D), economy (E), and time (F); and delta as a function of accuracy (G), economy (H), and time (I) per trial and participant obtained in Experiment 1 (n = 19). Rrm and p-values presented.

Discussion

Experiment 1 results exhibited a significant improvement in all behavioral measurements of performance, as well as a significant decrease in VC9 and delta activity (but not theta) between the three trials. In addition, VC9 and delta activity decreased with better individual performance, which was expressed by the significant correlations between all three behavioral measures and VC9 and delta activity. Theta also exhibited individual differences and decreased with higher accuracy and economy of movement, and marginally increased with longer time to complete the task. These results suggest that changes in cognitive load correspond to performance in a surgery simulator as shown by VC9, delta, and, to some extent, frontal theta.

Next, we aimed to explore the effect of nighttime sleep memory consolidation on task performance in the simulator and participants’ brain activity. Hence, we conducted a second experiment that included 10 participants who had also participated in Experiment 1. They performed an additional three-trial session on the consecutive day.

Experiment 2

Materials and Methods

Participants and Procedure

On the day following Experiment 1, 10 randomly chosen participants performed an additional three trials of the same procedure.

Results

Behavioral Measurements

Behavioral performance did not differ between the three trials of this session in any of the measurements (all p’s > 0.05; see Figure 5).

FIGURE 5
www.frontiersin.org

Figure 5. The means of accuracy (in percentage), economy of movement (in percentage), and time to exceed the task (in seconds) obtained in the three trials of Experiment 2, as a function of task repetitions: first trial (blue), second trial (red), and third trial (green), for participants who underwent Experiment 2 (n = 10). Error bars represent standard errors. *p < 0.05, **p < 0.01, and ***p < 0.001.

Electroencephalography Features

Activity levels of all EEG features (VC9, theta, and delta) did not differ between the trials of the session (p > 0.05; see Figure 6).

FIGURE 6
www.frontiersin.org

Figure 6. (A) The distribution of user VC9 activity (normalized between 1 and 100); (B) the distribution of theta (averaged power of 4–7 Hz in dBm); and (C) the distribution of delta (averaged power of 0.5–4 Hz in dBm), obtained in the three trials of Experiment 2, as a function of task repetitions: first trial (blue), second trial (red), and third trial (green), for all participants (n = 10). Dashed lines represent means and SDs. *p < 0.05, **p < 0.01, and ***p < 0.001.

Correlations Between Electroencephalography Features Activity and Behavioral Performance

VC9 activity significantly correlated with all three behavioral measurements and was found to decrease with better participant performance (i.e., negatively correlated with accuracy and economy of movement and positively correlated with time): rrm = −0.47, p = 0.033; rrm = −0.67, p = 0.001; and rrm = 0.64, p = 0.002 for accuracy, economy of movement, and time, respectively. Delta also significantly correlated with all three measurements: rrm = −0.69, p = 0.001; rrm = −0.65, p = 0.001; and rrm = 0.63, p = 0.002 for accuracy, economy of movement, and time, respectively. Theta did not correlate significantly with any of the behavioral measurements (all p’s > 0.05; see Figure 7).

FIGURE 7
www.frontiersin.org

Figure 7. Mean activity of VC9 as function of accuracy (A), economy (B), and time (C); theta as function of accuracy (D), economy (E), and time (F); and delta as function of accuracy (G), economy (H), and time (I) per trial and participant obtained in Experiment 2 (n = 10). Rrm and p-values presented.

Comparison Between Experiment 1 and Experiment 2

To explore the “offline gains” between Experiment 1 and Experiment 2, behavioral performance and brain activity of the 10 participants who participated in both experiments were studied. Paired t-tests on the activity of the 10 participants who underwent both experiments were calculated. The t-tests compared accuracy, economy of movement, time, VC9, theta, and delta activity for the last trial of Experiment 1 and the first trial of Experiment 2. Both behavioral performance and brain activity were the same in the first trial of Experiment 2 relative to the last trial of Experiment 1: all p’s > 0.05; see Figure 8 and Table 2.

FIGURE 8
www.frontiersin.org

Figure 8. (A) The distribution of VC9 (normalized between 1 and 100), theta (averaged power of 4–7 Hz in dBm), and delta (averaged power of 0.5–4 Hz in dBm), obtained in the last trial repetition of Experiment 1 (blue), and the first task repetition of Experiment 2 (red), for participants who participated in both Experiment 1 and Experiment 2 (n = 10). Dashed lines represent means and SDs. (B) The means of accuracy (in percentage), economy of movement (in percentage), and time to exceed the task (in seconds), obtained in the last trial repetition of Experiment 1 (blue), and the first task repetition of Experiment 2 (red), for participants who participated in both Experiment 1 and Experiment 2 (n = 10). Error bars represent standard errors.

TABLE 2
www.frontiersin.org

Table 2. t/W statistics and p-values for t-tests/Wilcoxon tests comparing Experiment 1 last trial and Experiment 2 first trial, for participants who participated in both Experiment 1 and Experiment 2 (n = 10).

Discussion

The results of Experiment 2 exhibited a clear correspondence between behavioral performance and EEG features. First, both behavioral performance and all EEG features’ activity did not exhibit significant differences between the three trials, meaning that participants’ performance did not improve, and theta, delta, and VC9 activity did not decrease between the three trials. Second, VC9 and delta activity decreased with better behavioral performance, as depicted by a significant correlation with all three behavioral measurements. Theta correlated with accuracy but not with economy of movement and time.

The comparison between the last trial of Experiment 1 and the first trial of Experiment 2 showed no difference between the two in any of the behavioral or neurophysiological measures. However, taken together with the null improvement during these three sessions, this lack of difference may be explained by participants having reached their asymptotic level (ceiling effect): that participants reached their maximum ability on the third trial of the first day, and maintained it through the first trials of the second day. To further reveal “offline gains” from nighttime sleep consolidation, we conducted a third experiment on an additional 19 medical students who had not participated in Experiments 1 and 2. Participants performed a single-task trial per day for three consecutive days. This was done to make sure participants did not reach asymptotic performance levels before the nighttime sleep consolidation, since they would only complete a single trial before nighttime sleep.

Experiment 3

Materials and Methods

Participants and Procedure

Nineteen (63% females) healthy medical students from their first to sixth years of studying (mean = 4), who did not participate in former Experiments 1 and 2, with mean age of 25.631 (SD = 2.532), participated in Experiment 3. All students had no prior experience using a surgical simulator. Participants underwent the same task as in Experiments 1 and 2, with one trial per day over three consecutive days.

Results

Behavioral Measurements

Averages and standard errors of the behavioral data in Experiment 3 are presented in Figure 9. A significant increase of the participants’ accuracy was observed between the first and second trials and between the first and third trials (W = 11, p = 0.002; W = 2, p < 0.001, respectively), as well as a significant increase in the economy of movement (p < 0.001, d = −1.762; p < 0.001, d = −1.83, respectively). The average time required for the completion of the task was also significantly reduced between the trials (p < 0.001, d = 1.014; p < 0.001, d = 3.171 for first vs. second trials and first vs. third trials, respectively).

FIGURE 9
www.frontiersin.org

Figure 9. The means of accuracy (in percentage), economy of movement (in percentage), and time to exceed the task (in seconds) obtained in the three trials of Experiment 3, as a function of task repetitions: first trial (blue), second trial (red), and third trial (green), for participants who underwent Experiment 3 (n = 19). Error bars represent standard errors. *p < 0.05, **p < 0.01, and ***p < 0.001.

Electroencephalography Features

Distributions of VC9, theta, and delta data in Experiment 3 are presented in Figure 10. There were no significant differences in VC9 and delta activity between the trials on consecutive days (all p’s > 0.05). Theta exhibited a significant difference between the first and second trials (p = 0.019, d = −0.354), but the difference between the first and third trials did not reach significance level (p > 0.05).

FIGURE 10
www.frontiersin.org

Figure 10. (A) The distribution of user VC9 activity (normalized between 1 and 100); (B) the distribution of theta (averaged power of 4–7 Hz in dBm); and (C) the distribution of delta (averaged power of 0.5–4 Hz in dBm), obtained in the three trials of Experiment 3, as a function of task repetitions: first trial (blue), second trial (red), and third trial (green), for all participants (n = 19). Dashed lines represent means and SDs. *p < 0.05, **p < 0.01, and ***p < 0.001.

Correlations Between Electroencephalography Features Activity and Behavioral Performance

No significant correlations were found between any of the EEG features and behavioral measurement (all p’s > 0.05; see Figure 11).

FIGURE 11
www.frontiersin.org

Figure 11. Mean activity of VC9 as a function of accuracy (A), economy (B), and time (C); theta as a function of accuracy (D), economy (E), and time (F); and delta as a function of accuracy (G), economy (H), and time (I) per trial and participant obtained in Experiment 3 (n = 19). Rrm and p-values presented.

Discussion

Experiment 3 results reflect “offline gains”: participants’ performances under the simulator improved with the trials’ repetition, although each occurred on the consecutive day. However, VC9, delta, and theta did not show significant differences between the trials, nor did they exhibit significant correlations with any of the behavioral measurements. This discrepancy might be explained by the different brain networks that take part in “online” versus “offline” learning; see general discussion below.

General Discussion

In this study, we continuously measured cognitive load levels using theta and band power and VC9 biomarker activity during task performance on a surgical simulator. A single-channel EEG device utilizing decomposition of the EEG signal via harmonic analysis was used to obtain the novel VC9 biomarker as well as the frequently used theta and delta bands. All EEG features (i.e., theta, delta, and VC9) were previously shown to correlate with cognitive load levels (Antonenko et al., 2010; Maimon et al., 2020; Bolton et al., 2021; Molcho et al., 2021). Experiment 1 consisted of three repeating trials on the surgery simulator, and its results showed that participants’ performances improved while VC9 and delta activity decreased. Additionally, VC9, delta, and, to some extent, theta band decreased with higher individual performance. Overall, these results support previous findings that as the participant becomes more proficient in performing a task, as depicted by the behavioral performance, the pre-frontal activity associated with cognitive load is reduced (Takeuchi et al., 2013).

To examine the effect of nighttime sleep memory consolidation, we conducted Experiment 2, which included three simulator trials on the consecutive day. Testing the “offline gains” refers to the improvements in skill acquisition preceding a consolidation during nighttime sleep between tasks trainings and without further practice (Walker et al., 2002). Results revealed no significant differences in performance or EEG features between the three trials. Although they did not exhibit significant decreases between the trials, the VC9, delta, and, to some extent, theta still exhibited significant correlations with individual performances. Further analysis revealed that both behavioral and EEG features results maintained the same levels between the last trial of the first day and the first trial on the second day. Taken together, these findings potentially indicate that “offline gains” were not detected, since participants reached their maximum performance levels on the last trial of the first day and maintained them throughout the trials on the following day.

Consequently, we conducted an additional experiment in which 19 medical students performed a single simulator trial on three consecutive days. This was done to prevent participants from reaching their maximum performance levels on the first day, and to reveal the nighttime sleep memory consolidation related to “offline gains.” Indeed, participants’ performances in all behavioral measurements improved significantly between the testing days. Interestingly, VC9 and delta showed no difference between the trials, and the theta band even showed a mild increase in activity between the trials. This lack of increase in activity can be explained by the difference in brain networks that are involved in fast vs. slow stages of motor skill learning (Dayan and Cohen, 2011). Similar to online learning, fast motor skill acquisition occurs during task training and could last minutes (Karni et al., 1995). In the slow stage, further gains are achieved across multiple sessions of training, mostly divided by nighttime sleep consolidation. The neural substrates during the fast stage show complex brain activation patterns. First, they include increasing activity in the Supplementary Motor Area (SMA), dorsomedial striatum (DMS), premotor cortex (PM), posterior parietal cortex (PPC), and posterior cerebellum (Honda et al., 1998; Grafton et al., 2002; Floyer-Lea and Matthews, 2005). This reflects the requirement of additional cortical brain activity during practice. At the same time, the fast stage also causes decreased activity of the dorsolateral prefrontal cortex (DLPFC), primary motor cortex (M1), and pre Supplementary Motor Area (preSMA, Poldrack, 2000). These decreases may suggest that with online practice, one uses fewer neuronal resources (Hikosaka et al., 2002). Conversely, the slow motor skill stage is characterized by increased activity in M1, S1, SMA, and DLS (Floyer-Lea and Matthews, 2005), and decreased activity in the lateral cerebellum (Lehéricy et al., 2005). Notably, the decrease in the frontal area (DLPFC), which commonly correlates with WM load (Manoach et al., 1997), is not visible during this slow learning stage. Thus, progress from fast to slow stages can be generalized to the shift from anterior to more posterior brain regions (Floyer-Lea and Matthews, 2005).

The current WM biomarkers that were used in the present research were extracted via frontal single-EEG channel, and although this electrode can extract additional brain activity beyond the frontal lobe, we aimed to focus on biomarkers that reflect frontal activity associated with WM load. Taken together, this may explain the results obtained in Experiment 3 of the present study, which revealed prominent “offline gains” within the behavioral performance of the task, but no decrease within the current EEG WM load biomarkers. Further research, however, should look for novel biomarkers that will be able to correlate with such posterior/limbic activity. Finally, this discrepancy may further support frontal theta, delta, and VC9 as continuous measurements of cognitive load, to monitor WM load during task completion and not as evaluators of cognitive load between tasks’ sessions. Therefore, these biomarkers may be adequate for assessing laparoscopic dexterity of non-experts during their first real-life surgeries. However, for the assessment of surgeons’ improvement in between laparoscopic practices, other biomarkers should be considered.

The activation patterns obtained in Experiments 1 and 2 are compatible with WM load, as previously described (Zarjam et al., 2011; Wang et al., 2016). Delta and VC9 activity decreased with the lower cognitive load that resulted from simulator trial repetition, but the theta band activity did not decrease. As shown in previous studies (Maimon et al., 2020; Molcho et al., 2021), VC9 was found to be more sensitive to lower loads than theta and delta, as its activity also showed a significant decrease between the first and third task trials. This further validates VC9 as an effective biological measurement for the assessment of cognitive load while performing laparoscopic tasks using the surgical simulator specifically and suggests that it may provide a measure of cognitive functioning during surgery. Due to the unobtrusive nature of mobile EEG devices, such devices can be used by surgeons during live operations. Intraoperative evaluation can provide an objective metric to assess surgeons’ performance in real-life scenarios and measure additional unobservable behaviors, such as mental readiness (Cha and Yu, 2021).

Since simulation-based training is an effective tool for acquiring practical skills, the question remains as to which methods should be utilized to optimize this process and achieve better assessment of improved manual dexterity. Specifically, it is not fully understood which brain processes during simulation-based training translate to better skill acquisition through practice. Here, we show that VC9 and delta, collected via single-channel EEG set, can be reliably used to monitor and assess participants’ WM levels during manual practice. Furthermore, the correlations between VC9, delta, and improved simulator scores can be translated to other areas and used in other procedures that require manual training. Importantly, as this device is portable and can be easily worn inside the operating room, it could potentially be used to predict participants’ WM loads and indirect performance during real-life operating procedures.

This study has several limitations. The analyses were performed on young medical internists, which may not accurately reflect the overall medical population. Additionally, evaluating finer differences between the experiments requires a larger cohort of participants. Future studies on a larger and more diverse population should further validate the findings presented in this work, as well as study the effect of prolonged breaks on the activity of these biomarkers. Moreover, a 24-h break period between the sessions may not be sufficient to study the effects of long-term memory on the prefrontal brain activity during the simulation. Future studies should evaluate longer time windows to assess the temporal changes in the activity of the working-memory-associated biomarkers and search for additional biomarkers to represent other cerebral regions beyond the frontal regions (like the limbic or motor systems).

To conclude, in this study, we extracted the previously validated cognitive load biomarkers—theta, delta, and VC9—from a novel single-channel EEG using advanced signal analysis. Results showed high correlations between the EEG features and participants’ individual performances using a surgical simulator. As surgical simulations allow doctors to gain important skills and experience needed to perform procedures without any patient risk, evaluation and optimization of these effects on medical staff are crucial. This could potentially be expanded to evaluate the efficacy of different medical simulations to provide more efficient training to medical staff and to measure cognitive and mental loads in real laparoscopic surgeries.

Data Availability Statement

The datasets presented in this article are not readily available because the data that support the findings of this study are available for simbionix and Neurosteer Ltd. Restrictions apply to the availability of these data, which were used under license for this study. Requests to access the datasets should be directed to bGlvckBuZXVyb3N0ZWVyLmNvbQ==.

Ethics Statement

The studies involving human participants were reviewed and approved by Galilee Medical Center Institutional Review Board. The patients/participants provided their written informed consent to participate in this study.

Author Contributions

AB, MB, NM, and NI conceived and planned the experiments. AB, MB, and DD carried out the experiments. NI and LM designed and provided tech support for the EEG device. MB, NM, and LM contributed to the interpretation of the results. MB and NM took the lead in writing the manuscript. All authors provided critical feedback and helped shape the research, analysis, and manuscript.

Conflict of Interest

NM is a part-time researcher at Neurosteer Ltd., which supplied the EEG system. LM is a clinical researcher at Neurosteer Ltd. NI is the founder of Neurosteer Ltd.

The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary Material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fnins.2021.694010/full#supplementary-material

References

Andersen, S. A. W., Konge, L., and Sørensen, M. S. (2018). The effect of distributed virtual reality simulation training on cognitive load during subsequent dissection training. Med. Teacher 40, 684–689.

Google Scholar

Andersen, S. A. W., Mikkelsen, P. T., Konge, L., Cayé-Thomasen, P., and Sørensen, M. S. (2016). The effect of implementing cognitive load theory-based design principles in virtual reality simulation training of surgical skills: a randomized controlled trial. Adv. Simul. 1, 1–8.

Google Scholar

Antonenko, P., Paas, F., Grabner, R., and van Gog, T. (2010). Using electroencephalography to measure cognitive load. Educ. Psychol. Rev. 22, 425–438. doi: 10.1007/s10648-010-9130-y

CrossRef Full Text | Google Scholar

Bakdash, J. Z., and Marusich, L. R. (2017). Repeated measures correlation. Front. Psychol. 8:456. doi: 10.3389/fpsyg.2017.00456

PubMed Abstract | CrossRef Full Text | Google Scholar

Bastiaansen, M. C., Posthuma, D., Groot, P. F., and De Geus, E. J. (2002). Event-related alpha and theta responses in a visuo-spatial working memory task. Clin. Neurophysiol. 113, 1882–1893. doi: 10.1016/S1388-2457(02)00303-6

CrossRef Full Text | Google Scholar

Boisgontier, M. P., and Cheval, B. (2016). The ANOVA to mixed model transition. Neurosci. Biobehav. Rev. 68, 1004–1005. doi: 10.1016/j.neubiorev.2016.05.034

PubMed Abstract | CrossRef Full Text | Google Scholar

Bolton, F., Te’Eni, D., Maimon, N. B., and Toch, E. (2021). “Detecting interruption events using EEG,” in Proceedings of the 2021 IEEE 3rd Global Conference on Life Sciences and Technologies (LifeTech) (Nara: IEEE), 33–34.

Google Scholar

Bong, C. L., Lightdale, J. R., Fredette, M. E., and Weinstock, P. (2010). Effects of simulation versus traditional tutorial-based training on physiologic stress levels among clinicians: a pilot study. Simul. Healthcare 5, 272–278. doi: 10.1097/sih.0b013e3181e98b29

PubMed Abstract | CrossRef Full Text | Google Scholar

Cha, J. S., and Yu, D. (2021). Objective measures of surgeon nontechnical skills in surgery: a scoping review. Hum. Fact. doi: 10.1177/0018720821995319

PubMed Abstract | CrossRef Full Text | Google Scholar

Coifman, R. R., and Wickerhauser, M. V. (1992). Entropy-based algorithms for best basis selection. IEEE Trans. Inform. Theory 38, 713–718. doi: 10.1109/18.119732

CrossRef Full Text | Google Scholar

Constantinidis, C., and Klingberg, T. (2016). The neuroscience of working memory capacity and training. Nat. Rev. Neurosci. 17:438. doi: 10.1038/nrn.2016.43

PubMed Abstract | CrossRef Full Text | Google Scholar

Dayan, E., and Cohen, L. G. (2011). Neuroplasticity subserving motor skill learning. Neuron 72, 443–454. doi: 10.1016/j.neuron.2011.10.008

PubMed Abstract | CrossRef Full Text | Google Scholar

Di Flumeri, G., Borghini, G., Aricò, P., Sciaraffa, N., Lanzi, P., and Pozzi, S. (2018). EEG-based mental workload neurometric to evaluate the impact of different traffic and road conditions in real driving settings. Front. Hum. Neurosci. 12:509. doi: 10.3389/fnhum.2018.00509

PubMed Abstract | CrossRef Full Text | Google Scholar

Dussault, C., Jouanin, J. C., Philippe, M., and Guezennec, C. Y. (2005). EEG and ECG changes during simulator operation reflect mental workload and vigilance. Aviation Space Environ. Med. 76, 344–351.

Google Scholar

Erdfelder, E., Faul, F., and Buchner, A. (1996). GPOWER: a general power analysis program. Behavior research methods, Instruments Comp. 28, 1–11.

Google Scholar

Floyer-Lea, A., and Matthews, P. M. (2005). Distinguishable brain activation networks for short- and long-term motor skill learning. J. Neurophysiol. 94, 512–518. doi: 10.1152/jn.00717.2004

PubMed Abstract | CrossRef Full Text | Google Scholar

Fraser, K. L., Ayres, P., and Sweller, J. (2015). Cognitive load theory for the design of medical simulations. Simul. Healthcare 10, 295–307. doi: 10.1097/sih.0000000000000097

PubMed Abstract | CrossRef Full Text | Google Scholar

Frederiksen, J.G., Sørensen, S. M. D., Konge, L. Svendsen, M. B. S., Nobel-Jørgensen M., Bjerrum F. et al. (2020). Cognitive load and performance in immersive virtual reality versus conventional virtual reality simulation training of laparoscopic surgery: a randomized trial. Surg. Endosc. 34, 1244–1252. doi: 10.1007/s00464-019-06887-8

PubMed Abstract | CrossRef Full Text | Google Scholar

Gallagher, A. G., Ritter, E. M., Champion, H., Higgins, G., Fried, M. P., and Moses, G. (2005). Virtual reality simulation for the operating room: proficiency-based training as a paradigm shift in surgical skills training. Ann. Surg. 241:364. doi: 10.1097/01.sla.0000151982.85062.80

CrossRef Full Text | Google Scholar

Gauger, P. G., Hauge, L. S., Andreatta, P. B., Hamstra, S. J., Hillard, M. L., and Arble, E. P., (2010). Laparoscopic simulation training with proficiency targets improves practice and performance of novice surgeons. Am. J. Surg. 199, 72–80. doi: 10.1016/j.amjsurg.2009.07.034

PubMed Abstract | CrossRef Full Text | Google Scholar

Gevins, A., Smith, M. E., McEvoy, L., and Yu, D. (1997). High-resolution EEG mapping of cortical activation related to working memory: effects of task difficulty, type of processing, and practice. Cereb. Cortex 7, 374–385. doi: 10.1093/cercor/7.4.374

PubMed Abstract | CrossRef Full Text | Google Scholar

Grafton, S. T., Hazeltine, E., and Ivry, R. B. (2002). Motor sequence learning with the nondominant left hand. Exp. Brain Res. 146, 369–378. doi: 10.1007/s00221-002-1181-y

PubMed Abstract | CrossRef Full Text | Google Scholar

Gurusamy, K. S., Aggarwal, R., Palanivelu, L., and Davidson, B. R. (2009). Virtual reality training for surgical trainees in laparoscopic surgery. Cochrane Datab. Syst. Rev. 1:CD006575.

Google Scholar

Harmony, T. (2013). The functional significance of delta oscillations in cognitive processing. Front. Integrat. Neurosci. 7:83. doi: 10.3389/fnint.2013.00083

PubMed Abstract | CrossRef Full Text | Google Scholar

Harmony, T., Fernández, T., Silva, J., Bernal, J., Díaz-Comas, L., Reyes, A., et al. (1996). EEG delta activity: an indicator of attention to internal processing during performance of mental tasks. Int. J. Psychophysiol. 24, 161–171. doi: 10.1016/S0167-8760(96)00053-0

CrossRef Full Text | Google Scholar

Hikosaka, O., Rand, M. K., Nakamura, K., Miyachi, S., Kitaguchi, K., Sakai, K., (2002). Long-term retention of motor skill in macaque monkeys and humans. Exp. Brain Res. 147, 494–504. doi: 10.1007/s00221-002-1258-7

PubMed Abstract | CrossRef Full Text | Google Scholar

Honda, M., Deiber, M. P., Ibánez, V., Pascual-Leone, A., Zhuang, P., and Hallett, M. (1998). Dynamic cortical involvement in implicit and explicit motor sequence learning. A PET study. Brain J. Neurol. 121, 2159–2173.

Google Scholar

Issenberg, S. B., Ringsted, C., Østergaard, D., and Dieckmann, P. (2011). Setting a research agenda for simulation-based healthcare education: a synthesis of the outcome from an Utstein style meeting. Simul. Healthcare 6, 155–167. doi: 10.1097/sih.0b013e3182207c24

PubMed Abstract | CrossRef Full Text | Google Scholar

Jolles, D. D., van Buchem, M. A., Crone, E. A., and Rombouts, S. A. (2013). Functional brain connectivity at rest changes after working memory training. Hum. Brain Mapp. 34, 396–406. doi: 10.1002/hbm.21444

PubMed Abstract | CrossRef Full Text | Google Scholar

Karni, A., Meyer, G., Jezzard, P., Adams, M. M., Turner, R., and Ungerleider, L. G. (1995). Functional MRI evidence for adult motor cortex plasticity during motor skill learning. Nature 377, 155–158. doi: 10.1038/377155a0

PubMed Abstract | CrossRef Full Text | Google Scholar

Kim, J. H., Kim, D. W., and Im, C. H. (2017). Brain areas responsible for vigilance: an EEG source imaging study. Brain Topograp. 30, 343–351. doi: 10.1007/s10548-016-0540-0

PubMed Abstract | CrossRef Full Text | Google Scholar

Kunkler, K. (2006). The role of medical simulation: an overview. Int. J. Med. Robot. Comp. Assisted Surg. 2, 203–210. doi: 10.1002/rcs.101

PubMed Abstract | CrossRef Full Text | Google Scholar

Larsen, C. R., Soerensen, J. L., Grantcharov, T. P., Dalsgaard, T., Schouenborg, L., Ottosen, C., and Ottesen, B. S. (2009). Effect of virtual reality training on laparoscopic surgery: randomised controlled trial. BMJ 338:b2074. doi: 10.1136/bmj.b2074

CrossRef Full Text | Google Scholar

Lehéricy, S., Benali, H., Van de Moortele, P. F., Pélégrini-Issac, M., Waechter, T., Ugurbil, K., and Doyon, J. (2005). Distinct basal ganglia territories are engaged in early and advanced motor sequence learning. Proc. Natl. Acad. Sci. U.S.A. 102, 12566–12571. doi: 10.1073/pnas.0502762102

PubMed Abstract | CrossRef Full Text | Google Scholar

Lugassy, D., Herszage, J., Pilo, R., Brosh, T., and Censor, N. (2018). Consolidation of complex motor skill learning: evidence for a delayed offline process. Sleep 41:zsy123. doi: 10.1093/sleep/zsy123

PubMed Abstract | CrossRef Full Text | Google Scholar

Maimon, N. B., Molcho, L., Intrator, N., and Lamy, D. (2020). Single-channel EEG features during n-back task correlate with working memory load. arXiv [Preprint]. arXiv:2008.04987.

Google Scholar

Maimon, N. B., Molcho, L., Intrator, N., and Lamy, D. (2021). “Novel single-channel EEG features correlate with working memory load,” in Proceedings of the 2021 IEEE 3rd Global Conference on Life Sciences and Technologies (LifeTech), Nara, 471–472. doi: 10.1109/LifeTech52111.2021.9391963

CrossRef Full Text | Google Scholar

Manoach, D. S., Schlaug, G., Siewert, B., Darby, D. G., Bly, B. M., Benfield, A., Edelman, R. R., and Warach, S. (1997). Prefrontal cortex fMRI signal changes are correlated with working memory load. Neuroreport 8, 545–549. doi: 10.1097/00001756-199701200-00033

PubMed Abstract | CrossRef Full Text | Google Scholar

Miller, J. G. (1956). General behavior systems theory and summary. J. Counsel. Psychol. 3:120.

Google Scholar

Molcho, L., Maimon, N. B., Pressburger, N., Regev-Plotnik, N., Rabinowicz, S., Intrator, N., and Sasson, A. (2021). Detection of cognitive decline using a single-channel EEG system with an interactive assessment tool. Res. Square [Preprint] doi: 10.21203/rs.3.rs-242345/v1

CrossRef Full Text | Google Scholar

Neretti, N., and Intrator, N. (2002). “An adaptive approach to wavelet filters design,” in Proceedings of the 12th IEEE Workshop on Neural Networks for Signal Processing, Martigny, 317–326.

Google Scholar

Paas, F., Tuovinen, J. E., Tabbers, H., and Van Gerven, P. W. (2003). Cognitive load measurement as a means to advance cognitive load theory. Educ. Psychol. 38, 63–71. doi: 10.1207/S15326985EP3801_8

PubMed Abstract | CrossRef Full Text | Google Scholar

Poldrack, R. A. (2000). Imaging brain plasticity: conceptual and methodological issues—a theoretical review. Neuroimage, 12, 1–13. doi: 10.1006/nimg.2000.0596

PubMed Abstract | CrossRef Full Text | Google Scholar

RStudio Team (2021). RStudio: Integrated Development Environment for R. Boston, MA: RStudio, PBC.

Google Scholar

Schapkin, S. A., Raggatz, J., Hillmert, M., and Böckelmann, I. (2020). EEG correlates of cognitive load in a multiple choice reaction task. Acta Neurobiol. Exp. 80, 76–89.

Google Scholar

Scharinger, C., Soutschek, A., Schubert, T., and Gerjets, P. (2017). Comparison of the working memory load in n-back and working memory span tasks by means of EEG frequency band power and P300 amplitude. Front. Hum. Neurosci. 11:6.

Google Scholar

Scheeringa, R., Petersson, K. M., Oostenveld, R., Norris, D. G., Hagoort, P., and Bastiaansen, M. C. (2009). Trial-by-trial coupling between EEG and BOLD identifies networks related to alpha and theta EEG power increases during working memory maintenance. Neuroimage, 44, 1224–1238. doi: 10.1016/j.neuroimage.2008.08.041

PubMed Abstract | CrossRef Full Text | Google Scholar

Schneider, W., and Shiffrin, R. M. (1977). Controlled and automatic human information processing: I. Detection, search, and attention. Psychol. Rev. 84:1.

Google Scholar

Seabold, S., and Perktold, J. (2010). “Statsmodels: econometric and statistical modeling with python”, in Proceedings of the 9th Python in Science Conference, Vol. 57, Austin, TX, 61

Google Scholar

Takeuchi, H., Taki, Y., Nouchi, R., Hashizume, H., Sekiguchi, A., Kotozaki, Y., and Kawashima, R. (2013). Effects of working memory training on functional connectivity and cerebral blood flow during rest. Cortex 49, 2106–2125. doi: 10.1016/j.cortex.2012.09.007

PubMed Abstract | CrossRef Full Text | Google Scholar

Thompson, J. E., Egger, S., Böhm, M., Haynes, A. M., Matthews, J., Rasiah, K., and Stricker, P. D. (2014). Superior quality of life and improved surgical margins are achievable with robotic radical prostatectomy after a long learning curve: a prospective single-surgeon study of 1552 consecutive cases. Eur. Urol. 65, 521–531. doi: 10.1016/j.eururo.2013.10.030

PubMed Abstract | CrossRef Full Text | Google Scholar

Tucker, M. A., Morris, C. J., Morgan, A., Yang, J., Myers, S., Pierce, J. G., and Scheer, F. A. (2017). The relative impact of sleep and circadian drive on motor skill acquisition and memory consolidation. Sleep 40:zsx036. doi: 10.1093/sleep/zsx036

PubMed Abstract | CrossRef Full Text | Google Scholar

Van Dillen, L. F., Heslenfeld, D. J., and Koole, S. L. (2009). Tuning down the emotional brain: an fMRI study of the effects of cognitive load on the processing of affective images. Neuroimage 45, 1212–1219. doi: 10.1016/j.neuroimage.2009.01.016

PubMed Abstract | CrossRef Full Text | Google Scholar

Van Gerven, P. W., Paas, F., Van Merriënboer, J. J., and Schmidt, H. G. (2004). Memory load and the cognitive pupillary response in aging. Psychophysiol. 41, 167–174. doi: 10.1111/j.1469-8986.2003.00148.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Van Gog, T., and Scheiter, K. (2010). Eye tracking as a tool to study and enhance multimedia learning. Learn. Instruction 20, 95–99. doi: 10.1016/j.learninstruc.2009.02.009

CrossRef Full Text | Google Scholar

Van Merriënboer, J. J., and Sweller, J. (2005). Cognitive load theory and complex learning: recent developments and future directions. Educ. Psychol. Rev. 17, 147–177. doi: 10.1007/s10648-005-3951-0

CrossRef Full Text | Google Scholar

Walker, M. P., Brakefield, T., Morgan, A., Hobson, J. A., and Stickgold, R. (2002). Practice with sleep makes perfect: sleep-dependent motor skill learning. Neuron 35, 205–211. doi: 10.1016/s0896-6273(02)00746-8

CrossRef Full Text | Google Scholar

Wang, S., Gwizdka, J., and Chaovalitwongse, W. A. (2016). Using wireless EEG signals to assess memory workload in the n-back task. IEEE Trans. Hum. Machine Syst. 46, 424–435. doi: 10.1109/THMS.2015.2476818

CrossRef Full Text | Google Scholar

Westfall, J., Kenny, D. A., and Judd, C. M. (2014). Statistical power and optimal design in experiments in which samples of participants respond to samples of stimuli. J. Exper. Psychol. Gen. 143:2020. doi: 10.1037/xge0000014

PubMed Abstract | CrossRef Full Text | Google Scholar

Ye, J., and Li, Q. (2005). A two-stage linear discriminant analysis via QR-decomposition. IEEE Trans. Pattern Anal. Machine Intell. 27, 929–941. doi: 10.1109/TPAMI.2005.110

PubMed Abstract | CrossRef Full Text | Google Scholar

Yiannakopoulou, E., Nikiteas, N., Perrea, D., and Tsigris, C. (2015). Virtual reality simulators and training in laparoscopic surgery. Int. J. Surg. 13, 60–64. doi: 10.1016/j.ijsu.2014.11.014

PubMed Abstract | CrossRef Full Text | Google Scholar

Zarjam, P., Epps, J., and Chen, F. (2011). “Characterizing working memory load using EEG delta activity,” in Proceedings of the 2011 19th European Signal Processing Conference, Barcelona, 1554–1558

Google Scholar

Keywords: cognitive load, surgical simulator training, EEG biomarker, laparoscopic operations, brain assessment, mental load assessment, machine learning

Citation: Maimon NB, Bez M, Drobot D, Molcho L, Intrator N, Kakiashvilli E and Bickel A (2022) Continuous Monitoring of Mental Load During Virtual Simulator Training for Laparoscopic Surgery Reflects Laparoscopic Dexterity: A Comparative Study Using a Novel Wireless Device. Front. Neurosci. 15:694010. doi: 10.3389/fnins.2021.694010

Received: 12 April 2021; Accepted: 22 November 2021;
Published: 20 January 2022.

Edited by:

Yuanpeng Zhang, Nantong University, China

Reviewed by:

Ji Won Bang, NYU Grossman School of Medicine, United States
Yufeng Yao, Changshu Institute of Technology, China

Copyright © 2022 Maimon, Bez, Drobot, Molcho, Intrator, Kakiashvilli and Bickel. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Neta B. Maimon, bmV0YWNvaDNAbWFpbC50YXUuYWMuaWw=

These authors have contributed equally to this work and share first authorship

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.