- 1Robotics Innovation Center, German Research Center for Artificial Intelligence GmbH, Bremen, Germany
- 2Institute of Medical Technology Systems, University of Duisburg-Essen, Duisburg, Germany
Robot learning based on implicitly extracted error detections (e.g., EEG-based error detections) has been well-investigated in human-robot interaction (HRI). In particular, the use of error-related potential (ErrP) evoked when recognizing errors is advantageous for robot learning when evaluation criteria cannot be explicitly defined, e.g., due to the complex behavior of robots. In most studies, erroneous behavior of robots were recognized visually. In some studies, visuo-tactile stimuli were used to evoke ErrPs or a tactile cue was used to indicate upcoming errors. To our knowledge, there are no studies in which ErrPs are evoked when recognizing errors only via the tactile channel. Hence, we investigated ErrPs evoked by tactile recognition of errors during HRI. In our scenario, subjects recognized errors caused by incorrect behavior of an orthosis during the execution of arm movements tactilely. EEG data from eight subjects was recorded. Subjects were asked to give a motor response to ensure error detection. Latency between the occurrence of errors and the response to errors was expected to be short. We assumed that the motor related brain activity is timely correlated with the ErrP and might be used from the classifier. To better interpret and test our results, we therefore tested ErrP detections in two additional scenarios, i.e., without motor response and with delayed motor response. In addition, we transferred three scenarios (motor response, no motor response, delayed motor response). Response times to error was short. However, high ErrP-classification performance was found for all subjects in case of motor response and no motor response condition. Further, ErrP classification performance was reduced for the transfer between motor response and delayed motor response, but not for the transfer between motor response and no motor response. We have shown that tactilely induced errors can be detected with high accuracy from brain activity. Our preliminary results suggest that also in tactile ErrPs the brain response is clear enough such that motor response is not relevant for classification. However, in future work, we will more systematically investigate tactile-based ErrP classification.
1 Introduction
In recent years, the human-in-the-loop approach has shown great impact on human-robot interaction, and the effects of both explicit and implicit evaluation in robot learning in human-robot interaction have also been well-studied (Daniel et al., 2014; Iturrate et al., 2015; Kim et al., 2017, 2020a,b). In particular, implicitly extracted error detection, e.g., by using electroencephalogram (EEG) is advantageous for robot learning, when evaluation criteria cannot be explicitly defined, e.g., due to the complex behavior of robots or subjective preferences of the interacting person (e.g., Kim et al., 2017).
For detection of implicit evaluation of errors, error-related potentials (ErrPs) have been widely used for various applications (see review, Chavarriaga et al., 2014). In the literature, ErrPs have been classified into four types: response ErrPs (ErrPs elicited by self-induced errors; Falkenstein et al., 2000), feedback ErrPs (ErrPs elicited when recognizing errors after receiving the feedback indicating erroneous events; Holroyd and Coles, 2002), observation ErrPs (ErrPs elicited when observing erroneous actions; van Schie et al., 2004), interaction ErrPs [ErrPs elicited during interaction (Ferrez and Millán, 2008), and i.e., originally interface errors caused by interaction with BCI interfaces].
In robotics, ErrPs have been evoked when observing erroneous actions of robots or systems (Iturrate et al., 2010; Kim and Kirchner, 2013, 2015) or interaction with robots (Kim et al., 2017, 2020a). In human-robot interaction, ErrPs have been used for robot learning (Iturrate et al., 2015; Kim et al., 2017), co-adaptation (Ehrlich and Cheng, 2018), or corrections of erroneous robot behavior (Salazar-Gomez et al., 2017). Furthermore, ErrP-based error detections have been widely used for adaptive classifier of P300 or motor imagery, e.g., by correcting misclassification of P300 (Dal Seno et al., 2010; Combaz et al., 2012; Margaux et al., 2012) or motor-related cortical potentials (MRCPs; Bhattacharyya et al., 2017; Tao et al., 2023) in brain-computer interface (BCIs) area.
In most ErrP-based studies, ErrPs were elicited when recognizing errors visually. For example, ErrPs were elicited when subjects recognized erroneous behavior of robots while observing (e.g., Iturrate et al., 2010; Kim and Kirchner, 2013) or interacting with robots (e.g., Kim et al., 2017). However, tactile-based ErrP classifications have not been widely investigated. Although tactile-based BCIs are not studied as intensively as visual-based BCIs, tactile-based BCIs have a potential for motor rehabilitation and exoskeleton control in rehabilitation applications (e.g., Yao et al., 2022; Lakshminarayanan et al., 2023). In ErrP-based studies, some studies investigated error detections using visuo-tactile stimuli. For example, visual and tactile stimuli were used together to detect errors and these visuo-tactile stimuli evoked ErrPs (Tessadori et al., 2017; Schiatti et al., 2019). In other studies, tactile stimuli were used to indicate upcoming errors, but the errors were recognized visually. Thus, ErrPs were evoked by visual recognition of errors indicated by the tactile cue (Chavarriaga et al., 2012; Ahkami and Ghassemi, 2021). In Perrin et al. (2008), various modalities of stimuli (visual, auditory, tactile) were compared, in which the effect of different modalities of stimuli on reaction time was systematically investigated, but only preliminary results of ErrP classification were shown due to the very small sample size and very large variability between subjects and between stimulus types. In this study, the modalities of stimuli (visual, auditory, tactile) were used to indicated upcoming errors, and the visual recognition of errors evoked ErrPs. In summary, to our knowledge, there were no studies in which ErrPs were evoked when recognizing errors only through tactile channel. In Section 1.1, ErrP-based BCI studies using visuo-tactile stimuli are described in detail.
1.1 Related works
In Tessadori et al. (2017), both visual and tactile stimuli were used together, in which the subjects received a vibration from the wristband (Myo armband) in two of three experiments. Here, the subjects wore a Myo armband on each of their left and right wrists. ErrPs were not evoked when a cursor (green square) moved to the target position (orange square), whereas ErrPs were evoked when the cursor moved in the opposite direction of the target position. The authors hypothesized that the additional use of tactile stimuli would increase ErrP classification performance. In the first experiment, errors were recognized visually without the use of tactile stimuli. In the second experiment, the Myo wristband vibrated according to the direction of the cursor movements, e.g., the Myo armband of the right wrist vibrated when the cursor move to the right or vice versa. In this case, the visual recognition of the cursor movement was congruent with the tactile stimulation. In the third experiment, the Myo armband vibrated not according to the direction of the cursor movements, e.g., the cursor moved to the target position (move to the right), but the Myo armband of the opposite site of the target vibrated (vibration of the Myo armband on the left wrist). That means, the visual recognition of the cursor movement was not congruent with the tactile stimulation. The authors found that classification performance was higher when both visual and tactile stimuli were used together than when only visual stimuli were used (first vs. second experiment or first vs. third experiment). However, there was no difference in classification performance depending on congruency of visual and tactile stimuli caused by a mismatch between the two types of feedback (second vs. third experiment).
In Schiatti et al. (2019), visuo-tactile stimuli were used to evoke ErrPs. ErrPs were not evoked when the cursor moved in the direction depicted by the arrow which indicated the correct direction (e.g., move to the left). When the cursor did not move in the direction pointed by the arrow (e.g., move to right, up, or down), ErrP were evoked. ErrP classification performance was slightly higher when visual and tactile stimuli were used together than when only visual stimuli were used.
In Chavarriaga et al. (2012), a visual (arrow) or tactile stimulus (vibration) was used to indicate the upcoming movement direction of the simulated robot, and ErrPs were evoked during visual recognition of the incongruence between the (visually or tactilely) cued movement direction of the simulated robot and the direction of the actually executed movement of the simulated robot. The authors performed two experiments (visual cue and tactile cue). In the first experiment, a visual cue was used to indicate the upcoming movement direction of the simulated robot (i.e., visually presented arrow). ErrPs were elicited during visual recognition of the incongruence between the visual cue and the direction of the actually executed movement of the simulated robot. In the second experiment, a tactile cue was used to indicate the upcoming movement direction of the simulated robot (vibration instead of visual arrow). ErrPs were elicited during visual recognition of the incongruence between the tactile cue and the direction of the actually executed movement of the simulated robot. In both experiments, the incorrectly chosen movement direction of the simulated robot (errors) was recognized visually and ErrPs were evoked when recognizing these errors.
In Ahkami and Ghassemi (2021), a visual or tactile stimulus was used as a cue to indicate the upcoming movements or the direction of upcoming movements. ErrPs were evoked during the recognition of (1) the incongruence between the cued movements and the absence of movements (i.e., errors) or (2) the incongruence between the cued movement direction and the direction of the actually executed movement (i.e., error). The correct and incorrect condition were predefined depending on movement direction or absence of movement: it was correct if the red square moved to the left (first, third, and fourth experiment), whereas it was incorrect when the red square moved to the right (first, third, and fourth experiment) or when the red square did not move at all (second experiment), indicating that there was no correct condition in the second experiment. In the first experiment, ErrPs were evoked when the red square moved to the right (incorrect condition) and ErrPs were not evoked when the red square moved to the left (correct condition), in which the green and red squares were visually presented. That is, the visual recognition of errors can evoke ErrPs. In the second experiment, according to the authors only tactile stimulation was used to evoke ErrPs. Here, the subjects were told that the red square would move after tactile stimulation and they observed the red square visually. That is, the tactile stimulation was used as a cue to anticipate the movement of the red square, but the errors (i.e., the absence of the movement of the red square) were visually recognized. Thus, visual recognition of the absence of movement of the square (error) did elicit ErrPs, not the recognition of tactile stimulation. In the third and fourth experiments, tactile stimulation was used as a cue to anticipate the movement direction of the red square (i.e., the red square moves to the right or left). If the direction of movement indicated by tactile stimulation (e.g., move to the right) does not match the visual recognition of the direction of movement of the red cursor (e.g., move to the left), ErrPs can be elicited.
In Perrin et al. (2008), six different types of stimulus used to indicate errors were compared in a simulated robot control: visual arrows, visual cursors, auditory tones, auditory words, and vibro-tactile actuator. Subjects were asked to press a button when recognizing error (i.e., incongruence between the visually, auditorily, tactilely cue and the executed action of the simulated robot). Here, the action of the simulated robot was visually recognized. The reaction time was shorter for visual stimuli (visual arrows, visual cursors) than auditory stimuli (auditory tones, auditory words) or tactile stimuli. There was no significant difference in reaction time between auditory and tactile stimuli, but the reaction time was longer for auditory stimuli than tactile stimuli. Especially, the voice cue had the longest reaction time and the largest variation between subjects. ErrP classification performance was very heterogeneous between subjects and between stimulus types so that the comparison between six different types of stimulus was not possible, e.g., the ranking of ErrP classification performance depending on stimulus types was completely different between subjects. Behavioral data (e.g., reaction time) was collected from twenty two subjects, but EEG data was recorded only from four subject of them. For this reason, the authors noted that the results of ErrP classification were preliminary.
In summary, visual and tactile stimuli were used together to evoke ErrPs (Tessadori et al., 2017; Schiatti et al., 2019) or tactile stimuli were used to indicate upcoming errors and the visual recognition of errors evoked ErrPs (Perrin et al., 2008; Chavarriaga et al., 2012; Ahkami and Ghassemi, 2021).
1.2 Approaches and goals
In our study, we used tactile stimuli directly to evoke ErrPs without the combination with the visual channel. We also did not use any other cues (visual, auditory, or tactile cue) to indicate upcoming errors. We aimed to evoke ErrPs when recognizing errors only through tactile channel.
In our scenario, the subjects wore an orthosis on their right arm and performed arm movements (details, see Section 2.2). Sometimes, the orthosis did not work correctly, i.e., the orthosis briefly moved in the opposite direction of the intended movements of subjects. This malfunction of the orthosis was preprogrammed to induce errors. Here, we expected ErrPs when subjects recognized such malfunction of the orthosis (i.e., interaction errors) only through the tactile channel.
Since we did not find studies in which ErrPs were evoked when tactile errors were recognized, we asked the subjects to press a button when they felt the malfunction of the orthosis to ensure that the subject can detect tactile based errors. However, in our previous studies (Kim et al., 2017, 2020a,b), we have successfully classified ErrPs without motor response (e.g., button press).
On the other hand, we assumed that the motor potential can also be used for feature extraction of ErrP classification because the latency between the occurrence of errors and the response to errors (e.g., button press) was short. Therefore, we performed additional tests, i.e., test experiments, in a scenario, in which a motor response was not requested after error detection to better interpret our results and to ensure that ErrP-classification performance might be increased or decreased, e.g., without motor response (see Figure 3 and Table 5). To this end, we additionally recorded 10 datasets from one subject (Subject 1) in the experiments, where the subject was not asked to press a button after error detection. We also conducted experiments in a scenario, in which a subject was asked to delay a motor response after error detection to avoid including features possibly used for ErrP detection. To this end we recorded three datasets from one subject (Subject 8) in both scenarios: (a) ErrP detections without motor response and (b) ErrP detections with delayed motor response. Here, three scenarios (motor response, no motor response, delayed motor response) were transferred to better interpret our assumption (see Figure 3 and Table 4). However, these additional experiments only serve us as a preview for future work will need further experiments that were out of the focus of this work.
2 Methods
2.1 Subjects
Eight healthy subjects (four male and four female; 21.8 ± 2.4 ages; right-handed; students) participated in the experiments. All experiments were carried out in accordance with the approved guidelines. Experimental protocols were approved by the ethics committee of the department of Computer Science and Applied Cognitive Science of the Faculty of Engineering at the University of Duisburg-Essen. Written informed consent was obtained from all participants that volunteered to perform the experiments.
2.2 Experimental setup
Figure 1 shows the experimental setup. The subjects wore an orthosis on their right arm and held a button with their left hand. The orthosis was developed at the DFKI (details, see Kueper et al., 2023). The subjects performed arm movements consisting of flexions and extensions. We programmed the orthosis to move in the opposite direction of the subjects' intended movements with a certain probability for 0.25s. For example, subjects intended to extend their arm wearing the orthosis, but the orthosis briefly moved in the opposite direction of the intended movements. We hypothesized that such simulated malfunction of the orthosis, i.e., erroneous actions of the orthosis (errors) would elicit ErrPs. The subjects were asked to press a button when they felt that the orthosis was not functioning properly, i.e., when subjects felt erroneous actions of the orthosis. The button press was intended to serve as a response to the error detection. The responses (i.e., button presses) of subjects were sent to the EEG recording system so that the responses were written in the EEG data in real time (see Figure 1 response markers). The onset of movements (the onset of extension and flexion) and the onset of stimulated errors (i.e., incorrect orthosis behavior) were also sent to the EEG recording system (see Figure 1 movement markers and error markers, i.e., markers for error trials). Note that the orthosis position is denoted as −10° when the arm is fully extended, whereas the orthosis position is denoted as −90° when the arm is fully flexed. Malfunction of the orthosis (errors) are simulated to make the orthosis move in the opposite direction with the mean error position of −42° (flexion) and −58° (extension), respectively. The experiment procedure is described in detail in the next section (Section 2.3 and Figure 2).
Figure 1. Experimental setup. The subjects wore an orthosis on their right arm and held a button with their left hand. In the main experiment, the subjects were instructed to press the button when they recognized tactile errors, i.e., incorrect behavior of the orthosis (e.g., the orthosis moved in the opposite direction of the intended movements of the subjects). We expect ErrPs when recognizing tactile errors.
Figure 2. Experimental design. A baseline set without tactile errors (a) and the practice set including tactile errors (b) were recorded before the main experiment (c, d), which divided into two parts: 5 sets (c) before the break and 5 sets after the break (e).
2.3 Experimental design and procedure
Figure 2 shows the experimental design. Before the main experiment (see Figure 2a), subjects performed 30 movements (15 flexions, 15 extensions) without errors to familiarize themselves with the system (orthosis). After that, the practice set was performed to learn the experiment procedure of the main experiment (see Figure 2b). The main experiment consists of three parts: recordings of five sets, break, and recordings of further five sets (see Figures 2c–e).
In the main experiment, it was implemented that errors occur randomly with a probability of 20%. Two facts emerged from this: (a) errors can occur during the last movement (the last, i.e., 30th flexion and the 30th extension) and (b) 6 errors (i.e., simulated malfunction of the orthosis) occur within 30 arm movements (i.e., 24 correct trials and 6 incorrect trials). In the main study, we added two movement (one flexion, one extension) to avoid that the final trials (flexion and extension) did not contain simulated errors. Thus, the subject performed additional two movements after performing 30 movements. This results in two facts: (a) each set contained 32 movements (16 flexions and 16 extensions) and (b) the last movements cannot contain errors, since the errors were simulated within the first 30 arm movements. Finally, each set contained 26 correct trials and 6 incorrect trials (i.e., 6 errors).
We defined a trial as an extension movement or a flexion movement. We calculated the median value of trial duration since the trial duration varied depending on sets and subjects. The median trial duration was ~7 s across all subjects (details, see Table 1). The interval between trials (i.e., inter-trial interval) also varied across subjects, e.g., some subjects had a longer break between trials than others. The inter-trial interval ranged between 0.5 and 1 s per set. Thus, the task duration was between 4 and 5 min per set (32 trials for each set). We measured the task duration of 22 min for each recording phase (see Figures 2c, e) across all subjects. Hence, the duration of the main experiment was 44 min excluding the break between two recording phases, the practice set and the baseline experiment. According to the experimental design (Figure 2), the whole experiment should take 1 h and 7 min. However, the actual duration of the experiment was ~2 h, since we checked impedance and correct transfer of the markers in the EEG data file (see Figure 1) after each measurement (i.e., after recording of each set), which took ~5–10 min per set, and from 50 min to 1 h for all 10 sets per subject. Furthermore, the preparation of EEG and EMG measurements took between 1 and 2 h depending on subjects, where impedance was kept below 5 kΩ. Thus, the duration of the whole experiment including experiment preparation was between 3 and 4 h.
In addition to our main scenario described above, one subject (Subject 8) participated in two additional scenarios after the main scenario (ButtonPress scenario), in which (a) the subject was instructed not to give motor response after error detection (NoButtonPress scenario) and (b) the subject was asked to artificially delay a motor response (DelayedButtonPress scenario). Here, we recorded three datasets for the DelayedButtonPress scenario and three datasets for the NoButtonPress scenario, because the main study was expected to take 2 h. We started the following experiment sequences: NoButtonPress scenario, DelayedButtonPress scenario, and ButtonPress scenario. Indeed, the main study took 1 h 50 min. Recording more than 16 EEG data sets (3 h and 30 min without the preparation time) was not realistic. Note that 10 datasets from the main scenario (ButtonPress scenario) were used to train a classifier to evaluate the datasets from DelayedButtonPress scenario or NoButtonPress scenario (details, see Section 2.8.4 and Figure 3).
Figure 3. Evaluation design. For the main study (a), a 10-fold cross validation was applied for evaluation. For the NoButtonPress scenario (b), a 10-fold cross validation was also applied for evaluation, where the same subject (Subject 1) participated in the ButtonPress scenario and the NoButtonPress scenario on different days. For transfer learning (c, d), the 10 datasets from the ButtonPress scenario (main study) was used to train a classifier. This trained classifier was used to evaluate the data from the NoButtonPress scenario (c) or DelayedResponse Press scenario (d).
Another subject (Subject 1) participated in our main scenario (ButtonPress scenario) on one day and in the scenario where motor response was not required after error detection (NoButtonPress scenario) on another day. Here, we recorded 10 datasets for the NoButtonPress scenario just like for the main scenario (ButtonPress scenario). Due to the long duration of main study (at least 2 h and 30 min), it was not possible to record 10 datasets for each scenario (ButtonPress and NoButtonPress) at the same day from the same subject. The reason for this additional data acquisition was explained in Section 1.2.
2.4 Dataset
In the main experiment, we recorded 10 EEG datasets per subject. Each dataset contained 26 correct trials and 6 incorrect trials. Thus, we collected a total of 260 correct trials and 60 incorrect trials per subject. As mentioned earlier, we carried out experiments in two additional scenarios to test whether the ErrP-classification performance might be increased or decreased, e.g., without motor response. Thus, for one subject (Subject 8), we additionally recorded three datasets for error detection without motor response (NoButtonPress) and three datasets for error detection with delayed motor response (DelayedButtonPress). In total, for Subject 8, we recorded 10 datasets for error detection with motor response (ButtonPress), three datasets for error detection without motor response (NoButtonPress), and three datasets for error detection with delayed motor response (DelayedButtonPress). In the end, we also recorded 10 EEG datasets for one subject (Subject 1) on another day for error detection without motor response (NoButtonPress). In this way, we recorded 10 EEG datasets for error detection with motor response (ButtonPress) and 10 EEG datasets for error detection without motor response (NoButtonPress). For our main analysis, we used 10 EEG datasets with motor response from eight subjects.
2.5 Behavioral data acquisition
We recorded the behavioral data of subjects (e.g., response to errors) to analyze the accuracy of responses and response time per subject (see Figure 1).
2.6 EEG data acquisition
As mentioned earlier in Section, 2.4 we recorded EEG datasets from eight subjects. EEGs were continuously recorded using the actiCapSlim system (Brain Products GmbH, Munich, Germany), in which 64 active electrodes were arranged in accordance to an extended 10-20 system with reference at FCz. Impedance was kept below 5 kΩ. EEG signals were sampled at 500 Hz, amplified by two 32 channel BrainLiveAmp amplifiers (Brain Products GmbH, Munich, Germany), and filtered with a low cut-off of 0.1 Hz and high cut-off of 1 kHz. We sent EEG markers (labels) for relevant events to the EEG recording system to write the EEG markers into the continuous EEGs (see Figure 1): (a) movement onset for flexion, (b) movement onset for extension, (c) onset of simulated erroneous behavior of the orthosis (i.e., ErrP label), and (d) onset of button press to erroneous behavior of the orthosis (errors). In addition, we also sent a marker after the start of movement (flexion or extension) to obtain the NoErrP label, in case that the orthosis performed correctly.
2.7 Behavioral data analysis
We calculated response time, i.e., reaction time (RT) to error detection. We also calculated the number of undetected errors, i.e., false negatives (FN) and the number of incorrect responses, i.e., false positive (FP) responses. The analysis of the behavioral data was later used to exclude the epochs with erroneous response for both classes, e.g., button press for correct behavior of the orthosis or no button press for incorrect behavior of the orthosis (Details, see Section 2.8.1).
2.8 EEG data analysis
2.8.1 EEG preprocessing
The EEG data was analyzed using a Python-based framework for preprocessing and classification (Krell et al., 2013) containing algorithms and external packages for feature extraction and classification [e.g., xDAWN (Rivet et al., 2009), pyRiemann (Barachant, 2015; Barachant et al., 2022)]. The continuous EEG signal was segmented into epochs from −0.1 to 1 s for each event type (correct/erroneous trial). Here, we excluded the epochs containing an incorrect response from EEG analysis. For example, epochs in which subjects pressed the button although the behavior of the orthosis was correct (false positive). On the other hand, we excluded epochs in which subjects did not press the button even though the behavior of the orthosis was incorrect (false negative). Hence, only the epochs with correct responses to errors (incorrect orthosis behavior) were labeled as incorrect, and the epochs with correct responses, i.e., no responses, to correct orthosis behavior were labeled as correct. All epochs were normalized to zero mean for each channel, decimated to 50 Hz, and band pass filtered (0.1–12 Hz).
2.8.2 Feature extraction and classification
Features were extracted per subject. All datasets (i.e., 10 datasets) were concatenated per subject. The xDAWN spatial filter (Rivet et al., 2009) was used to enhance the signal-to-noise ratio. By applying the xDAWN the number of 64 physical channels was reduced to 7 pseudo channels. In this way, the signal-to-noise ratio for the positive class, i.e., incorrect class was maximized. All epochs were projected into the pseudo channels, i.e., 350 data points (7 channels × 50 data points) were obtained after applying xDAWN.
After applying xDAWN, we used a Riemmanian manifold approach (for review Yger et al., 2016; Congedo et al., 2017). First, we generated extended epochs (Barachant and Congedo, 2014) and obtained 14 pseudo channels (7·2 = 14 channels). We estimated a 14 × 14-dimensional covariance matrix across the 50 data points for each extended epoch. To this end, we used the shrinkage regularized estimator of Ledoit-Wolf (Ledoit and Wolf, 2004), which ensures that the estimated covariance matrices are positively defined. After estimating the covariance matrices, we approximated their Riemannian center of mass (Fréchet mean; Cartan, 1929) or often called geometric mean in BCI applications. This Riemannian center of mass was used as reference point to append a tangent space. All training and testing data were projected into this tangent space and vectorized using Mandel notation. Using Mandel notation, we reduced the symmetric 14 × 14-dimensional matrices into 105-dimensional feature vectors. After that, we normalized the feature vectors. For classification, we used a linear Support Vector Machine (SVM; Cortes and Vapnik, 1995; Mangasarian and Musicant, 1999). We optimized the cost parameter of the SVM (i.e., regularization constant; Schölkopf et al., 2000). To this end, the hyperparameter C of the SVM was selected from {10−6, 10−5, 10−4, 10−3, 10−2, 10−1, 1} using a five-fold stratified cross validation, in which the correct (NoErrP) and incorrect (ErrP) classes were weighted as 1:2.
2.8.3 Event-related potential analysis
For ERP analysis, we analyzed the EEG data using EEGLAB1. For preprocessing, the raw EEGs were downsampled to 250 Hz, re-referenced to an average reference, and filtered between 0.1 and 15 Hz. The FCz channel, used as a reference in the EEG recording, was recalculated as an EEG channel for ERP analysis. After preprocessing, independent component analysis (ICA) was applied for artifact removal. We used Infomax ICA to remove artifacts (e.g., muscle, eye blinks, or eye movements) by subtracting ICA components containing eye artifacts. After artifact removal, EEG data were segmented into epochs from 0.1 to 1 s after each event type (correct/incorrect). Epochs were averaged within each event type with a baseline correction (-0.1 s until stimulus onset) per subject. For calculation of grand average ERP, the ERPs of all subjects were averaged.
2.8.4 Evaluation
Figure 3 shows the evaluation design for our study. For evaluation, we used two labels: correct behavior of the orthosis (NoErrP) and incorrect behavior of the orthosis (ErrP). As mentioned earlier, we concatenated the epochs of the 10 datasets and obtained a total of 260 correct trials and 60 incorrect trials per subject. A ten fold stratified cross validation was applied on the concatenated datasets per subject (see Figure 3a). Thus, we obtained the classification performance for each subject. For performance metric, we used a balanced accuracy, i.e., the arithmetic mean of true positive rate (TRP) and true negative rate (TNR). We also additionally reported TPR (recall), TNR, F measure, and precision (Details, see Section 3.2). Note that the positive class stands for incorrect behavior of the orthosis and negative class stands for correct behavior of the orthosis. For transfer learning, we trained a classifier based on the 10 datasets recorded in the scenario (ButtonPress scenario), in which motor response was required after error detection. This trained classifier was used to test the three datasets and three datasets, respectively recorded in the scenario (NoButtonPress scenario) where motor response was not required after error detection (see Figure 3c) or in the scenario (DelayedButtonPress scenario) where delayed response was required after error detection (see Figure 3d). For transfer learning, as mentioned earlier, we evaluated datasets from both scenarios for one subject (Subject 8). For the evaluation of NoButtonPress scenario, the same evaluation process was used for the evaluation of ButtonPress scenario (main study) to compare the two scenarios ButtonPress and NoButtonPress (see Figure 3b), where the same subject (Subject 1) participated in both scenarios on different days.
3 Results
3.1 Behavior data
Figure 4 and Table 2 shows the results of response time, i.e., reaction time (RT) to errors. Figure 4A and Table 2A show the median value of reaction time for each set across eight subjects whereas Figure 4B and Table 2A show the median value of reaction time for each subject across 10 sets. We obtained a mean of 0.746 s RT across all subjects (Table 2A) and a mean of 0.728 s RT across all sets (Table 2B). However, as shown in Figure 4A, a high variation in reaction time between subjects was observed for all sets. Similarly, the variation in reaction time between sets was observed for all subjects except for Subject 4 and Subject 7 (Figure 4B). For example, the reaction time after half of the experiments (from Set 6) was longer in Subject 3 (Figure 5A), whereas the reaction time was highly different between sets for Subject 8 (Figure 5B). Figure 6 shows the results of response time, i.e., reaction time (RT) to errors in the experiment, in which the subject (subject 8) was asked to delay his motor response after error recognition. We obtained a mean of 4.54 s RT across all sets for this subject.
Figure 4. Response time to erroneous behavior of the orthosis (i.e., reaction time) for all sets and all subjects. (A) Reaction time per set across eight subjects. (B) Reaction time per subject across 10 sets.
Figure 5. Response time to erroneous behavior of the orthosis (i.e., reaction time) for Subject 3 and Subject 8 as example. (A) Reaction time (subject 3). (B) Reaction time (subject 8).
Figure 6. Response time to erroneous behavior of the orthosis (i.e., reaction time) for Subject 8 in the experiment, where the subject was asked to delay his motor response after error detection (DelayedButtonPress experiment).
We also analyzed the accuracy of the response to errors. We have calculated the number of undetected errors, i.e., false negatives (FN). A total of 471 errors were detected from 480 errors (6 errors × 10 sets × 8 subjects). This results in an average error detection accuracy of 98.82% for all subjects. In other words, we obtained a mean of 1.88% FN across all subjects. We also calculated the number of incorrect responses, i.e., false positive (FP) responses. (i.e., the button was pressed even though there was no error). We obtained a mean of 0.83% FP for all subjects.
3.2 EEG data
3.2.1 EEG classification performance
Table 3 shows ErrP-classification performance. We obtained high ErrP classification performance for all subjects. The mean value of TNR was slightly higher compared to TPR, i.e., the false positive rate (1-TNR) was less than the false negative rate (1-TPR). We also found more variability between subjects for TPR than TNR.
Table 4 shows ErrP-classification performance for transfer learning. As mentioned earlier, we performed experiments in two additional scenarios to evaluate whether ErrP-classification performance might be increased or decreased, e.g., without motor response. We trained a classifier using the datasets recorded in the main scenario, in which motor response was required after error recognition (ButtonPress). This trained classifier was used to test (a) the datasets recorded in the scenario, in which the subjects were instructed not to give motor response after error recognition (NoButtonPress) or (b) the datasets recorded in the scenario, where subjects were asked to artificially delay motor response after error detection (DelayedButtonPress).
As shown in Table 4A, we still observed a high TPR, but a slightly lower TNR when transferring from the scenario with motor response to the scenario without motor response. That means, the rate of error detection (ErrP detection) was still high, but there were many false alarms, i.e., the false positive rate was high (1-TNR). In contrast, we observed a very low TPR, but a high TNR when transferring from the scenario with motor response to the scenario with delayed motor response (see Table 5B). That means, the false alarm (false positive rate) was not high, but the rate of error detection (ErrP detection) was very low.
Table 5 shows the comparison between two scenarios (motor response after error detection vs. no motor response after error detection) in ErrP-classification performance in the same subjects (subject 1). Note that the datasets were not recorded on the same day. We achieved high ErrP-classification performance for both scenarios. That means, we obtained high ErrP-classification performance in the scenario, in which motor response was not required after error detection.
3.2.2 Event-related potential analysis
Figure 7 shows the grand average ERPs in our main study (ButtonPress scenario) at Fz, FCz, Cz, and Pz, in which the ERPs averaged over all trials per condition (correct/incorrect) for each subject were in turn averaged over all subjects. We observed the first negativity around 250 ms followed by a positivity between 300 and 500 ms and a further negativity around 600 ms. Here, the ERP shape (i.e., temporal sequences of ErrPs) is identical to other ErrP studies (e.g. Chavarriaga et al., 2012). However, the first negativity was suppressed due to the strong positivity in its minus polarity. As expected from ERP pattern evoked by interaction errors, the first and second negativity was smaller at Cz and Pz than at Fz and FCz.
Figure 7. Grand average ERPs in our main study (ButtonPress scenario) at Fz, FCz, Cz, and Pz. The blue curve stands for the ERPs averaged over all correct trials (S48), whereas the red curve stands for the ERPs averaged over all incorrect trials (S96). Grand average ERP stands for the ERPs averaged across all subjects for each condition (correct/incorrect).
Figure 8 shows the ERP curve averaged across all correct trials (S48, blue curve) and all incorrect trials (S96, red curve) for each subject at FCz, in which we expect maximum ErrPs at fronto-central region (FCz). We observed that Subject 7 showed most clearly the expected classic ErrP form (first negativity followed by positivity and second negativity).
Figure 8. ERPs in our main study (ButtonPress scenario) for each subject at FCz. The blue curve stands for the ERPs averaged over all correct trials (S48), whereas the red curve stands for the ERPs averaged over all incorrect trials (S96).
Figure 9 shows the ERP curve averaged across all correct trials (S48, blue curve) and all incorrect trials (S96, red curve) for Subject 1 in comparison for both ButtonPress scenario and NoButtonPress scenario. The ErrP shape is similar in both scenarios.
Figure 9. ERPs for Subject 1 at Fz, FCz, Cz, and Pz in both ButtonPress scenario and NoButtonPress scenario. The blue curve stands for the ERPs averaged over all correct trials (S48), whereas the red curve stands for the ERPs averaged over all incorrect trials (S96).
4 Discussion
In our study, we developed a scenario in which tactilely triggered ErrPs were examined for tactilely perceived misbehavior of an orthosis. The behavioral results showed that interaction errors mediated by erroneous behavior of the orthosis via the tactile channel were detected with a high accuracy (98.82%) by all subjects. Our EEG results showed that it is feasible to detect ErrPs, which are elicited when interaction errors are detected only via the tactile channel without visual, auditory, or tactile cues. Thus, our results revealed that it was also feasible to detect tactile-based ErrPs without combining them with visual error recognition.
Further, the comparison in ErrP-classification performance between two scenarios (motor response after error detection and no motor response after error detection) revealed that ErrP detection is also feasible without making use of features from motor activity (Table 5). Note that we performed this evaluation on one subject (10 datasets recorded from the scenario motor response after error detection vs. 10 datasets recorded from the scenario no motor response after error detection) and the results are therefore preliminary. These results are consistent with our previous studies (Kim et al., 2017, 2020a,b), in which subjects observed the robot's behavior while interacting with the robot and did not perform a motor response (e.g., press a button) after detecting erroneous actions of a robot.
Furthermore, reaction-time analysis show that the mean response time across all subjects was 746 ms. This means that the motor response occurred very late, which can only affect the last part of the ERP evoked by errors (i.e., the second negativity). Moreover, response time is not very strongly time locked to the error, as reaction-time analysis has shown. For this reason, we expect that the effect of motor response should be small. As mentioned in Section 1, in many ErrP studies, the recognition of errors elicited ErrPs without any motor response. In most studies, errors were recognized visually while observing/monitoring errors (visual feedback) without a motor response, and this visual recognition of errors (error monitoring) elicited ErrPs. Furthermore, there was an ErrP study (Chavarriaga et al., 2012) conducted with motor response (KeyPress after error recognition) and without motor response (error monitoring) using visuo-tactile stimuli. In this study, ErrPs were evoked under both conditions, and the ERP shape was similar under both conditions. Although our study cannot prove that motor activity does not contribute to the observed ERP curve, the additional experiment with motor response in subject 1 showed that a similar ERP shape was observed in case of no motor response. We want to emphasize that we required motor response to ensure that the subjects recognized tactile-based errors.
An interesting result is that ErrP-classification performance was only slightly reduced (91%) when we applied the classifier trained on the datasets containing interaction errors with motor response to test the datasets containing interaction errors without motor response (Table 4A). These results suggest that the features used for motor potential detection are not very relevant for ErrP detection. In contrast, the ErrP-classification performance was substantially reduced (74%) when we used the classifier trained on the datasets containing interaction errors with motor response to test the datasets containing interaction errors with delayed motor response (Table 4B). These findings suggest that features used from motor related activity could have a negative effect on ErrP detection when the motor response was artificially delayed by subjects. Again, these results are preliminary and need further studies. We performed theses tests to develop more ideas into future possible experiments.
Another interesting preliminary finding is that we observed a very high true positive rate (100%) and an increased false positive rate (17%) when we transferred the classifier trained on the datasets containing interaction errors with motor response to the datasets containing interaction errors without motor response (Table 4A). Here, we assumed that the effect of maximizing the positive class (incorrect class, i.e., ErrP label) during feature extraction (details, spatial filtering in Section 2.8.2) might become stronger when the test data did not include features detecting motor potentials. For this reason, the true positive rates could be increased whereas the false positive rate could be decreased. On the other hand, we observed a very low true positive rate (53%) and a low false positive rate (4%) when we transferred the classifier trained on the datasets containing interaction errors with motor response into the datasets containing interaction errors with delayed motor response (Table 4B). Again, the results suggest that features used for motor potential detection could have a negative impact on ErrP detection when the motor response was artificially delayed by subjects.
However, we studied only one subject for transfer learning. Therefore, the results for transfer learning are preliminary and it is necessary to investigate this systematically with more subjects in future work. We have shown with our preliminary results that our main scenario that allows us to prove that subjects detect tactilely induced errors, can be adapted to gain more insight into the correlation between different EEG activities by making modifications such as those for the NoButtonPress or DelayedButtonPress conditions. Thus, our preliminary findings from classifier transfer and adaptation of response situation opens up interesting research questions for future experiments worth it to be investigated.
The ERP results (Figure 7) indicates that ErrPs could be also shown at frontal region (e.g., Fz) when errors were introduced by tactile stimuli. These results are consistent with the findings of Chavarriaga et al. (2012), which showed a similar ERP morphology (first negativity followed positivity and a further negativity) at frontal region (Fz) when errors were introduced by visuo-tactile stimuli. In our study, the positivity was stronger than in other ErrP studies (review see Chavarriaga et al., 2014). We first assumed that the task-relevant event (i.e., the button press when recognizing errors instead of passively observing errors) might induce a stronger positivity (i.e., P300) in our study. However, the test with one subject (Subject 1) in the NoButtonPress scenario compared to the data recorded for this subject in the ButtonPress scenario did not support this hypothesis. We found no strong differences for this one subject under the response and the no response condition (see Figure 9). This interesting preliminary result opens up future research questions on the processing of tactile stimuli. Further studies are needed in the future.
In recent years, there has been increased interest and impact of the human-in-the-loop approach to human-robot interaction and also to tele-rehabilitation (e.g., therapist-in-the-loop in assisted as needed approach). Here, tactile feedback (tactile stimulation or force feedback) has a great importance for a better cooperation with the users and a better support of the patients. Therefore, in the future, we want to investigate errors or resistance that are received via tactile channels from users or patients, which can occur during human-robot interaction in our robotic supported (tele-) rehabilitation scenarios.
Data availability statement
The datasets presented in this study can be found in online repositories. The names of the repository/repositories and accession number(s) can be found at: https://zenodo.org/record/8345429.
Ethics statement
The studies involving humans were approved by Ethics Committee of the Department of Computer Science and Applied Cognitive Science of the Faculty of Engineering at the University of Duisburg-Essen. The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their written informed consent to participate in this study.
Author contributions
SK: Conceptualization, Investigation, Methodology, Validation, Visualization, Writing—original draft. EK: Conceptualization, Funding acquisition, Project administration, Resources, Supervision, Writing—review & editing.
Funding
The author(s) declare financial support was received for the research, authorship, and/or publication of this article. This work was funded by the German Federal Ministry of Education and Research (BMBF) within the project M-ROCK (grant number: 01IW21002) and supported by the Institute of Medical Technology Systems at the University of Duisburg-Essen and German Research Center for Artificial Intelligence (DFKI).
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher's note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Footnotes
References
Ahkami, B., and Ghassemi, F. (2021). Adding tactile feedback and changing ISI to improve BCI systems' robustness: an error-related potential study. Brain Topogr. 34, 467–477. doi: 10.1007/s10548-021-00840-6
Barachant, A., Barthélemy, Q., King, J.-R., Gramfort, A., Chevallier, S., Rodrigues, P. L., et al. (2022). Release 0.3 Toggle Dropdown.
Barachant, A., and Congedo, M. (2014). A plug&play p300 BCI using information geometry. arXiv preprint arXiv:1409.0107.
Bhattacharyya, S., Konar, A., and Tibarewala, D. (2017). Motor imagery and error related potential induced position control of a robotic arm. IEEE/CAA J. Autom. Sin. 4, 639–650. doi: 10.1109/JAS.2017.7510616
Cartan, É. (1929). Groupes simples clos et ouverts et géométrie riemannienne. J. Math. Pures Appl. 8, 1–33.
Chavarriaga, R., Perrin, X., Siegwart, R., and Millán, J. d. R. (2012). “Anticipation-and error-related EEG signals during realistic human-machine interaction: a study on visual and tactile feedback,” in 2012 Annual International Conference of the IEEE Engineering in Medicine and Biology Society (San Diego, CA), 6723–6726. doi: 10.1109/EMBC.2012.6347537
Chavarriaga, R., Sobolewski, A., and Millán, J. d. R. (2014). Errare machinale EST: the use of error-related potentials in brain-machine interfaces. Front. Neurosci. 8:208. doi: 10.3389/fnins.2014.00208
Combaz, A., Chumerin, N., Manyakov, N. V., Robben, A., Suykens, J. A., and Van Hulle, M. M. (2012). Towards the detection of error-related potentials and its integration in the context of a p300 speller brain-computer interface. Neurocomputing 80, 73–82. doi: 10.1016/j.neucom.2011.09.013
Congedo, M., Barachant, A., and Bhatia, R. (2017). Riemannian geometry for eeg-based brain-computer interfaces; a primer and a review. Brain-Comput. Interfaces 4, 155–174. doi: 10.1080/2326263X.2017.1297192
Cortes, C., and Vapnik, V. (1995). Support-vector networks. Mach. Learn. 20, 273–297. doi: 10.1007/BF00994018
Dal Seno, B., Matteucci, M., Mainardi, L., et al. (2010). Online detection of p300 and error potentials in a BCI speller. Comput. Intell. Neurosci. 2010:307254. doi: 10.1155/2010/307254
Daniel, C., Viering, M., Metz, J., Kroemer, O., and Peters, J. (2014). “Active reward learning,” in Proceedings of Robotics: Science and Systems (Rome). doi: 10.15607/RSS.2014.X.031
Ehrlich, S. K., and Cheng, G. (2018). Human-agent co-adaptation using error-related potentials. J. Neural Eng. 15:066014. doi: 10.1088/1741-2552/aae069
Falkenstein, M., Hoormann, J., Christ, S., and Hohnsbein, J. (2000). ERP components on reaction errors and their functional significance: a tutorial. Biol. Psychol. 51, 87–107. doi: 10.1016/S0301-0511(99)00031-9
Ferrez, P. W., and Millán, J. d. R. (2008). Error-related EEG potentials generated during simulated brain-computer interaction. IEEE Trans. Biomed. Eng. 55, 923–929. doi: 10.1109/TBME.2007.908083
Holroyd, C. B., and Coles, M. G. (2002). The neural basis of human error processing: reinforcement learning, dopamine, and the error-related negativity. Psychol. Rev. 109:679. doi: 10.1037/0033-295X.109.4.679
Iturrate, I., Chavarriaga, R., Montesano, L., Minguez, J., and Millán, J. d. R. (2015). Teaching brain-machine interfaces as an alternative paradigm to neuroprosthetics control. Sci. Rep. 5:13893. doi: 10.1038/srep13893
Iturrate, I., Montesano, L., and Minguez, J. (2010). “Single trial recognition of error-related potentials during observation of robot operation,” in 2010 Annual International Conference of the IEEE Engineering in Medicine and Biology (Buenos Aires), 4181–4184. doi: 10.1109/IEMBS.2010.5627380
Kim, S. K., and Kirchner, E. A. (2013). “Classifier transferability in the detection of error related potentials from observation to interaction,” in 2013 IEEE International Conference on Systems, Man, and Cybernetics (Manchester: IEEE), 3360–3365. doi: 10.1109/SMC.2013.573
Kim, S. K., and Kirchner, E. A. (2015). Handling few training data: classifier transfer between different types of error-related potentials. IEEE Trans. Neural Syst. Rehabil. Eng. 24, 320–332. doi: 10.1109/TNSRE.2015.2507868
Kim, S. K., Kirchner, E. A., and Kirchner, F. (2020a). “Flexible online adaptation of learning strategy using EEG-based reinforcement signals in real-world robotic applications,” in 2020 IEEE International Conference on Robotics and Automation (ICRA), 4885–4891. doi: 10.1109/ICRA40945.2020.9197538
Kim, S. K., Kirchner, E. A., Schloßmüller, L., and Kirchner, F. (2020b). Errors in human-robot interactions and their effects on robot learning. Front. Robot. AI 7:558531. doi: 10.3389/frobt.2020.558531
Kim, S. K., Kirchner, E. A., Stefes, A., and Kirchner, F. (2017). Intrinsic interactive reinforcement learning-using error-related potentials for real world human-robot interaction. Sci. Rep. 7, 1–16. doi: 10.1038/s41598-017-17682-7
Krell, M. M., Straube, S., Seeland, A., Wöhrle, H., Teiwes, J., Metzen, J. H., et al. (2013). Pyspace-a signal processing and classification environment in python. Front. Neuroinform. 7:40. doi: 10.3389/fninf.2013.00040
Kueper, N., Chari, K., Bütefür, J., Habenicht, J., Kim, S. K., Rossol, T., et al. (2023). EEG and EMG dataset for the detection of errors introduced by an active orthosis device. arXiv preprint arXiv:2305.11996.
Lakshminarayanan, K., Shah, R., Daulat, S. R., Moodley, V., Yao, Y., Sengupta, P., et al. (2023). Evaluation of EEG oscillatory patterns and classification of compound limb tactile imagery. Brain Sci. 13:656. doi: 10.3390/brainsci13040656
Ledoit, O., and Wolf, M. (2004). A well-conditioned estimator for large-dimensional covariance matrices. J. Multivar. Anal. 88, 365–411. doi: 10.1016/S0047-259X(03)00096-4
Mangasarian, O. L., and Musicant, D. R. (1999). Successive overrelaxation for support vector machines. IEEE Trans. Neural Netw. 10, 1032–1037. doi: 10.1109/72.788643
Margaux, P., Emmanuel, M., Sébastien, D., Olivier, B., and Jérémie, M. (2012). Objective and subjective evaluation of online error correction during p300-based spelling. Adv. Hum. Comput. Interact. 2012:578295. doi: 10.1155/2012/578295
Perrin, X., Chavarriaga, R., Ray, C., Siegwart, R., and Millán, J. d. R. (2008). “A comparative psychophysical and EEG study of different feedback modalities for HRI,” in Proceedings of the 3rd ACM/IEEE International Conference on Human Robot Interaction, 41–48. doi: 10.1145/1349822.1349829
Rivet, B., Souloumiac, A., Attina, V., and Gibert, G. (2009). xdawn algorithm to enhance evoked potentials: application to brain-computer interface. IEEE Trans. Biomed. Eng. 56, 2035–2043. doi: 10.1109/TBME.2009.2012869
Salazar-Gomez, A. F., DelPreto, J., Gil, S., Guenther, F. H., and Rus, D. (2017). “Correcting robot mistakes in real time using EEG signals,” in 2017 IEEE international conference on robotics and automation (ICRA) (Singapore), 6570–6577. doi: 10.1109/ICRA.2017.7989777
Schiatti, L., Barresi, G., Tessadori, J., King, L. C., and Mattos, L. S. (2019). “The effect of vibrotactile feedback on ErrP-based adaptive classification of motor imagery,” in 2019 41st Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC) (Berlin), 6750–6753. doi: 10.1109/EMBC.2019.8857192
Schölkopf, B., Smola, A. J., Williamson, R. C., and Bartlett, P. L. (2000). New support vector algorithms. Neural Comput. 12, 1207–1245. doi: 10.1162/089976600300015565
Tao, T., Jia, Y., Xu, G., Liang, R., Zhang, Q., Chen, L., et al. (2023). Enhancement of motor imagery training efficiency by an online adaptive training paradigm integrated with error related potential. J. Neural Eng. 20:016029. doi: 10.1088/1741-2552/acb102
Tessadori, J., Schiatti, L., Barresi, G., and Mattos, L. S. (2017). “Does tactile feedback enhance single-trial detection of error-related EEG potentials?,” in 2017 IEEE International Conference on Systems, Man, and Cybernetics (SMC) (Banff), 1417–1422. doi: 10.1109/SMC.2017.8122812
van Schie, H. T., Mars, R. B., Coles, M. G., and Bekkering, H. (2004). Modulation of activity in medial frontal and motor cortices during error observation. Nat. Neurosci. 7, 549–554. doi: 10.1038/nn1239
Yao, L., Jiang, N., Mrachacz-Kersting, N., Zhu, X., Farina, D., and Wang, Y. (2022). Performance variation of a somatosensory BCI based on imagined sensation: a large population study. IEEE Trans. Neural Syst. Rehabil. Eng. 30, 2486–2493. doi: 10.1109/TNSRE.2022.3198970
Keywords: human-robot interaction, implicit error detection, error-related potentials, orthosis, EEG, tactile-based ErrP detection
Citation: Kim SK and Kirchner EA (2023) Detection of tactile-based error-related potentials (ErrPs) in human-robot interaction. Front. Neurorobot. 17:1297990. doi: 10.3389/fnbot.2023.1297990
Received: 20 September 2023; Accepted: 22 November 2023;
Published: 12 December 2023.
Edited by:
Guang Chen, Tongji University, ChinaReviewed by:
Wenbin Zhang, Hohai University, ChinaKishor Lakshminarayanan, Vellore Institute of Technology (VIT), India
Copyright © 2023 Kim and Kirchner. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Elsa Andrea Kirchner, ZWxzYS5raXJjaG5lciYjeDAwMDQwO3VuaS1kdWUuZGU=