Skip to main content

ORIGINAL RESEARCH article

Front. Hum. Neurosci., 15 May 2023
Sec. Brain-Computer Interfaces
This article is part of the Research Topic Neural Decoding for Direct Communication in Brain-Computer Interface View all 6 articles

Bridging the BCI illiteracy gap: a subject-to-subject semantic style transfer for EEG-based motor imagery classification

  • Department of Artificial Intelligence, Korea University, Seoul, Republic of Korea

Introduction: Brain-computer interfaces (BCIs) facilitate direct interaction between the human brain and computers, enabling individuals to control external devices through cognitive processes. Despite its potential, the problem of BCI illiteracy remains one of the major challenges due to inter-subject EEG variability, which hinders many users from effectively utilizing BCI systems. In this study, we propose a subject-to-subject semantic style transfer network (SSSTN) at the feature-level to address the BCI illiteracy problem in electroencephalogram (EEG)-based motor imagery (MI) classification tasks.

Methods: Our approach uses the continuous wavelet transform method to convert high-dimensional EEG data into images as input data. The SSSTN 1) trains a classifier for each subject, 2) transfers the distribution of class discrimination styles from the source subject (the best-performing subject for the classifier, i.e., BCI expert) to each subject of the target domain (the remaining subjects except the source subject, specifically BCI illiterates) through the proposed style loss, and applies a modified content loss to preserve the class-relevant semantic information of the target domain, and 3) finally merges the classifier predictions of both source and target subject using an ensemble technique.

Results and discussion: We evaluate the proposed method on the BCI Competition IV-2a and IV-2b datasets and demonstrate improved classification performance over existing methods, especially for BCI illiterate users. The ablation experiments and t-SNE visualizations further highlight the effectiveness of the proposed method in achieving meaningful feature-level semantic style transfer.

1. Introduction

Brain-computer interface (BCI) is a core technology that establishes a direct communication pathway between the human brain and an external device (Nicolas-Alonso and Gomez-Gil, 2012; Chaudhary et al., 2016; Jeong et al., 2022). Electroencephalography (EEG) has been actively used for current BCI systems in order to record the brain signals, due to its non-invasive, simple operation, relatively low-cost, and high temporal resolution (Lotte et al., 2007; Abiri et al., 2019). Over the last few decades, several neurophysiological studies (Decety, 1996; Munzert et al., 2009; Leeuwis et al., 2021) have shown that during motor imagery (MI), there are increased connections between sensorimotor areas in the human brain. In other words, MI can activate similar neural activities to those involved in actual motor movement (Case et al., 2015), and therefore it is feasible to identify the MI intention of the user by examining the EEG signals that exhibit unique patterns for each movement type.

In the MI-BCI paradigm, the user is instructed to imagine performing a specific body movement, such as the left hand or right hand, without actually executing the movement. During the MI period, the brain produces characteristic patterns of event-related desynchronization (ERD) and event-related synchronization (ERS), which can be observed in specific EEG frequency bands (Neuper et al., 2006; Pfurtscheller et al., 2006). Specifically, it is well-known that different ERD/ERS patterns are associated with different types of MI tasks, e.g., mu and beta frequency bands are reactive to imagined hand movements (Jeon et al., 2011). In this respect, EEG-based MI-BCI systems can be developed by using the ERD/ERS patterns to translate the imagined movement for recognizing the user's intention. As a result, MI-BCI has become one of the most promising sub-fields in BCI research and extensively studied over the past several years (Hamedi et al., 2016; Deng et al., 2021; Jeong et al., 2022).

Despite the great potential of MI-BCI, there is one major challenge to be addressed before using it in a real-world BCI application. In the MI-BCI research, it is widely recognized that there are substantial individual differences in the capability to perform a given MI task (Ren et al., 2020). Specifically, individual differences refer to the natural variation in personality traits, cognitive abilities, motivation, and other innate characteristics among individuals. Several studies (Allison et al., 2010; Volosyak et al., 2011; Wriessnegger et al., 2020) have found that these individual differences can influence MI-BCI performance and level of success in performing MI tasks. For instance, individuals who have low levels of motivation or who have difficulty maintaining attention may struggle with MI-BCI training and may not achieve the desired performance (Thompson, 2019). In the context of MI-BCI research, these aforementioned individuals are considered to have BCI illiteracy, which refers to the lack of knowledge and proficiency in using a BCI system within a standard training period (Lee et al., 2019; Volosyak et al., 2020). Based on previous studies (Becker et al., 2022; Tibrewal et al., 2022), around 15–30% of BCI users fail to produce the desired EEG patterns in order to control a BCI device accurately. Moreover, few studies (Zhang et al., 2020; Wang et al., 2021) have reported that BCI illiteracy subjects typically achieve classification accuracy below 70% and decrease the average performance of all subjects. This inability derived from BCI illiteracy can limit the applicability of the BCI system to be used by individuals who do not have specialized BCI knowledge or training. Thus, understanding and addressing BCI illiteracy is one of the major challenges for advancing the development of BCI technology (Ahn and Jun, 2015; Thompson, 2019).

In order to tackle the BCI illiteracy problem, various machine learning (ML)-based approaches (Vidaurre and Blankertz, 2010; Vidaurre et al., 2011; Tao et al., 2022) have been developed. One such approach is co-adaptive learning (Vidaurre and Blankertz, 2010), which uses the ML algorithm, i.e., linear discriminant analysis (LDA), to help users achieve closed-loop feedback. During the feedback process, both the user and the ML algorithm adapt to each other, thereby improving the overall performance of the BCI system. Another ML-based approach is based on multi-kernel learning (Tao et al., 2022) that aims to make the distribution of features closer to each other, while maximizing the divisibility of categories. Despite the reasonable performance achieved by ML-based approaches, they often rely on heuristic statistical reasoning and assumptions such as linear separability (Medin and Schwanenflugel, 1981) and same feature space assumption (Girolami, 2002). Moreover, these ML-based approaches may suffer from high computation costs when dealing with high dimensional data due to the curse of dimensionality (Bach, 2017). The EEG data are considered as inherently high-dimensional because they are typically collected using multiple electrodes with various spatial and temporal features. Thus, ML-based approaches are even more susceptible to these high-dimensionality issues.

Recently, a number of deep learning (DL)-based approaches (Tan et al., 2018; Gao et al., 2020; Zhao et al., 2020; Jeon et al., 2021; Sun et al., 2022) have been applied to the BCI illiteracy problem and achieved better performance compared to conventional ML-based approaches. In particular, deep transfer learning based on domain adaptation (Tan et al., 2018; Zhao et al., 2020; Jeon et al., 2021) has gained attention due to its capability to extract common feature representations. Specifically, Zhao et al. (2020) proposed a deep representation-based domain adaptation (DRDA) method that learns significant domain invariant features from multiple subjects (source domain) and uses that information to improve the performance on a single subject (target domain). Jeon et al. (2021) further developed it by utilizing mutual information to estimate the relevance of features and then extracting subject-invariant feature representations that are relevant to the classification task.

Even though these above-mentioned domain adaption methods have attained promising results, there are some disadvantages in their real-world BCI applications. First of all, these methods require a significant amount of labeled data in order to achieve good performance because they need to extract the common domain-invariant representations from multiple subjects (Sun et al., 2022). This can be problematic in BCI scenarios where large datasets do not exist or labeled data from multiple subjects are expensive to obtain in terms of time and cost. Second, they may suffer from negative transfer due to the large distributional discrepancy when extracting the common representation from multiple subjects (Jiménez-Guarneros and Gómez-Gil, 2021). Specifically, the negative transfer refers to a phenomenon that occurs when transferred knowledge or information from the source domain hinders the performance of the classifier on the target domain instead of improving it (Cao et al., 2018). Third, in real-world BCI applications, the EEG data from different domains vary substantially due to intra- and inter-subject variability (Saha and Baumert, 2020) that involves physiological noise level (Sanei and Chambers, 2013), signal quality (Ball et al., 2009), or emotions (Zhao et al., 2021). Therefore, these complex domain shifts derived from intra- and inter-subject variability can make it very difficult to find common domain-invariant feature representations across multiple subjects (Saha and Baumert, 2020). Hence, conventional domain adaption methods may fail to extract useful information from the source domain because of the large discrepancy within multiple sources. Besides, the EEG data of most BCI datasets are recorded over multiple sessions. So this session-to-session variability makes it even more difficult to construct a robust classifier across multiple subjects over multiple sessions.

Style transfer is also one approach for transferring between these domains. Traditional style transfer methods directly processed the image's content by applying various filters, machine learning, or probabilistic models (Ma et al., 2020). Gatys et al. (2015b, 2016a) proposed neural style transfer using convolutional neural networks, causing a paradigmatic revolution in the field, and this approach to neural style transfer became mainstream. Moreover, feedforward neural net-based approaches or iterative optimization methods (Isola et al., 2017; Yu et al., 2019) were proposed. These methods using style loss and content loss, effectively transfer the texture of a style image to a content image. However, they still have the limitation that they are mainly effective for the visual style and content. This is an important restriction because the nature of EEG data is such that differences are not visible.

To address these inter-subject variability issues, Sun et al. (2022) proposed another style transfer approach for the EEG classification tasks. Specifically, they introduced a subject transfer neural network (STNN) that directly transforms the data distribution from BCI-illiterate subject into BCI-expert, known as “golden subject”, by utilizing a subject-to-subject transfer approach. More precisely, the STNN model aims to learn a one-to-one style transfer between the golden subject (source domain) and the BCI-illiterate subject (target domain) without using any domain discriminator nor explicit regularizers. During training, the STNN uses only classification loss and perceptual loss, which compares the feature differences between source and translated target domains in order to facilitate the style transfer process. However, there are some limitations and drawbacks to the STNN. The main problem is that they only focus on transferring the style of the source domain and do not manage to preserve any content information from the target domain. Therefore, they may have difficulties in generating diverse and plausible data distributions containing desired content information from the target domain.

In this study, we first transform the high-dimensional EEG data into an image by using a continuous wavelet transform technique, and then we transfer the class-discriminative style of the source domain (BCI expert) to the target domain (BCI illiterates). In particular, we introduce a modified content loss to preserve the sementic content information of the target domain even after the style transfer process. Because the translated image of the target domain obtains class-discriminative characteristics from the source domain, we can improve the classification performance by leveraging the source classifier trained on the source domain data. Hence, our proposed method can alleviate the BCI illiteracy problem.

The main contributions of our proposed method are as follows:

1. To tackle the BCI illiteracy problem, we propose a subject-to-subject semantic style transfer network (SSSTN) that allows for effective and seamless interaction between users and the BCI system.

2. Unlike previous approaches that mainly focus on visible information, we introduce the semantic-aware style transfer loss function consisting of (i) a content loss to preserve the semantic identity information of the target domain, (ii) a style loss to transfer the semantic texture information of the source domain, and (iii) a semantic loss to further improve the classification performance.

3. By utilizing a subject-to-subject transfer strategy that performs a one-to-one mapping from the target domain to the source domain, our proposed method demonstrates high data efficiency, requiring only labeled data from a single subject in each domain dataset.

4. Our proposed method facilitates the construction of an ensemble classifier by integrating two subject-dependent classifiers—one from the target domain and the other from the source domain. This ensemble approach enables the fusion of diverse feature representations, resulting in a more robust classification model.

Our proposed method is evaluated on the BCI Competition IV-2a and IV-2b datasets, with experimental results showing that SSSTN outperforms other competing approaches in mean accuracy, particularly for BCI illiterates. Additionally, we performed an ablation study to assess the efficacy of each component within our proposed method. We also conducted ablation studies and visualization using t-SNE (Van der Maaten and Hinton, 2008) to evaluate the efficiency of each component within the proposed method. The experimental results demonstrated that the proposed method achieved meaningful feature-level semantic style transfer results.

2. Materials and methods

2.1. Dataset and preprocessing

2.1.1. Dataset description

In this work, we used two publicly available benchmark datasets, namely BCI Competition IV-2a (Brunner et al., 2008) and 2b (Leeb et al., 2008), to evaluate our proposed method. The BCI Competition IV-2a dataset consists of EEG recordings from 9 healthy subjects, each performing four-class motor imagery tasks involving left hand, right hand, both feet, and tongue movements. For each subject, there were two separate EEG recording sessions that took place on different days. In each session, there were 72 EEG trials for each motor imagery task, resulting in a total of 288 EEG trials per subject. Each EEG trial lasted for 6 s, starting with a fixation cross, a cue followed by the motor imagery task. The EEG signals were recorded using 22 Ag/AgCl electrodes on the scalp in the 10-20 system (Homan et al., 1987), and the EEG data was sampled at 250 Hz. The BCI Competition IV-2b dataset comprises MI task experiments for two classes (right- and left-hand movements). The competition's objective was to classify MI tasks using EEG signals recorded from C3, CZ, and C4 channels. A total of 9 subjects participated in the experiment, with five sessions recorded for each subject. In each session, we used 60 trials for each motor imagery task, resulting in a total of 120 EEG trials per subject. Otherwise, the sampling rate and recording method are the same as for the BCI Competition IV-2a.

2.1.2. Data preprocessing

For each trial, we obtained 4.5 s of EEG data by including 0.5 s prior to the start cue and 4.0 s after it (Schirrmeister et al., 2017). We then applied a bandpass filter ranging from 0.5 to 40 Hz and utilized exponential moving standardization to preprocess the raw EEG data. As a result, the preprocessed EEG signals consisted of a total of 1,125 time points (4.5 s × 250 Hz sampling rate) from 22 electrodes for the BCI Competition IV-2a dataset.

In this study, we decided to use a wavelet transform (Rioul and Duhamel, 1992) to transform the EEG data into images in order to use it as input data for our proposed SSSTN. Note that the use of wavelet transform is advantageous in our study for several reasons. Firstly, it allows for the effective representation of EEG signals in a multi-scale and multi-resolution manner, and this richer representation helps the algorithm to work with more informative data for the style transfer process. Secondly, transformation to images allows a convolutional neural network (CNN) to capture spatial-spectral-temporal features from multiple EEG representations from different electrodes, making it easier to perform semantic style transfer. Thus, after data preprocessing, we employed the continuous wavelet transform (CWT) technique (Rioul and Duhamel, 1992) to translate a one-dimensional signal from the temporal domain to a two-dimensional signal in the temporal-spectral domain. Specifically, the application of CWT to the EEG signal yielded a two-dimensional matrix, commonly referred to as a scalogram. Note that this scalogram matrix comprises the absolute values of the wavelet coefficients at different wavelet scales for the given EEG signals. Hence, this matrix can provide a detailed representation of the EEG signals in the temporal-spectral domain. By treating the scalogram matrix as an image, the 2D-CNN can be employed to extract spatial-spectral-temporal features and classify the MI tasks. Mathematically, the EEG signal x(t) can be transformed by CWT operation such as follow:

Xw(a,b)=1|a|1/2-x(t) ψ (t-ba)dt,    (1)

where a ∈ ℝ+ and b ∈ ℝ denote the scaling parameter for the spectral domain and the shifting parameter for the temporal domain, respectively (Rioul and Duhamel, 1992). Note that ψ(t) represents the Morlet wavelet (Grossmann and Morlet, 1984), which is one of the most widely used wavelet base functions because it provides good resolution in both temporal and spectral domains. To be more specific, the Morlet wavelet ψ(t) is expressed as follow:

 ψ (t)=exp (-β2t22)cos(πt),    (2)

where β represents the admissibility condition that determines the balance between the spectral resolution and the temporal resolution of the wavelet Morlet. The images of each subject obtained through the CWT process are used as input data for the generator in the style transfer process.

2.2. Methods

Here, we propose a Subject-to-subject Semantic Style Transfer Network (SSSTN) for tackling the BCI illiteracy problem in EEG-based MI classification tasks. Unlike conventional style transfer networks, our SSSTN not only performs style transfer but also transfers and generates class-discriminative data while taking into account the semantics of the input data. Since we transform an EEG signal into an image by using the CWT method, the stylistic differences between the source and target data may not be easily distinguishable as to be visible. Consequently, we think that existing style transfer approaches are not suitable for this task. To address this issue, we introduce a modified style loss and content loss that are tailored to this specific task of transferring semantic style and semantic content information in an effective manner. During the transformation of the target subject (especially, BCI illiterates) data into the source subject (BCI expert) data, our SSSTN effectively transfers and preserves the underlying semantic style and content information at feature-level in order to ensure that the source subject classifier is able to accurately classify the transformed data. The overall flow of our proposed method is depicted in Figure 1. The training process of SSSTN can be divided into three phases: (1) Pretraining, (2) Style Transfer, and (3) Prediction and Ensemble. Detailed explanations of each phase are presented below.

FIGURE 1
www.frontiersin.org

Figure 1. Overview of the proposed SSSTN framework. The SSSTN consists of three phases: (1) pretraining, (2) style transfer, (3) prediction and ensemble. The input variables, xT and xS, represent the target and source subject data, respectively, while xS signifies the target data transformed by the generator G. During the pretraining phase, the classifiers CT and CS are pretrained using xT and xS, respectively. In the style transfer phase, only G is trained. Content loss is computed using features hTl and hSl from the l-th convolutional layer after passing xT and xS through CT and CS, respectively. Style loss is calculated between hSl and hSl after passing xS and xS through CS, respectively. Semantic loss is computed between the predicted label ŷS, which is obtained by passing xS through CS, and the ground-truth label yT. In the last phase, the final prediction is generated using a soft voting ensemble of ŷS and ŷT, where ŷT is the predicted label for xT obtained from CT.

2.2.1. Pretraining

In the pretraining phase, we train a classifier for each subject using image data, i.e., obtained through the CWT method, and then perform a classification task under the subject-dependent scenario (subject-specific training). The primary objective of the pretraining process is to train the classifier, which can accurately classify input image data into one of the four distinct MI classes. The architecture of the classifiers utilized in this process is represented in Figure 2. The numerical values in the convolutional layer box represent the kernel size, number of output channels, stride, and padding, respectively. Note that the classifier architecture remains consistent across all subjects, comprising two convolutional layers and one dense layer. The convolutional layers are composed of batch normalization, LeakyReLU activation, max pooling operation, and dropout (Zhang et al., 2021). After each convolutional layer, a squeeze-and-excitation (SE) (Hu et al., 2018) module is employed. The SE module is responsible for capturing the important channel-wise relationships, which is crucial in EEG signals where interdependencies between channels are important, as each channel represents an electrode. After each SE operation, the feature hl is used to calculate the loss in the subsequent step of style transfer, where lL = {1, …, L} denotes the l-th convolutional layer of the classifier C. Our proposed model utilizes a total of two convolutional layers, denoted as L = 2 in our case. After the two convolutional layers and the SE modules, the dense layer produces the prediction. The dense layer is a fully-connected layer that flattens the input features and then performs fully-connected operations, returning them as predictions for the four classes. Finally, classifiers are trained with the following classification loss:

Lcls=k=1Ky(k) logy^(k),    (3)

where K denotes the total number of classes. In our case, each classifier performs a 4-class classification task, which sets the value of K to 4. Additionally, during this stage, we identify the subject with the highest classification performance to serve as the source subject for style transfer, while designating all other subjects as target subjects.

FIGURE 2
www.frontiersin.org

Figure 2. The architecture of the classifier C in the proposed framework. The dimensions of the input image are denoted by C, F, and T, corresponding to the number of channels, frequencies, and time points, respectively. The kernel size in each convolutional layer is followed by a number indicating the number of output channels. The variables s and p denote the stride and padding used in the convolutional layers, respectively. The activation function used in this architecture is LReLU, which stands for Leaky ReLU. The output features of the first and second convolutional blocks are represented as h1 and h2, respectively, and are utilized to calculate the loss. Finally, the number written in the dense layer indicates the dimension of the output feature, resulting in a prediction of size 4 at the end.

2.2.2. Style transfer

It is worth noting that the objective of the style transfer phase in our study differs from that of general style transfer approaches. In general style transfer, the style is typically represented by the texture or color of the image, while the content refers to the object depicted in the image. However, the semantic style and content in our study are not defined by visible textures, colors, objects, or general scenery. As a result, they may not be immediately apparent or easily recognizable in the image data. As previously mentioned, there exist substantial inter-individual differences in EEG characteristics within MI-BCI (Ren et al., 2020). Therefore, even if multiple subjects are trained with the same classifier structure, the features of each subject are represented in distinct feature spaces. Among these feature spaces, the feature space of the source subject with the highest classification performance, which we refer to as BCI experts, contains source features that are well-classified by the source classifier. It is reasonable to assume that this feature space is where the source classifier performs effectively. Thus, if the generator can accurately map the target features to the source feature space while retaining class-relevant feature representation, the resulting transformed features can be expected to be effectively classified by the source classifier. In this context, the class-relevant feature representation of the target feature is considered as the content that should be preserved during the transfer process, and the feature space where the source feature is located is treated as the style that should be applied to the target data.

In the style transfer phase, the generator G is responsible for subject-to-subject style transfer by transforming the target data from each target subject to align with the style of a single source data (in our specific case, subject 3 was selected as the source). The primary objective of G is to transform the input image in such a way that the resulting target data reflects the style of the source data at feature-level, while concurrently preserving the content of the target data. This is achieved by effectively mapping the target features to the source feature space, enabling the transformed features to be accurately classified by the source classifier. The architecture of G follows an encoder-decoder structure as illustrated in Figure 3. The encoder module consists of three convolutional layers, each with batch normalization, LeakyReLU activation, and dropout. The decoder module, on the other hand, consists of three transposed convolutional layers. Notably, the first and second transposed convolutional layers are followed by a self-attention layer (Vaswani et al., 2017; Zhang et al., 2019). The self-attention layer is incorporated into the generator to enhance its capability to selectively emphasize relevant features in the input data that are crucial for precise classification during the transformation of target data into source data (Sun et al., 2022). After each self-attention layer, batch normalization, LeakyReLU, and dropout are applied to the encoder. The output image generated by G is of the same size as the input image. It is noteworthy that during the entire style transfer process, the pretrained source classifier CS and target classifier CT remain fixed, while G is the only network being trained. The entire loss function of G consists of style loss, content loss, and semantic loss, which are combined to guide the network toward the desired outcome.

FIGURE 3
www.frontiersin.org

Figure 3. The architecture of the classifier G in the proposed framework is shown, with deconv representing the transposed convolutional layer. All other notations used are consistent with those shown in Figure 2.

The features hSl and hSl are obtained by passing the transformed and source data, xS and xS, respectively, through the l-th convolutional layer of CS. The objective of the style loss is to align the feature distributions of the transformed target features hSl and the style features hSl in the feature space. This is achieved by measuring the discrepancy between the Gram matrices of hSl and hSl (Gatys et al., 2015b, 2016a). By minimizing the style loss, the distribution of hSl becomes more similar to that of hSl. The style loss can be mathematically defined as follows:

Lstyle=i=1NKL(σ (Gr (hSl))  σ(Gr (hSl))),                                   Grijl=m=1Mhimlhjml,    (4)

where KL and σ represent the KL divergence and softmax activation function, respectively, while Gr stands for the Gram matrix and N represents the sample size. The Grijl is the inner product between vectorized feature maps i and j in layer l, and the M indicates the total number of spatial locations in the feature map (Gatys et al., 2015a, 2016b). In order to achieve further semantic style consistency between target and source data, we employed the Gram matrix-based style loss (Gatys et al., 2015a) in conjunction with the Kullback-Leibler divergence (KL divergence) (Kullback and Leibler, 1951). The application of Gram matrices has become a widely adopted technique for matching second-order statistics between different feature activations in various studies (Gatys et al., 2015a, 2016b; Huang and Belongie, 2017). Building upon previous studies using Gram matrices, we utilize the KL divergence with a softmax function applied to the Gram matrix representations of hSl and hSl as the style loss. Of note, the KL divergence represents the distance between two distributions. By minimizing the style loss based on KL divergence between the Gram matrix representations of hSl and hSl, the generator G can bring the distribution of hSl closer to that of hSl, both of which are based on the feature space of CS. Consequently, our proposed approach results in xS being better classified by CS, thereby achieving improved style transfer performance.

Our proposed style loss serves as a regularization technique for the target data (xT) by guiding G to transform it in a direction that incorporates more of the semantic style of the source data (xS) (Huang et al., 2022). However, unilaterally transferring only style information can result in negative transfers. To address this issue, we introduce a content loss term that enforces the preservation of the semantic content in xT before and after the transformation. The mathematical formulation of our content loss is provided below:

Lcont=1Nl=1L(hTlhSl)2,    (5)

where N represents the sample size again and the features hTl are obtained from the l-th convolutional layer of the target classifier CT. The feature representation hTl contains class-relevant information of the target image xT, and thus transferring it to xS ensures that G preserves the semantic content of xT during the transformation.

The ultimate objective of SSSTN is to ensure that the generated image xS is correctly classified by CS. To achieve this, we use semantic loss, which is designed to minimize the classification error of xS in the MI classification task. Specifically, we compute the cross-entropy loss between the prediction ŷS and true label yT of xT. The resulting semantic loss is shown below:

Lsem=-k=1KyT(k)logŷS(k),    (6)

where K represents the total number of classes, which is 4 in our specific case. The total loss is defined as follows:

Ltotal=αLstyle+βLcont+γLsem,    (7)

where α, β, and γ are weighting factors for style, content, and semantic loss, respectively. Recall that our style loss and content loss do not directly enforce xS to match xS, but instead leverage CS and CT to measure the similarity at feature-level between the two. The goal of the style loss and content loss is to ensure that hSl is mapped to a feature space that is similar to hSl, while preserving the class-relevant representation of hTi. Therefore, minimizing this total combined loss leads to xS having high classification performance on CS.

2.2.3. Prediction and ensemble

After completing the style transfer phase, two predictions for xT can be obtained. The first prediction, denoted as ŷS, is derived from CS for xS that is generated by G. The second prediction, denoted as ŷT, is derived from CT for xT. Next, we employ a soft voting-based ensemble to obtain the final prediction by combining the two predictions obtained for xT. A soft voting-based ensemble is a well-established method of combining multiple classifiers by taking into account both individual decisions and probability values to assign data to a specific class (Sherazi et al., 2021). Soft voting has been demonstrated to yield better performance and results than hard voting since it utilizes an average of probabilities (Saqlain et al., 2019). As such, soft voting has been widely used in many BCI studies due to its potential to improve the classification performance (Mussabayeva et al., 2021; Tasnim et al., 2022; Mehtiyev et al., 2023). In addition, soft voting-based ensembles are effective in compensating for the weaknesses of individual classifiers and can achieve even better performance when combining classifiers trained on different features. The increased diversity in feature space is the key factor behind the performance improvement of ensemble classifiers, as it enhances their robustness to both inter- and intra-subject variability (Corsi et al., 2022). Due to the fact that CT and CS were trained on distinct datasets, namely xT and xS, respectively, their feature spaces are different and complementary. Consequently, by leveraging both classifiers through an ensemble method, we can effectively utilize multiple feature spaces to enhance the classification performance. Mathematically, the final prediction ŷTvote is obtained by employing the soft voting-based ensemble method, as depicted below.

ŷTvote=ŷS+ŷT,    (8)

where ŷS and ŷT represent the predicted labels obtained from subject-specific classifiers CS and CT, respectively. The final prediction is determined by selecting the class with the highest logit value in ŷTvote. To sum up, our proposed style transfer approach leverages subject-specific classifiers to facilitate the use of the ensemble method, resulting in enhanced performance and classification accuracy.

3. Results

3.1. Competing methods

We evaluated the proposed SSSTN on accuracy by comparing it to the following competing methods on the BCI Competition IV-2a and the BCI Competition IV-2b datasets. For a more accurate and intuitive comparison, we categorized the competing methods into signal-based ([1]~[4]) and image-based ([5]~[8]), as shown in the Table 1.

[1] EEGNet (Lawhern et al., 2018) is tailored to exploit the unique spatial and temporal structure of EEG signals, combining temporal and spatial convolutions to capture relevant information. The model is lightweight and adaptive and has shown good performance for low computational cost in various BCI studies.

[2] DeepConvNet (Schirrmeister et al., 2017) focuses on exploiting the inherent spatial and temporal structure of the EEG data by employing specialized convolutional layers that capture the intricate patterns within the signals. Furthermore, the authors introduce a visualization technique that allows for the inspection and interpretation of the learned representations in the context of the underlying neural processes.

[3] DRDA (Zhao et al., 2020) proposes a deep representation-based domain adaptation approach to address non-stationary EEG classification challenges. The methodology focuses on learning transferable and discriminative representations by bridging the gap between source and target domains, while simultaneously preserving the class-discriminative information.

[4] MI (Jeon et al., 2021) presents a method that leverages mutual information for deep representation learning in brain-computer interface (BCI) applications. The proposed approach aims to achieve subject-invariant and class-relevant representations by optimizing mutual information between the learned features and class labels, while minimizing subject-related information. This method facilitates enhanced generalization across subjects and increased classification performance in BCI tasks.

[5] CWT-CNN (Mahamune and Laskar, 2021) employs two-dimensional images, generated through continuous wavelet transform (CWT) filter bank decomposition of pre-processed EEG data using the multi-class common spatial pattern (CSP) technique. These 2D images serve as the basis for training a convolutional neural network (CNN), enhancing classification accuracy.

[6] SE-CNN inspired by Zhang et al. (2021) is designed as the backbone classifier for our SSSTN in this study. It consists of two convolutional layers and SE modules, as described above. Therefore, the performance of SE-CNN corresponds to the version of the pretraining phase of SSSTN.

[7] STNN (Sun et al., 2022) proposes a style transfer approach that trains a golden subject-specific classifier and transforms other subjects into generators to fit that classifier. The generator is trained by BCE loss and perceptual loss. It is most similar to the proposed method, and the difference between the two methods will be discussed later in the ablation study.

[8] SSSTN (ours) introduces a novel subject-to-subject semantic style transfer network (SSSTN) designed to address the BCI illiteracy problem. Our method incorporates a semantic-aware style transfer loss function, which consists of content, style, and semantic losses to preserve and transfer essential information while improving classification performance. SSSTN consists of three main phases: pretraining, style transfer, prediction and ensemble.

TABLE 1
www.frontiersin.org

Table 1. Performance comparison of proposed and competing methods for subject-dependent scenario on the BCI Competition IV-2a dataset.

3.2. Experimental settings

Regarding dataset splitting, we used the first of two sessions of the BCI Competition IV-2a dataset as the training set and the second as the test set during the training of SSSTN and all competing methods. For the BCI Competition IV-2b dataset, the first three sessions were used for training, while the remaining two sessions served as test data. During the initial model training, we allocated 10% of the training set for use as a validation set. We then proceeded to tune the hyperparameter configuration based on the model's performance on this validation set. Due to the small size of the dataset, we incorporated the validation set back into the training set for the final model training. This allowed us to leverage the full dataset for training, thereby potentially improving our model's performance.

As previously mentioned, we designated subject 3 as the source subject and assigned the remaining subjects as target subjects for the BCI Competition IV-2a dataset. Similarly, we selected subject 5 as the source subject and allocated the remaining subjects as target subjects for the BCI Competition IV-2b dataset. We evaluated the performance of each model with classification accuracy. All competing methods were trained using the number of epochs and hyperparameters mentioned in each paper. The training process of the proposed SSSTN consists of pretraining and style transfer (since the model is not trained in the prediction and ensemble process). In the pretraining phase, each classifier was trained with a learning rate of 0.0002 for 3,000 epochs. For the style transfer phase, the generator was trained with a learning rate of 0.003 for 600 epochs for the BCI Competition IV-2a dataset. The experimental settings for the 2b dataset were consistent with those of the 2a dataset, with the sole difference being a reduced training duration, set at 200 epochs. Finally, we adopted α as 0.1, β as 1, and γ as 1 for style loss.

3.3. Performance evaluations

Table 1 presents the classification accuracy of each method under the subject-dependent scenario in the BCI Competition IV-2a dataset. The proposed SSSTN method demonstrated superior performance with the mean accuracy of 80.66% on the BCI Competition IV-2a dataset, outperforming all other competing methods. The SSSTN also exhibited the highest accuracy on all the individual subjects except subject 3. It is worth noting that since subject 3 was used as the source subject, it did not undergo any additional training beyond the pretraining phase. Among the subjects included in the study, subjects 2 and 6 were considered to be BCI illiterate as they exhibited particularly low classification accuracy in the dataset. Our proposed SSSTN method achieved significant improvements in classification accuracy on both BCI illiterate subjects 2 and 6, compared to other competing methods evaluated on the same dataset. Specifically, the SSSTN method demonstrated notable performance gains over our baseline model, i.e., SE-CNN [6], achieving improvements of 4.79 and 6.11% on subject 2 and subject 6, respectively. Moreover, when compared to the second-best performing method, i.e., MI [4] (signal-based), the SSSTN method continued to demonstrate superiority with improvements of 1.73 and 6.94% on subjects 2 and 6, respectively. Remarkably, the proposed SSSTN method demonstrated substantial improvement in performance for individuals, such as subject 5, who exhibited moderately low classification accuracy in the dataset. To be specific, the SSSTN achieved a significant enhancement of 12.77 and 8.68% in classification accuracy compared to our baseline model and the second-best performing method, respectively.

Among our competing methods, the STNN [7] was considered as the most analogous competitor to our proposed approach due to its comparable implementation of subject-to-subject feature-level style transfer based on source classifiers. The BCE loss used in the STNN can be aligned with our semantic loss in terms of conceptual similarity. However, unlike the perceptual loss function employed in the STNN that calculates the L2 loss between the source feature (hSl) and transformed target feature (hSl), our style loss function measures the difference between the probability distributions of the semantic styles of hSl and hSl by utilizing KL divergence. Furthermore, the STNN did not have the content loss designed to guarantee the retention of salient information derived from the target. These key differences in loss functions make the STNN a suitable comparison method for our proposed method when exploring the impact of different types of style loss and the absence of content loss. In comparison to the STNN, our SSSTN surpassed its performance across all subjects. Notably, our semantic style loss encourages target and source feature distribution alignment, while STNN's perceptual loss focuses on matching feature values. Experimental results convincingly demonstrated that our proposed style loss contributes to superior classification performance compared to the perceptual loss.

Overall, these findings suggest that our proposed method holds the potential for addressing the BCI illiteracy problem, especially for individuals who were previously difficult to classify MI tasks. In addition to the significant improvements observed on BCI illiterate subjects, our proposed SSSTN method achieved high levels of accuracy on top-performing subjects in the dataset. Particularly, our method exhibited an improvement in the performance of 3.55% on subject 7, when compared to our baseline model. Additionally, our method demonstrated robust performance on subjects with intermediate levels of classification accuracy. In a nutshell, we want to emphasize that our proposed method is effective across a range of BCI proficiency levels, and may have broad applicability for individuals with varying levels of BCI performance.

As illustrated in Table 2, we conducted an additional experiment utilizing the BCI Competition IV-2b dataset. It is essential to note that SE-CNN denotes our baseline network. Following SE-CNN's pretraining, subject 5 achieved the highest performance of 95.83%, serving as the source subject, while subject 2, with a performance of 67.71%, was identified as the BCI illiterate. Without the need for any additional hyperparameter tuning or alterations to network architectures, our SSSTN consistently outperformed the competing methods in terms of mean accuracy in this experiment. Remarkably, SSSTN was the sole method to achieve a classification accuracy exceeding 70% for all subjects, including subject 2, and displayed the highest accuracy for the majority of subjects. The analysis of the BCI Competition IV-2b dataset substantiates that our proposed SSSTN can be effectively applied to a diverse range of datasets, thereby demonstrating the generalizability of our approach.

TABLE 2
www.frontiersin.org

Table 2. Performance comparison of proposed and competing methods for subject-dependent scenario on the BCI Competition IV-2b dataset.

3.4. Ablation study

We performed an ablation study on the BCI Competition IV-2a to validate the effectiveness of our proposed method and its loss functions. Table 3 presents the variants of our SSSTN method utilized in this ablation study. Specifically, SSSTN-A represents the model that excludes the content loss, SSSTN-B is the model that omits the style loss, and SSSTN-C indicates the model without the semantic loss. Finally, SSSTN refers to the entire network we proposed. The remaining SSSTN and its variants all achieved higher mean accuracy than SE-CNN. These results suggest that our proposed SSSTN method with different types of losses has a positive impact on the classification performance of the model. In terms of mean accuracy, the SSSTN-A exhibited the most inferior performance, succeeded by SSSTN-B, subsequently, SSSTN-C. From these observations, we conclude that content loss, followed by style loss and semantic loss, contributes to improved classification performance. It is worth noting that our SSSTN method outperformed all other methods, displaying the highest performance improvement across nearly all subjects. Specifically, for subjects 2, 5, and 6, classified as BCI illiterate due to low classification accuracy, the SSSTN-B model exhibited the most significant performance improvement, surpassing all other models except for the complete SSSTN method. This outcome suggests that the content loss is one of the most critical factors in addressing the BCI illiteracy issue because it helps to retain the class-relevant feature representation of xT. As demonstrated in Section 3.3.1, our proposed style loss exhibited superior performance, which becomes evident when comparing the results of STNN and our SSSTN. Therefore, these findings demonstrate the effectiveness of our proposed SSSTN method in addressing the BCI illiteracy problem, which is achieved through leveraging classifier-based feature-level semantic style transfer with appropriate loss functions.

TABLE 3
www.frontiersin.org

Table 3. Ablation study results demonstrating the impact of removing specific components from the proposed SSSTN method on the BCI Competition IV-2a dataset.

In order to ascertain that our SSSTN can be adaptable to different source subjects, we conducted additional experiments on the BCI Competition IV-2a to validate this claim by adopting subject 7, which had the second-highest accuracy of 89.51% among the subjects in our baseline network (SE-CNN), as the source subject. As shown in Table 4, the experimental results demonstrated that our proposed method is not confined to a specific source subject, i.e., subject 3, but also exhibits robust performance when applied to subject 7, thereby confirming its adaptability and applicability to various source subjects. Employing subject 7 as the source subject, the SSSTN achieved a mean accuracy of 78.74%, surpassing the performance of the baseline network. However, it is worth mentioning that selecting subject 7 as the source subject led to a minor improvement in performance from the baseline compared to when subject 3 was chosen, which can be attributed to the relatively lower classification accuracy of subject 7 in relation to subject 3.

TABLE 4
www.frontiersin.org

Table 4. Additional results demonstrating the applicability of our SSSTN model to other source subjects beyond the single source subject on the BCI Competition IV-2a dataset.

4. Discussion

4.1. t-SNE visualization

To examine the impact of style transfer, we employed the t-SNE algorithm (Van der Maaten and Hinton, 2008) to visualize the feature h2 obtained from the second convolutional layer of C before and after the transformation. This experiment was carried out on a test set of Subjects 5 and 7 that were randomly selected. Figures 4A, C depict hT2 and hS2 before style transfer (ST), while Figures 4B, D show hS2 and hS2 after ST, both in two-dimensional embedding space. Each marker represents the subject (source or target) of the samples, and each color corresponds to the label of the samples. Before the transformation, distinct distributions were observed between the target and source subjects, evident for both Subjects 5 and 7 as illustrated in Figures 4A, C. However, after transformation, the target and source distributions merged into a single distribution, as depicted in Figures 4B, D. Note that the same labels are consistently clustered together, regardless of whether they belong to the source or target data. This result supports the idea that our proposed style transfer process effectively preserved the underlying class-relevant feature representation. In particular, before the transformation of Subject 5, the target data was not distinctly separated by the label as shown in Figure 4A. However, as shown in Figure 4B, after the transformation, the target data exhibited a significantly better separation by the label and were closely clustered with the source data samples that shared the same label. The obtained results suggest that the proposed SSSTN method effectively improves the discriminability of the transformed target features, leading to better classification by CS. Therefore, we verified that the proposed SSSTN method transfers the target data to a well-classified feature space of the source data, while preserving the feature class-relevant representation of the target data.

FIGURE 4
www.frontiersin.org

Figure 4. t-SNE (Van der Maaten and Hinton, 2008) visualization of the change in the h2 feature distribution of the target (S05, S07) and source (S03) before and after the style transfer. (A, B) Before/after style transfer from S05 to source. (C, D) Before/after style transfer from S07 to source.

4.2. Training loss analysis

As part of our analysis of the SSSTN network, we plotted the training loss graphs for subjects 5 and 7, which show the changes in style, content, and semantic loss over time. It is noteworthy that our selection of subjects 5 and 7 is consistent with our t-SNE visualization analysis. As depicted in Figure 5, the loss converges to a stable value after approximately 600 epochs of training. This convergence demonstrates the stability of the SSSTN during the training process and provides a profound understanding of the network's overall performance.

FIGURE 5
www.frontiersin.org

Figure 5. Training loss plot of the SSSTN for subject 5 and 7. Lstyle, Lcont, and Lsem denote style, content, and semantic loss, respectively.

4.3. Practical potential of our proposed method

In this section, we investigate the real-world application and the practical potential of our proposed method. Our proposed method is based on the assumption that we can identify a BCI expert (source subject) and BCI illiterates (target subjects) through a pretrained classification network. As mentioned in the introduction and supported by prior studies (Zhang et al., 2020; Wang et al., 2021), subjects with performance below 70% for a predefined period are classified as BCI illiterates. Our study seeks to enhance the performance of these illiterate subjects using a pretrained network. In this context, it could be feasible to distinguish between a BCI expert, who exhibits the highest classification performance, and BCI illiterates who score less than 70% based on the pretrained network. We then use a semantic style transfer process to transform the BCI illiterates' data into the expert's style, thereby improving their performance further. We believe that this approach offers potential for real-world applications in situations where an existing BCI system, specifically a pretrained classifier, is available to effectively distinguish between BCI expert and BCI illiterate subjects. Furthermore, our proposed method relies on prior knowledge, such as a pretrained network, but is not restricted to a specific network. This flexibility allows for its application to networks with varying architectures. We will continue to investigate strategies to enhance the generalizability and practicality of our method in future studies.

5. Conclusion

In this study, we proposed a subject-to-subject semantic style transfer network (SSSTN) to address the problem of BCI illiteracy in EEG-based motor imagery classification tasks. The proposed SSSTN leverages subject-specific classifier-based modified style loss and content loss to effectively transfer invisible feature-level semantic styles from source subject (BCI expert) to target subjects (specifically, BCI illiterates) while preserving their class-relevant semantic information of target subjects. Therefore, the transformed data from the target to the source retains the distribution of class-discriminative features from the source, leading to better classification performance by the source classifier. Experimental results on the BCI Competition IV-2a dataset show that our proposed method outperforms other competing methods, especially for the BCI illiterate. The ablation study and t-SNE visualization demonstrate the ability to achieve meaningful feature-level semantic style transitions by confirming the effectiveness of each component within the SSSTN. Furthermore, the ensemble approach used in this method contributes to improving classification performance by fusing different feature representations. This study paves the way for further research on subject-to-subject style transfer and BCI illiteracy mitigation.

Data availability statement

The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.

Author contributions

D-HK and D-HS designed the research and wrote the initial manuscript. D-HK designed and conducted the experiments and analyzed the results. D-HS preprocessed the data. T-EK supervised the research and revised the manuscript. All authors contributed to the article and approved the submitted version.

Funding

This work was supported by Institute of Information & Communications Technology Planning & Evaluation (IITP) grant funded by the Korea government (MSIT) (No. 2019-0-00079, Artificial Intelligence Graduate School Program (Korea University), No. 2017-0-00451, Development of BCI based Brain and Cognitive Computing Technology for Recognizing User's Intentions Using Deep Learning), and the National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIT) (No. RS202300212498).

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Abiri, R., Borhani, S., Sellers, E. W., Jiang, Y., and Zhao, X. (2019). A comprehensive review of EEG-based brain–computer interface paradigms. J. Neural Eng. 16, 011001. doi: 10.1088/1741-2552/aaf12e

PubMed Abstract | CrossRef Full Text | Google Scholar

Ahn, M., and Jun, S. C. (2015). Performance variation in motor imagery brain–computer interface: a brief review. J. Neurosci. Methods 243, 103–110. doi: 10.1016/j.jneumeth.2015.01.033

PubMed Abstract | CrossRef Full Text | Google Scholar

Allison, B., Luth, T., Valbuena, D., Teymourian, A., Volosyak, I., and Graser, A. (2010). BCI demographics: how many (and what kinds of) people can use an SSVEP BCI? IEEE Trans. Neural Syst. Rehabil. Eng. 18, 107–116. doi: 10.1109/TNSRE.2009.2039495

PubMed Abstract | CrossRef Full Text | Google Scholar

Bach, F. (2017). Breaking the curse of dimensionality with convex neural networks. J. Mach. Learn. Res. 18, 629–681. doi: 10.5555/3122009.3122028

CrossRef Full Text | Google Scholar

Ball, T., Kern, M., Mutschler, I., Aertsen, A., and Schulze-Bonhage, A. (2009). Signal quality of simultaneously recorded invasive and non-invasive EEG. Neuroimage 46, 708–716. doi: 10.1016/j.neuroimage.2009.02.028

PubMed Abstract | CrossRef Full Text | Google Scholar

Becker, S., Dhindsa, K., Mousapour, L., and Al Dabagh, Y. (2022). “BCI illiteracy: it's us, not them. Optimizing BCIs for individual brains,” in 2022 10th International Winter Conference on Brain-Computer Interface (BCI) (Gangwon-do: IEEE), 1–3.

Google Scholar

Brunner, C., Leeb, R., Müller-Putz, G., Schlögl, A., and Pfurtscheller, G. (2008). BCI Competition 2008–Graz data set A. Institute for Knowledge Discovery, Laboratory of Brain-Computer Interfaces, Graz University of Technology.

PubMed Abstract | Google Scholar

Cao, Z., Ma, L., Long, M., and Wang, J. (2018). “Partial adversarial domain adaptation,” in Proceedings of the European Conference on Computer Vision (ECCV) (Munich), 135–150.

PubMed Abstract | Google Scholar

Case, L. K., Pineda, J., and Ramachandran, V. S. (2015). Common coding and dynamic interactions between observed, imagined, and experienced motor and somatosensory activity. Neuropsychologia 79, 233–245. doi: 10.1016/j.neuropsychologia.2015.04.005

PubMed Abstract | CrossRef Full Text | Google Scholar

Chaudhary, U., Birbaumer, N., and Ramos-Murguialday, A. (2016). Brain–computer interfaces for communication and rehabilitation. Nat. Rev. Neurol. 12, 513–525. doi: 10.1038/nrneurol.2016.113

PubMed Abstract | CrossRef Full Text | Google Scholar

Corsi, M.-C., Chevallier, S., Fallani, F. D. V., and Yger, F. (2022). Functional connectivity ensemble method to enhance BCI performance (FUCONE). IEEE Trans. Biomed. Eng. 69, 2826–2838. doi: 10.1109/TBME.2022.3154885

PubMed Abstract | CrossRef Full Text | Google Scholar

Decety, J. (1996). The neurophysiological basis of motor imagery. Behav. Brain Res. 77, 45–52.

Google Scholar

Deng, X., Zhang, B., Yu, N., Liu, K., and Sun, K. (2021). Advanced TSGL-EEGNet for motor imagery EEG-based brain-computer interfaces. IEEE Access 9, 25118–25130. doi: 10.1109/ACCESS.2021.3056088

PubMed Abstract | CrossRef Full Text | Google Scholar

Gao, Z., Yuan, T., Zhou, X., Ma, C., Ma, K., and Hui, P. (2020). A deep learning method for improving the classification accuracy of SSMVEP-based BCI. IEEE Trans. Circ. Syst. II Express Briefs 67, 3447–3451. doi: 10.1109/TCSII.2020.2983389

CrossRef Full Text | Google Scholar

Gatys, L., Ecker, A. S., and Bethge, M. (2015a). “Texture synthesis using convolutional neural networks,” in Advances in Neural Information Processing Systems 28, 262–270.

Google Scholar

Gatys, L. A., Ecker, A. S., and Bethge, M. (2015b). A neural algorithm of artistic style. arXiv preprint arXiv:1508.06576. doi: 10.1167/16.12.326

CrossRef Full Text | Google Scholar

Gatys, L. A., Ecker, A. S., and Bethge, M. (2016a). “Image style transfer using convolutional neural networks,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (Las Vegas, NV), 2414–2423.

Google Scholar

Gatys, L. A., Ecker, A. S., and Bethge, M. (2016b). “Image style transfer using convolutional neural networks,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (Las Vegas, NV), 2414–2423.

Google Scholar

Girolami, M. (2002). Mercer kernel-based clustering in feature space. IEEE Trans. Neural Netw. 13, 780–784. doi: 10.1109/TNN.2002.1000150

PubMed Abstract | CrossRef Full Text | Google Scholar

Grossmann, A., and Morlet, J. (1984). Decomposition of hardy functions into square integrable wavelets of constant shape. SIAM J. Math. Anal. 15, 723–736.

Google Scholar

Hamedi, M., Salleh, S.-H., and Noor, A. M. (2016). Electroencephalographic motor imagery brain connectivity analysis for BCI: a review. Neural Comput. 28, 999–1041. doi: 10.1162/NECO_a_00838

PubMed Abstract | CrossRef Full Text | Google Scholar

Homan, R. W., Herman, J., and Purdy, P. (1987). Cerebral location of international 10–20 system electrode placement. Electroencephalogr. Clin. Neurophysiol. 66, 376–382.

Google Scholar

Hu, J., Shen, L., and Sun, G. (2018). “Squeeze-and-excitation networks,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (Salt Lake City, UT), 7132–7141.

Google Scholar

Huang, S., Xiong, H., Wang, T., Wen, B., Wang, Q., Chen, Z., et al. (2022). “Parameter-free style projection for arbitrary image style transfer,” in ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) (Singapore: IEEE), 2070–2074.

Google Scholar

Huang, X., and Belongie, S. (2017). “Arbitrary style transfer in real-time with adaptive instance normalization,” in Proceedings of the IEEE International Conference on Computer Vision (Venice), 1501–1510.

Google Scholar

Isola, P., Zhu, J.-Y., Zhou, T., and Efros, A. A. (2017). “Image-to-image translation with conditional adversarial networks,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (Honolulu, HI), 1125–1134.

PubMed Abstract | Google Scholar

Jeon, E., Ko, W., Yoon, J. S., and Suk, H.-I. (2021). Mutual information-driven subject-invariant and class-relevant deep representation learning in BCI. IEEE Trans. Neural Netw. Learn. Syst. 34, 739–749. doi: 10.1109/TNNLS.2021.3100583

PubMed Abstract | CrossRef Full Text | Google Scholar

Jeon, Y., Nam, C. S., Kim, Y.-J., and Whang, M. C. (2011). Event-related (De) synchronization (ERD/ERS) during motor imagery tasks: implications for brain–computer interfaces. Int. J. Indus. Ergon. 41, 428–436. doi: 10.1016/j.ergon.2011.03.005

CrossRef Full Text | Google Scholar

Jeong, J.-H., Cho, J.-H., Lee, Y.-E., Lee, S.-H., Shin, G.-H., Kweon, Y.-S., et al. (2022). 2020 International brain–computer interface competition: a review. Front. Hum. Neurosci. 16, 898300. doi: 10.3389/fnhum.2022.898300

PubMed Abstract | CrossRef Full Text | Google Scholar

Jiménez-Guarneros, M., and Gómez-Gil, P. (2021). A study of the effects of negative transfer on deep unsupervised domain adaptation methods. Expert Syst. Appl. 167, 114088. doi: 10.1016/j.eswa.2020.114088

CrossRef Full Text | Google Scholar

Kullback, S., and Leibler, R. A. (1951). On information and sufficiency. Ann. Math. Stat. 22, 79–86.

Google Scholar

Lawhern, V. J., Solon, A. J., Waytowich, N. R., Gordon, S. M., Hung, C. P., and Lance, B. J. (2018). EEGNet: a compact convolutional neural network for EEG-based brain–computer interfaces. J. Neural Eng. 15, 056013. doi: 10.1088/1741-2552/aace8c

PubMed Abstract | CrossRef Full Text | Google Scholar

Lee, M.-H., Kwon, O.-Y., Kim, Y.-J., Kim, H.-K., Lee, Y.-E., Williamson, J., et al. (2019). EEG dataset and OpenBMI toolbox for three BCI paradigms: an investigation into BCI illiteracy. GigaScience 8, giz002. doi: 10.1093/gigascience/giz002

PubMed Abstract | CrossRef Full Text | Google Scholar

Leeb, R., Brunner, C., Müller-Putz, G., Schlögl, A., and Pfurtscheller, G. (2008). BCI Competition 2008–Graz Data Set B. Graz University of Technology.

PubMed Abstract | Google Scholar

Leeuwis, N., Yoon, S., and Alimardani, M. (2021). Functional connectivity analysis in motor-imagery brain computer interfaces. Front. Hum. Neurosci. 15, 732946. doi: 10.3389/fnhum.2021.732946

PubMed Abstract | CrossRef Full Text | Google Scholar

Lotte, F., Congedo, M., Lécuyer, A., Lamarche, F., and Arnaldi, B. (2007). A review of classification algorithms for EEG-based brain–computer interfaces. J. Neural Eng. 4, R1. doi: 10.1088/1741-2560/4/2/R01

PubMed Abstract | CrossRef Full Text | Google Scholar

Ma, Z., Li, J., Wang, N., and Gao, X. (2020). Image style transfer with collection representation space and semantic-guided reconstruction. Neural Netw. 129, 123–137. doi: 10.1016/j.neunet.2020.05.028

PubMed Abstract | CrossRef Full Text | Google Scholar

Mahamune, R., and Laskar, S. H. (2021). Classification of the four-class motor imagery signals using continuous wavelet transform filter bank-based two-dimensional images. Int. J. Imaging Syst. Technol. 31, 2237–2248. doi: 10.1002/ima.22593

CrossRef Full Text | Google Scholar

Medin, D. L., and Schwanenflugel, P. J. (1981). Linear separability in classification learning. J. Exp. Psychol. Hum. Learn. Mem. 7, 355.

Google Scholar

Mehtiyev, A., Al-Najjar, A., Sadreazami, H., and Amini, M. (2023). “Deepensemble: a novel brain wave classification in MI-BCI using ensemble of deep learners,” in 2023 IEEE International Conference on Consumer Electronics (ICCE) (Las Vegas, NV: IEEE), 1–5.

Google Scholar

Munzert, J., Lorey, B., and Zentgraf, K. (2009). Cognitive motor processes: the role of motor imagery in the study of motor representations. Brain Res. Rev. 60, 306–326. doi: 10.1016/j.brainresrev.2008.12.024

PubMed Abstract | CrossRef Full Text | Google Scholar

Mussabayeva, A., Jamwal, P. K., and Akhtar, M. T. (2021). Ensemble voting-based multichannel EEG classification in a subject-independent P300 speller. Appl. Sci. 11, 11252. doi: 10.3390/app112311252

CrossRef Full Text | Google Scholar

Neuper, C., Wörtz, M., and Pfurtscheller, G. (2006). ERD/ERS patterns reflecting sensorimotor activation and deactivation. Prog. Brain Res. 159, 211–222. doi: 10.1016/S0079-6123(06)59014-4

PubMed Abstract | CrossRef Full Text | Google Scholar

Nicolas-Alonso, L. F., and Gomez-Gil, J. (2012). Brain computer interfaces, a review. Sensors 12, 1211–1279. doi: 10.3390/s120201211

PubMed Abstract | CrossRef Full Text | Google Scholar

Pfurtscheller, G., Brunner, C., Schlögl, A., and Da Silva, F. L. (2006). Mu rhythm (de) synchronization and EEG single-trial classification of different motor imagery tasks. Neuroimage 31, 153–159. doi: 10.1016/j.neuroimage.2005.12.003

PubMed Abstract | CrossRef Full Text | Google Scholar

Ren, S., Wang, W., Hou, Z.-G., Liang, X., Wang, J., and Shi, W. (2020). Enhanced motor imagery based brain-computer interface via FES and VR for lower limbs. IEEE Trans. Neural Syst. Rehabil. Eng. 28, 1846–1855. doi: 10.1109/TNSRE.2020.3001990

PubMed Abstract | CrossRef Full Text | Google Scholar

Rioul, O., and Duhamel, P. (1992). Fast algorithms for discrete and continuous wavelet transforms. IEEE Trans. Inform. Theory 38, 569–586.

Google Scholar

Saha, S., and Baumert, M. (2020). Intra-and inter-subject variability in EEG-based sensorimotor brain computer interface: a review. Front. Comput. Neurosci. 13, 87. doi: 10.3389/fncom.2019.00087

PubMed Abstract | CrossRef Full Text | Google Scholar

Sanei, S., and Chambers, J. A. (2013). EEG Signal Processing. John Wiley & Sons.

Google Scholar

Saqlain, M., Jargalsaikhan, B., and Lee, J. Y. (2019). A voting ensemble classifier for wafer map defect patterns identification in semiconductor manufacturing. IEEE Trans. Semiconductor Manufact. 32, 171–182. doi: 10.1109/TSM.2019.2904306

CrossRef Full Text | Google Scholar

Schirrmeister, R. T., Springenberg, J. T., Fiederer, L. D. J., Glasstetter, M., Eggensperger, K., Tangermann, M., et al. (2017). Deep learning with convolutional neural networks fo EEG decoding and visualization. Hum. Brain Mapp. 38, 5391–5420. doi: 10.1002/hbm.23730

PubMed Abstract | CrossRef Full Text | Google Scholar

Sherazi, S. W. A., Bae, J.-W., and Lee, J. Y. (2021). A soft voting ensemble classifier for early prediction and diagnosis of occurrences of major adverse cardiovascular events for stemi and nstemi during 2-year follow-up in patients with acute coronary syndrome. PLoS ONE 16, e0249338. doi: 10.1371/journal.pone.0249338

PubMed Abstract | CrossRef Full Text | Google Scholar

Sun, B., Wu, Z., Hu, Y., and Li, T. (2022). Golden subject is everyone: a subject transfer neural network for motor imagery-based brain computer interfaces. Neural Netw. 151, 111–120. doi: 10.1016/j.neunet.2022.03.025

PubMed Abstract | CrossRef Full Text | Google Scholar

Tan, C., Sun, F., and Zhang, W. (2018). “Deep transfer learning for EEG-based brain computer interface,” in 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) (Calgary, AB: IEEE), 916–920.

Google Scholar

Tao, L., Cao, T., Wang, Q., Liu, D., and Sun, J. (2022). Distribution adaptation and classification framework based on multiple kernel learning for motor imagery BCI illiteracy. Sensors 22, 6572. doi: 10.3390/s22176572

PubMed Abstract | CrossRef Full Text | Google Scholar

Tasnim, N., Halder, J., Ahmed, S., and Fattah, S. A. (2022). “An approach for analyzing cognitive behavior of autism spectrum disorder using P300 BCI data,” in 2022 IEEE Region 10 Symposium (TENSYMP) (Mumbai: IEEE), 1–6.

Google Scholar

Thompson, M. C. (2019). Critiquing the concept of BCI illiteracy. Sci. Eng. Ethics 25, 1217–1233. doi: 10.1007/s11948-018-0061-1

PubMed Abstract | CrossRef Full Text | Google Scholar

Tibrewal, N., Leeuwis, N., and Alimardani, M. (2022). Classification of motor imagery EEG using deep learning increases performance in inefficient BCI users. PLoS ONE 17, e0268880. doi: 10.1371/journal.pone.0268880

PubMed Abstract | CrossRef Full Text | Google Scholar

Van der Maaten, L., and Hinton, G. (2008). Visualizing data using t-SNE. J. Mach. Learn. Res. 9, 2579–2605.

Google Scholar

Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., et al. (2017). “Attention is all you need,” in Advances in Neural Information Processing Systems 30, 6000–6010.

Google Scholar

Vidaurre, C., and Blankertz, B. (2010). Towards a cure for BCI illiteracy. Brain Topogr. 23, 194–198. doi: 10.1007/s10548-009-0121-6

PubMed Abstract | CrossRef Full Text | Google Scholar

Vidaurre, C., Sannelli, C., Müller, K.-R., and Blankertz, B. (2011). Machine-learning-based coadaptive calibration for brain-computer interfaces. Neural Comput. 23, 791–816. doi: 10.1162/NECO_a_00089

PubMed Abstract | CrossRef Full Text | Google Scholar

Volosyak, I., Rezeika, A., Benda, M., Gembler, F., and Stawicki, P. (2020). Towards solving of the Illiteracy phenomenon for VEP-based brain-computer interfaces. Biomed. Phys. Eng. Exp. 6, 035034. doi: 10.1088/2057-1976/ab87e6

PubMed Abstract | CrossRef Full Text | Google Scholar

Volosyak, I., Valbuena, D., Luth, T., Malechka, T., and Graser, A. (2011). BCI demographics II: how many (and what kinds of) people can use a high-frequency SSVEP BCI? IEEE Trans. Neural Syst. Rehabil. Eng. 19, 232–239. doi: 10.1109/TNSRE.2011.2121919

PubMed Abstract | CrossRef Full Text | Google Scholar

Wang, T., Du, S., and Dong, E. (2021). A novel method to reduce the motor imagery BCI illiteracy. Med. Biol. Eng. Comput. 59, 2205–2217. doi: 10.1007/s11517-021-02449-0

PubMed Abstract | CrossRef Full Text | Google Scholar

Wriessnegger, S. C., Müller-Putz, G. R., Brunner, C., and Sburlea, A. I. (2020). Inter-and intra-individual variability in brain oscillations during sports motor imagery. Front. Hum. Neurosci. 14, 576241. doi: 10.3389/fnhum.2020.576241

PubMed Abstract | CrossRef Full Text | Google Scholar

Yu, J., Zhu, C., Zhang, J., Huang, Q., and Tao, D. (2019). Spatial pyramid-enhanced NetVLAD with weighted triplet loss for place recognition. IEEE Trans. Neural Netw. Learn. Syst. 31, 661–674. doi: 10.1109/TNNLS.2019.2908982

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhang, H., Goodfellow, I., Metaxas, D., and Odena, A. (2019). “Self-attention generative adversarial networks,” in International Conference on Machine Learning (Long Beach, CA: PMLR), 7354–7363.

Google Scholar

Zhang, H., Zhao, X., Wu, Z., Sun, B., and Li, T. (2021). Motor imagery recognition with automatic EEG channel selection and deep learning. J. Neural Eng. 18, 016004. doi: 10.1088/1741-2552/abca16

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhang, X., Guo, Y., Gao, B., and Long, J. (2020). Alpha frequency intervention by electrical stimulation to improve performance in mu-based BCI. IEEE Trans. Neural Syst. Rehabil. Eng. 28, 1262–1270. doi: 10.1109/TNSRE.2020.2987529

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhao, H., Zheng, Q., Ma, K., Li, H., and Zheng, Y. (2020). Deep representation-based domain adaptation for nonstationary EEG classification. IEEE Trans. Neural Netw. Learn. Syst. 32, 535–545. doi: 10.1109/TNNLS.2020.3010780

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhao, L.-M., Yan, X., and Lu, B.-L. (2021). “Plug-and-play domain adaptation for cross-subject EEG-based emotion recognition,” in Proceedings of the AAAI Conference on Artificial Intelligence, 863–870.

Google Scholar

Keywords: brain-computer interface, electroencephalogram, motor imagery, BCI illiteracy, style transfer, convolutional neural network

Citation: Kim D-H, Shin D-H and Kam T-E (2023) Bridging the BCI illiteracy gap: a subject-to-subject semantic style transfer for EEG-based motor imagery classification. Front. Hum. Neurosci. 17:1194751. doi: 10.3389/fnhum.2023.1194751

Received: 27 March 2023; Accepted: 25 April 2023;
Published: 15 May 2023.

Edited by:

Jiahui Pan, South China Normal University, China

Reviewed by:

Fangzhou Xu, Qilu University of Technology, China
Yong Peng, Hangzhou Dianzi University, China
Xin Deng, Chongqing University of Posts and Telecommunications, China

Copyright © 2023 Kim, Shin and Kam. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Tae-Eui Kam, kamte@korea.ac.kr

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.