Loading [MathJax]/jax/element/mml/optable/GreekAndCoptic.js

ORIGINAL RESEARCH article

Front. Neuroinform., 09 April 2025

Volume 19 - 2025 | https://doi.org/10.3389/fninf.2025.1559335

This article is part of the Research TopicAdvanced EEG Analysis Techniques for Neurological DisordersView all 8 articles

Recognition of MI-EEG signals using extended-LSR-based inductive transfer learning

Zhibin Jiang,Zhibin Jiang1,2Keli Hu,Keli Hu1,3Jia QuJia Qu4Zekang Bian,Zekang Bian5,6Donghua Yu,Donghua Yu1,2Jie Zhou,
Jie Zhou1,2*
  • 1Department of Computer Science and Engineering, Shaoxing University, Shaoxing, China
  • 2Institute of Artificial Intelligence, Shaoxing University, Shaoxing, China
  • 3Information Technology R&D Innovation Center of Peking University, Shaoxing, China
  • 4Department of Computer Science and Artificial Intelligence, Changzhou University, Changzhou, China
  • 5Department of AI & Computer Science, Jiangnan University, Wuxi, China
  • 6Department of Taihu Jiangsu Key Construction Lab of IoT Application Technologies, Wuxi, China

Introduction: Motor imagery electroencephalographic (MI-EEG) signal recognition is used in various brain–computer interface (BCI) systems. In most existing BCI systems, this identification relies on classification algorithms. However, generally, a large amount of subject-specific labeled training data is required to reliably calibrate the classification algorithm for each new subject. To address this challenge, an effective strategy is to integrate transfer learning into the construction of intelligent models, allowing knowledge to be transferred from the source domain to enhance the performance of models trained in the target domain. Although transfer learning has been implemented in EEG signal recognition, many existing methods are designed specifically for certain intelligent models, limiting their application and generalization.

Methods: To broaden application and generalization, an extended-LSR-based inductive transfer learning method is proposed to facilitate transfer learning across various classical intelligent models, including neural networks, Takagi-SugenoKang (TSK) fuzzy systems, and kernel methods.

Results and discussion: The proposed method not only promotes the transfer of valuable knowledge from the source domain to improve learning performance in the target domain when target domain training data are insufficient but also enhances application and generalization by incorporating multiple classic base models. The experimental results demonstrate the effectiveness of the proposed method in MI-EEG signal recognition.

1 Introduction

A brain–computer interface (BCI) is a technology that establishes connections between the brain and external devices, facilitating information exchange between them (Edelman et al., 2024). BCIs collect and analyze electrical signals generated by brain activity, transforming these signals into instructions that can be used to control external devices such as computers, prosthetics, and wheelchairs. As such, BCIs can assist, enhance, and repair human sensory and motor functions, improving human–computer interaction capabilities. BCIs do not rely on the peripheral nervous system or muscles, providing a new method for people who have lost their mobility due to illness or disability to communicate with the external environment and operate devices. BCIs not only open new possibilities for people with disabilities but also advance our understanding of the brain, ushering in a new era of human–computer interaction.

1.1 Motivation

Motor imagery electroencephalographic (MI-EEG) (Mohammadi et al., 2022) signal recognition is an important mechanism for brain-computer interfaces (BCIs). Moreover, with the advancement of machine learning, numerous classification methods based on machine learning have been proposed for MI-EEG signal recognition in the literature (Abbas et al., 2021; Ko et al., 2021; Zhang et al., 2024; Ghumman et al., 2021; Cover and Hart, 1967; Aldea et al., 2014; Kohavi, 1996; Wang and Zhang, 2016; Fisher, 1936; Li et al., 2022; Bennett and Demiriz, 1999; Fouad et al., 2020; Siddiqa et al., 2024; Siddiqa et al., 2023; Qureshi et al., 2022; Qureshi et al., 2023), including neural networks (NNs) (Abbas et al., 2021; Ko et al., 2021), fuzzy logic systems (FLSs) (Zhang et al., 2024; Ghumman et al., 2021), k-nearest neighbors (kNNs) (Cover and Hart, 1967; Aldea et al., 2014), naïve Bayes (NB) (Kohavi, 1996; Wang and Zhang, 2016), linear discriminant analysis (LDA) (Fisher, 1936; Li et al., 2022), support vector machines (SVMs) (Bennett and Demiriz, 1999; Fouad et al., 2020), and more. Although these methods have demonstrated varying degrees of success, they typically require a large amount of subject-specific training data to adjust their parameters. However, this data acquisition process can be time-consuming and not user-friendly. When calibration data is insufficient, the classification performance of these algorithms can significantly deteriorate. As highlighted in BCI Competition III (Blankertz et al., 2006), “a challenge is that more expectations of training a model with a good classification accuracy are becoming urgent in the case that only a small number of training samples are available.” Therefore, it is essential to develop advanced machine-learning methods for MI-EEG that perform effectively with small calibration datasets.

Transfer learning is a promising method for addressing the above problem. It can be used to transfer useful information from related scenes (i.e., source domains) to the current scene (i.e., target domain), which typically has limited training data (Pan and Yang, 2010). As a result, transfer learning is particularly effective in improving classification performance during the early stages of model training when there is not enough subject-specific training data. Figure 1 shows the differences between traditional machine learning and transfer learning. Since its introduction in 1995, transfer learning has been successfully applied in classification, clustering, and regression, with classification being the most extensively researched area. Some representative studies can be found in Zhang et al. (2022), Jiang et al. (2019), Xie et al. (2018), Pan et al. (2011), Wan et al. (2021), Li et al., 2019). Existing transfer learning methods can be categorized into three types: inductive transfer learning methods (Zhang et al., 2022; Jiang et al., 2019), which consider both supervised source and target domains; transductive transfer learning methods (Xie et al., 2018; Pan et al., 2011), which involve supervised source domains and unsupervised target domains; and unsupervised transfer learning methods, which account for both unsupervised source and target domains (Wan et al., 2021; Li et al., 2019). In MI-EEG signal recognition, when labeled MI-EEG samples in the target domain are insufficient, inductive transfer learning methods naturally become the preferred choice. Furthermore, since MI-EEG signals involve personal privacy information, inspired by Jiang et al. (2019), we investigate a knowledge-based inductive transfer learning method to ensure security without directly utilizing samples from the source domain.

Figure 1
www.frontiersin.org

Figure 1. Differences between traditional machine learning (a) and transfer learning (b).

Inductive transfer learning has recently attracted widespread attention and demonstrated strong performance in MI-EEG signal recognition. However, most existing inductive transfer learning methods are tailored to specific base models, rendering them inapplicable to other base models. As a result, they demonstrate poor performance in terms of application and generalization. To address this limitation, we propose an extended-LSR-based inductive transfer learning framework (ELSR-TL) that integrates neural networks, Takagi-Sugeno-Kang (TSK) fuzzy systems, and kernel methods. Figure 2 shows the framework of ELSR-TL.

Figure 2
www.frontiersin.org

Figure 2. Framework of ELSR-TL.

1.2 Contributions

The main contributions of this study can be highlighted as follows:

1. ELSR-TL has an inductive transfer learning mechanism that can be used to transfer useful knowledge from the source domain to enhance learning performance in the target domain when the training data in the target domain are insufficient.

2. ELSR-TL enhances LSR by integrating multiple classic base models, such as neural networks, TSK fuzzy systems, and kernel methods. As such, ELSR-TL is not only suited for a specific model but also demonstrates improved applicability and generalization.

3. Experimental studies were conducted to validate the applicability of the proposed method for MI-EEG signal identification.

The remainder of this paper is organized as follows: Section II describes related work, including studies on existing MI-EEG feature extraction and pattern recognition methods. Section III details the proposed extended-LSR-based inductive transfer learning method. Section IV provides the experimental results and analysis. Finally, Section VI presents the conclusions drawn.

2 Backgrounds

This section states the backgrounds underlying the proposed MI-EEG recognition method. It describes the datasets used to evaluate the method and reviews several classical feature extraction and pattern recognition methods.

2.1 Datasets

We used BCI Competition Data Set IVa, provided by Fraunhofer FIRST and Charité University Medicine Berlin. A detailed description of this dataset can be found in (Blankertz et al., 2006).

This MI-EEG dataset contains five subsets corresponding to five healthy testers (aa, al, av., aw, and ay). Each subset contains 280 EEG trials, which have 128 electrodes and a trial length of 3.5 s. Each subset was partitioned into a training set and a test set, as shown in Figure 3. Figure 4 shows the representative MI-EEG signals in the five subsets.

Figure 3
www.frontiersin.org

Figure 3. Distribution of each subset from the BCI Competition Data Set IVa.

Figure 4
www.frontiersin.org

Figure 4. Representative MI-EEG signals for each subset of the BCI Competition Data Set IVa.

2.2 Feature extraction methods

EEG signals are complex, nonlinear, and non-stationary. Effective feature extraction is critical to pattern recognition performance. Some of the most representative feature extraction methods have been proposed to manage raw MI-EEG signals. Typically, feature extraction methods can be classified into four main categories: time-domain analysis, frequency-domain analysis, time-frequency analysis, and space-domain analysis.

In time domain analysis, EEG signal features are analyzed in the time domain. Characteristics of the waveforms, such as mean, variance, amplitude, and kurtosis, can be used to extract features of MI-EEG signals (Greene et al., 2008).

In frequency domain analysis, the features of EEG signals are analyzed by investigating the relationship between their frequency and energy. The short-time Fourier transform (Schafer and Rabiner, 1973) is a classical power spectrum analysis method, and adaptive autoregression (Pfurtscheller et al., 1998) is an improved frequency domain analysis method.

In time-frequency analysis (Blanco et al., 1997), the features of EEG signals are extracted using the joint distribution information of the time and frequency domains. Wavelet transform analysis (Antonini et al., 1992) is the most representative method in this category.

In space-domain analysis, the features of EEG signals are extracted by analyzing the electrical activity of neurons in different brain spaces. Common spatial pattern (CSP) (Lotte and Guan, 2011) is a commonly used method in this category. In this method, labeled trials are used to produce a transformation that maximizes the variance of one class while minimizing the variance of the other.

2.3 Pattern recognition methods

Pattern recognition utilizes the extracted EEG features for classification. Some of the most representative pattern recognition methods include the following: (1) NNs (Abbas et al., 2021; Ko et al., 2021), which simulate the mechanism of the human nervous system. Feedforward NNs are the most commonly used in EEG classification. (2) FLSs (Zhang et al., 2024; Ghumman et al., 2021), which emulate the human reasoning process and excel at managing numerical and linguistic uncertainties. (3) kNNs (Cover and Hart, 1967; Aldea et al., 2014), which determine the class of a new sample by considering its k nearest neighbors. (4) NB (Kohavi, 1996; Wang and Zhang, 2016), a simple and efficient classification algorithm based on probability. By utilizing known conditional probability and a priori probability, NB calculates the posterior probability of each class and assigns the test sample to the class with the highest a posteriori probability. (5) LDA (Fisher, 1936; Li et al., 2022), which applies the Fisher criterion to find the optimal projective vector that maximizes the largest scatter between classes while minimizing the scatter within each class. (6) SVMs (Bennett and Demiriz, 1999; Fouad et al., 2020), which aim to maximize the margins between different classes.

Although existing MI-EEG classification methods have demonstrated their effectiveness in various applications, they all require a substantial amount of subject-specific training data. In practice, such training data may not be easy to obtain, and the classification accuracy of existing methods may drop significantly. To address this challenge, we use an inductive transfer learning-based MI-EEG classification method.

3 Extended-LSR-based inductive transfer learning

In this section, we provide a detailed description of the proposed extended-LSR-based inductive transfer learning (ELSR-TL) method. First, we extend LSR (Naseem et al., 2010) to its extended version, ELSR, by merging neural networks, TSK fuzzy systems, and kernel methods. Then, we develop the proposed ELSR-TL. Finally, we present the learning algorithm and theoretical analysis of ELSR-TL.

3.1 ELSR

3.1.1 Objective function of ELSR

ELSR is an extension of the basic LSR (Naseem et al., 2010). Given n d-dimensional samples {(xi,yi)}Ni=1 , where xid , yi{1;+1} , the objective function of ELSR can be expressed as follows:

minw12Xρwy2+λ2w2    (1)
where the matrix Xρ=[ρ(x1),,ρ(xN)]TN×dρ denotes all the given training samples, and ρ(x)dρis the hidden mapping function in the hidden mapping space. w represents the mapping matrix, λ is the given regularization parameter, and y is the corresponding label matrix.

The decision-making function of ELSR can be expressed as follows:

y=f(x)=ρ(x)Tw    (2)

Using different mapping functions ρ(x) , we can integrate multiple models, such as neural networks, TSK fuzzy systems, and kernel methods, into the proposed ELSR framework. In other words, ELSR can be developed for different base models, which improves its generalization and adaptability. We will describe its relationships with several base models next.

3.1.1.1 The relationship between ELSR and feedforward NNs

A multiple hidden layer feedforward network (MHFN) has an input layer, M hidden layers, and an output layer. The multiple hidden layers can be treated as a single complex hidden layer, allowing the overall activation function of these hidden layers to be represented by a single complex function. Therefore, an MHFN can be viewed as a generalized single hidden layer feedforward network (SHFN) with a more complex activation function. The output of a generalized SHFN can be expressed as follows:

y=f(x)=
where N M is the number of nodes in the last hidden layer of an MHFN. As demonstrated in Huang et al. (2006), if the activation function g i x θ i is piecewise continuous, then the hidden nodes can be randomly generated independently of the training data, and the corresponding NN still maintains its universal approximation capability. Let the hidden mapping function ρ x as Equation 4:
ρ x = g 1 x θ 1 , , g N M x θ N M T     (4)

Then, Equation 3 can be expressed as follows:

y = f x = ρ x T w     (5)

Comparing Equation 5 with Equation 2, we can see that Equation 5 is a special case of Equation 2, so Equation 1 can be used to optimize the corresponding MHFN.

3.1.1.2 The relationship between ELSR and TSK fuzzy systems

The Takagi–Sugeno–Kang fuzzy system (Gu et al., 2024; Bian et al., 2024) is the most widely used FLS due to its simplicity and flexibility. The rules in a TSK fuzzy system are typically represented as Equation 6:

TSK Fuzzy Rule R k :

IF x 1 is A 1 k x 2 is A 2 k x d is A d k Then f k x = p 0 k + p 1 k x 1 + + p d k x d k = 1 , , K     (6)

Here, A i k is a fuzzy set for the ith input variable in the kth rule, K is the number of fuzzy rules, and is a fuzzy conjunction operator. The output of the TSK fuzzy system is computed as Equation 7:

y = f x = Σ k = 1 K μ k x Σ k = 1 K μ k x f k x = Σ k = 1 K μ ˜ k x f k x     (7)
where μ k x is the firing level of Rule R k , and μ ˜ k x is the normalized μ k x , i.e., Equation 8:
μ k x = i = 1 d μ A i k x i     (8)
μ ˜ k x = μ k x / Σ k = 1 K μ k x     (8)

The parameters of the antecedent fuzzy sets are usually derived from clustering. The output of the TSK fuzzy system can subsequently be expressed as as Equations 9, 10:

y = f x = ρ x T w     (9)
where
ρ x = x ˜ 1 T x ˜ 2 T x ˜ K T T     (10)
x ˜ k = μ ˜ k x x e     (10)
x e = 1 x T T     (10)
w = p 1 T p 2 T p K T T     (10)
p k = p 0 k p 1 k p d k T     (10)

Equation 9 suggests that training the TSK fuzzy system can also be treated as a special case of ELSR, and thus, it can be addressed using Equation 1.

3.1.1.3 The relationship between ELSR and kernel methods

A kernel linear regression model is expressed as follows:

y = ρ x T w     (11)

The hidden mapping ρ x can be viewed as a kernel function; thus, Equation 11 can also be solved using Equation 1. In this case, ELSR also corresponds to the classical kernel ridge regression (Saunders et al., 1998).

3.1.2 Solution of ELSR

Depending on the condition of the hidden mapping, the objective function of ELSR in Equation 1 can be efficiently solved in various ways ρ x . Here, we discuss the different cases as follows:

Case 1: ρ x is known: In this case, we can obtain explicit values of the data ρ x in the hidden mapping space.

Let J w = min w 1 2 X ρ w y 2 + λ 2 w 2 ; according to the optimization theory (Qu et al., 2023a; Qu et al., 2023b), the solution for the model parameter w can then be obtained by taking the derivatives of Equation 1 and equating them to zero. That is,

J w w = 0 X ρ T X ρ w X ρ T y + λ w = 0 w = X ρ T X ρ + λ I d ρ 1 X ρ T y     (12)

The final decision function f x can then be expressed as Equation 13:

y = f x = ρ x T w     (13)
with w obtained in Equation 12.

Case 2: ρ x is unknown: In this case, the explicit formulation of the data ρ x in the hidden mapping space cannot be obtained, meaning that w cannot be specified explicitly. Therefore, the kernel trick is necessary to determine the final decision function f x . Although introducing the kernel trick into the solution strategy in Equation 12 is challenging, Equation 14, identity can be adopted to address this issue:

P 1 + Q T U 1 Q 1 Q T U 1 = P Q T Q P Q T + U 1     (14)

In Equation 14, P, Q, and U are three matrices. Let P = 1 λ I d ρ , Q = X ρ , and U = I N . With the identity of Equation 14, the solution in Equation 12 can then be expressed as follows:

w = X ρ T X ρ + λ I d ρ 1 X ρ T y = X ρ T X ρ X ρ T + λ I N 1 y     (15)

Define a Mercer kernel matrix as Equation 16:

Ω = X ρ X ρ T N × N ,     (16)
where Ω i , j = ρ x i T ρ x j = K x i x j , and K · is a kernel function.

The final decision function f x can then be expressed as follows:

y = f x = ρ x T w = ρ x T X ρ T X ρ X ρ T + λ I N 1 y = K x x 1 K x x N Ω + λ I N 1 y     (17)

3.2 ELSR-TL

3.2.1 Objective function of ELSR-TL

ELSR-TL integrates transfer learning and ELSR. Its objective function can be expressed as follows:

min w t 1 2 X ρ , t w t y 2 + λ 2 w t 2 + β 2 w t w s 2     (18)
where X ρ , t N t × d ρ represents N t training samples of d ρ dimensions in the target domain. w t and w s represent the mapping matrices of the target domain and source domain, respectively. λ and β are the given regularization parameters, and y is the corresponding label matrix of the target domain.

In Equation 18, the first two terms are inherited directly from ELSR for learning from the target domain data, while the third term is used to leverage knowledge from the source domain. In other words, ELSR-TL generalizes ELSR from the perspective of transfer learning.

Moreover, as a regularization parameter, β can be used to adjust the role of transfer learning. When β is large, it indicates that transfer learning has a significant impact, indicating that the knowledge obtained from the source domain has a significant positive effect on the target domain. In contrast, when β is very small, it indicates that its role in learning of the target domain is relatively small. In extreme cases, when β = 0 , it means that β has no effect on the learning of the target domain. In other words, we can control the effectiveness of transfer learning by making adjustments, thus effectively avoiding negative transfer.

3.2.2 Solution of ELSR-TL

ELSR-TL is solved differently in different scenarios:

Case 1: ρ x is known: In this case, we can obtain explicit values of the data ρ x in the hidden mapping space. The solution for the model parameter w t can then be obtained in a similar form as that shown in Equation 12, that is

J w t w t = 0 X ρ , t T X ρ , t w t X ρ , t T y + λ w t + β w t w s = 0 w t = X ρ , t T X ρ , t + λ + β I d ρ 1 X ρ , t T y + β w s     (19)

The final output of the proposed ELSR-TL is expressed as follows:

y = f x = ρ x T w t     (20)
with w t obtained in Equation 19.

Case 2: ρ x is unknown: In this case, the explicit formulation of the data ρ x in the hidden mapping space cannot be obtained, and thus, w t cannot be specified explicitly. Similar to the form shown in Equation 17, the output of the proposed ELSR-TL can be calculated using the kernel trick. From Equation 15, we know that w s can be expressed as Equation 21:

w s = X ρ , s T X ρ , s X ρ , s T + λ s I N s 1 y s     (21)

Here, w s is the parameter of ELSR in the source domain. For a similar scenario, let

α s = X ρ , s X ρ , s T + λ s I N s 1 y s     (22)

A Mercer kernel matrix is defined, and Equation 22 can then be re-expressed as Equation 23:

α s = Ω s + λ s I N s 1 y s     (23)
where Ω s = X ρ , s X ρ , s T = K x i , s x j , s N s × N s , in which K x i , s x j , s is the kernel function.

w s can then be written as follows:

w s = X ρ , s T α s     (24)

From Equation 19, we obtain the following:

w t = β λ + β w s + 1 λ + β X ρ , t T X ρ , t X ρ , t T + λ + β I N t 1 y β λ + β X ρ , t w s     (25)

Substituting Equation 24 into Equation 25 and defining a Mercer kernel matrix, the equation above can then be rewritten as follows:

w t = β λ + β w s + 1 λ + β X ρ , t T X ρ , t X ρ , t T + λ + β I N t 1 y β λ + β X ρ , t w s = β λ + β X ρ , s T α s + 1 λ + β X ρ , t T Ω t + λ + β I N t 1 y β λ + β Ω t , s α s     (26)
where Ω t = X ρ , t X ρ , t T = K x i , t x j , t N t × N t , Ω t , s = X ρ , t X ρ , s T = K x i , t x j , s N t × N s

Finally, by using w t obtained in Equation 26, the decision function of the proposed ELSR-TL can be expressed as follows:

f x = ρ x T w t = ρ x T β λ + β X ρ , s T α s + 1 λ + β X ρ , t T Ω t + λ + β I N t 1 y β λ + β Ω t , s α s = β λ + β K x 1 , s x K x N s , s x T α s + 1 λ + β K x 1 , t x K x N t , t x T Ω t + λ + β I N t 1 y β λ + β Ω t , s α s     (27)

3.3 Learning algorithm of ELSR-TL

Considering the above discussion, we summarize the learning algorithm of ELSR-TL in Algorithm 1. Below, we provide some remarks on ELSR-TL.

Algorithm 1
www.frontiersin.org

Algorithm 1. The ELSR-TL.

3.3.1 Remark 1

For the proposed ELSR-TL, if the hidden mapping is known and the amount of training data exceeds the dimensionality of the hidden mapping features (i.e., ( N t d ρ , t ), obtaining the solution using Equation 25 is more efficient than that using Equation 19, due to the computational complexity of matrix; otherwise, Equation 19 is more efficient.

3.3.2 Remark 2

When the hidden mapping is known, only the knowledge w s is used for transfer learning, and the data in the source domain are not required. This means that the proposed method provides good privacy protection. However, if the hidden feature mapping is unknown, the data in the source are also required, as shown in Equations 26, 27, to effectively implement transfer learning. In this case, the proposed method can no longer protect the privacy of the data in the source domain.

3.3.3 Computational Complexity

In this section, we discussed the computational complexity of Algorithm 1 as follows:

When the hidden mapping is known, the complexity of computing step 1 is about O d ρ , s 3 + d ρ , s 2 N s , where d ρ , s is the dimension of samples and N s is the number of samples in the source domain. The complexity of computing the target domain model parameters w t in step 2 is about O d ρ , t 3 + d ρ , t 2 N t , where d ρ , t is the dimension of samples and N t is the number of samples in the target domain. In this case, the computational complexity of Algorithm 1 is about O d ρ , s 3 + d ρ , s 2 N s + d ρ , t 3 + d ρ , t 2 N t . When the hidden mapping is unknown, the computational complexity of Algorithm 1 is about O N s 3 + N t 3 .

4 Experiments

In this section, we adopted a real MI-EEG dataset to evaluate the performance of the proposed ELSR-TL method. Moreover, we compared it with seven non-transfer learning methods—LSR (Naseem et al., 2010), KNN (Cover and Hart, 1967), SVM (Bennett and Demiriz, 1999), NB (Kohavi, 1996), CNN (Zhang et al., 2019), ELSR (NN), ELSR (TSK), and ELSR (Ker)—alongside two transfer learning methods—Au-SVM (Wu and Dietterich, 2004) and Tr-Adaboost (Dai et al., 2007). The comparison was conducted in terms of both average classification accuracy and standard deviation for 10 runs. The details of the experimental settings and the MI-EEG recognition results are provided as follows.

4.1 Data preparation and feature extraction

4.1.1 Configurations of source and target domains

To match the transfer learning task, we constructed 20 different transfer learning datasets by subject-to-subject transferring. Table 1 shows the 20 different configurations of source and target domains. All source domains have the same number of training data, but the target domains do not. Please note that in our experiments, non-transfer learning methods are only used on the target domain.

Table 1
www.frontiersin.org

Table 1. Settings of the source domain and target domain.

4.1.2 Feature extraction

As mentioned in the background section, effective feature extraction is critical to pattern recognition performance. Based on (Lotte and Guan, 2011), we primarily used the Tikhonov regularization-based common spatial pattern (TR-CSP) (Lotte and Guan, 2011) for feature extraction. Furthermore, we conducted simple experiments using two other feature extraction methods, namely Composite CSP (C-CSP) and Filter Bank CSP (FB-CSP), to compare with TR-CSP. The three feature extraction methods are briefly introduced as follows:

1. TR-CSP: It introduces a quadratic regularization into the CSP objective function and replaces the feature matrix of the new data with the prior knowledge matrix. This regularization prefers filters with smaller norms, reducing the influence of noise.

2. C-CSP: It aims to perform subject-to-subject transfer by regularizing the covariance matrices using data from other subjects. Within the framework of this study, it relies only on the β hyperparameter and defines the generic covariance matrices according to the covariance matrices of other subjects.

FB-CSP: This is a feature extraction method used for motor imagery classification in BCI. It improves the accuracy of motion imagery classification by combining CSP and filter bank techniques, optimizing the subject-specific frequency band for CSP.

We extracted features from the time segment between 0.5 and 2.5 s after the cue instructing the subject to perform MI. Each trial is bandpass filtered in the 8–30 Hz range using a fifth-order Butterworth filter. For TR-CSP, we applied three pairs of filters, as recommended in (Lotte and Guan, 2011). Some examples of features extracted from subset aa are shown in Figure 5.

Figure 5
www.frontiersin.org

Figure 5. Features extracted from subset aa by TR-CSP.

4.2 Adopted methods and parameters settings

All the adopted methods (without CNN) are listed in Table 2. Based on the guidelines in (Jiang et al., 2019; Xie et al., 2018; Zhang X. et al., 2023) and our experiments, we employ a grid search strategy to identify the appropriate parameters for all the adopted methods. Table 2 also includes a list of the grid search ranges for each parameter related to all the adopted methods.

Table 2
www.frontiersin.org

Table 2. The parameter setting of different methods.

4.3 Performance indices

The classification accuracy defined Equation 28 is used to evaluate the performances of different methods:

Accuracy = T P + T N T P + T N + F P + F N     (28)

4.4 Results and discussions

In all the experiments, each comparison method is implemented for 10 runs to report the average classification performance. The experimental results are shown in Tables 3, 4 and Figure 6. Please note that the feature extraction method used for these results is TR-CSP. We can make the following observations:

i. In general, the performance of the proposed ELSR-TL-based methods significantly surpasses that of the other methods used, whether they are non-transfer learning methods such as LSR, kNN, SVM, NB, CNN, and ELSR-based methods, or transfer learning methods such as Au-SVM and Tr-Adaboost. This provides experimental evidence that ELSR-TL effectively enhances MI-EEG recognition through knowledge transfer from the source domain to the target domain.

ii. Comparing the performances of the seven non-transfer learning methods, we can see that the performance of ELSR (TSK) is the best, while the performance of NB is inferior. Moreover, each method obtains significant performance differences on different datasets. Specifically, seven non-transfer learning methods obtain the best performance on dataset al but poor performance on datasets aw and ay. This is because these methods require a large amount of training samples to achieve satisfactory performance, while their performance decreases when there are few training samples.

iii. Table 4 shows the performances of five transfer learning methods, showing that ELSR-TL (TSK) performs the best while Tr-Adaboost performs the worst. Furthermore, each method achieves similar performances for each target subject, regardless of the auxiliary subject chosen as the source domain. Additionally, for a fixed configuration of the source and target domains, the three ELSR-TL-based methods yield similar classification accuracies.

iv. Comparing the performances of transfer learning methods (i.e., Au-SVM and ELSR-TL-based methods) with their corresponding non-transfer learning methods (i.e., SVM and ELSR-based methods), we can see that the transfer learning methods outperform the others. Therefore, transfer learning strategies are effective for MI-EEG signal recognition. Impressively, even the least effective transfer learning methods still perform better than or are comparable to the non-transfer learning methods.

v. When datasets aa, al, and av are used in the target domain, the performance improvements achieved by the ELSR-TL-based methods are not very significant. This is because the target domain contains sufficient data to train a good model; therefore, the knowledge from the source domain is not critical. However, when datasets aw and ay are used in the target domain, the ELSR-TL-based methods significantly outperform the other methods due to the limited training data available in the target domain.

vi. To visually compare the performances of all methods, Figure 6 illustrates the performance of each method across all datasets for visual comparison. Please note that, for transfer learning methods, we report the average performance of four different source domains with fixed target domains in Figure 6. For each dataset, the ELSR-TL-based methods achieve either the best accuracy or performance comparable to that of the other methods.

Table 3
www.frontiersin.org

Table 3. Classification accuracies of the non-transfer learning method.

Table 4
www.frontiersin.org

Table 4. Average accuracies of four different source domains for the transfer learning methods.

Figure 6
www.frontiersin.org

Figure 6. Classification accuracies of 13 different methods, where the accuracies of the transfer learning methods represent the average accuracies across four different source domains with a fixed target domain.

In summary, we show that ELSR-TL-based methods can outperform other methods, especially when the number of training samples in the target domain is limited.

4.5 On different feature extraction

In this section, we compare the effectiveness of three feature extraction methods. Figure 7 illustrates the classification results of these methods when using the same classification method. Specifically, we use ELSR-TL (TSK) as the classification method.

Figure 7
www.frontiersin.org

Figure 7. Classification results of three feature extraction methods.

4.6 On running time

In this section, we compared the average running times of all adopted methods (without CNN) over ten trials. Table 5 lists the average time (in seconds) for each method across all datasets. It is evident that LSR has the shortest computational time. However, among all transfer learning methods, the computational time of ELSR-TL (NN) is less than that of the other two transfer learning methods. Nevertheless, the running time of the proposed ELSR-TL is still not particularly small. Therefore, determining how to accelerate the proposed method for large-scale data remains an open problem that we should explore in the future.

Table 5
www.frontiersin.org

Table 5. Running time (Seconds) for all adopted methods on all datasets.

4.7 Statistical analysis

A nonparametric Friedman test (Zhang Y. et al., 2023) is used to validate whether the performance differences among different algorithms are statistically significant. This test uses the rankings of different algorithms in multiple comparisons. First, we calculate the sum ranking and average ranking of the accuracy of each algorithm (without CNN), as shown in Table 6, and find the best one. We then perform post hoc hypothesis testing.

Table 6
www.frontiersin.org

Table 6. Rankings of the 12 algorithms (Friedman test).

The Friedman test statistics are as Equation 29:

Q = 12 k n n + 1 Σ i = 1 k R i k n + 1 2 2 = 12 k n n + 1 Σ i = 1 k R i 2 3 k n + 1     (29)
where R i is the sum ranking of each algorithm, n is the number of algorithms, and k is the number of datasets.

From Table 6, we have Q = 26.25 , and the corresponding p-value is 0.005964. This suggests that the performance differences among the 12 methods are statistically significant, with ELSR-TL (TSK) performing the best. To further evaluate the performance differences between ELSR-TL (TSK) and the other 11 methods, we also conduct post hoc multiple comparison tests:

z = R ¯ 0 R ¯ i S E with S E = n n + 1 6 k

where z is subject to the standard normal distribution that will be used further to calculate the value of P.

Table 7 shows the post hoc comparison results for α = 0.05 (Friedman). The null hypothesis is rejected when p 0.00625 because p Holm . In summary, we conclude that there are significant performance differences between ELSR-TL (TSK) and other methods, confirming that transfer learning is effective in boosting classification accuracy.

Table 7
www.frontiersin.org

Table 7. The post-hoc comparison for α=0.05 (Friedman).

In addition, Figure 8 shows the specific differences between ELSR-TL (TSK) and other methods. Clearly, it is consistent with the above conclusion that there are significant performance differences between ELSR-TL (TSK) and the other methods.

Figure 8
www.frontiersin.org

Figure 8. Specific differences of ELSR-TL (TSK) compared to other methods.

4.8 Sensitivity analysis

We also conduct experiments to study the sensitivity of ELSR-TL to various parameters. Below, we use the AW dataset as an example of sensitivity analysis. Figure 9 illustrates how accuracy varies with different values of four parameters while the others remain fixed, based on the grid search detailed in section 4.2. Please note that, due to the limitations of this paper, we only use ELSR-TL (TSK) for the sensitivity analysis.

Figure 9
www.frontiersin.org

Figure 9. Accuracy changes with different values of four parameters: (a) M, (b) τ , (c) λ , (d) β .

4.9 Limitations

Although the proposed ELSR-TL demonstrates effectiveness in these experiments, it still has some limitations. For example, there are four hyperparameters in the proposed method, and the hyperparameter optimization procedure based on grid searching and cross-validation is computationally expensive. The running time of the proposed method is significant, especially compared to traditional simple methods, making it unsuitable for real-time scenarios. The proposed ELSR-TL operates offline and cannot be used in online scenarios. The datasets used in this paper contain only five subjects on a small scale, and the effectiveness of the proposed method needs to be validated on more extensive and larger datasets in future studies. In addition, it is also worth further investigating how to provide more theoretical justifications for knowledge transfer and how to avoid negative transfer. We primarily focused on binary MI tasks in this study; therefore, exploring how to extend the proposed ELSR-TL to multi-class MI tasks, multimodal integration, and cross-dataset transfers is worth studying.

5 Conclusion

In this study, an extended LSR-based inductive transfer learning method was proposed to facilitate transfer learning for several classical intelligent models, including neural networks, TSK fuzzy systems, and kernel methods. We applied this method to MI-EEG signal recognition in BCIs. ELSR-TL provides three distinctive advantages: (1) It features an inductive transfer learning mechanism that allows for the transfer of useful knowledge from the source domain to enhance learning performance in the target domain when the training data in the target domain are insufficient. (2) It enhances application and generalization by extending LSR while integrating multiple classic base models such as neural networks, TSK fuzzy systems, and kernel methods. (3) It uses knowledge extracted from the source domain to train the classification model in the target domain, ensuring security for MI-EEG signal recognition. Experimental studies indicate the effectiveness of the proposed method in MI-EEG signal recognition. Although the proposed ELSR-TL demonstrates effectiveness in these experiments, there is still room for further research. For example, the hyperparameter optimization procedure based on grid searching and cross-validation is computationally expensive, so future research should focus on addressing this issue. The proposed ELSR-TL operates offline and cannot be applied in real-time scenarios. The datasets used in this study are relatively small in scale; thus, the effectiveness of the proposed method needs validation on more extensive datasets in future studies. Additionally, it is also worth further examining how to provide more theoretical justifications for knowledge transfer to avoid negative transfer. While this study primarily focuses on binary MI tasks, extending the proposed ELSR-TL to multi-class MI tasks, multimodal integration, and cross-dataset transfers is also worth studying.

Data availability statement

Publicly available datasets were analyzed in this study. This data can be found at: https://www.bbci.de/competition/iii/desc_IVa.html.

Author contributions

ZJ: Writing – original draft, Writing – review & editing. KH: Formal analysis, Writing – review & editing. JQ: Methodology, Validation, Visualization, Writing – review & editing. ZB: Data curation, Methodology, Validation, Visualization, Writing – review & editing. DY: Methodology, Validation, Visualization, Writing – review & editing. JZ: Conceptualization, Methodology, Supervision, Writing – review & editing.

Funding

The author(s) declare that financial support was received for the research and/or publication of this article. This study was partially supported by the National Natural Science Foundation of China through Grants 62206177, 62106145, 62002227, and 62306126; from the Zhejiang Provincial Natural Science Foundation of China under Grants LY23F020007, LQ22F020024, LTY22F020003, and LZ24F020006; and from the Natural Science Foundation of Jiangsu Province through Grant BK20220621.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The authors declare that no Gen AI was used in the creation of this manuscript.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Abbas, A., Sutter, D., Zoufal, C., Lucchi, A., Figalli, A., and Woerner, S. (2021). The power of quantum neural networks. Nat. Comput. Sci. 1, 403–409. doi: 10.1038/s43588-021-00084-1

PubMed Abstract | Crossref Full Text | Google Scholar

Aldea, R., Fira, M., and Lazăr, A. (2014). Classifications of motor imagery tasks using k-nearest neighbors. In 12th symposium on Neural Network applications in electrical engineering (NEUREL), IEEE. 115–120.

Google Scholar

Antonini, M., Barlaud, M., Mathieu, P., and Daubechies, I. (1992). Image coding using wavelet transform. IEEE Trans. Image Process. 1, 205–220. doi: 10.1109/83.136597

PubMed Abstract | Crossref Full Text | Google Scholar

Bennett, K., and Demiriz, A. (1999). Semi-supervised support vector machines. Adv. Neural Inf. Proces. Syst. 11, 368–374.

Google Scholar

Bian, Z., Qu, J., Zhou, J., Jiang, Z., and Wang, S. (2024). Weighted adaptively ensemble clustering method based on fuzzy co-association matrix. Inform. Fusion 103:102099. doi: 10.1016/j.inffus.2023.102099

PubMed Abstract | Crossref Full Text | Google Scholar

Blanco, S., Kochen, S., Rosso, O., and Salgado, P. (1997). Applying time-frequency analysis to seizure EEG activity. Engin. Med. Biol. Magazine IEEE 16, 64–71. doi: 10.1109/51.566156

PubMed Abstract | Crossref Full Text | Google Scholar

Blankertz, B., Müller, K., Krusienski, D., Schalk, G., Wolpaw, J. R., Schlögl, A., et al. (2006). The BCI competition III: validating alternative approaches to actual BCI problems. IEEE Trans. Neural Syst. Rehabil. Eng. 14, 153–159. doi: 10.1109/TNSRE.2006.875642

PubMed Abstract | Crossref Full Text | Google Scholar

Cover, T., and Hart, P. (1967). Nearest neighbor pattern classification. IEEE Trans. Inf. Theory 13, 21–27. doi: 10.1109/TIT.1967.1053964

Crossref Full Text | Google Scholar

Dai, W., Yang, Q., Xue, G., and Yu, Y., “Boosting for transfer learning,” in Proc. 24th Int. Conf. Mach. Learning, (2007), pp. 193–200.

Google Scholar

Edelman, B. J., Zhang, S., Schalk, G., Brunner, P., Muller-Putz, G., Guan, C., et al. (2024). Non-invasive brain-computer interfaces: state of the art and trends. IEEE Rev. Biomed. Eng. 18, 26–49. doi: 10.1109/RBME.2024.3449790

PubMed Abstract | Crossref Full Text | Google Scholar

Fisher, R. (1936). The use of multiple measurements in taxonomic problems. Ann. Eugenics 7, 179–188. doi: 10.1111/j.1469-1809.1936.tb02137.x

Crossref Full Text | Google Scholar

Fouad, I. A., Labib, F. E. Z. M., Mabrouk, M. S., Sharawy, A. A., and Sayed, A. Y. (2020). Improving the performance of P300 BCI system using different methods. Network Model. Analysis Health Inform. Bioinform. 9, 1–13. doi: 10.1007/s13721-020-00268-1

PubMed Abstract | Crossref Full Text | Google Scholar

Ghumman, M. K., Singh, S., Singh, N., and Jindal, B. (2021). Optimization of parameters for improving the performance of EEG-based BCI system. J. Reliable Intelligent Environ. 7, 145–156. doi: 10.1007/s40860-020-00117-y

Crossref Full Text | Google Scholar

Greene, B. R., Faul, S., Marnane, W. P., Lightbody, G., Korotchikova, I., and Boylan, G. B. (2008). A comparison of quantitative EEG features for neonatal seizure detection. Clin. Neurophysiol. 119, 1248–1261. doi: 10.1016/j.clinph.2008.02.001

PubMed Abstract | Crossref Full Text | Google Scholar

Gu, S., Chou, Y., Zhou, J., Jiang, Z., and Lu, M. (2024). Takagi–sugeno–Kang fuzzy clustering by direct fuzzy inference on fuzzy rules. IEEE Transact. Emerg. Topics Comput. Intelligence 8, 1264–1279. doi: 10.1109/TETCI.2023.3336537

Crossref Full Text | Google Scholar

Huang, G., Chen, L., and Siew, C. (2006). Universal approximation using incremental constructive feedforward networks with random hidden nodes. IEEE Trans. Neural Netw. 17, 879–892. doi: 10.1109/TNN.2006.875977

PubMed Abstract | Crossref Full Text | Google Scholar

Jiang, Z., Chung, F. L., and Wang, S. (2019). Recognition of multiclass epileptic EEG signals based on knowledge and label space inductive transfer. IEEE Trans. Neural Syst. Rehabil. Eng. 27, 630–642. doi: 10.1109/TNSRE.2019.2904708

PubMed Abstract | Crossref Full Text | Google Scholar

Ko, W., Jeon, E., Jeong, S., and Suk, H-Il. (2021). Multi-scale neural network for EEG representation learning in BCI. IEEE Comput. Intell. Mag. 16, 31–45. doi: 10.1109/MCI.2021.3061875

Crossref Full Text | Google Scholar

Kohavi, R. (1996). Scaling up the accuracy of naive-Bayes classifiers: a decision-tree hybrid. KDD. 96, 202–207.

Google Scholar

Li, S., Jin, J., Daly, I., Liu, L., and Cichocki, A. (2022). Feature selection method based on Menger curvature and LDA theory for a P300 brain–computer interface. J. Neural Eng. 18:066050. doi: 10.1088/1741-2552/ac42b4

Crossref Full Text | Google Scholar

Li, J., Qiu, S., Shen, Y. Y., Liu, C. L., and He, H. (2019). Multisource transfer learning for cross-subject EEG emotion recognition. IEEE Transact. Cybernet. 50, 3281–3293. doi: 10.1109/TCYB.2019.2904052

PubMed Abstract | Crossref Full Text | Google Scholar

Lotte, F., and Guan, C. (2011). Regularizing common spatial patterns to improve BCI designs: unified theory and new algorithms. IEEE Trans. Biomed. Eng. 58, 355–362. doi: 10.1109/TBME.2010.2082539

PubMed Abstract | Crossref Full Text | Google Scholar

Mohammadi, E., Daneshmand, P. G., and Khorzooghi, S. M. S. M. (2022). Electroencephalography-based brain–computer interface motor imagery classification. J. Med. Signals Sensors 12, 40–47. doi: 10.4103/jmss.JMSS_74_20

PubMed Abstract | Crossref Full Text | Google Scholar

Naseem, I., Togneri, R., and Bennamoun, M. (2010). Linear regression for face recognition. IEEE Trans. Pattern Anal. Mach. Intell. 32, 2106–2112. doi: 10.1109/TPAMI.2010.128

PubMed Abstract | Crossref Full Text | Google Scholar

Pan, S., Tsang, I., Kwok, J., and Yang, Q. (2011). Domain adaptation via transfer component analysis. IEEE Trans. Neural Netw. 22, 199–210. doi: 10.1109/TNN.2010.2091281

PubMed Abstract | Crossref Full Text | Google Scholar

Pan, S., and Yang, Q. (2010). A survey on transfer learning. IEEE Trans. Knowl. Data Eng. 22, 1345–1359.

Google Scholar

Pfurtscheller, G., Neuper, C., Schlögl, A., and Lugger, K. (1998). Separability of EEG signals recorded during right and left motor imagery using adaptive autoregressive parameters. IEEE Trans. Rehabil. Eng. 6, 316–325. doi: 10.1109/86.712230

PubMed Abstract | Crossref Full Text | Google Scholar

Qu, J., Song, Z., Cheng, X., Jiang, Z., and Zhou, J. (2023a). A new integrated framework for the identification of potential virus–drug associations. Front. Microbiol. 14:1179414. doi: 10.3389/fmicb.2023.1179414

PubMed Abstract | Crossref Full Text | Google Scholar

Qu, J., Song, Z., Cheng, X., Jiang, Z., and Zhou, J. (2023b). Neighborhood-based inference and restricted Boltzmann machine for small molecule-miRNA associations prediction. PeerJ 11:e15889. doi: 10.7717/peerj.15889

PubMed Abstract | Crossref Full Text | Google Scholar

Qureshi, M. F., Mushtaq, Z., Rehman, M. Z. U., and Kamavuako, E. N. (2022). Spectral image-based multiday surface electromyography classification of hand motions using CNN for human–computer interaction. IEEE Sensors J. 22, 20676–20683. doi: 10.1109/JSEN.2022.3204121

Crossref Full Text | Google Scholar

Qureshi, M. F., Mushtaq, Z., Rehman, M. Z. U., and Kamavuako, E. N. (2023). E2cnn: an efficient concatenated cnn for classification of surface emg extracted from upper limb. IEEE Sensors J. 23, 8989–8996. doi: 10.1109/JSEN.2023.3255408

Crossref Full Text | Google Scholar

Saunders, C., Gammerman, A., and Vovk, V., “Ridge regression learning algorithm in dual variables,” (ICML-1998) Proceedings of the 15th International Conference on Machine Learning. Morgan Kaufmann, pp. 515–521 (1998)

Google Scholar

Schafer, R., and Rabiner, L. (1973). Design and simulation of a speech analysis-synthesis system based on short-time Fourier analysis. IEEE Trans. Audio Electroacoust. 21, 165–174. doi: 10.1109/TAU.1973.1162474

Crossref Full Text | Google Scholar

Siddiqa, H. A., Irfan, M., Abbasi, S. F., and Chen, W. (2023). Electroencephalography (EEG) based neonatal sleep staging and detection using various classification algorithms. Comput. Materials Continua 77, 1759–1778. doi: 10.32604/cmc.2023.041970

Crossref Full Text | Google Scholar

Siddiqa, H. A., Tang, Z., Xu, Y., Wang, L., Irfan, M., Abbasi, S. F., et al. (2024). Single-Channel EEG data analysis using a multi-branch CNN for neonatal sleep staging. IEEE Access 12, 29910–29925. doi: 10.1109/ACCESS.2024.3365570

Crossref Full Text | Google Scholar

Wan, Z., Yang, R., Huang, M., Zeng, N., and Liu, X. (2021). A review on transfer learning in EEG signal analysis. Neurocomputing 421, 1–14. doi: 10.1016/j.neucom.2020.09.017

Crossref Full Text | Google Scholar

Wang, H., and Zhang, Y. (2016). Detection of motor imagery EEG signals employing Naïve Bayes based learning process. Measurement 86, 148–158. doi: 10.1016/j.measurement.2016.02.059

PubMed Abstract | Crossref Full Text | Google Scholar

Wu, P., and Dietterich, T. G., “Improving SVM accuracy by training on auxiliary data sources,” in Proc. 21st Int. Conf. Mach. Learning, (2004)

Google Scholar

Xie, L., Deng, Z., Xu, P., Choi, K. S., and Wang, S. (2018). Generalized hidden-mapping transductive transfer learning for recognition of epileptic electroencephalogram signals. IEEE Transact. Cybernet. 49, 2200–2214. doi: 10.1109/TCYB.2018.2821764

PubMed Abstract | Crossref Full Text | Google Scholar

Zhang, X., Dong, S., Shen, Q., Zhou, J., and Min, J. (2023). Deep extreme learning machine with knowledge augmentation for EEG seizure signal recognition. Front. Neuroinform. 17:1205529. doi: 10.3389/fninf.2023.1205529

PubMed Abstract | Crossref Full Text | Google Scholar

Zhang, J., Li, Y., Liu, B., Chen, H., Zhou, J., Yu, H., et al. (2024). A dynamic broad TSK fuzzy classifier based on iterative learning on progressively rebalanced data. Inf. Sci. 677, 120976.

Google Scholar

Zhang, Y., Wang, G., Huang, X., and Ding, W. (2023). TSK fuzzy system fusion at sensitivity-ensemble-level for imbalanced data classification. Inform. Fusion 92, 350–362. doi: 10.1016/j.inffus.2022.12.014

Crossref Full Text | Google Scholar

Zhang, Y., Xia, K., Jiang, Y., Qian, P., Cai, W., Qiu, C., et al. (2022). Multi-modality Fusion & Inductive Knowledge Transfer Underlying non-Sparse Multi-Kernel Learning and distribution adaption. IEEE/ACM Trans. Comput. Biol. Bioinform. 20, 2387–2397. doi: 10.1109/TCBB.2022.3142748

PubMed Abstract | Crossref Full Text | Google Scholar

Zhang, L., Song, J., Gao, A., Chen, J., Bao, C., and Ma, K. “Be your own teacher: Improve the performance of convolutional neural networks via self distillation,” In Proc. IEEE/CVF Int. Conf. Comput. Vis., Seoul, Korea (South), (2019), pp. 3713–3722.

Google Scholar

Keywords: motor imagery, EEG, brain-computer interface, LSR, inductive transfer learning

Citation: Jiang Z, Hu K, Qu J, Bian Z, Yu D and Zhou J (2025) Recognition of MI-EEG signals using extended-LSR-based inductive transfer learning. Front. Neuroinform. 19:1559335. doi: 10.3389/fninf.2025.1559335

Received: 12 January 2025; Accepted: 06 March 2025;
Published: 09 April 2025.

Edited by:

Saadullah Farooq Abbasi, University of Birmingham, United Kingdom

Reviewed by:

Muhammad Farrukh Qureshi, Namal College, Pakistan
Hafza Ayesha Siddiqa, Fudan University, China

Copyright © 2025 Jiang, Hu, Qu, Bian, Yu and Zhou. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Jie Zhou, sxuj_zhou@163.com

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.

Research integrity at Frontiers

94% of researchers rate our articles as excellent or good

Learn more about the work of our research integrity team to safeguard the quality of each article we publish.


Find out more