Skip to main content

ORIGINAL RESEARCH article

Front. Aging Neurosci., 24 June 2022
Sec. Alzheimer's Disease and Related Dementias
This article is part of the Research Topic Mild Cognitive Impairment Recognition Via Gene Expression Mining and Neuroimaging Techniques View all 16 articles

Discriminant Subspace Low-Rank Representation Algorithm for Electroencephalography-Based Alzheimer’s Disease Recognition

  • 1School of Computer Science and Information Engineering, Changzhou Institute of Technology, Changzhou, China
  • 2School of Information Engineering, Changzhou Institute of Industry Technology, Changzhou, China
  • 3School of Computer Science and Artificial Intelligence, Changzhou University, Changzhou, China
  • 4Department of Nephrology, The Affiliated Wuxi People’s Hospital of Nanjing Medical University, Wuxi, China

Alzheimer’s disease (AD) is a chronic progressive neurodegenerative disease that often occurs in the elderly. Electroencephalography (EEG) signals have a strong correlation with neuropsychological test results and brain structural changes. It has become an effective aid in the early diagnosis of AD by exploiting abnormal brain activity. Because the original EEG has the characteristics of weak amplitude, strong background noise and randomness, the research on intelligent AD recognition based on machine learning is still in the exploratory stage. This paper proposes the discriminant subspace low-rank representation (DSLRR) algorithm for EEG-based AD and mild cognitive impairment (MCI) recognition. The subspace learning and low-rank representation are flexibly integrated into a feature representation model. On the one hand, based on the low-rank representation, the graph discriminant embedding is introduced to constrain the representation coefficients, so that the robust representation coefficients can preserve the local manifold structure of the EEG data. On the other hand, the least squares regression, principle component analysis, and global graph embedding are introduced into the subspace learning, to make the model more discriminative. The objective function of DSLRR is solved by the inexact augmented Lagrange multiplier method. The experimental results show that the DSLRR algorithm has good classification performance, which is helpful for in-depth research on AD and MCI recognition.

Introduction

Alzheimer’s disease (AD) is a disease characterized by memory loss, slow and gradual changes in brain function, and the manifestations of intellectual loss (Zhang et al., 2021). With the advancement of global aging, AD has now become a major public health problem affecting the world. The existing treatment of AD can only temporarily help relieve memory and cognition, but not a cure. To obtain disease-controlling treatments, it is an urgent need to classify the course of AD for early diagnosis. And especially, the National Institutes of Health revised the clinical diagnostic criteria for AD, characterizing research guidelines for early diagnosis and treatment (Cummings, 2021). The progression of AD is mainly divided into three stages. The first is the early clinical stage with no symptoms; the second is the intermediate stage with mild cognitive impairment (MCI); and the final stage with dementia symptoms (Mirzaei and Adeli, 2022).

More researchers are studying methods that can sensitively and conveniently monitor AD, involving cognitive neuropsychological detection, biochemical detection, neuroimaging detection, and so on. In recent years, electroencephalography (EEG) has become an important tool for studying human brain activity (Ghorbanian et al., 2015). Noninvasive EEG imaging methods are directly related to neural local field potentials and have a high temporal resolution. The millisecond-level temporal resolution and direct electrophysiological information provided by EEG can accurately reflect cognitive behaviors related to human neural activity. Therefore, more studies are beginning to use EEG for the diagnosis and prediction of early AD. For example, EEG spectral studies have revealed that EEG diffuse slow waves are a major feature of AD. EEG studies of AD patients have shown that the reduced power in the alpha (8–15 Hz) band and the increased power in the delta (0.5–4 Hz) band are significant features of AD (Fröhlich et al., 2021). The increase in power in the theta (4–8 Hz) band and the decrease in power in the beta (15–30 Hz) band also indicate that they can be useful for detecting MCI to AD transitions (Maturana-Candelas et al., 2020). Recently, machine learning technology has been widely used in the analysis of brain imaging data, which has greatly promoted the development of cognitive neuroscience. Most of the research revolves around feature extraction and classifier optimization. In terms of feature extraction, Wen et al. (2020) first converted the EEG signals into multispectral images and then used a deep convolutional neural network learning model for EEG classification. Similarly, Ieracitano et al. (2019a) drew the power spectral density of the EEG into the form of a spectrogram, and converted the EEG signal classification into a CNN-based image classification problem. Ieracitano et al. (2019b) spliced the continuous wavelet transform features and bispectral features of EEG signals to achieve the fusion of the two types of features. The advantage of this algorithm is that the fused features can obtain higher accuracy than only using one type of feature. The disadvantage is that the correlation between features is not considered enough. At the same time, the dimension of fusion features is greatly increased, which is easy causing the over-fitting problem.

In terms of classification algorithms, Miltiadous et al. (2021) compared six classification algorithms for EEG analysis for frontotemporal dementia in AD and verified the effectiveness of these algorithms. This study provided solutions for the early diagnosis of frontotemporal dementia. Anuradha and Jamal (2021) detected the progression of AD by detecting abnormal behavior in EEG. The authors used a feed-forward artificial neural network as a classifier to perform EEG feature analysis on abnormal and normal subjects and obtained a classification accuracy of 94.4%. Ge et al. (2020) exploited the robust biomarkers in EEG, combined linear discriminant analysis as a classifier, and proposed a systematic identification framework based on signal processing and computer-aided techniques for the detection of AD. Araujo et al. (2022) developed an intelligent system that can distinguish various stages of AD through EEG signals. The system used wavelet packet to extract multi-band features of EEG signals and used multiple machine learning methods as classification models.

Electroencephalography signals can reflect the functional state of the brain and the activity of brain physiological structures. The difficulties in classifying EEG signals using machine learning algorithms are as follows: first, the amplitude of the EEG signals is usually around 50 μv. The EEG signals are very weak, and their background noise is usually very strong. Second, EEG signals have strong randomness. In the process of acquisition, EEG signals will not only be stimulated by the outside world but also produce interference signals due to their own blinking and other actions. Therefore, it is still a challenging task to use machine learning methods to identify AD based on EEG signals. To solve this problem, the researchers usually reduce the dimension of EEG high-dimensional data and extract a small amount of the most valuable compact information, which not only saves storage space and processing time but also enables learning a robust model (Lei et al., 2021). Subspace learning and low-rank representation can well achieve this goal. Subspace learning is a well-known dimension reduction method in machine learning. Its main goal is to adopt appropriate strategies to map high-dimensional original data into the low-dimensional subspace to reduce the data dimension. Low-rank representation (LRR) can effectively separate the noise in the EEG signals to restore clean data and obtain accurate subspace segmentation of data.

Inspired by the strong theory of subspace learning and low-rank representations, this paper proposes an EEG-based discriminant subspace low-rank representation learning algorithm (DSLRR) for AD recognition. On the one hand, based on the low-rank representation, DSLRR utilizes the supervised information and local manifold information by least squares regression (LSR) and graph discriminant embedding. On the other hand, DSLRR introduces principal component analysis (PCA) and global preserved constraints into the subspace of learning. The algorithm optimization adopts a strategy of alternating parameter updates using the inexact augmented Lagrange multiplier method. Our contribution is as follows: (1) The DSLRR algorithm combines subspace learning and low-rank representation in a flexible manner. (2) By introducing global graph embedding and PCA term, the data projection can preserve the global structure information of EEG data in the discriminant subspace. (3) The learned low-rank representation coefficient can effectively avoid the negative effects of the original data’s redundant features and noise information. (4) By introducing LSR and graph discriminant embedding, the learned low-rank representation coefficient can explicitly contain the intrinsic local manifold structure and discriminant information of EEG data. The experiments on four EEG datasets verify that the DSLRR algorithm can be effectively used for the recognition of AD, MCI, and healthy control (HC).

Background

Electroencephalography Dataset for Alzheimer’s Disease and Mild Cognitive Impairment Recognition

The EEG data were obtained from 109 participants recruited at the IRCCS Centro Neurolesi Bonino-Pulejo in Italy, including 23 HC, 49 AD, and 37 MCI (Fiscon et al., 2018). The age of men and women and the proportion of genders are shown in Figure 1. The EEG data collection time was from 2012 to 2013. The scalp electrode position was determined using the international 10–20 system, and EEG data from 19 electrodes were collected. The sampling frequency was 256 or 1,024, and the acquisition time of EEG signals was 300 s. To reduce the effect of the artifact, the EEG signals from 60 to 240 s were selected, and the adopted normalized sampling frequency was 256 Hz. Feature extraction adopted the fast Fourier transform, which divided 180 s of data into six epochs of 30 s, and extracted 16 Fourier coefficients. Therefore, 304 features (19 electrodes × 16 Fourier coefficients) were available for each sample.

FIGURE 1
www.frontiersin.org

Figure 1. The basic information of EEG data used in this study, (A) age of men and women, and (B) proportion of gender.

Subspace Learning

We have a labeled dataset with n samples Y = [y1,…,yn] ∈ Rd×n, where yi represents the ith training sample, and its class label matrix is Y¯=[y¯1,,y¯n]RC×n. The dimension of the sample is d, and n samples are divided into C classes. When the dimensionality of the original EEG data is high, the data computational and storage costs will be very large. Thus, a common solution is to project the high-dimensional data into a low-dimensional space (Lei et al., 2021). Let QRd×C be the projection matrix, the projection data can be represented as V = [v1,…,vn] ∈ RC×n in the label space, where V = QTY.

Generally speaking, the premise of manifold subspace learning is that the data exists in high-dimensional space in the form of manifold embedding from low-dimensional space data. The key point of manifold learning is to ensure that low-dimensional data can reflect the inherent structural information contained in high-dimensional space (Zhang et al., 2020). As a commonly used manifold learning algorithm, locality-preserving projection (LPP) preserves the local neighbor relationship of the data by using an adjacency graph and affinity matrix (Weng and Shen, 2008). The LPP algorithm consists of three steps. Step 1 is to construct an adjacency graph. For example, we construct an adjacency graph using the k-nearest neighbor algorithm. The nearest neighbors of each point connected to it are known as neighbor nodes. Step 2 is to assign weights to each edge. In the adjacency graph, the affinity matrix represents the similarity between sample points, which can generally be calculated using the two-value method, cosine distance or Gaussian kernel function. For example, the affinity matrix E constructed by the two-value method can be defined as follows:

E ij = { 1 , if y i N k ( y j ) or y j N k ( y i ) 0 , otherwise (1)

where Nk(yi) represents the k nearest neighbor nodes of yi.

Low-Rank Representation

Low-rank representation aims to exploit the sparsity of matrix singular values to model high-dimensional data in multi subspace (Li et al., 2017; Jiang et al., 2021). Given a dataset Y, the LRR algorithm regards the input data itself as a dictionary and uses the basis in the dictionary to linearly represent the sample points, while minimizing its rank. The optimization problem of LRR can be described as follows:

min L rank ( L ) , s . t . Y = Y L , (2)

where LRn×n is the representation coefficients of Y, which reflects the global correlation between the original data samples. In theory, the coefficient matrix L obtained by the LRR should be a block diagonal matrix. That is to say, each block corresponds to a subspace, the number of blocks represents the number of data subspaces, and the size of the block corresponds to the dimension of the subspace.

Eq. (2) is not a convex optimization problem due to its discrete. Using the nuclear norm instead of rank(L), Eq. (2) can be transformed into the convex optimization problem as:

min L || L || * , s . t . Y = Y L , (3)

where ||||* is the nuclear norm.

Considering the noise or sparse error in Y, LRR enhances the model’s robustness by improving the correlation between the individual columns of L, and the problem of LRR can be written as:

min L , S || L || * + θ || S || 1 , s . t . Y = Y L + S , (4)

where SRd×n is sparse component of Y. θ is the regularization parameter.

Obviously, LRR decomposes the data Y into low-rank representation YL and sparse representation S. The former component YL generally represents the main features contained in Y, and the latter generally represents the redundant features and noise information contained in Y. In the clean data scenario, S represents the reconstruction error. Therefore, L can accurately indicate the subspace segmentation of Y, which ensures the robustness of the learned model. However, LRR ignores the role of local structure information in data and does not exploit the supervised information in the training data. Therefore, LRR cannot reflect the intra-class identity and inter-class dissimilarity in low-rank representation.

Discriminant Subspace Low-Rank Representation Algorithm

Objective Function

Discriminant Margin Term on Representation Coefficients

To learn the discriminant low-rank representations, we introduce graph discriminant embedding (Huang et al., 2018) into our algorithm, which combines supervised information to define intra-class and inter-class graph affinity matrices. We think if two EEG samples are closer in the original space, their representation coefficients will be close to each other. The compactness between samples of the same class and the separability between samples of different classes is the important knowledge in discriminant low-rank representations. To this end, we define affinity matrices Ecom and Esep to represent the similar relationship between intra-class and inter-class, respectively:

E i , j com = { e - || y i - y j || 2 t , if y i N ^ k ( y i ) or y j N ^ k ( y i ) , y ¯ j = y ¯ i e - t || y i - y j || 2 , if y i N ^ k ( y i ) or y j N ^ k ( y i ) , y ¯ j = y ¯ i 0 , if y ¯ j y ¯ i (5)
E i , j sep = { e - || y i - y j || 2 t , if y i N ~ k ( y i ) or y j N ~ k ( y i ) , y ¯ j y ¯ i e - t || y i - y j || 2 , i f y i N ~ k ( y i ) or y j N ~ k ( y i ) , y ¯ j y ¯ i 0 , if y ¯ j = y ¯ i (6)

where N^k() and N~k() represent the k-nearest neighbor samples of intra-class and inter-class, respectively. The parameter t (t > 0) is the weight parameter used to adjust the correlation between two samples. We set t = 1 in this study.

Then we define the discriminant margin term ς1(L) on representation coefficients:

ς 1 ( L ) = i = 1 n ς i ( L i ) = i = 1 n j = 1 n ( || L i - L j || 2 E i , j com - || L i - L j || 2 E i , j sep ) = Tr ( L T U L ) (7)

where U = EcomEsep + εI, ε is a very small positive. Eq. (7) represents the intra-class compactness and the inter-class dissimilarity in representation coefficients. Its essence is to excavate the local structural information representation coefficients. In addition, Eq. (7) can avoid the influence of the redundant information and noise of the original data.

Global Structure Term on Projection

We adopt the affinity matrix E to represent the correlation between two samples using supervised information. The element eij in E is computed as:

e i j = { 1 , if y i and y j are of the same class 0 , otherwise (8)

To preserve the global discriminant information of the original data in the subspace, we introduce the global structure term on projection:

ς 2 ( Q ) = 1 2 i , j e ij || Q T y i - Q T y j || 2 2 - β Tr ( Q T Y Y T Q ) = Tr ( Q T Y E Y T Q ) - β Tr ( Q T Y Y T Q ) = Tr ( Q T Y ( E - β I ) Y T Q ) (9)

where β is the regularization parameter.

The first factor i,jeij||QTyi-QTyj||22 in Eq. (9) is the global preserved component on projection. Obviously, when this component reaches the minimum, the distance of samples of the same class will be as close as possible in the projection subspace. The second component Tr(QTYYTQ) in Eq. (9) is the PCA component on projection. Its goal is to ensure that the projecting data in the low-dimensional subspace can depict the inherent structure information contained in the original space.

Least Squares Regression Term

As an effective supervised learning method, LSR learns the linear projection that transforms the sample to the label space, and obtains the regression vector as the data representation in the label space (Zhao et al., 2022). Therefore, we try to find a projection matrix with the help of LSR in the low-rank representation. Different from the traditional projection method on the original data, the DSLRR algorithm only uses clean data representation to learn the projection matrix in the low-rank representation framework, which can not be affected by the redundant information of EEG data. This idea can be obtained as:

ς 3 ( Q , L , S ) = || L || * + θ || S || 1 + γ || V - Y ¯ || F 2 + η || V || F 2 , s . t . Y = Y L + S , V = Q T Y L ,      1 n T L = 1 n T . (10)

where γ and η are regularization parameters.

Equation (10) tries to minimize the least squares loss between the regression results and the corresponding regression target. In addition, in the low-rank representation framework, the compact representation of the data can be learned through subspace projection.

The Objective Function

We integrate Eqs (7), (9), and (10) into a learning model, and obtain the objective function of the DSLRR algorithm:

min ς 1 ( L ) + ς 2 ( Q ) + ς 3 ( Q , L , S ) = min Q , L , S μ Tr ( L T U L ) + α Tr ( Q T Y ( E - β I ) Y T Q ) + || L || * + θ || S || 1 + γ || V - Y ¯ || F 2 + η || V || F 2 , s . t . Y = Y L + S , V = Q T Y L ,          1 n T L = 1 n T . (11)

where α and μ are regularization parameters.

From Eq. (11), we can see that the DSLRR algorithm combines subspace learning and low-rank representation into a learning model. Based on low-rank representation learning, the compact and discriminant low-rank representation can be reinforced by graph discriminant embedding. Based on subspace learning, the discriminant projection can be obtained by LSR, global structure preserved, and PCA technologies.

Optimization

There are three unsolved parameters {Q, L, S} in Eq. (11). To make Eq. (11) separable, the relaxation matrix Λ is introduced to represent L. Substitute the constraint V = QTYL into Eq. (11), Eq. (11) can be re-written as:

min Q , L , S || Λ || * + θ || S || 1 + γ || Q T Y L - Y ¯ || F 2 + η || Q T Y L || F 2 + μ Tr ( L T U L ) + α Tr ( Q T Y ( E - β I ) Y T Q ) , s . t . Y = Y L + S ,      1 n T L = 1 n T , L = Λ . (12)

We optimize three parameters by the inexact augmented Lagrange multiplier algorithm in an iterative optimization strategy (Kang et al., 2015). Eq. (12) has the following form:

min Q , L , S , Λ || Λ || * + θ || S || 1 + γ || Q T Y L - Y ¯ || F 2 + η || Q T Y L || F 2 + μ Tr ( L T U L ) + α Tr ( Q T Y ( E - β I ) Y T Q ) + Tr ( τ a T ( Y - Y L - S ) ) + Tr ( τ b T ( Λ - L ) ) + Tr ( τ c T ( 1 n T L - 1 n T ) ) + δ 2 ( | | Y-YL - S | | F 2 + | | Λ - L | | F 2 + | | 1 n T L - 1 n T | | F 2 ) , (13)

where δ is a trade-off parameter. The matrices τaRd×n, τbRd×n, τcRd×n, and τdRd×n are the Lagrange multipliers.

1) Optimize Q, while fixing the other parameters. Eq. (13) can be written as:

min Q γ || Q T Y L - Y ¯ || F 2 + η || Q T Y L || F 2 + α Tr ( Q T Y ( E - β I ) Y T Q ) . (14)

We can get the closed-solution of Q as:

Q = [ Y ( ( α ( E - β I ) ) + ( γ + η ) L L T ) Y T ] - 1 Y L Y ¯ T . (15)

2) Optimize Λ, while fixing the other parameters. Eq. (13) can be written as:

min Λ || Λ || * + Tr ( τ c T ( L - Λ ) ) + δ 2 || L - Λ || F 2 = min Λ 1 δ || Λ || * + 1 2 || Λ - ( L + 1 δ τ c ) || F 2 . (16)

We use the singular value thresholding operator (Cai et al., 2010; Li et al., 2017) to solve Eq. (16). We employ the singular value decomposition algorithm on L+1δτc as L+1δτc=HΣΔ, where H is the diagonal matrix with its element being a group of singular values {Θk},1≤kp, p is the rank. The matrix Λ can be computed by Λ = HΩ(1/δ)ΣΔ, in which Ω(1/δ)Σ=diag({Θk-1δ}+), where “+” means the positive part.

3) Optimize L, while fixing the other parameters. Eq. (13) can be written as:

min L γ || Q T Y L - Y ¯ || F 2 + η || Q T Y L || F 2 + μ Tr ( L T U L ) + Tr ( τ a T ( Y - Y L - S ) ) + Tr ( τ b T ( Λ - L ) ) + Tr ( τ c T ( 1 n T L - 1 n T ) ) + δ 2 ( || Y-YL - S || F 2 + || Λ - L || F 2 + || 1 n T L - 1 n T || F 2 ) , (17)

Let the first derivative of L in Eq. (16) be zero, we have,

Θ a L = Θ b + Θ c , (18)
Θ a = 2 U + 2 ( γ + η ) Y T Q Q T Y + δ ( Y T Y + 1 n 1 n T + I n ) , (19)
Θ b = 2 γ Y T Q Y ¯ + δ ( Y T Y-Y T S + Λ + 1 n 1 n T ) , (20)
Θ c = Y T τ a - τ b - 1 n τ c , (21)

We can get the closed-solution of L as:

L = Θ a - 1 ( Θ b + Θ c ) . (22)

4) Optimize S, while fixing the other parameters. Eq. (13) can be written as:

min S θ δ || S || 2 , 1 + 1 2 || S - ( Y - Y L + 1 δ τ a ) || F 2 , (23)

According to the theory of (Liu et al., 2013), we can obtain the S by

S ( : , i ) = { || τ i || - θ || τ i || || τ i || , i f θ δ < || τ i || 0 , otherwise (24)

where τi is the ith column vector of the matrix τa.

Testing

Given test EEG data Ytest, we first compute its low-rank representation Ltest using Eq. (11), while setting parameters γ = 0, α = 0, and μ = 0. Second, we construct the new training set YL and test set YtestLtest. Third, we use the training set YL to train a classifier and build a classifier to predict the label of YtestLtest. In this study, we used nearest neighbor (NN) algorithm as the classifier. The whole training and testing procedure for EEG data recognition are summarized in Algorithm 1.

Algorithm 1. DSLRR algorithm for EEG data recognition.

www.frontiersin.org

Experiment

Experimental Settings

To verify the effectiveness of the DSLRR algorithm, we compared the DSLRR algorithm with the SPCA (Jiang, 2011), LRR (Liu et al., 2013), LRDLSR (Chen and Yang, 2014), JSLC (Lu et al., 2021), and NRLRL (Gao et al., 2021) in the experiment. The LRR algorithm is the baseline algorithm of the DSLRR algorithm. SPCA and JSLC algorithms are subspace learning algorithms. LRDLSR and NRLRL are low-rank representation algorithms. For SPCA, the weight parameters are set in the covariance mixture, α is set inversely proportional to the sample size, and η is searched in [2−5,2−4,…,25]. For LRR, the parameter λ is searched in [1,4,,30]/d, where d is the data dimension. For LRDLSR, the parameters α and β are searched in [10−4,10−3,…,1], and the parameters γ and λ are set to be 0.01. For JSLC, subspace dimension and the size of the dictionary are searched in [50, 100,…, 300]. The regularization parameters are searched in [0.5, 1,…, 5]. For NRLRL, the size of the dictionary is searched in [50, 100,…, 300], and λ, γ, and η are searched in [2−4,2−3,…,24]. For DSLRR, all regularization parameters are searched in[2−4,2−3,…,24], and k-nearest neighbors in N^k() and N~k() are searched in [1,…, 11].

Due to the limited training EEG samples, we expand the EEG data with the data augmentation strategy. The number of EEG samples in HC, MCI, and AD is 69, 74, and 98, respectively. In this section, the experiments are conducted on four EEG datasets for AD and MCI recognition, namely, (1) HC & AD, (2) HC & MCI, (3) HC & (MCI+AD), and (4) MCI & AD. The ratio of the two classes of samples is 1:1. We randomly select 50 samples in each class for model training, and the rest samples are used for testing. We perform our experiments 10 times and record the classification performance in terms of accuracy, sensitivity, specificity, precision, F-measure, G-mean, and Jaccard. All experiments are conducted by MATLAB on a Windows machine.

Classification Results

The classification results in four EEG datasets are reported in Tables 14, where the best results are highlighted in bold. These four data sets are binary classification problems. According to the results in Tables 14, we can see that:

TABLE 1
www.frontiersin.org

Table 1. Classification results of the comparison algorithms in HC and AD dataset.

TABLE 2
www.frontiersin.org

Table 2. Classification results of the comparison algorithms in MCI and AD dataset.

TABLE 3
www.frontiersin.org

Table 3. Classification results of the comparison algorithms in HC and (MCI+AD) dataset.

TABLE 4
www.frontiersin.org

Table 4. Classification results of the comparison algorithms in HC and MCI dataset.

(1) Alzheimer’s disease is a population suffering from AD, which has shown clinical symptoms. The EEG signal differentiation between AD and healthy people is the most significant, and the difference between EEG features is more obvious. Therefore, the classification performance in the dataset of AD and HC is high. Although the symptoms of MCI are not as significant as those of AD, there is a certain probability of AD. The difference between the EEG features and those of healthy people is also significant, and the difference between EEG features is also obvious, so the classification performance in the dataset of MCI and HC is also high. In addition, AD and MCI are mixed into one class in the third dataset of HC and (AD+MCI), which is significantly distinguishable from healthy EEG signals. Therefore, its classification performance is expectable. The classification accuracy of DSLRR algorithm in AD and HC is 97.74%. The classification accuracy of the DSLRR algorithm in MCI & and HC is 95.61%. The classification accuracy of the DSLRR algorithm in HC and MCI+AD is 98.42%. The classification accuracy of these three datasets is above 97.26%. The experimental results illustrate that DSLRR can better identify MCI and AD from HC.

(2) Compared with the first three datasets, the difference between EEG features between MCI and AD is relatively low. Therefore, the classification performance of each algorithm decreases to a certain extent in the MCI & AD dataset. However, we can see that the DSLRR algorithm still achieves the best values of accuracy, F-measure, G-mean, and Jaccard. On the one hand, through the joint learning of subspace and low-rank representation, the DSLRR algorithm can learn the robust and discriminant projection subspace. On the other hand, by making full use of Laplace manifold and LSR technologies, the DSLRR algorithm can exploit the structure knowledge and manifold structure information of EEG signals. Furthermore, the sum of the columns of each low-rank coefficient matrix L of 1 has a positive effect on the classification.

(3) The LRR algorithm can describe the correlation of data, and the coefficient matrix is low rank. However, this algorithm doesn’t consider the local structural characteristics of the data, and often cannot effectively exploit the discriminant information in the data. In this case, the LRR algorithm is not directly applicable to the EEG classification for AD recognition. The JSLC algorithm achieves good results in four datasets. JSLC is a low-rank representation model based on dictionary learning, which integrates discriminant information of samples into dictionary learning, and can also eliminate the influence of noise information on the classification model. This result shows that joint learning of low-rank representation and subspace learning is an effective means to solve EEG classification. The NRLRL algorithm conducts low-rank learning in the original data space. Its classification performance is lower than DSLRR in four datasets, which further shows that more data dimensions may not improve model performance. Due to the redundant information and noise in EEG data, it is effective to obtain the compact and discriminant feature representation through subspace learning and low-rank representation.

Ablation Experiment

The DSLRR algorithm integrates discriminant margin term, global structure term, and LSR term on the basis of the LRR algorithm. To verify the role of these terms, we performed ablation experiments on four EEG datasets. For discriminant margin term, its purpose is to use supervised information to establish graph embedding, to improve the distinguishing ability of the model. To verify its effect, we remove this item from Eq. (11), that is, set the parameter μ = 0. For global structure terms, their purpose is to preserve the structure information of data in subspace. To verify its effect, we remove this item from Eq. (11) by setting the parameter α = 0. For the LSR term, its purpose is to use the least square constraint to utilize the discriminant information in the data. Similarly, to verify its effect, we remove this item from Eq. (11), that is, set the parameter γ = 0. The classification accuracy, F-measure, and G-means of DSLRR with an ablation experiment in four EEG datasets are shown in Figures 24, respectively. From the results in Figure 2, we can see that if any one of three terms is removed from Eq. (11), the classification accuracy in the four EEG datasets has decreased to varying degrees. This is because each term has a corresponding contribution to the EEG classification task, which also illustrates the necessity of the coexistence of these three terms from another perspective. The results in Figures 3, 4 show that this conclusion is well verified. Therefore, the lack of any term will degrade the performance of the DSLRR algorithm.

FIGURE 2
www.frontiersin.org

Figure 2. Classification accuracy of DSLRR with ablation experiment in four EEG datasets.

FIGURE 3
www.frontiersin.org

Figure 3. F-measure of DSLRR with ablation experiment in four EEG datasets.

FIGURE 4
www.frontiersin.org

Figure 4. G-mean of DSLRR with ablation experiment in four EEG datasets.

Parameter Analysis

To show the convergence of the DSLRR algorithm, we plot its convergence curve in Figure 5. As shown in Figure 5, the DSLRR algorithm converges quickly in several iterations across four EEG datasets. The results show that the DSLRR algorithm is acceptable in the running time, which shows that the DSLRR algorithm has high practical worthiness.

FIGURE 5
www.frontiersin.org

Figure 5. The convergence of the DSLRR algorithm in four datasets, (A) HC and AD, (B) HC and MCI, (C) HC and (MCI+AD), and (D) MCI and AD.

We plot the classification accuracy of the DSLRR algorithm with different k-nearest neighbors in Figure 6. Figure 6 visually shows that the classification is mildly sensitive to k. The DSLRR algorithm can achieve good classification accuracy when the parameter k is in the range of [5, 7, 9]. When k is <5 or k is greater than 9, the classification accuracy is slightly lower. Therefore, we can fix k = 7 in the experiment.

FIGURE 6
www.frontiersin.org

Figure 6. The accuracy of the DSLRR algorithm with different k in four datasets, (A) HC and AD, (B) HC and MCI, (C) HC and (MCI+AD), and (D) MCI and AD.

Conclusion

With the emergence of global aging, the prediction and diagnosis of AD have attracted extensive attention. In recent years, EEG technology has been developed and has become an important means to detect abnormal brain activity in patients with AD. To realize the early diagnosis of AD, we propose the DSLRR learning algorithm. The DSLRR algorithm inherits the advantages of low-rank representation, removes redundant information and noise, and improves the discriminant ability of low-rank representation through graph discriminant embedding. Meanwhile, based on subspace learning, the DSLRR algorithm introduces LSR and global structure preserving constraints to further improve the discriminative ability of the model. Extensive experimental results on real EEG data verify the effectiveness of the DSLRR algorithm.

In the future, we will continue to explore our work in the following aspects. First, the DSLRR algorithm is essentially a linear learning method. The brain is a nonlinear system with the ability of self-adaptation and self-regulation. Under some internal or external stimuli, the regulation and application functions of biological tissue will inevitably affect the electrophysiological signals, so that neurons have chaotic discharge phenomena, which present nonlinear characteristics. This makes the DSLRR algorithm unable to exert its performance in complex EEG data. To this end, we consider introducing a nonlinear learning model to improve the stability and accuracy of the DSLRR algorithm, so that it can be better suitable for various complex application scenarios. Second, the DSLRR algorithm is suitable for EEG classification using single-feature information. At present, the technologies of feature processing and feature exaction are more mature, and the obtained feature information is correspondingly more diverse. In the next stage, we will extend the proposed algorithm to multi-feature scenarios to form a richer AD recognition system. Third, with the popularization of EEG acquisition equipment, using the existing labeled samples to analyze the unlabeled samples in multiple domains is a difficult problem in EEG-based AD recognition. We will use transfer learning technology to extend our algorithm in the future, to further enhance the generalization of the algorithm.

Data Availability Statement

Publicly available datasets were analyzed in this study. The EEG dataset analyzed in this study can be found in: https://github.com/tsyoshihara/Alzheimer-s-Classification-EEG.

Author Contributions

TT, XG, and JX conceived and developed the model and wrote the manuscript. HL and GZ ran the experiment and analyzed the results. All authors read, edited, and approved the manuscript.

Funding

This work was supported in part by the Science and Technology Project of Changzhou city under grant CE20215032.

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Anuradha, G., and Jamal, D. N. (2021). Classification of dementia in EEG with a two-layered feed forward artificial neural network. Eng. Technol. Appl. Sci. Res. 11, 7135–7139. doi: 10.48084/etasr.4112

CrossRef Full Text | Google Scholar

Araujo, T., Teixeira, J. P., and Rodrigues, P. M. (2022). Smart-data-driven system for Alzheimer disease detection through electroencephalographic signals. Bioengineering (Basel) 9:141. doi: 10.3390/bioengineering9040141

PubMed Abstract | CrossRef Full Text | Google Scholar

Cai, J., Candèsm, E. J., and Shen, Z. (2010). A singular value thresholding algorithm for matrix completion. SIAM J. Optim. 20, 1956–1982. doi: 10.1137/080738970

CrossRef Full Text | Google Scholar

Chen, J., and Yang, J. (2014). Robust subspace segmentation via low-rank representation. IEEE Trans. Cybern. 44, 1432–1445. doi: 10.1109/TCYB.2013.2286106

PubMed Abstract | CrossRef Full Text | Google Scholar

Cummings, J. (2021). The role of neuropsychiatric symptoms in research diagnostic criteria for neurodegenerative diseases. Am. J. Geriatr. Psychiatry 29, 375–383. doi: 10.1016/j.jagp.2020.07.011

PubMed Abstract | CrossRef Full Text | Google Scholar

Fiscon, G., Weitschek, E., Cialini, A., Felici, G., Bertolazzi, P., Salvo, S. D., et al. (2018). Combining EEG signal processing with supervised methods for alzheimer’s patients classification. BMC Med. Inform. Decis. Mak. 18:35. doi: 10.1186/s12911-018-0613-y

PubMed Abstract | CrossRef Full Text | Google Scholar

Fröhlich, S., Kutz, D. F., Müller, K., and Voelcker-Rehage, C. (2021). Characteristics of resting state EEG power in 80+-year-olds of different cognitive status. Front. Aging Neurosci. 13:675689. doi: 10.3389/fnagi.2021.675689

PubMed Abstract | CrossRef Full Text | Google Scholar

Gao, M., Liu, R., and Mao, J. (2021). Noise robustness low-rank learning algorithm for electroencephalogram signal classification. Front. Neurosci. 15:797378. doi: 10.3389/fnins.2021.797378

PubMed Abstract | CrossRef Full Text | Google Scholar

Ge, Q., Lin, Z. C., Gao, Y. X., and Zhang, J. X. (2020). A robust discriminant framework based on functional biomarkers of EEG and its potential for diagnosis of Alzheimer’s disease. Healthcare 8:476. doi: 10.3390/healthcare8040476

PubMed Abstract | CrossRef Full Text | Google Scholar

Ghorbanian, P., Ramakrishnan, S., and Ashrafiuon, H. (2015). Stochastic non-linear oscillator models of EEG: the Alzheimer’s disease case. Front. Comput. Neurosci. 9:48. doi: 10.3389/fncom.2015.00048

PubMed Abstract | CrossRef Full Text | Google Scholar

Huang, P., Li, T., Cao, G., and Geng, Y. (2018). Feature extraction based on graph discriminant embedding and its applications to face recognition. Soft Comput. 23, 7015–7028. doi: 10.1007/s00500-018-3340-5

CrossRef Full Text | Google Scholar

Ieracitano, C., Mammone, N., Bramanti, A., Hussaind, A., and Morabitoa, F. C. (2019a). A convolutional neural network approach for classification of dementia stages based on 2Dspectral representation of EEG recordings. Neurocomputing 323, 96–107. doi: 10.1016/j.neucom.2018.09.071

CrossRef Full Text | Google Scholar

Ieracitano, C., Mammone, N., Hussain, A., and Francesco, M. (2019b). A novel multi-modal machine learning based approach for automatic classification of EEG recordings in dementia. Neural Netw. 123, 176–190. doi: 10.1016/j.neunet.2019.12.006

PubMed Abstract | CrossRef Full Text | Google Scholar

Jiang, X. (2011). Linear subspace learning-based dimensionality reduction. IEEE Signal Process. Mag. 28, 16–26. doi: 10.1109/MSP.2010.939041

CrossRef Full Text | Google Scholar

Jiang, Y., Gu, X., Wu, D., Hang, W., Xue, J., Qiu, S., et al. (2021). A novel negative-transfer-resistant fuzzy clustering model with a shared cross domain transfer latent space and its application to brain CT image segmentation. IEEE ACM Trans. Comput. Biol. Bioinform. 18, 40–52. doi: 10.1109/TCBB.2019.2963873

PubMed Abstract | CrossRef Full Text | Google Scholar

Kang, M., Kang, M., and Jung, M. (2015). Inexact accelerated augmented Lagrangian methods. Comput. Optim. Appl. 62, 373–404. doi: 10.1007/s10589-015-9742-8

CrossRef Full Text | Google Scholar

Lei, W., Ma, Z., Liu, S., and Lin, Y. (2021). EEG mental recognition based on RKHS learning and source dictionary regularized RKHS subspace learning. IEEE Access 9, 150545–150559. doi: 10.1109/ACCESS.2021.3124028

CrossRef Full Text | Google Scholar

Li, P., Yu, J., Wang, M., Zhang, L., Cai, D., and Li, X. (2017). Constrained low-rank learning using least squares-based regularization. IEEE Trans. Cybern. 47, 4250–4262. doi: 10.1109/TCYB.2016.2623638

PubMed Abstract | CrossRef Full Text | Google Scholar

Liu, G., Lin, Z., Yan, S., Sun, J., Yu, Y., and Ma, Y. (2013). Robust recovery of subspace structures by low-rank representation. IEEE Trans. Pattern Anal. Mach. Intell. 35, 171–184. doi: 10.1109/TPAMI.2012.88

PubMed Abstract | CrossRef Full Text | Google Scholar

Lu, J., Zhou, G., Zhu, J., and Xue, L. (2021). Joint subspace and low-rank coding method for makeup face recognition. Math. Prob. Eng. 2021, 1–8. doi: 10.1155/2021/9914452

CrossRef Full Text | Google Scholar

Maturana-Candelas, A., Gómez, C., Poza, J., Ruiz-Gómez, S. J., and Hornero, R. (2020). Inter-band bispectral analysis of EEG background activity to characterize Alzheimer’s disease continuum. Front. Comput. Neurosci. 14:70. doi: 10.3389/fncom.2020.00070

PubMed Abstract | CrossRef Full Text | Google Scholar

Miltiadous, A., Tzimourta, K. D., Giannakeas, N., Tsipouras, M. G., Afrantou, T., Ioannidis, P., et al. (2021). A1zheimer’s disease and frontotempora1 dementia a robust c1assification method of EEG signa1s and a comparison of va1idation methods. Diagnostics 11:1437. doi: 10.3390/diagnostics11081437

PubMed Abstract | CrossRef Full Text | Google Scholar

Mirzaei, G., and Adeli, H. (2022). Machine learning techniques for diagnosis of alzheimer disease, mild cognitive disorder, and other types of dementia. Biomed. Signal Process. Control 72:103293. doi: 10.1016/j.bspc.2021.103293

CrossRef Full Text | Google Scholar

Wen, D., Li, P., Li, X., Wei, Z., Zhou, Y., Pei, H., et al. (2020). The feature extraction of resting-state EEG signal from amnestic mild cognitive impairment with type 2 diabetes mellitus based on feature-fusion multispectral image method. Neural Netw. 4, 373–382. doi: 10.1016/j.neunet.2020.01.025

PubMed Abstract | CrossRef Full Text | Google Scholar

Weng, B., and Shen, J. (2008). Classification of multivariate time series using locality preserving projections. Knowl. Based Syst. 21, 581–587. doi: 10.1016/j.knosys.2008.03.027

CrossRef Full Text | Google Scholar

Zhang, Y., Chung, F., and Wang, S. (2020). Clustering by transmission learning from data density to label manifold with statistical diffusion. Knowl. Based Syst. 193:105330. doi: 10.1016/j.knosys.2019.105330

CrossRef Full Text | Google Scholar

Zhang, Y., Wang, S., Xia, K., Jiang, Y., and Qian, P. (2021). Alzheimer’s disease multiclass diagnosis via multimodal neuroimaging embedding feature selection and fusion. Inf. Fusion 66, 170–183. doi: 10.1016/j.inffus.2020.09.002

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhao, S., Wu, J., Zhang, B., and Fei, L. (2022). Low-rank inter-class sparsity based semi-flexible target least squares regression for feature representation. Pattern Recognit. 123:108346. doi: 10.1016/j.patcog.2021.108346

CrossRef Full Text | Google Scholar

Keywords: electroencephalography, Alzheimer’s disease, low-rank representation, subspace learning, classification

Citation: Tang T, Li H, Zhou G, Gu X and Xue J (2022) Discriminant Subspace Low-Rank Representation Algorithm for Electroencephalography-Based Alzheimer’s Disease Recognition. Front. Aging Neurosci. 14:943436. doi: 10.3389/fnagi.2022.943436

Received: 13 May 2022; Accepted: 06 June 2022;
Published: 24 June 2022.

Edited by:

Mohammad Khosravi, Persian Gulf University, Iran

Reviewed by:

Ximing Xia, Nanjing Institute of Technology (NJIT), China
Runmin Liu, Wuhan Sports University, China

Copyright © 2022 Tang, Li, Zhou, Gu and Xue. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Jing Xue, xuejing@njmu.edu.cn

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.