Skip to main content

ORIGINAL RESEARCH article

Front. Med., 14 November 2024
Sec. Precision Medicine
This article is part of the Research Topic Unveiling Complex Medical Interdependencies Through High-Order Correlation Mining View all 5 articles

A hypergraph transformer method for brain disease diagnosis

\r\nXiangmin HanXiangmin Han1Jingxi FengJingxi Feng2Heming XuHeming Xu2Shaoyi DuShaoyi Du2Junchang Li
Junchang Li3*
  • 1School of Software, Tsinghua University, Beijing, China
  • 2The Institute of Artificial Intelligence and Robotics, College of Artificial Intelligence, Xi'an Jiaotong University, Xi'an, China
  • 3Shenzhen Clinical Research Center for Mental Disorders, Shenzhen Kangning Hospital and Shenzhen Mental Health Center, Shenzhen, China

Objective: To address the high-order correlation modeling and fusion challenges between functional and structural brain networks.

Method: This paper proposes a hypergraph transformer method for modeling high-order correlations between functional and structural brain networks. By utilizing hypergraphs, we can effectively capture the high-order correlations within brain networks. The Transformer model provides robust feature extraction and integration capabilities that are capable of handling complex multimodal brain imaging.

Results: The proposed method is evaluated on the ABIDE and ADNI datasets. It outperforms all the comparison methods, including traditional and graph-based methods, in diagnosing different types of brain diseases. The experimental results demonstrate its potential and application prospects in clinical practice.

Conclusion: The proposed method provides new tools and insights for brain disease diagnosis, improving accuracy and aiding in understanding complex brain network relationships, thus laying a foundation for future brain science research.

1 Introduction

The structural and functional connections of brain networks reflect the interaction and collaboration between different brain regions (1, 2). Structural connectivity is typically represented by the distribution of neural fiber tracts (3), while functional connectivity describes the synchronous activity of different brain regions during specific tasks or at rest state (4). Understanding the structural and functional connectivity of brain networks is crucial for comprehending both normal brain function and pathological states (5). For instance, abnormalities in structural connectivity may be associated with brain tissue damage, while disruptions in functional connectivity could indicate communication issues between neurons. Therefore, studying functional and structural brain networks is essential for uncovering the mechanisms underlying brain disease diagnosis.

Current research on brain networks primarily relies on techniques such as functional magnetic resonance imaging (fMRI) and diffusion tensor imaging (DTI). fMRI captures brain activity during specific tasks or at rest, revealing functional connectivity between different brain regions. DTI, on the other hand, tracks the diffusion paths of water molecules within neural fibers, providing information on the structural connectivity of the brain's white matter. Integrating data from fMRI and DTI offers a more comprehensive and enriched perspective for diagnosing brain diseases. For example, in Alzheimer's disease research, fMRI can reveal changes in functional connectivity, while DTI can demonstrate the degradation of white matter structure. In recent years, multimodal imaging techniques that combine fMRI and DTI have become mainstream in brain network research, further enhancing diagnostic accuracy and depth of understanding regarding brain diseases.

Artificial intelligence (AI) has achieved great success in various fields (6). For the brain network analysis task, graph and hypergraph methods (79) have shown great potential in brain network research. Graph methods represent brain networks as vertices and edges, allowing for the analysis of pairwise, low-order relationships. However, these methods have limitations, as they fail to effectively capture higher-order relationships within brain networks. For instance, traditional graph neural networks (GNNs) (10, 11) often underperform in handling complex high-order interactions (12, 13), neglecting the interactions among multiple vertices. Hypergraph methods (14) introduce hyperedges, which better model higher-order relationships in brain networks, but challenges remain in integrating functional and structural brain networks (15). Although hypergraphs can represent high-order relationships among multiple vertices, existing methods lack effective strategies for integrating information from different modalities, making it difficult to fully leverage the advantages of multimodal data. Thus, new methods are needed to address these issues and improve the accuracy and reliability of brain disease diagnosis.

This paper proposes a hypergraph Transformer (HGTrans) method for calculating high-order correlations between functional and structural brain networks. By utilizing hypergraphs, we can effectively model the high-order interactions within brain networks. The Transformer model provides robust feature extraction and integration capabilities, capable of handling complex multimodal data. Specifically, we use hypergraphs to represent high-order correlations in brain networks, including both functional and structural connectivity. Then, we propose the cross-attention Transformer module to extract features and integrate information from the hypergraphs, constructing a joint representation of the functional-structural brain network. This approach not only captures high-order functional and structural correlations but also effectively integrates information from different modalities, enhancing brain disease diagnosis performance. The main contributions of this paper are as follows:

• We propose a hypergraph-based method for modeling and computing the integration of functional and structural brain networks, effectively capturing high-order correlations. By using hypergraph modeling, we can accurately represent high-order interactions among multiple regions within brain networks, thereby enhancing our understanding and diagnosis performance of brain diseases.

• We introduce the Transformer model into hypergraph-based multimodal brain disease diagnosis, integrating diverse information from fMRI and DTI to improve diagnostic accuracy. The Transformer is conducted to refine the structural embeddings by incorporating high-order relationships derived from the functional network, thereby enhancing the diagnosis of brain diseases.

• We validated our method on the ABIDE and ADNI datasets, showing that our approach outperforms all the traditional and graph-based methods for different types of brain diseases, demonstrating its potential and application prospects in brain disease diagnosis.

2 Materials and methods

2.1 Datasets and preprocessing

The proposed method is evaluated on the ABIDE (16) and ADNI (17) datasets. We utilized the NYU1 and TCD sites of the ABIDE database in this work. Specifically, the NYU1 dataset contains 55 subjects, of which 33 subjects are autism spectrum disorder (ASD) patients and 22 subjects are normal controls (NCs). The TCD site contains 40 subjects, of which 20 subjects are ASD subjects and 20 subjects are NCs. The ADNI dataset is collected from multiple sites that study for improving the clinical trials for the prevention and treatment of Alzheimer's disease (AD). We used a subset of ADNI in this work, consisting of 39 AD patients, 62 MCI patients, and 61 NCs. Each subject has both rs-fMRI and DTI data in this work. The AAL (18) brain atlas was used to segment the regions of interest (ROIs) of the brain network. We preprocessed the original rs-fMRI via DPARSF,1 and the original DTI via PANDAS.

2.2 Method

2.2.1 Preliminaries of hypergraph computation

The hypergraph computation framework models high-order correlations by using hyperedges, which represent complex relationships beyond pairwise connections, and performs collaborative computation on these high-order interactions. Each hyperedge can connect multiple vertices, allowing it to capture both low-order (pairwise) correlations and high-order correlations across larger vertex sets. This approach leverages these high-order interactions to optimize data usage and improve overall task performance.

Given a hypergraph H={V,E,W}, where V and E represent the vertex set and the hyperedge set, respectively, and W denotes the weight matrix of the hyperedges. The incidence matrix of the hypergraph is defined as a |V|×|E| matrix, with each entry defined as

H(v,e)={we(v),if ve0,if ve,    (1)

where we(v)∈W represents the weight of vertex v within the hyperedge e.

2.2.2 HGTrans framework

As shown in Figure 1, the proposed HGTrans Framework consists of two main modules: the hypergraph computation module based on brain imaging and the structure-function Transformer module. The former constructs high-order relational structures from the information embedded in fMRI and DTI, exploring the complex relationships between different brain regions under fMRI and DTI, and generating high-order feature representations for fMRI and DTI. Then, semantic computations are performed using a hypergraph neural network to generate high-order feature representations. The latter uses the high-order features of the functional brain network as keys (K) and values (V), and the high-order features of the structural brain network as queries (Q) to achieve information interaction and fusion within the Transformer module. Finally, the fused features are fed into a classifier to enable brain disease diagnosis.

Figure 1
www.frontiersin.org

Figure 1. The pipeline of the proposed HGTrans framework.

2.2.3 Hypergraph computation for fMRI and DTI

2.2.3.1 Higher-order functional brain network representation

To model the complex interactions within functional brain networks, we utilize hypergraphs, which allow for the connection of multiple ROIs in the brain, rather than just pairs of regions. This structure facilitates the representation of high-order associations that arise in functional brain activity. The time series data of 116 ROIs from each subject's resting-state fMRI (rs-fMRI) data is extracted, followed by calculating the Pearson correlation coefficient between each pair of ROIs. This correlation coefficient quantifies the degree of linear relationship, ranging from –1 (perfect negative correlation) to 1 (perfect positive correlation), with 0 indicating no linear association. Using this approach, a functional connectivity (FC) matrix of size 116 × 116 is then generated, where each element represents the pairwise linear correlation between two ROIs.

In the hypergraph model, each of the 116 ROIs is treated as a vertex in the set V={v1,v2,,v116}, with vi representing the i-th vertex. In this work, we fix the K value as 3. The vertex feature set Xf = {xf1, xf2, …, xf116} describes the Pearson correlation values between the i-th ROI and all other ROIs. To capture the structural relationships between ROIs, we apply a K-Nearest Neighbors (KNN) algorithm to identify the k1−1 nearest neighbors for each vertex vi. A hyperedge is then formed for each vertex, connecting it with its nearest neighbors. Each hyperedge ej, constructed using KNN with a specified k-value, can be expressed as ej = {v1, v2, …, vk}, where k represents the number of vertices in the hyperedge. The similarity between vertices is measured using Euclidean distance, calculated as follows:

Edist(vi,vj)=p=1d(l)(zi,p-zj,p)2    (2)

where Edist(vi, vj) denotes the Euclidean distance between vertices vi and vj, and d(l) represents the number of feature dimensions in layer l.

By incorporating KNN with multiple values of k, representing local and global scales, the resulting hyperedges reflect complex high-order interactions between the vertices. The functional brain network hypergraph is then used for hypergraph convolution, allowing the learning of vertex representations. The HGNN+ convolution operation (19) consists of a two-step message-passing scheme. The process is formalized as follows:

       Zt=WHTDe1XtXt+1=σ(Dv1HZθt+1)    (3)

where Xt|V|×Mt is the vertex feature matrix at layer t, and Zt|E|×Mt is the corresponding hyperedge feature matrix. The learnable parameter matrix θt+1Mt×Mt+1 defines the transformation for the subsequent layer. Initially, the incidence matrix H guides the aggregation of vertex features to generate the hyperedge feature matrix Zt. These features are then combined with vertex-specific hyperedge features using the learnable parameters θt, updating the vertex feature matrix Xt+1. A nonlinear activation function σ(·) is applied to facilitate the transformation of features.

The vertex embeddings derived from multiple layers of hypergraph convolution effectively capture high-order relationships between ROIs within the functional brain network. This modeling approach provides a superior representation of complex brain activity patterns.

2.2.3.2 High-order structural brain networks

DTI data is utilized to derive the structural connectivity (SC) matrix, which quantifies the fiber tract connections between various ROIs in the brain. This method facilitates a comprehensive evaluation of potential alterations in the structural brain network that may be associated with ASD, offering a holistic perspective on how the disease may impact brain function.

The structural brain network is characterized by features such as small-world architecture and rich-club organization, both of which are critical for understanding network efficiency and communication. High-order structural characteristics are captured by computing the clustering coefficient ci and degree centrality di for each ROI, based on the SC matrix. The clustering coefficient assesses the extent of local interconnectivity, while degree centrality indicates the relative importance of each region within the broader network. These metrics provide valuable insights into the efficiency of information processing and communication within and between local brain regions.

The feature representation for each vertex in the network is defined as Xs = {xs1, xs2, …, xs116}, where xsi represents the feature vector for the i-th ROI, with xsi = {csi, dsi}. These initial features serve as input for subsequent analysis and modeling. To capture the higher-order relationships between ROIs, a K-Nearest Neighbors (KNN) algorithm is employed to construct a hypergraph representation of the structural brain network. This hypergraph captures multi-dimensional interactions that extend beyond simple pairwise connections, allowing for a more detailed representation of the complex inter-regional relationships in the brain.

Following the construction of the hypergraph, HGNN+ (19) is applied for feature learning and information integration. The hypergraph convolution process mirrors the procedure used for the functional brain network, as indicated in Equation 3. After two layers of hypergraph convolution, the resulting vertex embeddings encode high-order structural features, which are used as the final representations of each brain region. These embeddings enable a more nuanced analysis of the structural brain network, particularly in understanding the structural alterations associated with ASD. This approach provides a rigorous framework for examining both local and global connectivity patterns within the brain, offering valuable insights into the structural mechanisms underlying ASD.

2.2.3.3 Cross-attention transformer for multimodal integration

To effectively fuse functional and structural brain network features, a cross-attention Transformer module is introduced. This module leverages the Transformer architecture to model long-range dependencies between multimodal features, using the structural embeddings after hypergraph convolution as the query (Q) and the functional embeddings as the key (K) and value (V), enabling the integration of both modalities.

First, the structural embedding matrix Xs and the functional embedding matrix Xf obtained from hypergraph convolution are projected into Q, K, and V representations as follows:

Qs=WqsXs, Kf=WkfXf, Vf=WvfXf,    (4)

where Wqs, Wkf, and Wvf are learnable weight matrices that linearly project the structural and functional embeddings. This step maps both sets of features into a shared feature space, preparing them for cross-attention.

Next, through the cross-attention mechanism, the query matrix Qs from the structural features attends to the key matrix Kf from the functional features, generating the attention weight matrix:

A=softmax(QsKfTdk),    (5)

where A represents the attention weight matrix, and dk is the dimensionality of the key, used for scaling. These attention weights are then applied to the value matrix Vf from the functional features to generate updated structural feature embeddings:

Z=AVf.    (6)

This process refines the structural embeddings by incorporating high-order relationships derived from the functional network, enabling a more comprehensive representation of brain activity.

2.2.3.4 Brain disease diagnosis

The learned feature representations from the cross-attention Transformer module are then fed into the output layer for classification. The output layer consists of a fully connected layer and a log_softmax activation function to facilitate the final classification prediction.

Let xi and bi represent the input and bias for the i-th hidden layer, respectively, while Wi denotes the weight matrix facilitating connections from the i-th to the i+1-th hidden layer. Subsequently, the activation of the i+1-th hidden layer is computed using the equation below:

Zi+1=f(Wixi+bi),    (7)

Zi+1 is the activated output of layer i+1.f denotes the activation function.

f(x)=max(0,x),    (8)

The ReLU activation function constrains its output to the range [0, ∞). For the final layer, a fully connected dense layer is paired with a log_softmax activation functionactivation function to execute the terminal classification predictions.

log-softmax(zi)=zi-log(jezj),    (9)

The log-softmax function, which uses Euler's number e as the base for the natural logarithm. In a binary classification setting, it provides log probabilities as the output. We utilize the Adam optimizer for the optimization process, setting a relatively low learning rate of 1 × 10−5. The negative log-likelihood loss function is utilized for the binary classification task, the loss function is defined as follows:

L=-log(py)    (10)

where py represents the probability of the correct class y.

3 Results and discussion

The proposed method is compared against four categories of methods:

• Single-modality-based baseline: SVM (20), MLP (21)

• Single-modality-based graph methods: GCN (10), GAT (11), and GraphSage (22).

• Single-modality-based hypergraph method: HGNN+ (19).

• Multi-modality-based methods: BrainNN (23) and MVGCN (24).

A three-fold cross-validation approach was utilized to evaluate each method, quantifying the accuracy of ASD disease classification predictions using metrics such as accuracy, sensitivity, specificity, and F1 score. The final results is given by mean ± standard error. Tables 1, 2 show the experimental results of ABIDE and ADNI, respectively.

Table 1
www.frontiersin.org

Table 1. The comparison results on the two ABIDE datasets.

Table 2
www.frontiersin.org

Table 2. The comparison results on the ADNI dataset.

3.1 Comparison with single-modality baseline methods

In Tables 1, 2, single-modality baseline methods include traditional machine learning approaches such as Support Vector Machines (SVM) and Multilayer Perceptron (MLP). While these methods are widely used for classification tasks, they are limited to features from a single modality and cannot capture the complex interactions within brain networks. Specifically, on the ABIDE-NYU dataset, SVM achieved an accuracy of 0.746, and MLP achieved 0.655, both lower than the accuracy of 0.799 achieved by our proposed HGTrans. Similarly, on the ABIDE-TCD dataset, SVM and MLP achieved 0.625 and 0.646, respectively, which are significantly lower than HGTrans's 0.749. These results indicate that single-modality baseline methods are insufficient for effectively addressing the complexity of ASD data. By integrating both fMRI and DTI data, HGTrans can capture more informative features from different perspectives of the brain network, leading to superior classification performance. This highlights the necessity and effectiveness of multimodal data fusion.

3.2 Comparison with graph-based methods

Graph-based methods, including GCN, GAT, and GraphSAGE, utilize the graph structure of brain networks to model relationships between regions of interest (ROIs). These methods can capture more complex spatial topological features than traditional single-modality methods. However, as shown in Table 1, on the ABIDE-NYU dataset, GCN achieved an accuracy of 0.618, GAT 0.617, and GraphSAGE 0.582, all significantly lower than HGTrans's 0.799. Similarly, on the ABIDE-TCD dataset, GCN, GAT, and GraphSAGE achieved accuracies of 0.625, 0.623, and 0.623, respectively, which are lower than HGTrans's 0.749. Although graph-based methods can capture the topological information within brain networks, they are limited to modeling pairwise relationships and cannot fully represent higher-order interactions between brain regions. In contrast, HGTrans leverages hypergraph modeling to capture more complex higher-order relationships in multimodal settings, which significantly improves classification performance over traditional graph methods.

3.3 Comparison with hypergraph-based methods

Hypergraph-based methods, such as HGNN+, extend the capabilities of graph models by capturing higher-order relationships between multiple brain regions through hypergraph structures. On the ABIDE-NYU dataset, HGNN+ achieved an accuracy of 0.707, and on the ABIDE-TCD dataset, it achieved 0.676, both close to but lower than HGTrans's 0.799 and 0.749, respectively. These results show that while hypergraph methods can capture more complex brain region interactions, performance remains limited when using single-modality data. HGTrans outperforms HGNN+ primarily due to its ability to not only capture higher-order spatial topological structures through hypergraphs but also effectively integrate functional and structural brain network features using cross-attention mechanisms. By jointly modeling multimodal data, HGTrans generates more robust embeddings, leading to superior performance compared to single-modality hypergraph methods. On the other hand, When we engage in cognitive activities such as reading, writing, and listening, multiple brain regions cooperate to complete the tasks (25, 26), rather than a single brain region or pairs of brain regions working independently. Traditional methods find it difficult to model such group high-order correlations. However, high-order correlation modeling and semantic computation based on hypergraphs can achieve high-order correlation-driven local brain region cooperative message passing, which is more efficient than traditional graph neural networks and contains richer information. Therefore, for brain disease diagnosis tasks, the hypergraph computation model can provide more abundant semantic information, thereby improving diagnostic performance.

3.4 Comparison with multimodal methods

Multimodal methods, such as MVGNN and BrainNN, integrate both fMRI and DTI data to capture complementary information from different brain modalities. As shown in Tables 1, 2, while these multimodal methods outperform single-modality and graph-based methods, HGTrans still achieves the highest accuracy across both datasets. On the ABIDE-NYU dataset, MVGNN achieved an accuracy of 0.748, and BrainNN 0.688, both lower than HGTrans's 0.799. On the ABIDE-TCD dataset, MVGNN, and BrainNN achieved accuracies of 0.698 and 0.672, respectively, also lower than HGTrans's 0.749. HGTrans's advantage lies in its ability to not only fuse multimodal data but also effectively capture the complex interactions between functional and structural brain networks through hypergraph structures and cross-attention mechanisms. This mechanism allows the model to fully leverage the relationships between functional and structural brain networks, resulting in more expressive features and higher classification accuracy. There are also some domain adaption methods (27, 28) that can be used to transfer knowledge between structural and functional brain imaging. Although these cross-modal information transfer methods can achieve inference with only one modality in the testing phase, the performance is greatly limited by the lack of shared labels to guide the cross modality fusion.

4 Conclusion

In this study, we proposed a hypergraph Transformer-based approach to model and compute high-order associations between functional and structural brain networks. Our method effectively integrates multimodal data from fMRI and DTI, overcoming the limitations of traditional graph methods that can only capture pairwise relationships. By leveraging hypergraphs to model complex higher-order interactions and employing the Transformer architecture for feature extraction and integration, our approach has demonstrated significant improvements in brain disease diagnosis. The experimental results on the ABIDE and ADNI datasets show that the proposed method consistently outperforms existing approaches, confirming its effectiveness in enhancing the accuracy of brain disease classification. The introduction of a hypergraph-based model and the application of Transformer networks provide a robust framework for multimodal brain network analysis, advancing our understanding of the relationship between structural and functional connectivity.

Data availability statement

The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.

Author contributions

XH: Conceptualization, Formal analysis, Methodology, Writing – original draft, Writing – review & editing. JF: Formal analysis, Investigation, Methodology, Writing – original draft, Writing – review & editing. HX: Formal analysis, Writing – original draft, Writing – review & editing. SD: Project administration, Supervision, Writing – original draft, Writing – review & editing. JL: Funding acquisition, Project administration, Supervision, Writing – original draft, Writing – review & editing.

Funding

The author(s) declare financial support was received for the research, authorship, and/or publication of this article. This work was supported by the Shenzhen Fund for Guangdong Provincial High-level Clinical Key Specialties (No.SZGSP013), the Science and Technology Planning Project of Shenzhen Municipality (20210617155253001), and the National Natural Science Foundation of China (62401330).

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Footnotes

References

1. Ju R, Hu C, Li Q. Early diagnosis of Alzheimer's disease based on resting-state brain networks and deep learning. IEEE/ACM Trans Comput Biol Bioinform. (2017) 16:244–57. doi: 10.1109/TCBB.2017.2776910

PubMed Abstract | Crossref Full Text | Google Scholar

2. Katti G, Ara SA, Shireen A. Magnetic resonance imaging (MRI)-a review. Int J Dental Clin. (2011) 3:65–70.

Google Scholar

3. Assaf Y, Pasternak O. Diffusion tensor imaging (DTI)-based white matter mapping in brain research: a review. J Molec Neurosci. (2008) 34:51–61. doi: 10.1007/s12031-007-0029-0

PubMed Abstract | Crossref Full Text | Google Scholar

4. Van Den Heuvel MP, Pol HEH. Exploring the brain network: a review on resting-state fMRI functional connectivity. Eur Neuropsychopharmacol. (2010) 20:519–34. doi: 10.1016/j.euroneuro.2010.03.008

PubMed Abstract | Crossref Full Text | Google Scholar

5. Yang Y, Ye C, Guo X, Wu T, Xiang Y, Ma T. Mapping multi-modal brain connectome for brain disorder diagnosis via cross-modal mutual learning. IEEE Trans Med Imaging. (2023). doi: 10.1109/TMI.2023.3294967

PubMed Abstract | Crossref Full Text | Google Scholar

6. Yin J, Hu Z, Du X. Uncertainty quantification with mixed data by hybrid convolutional neural network for additive manufacturing. ASCE-ASME J Risk and Uncert in Engrg Sys Part B Mech Engr. (2024) 10:031103. doi: 10.1115/1.4065444

Crossref Full Text | Google Scholar

7. Han X, Xue R, Du S, Gao Y. Inter-intra high-order brain network for ASD diagnosis via functional MRIs. In: Proceedings of the International Conference on Medical Image Computing and Computer Assisted Intervention. (2024). doi: 10.1007/978-3-031-72069-7_21

Crossref Full Text | Google Scholar

8. Xiao L, Wang J, Kassani PH, Zhang Y, Bai Y, Stephen JM, et al. Multi-hypergraph learning-based brain functional connectivity analysis in fMRI data. IEEE Trans Med Imaging. (2019) 39:1746–58. doi: 10.1109/TMI.2019.2957097

PubMed Abstract | Crossref Full Text | Google Scholar

9. Xiao L, Stephen JM, Wilson TW, Calhoun VD, Wang YP. A hypergraph learning method for brain functional connectivity network construction from fMRI data. In: Medical Imaging 2020: Biomedical Applications in Molecular, Structural, and Functional Imaging. (2020). p. 254–259. doi: 10.1117/12.2543089

Crossref Full Text | Google Scholar

10. Kipf TN, Welling M. Semi-supervised classification with graph convolutional networks. arXiv preprint arXiv:160902907. (2016).

Google Scholar

11. Velickovic P, Cucurull G, Casanova A, Romero A, Lio P, Bengio Y, et al. Graph attention networks. Stat. (2017) 1050:10–48550.

Google Scholar

12. Lee MH, Smyser CD, Shimony JS. Resting-state fMRI: a review of methods and clinical applications. Am J Neuroradiol. (2013) 34:1866–72. doi: 10.3174/ajnr.A3263

PubMed Abstract | Crossref Full Text | Google Scholar

13. Cao P, Liu X, Liu H, Yang J, Zhao D, Huang M, et al. Generalized fused group lasso regularized multi-task feature learning for predicting cognitive outcomes in Alzheimer's disease. Comput Methods Programs Biomed. (2018) 162:19–45. doi: 10.1016/j.cmpb.2018.04.028

PubMed Abstract | Crossref Full Text | Google Scholar

14. Gao Y, Ji S, Han X, Dai Q. Hypergraph computation. Engineering. (2024) 40:188–201. doi: 10.1016/j.eng.2024.04.017

Crossref Full Text | Google Scholar

15. Wang J, Li H, Qu G, Cecil KM, Dillman JR, Parikh NA, et al. Dynamic weighted hypergraph convolutional network for brain functional connectome analysis. Med Image Anal. (2023) 87:102828. doi: 10.1016/j.media.2023.102828

PubMed Abstract | Crossref Full Text | Google Scholar

16. Craddock C, Benhajali Y, Chu C, Chouinard F, Evans A, Jakab A, et al. The neuro bureau preprocessing initiative: open sharing of preprocessed neuroimaging data and derivatives. Front Neuroinform. (2013) 7:5. doi: 10.3389/conf.fninf.2013.09.00041

Crossref Full Text | Google Scholar

17. Jack CR Jr, Bernstein MA, Fox NC, Thompson P, Alexander G, Harvey D, et al. The Alzheimer's disease neuroimaging initiative (ADNI): MRI methods. J Magnetic Reson Imag. (2008) 27:685–91. doi: 10.1002/jmri.21049

PubMed Abstract | Crossref Full Text | Google Scholar

18. Tzourio-Mazoyer N, Landeau B, Papathanassiou D, Crivello F, Etard O, Delcroix N, et al. Automated anatomical labeling of activations in SPM using a macroscopic anatomical parcellation of the MNI MRI single-subject brain. Neuroimage. (2002) 15:273–89. doi: 10.1006/nimg.2001.0978

PubMed Abstract | Crossref Full Text | Google Scholar

19. Gao Y, Feng Y, Ji S, Ji R. HGNN+: general hypergraph neural networks. IEEE Trans Pattern Anal Mach Intell. (2022) 45:3181–99. doi: 10.1109/TPAMI.2022.3182052

PubMed Abstract | Crossref Full Text | Google Scholar

20. Hearst MA, Dumais ST, Osuna E, Platt J, Scholkopf B. Support vector machines. IEEE Intell Syst Their Applic. (1998) 13:18–28. doi: 10.1109/5254.708428

Crossref Full Text | Google Scholar

21. Subah FZ, Deb K, Dhar PK, Koshiba T, A. deep learning approach to predict autism spectrum disorder using multisite resting-state fMRI. Appl Sci. (2021) 11:3636. doi: 10.3390/app11083636

PubMed Abstract | Crossref Full Text | Google Scholar

22. Hamilton W, Ying Z, Leskovec J. Inductive representation learning on large graphs. In: Advances in Neural Information Processing Systems. (2017). p. 30.

Google Scholar

23. Zhu Y, Cui H, He L, Sun L, Yang C. Joint embedding of structural and functional brain networks with graph neural networks for mental illness diagnosis. In: 2022 44th Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), IEEE (2022). p. 272–276. doi: 10.1109/EMBC48229.2022.9871118

PubMed Abstract | Crossref Full Text | Google Scholar

24. Fu H, Huang F, Liu X, Qiu Y, Zhang W. MVGCN data integration through multi-view graph convolutional network for predicting links in biomedical bipartite networks. Bioinformatics. (2022) 38:426–34. doi: 10.1093/bioinformatics/btab651

PubMed Abstract | Crossref Full Text | Google Scholar

25. Shinn M, Hu A, Turner L, Noble S, Preller KH, Ji JL, et al. Functional brain networks reflect spatial and temporal autocorrelation. Nat Neurosci. (2023) 26:867–78. doi: 10.1038/s41593-023-01299-3

PubMed Abstract | Crossref Full Text | Google Scholar

26. Park HJ, Friston K. Structural and functional brain networks: from connections to cognition. Science. (2013) 342:1238411. doi: 10.1126/science.1238411

PubMed Abstract | Crossref Full Text | Google Scholar

27. Song R, Cao P, Wen G, Zhao P, Huang Z, Zhang X, et al. BrainDAS: structure-aware domain adaptation network for multi-site brain network analysis. Med Image Anal. (2024) 96:103211. doi: 10.1016/j.media.2024.103211

PubMed Abstract | Crossref Full Text | Google Scholar

28. Xu B, Yin J, Lian C, Su Y, Zeng Z. Low-rank optimal transport for robust domain adaptation. IEEE/CAA J Autom Sinica. (2024) 11:1667–80. doi: 10.1109/JAS.2024.124344

Crossref Full Text | Google Scholar

Keywords: hypergraph computation, brain network, high-order correlation, brain disease diagnosis, transformer

Citation: Han X, Feng J, Xu H, Du S and Li J (2024) A hypergraph transformer method for brain disease diagnosis. Front. Med. 11:1496573. doi: 10.3389/fmed.2024.1496573

Received: 14 September 2024; Accepted: 30 October 2024;
Published: 14 November 2024.

Edited by:

Donglin Di, Li Auto, China

Reviewed by:

Zhikang Xu, Shanxi University, China
Jianhua Yin, Wuhan University of Technology, China
Saisai Ding, Shanghai University, China

Copyright © 2024 Han, Feng, Xu, Du and Li. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Junchang Li, bGlqYy4yMDA1JiN4MDAwNDA7dHNpbmdodWEub3JnLmNu

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.