<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article article-type="research-article" dtd-version="2.3" xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Physiol.</journal-id>
<journal-title>Frontiers in Physiology</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Physiol.</abbrev-journal-title>
<issn pub-type="epub">1664-042X</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">1439607</article-id>
<article-id pub-id-type="doi">10.3389/fphys.2024.1439607</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Physiology</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Multi-branch fusion graph neural network based on multi-head attention for childhood seizure detection</article-title>
<alt-title alt-title-type="left-running-head">Li et al.</alt-title>
<alt-title alt-title-type="right-running-head">
<ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3389/fphys.2024.1439607">10.3389/fphys.2024.1439607</ext-link>
</alt-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name>
<surname>Li</surname>
<given-names>Yang</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/2750683/overview"/>
<role content-type="https://credit.niso.org/contributor-roles/data-curation/"/>
<role content-type="https://credit.niso.org/contributor-roles/investigation/"/>
<role content-type="https://credit.niso.org/contributor-roles/methodology/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-original-draft/"/>
<role content-type="https://credit.niso.org/contributor-roles/Writing - review &#x26; editing/"/>
<role content-type="https://credit.niso.org/contributor-roles/validation/"/>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Yang</surname>
<given-names>Yang</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="corresp" rid="c001">&#x2a;</xref>
<role content-type="https://credit.niso.org/contributor-roles/funding-acquisition/"/>
<role content-type="https://credit.niso.org/contributor-roles/methodology/"/>
<role content-type="https://credit.niso.org/contributor-roles/resources/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-original-draft/"/>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Song</surname>
<given-names>Shangling</given-names>
</name>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<xref ref-type="corresp" rid="c001">&#x2a;</xref>
<role content-type="https://credit.niso.org/contributor-roles/investigation/"/>
<role content-type="https://credit.niso.org/contributor-roles/resources/"/>
<role content-type="https://credit.niso.org/contributor-roles/supervision/"/>
<role content-type="https://credit.niso.org/contributor-roles/Writing - review &#x26; editing/"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Wang</surname>
<given-names>Hongjun</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<role content-type="https://credit.niso.org/contributor-roles/funding-acquisition/"/>
<role content-type="https://credit.niso.org/contributor-roles/methodology/"/>
<role content-type="https://credit.niso.org/contributor-roles/Writing - review &#x26; editing/"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Sun</surname>
<given-names>Mengzhou</given-names>
</name>
<xref ref-type="aff" rid="aff3">
<sup>3</sup>
</xref>
<role content-type="https://credit.niso.org/contributor-roles/investigation/"/>
<role content-type="https://credit.niso.org/contributor-roles/methodology/"/>
<role content-type="https://credit.niso.org/contributor-roles/Writing - review &#x26; editing/"/>
<role content-type="https://credit.niso.org/contributor-roles/validation/"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Liang</surname>
<given-names>Xiaoyun</given-names>
</name>
<xref ref-type="aff" rid="aff4">
<sup>4</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/2079122/overview"/>
<role content-type="https://credit.niso.org/contributor-roles/investigation/"/>
<role content-type="https://credit.niso.org/contributor-roles/methodology/"/>
<role content-type="https://credit.niso.org/contributor-roles/Writing - review &#x26; editing/"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Zhao</surname>
<given-names>Penghui</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<role content-type="https://credit.niso.org/contributor-roles/methodology/"/>
<role content-type="https://credit.niso.org/contributor-roles/Writing - review &#x26; editing/"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Wang</surname>
<given-names>Baiyang</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/1859375/overview"/>
<role content-type="https://credit.niso.org/contributor-roles/methodology/"/>
<role content-type="https://credit.niso.org/contributor-roles/Writing - review &#x26; editing/"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Wang</surname>
<given-names>Na</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<role content-type="https://credit.niso.org/contributor-roles/methodology/"/>
<role content-type="https://credit.niso.org/contributor-roles/Writing - review &#x26; editing/"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Sun</surname>
<given-names>Qiyue</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<role content-type="https://credit.niso.org/contributor-roles/methodology/"/>
<role content-type="https://credit.niso.org/contributor-roles/Writing - review &#x26; editing/"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Han</surname>
<given-names>Zijuan</given-names>
</name>
<xref ref-type="aff" rid="aff5">
<sup>5</sup>
</xref>
<role content-type="https://credit.niso.org/contributor-roles/methodology/"/>
<role content-type="https://credit.niso.org/contributor-roles/Writing - review &#x26; editing/"/>
</contrib>
</contrib-group>
<aff id="aff1">
<sup>1</sup>
<institution>School of Information Science and Engineering</institution>, <institution>Shandong University</institution>, <addr-line>Qingdao</addr-line>, <country>China</country>
</aff>
<aff id="aff2">
<sup>2</sup>
<institution>Bidding Office</institution>, <institution>The Second Hospital of Shandong University</institution>, <addr-line>Jinan</addr-line>, <country>China</country>
</aff>
<aff id="aff3">
<sup>3</sup>
<institution>Institute of Research and Clinical Innovations</institution>, <institution>Neusoft Medical Systems Co., Ltd.</institution>, <addr-line>Beijing</addr-line>, <country>China</country>
</aff>
<aff id="aff4">
<sup>4</sup>
<institution>Institute of Research and Clinical Innovations</institution>, <institution>Neusoft Medical Systems Co., Ltd.</institution>, <addr-line>Shanghai</addr-line>, <country>China</country>
</aff>
<aff id="aff5">
<sup>5</sup>
<institution>Center for Optics Research and Engineering</institution>, <institution>Shandong University</institution>, <addr-line>Qingdao</addr-line>, <country>China</country>
</aff>
<author-notes>
<fn fn-type="edited-by">
<p>
<bold>Edited by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/75268/overview">Ahsan H. Khandoker</ext-link>, Khalifa University, United Arab Emirates</p>
</fn>
<fn fn-type="edited-by">
<p>
<bold>Reviewed by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/101627/overview">Qingyun Wang</ext-link>, Beihang University, China</p>
<p>
<ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/2054538/overview">Xu Huang</ext-link>, Nanjing University of Science and Technology, China</p>
<p>
<ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/2667833/overview">Omar Farooq</ext-link>, Aligarh Muslim University, India</p>
</fn>
<corresp id="c001">&#x2a;Correspondence: Yang Yang, <email>yyang@sdu.edu.cn</email>; Shangling Song, <email>690496013@qq.com</email>
</corresp>
</author-notes>
<pub-date pub-type="epub">
<day>31</day>
<month>10</month>
<year>2024</year>
</pub-date>
<pub-date pub-type="collection">
<year>2024</year>
</pub-date>
<volume>15</volume>
<elocation-id>1439607</elocation-id>
<history>
<date date-type="received">
<day>28</day>
<month>05</month>
<year>2024</year>
</date>
<date date-type="accepted">
<day>07</day>
<month>10</month>
<year>2024</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#xa9; 2024 Li, Yang, Song, Wang, Sun, Liang, Zhao, Wang, Wang, Sun and Han.</copyright-statement>
<copyright-year>2024</copyright-year>
<copyright-holder>Li, Yang, Song, Wang, Sun, Liang, Zhao, Wang, Wang, Sun and Han</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/">
<p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p>
</license>
</permissions>
<abstract>
<p>The most common manifestation of neurological disorders in children is the occurrence of epileptic seizures. In this study, we propose a multi-branch graph convolutional network (MGCNA) framework with a multi-head attention mechanism for detecting seizures in children. The MGCNA framework extracts effective and reliable features from high-dimensional data, particularly by exploring the relationships between EEG features and electrodes and considering the spatial and temporal dependencies in epileptic brains. This method incorporates three graph learning approaches to systematically assess the connectivity and synchronization of multi-channel EEG signals. The multi-branch graph convolutional network is employed to dynamically learn temporal correlations and spatial topological structures. Utilizing the multi-head attention mechanism to process multi-branch graph features further enhances the capability to handle local features. Experimental results demonstrate that the MGCNA exhibits superior performance on patient-specific and patient-independent experiments. Our end-to-end model for automatic detection of epileptic seizures could be employed to assist in clinical decision-making.</p>
</abstract>
<kwd-group>
<kwd>childhood seizure detection</kwd>
<kwd>graph convolutional network</kwd>
<kwd>adjacency matrix</kwd>
<kwd>EEG</kwd>
<kwd>multi-head attention</kwd>
</kwd-group>
<custom-meta-wrap>
<custom-meta>
<meta-name>section-at-acceptance</meta-name>
<meta-value>Computational Physiology and Medicine</meta-value>
</custom-meta>
</custom-meta-wrap>
</article-meta>
</front>
<body>
<sec id="s1">
<title>1 Introduction</title>
<p>Epilepsy is a neurological disorder characterized by abnormal synchronous discharges of neurons. Childhood seizures carry a risk for the presence of cognitive impairment and behavioral disorders (<xref ref-type="bibr" rid="B42">Rennie et al., 2004</xref>). Therefore, accurate detection of seizures in children is of paramount importance for determining the best treatment plans and preventing adverse conditions. The diagnosis of epilepsy typically relies on the analysis of electroencephalogram (EEG), which is abnormal in the majority of patients. However, this task requires highly experienced experts who must invest a significant amount of time and effort in inspecting lengthy EEG recordings (<xref ref-type="bibr" rid="B26">Kannathal et al., 2005</xref>). This process is also susceptible to the subjective influence of epilepsy experts.The occurrence of epileptic seizures significantly impacts the physical development of children, emphasizing the necessity for early detection and intervention in pediatric epilepsy to mitigate its effects. Therefore, designing a reliable childhood seizure detection model can facilitate the automation of epilepsy diagnosis, holding significant importance in enhancing the quality of life for children.</p>
<p>There is a significant need to expand the utilization of machine learning, particularly within the emerging realm of deep learning, for automating the detection of epilepsy through EEG signal classification. Among the extensively explored techniques for EEG seizure detection, feature extraction rooted in machine learning stands as one of the most prominent approachs. For instance, multiscale entropy were extracted using extreme learning machines (<xref ref-type="bibr" rid="B10">Cui et al., 2017</xref>), nonlinear features were extracted using Gradient Boosting Decision Trees (GBDT) (<xref ref-type="bibr" rid="B59">Xu et al., 2022</xref>), various features were extracted using Empirical Mode Decomposition (EMD) (<xref ref-type="bibr" rid="B44">Singh et al., 2019</xref>), multi-scale features were extracted using wavelet transformation (<xref ref-type="bibr" rid="B60">Zhang et al., 2010</xref>), and the Comprehensive Representation of K Nearest Neighbors (CRMKNN) (<xref ref-type="bibr" rid="B38">Na et al., 2021</xref>) approach is proposed for epilepsy diagnosis. The traditional machine learning algorithms often struggle to achieve automatic detection of epilepsy, and experimental results are influenced by empirical parameters, making it difficult to stabilize the algorithm&#x2019;s performance. In essence, time series data is nonlinear and dynamic, making it challenging for traditional machine learning algorithms to effectively capture these complex signal characteristics. Moreover, there are large variations between different patients. Therefore, traditional machine learning faces challenges in learning the hidden features of EEG signals and lacks generalization.</p>
<p>Compared to traditional machine learning, neural networks, as a more promising algorithm with greater capacity for learning from complex data, have been applied to various research fields (<xref ref-type="bibr" rid="B61">Zhao P. et al., 2022</xref>; <xref ref-type="bibr" rid="B47">Sun and Yang, 2023</xref>; <xref ref-type="bibr" rid="B6">Chen J. et al., 2024</xref>; <xref ref-type="bibr" rid="B2">Abu and Diamant, 2023</xref>). There have been numerous advancements in the detection of epilepsy EEG signals as well (<xref ref-type="bibr" rid="B51">Wang et al., 2023a</xref>; <xref ref-type="bibr" rid="B64">Zhao et al., 2023b</xref>; <xref ref-type="bibr" rid="B20">He et al., 2022</xref>; <xref ref-type="bibr" rid="B58">Xiao et al., 2024</xref>). A novel deep network called Two-Stream 3-D Attention Module (TSA3-D) (<xref ref-type="bibr" rid="B4">Cao et al., 2022</xref>) was introduced to leverage the multichannel time-frequency and frequency-space features of interictal EEGs for epilepsy classification. In (<xref ref-type="bibr" rid="B18">Feng et al., 2022</xref>) a 3D deep network combined with residual attention modules was proposed to explore the spatial and time-frequency features of multi-channel EEG. In (<xref ref-type="bibr" rid="B12">Cui et al., 2022b</xref>), a fusion model based on transfer learning and time-frequency features was proposed for the effective detection of childhood epilepsy syndrome. In (<xref ref-type="bibr" rid="B11">Cui et al., 2022a</xref>), an analysis of the correlation between time-frequency features and EEG signals was conducted, and a childhood epilepsy syndrome classification model based on transfer networks was proposed.</p>
<p>Attention mechanisms have gained widespread applications in the field of signal recognition (<xref ref-type="bibr" rid="B64">Zhao et al., 2023b</xref>; <xref ref-type="bibr" rid="B32">Lian and Xu, 2023</xref>; <xref ref-type="bibr" rid="B39">Peh et al., 2023</xref>; <xref ref-type="bibr" rid="B40">Qiu et al., 2023</xref>; <xref ref-type="bibr" rid="B55">Wang Z. et al., 2023</xref>; <xref ref-type="bibr" rid="B33">Liu et al., 2023</xref>), emerging as a pivotal technology attracting significant attention and in-depth exploration within the realm of deep learning. Many researchers have integrated attention mechanisms with neural network models, resulting in the creation of a series of innovative models. The emergence of these models has introduced new possibilities for enhancing the accuracy and efficiency of EEG signal recognition, thus steering the direction of development in this field. <xref ref-type="bibr" rid="B15">Ding et al. (2023)</xref> proposed a novel seizure prediction model that utilizes a CNN to automatically capture features from EEG signals. This model combines multiple head attention mechanisms to identify relevant information within these features for the recognition of EEG signal segments. <xref ref-type="bibr" rid="B14">Deng et al. (2023)</xref> introduced a novel hybrid vision transformer (HViT) model that could enhance the multi-head attention mechanism by augmenting the capability of convolution to process local features, thereby achieving data uncertainty learning. Zhao et al. (<xref ref-type="bibr" rid="B62">Zhao X. et al., 2022</xref>) proposed a recommendation detector based on multi-head attention mechanism, utilized for detecting pathological high-frequency oscillations (HFOs) associated with epilepsy to locate the epileptogenic zones. The attention mechanism aids the network in capturing dependencies among features and enhancing the model&#x2019;s sensitivity to local information. In the field of seizure detection, the potential of attention mechanism remains to be further explored.</p>
<p>In the epilepsy detection process using deep learning, EEG signals are represented as two-dimensional signals, considering only channel-based features and disregarding information about the physical distribution of channels. The electrode distribution in EEG exhibits a non-Euclidean topological structure, which can lead to the loss of connectivity information between brain functional regions, neglecting the long-term interdependencies among EEG signals from different channels. The graph convolutional network (GCN) algorithms can effectively leverage the implicit graph representation information within EEG signals. GCN algorithms utilize graph structures and update graph representations through node aggregation. GCN has been widely applied in numerous EEG signals, which have demonstrated excellent performance, such as emotion recognition (<xref ref-type="bibr" rid="B45">Song et al., 2021</xref>; <xref ref-type="bibr" rid="B34">Liu et al., 2022</xref>; <xref ref-type="bibr" rid="B7">Chen Y. et al., 2024</xref>; <xref ref-type="bibr" rid="B31">Li Y. et al., 2022</xref>), Alzheimer&#x2019;s disease (<xref ref-type="bibr" rid="B35">Lopez et al., 2023</xref>), automatic seizure detection (<xref ref-type="bibr" rid="B49">Wagh and Varatharajah, 2020</xref>; <xref ref-type="bibr" rid="B37">Meng et al., 2022</xref>; <xref ref-type="bibr" rid="B21">Ho and Armanfard, 2023</xref>), driver state monitoring (<xref ref-type="bibr" rid="B25">Kalaganis et al., 2020</xref>), motor imagery (<xref ref-type="bibr" rid="B3">Cai et al., 2022</xref>), and sleep stage classification (<xref ref-type="bibr" rid="B29">Li M. et al., 2022</xref>; <xref ref-type="bibr" rid="B24">Jia et al., 2020</xref>; <xref ref-type="bibr" rid="B28">Lee et al., 2024</xref>; <xref ref-type="bibr" rid="B22">Ji et al., 2022</xref>; <xref ref-type="bibr" rid="B23">Jia et al., 2021</xref>). <xref ref-type="bibr" rid="B54">Wang et al. (2022)</xref> proposed a spatiotemporal graph attention network (STGAT) based on phase locking value (PLV) to extract spatial and functional connectivity information. <xref ref-type="bibr" rid="B41">Raeisi et al. (2022)</xref> constructed graph representations using three different types of spatial information and assessed the performance of neonatal seizure detection. <xref ref-type="bibr" rid="B20">He et al. (2022)</xref> utilized the graph attention network (GAT) to extract spatial features and employed a bi-directional long short-term memory network (BiLSTM) to capture temporal relationships before and after the current time frame for epilepsy detection.</p>
<p>The integration of attention mechanisms with GCN has proven effective in enhancing model performance on graph-structured data (<xref ref-type="bibr" rid="B57">Wu et al., 2024</xref>; <xref ref-type="bibr" rid="B30">Li et al., 2023</xref>; <xref ref-type="bibr" rid="B8">Cheng et al., 2023</xref>; <xref ref-type="bibr" rid="B16">Dong et al., 2022</xref>; <xref ref-type="bibr" rid="B53">Wang Y. et al., 2023</xref>; <xref ref-type="bibr" rid="B50">2020</xref>; <xref ref-type="bibr" rid="B19">Grattarola et al., 2022</xref>). Attention helps highlight important nodes or features, improving the ability of GCN to capture both global and local relationships, which is crucial for EEG classification. The dynamic temporal graph convolutional network (DTGCN) (<xref ref-type="bibr" rid="B57">Wu et al., 2024</xref>) is proposed for seizure detection and classification, incorporating a seizure attention layer to capture the distribution patterns of epilepsy and a graph structure learning layer to represent the dynamically evolving graph structure in the data. A spatiotemporal hypergraph convolutional network (STHGCN) (<xref ref-type="bibr" rid="B30">Li et al., 2023</xref>) is designed to capture higher-order relationships in EEG recordings, construct feature hypergraphs across the spectral, spatial, and temporal domains to focus on EEG channel correlations and dynamic temporal relationships, and integrate self-attention mechanisms to initialize and update relationships within EEG sequences. A hybrid network (<xref ref-type="bibr" rid="B8">Cheng et al., 2023</xref>) is proposed, consisting of a Dynamic Graph Convolution (DGC) module and a Temporal Self-Attention Representation (TSAR) module. This network simultaneously integrates representative knowledge of spatial topology and temporal context into the EEG emotion recognition task. In summary, the integration of attention mechanisms with GCN allows for more effective feature representation by dynamically weighting the importance of spatial and temporal relationships, thereby enhancing the ability of model to capture complex dependencies in EEG data for emotion recognition tasks.</p>
<p>During the training process, traditional GCN describes the dynamic process of epileptic seizures using a single graph representation. During seizures, various complex interactions of neural activities occur, including different types of brainwave changes and alterations in connectivity patterns between brain regions. Therefore, the dynamic process of epileptic seizures cannot be exhaustively represented by a single static graph. To comprehend and describe seizures, it is necessary to consider the changes in the temporal dimension, and the interactions between different brain regions in the spatial dimension. Therefore, this study introduces a multi-branch graph convolutional model with multi-head attention (MGCNA) for childhood seizure detection. Specifically, the MGCNA employs three graph representation approaches to characterize the feature representation of EEG data from multiple dimensions. By incorporating a multi-head attention mechanism, it combines spatial topological information from multi-channel electrodes with dynamic temporal information, enhancing the global contextual awareness and recognition capabilities of model. The major contributions of MGCNA can be summarized as follows:<list list-type="simple">
<list-item>
<p>&#x2022; A multi-branch GCN model is proposed. It utilizes Euclidean distance to capture spatial information between channels, employs Pearson correlation coefficient to gather functional connectivity information among channels, and supplements latent spatiotemporal correlations through a trainable adjacency matrix. The multi-channel EEG signals are modeled as graph signals, enabling the extraction of synchrony relationships within EEG signals.</p>
</list-item>
<list-item>
<p>&#x2022; By integrating graph signals with a multi-head attention mechanism, attention weights for graph features are obtained, and the hidden vector representation of graph signals is derived through the summation of weighted values.</p>
</list-item>
<list-item>
<p>&#x2022; We conduct patient-specific and patient-independent experiments to assist doctors in rapidly identifying the onset period in complex scenarios. The outcomes from these two experiments showcase the effectiveness of our method. Through comparison with other methods, the sensitivity, which is of utmost clinical concern, is the highest in two experiments.</p>
</list-item>
</list>
</p>
<p>The organization of the rest of this article is as follows. In <xref ref-type="sec" rid="s2">Section 2</xref>, the proposed seizure detection model is introduced in terms of the overall MGCNA structure, extraction of intrachannels features, multi-branch GCN, multi-head attention mechanism, and classifier module. <xref ref-type="sec" rid="s3">Section 3</xref> presents the implementation details, performance evaluation metrics and experimental results of the CHB-MIT dataset. <xref ref-type="sec" rid="s4">Section 4</xref> interprets the results and emphasizes the limitations of the current study, and the conclusion is provided in <xref ref-type="sec" rid="s5">Section 5</xref>.</p>
</sec>
<sec sec-type="methods" id="s2">
<title>2 Methods</title>
<p>In this section, we introduce our proposed MGCNA model for epilepsy classification using multi-channel EEG signal.</p>
<sec id="s2-1">
<title>2.1 General structure</title>
<p>
<xref ref-type="fig" rid="F1">Figure 1</xref> provides an overview of the overall process for classifying epileptic EEG signals using the MGCNA proposed in this article. First, intrachannels features are extracted from the raw EEG signals to create graph signals. Second, three different time-series-to-graph representation are employed and graph features are extracted through GCN. Third, a multi-head attention is utilized to learn dependencies of different graph features. Finally, the combined graph features are processed to output classification results.</p>
<fig id="F1" position="float">
<label>FIGURE 1</label>
<caption>
<p>The overall architecture of MGCNA. There are mainly four steps including extraction of intrachannels features module; Multi-branch GCN; Multi-head attention module and Classifier module.</p>
</caption>
<graphic xlink:href="fphys-15-1439607-g001.tif"/>
</fig>
</sec>
<sec id="s2-2">
<title>2.2 Extraction of intrachannels features</title>
<p>Convolution has been proven to be highly effective in capturing features of EEG signals. Inspired by the EEGWaveNet (<xref ref-type="bibr" rid="B48">Thuwajit et al., 2021</xref>) model, this module utilizes depthwise convolution to compress the input signals&#x2019; resolution in a channel-wise manner and capture features. Depthwise convolution does not cross information between channels and extracts features independently. This module comprises <inline-formula id="inf4">
<mml:math id="m4">
<mml:mrow>
<mml:mi>k</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> consecutive layers, where each layer captures valuable features within channels at half the scale of the resolution of the previous layer.</p>
<p>In this module, the input data <inline-formula id="inf5">
<mml:math id="m5">
<mml:mrow>
<mml:mi>X</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> are shaped as <inline-formula id="inf6">
<mml:math id="m6">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>n</mml:mi>
<mml:mo>,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:mi>C</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, where <inline-formula id="inf7">
<mml:math id="m7">
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf8">
<mml:math id="m8">
<mml:mrow>
<mml:mi>C</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> represent the number of channels and the number of sampling points, respectively. <inline-formula id="inf9">
<mml:math id="m9">
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> is set to 22, and <inline-formula id="inf10">
<mml:math id="m10">
<mml:mrow>
<mml:mi>C</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> is defined as <inline-formula id="inf11">
<mml:math id="m11">
<mml:mrow>
<mml:mi>f</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, where <inline-formula id="inf12">
<mml:math id="m12">
<mml:mrow>
<mml:mi>f</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> denotes the EEG&#x2019;s sampling frequency, and <inline-formula id="inf13">
<mml:math id="m13">
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> is the duration of each epoch&#x2019;s segmentation. In each layer, depthwise convolution is applied to each channel with a kernel size of (1, 2), a stride of 2, and no padding. The output size of each <inline-formula id="inf14">
<mml:math id="m14">
<mml:mrow>
<mml:mi>k</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>th layer is (<inline-formula id="inf15">
<mml:math id="m15">
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, 1, <inline-formula id="inf16">
<mml:math id="m16">
<mml:mrow>
<mml:mi>C</mml:mi>
<mml:mo>/</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>k</mml:mi>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula>).</p>
<p>The module of extraction of intrachannels features can be represented as:<disp-formula id="e1">
<mml:math id="m17">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi>F</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>k</mml:mi>
</mml:mrow>
</mml:msup>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>D</mml:mi>
<mml:mi>W</mml:mi>
<mml:mi>C</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>n</mml:mi>
<mml:mi>v</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi>F</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:math>
<label>(1)</label>
</disp-formula>where <inline-formula id="inf17">
<mml:math id="m18">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi>F</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> is the input signal <inline-formula id="inf18">
<mml:math id="m19">
<mml:mrow>
<mml:mi>X</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="double-struck">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mi>C</mml:mi>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf19">
<mml:math id="m20">
<mml:mrow>
<mml:mi>D</mml:mi>
<mml:mi>W</mml:mi>
<mml:mi>C</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>n</mml:mi>
<mml:mi>v</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mo>&#x22c5;</mml:mo>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> is the depthwise convolution. After reshape, the features extracted by this module <inline-formula id="inf20">
<mml:math id="m21">
<mml:mrow>
<mml:mi>N</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="double-struck">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>C</mml:mi>
<mml:mo>/</mml:mo>
<mml:mn>2</mml:mn>
<mml:mi>K</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> will be fed into the next module to obtain multi-scale graph features. <xref ref-type="fig" rid="F2">Figure 2</xref> depicts the process of extracting intrachannels features.</p>
<fig id="F2" position="float">
<label>FIGURE 2</label>
<caption>
<p>The data flow for extracting intrachannels features.</p>
</caption>
<graphic xlink:href="fphys-15-1439607-g002.tif"/>
</fig>
</sec>
<sec id="s2-3">
<title>2.3 Multi-branch graph convolutional network</title>
<p>To design an appropriate graph representation for EEG signals, it is necessary to construct a graph based on the physical distribution of channels and the spectral-temporal domain characteristics of the signals. Each channel was considered as a node and the connection of each node is characterized in terms of spatial position information, EEG signal similarity features, and a learnable graph representation to dynamically capture the topological relationship between EEG signal channels. These three types of graph signals are input into GCN, allowing us to leverage the structural characteristics of EEG signals to capture spatial dependencies between nodes. <xref ref-type="fig" rid="F3">Figure 3</xref> illustrates the overall process of multi-branch GCN. The graph convolution processes at the top, middle, and bottom are based on spatial distance graph representation, functional connectivity graph representation, and adaptive graph representation, respectively.</p>
<fig id="F3" position="float">
<label>FIGURE 3</label>
<caption>
<p>The pipeline of multi-branch graph convolutional network.</p>
</caption>
<graphic xlink:href="fphys-15-1439607-g003.tif"/>
</fig>
<sec id="s2-3-1">
<title>2.3.1 Graph generator</title>
<p>Graph representation consists of graph signal and graph. Graph signals are extracted from extraction of intrachannels features module. A graph is denoted as <inline-formula id="inf21">
<mml:math id="m22">
<mml:mrow>
<mml:mi>G</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">{</mml:mo>
<mml:mrow>
<mml:mi mathvariant="script">V</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>&#x3f5;</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold-italic">A</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">}</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>, where <inline-formula id="inf22">
<mml:math id="m23">
<mml:mrow>
<mml:mi mathvariant="script">V</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> is the set of nodes, <inline-formula id="inf23">
<mml:math id="m24">
<mml:mrow>
<mml:mi>&#x3f5;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> is the set of edges, and <inline-formula id="inf24">
<mml:math id="m25">
<mml:mrow>
<mml:mi mathvariant="bold-italic">A</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> is the adjacency matrix. Graph nodes represent EEG channels. The connections between nodes are controlled by the adjacency matrix <inline-formula id="inf25">
<mml:math id="m26">
<mml:mrow>
<mml:mi mathvariant="bold-italic">A</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>. To construct the adjacency matrix <inline-formula id="inf26">
<mml:math id="m27">
<mml:mrow>
<mml:mi mathvariant="bold-italic">A</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, this paper employs three graph learning methods: information based on spatial distance between nodes, measure based on functional connectivity, and adaptive graph learning methods.</p>
<sec id="s2-3-1-1">
<title>2.3.1.1 Spatial distance graph learning</title>
<p>Seizures are produced by abnormal nerve discharges in different areas of the cerebral cortex and cause significant changes in the EEG. And this abnormal discharge spreads and affects neighboring brain regions, making strong correlation between adjacent areas. So the spatial distance between nodes of EEG signals is used for graph representation in EEG signal analysis.</p>
<p>In this approach, we consider the electrode positions of the EEG as a three-dimensional grid model. To represent the spatial distance of EEG electrodes, we compute adjacency matrix <inline-formula id="inf27">
<mml:math id="m28">
<mml:mrow>
<mml:mi mathvariant="bold-italic">A</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> by applying Euclidean distance between pairs of channels in the bipolar montage. The distance between pairs of channels <inline-formula id="inf28">
<mml:math id="m29">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>v</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf29">
<mml:math id="m30">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>v</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>, which can be denoted as:<disp-formula id="e2">
<mml:math id="m31">
<mml:mrow>
<mml:mi>D</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>s</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>s</mml:mi>
<mml:mi>q</mml:mi>
<mml:mi>r</mml:mi>
<mml:mi>t</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo>&#x2b;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:math>
<label>(2)</label>
</disp-formula>where <inline-formula id="inf30">
<mml:math id="m32">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>v</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf31">
<mml:math id="m33">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>v</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> are the centers of the bipolar derivations. The spatial distances obtained from Euclidean distance are transformed into the adjacency matrix while taking into account the strength of connections between nodes. To calculate the adjacency matrix, we apply the following rules:<disp-formula id="e3">
<mml:math id="m34">
<mml:mrow>
<mml:mtable class="aligned">
<mml:mtr>
<mml:mtd columnalign="right">
<mml:mi mathvariant="bold-italic">A</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mfenced open="{" close="">
<mml:mrow>
<mml:mtable class="cases">
<mml:mtr>
<mml:mtd columnalign="left">
<mml:mi>e</mml:mi>
<mml:mi>x</mml:mi>
<mml:mi>p</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mi>D</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>s</mml:mi>
<mml:msubsup>
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msubsup>
</mml:mrow>
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi>&#x3b4;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:mfenced>
<mml:mo>,</mml:mo>
<mml:mspace width="1em"/>
</mml:mtd>
<mml:mtd columnalign="left">
<mml:mi>i</mml:mi>
<mml:mi>f</mml:mi>
<mml:mi>D</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>s</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2264;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd columnalign="left">
<mml:mn>0</mml:mn>
<mml:mo>,</mml:mo>
<mml:mspace width="1em"/>
</mml:mtd>
<mml:mtd columnalign="left">
<mml:mi>o</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>h</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>r</mml:mi>
<mml:mi>w</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>s</mml:mi>
<mml:mi>e</mml:mi>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
</mml:mfenced>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
</mml:math>
<label>(3)</label>
</disp-formula>where <inline-formula id="inf32">
<mml:math id="m35">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> is the threshold for sparsity and <inline-formula id="inf33">
<mml:math id="m36">
<mml:mrow>
<mml:mi>&#x3b4;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> is the standard deviation of the distances. This results in a universally undirected, weighted graph for all EEG graph representation.</p>
</sec>
<sec id="s2-3-1-2">
<title>2.3.1.2 Functional connectivity graph learning</title>
<p>The strength of association between channels varies during seizures. and the synchronization between brain regions can be assessed by calculating the connectivity between channels. Therefore, we investigate the relationship of information interaction between channels during seizures.</p>
<p>The functional connectivity between nodes of EEG signals is used for graph representation in EEG signal analysis. To represent the functional connectivity of EEG electrodes, we compute adjacency matrix <inline-formula id="inf34">
<mml:math id="m37">
<mml:mrow>
<mml:mi mathvariant="bold-italic">A</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> by applying Pearson correlation coefficient between the preprocessed signals in <inline-formula id="inf35">
<mml:math id="m38">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>v</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf36">
<mml:math id="m39">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>v</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>. The Pearson correlation coefficient between pairs of channels <inline-formula id="inf37">
<mml:math id="m40">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>v</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf38">
<mml:math id="m41">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>v</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, which can be denoted as:<disp-formula id="e4">
<mml:math id="m42">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3c1;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mo>&#x2211;</mml:mo>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>&#x304;</mml:mo>
</mml:mover>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>&#x304;</mml:mo>
</mml:mover>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mrow>
<mml:msqrt>
<mml:mrow>
<mml:mo>&#x2211;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>&#x304;</mml:mo>
</mml:mover>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo>&#x2211;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>&#x304;</mml:mo>
</mml:mover>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:msqrt>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
<label>(4)</label>
</disp-formula>where <inline-formula id="inf39">
<mml:math id="m43">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf40">
<mml:math id="m44">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> are graph signals extracted from extraction of intrachannels features module in <inline-formula id="inf41">
<mml:math id="m45">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>v</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf42">
<mml:math id="m46">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>v</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>. The computed correlation coefficients are normalized to the range [0, 1]. The functional connectivity obtained from Pearson correlation coefficient are transformed into the adjacency matrix while taking into account the strength of connections between nodes. To calculate the adjacency matrix, we apply the following rules:<disp-formula id="e5">
<mml:math id="m47">
<mml:mrow>
<mml:mtable class="aligned">
<mml:mtr>
<mml:mtd columnalign="right">
<mml:mi mathvariant="bold-italic">A</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mfenced open="{" close="">
<mml:mrow>
<mml:mtable class="cases">
<mml:mtr>
<mml:mtd columnalign="left">
<mml:msub>
<mml:mrow>
<mml:mi>&#x3c1;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mspace width="1em"/>
</mml:mtd>
<mml:mtd columnalign="left">
<mml:mi>i</mml:mi>
<mml:mi>f</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3c1;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2265;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd columnalign="left">
<mml:mn>0</mml:mn>
<mml:mo>,</mml:mo>
<mml:mspace width="1em"/>
</mml:mtd>
<mml:mtd columnalign="left">
<mml:mi>o</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>h</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>r</mml:mi>
<mml:mi>w</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>s</mml:mi>
<mml:mi>e</mml:mi>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
</mml:mfenced>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
</mml:math>
<label>(5)</label>
</disp-formula>where <inline-formula id="inf43">
<mml:math id="m48">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> is the threshold. This results in an undirected, weighted graph for all EEG graph representation.</p>
</sec>
<sec id="s2-3-1-3">
<title>2.3.1.3 Adaptive graph learning</title>
<p>The human brain possesses extremely complex structure and functionality, and analyzing it merely through spatial distance and functional connectivity graph structures does not fully capture its functions and behavioral manifestations. The neural network model aims to simulate the interconnection and information transfer between neurons in the human brain, enabling learning and reasoning of complex tasks. Therefore, we incorporate graph structures as part of the parameters in neural networks, training them to capture the coupling relationships with EEG signals.</p>
<p>The adaptive graph learning method can learn the intrinsic connections between EEG signals. In the adaptive graph learning method we employed, the adjacency matrix <inline-formula id="inf44">
<mml:math id="m49">
<mml:mrow>
<mml:mi mathvariant="bold-italic">A</mml:mi>
<mml:mn>3</mml:mn>
<mml:mo>&#x2208;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="double-struck">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula>, which characterizes the relationships between individual vertex nodes, is dynamically learned rather than predetermined. During the model training process, the adjacency matrix <inline-formula id="inf45">
<mml:math id="m50">
<mml:mrow>
<mml:mi mathvariant="bold-italic">A</mml:mi>
<mml:mn>3</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> of the adaptive graph representation is updated during training through back propagation.During the training process, <inline-formula id="inf46">
<mml:math id="m51">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold">A</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>3</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> will be constrained by the following formula:<disp-formula id="e6">
<mml:math id="m52">
<mml:mrow>
<mml:mi mathvariant="bold">A</mml:mi>
<mml:msup>
<mml:mrow>
<mml:mn>3</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>l</mml:mi>
</mml:mrow>
</mml:msup>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>R</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>L</mml:mi>
<mml:mi>U</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>B</mml:mi>
<mml:mi>N</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi mathvariant="bold">a</mml:mi>
<mml:msup>
<mml:mrow>
<mml:mn>3</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>l</mml:mi>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:math>
<label>(6)</label>
</disp-formula>where <inline-formula id="inf47">
<mml:math id="m53">
<mml:mrow>
<mml:mi mathvariant="bold">a</mml:mi>
<mml:msup>
<mml:mrow>
<mml:mn>3</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>l</mml:mi>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> is the adjacency matrix obtained in the <inline-formula id="inf48">
<mml:math id="m54">
<mml:mrow>
<mml:mi>l</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>-th layer, using BN and ReLU to prevent overfitting and improve stability. Due to the fact that the closer spatial connection between channels does not accurately reflect a closer functional relationship between them, and functional connectivity does not fully characterize the intrinsic relationship between channels (<xref ref-type="bibr" rid="B46">Song et al., 2018</xref>), we utilize the adaptive graph learning method to capture the inherent connections among EEG channels and enhance EEG recognition.</p>
</sec>
</sec>
<sec id="s2-3-2">
<title>2.3.2 Graph convolution</title>
<p>A general GCN model (<xref ref-type="bibr" rid="B27">Kipf and Welling, 2016</xref>; <xref ref-type="bibr" rid="B13">Defferrard et al., 2016</xref>) takes graph signals <inline-formula id="inf49">
<mml:math id="m55">
<mml:mrow>
<mml:mi>N</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="double-struck">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>C</mml:mi>
<mml:mo>/</mml:mo>
<mml:mn>2</mml:mn>
<mml:mi>K</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> and adjacency matrices <inline-formula id="inf50">
<mml:math id="m56">
<mml:mrow>
<mml:mi mathvariant="bold-italic">A</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="double-struck">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> as inputs and a graph feature matrix <inline-formula id="inf51">
<mml:math id="m57">
<mml:mrow>
<mml:mi>Z</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="double-struck">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>M</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mi>P</mml:mi>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> is generated through node aggregation. <inline-formula id="inf52">
<mml:math id="m58">
<mml:mrow>
<mml:mi>P</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> represents the dimension of the output feature vector for each node. GCN updates node features by aggregating features from neighboring nodes. The graph convolutional layer can be represented as follows:<disp-formula id="e7">
<mml:math id="m59">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi>H</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>l</mml:mi>
</mml:mrow>
</mml:msup>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>R</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>L</mml:mi>
<mml:mi>U</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi>D</mml:mi>
</mml:mrow>
<mml:mo>&#x303;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:msup>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi>A</mml:mi>
</mml:mrow>
<mml:mo>&#x303;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi>D</mml:mi>
</mml:mrow>
<mml:mo>&#x303;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:msup>
<mml:msup>
<mml:mrow>
<mml:mi>H</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>l</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msup>
<mml:msup>
<mml:mrow>
<mml:mi>&#x3b8;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>l</mml:mi>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:math>
<label>(7)</label>
</disp-formula>where <inline-formula id="inf53">
<mml:math id="m60">
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi>A</mml:mi>
</mml:mrow>
<mml:mo>&#x303;</mml:mo>
</mml:mover>
</mml:mrow>
</mml:math>
</inline-formula> represents adjacency matrix containing diagonal matrix. <inline-formula id="inf54">
<mml:math id="m61">
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi>D</mml:mi>
</mml:mrow>
<mml:mo>&#x303;</mml:mo>
</mml:mover>
</mml:mrow>
</mml:math>
</inline-formula> denotes the degree matrix for each node in the graph. <inline-formula id="inf55">
<mml:math id="m62">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi>H</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>l</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> represents the graph feature matrix of the previous output, and <inline-formula id="inf56">
<mml:math id="m63">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi>H</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo>&#x2208;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="double-struck">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>C</mml:mi>
<mml:mo>/</mml:mo>
<mml:mn>2</mml:mn>
<mml:mi>K</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> is the graph signal <inline-formula id="inf57">
<mml:math id="m64">
<mml:mrow>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>. <inline-formula id="inf58">
<mml:math id="m65">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi>&#x3b8;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>l</mml:mi>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> represents the matrix of learnable parameters.</p>
<p>Notably, the structures of the three GCN branches are identical, but their parameters are not shared. After each GCN, Batch Normalization <inline-formula id="inf59">
<mml:math id="m66">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>B</mml:mi>
<mml:mi>N</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> is applied to normalize the graph features at each layer, which helps improve the generalization capability of network. And it is presented as the following form:<disp-formula id="e8">
<mml:math id="m67">
<mml:mrow>
<mml:msubsup>
<mml:mrow>
<mml:mi mathvariant="script">Z</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>l</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>B</mml:mi>
<mml:mi>N</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msubsup>
<mml:mrow>
<mml:mi>H</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>l</mml:mi>
</mml:mrow>
</mml:msubsup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:math>
<label>(8)</label>
</disp-formula>where <inline-formula id="inf60">
<mml:math id="m68">
<mml:mrow>
<mml:msubsup>
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>l</mml:mi>
</mml:mrow>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula> is the input of the next graph convolution layer. <inline-formula id="inf61">
<mml:math id="m69">
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> indicated as the <inline-formula id="inf62">
<mml:math id="m70">
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>th branch of the GCN. The final output of multi-branch GCN is expressed as the following formula:<disp-formula id="e9">
<mml:math id="m71">
<mml:mrow>
<mml:mi mathvariant="script">Z</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mfenced open="[" close="]">
<mml:mrow>
<mml:msubsup>
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>m</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:mo>,</mml:mo>
<mml:msubsup>
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>m</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:mo>,</mml:mo>
<mml:msubsup>
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>3</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>m</mml:mi>
</mml:mrow>
</mml:msubsup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:math>
<label>(9)</label>
</disp-formula>where <inline-formula id="inf63">
<mml:math id="m72">
<mml:mrow>
<mml:mi>m</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> represents the number of layers of the GCN, and <inline-formula id="inf64">
<mml:math id="m73">
<mml:mrow>
<mml:mo stretchy="false">[</mml:mo>
<mml:mrow>
<mml:mo>&#x22c5;</mml:mo>
</mml:mrow>
<mml:mo stretchy="false">]</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> represents concatenation.</p>
</sec>
</sec>
<sec id="s2-4">
<title>2.4 Multi-head attention mechanism</title>
<p>The attention mechanism, which is a deep learning model, emulates the pivotal information and critical elements that individuals concentrate on during observations. The multi-head attention mechanism makes the output of attention to incorporate encoding from different spatial locations, thereby enhancing the expressive power of the model. In standard multi-head attention with input matrix <inline-formula id="inf65">
<mml:math id="m74">
<mml:mrow>
<mml:mi>Z</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="double-struck">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>m</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mi>n</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mi>D</mml:mi>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula>, the self-attention for each head are computed according to the following formula:<disp-formula id="e10">
<mml:math id="m75">
<mml:mrow>
<mml:mi>H</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>d</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi mathvariant="bold-italic">Q</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold-italic">K</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold-italic">V</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>S</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>f</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>m</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>x</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="bold-italic">Q</mml:mi>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="bold-italic">K</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>T</mml:mi>
</mml:mrow>
</mml:msup>
</mml:mrow>
<mml:mrow>
<mml:msqrt>
<mml:mrow>
<mml:mi>d</mml:mi>
</mml:mrow>
</mml:msqrt>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:mfenced>
<mml:mi mathvariant="bold-italic">V</mml:mi>
</mml:mrow>
</mml:math>
<label>(10)</label>
</disp-formula>where <inline-formula id="inf66">
<mml:math id="m76">
<mml:mrow>
<mml:mi mathvariant="bold-italic">Q</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf67">
<mml:math id="m77">
<mml:mrow>
<mml:mi mathvariant="bold-italic">K</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf68">
<mml:math id="m78">
<mml:mrow>
<mml:mi mathvariant="bold-italic">V</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> represent the query vector, key vector and value vector, respectively. <inline-formula id="inf69">
<mml:math id="m79">
<mml:mrow>
<mml:mi>h</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> is the number of heads, <inline-formula id="inf70">
<mml:math id="m80">
<mml:mrow>
<mml:mi>d</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>m</mml:mi>
<mml:mo>/</mml:mo>
<mml:mi>h</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>. The original multi-head attention mechanism has a high computational cost, with a quadratic complexity of <inline-formula id="inf71">
<mml:math id="m81">
<mml:mrow>
<mml:mi mathvariant="script">O</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi>L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mi>d</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>, where <inline-formula id="inf72">
<mml:math id="m82">
<mml:mrow>
<mml:mi>L</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> is <inline-formula id="inf73">
<mml:math id="m83">
<mml:mrow>
<mml:mi>n</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mi>D</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>.</p>
<p>Inspired by MobileViT v2 (<xref ref-type="bibr" rid="B36">Mehta and Rastegari, 2022</xref>) and HViT-DUL (<xref ref-type="bibr" rid="B14">Deng et al., 2023</xref>), the multi-head design is incorporated into the separable self-attention mechanism, which is applied to the proposed model for the reduction of computational overhead. The separable multi-head attention mechanism employed in this paper has a linear complexity of <inline-formula id="inf74">
<mml:math id="m84">
<mml:mrow>
<mml:mi mathvariant="script">O</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>L</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>, which is lower compared to the standard multi-head attention mechanism. When multiple self-attention mechanisms are sequentially linked and subsequently subjected to a final projection, it results in the generation of the ultimate values for the multi-head attention mechanism. The spatial topological information and dynamic variation information of multi-channel electrodes are obtained through a multi-head attention mechanism to combine representations of multiple graph signals, thereby enhancing the discrimination ability. <xref ref-type="fig" rid="F4">Figure 4</xref> illustrates the process of separable multi-head attention. <bold>
<italic>Q</italic>
</bold>, <bold>
<italic>K</italic>
</bold>, <bold>
<italic>V</italic>
</bold> is calculated using the following formula:<disp-formula id="e11">
<mml:math id="m85">
<mml:mrow>
<mml:mtable class="aligned">
<mml:mtr>
<mml:mtd columnalign="right">
<mml:mfenced open="{" close="">
<mml:mrow>
<mml:mtable class="cases">
<mml:mtr>
<mml:mtd columnalign="left">
<mml:mi>Q</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>C</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>n</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mi>v</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi mathvariant="script">Z</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mspace width="1em"/>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd columnalign="left">
<mml:mi>K</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>C</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>n</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mi>v</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi mathvariant="script">Z</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mspace width="1em"/>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd columnalign="left">
<mml:mi>V</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>C</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>n</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mi>v</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>3</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi mathvariant="script">Z</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mspace width="1em"/>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
</mml:mfenced>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
</mml:math>
<label>(11)</label>
</disp-formula>where <inline-formula id="inf75">
<mml:math id="m86">
<mml:mrow>
<mml:mi mathvariant="script">Z</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> is the output of multi-branch GCN features. <inline-formula id="inf76">
<mml:math id="m87">
<mml:mrow>
<mml:mi>C</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>n</mml:mi>
<mml:mi>v</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> is two-dimensional convolution. We employ three <inline-formula id="inf77">
<mml:math id="m88">
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> convolutional layers, with parameters not shared, to compute the <inline-formula id="inf78">
<mml:math id="m89">
<mml:mrow>
<mml:mi mathvariant="bold-italic">Q</mml:mi>
<mml:mspace width="0.3333em"/>
<mml:mi mathvariant="bold-italic">K</mml:mi>
<mml:mspace width="0.3333em"/>
<mml:mi mathvariant="bold-italic">V</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> of the input graph features. After the reshaping process, the outputs <inline-formula id="inf79">
<mml:math id="m90">
<mml:mrow>
<mml:mi mathvariant="bold-italic">Q</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf80">
<mml:math id="m91">
<mml:mrow>
<mml:mi mathvariant="bold-italic">K</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf81">
<mml:math id="m92">
<mml:mrow>
<mml:mi mathvariant="bold-italic">V</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> are in the following dimensions: <inline-formula id="inf82">
<mml:math id="m93">
<mml:mrow>
<mml:mi mathvariant="bold-italic">Q</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="double-struck">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>h</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mi>n</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mi>D</mml:mi>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf83">
<mml:math id="m94">
<mml:mrow>
<mml:mi mathvariant="bold-italic">K</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="double-struck">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>h</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mi>d</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mi>n</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mi>D</mml:mi>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf84">
<mml:math id="m95">
<mml:mrow>
<mml:mi mathvariant="bold-italic">V</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="double-struck">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>h</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mi>d</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mi>n</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mi>D</mml:mi>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula>.</p>
<fig id="F4" position="float">
<label>FIGURE 4</label>
<caption>
<p>The process of self-attention.</p>
</caption>
<graphic xlink:href="fphys-15-1439607-g004.tif"/>
</fig>
<p>To calculate self attention scores, we apply the <inline-formula id="inf85">
<mml:math id="m96">
<mml:mrow>
<mml:mi>S</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>f</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>m</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>x</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> operation to the <inline-formula id="inf86">
<mml:math id="m97">
<mml:mrow>
<mml:mi mathvariant="bold-italic">Q</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> and taking the element-wise multiplication with <inline-formula id="inf87">
<mml:math id="m98">
<mml:mrow>
<mml:mi mathvariant="bold-italic">K</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, After that, we perform element-wise multiplication with <inline-formula id="inf88">
<mml:math id="m99">
<mml:mrow>
<mml:mi mathvariant="bold-italic">V</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> after passing it through ReLU. Self attention scores <inline-formula id="inf89">
<mml:math id="m100">
<mml:mrow>
<mml:mi>A</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>e</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> are computed based on the following formula:<disp-formula id="e12">
<mml:math id="m101">
<mml:mrow>
<mml:mi>a</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>e</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>R</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>L</mml:mi>
<mml:mi>U</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi mathvariant="bold-italic">V</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2217;</mml:mo>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi mathvariant="bold-italic">K</mml:mi>
<mml:mo>&#x2217;</mml:mo>
<mml:mi>S</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>f</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>m</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>x</mml:mi>
<mml:mi mathvariant="bold-italic">Q</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:math>
<label>(12)</label>
</disp-formula>where <inline-formula id="inf90">
<mml:math id="m102">
<mml:mrow>
<mml:mi>a</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> is the self-attention scores. <inline-formula id="inf91">
<mml:math id="m103">
<mml:mrow>
<mml:mi>R</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>L</mml:mi>
<mml:mi>U</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> is activation function. After concatenating the scores of multiple self-attention mechanisms, we perform a <inline-formula id="inf92">
<mml:math id="m104">
<mml:mrow>
<mml:mn>3</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>3</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> convolution on the resulting features, normalize it using BN, and then apply ReLU to enhance generalization ability. The computed multi-head attention scores, matching the dimensions of the input <inline-formula id="inf93">
<mml:math id="m105">
<mml:mrow>
<mml:mi mathvariant="script">Z</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, are added to the output of multi-branch GCN, expressed as:<disp-formula id="e13">
<mml:math id="m106">
<mml:mrow>
<mml:mi>A</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>n</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mfenced open="[" close="]">
<mml:mrow>
<mml:mi>a</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>e</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mi>a</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>e</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>,</mml:mo>
<mml:mi>a</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>e</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>h</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:math>
<label>(13)</label>
</disp-formula>
<disp-formula id="e14">
<mml:math id="m107">
<mml:mrow>
<mml:mi>M</mml:mi>
<mml:mi>A</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>R</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>L</mml:mi>
<mml:mi>U</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>B</mml:mi>
<mml:mi>N</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>C</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>n</mml:mi>
<mml:mi>v</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>A</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:math>
<label>(14)</label>
</disp-formula>
<disp-formula id="e15">
<mml:math id="m108">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="script">Z</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2032;</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>B</mml:mi>
<mml:mi>N</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi mathvariant="script">Z</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mi>M</mml:mi>
<mml:mi>A</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:math>
<label>(15)</label>
</disp-formula>where <inline-formula id="inf94">
<mml:math id="m109">
<mml:mrow>
<mml:mo stretchy="false">[</mml:mo>
<mml:mrow>
<mml:mo>&#x22c5;</mml:mo>
</mml:mrow>
<mml:mo stretchy="false">]</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> represents the concatenation.</p>
</sec>
<sec id="s2-5">
<title>2.5 Classifier module</title>
<p>A classifier module is placed after the multi-head attention mechanism for the final class inference. Initially, it involves two layers of 2D convolution, with a <inline-formula id="inf95">
<mml:math id="m110">
<mml:mrow>
<mml:mi>R</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>L</mml:mi>
<mml:mi>U</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> activation function applied after each convolution to enhance the robustness of model. This process can be described as:<disp-formula id="e16">
<mml:math id="m111">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="script">F</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2032;</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>R</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>L</mml:mi>
<mml:mi>U</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>C</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>n</mml:mi>
<mml:mi>v</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>R</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>L</mml:mi>
<mml:mi>U</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>C</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>n</mml:mi>
<mml:mi>v</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="script">Z</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2032;</mml:mo>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:math>
<label>(16)</label>
</disp-formula>where <inline-formula id="inf96">
<mml:math id="m112">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="script">Z</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2032;</mml:mo>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> is the output of multi-head attention mechanism. Global average pooling pools each channel of the convolution result, followed by a linear layer that reduces the feature dimensionality to match the number of classes. After passing through a sigmoid operation, the index with the highest probability is selected as the classification result. This process can be described as:<disp-formula id="e17">
<mml:math id="m113">
<mml:mrow>
<mml:mi>P</mml:mi>
<mml:mi>r</mml:mi>
<mml:msup>
<mml:mrow>
<mml:mi>e</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>s</mml:mi>
</mml:mrow>
</mml:msup>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>&#x3c3;</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>L</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>n</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>r</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>A</mml:mi>
<mml:mi>v</mml:mi>
<mml:mi>g</mml:mi>
<mml:mi>P</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>l</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="script">F</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2032;</mml:mo>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:math>
<label>(17)</label>
</disp-formula>where <inline-formula id="inf97">
<mml:math id="m114">
<mml:mrow>
<mml:mi>A</mml:mi>
<mml:mi>v</mml:mi>
<mml:mi>g</mml:mi>
<mml:mi>P</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>l</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mo>&#x22c5;</mml:mo>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> denotes global average pooling, <inline-formula id="inf98">
<mml:math id="m115">
<mml:mrow>
<mml:mi>&#x3c3;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> is sigmoid operation, <inline-formula id="inf99">
<mml:math id="m116">
<mml:mrow>
<mml:mi>L</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>n</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>r</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mo>&#x22c5;</mml:mo>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> is the Linear operation, and <inline-formula id="inf100">
<mml:math id="m117">
<mml:mrow>
<mml:mi>P</mml:mi>
<mml:mi>r</mml:mi>
<mml:msup>
<mml:mrow>
<mml:mi>e</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>s</mml:mi>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> is the final classification result. <xref ref-type="fig" rid="F5">Figure 5</xref> illustrates the structure of the classifier module.</p>
<fig id="F5" position="float">
<label>FIGURE 5</label>
<caption>
<p>The structure of classifier module.</p>
</caption>
<graphic xlink:href="fphys-15-1439607-g005.tif"/>
</fig>
</sec>
<sec id="s2-6">
<title>2.6 Training procedure</title>
<p>In order to achieve optimal network parameters during the training process, we employ the backpropagation (BP) algorithm, which iteratively updates the network parameters until an optimal or suboptimal solution is reached. In this context, we introduce a loss function based on binary cross entropy, defined as follows:<disp-formula id="e18">
<mml:math id="m118">
<mml:mrow>
<mml:mi>l</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>s</mml:mi>
<mml:mi>s</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>c</mml:mi>
<mml:mi>r</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>s</mml:mi>
<mml:mi>s</mml:mi>
<mml:mtext>_</mml:mtext>
<mml:mi>e</mml:mi>
<mml:mi>n</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>r</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>p</mml:mi>
<mml:mi>y</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>P</mml:mi>
<mml:mi>r</mml:mi>
<mml:mi>e</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>P</mml:mi>
<mml:mi>r</mml:mi>
<mml:msup>
<mml:mrow>
<mml:mi>e</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>s</mml:mi>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2b;</mml:mo>
<mml:mi>&#x3c4;</mml:mi>
<mml:mo stretchy="false">&#x2016;</mml:mo>
<mml:mi mathvariant="normal">&#x3a9;</mml:mi>
<mml:mo stretchy="false">&#x2016;</mml:mo>
</mml:mrow>
</mml:math>
<label>(18)</label>
</disp-formula>
<disp-formula id="e19">
<mml:math id="m119">
<mml:mrow>
<mml:mi>c</mml:mi>
<mml:mi>r</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>s</mml:mi>
<mml:mi>s</mml:mi>
<mml:mtext>_</mml:mtext>
<mml:mi>e</mml:mi>
<mml:mi>n</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>r</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>p</mml:mi>
<mml:mi>y</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>P</mml:mi>
<mml:mi>r</mml:mi>
<mml:mi>e</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>P</mml:mi>
<mml:mi>r</mml:mi>
<mml:msup>
<mml:mrow>
<mml:mi>e</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>s</mml:mi>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#x2212;</mml:mo>
<mml:mfenced open="[" close="]">
<mml:mrow>
<mml:mi>P</mml:mi>
<mml:mi>r</mml:mi>
<mml:mi>e</mml:mi>
<mml:mo>&#x2061;</mml:mo>
<mml:mi>log</mml:mi>
<mml:mo>&#x2061;</mml:mo>
<mml:mi>P</mml:mi>
<mml:mi>r</mml:mi>
<mml:msup>
<mml:mrow>
<mml:mi>e</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>S</mml:mi>
</mml:mrow>
</mml:msup>
<mml:mo>&#x2b;</mml:mo>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>P</mml:mi>
<mml:mi>r</mml:mi>
<mml:mi>e</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mi>log</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>P</mml:mi>
<mml:mi>r</mml:mi>
<mml:msup>
<mml:mrow>
<mml:mi>e</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>S</mml:mi>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:math>
<label>(19)</label>
</disp-formula>where <inline-formula id="inf101">
<mml:math id="m120">
<mml:mrow>
<mml:mi>P</mml:mi>
<mml:mi>r</mml:mi>
<mml:mi>e</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf102">
<mml:math id="m121">
<mml:mrow>
<mml:mi>P</mml:mi>
<mml:mi>r</mml:mi>
<mml:msup>
<mml:mrow>
<mml:mi>e</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>s</mml:mi>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> represent the actual label and the predicted label, respectively. The binary cross entropy, denoted as <inline-formula id="inf103">
<mml:math id="m122">
<mml:mrow>
<mml:mi>c</mml:mi>
<mml:mi>r</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>s</mml:mi>
<mml:mi>s</mml:mi>
<mml:mtext>_</mml:mtext>
<mml:mi>e</mml:mi>
<mml:mi>n</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>r</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>p</mml:mi>
<mml:mi>y</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>P</mml:mi>
<mml:mi>r</mml:mi>
<mml:mi>e</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>P</mml:mi>
<mml:mi>r</mml:mi>
<mml:msup>
<mml:mrow>
<mml:mi>e</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>s</mml:mi>
</mml:mrow>
</mml:msup>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>, quantifies the disparity between the true labels and the predicted labels. <inline-formula id="inf104">
<mml:math id="m123">
<mml:mrow>
<mml:mi>&#x3c4;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> represents the trade-off parameter, while <inline-formula id="inf105">
<mml:math id="m124">
<mml:mrow>
<mml:mo stretchy="false">&#x2016;</mml:mo>
<mml:mo>&#x22c5;</mml:mo>
<mml:mo stretchy="false">&#x2016;</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> denotes the <inline-formula id="inf106">
<mml:math id="m125">
<mml:mrow>
<mml:mi>l</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>2-norm to prevent overfitting. <inline-formula id="inf107">
<mml:math id="m126">
<mml:mrow>
<mml:mi mathvariant="normal">&#x3a9;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> represents all the parameters within this model.</p>
<p>For adaptive graph learning, the adjacency matrix A3 is a trainable parameter within the network that optimizes with model optimization. The partial derivatives of the loss function with respect to the optimal adjacency matrix A3 and the loss are expressed as follows:<disp-formula id="e20">
<mml:math id="m127">
<mml:mrow>
<mml:mfrac>
<mml:mrow>
<mml:mi>&#x2202;</mml:mi>
<mml:mi>l</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>s</mml:mi>
<mml:mi>s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x2202;</mml:mi>
<mml:mi>A</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mi>c</mml:mi>
<mml:mi>r</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>s</mml:mi>
<mml:mi>s</mml:mi>
<mml:mtext>_</mml:mtext>
<mml:mi>e</mml:mi>
<mml:mi>n</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>r</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>p</mml:mi>
<mml:mi>y</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>P</mml:mi>
<mml:mi>r</mml:mi>
<mml:mi>e</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>P</mml:mi>
<mml:mi>r</mml:mi>
<mml:msup>
<mml:mrow>
<mml:mi>e</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>s</mml:mi>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x2202;</mml:mi>
<mml:mi>A</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:mo>&#x2b;</mml:mo>
<mml:mi>&#x3c4;</mml:mi>
<mml:mfrac>
<mml:mrow>
<mml:mo stretchy="false">&#x2016;</mml:mo>
<mml:mi mathvariant="normal">&#x3a9;</mml:mi>
<mml:mo stretchy="false">&#x2016;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x2202;</mml:mi>
<mml:mi>A</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
<label>(20)</label>
</disp-formula>
</p>
</sec>
<sec id="s2-7">
<title>2.7 Dataset</title>
<p>The CHB-MIT dataset (<xref ref-type="bibr" rid="B43">Shoeb, 2009</xref>) was collected by Boston Children&#x2019;s Hospital and is currently the most widely used public dataset. The multichannel scalp EEG signals consists of EEG signal recordings from 23 children with epilepsy at a sampling rate of 256 Hz. The EEG signals are collected using EEG electrodes placed according to the International 10&#x2013;20 system. The dataset comprises approximately 983 h of EEG recordings, including 198 seizure onset events, with a total duration of 3 h 15 min.</p>
</sec>
<sec id="s2-8">
<title>2.8 Implementation details</title>
<p>In the CHB-MIT dataset, most records have 23 channels, while some records have missing or duplicated channels. In this study, 22 channels are selected in this study to maintain consistency of channels among all patients. The selected 22 channels are FP1-F7, F7-T7, T7-P7, P7-O1, FP1-F3, F3-C3, C3-P3, P3-O1, FP2-F4, F4-C4, C4-P4, P4-O2, FP2-F8, F8-T8, T8-P8, P8-O2, FZ-CZ, CZ-PZ, P7-T7, T7-FT9, FT9-FT10, FT10-T8. This dataset is classed into two categories, including interictal and ictal. To reduce the effect of noise, a fifth-order Butterworth band-pass filter ranging from 0.5 Hz to 70 Hz was used. The filtered EEG signals are segmented into 3-s windows. Due to the scarcity of ictal EEG data compared to interictal EEG data, the windows overlap by 50<inline-formula id="inf108">
<mml:math id="m128">
<mml:mrow>
<mml:mi>%</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> to obtain more ictal data.</p>
<p>For a comprehensive evaluation of the MGCNA, we train both a patient-independent model and a patient-specific model. In a patient-specific approach, the model is trained, validated, and tested using data from an individual subject. Interictal signals are randomly discarded for each subject, and the ratio of interictal signals to ictal signals is maintained at 5:1. As the number of ictal signals is significantly lower than that of interictal signals, which is detrimental to model training, we employ re-sampling to create a balanced training dataset. In each epoch of training, a random selection of an equal number of interictal signals is made to match the ictal signals in the training set, followed by random shuffling. Consequently, we obtain a balanced training subset for each epoch. We employ a 5-fold cross-validation approach and report the average performance. In a patient-independent approach, the model was trained and validated using data from multiple subjects, and then tested on data from individual subject. We utilized the leave-one-subject-out cross-validation as evaluation method. The data in both categories are kept balanced. The patient-specific approach emphasizes individual differences and personalization, and the patient-independent approach focuses more on overall trends and general patterns.</p>
<p>We employ the Adam optimizer with a learning rate of 3e-4, a weight decay of 1e-3.The dropout rate is experimentally set to 0.5. <inline-formula id="inf109">
<mml:math id="m129">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> is set to 0.4 and <inline-formula id="inf110">
<mml:math id="m130">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> is set to 0.25. <inline-formula id="inf111">
<mml:math id="m131">
<mml:mrow>
<mml:mi>k</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> is set to 2. All the above experiments were performed and implemented by Pytorch 1.7.1 in the NVIDIA GTX3090 and CUDA11.0 environment.</p>
</sec>
<sec id="s2-9">
<title>2.9 Performance evaluation metrics</title>
<p>We use eight different performance metrics to evaluate the performance of MGCNA including Accuracy (Acc), Sensitivity (Sen), Specificity (Spe), <inline-formula id="inf112">
<mml:math id="m132">
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> Score <inline-formula id="inf113">
<mml:math id="m133">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, and Area Under Curve (AUC). These metrics are obtained using true positives (TP), true negatives (TN), false positives (FP), and false negatives (FN) values.</p>
<p>Accuracy represents the proportion of samples correctly classified by the model out of the total sample count. Sensitivity and Specificity measure the model&#x2019;s ability to classify positive and negative samples. F1 Score combines precision and recall to provide a comprehensive performance measure. AUC is based on the area under the ROC curve and provides a comprehensive performance metric for unbalanced category distributions.</p>
</sec>
</sec>
<sec id="s3">
<title>3 Experimental results</title>
<sec id="s3-1">
<title>3.1 Patient-specific experiments</title>
<p>Patient-specific experimental results on CHB-MIT dataset using the MGCNA are shown in <xref ref-type="table" rid="T1">Table 1</xref>. According to the table, the model demonstrates an average accuracy of 99.32<inline-formula id="inf114">
<mml:math id="m134">
<mml:mrow>
<mml:mi>%</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, specificity of 98.4<inline-formula id="inf115">
<mml:math id="m135">
<mml:mrow>
<mml:mi>%</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, sensitivity of 98.74<inline-formula id="inf116">
<mml:math id="m136">
<mml:mrow>
<mml:mi>%</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, an F1 score of 97.76<inline-formula id="inf117">
<mml:math id="m137">
<mml:mrow>
<mml:mi>%</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, and an AUC of 98.51<inline-formula id="inf118">
<mml:math id="m138">
<mml:mrow>
<mml:mi>%</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>. For the majority of patients, the recognition accuracy exceeds 97<inline-formula id="inf119">
<mml:math id="m139">
<mml:mrow>
<mml:mi>%</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, with one patient achieving 100<inline-formula id="inf120">
<mml:math id="m140">
<mml:mrow>
<mml:mi>%</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> accuracy. The model achieves a specificity of 99<inline-formula id="inf121">
<mml:math id="m141">
<mml:mrow>
<mml:mi>%</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> for 12 patients, which represents 50<inline-formula id="inf122">
<mml:math id="m142">
<mml:mrow>
<mml:mi>%</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> of all test subjects, and a sensitivity of 99<inline-formula id="inf123">
<mml:math id="m143">
<mml:mrow>
<mml:mi>%</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> for 17 patients, covering 70<inline-formula id="inf124">
<mml:math id="m144">
<mml:mrow>
<mml:mi>%</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> of all test subjects. F1 scores for all patients are greater than 92<inline-formula id="inf125">
<mml:math id="m145">
<mml:mrow>
<mml:mi>%</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, and the AUC is above 95<inline-formula id="inf126">
<mml:math id="m146">
<mml:mrow>
<mml:mi>%</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>. These results indicate the stability and high performance of MGCNA proposed in this study, which can aid medical professionals in diagnosis.</p>
<table-wrap id="T1" position="float">
<label>TABLE 1</label>
<caption>
<p>Patient-specific experimental results on CHB-MIT dataset using the proposed architecture.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="center">Case</th>
<th align="center">Acc<inline-formula id="inf127">
<mml:math id="m147">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>%</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>
</th>
<th align="center">Sens<inline-formula id="inf128">
<mml:math id="m148">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>%</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>
</th>
<th align="center">Spec<inline-formula id="inf129">
<mml:math id="m149">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>%</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>
</th>
<th align="center">F1<inline-formula id="inf130">
<mml:math id="m150">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>%</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>
</th>
<th align="center">AUC<inline-formula id="inf131">
<mml:math id="m151">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>%</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>
</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="center">CHB01</td>
<td align="center">99.79</td>
<td align="center">99.12</td>
<td align="center">100.00</td>
<td align="center">99.33</td>
<td align="center">99.52</td>
</tr>
<tr>
<td align="center">CHB02</td>
<td align="center">99.56</td>
<td align="center">96.96</td>
<td align="center">100.00</td>
<td align="center">98.01</td>
<td align="center">98.44</td>
</tr>
<tr>
<td align="center">CHB03</td>
<td align="center">99.75</td>
<td align="center">99.74</td>
<td align="center">99.74</td>
<td align="center">99.20</td>
<td align="center">99.74</td>
</tr>
<tr>
<td align="center">CHB04</td>
<td align="center">99.32</td>
<td align="center">96.58</td>
<td align="center">99.50</td>
<td align="center">94.81</td>
<td align="center">98.05</td>
</tr>
<tr>
<td align="center">CHB05</td>
<td align="center">99.48</td>
<td align="center">97.40</td>
<td align="center">99.90</td>
<td align="center">98.34</td>
<td align="center">98.65</td>
</tr>
<tr>
<td align="center">CHB06</td>
<td align="center">99.94</td>
<td align="center">100.00</td>
<td align="center">99.94</td>
<td align="center">98.84</td>
<td align="center">98.63</td>
</tr>
<tr>
<td align="center">CHB07</td>
<td align="center">99.79</td>
<td align="center">100.00</td>
<td align="center">98.82</td>
<td align="center">99.39</td>
<td align="center">99.42</td>
</tr>
<tr>
<td align="center">CHB08</td>
<td align="center">99.62</td>
<td align="center">99.82</td>
<td align="center">98.57</td>
<td align="center">98.83</td>
<td align="center">99.20</td>
</tr>
<tr>
<td align="center">CHB09</td>
<td align="center">100.00</td>
<td align="center">100.00</td>
<td align="center">100.00</td>
<td align="center">100.00</td>
<td align="center">100.00</td>
</tr>
<tr>
<td align="center">CHB10</td>
<td align="center">99.96</td>
<td align="center">100.00</td>
<td align="center">99.74</td>
<td align="center">99.87</td>
<td align="center">99.87</td>
</tr>
<tr>
<td align="center">CHB11</td>
<td align="center">99.83</td>
<td align="center">99.92</td>
<td align="center">99.39</td>
<td align="center">99.51</td>
<td align="center">99.66</td>
</tr>
<tr>
<td align="center">CHB12</td>
<td align="center">98.63</td>
<td align="center">99.71</td>
<td align="center">93.77</td>
<td align="center">95.67</td>
<td align="center">96.74</td>
</tr>
<tr>
<td align="center">CHB13</td>
<td align="center">97.88</td>
<td align="center">99.26</td>
<td align="center">93.51</td>
<td align="center">95.07</td>
<td align="center">96.38</td>
</tr>
<tr>
<td align="center">CHB14</td>
<td align="center">98.11</td>
<td align="center">99.07</td>
<td align="center">93.45</td>
<td align="center">94.18</td>
<td align="center">96.26</td>
</tr>
<tr>
<td align="center">CHB15</td>
<td align="center">98.31</td>
<td align="center">98.66</td>
<td align="center">97.91</td>
<td align="center">98.21</td>
<td align="center">98.29</td>
</tr>
<tr>
<td align="center">CHB16</td>
<td align="center">99.61</td>
<td align="center">100.00</td>
<td align="center">97.50</td>
<td align="center">98.67</td>
<td align="center">98.75</td>
</tr>
<tr>
<td align="center">CHB17</td>
<td align="center">99.29</td>
<td align="center">99.79</td>
<td align="center">96.90</td>
<td align="center">97.82</td>
<td align="center">98.34</td>
</tr>
<tr>
<td align="center">CHB18</td>
<td align="center">99.55</td>
<td align="center">99.80</td>
<td align="center">98.28</td>
<td align="center">98.60</td>
<td align="center">99.04</td>
</tr>
<tr>
<td align="center">CHB19</td>
<td align="center">99.78</td>
<td align="center">100.00</td>
<td align="center">98.57</td>
<td align="center">99.26</td>
<td align="center">99.29</td>
</tr>
<tr>
<td align="center">CHB20</td>
<td align="center">98.91</td>
<td align="center">93.61</td>
<td align="center">99.50</td>
<td align="center">94.32</td>
<td align="center">96.56</td>
</tr>
<tr>
<td align="center">CHB21</td>
<td align="center">97.98</td>
<td align="center">92.68</td>
<td align="center">98.82</td>
<td align="center">92.81</td>
<td align="center">95.75</td>
</tr>
<tr>
<td align="center">CHB22</td>
<td align="center">99.43</td>
<td align="center">99.53</td>
<td align="center">99.41</td>
<td align="center">98.15</td>
<td align="center">99.47</td>
</tr>
<tr>
<td align="center">CHB23</td>
<td align="center">99.58</td>
<td align="center">99.80</td>
<td align="center">98.55</td>
<td align="center">98.78</td>
<td align="center">99.18</td>
</tr>
<tr>
<td align="center">CHB24</td>
<td align="center">99.48</td>
<td align="center">98.20</td>
<td align="center">99.77</td>
<td align="center">98.57</td>
<td align="center">98.99</td>
</tr>
<tr>
<td align="center">Mean</td>
<td align="center">99.32</td>
<td align="center">98.74</td>
<td align="center">98.40</td>
<td align="center">97.76</td>
<td align="center">98.51</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="s3-2">
<title>3.2 Patient-independent experiments</title>
<p>In contrast to the patient-specific experiments, patient-independent experiments involve training a general model for all patients. The experiment employed the leave-one-out method. Patient-independent experiments separate one patient&#x2019;s signal as a test set, using the EEG data of other patients as the training set and Validation set. This allows the model to detect epileptic activity in the test set by learning common seizure characteristics from the training set. This experiment is more clinically meaningful but requires to address differences in EEG signals among different patients, which may result from physiological variations, equipment noise, etc. Compared to patient-specific experiments, patient-independent experiments make it more challenging for the algorithm to identify epileptic seizures and lead to poorer detection results.</p>
<p>Patient-independent experimental results on CHB-MIT dataset using the MGCNA are shown in <xref ref-type="table" rid="T2">Table 2</xref>. As seen in the table, the average values for Accuracy, Sensitivity, Specificity, and F1 are all above 80<inline-formula id="inf132">
<mml:math id="m152">
<mml:mrow>
<mml:mi>%</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>. There is a significant individual variation among different patients. Patients <inline-formula id="inf133">
<mml:math id="m153">
<mml:mrow>
<mml:mi>&#x266f;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>4, <inline-formula id="inf134">
<mml:math id="m154">
<mml:mrow>
<mml:mi>&#x266f;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>5, <inline-formula id="inf135">
<mml:math id="m155">
<mml:mrow>
<mml:mi>&#x266f;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>7, <inline-formula id="inf136">
<mml:math id="m156">
<mml:mrow>
<mml:mi>&#x266f;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>9, <inline-formula id="inf137">
<mml:math id="m157">
<mml:mrow>
<mml:mi>&#x266f;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>10, <inline-formula id="inf138">
<mml:math id="m158">
<mml:mrow>
<mml:mi>&#x266f;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>15, <inline-formula id="inf139">
<mml:math id="m159">
<mml:mrow>
<mml:mi>&#x266f;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>17, <inline-formula id="inf140">
<mml:math id="m160">
<mml:mrow>
<mml:mi>&#x266f;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>18, <inline-formula id="inf141">
<mml:math id="m161">
<mml:mrow>
<mml:mi>&#x266f;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>19, <inline-formula id="inf142">
<mml:math id="m162">
<mml:mrow>
<mml:mi>&#x266f;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>22, <inline-formula id="inf143">
<mml:math id="m163">
<mml:mrow>
<mml:mi>&#x266f;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>23 achieved accuracy rates exceeding 90<inline-formula id="inf144">
<mml:math id="m164">
<mml:mrow>
<mml:mi>%</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, whereas patients <inline-formula id="inf145">
<mml:math id="m165">
<mml:mrow>
<mml:mi>&#x266f;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>8, <inline-formula id="inf146">
<mml:math id="m166">
<mml:mrow>
<mml:mi>&#x266f;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>12, <inline-formula id="inf147">
<mml:math id="m167">
<mml:mrow>
<mml:mi>&#x266f;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>13, and <inline-formula id="inf148">
<mml:math id="m168">
<mml:mrow>
<mml:mi>&#x266f;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>24 had comparatively lower accuracy rates. Compared to patient-specific experiments, patient-independent experiments exhibited a 13.95<inline-formula id="inf149">
<mml:math id="m169">
<mml:mrow>
<mml:mi>%</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> lower accuracy rate. This is due to the differences in seizures between patients and the presence of different types of symptoms during seizures, which vary in frequency and duration.</p>
<table-wrap id="T2" position="float">
<label>TABLE 2</label>
<caption>
<p>Patient-independent experimental results on CHB-MIT dataset using the proposed architecture.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="center">Case</th>
<th align="center">Acc<inline-formula id="inf150">
<mml:math id="m170">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>%</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>
</th>
<th align="center">Sens<inline-formula id="inf151">
<mml:math id="m171">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>%</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>
</th>
<th align="center">Spec<inline-formula id="inf152">
<mml:math id="m172">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>%</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>
</th>
<th align="center">F1<inline-formula id="inf153">
<mml:math id="m173">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>%</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>
</th>
<th align="center">AUC<inline-formula id="inf154">
<mml:math id="m174">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>%</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>
</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="center">CHB01</td>
<td align="center">91.51</td>
<td align="center">83.73</td>
<td align="center">99.28</td>
<td align="center">90.79</td>
<td align="center">91.51</td>
</tr>
<tr>
<td align="center">CHB02</td>
<td align="center">82.10</td>
<td align="center">64.20</td>
<td align="center">100.00</td>
<td align="center">78.20</td>
<td align="center">82.10</td>
</tr>
<tr>
<td align="center">CHB03</td>
<td align="center">81.33</td>
<td align="center">62.67</td>
<td align="center">100.00</td>
<td align="center">77.05</td>
<td align="center">81.33</td>
</tr>
<tr>
<td align="center">CHB04</td>
<td align="center">99.45</td>
<td align="center">99.18</td>
<td align="center">99.73</td>
<td align="center">99.45</td>
<td align="center">99.45</td>
</tr>
<tr>
<td align="center">CHB05</td>
<td align="center">96.49</td>
<td align="center">97.60</td>
<td align="center">95.38</td>
<td align="center">96.53</td>
<td align="center">96.49</td>
</tr>
<tr>
<td align="center">CHB06</td>
<td align="center">85.04</td>
<td align="center">92.31</td>
<td align="center">77.78</td>
<td align="center">86.06</td>
<td align="center">85.04</td>
</tr>
<tr>
<td align="center">CHB07</td>
<td align="center">98.89</td>
<td align="center">98.41</td>
<td align="center">99.36</td>
<td align="center">98.88</td>
<td align="center">98.89</td>
</tr>
<tr>
<td align="center">CHB08</td>
<td align="center">92.34</td>
<td align="center">91.90</td>
<td align="center">92.79</td>
<td align="center">92.31</td>
<td align="center">92.34</td>
</tr>
<tr>
<td align="center">CHB09</td>
<td align="center">100.00</td>
<td align="center">100.00</td>
<td align="center">100.00</td>
<td align="center">100.00</td>
<td align="center">100.00</td>
</tr>
<tr>
<td align="center">CHB10</td>
<td align="center">99.76</td>
<td align="center">99.76</td>
<td align="center">99.76</td>
<td align="center">99.76</td>
<td align="center">99.76</td>
</tr>
<tr>
<td align="center">CHB11</td>
<td align="center">93.89</td>
<td align="center">93.32</td>
<td align="center">94.46</td>
<td align="center">93.86</td>
<td align="center">93.89</td>
</tr>
<tr>
<td align="center">CHB12</td>
<td align="center">61.46</td>
<td align="center">46.21</td>
<td align="center">76.70</td>
<td align="center">54.52</td>
<td align="center">61.46</td>
</tr>
<tr>
<td align="center">CHB13</td>
<td align="center">70.52</td>
<td align="center">81.09</td>
<td align="center">59.95</td>
<td align="center">73.34</td>
<td align="center">70.52</td>
</tr>
<tr>
<td align="center">CHB14</td>
<td align="center">70.44</td>
<td align="center">58.39</td>
<td align="center">82.48</td>
<td align="center">66.39</td>
<td align="center">70.44</td>
</tr>
<tr>
<td align="center">CHB15</td>
<td align="center">91.93</td>
<td align="center">82.93</td>
<td align="center">93.33</td>
<td align="center">73.52</td>
<td align="center">91.93</td>
</tr>
<tr>
<td align="center">CHB16</td>
<td align="center">86.90</td>
<td align="center">73.81</td>
<td align="center">100.00</td>
<td align="center">84.93</td>
<td align="center">86.90</td>
</tr>
<tr>
<td align="center">CHB17</td>
<td align="center">98.05</td>
<td align="center">97.87</td>
<td align="center">98.23</td>
<td align="center">98.05</td>
<td align="center">98.05</td>
</tr>
<tr>
<td align="center">CHB18</td>
<td align="center">93.20</td>
<td align="center">86.39</td>
<td align="center">100.00</td>
<td align="center">92.70</td>
<td align="center">93.20</td>
</tr>
<tr>
<td align="center">CHB19</td>
<td align="center">94.69</td>
<td align="center">90.27</td>
<td align="center">99.12</td>
<td align="center">94.44</td>
<td align="center">94.69</td>
</tr>
<tr>
<td align="center">CHB20</td>
<td align="center">79.25</td>
<td align="center">58.49</td>
<td align="center">100.00</td>
<td align="center">73.81</td>
<td align="center">79.25</td>
</tr>
<tr>
<td align="center">CHB21</td>
<td align="center">95.41</td>
<td align="center">91.89</td>
<td align="center">98.92</td>
<td align="center">95.24</td>
<td align="center">95.41</td>
</tr>
<tr>
<td align="center">CHB22</td>
<td align="center">98.70</td>
<td align="center">97.93</td>
<td align="center">99.48</td>
<td align="center">98.69</td>
<td align="center">98.70</td>
</tr>
<tr>
<td align="center">CHB23</td>
<td align="center">97.61</td>
<td align="center">95.98</td>
<td align="center">99.25</td>
<td align="center">97.57</td>
<td align="center">97.61</td>
</tr>
<tr>
<td align="center">CHB24</td>
<td align="center">87.80</td>
<td align="center">88.91</td>
<td align="center">86.70</td>
<td align="center">87.94</td>
<td align="center">87.80</td>
</tr>
<tr>
<td align="center">Mean</td>
<td align="center">89.45</td>
<td align="center">84.72</td>
<td align="center">93.86</td>
<td align="center">87.67</td>
<td align="center">89.45</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="s3-3">
<title>3.3 Ablation study</title>
<p>To assess the contributions of different components in our model to classification performance, we conduct comparative experiments using model based only on similarity graph representation, model based only on distance graph representation, model based only on trainable graph representation, and model without multi-head attention. To ensure the fairness of the experiments, the model settings are kept consistent. The results are shown in <xref ref-type="fig" rid="F6">Figures 6</xref>, <xref ref-type="fig" rid="F7">7</xref>.</p>
<fig id="F6" position="float">
<label>FIGURE 6</label>
<caption>
<p>Ablation study for epilepsy recognition CHB-MIT dataset of different graph learning.&#x2217;h1 represents the use of spatial distance graph learning only. &#x2217;h2 indicates the use of functional connectivity graph learning only. &#x2217;h3 denotes the use of adaptive graph learning only.</p>
</caption>
<graphic xlink:href="fphys-15-1439607-g006.tif"/>
</fig>
<fig id="F7" position="float">
<label>FIGURE 7</label>
<caption>
<p>Ablation study for epilepsy recognition CHB-MIT dataset of multi-head attention.</p>
</caption>
<graphic xlink:href="fphys-15-1439607-g007.tif"/>
</fig>
<p>Graph convolutional network uses only spatial distance graph learning, denoted as &#x2217;h1. Graph convolutional network uses only functional connectivity graph learning, denoted as &#x2217;h2. Graph convolutional network uses adaptive graph learning, denoted as &#x2217;h3. Non-attention graph representation models refers to the model that does not employ self-attention mechanisms to obtain self-attention scores for the three graph representations. We directly concatenate the representations of three graphs into a graph feature, denoted as &#x2217;self-attention.</p>
<p>From the <xref ref-type="fig" rid="F6">Figures 6</xref>, <xref ref-type="fig" rid="F7">7</xref>, it is evident that MGCNA outperforms the other four ablation experiments. Compared to our model in this paper, the accuracy of the four ablation experiment models is lower by 5.22<inline-formula id="inf155">
<mml:math id="m175">
<mml:mrow>
<mml:mi>%</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, 5.30<inline-formula id="inf156">
<mml:math id="m176">
<mml:mrow>
<mml:mi>%</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, 3.37<inline-formula id="inf157">
<mml:math id="m177">
<mml:mrow>
<mml:mi>%</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, and 3.99<inline-formula id="inf158">
<mml:math id="m178">
<mml:mrow>
<mml:mi>%</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, respectively. As illustrated in <xref ref-type="fig" rid="F6">Figure 6</xref>, the trainable graph representation model demonstrates superior performance compared to the other two graph representation learning methods, particularly for patients <inline-formula id="inf159">
<mml:math id="m179">
<mml:mrow>
<mml:mi>&#x266f;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>4, <inline-formula id="inf160">
<mml:math id="m180">
<mml:mrow>
<mml:mi>&#x266f;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>12, <inline-formula id="inf161">
<mml:math id="m181">
<mml:mrow>
<mml:mi>&#x266f;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>13, <inline-formula id="inf162">
<mml:math id="m182">
<mml:mrow>
<mml:mi>&#x266f;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>14, <inline-formula id="inf163">
<mml:math id="m183">
<mml:mrow>
<mml:mi>&#x266f;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>17, and <inline-formula id="inf164">
<mml:math id="m184">
<mml:mrow>
<mml:mi>&#x266f;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>21. Between spatial distance graph learning and functional connectivity graph learning, the latter more effectively captures the interrelationships among EEG signals.</p>
<p>The effect of the self-attention mechanism on epilepsy detection approximates that of the model using only distance graph representation. <xref ref-type="fig" rid="F7">Figure 7</xref> illustrates that allocating different attention weights to graph features through self-attention mechanism helps to capture dependencies between graph features and enhances the model&#x2019;s ability to model relationships between graph features, which can achieve good performance in epilepsy detection. The model can effectively utilize the spatiotemporal relationships among EEG signals, supplement information through learnable graph representations, and obtain attention scores through self-attention mechanisms, and therefore the model&#x2019;s learning capability has been significantly enhanced.</p>
<p>To further validate the superiority of MGCNA, t-SNE is applied to visualize and analyze the features extracted from the ablation study. As depicted in <xref ref-type="fig" rid="F8">Figure 8</xref>, the t-SNE visualization in two-dimensional embedding space illustrates the interictal and ictal features for both patient-specific and patient-independent experiments. We can see that in both patient-specific and patient-independent experiments, our approach exhibits superior recognition capabilities compared to the ablated models. Particularly, models utilizing only one graph construction method tend to confuse some interictal and ictal features. In contrast, better discriminative features were obtained using MGCNA, mainly in terms of significant inter-ictal distances and dense intra-ictal distributions. These observations indicate that combining multi-branch GCN with self-attention mechanisms can yield the optimal performance for epileptic seizure classification.</p>
<fig id="F8" position="float">
<label>FIGURE 8</label>
<caption>
<p>The t-SNE visualization in 2D embedding space of interictal and ictal features by comparing the models from ablation study.</p>
</caption>
<graphic xlink:href="fphys-15-1439607-g008.tif"/>
</fig>
</sec>
</sec>
<sec sec-type="discussion" id="s4">
<title>4 Discussion</title>
<sec id="s4-1">
<title>4.1 The influence of thresholds on graph representations</title>
<p>In both functional connectivity graph learning and spatial distance graph learning methods, it is necessary to set a threshold for constructing an adjacency matrix. The constructed adjacency matrix must not only ensure the sparsity of the graph but also be capable of distinguishing temporal and spatial characteristics of different types of EEG data to enhance the accuracy of the model in identifying epileptic seizures.</p>
<p>
<xref ref-type="fig" rid="F9">Figure 9A</xref> depicts the average epileptic seizure detection results of the CHB-MIT dataset in patient-specific and patient-independent experiments, under varying thresholds <inline-formula id="inf165">
<mml:math id="m185">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> for spatial distance graph learning. The threshold range selected spans from 0.2 to 0.7. Notably, at a threshold of 0.2, the patient-specific epileptic seizure detection accuracy is the lowest, as lower thresholds tend to introduce excessive irrelevant physical connections between unrelated nodes in the spatial distance graph. Therefore, to maintain the sparsity of the adjacency matrix and achieve optimal seizure detection results, configuring the threshold <inline-formula id="inf166">
<mml:math id="m186">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> for spatial distance graph learning to 0.4 is advocated.</p>
<fig id="F9" position="float">
<label>FIGURE 9</label>
<caption>
<p>Performance comparison of patient-specific and patient-independent experiments with different thresholds. <bold>(A)</bold> spatial distance graph; <bold>(B)</bold> functional connectivity graph.</p>
</caption>
<graphic xlink:href="fphys-15-1439607-g009.tif"/>
</fig>
<p>
<xref ref-type="fig" rid="F9">Figure 9B</xref> illustrates the average epileptic seizure detection results in both patient-specific and patient-independent experiments under various threshold values <inline-formula id="inf167">
<mml:math id="m187">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> for functional connectivity graph learning. The threshold range selected spans from 0.1 to 0.5, with a stride of 0.05. From <xref ref-type="fig" rid="F9">Figure 9</xref>, it is evident that <inline-formula id="inf168">
<mml:math id="m188">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> has a more pronounced impact on patient-independent epileptic seizure detection results. In the patient-specific experiment, where the training and testing datasets originate from the same patient, they exhibit similar data distributions. However, in the patient-independent experiment, differences exist between various patients, leading to inconsistent effects of different thresholds on the patients. Appropriate thresholds are applied to eliminate unrelated channel signals, utilizing functional connectivity to capture neuronal synchronized discharge during epileptic seizures (<xref ref-type="bibr" rid="B1">Abbas et al., 2021</xref>). To maximize the spatial discriminative power of the functional connectivity graph and achieve optimal results for seizure detection, the threshold <inline-formula id="inf169">
<mml:math id="m189">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> for functional connectivity graph learning is set at 0.25.</p>
<p>The spatial distance graph learning based on <xref ref-type="disp-formula" rid="e3">Equation 3</xref> with the threshold <inline-formula id="inf170">
<mml:math id="m190">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> set to 0.4 is illustrated in <xref ref-type="fig" rid="F10">Figure 10</xref>. The distance between two bipolar derivations was determined by measuring the separation between the centers of the bipolar derivations. For instance, the coordinates of FP1-F3 are the centers of the FP1 coordinates and the F3 coordinates. After calculating the Euclidean distances, these distances are normalized to create the distance graph representation, and self-connections are added, i.e., the graph representation of the Euclidean distances plus the diagonal matrix.</p>
<fig id="F10" position="float">
<label>FIGURE 10</label>
<caption>
<p>The spatial distance graph learning of EEG signals under thresholds <inline-formula id="inf171">
<mml:math id="m191">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> set as 0.4.</p>
</caption>
<graphic xlink:href="fphys-15-1439607-g010.tif"/>
</fig>
<p>
<xref ref-type="fig" rid="F11">Figure 11</xref> depicts the functional connectivity graph learning of interictal and ictal signals under thresholds set at 0.25. In the figure, we have selected and described the functional connectivity graph learning that represents interictal and ictal signals with the highest degree of similarity. The values in the graph are calculated as the Pearson correlation coefficients, which are then normalized into a similarity adjacency matrix ranging from 0 to 1. From <xref ref-type="fig" rid="F11">Figure 11</xref>, it can be observed that functional connectivity graph learning for interictal signals exhibits a strong degree of similarity, while the ictal signal displays weaker inter-correlations across many bipolar derivations. Graph representations based on functional connectivity can depict the functional connections between brain regions as interdependencies among EEG signals.</p>
<fig id="F11" position="float">
<label>FIGURE 11</label>
<caption>
<p>The functional connectivity graph learning of interictal and ictal signals under thresholds <inline-formula id="inf172">
<mml:math id="m192">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> set as 0.25. <bold>(A)</bold> ictal signal; <bold>(B)</bold> interictal signal.</p>
</caption>
<graphic xlink:href="fphys-15-1439607-g011.tif"/>
</fig>
</sec>
<sec id="s4-2">
<title>4.2 Comparisons with state-of-the-art methods</title>
<p>
<xref ref-type="table" rid="T3">Table 3</xref> presents a performance comparison between the proposed MGCNA and state-of-the-art epilepsy detection algorithms in the patient-specific experimental setting. Most of these methods involve feature extraction and deep learning algorithms for epilepsy detection. It is evident that the advancement of deep learning methods holds significant importance for signal recognition and detection.</p>
<table-wrap id="T3" position="float">
<label>TABLE 3</label>
<caption>
<p>Comparisons of performance between proposed method and recent works (patient-specific experimental setting).</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left">Method</th>
<th align="left">Acc<inline-formula id="inf173">
<mml:math id="m193">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>%</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>
</th>
<th align="left">Sens<inline-formula id="inf174">
<mml:math id="m194">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>%</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>
</th>
<th align="left">Spec<inline-formula id="inf175">
<mml:math id="m195">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>%</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>
</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">Channel selection &#x2b; KNN (<xref ref-type="bibr" rid="B17">Ein Shoka et al., 2021</xref>)</td>
<td align="left">85</td>
<td align="left">86.04</td>
<td align="left">83.78</td>
</tr>
<tr>
<td align="left">Multiscale spectral features &#x2b; RF (<xref ref-type="bibr" rid="B5">Sukriti et al., 2021</xref>)</td>
<td align="left">98.9</td>
<td align="left">98.12</td>
<td align="left">
<bold>99.17</bold>
</td>
</tr>
<tr>
<td align="left">normalization &#x2b; CNN (<xref ref-type="bibr" rid="B9">Cimr et al., 2023</xref>)</td>
<td align="left">96.99</td>
<td align="left">97.06</td>
<td align="left">96.89</td>
</tr>
<tr>
<td align="left">WNG-TS-1DCNN (<xref ref-type="bibr" rid="B52">Wang et al., 2023b</xref>)</td>
<td align="left">93.1</td>
<td align="left">91.8</td>
<td align="left">96.3</td>
</tr>
<tr>
<td align="left">HAN (<xref ref-type="bibr" rid="B64">Zhao et al., 2023b</xref>)</td>
<td align="left">98.3</td>
<td align="left">97.34</td>
<td align="left">96.07</td>
</tr>
<tr>
<td align="left">GAT &#x2b; BiLSTM (<xref ref-type="bibr" rid="B20">He et al., 2022</xref>)</td>
<td align="left">98.52</td>
<td align="left">97.75</td>
<td align="left">94.34</td>
</tr>
<tr>
<td align="left">CNN &#x2b; Transformer (<xref ref-type="bibr" rid="B63">Zhao et al., 2023a</xref>)</td>
<td align="left">98.76</td>
<td align="left">97.70</td>
<td align="left">97.60</td>
</tr>
<tr>
<td align="left">proposed MGCNA</td>
<td align="left">
<bold>99.32</bold>
</td>
<td align="left">
<bold>98.74</bold>
</td>
<td align="left">98.40</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn>
<p>The bold values represent the highest values in each column.</p>
</fn>
</table-wrap-foot>
</table-wrap>
<p>Machine learning methods are classical approaches for epilepsy detection. <xref ref-type="bibr" rid="B17">Ein Shoka et al. (2021)</xref> conducted feature extraction after channel selection and evaluated the performance of different classifiers for epilepsy detection, with KNN exhibiting the best classification performance. Sukriti et al. (<xref ref-type="bibr" rid="B5">Sukriti et al., 2021</xref>) extracted multiscale spectral features (MSSFs) and employed a random forest classifier for epilepsy classification exhibits a higher specificity compared to MGCNA (99.17<inline-formula id="inf176">
<mml:math id="m196">
<mml:mrow>
<mml:mi>%</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> <italic>versus</italic> 98.4<inline-formula id="inf177">
<mml:math id="m197">
<mml:mrow>
<mml:mi>%</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>). However, the MGCNA achieves higher accuracy and sensitivity than that method. <xref ref-type="bibr" rid="B9">Cimr et al. (2023)</xref> achieved EEG classification by normalizing input signals and an 8-layer depth CNN model. The MGCNA is superior to that method in terms of accuracy, specificity and sensitivity. <xref ref-type="bibr" rid="B63">Zhao et al. (2023a)</xref> proposed using CNNs to extract local features and transformers to capture global information. Although MGCNA outperforms this approach, the differences in preprocessing methods also affect the results of epilepsy detection.</p>
<p>GCN can analyze signals by considering the three-dimensional spatial positions of EEG electrodes, thereby compensating for potential spatial information loss in deep learning methods such as CNN. The introduction of GCN can be utilized to analyze the spatiotemporal correlations among channels. Methods based on GCN have already demonstrated excellent performance for the detection of epileptic signals. <xref ref-type="bibr" rid="B52">Wang et al. (2023b)</xref> introduced a two-stream graph-based framework for learning the Weighted Neighbour Graph (WNG) representation in both the frequency and time domains. However, this approach achieved inferior accuracy, sensitivity, and specificity, with values of 93.1<inline-formula id="inf178">
<mml:math id="m198">
<mml:mrow>
<mml:mi>%</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, 91.8<inline-formula id="inf179">
<mml:math id="m199">
<mml:mrow>
<mml:mi>%</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, and 96.3<inline-formula id="inf180">
<mml:math id="m200">
<mml:mrow>
<mml:mi>%</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, respectively, in comparison to the method proposed in the current study. <xref ref-type="bibr" rid="B64">Zhao et al. (2023b)</xref> introduced a hybrid Attention Network that utilizes the GAT to extract spatial features and the Transformer to extract temporal features addresses the issue of imbalanced data. However, in the patient-specific experiment, it achieved slightly lower accuracy, specificity, and sensitivity compared to MGCNA. <xref ref-type="bibr" rid="B20">He et al. (2022)</xref> utilized the GAT to extract spatial features and employed a BiLSTM network to capture temporal features, achieving an accuracy of 98.52<inline-formula id="inf181">
<mml:math id="m201">
<mml:mrow>
<mml:mi>%</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> on the CHB-MIT dataset, slightly lower than MGCNA.</p>
<p>
<xref ref-type="table" rid="T4">Table 4</xref> presents comparisons of performance between the proposed method and state-of-the-art epilepsy detection algorithms in the patient-independent experimental setting. Due to the substantial variations in signals among different patients, the overall performance of the patient-independent experiment is lower than that of the patient-specific experiment. The accuracy of the patient-independent experiment based on SVM (<xref ref-type="bibr" rid="B43">Shoeb, 2009</xref>) only reaches 58.32<inline-formula id="inf182">
<mml:math id="m202">
<mml:mrow>
<mml:mi>%</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, indicating a significant performance gap compared to MGCNA. <xref ref-type="bibr" rid="B56">Wei et al. (2019)</xref> proposed an epilepsy detection algorithm by combining CNN and Wasserstein Generative Adversarial Nets (WGANs), achieving lower sensitivity than MGCNA, i.e. 72.11<inline-formula id="inf183">
<mml:math id="m203">
<mml:mrow>
<mml:mi>%</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> vs. 83.54<inline-formula id="inf184">
<mml:math id="m204">
<mml:mrow>
<mml:mi>%</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, respectively. <xref ref-type="bibr" rid="B64">Zhao et al. (2023b)</xref> introduced the HAN model, which performs epilepsy detection not only in patient-specific experiments but also in patient-independent experiments. In both types of experiments, the three performance metrics were lower than MGCNA. Additionally, the ablation studies conducted in the patient-independent experimental setting are showcased herein, demonstrating that the MGCNA method exhibits superior performance compared to each ablation study. Since the three graph learning models can better extract the structural information of EEG and perform graph feature fusion, the overall performance is better than other epilepsy detection algorithms.</p>
<table-wrap id="T4" position="float">
<label>TABLE 4</label>
<caption>
<p>Comparisons of performance between proposed method and recent works (patient-independent experimental setting).</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left">Method</th>
<th align="left">Acc<inline-formula id="inf185">
<mml:math id="m205">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>%</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>
</th>
<th align="left">Sens<inline-formula id="inf186">
<mml:math id="m206">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>%</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>
</th>
<th align="left">Spec<inline-formula id="inf187">
<mml:math id="m207">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>%</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>
</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">SVM (<xref ref-type="bibr" rid="B43">Shoeb, 2009</xref>)</td>
<td align="left">58.32</td>
<td align="left">74.6</td>
<td align="left">42.19</td>
</tr>
<tr>
<td align="left">WGANs (<xref ref-type="bibr" rid="B56">Wei et al., 2019</xref>)</td>
<td align="left">84.00</td>
<td align="left">72.11</td>
<td align="left">
<bold>95.89</bold>
</td>
</tr>
<tr>
<td align="left">HAN (<xref ref-type="bibr" rid="B64">Zhao et al., 2023b</xref>)</td>
<td align="left">73.15</td>
<td align="left">72.75</td>
<td align="left">75.7</td>
</tr>
<tr>
<td align="left">&#x2217;h1</td>
<td align="left">82.49</td>
<td align="left">80.33</td>
<td align="left">85.74</td>
</tr>
<tr>
<td align="left">&#x2217;h2</td>
<td align="left">85.21</td>
<td align="left">79.83</td>
<td align="left">91.27</td>
</tr>
<tr>
<td align="left">&#x2217;h3</td>
<td align="left">84.72</td>
<td align="left">80.83</td>
<td align="left">89.16</td>
</tr>
<tr>
<td align="left">&#x2217;self-attention</td>
<td align="left">84.19</td>
<td align="left">79.49</td>
<td align="left">89.56</td>
</tr>
<tr>
<td align="left">proposed MGCNA</td>
<td align="left">
<bold>85.45</bold>
</td>
<td align="left">
<bold>84.72</bold>
</td>
<td align="left">93.86</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn>
<p>The bold values represent the highest values in each column.</p>
</fn>
</table-wrap-foot>
</table-wrap>
<p>Through the above analysis, the MGCNA framework offers several key advantages. First, it employs a multi-branch graph convolutional network structure that dynamically learns temporal correlations and spatial topological information, enhancing the ability to process complex EEG signals, particularly in capturing spatial and temporal dependencies in epileptic brains. Second, the use of three graph learning approaches allows for a comprehensive evaluation of connectivity and synchronization across multiple channels, improving adaptability in different patient scenarios. Additionally, the multi-head attention mechanism further strengthens the framework&#x2019;s ability to handle local features and complex EEG patterns. Experimental results demonstrate that MGCNA outperforms other methods in both patient-specific and patient-independent tasks, highlighting its strong generalization capabilities. Lastly, as an end-to-end automatic seizure detection model, MGCNA can be applied in clinical decision-making, helping clinicians diagnose childhood epileptic seizures more quickly and accurately, providing significant practical value.</p>
</sec>
<sec id="s4-3">
<title>4.3 Limitations and future work</title>
<p>While the MGCNA achieved satisfactory results in both patient-specific and patient-independent experiments, the MGCNA has several limitations. Firstly, spatial distance graph learning requires the prior determination of channel locations and is not robust to variations in EEG channel count. If channels are missing or if channel locations change, it can adversely affect recognition performance. Secondly, our model has a relatively long computation time because it involves calculating Pearson correlation coefficients for each sample to extract features for constructing EEG graph representations. Finally, the performance of model may depend on the quality and diversity of the data it has been trained on, and it is crucial to validate it on larger, more diverse datasets to ensure its generalizability. Specifically, both functional connectivity graph learning and spatial distance graph learning employ thresholds.</p>
<p>In future work, it is essential to explore graph representation methods that are more suitable for epileptic signals. While many scholars have already employed various graph representations (<xref ref-type="bibr" rid="B41">Raeisi et al., 2022</xref>), these often involve manual feature engineering and have long computation times. In the future, it&#x2019;s possible to explore alternative composition methods or utilize clustering algorithms to identify the most suitable adjacency matrix. In the future, we will explore alternative graph generation techniques, such as imposing appropriate constraints on trainable adjacency matrix or using clustering algorithms to identify the most suitable graph generator. Identifying the most suitable graph representation method for epileptic signal recognition is of paramount importance. In this study, patient-independent experiments hold more clinical relevance, and there is significant room for improvement in accuracy. To capture common seizure characteristics among different patients, techniques such as transfer learning will be considered to enhance the accuracy of patient-independent experiments. In current research, considerable attention has been devoted to the occurrence of seizures, yet there exist variations in seizure types among individuals. Consequently, in future investigations, emphasis will be placed on the analysis of seizure types in epilepsy research.</p>
</sec>
</sec>
<sec sec-type="conclusion" id="s5">
<title>5 Conclusion</title>
<p>This study proposes a children epilepsy detection model named as MGCNA that combines a multi-branch GCN with multi-head attention. The MGCNA leverages three graph structures to learn spatiotemporal features among channels. It uses spatial graph representations to capture spatial distances between channels, functional connections between channels to learn spatial dependencies in the signals, and employs learnable graph representations to complement spatiotemporal features. The model employs a multi-head attention to assign importance weights to graph signals, learning relationships between graph representations. The model&#x2019;s performance in classifying epileptic EEG signals is validated on the CHB-MIT dataset through patient-specific and patient-independent experiments. The experimental results indicate that the MGCNA shows excellent performance of childhood seizure detection surpassing other existing methods. This method can be used to assist in the childhood seizure detection and effectively reduce the workload of physicians. The EEG classification algorithm introduced in this research provides the potential to establish an EEG monitoring system for children with epilepsy.</p>
</sec>
</body>
<back>
<sec sec-type="data-availability" id="s7">
<title>Data availability statement</title>
<p>The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding authors.</p>
</sec>
<sec sec-type="author-contributions" id="s8">
<title>Author contributions</title>
<p>YL: Data curation, Investigation, Methodology, Writing&#x2013;original draft, Writing&#x2013;review and editing, Validation. YY: Funding acquisition, Methodology, Resources, Writing&#x2013;original draft. SS: Investigation, Resources, Supervision, Writing&#x2013;review and editing. HW: Funding acquisition, Methodology, Writing&#x2013;review and editing. MS: Investigation, Methodology, Writing&#x2013;review and editing, Validation. XL: Investigation, Methodology, Writing&#x2013;review and editing. PZ: Methodology, Writing&#x2013;review and editing. BW: Methodology, Writing&#x2013;review and editing. NW: Methodology, Writing&#x2013;review and editing. QS: Methodology, Writing&#x2013;review and editing. ZH: Methodology, Writing&#x2013;review and editing.</p>
</sec>
<sec sec-type="funding-information" id="s9">
<title>Funding</title>
<p>The author(s) declare that financial support was received for the research, authorship, and/or publication of this article. This study was supported by the Shandong Provincial Natural Science Foundation of China, grant numbers ZR2019ZD01, ZR2020MF027, and ZR2020MF143.</p>
</sec>
<sec sec-type="COI-statement" id="s10">
<title>Conflict of interest</title>
<p>Authors MS and XL were employed by Neusoft Medical Systems Co., Ltd.</p>
<p>The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="disclaimer" id="s11">
<title>Publisher&#x2019;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<ref-list>
<title>References</title>
<ref id="B1">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Abbas</surname>
<given-names>A. K.</given-names>
</name>
<name>
<surname>Azemi</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Ravanshadi</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Omidvarnia</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>An eeg-based methodology for the estimation of functional brain connectivity networks: application to the analysis of newborn eeg seizure</article-title>. <source>Biomed. Signal Process. Control</source> <volume>63</volume>, <fpage>102229</fpage>. <pub-id pub-id-type="doi">10.1016/j.bspc.2020.102229</pub-id>
</citation>
</ref>
<ref id="B2">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Abu</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Diamant</surname>
<given-names>R.</given-names>
</name>
</person-group> (<year>2023</year>). <article-title>Underwater object classification combining sas and transferred optical-to-sas imagery</article-title>. <source>arXiv Prepr. arXiv:2304.11875</source> <volume>144</volume>, <fpage>109868</fpage>. <pub-id pub-id-type="doi">10.1016/j.patcog.2023.109868</pub-id>
</citation>
</ref>
<ref id="B3">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Cai</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Wu</surname>
<given-names>Q.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>Y.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>Motor imagery decoding in the presence of distraction using graph sequence neural networks</article-title>. <source>IEEE Trans. Neural Syst. Rehabilitation Eng.</source> <volume>30</volume>, <fpage>1716</fpage>&#x2013;<lpage>1726</lpage>. <pub-id pub-id-type="doi">10.1109/TNSRE.2022.3183023</pub-id>
</citation>
</ref>
<ref id="B4">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Cao</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Feng</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Zheng</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Cui</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Zhao</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Jiang</surname>
<given-names>T.</given-names>
</name>
<etal/>
</person-group> (<year>2022</year>). <article-title>Two-stream attention 3-d deep network-based childhood epilepsy syndrome classification</article-title>. <source>IEEE Trans. Instrum. Meas.</source> <volume>72</volume>, <fpage>1</fpage>&#x2013;<lpage>12</lpage>. <pub-id pub-id-type="doi">10.1109/tim.2022.3220287</pub-id>
</citation>
</ref>
<ref id="B6">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Chen</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Huang</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Debattista</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Han</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2024a</year>). <article-title>Dynamic contrastive learning guided by class confidence and confusion degree for medical image segmentation</article-title>. <source>Pattern Recognit.</source> <volume>145</volume>, <fpage>109881</fpage>. <pub-id pub-id-type="doi">10.1016/j.patcog.2023.109881</pub-id>
</citation>
</ref>
<ref id="B7">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Chen</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Xu</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Bian</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Qin</surname>
<given-names>X.</given-names>
</name>
</person-group> (<year>2024b</year>). <article-title>Eeg emotion recognition based on ordinary differential equation graph convolutional networks and dynamic time wrapping</article-title>. <source>Appl. Soft Comput.</source> <volume>152</volume>, <fpage>111181</fpage>. <pub-id pub-id-type="doi">10.1016/j.asoc.2023.111181</pub-id>
</citation>
</ref>
<ref id="B8">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Cheng</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Yu</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Feng</surname>
<given-names>L.</given-names>
</name>
</person-group> (<year>2023</year>). <article-title>Hybrid network using dynamic graph convolution and temporal self-attention for eeg-based emotion recognition</article-title>. <source>IEEE Trans. Neural Netw. Learn. Syst.</source>, <fpage>1</fpage>&#x2013;<lpage>11</lpage>. <pub-id pub-id-type="doi">10.1109/tnnls.2023.3319315</pub-id>
</citation>
</ref>
<ref id="B9">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Cimr</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Fujita</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Tomaskova</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Cimler</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Selamat</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2023</year>). <article-title>Automatic seizure detection by convolutional neural networks with computational complexity analysis</article-title>. <source>Comput. Methods Programs Biomed.</source> <volume>229</volume>, <fpage>107277</fpage>. <pub-id pub-id-type="doi">10.1016/j.cmpb.2022.107277</pub-id>
</citation>
</ref>
<ref id="B10">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Cui</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Xia</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Tu</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Liang</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2017</year>). <article-title>Automatic classification of epileptic electroencephalogram based on multiscale entropy and extreme learning machine</article-title>. <source>J. Med. Imaging Health Inf.</source> <volume>7</volume>, <fpage>949</fpage>&#x2013;<lpage>955</lpage>. <pub-id pub-id-type="doi">10.1166/jmihi.2017.2121</pub-id>
</citation>
</ref>
<ref id="B11">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Cui</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Cao</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Hu</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Jiang</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Gao</surname>
<given-names>F.</given-names>
</name>
</person-group> (<year>2022a</year>). <article-title>Regional scalp eegs analysis and classification on typical childhood epilepsy syndromes</article-title>. <source>IEEE Trans. Cognitive Dev. Syst.</source> <volume>15</volume>, <fpage>662</fpage>&#x2013;<lpage>674</lpage>. <pub-id pub-id-type="doi">10.1109/tcds.2022.3175636</pub-id>
</citation>
</ref>
<ref id="B12">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Cui</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Hu</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Lin</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Cao</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Lai</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>T.</given-names>
</name>
<etal/>
</person-group> (<year>2022b</year>). <article-title>Deep feature fusion based childhood epilepsy syndrome classification from electroencephalogram</article-title>. <source>Neural Netw.</source> <volume>150</volume>, <fpage>313</fpage>&#x2013;<lpage>325</lpage>. <pub-id pub-id-type="doi">10.1016/j.neunet.2022.03.014</pub-id>
</citation>
</ref>
<ref id="B13">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Defferrard</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Bresson</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Vandergheynst</surname>
<given-names>P.</given-names>
</name>
</person-group> (<year>2016</year>). <article-title>Convolutional neural networks on graphs with fast localized spectral filtering</article-title>. <source>Adv. neural Inf. Process. Syst.</source> <volume>29</volume>.</citation>
</ref>
<ref id="B14">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Deng</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Song</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Qian</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>X.</given-names>
</name>
</person-group> (<year>2023</year>). <article-title>Eeg-based seizure prediction via hybrid vision transformer and data uncertainty learning</article-title>. <source>Eng. Appl. Artif. Intell.</source> <volume>123</volume>, <fpage>106401</fpage>. <pub-id pub-id-type="doi">10.1016/j.engappai.2023.106401</pub-id>
</citation>
</ref>
<ref id="B15">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ding</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Nie</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Yuan</surname>
<given-names>Q.</given-names>
</name>
</person-group> (<year>2023</year>). <article-title>Compact convolutional neural network with multi-headed attention mechanism for seizure prediction</article-title>. <source>Int. J. Neural Syst.</source> <volume>33</volume>, <fpage>2350014</fpage>. <pub-id pub-id-type="doi">10.1142/S0129065723500144</pub-id>
</citation>
</ref>
<ref id="B16">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Dong</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Zhao</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Xue</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Chu</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>He</surname>
<given-names>J.</given-names>
</name>
<etal/>
</person-group> (<year>2022</year>). <article-title>Attention-based graph resnet with focal loss for epileptic seizure detection</article-title>. <source>J. Ambient Intell. Smart Environ.</source> <volume>14</volume>, <fpage>61</fpage>&#x2013;<lpage>73</lpage>. <pub-id pub-id-type="doi">10.3233/ais-210086</pub-id>
</citation>
</ref>
<ref id="B17">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ein Shoka</surname>
<given-names>A. A.</given-names>
</name>
<name>
<surname>Alkinani</surname>
<given-names>M. H.</given-names>
</name>
<name>
<surname>El-Sherbeny</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>El-Sayed</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Dessouky</surname>
<given-names>M. M.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>Automated seizure diagnosis system based on feature extraction and channel selection using eeg signals</article-title>. <source>Brain Inf.</source> <volume>8</volume>, <fpage>1</fpage>&#x2013;<lpage>16</lpage>. <pub-id pub-id-type="doi">10.1186/s40708-021-00123-7</pub-id>
</citation>
</ref>
<ref id="B18">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Feng</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Zheng</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Cui</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Jiang</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Gao</surname>
<given-names>F.</given-names>
</name>
<etal/>
</person-group> (<year>2022</year>). <article-title>3d residual-attention-deep-network-based childhood epilepsy syndrome classification</article-title>. <source>Knowledge-Based Syst.</source> <volume>248</volume>, <fpage>108856</fpage>. <pub-id pub-id-type="doi">10.1016/j.knosys.2022.108856</pub-id>
</citation>
</ref>
<ref id="B19">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Grattarola</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Livi</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Alippi</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Wennberg</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Valiante</surname>
<given-names>T. A.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>Seizure localisation with attention-based graph neural networks</article-title>. <source>Expert Syst. Appl.</source> <volume>203</volume>, <fpage>117330</fpage>. <pub-id pub-id-type="doi">10.1016/j.eswa.2022.117330</pub-id>
</citation>
</ref>
<ref id="B20">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>He</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Cui</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Xue</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Chu</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Zhao</surname>
<given-names>Y.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>Spatial&#x2013;temporal seizure detection with graph attention network and bi-directional lsm architecture</article-title>. <source>Biomed. Signal Process. Control</source> <volume>78</volume>, <fpage>103908</fpage>. <pub-id pub-id-type="doi">10.1016/j.bspc.2022.103908</pub-id>
</citation>
</ref>
<ref id="B21">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ho</surname>
<given-names>T. K. K.</given-names>
</name>
<name>
<surname>Armanfard</surname>
<given-names>N.</given-names>
</name>
</person-group> (<year>2023</year>). <article-title>Self-supervised learning for anomalous channel detection in eeg graphs: application to seizure analysis</article-title>. <source>Proc. AAAI Conf. Artif. Intell.</source> <volume>37</volume>, <fpage>7866</fpage>&#x2013;<lpage>7874</lpage>. <pub-id pub-id-type="doi">10.1609/aaai.v37i7.25952</pub-id>
</citation>
</ref>
<ref id="B22">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ji</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Wen</surname>
<given-names>P.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>Jumping knowledge based spatial-temporal graph convolutional networks for automatic sleep stage classification</article-title>. <source>IEEE Trans. Neural Syst. Rehabilitation Eng.</source> <volume>30</volume>, <fpage>1464</fpage>&#x2013;<lpage>1472</lpage>. <pub-id pub-id-type="doi">10.1109/tnsre.2022.3176004</pub-id>
</citation>
</ref>
<ref id="B23">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Jia</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Lin</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Ning</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>He</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Zhou</surname>
<given-names>R.</given-names>
</name>
<etal/>
</person-group> (<year>2021</year>). <article-title>Multi-view spatial-temporal graph convolutional networks with domain generalization for sleep stage classification</article-title>. <source>IEEE Trans. Neural Syst. Rehabilitation Eng.</source> <volume>29</volume>, <fpage>1977</fpage>&#x2013;<lpage>1986</lpage>. <pub-id pub-id-type="doi">10.1109/TNSRE.2021.3110665</pub-id>
</citation>
</ref>
<ref id="B24">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Jia</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Lin</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Zhou</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Ning</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>He</surname>
<given-names>Y.</given-names>
</name>
<etal/>
</person-group> (<year>2020</year>). <article-title>Graphsleepnet: adaptive spatial-temporal graph convolutional networks for sleep stage classification</article-title>. <source>IJCAI</source> <volume>2021</volume>, <fpage>1324</fpage>&#x2013;<lpage>1330</lpage>. <pub-id pub-id-type="doi">10.24963/ijcai.2020/184</pub-id>
</citation>
</ref>
<ref id="B25">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kalaganis</surname>
<given-names>F. P.</given-names>
</name>
<name>
<surname>Laskaris</surname>
<given-names>N. A.</given-names>
</name>
<name>
<surname>Chatzilari</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>Nikolopoulos</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Kompatsiaris</surname>
<given-names>I.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>A data augmentation scheme for geometric deep learning in personalized brain&#x2013;computer interfaces</article-title>. <source>IEEE access</source> <volume>8</volume>, <fpage>162218</fpage>&#x2013;<lpage>162229</lpage>. <pub-id pub-id-type="doi">10.1109/access.2020.3021580</pub-id>
</citation>
</ref>
<ref id="B26">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kannathal</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Choo</surname>
<given-names>M. L.</given-names>
</name>
<name>
<surname>Acharya</surname>
<given-names>U. R.</given-names>
</name>
<name>
<surname>Sadasivan</surname>
<given-names>P.</given-names>
</name>
</person-group> (<year>2005</year>). <article-title>Entropies for detection of epilepsy in eeg</article-title>. <source>Comput. methods programs Biomed.</source> <volume>80</volume>, <fpage>187</fpage>&#x2013;<lpage>194</lpage>. <pub-id pub-id-type="doi">10.1016/j.cmpb.2005.06.012</pub-id>
</citation>
</ref>
<ref id="B27">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kipf</surname>
<given-names>T. N.</given-names>
</name>
<name>
<surname>Welling</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2016</year>). <article-title>Semi-supervised classification with graph convolutional networks</article-title>
</citation>
</ref>
<ref id="B28">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Lee</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Yu</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Back</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Seo</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Lee</surname>
<given-names>K.</given-names>
</name>
</person-group> (<year>2024</year>). <article-title>Sleepy: automatic sleep scoring with feature pyramid and contrastive learning</article-title>. <source>Expert Syst. Appl.</source> <volume>240</volume>, <fpage>122551</fpage>. <pub-id pub-id-type="doi">10.1016/j.eswa.2023.122551</pub-id>
</citation>
</ref>
<ref id="B29">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Li</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Cheng</surname>
<given-names>Z.</given-names>
</name>
</person-group> (<year>2022a</year>). <article-title>An attention-guided spatiotemporal graph convolutional network for sleep stage classification</article-title>. <source>Life</source> <volume>12</volume>, <fpage>622</fpage>. <pub-id pub-id-type="doi">10.3390/life12050622</pub-id>
</citation>
</ref>
<ref id="B30">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Li</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Qiu</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Zhu</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Kong</surname>
<given-names>W.</given-names>
</name>
</person-group> (<year>2023</year>). <article-title>Feature hypergraph representation learning on spatial-temporal correlations for eeg emotion recognition</article-title>. <source>Cogn. Neurodynamics</source> <volume>17</volume>, <fpage>1271</fpage>&#x2013;<lpage>1281</lpage>. <pub-id pub-id-type="doi">10.1007/s11571-022-09890-3</pub-id>
</citation>
</ref>
<ref id="B31">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Li</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Fu</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Wu</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Ji</surname>
<given-names>Y.</given-names>
</name>
<etal/>
</person-group> (<year>2022b</year>). <article-title>Gmss: graph-based multi-task self-supervised learning for eeg emotion recognition</article-title>. <source>IEEE Trans. Affect. Comput.</source> <volume>14</volume>, <fpage>2512</fpage>&#x2013;<lpage>2525</lpage>. <pub-id pub-id-type="doi">10.1109/taffc.2022.3170428</pub-id>
</citation>
</ref>
<ref id="B32">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Lian</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Xu</surname>
<given-names>F.</given-names>
</name>
</person-group> (<year>2023</year>). <article-title>Epileptic eeg classification via graph transformer network</article-title>. <source>Int. J. neural Syst.</source> <volume>33</volume>, <fpage>2350042</fpage>. <pub-id pub-id-type="doi">10.1142/S0129065723500429</pub-id>
</citation>
</ref>
<ref id="B33">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Liu</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Dong</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Bian</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Zhou</surname>
<given-names>W.</given-names>
</name>
</person-group> (<year>2023</year>). <article-title>Epileptic seizure prediction using attention augmented convolutional network</article-title>. <source>Int. J. Neural Syst.</source> <volume>33</volume>, <fpage>2350054</fpage>. <pub-id pub-id-type="doi">10.1142/S0129065723500545</pub-id>
</citation>
</ref>
<ref id="B34">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Liu</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>Q.</given-names>
</name>
<name>
<surname>Cao</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>Minimum spanning tree based graph neural network for emotion classification using eeg</article-title>. <source>Neural Netw.</source> <volume>145</volume>, <fpage>308</fpage>&#x2013;<lpage>318</lpage>. <pub-id pub-id-type="doi">10.1016/j.neunet.2021.10.023</pub-id>
</citation>
</ref>
<ref id="B35">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Lopez</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Del Percio</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Lizio</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Noce</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Padovani</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Nobili</surname>
<given-names>F.</given-names>
</name>
<etal/>
</person-group> (<year>2023</year>). <article-title>Patients with alzheimer&#x2019;s disease dementia show partially preserved parietal &#x2018;hubs&#x2019; modeled from resting-state alpha electroencephalographic rhythms</article-title>. <source>Front. Aging Neurosci.</source> <volume>15</volume>, <fpage>780014</fpage>. <pub-id pub-id-type="doi">10.3389/fnagi.2023.780014</pub-id>
</citation>
</ref>
<ref id="B36">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Mehta</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Rastegari</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>Separable self-attention for mobile vision transformers</article-title>. <source>arXiv Prepr. arXiv:2206.02680</source>. <pub-id pub-id-type="doi">10.48550/arXiv.2206.02680</pub-id>
</citation>
</ref>
<ref id="B37">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Meng</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Hu</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Deng</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Hu</surname>
<given-names>Y.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>Electrical status epilepticus during sleep electroencephalogram waveform identification and analysis based on a graph convolutional neural network</article-title>. <source>Biomed. Signal Process. Control</source> <volume>77</volume>, <fpage>103788</fpage>. <pub-id pub-id-type="doi">10.1016/j.bspc.2022.103788</pub-id>
</citation>
</ref>
<ref id="B38">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Na</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Lv</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Xu</surname>
<given-names>Z.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>An extended k nearest neighbors-based classifier for epilepsy diagnosis</article-title>. <source>IEEE Access</source> <volume>9</volume>, <fpage>73910</fpage>&#x2013;<lpage>73923</lpage>. <pub-id pub-id-type="doi">10.1109/access.2021.3081767</pub-id>
</citation>
</ref>
<ref id="B39">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Peh</surname>
<given-names>W. Y.</given-names>
</name>
<name>
<surname>Thangavel</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Yao</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Thomas</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Tan</surname>
<given-names>Y.-L.</given-names>
</name>
<name>
<surname>Dauwels</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2023</year>). <article-title>Six-center assessment of conn-transformer with belief matching loss for patient-independent seizure detection in eeg</article-title>. <source>Int. J. Neural Syst.</source> <volume>33</volume>, <fpage>2350012</fpage>. <pub-id pub-id-type="doi">10.1142/S0129065723500120</pub-id>
</citation>
</ref>
<ref id="B40">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Qiu</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Yan</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>H.</given-names>
</name>
</person-group> (<year>2023</year>). <article-title>A difference attention resnet-lstm network for epileptic seizure detection using eeg signal</article-title>. <source>Biomed. Signal Process. Control</source> <volume>83</volume>, <fpage>104652</fpage>. <pub-id pub-id-type="doi">10.1016/j.bspc.2023.104652</pub-id>
</citation>
</ref>
<ref id="B41">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Raeisi</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Khazaei</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Croce</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Tamburro</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Comani</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Zappasodi</surname>
<given-names>F.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>A graph convolutional neural network for the automated detection of seizures in the neonatal eeg</article-title>. <source>Comput. Methods Programs Biomed.</source> <volume>222</volume>, <fpage>106950</fpage>. <pub-id pub-id-type="doi">10.1016/j.cmpb.2022.106950</pub-id>
</citation>
</ref>
<ref id="B42">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Rennie</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Chorley</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Boylan</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Pressler</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Nguyen</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Hooper</surname>
<given-names>R.</given-names>
</name>
</person-group> (<year>2004</year>). <article-title>Non-expert use of the cerebral function monitor for neonatal seizure detection</article-title>. <source>Archives Dis. Childhood-Fetal Neonatal Ed.</source> <volume>89</volume>, <fpage>37</fpage>&#x2013;<lpage>40</lpage>. <pub-id pub-id-type="doi">10.1136/fn.89.1.f37</pub-id>
</citation>
</ref>
<ref id="B43">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Shoeb</surname>
<given-names>A. H.</given-names>
</name>
</person-group> (<year>2009</year>). <article-title>Application of machine learning to epileptic seizure onset detection and treatment</article-title>. <source>Ph.D. thesis, Mass. Inst. Technol</source>.</citation>
</ref>
<ref id="B44">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Singh</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Singh</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Kaur</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>Grasshopper optimization algorithm&#x2013;based approach for the optimization of ensemble classifier and feature selection to classify epileptic eeg signals</article-title>. <source>Med. and Biol. Eng. and Comput.</source> <volume>57</volume>, <fpage>1323</fpage>&#x2013;<lpage>1339</lpage>. <pub-id pub-id-type="doi">10.1007/s11517-019-01951-w</pub-id>
</citation>
</ref>
<ref id="B45">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Song</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Zheng</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Zong</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Cui</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>Y.</given-names>
</name>
<etal/>
</person-group> (<year>2021</year>). <article-title>Variational instance-adaptive graph for eeg emotion recognition</article-title>. <source>IEEE Trans. Affect. Comput.</source> <volume>14</volume>, <fpage>343</fpage>&#x2013;<lpage>356</lpage>. <pub-id pub-id-type="doi">10.1109/taffc.2021.3064940</pub-id>
</citation>
</ref>
<ref id="B46">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Song</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Zheng</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Song</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Cui</surname>
<given-names>Z.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>Eeg emotion recognition using dynamical graph convolutional neural networks</article-title>. <source>IEEE Trans. Affect. Comput.</source> <volume>11</volume>, <fpage>532</fpage>&#x2013;<lpage>541</lpage>. <pub-id pub-id-type="doi">10.1109/taffc.2018.2817622</pub-id>
</citation>
</ref>
<ref id="B5">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Sukriti</surname>
</name>
<name>
<surname>Chakraborty</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Mitra</surname>
<given-names>D.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>A computationally efficient automated seizure detection method based on the novel idea of multiscale spectral features</article-title>. <source>Biomed. Signal Process. Control</source> <volume>70</volume>, <fpage>102990</fpage>. <pub-id pub-id-type="doi">10.1016/j.bspc.2021.102990</pub-id>
</citation>
</ref>
<ref id="B47">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Sun</surname>
<given-names>Q.</given-names>
</name>
<name>
<surname>Yang</surname>
<given-names>Y.</given-names>
</name>
</person-group> (<year>2023</year>). <article-title>Unsupervised video anomaly detection based on multi-timescale trajectory prediction</article-title>. <source>Comput. Vis. Image Underst.</source> <volume>227</volume>, <fpage>103615</fpage>. <pub-id pub-id-type="doi">10.1016/j.cviu.2022.103615</pub-id>
</citation>
</ref>
<ref id="B48">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Thuwajit</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Rangpong</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Sawangjai</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Autthasan</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Chaisaen</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Banluesombatkul</surname>
<given-names>N.</given-names>
</name>
<etal/>
</person-group> (<year>2021</year>). <article-title>Eegwavenet: multiscale conn-based spatiotemporal feature extraction for eeg seizure detection</article-title>. <source>IEEE Trans. Industrial Inf.</source> <volume>18</volume>, <fpage>5547</fpage>&#x2013;<lpage>5557</lpage>. <pub-id pub-id-type="doi">10.1109/tii.2021.3133307</pub-id>
</citation>
</ref>
<ref id="B49">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Wagh</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Varatharajah</surname>
<given-names>Y.</given-names>
</name>
</person-group> (<year>2020</year>). &#x201c;<article-title>Eeg-gcnn: augmenting electroencephalogram-based neurological disease diagnosis using a domain-guided graph convolutional neural network</article-title>,&#x201d; in <source>
<italic>Machine Learning for health</italic> (PMLR)</source>, <fpage>367</fpage>&#x2013;<lpage>378</lpage>.</citation>
</ref>
<ref id="B50">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Xu</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Bezerianos</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>Z.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Linking attention-based multiscale conn with dynamical gcn for driving fatigue detection</article-title>. <source>IEEE Trans. Instrum. Meas.</source> <volume>70</volume>, <fpage>1</fpage>&#x2013;<lpage>11</lpage>. <pub-id pub-id-type="doi">10.1109/tim.2020.3047502</pub-id>
</citation>
</ref>
<ref id="B51">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Gao</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Zheng</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Zhu</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Shi</surname>
<given-names>C.-J. R.</given-names>
</name>
</person-group> (<year>2023a</year>). <article-title>Ssgcnet: a sparse spectra graph convolutional network for epileptic eeg signal classification</article-title>. <source>IEEE Trans. Neural Netw. Learn. Syst.</source> <volume>35</volume>, <fpage>12157</fpage>&#x2013;<lpage>12171</lpage>. <pub-id pub-id-type="doi">10.1109/tnnls.2023.3252569</pub-id>
</citation>
</ref>
<ref id="B52">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Liang</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Wu</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Gao</surname>
<given-names>R.</given-names>
</name>
<etal/>
</person-group> (<year>2023b</year>). <article-title>EEG signal epilepsy detection with a weighted neighbor graph representation and two-stream graph-based framework</article-title>. <source>IEEE Trans. Neural Syst. Rehabilitation Eng.</source> <volume>31</volume>, <fpage>3176</fpage>&#x2013;<lpage>3187</lpage>. <pub-id pub-id-type="doi">10.1109/tnsre.2023.3299839</pub-id>
</citation>
</ref>
<ref id="B53">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Cui</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Yu</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Liao</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>Y.</given-names>
</name>
</person-group> (<year>2023c</year>). <article-title>Dynamic multi-graph convolution based channel-weighted transformer feature fusion network for epileptic seizure prediction</article-title>. <source>IEEE Trans. Neural Syst. Rehabilitation Eng.</source> <volume>31</volume>, <fpage>4266</fpage>&#x2013;<lpage>4277</lpage>. <pub-id pub-id-type="doi">10.1109/tnsre.2023.3321414</pub-id>
</citation>
</ref>
<ref id="B54">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Shi</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Cheng</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>He</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Wei</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>Z.</given-names>
</name>
<etal/>
</person-group> (<year>2022</year>). <article-title>A spatiotemporal graph attention network based on synchronization for epileptic seizure prediction</article-title>. <source>IEEE J. Biomed. Health Inf.</source> <volume>27</volume>, <fpage>900</fpage>&#x2013;<lpage>911</lpage>. <pub-id pub-id-type="doi">10.1109/JBHI.2022.3221211</pub-id>
</citation>
</ref>
<ref id="B55">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Hou</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Xiao</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Lv</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>J.</given-names>
</name>
<etal/>
</person-group> (<year>2023d</year>). <article-title>Lightweight seizure detection based on multi-scale channel attention</article-title>. <source>Int. J. Neural Syst.</source> <volume>33</volume>, <fpage>2350061</fpage>. <pub-id pub-id-type="doi">10.1142/S0129065723500612</pub-id>
</citation>
</ref>
<ref id="B56">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wei</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Zou</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Xu</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>Automatic epileptic eeg detection using convolutional neural network with improvements in time-domain</article-title>. <source>Biomed. Signal Process. Control</source> <volume>53</volume>, <fpage>101551</fpage>. <pub-id pub-id-type="doi">10.1016/j.bspc.2019.04.028</pub-id>
</citation>
</ref>
<ref id="B57">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wu</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Yu</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Zhou</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Wu</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Su</surname>
<given-names>S.</given-names>
</name>
</person-group> (<year>2024</year>). <article-title>Time-series anomaly detection based on dynamic temporal graph convolutional network for epilepsy diagnosis</article-title>. <source>Bioengineering</source> <volume>11</volume>, <fpage>53</fpage>. <pub-id pub-id-type="doi">10.3390/bioengineering11010053</pub-id>
</citation>
</ref>
<ref id="B58">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Xiao</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Feng</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Zhao</surname>
<given-names>Y.</given-names>
</name>
<etal/>
</person-group> (<year>2024</year>). <article-title>Self-supervised learning with attention mechanism for eeg-based seizure detection</article-title>. <source>Biomed. Signal Process. Control</source> <volume>87</volume>, <fpage>105464</fpage>. <pub-id pub-id-type="doi">10.1016/j.bspc.2023.105464</pub-id>
</citation>
</ref>
<ref id="B59">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Xu</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Lin</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Xu</surname>
<given-names>T.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>Epilepsy seizures prediction based on nonlinear features of eeg signal and gradient boosting decision tree</article-title>. <source>Int. J. Environ. Res. Public Health</source> <volume>19</volume>, <fpage>11326</fpage>. <pub-id pub-id-type="doi">10.3390/ijerph191811326</pub-id>
</citation>
</ref>
<ref id="B60">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhang</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Jiang</surname>
<given-names>N.</given-names>
</name>
</person-group> (<year>2010</year>). <article-title>Multi-scale phase average waveform of electroencephalogram signals in childhood absence epilepsy using wavelet transformation</article-title>. <source>NEURAL Regen. Res.</source> <volume>5</volume>, <fpage>774</fpage>&#x2013;<lpage>780</lpage>. <pub-id pub-id-type="doi">10.3969/j.issn.1673-5374.2010.10.010</pub-id>
</citation>
</ref>
<ref id="B61">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhao</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Zheng</surname>
<given-names>Q.</given-names>
</name>
<name>
<surname>Ding</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Yang</surname>
<given-names>Y.</given-names>
</name>
</person-group> (<year>2022a</year>). <article-title>A high-dimensional and small-sample submersible fault detection method based on feature selection and data augmentation</article-title>. <source>Sensors</source> <volume>22</volume>, <fpage>204</fpage>. <pub-id pub-id-type="doi">10.3390/s22010204</pub-id>
</citation>
</ref>
<ref id="B62">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhao</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Peng</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Niu</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>He</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Yang</surname>
<given-names>F.</given-names>
</name>
<etal/>
</person-group> (<year>2022b</year>). <article-title>A multi-head self-attention deep learning approach for detection and recommendation of neuromagnetic high frequency oscillations in epilepsy</article-title>. <source>Front. Neuroinformatics</source> <volume>16</volume>, <fpage>771965</fpage>. <pub-id pub-id-type="doi">10.3389/fninf.2022.771965</pub-id>
</citation>
</ref>
<ref id="B63">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhao</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Chu</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>He</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Xue</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Jia</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Xu</surname>
<given-names>F.</given-names>
</name>
<etal/>
</person-group> (<year>2023a</year>). <article-title>Interactive local and global feature coupling for eeg-based epileptic seizure detection</article-title>. <source>Biomed. Signal Process. Control</source> <volume>81</volume>, <fpage>104441</fpage>. <pub-id pub-id-type="doi">10.1016/j.bspc.2022.104441</pub-id>
</citation>
</ref>
<ref id="B64">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhao</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>He</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Zhu</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Xiao</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>Z.</given-names>
</name>
<etal/>
</person-group> (<year>2023b</year>). <article-title>Hybrid attention network for epileptic eeg classification</article-title>. <source>Int. J. Neural Syst.</source> <volume>33</volume>, <fpage>2350031</fpage>. <pub-id pub-id-type="doi">10.1142/S0129065723500314</pub-id>
</citation>
</ref>
</ref-list>
</back>
</article>