Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Neurol., 29 August 2025

Sec. Artificial Intelligence in Neurology

Volume 16 - 2025 | https://doi.org/10.3389/fneur.2025.1626922

PlgFormer: parallel extraction of local-global features for AD diagnosis on sMRI using a unified CNN-transformer architecture


Guoxin Wang&#x;Guoxin Wang1Yuxia Li&#x;Yuxia Li2Zhiyi Zhou&#x;Zhiyi Zhou3Shan An
Shan An4*Xuyang CaoXuyang Cao5Yuxin JinYuxin Jin5Zhengqin SunZhengqin Sun2Guanqun ChenGuanqun Chen6Mingkai ZhangMingkai Zhang7Zhixiong Li
Zhixiong Li8*Feng Yu
Feng Yu1*
  • 1College of Biomedical Engineering and Instrument Science, Zhejiang University, Hangzhou, China
  • 2Tangshan Central Hospital, Tangshan, Hebei, China
  • 3Electrical and Information Engineering, Tianjin University, Tianjin, China
  • 4School of Electrical and Information Engineering, Tianjin University, Tianjin, China
  • 5JD Health International Inc., Beijing, China
  • 6Department of Neurology, Beijing Chao-Yang Hospital, Capital Medical University, Beijing, China
  • 7Department of Neurology, XuanWu Hospital of Capital Medical University, Beijing, China
  • 8Karamay Integrated Traditional Chinese and Western Medicine Hospital (People's Hospital of Karamay), Karamay, China

Introduction: Structural magnetic resonance imaging (sMRI) is an important tool for the early diagnosis of Alzheimer's disease (AD). Previous methods based on voxel, region of interests (ROIs) or patch have limitations in characterizing discriminative features in sMRI for AD as they can only focus on specific local or global features.

Methods: We propose a computer-aided AD diagnosis method based on sMRI, named PlgFormer, which considers the extraction of both local and global features. By using a combination of convolution and self-attention, we can extract context features at both local and global levels. In the decision-making layer of the model, we design a feature fusion module that adaptively selects context features through a gating mechanism. Additionally, to account for changes in image input resolution during the downsampling operation, we embed a dynamic embedding block at each stage of the network, which can adaptively adjust the weights of the inputs with different resolutions.

Results: We evaluated the performance of our method on dichotomous AD vs. normal control (NC) and mild cognitive impairment (MCI) vs. NC, as well as trichotomous AD vs. MCI vs. NC classification tasks, using publicly available ADNI and XWNI datasets that we collected. On the ADNI dataset, the proposed method achieves classification accuracies of 0.9431 for AD vs. NC, 0.8216 for MCI vs. CN, and 0.6228 for the AD vs. MCI vs. CN task. On the XWNI dataset, the corresponding accuracies are 0.9307, 0.8600, and 0.8672, respectively. The experimental results demonstrate the high precision and robustness of our method in diagnosing people with different stages of cognitive impairment.

Conclusion: The findings in our experimental results underscore the clinical potential of our proposed PlgFormer as a reliable and interpretable framework for supporting early and accurate diagnosis of AD using sMRI.

1 Introduction

Alzheimer's disease (AD) is a neurodegenerative disorder that progressively impairs cognitive functions such as memory, thinking, and behavior. It is the most common cause of dementia and affects the normal lives of over 30 million people worldwide (1), primarily older adults. As the disease advances, individuals may experience difficulties with language, disorientation, mood changes, and eventually lose the ability to perform daily activities independently. These challenges not only degrade the quality of life but also place a heavy burden on families and healthcare systems. Early and accurate diagnosis of AD is therefore essential for effective intervention and management.

Diagnosis AD is based primarily on clinical evaluation, including medical history, physical examination, and cognitive tests (2, 3). Unfortunately, effective treatments for AD have yet to be discovered. Therefore, early diagnosis is imperative not only for improving the quality of life for patients but also for supporting the development of more effective treatment methods and intervention measures.

Structural magnetic resonance imaging (sMRI) is a noninvasive medical imaging technique that captures the physical structure of the brain and provides detailed images of its tissues, including gray and white matter. Unlike functional magnetic resonance imaging (fMRI), which measures changes in blood flow and neural activity, sMRI uses strong magnetic fields and radio waves. sMRI has become a valuable tool in diagnosing AD due to its ability to detect changes in brain structure associated with the disease (47), such as shrinkage in the hippocampus and other areas of the brain. However, its use is often limited by physician experience and time-consuming processes.

To solve this dilemma, hopes have gradually been placed on computer-aided diagnosis of AD, which has a long history and has achieved good performance. Existing methods for computer-aided diagnosis of AD using sMRI can be broadly classified into three categories: (1) Voxel-based methods; (2) Regions of interest (ROIs)-based methods; and (3) Patch-based methods. However, these methods face challenges due to the high dimensionality of sMRI data and the discrete distribution of AD lesions.

Voxel-based methods for computer-aided diagnosis of AD utilize the entire sMRI as input and extract global features for AD diagnosis. Hinrichs et al. (8) used gray matter density to extract discriminative features and employed a linear programming boosting method to classify AD and normal control (NC). Kao et al. (9) detected white matter changes throughout the whole sMRI to diagnose AD. Vounou et al. (10) detected markers associated with longitudinal changes in brain voxels caused by AD and used a sparse reduced-rank regression model to classify them. However, these methods extract features across the whole sMRI, resulting in high computational complexity and overfitting of the model due to limited data.

In contrast, ROI-based methods focus on extracting features from pre-segmented regions with lower feature dimensionality compared to voxel-based methods. Zhang et al. (11) adaptively extracted 93 ROIs using the atlas warping algorithm and identified AD with support vector machines (SVM). Liu et al. (12) proposed an ensemble classification model to construct gray matter density features within regions by using multiple spatially normalized templates. However, ROI selection is heavily dependent on specialist knowledge, making it difficult for developers to master. Additionally, ROI-based methods can only capture local discriminative features, making it difficult to capture global features such as ventricular volume and gyral sulcus morphological changes, which is insufficient for computer-aided AD diagnosis.

Patch-based methods extract features with intermediate feature dimensions, focusing more effectively on local discriminative features. Qiu et al. (13) trained a FCN to adaptively and randomly select meaningful patches, which were fed to a multilayer perceptron (MLP) for individual-level AD diagnosis. Zhang et al. (14) selected discriminative patches by computing shapley values and extracted local-global context features in sMRI using Convolutional Neural Network (CNN). However, how to combine local patch features with global representations is still a problem that needs to be explored.

We propose PlgFormer, a new method for AD diagnosis using sMRI. PlgFormer utilizes convolution modules and transformer modules in parallel to extract discriminative local-global context features in a unified manner. To adaptively adjust parameter size for inputs of different sizes, we embed dynamic convolutional layers in early stages of the model. This operation also reduces the number of parameters 72 compared to conventional convolution, as shown in Figure 1.

Figure 1
Scatter plot illustrating model accuracy versus parameters in megabytes. Each point represents a model, differentiated by color and shape according to a legend. Accuracy ranges from 86% to 96%, and parameters range from 0 to 70MB. The PlgFormer model, represented by a pink circle, achieves the highest accuracy of approximately 95% with 4.36MB. The largest model, 3D-ResNet-34, is marked by a large green triangle with 60.53MB.

Figure 1. Various existing models were evaluated on ADNI dataset for the binary classification of AD and CN, and their performance was compared in relation to the number of model parameters. The graphical area corresponded to the number of model parameters. Our PlgFormer achieved optimal classification performance with the lightest number of parameters. The classification performance is illustrated in Table 1.

In summary, the major contributions of this paper can be summarized as follows.

1. We design a novel dynamic embedding block (DEB) in our model, which is a combination of traditional convolution and dynamic convolution. Compared to traditional convolution, introducing dynamic convolution can help the model adaptively adjust the size of parameters based on the size of input, which reduces the model parameters while enhancing its robustness.

2. A dual-branch structure extracts local-global context features in sMRI in parallel. Considering feature alignment, we design both dual-branch operations as a multi-head architecture. Additionally, we introduce a feature fusion module (F2M) that adaptively selects local or global features at the decision-making layer of the model.

3. We collected a large-scale sMRI dataset called the XWNI dataset. Our extensive experiments, which included both the publicly available ADNI dataset and the collected XWNI dataset, show that our proposed PlgFormer substantially outperforms the existing baseline models while utilizing minimal parameters.

2 Related works

In this section, we provide a summary of current methods for computer-aided diagnosis of AD using sMRI data.

2.1 Traditional machine learning for AD diagnosis

Early methods for computer-aided diagnosis of AD using sMRI data mainly focused on extracting features and using traditional machine learning methods to analyze and classify them. Ashburner and Friston (15) compared differences in brain structure between individuals or groups to identify brain regions associated with specific diseases or cognitive functions. Klöppel et al. (16) transformed brain MRI into statistical features and employed SVM to identify early structural changes in the brain of AD patients. Similarly, Fan et al. (17) split MRI into specified regions and selected the most discriminative regions for AD diagnosis based on the statistical characteristics of each region and the inter-relationship between them, and performed the classification using SVM. Hinrichs et al. (8) randomly augmented the raw ADNI data and adopted a linear programming boosting method for classification, achieving more robust results. sMRI data are always over dimensional. Cao et al. (18) proposed a multi-kernel-based method combined with marginal fisher analysis to reduce the feature dimensionality of MRI and establish a complex mapping relationship from image to disease. Zhang et al. (11) used principal component analysis (PCA) for feature extraction and dimensionality reduction and conducted classification of AD and CN instances using SVM. Abuhmed et al. (19) proposed two novel hybrid deep learning models, DFBL and MRBL, which integrate multivariate BiLSTM architecture with traditional machine learning models to enhance the prediction of Alzheimer's disease progression using multimodal time-series data.

However, these methods mentioned above present two limitations: (1) manual feature extraction is reliant on human experience and may ignore discriminative features; (2) high-dimensional features may lead to overfitting.

2.2 CNN-based methods for AD diagnosis

CNNs have achieved considerable amount of success in the field of computer vision, with their strong inductive bias performs well in tasks including image classification (2022), object detection, semantic segmentation, and video understanding. Naturally, these CNN-based methods were applied to AD diagnosis using MRI data, achieving satisfactory performances. Lian et al. (23) proposed a hierarchical full convolutional network (H-FCN) to automatically identify discriminative local patches associated with AD in the whole brain sMRI. Zhu et al. (24) proposed a dual attention multi-instance deep learning network to extract discriminative features from local patches and aggregate these features by attention-aware weights. Wu et al. (25) proposed a 3D CNN model to extract and integrate robust multiscale spatial features to promote AD computer-aided diagnosis. Zhang et al. (26) proposed a residual self-attention deep neural network to capture local spatial features in sMRI, while attention mechanisms are still necessary to jointly construct global representations. A multi-branch convolutional network was proposed in (27), where each branch extracts its own features independently and uses a fully connected layer for feature aggregation to obtain global representations. Zhang et al. (28) proposed a multi-relation reasoning network (MRN) that constructs brain graphs from sMRI data to capture spatial and topological relationships, improving Alzheimer's disease diagnosis through enhanced feature representation and global reasoning.

In addition to specialized CNN-based models, general models such as VGGNet (21) and ResNet (22) have also been applied to AD diagnosis due to their successful performance in natural visual scene analysis. Billones et al. (29) utilized a modified VGGNet model for AD diagnosis using sMRI data. Li et al. (30) employed the ResNet model to the ADNI dataset, focusing specifically on local discriminative features in hippocampal regions. These methods all use the strong inductive bias of CNNs to extend the conventional 2D convolutional kernels to 3D and extract the local representations associated with AD in sMRI.

The strong inductive bias of CNNs enables them to capture local features effectively even with a limited number of samples. However, when constructing global representations by integrating local information, attention mechanisms are typically still required.

2.3 Attention mechanism-based methods for AD diagnosis

Attention mechanisms excel at uniting different local features to construct global representations. Thanks to this, Wu et al. (25) aggregated the local features extracted by the convolution module under the supervision of the attention weights. similarly, Zhang et al. (26) directed the network to focus on critical information in the CNN feature maps and suppress non-essential features by using self-attention scores. Liu et al. (31) aggregated the semantic features between different channels by the classical Squeeze and Excitation (SE) module. Zhang et al. (32) embedded a connection-wise attention mechanism in DesNet for associating context features of the model. Jin et al. (33) constructed a simple convolutional layer for computing attention weights to retrace global features for local features extracted by convolution. Transformer is a pure attention mechanism architecture, originally proposed for natural language processing tasks (34) and has since been extended to the field of computer vision (35, 36) demonstrating progressively satisfactory performance.

However, Jang and Hwang (37) have experimentally demonstrated that the coarse embedding of Transformer into a mature CNN architecture may not always yield satisfactory performances due to its lack of focus on local features, implying that attention may not be all need for AD diagnosis with sMRI data.

3 Method

In this section, we present the details of the proposed method. The overview of the method is elaborated in Section 3.1. After that, dynamic embedding block, convolution and self-attention parallel extraction of local-global context features and the F2M are described in Section 3.2, Section 2.3, and Section 3.4, respectively.

3.1 Overview of the proposed method

The strong inductive bias of the convolution is beneficial for local features extraction, while the self-attention is significant for the representation of global features. Both local and global features are important for AD diagnosis from sMRI, which means we cannot simply conduct convolution or self-attention operations. To achieve this, we propose an architecture that extracts local-global context features in parallel with a uniform multi-head self-attention mechanism. Our designed F2M in the decision-making layer of the network selects discriminative local-global context features through a gating mechanism. Additionally, since downsampling changes the input resolution at each stage, we embed a dynamic embedding block in each stage of the network to adjust the weights for different input resolutions adaptively.

Figure 2 provides an overview of our proposed PlgFormer algorithm. Simply described, we take a vanilla 3-D sMRI X∈ℝD × H × W as an example. The training process of PlgFormer is depicted in Algorithm 1. The proposed PlgFormer consists of multi-stages, each of which conducts a downsampling operation to reduce the dimensionality of the data. To adapt to the change of input size, DEBs are introduced in each stage of the model to adaptively adjust the weights of convolution kernels for various resolutions. In the early stages of the model, Multi-Head-Self-Attention (MHSA) was introduced in a purely convolutional operation paired with multi-head style, without considering the computationally burdensome self-attention mechanism due to the large input size. Local and global discriminative features are then extracted using the convolution and self-attentive modules. To facilitate feature alignment before fusion, the convolution operation is designed as a multi-head structure unified with the self-attention mechanism. Local-global context features are fused with F2M and passed through global average pooling and full connectivity to obtain the one-hot tensor at the decision-making layer of the network.

Figure 2
Flowchart showing a neural network architecture for image processing. It begins with a stack of four grayscale brain scans labeled as dimensions 1  ×  D × H × W. These pass through four stages, each with patch embedding and MHS-A with FFN components. The stages progressively halve the dimensions from 16×D4×H4×W4 to 128×D32×H32×W32. Outputs are combined and fed into a fully connected layer for analysis.

Figure 2. Overview of the proposed PlgFormer. This method comprises four stages of feature extraction. Each stage employs dynamic convolutional style to encode location information, MHSAl or MHSAg to extract local or global features, and a feature fusion module to combine these features and produce classification results. For further elaboration, please refer to Section 3. The figure schematically illustrates the data dimensions as a binary classification of AD and CN.

Algorithm 1
www.frontiersin.org

Algorithm 1. The training procedure of PlgFormer.

The key factors of our proposed method for effective AD diagnosis include: (1) how to encode dynamic positions for inputs of different resolutions; (2) how to extract local and global features in a uniform manner using convolution and self-attention operations for subsequent feature alignment; and (3) how to effectively fuse local-global context features. These solutions will be elaborated in Sections 3.2–3.4.

3.2 Dynamic embedding block

Dynamic convolution can adjust the size of the convolution kernel according to the different sizes of the input data, effectively reducing the parameter count of the model. Unlike traditional convolutions that only use fixed kernel sizes, this type of operator also improves the robustness of convolution nerual nerworks, as the model can more quickly focus on discriminative features. Before executing the multi-head attention in each stage, the DEB uses a regular convolution operation to non-overlappingly divide the feature map of the original input into many patches (kernel_size, stride = patch_size). As shown in Figure 3, before and after this regular convolution operation, the DEB introduces a dynamic convolutional layer to guide the subsequent multi-head module to focus on more meaningful features:

H=DEB(X),    (1)

Where DEB represents a single dynamic embedding block, consisting of two dynamic convolution layers and a vanilla convolution layer with a stride of 2 or 4 for downsampling. We introduce this block to integrate of all tokens before feeding them to the multi-headed self-attentive module. Such a design combining dynamic convolution has two benefits. First, the dynamic convolution operation is very friendly to varying input resolutions. Second, dynamic convolution is light-weight, which can largely alleviate the over-fitting problem encountered because of the small amount of data.

Figure 3
Diagram of a Dynamic Embedding Block showing a sequence of processes: Dynamic Conv 3D with GELU and batch normalization, Conv 3D with stride equal to patch size, another Dynamic Conv 3D with GELU and batch normalization, followed by Flatten, Layer Norm, and Reshape. Inputs are X and outputs are H.

Figure 3. Detailed structure of the proposed DEB. The kernel size of 3D dynamic convolution is 1, while the normal 3D convolution with patch size as the kernel size and stride is used to downsample at a specific ratio. This structure enables the block to be more adaptive and reduces the number of parameters as the input image size decreases.

3.3 Parallel local-global feature extraction

Positional encoding introduces relative position relationships between multiple tokens, which is essential for ViTs to capture sequential information in the sequence. Traditionally, absolute positional encoding was first introduced into ViTs (35, 36), and this style is not friendly for different input resolutions. And relative positional encoding also does not always perform well due to its heavy computational burden (38). To improve efficiency, several recent works introducing convolutional positional encoding have been proposed to flexibly embed position relations (39, 40), and we follow them.

As previously discussed, convolution operations are effective at capturing local contextual relationships, while self-attentive mechanisms excel at capturing global dependencies. Therefore, we propose a two-branch structure that concurrently extracts local and global context features, with one branch utilizing convolution operations and the other branch calculating attention distributions. This design enables efficient and effective hierarchical representation learning of local-global context features through multi-stage stacking. Additionally, for feature alignment, we implement both branches as unified multi-head styles, i.e., local MHSA: MHSAl and global MHSA: MHSAg,

Vn=AnHn,    (2)
Hk=W*Concat[V1,V2,...,Vn].    (3)

Given a series of tokens Hn, the token relation aggregators An capture the dependencies between them. Then, these relations are concatenated in the channel dimension, where W represents the learnable parameter matrix.

1) Local MHSA: benefiting from the fact that convolution neural networks are specialized in focusing on local detailed features within a small region, MHSAl is designed as a pure convolutional structure without introducing self-attention operations, as described in Figure 4. In particular, unlike the previous convolutional blocks, MHSAl follows a transformer-like style. We extract features from various “heads” using group convolution, and aggregate multi-head features in the channel dimension. Concretely, given a series of anchor tokens {Hi}1:n obtained by group convolution, the MHSAl learns the affinities between them by a local convolution operation in a small neighborhood Ωd × h × w:

Vn=Wni-j*(Hj,Hi),jΩd×h×w,    (4)

where Wni-jd×h×w is the learnable convolutional kernel parameter matrix, and (ij) denotes the relative position between raw tokens Hi and Hj. By doing so, each head corresponds to each channel of the feature map, enabling us to extract local discriminative features in sMRI within a limited receptive field while preserving the spatial structure of the feature map.

Figure 4
Diagram illustrating a neural network architecture. It starts with convolutional positional encoding, followed by batch normalization (BN). Grouped convolution transforms input, it's concatenated, then batch normalized again. A feedforward network (FFN) follows with linear transformations, a GELU activation, and ends with addition operations connecting back to earlier inputs.

Figure 4. The detailed structure diagram of MHSAl. Multi-head feature extraction is achieved by grouped convolution to extract local features, followed by point convolution for feature aggregation. Similar to traditional ViTs, we add residual connections for the model to trace back to previous features. MHSAg, on the other hand, extracts and aggregates global features through fully connected layers.

2) Global MHSA: self-attention is a natural method for capturing long-range dependencies between features and constructing discriminative global representations. In Vision Transformers (ViTs), a multi-head structure is frequently employed to process sequence information. Each “head” learns independent feature mappings and, ultimately, obtains a global view via linear aggregation:

Hn=eQn(Hi)TKn(Hj)j'ΩD×H×WeQn(Hi)TKn(Hj'),    (5)

where Qn(·) and Kn(·) are chosen as normal fully connected transformations and the above equation is a standard softmax operator. Note that j′∈ΩD × H × W belongs to the global, signaling the concern of the MHSAg for global dependencies. To process the spatial dimensions of sMRI (depth D, height H, and width W), we convert them into a one-dimensional tensor and input it to the fully connected layer. Although traditionally this operator introduces significant computational burden, our MHSAg is located at the post-stage of the network, where it processes feature maps that have been downsampled through multiple stages. This allows for a balance between computational efficiency and accuracy.

Local-global context features in sMRI are aggregated in multiple stages through parallel feature extraction using MHSAl and MHSAg. Similarly to conventional ViTs, we introduce a feed-forward network (FFN) after each MHSA block. Our FFN has a specific architecture consisting of two linear layers wrapped around a nonlinear activation function (GELU). The first linear layer expands the channel dimension by the ratio of 4, while the next linear layer reduces it back to its original level. This operation facilitates the further filtering of features and improves the nonlinear expression of the model.

Considering the excessively large dimensions of the feature maps in the first two stages, it imposes a significant computational burden on the calculation of self-attention. Therefore, in the initial two stages of PlgFormer, the computation of MHSA adopts a locally based convolutional feature extraction module. Subsequently, a parallel local-global context extraction structure is employed to independently extract local features and aggregate global representations.

3.4 Feature fusion module

The F2M introduces a simple gating mechanism to enable adaptive feature selection, as depicted in Figure 5. For local feature maps, a convolution with a larger kernel size of 5 × 5 × 5 is applied to focus on the more global features (the size of feature maps is 5 × 6 × 5 at this stage). Conversely, for global feature maps, point-by-point convolution is employed to preserve the global features of interest to self-attention. Finally, Zl is passed through a sigmoid activation function, resulting in a mapping to values between 0 and 1 that controls the flow of local-global context features:

Zl=Wl*Hlm,    (6)
Zg=Wg*Hgm+Hgm,    (7)
Zaug=Zg*σ(Zl)+Zl*(1-σ(Zl)).    (8)

It's worth noting that, in the computation of multi-head self-attention, we drew inspiration from Zhang et al. (14) and introduced residual connections for the MHSAg module, while this design consideration was not applied to MHSAl. This choice is primarily made because residual connections assist deep networks in rapidly backpropagating shallow features. The convolutional module, having fewer network layers, does not necessitate the introduction of residual connections, as gradients can quickly propagate to the lower layers without them. However, for the globally self-attention operation based on fully connected layers, residual connections play a crucial role in facilitating better information propagation. They help alleviate the issue of gradient decay, ensuring improved preservation and dissemination of vital global information during global modeling. Here, Wl and Wg represent the learnable convolutional kernel parameter matrices, and σ(·) denoting the sigmoid activation function. The resulting feature set Zaug contains both global features aggregated by self-attention and local features extracted through convolution. We obtain Zaug as a one-hot tensor following global average pooling of aggregated spatial features and projection using a fully connected layer.

Figure 5
www.frontiersin.org

Figure 5. The detailed structure of the F2M.

4 Experiments

In this section, we first present the two datasets used in this study in Section 4.1. Then, we present the specific experimental setup and evaluation metrics in Section 4.2. Subsequently, we introduce the comparison experiments with other existing methods in Section 4.3, and the ablation studies to validate the key components of the proposed model in Section 4.4.

4.1 Datasets

In this study, we employed two independent datasets to validate the performance of our proposed method, i.e., the publicly available Alzheimer's Disease Neuroimaging Initiative (ADNI) dataset and the privately available Xuanwu Neuroimaging (XWNI) dataset.

1) ADNI dataset: the ADNI is a large-scale, multicenter research study that aims to identify clinical, imaging, genetic, and biochemical biomarkers for the early detection and tracking of AD (41). This dataset has been extensively used in AD research, including studies on disease progression, diagnosis, and treatment. The availability of longitudinal data from multiple modalities makes it an invaluable resource for developing and evaluating machine learning algorithms for AD detection, prediction, and diagnosis. In this study, we utilized T1-weighted MRI scans from the ADNI dataset, which consists of various types of data, including clinical assessments, neuropsychological tests, MRI, PET, and genetic data. We only select data from ADNI1 dataset, and all sMRI data are acquired using a 1.5T MRI scanner. The ADNI dataset used in our study includes 1,779 samples, comprising 546 cases of NC (283 females, 263 males, 76.5 ± 5.1 years), 840 cases of mild cognitive impairment (MCI; 336 females, 504 males, 75.5 ± 7.1 years), and 393 cases of AD (198 females, 195 males, 75.3 ± 7.6 years). The demographics of participants in ADNI dataset are shown in Table 1. Moreover, we use Table 2 to show our training-testing splits number of ADNI datasets.

Table 1
www.frontiersin.org

Table 1. Demographics of participants in ADNI and XWNI datasets.

Table 2
www.frontiersin.org

Table 2. Training and testing split numbers of ADNI and XWNI datasets.

2) XWNI dataset: the XWNI dataset was obtained from Xuanwu Hospital Capital Medical University, located in Beijing, China.1 This dataset includes sMRI data from patients diagnosed with AD, MCI, and NC. It stands out due to its large sample size and high-quality sMRI images. The sMRI data were acquired using a 3.0T MRI scanner and were preprocessed to improve image quality. The dataset contains both raw and preprocessed sMRI data, including skull-stripped and segmented images. A total of 711 samples were collected, including 515 cases of CN, 47 cases of MCI, and 149 cases of AD. The XWNI database is anticipated to be a valuable resource for researchers involved in AD diagnosis and related studies. The demographics of participants in XWNI dataset are shown in Table 1. Moreover, we use Table 2 to show our training-testing splits number of XWNI datasets.

3) Preprocessing: MNI152_T1 is a commonly used standard template for brain MRI image analysis and research, featuring a standard spatial coordinate system and brain structure information. Considering spatial resolution as a vital parameter affecting image quality and resolution, we initially paired the T1-weighted sMRI with the MNI152_T1_1mm template to obtain a more precise description of brain structure. Moreover, to avoid the computational burden from excessive sMRI data dimensionality, we cropped all images to D × H × W:148 × 192 × 156. These preprocessing methods can enhance the quality and effectiveness of sMRI data and lay the groundwork for subsequent analysis and research. Figure 6 illustrates the difference between sMRI images before and after preprocessing, indicating that the preprocessed images better capture the structural organization of brain tissue and demonstrate improved image quality.

Figure 6
www.frontiersin.org

Figure 6. Comparison of sMRI before and after preprocessing. (a–c) Are raw sMRI from the axial view, the coronal view, and the sagittal view, respectively; (d–f) are preprocessed brain images from the corresponding view that have been cropped after matching to the MNI152_T1_1mm template.

We utilized the medical image processing tool, MONAI, to transform and augment the dataset to adapt to the model input and enhance the robustness of model training, as described in Table 3. In the table, “Augmentation” refers to the functions encapsulated in MONAI, and “Values” refers to the specific parameters corresponding to the functions.

Table 3
www.frontiersin.org

Table 3. The dataset was transformed and augmented using MONAI.

4.2 Experimental setup and evaluation metrics

The proposed method was implemented using Python 3.7.0 and PyTorch 1.10.0. The GPU we used is one P40 with a memory of 24GB. We trained the network end-to-end using AdamW optimizer and optimize under the supervision of the cross-entropy loss function, without additional data for pre-training. PlgFormer consists of four stages, with 2, 2, 3, and 4 Multi-Head Self-Attention (MHSA) mechanisms allocated to each stage respectively. The DEB module uses 4 dynamic convolution kernels to enhance feature representation, an attention hidden ratio of 0.25 to balance performance and computational efficiency, and a temperature parameter initialized at 34 and decreased by 3 during training to progressively sharpen the attention distribution. The initial learning rate was set to 5e-4 and decayed with a cosine annealing strategy. To overcome early optimization difficulties, we performed a linear warm-up for the first 100 epochs, while the total training epochs were 400. Considering the extremely unbalanced number of samples contained in each category of the dataset, especially in the XWNI dataset, we used weighted random sampling to balance the number of samples contained in each category when loading the dataset to pursue better diagnostic performances. We conducted dichotomous (AD vs. CN and MCI vs. CN) and trichotomous (AD vs. MCI vs. CN) classifications to comprehensively evaluate the diagnostic performance of our model with varying degrees of cognitive impairment.

For model evaluation, we select a comprehensive set of metrics including accuracy (ACC), recall (REC), precision (PRE), F1-score (F1), and specificity (SPE). The calculation of each metric is as follows:

ACC=TP+TNTP+TN+FP+FN    (9)
REC=TPTP+FN    (10)
PRE=TPTP+FP    (11)
F1=2×REC×PREREC+PRE    (12)
SPE=TNTN+FP    (13)

where TP, FP, TN, and FN denote true positive, false positive, true negative, and false negative, respectively. In practice, REC denotes the probability that a positive sample in the data set is correctly discriminated, and PRE denotes the proportion of all examples diagnosed as positive by the model that are in fact positive. Obviously, REC and PRE are numerically a pair of mutually exclusive indicators, so we calculate the F1-score to evaluate the performance of the model in a comprehensive way. And SPE indicates the ability of the model to determine negative samples. In addition, we also report the Area Under the Curve (AUC), which is the area under the Receiver Operating Characteristic Curve (ROC) curve, to measure the performance of the dichotomous model. is equal to 0.5, the model is equivalent to a random guess.

4.3 Comparison experiments

We compared our proposed method with several existing classification approaches fine-tuned on large-scale multimodal medical imaging datasets, including Med3D-ResNet-10, Med3D-ResNet-18, and Med3D-ResNet-34 (42). Additionally, we extended the original ViT model (35) to support 3D medical image classification, referred to as ViT-ori in this study. We also included methods specifically designed for Alzheimer's disease classification, covering both CNN-based and Transformer-based architectures, such as DA-MIDL (24), AMSNet (25), ResAttNet-10 (26), ResAttNet-18 (26), ViT-for-AD (43), and MCNEL (44). To comprehensively evaluate the performance of our approach, we conducted three sets of experiments on both datasets: (1) binary classification between AD and CN, (2) binary classification between MCI and CN, and (3) multi-class classification among AD, MCI, and CN subjects.

1) Binary classification on AD and CN subjects: accurately identifying AD plays a crucial role in clinical diagnosis as well as computer-aided diagnosis. AD is characterized by brain atrophy, gray matter atrophy, reduced brain tissue, white matter damage, enlarged sulcal gyrus, and ventricles, which can be easily detected by physicians or computers on sMRI. In this study, we employed PlgFormer to set the embedded dimension of each stage to 16, 32, 64, and 128, and each head dimension to 16. The weight decay, learning rate, and batch size were set to 5e-4, 2e-4, and 16, respectively. We conducted comparison experiments on ADNI and XWNI datasets, and the results are presented in Table 4.

Table 4
www.frontiersin.org

Table 4. Quantitative comparison of our proposed PlgFormer and other existing methods for AD and CN binary classification on ADNI and XWNI datasets.

As elaborated in Table 4, our proposed PlgFormer achieves superior performance on ADNI and XWNI datasets in most cases. PlgFormer outperforms other methods in most of the evaluation metrics on both datasets, with ACC = 0.9431, PRE = 0.9016, F1 = 0.8871, SPE = 0.9672, and AUC = 0.9201 on ADNI dataset and ACC = 0.9407, REC = 0.9184, PRE = 0.8036, F1 = 0.8517, and AUC = 0.9021 on XWNI dataset. Although PlgFormer did not achieve the best performance in some metrics, such as REC on ADNI dataset and SPE on XWNI dataset, it notably outperformed other existing methods on several metrics. Notably, PlgFormer achieved the highest accuracy, suggesting its excellent performance in diagnosing AD patients. These results demonstrate the effectiveness and potential of using PlgFormer for binary classification in AD and CN, which can aid in early diagnosis and intervention for AD.

2) Binary classification on MCI and CN subjects: identifying individuals with MCI is a crucial task for early intervention in AD, especially since there are currently no well-established treatment options for AD. However, using sMRI alone to perform this task is challenging, as the brain regions of patients with MCI usually do not undergo significant morphological changes. In this study, we set the embedding dimension of each stage of PlgFormer to 16, 32, 64, 96, while keeping the other hyperparameters the same as those for the AD and CN binary classification. The results of a detailed comparison with other existing methods are presented in Table 5.

Table 5
www.frontiersin.org

Table 5. Quantitative comparison of our proposed PlgFormer and other existing methods for MCI and CN binary classification on ADNI and XWNI datasets.

The results presented in Table 5 demonstrate that our proposed PlgFormer outperforms other existing methods on both ADNI and XWNI datasets, with the highest metrics achieved in most cases. For example, on the ADNI dataset, our method achieved a PRE of 0.9290 and an AUC of 0.7654, while on the XWNI dataset, our method achieved a PRE of 0.7778 and an AUC of 0.8512. It is worth noting that MCI patients exhibit less significant structural alterations in brain regions on sMRI compared to AD patients. Consequently, the binary performance of the MCI and CN subjects was not as good as that of the AD and CN subjects on both data sets. As illustrated in Table 5, our PlgFormer achieves notable performance in F1 scores on both datasets, thus demonstrating its ability to distinguish between MCI and CN subjects.

3) Triple classification on AD, MCI, and CN subjects: performing a triple classification experiment using sMRI on AD, MCI, and CN subjects is of great significance for computer-aided AD diagnosis. Currently, there is no effective cure for AD, underscoring the importance of early definitive detection to enable early intervention and treatment. MCI, as a transitional stage between normal aging and dementia, significantly increases the risk of developing AD. Therefore, accurately diagnosing and differentiating individuals with MCI and AD from those with normal cognitive function (CN) is crucial to facilitate the early detection and management of AD.

Table 6 presents quantitative comparisons, revealing that our proposed PlgFormer outperforms other existing methods, particularly in the recognition of AD subjects, notably in the ADNI dataset. We attribute this to the flexibility of PlgFormer in addressing both local and global requirements, making it more sensitive to significant structural changes in brain regions. In general, our method demonstrated remarkable performance on most metrics, providing compelling evidence of its potential use in clinical practice for computer-aided diagnosis. Furthermore, our review of existing studies utilizing sMRI for AD diagnosis reveals that limited attention has been given to the three-way classification of AD, MCI, and CN. We acknowledge the inherent difficulty of this task, as current models often struggle to capture the subtle and discriminative features present in sMRI. Nevertheless, given its substantial clinical relevance, we strongly encourage future research to further explore and address this challenging problem.

Table 6
www.frontiersin.org

Table 6. Quantitative comparison of our proposed PlgFormer and other existing methods for AD MCI and CN triple classification on ADNI and XWNI datasets.

4.4 Ablation studies

In this subsection, we conducted ablation studies on AD and CN binary classification on ADNI dataset to evaluate the impact of various key components in PlgFormer on model representation capacity. Similarly, we selected ACC, REC, PRE, F1, SPE, and AUC as evaluation metrics. The corresponding results are presented in Table 7, L and G respectively represent MHSAl and MHSAg, ✓denotes leaving the corresponding component in place, whereas × denotes replacing it with another multi-head self-attention module.

Table 7
www.frontiersin.org

Table 7. Ablation studies on individual components of the proposed PlgFormer.

According to the results presented in Table 7, it is evident that convolutional operations play a crucial role in the PlgFormer model we designed. When MHSAl was replaced with MHSAg, the performance of the model dropped significantly (ACC decreased from 0.9431 to 0.7480). We attribute this phenomenon to the small sample size of our dataset, as pure self-attention modules require a large amount of data to demonstrate their effectiveness. We also observed a slight drop in performance when MHSAg was replaced with a local MHSAl, suggesting that global features are still necessary for AD classification using sMRI. This conclusion was supported by the fact that attention modules were embedded in the DA-MIDL and ResAttNet-10 architectures. Furthermore, when we removed the designed DEB, the performance slightly decreased, indicating that encoding image patch sequences dynamically is a meaningful operation. In addition, F2M is also an important and effective module for feature fusion, as replacing it with a simple concatenation caused a decrease in all evaluation metrics.

We evaluated the effects of dynamic convolution on overfitting and generalization, as shown in Table 8. The results indicate that when dynamic convolution is used, the model demonstrates similar training errors but reduces validation errors, highlighting its capability to mitigate overfitting. Furthermore, dynamic convolution leads to lower errors in the testing set of the XWNI dataset, suggesting improved generalization on other datasets. It is worth noting that we conducted experiments across 5 independent runs with different random seeds and reported the mean and standard deviation of the final validation loss, in order to evaluate the training stability of DEB. The results show that the DEB-enhanced model achieves a lower average loss with reduced variance, demonstrating improved robustness and stability.

Table 8
www.frontiersin.org

Table 8. Ablation studies of DEB.

4.5 Visualization

To provide human physicians with reliable and accurate computer-aided diagnostic results, we employed Grad-CAM (45) to generate sMRI slice heat maps in sagittal, axial, and coronal planes, as illustrated in Figure 7. To produce high-resolution heat maps that are easy to interpret, we applied 3D Grad-CAM at a lower layer with a resolution of 36 × 44 × 36 (D × H × W). In the process of visualization, the reshaped tensor of dimensions 36 × 44 × 36 was restored to its original resolution to facilitate a more detailed examination of features attended to by the neural network. The Grad-CAM procedure commenced with the loading of a pre-trained model alongside the original sMRI input. Subsequently, an intermediate layer's output and its gradients with respect to the output were selectively chosen. The visual representation of the feature maps derived from this output was then obtained. Following this, the impact of the gradients on the output of the target layer was quantified, yielding weight factors indicative of the significance of the feature maps. These weights were employed to project the importance of the feature maps back onto the input image. Finally, the application of these weights to the target layer's output resulted in the generation of a heatmap. This heatmap provides insights into which regions of the input image hold critical information for the model's predictive capacity.

Figure 7
www.frontiersin.org

Figure 7. Salient maps of sagittal, axial and coronal sMRI slices generated by Grad-CAM. The visualization results allow to observe the features of network interest.

We have drawn GradCAM using six patients with AD, as illustrated in Figure 7, all of whom are from the test set of the ADNI dataset. Additionally, We have analyzed the visualization results for these patients using the ICBM 152 template. First, we aligned the sMRI data of AD patients with the template, and then analyzed the brain regions corresponding to the areas of interest (high-luminance areas) of our model. After analysis, we found that the brain regions that our model focuses on include: Posterior Cingulate, Precuneus, Isthmus Cingulate, and Lateral Ventricle. We have consulted with doctors who diagnose AD clinically, and found that the Posterior Cingulate and Precuneus regions are consistent with the areas that doctors focus on during clinical diagnosis. This further confirms that the features extracted by our method are not only meaningful for deep neural network models but also provide credible diagnoses of AD via sMRI for human physicians (46).

Moreover, the visualization results obtained using Grad-CAM empower physicians to better understand and interpret the classification decision of our model. By providing a visual reference, physicians can easily validate the reasoning behind a diagnosis and identify any potential shortcomings or biases in the model. This serves as a valuable tool for improving the interpretability and transparency of Computer-aided diagnosis (CAD), and ultimately helps build trust between physicians and machine learning models.

5 Conclusions

Using sMRI for computer-aided diagnosis is significant for early detection and timely intervention of AD. In this paper, we propose PlgFormer, a unified and parallel approach that combines CNNs and pure self-attention mechanisms to extract local-global context features in sMRI with discriminative value for AD diagnosis. Our designed DEB introduces dynamic convolutions that adaptively adjust the kernel size based on the input size, while our designed F2M adaptively fuses the extracted local and global features through a gating mechanism. On publicly available ADNI and privately held XWNI datasets, our PlgFormer achieved state-of-the-art performance compared to existing methods in AD vs. CN binary classification, MCI vs. CN binary classification, and AD vs. MCI vs. CN triple classification tasks. Saliency maps generated by Grad-CAM confirmed that our proposed method can help human experts identify lesions quickly in sMRI. Further studies could investigate the potential of PlgFormer on other datasets and explore its application in other areas of medical image analysis. Our research anticipates practical clinical applications.

Data availability statement

The datasets presented in this article are not readily available because the XWNI dataset used in this study is not publicly available due to patient privacy concerns and institutional data sharing restrictions. Access to the dataset is limited by the policies of Xuanwu Hospital Capital Medical University and requires specific institutional approvals. Requests to access the datasets should be directed to Zhixiong Li, ODY1ODE4NjgzQHFxLmNvbQ==.

Ethics statement

The studies involving humans were approved by Ethics Committee of Xuanwu Hospital Capital Medical University. The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their written informed consent to participate in this study.

Author contributions

GW: Conceptualization, Formal analysis, Investigation, Methodology, Validation, Writing – original draft, Writing – review & editing. YL: Data curation, Investigation, Writing – review & editing. ZZ: Data curation, Formal analysis, Software, Writing – original draft. SA: Project administration, Validation, Writing – review & editing. XC: Formal analysis, Methodology, Writing – review & editing. YJ: Writing – review & editing. ZS: Resources, Writing – review & editing. GC: Resources, Writing – review & editing. MZ: Resources, Writing – review & editing. ZL: Resources, Writing – review & editing. FY: Project administration, Writing – review & editing.

Funding

The author(s) declare that financial support was received for the research and/or publication of this article. The authors gratefully acknowledge the support of the National Key Research and Development Program of China (Grant No. 2023YFC3603601), the National Natural Science Foundation of China (Grant No. 82001773), the Medical Science Research Project of Hebei Province (20221842), and the Construction Project of Academician Cooperation Key Unit of Hebei Province.

Conflict of interest

XC and YJ were employed by JD Health International Inc.

The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declare that no Gen AI was used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Footnotes

1. ^This study was performed in line with the principles of the Declaration of Helsinki. Ethical Approval was granted by the Ethics Committee of Xuanwu Hospital Capital Medical University (No. 2017046). All subjects signed informed consent forms.

References

1. Association A. 2023 Alzheimer's disease facts and figures. Alzheimers Dement. (2023) 19:1598–695. doi: 10.1002/alz.13016

PubMed Abstract | Crossref Full Text | Google Scholar

2. Breijyeh Z, Karaman R. Comprehensive review on Alzheimer's disease: causes and treatment. Molecules. (2020) 25:5789. doi: 10.3390/molecules25245789

PubMed Abstract | Crossref Full Text | Google Scholar

3. Weller J, Budson A. Current understanding of Alzheimer's disease diagnosis and treatment. F1000Research. (2018) 7:F1000-Faculty. doi: 10.12688/f1000research.14506.1

PubMed Abstract | Crossref Full Text | Google Scholar

4. Frisoni GB, Fox NC, Jack Jr CR, Scheltens P, Thompson PM. The clinical use of structural MRI in Alzheimer disease. Nat Rev Neurol. (2010) 6:67–77. doi: 10.1038/nrneurol.2009.215

PubMed Abstract | Crossref Full Text | Google Scholar

5. Jack Jr PRCXYCea C R. Prediction of AD with MRI-based hippocampal volume in mild cognitive impairment. Neurology. (1999) 52:1397. doi: 10.1212/WNL.52.7.1397

PubMed Abstract | Crossref Full Text | Google Scholar

6. Convit A, De Leon M, Tarshish C, De Santi S, Tsui W, Rusinek H, et al. Specific hippocampal volume reductions in individuals at risk for Alzheimer's disease. Neurobiol Aging. (1997) 18:131–8. doi: 10.1016/S0197-4580(97)00001-8

PubMed Abstract | Crossref Full Text | Google Scholar

7. Apostolova LG, Dinov ID, Dutton RA, Hayashi KM, Toga AW, Cummings JL, et al. 3D comparison of hippocampal atrophy in amnestic mild cognitive impairment and Alzheimer's disease. Brain. (2006) 129:2867–73. doi: 10.1093/brain/awl274

PubMed Abstract | Crossref Full Text | Google Scholar

8. Hinrichs C, Singh V, Mukherjee L, Xu G, Chung MK, Johnson SC, et al. Spatially augmented LPboosting for AD classification with evaluations on the ADNI dataset. Neuroimage. (2009) 48:138–49. doi: 10.1016/j.neuroimage.2009.05.056

PubMed Abstract | Crossref Full Text | Google Scholar

9. Kao YH, Chou MC, Chen CH, Yang YH. White matter changes in patients with Alzheimer's disease and associated factors. J Clin Med. (2019) 8:167. doi: 10.3390/jcm8020167

PubMed Abstract | Crossref Full Text | Google Scholar

10. Vounou M, Janousova E, Wolz R, Stein JL, Thompson PM, Rueckert D, et al. Sparse reduced-rank regression detects genetic associations with voxel-wise longitudinal phenotypes in Alzheimer's disease. Neuroimage. (2012) 60:700–16. doi: 10.1016/j.neuroimage.2011.12.029

PubMed Abstract | Crossref Full Text | Google Scholar

11. Zhang D, Wang Y, Zhou L, Yuan H, Shen D, Initiative ADN, et al. Multimodal classification of Alzheimer's disease and mild cognitive impairment. Neuroimage. (2011) 55:856–67. doi: 10.1016/j.neuroimage.2011.01.008

PubMed Abstract | Crossref Full Text | Google Scholar

12. Liu M, Zhang D, Shen D. Relationship induced multi-template learning for diagnosis of Alzheimer's disease and mild cognitive impairment. IEEE Trans Med Imaging. (2016) 35:1463–74. doi: 10.1109/TMI.2016.2515021

PubMed Abstract | Crossref Full Text | Google Scholar

13. Qiu S, Joshi PS, Miller MI, Xue C, Zhou X, Karjadi C, et al. Development and validation of an interpretable deep learning framework for Alzheimer's disease classification. Brain. (2020) 143:1920–33. doi: 10.1093/brain/awaa137

PubMed Abstract | Crossref Full Text | Google Scholar

14. Zhang X, Han L, Han L, Chen H, Dancey D, Zhang D. MRI-PatchNet: a novel efficient explainable patch-based deep learning network for Alzheimer's disease diagnosis with Structural MRI. IEEE Access. (2023) 11:108603–16. doi: 10.1109/ACCESS.2023.3321220

Crossref Full Text | Google Scholar

15. Ashburner J, Friston KJ. Voxel-based morphometry—the methods. Neuroimage. (2000) 11:805–21. doi: 10.1006/nimg.2000.0582

PubMed Abstract | Crossref Full Text | Google Scholar

16. Klöppel S, Stonnington CM, Chu C, Draganski B, Scahill RI, Rohrer JD, et al. Automatic classification of MR scans in Alzheimer's disease. Brain. (2008) 131:681–9. doi: 10.1093/brain/awm319

PubMed Abstract | Crossref Full Text | Google Scholar

17. Fan Y, Shen D, Gur RC, Gur RE, Davatzikos C. COMPARE: classification of morphological patterns using adaptive regional elements. IEEE Trans Med Imaging. (2006) 26:93–105. doi: 10.1109/TMI.2006.886812

PubMed Abstract | Crossref Full Text | Google Scholar

18. Cao P, Liu X, Yang J, Zhao D, Huang M, Zhang J, et al. Nonlinearity-aware based dimensionality reduction and over-sampling for AD/MCI classification from MRI measures. Comput Biol Med. (2017) 91:21–37. doi: 10.1016/j.compbiomed.2017.10.002

PubMed Abstract | Crossref Full Text | Google Scholar

19. Abuhmed T, El-Sappagh S, Alonso JM. Robust hybrid deep learning models for Alzheimer's progression detection. Knowl Based Syst. (2021) 213:106688. doi: 10.1016/j.knosys.2020.106688

Crossref Full Text | Google Scholar

20. Krizhevsky A, Sutskever I, Hinton GE. ImageNet classification with deep convolutional neural networks. Commun ACM. (2017) 60:84–90. doi: 10.1145/3065386

Crossref Full Text | Google Scholar

21. Simonyan K, Zisserman A. Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:14091556. (2014). doi: 10.48550/arXiv.1409.1556

Crossref Full Text | Google Scholar

22. He K, Zhang X, Ren S, Sun J. Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Las Vegas, NV: IEEE (2016). p. 770–8. doi: 10.1109/CVPR.2016.90

Crossref Full Text | Google Scholar

23. Lian C, Liu M, Zhang J, Shen D. Hierarchical fully convolutional network for joint atrophy localization and Alzheimer's disease diagnosis using structural MRI. IEEE Trans Pattern Anal Mach Intell. (2018) 42:880–93. doi: 10.1109/TPAMI.2018.2889096

PubMed Abstract | Crossref Full Text | Google Scholar

24. Zhu W, Sun L, Huang J, Han L, Zhang D. Dual attention multi-instance deep learning for Alzheimer's disease diagnosis with structural MRI. IEEE Trans Med Imaging. (2021) 40:2354–66. doi: 10.1109/TMI.2021.3077079

PubMed Abstract | Crossref Full Text | Google Scholar

25. Wu Y, Zhou Y, Zeng W, Qian Q, Song M. An attention-based 3D CNN with multi-scale integration block for Alzheimer's disease classification. IEEE J Biomed Health Inform. (2022) 26:5665–73. doi: 10.1109/JBHI.2022.3197331

PubMed Abstract | Crossref Full Text | Google Scholar

26. Zhang X, Han L, Zhu W, Sun L, Zhang D. An explainable 3D residual self-attention deep neural network for joint atrophy localization and Alzheimer's disease diagnosis using structural MRI. IEEE J Biomed Health Inform. (2021) 26:5289–97. doi: 10.1109/JBHI.2021.3066832

PubMed Abstract | Crossref Full Text | Google Scholar

27. Ashtari-Majlan M, Seifi A, Dehshibi MM. A multi-stream convolutional neural network for classification of progressive MCI in Alzheimer's disease using structural MRI images. IEEE J Biomed Health Inform. (2022) 26:3918–26. doi: 10.1109/JBHI.2022.3155705

PubMed Abstract | Crossref Full Text | Google Scholar

28. Zhang J, He X, Qing L, Chen X, Liu Y, Chen H. Multi-relation graph convolutional network for Alzheimer's disease diagnosis using structural MRI. Knowl Based Syst. (2023) 270:110546. doi: 10.1016/j.knosys.2023.110546

Crossref Full Text | Google Scholar

29. Billones CD, Demetria OJLD, Hostallero DED, Naval PC. DemNet: a convolutional neural network for the detection of Alzheimer's disease and mild cognitive impairment. In: 2016 IEEE Region 10 Conference (TENCON). Singapore: IEEE (2016). p. 3724–7. doi: 10.1109/TENCON.2016.7848755

Crossref Full Text | Google Scholar

30. Li H, Habes M, Fan Y. Deep ordinal ranking for multi-category diagnosis of Alzheimer's disease using hippocampal MRI data. arXiv preprint arXiv:170901599. (2017). doi: 10.48550/arXiv.1709.01599

Crossref Full Text | Google Scholar

31. Liu Z, Lu H, Pan X, Xu M, Lan R, Luo X. Diagnosis of Alzheimer's disease via an attention-based multi-scale convolutional neural network. Knowl-Based Syst. (2022) 238:107942. doi: 10.1016/j.knosys.2021.107942

Crossref Full Text | Google Scholar

32. Zhang J, Zheng B, Gao A, Feng X, Liang D, Long X, et al. 3D densely connected convolution neural network with connection-wise attention mechanism for Alzheimer's disease classification. Magn Reson Imaging. (2021) 78:119–26. doi: 10.1016/j.mri.2021.02.001

PubMed Abstract | Crossref Full Text | Google Scholar

33. Jin D, Xu J, Zhao K, Hu F, Yang Z, Liu B, et al. Attention-based 3D convolutional network for Alzheimer's disease diagnosis and biomarkers exploration. In: 2019 IEEE 16Th International Symposium on biomedical imaging (ISBI 2019). Venice: IEEE (2019). p. 1047–51. doi: 10.1109/ISBI.2019.8759455

Crossref Full Text | Google Scholar

34. Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, et al. Attention is all you need. Adv Neural Inf Process Syst. (2017) 30:6000–10. doi: 10.5555/3295222.3295349

Crossref Full Text | Google Scholar

35. Dosovitskiy A, Beyer L, Kolesnikov A, Weissenborn D, Zhai X, Unterthiner T, et al. An image is worth 16 × 16 words: transformers for image recognition at scale. arXiv preprint arXiv:201011929. (2020). doi: 10.48550/arXiv.2010.11929

Crossref Full Text | Google Scholar

36. Liu Z, Lin Y, Cao Y, Hu H, Wei Y, Zhang Z, et al. Swin transformer: hierarchical vision transformer using shifted windows. In: Proceedings of the IEEE/CVF International Conference on Computer Vision. Montreal, QC: IEEE (2021). p. 10012–22. doi: 10.1109/ICCV48922.2021.00986

Crossref Full Text | Google Scholar

37. Jang J, Hwang D. M3T: three-dimensional medical image classifier using multi-plane and multi-slice transformer. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. New Orleans, LA: IEEE (2022). p. 20718–29. doi: 10.1109/CVPR52688.2022.02006

Crossref Full Text | Google Scholar

38. Chu X, Zhang B, Tian Z, Wei X, Xia H. Do we really need explicit position encodings for vision transformers. arXiv preprint arXiv:210210882. (2021). doi: 10.48550/arXiv.2102.10882

Crossref Full Text | Google Scholar

39. Chu X, Tian Z, Wang Y, Zhang B, Ren H, Wei X, et al. Twins: revisiting the design of spatial attention in vision transformers. Adv Neural Inf Process Syst. (2021) 34:9355–66. doi: 10.5555/3540261.3540977

Crossref Full Text | Google Scholar

40. Dong X, Bao J, Chen D, Zhang W, Yu N, Yuan L, et al. CSWin transformer: a general vision transformer backbone with cross-shaped windows. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. New Orleans, LA: IEEE (2022). p. 12124–34. doi: 10.1109/CVPR52688.2022.01181

Crossref Full Text | Google Scholar

41. Jack Jr CR, Bernstein MA, Fox NC, Thompson P, Alexander G, Harvey D, et al. The Alzheimer's disease neuroimaging initiative (ADNI): MRI methods. J Magn Reson Imaging. (2008) 27:685–91. doi: 10.1002/jmri.21049

PubMed Abstract | Crossref Full Text | Google Scholar

42. Chen S, Ma K, Zheng Y. Med3D: transfer learning for 3D medical image analysis. arXiv preprint arXiv:190400625. (2019). doi: 10.48550/arXiv.1904.00625

Crossref Full Text | Google Scholar

43. Kunanbayev K, Shen V, Kim DS. Training ViT with limited data for Alzheimer's disease classification: an empirical study. In: Proceedings of Medical Image Computing and Computer Assisted Intervention-MICCAI 2024. Vol. LNCS 15012. Springer Nature Switzerland: New York (2024). doi: 10.1007/978-3-031-72390-2_32

Crossref Full Text | Google Scholar

44. Yan F, Peng L, Dong F, Hirota K. MCNEL: a multi-scale convolutional network and ensemble learning for Alzheimer's disease diagnosis. Comput Methods Programs Biomed. (2025) 264:108703. doi: 10.1016/j.cmpb.2025.108703

PubMed Abstract | Crossref Full Text | Google Scholar

45. Selvaraju RR, Cogswell M, Das A, Vedantam R, Parikh D, Batra D. Grad-cam: visual explanations from deep networks via gradient-based localization. In: Proceedings of the IEEE International Conference on Computer Vision. Venice: IEEE (2017). p. 618–26. doi: 10.1109/ICCV.2017.74

Crossref Full Text | Google Scholar

46. Ott BR, Cohen RA, Gongvatana A, Okonkwo OC, Johanson CE, Stopa EG, et al. Brain ventricular volume and cerebrospinal fluid biomarkers of Alzheimer's disease. J Alzheimers Dis. (2010) 20:647–57. doi: 10.3233/JAD-2010-1406

PubMed Abstract | Crossref Full Text | Google Scholar

Keywords: Alzheimer's disease diagnosis, attention mechanism, computer-aided diagnosis, sMRI, multi-level feature fusion

Citation: Wang G, Li Y, Zhou Z, An S, Cao X, Jin Y, Sun Z, Chen G, Zhang M, Li Z and Yu F (2025) PlgFormer: parallel extraction of local-global features for AD diagnosis on sMRI using a unified CNN-transformer architecture. Front. Neurol. 16:1626922. doi: 10.3389/fneur.2025.1626922

Received: 12 May 2025; Accepted: 07 August 2025;
Published: 29 August 2025.

Edited by:

Shang-Ming Zhou, University of Plymouth, United Kingdom

Reviewed by:

Rakeshkumar Mahto, California State University, Fullerton, United States
Wang Liao, The Second Affiliated Hospital of Guangzhou Medical University, China

Copyright © 2025 Wang, Li, Zhou, An, Cao, Jin, Sun, Chen, Zhang, Li and Yu. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Shan An, YW5zaGFuQHRqdS5lZHUuY24=; Zhixiong Li, ODY1ODE4NjgzQHFxLmNvbQ==; Feng Yu, b3NmZW5neXVAemp1LmVkdS5jbg==

These authors have contributed equally to this work

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.