Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Neurol., 05 November 2025

Sec. Movement Disorders

Volume 16 - 2025 | https://doi.org/10.3389/fneur.2025.1706317

This article is part of the Research TopicAI-powered Advances in Diagnosis and Management of Movement DisordersView all articles

Parkinson's disease detection using spectrogram-based multi-model feature fusion networks


Wenna Chen
&#x;Wenna Chen1*Rongfu Lv&#x;Rongfu Lv2Xiaowei DuXiaowei Du1Xiangyu ChenXiangyu Chen1Hao WangHao Wang1Jincan ZhangJincan Zhang2Ganqin Du
Ganqin Du1*
  • 1The First Affiliated Hospital and College of Clinical Medicine of Henan University of Science and Technology, Luoyang, China
  • 2College of Information Engineering, Henan University of Science and Technology, Luoyang, China

Introduction: Parkinson's disease (PD) is a common neurodegenerative disorder. Traditional diagnostic methods, relying on clinical assessment and imaging, are often invasive, costly, and require specialized personnel, posing barriers to early detection. As approximately 90% of PD patients develop vocal impairments, vocal analysis emerges as a promising non-invasive diagnostic tool. However, individual deep learning models are often limited by overfitting and poor generalizability.

Methods: This study proposes a PD classification method using spectrogram feature fusion with pre-trained convolutional neural networks (CNNs). Voice recordings were obtained from 61 PD patients and 70 healthy controls (HC) at the First Affiliated Hospital of Henan University of Science and Technology. Preprocessing the raw speech signals yielded 2,476 spectrograms. Three pre-trained models, DenseNet121, MobileNetV3-Large, and ShuffleNetV2, were used for feature extraction. The output of MobileNetV3-Large was adjusted using a 1 × 1 convolutional layer to ensure dimensional alignment before features were fused via summation.

Results: Evaluation using 5-fold cross-validation demonstrated that models employing feature fusion consistently outperformed individual models across all metrics. Specifically, the fusion of MobileNetV3-Large and ShuffleNetV2 achieved the highest accuracy of 95.56% and an AUC of 0.99. Comparative experiments with existing state-of-the-art methods confirmed the competitive performance of the proposed approach.

Discussion: The fusion of multi-model features more effectively captures subtle pathological signatures in PD speech, overcoming the limitations of single models. This method provides a reliable, low-cost, and non-invasive tool for auxiliary PD diagnosis, with significant potential for clinical application. The code is available at https://github.com/lvrongfu/pjs.

1 Introduction

Parkinson's disease (PD) is a disorder of the central nervous system. It causes involuntary and uncontrollable body movements such as tremors, stiffness, or difficulties with balance and control. People with PD may also experience behavioral or mental changes, such as depression or memory decline. Currently, there is no cure for PD, but medications can alleviate symptoms. Regardless, it is preferable to intervene and prevent the progressive onset of PD rather than treat it in its most severe state. However, traditional diagnostic methods often rely on clinical assessments and imaging techniques, which can be invasive, costly, and require specialized medical expertise. In recent years, the emergence of artificial intelligence has opened new opportunities for diagnosis, particularly through voice analysis (1). Notably, since approximately 90% of PD patients exhibit speech difficulties or dysphonia, this can distinguish them from healthy individuals. Consequently, detection based on voice disorders has become a valuable, non-invasive tool for early screening of PD (2, 3).

PD affects approximately 10 million people worldwide, making it the second most common neurodegenerative disorder after Alzheimer's disease. The risk of developing PD increases with age, particularly after the age of 65 (4). Early symptoms such as loss of smell, constipation, and sleep disturbances often precede motor symptoms. Timely intervention in the early stages is crucial for slowing or halting disease progression. However, diagnosing PD based solely on clinical symptoms remains challenging and complex. Given that 90% of PD patients experience speech disorders, utilizing speech data for detection has gained prominence in recent years (5, 6). Audio signals play a pivotal role in early PD diagnosis, as speech abnormalities detectable through voice signal analysis may remain imperceptible to the human ear during initial stages. Furthermore, since voice samples can be easily recorded both clinically and non-clinically, voice changes can also serve as a means to track disease progression (7). This paper investigates the application of Deep Learning (DL) techniques in the early diagnosis of PD through the analysis of vocal characteristics to distinguish between PD patients and healthy controls (HC).

Machine learning (ML) and DL have revolutionized medical data analysis and image processing, driving significant advancements in disease diagnosis, identification, and treatment planning. Furthermore, ML enables algorithms to learn from data. As a subfield of ML, DL uses multi-layered neural networks to excel at simulating complex patterns (8). Within medical image analysis, DL approaches demonstrate exceptional efficiency in tasks related to image classification, segmentation, and anomaly detection. A key advantage of DL models lies in their ability to extract relevant features and biomarkers from diverse data sources. In voice recordings, variations in speech patterns—such as changes in pitch, rhythm, and clarity—serve as valuable biomarkers for detecting PD (9). DL, particularly convolutional neural networks (CNNs), has been extensively applied across various computer-aided diagnostic methods for medical imaging. This is because CNNs can achieve or even surpass human performance in general object recognition. Unlike traditional ML algorithms, CNNs can automatically extract meaningful features from images. Traditional algorithms, such as support vector machines and backpropagation neural networks, typically require manual feature engineering. CNNs eliminate this step, making the process more efficient and less reliant on expert knowledge. Consequently, the development of CNNs models has significantly improved the performance of computer-aided diagnostic systems. Several classification network models have been extensively applied to medical image classification tasks, including RegNet (10), ResNet (11), ConvNext (12), Vision Transformer (13), and Swin Transformer (14).

However, relying solely on a single model often leads to overfitting to the training set and poor generalization to the test set, thereby reducing the model's performance. To address this limitation, this paper proposes using model ensemble techniques. This paper utilizes three pre-trained models, namely DenseNet-121, MobileNetV3, and ShuffleNetV2, to extract features from the spectrograms of patients with PD. Subsequently, the extracted features are fused using a summation method, followed by classification of the fused features. The main contributions of this paper are as follows:

• This study proposes a spectrogram-based deep learning framework for PD classification, leveraging time-frequency representations of speech signals to capture discriminative acoustic patterns.

• The proposed novel network model enhances feature extraction capabilities by fusing features from three baseline networks.

• In extensive experimental evaluations, the proposed method has been demonstrated to outperform current techniques.

2 Related work

In recent years, the application of ML in PD research has advanced significantly, thanks to collaborative efforts between biomedicine and computer science. However, achieving timely and accurate diagnosis remains a major challenge. Extensive research has focused on enhancing classification performance via ML methods, using speech samples to enable automated detection of PD-related speech disorders. Typically, sustained vowel production evaluates phonetic characteristics, while continuous speech assesses articulation and intonation features (15). Numerous algorithms have been developed, and researchers have evaluated the performance metrics of different approaches (16).

Gómez-García et al. investigated the relationship between pathological voice perception and certain voice characteristics (17). Typically, in the analysis of vocalization, acoustic features are manually extracted from recordings of sustained vowels, syllable repetitions, and the reading of words, sentences, or free monologs. Selecting the most appropriate features is crucial for obtaining precise results. For instance, Mekyska et al. introduced 36 novel speech features in their study, achieving 67.9% accuracy using SVM and random forest classifiers on PD data solely through/a/vowel subset analysis (18). The Almeida team evaluated 18 feature extraction methods and four ML classifiers for PD detection and classification by analyzing sustained vowel production and other speech tasks (16). They achieved the highest accuracy of 94.55% by analyzing sustained vowel production. Carron et al. proposed a mobile-assisted speech status analysis system for PD detection, based on a server-client architecture (19). On the server side, feature extraction and ML algorithms were designed and implemented to distinguish between PD and HC. This Android application allows patients to submit voice samples, enabling physicians to review each patient's complete record. Saleh et al. (20) proposed a PD pre-diagnosis system based on Artificial Intelligence of Things (AIoT) and speech analysis. Through the sequential forward feature selection method, 10 key features were selected from 22 speech features, and the K-nearest neighbor algorithm was used for classification. Cherradi et al. (21) proposed an embedded medical system based on AIoT, which predicts PD by analyzing sound signals. This system integrates multiple machine learning algorithms and achieves an accuracy rate of 95.38%.

The aforementioned studies all employ traditional ML techniques, characterized by the need for manual feature selection. Consequently, their results are highly dependent on the quality of feature selection. To overcome this limitation, new approaches have emerged. Recent studies explored the use of spectrograms as robust inputs for deep learning models in PD detection. Spectrograms provide a time-frequency representation that preserves both temporal dynamics and spectral details, which are crucial for capturing subtle vocal impairments. For instance, Iyer et al. (7) proposed a deep learning method based on spectrograms to identify PD. The researchers used a pre-trained Inception V3 to analyze the spectrograms of continuous vowels “/a/” produced by patients and healthy individuals, achieving high-precision classification. The Berus team employed multiple neural networks to identify vocal characteristics in PD patients (22), achieving 86.47% accuracy by analyzing various speech tasks constrained by linguistic skills and experimental settings. The Rios-Urrego team utilized a CNNs-based transfer learning approach to detect PD through Mel-scale spectrogram analysis of speech data, achieving 82% accuracy (23). Hireš et al. employed an ensemble model composed of multiple CNNs to detect PD lesions in spectrogram images of vowel recordings obtained under controlled conditions (15). Optimal performance was achieved on sustained/a/vowels, yielding an AUC value of 0.89. However, most spectrogram-based approaches have focused on sustained phonations, which may not fully capture the articulatory variability present in connected speech. In contrast, our work utilizes connected speech spectrograms, which encompass a wider range of phonetic and prosodic variations, thereby offering a more naturalistic and challenging scenario for robust PD classification.

Although deep learning methods avoid the need for manual feature engineering, current research often relies on a single model architecture. This limits their ability to comprehensively capture the multi-scale and subtle pathological patterns in PD speech. Therefore, in this paper, to address the limitations of relying solely on a single model, a feature fusion technique is proposed. Three pre-trained network models are used for feature fusion, and the fused features are then classified. Experimental results show that this network achieves the expected performance in the PD classification task.

3 Methodology

This paper employs three pre-trained models: MobileNetV3-large, DenseNet121, and ShuffleNetV2. The outputs of these models are adjusted to ensure consistent data size, followed by feature fusion from the extracted features. Feature classification was subsequently performed. To achieve consistent output across all models' feature extraction modules, MobileNet3's feature extraction module was aligned with DenseNet121 and ShuffleNetV2 by incorporating 1 × 1 convolutional layers.

3.1 Dataset and pre-processing

The dataset used in this study originates from a private collection gathered by the First Affiliated Hospital of Henan University of Science and Technology. Table 1 provides demographics of the participants. All PD patients were diagnosed in accordance with the Movement Disorder Society (MDS) Clinical Diagnostic Criteria for PD (24), excluding those with comorbid neurological disorders or severe hearing/speech impairments. Participants read a passage aloud, with their voices captured to form the speech dataset. The passage reads: “On a hot summer day, we sat together under a tree by the river, enjoying the cool breeze blowing through…” Each volunteer's reading session lasted approximately 1 min and 30 s. All recordings were captured in stereo mode and saved in WAV format. In the continuous speech task, the natural pauses and silences that occur when the subjects read the paragraphs are inherent components of the speech signal. In this study, these parts were not actively removed but were retained as part of the original speech signal and used for subsequent spectrogram generation. Specifically, during the speech segmentation process, we divided the speech based on sentence pauses and paragraph endings, ensuring that each speech segment contains complete speech units and their natural pauses. Although the silent segments appear as low-frequency energy regions in the spectrogram, they, together with the speech segments, constitute the overall expression of the time-frequency characteristics. The segmentation strategy was inspired by the approach of Almeida et al. (16) in sustained vowel analysis, using sentence pauses and paragraph endings as segmentation points to preserve the naturalness and representativeness of speech segments.

Table 1
www.frontiersin.org

Table 1. Demographics of participants considered in this study.

A spectrogram is a two-dimensional heatmap depicting the frequency components of a signal over time. In speech processing, the Short-Time Fourier Transform (STFT) is commonly applied to time-domain audio signals to capture the temporal distribution of spectral characteristics. In this paper, speech segments from the dataset underwent preprocessing. An acoustic feature extraction method based on the STFT is employed to convert the raw speech signal into a spectrogram representing its time-frequency characteristics, as shown in Figure 1. Spectrograms of HC subjects (Figure 1A) exhibit clear, regular harmonic stripes, indicating stable vocal fold vibration and consistent pitch control. Their formant structures (dark horizontal bands) are equally distinct and continuous, reflecting normal phonetic precision and breath support during sustained, continuous speech. In contrast, the spectrogram of PD patients (Figure 1B) exhibits reduced regularity in harmonic structure, accompanied by increased spectral noise and formant blurring. These features indicate irregular perturbations in fundamental frequency and amplitude due to impaired neuromuscular control in PD patients; Incomplete glottic closure increases non-periodic noise in the high-frequency range, while imprecise articulation blurs the transition zones between formants. These visual patterns align with PD pathophysiological mechanisms, positioning the spectrogram as a non-invasive surrogate marker for assessing speech function deterioration. Consequently, the proposed deep learning model can learn these discriminative time-frequency features to enable automated PD detection.

Figure 1
Two side-by-side color spectrograms labeled A and B. Spectrogram A shows distinct, horizontal yellow and green bands, indicating consistent sound frequencies. Spectrogram B has more varied, irregular patterns with diffuse colors, suggesting fluctuating frequencies.

Figure 1. Spectrograms derived from raw audio signals. (A) Healthy control group, (B) Parkinson's disease patients.

Let the discrete-time audio signal be x[n], with each input file resampled to 8 kHz. Divide x[n] into overlapping frames of duration Tw and stride Ts. Specifically, each frame contains a fixed number of samples Nw, and frames are separated by a sliding stride Ns, as shown in the following formula:

Nw=Tw×fs    (1)
NS=TS×fs    (2)

Where fs = 8,000 Hz, Tw = 2.0 s, Ts = 1.0 s. This setting aims to balance the temporal resolution and the integrity of frequency information. The fixed frame length helps the model capture speech features on a consistent time scale, while the STFT retains sufficient details in the frequency dimension to reflect the acoustic abnormalities commonly seen in PD patients, such as minor pitch perturbations and amplitude disturbances. The formula for calculating the total number of frames in each file is:

Nframes=max(1,L-NsNs+1)    (3)

Here, L denotes the total length of the waveform signal. If the frame extends beyond the signal's end, zero padding is applied to maintain a consistent length. For each frame xi[n] (I = 1,…, Nframes), the STFT is computed using a Hamming window of length N=256 (i.e., 32 ms) and a shift length Nh = 128 (50% overlap), with NFFT = 1,024 Fourier transform points. The STFT parameter settings were aligned with the recommendations of Moro-Velázquez et al. (25) for PD speech analysis to ensure discriminative time-frequency representations. For the i-th audio segment, the Fourier coefficient (complex number) corresponding to the k-th frequency point in the m-th frame is given by the following formula:

Xi(k,m)=n=0Nh-1xi[n+mNh]w[n]e-j2πkn/NFFT    (4)

Here, w[n] represents the Hamming window, where k = 0, 1,…, NFFT-1. n denotes the index of the current sample point within the window, with a range of 0<n<N. The resulting amplitude spectrum matrix is:

Si(k,m)=|Xi(k,m)|    (5)

To enhance the dynamic range and for file normalization, each spectrogram is converted to the decibel (dB) scale:

Si,dB(k,m)=10log10(Si(k,m)maxk,mSi(k,m)+ε)    (6)

Here, ε = 10−10 is used to avoid numerical instability.

Each segment's Si, dB data is rendered as a color image and saved in JPEG format. This process generates multiple spectrograms per recording, containing both time-domain and frequency-domain information for subsequent neural network model training. Through this procedure, a total of 2,476 spectrograms were obtained: 1,070 from PD and 1,406 from HC. The images have dimensions of 224 × 224 pixels. Subsequently, image intensity values were normalized to the range [0,1]. Finally, the data was divided into training and test set in an 8:2 ratio. A stratified sampling strategy was employed to ensure that the proportion of PD patients to HC in both the training and test set matched that of the original dataset.

It is worth noting that the speech material used in this study consists of read passages rather than isolated sustained vowels. Although read passages contain diverse phonemes and syllable structures that may introduce variability inherent to speech content, the purpose of employing a multi-frame sampling strategy was to capture patients' overall articulation characteristics within continuous speech, rather than relying on the acoustic properties of specific phonemes. This approach offers the advantage of reflecting the patient's articulatory capabilities in a more natural and representative speech task, encompassing aspects such as pitch control, speech fluency, and syllable transitions—all of which may be affected by PD. Furthermore, despite inherent spectral differences among phonemes, deep learning models possess the ability to learn high-level abstract features from diverse inputs, enabling them to distinguish pathological from non-pathological speech patterns rather than relying solely on the spectral characteristics of specific phonemes. Existing research also supports the use of continuous speech or read passages for PD detection. For instance, Almeida et al. (16) employed both sustained vowels and continuous speech tasks in their study, noting that continuous speech provides richer articulatory dynamics information in certain scenarios. Similarly, Vásquez-Correa et al. (23) incorporated multiple speech tasks, including reading aloud, into their multimodal PD detection system, demonstrating the effectiveness of continuous speech in PD identification. Therefore, while speech content diversity may introduce additional variability, this variability does not significantly impair the model's ability to recognize PD-related speech features when addressed through appropriate model design and training strategies.

3.2 The architecture of the proposed method

Transfer learning is a ML technique that leverages knowledge accumulated while solving one problem to train a model for another task or domain. This transfer learning approach utilizes pre-trained network knowledge acquired from vast amounts of visual data, offering significant advantages in time savings and improved accuracy compared to training models from scratch (26). Transfer learning is particularly advantageous when dealing with limited data, as pre-trained models bring robust feature representations learned from large-scale datasets that capture generic visual patterns such as edges, textures, and shapes (27). These representations can be effectively adapted to domain-specific tasks, even with a modest number of samples.

CNNs have been widely applied to image-related tasks such as image recognition and image classification. The use of CNNs has effectively improved the performance of many image-related tasks (28). This study employs a CNNs based on feature fusion, thereby enhancing the network's ability to extract features. Figure 2 illustrates the framework of the proposed method. First, PD data undergoes processing, and images are adjusted. Next, feature extraction from spectrograms is performed using pre-trained models. Finally, extracted features are aggregated for feature fusion before classification. Specifically, DenseNet121, MobileNetV3, and ShuffleNetV2 serve as pre-trained network models. The output from the feature extraction layer of MobileNetV3 is reshaped to dimensions (1024, 7, 7). Feature fusion via summation was selected after empirical comparison with alternatives including concatenation, weighted averaging, and attention-based fusion. Summation preserves the spatial dimensions of feature maps while reducing computational complexity compared to concatenation. It also encourages the model to integrate complementary information from different networks without introducing additional parameters, as would be required by attention mechanisms. After extracting features from the spectrogram, the features of each model are summed up pairwise to form a fused feature representation. In this study, a pairwise model fusion approach was adopted instead of a three-model fusion approach. The main reason was to avoid a significant increase in computational burden and potential feature redundancy. The experiments demonstrated that pairwise fusion was able to fully capture complementary information and achieve excellent performance. This fused feature map is then passed to a linear classifier consisting of a fully connected layer followed by a softmax activation function, which outputs the final probability distribution over the two classes. The fusion strategy employed is thus feature-level summation. This approach allows the model to leverage complementary information from different architectures directly in the feature space, enhancing the discriminative power for PD classification.

Figure 2
Diagram of a neural network architecture featuring three pathways: DenseNet121, MobileNetV3, and ShuffleNetV2. Each pathway starts with convolutional and pooling layers and progresses through various blocks. Outputs from each pathway are combined using addition operators, leading to linear layers producing outputs for PD or HC classification. The input image is sized at 224x224x3.

Figure 2. The proposed network model architecture.

3.3 Pre-trained models

In this study, pre-trained refers to CNN models that were initially trained on large-scale external datasets unrelated to PD or medical voice data. The three pre-trained models employed in this study were all pre-trained on the ImageNet dataset. ImageNet is a large-scale visual recognition dataset comprising over 14 million annotated images across 1,000 object categories, widely used for model pre-training in image classification tasks (29).

In typical CNNs, the convolutional layer extracts local features by sliding a fixed-size convolution kernel over the input tensor and performing element-wise multiplication of the kernel weights with corresponding input regions. Beyond the kernel parameters, the stride, padding, and kernel size jointly determine the spatial transformation of the feature map. The output spatial size of the convolutional layer is given by the following formula:

Hout=Hin+2p-ks+1,Wout=Win+2p-ks+1    (7)

Here, Hin and Win represent the height and width of the input feature map, respectively; p denotes the number of zero-padding pixels on each side; k indicates the size of the convolutional kernel; and s signifies the stride of the convolution. When k = 1, s = 1, and p = 0, the spatial dimensions of the output feature map remain unchanged, thereby helping to preserve high-resolution details.

The selection of DenseNet121, MobileNetV3-Large, and ShuffleNetV2 is based on their complementary architectural strengths for spectrogram analysis. DenseNet121 excels in feature propagation and reuse, ideal for capturing complex spectral patterns. MobileNetV3-Large offers a superior balance between accuracy and computational efficiency via depthwise convolutions and attention mechanisms. ShuffleNetV2 is a highly efficient model designed for low-resource environments, using channel shuffling to maintain representational power. Integrating these diverse models allows our fusion method to leverage their distinct advantages, thereby more comprehensively capturing the multi-scale pathological features in PD speech.

DenseNet was proposed by Huang et al. (30). The core idea of DenseNet is to introduce dense connections within the network. This design significantly enhances the capabilities of feature propagation and reuse, while also effectively mitigating the vanishing gradient problem. Figure 3 illustrates the basic framework of the DenseNet121 model. In the DenseNet architecture, each layer is directly connected to all preceding layers. Specifically, layer l receives all feature maps from the previous l1 layers as input, and its output is also fed to all subsequent layers. The output dimension of the feature extraction layer is (1024, 7, 7). This structure can be represented as:

xl=Hl([x0,x1,,xl-1])    (8)

Here, xl denotes the output of layer l; Hl(·) represents the nonlinear transformation function comprising batch normalization, the ReLU activation function, and convolution; [x0,x1,…,xl−1] denotes the concatenation of feature maps.

Figure 3
Diagram of a neural network architecture showing sequential layers: Conv 7x7 stride 2, Max Pool 3x3 stride 2, Dense Block 1, Conv 1x1 Average Pool, Dense Block 2, Conv 1x1 Average Pool, Dense Block 3, Conv 1x1 Average Pool, Dense Block 4, AdaptiveAvgPool, and Linear layer.

Figure 3. Structure of the DenseNet121 model.

MobileNetV3 is a lightweight network architecture proposed by the Google team (31), representing a significant upgrade following MobileNetV1 and V2. Figure 4 illustrates the fundamental framework of the MobileNetV3 model. Its design comprehensively integrates multiple optimization strategies. These include depthwise separable convolutions, inverted residual structures, squeeze-and-excitation channel attention mechanisms, and an efficient network architecture. This architecture compresses parameters and reduces computational complexity while preserving strong feature representation capabilities. As one of the representative models in the current edge computing field, it achieves a balance between precision and efficiency. The feature output dimension from the feature extraction layer of the MobileNetV3 model is (960, 7, 7). To enable feature fusion with DenseNet121 and ShuffleNetV2, a 1 × 1 convolutional layer with 1,024 convolutional kernels is added to the base model, adjusting the feature extraction layer output dimension to (1024, 7, 7).

Figure 4
Diagram of a neural network architecture. It includes a Conv 3x3 layer with stride 2, followed by Bneck 3x3 repeated three times, Bneck 5x5 three times, Bneck 3x3 six times, Bneck 5x5 three times, Conv 1x1, AvgPool, and finally a Linear layer. Arrows indicate the flow of data between layers.

Figure 4. Structure of the MobileNetV3-Large model.

ShuffleNetV2 is an efficient, lightweight CNNs proposed by Megvii Technology team (32), designed to achieve faster inference speeds and lower resource consumption on mobile and embedded devices. Its core concept lies in breaking communication barriers between different groups through channel grouping and feature channel shuffling operations, thereby enhancing feature flow efficiency and network expressiveness. Structurally, the basic unit of ShuffleNetV2 comprises two branches. The outputs from both branches undergo channel concatenation and are combined with channel shuffle operations to facilitate information exchange, achieving parameter efficiency and computational friendliness. Figure 5 illustrates the basic framework of the ShuffleNetV2 model. This structure is stacked repeatedly to build the complete network, offering excellent scalability. The output dimension of the feature extraction layer is (1024, 7, 7).

Figure 5
Diagram showing a sequence of neural network layers: a convolutional layer (Conv 3x3 stride 2), max pooling, stage 2 repeated four times, stage 3 repeated eight times, stage 4 repeated four times, a 1x1 convolution, global pooling, and a linear layer.

Figure 5. Structure of the ShuffleNetV2 model.

4 Results

The experimental environment in this study utilized a Windows 10 operating system with 64GB of random access memory, Python 3.10.15, and PyTorch 2.1.0. Employing a cosine annealing scheduling strategy. Hyperparameters were primarily selected based on established defaults for the Adamw optimizer and adjusted through manual trial and error guided by performance on a held-out test set. Detailed information regarding the hardware configuration and model parameters is provided in Table 2.

Table 2
www.frontiersin.org

Table 2. Hardware configuration and model parameters.

4.1 Evaluation metric

To comprehensively evaluate the model's effectiveness, this paper employs evaluation metrics including accuracy, precision, recall, F1-score, and specificity. The expressions for these evaluation metrics are shown in Equations 913.

Accuracy=TP+TNTP+TN+FP+FN    (9)
Precision=TPTP+FP    (10)
Recall=TPTP+FN    (11)
F1-score=2×Precision×RecallPrecision+Recall    (12)
Specificity=TNTN+FP    (13)

Among these, TP (True Positives) represents the number of correctly predicted positives; TN (True Negatives) represents the number of correctly predicted negatives; FP (False Positives) represents the number of incorrectly predicted positives; and FN (False Negatives) represents the number of incorrectly predicted negatives.

4.2 Experimental results

This section presents the classification results obtained by the proposed method through 5-fold cross-validation, and conducts a comparative analysis between the use and non-use of the feature fusion method.

4.2.1 The classification results of a single model

To more intuitively reflect the overall classification performance of the models, the confusion matrices of the three pre-trained models on the dataset are plotted as shown in Figure 6. Additionally, Table 3 lists the specific values of accuracy, precision, recall, F1-score, and specificity calculated using Formulas 9–13. Table 3 indicates that DenseNet121 demonstrates the best PD classification performance, achieving an accuracy of 92.73%.

Figure 6
Three confusion matrix images labeled A, B, and C show true versus predicted labels for HC and PD. Image A: 263 true positives for HC, 196 for PD, 18 false positives for both. Image B: 260 HC, 195 PD true positives, 21 HC false, 19 PD false. Image C: 262 HC, 192 PD true positives, 19 HC, 22 PD false predictions. Each has a color gradient indicating frequency counts.

Figure 6. Confusion matrix of predicted results for a single model on the test set. (A) DenseNet121 (B) MobileNetV3, and (C) ShuffleNetV2.

Table 3
www.frontiersin.org

Table 3. Indicators for the classification of a single model.

4.2.2 The classification results after feature fusion

Figure 7 displays the confusion matrix for PD classification results achieved through feature fusion. The diagonal region representing correctly classified cases exhibits distinct dark clustering characteristics, with numerical values highly concentrated along the diagonal. Additionally, Table 4 provides detailed metric values evaluated on the dataset. It can be observed that MobileNetV3+ShuffleNetV2 achieved the best classification results on the dataset with an accuracy of 95.56% and an Area Under Curve of 99%. This validates the effectiveness and reliability of this model for PD classification tasks.

Figure 7
Three confusion matrices labeled A, B, and C compare predicted and true labels for HC and PD. Matrix A shows 267 true positives, 14 false positives, 13 false negatives, and 201 true negatives. Matrix B has 270 true positives, 11 false positives, 13 false negatives, and 201 true negatives. Matrix C presents 271 true positives, 10 false positives, 12 false negatives, and 202 true negatives. Color intensity varies with counts.

Figure 7. Classification results of PD on the test set. (A) DenseNet121+ShuffleNetV2 (B) DenseNet121+MobileNetV3, and (C) MobileNetV3+ShuffleNetV2.

Table 4
www.frontiersin.org

Table 4. The classification results of feature fusion methods.

In order to more intuitively demonstrate the advantages of the fusion model, Figure 8 shows the average evaluation metrics for PD classification across each model. It can be observed that the combination of MobileNetV3 and ShuffleNetV2 (MobileNetV3+ShuffleNetV2) achieves the best classification accuracy, precision, recall, and F1-score, at 95.56%, 95.42%, 95.52%, and 95.47%, respectively. Additionally, among individual models, DenseNet121 demonstrated the best classification results for PD. All performance metrics of the feature-fused hybrid models outperformed the respective metrics of the individual models. The experimental results validate the effectiveness of combining features from different models through feature fusion. This provides a more reliable approach for PD classification than relying on a single model.

Figure 8
Bar chart comparing the performance of six models: DenseNet121, ShuffleNetV2, DenseNet+ShuffleNetV2, MobileNetV3, DenseNet+MobileNetV3, and MobileNetV3+ShuffleNetV2. Metrics include accuracy, precision, recall, and F1-score, with all values centered around 90-96%. DenseNet+MobileNetV3 performs slightly better across most metrics.

Figure 8. Visualization of PD classification metrics.

4.2.3 Robustness validation

Based on the foregoing, it is evident that the feature fusion model yields superior classification results across the private dataset. This section aims to evaluate the model's robustness by validating it on the publicly available PD speech dataset from Figshare (33). Using the same data processing methodology, a total of 1,034 spectrograms were obtained. Among these, 536 were from PD patients and 498 from HC. Results from 5-fold cross-validation are presented in Table 5. The findings indicate that high accuracy is maintained even on the public dataset, demonstrating the model's robustness.

Table 5
www.frontiersin.org

Table 5. Classification results of feature fusion methods on public datasets.

4.3 Comparison with other state-of-the-art methods

To further validate the effectiveness of the proposed method, this study conducted comprehensive comparative experiments between the proposed network and a series of state-of-the-art models. In order to conduct fair and direct comparisons under the same conditions, we implemented the core algorithms and architectural frameworks from each literature, and applied them to our own dataset for training and evaluation. The experimental results are shown in Table 6. The results demonstrate that the proposed model exhibits outstanding performance, achieving an accuracy of 95.56%, precision of 95.42%, recall of 95.52%, F1-score of 95.47%, and specificity of 95.52%, outperforming all other models. The comparative results presented in Table 5 indicate that our research achieves competitive classification performance compared to the state-of-the-art methods in the current literature.

Table 6
www.frontiersin.org

Table 6. Comparison with other state-of-the-art models.

5 Discussion

This study investigates the effectiveness of pre-trained CNNs and feature fusion strategies in PD classification based on spectrograms, yielding several important findings worthy of further exploration.

The superior performance of the DenseNet121 model, which achieved a peak accuracy of 92.73%, is attributable to its densely connected architecture. As detailed in Section 3.3.1, this design enhances feature propagation and reuse, effectively mitigates the vanishing gradient problem, and promotes robust feature integration. These properties likely explain its efficacy in capturing the subtle acoustic patterns indicative of PD voice impairment. In contrast, the lightweight architectures of MobileNetV3-Large and ShuffleNetV2—optimized for computational efficiency on edge devices—achieved slightly lower accuracies of 91.92% and 91.72%, respectively. This performance discrepancy illustrates a fundamental trade-off between model efficiency and representational capacity. While MobileNet3's squeeze-and-excite modules and ShuffleNet2's channel shuffling operation prioritize inference speed, DenseNet11's structure favors comprehensive feature extraction.

The experimental results demonstrate an improvement in the performance of the feature fusion model. This enhancement indicates that integrating features from models based on distinct design philosophies can strengthen discriminative capabilities. The fusion model consistently outperforms results from individual models, highlighting the limitations of relying on features extracted from a single model architecture for PD classification. Pathological voice characteristics in PD may manifest across a range of spectral bands (e.g., low-frequency harmonics, high-frequency noise components) and temporal segments (e.g., vowel pronunciation, transitional speech segments), which are not equally captured by all models.

Despite the aforementioned strengths of this study, limitations remain. The dataset originates from a single institution, comprising 61 PD patients and 70 HC, with speech recordings confined to a single text. This may limit the study's generalizability, as speech patterns can vary across dialects, age groups, and task complexity. Future research should focus on expanding datasets to incorporate diverse linguistic materials and populations, exploring cross-lingual transfer learning, and integrating additional modalities (such as gait or handwriting.) to achieve multimodal PD diagnosis (34, 35).

In conclusion, this study demonstrates that the feature fusion of pre-trained CNNs can improve the performance of PD classification based on spectrograms, providing a promising approach for auxiliary diagnosis. The research results emphasize the importance of multi-model integration in capturing complex pathological biomarkers and also lay the foundation for the development of deployable PD screening tools.

6 Conclusion

This study aims to overcome the limitations of traditional PD diagnostic methods, including their invasiveness, high cost, and dependence on specialist expertise. Additionally, it addresses the drawbacks of single DL models in voice-based diagnosis to pursue a non-invasive and reliable auxiliary diagnostic solution. To this end, this study utilized the dataset comprising 61 PD patients and 70 HC. The raw speech signals were preprocessed using STFT to generate 2,476 spectrograms. Three pre-trained CNNs—DenseNet121, MobileNetV3-Large, and ShuffleNetV2—were selected for feature extraction. To ensure consistent output dimensions with the other two models, the output dimension of MobileNetV3-Large was adjusted to (1024, 7, 7) via a 1 × 1 convolutional layer, followed by feature fusion through summation. Experimental results demonstrated that all feature fusion models outperformed single models. Among the single models, DenseNet121 achieved the highest accuracy at 92.73%. The fusion model combining MobileNetV3-Large and ShuffleNetV2 achieved optimal classification performance with an accuracy of 95.56% and an AUC value of 0.99, outperforming existing state-of-the-art methods reported in the literature. This result confirms that integrating CNNs models with distinct architectural strengths enables more comprehensive capture of subtle pathological speech patterns in PD patients, thereby enhancing diagnostic accuracy. Although this study contributes to providing a non-invasive PD diagnostic aid, limitations exist. For instance, speech recordings were restricted to a single language, potentially affecting the study's generalizability. Future work will focus on expanding datasets to include diverse populations and linguistic materials. Cross-lingual transfer learning and integrating multimodal data will be explored to further improve the model's diagnostic performance and clinical applicability.

Data availability statement

The datasets presented in this article are not readily available because the data that has been used is confidential. Requests to access the datasets should be directed to Wenna Chen: Y2hlbndlbm5hMDQwOEAxNjMuY29t.

Ethics statement

The studies involving humans were approved by the Ethics Committee of the First Affiliated Hospital of Henan University of Science and Technology. The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their written informed consent to participate in this study. Written informed consent was obtained from the individual(s) for the publication of any potentially identifiable images or data included in this article.

Author contributions

WC: Conceptualization, Funding acquisition, Methodology, Project administration, Resources, Writing – original draft, Writing – review & editing. RL: Data curation, Investigation, Software, Supervision, Validation, Writing – original draft, Writing – review & editing. XD: Data curation, Investigation, Validation, Writing – original draft, Writing – review & editing. XC: Data curation, Investigation, Writing – original draft. HW: Data curation, Investigation, Writing – original draft. JZ: Formal analysis, Validation, Writing – original draft. GD: Software, Validation, Writing – original draft.

Funding

The author(s) declare that financial support was received for the research and/or publication of this article. This research was supported by the Major Science and Technology Projects of Henan Province (No. 221100210500).

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declare that no Gen AI was used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

1. Shen M, Mortezaagha P, Rahgozar A. Explainable artificial intelligence to diagnose early Parkinson's disease via voice analysis. Sci Rep. (2025) 15:11687. doi: 10.1038/s41598-025-96575-6

PubMed Abstract | Crossref Full Text | Google Scholar

2. Bhoomika R, Shahane S, Siri TC, Rao T, Kodipalli A, Chodon PK. Ensemble learning approaches for detecting Parkinson's disease. In: 2022 IEEE 9th Uttar Pradesh Section International Conference on Electrical, Electronics and Computer Engineering (UPCON). Prayagraj: IEEE (2022). p. 1–6. doi: 10.1109/UPCON56432.2022.9986450

Crossref Full Text | Google Scholar

3. Kanakaprabha S, Arulprakash P, Srikanth R. Parkinson disease detection using various machine learning algorithms. In: 2022 International Conference on Advanced Computing Technologies and Applications (ICACTA) Coimbatore: IEEE (2022). p. 1–6. doi: 10.1109/ICACTA54488.2022.9752925

Crossref Full Text | Google Scholar

4. Singh K, Khare M, Khare A, Kohli N. Review on computational methods for the detection and classification of Parkinson's Disease. Comput Biol Med. (2025) 187:109767. doi: 10.1016/j.compbiomed.2025.109767

PubMed Abstract | Crossref Full Text | Google Scholar

5. Umar MM, Naaz S. Classification using radial basis function for prediction of Parkinson's disease. In: 2022 IEEE 3rd Global Conference for Advancement in Technology (GCAT). Bangalore: IEEE (2022). p. 1–6. doi: 10.1109/GCAT55367.2022.9971828

Crossref Full Text | Google Scholar

6. Kanagaraj S, Hema MS, Guptha MN, Namitha V. Detecting Parkinsons disease with image classification. In: 2022 IEEE 3rd Global Conference for Advancement in Technology (GCAT). Bangalore: IEEE (2022). p. 1–5. doi: 10.1109/GCAT55367.2022.9971993

Crossref Full Text | Google Scholar

7. Iyer A, Kemp A, Rahmatallah Y, Pillai L, Glover A, Prior F, et al. machine learning method to process voice samples for identification of Parkinson's disease. Sci Rep. (2023) 13:20615. doi: 10.1038/s41598-023-47568-w

PubMed Abstract | Crossref Full Text | Google Scholar

8. Gene L, Alim A, Imtiaz MH, Schuckers S. A machine learning model for early prediction of Parkinson's disease from wearable sensors. In: 2023 IEEE 13th Annual Computing and Communication Workshop and Conference (CCWC). Las Vegas, NV: IEEE (2023). p. 734–7. doi: 10.1109/CCWC57344.2023.10099230

Crossref Full Text | Google Scholar

9. Alshammri R, Alharbi G, Alharbi E, Almubark I. Machine learning approaches to identify Parkinson's disease using voice signal features. Front Artif Intell. (2023) 6:1084001. doi: 10.3389/frai.2023.1084001

PubMed Abstract | Crossref Full Text | Google Scholar

10. Xu J, Pan Y, Pan X, Hoi S, Yi Z, Xu Z. RegNet: self-regulated network for image classification. IEEE Trans Neural Netw Learn Syst. (2023) 34:9562–7. doi: 10.1109/TNNLS.2022.3158966

PubMed Abstract | Crossref Full Text | Google Scholar

11. Xu W, Fu Y-L, Zhu D. ResNet and its application to medical image processing: research progress and challenges. Comput Methods Programs Biomed. (2023) 240:107660. doi: 10.1016/j.cmpb.2023.107660

PubMed Abstract | Crossref Full Text | Google Scholar

12. Liu Z, Mao H, Wu C-Y, Feichtenhofer C, Darrell T, Xie S. A ConvNet for the 2020s. In: 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). New Orleans, LA: IEEE (2022). p. 11966–76. doi: 10.1109/CVPR52688.2022.01167

Crossref Full Text | Google Scholar

13. Han K, Wang Y, Chen H, Chen X, Guo J, Liu Z, et al. A survey on vision transformer. IEEE Trans Pattern Anal Mach Intell. (2023) 45:87–110. doi: 10.1109/TPAMI.2022.3152247

PubMed Abstract | Crossref Full Text | Google Scholar

14. Liu Z, Lin Y, Cao Y, Hu H, Wei Y, Zhang Z, et al. Swin transformer: hierarchical vision transformer using shifted windows. In: 2021 IEEE/CVF International Conference on Computer Vision (ICCV). Montreal, QC: IEEE (2021). p. 9992–10002 doi: 10.1109/ICCV48922.2021.00986

Crossref Full Text | Google Scholar

15. Hireš M, Gazda M, Drotár P, Pah ND, Motin MA, Kumar DK. Convolutional neural network ensemble for Parkinson's disease detection from voice recordings. Comput Biol Med. (2022) 141:105021. doi: 10.1016/j.compbiomed.2021.105021

PubMed Abstract | Crossref Full Text | Google Scholar

16. Almeida JS, Filho PPR, Carneiro T, Wei W, Damaševičius R, Maskeliunas R, et al. Detecting Parkinson's disease with sustained phonation and speech signals using machine learning techniques. Pattern Recognit Lett. (2019) 125:55–62. doi: 10.1016/j.patrec.2019.04.005

Crossref Full Text | Google Scholar

17. Gómez-García JA, Moro-Velázquez L, Godino-Llorente JI. On the design of automatic voice condition analysis systems. Part I: Review of concepts and an insight to the state of the art. Biomed Signal Process Control. (2019) 51:181–99. doi: 10.1016/j.bspc.2018.12.024

Crossref Full Text | Google Scholar

18. Mekyska J, Janousova E, Gomez-Vilda P, Smekal Z, Rektorova I, Eliasova I, et al. Robust and complex approach of pathological speech signal analysis. Neurocomputing. (2015) 167:94–111. doi: 10.1016/j.neucom.2015.02.085

Crossref Full Text | Google Scholar

19. Carrón J, Campos-Roca Y, Madruga M, Pérez CJ, A. mobile-assisted voice condition analysis system for Parkinson's disease: assessment of usability conditions. BioMed Eng Online. (2021) 20:114. doi: 10.1186/s12938-021-00951-y

PubMed Abstract | Crossref Full Text | Google Scholar

20. Saleh S, Alouani Z, Daanouni O, Hamida S, Cherradi B, Bouattane O. AIoT-based embedded systems optimization using feature selection for Parkinson's disease diagnosis through speech disorders. Intell-Based Med. (2024) 10:100184. doi: 10.1016/j.ibmed.2024.100184

Crossref Full Text | Google Scholar

21. Saleh S, Cherradi B, Laghmati S, Gannour OE, Hamida S, Bouattane O. Healthcare embedded system for predicting Parkinson's disease based on ai of things. In: 2023 3rd International Conference on Innovative Research in Applied Science, Engineering and Technology (IRASET). Mohammedia: IEEE (2023). p. 1–7 doi: 10.1109/IRASET57153.2023.10153040

Crossref Full Text | Google Scholar

22. Berus L, Klancnik S, Brezocnik M, Ficko M. Classifying Parkinson's disease based on acoustic measures using artificial neural networks. Sensors. (2018) 19:16. doi: 10.3390/s19010016

PubMed Abstract | Crossref Full Text | Google Scholar

23. Rios-Urrego CD, Vásquez-Correa JC, Orozco-Arroyave JR, Nöth E. Transfer learning to detect Parkinson's disease from speech in different languages using convolutional neural networks with layer freezing. In:Sojka P, Kopeček I, Pala K, Horák A, , editors. Text, Speech, and Dialogue. Cham: Springer International Publishing (2020). p. 331–9. doi: 10.1007/978-3-030-58323-1_36

Crossref Full Text | Google Scholar

24. Postuma RB, Berg D, Stern M, Poewe W, Olanow CW, Oertel W, et al. MDS clinical diagnostic criteria for Parkinson's disease: MDS-PD clinical diagnostic criteria. Mov Disord. (2015) 30:1591–601. doi: 10.1002/mds.26424

PubMed Abstract | Crossref Full Text | Google Scholar

25. Moro-Velazquez L, Gomez-Garcia JA, Arias-Londoño JD, Dehak N, Godino-Llorente JI. Advances in Parkinson's disease detection and assessment using voice and speech: a review of the articulatory and phonatory aspects. Biomed Signal Process Control. (2021) 66:102418. doi: 10.1016/j.bspc.2021.102418

Crossref Full Text | Google Scholar

26. Kim HE, Cosa-Linan A, Santhanam N, Jannesari M, Maros ME, Ganslandt T. Transfer learning for medical image classification: a literature review. BMC Med Imaging. (2022) 22:69. doi: 10.1186/s12880-022-00793-7

PubMed Abstract | Crossref Full Text | Google Scholar

27. Iman M, Arabnia HR, Rasheed K. A review of deep transfer learning and recent advancements. Technologies. (2023) 11:40. doi: 10.3390/technologies11020040

Crossref Full Text | Google Scholar

28. Hussain SS, Degang X, Shah PM, Islam SU, Alam M, Khan IA, et al. Classification of Parkinson's disease in patch-based MRI of Substantia Nigra. Diagnostics. (2023) 13:2827. doi: 10.3390/diagnostics13172827

PubMed Abstract | Crossref Full Text | Google Scholar

29. Hosna A, Merry E, Gyalmo J, Alom Z, Aung Z, Azim MA. Transfer learning: a friendly introduction. J Big Data. (2022) 9:102. doi: 10.1186/s40537-022-00652-w

PubMed Abstract | Crossref Full Text | Google Scholar

30. Huang G, Liu Z, Van Der Maaten L, Weinberger KQ. Densely connected convolutional networks. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). Honolulu, HI: IEEE (2017). doi: 10.1109/CVPR.2017.243

Crossref Full Text | Google Scholar

31. Howard A, Sandler M, Chen B, Wang W, Chen L-C, Tan M, et al. Searching for MobileNetV3. In: 2019 IEEE/CVF International Conference on Computer Vision (ICCV). Seoul: IEEE (2019). p. 1314–24. doi: 10.1109/ICCV.2019.00140

Crossref Full Text | Google Scholar

32. ShuffleNet V2: Practical guidelines for efficient CNN architecture design. In: Lecture Notes in Computer Science. Cham: Springer International Publishing (2018). p. 122–38. doi: 10.1007/978-3-030-01264-9_8

Crossref Full Text | Google Scholar

33. Prior F, Virmani T, Iyer A, Larson-Prior L, Kemp A, Rahmatallah Y, et al. Voice samples for patients with Parkinson's disease and healthy controls (2023). doi: 10.6084/m9.figshare.23849127.v1

Crossref Full Text | Google Scholar

34. Saleh S, Ouhmida A, Cherradi B, Al-Sarem M, Hamida S, Alblwi A, et al. novel hybrid CNN-KNN ensemble voting classifier for Parkinson's disease prediction from hand sketching images. Multimed Tools Appl. (2024) 84:36401–33. doi: 10.1007/s11042-024-19314-5

Crossref Full Text | Google Scholar

35. Saleh S, Cherradi B, El Gannour O, Hamida S, Bouattane O. Predicting patients with Parkinson's disease using machine learning and ensemble voting technique. Multimed Tools Appl. (2023) 83:33207–34. doi: 10.1007/s11042-023-16881-x

Crossref Full Text | Google Scholar

36. Han K, Wang Y, Tian Q, Guo J, Xu C, Xu C. GhostNet: more features from cheap operations. In: 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). Seattle, WA: IEEE (2020). p. 1577–86. doi: 10.1109/CVPR42600.2020.00165

Crossref Full Text | Google Scholar

37. Radosavovic I, Kosaraju RP, Girshick R, He K, Dollar P. Designing network design spaces. In: 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). Seattle, WA: IEEE (2020). p. 10425–33. doi: 10.1109/CVPR42600.2020.01044

Crossref Full Text | Google Scholar

38. Sunkara R, Luo T. No more strided convolutions or pooling: a new CNN building block for low-resolution images and small objects. In:Amini M-R, Canu S, Fischer A, Guns T, Kralj Novak P, Tsoumakas G, , editors. Machine Learning and Knowledge Discovery in Databases. Cham: Springer Nature Switzerland (2023). p. 443–59. doi: 10.1007/978-3-031-26409-2_27

Crossref Full Text | Google Scholar

39. Yu Y, Zhang Y, Cheng Z, Song Z, Tang C, MCA. Multidimensional collaborative attention in deep convolutional neural networks for image recognition. Eng Appl Artif Intell. (2023) 126:107079. doi: 10.1016/j.engappai.2023.107079

Crossref Full Text | Google Scholar

40. Woo S, Debnath S, Hu R, Chen X, Liu Z, Kweon IS, et al. ConvNeXt V2: co-designing and scaling ConvNets with masked autoencoders. In: 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). Vancouver, BC: IEEE (2023). p. 16133–42. doi: 10.1109/CVPR52729.2023.01548

Crossref Full Text | Google Scholar

41. Ma X, Dai X, Bai Y, Wang Y, Fu Y. Rewrite the stars. In: 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). Seattle, WA: IEEE (2024). p. 5694–703. doi: 10.1109/CVPR52733.2024.00544

Crossref Full Text | Google Scholar

42. Huo X, Sun G, Tian S, Wang Y, Yu L, Long J, et al. HiFuse: hierarchical multi-scale feature fusion network for medical image classification. Biomed Signal Process Control. (2024) 87:105534. doi: 10.1016/j.bspc.2023.105534

Crossref Full Text | Google Scholar

43. Hou Q, Lu C-Z, Cheng M-M, Feng J. Conv2Former: a simple transformer-style ConvNet for visual recognition. IEEE Trans Pattern Anal Mach Intell. (2024) 46:8274–83. doi: 10.1109/TPAMI.2024.3401450

PubMed Abstract | Crossref Full Text | Google Scholar

44. Yu Y, Zhang Y, Cheng Z, Song Z, Tang C. Multi-scale spatial pyramid attention mechanism for image recognition: an effective approach. Eng Appl Artif Intell. (2024) 133:108261. doi: 10.1016/j.engappai.2024.108261

Crossref Full Text | Google Scholar

45. Ding X, Zhang Y, Ge Y, Zhao S, Song L, Yue X, et al. UniRepLKNet: a universal perception large-kernel ConvNet for audio, video, point cloud, time-series and image recognition. In: 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2024). p. 5513–24. doi: 10.1109/CVPR52733.2024.00527

Crossref Full Text | Google Scholar

46. Li S, Wang Z, Liu Z, Tan C, Lin H, Wu D, et al. MogaNet: multi-order gated aggregation network. In: The Twelfth International Conference on Learning Representations (2024). Available online at: https://openreview.net/forum?id=XhYWgjqCrV (Accessed October 25, 2025).

Google Scholar

Keywords: Parkinson's detection, deep learning, feature fusion, convolutional neural networks, transfer learning

Citation: Chen W, Lv R, Du X, Chen X, Wang H, Zhang J and Du G (2025) Parkinson's disease detection using spectrogram-based multi-model feature fusion networks. Front. Neurol. 16:1706317. doi: 10.3389/fneur.2025.1706317

Received: 16 September 2025; Accepted: 14 October 2025;
Published: 05 November 2025.

Edited by:

Gaowei Xu, Tongji University, China

Reviewed by:

Dafni Rose, St. Joseph's Institute of Technology, India
Jiu-Cheng Xie, Nanjing University of Posts and Telecommunications, China

Copyright © 2025 Chen, Lv, Du, Chen, Wang, Zhang and Du. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Wenna Chen, Y2hlbndlbm5hMDQwOEAxNjMuY29t; Ganqin Du, ZGdxOTlAMTYzLmNvbQ==

These authors have contributed equally to this work

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.