Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Neurol., 10 September 2025

Sec. Neuromuscular Disorders and Peripheral Neuropathies

Volume 16 - 2025 | https://doi.org/10.3389/fneur.2025.1640428

Development of a deep learning model for automated diagnosis of neuromuscular diseases using ultrasound imaging

  • 1School of Physical Education, Central China Normal University, Wuhan, China
  • 2Department of Physical Education and Military Training, Zhejiang Agriculture and Forestry University, Hangzhou, China

Background: Neuromuscular diseases (NMDs) pose significant diagnostic challenges due to their heterogeneous clinical manifestations and the limitations of traditional diagnostic tools. While musculoskeletal ultrasound has become a promising non-invasive modality for evaluating muscle pathology, its diagnostic accuracy remains heavily dependent on the operator’s expertise. To address this, we propose a lightweight and interpretable deep learning model to enable automated classification of ultrasound images in NMD screening.

Method: We developed a novel model, termed NMD-AssistNet, which integrates GhostNet as the backbone with CBAM attention modules and depthwise separable convolutions to enhance both efficiency and discriminative capacity. The model was trained and evaluated on a public dataset containing 3,917 annotated ultrasound images of various muscle groups. Mixup augmentation, label smoothing, and SWALR learning rate scheduling were applied to improve generalizability. Performance was benchmarked against CSPNet, EfficientNet, GhostNet, HRNet, and Vision Transformer.

Results: NMD-AssistNet achieved the highest performance among the evaluated models, reaching a classification accuracy of 0.9502 and an area under the curve (AUC) of 0.9776. Grad-CAM visualizations revealed that the model effectively focused on clinically relevant muscle regions, highlighting its potential interpretability.

Conclusion: NMD-AssistNet demonstrates strong diagnostic capability, computational efficiency, and model interpretability and offers a promising solution for real-time, automated NMD screening. This framework has the potential to be deployed in portable ultrasound systems or edge AI devices to assist clinicians in both hospital and community healthcare settings.

1 Introduction

Neuromuscular diseases (NMDs) are a class of complex diseases that originate from the anterior horn of the spinal cord, peripheral nerves, neuromuscular junctions, or skeletal muscles. Their clinical manifestations include a variety of symptoms such as muscle atrophy, decreased muscle strength, impaired motor function, and even respiratory insufficiency (1). Different types of NMDs vary significantly in disease progression, prognosis, and treatment response. Clinical intervention strategies are also highly dependent on the accurate identification of disease types and their stages (2). While a detailed neurological examination is the cornerstone of diagnosis and can often differentiate the level of the lesion, the diagnostic process can still be challenging and protracted. This is particularly true in the early stages of the disease, where symptoms may be subtle or overlap with other neurological conditions, necessitating objective diagnostic aids to supplement clinical assessment (3).

Ultrasound imaging, as a non-invasive, dynamic and repeatable imaging technology, has been widely used in the clinical evaluation of muscle diseases in recent years. Compared with traditional methods such as electromyography (EMG) or muscle biopsy, ultrasound can not only reflect changes in muscle structure in real time, such as changes in muscle thickness, echo intensity and texture pattern, but can also be used to dynamically observe the muscle’s behavior during movement, thereby assisting in the assessment of disease activity and treatment effect (4). However, this imaging technology is highly dependent on the operator’s technical level and clinical experience, and different doctors may have different interpretations of the same image. In addition, the performance of muscle tissue in ultrasound images is often complex and variable due to disease type, individual differences and different scanning angles. Factors such as blurred lesion boundaries and overlapping tissue echoes may interfere with diagnostic judgment (5).

To address the shortcomings of traditional image interpretation methods, the application of artificial intelligence (AI), especially deep learning, in the field of medical image processing has gradually deepened in recent years. Recent comprehensive reviews have highlighted that AI, and particularly deep learning, holds significant promise for transforming diagnostic workflows in neuromuscular medicine (6). A convolutional neural network (CNN) is a specialized deep learning architecture for processing grid-like data, such as an image. Its key advantage is the ability to automatically learn relevant features from the input data through the use of convolutional filters. In the imaging diagnosis of neuromuscular diseases, CNNs are widely used for automatic feature extraction and image classification tasks. Their advantage is that they can autonomously learn and extract multi-level texture information from complex image data (7). At the same time, transfer learning strategies have been introduced to overcome the limitations of small sample data sets on model training, and have shown good adaptability under the realistic conditions of scarce medical images (8). The multi-task learning framework improves the spatial sensitivity and discrimination of the model to the lesion area by simultaneously optimizing target tasks such as classification and segmentation, thereby improving the overall diagnostic performance (9).

Although the above-mentioned technological progress is encouraging, there are still many technical bottlenecks that limit its clinical promotion. First, the performance of existing models in real ultrasound scenes facing high noise, low contrast and complex backgrounds is still unstable, and it is prone to misjudgment and missed diagnosis (10). More importantly, the vast majority of studies are based on small, single-center datasets, and the generalization ability and robustness of the model have not been fully verified in a multi-institutional and heterogeneous equipment environment, limiting its translation from scientific research to clinical practice. Therefore, developing a lightweight, high-performance neural network architecture with good generalization ability to improve the efficiency and stability of auxiliary diagnosis of neuromuscular diseases in ultrasound images has become an important direction of current medical artificial intelligence research.

To address the aforementioned challenges, this paper proposes NMD-AssistNet, a novel lightweight deep learning model for the efficient and automated classification of neuromuscular diseases from ultrasound images. Based on the GhostNet architecture (11), this model introduces the CBAM (Convolutional Block Attention Module) that integrates channel attention and spatial attention to enhance the feature focusing ability of key lesion areas. At the same time, it combines depthwise separable convolution and a channel shuffling mechanism to significantly reduce the number of model parameters and computational complexity, thereby adapting to the characteristics of complex texture details and blurred boundaries in ultrasound images. In addition, in order to improve the generalization ability and training stability of the model, we also introduced the Mixup data enhancement strategy, label smoothing regularization, and SWALR (Stochastic Weight Averaging with Learning Rate Scheduling). The experimental part is based on a real clinical ultrasound image dataset, and systematically compares its performance with multiple mainstream models on multiple classification performance indicators, and analyzes its diagnostic accuracy and clinical interpretability through attention heat maps (Figure 1).

Figure 1
Flowchart illustrating a medical imaging analysis process: input ultrasound images undergo data preprocessing with standardization, noise removal, and data augmentation. NMD-AssistNet processes data using GhostNet, CBAM, and convolutions. The model predicts healthy or pathological states, with label smoothing and SWALR. Interpretation is shown with heatmaps, and evaluation includes an ROC curve. Results are displayed as a matrix of true positives, false positives, true negatives, and false negatives, along with bar charts.

Figure 1. The overall technical route of this study.

2 Methods

2.1 Data collection

This study utilized a public ultrasound image dataset from the Mendeley data repository (DOI: 10.17632/3jykz7wz8d.1), originally collected in the Arnhem-Nijmegen region. The dataset is governed by the Creative Commons Attribution 4.0 (CC BY 4.0) license, which explicitly permits data reuse for any research purpose with proper citation, thus no separate institutional review board approval was required for our secondary analysis (12). The dataset contains data from a total of 1,283 subjects (average age 50 ± 21 years, 729 males), including 3,917 cross-sectional ultrasound images, covering three common skeletal muscle groups: biceps brachii, tibialis anterior, and medial gastrocnemius. All images were manually annotated by experienced clinical ultrasound experts to ensure the accuracy of anatomical structure identification and classification labels (13). To classify muscle images into healthy and pathological categories, the original research team used a z-score method based on grayscale values. The z-score was calculated using linear regression models derived from a healthy population, incorporating variables such as age, gender, and BMI. An image was labeled as “abnormal” if its grayscale value z-score was greater than 2, and “normal” otherwise. This classification resulted in the following distribution per muscle group: biceps brachii (287 normal, 158 abnormal subjects), medial gastrocnemius (266 normal, 110 abnormal subjects), and tibialis anterior (296 normal, 166 abnormal subjects). In total, the dataset comprises images from 849 normal and 434 abnormal subjects (Table 1 and Supplementary Table S1).

Table 1
www.frontiersin.org

Table 1. Description of the public neuromuscular ultrasound dataset.

2.2 Data preprocessing

Before model training, we performed a systematic data preprocessing pipeline on all ultrasound images to ensure input data quality, structural consistency, and to promote model training stability and convergence. This pipeline included size unification, noise suppression, and image enhancement. First, we performed image standardization and cleaning. To meet the input requirements of the deep learning model, all images were uniformly resized to 224 × 224 pixels. Given that ultrasound images are inherently affected by speckle noise, we employed a median filter with a 5 × 5 kernel (14). This method was chosen for its effectiveness in suppressing salt-and-pepper-like noise while preserving critical muscle boundary details better than linear filters like Gaussian blur. Subsequently, to address the issue of variable contrast across different images, which can arise from different equipment settings or patient tissues, we applied Contrast Limited Adaptive Histogram Equalization (CLAHE) instead of global histogram equalization. CLAHE enhances local contrast without over-amplifying noise in relatively uniform regions, making it particularly suitable for medical image analysis.

Considering the high heterogeneity of clinical ultrasound images, we implemented a comprehensive data augmentation strategy to enhance the model’s robustness and mitigate overfitting (15). For the training set, we applied several transformations with specified probabilities: resized cropping (to 224 × 224), horizontal flipping (p = 0.5), and perturbations in brightness, contrast, and saturation (with a jitter factor of 0.2). On this basis, we further introduced the Mixup data augmentation strategy, with the mixing coefficient λ drawn from a Beta distribution (α = 0.2, β = 0.2). This method generates virtual training samples by linearly combining two images and their corresponding labels, which has been shown to improve model generalization and performance on imbalanced datasets (15).

For label processing, we encoded the original diagnostic labels (“normal” and “abnormal”) into a binary format: 0 for normal and 1 for pathological. This was based on the z-score grading results, and a mapping dictionary was constructed to ensure consistent annotation during training and validation. Additionally, to meet the input requirements of NMD-AssistNet, the images were converted to tensors and normalized at the channel level using PyTorch’s built-in Normalize function. The mean and standard deviation were set to the standard values used by the ImageNet pre-trained model ([0.485, 0.456, 0.406] and [0.229, 0.224, 0.225]), respectively, to facilitate the application of transfer learning strategies (16). To systematically validate the contribution of each component within this preprocessing pipeline, we designed an ablation study. This study involved incrementally adding each processing step to a baseline model to quantify its specific impact on performance.

2.3 Region of interest segmentation

Although the main task of this study focuses on the classification of ultrasound images of neuromuscular diseases, due to the fact that muscle tissue in ultrasound images often has features such as blurred boundaries, uneven grayscale distribution, and complex background structures, direct classification training on the entire image may cause the model’s attention to diverge and interfere with the accuracy of feature extraction. Therefore, before the image is input into the model, we introduced the muscle contour information based on expert annotation to extract the region of interest (ROI) to weaken irrelevant background interference and enhance the model’s ability to focus on key structures. Each image in the dataset is accompanied by a muscle cross-sectional area (CSA) contour mask drawn by experts. These masks were originally used for grayscale analysis and muscle volume estimation. In this study, we converted it into the cropped area required for the classification task, extracted the minimum rectangular area containing the muscle body by calling the contour boundary coordinates, and performed appropriate upper, lower, left, and right edge expansion on this basis to retain the surrounding important structural information. The processed ROI images were uniformly adjusted to the standard input size and replaced the original full images in the subsequent model training process. This strategy effectively reduces the risk of model learning background noise while retaining the core muscle tissue features, and improves the focus and discriminability of feature extraction. It is worth pointing out that this study did not train a pixel-level segmentation model separately, but made full use of the existing manually annotated segmentation mask as a guide to construct better classification input samples through ROI cropping. The reliability of these masks is critical. The ground-truth segmentation was performed by two experienced clinicians (a physiotherapist and a technical physician). They first annotated images independently, then resolved discrepancies via consensus discussion (12).

2.4 Diagnostic model construction

2.4.1 NMD-AssistNet

In order to achieve efficient and accurate identification of neuromuscular diseases, this paper designs a lightweight deep learning model tailored for ultrasound image diagnosis tasks, named NMD-AssistNet (Neuromuscular Disease Assistive Network). The model uses an improved GhostNet as the backbone architecture, combined with an attention mechanism and a lightweight convolution strategy, to significantly reduce the number of model parameters and computational overhead while ensuring diagnostic performance, thereby meeting the comprehensive requirements of medical image analysis for efficiency, generalization ability, and clinical deployability.

NMD-AssistNet consists of three key modules: the basic backbone adopts the GhostNet framework to obtain efficient sparse feature representation; the feature enhancement part introduces the channel and space dual attention mechanism—CBAM to enhance the model’s ability to pay attention to the texture and structure of the lesion area (17); at the same time, the model embeds depthwise separable convolution and channel shuffle operations in multiple convolutional layers to further compress parameters, improve operation speed, and enhance the fusion and flow of multi-scale features (12). The final output layer of the model uses global average pooling combined with a fully connected layer to map to a binary classification label space to achieve automatic judgment of “healthy” and “pathological” images. We introduced label smoothing technology at the loss function level to effectively alleviate the risk of overfitting during training. At the same time, in terms of training scheduling strategy, the SWALR method is used, combined with periodic learning rate adjustment and model weight averaging to improve the optimization convergence speed and obtain more stable model performance (18).

2.4.2 Model comparison

In order to systematically evaluate the performance advantages of NMD-AssistNet, this study selected five representative deep learning models for comparison, including CSPNet (12), EfficientNet (12), GhostNet (19), HRNet (20), and Vision Transformer (ViT) (21). These models cover different technical paths such as lightweight design, multi-scale modeling and global attention mechanism, representing the current mainstream modeling strategies for medical image classification. CSPNet and GhostNet emphasize high efficiency, but have deficiencies in feature expression and lesion perception; although EfficientNet has strong performance, its large network size limits clinical deployment; HRNet maintains high-resolution features and has ideal boundary preservation effects, but has high computational complexity; ViT has global modeling capabilities and is suitable for large-scale natural image classification, but is prone to overfitting in small and medium sample medical scenarios.

2.5 Experimental setup

To ensure a rigorous and clinically meaningful evaluation, the dataset was partitioned at the subject level to prevent data leakage, where all images from a single individual were exclusively allocated to either the training or the validation set. We employed a stratified random sampling strategy based on the diagnostic label (normal vs. abnormal) to maintain a consistent class distribution across the sets. From the total of 1,283 subjects, 898 subjects (70%) were allocated to the training set, and the remaining 385 subjects (30%) were assigned to the validation set. This subject-level split resulted in the following data distribution: (1) Training set: Comprised 898 subjects (595 normal, 303 abnormal), totaling approximately 2,742 images. (2) Validation set: Comprised 385 subjects, totaling 783 images. The model was trained on an Ubuntu 20.04 platform configured with an Intel i9-12900K CPU, 128GB RAM, and an NVIDIA RTX 3090 GPU (24GB). The development environment included PyTorch 1.13.1, CUDA 11.7, and cuDNN 8.4.1 (22). The training used the Adam optimizer (initial learning rate 0.001, batch size 32, weight decay 1e-5), and combined with the SWALR strategy to dynamically adjust the learning rate to improve convergence efficiency. The total number of training epoch was set to 100, and the early stopping mechanism was enabled to prevent overfitting. All comparison models (CSPNet, EfficientNet, GhostNet, HRNet, and ViT) were uniformly trained on the same data and environment to ensure experimental fairness.

2.6 Model evaluation

We evaluated model performance using several quantitative metrics, including accuracy, precision, recall, F1-score, and the area under the curve (AUC). Accuracy measures the proportion of correct classifications and is the basic indicator for evaluating overall performance. Precision (also known as positive predictive value, PPV) is used to evaluate the proportion of the model’s predictions of the “pathology” category that are actually pathological, reflecting its ability to control false positives. Recall measures the model’s ability to identify pathological samples and reflects the risk of missed diagnosis. The F1-score, as the harmonic mean of precision and recall, is suitable for scenarios with uneven sample distribution. Finally, the AUC value reflects the stability and discrimination ability of the model at different decision thresholds and is an important basis for evaluating the generalization performance of the model (23). In addition to these quantitative metrics, we qualitatively assessed the model’s interpretability using Gradient-weighted Class Activation Mapping (Grad-CAM), a technique that generates a visual heatmap to identify the image regions most influential for the model’s classification decision. In these visualizations, the heatmap is used for a qualitative assessment of the model’s focus. Warmer colors (e.g., red) indicate regions that were more influential in the model’s classification decision, while cooler colors (e.g., blue) represent areas of lesser importance. This approach helps to interpret where the model is looking, rather than quantifying the precise importance of each pixel.

3 Results

3.1 NMD-AssistNet performance

3.1.1 Overall performance

NMD-AssistNet achieved high performance across all evaluated metrics (Table 2). The model correctly classified 738 out of 783 total samples, with the full confusion matrix detailed in Table 2. Figure 2 shows the receiver operating characteristic (ROC) curve for NMD-AssistNet, which is positioned close to the upper-left corner with an AUC value of 0.9776.

Table 2
www.frontiersin.org

Table 2. Detailed performance metrics and confusion matrix.

Figure 2
ROC curve graph with the true positive rate on the y-axis and false positive rate on the x-axis. The curve, shown in blue, indicates strong model performance with an area of 0.9776. A diagonal red dashed line represents random chance.

Figure 2. ROC curve and corresponding AUC value of the NMD-AssistNet model on the validation set.

3.1.2 Ablation study on preprocessing methods

An ablation study was conducted to validate the effectiveness of each component in the data preprocessing pipeline. As detailed in Table 3, each added step incrementally improved model performance. The inclusion of standard data augmentation (Exp. D) yielded the most substantial increase in F1-score, while the complete pipeline with Mixup (Exp. E) achieved the highest scores across all metrics.

Table 3
www.frontiersin.org

Table 3. Ablation study on the contribution of each data preprocessing component to the model’s performance.

3.2 Comparison of model results

To evaluate its performance relative to other models, NMD-AssistNet was compared with five mainstream architectures under identical conditions. As shown in Figure 3, NMD-AssistNet achieved the highest values across all four metrics (accuracy, recall, precision, and F1-score) when compared to CSPNet, EfficientNet, GhostNet, HRNet, and ViT. Figure 4 illustrates that the parameter count of NMD-AssistNet (2.4 M) is substantially lower than that of the other models. The ROC analysis further shows that the AUC for NMD-AssistNet (0.979) was higher than all other tested models (Figure 5).

Figure 3
Bar chart comparing five models: CSPNet, EfficientNet, GhostNet, HRNet, ViT, and NMD-AssistNet. Each model's performance is evaluated by four metrics: Accuracy, Recall, F1-score, and Precision. NMD-AssistNet shows the highest scores across all metrics, notably achieving an accuracy of 0.950.

Figure 3. Performance comparison of classification metrics. Bar chart comparing accuracy, recall, precision, and F1-score across six models.

Figure 4
Bar chart showing the number of parameters in millions for different networks. CSPNet has 21.6 million, EfficientNet has 4 million, GhostNet has 3.9 million, HRNet has 19.3 million, ViT has 85.8 million, and NMD-AssistNet has 2.4 million.

Figure 4. Model size comparison by trainable parameters, showing the number of trainable parameters for each model. NMD-AssistNet has the smallest model size (2.4 M) while maintaining top performance, demonstrating strong potential for edge deployment.

Figure 5
Five Receiver Operating Characteristic (ROC) curves labeled A to E, each showing a blue curve representing the model's performance against a diagonal red dashed line indicating random guessing. The Area Under the Curve (AUC) scores are: A=0.8904, B=0.9593, C=0.9724, D=0.9676, and E=0.9495. Each graph plots True Positive Rate on the Y-axis against False Positive Rate on the X-axis.

Figure 5. ROC curve and AUC comparison among models on the validation set. NMD-AssistNet achieved the highest AUC (0.9776), outperforming CSPNet (A, 0.8904), EfficientNet (B, 0.9593), GhostNet (C, 0.9724), HRNet (D, 0.9676), and ViT (E, 0.9495).

3.2.1 Grad-CAM display and analysis of the model

Grad-CAM was used to visualize the model’s focus areas on typical healthy and pathological muscle images (Figure 6). For healthy samples (Figure 6A), the high-response areas in the heat map concentrated on the transverse stripe-like texture corresponding to neatly arranged muscle fibers. For the pathological sample (Figure 6B), the high-response areas focused on regions with disordered echoes, blurred edges, or focal hypoechoic areas. These visualized regions correspond to locations of known pathological features.

Figure 6
Two rows labeled A and B show ultrasound images. Each row has grayscale and color-coded elastography images. The left images display tissue structure, while the right images show elasticity variations with colors representing different stiffness levels, indicated by blue to red gradients.

Figure 6. Grad-CAM activation heatmap visualization, where (A) represents healthy samples and (B) represents abnormal samples.

4 Discussion

This study developed NMD-AssistNet, an efficient and lightweight deep learning model designed to meet the clinical need for auxiliary diagnosis of neuromuscular diseases by automatically classifying muscle tissue in ultrasound images. The model demonstrated strong performance, achieving an accuracy of 0.9502 and an AUC of 0.9776. This success is attributed to its hybrid architecture. By integrating the lightweight GhostNet framework with a CBAM attention mechanism, depthwise separable convolution, and channel shuffle strategies, NMD-AssistNet effectively focuses on key pathological features while maintaining a remarkably low parameter count of 2.4 M, which is significantly lower than larger models like ViT and HRNet. Furthermore, advanced training strategies, including Mixup augmentation, label smoothing, and SWALR, were employed to enhance the model’s generalization and stability, proving the effectiveness of our design.

A critical analysis of the Grad-CAM visualizations, however, warrants a more nuanced discussion. As observed, the heatmaps occasionally highlight the boundary between the ROI and the masked background. This phenomenon can be attributed to the nature of Grad-CAM interacting with the pre-segmented dataset used in this study (24). The artificial, sharp edge resulting from the expert-provided segmentation masks creates a high-gradient region, which can become a prominent focus for the visualization technique (25). While this may suggest the model is learning from edge artifacts, it is crucial to interpret this alongside the model’s high accuracy. The strong performance indicates that the model successfully learns discriminative features from within the muscle tissue itself, as these are essential for distinguishing between healthy and pathological patterns. The boundary highlighting is more likely a byproduct of the visualization technique interacting with pre-processed data, rather than the sole basis for the model’s decision. This observation underscores a key challenge in applying post-hoc interpretability methods to pre-segmented images and reinforces the importance of our future research direction (26). Developing an end-to-end model that operates on unsegmented, raw ultrasound images would not only streamline the clinical workflow but also yield more authentic interpretability results, ensuring the model’s focus is entirely on natural anatomical and pathological features.

It is crucial to contextualize the role of NMD-AssistNet within the broader clinical diagnostic pathway. Neuromuscular diseases are characterized by hidden onset, diverse symptoms, and high misdiagnosis rates (27). Our model’s function is to provide a rapid, objective classification of a single muscle’s ultrasound image as ‘pathological’ or ‘healthy’; it does not, by itself, identify the specific type of NMD. The definitive diagnosis of NMDs relies on a comprehensive evaluation, where identifying the specific pattern of affected and spared muscle groups is paramount (28). For instance, patterns of selective muscle involvement are critical clues that help clinicians narrow the differential diagnosis among various muscular dystrophies or myopathies before proceeding to genetic testing or muscle biopsy.

Therefore, the primary clinical impact of NMD-AssistNet is its potential as a powerful adjunctive screening tool, rather than a standalone diagnostic solution. First, it can provide an objective, quantitative assessment of muscle echogenicity, reducing the inter-observer variability common in ultrasound interpretation and helping less experienced clinicians make more reliable initial judgments. Second, in a clinical setting, a rapid “abnormal” finding from NMD-AssistNet can help prioritize patients for further, more invasive and costly investigations like electromyography (EMG) or genetic panels. Finally, due to its non-invasive nature, the tool could be used longitudinally to objectively monitor disease progression or treatment response. The model’s low computational overhead makes it highly suitable for deployment on portable ultrasound equipment or edge computing devices, bringing intelligent diagnostic support to primary care facilities and community screening scenarios.

Although this study has achieved good results in model design and performance, there are still several limitations that cannot be ignored in the process of clinical transformation and practical application. Firstly, a primary limitation is the model’s reliance on pre-segmented ROIs. This approach hinders clinical efficiency. Furthermore, it can create visual artifacts at the ROI boundaries in Grad-CAM maps, complicating interpretability. Therefore, future work will focus on developing an end-to-end model to analyze unsegmented images, which would simultaneously streamline the diagnostic process and provide more authentic visual explanations. Secondly, the use of single-center data is a key limitation, necessitating external validation to ensure the model can generalize across different equipment and patient populations. To this end, validating our model on public datasets is our crucial next step to prove its potential as a clinically viable tool. Thirdly, the current model only performs a binary classification (healthy vs. pathological) and has not yet been refined to differentiate specific subtypes of neuromuscular diseases, such as muscular dystrophy or myasthenia gravis. This limits the model’s immediate applicability for refined diagnosis and individualized treatment planning. Finally, our dataset was partitioned into only training and test sets, without a separate, dedicated validation set for hyperparameter tuning. While we mitigated this by performing model selection using cross-validation within the training data, future work could benefit from a three-way split (training, validation, and test). This would provide an even more robust framework for model development and evaluation, further ensuring the generalizability of the results.

5 Conclusion

This study introduces NMD-AssistNet, a lightweight deep learning model for the automated classification of neuromuscular diseases from ultrasound images. The primary significance of our work lies in achieving high diagnostic accuracy while maintaining exceptional computational efficiency. This balance is critical for clinical translation, as it enables the model’s potential deployment on portable ultrasound systems or edge devices, thereby facilitating point-of-care screening. Furthermore, our interpretability analysis using Grad-CAM confirms that the model’s decisions are based on clinically relevant tissue features, a crucial step for building trust and acceptance among clinicians. While the results are promising, we acknowledge that future work must validate the model’s generalizability on larger, multi-center datasets. The next logical step is to evolve this framework from a binary classifier into a multi-class system capable of differentiating specific disease types. In summary, NMD-AssistNet provides a viable technical approach for developing intelligent, efficient, and trustworthy AI-powered tools to augment the clinical workflow for neuromuscular disease diagnosis.

Data availability statement

The original contributions presented in the study are included in the article/Supplementary material, further inquiries can be directed to the corresponding author.

Author contributions

JX: Conceptualization, Data curation, Formal analysis, Funding acquisition, Investigation, Methodology, Project administration, Resources, Software, Supervision, Validation, Visualization, Writing – original draft, Writing – review & editing. ZZ: Conceptualization, Data curation, Investigation, Methodology, Software, Supervision, Writing – original draft, Writing – review & editing.

Funding

The author(s) declare that no financial support was received for the research and/or publication of this article.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The authors declare that no Gen AI was used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fneur.2025.1640428/full#supplementary-material

References

1. Mary, P, Servais, L, and Vialle, R. Neuromuscular diseases: diagnosis and management. Orthop Traumatol Surg Res. (2018) 104:S89–95. doi: 10.1016/j.otsr.2017.04.019

Crossref Full Text | Google Scholar

2. McDonald, CM. Clinical approach to the diagnostic evaluation of hereditary and acqivuired neuromuscular diseases. Phys Med Rehabil Clin N Am. (2012) 23:495–563. doi: 10.1016/j.pmr.2012.06.011

PubMed Abstract | Crossref Full Text | Google Scholar

3. Padua, L, Liotta, G, Di Pasquale, A, Granata, G, Pazzaglia, C, Caliandro, P, et al. Contribution of ultrasound in the assessment of nerve diseases. Eur J Neurol. (2012) 19:47–54. doi: 10.1111/j.1468-1331.2011.03421.x

PubMed Abstract | Crossref Full Text | Google Scholar

4. Pillen, S, and van Alfen, N. Skeletal muscle ultrasound. Neurol Res. (2011) 33:1016–24. doi: 10.1179/1743132811Y.0000000010

PubMed Abstract | Crossref Full Text | Google Scholar

5. Mayans, D, Cartwright, MS, and Walker, FO. Neuromuscular ultrasonography: quantifying muscle and nerve measurement. Phys Med Rehabil Clin N Am. (2012) 23:133–48. doi: 10.1016/j.pmr.2011.11.009

PubMed Abstract | Crossref Full Text | Google Scholar

6. Taha, MA, and Morren, JA. The role of artificial intelligence in electrodiagnostic and neuromuscular medicine: current state and future directions. Muscle Nerve. (2024) 69:260–72. doi: 10.1002/mus.28023

PubMed Abstract | Crossref Full Text | Google Scholar

7. LeCun, Y, Bengio, Y, and Hinton, G. Deep learning. Nature. (2015) 521:436–44. doi: 10.1038/nature14539

PubMed Abstract | Crossref Full Text | Google Scholar

8. Shin, HC, Roth, HR, Gao, M, Lu, L, Xu, Z, Nogues, I, et al. Deep convolutional neural networks for computer-aided detection: CNN architectures, dataset characteristics and transfer learning. IEEE Trans Med Imaging. (2016) 35:1285–98. doi: 10.1109/TMI.2016.2528162

PubMed Abstract | Crossref Full Text | Google Scholar

9. Zhang, Y, and Yang, Q. A survey on multi-task learning. IEEE Trans Knowl Data Eng. (2022) 34:5586–609. doi: 10.1109/TKDE.2021.3070203

Crossref Full Text | Google Scholar

10. de Filippis, RD, and Foysal, AA. Harnessing the power of artificial intelligence in neuromuscular disease rehabilitation: a comprehensive review and algorithmic approach. Adv Biosci Biotechnol. (2024) 15:289–309. doi: 10.4236/abb.2024.155018

Crossref Full Text | Google Scholar

11. Siddique, N, Paheding, S, Elkin, CP, and Devabhaktuni, V. U-net and its variants for medical image segmentation: a review of theory and applications. IEEE Access. (2021) 9:82031–57. doi: 10.1109/ACCESS.2021.3086020

Crossref Full Text | Google Scholar

12. Khan, A, Sohail, A, Zahoora, U, and Qureshi, AS. A survey of the recent architectures of deep convolutional neural networks. Artif Intell Rev. (2020) 53:5455–516. doi: 10.1007/s10462-020-09825-6

Crossref Full Text | Google Scholar

13. Yuan, G, Chen, Z, Zhou, Y, and Du, L. Quantifying skeletal muscle structure parameters in ultrasound images using deep learning without linear assumptions. Biomed Signal Process Control. (2024) 91:106050. doi: 10.1016/j.bspc.2024.106050

Crossref Full Text | Google Scholar

14. Liu, J, Li, C, Liu, L, Chen, H, Han, H, Zhang, B, et al. Speckle noise reduction for medical ultrasound images based on cycle-consistent generative adversarial network. Biomed Signal Process Control. (2023) 86:105150. doi: 10.1016/j.bspc.2023.105150

Crossref Full Text | Google Scholar

15. Shorten, C, and Khoshgoftaar, TM. A survey on image data augmentation for deep learning. J Big Data. (2019) 6:60. doi: 10.1186/s40537-019-0197-0

Crossref Full Text | Google Scholar

16. He, K, Zhang, X, Ren, S, and Sun, J. Deep residual learning for image recognition. In: Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2016), Las Vegas, NV. 770–8.

Google Scholar

17. Niu, Z, Zhong, G, and Yu, H. A review on the attention mechanism of deep learning. Neurocomputing. (2021) 452:48–62. doi: 10.1016/j.neucom.2021.03.091

Crossref Full Text | Google Scholar

18. Szegedy, C, Vanhoucke, V, Ioffe, S, Shlens, J, and Wojna, Z. Rethinking the inception architecture for computer vision. In: Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV (2016). 2818–2826.

Google Scholar

19. Han, K, Wang, Y, Tian, Q, Guo, J, Xu, C, and Xu, C. GhostNet: more features from cheap operations. In: Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA (2020), 1577–1586.

Google Scholar

20. Sun, K, Xiao, B, Liu, D, and Wang, J Deep high-resolution representation learning for human pose estimation. In: Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (June 15–20, 2019), Long Beach, CA (2019), 5686–5696.

Google Scholar

21. Khan, S, Naseer, M, Hayat, M, Zamir, SW, Khan, FS, and Shah, M. Transformers in vision: a survey. ACM Comput Surv. (2022) 54:200. doi: 10.1145/3505244

Crossref Full Text | Google Scholar

22. Paszke, A, Gross, S, Massa, F, Lerer, A, Bradbury, J, Chanan, G, et al. Pytorch: an imperative style, high-performance deep learning library. arXiv [Preprint] (2019) (Accessed August 7, 2025). Available online at: http://arxiv.org/abs/1912.01703

Google Scholar

23. Fawcett, T. An introduction to ROC analysis. Pattern Recogn Lett. (2006) 27:861–74. doi: 10.1016/j.patrec.2005.10.010

Crossref Full Text | Google Scholar

24. Hao, J. Deep learning-based medical image analysis with explainable transfer learning. In: Proceedings of the 2023 International Conference on Computer Engineering and Distance Learning (CEDL), June 29–July 1, 2023, Kunming, China.

Google Scholar

25. Adebayo, J, Gilmer, J, Muelly, M, Goodfellow, I, Hardt, M, and Kim, B. Sanity checks for saliency maps. In: Proceedings of the 32nd International Conference on Neural Information Processing Systems; December 3–8, 2018, Montréal, Canada. 9525–36.

Google Scholar

26. Ghassemi, M, Oakden-Rayner, L, and Beam, AL. The false hope of current approaches to explainable artificial intelligence in health care. Lancet Digit Health. (2021) 3:e745–50. doi: 10.1016/S2589-7500(21)00208-9

PubMed Abstract | Crossref Full Text | Google Scholar

27. Rafii, MS. Neurological history and examination In: L Ganti, M Gentile, and P Tadi, editors. First aid for the neurology clerkship. 3rd ed. New York, NY: McGraw Hill (2023)

Google Scholar

28. Mercuri, E, Bönnemann, CG, and Muntoni, F. Muscular dystrophies. Lancet. (2019) 394:2025–38. doi: 10.1016/S0140-6736(19)32910-1

PubMed Abstract | Crossref Full Text | Google Scholar

Keywords: neuromuscular disease, ultrasound image classification, deep learning, lightweight neural network, model interpretability

Citation: Xie J and Zhang Z (2025) Development of a deep learning model for automated diagnosis of neuromuscular diseases using ultrasound imaging. Front. Neurol. 16:1640428. doi: 10.3389/fneur.2025.1640428

Received: 03 June 2025; Accepted: 28 August 2025;
Published: 10 September 2025.

Edited by:

Edoardo Malfatti, Hôpitaux Universitaires Henri Mondor, France

Reviewed by:

Kristen M. Meiburger, Polytechnic University of Turin, Italy
Helga Silva, Federal University of São Paulo, Brazil

Copyright © 2025 Xie and Zhang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Zhenying Zhang, enp5emFmdTU1MzZAMTYzLmNvbQ==

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.