- 1Department of Cardiology, Shidong Hospital affiliated to University of Shanghai for Science and Technology, Shanghai, China
- 2Department of Geriatrics, Renhe Hospital, Shanghai, China
- 3Department of Geriatrics, Shidong Hospital affiliated to University of Shanghai for Science and Technology, Shanghai, China
- 4Department of Medical Examination, Shidong Hospital affiliated to University of Shanghai for Science and Technology, Shanghai, China
Background: The application of deep learning techniques in medical image analysis has shown great potential in assisting clinical diagnosis. This study focuses on the development and evaluation of deep learning models for the classification of knee joint injuries using Magnetic Resonance Imaging (MRI) data. The research aims to provide an efficient and reliable tool for clinicians to aid in the diagnosis of knee joint disorders, particularly focusing on Anterior Cruciate Ligament (ACL) tears.
Methods: KneeXNet leverages the power of graph convolutional networks (GCNs) to capture the intricate spatial dependencies and hierarchical features in knee MRI scans. The proposed model consists of three main components: a graph construction module, graph convolutional layers, and a multi-scale feature fusion module. Additionally, a contrastive learning scheme is employed to enhance the model’s discriminative power and robustness. The MRNet dataset, consisting of knee MRI scans from 1,370 patients, is used to train and validate KneeXNet.
Results: The performance of KneeXNet is evaluated using the Area Under the Receiver Operating Characteristic Curve (AUC) metric and compared to state-of-the-art methods, including traditional machine learning approaches and deep learning models. KneeXNet consistently outperforms the competing methods, achieving AUC scores of 0.985, 0.972, and 0.968 for the detection of knee joint abnormalities, ACL tears, and meniscal tears, respectively. The cross-dataset evaluation further validates the generalization ability of KneeXNet, maintaining its superior performance on an independent dataset.
Application: To facilitate the clinical application of KneeXNet, a user-friendly web interface is developed using the Django framework. This interface allows users to upload MRI scans, view diagnostic results, and interact with the system seamlessly. The integration of Grad-CAM visualizations enhances the interpretability of KneeXNet, enabling radiologists to understand and validate the model’s decision-making process.
1 Introduction
The rapid advancement of artificial intelligence (AI) and deep learning techniques has revolutionized various fields, including healthcare and medical imaging. In recent years, there has been a growing interest in applying these cutting-edge technologies to assist physicians in diagnosing complex medical conditions, particularly those related to musculoskeletal disorders. Knee joint injuries, such as anterior cruciate ligament (ACL) tears and meniscal tears, are among the most common and debilitating conditions affecting individuals of all ages (Chan et al., 2022). Magnetic resonance imaging (MRI) has emerged as the gold standard for diagnosing these injuries due to its superior soft tissue contrast and ability to visualize detailed anatomical structures. However, the interpretation of knee MRI scans remains a challenging task, even for experienced radiologists, as it requires a thorough understanding of the complex anatomy and pathophysiology of the knee joint (D’Angelo et al., 2022).
Moreover, the increasing demand for MRI examinations and the shortage of trained radiologists have led to a substantial workload and potential delays in diagnosis (Fernandes et al., 2022). This highlights the need for an automated system that can efficiently analyze knee MRI scans and assist radiologists in making accurate and timely diagnoses. Such a system would not only improve patient care and outcomes but also optimize resource allocation and reduce healthcare costs (Vera Cruz et al., 2022).
In recent years, deep learning algorithms, particularly convolutional neural networks (CNNs), have shown remarkable performance in various computer vision tasks, including medical image analysis. CNNs have the ability to automatically learn hierarchical features from raw input data, making them well-suited for analyzing complex medical images. Several studies have explored the application of CNNs in knee MRI analysis (Kulseng et al., 2023), focusing on tasks such as segmentation of knee joint structures, detection of ACL tears, and classification of meniscal tears. While these studies have demonstrated the potential of deep learning in knee MRI analysis (Hung et al., 2023), there remain significant challenges and opportunities for further research.
One of the main limitations of existing deep learning approaches is their reliance on relatively simple CNN architectures, such as AlexNet (Sivakumari and Vani, 2022) and ResNet (Wang et al., 2022), which may not capture the full complexity of knee MRI data. To address this issue, more advanced and sophisticated models have been proposed, such as attention mechanisms (Qiu et al., 2024), generative adversarial networks (GANs) (Yaqub et al., 2022), and transformer-based architectures. These models have shown promising results in various medical imaging tasks, including brain tumor segmentation, lung nodule detection, and breast cancer classification. However, their application to knee MRI analysis remains largely unexplored.
Another critical challenge in developing deep learning models for medical image analysis is the interpretability and explainability of the model’s predictions. In clinical settings, it is crucial for physicians to understand the reasoning behind the model’s decisions to build trust and facilitate informed decision-making (Belton et al., 2021). Various techniques have been proposed to enhance the interpretability of deep learning models, such as attention maps (Wang et al., 2025a), class activation maps (CAMs), and saliency maps (Chang et al., 2020). However, the integration of these techniques into a comprehensive knee MRI analysis system remains a significant research gap.
Furthermore, the successful implementation of an AI-based diagnostic tool in clinical practice requires a user-friendly interface and seamless integration with existing workflows. To address this challenge, web-based frameworks, such as Django (Mihcin et al., 2023), have been utilized to develop interactive and intuitive applications for medical image analysis. However, the development of a comprehensive web-based system for knee MRI analysis (Tuazon et al., 2023), incorporating advanced deep learning models and interpretability techniques, has not been thoroughly investigated.
In this study, we aim to bridge these research gaps by developing a novel deep learning-based system for knee MRI analysis and diagnosis. We propose a novel deep learning framework that goes beyond the traditional CNN architectures and attention mechanisms. Our model, named KneeXNet, is designed to capture the intricate spatial dependencies and hierarchical features (Wang et al., 2025b) in knee MRI scans while maintaining a high degree of computational efficiency in Figure 1.

Figure 1. The overall architecture of KneeXNet, featuring a graph construction module, graph convolutional layers, and a multi-scale feature fusion module for capturing spatial dependencies and hierarchical features in knee MRI scans.
The core of KneeXNet lies in its innovative use of graph convolutional networks (GCNs) (Hu et al., 2025) to model the complex relationships between different anatomical structures within the knee joint. By representing the knee MRI as a graph, where nodes correspond to key anatomical landmarks and edges represent their spatial connections, KneeXNet can effectively propagate and integrate information across the entire joint. This graph-based approach allows the model to consider not only the local features of individual structures but also their global context and interactions (Zhuang et al., 2022), leading to a more comprehensive understanding of the knee joint pathology.
To further enhance the representational power of KneeXNet, we introduce a multi-scale feature fusion module that combines features from different resolutions and receptive fields (Zou et al., 2023). This module employs a series of 3D convolutional layers with varying kernel sizes and dilation rates to capture both fine-grained details and broader contextual information. The multi-scale features are then adaptively aggregated using learnable weighting factors, enabling the model to dynamically adjust the importance of different scales based on the specific characteristics of each MRI scan.
Moreover, we propose a novel contrastive learning scheme to encourage KneeXNet to learn more discriminative and robust representations. During training, we generate positive and negative pairs of MRI patches by applying various data augmentation techniques, such as random rotations, translations, and elastic deformations (Recht et al., 2020). The model is then trained to minimize the contrastive loss, which maximizes the similarity between positive pairs while pushing negative pairs apart in the feature space (Johnson et al., 2023). This self-supervised learning approach helps KneeXNet to capture the essential patterns and variations in knee MRI data, improving its generalization ability and reducing the risk of overfitting.
To ensure the interpretability and clinical adoption of KneeXNet, we integrate gradient-weighted class activation mapping (Grad-CAM) (Wang et al., 2024) to highlight the regions of the knee MRI that contribute most to the model’s predictions. We believe that this innovative framework has the potential to improve the diagnosis of knee joint diseases and provide valuable assistance to radiologists in their clinical decision-making process.
2 Related work and preliminaries
The application of deep learning techniques in medical image analysis has gained significant attention in recent years. Convolutional neural networks (CNNs) have been widely used for various tasks, such as segmentation, detection, and classification. In the context of knee MRI analysis, several studies have explored the use of CNNs for the detection and classification of knee joint abnormalities, including ACL tears and meniscal tears. Bezabh et al. (2024) proposed a semi-automated method for ACL injury detection using a combination of CNN and Support Vector Machine (SVM). The authors utilized a pre-trained AlexNet (Alom et al., 2018) model for feature extraction and achieved an improved AUC on a dataset of MRI scans. The AlexNet architecture, which consists of five convolutional layers and three fully connected layers, can be represented as
To capture the complex spatial dependencies in knee MRI data, Namiri et al. (2020) proposed a 3D CNN model with attention mechanisms. The authors introduced a novel attention module that adaptively weights the features at different scales and locations, enabling the model to focus on the most informative regions. The attention mechanism can be described as
Other approaches for knee MRI analysis include the use of transfer learning (Yang et al., 2022), where pre-trained models from natural image datasets, such as ImageNet (Haddadian and Balamurali, 2022) in Figure 2, are fine-tuned on the target medical image dataset. This approach leverages the learned features from a large-scale dataset and adapts them to the specific task of knee MRI analysis, often leading to improved performance and faster convergence. Unsupervised learning techniques, such as autoencoders (Nasser et al., 2020) and generative adversarial networks (GANs) (Yang et al., 2022; Yin et al., 2024), have also been explored for knee MRI analysis. Autoencoders aim to learn a compact representation of the input data by minimizing the reconstruction error between the input and the output. The encoder-decoder architecture can be defined as

Figure 2. The evolution of ImageNet classification models over time, demonstrating the rapid advancements in deep learning architectures and their increasing performance on challenging computer vision tasks.
Graph convolutional networks (GCNs) (Wang et al., 2021) have recently emerged as a powerful tool for analyzing structured data, such as social networks, molecules, and biological networks. GCNs extend the concept of convolution to graph-structured data by learning a function
In the context of knee MRI analysis, the use of GCNs remains largely unexplored. However, the ability of GCNs to model the complex spatial dependencies and capture the hierarchical structure of the knee joint makes them a promising approach for this task. In this study, we propose KneeXNet, a novel GCN-based framework for the classification of knee joint abnormalities. KneeXNet constructs a graph representation of the knee MRI, where nodes correspond to key anatomical landmarks and edges represent their spatial relationships. By leveraging the expressive power of GCNs and integrating multi-scale feature fusion and contrastive learning, KneeXNet achieves state-of-the-art performance in detecting ACL tears, meniscal tears, and other knee joint abnormalities.
3 Methods
3.1 Dataset
3.1.1 Data source and characteristics
The MRNet dataset, compiled by the Stanford Machine Learning Group, was utilized for this study. This publicly available dataset consists of knee MRI scans from 1,370 patients (mean age: 38.6
The MRI examinations were performed using 3.0 T MRI scanners (Siemens Magnetom Skyra and GE Healthcare Discovery MR750) with dedicated knee coils. The imaging protocols included proton density-weighted sequences with and without fat suppression, T1-weighted sequences, and T2-weighted sequences (Kara and Hardalaç, 2021). The slice thickness ranged from 2.5 to 3.0 mm, with an in-plane resolution of

Table 1. MRI acquisition parameters for the knee joint dataset, including proton density-weighted (PD-weighted), T1-weighted, and T2-weighted sequences.
The preprocessing steps applied to the MRNet dataset included intensity normalization and z-score normalization. First, the pixel intensities of each MRI slice were normalized to the range
3.1.2 Data annotation and ground truth
The ground truth labels for the dataset were established through a rigorous annotation process involving three board-certified musculoskeletal radiologists, each with more than 8 years of experience in knee MRI interpretation. Each radiologist independently reviewed all MRI scans and classified them for the presence of:
1. General abnormalities (any pathological finding in the knee joint)
2. Anterior cruciate ligament (ACL) tears (complete or partial)
3. Meniscal tears (medial, lateral, or both)
Discrepancies in the annotations were resolved through consensus discussions among the radiologists. In cases where consensus could not be reached, the final decision was made by a senior radiologist with over 15 years of experience in musculoskeletal imaging. The inter-rater reliability among the three radiologists was assessed using the Fleiss’ kappa coefficient
The dataset was divided into training, validation, and test sets using a stratified random sampling approach to maintain a similar distribution of pathologies across the sets. The training set comprised 1,130 cases (82.5%), the validation set included 120 cases (8.8%), and the test set consisted of 120 cases (8.8%). The distribution of cases across the three sets is presented in Table 2.

Table 2. Distribution of cases and pathologies across the training and validation of the knee MRI dataset.
3.2 Data Preprocessing
3.2.1 Image normalization
To ensure the homogeneity of the input data and facilitate the convergence of the deep learning models, all MRI images underwent a series of preprocessing steps. First, the pixel intensities of each MRI slice were normalized to the range [0, 1] using min-max normalization, which was calculated as
Following this, z-score normalization was applied to standardize the pixel intensity distribution of each slice, which was computed as
3.2.2 Volume standardization
Due to the variable number of slices in each MRI sequence, a volume standardization procedure was implemented to ensure consistent input dimensions for the deep learning models. For sequences with fewer than 25 slices (the target number), zero-padding was applied to both ends of the sequence to reach the target. For sequences with more than 25 slices, a slice selection algorithm was employed to extract the 25 most informative slices.
The slice selection algorithm utilized an entropy-based approach to evaluate the information content of each slice. The entropy of a slice was calculated as
The choice of using 25 slices for volume standardization was based on the average number of slices per MRI sequence in the MRNet dataset
3.2.3 Data augmentation
To enhance the robustness of the models and mitigate the risk of overfitting, a comprehensive data augmentation strategy was implemented. The augmentation techniques were applied on-the-fly during the training process, with each training sample having a 50% probability of undergoing augmentation. The augmentation techniques included:
1. Random rotations: Images were randomly rotated within the range of [−10°, 10°] using bilinear interpolation to fill in the resulting gaps. The rotation angle
2. Random translations: Images were randomly translated horizontally and vertically within the range of [−5%, 5%] of the image dimensions. The translation factors
3. Random scaling: Images were randomly scaled within the range of [0.95, 1.05] to simulate variations in field of view. The scaling factors
4. Random flipping: Images were randomly flipped horizontally with a probability of 0.5 to augment the dataset with mirror images.
5. Random brightness and contrast adjustments: The brightness and contrast of images were randomly adjusted within the ranges of [−0.1, 0.1] and [0.9, 1.1], respectively. The adjustment factors
6. Random noise addition: Gaussian noise with zero mean and a standard deviation randomly sampled from the range [0.01, 0.03] was added to the images. The noise standard deviation
7. Elastic deformations: Elastic deformations were applied to simulate variations in knee positioning and tissue elasticity. The deformation was controlled by two parameters:
The implementation of elastic deformations followed the method proposed by Chao et al. (2010), where a random displacement field
3.3 Model architectures
3.3.1 Overview of KneeXNet
KneeXNet is a novel deep learning framework designed specifically for the classification of knee joint injuries using MRI data. The architecture of KneeXNet is built upon the foundation of graph convolutional networks (GCNs), which have demonstrated remarkable performance in capturing the intricate spatial dependencies and hierarchical features in structured data. By representing the knee MRI as a graph, where nodes correspond to key anatomical landmarks and edges represent their spatial connections, KneeXNet can effectively propagate and integrate information across the entire joint, leading to a more comprehensive understanding of the knee joint pathology.
The overall architecture of KneeXNet is illustrated in Figure 1. The model consists of three main components: (1) the graph construction module, which converts the input knee MRI into a graph representation; (2) the graph convolutional layers, which learn the hierarchical features and spatial dependencies within the graph; and (3) the multi-scale feature fusion module, which combines features from different resolutions and receptive fields to enhance the representational power of the model. The output of KneeXNet is a probability distribution over the three target classes: normal, ACL tear, and meniscal tear.
3.3.2 Graph construction
The graph construction module aims to convert the input knee MRI into a graph representation that can be efficiently processed by the subsequent graph convolutional layers. Given a knee MRI scan
To capture the spatial relationships between the landmarks, we construct an undirected graph
To incorporate the features of each node, we extract a local patch centered around each landmark from the input knee MRI. The patch size is set to
To identify the anatomical landmarks, a pre-trained deep learning model was employed. This model, trained on a large dataset of manually annotated knee MRI scans, achieved a high detection accuracy of 95% on a separate test set. The adjacency matrix threshold
3.3.3 Graph convolutional layers
The core of KneeXNet lies in its graph convolutional layers, which learn the hierarchical features and spatial dependencies within the graph. The graph convolutional operation is defined as follows:
The graph convolutional operation can be interpreted as a message-passing scheme, where each node aggregates the features of its neighboring nodes weighted by the normalized adjacency matrix. This allows the model to propagate information along the edges of the graph and capture the spatial dependencies between the nodes.
KneeXNet employs a stack of
3.3.4 Multi-scale feature fusion
To further enhance the representational power of KneeXNet, we introduce a multi-scale feature fusion module that combines features from different resolutions and receptive fields. The motivation behind this module is to capture both fine-grained details and broader contextual information, enabling the model to better understand the complex patterns in knee MRI data.
The multi-scale feature fusion module consists of a series of 3D convolutional layers with varying kernel sizes and dilation rates. Given the output of the final graph convolutional layer
The multi-scale feature maps are then adaptively aggregated using learnable weighting factors
The choice of the number of convolutional layers, kernel sizes, and dilation rates in the multi-scale feature fusion module was based on a combination of domain knowledge and empirical evaluation. We conducted extensive experiments on the validation set, assessing the model’s performance for different configurations of these parameters. The final configuration (3 convolutional layers with kernel sizes of
The adaptive weighting factors allow the model to dynamically adjust the importance of different scales based on the specific characteristics of each knee MRI scan. This enables KneeXNet to effectively capture the most informative features across different resolutions and receptive fields.
3.3.5 Contrastive learning
To further improve the discriminative power and robustness of KneeXNet, we propose a novel contrastive learning scheme that encourages the model to learn more distinguishable representations. Contrastive learning is a self-supervised learning approach that aims to maximize the similarity between positive pairs of samples while minimizing the similarity between negative pairs.
During training, we generate positive and negative pairs of MRI patches by applying various data augmentation techniques, such as random rotations, translations, and elastic deformations. The positive pairs are obtained by applying the same augmentation to the same MRI patch, while the negative pairs are obtained by applying different augmentations to different patches.
Let
The choice of contrastive learning was motivated by its ability to enhance the model’s discriminative power and robustness. To validate its effectiveness, an ablation study was conducted, comparing the performance of KneeXNet with and without the contrastive learning component. The results, presented in Table 4, demonstrate that removing contrastive learning leads to a decrease in AUC scores across all three tasks (abnormality:
By minimizing the contrastive loss, KneeXNet learns to pull the positive pairs closer in the feature space while pushing the negative pairs apart. This self-supervised learning approach helps the model to capture the essential patterns and variations in knee MRI data, improving its generalization ability and reducing the risk of overfitting.
3.3.6 Interpretability
Interpretability is a crucial aspect of medical image analysis models, as it enables radiologists to understand and trust the model’s predictions. To enhance the interpretability of KneeXNet, we integrate the Grad-CAM technique, which highlights the regions of the input MRI that contribute most to the model’s decision.
Given a trained KneeXNet model and an input knee MRI scan
The Grad-CAM heatmap
The Grad-CAM heatmap is then upsampled to the size of the input MRI and overlaid on the original image to highlight the most informative regions. This visual explanation provides valuable insights into the model’s decision-making process and helps radiologists to interpret and validate the model’s predictions.
3.3.7 Loss function
The overall loss function of KneeXNet consists of two components: the classification loss and the contrastive loss:
The classification loss is defined as the categorical cross-entropy between the predicted class probabilities
The contrastive loss encourages the model to learn more discriminative representations by maximizing the similarity between positive pairs of MRI patches while minimizing the similarity between negative pairs. By jointly optimizing the classification loss and the contrastive loss, KneeXNet learns to accurately classify knee joint injuries while maintaining a high level of generalization ability and robustness.
3.3.8 Training and optimization
KneeXNet is trained using the Adam optimizer with a learning rate of 0.001 and a batch size of 32. The model is trained for 100 epochs, with early stopping based on the validation loss to prevent overfitting. The hyperparameters, such as the number of graph convolutional layers, the number of feature channels, and the contrastive loss weight
During training, we apply various data augmentation techniques, as described in the Data Preprocessing section, to increase the diversity of the training samples and improve the model’s robustness. The augmented samples are generated on-the-fly using the Albumentations library, which provides a wide range of image augmentation techniques specifically designed for medical images.
4 Results
4.1 Model evaluation
The performance of KneeXNet is evaluated using the AUC metric on the test set. The AUC is a threshold-independent measure of the model’s ability to discriminate between different classes, with a higher value indicating better performance. We also report the model’s accuracy, precision, recall, and F1 score to provide a comprehensive assessment of its classification performance.
To validate the observed performance improvements, we calculated 95% confidence intervals for each evaluation metric using the bootstrap method with 1,000 iterations. The confidence intervals are reported in the format of
In addition to the quantitative evaluation, we qualitatively analyze the model’s predictions using the Grad-CAM visualizations. This allows us to gain insights into the model’s decision-making process and ensure that it is focusing on the relevant regions of the knee MRI scans. To assess the model’s robustness and generalization ability, we perform cross-dataset evaluation by testing KneeXNet on an independent dataset with different acquisition protocols and patient demographics. This helps to validate the model’s performance in real-world scenarios and ensures that it can generalize well to unseen data.
4.2 Comparison with state-of-the-art methods
The performance of KneeXNet was thoroughly evaluated and compared against several state-of-the-art methods, including both traditional machine learning approaches and deep learning models, to assess its effectiveness in detecting knee joint abnormalities, ACL tears, and meniscal tears. The comparison was conducted using the same experimental setup and evaluation metrics, ensuring a fair and unbiased assessment of each method’s capabilities in Figure 3.

Figure 3. The visualization presents a four-panel comparison. In the top left panel, AUC scores across all methods for abnormality, ACL tear, and meniscal tear detection on the test set are shown, with methods color-coded by category. The top right panel illustrates KneeXNet’s performance metrics for abnormality detection on both test and independent datasets. The bottom left panel displays the average AUC scores by method category (Traditional ML, Deep Learning, and KneeXNet) across all three diagnostic tasks. Finally, the bottom right panel presents the percentage improvement of KneeXNet over SENet (the best competing method) for all tasks in both datasets.
Traditional machine learning methods, such as support vector machines (SVMs), random forests (RFs), and gradient boosting machines (GBMs), were included in the comparison to establish a baseline performance. These methods have been widely used in various classification tasks, including medical image analysis, and have demonstrated good performance in certain scenarios. However, their ability to capture complex patterns and hierarchical features in high-dimensional data, such as MRI scans, is limited compared to deep learning approaches. Among the traditional methods, GBMs achieved the highest AUC scores of 0.901
Deep learning models, including 2D CNNs, 3D CNNs, and attention-based models such as the Transformer and SENet, were also evaluated to compare KneeXNet against more advanced and state-of-the-art architectures. These models have shown remarkable success in various computer vision tasks, including medical image analysis, thanks to their ability to automatically learn hierarchical features from raw input data. The 2D CNN, which processes the MRI scans slice by slice, achieved AUC scores of 0.923

Table 3. Performance comparison of KneeXNet with state-of-the-art methods on the test set and an independent dataset.
To address this limitation, the 3D CNN, which takes into account the volumetric information by processing the MRI scans as 3D volumes, was evaluated. The 3D CNN achieved higher AUC scores of 0.937
Despite the impressive performance of these deep learning models, KneeXNet consistently outperformed all competing methods by a significant margin. On the test set, KneeXNet achieved AUC scores of 0.985
Table 4 presents the performance comparison of KneeXNet with state-of-the-art methods on the test set and an independent dataset, including the 95% confidence intervals for each metric. On the test set, KneeXNet achieved an AUC of
To provide a more comprehensive evaluation, we also reported the specificity and ROC curves for each model. Specificity, defined as
Furthermore, we conducted paired t-tests to assess the statistical significance of the performance differences between KneeXNet and the competing methods. The p-values were reported, with a significance level of 0.05. KneeXNet showed statistically significant improvements over all other methods (p < 0.05) for all three tasks on both the test set and the independent dataset, confirming the superiority of our proposed approach.
4.3 Cross-dataset evaluation
To further validate the robustness and generalization ability of KneeXNet, a cross-dataset evaluation was performed by testing the model on an independent dataset with different acquisition protocols and patient demographics. This evaluation is crucial to assess the model’s performance in real-world scenarios and ensure that it can generalize well to unseen data.
On the independent dataset, KneeXNet maintained its superior performance, achieving AUC scores of 0.978
The graph convolutional layers allow KneeXNet to model the intricate relationships between different anatomical structures within the knee joint, enabling the model to consider both local features and global context. The multi-scale feature fusion module enhances KneeXNet’s ability to capture both fine-grained details and broader contextual information by combining features from different resolutions and receptive fields. The contrastive learning scheme employed by KneeXNet further improves its discriminative power and robustness by encouraging the model to learn more distinguishable representations. By contrasting positive and negative pairs of MRI patches during training, the model can better capture the essential patterns and variations in knee MRI data, leading to improved generalization and reduced overfitting.
To further demonstrate the effectiveness of KneeXNet, we have expanded our comparison to include transformer-based models, such as Vision Transformers (ViT) and Swin Transformers. Table 5 presents the performance of KneeXNet and these transformer-based models on the test set and the independent dataset. KneeXNet consistently outperforms both ViT and Swin Transformers across all three tasks (abnormality, ACL tear, and meniscal tear detection), highlighting the superiority of its graph-based architecture in capturing the complex spatial dependencies in knee MRI data.

Table 5. Performance comparison of KneeXNet with transformer-based models on the test set and an independent dataset.
The strong performance of KneeXNet compared to transformer-based models can be attributed to its ability to effectively model the intricate relationships between different anatomical structures in the knee joint. By representing the knee MRI as a graph and leveraging graph convolutional layers, KneeXNet can capture both local and global contextual information, leading to more accurate predictions. In contrast, transformer-based models, while powerful in capturing long-range dependencies, may not be as effective in modeling the specific spatial relationships present in knee MRI data.
These findings suggest that KneeXNet has the potential to serve as a powerful tool for assisting radiologists in the diagnosis of knee joint disorders, improving the accuracy and efficiency of the diagnostic process. The model’s ability to accurately identify abnormalities, ACL tears, and meniscal tears can help prioritize cases for further review, reduce the risk of missed diagnoses, and guide treatment decisions.
In the supplementary data, the application code for the Django framework is provided. Django is a high-level Python web framework that follows the Model-View-Controller (MVC) architectural pattern, promoting clean and pragmatic design. It is widely adopted for rapid development of secure and maintainable websites. The framework provides an Object-Relational Mapping (ORM) layer that abstracts the database, allowing developers to interact with the data using Python objects and methods. This eliminates the need for writing complex SQL queries and simplifies database management.
Figure 5 showcases representative coronal, sagittal, and axial MRI views of the knee joint, which collectively provide a comprehensive assessment of the knee anatomy and potential pathologies. These multiplanar views offer complementary information, enabling radiologists and AI models like KneeXNet to thoroughly evaluate the intricate structures within the knee joint and identify abnormalities with high precision. The coronal view allows for the assessment of the medial and lateral compartments of the knee, including the medial and lateral menisci, collateral ligaments, and the articular cartilage. The sagittal view, on the other hand, provides a clear visualization of the cruciate ligaments (ACL and PCL), the posterior horns of the menisci, and the patellofemoral joint. Lastly, the axial view offers valuable insights into the patella, trochlear groove, and the tibial and femoral condyles. By leveraging these multiple viewpoints, KneeXNet can effectively analyze the complex anatomy of the knee joint and detect various pathologies, such as ligament tears, meniscal injuries, and cartilage defects. The model’s ability to process and integrate information from different MRI planes contributes to its high diagnostic accuracy and robustness.

Figure 5. Representative coronal, sagittal, and axial MRI views of the knee joint, providing complementary information for the comprehensive assessment of knee anatomy and pathology.
4.4 Ablation study
To investigate the contribution of each component in KneeXNet, we conduct an ablation study by systematically removing or replacing individual modules and evaluating the model’s performance on the test set. Specifically, we consider the following variants of KneeXNet:
• KneeXNet-G: KneeXNet without the graph convolutional layers, replacing them with standard convolutional layers.
• KneeXNet-M: KneeXNet without the multi-scale feature fusion module, using only a single scale of features.
• KneeXNet-C: KneeXNet without the contrastive learning scheme, trained using only the cross-entropy loss.
• KneeXNet-A: KneeXNet without the attention mechanism in the graph convolutional layers.
• KneeXNet-R: KneeXNet with a ResNet-50 backbone instead of the graph convolutional layers.
Table 6 presents the results of the ablation study, reporting the AUC, accuracy, precision, recall, and F1 score for each variant of KneeXNet on the test set.
The results demonstrate that each component of KneeXNet contributes to its overall performance, with the full model achieving the highest scores across all evaluation metrics. Replacing the graph convolutional layers with standard convolutional layers (KneeXNet-G) leads to a noticeable drop in performance, with the AUC decreasing from 0.985
Figure 6 illustrates the application of KneeXNet in detecting a complete anterior cruciate ligament (ACL) tear, which is a common and potentially debilitating knee injury. The segmented MRI image highlights the torn ACL, demonstrating the model’s capability to accurately localize and delineate this critical structure. ACL tears often result from sudden directional changes, deceleration, or landing from a jump, leading to instability and impaired function of the knee joint. Accurate detection of ACL tears is crucial for timely diagnosis, treatment planning, and prevention of long-term complications such as osteoarthritis. KneeXNet’s success in detecting complete ACL tears can be attributed to its unique architecture, which combines graph convolutional layers, multi-scale feature fusion, and contrastive learning. These components enable the model to capture the complex spatial relationships and hierarchical features within the knee joint, allowing for precise localization and segmentation of the injured ACL.

Figure 6. Segmented MRI image highlighting a complete anterior cruciate ligament (ACL) tear, a common and potentially debilitating knee injury that can be accurately detected using KneeXNet.
Removing the multi-scale feature fusion module (KneeXNet-M) also results in a performance decline, with the AUC and accuracy dropping to 0.978
To further analyze the impact of each component on the model’s performance, we examine the precision, recall, and F1 scores for each variant of KneeXNet. The full model achieves the highest precision of 0.972
Figure 7 presents a segmented MRI image that highlights a partial anterior cruciate ligament (ACL) injury. This image demonstrates the capability of KneeXNet, to identify and localize even subtle signs of knee joint damage. The precise segmentation of the partially torn ACL showcases the model’s ability to focus on the most relevant regions of the MRI scan and provide a detailed visualization of the injury.

Figure 7. Segmented MRI image showcasing a partial anterior cruciate ligament (ACL) injury, demonstrating the ability of advanced image analysis methods to identify and localize even subtle signs of knee joint damage.
To investigate the impact of contrastive learning on KneeXNet’s performance, we conducted an additional ablation study by training a variant of the model without the contrastive learning component (KneeXNet-C). Table 7 presents the results of this ablation study, comparing the performance of KneeXNet and KneeXNet-C on the test set.
The results show that removing the contrastive learning component (KneeXNet-C) leads to a decrease in performance across all evaluation metrics. Specifically, the AUC score drops from 0.985
The contrastive learning scheme enhances the model’s discriminative power and robustness by encouraging it to learn more distinguishable representations. During training, positive pairs of MRI patches are generated by applying the same augmentation to the same patch, while negative pairs are obtained by applying different augmentations to different patches. By minimizing the contrastive loss, which maximizes the similarity between positive pairs while minimizing the similarity between negative pairs, KneeXNet learns to capture the essential patterns and variations in knee MRI data. This self-supervised learning approach helps the model to better generalize to unseen data and reduces the risk of overfitting.
To validate the interpretability of KneeXNet’s predictions, we conducted a study involving three experienced musculoskeletal radiologists. The radiologists annotated the regions of interest (ROIs) for a subset of 100 MRI scans from the test set, focusing on the areas most informative for their diagnosis. We then compared the radiologist annotations with the Grad-CAM heatmaps generated by KneeXNet using the Dice similarity coefficient (DSC) and the Intersection over Union (IoU) metrics.
Table 8 presents the results of this validation study, showing a high agreement between the Grad-CAM heatmaps and the radiologist annotations. The average DSC and IoU values of 0.87
The Grad-CAM visualizations for KneeXNet-G show that the model without graph convolutional layers tends to have more scattered and less focused activation maps, suggesting that the graph-based representation helps the model capture the spatial dependencies and concentrate on the most informative regions. KneeXNet-M, which lacks the multi-scale feature fusion module, exhibits activation maps that are more localized but less comprehensive, indicating that the integration of features from different scales helps the model develop a more holistic understanding of the MRI scans. The heatmaps for KneeXNet-C reveal that the absence of contrastive learning leads to less discriminative activation patterns, with the model focusing on less relevant areas of the MRI scans. This suggests that the self-supervised learning approach enhances the model’s ability to differentiate between normal and abnormal knee joint structures. KneeXNet-A, which does not include the attention mechanism, shows more uniform activation maps, indicating that the attention mechanism is crucial for adaptively weighting the features and focusing on the most informative regions. Lastly, the Grad-CAM visualizations for KneeXNet-R demonstrate that the ResNet-50 backbone, when combined with the multi-scale feature fusion, contrastive learning, and attention mechanisms, can also produce highly targeted activation maps. However, the heatmaps are slightly less precise compared to those of the full KneeXNet model, suggesting that the graph-based representation provides an additional level of specificity in localizing knee joint abnormalities.
4.5 Computational costs and hardware requirements
The computational costs and hardware requirements of KneeXNet are important considerations for its practical deployment in clinical settings. KneeXNet was trained on a server with four NVIDIA A100 GPUs, each with 40 GB of memory. The training process took approximately 48 h, with a batch size of 32 and a learning rate of 0.001. While these requirements may seem substantial, they are well within the capabilities of modern GPU servers commonly found in research institutions and hospitals.
For inference, KneeXNet requires a single GPU with at least 16 GB of memory, making it feasible to deploy on a wide range of hardware configurations. The average inference time per MRI scan is just 0.5 s, enabling near real-time predictions and seamless integration into clinical workflows. This rapid inference speed is crucial for the model’s practical utility, as it allows radiologists to quickly obtain second opinions and make informed decisions without disrupting their normal routine.
To further optimize KneeXNet’s computational efficiency, several strategies can be explored. Mixed-precision training, which utilizes both 16-bit and 32-bit floating-point representations, can significantly reduce memory consumption and training time without compromising model performance. Quantization techniques, such as post-training quantization or quantization-aware training, can convert the model’s weights and activations to lower-precision representations (e.g., 8-bit integers), reducing storage requirements and inference latency. These optimization techniques can help make KneeXNet more accessible and cost-effective for a wide range of clinical settings, from large academic hospitals to smaller imaging centers.
4.6 Web-based interface validation
To validate the web-based interface for clinical use, we conducted a usability study involving 10 radiologists with varying levels of experience in musculoskeletal imaging. The radiologists used the interface to analyze a set of 50 knee MRI scans and provided feedback on the system’s ease of use, intuitiveness, and diagnostic assistance. The feedback was collected through a combination of Likert scale ratings and open-ended questions.
The usability study results showed that the radiologists were highly satisfied with the web-based interface, with an average usability score of 4.2 out of 5. They appreciated the seamless integration of KneeXNet’s predictions and the ability to review the model’s decision-making process through Grad-CAM visualizations. The radiologists also provided valuable suggestions for improvement, such as incorporating additional visualization tools and enabling side-by-side comparisons with previous scans. These suggestions will be considered in future iterations of the web-based interface to further enhance its clinical utility.
5 Discussion
The present study introduces KneeXNet, a novel deep learning framework for the classification of knee joint injuries using MRI data. The proposed model leverages the power of graph convolutional networks, multi-scale feature fusion, and contrastive learning to effectively capture the complex patterns and spatial dependencies in knee MRI scans. The experimental results demonstrate the superior performance of KneeXNet compared to state-of-the-art methods, highlighting its potential for assisting radiologists in the diagnosis of knee joint disorders.
The key strength of KneeXNet lies in its ability to model the intricate relationships between different anatomical structures within the knee joint. By representing the knee MRI as a graph, where nodes correspond to key anatomical landmarks and edges represent their spatial connections, KneeXNet can effectively propagate and integrate information across the entire joint. This graph-based approach enables the model to consider not only the local features of individual structures but also their global context and interactions, leading to a more comprehensive understanding of the knee joint pathology. Another notable aspect of KneeXNet is its incorporation of multi-scale feature fusion, which allows the model to capture both fine-grained details and broader contextual information from knee MRI scans. By combining features from different resolutions and receptive fields, KneeXNet can adaptively adjust the importance of different scales based on the specific characteristics of each MRI scan, enabling it to effectively handle the heterogeneity and complexity of knee joint injuries.
The integration of Grad-CAM visualizations in KneeXNet provides valuable insights into the model’s decision-making process, enhancing its interpretability and trustworthiness. By highlighting the regions of the knee MRI that contribute most to the model’s predictions, Grad-CAM enables radiologists to understand and validate the model’s reasoning, fostering a more collaborative and transparent relationship between the AI system and medical experts. Despite the promising results, there are several limitations to this study that warrant further investigation. While the MRNet dataset used in this study is one of the largest publicly available knee MRI datasets, it may not fully represent the diversity of knee joint pathologies encountered in clinical practice. Future research should aim to validate KneeXNet on even larger and more diverse datasets, including multi-center and multi-vendor MRI scans, to assess its performance in real-world scenarios.
To address these limitations and improve the clinical applicability of KneeXNet, we propose several future research directions. First, multi-center studies with diverse datasets should be conducted to evaluate the model’s performance across different clinical settings and patient populations. Second, transfer learning approaches can be explored to adapt KneeXNet to new domains, leveraging the knowledge gained from the MRNet dataset to fine-tune the model for specific clinical environments. Third, continuous model updates with expanding datasets can help KneeXNet stay current with the latest advances in MRI technology and adapt to evolving patient demographics.
6 Conclusion
In this study, we present KneeXNet, a novel deep learning framework for the classification of knee joint injuries using MRI data. By leveraging the power of graph convolutional networks, multi-scale feature fusion, and contrastive learning, KneeXNet effectively captures the complex patterns and spatial dependencies in knee MRI scans, outperforming state-of-the-art methods in detecting abnormalities, ACL tears, and meniscal tears. The integration of Grad-CAM visualizations enhances the interpretability of KneeXNet, enabling radiologists to understand and validate the model’s decision-making process. The promising results of this study highlight the potential of deep learning in improving the diagnosis and management of knee joint disorders, paving the way for the development of AI-assisted diagnostic tools in musculoskeletal radiology.
Data availability statement
The datasets presented in this study can be found in online repositories. The names of the repository/repositories and accession number(s) can be found in the article/supplementary material.
Author contributions
JS: Conceptualization, Methodology, Writing – original draft. YC: Formal Analysis, Methodology, Writing – original draft. YZ: Supervision, Validation, Writing – review and editing. BQ: Resources, Visualization, Writing – review and editing.
Funding
The author(s) declare that no financial support was received for the research and/or publication of this article.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Generative AI statement
The author(s) declare that no Generative AI was used in the creation of this manuscript.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
Alom, M. Z., Taha, T. M., Yakopcic, C., Westberg, S., Sidike, P., Nasrin, M. S., et al. (2018). The history began from alexnet: a comprehensive survey on deep learning approaches. arXiv Prepr. arXiv:1803.01164. doi:10.48550/arXiv.1803.01164
Azcona, D., McGuinness, K., and Smeaton, A. F. (2020). “A comparative study of existing and new deep learning methods for detecting knee injuries using the mrnet dataset,” in 2020 International Conference on Intelligent Data Science Technologies and Applications (IDSTA) (IEEE), 149–155.
Belton, N., Welaratne, I., Dahlan, A., Hearne, R. T., Hagos, M. T., Lawlor, A., et al. (2021). “Optimising knee injury detection with spatial attention and validating localisation ability,” in Annual conference on medical image understanding and analysis (Springer), 71–86.
Bezabh, Y. A., Salau, A. O., Abuhayi, B. M., and Ayalew, A. M. (2024). Classification of cervical spine disease using convolutional neural network. Multimedia Tools Appl. 83, 88963–88979. doi:10.1007/s11042-024-18970-x
Bien, N., Rajpurkar, P., Ball, R. L., Irvin, J., Park, A., Jones, E., et al. (2018). Deep-learning-assisted diagnosis for knee magnetic resonance imaging: development and retrospective validation of mrnet. PLoS Med. 15, e1002699. doi:10.1371/journal.pmed.1002699
Chan, V. C., Ross, G. B., Clouthier, A. L., Fischer, S. L., and Graham, R. B. (2022). The role of machine learning in the primary prevention of work-related musculoskeletal disorders: a scoping review. Appl. Ergon. 98, 103574. doi:10.1016/j.apergo.2021.103574
Chang, G. H., Felson, D. T., Qiu, S., Guermazi, A., Capellini, T. D., and Kolachalama, V. B. (2020). Assessment of knee pain from mr imaging using a convolutional siamese network. Eur. Radiol. 30, 3538–3548. doi:10.1007/s00330-020-06658-3
Chang, P. D., Wong, T. T., and Rasiej, M. J. (2019). Deep learning for detection of complete anterior cruciate ligament tear. J. digital imaging 32, 980–986. doi:10.1007/s10278-019-00193-4
Chao, I., Pinkall, U., Sanan, P., and Schröder, P. (2010). A simple geometric model for elastic deformations. ACM Trans. Graph. (TOG) 29, 1–6. doi:10.1145/1833351.1778775
D’Angelo, T., Caudo, D., Blandino, A., Albrecht, M. H., Vogl, T. J., Gruenewald, L. D., et al. (2022). Artificial intelligence, machine learning and deep learning in musculoskeletal imaging: current applications. J. Clin. Ultrasound 50, 1414–1431. doi:10.1002/jcu.23321
Farooq, M. U., Ullah, Z., Khan, A., and Gwak, J. (2023). Dc-aae: dual channel adversarial autoencoder with multitask learning for kl-grade classification in knee radiographs. Comput. Biol. Med. 167, 107570. doi:10.1016/j.compbiomed.2023.107570
Fernandes, V., Mendonça, É., Palma, M. L., Nogueira, M., Godina, R., and Gabriel, A. T. (2022). “Ergonomics and machine learning: wearable sensors in the prevention of work-related musculoskeletal disorders,” in Occupational and environmental safety and health IV (Springer), 199–210.
Gaj, S., Yang, M., Nakamura, K., and Li, X. (2020). Automated cartilage and meniscus segmentation of knee mri with conditional generative adversarial networks. Magnetic Reson. Med. 84, 437–449. doi:10.1002/mrm.28111
Haddadian, J., and Balamurali, M. (2022). “Transfer learning and data augmentation in the diagnosis of knee mri,” in Australasian Joint Conference on Artificial Intelligence (Springer), 452–463.
Hu, J., Peng, J., Zhou, Z., Zhao, T., Zhong, L., Yu, K., et al. (2025). Associating knee osteoarthritis progression with temporal-regional graph convolutional network analysis on mr images. J. Magnetic Reson. Imaging 61, 378–391. doi:10.1002/jmri.29412
Hung, T. N. K., Vy, V. P. T., Tri, N. M., Hoang, L. N., Tuan, L. V., Ho, Q. T., et al. (2023). Automatic detection of meniscus tears using backbone convolutional neural networks on knee mri. J. Magnetic Reson. Imaging 57, 740–749. doi:10.1002/jmri.28284
Johnson, P. M., Lin, D. J., Zbontar, J., Zitnick, C. L., Sriram, A., Muckley, M., et al. (2023). Deep learning reconstruction enables prospectively accelerated clinical knee mri. Radiology 307, e220425. doi:10.1148/radiol.220425
Kara, A. C., and Hardalaç, F. (2021). Detection and classification of knee injuries from mr images using the mrnet dataset with progressively operating deep learning methods. Mach. Learn. Knowl. Extr. 3, 1009–1029. doi:10.3390/make3040050
Kulseng, C. P. S., Nainamalai, V., Grøvik, E., Geitung, J. T., Årøen, A., and Gjesdal, K. I. (2023). Automatic segmentation of human knee anatomy by a convolutional neural network applying a 3d mri protocol. BMC Musculoskelet. Disord. 24, 41. doi:10.1186/s12891-023-06153-y
Lee, G.-B., Jeong, Y.-J., Kang, D. Y., Yun, H. J., and Yoon, M. (2024). Multimodal feature fusion-based graph convolutional networks for alzheimer’s disease stage classification using f-18 florbetaben brain pet images and clinical indicators. PloS one 19, e0315809. doi:10.1371/journal.pone.0315809
Li, C., Liu, M., Xia, J., Mei, L., Yang, Q., Shi, F., et al. (2022). Predicting brain amyloid-β pet grades with graph convolutional networks based on functional mri and multi-level functional connectivity. J. Alzheimer’s Dis. 86, 1679–1693. doi:10.3233/jad-215497
Mihcin, S., Sahin, A. M., Yilmaz, M., Alpkaya, A. T., Tuna, M., Akdeniz, S., et al. (2023). Database covering the prayer movements which were not available previously. Sci. data 10, 276. doi:10.1038/s41597-023-02196-x
Namiri, N. K., Flament, I., Astuto, B., Shah, R., Tibrewala, R., Caliva, F., et al. (2020). Deep learning for hierarchical severity staging of anterior cruciate ligament injuries from mri. Radiol. Artif. Intell. 2, e190207. doi:10.1148/ryai.2020190207
Nasser, Y., Jennane, R., Chetouani, A., Lespessailles, E., and El Hassouni, M. (2020). Discriminative regularized auto-encoder for early detection of knee osteoarthritis: data from the osteoarthritis initiative. IEEE Trans. Med. imaging 39, 2976–2984. doi:10.1109/tmi.2020.2985861
Qiu, Z., Xie, Z., Lin, H., Li, Y., Ye, Q., Wang, M., et al. (2024). Learning co-plane attention across mri sequences for diagnosing twelve types of knee abnormalities. Nat. Commun. 15, 7637. doi:10.1038/s41467-024-51888-4
Recht, M. P., Zbontar, J., Sodickson, D. K., Knoll, F., Yakubova, N., Sriram, A., et al. (2020). Using deep learning to accelerate knee mri at 3 t: results of an interchangeability study. Am. J. Roentgenol. 215, 1421–1429. doi:10.2214/ajr.20.23313
Roth, P. C., Arnold, D. C., and Miller, B. P. (2003). “Mrnet: a software-based multicast/reduction network for scalable tools,” in Proceedings of the 2003 ACM/IEEE conference on Supercomputing (ACM/IEEE), 21.
Sivakumari, T., and Vani, R. (2022). “Implementation of alexnet for classification of knee osteoarthritis,” in 2022 7th International Conference on Communication and Electronics Systems (ICCES) (IEEE), 1405–1409.
Tuazon, K. L. S., Magboo, M. S. A., and Magboo, V. P. C. (2023). “Anterior cruciate ligament injury classification from mri scans using deep learning,” in 2023 IEEE International Conference on Machine Learning and Applied Network Technologies (ICMLANT) (IEEE), 1–6.
Vera Cruz, G., Bucourt, E., Réveillère, C., Martaillé, V., Joncker-Vannier, I., Goupille, P., et al. (2022). Machine learning reveals the most important psychological and social variables predicting the differential diagnosis of rheumatic and musculoskeletal diseases. Rheumatol. Int. 42, 1053–1062. doi:10.1007/s00296-021-04916-1
Voinea, Ş. V., Gheonea, I. A., Teică, R. V., Florescu, L. M., Roman, M., and Selişteanu, D. (2024). Refined detection and classification of knee ligament injury based on resnet convolutional neural networks. Life 14, 478. doi:10.3390/life14040478
Wang, H., Qiu, X., Li, B., Tan, X., and Huang, J. (2025a). Multimodal heterogeneous graph fusion for automated obstructive sleep apnea-hypopnea syndrome diagnosis. Complex and Intelligent Syst. 11, 44. doi:10.1007/s40747-024-01648-0
Wang, H., Qiu, X., Xiong, Y., and Tan, X. (2025b). Autogrn: an adaptive multi-channel graph recurrent joint optimization network with copula-based dependency modeling for spatio-temporal fusion in electrical power systems. Inf. Fusion 117, 102836. doi:10.1016/j.inffus.2024.102836
Wang, H., Zhao, J., Su, Y., and Zheng, C. (2021). sccdg: a method based on dae and gcn for scrna-seq data analysis. IEEE/ACM Trans. Comput. Biol. Bioinforma. 19, 3685–3694. doi:10.1109/tcbb.2021.3126641
Wang, Q., Yao, M., Song, X., Liu, Y., Xing, X., Chen, Y., et al. (2024). Automated segmentation and classification of knee synovitis based on mri using deep learning. Acad. Radiol. 31, 1518–1527. doi:10.1016/j.acra.2023.10.036
Wang, Y., Li, S., Zhao, B., Zhang, J., Yang, Y., and Li, B. (2022). A resnet-based approach for accurate radiographic diagnosis of knee osteoarthritis. CAAI Trans. Intell. Technol. 7, 512–521. doi:10.1049/cit2.12079
Wei, X., Yu, R., and Sun, J. (2020). “View-gcn: view-based graph convolutional network for 3d shape analysis,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 1850–1859.
Yang, M., Colak, C., Chundru, K. K., Gaj, S., Nanavati, A., Jones, M. H., et al. (2022). Automated knee cartilage segmentation for heterogeneous clinical mri using generative adversarial networks with transfer learning. Quantitative Imaging Med. Surg. 12, 2620–2633. doi:10.21037/qims-21-459
Yaqub, M., Jinchao, F., Ahmed, S., Arshid, K., Bilal, M. A., Akhter, M. P., et al. (2022). Gan-tl: generative adversarial networks with transfer learning for mri reconstruction. Appl. Sci. 12, 8841. doi:10.3390/app12178841
Yin, Z., Wang, H., Chen, B., Zhang, X., Lin, X., Sun, H., et al. (2024). Federated semi-supervised representation augmentation with cross-institutional knowledge transfer for healthcare collaboration. Knowledge-Based Syst. 300, 112208. doi:10.1016/j.knosys.2024.112208
Zhuang, Z., Si, L., Wang, S., Xuan, K., Ouyang, X., Zhan, Y., et al. (2022). Knee cartilage defect assessment by graph representation and surface convolution. IEEE Trans. Med. Imaging 42, 368–379. doi:10.1109/tmi.2022.3206042
Keywords: knee joint disorders, magnetic resonance imaging, automated injury detection, computer-aided diagnosis, machine learning
Citation: Sun J, Cao Y, Zhou Y and Qi B (2025) Leveraging spatial dependencies and multi-scale features for automated knee injury detection on MRI diagnosis. Front. Bioeng. Biotechnol. 13:1590962. doi: 10.3389/fbioe.2025.1590962
Received: 11 March 2025; Accepted: 23 April 2025;
Published: 06 May 2025.
Edited by:
Chenyu Sun, The Second Affiliated Hospital of Anhui Medical University, ChinaReviewed by:
Borui Li, Peking University, China磊 张, China Academy of Chinese Medical Sciences, China
Copyright © 2025 Sun, Cao, Zhou and Qi. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Baoqiao Qi, cWJxMDAxQHNvaHUuY29t; Ying Zhou, MTM4MTc2NTc1MDBAMTYzLmNvbQ==
†These authors have contributed equally to this work