Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Physiol., 27 January 2026

Sec. Computational Physiology and Medicine

Volume 16 - 2025 | https://doi.org/10.3389/fphys.2025.1704428

Predicting the growth of asymptomatic small abdominal aortic aneurysms (AAA) based on deep learning

Jiaxin Cheng,&#x;Jiaxin Cheng1,2Zhiqiang ZhangZhiqiang Zhang2Yasong WangYasong Wang2Yu SunYu Sun3Nan WangNan Wang3Xiaozeng Wang
&#x;Xiaozeng Wang2*Sihan Wang
&#x;Sihan Wang4*
  • 1College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, Liaoning, China
  • 2National Key Laboratory of Frigid Zone Cardiovascular Disease, Cardiovascular Research Institute and Department of Cardiology, General Hospital of Northern Theater Command, Shenyang, Liaoning, China
  • 3Department of Radiology, General Hospital of Northern Theater Command, Shenyang, Liaoning, China
  • 4State Key Laboratory of Robotics, Shenyang Institute of Automation, Chinese Academy of Sciences, Shenyang, Liaoning, China

Accurate prediction of asymptomatic small abdominal aortic aneurysm (AAA) growth is crucial for risk stratification and personalized surveillance. This study developed an end-to-end deep learning framework to predict rapid expansion (≥0.5 cm/6 months) using computed tomography angiography (CTA) images from 81 asymptomatic patients with small AAA (30 rapid-growth and 51 stable patients). The pipeline integrated three core components: a ResNet50 classifier for identifying aortic images (99.86% accuracy, 99.91% F1-score), a YOLOv11 detector for localizing aneurysms (precision–recall: 0.902), and a MedMamba-based feature fusion model that combined imaging features with clinical metadata via multi-head self-attention. Model robustness was ensured through stratified 5-fold cross-validation and comprehensive data augmentation. The fusion model achieved a predictive accuracy of 98.75% and an F1-score of 97.78, outperforming seven classical deep learning backbones. Furthermore, explainability analyses confirmed the model’s reliance on established clinical risk factors and highlighted biologically plausible imaging regions for prediction. The proposed ResNet50–YOLOv11–MedMamba framework demonstrates the feasibility of automating AAA growth prediction directly from CTA and shows promising potential to enhance clinical decision-making.

Introduction

Abdominal aortic aneurysm (AAA) constitutes a pathological dilation of the infrarenal abdominal aorta and is often subclinical until the time of acute rupture, a complication with prehospital mortality exceeding 80% (Schanzer et al., 2021; Isselbacher et al., 2022). Current clinical management relies on serial diameter monitoring, with intervention typically recommended when the maximum diameter exceeds 50 mm for men or a lower threshold for women (Wanhainen et al., 2024). The diameter is the most commonly used risk marker in AAA disease. It is manually measured clinically by ultrasound or multiplanar reconstruction of computed tomography angiography (CTA) perpendicular to the centerline. Current predictions of the AAA rupture risk, and consequently the indications for preventive treatment, are based on the maximum anterior-posterior diameter, measured perpendicular to the centerline with three-dimensional reconstructed CTA images, and growth rate. However, this size-only paradigm is an imperfect predictor of risk. A significant proportion of ruptures occur in aneurysms below this surgical threshold (Collaborators, 2013; Spanos et al., 2020), with large cohort studies reporting an annual rupture risk of 0.03% for small AAAs (Oliver-Williams et al., 2019). This critical limitation underscores the urgent need for better predictors of aneurysm behavior. Consequently, accurately forecasting the AAA growth rate has become a central research priority, as it is crucial for surgical planning and personalized surveillance, with guidelines recommending intervention when growth exceeds 0.5 cm per 6 months (Ullery et al., 2018; Isselbacher et al., 2022).

The pursuit of improved prediction has incorporated biological variables (e.g., C-reactive protein and D-dimers) and morphological parameters from CTA (Zhu et al., 2020; Cersit et al., 2021; Fernandez-Alonso et al., 2022; Kontopodis et al., 2022; Ristl et al., 2023; Vanmaele et al., 2025). Furthermore, the pathobiology of the perivascular environment is now recognized as a key contributor to AAA progression (Wang et al., 2022; Zhang et al., 2023; Lv et al., 2024). Deep learning offers a powerful approach to automatically extract prognostic features from images. Previous efforts have largely relied on radiomics derived from manual segmentations or reconstructed geometries for hemodynamic modeling. While valuable, this dependence on manual annotation introduces observer variability, limits scalability, and confines models to a predefined set of human-engineered features, potentially overlooking subtler prognostic patterns in the raw data (Kim et al., 2023; Wang et al., 2023).

To overcome these limitations, we propose a novel, end-to-end deep learning framework that automates the entire pipeline from raw CTA images to growth prediction. Our framework, ResNet50–YOLOv11–MedMamba, is designed to eliminate manual intervention through a three-stage cascade: it first identifies and filters relevant aortic images, then precisely localizes the aneurysm region, and finally fuses the automatically extracted imaging features with clinical metadata for a holistic assessment. This study aims to validate whether this fully automated approach can accurately predict rapid expansion (≥0.5 cm/6 months) of small AAAs, thereby offering a scalable, potentially more robust alternative to existing methods.

Methods

This section provides an overview of the methods employed in this study, with a particular focus on the feature fusion strategy and the application of the ResNet50, YOLOv11, and MedMamba models. We have detailed the data collection process, CTA data acquisition, definition, and the performance metrics used to evaluate our method. By integrating clinical data and CTA image feature-extraction techniques, we aim to improve the accuracy of predicting AAA growth in asymptomatic patients. Each subsection provides a comprehensive overview of the process used, ensuring the clarity of our research methods and facilitating the reproducibility of this work. The framework of the proposed algorithm for our prediction of AAA growth is shown in Figure 1.

Figure 1
Diagram illustrating a pipeline for processing CTA images. The process starts with groups of people labeled blue and brown. It progresses through stages including CTA images acquisition, pre-processing, and ResNet50 analysis, followed by YOLO model application. Clinical features are extracted for a feature fusion model, leading to classification into growth and stable groups.

Figure 1. Framework of our prediction of AAA growth. AAA, abdominal aortic aneurysm; YOLO, You Only Look Once.

Data collection

This retrospective study consecutively screened 289 patients with small AAAs who were treated in our center from January 2019 to December 2023. Among them, 135 patients who underwent intervention due to symptoms were excluded, 61 patients who did not undergo imaging follow-up were excluded, 10 patients were lost to follow-up, and two patients died. Finally, 81 patients were included in this study. According to whether the growth rate was >0.5 cm/6 months, they were divided into the rapid-growth group (n = 30) and the stable group (n = 51). The CTA images of these patients were obtained from the institutional Picture Archiving and Communication Systems at our center for the study, and the patients’ clinical data were obtained from the electronic medical record system. This study followed the ethical guidelines of the Helsinki Declaration and was approved by the ethics committee. The Ethics Committee of General Hospital of Northern Theater Command approved this study with an Ethics Batch Number Y (2024)356. Informed consent was waived due to the retrospective design and the use of fully anonymized, de-identified data.

CTA data acquisition

CTA examinations were conducted via a 256-slice multidetector computed tomography system (Brilliance iCT, Philips Healthcare). The scanning protocol included the following technical specifications: detector collimation of 128 mm × 0.625 mm, a gantry rotation time of 270 ms, automated tube voltage selection ranging from 100 kVp to 120 kVp based on body mass index, and tube current modulation between 500 mAs and 700 mAs. Image acquisition utilized retrospective electrocardiogram gating with intravenous administration of iodinated contrast medium (ioversol 320 mgI/mL) at weight-adjusted doses of 1–1.5 mL/kg and injection rates of 4–6 mL/s. CTA images were reconstructed at a window centered at the peak aortic enhancement phase, which can be determined by a test bolus or bolus-tracking technique. For most patients, this phase may occur at approximately 20–30 s after the start of contrast injection, but it can vary depending on the patient’s cardiovascular status and the injection parameters. CTA images were reconstructed at a window centered at 75% of the R‒R interval, with a section thickness of 0.625 mm and a reconstruction increment of 0.5 mm.

Definition

A small AAA is defined as the diameter of the abdominal aorta between 3 cm and 5 cm in women and between 3 cm and 5.5 cm in men (Isselbacher et al., 2022), and the maximum diameter was measured by two radiologists based on multiplanar reconstruction. The rapid growth of AAA is defined as an increase of more than 5 mm every 6 months (Ullery et al., 2018). Smoking history was defined as an individual’s lifetime consumption of 100 or more cigarettes (Rigotti et al., 2024), and drinking history was defined as daily consumption of at least 50 mL of liquor, at least once a week, for a duration of at least 6 months, including current drinkers and those who stopped drinking but met the aforementioned criteria (Herzog et al., 2024).

Model framework strategy

The CTA data collection adopts the RGB image format, with variable anatomical semantics. To ensure consistent model input, we used a classifier based on ResNet50 pre-trained on ImageNet to automatically identify CTA images containing the abdominal aorta. 182,291 CTA slices were uniformly preprocessed to a resolution of 224 × 224. The ImageNet standard normalization was applied (mean = [0.485, 0.456, 0.406], standard deviation = [0.229, 0.224, 0.225]), and the center cropping strategy was adopted to maintain the integrity of the anatomical structure. This model has undergone over 100 training epochs using a batch size of 64 and an initial learning rate of 0.001, and the Adam optimizer was employed with step-decayed learning rate scheduling (multiplied by 0.1 every 50 epochs), enabling it to reliably perform anatomical filtering for image selection. The loss function used is the cross-entropy loss.

The CTA images that contain the abdominal aorta were labeled by LabelMe and used as coordinates for feature extraction by the You Only Look Once (YOLO) models (Aly et al., 2021; Hechkel and Helali, 2025), located in the local area of the abdominal aorta, with labels set as normal aorta and AAA images. The local abdominal aorta features extracted by YOLO were used for the feature fusion. A comparative analysis framework was implemented to evaluate four YOLO variants (v5, v8, v10, and v11) under identical experimental conditions. Each model was trained for 100 epochs with a batch size of 16 and an input resolution of 640 pixels × 640 pixels. The optimizer is Stochastic Gradient Descent (SGD) with Nesterov momentum (μ = 0.937). The initialized learning rate was set at 0.01 with fixed scheduling and L2 regularization weight decay (λ = 5 × 104). Detection performance was assessed using mean average precision (mAP) at an intersection over union (IoU) = 0.5 threshold (mAP@50) and mAP averaged over IoU thresholds 0.5 to 0.95 (mAP@50–95). Computational efficiency was measured by total training time.

We propose a novel multimodal model that integrates CTA image slices and clinical features via a feature fusion mechanism, as shown in Figure 2. The architecture adopts a dual-pipeline architecture and comprises three innovative components: an image feature encoder, a clinical feature encoder, and a cross-modal attention fusion module. The dataset was randomly partitioned into training, validation, and test sets at an 8:1:1 ratio at the patient level to ensure independence. For the image feature encoder, the MedMamba backbone (Yue and Li, 2024) was selected for its specific design to model long-range dependencies in medical image sequences, a capability highly relevant to our task. The output of MedMamba from all slices was then treated as a sequence and passed into a transformer encoder, which was fed into the image encoder backbone to extract slice-level feature representations. The resulting features were then projected into a shared 512-dimensional latent space via a projection head consisting of linear layers, batch normalization, and dropout. Mean pooling is applied to produce a fixed-length image embedding for each patient.

Figure 2
Flowchart of a feature fusion model with three main sections: Image Feature Encoder, Clinical Feature Encoder, and Evaluation. The Image Encoder processes CTA image slices using MedMamba Backbone, FC, Transformer Encoder, and Mean Pooling. The Clinical Encoder processes clinical features using MLP and MHA, combines with Image Encoder output using Concat and Classifier. Evaluation includes accuracy, confusion matrix, and F1-score. The classification results in either Stable Group or Growth Group.

Figure 2. Overall framework of the feature fusion model. CTA, computed tomography angiography; FC, fully connected; MLP, multilayer perceptron; MHA, multi-head self-attention.

To prevent data leakage, all preprocessing steps were performed independently on each partition. Clinical features with missing values exceeding 50% were excluded from the analysis. For the remaining missing data, four imputation methods were evaluated: k-nearest neighbors (KNN), median imputation, random forest imputation, and multiple imputation (Lee and Styczynski, 2018). KNN imputation (k = 5) was selected because it best preserved the original data distribution and was subsequently applied to each subset (training, validation, test) separately. Continuous features were standardized to a mean of zero and a unit variance using the StandardScaler method. The categorical outcome variable (intact vs. ruptured) was binarized as 0 and 1, respectively. Prior to feature selection, multicollinearity was addressed by identifying and eliminating highly correlated features (Pearson’s |r| > 0.8). We implemented extreme gradient boosting (XGBoost) to screen clinical features. The top seven most discriminative features from each method were selected to form candidate feature subsets. The Shapley Additive exPlanations (SHAP) value is shown in Figure 3.

Figure 3
Bar chart showing the mean SHAP values, indicating the average impact on model output magnitude. Age has the highest impact, followed by MaxDiameter, SBP, Smoking History, CHD, Drinking History, and Male.

Figure 3. Feature importance ranking obtained through feature selection based on extreme gradient boosting. MaxDiameter, maximum diameter of abdominal aortic aneurysm; SBP, systolic blood pressure; CHD, coronary atherosclerotic heart disease.

The clinical features were analyzed via a two-layer multilayer perceptron (MLP), with intermediate batch normalization and dropout layers, to improve generalizability. The final output is mapped into the same latent space as the image embedding (i.e., 1,024 dimensions), enabling subsequent multimodal fusion. A key innovation of our architecture lies in the cross-modal attention fusion module. Here, we treat the image-derived and clinical-derived embeddings as two tokens and input them into a multi-head self-attention (MHA) block. This allows the model to explicitly learn modality-aware representations by attending to both intra- and intermodal interactions. The output vectors from each modality are then concatenated and passed through a multilayer classifier for binary classification. The thickness of the image layer is 5 mm, and the 2D images are all 224×224. All images are preprocessed using the low threshold zero processing method, with the threshold experimentally set at 200. During the training process, the Adam optimizer is used. In the training process, the 5-fold cross-validation method with category balance weights is used, and robustness and adaptive learning rate adjustment (initial 1e−4, reduced by 10% on the platform) are implemented. The loss function is also the cross-entropy loss. The batch size is set to 4, adjusted according to memory requirements for comparing models. Through normalization of shear (max = 1.0), supplementation, and maintaining gradient stability through strict CTA data augmentation, including random planar rotation (±15°), horizontal flipping (50%), ±20% brightness, contrast, and saturation, it is made applicable to medical data. To ensure the reduction of overfitting and underfitting of the model, the early stopping strategy is a common method to prevent model overfitting, with patience = 10. We adopted a stratified 5-fold cross-validation for dataset division to ensure that each fold maintains the original class distribution ratio. To address the class imbalance problem (with a class ratio of 30:51), we used the weighted cross-entropy loss function based on class frequency.

Implementation configurations

Our architecture is implemented on PyTorch 2.7.0, and experiments are conducted on a workstation equipped with an NVIDIA GeForce RTX 4090 GPU and providing up to 24 GB of VRAM, NVIDIA driver version 576.52, and CUDA 12.8.

Comparative experiment

In the comparative experiment, we set up three types of experiments. The first one is the comparison between the feature-extraction module MedMamba and the classical models, including VGG16, ResNet18, ResNet50, ResNet101, DenseNet121, ViT-B/16, and MedViT. The second one is comparing the results of feature fusion achieved by combining other attention mechanisms with the MedMamba model, including the convolutional block attention module (CBAM), channel attention, cross attention, the squeeze-and-excitation (SE) module, and the global attention mechanism (GAM). Additionally, we explored the number of heads in the MHA mechanism and compared the results of 4 heads, 8 heads, 16 heads, and 32 heads, respectively.

Model interpretability and visualization

To interpret the predictions of our trained model, we employed gradient-weighted class activation mapping (Grad-CAM) to generate visual explanations of its decision-making process (Wang et al., 2025; Zhao et al., 2025). This technique produces coarse localization heatmaps by leveraging gradient information flowing into the final convolutional layer of the network, highlighting regions of the input image that are most influential for predicting a specific class while preserving spatial information. A major advantage of Grad-CAM is that it requires no modifications to the model architecture and involves no additional training. In this study, we applied Grad-CAM to the final convolutional layer to visualize activation patterns corresponding to the model’s prediction of the AAA growth.

Statistical analysis

To evaluate the classification results of the deep learning models for the overall group, we used sensitivity, specificity, accuracy, F1-score, precision, the mAP@50, and the area under the curve (AUC). The evaluation formula is available in the supplementary materials. The statistical analyses were performed using IBM® SPSS® version 26 (IBM, Armonk, NY) and Python 3.12.9.

Results

Baseline characteristics

Overall, we evaluated 81 patients with asymptomatic and image-monitored small AAA without intervention. Most were men (90.12%), with a median age of 68 years. Among them, 30 patients met the criteria for rapid growth (0.5 cm/6 months), and the other 51 patients were of the stable type. The proportion of smoking history among the patients in the rapid-growth group was significantly higher than that of the patients in the stable group (P = 0.016). No statistically significant differences in other baseline characteristics were found between the two groups (all P > 0.05). Detailed baseline characteristics are presented in Table 1.

Table 1
www.frontiersin.org

Table 1. Comparison of baseline characteristics between the stable group and the rapid-growth group.

CTA images containing abdominal aorta classification

The CTA data were randomly split into training, validation, and test sets in a ratio of 8:1:1 at the patient level, which is displayed in the Supplementary Table S1. We compared eight models suitable for image classification, mainly to identify CTA images containing the abdominal aorta. The results of the model comparison are shown in Table 2. Among these models, the ResNet50 model performed best, with an accuracy (ACC) and F1-score of 99.86% and 99.91%, respectively. The ResNet50 model classified the CTA images suitable for the next step, that is, abdominal aorta detection, and the results of model training and validation are shown in Figure 4.

Table 2
www.frontiersin.org

Table 2. Performance analysis of different models for classifying CTA images containing the abdominal aorta on the test set.

Figure 4
Graphs showing ResNet50 training results over 100 epochs. Top left: Training and validation loss decrease, with validation loss slightly higher. Top right: Training and validation accuracy increase, nearly converging. Bottom left: Validation precision and recall remain high with minor fluctuations. Bottom right: Validation F1 score and specificity stabilize at high values.

Figure 4. Results of the ResNet50 model for classifying CTA images containing the abdominal aorta. CTA, computed tomography angiography.

Detecting and classifying the abdominal aorta

A total of 31,457 CTA images containing the abdominal aorta were trained with 100 epochs and were completed in 13.479 h. The YOLOv11 model detected the abdominal aorta and classified the normal aortae and AAAs with the highest precision (p = 0.902), which indicates that the model has a strong ability to correctly identify true positives. Its robust performance across four YOLO models is demonstrated in Table 3. This implementation demonstrates a high effectiveness in distinguishing between the normal aorta and the AAA, and the validation set is shown in Figure 5. The model correctly identifies the regions of interest and highlights them with bounding boxes in distinct colors. Figure 6 depicts key performance metrics of the classification result between the normal aorta and AAA. The precision–confidence curve (top left) shows how precision varies with confidence thresholds, indicating consistently high precision near the threshold of 0.902 for all categories. The precision–recall curve (top right) illustrates the relationship between precision and recall, demonstrating stable performance, with the overall mAP50 reaching 0.975. The recall–confidence curve (bottom left) highlights a gradual decline in recall as confidence thresholds increase, reflecting the trade-off between higher confidence and recall. Lastly, the F1–confidence curve (bottom right) provides insights into the harmonic means of precision and recall, maintaining a high F1-score across confidence levels. These curves collectively validate the model’s robustness and balanced performance across the different classification groups. Figure 7 illustrates how the model performed during training and validation over 100 epochs. In the top row, we see the training losses for bounding box regression, classification, and distribution focal loss. All these losses show a steady decline, which suggests that the model is being effectively optimized. Additionally, metrics like precision, recall, mAP50, and mAP50-95 consistently improve, reaching high levels by the end of training. The results of the other three YOLO models are presented in the supplementary materials.

Table 3
www.frontiersin.org

Table 3. Performance comparison of four YOLO models.

Figure 5
CT scan images showing cross-sectional views of an abdomen. The first image highlights the aorta, while the remaining images indicate the presence of an abdominal aortic aneurysm (AAA) with labels pointing to the affected area.

Figure 5. Validation for classifying the normal aorta and abdominal aortic aneurysm of the YOLOv11 model on CTA images. YOLO, You Only Look Once; CTA, computed tomography angiography.

Figure 6
Four graphs comparing model performance. Top left: Precision-Confidence Curve shows precision vs. confidence for aorta, AAA, and all classes with high precision at higher confidence levels. Top right: Precision-Recall Curve exhibits precision vs. recall, with AAA showing the highest at 0.986. Bottom left: Recall-Confidence Curve presents recall vs. confidence, AAA maintains higher recall across confidence levels. Bottom right: F1-Confidence Curve depicts F1 score vs. confidence, all classes peak at 0.91 at 0.280 confidence. Each curve is color-coded for clarity.

Figure 6. Performance curves of the YOLOv11 model for detecting and classifying the abdominal aorta. YOLO, You Only Look Once.

Figure 7
Nine line graphs show various metrics over 100 epochs. The top row includes training box loss, classification loss, distribution focal loss, precision, and recall. The bottom row shows validation box loss, classification loss, distribution focal loss, mean average precision at fifty, and mean average precision from fifty to ninety-five. Each graph features a blue line for results and an orange dotted line for smoothing.

Figure 7. Training and validation losses with performance metrics of the YOLOv11 model for detecting and classifying the abdominal aorta. YOLO, You Only Look Once.

Feature fusion model

A total of 15,704 CTA images of the local abdominal aorta were provided by the YOLOv11 model, which were utilized for the feature fusion model. Among the eight models, the MedMamba was considered the most suitable backbone for further feature fusion to predict AAA growth, which achieved an accuracy of 98.75% and an F1-score of 97.78%. The training and validation loss curves in Figure 8 indicate that the model has converged. The confusion matrices suggest that misclassifications are rare, reinforcing the model’s reliability, as shown in Figure 9. Compared with the VGG16, ResNet18, ResNet50, ResNet101, DenseNet121, ViT-B/16, and MedViT models, the feature fusion model based on MedMamba achieved the best performance, detailed in Table 4 and Figure 10. The training and validation loss curves and confusion matrices of feature fusion models based on seven other backbones are shown in the supplementary materials.

Figure 8
Five line graphs showing the loss curves for five different data folds. Each graph displays training and validation loss over epochs. Training loss is represented by a blue line, and validation loss by an orange line. All graphs depict a general downward trend, indicating a decrease in loss over time, with some fluctuations in the curves.

Figure 8. Training and validation losses of the feature fusion model based on the MedMamba backbone.

Figure 9
Five confusion matrices display classification results for k-fold cross-validation. Fold 1 shows 11 true negatives, 4 true positives, 1 false negative. Folds 2 to 5 each show 10 true negatives, 6 true positives, no false negatives or positives. Each matrix includes a color scale from light to dark blue, indicating value intensity.

Figure 9. Confusion matrices for the feature fusion model based on the MedMamba backbone.

Table 4
www.frontiersin.org

Table 4. Performance analysis of the MedMamba model and the traditional models to predict abdominal aortic aneurysm growth.

Figure 10
Four bar charts compare different models on various metrics. Top left shows Validation AUC, with models achieving high scores. Top right depicts F1 Score, with medmnamba performing best. Bottom left illustrates Accuracy, led by medmnamba. Bottom right shows Sensitivity, with medvit scoring highest. Error bars indicate variability.

Figure 10. Performance comparisons of the MedMamba backbone with the traditional models to predict abdominal aortic aneurysm growth.

Experiments to explore attention modules on model performance

We compared the prediction performance effects of applying different attention modules based on the MedMamba model. From the analysis of Table 5, the MHA mechanism has excellent performance compared to the other attention modules. When applying the MHA mechanism for feature fusion, we compared the effects of 4 heads, 8 heads, 16 heads, and 32 heads on the model results, as shown in Table 6. The 8-head self-attention mechanism achieved the optimal balance and resulted in the best prediction of AAA growth.

Table 5
www.frontiersin.org

Table 5. Evaluation of different attention modules on CTA images to predict abdominal aortic aneurysm growth.

Table 6
www.frontiersin.org

Table 6. Evaluating the number of heads of multi-head self-attention for predicting abdominal aortic aneurysm growth.

Model interpretability and visualization

Grad-CAM was used to generate heatmaps that visually interpreted the model’s reasoning process by highlighting regions within the input CTA images that most strongly influenced its classification decision for ruptured abdominal aortic aneurysm (RAAA), as shown in Figure 11. These visual results indicate that the model consistently focuses on anatomically significant areas, such as suggesting focal enhancement and activation in the perivascular regions, which have a strong correlation with the established imaging markers of AAA growth. Unexpectedly, the areas of thickened blood vessel walls that might reflect vascular inflammatory responses and characteristics of atherosclerosis were not activated.

Figure 11
Six-panel image showing comparisons of growth and stability in original, heatmap, and overlay forms. Panels (a) and (d) are grayscale images labeled Growth-Original and Stable-Original, respectively. Panels (b) and (e) are color-coded heatmaps labeled Growth-Heatmap and Stable-Heatmap, showing intensity variations. Panels (c) and (f) are overlays labeled Growth-Overlay and Stable-Overlay, highlighting specific areas with color contrasts.

Figure 11. Representative Grad-CAM visualizations for a rapid-growth AAA case (a–c) and a stable AAA case (d–f). (a–d) Original CTA image slices. (b–e) Gradient-weighted class activation mapping (Grad-CAM) heatmaps. (c–f) Heatmaps superimposed on the original images. The model’s attention in the rapid-growth case is localized to the perivascular region, while such focused activation is absent in the stable case. Grad-CAM, gradient-weighted class activation mapping; AAA, abdominal aortic aneurysm; CTA, computed tomography angiography.

Discussion

The accurate prediction of small AAA growth remains a significant challenge in clinical practice. Current management, guided by consensus guidelines, primarily relies on periodic surveillance using the maximum aortic diameter as a surrogate for rupture risk. While practical, this conventional approach often fails to capture the complex, multifactorial nature of aneurysm progression, which involves a dynamic interplay of biomechanical stress, chronic inflammation, and proteolytic degradation. This study establishes a novel, automated deep learning framework that integrates ResNet50, YOLOv11, and MedMamba to predict the rapid expansion of asymptomatic small AAAs directly from CTA images and clinical data. The end-to-end pipeline achieved an exceptionally high predictive performance, with an accuracy of 98.75% and an AUC of 99.64% in our cohort of 81 asymptomatic patients. Critically, despite the cohort size, this performance was consistent across rigorous internal validation, demonstrating the substantial potential of leveraging these architectures to automate the entire workflow, from image screening and precise, bounding-box-level aneurysm localization to multimodal feature fusion. This design crucially circumvents the manual segmentation and expert-dependent feature engineering required by traditional radiomics and hemodynamic models (Ristl et al., 2021; Kontopodis et al., 2022; Kim et al., 2023; Wang et al., 2023; Starck et al., 2025). This performance likely stems from both the architectural synergy between detection and sequence modeling components and the framework’s capacity to learn discriminative features directly from raw data, potentially capturing subtle prognostic image patterns that may serve as surrogates for underlying pathophysiological processes, such as inflammation or wall stress, which are lost in manual preprocessing. This study thereby suggests a new performance benchmark for AAA progression prediction and highlights a clinically translatable, scalable pathway for future decision-support tools.

The strong predictive performance of our framework is likely attributable to the synergistic design of its constituent architectures, each addressing specific limitations in conventional AAA analysis. This integrated design effectively establishes a fully automated pipeline that progresses from the identification of aortic images (ResNet50) to precise, bounding-box-level localization of the aneurysm (YOLOv11), and culminates in the fusion of imaging features based on the MedMamba backbone with clinical data. A key advantage of this design is its circumvention of manual segmentation, a step typically required in radiomics and biomechanical modeling, which can introduce observer variability and hinder scalability. The choice of YOLOv11 for localization ensures that subsequent feature extraction is focused exclusively on the pathologically relevant region, minimizing contamination from adjacent anatomic structures. However, the most substantial performance improvements, as evidenced by our comparative ablation studies, appear to originate from the feature fusion stage. Here, the MedMamba backbone was selected for its potential to model long-range dependencies across sequential CTA slices, potentially capturing a more global spatiotemporal context of the aneurysms than standard CNNs. This was complemented by the optimized 8-head MHA mechanism, which acts as a sophisticated information integrator. Our ablation studies (Table 6) demonstrate that this specific configuration offered an optimal balance, enabling the model to integrate multi-scale features, from local textural variations to broader morphological changes. The synergy between MedMamba’s sequential processing and MHA’s cross-modal alignment may allow the model to learn a hierarchical representation that captures morphological and textural heterogeneities, which, although recognized in the literature as markers of progression, are elusive to manual quantification. Consequently, by moving beyond the simplistic reliance on maximum diameter, which is the current cornerstone of clinical guidelines (Isselbacher et al., 2022), our framework suggests a path toward a more holistic, pathophysiologically grounded assessment of AAA progression risk. This could potentially enable more personalized surveillance strategies, ensuring that high-risk patients receive timely intervention while reducing unnecessary imaging for those with stable disease.

When contextualizing our results within the existing landscape of AAA prediction, our automated deep learning framework addresses several persistent methodological challenges. The field has progressively identified key prognostic indicators underpinned by specific pathophysiological mechanisms: for instance, the intraluminal thrombus is not merely a structural feature but an active source of proteolytic activity and wall inflammation that fuels progression (Zhu et al., 2020); radiomic signatures of the perivascular adipose tissue (Lv et al., 2024) may reflect a paracrine, inflammatory “outside-to-inside” signaling pathway; and the role of vascular calcification (Klopf et al., 2022) involves complex biomechanical alterations to wall stress distribution. However, translating these mechanistic insights into clinical tools has been constrained by methodology. Extracting these features typically relies on expert-dependent manual segmentation, a process that not only introduces variability but may also restrict analysis to a predefined set of human-engineered features, as seen in radiomics studies (Kontopodis et al., 2022; Wang et al., 2023), potentially missing subtler, synergistic patterns. Further compounding this, other approaches seek to integrate more complex data, such as combining anatomical imaging with hemodynamic simulations derived from reconstructed geometries to estimate wall stress (Kim et al., 2023). While physiologically insightful, the reliance on manual segmentation and complex multi-physics modeling poses a significant barrier to clinical scalability. In contrast, our ResNet50-YOLOv11-MedMamba framework addresses these bottlenecks by learning directly from raw data. The pipeline automatically transitions from CTA images to a prediction. We hypothesize that by doing so, the model does not merely “see” thrombus or calcification in a conventional sense but may learn complex imaging signatures that are surrogate markers for the underlying inflammatory and biomechanical processes these established features represent. This end-to-end strategy is posited to allow the model to learn a rich, hierarchical representation that can capture the prognostic information contained in the complex interplay of these mechanisms, without being constrained to predefined feature sets. This capability is a plausible explanation for its competitive performance when compared to models utilizing manually segmented features (Kontopodis et al., 2022; Wang et al., 2023) or those relying on a more limited set of traditional parameters (Ristl et al., 2021). Consequently, our work suggests a scalable pathway that leverages deep learning to integrate the pathophysiological complexity of AAA progression into a unified predictive assessment.

The biological plausibility of our model’s predictions was further interrogated using SHAP analysis and Grad-CAM visualizations. SHAP analysis confirmed that the model’s decision-making aligns with established clinical knowledge (Ullery et al., 2018), identifying smoking history, larger baseline aneurysm diameter, and advanced age as the most influential features, which are all well-documented risk factors for AAA progression. This concordance lends critical face validity, demonstrating that the model rationally leverages clinically pertinent information. The interpretation of Grad-CAM visualizations, however, must be contextualized within our model’s architecture. Given that the YOLOv11 pipeline first precisely localizes the aneurysm, Grad-CAM does not function in a primary diagnostic localization role. Instead, it elucidates which sub-regions within and surrounding the delineated aneurysm most strongly drive the prediction. A striking and consistent pattern emerged: in cases of rapid expansion, the model’s attention was frequently mapped not to the aortic wall itself, but to the adjacent perivascular adipose tissue (PVAT). We hypothesize that this pattern may indicate the model’s sensitivity to imaging signatures of perivascular inflammation. PVAT is a recognized secretory organ, modulating vascular inflammation, which is a cornerstone of AAA pathogenesis, via the paracrine release of adipokines and cytokines. It is plausible that our model detects subtle textural or radiodensity alterations within this tissue, capturing a novel imaging biomarker of a pro-inflammatory state. This hypothesis conceptually bridges the imaging findings with the systemic risk profiles highlighted by SHAP (e.g., smoking). We emphasize that this remains a hypothesis-generating observation that warrants future validation through direct correlation with histopathological analysis or specific inflammatory serological markers. Collectively, these explainability techniques present a coherent, biphasic narrative. The model initially anchors its decision in fundamental, patient-specific clinical risks (SHAP) and subsequently appears to refine its prognostic assessment by integrating a localized, imaging-based evaluation of the perivascular milieu (Grad-CAM). This synergy suggests that our framework transcends a mere “black-box” predictor, instead synthesizing established clinical wisdom with potential novel imaging biomarkers to enable a more holistic and mechanistically informed risk stratification for AAA progression.

A recent study by Oh et al. (2025) also developed a deep learning model for predicting AAA progression, providing a valuable benchmark for the field. Our work complements and extends this effort in several key aspects that may contribute to the differing performance profiles. First, the clinical targets differ. Our model was trained to predict rapid growth defined as ≥0.5 cm per 6 months, a threshold directly aligned with guideline recommendations for considering intervention. In contrast, Oh et al. employed a threshold of 2.5 mm/year. Predicting this more abrupt, clinically decisive growth event represents a distinct and potentially more challenging task, which may partly account for the differing performance metrics observed. Second, the feature-extraction paradigms diverge. Their approach relied on manually segmenting aortic geometries to derive predefined radiomic features, a methodologically rigorous but inherently limited process. Our fully automated YOLOv11-based localization, while providing less geometrically precise segmentation, preserves and analyzes the entire periaortic imaging context. This approach may have enabled our model to capture the prognostic information from the perivascular environment (Lv et al., 2024), as directly suggested by our Grad-CAM findings, which is typically excluded during manual contouring. Therefore, our framework explores a different trade-off, prioritizing automated, context-aware feature discovery over manual, lesion-specific measurement. When viewed together, these studies illustrate two parallel and complementary paths forward: one focused on creating robust, manually verified models from large cohorts, and the other pushing the boundaries of fully automated performance.

Beyond the methodological comparisons within the domain of CT-based modeling, our study also lays the groundwork for future integration with emerging, non-invasive biomarkers that probe different aspects of AAA pathophysiology. For instance, circulating biomarkers such as soluble glycoprotein VI (Benson et al., 2024) and specific microRNAs (Thanigaimani et al., 2022) have shown promise in reflecting platelet activity and cellular stress associated with aneurysm growth. Similarly, advanced MRI techniques, including MR elastography to assess wall stiffness (Dong et al., 2022) and 4D flow to quantify wall shear stress (Trenti et al., 2022), provide unique biomechanical and functional insights not directly accessible by standard CTA. While these modalities are not yet part of routine clinical practice for AAA surveillance, they represent a rich source of complementary data. The automated architecture presented in this study is, in principle, amenable to incorporating such multimodal inputs in the future. A compelling long-term goal would be to fuse our automated imaging-derived predictions with these serological and functional biomarkers, potentially creating a supremely robust, multi-parametric risk assessment tool that captures the biological, morphological, and biomechanical drivers of AAA progression in concert.

Limitations

Our study has several limitations that should be considered. First, the performance of our fully automated framework was evaluated on a single-center cohort with a limited sample size. Although we employed rigorous internal validation strategies, including 5-fold cross-validation and comprehensive data augmentation, to ensure robustness and mitigate overfitting, the generalizability of the model requires confirmation in larger, multi-center populations. Our findings may thus be viewed as complementary to recent large-scale, multi-center efforts, such as the study by Oh et al. (2025), which established a robust benchmark for generalizability. While their study exemplifies a pathway toward clinical deployment, our work explores the potential performance gains achievable with a novel architecture and full automation, suggesting a promising direction for future technical development. Second, our cohort was predominantly male, which aligns with the epidemiological prevalence of AAA but limits the model’s applicability to female patients, a population with recognized differences in rupture risk. Third, the retrospective design precluded the inclusion of genetic or novel biomarker data. Finally, while our explainability analyses provide strong, hypothesis-generating insights into the model’s decision-making, for instance, by highlighting the role of perivascular tissue, a comprehensive mapping of the full hierarchy of deep features to their specific biological correlates remains a key goal for future research. This includes linking these features to underlying biological processes such as specific inflammatory cell infiltrates or proteolytic activity levels. These considerations collectively outline a clear and logical path for future research. The initial performance achieved here justifies the next critical step: a prospective, multi-center validation study. Such an endeavor would be essential to confirm generalizability and would also provide a platform to integrate imaging-based predictions with multimodal data. Building upon the automated architecture presented here, future work that leverages large, diverse cohorts will be crucial to translating these technical advances into clinically reliable tools for personalized AAA management.

Conclusion

This study developed and validated a novel, end-to-end deep learning framework, ResNet50–YOLOv11–MedMamba, for predicting the growth of small asymptomatic AAAs. By integrating clinical data with automatically extracted imaging features from CTA, our model achieved a high predictive performance (98.75% accuracy, 97.78% F1-score) in internal validation, outperforming several classical benchmarks. The framework’s design, which leverages YOLOv11 for precise lesion localization and an 8-head MHA mechanism for effective feature fusion, demonstrates the feasibility of fully automating growth prediction. Future work, focused on multi-center prospective validation and the inclusion of biomechanical markers, is warranted to confirm these findings and pave the way for more personalized, image-based surveillance strategies.

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Ethics statement

The studies involving humans were approved by the Ethics Committee of General Hospital of Northern Theater Command with an Ethics Batch Number Y (2024)356. The studies were conducted in accordance with the local legislation and institutional requirements. The ethics committee/institutional review board waived the requirement of written informed consent for participation from the participants or the participants’ legal guardians/next of kin due to the retrospective design and the use of fully anonymized, de-identified data.

Author contributions

JC: Conceptualization, Investigation, Methodology, Visualization, Writing – original draft. ZZ: Data curation, Formal Analysis, Investigation, Writing – review and editing. YW: Data curation, Formal Analysis, Investigation, Writing – review and editing. YS: Resources, Software, Visualization, Writing – review and editing. NW: Resources, Software, Visualization, Writing – review and editing. XW: Conceptualization, Funding acquisition, Project administration, Supervision, Writing – review and editing. SW: Conceptualization, Project administration, Supervision, Writing – review and editing.

Funding

The author(s) declared that financial support was received for this work and/or its publication. This research article was supported by Science and Technology Plan Joint Plan Project of Liaoning Province (2025JH2/101800031).

Acknowledgements

The authors are grateful to the subjects who participated in the study and to the physicians’ assistance in this study.

Conflict of interest

The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declared that generative AI was not used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fphys.2025.1704428/full#supplementary-material

References

Aly G. H., Marey M., El-Sayed S. A., Tolba M. F. (2021). YOLO based breast masses detection and classification in full-field digital mammograms. Comput. Methods Programs Biomed. 200, 105823. doi:10.1016/j.cmpb.2020.105823

PubMed Abstract | CrossRef Full Text | Google Scholar

Benson T. W., Pike M. M., Spuzzillo A., Hicks S. M., Ali S., Pham M., et al. (2024). Soluble glycoprotein VI predicts abdominal aortic aneurysm growth rate and is a novel therapeutic target. Blood 144 (16), 1663–1678. doi:10.1182/blood.2023021655

PubMed Abstract | CrossRef Full Text | Google Scholar

Cersit S., Ocal L., Keskin M., Gursoy M. O., Kalcik M., Bayam E., et al. (2021). Association of C-Reactive protein-to-albumin ratio with the presence and progression of abdominal aortic aneurysm. Angiology 72 (2), 153–158. doi:10.1177/0003319720954084

PubMed Abstract | CrossRef Full Text | Google Scholar

Collaborators R., Bown M. J., Sweeting M. J., Brown L. C., Powell J. T., Thompson S. G. (2013). Surveillance intervals for small abdominal aortic aneurysms: a meta-analysis. Jama 309 (8), 806–813. doi:10.1001/jama.2013.950

PubMed Abstract | CrossRef Full Text | Google Scholar

Dong H., Raterman B., White R. D., Starr J., Vaccaro P., Haurani M., et al. (2022). MR elastography of abdominal aortic aneurysms: relationship to aneurysm events. Radiology 304 (3), 721–729. doi:10.1148/radiol.212323

PubMed Abstract | CrossRef Full Text | Google Scholar

Fernandez-Alonso S., Martinez-Aguilar E., Ravassa S., Orbe J., Paramo J. A., Fernandez-Alonso L., et al. (2022). Hemostatic biomarkers and volumetry help to identify high-risk abdominal aortic aneurysms. Life (Basel) 12 (6), 823. doi:10.3390/life12060823

PubMed Abstract | CrossRef Full Text | Google Scholar

Hechkel W., Helali A. (2025). Early detection and classification of Alzheimer’s disease through data fusion of MRI and DTI images using the YOLOv11 neural network. Front. Neurosci. 19, 1554015. doi:10.3389/fnins.2025.1554015

PubMed Abstract | CrossRef Full Text | Google Scholar

Herzog J., Skopp G., Musshoff F. (2024). Monitoring of phosphatidylethanol in dried blood spots and of ethyl glucuronide in hair over 6 months of alcohol consumption. Drug Test. Analysis 16 (4), 359–368. doi:10.1002/dta.3553

PubMed Abstract | CrossRef Full Text | Google Scholar

Isselbacher E. M., Preventza O., Black J. H., Augoustides J. G., Beck A. W., Bolen M. A., et al. (2022). 2022 ACC/AHA guideline for the diagnosis and management of aortic disease. J. Am. Coll. Cardiol. 80 (24), E223–E393. doi:10.1016/j.jacc.2022.08.004

PubMed Abstract | CrossRef Full Text | Google Scholar

Kim S., Jiang Z., Zambrano B. A., Jang Y., Baek S., Yoo S., et al. (2023). Deep learning on multiphysical features and hemodynamic modeling for abdominal aortic aneurysm growth prediction. IEEE Trans. Med. Imaging 42 (1), 196–208. doi:10.1109/tmi.2022.3206142

PubMed Abstract | CrossRef Full Text | Google Scholar

Klopf J., Fuchs L., Schernthaner R., Domenig C. M., Gollackner B., Brostjan C., et al. (2022). The prognostic impact of vascular calcification on abdominal aortic aneurysm progression. J. Vasc. Surg. 75 (6), 1926–1934. doi:10.1016/j.jvs.2021.11.062

PubMed Abstract | CrossRef Full Text | Google Scholar

Kontopodis N., Klontzas M., Tzirakis K., Charalambous S., Marias K., Tsetis D., et al. (2022). Prediction of abdominal aortic aneurysm growth by artificial intelligence taking into account clinical, biologic, morphologic, and biomechanical variables. Vascular 31 (3), 409–416. doi:10.1177/17085381221077821

PubMed Abstract | CrossRef Full Text | Google Scholar

Lee J. Y., Styczynski M. P. (2018). NS-kNN: a modified k-nearest neighbors approach for imputing metabolomics data. METABOLOMICS 14 (12), 153. doi:10.1007/s11306-018-1451-8

PubMed Abstract | CrossRef Full Text | Google Scholar

Lv R., Hu G., Zhang S., Zhang Z., Chen J., Wang K., et al. (2024). Assessing abdominal aortic aneurysm growth using radiomic features of perivascular adipose tissue after endovascular repair. Insights Imaging 15 (1), 232. doi:10.1186/s13244-024-01804-7

PubMed Abstract | CrossRef Full Text | Google Scholar

Oh S. J., Shin J. I., Kim E. N., Widiastini A., Hong Y., Sohn I., et al. (2025). Deep learning algorithm for predicting rapid progression of abdominal aortic aneurysm by integrating CT images and clinical features. Sci. Rep. 15 (1), 38413. doi:10.1038/s41598-025-22167-z

PubMed Abstract | CrossRef Full Text | Google Scholar

Oliver-Williams C., Sweeting M. J., Jacomelli J., Summers L., Stevenson A., Lees T., et al. (2019). Safety of men with small and medium abdominal aortic aneurysms under surveillance in the NAAASP. Circulation 139 (11), 1371–1380. doi:10.1161/CIRCULATIONAHA.118.036966

PubMed Abstract | CrossRef Full Text | Google Scholar

Rigotti N. A., Benowitz N. L., Prochaska J. J., Cain D. F., Ball J., Clarke A., et al. (2024). Cytisinicline for vaping cessation in adults using nicotine e-cigarettes: the ORCA-V1 randomized clinical trial. JAMA Intern. Med. 184 (8), 922–930. doi:10.1001/jamainternmed.2024.1313

PubMed Abstract | CrossRef Full Text | Google Scholar

Ristl R., Klopf J., Scheuba A., Wolf F., Funovics M., Gollackner B., et al. (2021). Growth prediction model for abdominal aortic aneurysms. Br. J. Surg. 109 (2), 211–219. doi:10.1093/bjs/znab407

PubMed Abstract | CrossRef Full Text | Google Scholar

Ristl R., Klopf J., Scheuba A., Sotir A., Wolf F., Domenig C. M., et al. (2023). Comparing maximum diameter and volume when assessing the growth of small abdominal aortic aneurysms using longitudinal CTA data: cohort study. Int. J. Surg. 109 (8), 2249–2257. doi:10.1097/js9.0000000000000433

PubMed Abstract | CrossRef Full Text | Google Scholar

Schanzer A., Solomon C. G., Oderich G. S. (2021). Management of abdominal aortic aneurysms. N. Engl. J. Med. 385 (18), 1690–1698. doi:10.1056/NEJMcp2108504

PubMed Abstract | CrossRef Full Text | Google Scholar

Spanos K., Eckstein H. H., Giannoukas A. D. (2020). Small abdominal aortic aneurysms are not all the same. Angiology 71 (3), 205–207. doi:10.1177/0003319719862965

PubMed Abstract | CrossRef Full Text | Google Scholar

Starck J., Brunkwall S., Lundgren F., Pärsson H., Gottsäter A., Holst J. (2025). Predictors of abdominal aortic aneurysm progression in men with small infrarenal aortic diameters at screening. J. Vasc. Surg. 81 (6), 1309–1318. doi:10.1016/j.jvs.2025.01.214

PubMed Abstract | CrossRef Full Text | Google Scholar

Thanigaimani S., Phie J., Quigley F., Bourke M., Bourke B., velu R., et al. (2022). Association of diagnosis of depression and small abdominal aortic aneurysm growth. Ann. Vasc. Surg. 79, 256–263. doi:10.1016/j.avsg.2021.06.038

PubMed Abstract | CrossRef Full Text | Google Scholar

Trenti C., Ziegler M., Bjarnegard N., Ebbers T., Lindenberger M., Dyverfeldt P. (2022). Wall shear stress and relative residence time as potential risk factors for abdominal aortic aneurysms in males: a 4D flow cardiovascular magnetic resonance case-control study. J. Cardiovasc Magn. Reson 24 (1), 18. doi:10.1186/s12968-022-00848-2

PubMed Abstract | CrossRef Full Text | Google Scholar

Ullery B. W., Hallett R. L., Fleischmann D. (2018). Epidemiology and contemporary management of abdominal aortic aneurysms. Abdom. Radiol. 43 (5), 1032–1043. doi:10.1007/s00261-017-1450-7

PubMed Abstract | CrossRef Full Text | Google Scholar

Vanmaele A., Bouwens E., Hoeks S. E., de Bruin J. L., Ten Raa S., Akkerhuis K. M., et al. (2025). The value of volume over maximum diameter for following abdominal aortic aneurysm growth and reducing surveillance visits in patients with a subthreshold aneurysm. J. Vasc. Surg. 82 (1), 102–110 e103. doi:10.1016/j.jvs.2025.03.395

PubMed Abstract | CrossRef Full Text | Google Scholar

Wang X. M., He B., Deng Y. S., Liu J. W., Zhang Z. H., Sun W. L., et al. (2022). Identification of a biomarker and immune infiltration in perivascular adipose tissue of abdominal aortic aneurysm. Front. Physiology 13, 977910. doi:10.3389/fphys.2022.977910

PubMed Abstract | CrossRef Full Text | Google Scholar

Wang Y., Xiong F., Leach J., Kao E., Tian B., Zhu C., et al. (2023). Contrast-enhanced CT radiomics improves the prediction of abdominal aortic aneurysm progression. Eur. Radiol. 33 (5), 3444–3454. doi:10.1007/s00330-023-09490-7

PubMed Abstract | CrossRef Full Text | Google Scholar

Wang L. X., Lin N. E., Chen W., Xiao H. Y., Zhang Y. Y., Sha Y. (2025). Deep learning models for differentiating three sinonasal malignancies using multi-sequence MRI. BMC Med. Imaging 25 (1), 56. doi:10.1186/s12880-024-01517-9

PubMed Abstract | CrossRef Full Text | Google Scholar

Wanhainen A., Van Herzeele I., Bastos Goncalves F., Bellmunt Montoya S., Berard X., Boyle J. R., et al. (2024). Editor's choice -- European society for vascular surgery (ESVS) 2024 clinical Practice Guidelines on the management of abdominal aorto-iliac artery aneurysms. Eur. J. Vasc. Endovascular Surg. 67 (2), 192–331. doi:10.1016/j.ejvs.2023.11.002

PubMed Abstract | CrossRef Full Text | Google Scholar

Yue Y., Li Z. (2024). Medmamba: vision mamba for medical image classification. arXiv Preprint arXiv:2403.03849. doi:10.48550/arXiv.2403.03849

CrossRef Full Text | Google Scholar

Zhang S., Gu H., Chang N., Li S., Xu T. Q., Liu M. H., et al. (2023). Assessing abdominal aortic aneurysm progression by using perivascular adipose tissue attenuation on computed tomography angiography. Korean J. Radiology 24 (10), 974–982. doi:10.3348/kjr.2023.0339

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhao Y., Zhang J. H., Li H. J., Wang Q. Y., Li Y. G., Wang Z. T. (2025). Automatic skeletal maturity grading from pelvis radiographs by deep learning for adolescent idiopathic scoliosis. Med. & Biol. Eng. & Comput. 63 (6), 1571–1583. doi:10.1007/s11517-025-03283-4

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhu C., Leach J. R., Wang Y., Gasper W., Saloner D., Hope M. D. (2020). Intraluminal Thrombus predicts rapid growth of abdominal aortic aneurysms. Radiology 294 (3), 707–713. doi:10.1148/radiol.2020191723

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: multi-head self-attention, computed tomography angiography, growth prediction, deep learning, abdominal aortic aneurysm

Citation: Cheng J, Zhang Z, Wang Y, Sun Y, Wang N, Wang X and Wang S (2026) Predicting the growth of asymptomatic small abdominal aortic aneurysms (AAA) based on deep learning. Front. Physiol. 16:1704428. doi: 10.3389/fphys.2025.1704428

Received: 13 September 2025; Accepted: 03 December 2025;
Published: 27 January 2026.

Edited by:

Vladimir Tadic, Technical College of Applied Sciences, Serbia

Reviewed by:

Mirzat Turhon, Capital Medical University, China
Wenqing Yuan, Chongqing Medical University, China

Copyright © 2026 Cheng, Zhang, Wang, Sun, Wang, Wang and Wang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Xiaozeng Wang, d3hpYW96ZW5nQDE2My5jb20=; Sihan Wang, d2FuZ3NpaGFuQHNpYS5jbg==

ORCID: Jiaxin Cheng, orcid.org/0009-0009-6918-688X; Xiaozeng Wang, orcid.org/0000-0003-1979-0287; Sihan Wang, orcid.org/0009-0004-6486-5924

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.