- 1Department of Ultrasound, Union Hospital, Tongji Medical College, Huazhong University of Science and Technology, Wuhan, China
- 2Department of Pathology, Union Hospital, Tongji Medical College, Huazhong University of Science and Technology, Wuhan, China
Background: Accurate assessment of axillary lymph node status is essential for the management of breast cancer. Recent advancements in deep learning (DL) have shown promising results in medical image analysis. This study aims to develop a multimodal DL model that integrates preoperative ultrasound images and hematoxylin and eosin (H&E)-stained core needle biopsy pathology images of primary breast cancer to predict axillary lymph node metastasis (ALNM).
Materials and methods: This study included 211 patients with histologically confirmed breast cancer, conducted between February 2023 and March 2024. For each patient, one ultrasound image and one histopathological image of the primary breast cancer lesion were collected. Various DL architectures were applied to extract tumor features from the ultrasound and histopathology images, respectively. Multiple fusion strategies, combining features from both ultrasound and pathology images, were developed to enhance the comprehensiveness and accuracy of predictions. The performance of the single-modality models, multi-modality models, and different fusion strategies were compared. Evaluation metrics included precision, accuracy, recall, F1-score, and area under the curve (AUC).
Results: PLNeT and ULNet were identified as the most effective feature extractors for histopathological and ultrasound image analysis, respectively. Overall, the multilayer fusion model outperformed single-modality models in predicting ALNM, achieving an accuracy of 0.7353, precision of 0.7344, recall of 0.7576, F1-score of 0.7463, and AUC of 0.7019.
Conclusion: Our study provides a multilayer fusion strategy using ultrasound and pathology images of the primary tumor to predict ALNM in breast cancer patients. Although achieving suboptimal performance, this model has the potential to determine appropriate axillary treatment options for patients with breast cancer.
1 Introduction
Breast cancer (BC) is the most common malignant tumor among women globally and has become the second leading cause of cancer-related mortality in this population, with a reported mortality of approximately 15% (1). Axillary lymph node metastasis (ALNM) is one of the most significant predictors of overall recurrence and survival in breast cancer patients (2), with each positive lymph node associated with an increased risk of death by approximately 6% (3). Accurately assessing whether axillary lymph nodes are metastatic is a critical step in the management of breast cancer (4, 5). Misdiagnosis of axillary lymph nodes can lead to missed surgical opportunities and may increase unnecessary surgical trauma and complications (6). Consequently, precise prediction of lymph node status in breast cancer patients is essential for effective management of surgical strategies.
Currently, the assessment of axillary lymph nodes primarily relies on imaging studies (such as ultrasound and MRI) and sentinel lymph node biopsy (SLNB) (7, 8). While imaging can provide valuable insights regarding metastasis, its accuracy is typically limited by operator experience, ranging from 70% to 85%, which can lead to misdiagnosis and delayed treatment (9). SLNB, as a minimally invasive procedure, can enhance accuracy to over 90%, but it may also overlook metastases in non-sentinel lymph nodes (6). Additionally, there are potential side effects, including lymphedema and upper limb numbness (10). Therefore, there is an urgent need to identify a reliable and effective alternative method for accurately assessing the axillary lymph node status in breast cancer patients, facilitating informed decisions regarding axillary management.
Radiomics, through high-throughput extraction of numerous features from medical images, has been applied in the prediction of ALNM in breast cancer patients (11–13). Some studies have constructed deep learning (DL) models based on clinicopathological data, ultrasound or MRI images, aimed at improving the detection accuracy of ALNM, with the area under the curve (AUC) ranging from 0.74 to 0.89 (14–16). However, as single-modal models, these methods have certain limitations, including information constraints (which may not comprehensively capture the biological characteristics of the lesions), lack of diversity (where different patients may exhibit similar imaging features despite significant differences in their pathological status), and insufficient model validation (with a lack of external validation data potentially affecting the generalizability of the results). With the rapid development of deep learning technology, pathological information has also gradually been integrated into high-throughput analysis. Xu et al. studied a cohort of early breast cancer patients who underwent preoperative core needle biopsy (CNB) and found that DL models based on primary tumor biopsy slices could effectively refine the prediction of ALNM, which achieved an AUC of 0.816 (17). This finding underscores the importance of pathological slices in predicting ALNM. Multimodal fusion model contains more comprehensive information than single-modal models. For instance, Bove et al. developed a model that combined clinical and radiomic features provided the best performances to predict the nodal status in clinically negative breast cancer patients, achieving an AUC value of 0.886 (18). If radiomics, which focuses on macro-level imaging features, can be combined with histopathology, which addresses micro-level characteristics, this multimodal approach will facilitate the integration of diverse information, thereby further enhancing the accuracy of predictions.
Therefore, we hypothesize that the combination of breast ultrasound and puncture pathology analysis may yield encouraging results in preoperatively distinguishing and predicting ALNM. In this research, we seek to establish a predictive model for ALNM in breast cancer using a multimodal fusion strategy, incorporating pre-treatment ultrasound images and hematoxylin and eosin (H&E)-stained core needle biopsy pathology of primary lesion.
2 Materials and methods
2.1 Patients
A retrospective collection of imaging and clinical data was conducted for patients diagnosed with breast cancer via CNB at Union Hospital of Tongji Medical College of Huazhong University of Science and Technology between February 2023 and March 2024. The pathological diagnosis methods for axillary lymph nodes included ultrasound-guided CNB, sentinel lymph node biopsy (SLNB), and axillary lymph node dissection (SLND). The inclusion criteria were: (1) availability of preoperative ultrasound images of the breast and axillary lymph nodes from our hospital, with complete imaging; (2) ultrasound-guided CNB performed following breast ultrasound examination; (3) breast surgery conducted 1–2 weeks after the biopsy. The exclusion criteria were: (1) patients with a history of contralateral breast cancer or who had undergone preoperative chemotherapy or radiotherapy; (2) patients with bilateral breast cancer or multiple lesions that made target delineation difficult; (3) tumors that were excessively large, exceeding the measurable range of the ultrasound probe. For each patient, one ultrasound image and one histopathological image of the primary breast cancer lesion were obtained.
This study was approved by the Ethics Committee of Union Hospital at Huazhong University of Science and Technology, and the methods were applied in accordance with the approved guidelines. Informed consent was obtained from all patients.
2.2 Data collection
2.2.1 Ultrasound examination
Breast ultrasound examinations were performed by one of ten experienced radiologists following standard practice protocols. Several ultrasound devices from manufacturers such as GE Healthcare, Mindray, and Philips were equipped with linear transducers (frequency range of 10–18 MHz) to capture breast and ultrasound images. Unlabeled two-dimensional images stored in the maximum longitudinal plane of the breast lesions were utilized for subsequent analysis. Additionally, axillary ultrasound reports for each patient were recorded and categorized as either suspicious or unsuspicious. Suspicious sonographic features of axillary lymph node metastasis (ALNM) included irregular morphology, absence of the fatty hilum, eccentric cortical thickening (>3 mm), and microcalcifications (7, 8).
2.2.2 Biopsy and pathological image collection
Core biopsy samples were obtained using a 16-gauge hollow needle under ultrasound guidance. Each lesion was punctured two to three times, with harvested tissue immediately fixed in 10% neutral buffered formalin for 6–72 hours. Surgical specimens were directly sampled from cancerous lesions and fixed for 24–72 hours. All tissues were subsequently embedded in paraffin, sectioned at 5 μm intervals, and stained with hematoxylin and eosin (H&E). Following the methodology described by Yang et al. for oral squamous cell carcinoma analysis (19), two senior breast pathologists evaluated the H&E slides and selected representative tumor regions for microscopic imaging independently. Image acquisition was performed at 10× objective (NA=0.30) and 10× eyepiece magnification using Olympus BX53 and Nikon Eclipse Ni microscopes equipped with DP27 digital cameras (2048 × 1536 pixel resolution). Spatial resolution was calibrated to 1.25 μm per pixel using a NIST-traceable stage micrometer, with each image covering a 2.0 × 1.5 mm tissue region. Representative region selection followed established pathological protocols requiring ≥30% tumor cellularity and avoidance of necrosis/artifact zones.
2.2.3 Clinicopathological characteristics
Clinical data recorded included patient age, menopausal status and tumor location. Histological type and immunohistochemical results for estrogen receptor (ER) status, progesterone receptor (PR) status, human epidermal growth factor receptor 2 (HER2) status, and Ki-67 levels were also documented based on the entire tumor surgical specimen. Ki-67 positive was defined as proliferation index at least 14% (20). All BC patients were categorized into three molecular subtypes according to their immunohistochemical results: Luminal (ER and/or PR-positive), HER2 overexpression (ER and PR-negative, HER2-positive) and triple-negative breast carcinoma (TNBC, ER, PR, and HER2-negative).
2.3 Image data processing
2.3.1 Manual annotation of ROI
This study required further processing of paired ultrasound images for each lesion to ensure the model focused on critical feature areas. Specifically, one ultrasound physician with over 10 years of clinical experience manually annotated the regions of interest (ROIs) in the ultrasound images. Utilizing their extensive clinical expertise, this physician accurately identified and localized the tumor areas, subsequently drawing the smallest bounding rectangles that encompassed the tumor regions. Another ultrasound physician, also with over 10 years of clinical experience, was responsible for the review process. This manual annotation procedure strictly adhered to a standardized annotation protocol to ensure consistency and accuracy in the annotations.
2.3.2 Data preprocessing
The data preprocessing in this study involved the processing of pathological images and ultrasound data, with the aim of providing standardized data input for subsequent model training. Pathological images underwent data cleaning and preprocessing. All pathological images were uniformly resized to 224x224 pixels to meet the input requirements of the deep learning model. Data augmentation techniques were employed to increase image diversity, primarily involving random resizing and cropping (RandomResizedCrop) and random horizontal flipping (RandomHorizontalFlip) to simulate different angles and positions of capture. Additionally, the image data were standardized to a distribution with a mean of 0.5 and a standard deviation of 0.5 to enhance the stability and generalization capability of model training.
In the preprocessing of ultrasound images, a similar pipeline data augmentation was followed, including resizing all images to 224×224 pixels and applying random cropping and horizontal flipping for data augmentation. Vertical flipping was intentionally avoided due to anatomical orientation constraints. To address the inherent heterogeneity caused by the use of ultrasound scanners from different manufacturers—namely GE Healthcare, Mindray, and Philips—we implemented additional harmonization techniques aimed at reducing inter-scanner variability. Specifically, we applied z-score normalization to standardize the intensity distribution of each image, followed by histogram equalization to align contrast and grayscale across different devices. These procedures were intended to minimize variations stemming from differences in hardware resolution and signal processing. Additionally, mask-guided annotation was utilized to highlight tumor-specific regions, enabling the model to concentrate on lesion-relevant features while reducing the influence of peripheral artifacts and device-specific noise. By combining normalization, histogram alignment, and spatial focus through masking, the ultrasound images were standardized to better support consistent feature learning across multi-source inputs.
2.4 Construction of deep learning model
This study established a multimodal deep learning model to perform feature extraction and fusion on ultrasound imaging data and pathological image data, aiming to predict the likelihood of axillary lymph node metastasis in breast cancer.
2.4.1 Construction of imaging model
The imaging model utilized classic convolutional neural network (CNN) architectures, ResNet50 and DenseNet, to extract tumor features from ultrasound images. After data preprocessing, the input ultrasound images were first processed through a series of convolutional and pooling layers to extract embedding representations containing high-level features. To prevent model overfitting, a Dropout layer (with a value set to 0.5) was applied before the fully connected layer of the network. During the model training process, to accelerate convergence and improve accuracy, transfer learning was employed using pretrained weights from ResNet50 and DenseNet models. The learning rate was set to 0.001 and gradually decayed by half every 10 epochs. Model parameters were optimized using the Adam optimizer, chosen for its adaptive learning rate adjustment and stable convergence, especially when training from pretrained backbones with moderate data sizes. We designate the name ULNet to represent the best-performing standard networks for ultrasound image analysis. ULNet corresponds to a modified ResNet50 architecture with strategic adaptations for ultrasound images analysis, selected due to its strong performance in image-based classification tasks and proven ability to capture spatial and edge-related features—especially suitable for ultrasound’s texture and boundary information.
2.4.2 Construction of pathological model
Due to the high-resolution features of pathological images, more detailed feature extraction methods are required. In this study, the Vision Transformer (ViT) and PLNet models were selected, as they are suitable for processing high-resolution images. ViT segments the pathological images into 16x16 image patches and feeds these patches into multiple layers of Transformer encoders to obtain a more global feature representation. It should be noted that PLNet is not a novel architecture, but rather a naming convention adopted to represent the best-performing standard networks for pathological images. PLNet is a lightweight CNN architecture optimized for pathology images, featuring four convolutional blocks with progressive filter expansion and global pooling. It was selected over more complex architectures like ViT due to its better balance between global context and convergence stability on high-resolution image data. The input pathological images were normalized and resized to 224x224 pixels, and features were extracted using both the ViT and PLNet models, resulting in embeddings that represent the pathological images. During training, the learning rate for the ViT model was set at 3e-5, while for PLNet it was set at 1e-4, with both utilizing the Adam optimizer. Batch Normalization layers were incorporated during feature extraction to eliminate channel discrepancies and enhance the model’s robustness on pathological images.
2.4.3 The strengths and weaknesses of ULNet and PLNet models
In terms of strengths, ULNet offers deep feature hierarchies and residual learning that mitigate vanishing gradients, while PLNet excels in efficient representation learning with fewer parameters, particularly helpful under limited data conditions. A noted limitation is that ULNet may be sensitive to image artifacts or probe differences, and PLNet, as a simpler CNN, may underperform on highly heterogeneous pathology data unless supported by proper normalization and augmentation.
2.4.4 Construction of multimodal fusion model
The multimodal fusion model combines features from both ultrasound images and pathological images to enhance the comprehensiveness and accuracy of model predictions. Initially, feature embeddings are extracted separately through the imaging and pathological models, followed by the fusion of information from the two modalities at the feature level. Various fusion strategies were explored, including concatenation, weighted fusion, attention mechanisms, and multilayer fusion. In the multilayer fusion strategy, the fusion module operates as follows: First, each modality-specific feature vector is individually processed through batch normalization layers to reduce scale discrepancies. The normalized features are then concatenated into a single vector. This combined vector is passed through a fully connected (dense) layer with 256 neurons, followed by ReLU activation. A subsequent Batch normalization layer and Squeeze-and-Excitation (SE) module are applied to reweight channel-wise dependencies and enhance feature importance adaptively. The output is then fed into a final classification head consisting of a single neuron with sigmoid activation for binary classification. All activation functions used are standard ReLU (except for the final sigmoid), and dropout was not applied within the fusion module to retain feature integrity. The learning rate for the multimodal fusion model was set at 1e-4, utilizing the AdamW optimizer, with the learning rate decayed every 5 epochs to stabilize the fusion effect.
For the multimodal fusion model, we employed AdamW, an improved variant that decouples weight decay from gradient updates. This choice was based on the empirical observation that AdamW offers more robust generalization in joint optimization settings, where fused features from two distinct modalities may introduce higher gradient variance and overfitting risks. AdamW’s stronger regularization capability helped stabilize fusion training and improve convergence without extensive fine-tuning. The detailed DL flowchart is illustrated in Figure 1.

Figure 1. Flowchart of deep learning model to predictive axillary lymph node metastasis in breast cancer.
2.5 Experimental setup
In the experimental setup of this study, to ensure an effective evaluation of model performance, the dataset was randomly divided into training and testing sets at a ratio of 8:2. The model training utilized the AdamW optimizer, and hyperparameter tuning was performed through grid search, including parameters such as learning rate and batch size, to ensure optimal model performance. During training, the initial learning rate was set to 0.001, with a reduction by half every 5 epochs to accommodate the convergence requirements of the model. Model training and validation were conducted on an NVIDIA RTX 3090 GPU cluster, with each GPU equipped with 24GB of memory, utilizing a total of 5 GPUs, and the PyTorch deep learning framework for model construction and training. The programming language used was primarily Python 3.8, with the experimental environment comprising the Ubuntu 20.04 operating system, supporting CUDA 11.1 to accelerate GPU computations. Key Python libraries required for the experiments included torch, torchvision, numpy, pandas, and scikit-learn. Additionally, to ensure the reproducibility of the experiments, the random seed was uniformly set during both data splitting and model training processes.
2.6 Model evaluation
For model evaluation, we employed multiple metrics to comprehensively assess model performance, including accuracy, precision, recall, and F1-score, to evaluate the classification effectiveness of the model in predicting axillary lymph node metastasis. For the binary classification task (metastatic vs. non-metastatic), a confusion matrix was used to analyze the model’s classification capabilities and misclassification instances. Furthermore, to further validate the model’s discriminative ability, we plotted the ROC curve and calculated the AUC value, with an AUC value closer to 1 indicating better discriminative performance. To assess the enhancement effect of multimodal fusion on the model, we compared the performance differences between the unimodal (imaging or pathological) models and the fusion model, focusing on improvements in recall and accuracy.
Statistical analysis of experimental results was conducted using SPSS software (V.27.0). Continuous variables were compared using either a student’s t-test or the Mann-Whitney U test, depending on the normality of the distribution. These variables are presented as mean ± standard deviation (SD) or median (interquartile range), as appropriate. The chi-square test was used to assess differences between categorical variables, which are presented as frequencies and percentages. A p-value of < 0.05 was considered statistically significant.
3 Results
3.1 Clinicopathological characteristics of patients
This study enrolled a total of 211 female BC patients between February 2023 and March 2024, with a mean age of 53.1 ± 11.7 years (ranging from 23 to 85 years). All participants underwent axillary lymph node biopsy and/or dissection, which identified ALNM in 107 patients, while the remaining 104 cases were node-negative. Patients were categorized into two groups according to axillary lymph node status. Table 1 summarizes the demographic and clinicopathological characteristics of the study population. Univariate analysis showed no significant differences between the two groups in terms of age, menopausal status, tumor location, histological types, hormone receptor status, or molecular subtypes (P > 0.05). However, significant differences were observed in the Ki-67 index, HER2 status, and preoperative axillary ultrasound findings (P < 0.05). Axillary ultrasound demonstrated a true positive rate of 52.3% (56/107), a false negative rate of 47.7% (51/107), a true negative rate of 90.4% (94/104), and a false positive rate of 9.6% (10/104).
3.2 Predictive performance of different DL models
3.2.1 Pathological model
To assess the predictive performance of DL models using histopathological images of primary breast cancer, we evaluated several architectures, including ResNet50, DenseNet, GoogLeNet, SqueezeNet, ViT, and PLNet. The results, summarized in Table 2, indicate that PLNet exceeds the other five architectures across almost all metrics, except for a slightly lower recall value. The confusion matrices and ROC curves for all six models in predicting ALNM based on histopathological images are shown in Figure 2. Based on these findings, we selected PLNet as the primary DL architecture for histopathological images analysis in this study, with an AUC of 0.7195, accuracy of 0.7186, precision of 0.7188, recall of 0.6667, and F1-score of 0.7091.

Table 2. Predictive performance of deep learning models based on histopathological images for predicting ALN metastasis.

Figure 2. Receiver operating characteristic curves and confusion matrix of six different models using histopathological images for predicting axillary lymph node metastasis in breast cancer. 0: negative; 1: positive. (a–f) represent ResNet50, DenseNet, GoogLeNet, SqueezeNet, ViT, and PLNet model, respectively.
3.2.2 Imaging model
We trained six distinct models to extract DL features from primary breast cancer ultrasound images for the prediction of ALNM. As shown in Table 3, the ULNet model outperforms other models in terms of AUC (0.6979), precision (0.6875), recall (0.7488) and F1-score (0.7436). It achieved the best overall performance and was therefore selected as the primary DL architecture for ultrasound image analysis. Compared to the PLNet model based on histopathological images, the ULNet model, which is based on ultrasound images, demonstrated lower AUC, accuracy, and precision. However, the ULNet model improved the recall to 0.7488, demonstrating an improvement in recall after incorporating mask information. The confusion matrices and ROC curves for all six models are presented in Figure 3.

Table 3. Predictive performance of deep learning models based on ultrasound images for predicting ALN metastasis.

Figure 3. Receiver operating characteristic curves and confusion matrix of six different models using ultrasound images for predicting axillary lymph node metastasis in breast cancer. 0: negative; 1: positive. (a–f) represent ResNet50, DenseNet, GoogLeNet, SqueezeNet, ViT, and ULNet, respectively.
It’s worth noting that the SqueezeNet model produced a recall and F1-score of 0 in both the ultrasound-based and pathology-based tasks, indicating that the model failed to correctly identify any breast cancer patients with ALNM. We attribute this failure to three interdependent factors: 1) Representational capacity deficit: The model’s extreme parameter compression (1.24M parameters vs. ResNet50’s 25.6M) critically limits its ability to capture histo-morphological discriminators of metastasis (e.g., micrometastases in capsule-distorted lymph nodes); 2) Feature space imbalance: Metastatic manifestations occupy sparse regions in the feature space, requiring specialized architectural components (e.g., attention mechanisms) absent in SqueezeNet; 3) Gradient dissipation: Vanishing gradients during backpropagation prevent effective weight updates. SqueezeNet, being an extremely lightweight architecture originally designed for resource-constrained environments, may lack sufficient capacity to capture the complex and high-dimensional patterns in medical imaging data-especially in the context of relatively subtle differences between metastatic and non-metastatic lymph node profiles. We intentionally chose to include SqueezeNet in our comparison to illustrate the limitations of overly simplified networks when applied to challenging clinical prediction tasks.
3.2.3 Multimodal fusion model
We developed a multimodal approach that integrates features from both ultrasound and histopathological images of primary breast cancer to predict ALNM. The ULNet and PLNet were chosen as the base DL models for analyzing ultrasound and histopathological images, respectively. We experimented with multiple fusion strategies, including concatenation, weighting, attention, voting, and multilayer fusion. Among these, the multilayer-fusion model delivered the best performance, with an accuracy of 0.7353, precision of 0.7344, recall of 0.7576, and an F1-score of 0.7463 (Table 4). Performance was assessed using the ROC curve and a confusion matrix (Figure 4). Compared to single-modality models (based on either imaging or pathology), the multilayer-fusion model demonstrated an AUC of 0.7019, which was comparable to or slightly lower than that of the PLNet model (AUC = 0.7195). The combination of ultrasound and histopathological images does not improve the AUC, which may be due to factors such as architectural constraints or label noise. However, the multilayer-fusion model outperformed the single-modality models in terms of accuracy, precision, recall, and F1-score. when all performance metrics were considered, the multilayer-fusion model surpassed single-modality models based solely on ultrasound or histopathological images, demonstrating its effectiveness for predicting ALNM.

Table 4. Predictive performance of an integrated deep learning model based on ultrasound and histopathological images for predicting ALN metastasis.

Figure 4. Receiver operating characteristic curves and confusion matrix of 5 different models using ultrasound and histopathological images for predicting axillary lymph node metastasis in breast cancer. 0: negative; 1: positive. (a–e) represent concatenation, weighting, attention, voting, and multilayer fusion, respectively.
3.3 Modal visualization of ALNM prediction
We employed a heatmap visualization technique for the multilayer-fusion model to assess the confidence of its predictions, offering an intuitive view of model focus areas. The heatmap highlights regions activated by the model that are most indicative of lymph node metastasis, with blue and red areas representing regions that received higher attention and carry the greatest predictive significance. The deeper the color, the higher the likelihood of predicting ALNM.
To pathologically validate our model’s attention mechanisms, two senior breast pathologists (>10 years specialty experience) conducted a blinded retrospective analysis of 50 randomly selected cases. Following standardized diagnostic protocols, they independently evaluated regions highlighted by the multilayer-fusion heatmaps against established histopathological criteria. High-attention areas were systematically assessed for correlation with clinically significant features including invasive carcinoma nests demonstrating aggressive morphological patterns (e.g., irregular infiltrative borders, desmoplastic stromal reactions) and dense collagenous matrix formations associated with metastatic propensity. The pathologists confirmed consistent spatial correspondence between model attention foci and these high-risk histological characteristics, particularly noting alignment in zones exhibiting lymphatic tumor emboli and peritumoral fibroblast proliferation. This qualitative concordance substantiates that our model prioritizes morphologically aggressive tumor phenotypes in its decision-making process. To illustrate these findings, we have incorporated representative pathological-thermal correlations with annotation markers in Figure 5, demonstrating the clinical relevance of attention mechanisms in identifying high-risk tumor regions.

Figure 5. Representative original histopathological images and heatmaps. Histopathological images of primary breast with annotation markers (left column) and their corresponding attention heatmaps (right column), the darker the feature color, the higher the attention of the model. (a) The marked regions on the H&E slides represent the tumor-stroma interface, containing both apparently invasive carcinoma and abundant collagenous stroma (left column); corresponding heatmaps are shown in the right column. (b) The marked regions on the H&E slides represents invasive carcinoma nests (left column), with their corresponding heatmaps shown in the right column. (c) The image in the left column represents H&E slides from a breast cancer patient without axillary lymph node metastasis, corresponding heatmaps are shown in the right column.
4 Discussion
In recent years, DL has been applied to breast cancer diagnosis, risk stratification, prognosis prediction, and treatment response assessment (21–23). Common DL network architectures, including CNN, recurrent neural network, and ViT, have been explored for these tasks. Among them, ViT has demonstrated state-of-the-art performance on several image classification datasets (24). To address the insufficient diagnostic accuracy of single-modality imaging in cancer, researchers have developed DL-based multimodal fusion models that leverage multiple data types commonly found in medical records. For instance, Ishak Pacal et al. developed InceptionNeXt-Transformer, a novel hybrid DL architecture combining CNNs and ViTs for multi-modal breast cancer image analysis (25). A hybrid deep learning model that integrated InceptionNeXt blocks, enhanced Swin Transformer blocks, and a Residual Multi-Layer Perceptron for automated colorectal cancer detection (26). Additionally, the same team designed XtBrain, another novel hybrid architecture that combined local and global feature learning for brain tumor classification. XtBrain leverages the NeXt Convolutional Block and the NeXt Transformer Block synergistically to enhance feature learning (27). Suat Ince et al. proposed a novel U-Net architecture enhanced with ConvNeXtV2 blocks and GRN-based Multi-Layer for cerebral vascular occlusion segmentation (28).
In this study, we developed a multimodal fusion model integrating ultrasound and histopathology images to predict ALNM in breast cancer. HE-stained biopsy pathology images provide micro-level information on tumor tissue and cellular structures, though limited by sampling scope, while ultrasound captures macro-level tissue characteristics. The complementary integration of these data using a fusion model offers a more comprehensive view of tumor biology. Huang et al. reported an AUC of 0.900 (95% CI 0.819–0.953) for early breast cancer subtype differentiation using a fusion model of preoperative ultrasound and whole-slide images, highlighting the advantages of image fusion over single-modality models (29). Our results showed that, compared to single-modality models and other fusion strategies, the multilayer fusion model achieved better predictive performance (accuracy: 0.7353, precision: 0.7344, recall: 0.7576, F1-score: 0.7463, AUC: 0.7019). Our findings further validate the benefits of integrating preoperative ultrasound with biopsy pathology images. Although the multimodal fusion model achieved an AUC of 0.7019, indicating modest diagnostic performance. It serves as a complementary tool for guiding personalized treatment decisions. Future work is needed to optimize the model. Complementary to our multimodal imaging approach, Bove et al. demonstrated that combining clinical parameters with radiomic features from ultrasound images significantly improves nodal status prediction in clinically negative patients (18). While their methodology leverages handcrafted radiomic features and clinical integration, our deep learning-based fusion of raw ultrasound and histopathology images represents a distinct technical pathway toward the shared goal of refining axillary management decisions. Both studies underscore the critical value of multidimensional data integration in breast cancer staging.
Axillary lymph node involvement typically indicates a poorer prognosis and higher recurrence risk in breast cancer patients (8). Thus, preoperative accurate assessment of axillary lymph node status helps clinicians devise appropriate axillary treatment strategies, reducing postoperative complications and improving outcomes. Axillary ultrasound is commonly recommended as part of the standard diagnostic workup for patients with invasive breast cancer (30). However, its diagnostic accuracy is limited by high operator dependency. In our study, axillary ultrasound demonstrated a true positive rate of 52.3% and a false negative rate of 47.7%. This false negative rate is notably higher than that reported in previous studies, which range from 23.4% to 25% (31, 32), potentially leading to the underdiagnosis of patients with clinically significant nodal metastasis. Our multilayer fusion model presents an alternative approach for detecting ALNM in breast cancer patients, offering clinicians an alternative tool to inform treatment decisions. For patients without evidence of nodal involvement, axillary surgery could potentially be avoided, whereas sentinel lymph node dissection (SLND) or axillary lymph node dissection (ALND) is typically performed in cases with confirmed ALNM (33). Although ALND remains the gold standard for diagnosing axillary lymph node status and preventing axillary recurrence, it is associated with significant complications, including lymphedema, restricted mobility, and sensory abnormalities (33). Over the past few decades, axillary management has become less invasive, with SLND largely replacing ALND, thereby minimizing physical harm. The Z0011 trial demonstrated that for patients with one or two positive sentinel nodes, survival outcomes are comparable between those undergoing SLND alone and those undergoing ALND (34). As research increasingly emphasizes individualized and minimally invasive treatments (35, 36), it is crucial to accurately assess axillary lymph node status preoperatively, particularly in cases where clinical and imaging examinations show no suspicious signs of ALNM. SLND could be avoided if reliable preoperative evaluation of ALN status were available.
In our study, the deep learning model based on ultrasound achieved an AUC of 0.6979, while the histopathology-based model achieved an AUC of 0.7195, both of which are lower than those reported in previous studies (15, 17), likely due to the smaller sample size. In clinical practice, multi-modal imaging offers more complementary information compared to single-modal imaging. The multimodal fusion model integrating ultrasound and histopathological images achieved an AUC of 0.7019 - comparable to, or slightly lower than that of the histopathology-based model, suggesting no significant diagnostic improvement. However, in medical imaging, evaluating model performance based solely on the AUC value is insufficient. Other critical metrics, such as precision, accuracy, recall, and F1-score, must also be considered. When evaluating these comprehensive performance metrics, the multilayer fusion model outperformed all single-modality models in our study, demonstrating improvements in precision, accuracy, recall, and F1-score. Specifically, the multilayer fusion model showed superior precision and accuracy, indicating that its positive predictions were primarily true positives, and that it achieved a higher overall correct prediction rate, with fewer false positives and false negatives. In terms of recall, the multilayer fusion model also exhibited superior performance, identifying more true positive cases and thereby reducing the risk of missed diagnoses. Importantly, the F1-score, which balances precision and recall, was significantly higher for the multilayer fusion model, underscoring its advantage in comprehensively evaluating positive cases and minimizing missed diagnoses. These findings suggest that the multilayer-fusion model is proposed not solely based on AUC, but on its balanced and consistent advantage across multiple performance dimensions, demonstrating its comprehensive advantage in predicting ALNM. In our research, the multilayer-fusion model based on ultrasound and histopathological images does not improve the AUC, which may be due to factors such as architectural constraints or label noise. Future work will employ cross-modal attention mechanisms to better isolate complementary features across modalities. This could include incorporating more advanced models, such as EfficientNetV2 and Swin Transformer, to validate performance improvements (37).
In our multilayer fusion model strategy, a Squeeze-and-Excitation (SE) attention mechanism was incorporated to adaptively recalibrate feature channel weights, thereby enhancing the representation of critical features while suppressing irrelevant or noisy features. This mechanism increases the model’s sensitivity to key features, thus improving classification accuracy (38). After feature scaling, individual model features were concatenated, ensuring a more uniform feature value distribution and reducing instability during training due to large feature value discrepancies. To address significant feature differences across modalities, batch normalization (BN) layers were applied to both models before feature fusion. Given the substantial variation in the distributions of ultrasound and histopathology features, BN layers effectively standardize the features, making the fusion process smoother and more stable (39). After integration, the recall and accuracy of the fusion model were improved compared to the unimodal (imaging or pathological) models. The improvements in recall (likewise recognized as sensitivity) with our fusion model suggest it may minimize missed diagnoses of lymph node metastasis, proving particularly valuable for identifying high-risk patients. By providing reliable risk assessments, the fusion model can support clinical decisions on further invasive evaluation or personalized treatment planning, underscoring its potential as a clinical decision support tool.
To enhance the interpretability of our multilayer fusion model, we employed heatmap visualization to highlight the region’s most influential in the model’s predictions. For ultrasound images, the key regions for distinguishing ALNM in breast cancer were found to be the tumor boundary and the low-echo areas within the tumor. In histopathological slides, the regions exhibiting prominent invasive cancer cells and dense collagen stroma were identified as the most predictive features. These findings are consistent with previous studies that have linked ultrasonographic characteristics—such as tumor size, echogenicity, and lesion boundary—as well as histological features like grading and lymphatic vascular invasion, to axillary lymph node involvement in primary breast cancer (40–44). Additionally, our study highlighted that high Ki-67 expression and positive HER2 status may also serve as potential risk factors for ALNM, aligning with prior research (45–47). High Ki-67 expression, which reflects the proliferation activity of cancer cells, has been established as a significant predictor of ALNM (48, 49). Collectively, these findings further support the clinical reliability of our model’s predictions, providing a framework for predicting ALNM in breast cancer patients.
This study has some limitations. First, the small sample size increases the risk of overfitting and may affect the model’s generalizability. The model’s performance may not hold up in a larger, more diverse patient cohort. Future research should involve larger and more diverse cohorts to enhance the model’s robustness across different patient demographics, tumor types, and pathological stages. Second, selection bias is an inherent limitation of this retrospective research. The stability, generalizability, and clinical utility of the model requires further validation through prospective, multi-center studies. Additionally, while the multilayer fusion model outperformed other models, its specificity can be further improved to reduce false positives. Enhancing specificity may require more complex model structures or higher-resolution imaging, potentially achieving a better balance between sensitivity and precision.
5 Conclusion
Overall, this study developed a multimodal fusion model that integrates ultrasound images and pathology slides of the primary tumor to preoperatively predict ALNM in breast cancer patients. The model demonstrated a moderate diagnostic performance, outperforming models based solely on single-modality images. This approach has the potential to assist clinicians in lymph node staging and personalized treatment decisions. Future improvements could focus on enhancing predictive accuracy by incorporating larger datasets, leveraging more advanced architectures, and comprehensively integrating clinical features with multiple image data.
Data availability statement
The original contributions presented in the study are included in the article/supplementary material. Further inquiries can be directed to the corresponding authors.
Ethics statement
The studies involving humans were approved by the Ethics Committee of Union Hospital, Tongji Medical College, Huazhong University of Science and Technology. The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their written informed consent to participate in this study.
Author contributions
LP: Data curation, Formal analysis, Software, Writing – original draft. LY: Data curation, Formal analysis, Writing – original draft, Methodology. BL: Data curation, Funding acquisition, Software, Writing – original draft. FX: Conceptualization, Investigation, Methodology, Supervision, Writing – review & editing, Funding acquisition. YW: Conceptualization, Resources, Supervision, Visualization, Writing – review & editing.
Funding
The author(s) declare financial support was received for the research and/or publication of this article. This work was supported by the Hubei Provincial Natural Science Foundation (Grant Nos. 2025AFD870. Principal Investigator: Feixiang Xiang) and the National Natural Science Foundation of China (Grant Nos. 82200648. Principal Investigator: Beibei Liu).
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Generative AI statement
The author(s) declare that no Generative AI was used in the creation of this manuscript.
Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Abbreviations
DL, deep learning; ALNM, axillary lymph node metastasis; AUC, area under the curve; BC, breast cancer; SLNB, sentinel lymph node biopsy; CNB, core needle biopsy; ER, estrogen receptor; PR, progesterone receptor; HER2, human epidermal growth factor receptor 2; TNBC, triple-negative breast carcinoma; ROI, regions of interest; SE, Squeeze-and-Excitation; BN, batch normalization; SLND, sentinel lymph node dissection; ALND, axillary lymph node dissection.
References
1. Siegel R, Giaquinto A, and Jemal A. Cancer statistics, 2024. CA: Cancer J Clin. (2024) 74:12–49. doi: 10.3322/caac.21820
2. Nikki T, Mireille A, Gauri K, Stephanie R, Jasleen C, Sergio D, et al. Breast cancer tissue markers, genomic profiling, and other prognostic factors: A primer for radiologists. Radiographics. (2018) 38:1902–20. doi: 10.1148/rg.2018180047
3. Michaelson J, Silverstein M, Sgroi D, Cheongsiatmoy J, Taghian A, Powell S, et al. The effect of tumor size and lymph node status on breast carcinoma lethality. Cancer. (2003) 98:2133–43. doi: 10.1002/cncr.v98:10
4. Armando EG, James LC, Stephen BE, Elizabeth AM, Hope SR, Lawrence JS, et al. Breast Cancer-Major changes in the American Joint Committee on Cancer eighth edition cancer staging manual. CA Cancer J Clin. (2017) 67:290–303.
5. Blumgart EI, Uren RF, Nielsen PM, Nash MP, and Reynolds HM. Predicting lymphatic drainage patterns and primary tumour location in patients with breast cancer. Breast Cancer Res Treat. (2011) 130:699–705. doi: 10.1007/s10549-011-1737-2
6. Chang J, Leung J, Moy L, Ha S, and Moon W. Axillary nodal evaluation in breast cancer: state of the art. Radiology. (2020) 295:500–15. doi: 10.1148/radiol.2020192534
7. Christophe VB, Manon H, Mireille VG, Andrew V, Konstantinos P, Inge V, et al. Preoperative ultrasound staging of the axilla make’s peroperative examination of the sentinel node redundant in breast cancer: saving tissue, time and money. Eur J Obstet Gynecol Reprod Biol. (2016) 206:164–71.
8. Jung Min C, Jessica WTL, Linda M, Su Min H, and Woo Kyung M. Axillary nodal evaluation in breast cancer: state of the art. Radiology. (2020) 295:500–15. doi: 10.1148/radiol.2020192534
9. de Boer M, van Deurzen C, van Dijck J, Borm G, van Diest P, Adang E, et al. Micrometastases or isolated tumor cells and the outcome of breast cancer. New Engl J Med. (2009) 361:653–63. doi: 10.1056/NEJMoa0904832
10. Gary HL, Mark RS, Linda DB, Cheryl LP, Donald LW, and Armando EG. Sentinel lymph node biopsy for patients with early-stage breast cancer: american society of clinical oncology clinical practice guideline update. J Clin Oncol. (2016) 35:561–4.
11. Wei S, Yingshi S, Rui Z, Wei X, Zhenqiang L, Ning M, et al. Prediction of axillary lymph node metastasis using a magnetic resonance imaging radiomics model of invasive breast cancer primary tumor. Cancer Imaging. (2024) 24:122. doi: 10.1186/s40644-024-00771-y
12. Ye X, Zhang X, Lin Z, Liang T, Liu G, and Zhao P. Ultrasound-based radiomics nomogram for predicting axillary lymph node metastasis in invasive breast cancer. Am J Trans Res. (2024) 16:2398–410. doi: 10.62347/KEPZ9726
13. Song B. A machine learning-based radiomics model for the prediction of axillary lymph-node metastasis in breast cancer. Breast Cancer (Tokyo Japan). (2021) 28:664–71. doi: 10.1007/s12282-020-01202-z
14. Shahriarirad R, Meshkati Yazd S, Fathian R, Fallahi M, Ghadiani Z, and Nafissi N. Prediction of sentinel lymph node metastasis in breast cancer patients based on preoperative features: a deep machine learning approach. Sci Rep. (2024) 14:1351. doi: 10.1038/s41598-024-51244-y
15. Zhou L, Wu X, Huang S, Wu G, Ye H, Wei Q, et al. Lymph node metastasis prediction from primary breast cancer US images using deep learning. Radiology. (2020) 294:19–28. doi: 10.1148/radiol.2019190372
16. Yi-Jun G, Rui Y, Qian Z, Jun-Qi H, Zhao-Xiang D, Peng-Bo W, et al. MRI-based kinetic heterogeneity evaluation in the accurate access of axillary lymph node status in breast cancer using a hybrid CNN-RNN model. J Magn Reson Imaging. (2024) 60:1352–64.
17. Xu F, Zhu C, Tang W, Wang Y, Zhang Y, Li J, et al. Predicting axillary lymph node metastasis in early breast cancer using deep learning on primary tumor biopsy slides. Front Oncol. (2021) 11:759007. doi: 10.3389/fonc.2021.759007
18. Samantha B, Maria Colomba C, Vito L, Cristian C, Vittorio D, Gianluca G, et al. A ultrasound-based radiomic approach to predict the nodal status in clinically negative breast cancer patients. Sci Rep. (2022) 12:7914. doi: 10.1038/s41598-022-11876-4
19. Yang SY, Li SH, Liu JL, Sun XQ, Cen YY, Ren RY, et al. Histopathology-based diagnosis of oral squamous cell carcinoma using deep learning. J Dent Res. (2022) 101:1321–7. doi: 10.1177/00220345221089858
20. Goldhirsch A, Wood W, Coates A, Gelber R, Thürlimann B, and Senn H. Strategies for subtypes–dealing with the diversity of breast cancer: highlights of the St. Gallen International Expert Consensus on the Primary Therapy of Early Breast Cancer 2011. Ann oncology: Off J Eur Soc Med Oncol. (2011) 22:1736–47. doi: 10.1093/annonc/mdr304
21. Di Z, Wang Z, Wen-Wu L, Xia-Chuan Q, Xian-Ya Z, Yan-Hong L, et al. Ultrasound-based deep learning radiomics for enhanced axillary lymph node metastasis assessment: a multicenter study. Oncologist. (2025) 30:oyaf090.
22. Min-Yi C, Can-Gui W, Ying-Yi L, Jia-Chen Z, Dong-Qing W, Bruce GH, et al. Development and validation of a multivariable risk model based on clinicopathological characteristics, mammography, and MRI imaging features for predicting axillary lymph node metastasis in patients with upgraded ductal carcinoma in situ. Gland Surg. (2025) 14:738–53. doi: 10.21037/gs-2025-89
23. Yizhou C, Xiaoliang S, Kuangyu S, Axel R, and Federico C. AI in breast cancer imaging: an update and future trends. Semin Nucl Med. (2025) 55:358–70. doi: 10.1053/j.semnuclmed.2025.01.008
24. Kai H, Yunhe W, Hanting C, Xinghao C, Jianyuan G, Zhenhua L, et al. A survey on vision transformer. IEEE Trans Pattern Anal Mach Intell. (2022) 45:87–110.
25. Ishak P and Omneya A. InceptionNeXt-Transformer: A novel multi-scale deep feature learning architecture for multimodal breast cancer diagnosis. Biomed Signal Process Control. (2025) 110:108116. doi: 10.1016/j.bspc.2025.108116
26. Ishak P and Omneya A. Hybrid deep learning model for automated colorectal cancer detection using local and global feature extraction. Knowledge-Based Systems. (2025) 319:113625. doi: 10.1016/j.knosys.2025.113625
27. Ishak P, Ozan A, Rumeysa Tuna D, and Muhammet D. NeXtBrain: Combining local and global feature learning for brain tumor classification. Brain Res. (2025) 1863:149762. doi: 10.1016/j.brainres.2025.149762
28. Suat I, Ismail K, Ali A, Bilal B, and Ishak P. Deep learning for cerebral vascular occlusion segmentation: A novel ConvNeXtV2 and GRN-integrated U-Net framework for diffusion-weighted imaging. Neuroscience. (2025) 574:42–53. doi: 10.1016/j.neuroscience.2025.04.010
29. Yini H, Zhao Y, Lingling L, Rushuang M, Weijun H, Zhengming H, et al. Deep learning radiopathomics based on preoperative US images and biopsy whole slide images can distinguish between luminal and non-luminal tumors in early-stage breast cancers. EBioMedicine. (2023) 94:104706. doi: 10.1016/j.ebiom.2023.104706
30. Bevers T, Anderson B, Bonaccio E, Buys S, Daly M, Dempsey P, et al. NCCN clinical practice guidelines in oncology: breast cancer screening and diagnosis. J Natl Compr Cancer Network: JNCCN. (2009) 7:1060–96. doi: 10.6004/jnccn.2009.0070
31. Keelan S, Heeney A, Downey E, Hegarty A, Roche T, Power C, et al. Breast cancer patients with a negative axillary ultrasound may have clinically significant nodal metastasis. Breast Cancer Res Treat. (2021) 187:303–10. doi: 10.1007/s10549-021-06194-8
32. Neal C, Daly C, Nees A, and Helvie M. Can preoperative axillary US help exclude N2 and N3 metastatic breast cancer? Radiology. (2010) 257:335–41.
33. Masakuni N, Masafumi I, Miki N, Emi M, Yukako O, and Tomoko K. Axillary surgery for breast cancer: past, present, and future. Breast Cancer. (2020) 28:9–15.
34. Armando EG, Karla VB, Linda M, Peter DB, Meghan BB, Pond RK, et al. Effect of axillary dissection vs no axillary dissection on 10-year overall survival among women with invasive breast cancer and sentinel node metastasis: the ACOSOG Z0011 (Alliance) randomized clinical trial. JAMA. (2017) 318:918–26. doi: 10.1001/jama.2017.11470
35. Galimberti V, Cole B, Viale G, Veronesi P, Vicini E, Intra M, et al. Axillary dissection versus no axillary dissection in patients with breast cancer and sentinel-node micrometastases (IBCSG 23-01): 10-year follow-up of a randomised, controlled phase 3 trial. Lancet Oncol. (2018) 19:1385–93. doi: 10.1016/S1470-2045(18)30380-2
36. Sávolt Á, Péley G, Polgár C, Udvarhelyi N, Rubovszky G, Kovács E, et al. Eight-year follow up result of the OTOASOR trial: The Optimal Treatment Of the Axilla - Surgery Or Radiotherapy after positive sentinel lymph node biopsy in early-stage breast cancer: A randomized, single centre, phase III, non-inferiority trial. Eur J Surg Oncol. (2017) 43:672–9.
37. Pradeepa M, Sharmila B, and Nirmala M. A hybrid deep learning model EfficientNet with GRU for breast cancer detection from histopathology images. Sci Rep. (2025) 15:24633. doi: 10.1038/s41598-025-00930-6
38. Jie H, Li S, Samuel A, Gang S, and Enhua W. Squeeze-and-excitation networks. IEEE Trans Pattern Anal Mach Intell. (2019) 42:2011–23.
39. Bjorck J, Gomes C, and Selman B. Understanding batch normalization. Advances in Neural Information Processing Systems. (2018) 31:7694–705.
40. Takaaki F, Reina Y, Hironori T, Toshinaga S, Hiroki M, Soichi T, et al. Significance of lymphatic invasion combined with size of primary tumor for predicting sentinel lymph node metastasis in patients with breast cancer. Anticancer Res. (2015) 35:3581–4.
41. Xu S, Wang Q, and Hong Z. The correlation between multi-mode ultrasonographic features of breast cancer and axillary lymph node metastasis. Front Oncol. (2024) 14:1433872. doi: 10.3389/fonc.2024.1433872
42. Bai X, Wang Y, Song R, Li S, Song Y, Wang H, et al. Ultrasound and clinicopathological characteristics of breast cancer for predicting axillary lymph node metastasis. Clin hemorheology microcirculation. (2023) 85:147–62. doi: 10.3233/CH-231777
43. Ding J, Jiang L, and Wu W. Predictive value of clinicopathological characteristics for sentinel lymph node metastasis in early breast cancer. Med Sci monitor: Int Med J Exp Clin Res. (2017) 23:4102–8. doi: 10.12659/MSM.902795
44. Yenidunya S, Bayrak R, and Haltas H. Predictive value of pathological and immunohistochemical parameters for axillary lymph node metastasis in breast carcinoma. Diagn pathology. (2011) 6:18. doi: 10.1186/1746-1596-6-18
45. Wuyue Z, Siying W, Yichun W, Jiawei S, Hong W, Weili X, et al. Ultrasound-based radiomics nomogram for predicting axillary lymph node metastasis in early-stage breast cancer. Radiol Med. (2024) 129:211–21. doi: 10.1007/s11547-024-01768-0
46. Ying S, Jinjin L, Chenyang J, Yan Z, Yingying Z, Kairen Z, et al. Value of contrast-enhanced ultrasound combined with immune-inflammatory markers in predicting axillary lymph node metastasis of breast cancer. Acad Radiol. (2024) 31:3535–45. doi: 10.1016/j.acra.2024.06.013
47. Wei L, Xian W, Sunwang X, Dexing W, Jinshu C, Linying C, et al. Expression of histone methyltransferase WHSC1 in invasive breast cancer and its correlation with clinical and pathological data. Pathol Res Pract. (2024) 263:155647. doi: 10.1016/j.prp.2024.155647
48. Qiucheng W, Bo L, Zhao L, Haitao S, Hui J, Hua S, et al. Prediction model of axillary lymph node status using automated breast ultrasound (ABUS) and ki-67 status in early-stage breast cancer. BMC Cancer. (2022) 22:929. doi: 10.1186/s12885-022-10034-3
Keywords: breast cancer, deep learning, axillary lymph node metastasis, ultrasound, core-needle biopsy
Citation: Peng L, Yu L, Liu B, Xiang F and Wu Y (2025) Construction of a prediction model for axillary lymph node metastasis in breast cancer patients based on a multimodal fusion strategy of ultrasound and pathological images. Front. Oncol. 15:1591858. doi: 10.3389/fonc.2025.1591858
Received: 14 March 2025; Accepted: 12 August 2025;
Published: 09 September 2025.
Edited by:
L. J. Muhammad, Bayero University Kano, NigeriaReviewed by:
Raffaella Massafra, National Cancer Institute Foundation (IRCCS), ItalyIshak Pacal, Iğdır Üniversitesi, Türkiye
Copyright © 2025 Peng, Yu, Liu, Xiang and Wu. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Yu Wu, d3l3MjAyNTAwMjVAMTI2LmNvbQ==; Feixiang Xiang, eGlhbmdmeEBodXN0LmVkdS5jbg==
†These authors have contributed equally to this work and share first authorship