Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Oncol., 08 October 2025

Sec. Breast Cancer

Volume 15 - 2025 | https://doi.org/10.3389/fonc.2025.1558880

This article is part of the Research TopicAdvancing Breast Cancer Care Through Transparent AI and Federated Learning: Integrating Radiological, Histopathological, and Clinical Data for Diagnosis, Recurrence Prediction, and SurvivorshipView all 12 articles

Integrating multimodal ultrasound imaging and machine learning for predicting luminal and non-luminal breast cancer subtypes

Yan Fu&#x;Yan Fu1†Huang Jing Chen&#x;Huang Jing Chen1†Hao ZhangHao Zhang1Dong Jie LiuDong Jie Liu1Xi ChenXi Chen1Cheng Yu QiuCheng Yu Qiu2Wen Wu LuWen Wu Lu3Hao Miao BaiHao Miao Bai1Qiu Wei LiQiu Wei Li1Guo Xue LiGuo Xue Li1Zi Jun ShenZi Jun Shen1Chang Jiang Gu*Chang Jiang Gu4*Yuan Peng Zhang*Yuan Peng Zhang2*Xue Jun Ni*Xue Jun Ni1*
  • 1Department of Ultrasound, Affiliated Hospital of Nantong University, Nantong, Jiangsu, China
  • 2Department of Medical Informatics, School of Medicine, Nantong University, Nantong, Jiangsu, China
  • 3Department of Ultrasound, The First Affiliated Hospital of Anhui Medical University, Hefei, Anhui, China
  • 4Department of General Surgery, Affiliated Hospital of Nantong University, Nantong, Jiangsu, China

Rationale and Objectives: Breast cancer molecular subtypes significantly influence treatment outcomes and prognoses, necessitating precise differentiation to tailor individualized therapies. This study leverages multimodal ultrasound imaging combined with machine learning to preoperatively classify luminal and non-luminal subtypes, aiming to enhance diagnostic accuracy and clinical decision-making.

Methods: This retrospective study included 247 patients with breast cancer, with 192 meeting the inclusion criteria. Patients were randomly divided into a training set (134 cases) and a validation set (58 cases) in a 7:3 ratio. Image segmentation was conducted using 3D Slicer software, adhering to IBSI-standardized radiomics feature extraction. We constructed four model configurations—monomodal, dual-modal, trimodal, and four-modal—through optimized feature selection. These included monomodal datasets comprising 2D ultrasound (US) images, dual-modal datasets integrating 2D US with color Doppler flow imaging (CDFI) (US+CDFI), trimodal datasets incorporating strain elastography (SE) alongside 2D US and CDFI (US+CDFI+SE), and four-modal datasets combining all modalities, including ABVS coronal imaging (US+CDFI+SE+ABVS). Machine learning classifiers such as logistic regression (LR), support vector machines (SVM), AdaBoost (adaptive boosting), random forests(RF), linear discriminant analysis(LDA), and ridge regression were utilized.

Results: The four-modal model achieved the highest performance (AUC: 0.947, 95% CI: 0.884-0.986), significantly outperforming the monomodal model (AUC 0.758, ΔAUC +0.189). Multimodal integration progressively enhanced performance: trimodal models surpassed dual-modal and monomodal approaches (AUC 0.865 vs 0.741 and 0.758), and the four-modal framework showed marked improvements in sensitivity (88.4% vs 71.1% for monomodal), specificity (92.7% vs 70.1%), and F1 scores (0.905).

Conclusion: This study establishes a multimodal machine learning model integrating advanced ultrasound imaging techniques to preoperatively distinguish luminal from non-luminal breast cancers. The model demonstrates significant potential to improve diagnostic accuracy and generalization, representing a notable advancement in non-invasive breast cancer diagnostics.

1 Introduction

Breast cancer remains the leading cause of cancer-related mortality among women worldwide (1). It encompasses several molecular subtypes with distinct differences in treatment responses and prognoses, making accurate subtype identification essential for personalized therapy. The luminal subtype, characterized by high levels of estrogen (ER) and progesterone (PR) receptors, generally responds well to hormonal therapies (2). Conversely, non-luminal subtypes, including HER2 (human epidermal growth factor receptor 2)-enriched and triple-negative breast cancers(TNBC), exhibit aggressive behavior, reduced hormonal treatment responsiveness, and poorer outcomes (3, 4). Luminal cancers typically undergo endocrine therapy ± surgery, whereas Non-luminal subtypes require neoadjuvant chemotherapy—a decision that must be made preoperatively. In resource-limited settings, patients wait >2 weeks for IHC results, delaying time-critical therapy. Thus, precise and timely molecular subtype classification is critical for optimizing individualized treatment strategies.

While Imaging techniques such as magnetic resonance imaging (MRI), ultrasound, and mammography are widely used in breast cancer detection (5), core needle biopsy (CNB) with immunohistochemistry (IHC) remains the diagnostic gold standard for molecular classification. However, CNB is invasive, costly, time-intensive, and may fail to capture tumor heterogeneity, potentially leading to diagnostic inaccuracies (6). Radiomics, an emerging technology providing quantitative tumor assessments beyond human visual interpretation, offers promising solutions to these challenges (7, 8).

MRI-based radiomics has effectively evaluated malignancy and molecular subtypes (911). Free from ionizing radiation, ultrasound is particularly suitable for young and pregnant women and demonstrates higher sensitivity than mammography for detecting intraductal and nodular lesions (12). In populations with smaller, denser breasts, such as younger and Chinese women, ultrasound is preferred for breast lesion screening and preoperative evaluation (13). However, grayscale ultrasound’s dependence on radiologist interpretation introduces variability and subjectivity, prompting the development of radiomics-based approaches to address these limitations but are frequently confined to the single modality of two-dimensional (2D) ultrasound (1417). Recent advances in deep learning (e.g., ResNet-101) show promise in unimodal ultrasound classification (18), yet remain constrained by single-modality data. Multimodal fusion may address critical diagnostic gaps in clinical practice, particularly for lesions with ambiguous imaging phenotypes (19).

Advancements in ultrasound technology have introduced modalities such as CDFI, SE, and ABVS, which offer distinct advantages in breast lesion assessment (20, 21). CDFI can quantify the vascular distribution of tumors, and SE is capable of measuring tissue hardness. Existing studies have shown that both elastography and CDFI are correlated with the Luminal A cancer subtype (22). ABVS, in particular, enables coronal imaging that enhances the visualization of lesions and adjacent tissues (23), facilitating detailed characterization and improved assessment of tissue architecture and spatial relationships. Additionally, ABVS produces standardized images, minimizing variability and making it ideal for radiomics analysis, where consistency and reproducibility are key for extracting meaningful features (24).

ABVS coronal combined with SE has demonstrated efficacy in distinguishing breast lesions (25), but radiomics studies utilizing ABVS for molecular subtype classification are still scarce. Manual feature selection methods, commonly used in healthcare data analysis, are inadequate for handling the growing complexity of multimodal imaging data. Unstable selection processes can lead to, substantial variability in selected feature subsets, compromising model reliability (26, 27).

This study aims to integrate multiple ultrasound modalities—including 2D ultrasound, CDFI, SE, and ABVS coronal imaging—into a machine-learning framework designed to preoperatively differentiate luminal from non-luminal breast cancer subtypes. Unlike prior research predominantly focused on single-modality radiomics, this study adopts a multimodal fusion approach to achieve superior accuracy and reliability in subtype differentiation. To further enhance model stability and performance, diverse feature selection strategies are employed to address the limitations of unstable selection methods, thereby improving the overall robustness and reliability of the analysis.

2 Methodology

This retrospective study included 247 breast cancer cases diagnosed through histopathological evaluation of CNB samples or surgical specimens at our hospital between January 2020 and June 2024. Ethical approval was obtained from the hospital review board, and informed consent was waived. Patient information was anonymized to ensure privacy.

The inclusion criteria were (1): complete and high-quality ultrasound images with mass-like lesions suitable for tumor segmentation and (2) solitary malignant tumors. Exclusion criteria included (1): poor-quality ultrasound images, (2) multiple lesions, (3) incomplete clinical or ultrasound data, (4) lack of puncture or surgical IHC results, (5) tumor diameters exceeding 50 mm(Large tumors frequently exceed standard ultrasound probe fields-of-view, resulting in discontinuous ROI segmentation that compromises radiomic feature stability.), and (6) prior local or systemic therapy (e.g., chemotherapy, radiation, ablation, or excision) before breast ultrasonography. Of the 247 cases, 192 patients met the criteria, including 140 ductal and 52 non-ductal breast cancers. Patients were randomly divided into a test set (134 cases) and a validation set (58 cases) in a 7:3 ratio (Figure 1).

Figure 1
Flowchart outlining the selection process for patients who underwent two-dimensional ultrasound and ABVS examination from January 2020 to June 2024 at the Affiliated Hospital of Nantong University. From 247 patients, inclusion criteria require mass-like lesions, clear ultrasonic images, and single malignant tumors. Exclusion criteria include poor image quality, multiple lesions, lack of biopsy data, tumors over fifty millimeters, and prior treatment. This results in 192 selected patients, divided randomly 7:3 into 134 for the training cohort and 58 for the testing cohort.

Figure 1. Flowchart illustrating the selection process for breast cancer patients, detailing the inclusion and exclusion criteria and the allocation to training and testing cohorts.

2.1 Clinical information

Clinical data collected for training and validation sets included age, ultrasound-reported tumor size, microcalcifications, convergence sign, breast imaging reporting and data system (BI-RADS) classification, pathology, strain elasticity score, ER/PR/HER2 status, Ki-67, and molecular subtype. Tumors were classified as luminal or non-luminal based on hormone receptor (HR) status from IHC results. Tumors with ≥1% of cells staining positive for ER or PR were categorized as luminal, while HR-negative tumors were classified as non-luminal (28). Images were labeled accordingly.

2.2 Image acquisition

Breast ultrasound examinations were performed by two experienced physicians, each with over 5 years of expertise in breast imaging. A Siemens Acuson Oxana 2 ABVS (Siemens Healthineers, Erlangen, Germany) equipped with 9L4 and 14L5B line-array probes performing radial, transverse, and longitudinal scans. The largest ultrasonic area was evaluated, and ultrasound features such as BI-RADS classification, size, location, shape, margins, internal echoes, microcalcifications, strain elasticity score, convergence sign, vascularity, and axillary lymph node involvement were documented. All modalities were acquired consecutively within 2 hours using the same scanner (Siemens Acuson Oxana 2). ROIs were segmented from synchronized ABVS/2D-US images.

2.3 Image segmentation and feature extraction

Figure 2 outlines the radiomics workflow, including segmentation, feature extraction and selection, image preprocessing, feature analysis, and model construction. Image preprocessing was performed following IBSI(Imaging Biomarker Standardization Initiative) guidelines, including voxel resampling for spatial normalization, intensity discretization, and z-score normalization of feature values. The target area was delineated using the open-source software 3D Slicer (version 5.6.1), and features—such as first-order, morphological, grayscale histogram, and wavelet transforms—were extracted using the SlicerRadiomics extension.

Figure 2
Diagram illustrating a workflow comprising image segmentation, feature extraction, feature selection, radiomics signature creation, model development, and evaluation. Steps include segmenting images, extracting shape and texture features, selecting and evaluating features, developing models with techniques like SVM and AdaBoost, and assessing results using ROC and calibration plots.

Figure 2. Workflow diagram of the study, outlining the key steps from data acquisition to model evaluation.

Forty ultrasound images (20 luminal and 20 non-luminal cases) were randomly selected (29), with regions of interest (ROIs) independently delineated by two physicians for intergroup consistency. Two weeks later, Physician 1 repeated the ROI delineation for intragroup consistency testing. The intraclass correlation coefficients (ICCs) for intergroup and intragroup tests exceeded 0.75, confirming high feature consistency. Subsequently, Physician 1 segmented the remaining images, retaining only features with an ICC above 0.75 for further analysis.

2.4 Feature selection and dataset construction

When the number of patients is substantially smaller than the number of extracted features, the data becomes sparse within a high-dimensional space (26), undermining machine learning models’ accuracy and generalizability. To reduce redundancy, exclude irrelevant features, and minimize the risk of overfitting, the method proposed by Li et al. (30) was employed, incorporating six unsupervised feature selection (FS) algorithms—Lap_score, SPEC, MCFS, NDFS, UDFS, and person score—alongside four supervised FS algorithms—F score, Tscore, ReliefF, and Fish_score— yielding a total of 24 FS combinations. Supervised FS: Depends on data labels (such as the categories in classification tasks and the continuous values in regression tasks). The goal is to select features that make the model more effective in predicting labels. Unsupervised FS: Does not rely on labels. The goal is to optimize the internal structure of the feature set (such as reducing redundancy and retaining key patterns). It is often used for subsequent unsupervised tasks (like clustering, dimensionality reduction) or as a preprocessing step for supervised tasks. After evaluation, we finally selected the dataset with AUC(Area Under the Curve)*stability greater than 0.45 for subsequent analysis. Feature selection was independently conducted for each dataset grouping to eliminate potential cross-group interference and to preserve methodological integrity. The overarching FS strategy is illustrated in Figure 3. This process facilitated the generation of multiple datasets with increasing complexity. These included monomodal datasets comprising 2D US images, dual-modal datasets integrating 2D US with CDFI (US+CDFI), trimodal datasets incorporating SE alongside 2D US and CDFI (US+CDFI+SE), and four-modal datasets combining all modalities, including ABVS coronal imaging (US+CDFI+SE+ABVS).

Figure 3
Flowchart depicting a data processing workflow. It begins with data undergoing bootstrapping, followed by “boostData,” then supervised and unsupervised selection, resulting in ranking outcomes. The selection involves scores like F Score and Lap_score. A decision node follows with a condition \(i >= M\). If true, it proceeds to stability evaluation, linking back to supervised selection. If false, it progresses to OOBData. The workflow loops incrementally through \(i++\).

Figure 3. Overview of feature selection methods and workflow.

Feature fusion was performed by channel concatenation:

Ffused=[FUSFCDFIFSEFABVS]

where FUSRn×15,FCDFIRn×15,FSERn×15,FABVSRn×15 … denote modality-specific feature matrices(e.g., ultrasound, color Doppler, shear wave elastography, etc.), ∣ represents column-wise concatenation (stacking matrices along the feature channel dimension), and n is the sample size (number of data instances).

2.5 ML model derivation and validation

Six classifiers—LR, AdaBoost, LDA, ridge regression, SVM, and RF—were used for prediction. Synthetic Minority Over-sampling Technique (SMOTE) oversampling balanced the dataset, and 10-fold cross-validation with grid search was used for hyperparameter optimization(e.g. SVM (kernel: linear/RBF; penalty C: [0.1, 1, 10]; γ: [0.001, 0.01, 0.1]; optimal: RBF kernel, C = 1, γ=0.01)). This approach reduces evaluation bias caused by uneven training data division, providing a more stable performance assessment. The Bootstrap method was used to calculate the confidence interval of AUC, and the Dunn-Sidak correction was applied to control the multiple comparison errors. All workflows were implemented in Python (version 3.8). Calibration curves assessed alignment between predictions and outcomes, while decision curve analysis (DCA) evaluated clinical utility. Calibration curves were generated by binning predicted probabilities (10 quantile bins) and plotting mean predictions against observed event rates, with perfect calibration indicated by a 45° line.

2.6 SHAP model

Using game theory principles, SHAP (Shapley Additive Explanations) quantified each feature’s contribution to the model’s output. This interpretability framework visualized feature importance, highlighting the impact of individual variables on predictions (31). We applied SHAP values to assess each feature’s contribution to the optimal model and its influence on decision-making in specific scenarios.

2.7 Model evaluation and statistical analysis

Continuous variables were analyzed using the Mann-Whitney U or Student’s t-test, depending on normality assumptions, while Pearson’s chi-square test assessed categorical differences. Model performance metrics included AUC, sensitivity, specificity, positive and negative predictive values, and F1-score. Statistical significance was set at p< 0.05.

3 Results

3.1 Patient cohort distribution

Table 1 summarizes the characteristics of patients in the training and validation cohorts, revealing no statistically significant differences in age, tumor size, microcalcification, ultrasound convergence signs, BI-RADS classification, pathological tumor type, SE score, or molecular subtype between the two groups (p > 0.05).

Table 1
www.frontiersin.org

Table 1. Comparative analysis of demographic and clinical parameters across the training and testing cohorts.

3.2 Radiomics feature extraction and selection

For each patient, 863 features were uniformly extracted from the ROIs across the four imaging modalities: US, CDFI, SE, and ABVS coronal imaging. These features included radiomics metrics such as gray-level co-occurrence matrices (GLCM) and gray-level dependence matrices (GLDM). During feature selection, 24 distinct methods were applied to independently screen features within each modality, resulting in a 192 x 15 unimodal dataset for each modality (Figure 4). Optimal feature selection combinations included Fish score and UDFS, T-score and UDFS, F score and NDFS, and ReliefF and NDFS. The data from each modality were then fused at the feature level to generate multimodal datasets corresponding to dimensions of 192 x 15 (monomodal), 192 x 30 (dual-model), 192 x 45 (trimodal), and 192 x 60 (four-modal).

Figure 4
Four scatter plots display “Stability against AUC”. Each plot shows red data points with labels indicating values and identifiers. The x-axis ranges from approximately 0.5 to 0.9, and the y-axis shows AUC values. Green and black lines represent “auc_median” and “stability_median”. The data points cluster variably around the median lines, demonstrating different stability levels in relation to AUC.

Figure 4. Relationship between feature selection method stability and AUC. Red dots represent the results of each method, with stability plotted on the x-axis and AUC on the y-axis. The green and black lines indicate the median AUC and stability, respectively. This figure highlights variations in stability and classification performance across methods. AUC, area under the curve.

3.3 ML model derivation and evaluation

Four-modal models demonstrated significantly better performance in the test set than monomodal models across all machine learning algorithms (Figure 5). The SVM four-modal model achieved the highest performance, with an AUC of 0.947 (95% CI: [0.884, 0.986]), which was higher than the AUC of 0.758 (95% CI: [0.637, 0.853]) observed in the monomodal model. Other algorithms, including AdaBoost, LR, RF, and ridge regression, also achieved higher AUC values in four-modal configurations than monomodal ones. The trimodal SVM model had an AUC of 0.865 (95% CI: [0.778, 0.938]), which was higher than that in the dual-modal (AUC 0.741, 95% CI: [0.648, 0.854]) and monomodal models (AUC 0.758, 95% CI: [0.637, 0.853]). The specificity and sensitivity of the trimodal model were higher than those of the dual-modal and monomodal models. However, both metrics remained lower than those observed in the four-modal model, indicating improved diagnostic performance.

Figure 5
Four ROC curves labeled A to D compare model performances. A: Four-modal, highest AUC 0.947 (SVM). B: Trimodal, highest AUC 0.859 (Random Forest). C: Dual-modal, highest AUC 0.801 (LDA). D: Monomodal, highest AUC 0.793 (AdaBoost). Each panel includes a diagonal line representing random performance.

Figure 5. ROC curves and AUC values for various modal models using machine learning classifiers on the test set. Panels (A-D) sequentially present the performance of monomodal, dual-modal, trimodal, and four-modal models. AUC, area under the curve; ROC, receiver operating characteristic; SVM, support vector machine; AdaBoost, adaptive enhancement; LDA, linear discriminant analysis; Ridge, ridge regression.

Table 2 highlights the superior predictive performance of four-modal models across all machine learning algorithms. The AUC of the SVM four-modal model (0.947 (95% CI: [0.884, 0.986]) was higher than that in any other configuration. The LR four-modal model achieved the highest sensitivity (0.884, 95% CI: [0.778, 0.976]), specificity (0.927, 95% CI: [0.841, 1.000]), and F1-score (0.905, 95% CI: [0.829, 0.965]). Calibration curves and DCA confirmed the accuracy and clinical applicability of the four-modal models (Figure 6). Overall, these findings demonstrate that multimodal fusion significantly enhances the predictive capacity of each model.

Table 2
www.frontiersin.org

Table 2. Predictive performance of SVM and LR models in classifying luminal breast cancer subtypes.

Figure 6
Four panels show decision curve and calibration analyses. Panel A shows decision curve analysis for train datasets, with various models plotted against threshold probability. Panel B shows similar analysis for test datasets. Both panels include lines for “treat all” and “treat none” strategies. Panels C and D display calibration curves for train and test datasets, respectively, with predicted probabilities against the fraction of positives. Each panel uses different colors for four-modal, tri-modal, dual-modal, and mono-modal approaches, and includes a line representing perfect calibration.

Figure 6. Calibration curves and DCA results using the SVM classifier, comparing the four-modal, trimodal, dual-modal, and monomodal models. Panels (A, B) display calibration curves for the training and test sets, respectively, while panels (C, D) present the DCA results. The four-modal model demonstrates superior calibration accuracy and clinical decision-making utility. SVM, support vector machine; DCA, decision curve analysis.

3.4 Model interpretation (SHAP)

The SVM and LR four-modal models demonstrated superior performance based on a comprehensive evaluation. SHAP analysis quantified the contribution of each feature, with average absolute SHAP values serving as the primary metric. Figure 7 visualizes the cumulative impact of each feature, where higher eigenvalues (in red) indicate a stronger positive influence on predictions. In Figure I, the features wavelet-LLH-glrlm-LongRunLowGrayLevelEmphasis3 and wavelet-LLH-glrlm-LongRunEmphasis3 had a greater influence on predictions in the SVM model than other features. In contrast, Figure G shows that wavelet-HLL-gldm-DependenceNonUniformityNormalized4 was the most significant feature for predictions in the LR model, followed by wavelet-LLH-glrlm-RunVariance3. Overall, features belonging to GLDM (Gray Level Dependency Matrix), GLRLM (Gray Level Run Length Matrix), GLSZM (Gray Level Size Zone Matrix), first-order statistical features, and wavelet features contributed the most to the predictions in both models.

Figure 7
Two SHAP summary plots for SVM and Logistic Regression models. Plot A shows SHAP values indicating feature impact on the SVM model output; features like “wavelet-LLH-glrlm-LongRunLowGrayLevelEmphasis3” have high impact. Plot B shows similar analysis for Logistic Regression, highlighting “wavelet-HLL-gldm-DependenceNonUniformityNormalized4” as significant. Both plots include feature importance bar graphs on the right, showing mean SHAP values.

Figure 7. SHAP analysis for the SVM and LR classifiers. (A) illustrates the SVM results, while (B) shows the LR results. Features are ranked by their mean absolute SHAP values, with top-ranked features significantly impacting predictions. The summary plot visualizes how feature values influence model output, with blue representing lower feature values and red representing higher feature values. The horizontal position of each dot reflects the SHAP value, indicating the feature’s effect on individual predictions. LR, logistic regression; SVM, support vector machine; SHAP, Shapley Additive Explanations.

4 Discussion

This study successfully developed a multimodal machine learning model to classify luminal and non-luminal breast cancer subtypes preoperatively. The progressive integration of modalities—from 2D-US to US+CDFI+SE+ABVS—yielded significant AUC gains (monomodal:0.758 → four-modal:0.947, Δ+0.189). This demonstrates that combining structural and functional data outperforms single-modality assessment. These findings underscore the critical role of multimodal data fusion in enhancing model generalizability and accuracy. Given the diagnostic challenges posed by tumor heterogeneity and biopsy limitations, this approach provides a promising complementary tool for personalized treatment strategies in early-stage breast cancer.

As non-luminal breast cancer often necessitates neoadjuvant therapy, the developed model provides a promising complementary tool for personalized treatment strategies in patients with early-stage breast cancer. Tumor heterogeneity, which can lead to inaccurate IHC results from preoperative biopsies, highlights the value of imaging-based approaches for molecular subtype classification (32, 33). Radiomics techniques have effectively addressed diagnostic gaps inherent in biopsy limitations, supporting their integration into clinical workflows.

Previous studies have confirmed the potential of mammographic radiomics and MRI-based analyses in predicting molecular subtypes. Early studies reported AUC values up to 0.836 (34, 35), while recent research advancements have shown that mammographic radiomics models (36) have achieved AUC values of 0.855, and multi-parametric MRI (mpMRI)-based feature fusion models (37) have reported AUCs of over 0.81.These findings indicate that advanced imaging techniques hold significant potential as effective complementary tools for accurately classifying breast cancer molecular subtypes, thereby augmenting existing diagnostic approaches. In comparison, ABVS offers distinct advantages by combining standard ultrasound and ABVS imaging into a single examination, thereby improving clinical efficiency, diagnostic accuracy, and patient management (3840).

Despite its utility, the correlation between ABVS coronal imaging and breast cancer molecular subtypes remains underexplored. This study addresses this gap by integrating ABVS with other imaging modalities, demonstrating its effectiveness in multimodal classification.

Our study demonstrated that LR and SVM models were the most effective in distinguishing between luminal and non-luminal breast cancers, consistent with findings from prior research. For instance, a study utilizing 2D ultrasound to classify PR-positive and PR-negative breast cancers achieved an AUC of 0.879 for the LR classifier, surpassing the performance of our monomodal model (41). This disparity may be attributed to the larger dataset size in the prior study, which likely enhanced the generalizability of the model in distinguishing between different molecular subtypes. Our multimodal approach demonstrated improved performance in our cohort, particularly with the SVM model. The SVM’s ability to handle complex nonlinear relationships through its kernel method enabled more accurate classification, underscoring its suitability for multimodal datasets (42, 43).

As Jiangfeng Wu et al. noted, texture, wavelet, and first-order features are crucial for differentiating luminal and non-luminal subtypes, with features such as wavelet-HLL-gldm-DependenceNonUniformityNormalized, original-glrlm-HighGrayLevelRunEmphasis, and wavelet-LHL-glszm-SizeZoneNonUniformity showing particular significance. Shape-based features, however, were omitted due to their lower diagnostic relevance (17). Our findings align closely with these observations and suggest that refining the extraction and application of these features could further enhance diagnostic accuracy.

The multimodal model showed strong performance for HER2-enriched subtypes (n=14 in test set), but TNBC predictions (n=5) require validation due to limited samples. Caution is warranted when generalizing TNBC results. The SVM and LR classifiers demonstrated outstanding performance within our multimodal model, particularly when ABVS and SE images were incorporated. During the feature selection process, the majority of the ten most important features originated from ABVS and SE images, while fewer key features were derived from CDFI images. These results highlight the pivotal role of ABVS and SE imaging in accurately classifying molecular subtypes of breast cancer, emphasizing their integration as essential components of multimodal radiomics workflows.

The Random Forest (RF) model's perfect AUC of 1.0 on the unimodal training set strongly indicates overfitting. Furthermore, the integration of a second modality did not consistently yield improvements; conversely, the dual-modal configuration (US+CDFI) exhibited a marginally lower AUC than the monomodal US in some classifiers (e.g., SVM: 0.741 vs. 0.758). We attribute this to feature redundancy between vascularity (CDFI) and parenchymal texture (US), which introduced noise without augmenting discriminatory power. Nevertheless, incorporating SE and ABVS coronal imaging resolved this limitation by adding orthogonal biological information—tissue stiffness and 3D architectural distortion—yielding significant improvements in four-modal models. Overfitting typically occurs when models are trained on a limited set of features, resulting in the fitting of noise within the training data and compromised performance on the unseen test data. In this study, the multimodal model leveraged a more diverse and comprehensive feature set by integrating multiple imaging modalities, thereby improving robustness and stabilizing metrics such as AUC. These findings underscore the value of multimodal fusion in mitigating overfitting and enhancing the generalizability of machine learning models for complex diagnostic tasks.

The clinical value of this model may extend beyond diagnostic accuracy to optimizing diagnostic and therapeutic pathways. For instance, in patients with a high-confidence prediction of a low-risk Luminal A-type profile (e.g., predicted probability >0.90), the model could potentially support a discussion about proceeding directly to surgery, with definitive diagnosis and full molecular subtyping (including Ki-67 and HER2 status) confirmed postoperatively on the surgical specimen. However, it is crucial to emphasize that this approach would be entirely inappropriate for cases where clinical or imaging features suggest a more aggressive phenotype, or specifically for patients who are potential candidates for neoadjuvant therapy (e.g., those with Luminal B2/HER2+ or triple-negative subtypes). In these scenarios, a core needle biopsy remains the absolute standard of care to obtain essential biomarker information (most critically, HER2 status) necessary to guide neoadjuvant treatment decisions. For patients contraindicated for biopsy (e.g., coagulation disorders), the model could provide a non-invasive risk assessment to aid in clinical planning. Overall, the model is intended as a complementary decision-support tool, not a replacement for standard pathological diagnosis.

Future investigations should prioritize the refinement of feature extraction and selection methodologies to further enhance the diagnostic precision of multimodal models. Evaluating the impact of different modality combinations on model performance and comparing their respective AUC values could inform strategies to optimize clinical workflows, reduce examination burdens, and improve diagnostic efficiency. These advancements would bolster the efficacy of multimodal fusion and provide clinicians with streamlined and reliable tools for decision-making in complex oncologic cases.

This study has inherent limitations. Conducting the research within a single facility and relying on a relatively small sample size restricts the generalizability of these findings, underscoring the necessity of validating the model in larger, multicenter cohorts. To address the lack of external validation, we implemented rigorous cross-validation procedures. However, obtaining four-modal external validation datasets proved particularly challenging. Despite contacting multiple hospitals, insufficient sample sizes were available. We plan to expand our collaborative efforts to collect multi-institutional validation cohorts in future studies. Additionally, manual delineation of tumor ROIs by sonographers introduced potential variability due to subjective interpretation. While inter-observer agreement assessments partially mitigate this concern, future studies should explore automated segmentation techniques to reduce operatory bias and ensure reproducibility. Finally, the dataset imbalance, with non-luminal breast cancers (e.g., HER2-enriched and triple-negative types) comprising a smaller proportion of cases, necessitated oversampling, which may have introduced additional complexities. A comprehensive evaluation of all available machine learning algorithms was beyond the scope of this study and represents a worthwhile area for future research.

The deployment of this study across hospitals using diverse ultrasound platforms faces three primary barriers: inter-system variability—where differences in image resolution, dynamic range, and post-processing algorithms (e.g., Siemens vs. GE vs. Philips systems) may alter radiomic feature values and compromise model generalizability; lack of protocol standardization—where variations in scanning parameters (e.g., frequency, depth, gain settings) and operator-dependent acquisition techniques (e.g., probe pressure variations during SE) introduce feature instability; and ABVS dependence—given the model’s heavy reliance on ABVS coronal features (Section 3.4, SHAP analysis), leading to reduced accuracy in trimodal/four-modal configurations at hospitals without ABVS capabilities. To address these challenges, future work will focus on developing vendor-agnostic feature normalization pipelines, establishing IBSI-compliant imaging acquisition guidelines, and exploring domain adaptation techniques to align heterogeneous data distributions.

4.1 Conclusion

This study developed a multimodal machine learning model for preoperative differentiation of luminal and non-luminal breast cancer subtypes. By incorporating ABVS and SE, the model demonstrated enhanced generalization and predictive accuracy, highlighting the clinical value of multimodal ultrasound imaging. However, external validation in diverse populations is essential prior to clinical deployment.

Data availability statement

The data analyzed in this study is subject to the following licenses/restrictions: No, the datasets are not included in the article/supplementary material to protect participant privacy. Requests to access these datasets should be directed to Yan Fu, MTE1OTQ0OTAyMUBxcS5jb20=.

Ethics statement

The studies involving humans were approved by Full Name: Ethics Committee of Nantong University Affiliated Hospital Affiliation: Nantong University Affiliated Hospital, Nantong, Jiangsu Province, China. The studies were conducted in accordance with the local legislation and institutional requirements. Written informed consent for participation was not required from the participants or the participants’ legal guardians/next of kin in accordance with the national legislation and institutional requirements. Written informed consent was not obtained from the individual(s) for the publication of any potentially identifiable images or data included in this article because This retrospective study waived informed consent and was approved by the hospital review board. All patient-related information has been anonymized to safeguard privacy.

Author contributions

YF: Writing – review & editing, Conceptualization, Investigation, Methodology, Writing – original draft. HC: Writing – review & editing. HZ: Software, Writing – review & editing, Data curation. DL: Writing – review & editing, Data curation. XC: Data curation, Writing – review & editing. CQ: Data curation, Methodology, Software, Writing – review & editing. WL: Methodology, Writing – review & editing. HB: Writing – review & editing. QL: Writing – review & editing. GL: Writing – review & editing. ZS: Writing – review & editing. CG: Writing – review & editing. YZ: Writing – review & editing, Data curation, Methodology. XN: Supervision, Writing – review & editing.

Funding

The author(s) declare financial support was received for the research and/or publication of this article. This work was funded by the Nantong Health and Health Commission Project (No. MSZ2024010).

Acknowledgments

We sincerely thank our colleagues and friends for their invaluable support and constructive feedback throughout the research process. Special thanks to YuanPeng Zhang, WenWu Lu, and ChengYu Qiu for their assistance with linguistic editing and proofreading of this manuscript.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declare that no Generative AI was used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

1. Sung H, Ferlay J, Siegel RL, Laversanne M, Soerjomataram I, Jemal A, et al. Global cancer statistics 2020: Globocan estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA Cancer J Clin. (2021) 71:209–49. doi: 10.3322/caac.21660

PubMed Abstract | Crossref Full Text | Google Scholar

2. Ignatiadis M and Sotiriou C. Luminal breast cancer: From biology to treatment. Nat Rev Clin Oncol. (2013) 10:494–506. doi: 10.1038/nrclinonc.2013.124

PubMed Abstract | Crossref Full Text | Google Scholar

3. Rubin I and Yarden Y. The basic biology of Her2. Ann Oncol. (2001) 12 Suppl 1:S3–8. doi: 10.1093/annonc/12.suppl_1.s3

PubMed Abstract | Crossref Full Text | Google Scholar

4. Li X, Yang J, Peng L, Sahin AA, Huo L, Ward KC, et al. Triple-negative breast cancer has worse overall survival and cause-specific survival than non-triple-negative breast cancer. Breast Cancer Res Treat. (2017) 161:279–87. doi: 10.1007/s10549-016-4059-6

PubMed Abstract | Crossref Full Text | Google Scholar

5. Mann RM, Hooley R, Barr RG, and Moy L. Novel approaches to screening for breast cancer. Radiology. (2020) 297:266–85. doi: 10.1148/radiol.2020200172

PubMed Abstract | Crossref Full Text | Google Scholar

6. Pölcher M, Braun M, Tischitz M, Hamann M, Szeterlak N, Kriegmair A, et al. Concordance of the molecular subtype classification between core needle biopsy and surgical specimen in primary breast cancer. Arch Gynecol Obstet. (2021) 304:783–90. doi: 10.1007/s00404-021-05996-x

PubMed Abstract | Crossref Full Text | Google Scholar

7. Liu Z, Wang S, Dong D, Wei J, Fang C, Zhou X, et al. The applications of radiomics in precision diagnosis and treatment of oncology: Opportunities and challenges. Theranostics. (2019) 9:1303–22. doi: 10.7150/thno.30309

PubMed Abstract | Crossref Full Text | Google Scholar

8. Aerts HJWL. The potential of radiomic-based phenotyping in precision medicine: A review. JAMA Oncol. (2016) 2:1636–42. doi: 10.1001/jamaoncol.2016.2631

PubMed Abstract | Crossref Full Text | Google Scholar

9. Leithner D, Horvat JV, Marino MA, Bernard-Davila B, Jochelson MS, Ochoa-Albiztegui RE, et al. Radiomic signatures with contrast-enhanced magnetic resonance imaging for the assessment of breast cancer receptor status and molecular subtypes: Initial results. Breast Cancer Res. (2019) 21:106. doi: 10.1186/s13058-019-1187-z

PubMed Abstract | Crossref Full Text | Google Scholar

10. Lafcı O, Celepli P, Seher Öztekin P, and Koşar PN. Dce-Mri radiomics analysis in differentiating luminal a and luminal B breast cancer molecular subtypes. Acad Radiol. (2023) 30:22–9. doi: 10.1016/j.acra.2022.04.004

PubMed Abstract | Crossref Full Text | Google Scholar

11. Shi Z, Huang X, Cheng Z, Xu Z, Lin H, Liu C, et al. Mri-based quantification of intratumoral heterogeneity for predicting treatment response to neoadjuvant chemotherapy in breast cancer. Radiology. (2023) 308:e222830. doi: 10.1148/radiol.222830

PubMed Abstract | Crossref Full Text | Google Scholar

12. von Euler-Chelpin M, Lillholm M, Vejborg I, Nielsen M, and Lynge E. Sensitivity of screening mammography by density and texture: A cohort study from a Population-Based Screening Program in Denmark. Breast Cancer Res. (2019) 21:111. doi: 10.1186/s13058-019-1203-3

PubMed Abstract | Crossref Full Text | Google Scholar

13. He J, Chen WQ, Li N, Shen HB, Li J, Wang Y, et al. China guideline for the screening and early detection of female breast cancer(2021, Beijing). Zhonghua Zhong Liu Za Zhi. (2021) 43:357–82. doi: 10.3760/cma.j.cn112152-20210119-00061

PubMed Abstract | Crossref Full Text | Google Scholar

14. Rashmi S, Kamala S, Murthy SS, Kotha S, Rao YS, and Chaudhary KV. Predicting the molecular subtype of breast cancer based on mammography and ultrasound findings. Indian J Radiol Imaging. (2018) 28:354–61. doi: 10.4103/ijri.IJRI_78_18

PubMed Abstract | Crossref Full Text | Google Scholar

15. Irshad A, Leddy R, Pisano E, Baker N, Lewis M, Ackerman S, et al. Assessing the role of ultrasound in predicting the biological behavior of breast cancer. AJR Am J Roentgenol. (2013) 200:284–90. doi: 10.2214/AJR.12.8781

PubMed Abstract | Crossref Full Text | Google Scholar

16. Sultan LR, Bouzghar G, Levenback BJ, Faizi NA, Venkatesh SS, Conant EF, et al. Observer variability in bi-rads ultrasound features and its influence on computer-aided diagnosis of breast masses. Adv Breast Cancer Res. (2015) 4:1–8. doi: 10.4236/abcr.2015.41001

PubMed Abstract | Crossref Full Text | Google Scholar

17. Wu J, Ge L, Jin Y, Wang Y, Hu L, Xu D, et al. Development and validation of an ultrasound-based radiomics nomogram for predicting the luminal from non-luminal type in patients with breast carcinoma. Front Oncol. (2022) 12:993466. doi: 10.3389/fonc.2022.993466

PubMed Abstract | Crossref Full Text | Google Scholar

18. Yadav A, Kolekar M, and Zope M. (2024). ResNet-101 empowered deep learning for breast cancer ultrasound image classification, in: Proceedings of the 17th International Joint Conference on Biomedical Engineering Systems and Technologies - BIOSIGNALS, . pp. 763–9. SciTePress, ISBN: ISBN 978-989-758-688-0. doi: 10.5220/0012377800003657

Crossref Full Text | Google Scholar

19. Haribabu M, Guruviah V, and Yogarajah P. Recent advancements in multimodal medical image fusion techniques for better diagnosis: an overview. Curr Med Imaging. (2023) 19:673–94. doi: 10.2174/1573405618666220606161137

PubMed Abstract | Crossref Full Text | Google Scholar

20. Sigrist RMS, Liau J, Kaffas AE, Chammas MC, and Willmann JK. Ultrasound elastography: Review of techniques and clinical applications. Theranostics. (2017) 7:1303–29. doi: 10.7150/thno.18650

PubMed Abstract | Crossref Full Text | Google Scholar

21. Liu H, Wan J, Xu G, Xiang LH, Fang Y, Ding SS, et al. Conventional us and 2-D shear wave elastography of virtual touch tissue imaging quantification: Correlation with immunohistochemical subtypes of breast cancer. Ultrasound Med Biol. (2019) 45:2612–22. doi: 10.1016/j.ultrasmedbio.2019.06.421

PubMed Abstract | Crossref Full Text | Google Scholar

22. Zhu JY, HL H, XC J, HW B, and Chen F. Multimodal ultrasound features of breast cancers: correlation with molecular subtypes. BMC Med Imaging. (2023) 23:57. doi: 10.1186/s12880-023-00999-3

PubMed Abstract | Crossref Full Text | Google Scholar

23. Vourtsis A. Three-dimensional automated breast ultrasound: Technical aspects and first results. Diagn Interv Imaging. (2019) 100:579–92. doi: 10.1016/j.diii.2019.03.012

PubMed Abstract | Crossref Full Text | Google Scholar

24. Ma Q, Wang J, Xu D, Zhu C, Qin J, Wu Y, et al. Automatic breast volume scanner and B-ultrasound-based radiomics nomogram for clinician management of bi-rads 4a lesions. Acad Radiol. (2023) 30:1628–37. doi: 10.1016/j.acra.2022.11.002

PubMed Abstract | Crossref Full Text | Google Scholar

25. Ma Q, Shen C, Gao Y, Duan Y, Li W, Lu G, et al. Radiomics analysis of breast lesions in combination with coronal plane of Abvs and strain elastography. Breast Cancer (Dove Med Press). (2023) 15:381–90. doi: 10.2147/BCTT.S410356

PubMed Abstract | Crossref Full Text | Google Scholar

26. Li J, Cheng K, Wang S, Morstatter F, Trevino RP, Tang J, et al. Feature selection: A data perspective. ACM Comput Surv. (2018) 50:94. doi: 10.1145/3136625

Crossref Full Text | Google Scholar

27. Nogueira S, Sechidis K, and Brown G. On the stability of feature selection algorithms. J Mach Learn Res. (2017) 18:6345–98.

Google Scholar

28. Goldhirsch A, Wood WC, Coates AS, Gelber RD, Thürlimann B, Senn HJ, et al. Strategies for subtypes—Dealing with the diversity of breast cancer: Highlights of the St. Gallen international expert consensus on the primary therapy of early breast cancer 2011. Ann Oncol. (2011) 22:1736–47. doi: 10.1093/annonc/mdr304

PubMed Abstract | Crossref Full Text | Google Scholar

29. Koo TK and Li MY. A guideline of selecting and reporting intraclass correlation coefficients for reliability research. J Chiropr Med. (2016) 15:155–63. doi: 10.1016/j.jcm.2016.02.012

PubMed Abstract | Crossref Full Text | Google Scholar

30. Li B, Zheng X, Guo W, Wang Y, Mao R, Cheng X, et al. Radiation pneumonitis prediction using multi-omics fusion based on a novel machine learning pipeline. Hum-Cent Comput Inf Sci. (2022) 12:49.

Google Scholar

31. Lundberg SM and Lee S-I. A unified approach to interpreting model predictions. In: Neural information processing systems. United States: Curran Associates Inc. (2017).

Google Scholar

32. Li C, Huang H, Chen Y, Shao S, Chen J, Wu R, et al. Preoperative non-invasive prediction of breast cancer molecular subtypes with a deep convolutional neural network on ultrasound images. Front Oncol. (2022) 12:848790. doi: 10.3389/fonc.2022.848790

PubMed Abstract | Crossref Full Text | Google Scholar

33. Davey MG, Davey MS, Boland MR, Ryan ÉJ, Lowery AJ, and Kerin MJ. Radiomic differentiation of breast cancer molecular subtypes using preoperative breast imaging – A systematic review and meta-analysis. Eur J Radiol. (2021) 144:109996. doi: 10.1016/j.ejrad.2021.109996

PubMed Abstract | Crossref Full Text | Google Scholar

34. Ma W, Zhao Y, Ji Y, Guo X, Jian X, Liu P, et al. Breast cancer molecular subtype prediction by mammographic radiomic features. Acad Radiol. (2019) 26:196–201. doi: 10.1016/j.acra.2018.01.023

PubMed Abstract | Crossref Full Text | Google Scholar

35. Sun R, Hou X, Li X, Xie Y, and Nie S. Transfer learning strategy based on unsupervised learning and ensemble learning for breast cancer molecular subtype prediction using dynamic contrast-enhanced mri. J Magn Reson Imaging. (2022) 55:1518–34. doi: 10.1002/jmri.27955

PubMed Abstract | Crossref Full Text | Google Scholar

36. Bakker MAG, MdL O, Matela N, and Mota AM. Decoding breast cancer: using radiomics to non-invasively unveil molecular subtypes directly from mammographic images. J Imaging.(2024) 10:218. doi: 10.3390/jimaging10090218

PubMed Abstract | Crossref Full Text | Google Scholar

37. Lai S, Liang F, Zhang W, Zhao Y, Li J, Zhao Y, et al. Evaluation of molecular receptors status in breast cancer using an mpMRI-based feature fusion radiomics model: mimicking radiologists’ diagnosis. Front Oncol. (2023) 13:1219071. doi: 10.3389/fonc.2023.1219071

PubMed Abstract | Crossref Full Text | Google Scholar

38. Rella R, Belli P, Giuliani M, Bufi E, Carlino G, Rinaldi P, et al. Automated breast ultrasonography (Abus) in the screening and diagnostic setting: Indications and practical use. Acad Radiol. (2018) 25:1457–70. doi: 10.1016/j.acra.2018.02.014

PubMed Abstract | Crossref Full Text | Google Scholar

39. Girometti R, Zanotel M, Londero V, Linda A, Lorenzon M, and Zuiani C. Automated breast volume scanner (Abvs) in assessing breast cancer size: A comparison with conventional ultrasound and magnetic resonance imaging. Eur Radiol. (2018) 28:1000–8. doi: 10.1007/s00330-017-5074-7

PubMed Abstract | Crossref Full Text | Google Scholar

40. D’Angelo A, Orlandi A, Bufi E, Mercogliano S, Belli P, and Manfredi R. Automated breast volume scanner (Abvs) compared to handheld ultrasound (Hhus) and contrast-enhanced magnetic resonance imaging (Ce-Mri) in the early assessment of breast cancer during neoadjuvant chemotherapy: An emerging role to monitoring tumor response? Radiol Med. (2021) 126:517–26. doi: 10.1007/s11547-020-01319-3

PubMed Abstract | Crossref Full Text | Google Scholar

41. Ma M, Liu R, Wen C, Xu W, Xu Z, Wang S, et al. Predicting the molecular subtype of breast cancer and identifying interpreta ble imaging features using machine learning algorithms. Eur Radiol. (2022) 32:1652–62. doi: 10.1007/s00330-021-08271-4

PubMed Abstract | Crossref Full Text | Google Scholar

42. Ghosh S, Dasgupta A, and Swetapadma A. A study on support vector machine based linear and nonlinear pattern classification International Conference on Intelligent Sustainable Systems (ICISS. New Jersey, USA: Institute of Electrical and Electronics Engineers (IEEE) (2019). doi: 10.1109/ISS1.2019.8908018.

Crossref Full Text | Google Scholar

43. Gholami R and Fakhari N. Chapter 27. Support vector machine: Principles, parameters, and applications. In: Samui P, Sekhar S, and Balas VE, editors. Handbook of neutral computation. Academic Press (2017). p. 515–35. doi: 10.1016/B978-0-12-811318-9.00027-2

Crossref Full Text | Google Scholar

Keywords: breast cancer, machine learning, ultrasound diagnostics, radiomics, molecular subtypes, multimodal imaging, automated breast volume scanner (ABVS)

Citation: Fu Y, Chen HJ, Zhang H, Liu DJ, Chen X, Qiu CY, Lu WW, Bai HM, Li QW, Li GX, Shen ZJ, Gu CJ, Zhang YP and Ni XJ (2025) Integrating multimodal ultrasound imaging and machine learning for predicting luminal and non-luminal breast cancer subtypes. Front. Oncol. 15:1558880. doi: 10.3389/fonc.2025.1558880

Received: 13 April 2025; Accepted: 10 September 2025;
Published: 08 October 2025.

Edited by:

Hatem A. Rashwan, University of Rovira i Virgili, Spain

Reviewed by:

Ana Margarida Mota, University of Lisbon, Portugal
Agnesh Yadav, Indian Institute of Technology Patna, India
Dinghao Guo, Northeastern University, China

Copyright © 2025 Fu, Chen, Zhang, Liu, Chen, Qiu, Lu, Bai, Li, Li, Shen, Gu, Zhang and Ni. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Chang Jiang Gu, Y2hhbmdqaWFuZzA1OEBmb3htYWlsLmNvbQ==; Yuan Peng Zhang, eS5wLnpoYW5nQGllZWUub3Jn; Xue Jun Ni, ZHlmbnhqMjEzQDE2My5jb20=

These authors have contributed equally to this work and share first authorship

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.