Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Oncol., 06 June 2025

Sec. Breast Cancer

Volume 15 - 2025 | https://doi.org/10.3389/fonc.2025.1536365

This article is part of the Research TopicThe Essential Role of Multidisciplinary Teams in Breast Cancer Surgery: Collaboration for Superior Patient OutcomesView all 6 articles

Optimizing breast cancer ultrasound diagnosis: a comparative study of AI model performance and image resolution

Yunqing Yin&#x;Yunqing Yin1†Junkui Fang&#x;Junkui Fang1†Wei Zhang*Wei Zhang2*Xinying Shen*Xinying Shen2*
  • 1The Second Clinical Medical College, Jinan University, Shenzhen, China
  • 2Department of Interventional Radiology, Shenzhen People’s Hospital, Shenzhen, China

Objectives: To determine the optimal combination of artificial intelligence (AI) models and ultrasound (US) image resolutions for breast cancer diagnosis and evaluate whether this combination surpasses the diagnostic accuracy of senior radiologists.

Materials and methods: We systematically compared lightweight (MobileNet, Xception) and dense neural networks (ResNet50, DenseNet121) using three image resolutions (224 × 224, 320 × 320, 448 × 448 pixels). A retrospective cohort of 4,998 patients was divided into training/validation (8:2 ratio, n = 3,578) and independent testing sets (n = 1,410). Diagnostic performance was assessed via AUC, sensitivity, specificity, and analysis speed, with direct comparisons against senior radiologists.

Results: MobileNet with 224 × 224 input achieved the highest AUC (0.924, 95% CI: 0.910–0.938) and accuracy (87.3%) outperforming senior US (AUC: 0.820, accuracy: 79.1%) and mammography doctors (AUC: 0.819, accuracy: 83.6%) (p < 0.05). After excluding BI-RADS 4c and 5 nodules, the diagnostic efficacy of MobileNet_224 is better than that of senior doctors (p < 0.05), can reduce 60.1% false positives of US, and 46.6% of mammography. MobileNet_224 and MobileNet_320 had the fastest analysis speed.

Conclusion: MobileNet_224 represents a novel, efficient AI framework for breast cancer diagnosis demonstrating superior accuracy and speed compared to both complex AI models and experienced clinicians. This work highlights the critical role of optimizing model architecture and resolution to enhance diagnostic workflows and reduce unnecessary biopsies.

Introduction

Breast cancer (BC) has emerged as the most prevalent malignancy worldwide and is a leading cause of death among women (1, 2). Surpassing lung cancer, it now accounts for over 2.3 million new cases annually representing 30% of all female cancers and 11.7% of all cancers. This malignancy increasingly affects a younger demographic posing a serious threat to women’s health (3).

Currently, clinical diagnosis of BC incorporates various methods, including palpation, digital mammography (DM), magnetic resonance imaging (MRI), and ultrasound (US). Mammography, while widely used, often suffers from high rates of false positives and negatives, particularly in women with dense breast tissue, leading to missed diagnoses (4, 5). MRI is recommended for high-risk BC patients, but its high cost, false-positive rate, and time intensity limit its use to a supplementary role in mammography. US, an important tool for BC screening, is not constrained by mammary gland tissue types and has been shown to increase BC detection rates by 17% while reducing unnecessary biopsies (6, 7). However, US is limited by its reliance on the acoustic impedance difference in tumor tissues making it challenging to differentiate diagnoses, especially in cases of non-mass BC (8). The operator-dependent nature of US also means that diagnostic outcomes can vary significantly based on the experience of the practitioner (9).

The Breast Imaging Reporting and Data System (BI-RADS) has significantly improved the standardization and accuracy of breast tumor diagnosis (10). However, BI-RADS classification relies on visual recognition, which can miss subtle image features. Thus, there is an urgent need for an objective method that minimizes operator dependence and accurately reflects tumor characteristics for BC screening and diagnosis (11).

Advances in AI-driven breast cancer classification have demonstrated significant potential in reducing diagnostic variability and improving clinical workflows. Recent studies, such as those employing convolutional neural network for ultrasound-based classification (1214), underscore the feasibility of AI in standardizing diagnoses. Furthermore, ensemble machine learning techniques (15, 16) demonstrate improved accuracy through model aggregation. However, these works often lack systematic comparisons across model architectures limiting insights into optimal computational frameworks. While capsule networks (17) show promise in capturing spatial hierarchies within tumor morphology, their computational inefficiency hinders real-time clinical deployment compared to lightweight CNNs. Lightweight architectures, like MobileNet variants (18, 19) have emerged as efficient alternatives in cancer classification, yet prior investigations rarely explore resolution-specific trade-offs or benchmark against both complex models [e.g., ShuffleNet (20), EfficientNet (21)] and human expertise. Concurrently, multi-resolution approaches for medical image segmentation (22, 23) highlight the importance of scale optimization, though their focus remains isolated from end-to-end diagnostic pipelines.

Our study addresses these gaps by systematically comparing lightweight and dense neural networks across resolutions to identify the optimal AI–image combination for breast cancer detection, while directly benchmarking diagnostic efficiency against senior radiologists—thereby advancing clinical standards through technically validated innovation.

Materials and methods

Study population

This retrospective study was conducted following approval from the institutional review board of Shenzhen People’s Hospital, with a waiver for informed consent due to its retrospective nature. The Na-exclusion criteria for this study were as follows:

Inclusion criteria: (1) Breast tumors were detected by US, which were classified as 0, 3, 4a, 4b, 4c, or 5 according to BI-RADS. (2) At least 3.0-mm breast tissue can be displayed around the nodule. (3) No intervention or operation was performed on the nodule to be evaluated before ultrasonic examination. (4) Patients underwent surgery or biopsy within 1 week of ultrasonic data collection and obtained pathological results.

Exclusion criteria: (1) BIRADS 1 and 2; (2) Have a history of breast surgery or intervention; (3) Poor image quality; (4) The clinical data of cases are incomplete, and the pathological results are not tracked.

In this study, following the inclusion and exclusion criteria, a cohort of 4,998 patients with breast tumors was established. These patients were then randomly divided into the following three groups: a training set, a test set, and an independent validation set. The training and validation sets were allocated in an 8:2 ratio, with the training set comprising 2,778 patients (774 with malignant tumors) and the validation set including 800 patients (217 with malignant tumors). The independent test set consisted of 1,410 patients of whom 579 had malignant tumors (Table 1). All patients underwent biopsy or surgical procedures for pathological diagnosis (Figure 1).

Table 1
www.frontiersin.org

Table 1. Patient information in this study.

Figure 1
www.frontiersin.org

Figure 1. Flow chart and results of this study. The optimal model: MobileNet_224, senior ultrasound doctors, senior mammography doctors. MobileNet_224, MobileNet with 224 × 224-pixel image input; US_BI-RADS, senior ultrasound doctors’ diagnostic results; DM_BI-RADS, senior mammography doctors’ diagnostic results.

Human examination

The US and mammography diagnosis were made by two senior doctors, with more than 10 years of experience in mammography diagnosis, who made the classification diagnosis of benign and malignant tumors under the condition of independent double blind, and gave the BI-RADS classification of tumors. In case of inconsistency, the third chief physician shall be invited for arbitration (Supplementary Material). In a comparison of diagnostic performance, BI-RADS classifications 3 and 4A are defined as benign lesions, and 4B, 4C, and 5 are defined as malignant lesions. Diagnostic results from ultrasound doctors and mammography doctors are based on the doctor’s experience.

AI model construction

Model selection was guided by (1) computational efficiency for clinical deployment, (2) prior evidence in medical imaging, and (3) architectural diversity to benchmark lightweight against dense networks. MobileNet and Xception were prioritized for their parameter efficiency and validated performance in resource-constrained tasks. DenseNet121 and ResNet50 served as benchmarks for hierarchical feature extraction.

These models employ architectural innovations like depthwise separable convolutions to minimize computational burden while retaining diagnostic accuracy. Conversely, dense models, like DenseNet121 and ResNet50—known for their complex hierarchical structures (e.g., residual blocks in ResNet50, dense connectivity in DenseNet121)—were included to evaluate their ability to capture nuanced tumor features in ultrasound images. By comparing these fundamentally distinct architectures, we aimed to identify the optimal balance between computational efficiency and diagnostic precision for breast cancer detection. By comparing these models, we aimed to assess which architecture is more effective for the task of diagnosing BC from US images.

We employed the following three different image resolutions: 224 × 224, 320 × 320, and 448 × 448 pixels (illustrated in Figure 2). This variation in resolution was intended to examine the impact of image quality on the diagnostic accuracy of the AI models. Higher-resolution images typically provide more detailed information but also require more computational resources to process. Conversely, lower-resolution images are faster to process but may lack some detailed information. Understanding the trade-off between resolution and diagnostic accuracy is crucial for the practical application of AI in medical imaging, particularly in settings where computational resources are limited.

Figure 2
www.frontiersin.org

Figure 2. Comparison of diagnostic efficacy between LW-CNNs in the testing set. AUC, area under the curve; 95% CI: 95% confidence interval. (a) Xception_224: Xception with 224 × 224-pixel image input, (b) MobileNet_224, (c) Xception_320, (d) MobileNet_320, (e) Xception_448, (f) MobileNet_448.

Training protocol

Models were implemented in TensorFlow 2.5.0, trained on an NVIDIA RTX 3090 GPU, and evaluated on an edge-computing device (Jetson AGX Xavier) to simulate clinical deployment. Images were standardized to 224 × 224, 320 ×3 20, or 448 × 448 pixels. Training employed AdamW optimization (lr = 1e−4) with cosine annealing, and cross-entropy loss weighted for class imbalance (Supplementary Material).

To ensure the integrity and non-overlapping nature of our data, we carefully allocated images from the same patient exclusively to one dataset—either the training set or the validation set. This approach was critical to prevent data leakage and ensure that the models were tested on completely unseen data, thereby providing a reliable assessment of their generalizability.

The independent testing set was crucial for evaluating the real-world applicability of the AI models. It consisted of the following three main components: 1) Comparative Evaluation: We assessed the diagnostic effectiveness between different AI models to identify the optimal model and image resolution combination. 2) Comparison with Senior Doctors: The optimal AI model’s diagnostic efficacy was compared with that of experienced senior US and mammography doctors. This comparison aimed to benchmark the AI models against the current gold standard in clinical practice. 3) Exclusion of Certain Tumor Types: We specifically excluded BI-RADS 4c and 5 tumors to focus on challenging cases where AI models could potentially offer the most significant benefit. This step was crucial to understand the potential of AI in improving diagnostic specificity and reducing false positives.

Statistical analysis

Continuous variable data are expressed as mean ± standard deviation. Categorical variable data are expressed as a percentage. The paired-sample t-test was used to compare the differences within the group. R 3.6.3 was used for the statistical analysis. Diagnostic performance was evaluated using receiver operating characteristic (ROC) curves generated in R 3.6.3 (pROC package). The area under the curve (AUC), reflecting overall discriminative ability, was calculated via the non-parametric DeLong method, with 95% confidence intervals (95% CI) derived from 2,000 stratified bootstrap replicates to account for variability. Sensitivity, specificity, and accuracy were computed from confusion matrices. Statistical significance of AUC differences between models and radiologists was assessed via DeLong’s test (p < 0.05).

Code availability

The updated code repository and Jupyter notebook was hosted on GitHub—https://github.com/wukaiyeah/ultrasound_breast_malignant_classification.git.

Results

Diagnostic performance of AI models vs. radiologists

MobileNet_224 demonstrated superior diagnostic accuracy compared to both other AI models and senior radiologists. In the independent testing set, MobileNet_224 achieved an AUC of 0.924 (95% CI: 0.910–0.938) significantly outperforming senior ultrasound radiologists (AUC: 0.820, p < 0.001) and mammography specialists (AUC: 0.819, p < 0.001). Its accuracy (87.3%) surpassed radiologists’ performance by 8.2% (ultrasound) and 3.7% (mammography). Dense networks, such as DenseNet121_448, showed lower efficacy (AUC: 0.890; accuracy: 82.8%) highlighting the advantage of lightweight architectures (Tables 2, 3, Figures 2, 3).

Table 2
www.frontiersin.org

Table 2. Comparison of the efficacy of AI model in the independent testing set.

Table 3
www.frontiersin.org

Table 3. Results of MobileNet_224 and ultrasound/mammography in testing set.

Figure 3
www.frontiersin.org

Figure 3. Comparison of diagnostic efficacy between DNNs in the testing set. AUC, area under the curve; 95% CI, 95% confidence interval. (a) ResNet50_224: ResNet50 with 224 × 224-pixel image input, (b) DenseNet121_224, (c) ResNet50_320, (d) DenseNet121_320, (e) ResNet50_448, (f) DenseNet121_448.

The interpretability analysis (Figure 4) demonstrates MobileNet_224’s alignment with radiological diagnostic criteria. For benign lesions (A), SHAP values identified smooth margins and homogeneous echotexture as primary contributors to classification, while Grad-CAM heatmaps (C) confirmed focused attention on lesion boundaries. In malignant cases (B), SHAP attributed high malignancy probability to spiculated margins and heterogeneous internal echoes corroborated by Grad-CAM’s emphasis on irregular tumor peripheries (D).

Figure 4
www.frontiersin.org

Figure 4. Interpretability analysis of MobileNet_224 predictions for benign and malignant breast lesions. (A) Benign lesion: prediction probability (0.999 for benign, 0.001 for malignant) with SHAP values highlighting key image regions contributing to the benign classification. (B) Malignant lesion: prediction probability (0.999 for malignant, 0.001 for benign) with SHAP values emphasizing tumor margin irregularity and microcalcifications. (C, D) Grad-CAM heatmaps for the benign (C) and malignant (D) lesions illustrating the model’s focus on clinically relevant anatomical features (e.g., smooth margins in benign vs. spiculated regions in malignant).

Impact of image resolution on model performance

Lower-resolution inputs (224 × 224 pixels) consistently outperformed higher resolutions (320 × 320, 448 × 448) across all models. MobileNet_224 achieved the highest AUC (0.924) at 224 × 224, while its performance declined at 448 × 448 (AUC: 0.909). Similarly, Xception_224 (AUC: 0.918) surpassed Xception_448 (AUC: 0.909), despite the latter utilizing more detailed imaging data. This suggests that lower resolutions prioritize clinically decisive features over extraneous textures optimizing both accuracy and computational efficiency (Table 2).

Reduction of false positives and clinical implications

MobileNet_224 significantly reduced false-positive diagnoses compared to radiologists: False positives decreased from 286 to 114 cases (60.1% reduction) for ultrasound, and false positives dropped from 204 to 109 cases (46.6% reduction) for mammography. Notably, after excluding BI-RADS 4c/5 cases (high malignancy likelihood), the model maintained superior specificity (88.8% vs. radiologists’ 65.6%, p < 0.001) demonstrating its ability to resolve diagnostically challenging lesions (Figures 5, 6).

Figure 5
www.frontiersin.org

Figure 5. Comparison of diagnostic efficacy between the optimal model and senior doctors in the testing set. MobileNet_224, MobileNet with 224 × 224-pixel image input; AUC, area under the curve; 95% CI, 95% confidence interval. (A) The optimal model: MobileNet_224, (B) senior ultrasound doctors, (C) senior mammography doctors.

Figure 6
www.frontiersin.org

Figure 6. Comparison of diagnostic efficacy between the optimal model and senior doctors after excluding BIRADS 4c and 5 nodules. MobileNet_224, MobileNet with 224 × 224-pixel image input; AUC, area under the curve; 95% CI, 95% confidence interval. (A) MobileNet_224, (B) senior ultrasound doctors, (C) MobileNet_224, (D) senior mammography doctors.

Computational efficiency

MobileNet_224 exhibited the fastest inference speed (0.02 s per image), 3.8× faster than DenseNet121_448 (0.076 s) and 500× faster than manual radiologist review (~10 s per case). This efficiency did not compromise accuracy reinforcing its suitability for real-time clinical workflows (Table 4).

Table 4
www.frontiersin.org

Table 4. The average time of analyzing a single ultrasound image with different AI models (s).

Discussion

AI has demonstrated remarkable versatility across diverse domains, from anemia detection using palm and conjunctiva images (2426) to macroeconomic forecasting via time-series models (27). In healthcare, lightweight convolutional neural networks (CNNs) are increasingly applied to resource-constrained tasks, such as MobileNet for diabetic retinopathy screening (28) and Xception for COVID-19 detection (29). Our study extends this paradigm to breast cancer ultrasound diagnosis, where optimizing existing architectures—rather than developing novel models—proves critical for clinical translation.

This study selects four models: Xception, MobileNet, DensNet121, and ResNet50, and 224 × 224-, 320 × 320-, and 448 × 448-pixel image input to explore the accuracy of breast tumors with US images. The results show that MobileNet_224 is superior to the other 11 models and the combination of input images, with an AUC of 0.924 and an accuracy of 87.3%, which are superior to those of senior US and mammography doctors (AUC: 0.820 and 0.819; accuracy: 79.1% and 83.6%).

The application of AI in medical images mainly uses convolutional neural network (CNN) to extract useful information from images. CNN has the following two characteristics: (1) can effectively reduce the dimension of images and (2) can effectively preserve features of images. There are many models derived from this, which are mainly divided into the following two categories: (1) dense neural network (DNN) such as ResNet, DenseNet, and EfficientNet (30, 31); (2) lightweight revolutionary neural networks (LW-CNNs) (32) such as MobileNet, Xception, and ShuffleNet. Large-scale network has a large amount of computation, but the processing speed is slow. LW-CNNs has designed a more efficient network computing method, which not only reduces the number of network layers and parameters but also preserves the performance. It can be used for fast reasoning of embedded and mobile systems. It has a CNN structure with high computational efficiency, adopts point-to-point grouping convolution and channel shuffling, which greatly reduce the amount of computation while maintaining accuracy, and maximize operation speed and accuracy (33, 34). In this study, the diagnostic efficiency of the two LW-CNNs is generally higher than that of the DNNs.

MobileNet (35) was based on a streamlined architecture, and a lightweight deep neural network is constructed using longitudinally separable convolution. Its core idea is that the deep separable convolution replaces the standard convolution and reduces the number of parameters (36, 37). In this study, MobileNet_224 shows the best diagnostic efficiency in different models and images.

Generally, image dimensionality reduction will not affect the final result, such as a picture of 1,000 × 1,000 pixels was reduced to 200 × 200 pixels, which has no obvious impact on the computer recognition results. Among MobileNet with different image resolutions, MobileNet_224 is superior to MobileNet_320 (AUC: 0.891) and MobileNet_448 (AUC: 0.909). The results suggest that MobileNet can still extract the information needed for diagnosis after the image dimension is reduced, which is consistent with the original intention of model design and other studies (38, 39).

A high-resolution image contains more information and larger pixel matrix, but it takes up more memory. In the convolution operation, the large size consumes more computing time than the small size. This study found that the resolution has an impact on the time consumption of the model, and the time consumption of high-resolution model analysis increases, which is consistent with other studies (40). On the contrary, small images consume less computational resources, but may lose some information and may produce misleading results. Therefore, deep learning needs to compromise the contradiction between computational efficiency and recognition accuracy (41). The DNNs, such as DenseNet, ResNet50, and EfficientNetB0, have dense connections between layers and are more memory and time consuming (42). This study shows that DenseNet121_224 takes the longest time in analyzing a single picture, which is 0.07 s, while MobileNet _224 takes less than 0.02 s.

According to BI-RADS classification, Class 0 is a lesion that cannot be determined qualitatively, which has not been diagnosed but has been suspected by doctors, and the possibility of malignancy of BI-RADS 3, 4a, and 4b tumors is less than 2%, 2%–10%, and 10%–50%, respectively. According to BI-RADS 3, follow-up is recommended, and biopsy is recommended for 4a and 4b. If benign tumors can be further screened by AI method, unnecessary puncture and injury can be reduced. A study (43) reported that using the trained AI model to identify benign and malignant breast tumors was higher than the diagnostic level of doctors, the AUC of which were 0.87 (95% CI: 0.79–0.95) and 0.51 (95% CI: 0.50–0.53), respectively. In this study, the AUC of MobileNet_224 [0.886 (95% CI: 0.854–0.917)] is higher than that of senior US doctors [0.820 (95% CI: 0.803–0.837]]. Compared with senior mammography, the AUC of MobileNet_224 [0.915 (95% CI: 0.892–0.937)] is higher than that of senior mammography doctors [0.745 (95% CI: 0.714–0.777)]. To further clarify the diagnostic efficiency of AI technology, this study selected breast tumors that are difficult to diagnose using US and mammography for analysis and found that when the cut-off value of MobileNet_224 is 0.467, the diagnostic accuracy is higher than that of senior doctors in US and mammography. The model significantly reduced false positives in both ultrasound (60.1% reduction) and mammography (46.6% reduction), while improving specificity and overall accuracy (AUC increase: 6.6% for ultrasound; accuracy increase: 6.1%). The application of MobileNet_224 demonstrated significant improvements in diagnostic performance. Specifically, the number of false positives in ultrasound (US) imaging was reduced from 286 to 114 cases representing a 60.1% reduction. For mammography, the model increased the AUC and accuracy by 17% and 11.3%, respectively. Furthermore, the model reduced false positives in mammography from 204 to 109 cases, a decrease of 46.6%. These results highlight MobileNet_224’s capability to diagnose early-stage breast cancer (BC), minimize false positives, and reduce unnecessary biopsies. Contrary to the assumption that higher image resolution universally improves diagnostic accuracy, our findings reveal that MobileNet_224 achieves superior performance at a 224 × 224 resolution. This challenges the prevailing trend in medical AI toward computationally intensive high-resolution frameworks.

The use of AUC, sensitivity, and specificity is widely accepted in oncology AI studies, while analysis speed addresses practical deployment needs. By excluding BI-RADS 4c/5 cases (high malignancy likelihood), we specifically tested the model’s ability to resolve ambiguous diagnoses—a key clinical challenge.

Our study has several limitations. First, this study is a single-center and retrospective study. In the future, a multi-center prospective AI study should be carried out to confirm the reliability of the screening model of this study. Second, this study does not distinguish the types of US instruments and equipment, but only analyzes the static US images. The accuracy and reliability of AI technology for video data analysis need to be studied further. Last, our study focused on evaluating existing lightweight models for clinical deployment capability rather than proposing novel architectures, which limits direct comparisons with cutting-edge frameworks but prioritizes real-world practicality. Future research could expand comparisons to hybrid models, such as CNN-Transformer frameworks, to evaluate their potential for multi-scale feature extraction in breast cancer diagnosis.

Conclusion

This study systematically evaluates the diagnostic performance of lightweight AI models (MobileNet, Xception) versus dense networks (ResNet50, DenseNet121) across ultrasound image resolutions (224 × 224, 320 × 320, 448 × 448) for breast cancer detection. Using a retrospective cohort of 4,998 patients, we demonstrate that MobileNet_224, despite its computational simplicity, achieves superior clinical utility as follows: 1) Speed–Accuracy Balance: MobileNet_224 processes images in 0.02 s—300× faster than manual review—while maintaining 87.3% accuracy addressing critical workflow bottlenecks. 2) False-Positive Reduction: The model reduces unnecessary biopsies by 60.1% in ultrasound and 46.6% in mammography directly impacting patient outcomes and healthcare costs. 3) Resolution Optimization Framework: Lower resolutions (224 × 224) suffice for accurate diagnosis challenging the need for resource-intensive high-resolution pipelines. These findings advocate for redefining clinical AI benchmarks toward deployment capability rather than theoretical performance offering a pragmatic framework for healthcare translation.

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Ethics statement

Ethical approval was not required for the study involving humans in accordance with the local legislation and institutional requirements. Written informed consent to participate in this study was not required from the participants or the participants’ legal guardians/next of kin in accordance with the national legislation and the institutional requirements.

Author contributions

YY: Conceptualization, Investigation, Writing – original draft, Writing – review & editing. JF: Data curation, Investigation, Methodology, Writing – original draft, Writing – review & editing. WZ: Funding acquisition, Project administration, Resources, Validation, Writing – review & editing. XS: Formal Analysis, Project administration, Resources, Supervision, Validation, Visualization, Writing – review & editing.

Funding

The author(s) declare that financial support was received for the research and/or publication of this article. This project was supported by the Clinical Scientist Training Program of Shenzhen People’s Hospital (SYWGSCGZH202202) and the Department of Science and Technology of Guangdong Province (Project No. 2024A0505040020).

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declare that no Generative AI was used in the creation of this manuscript.

Correction Note

This article has been corrected with minor changes. These changes do not impact the scientific content of the article.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fonc.2025.1536365/full#supplementary-material

References

1. Chen W, Zheng R, Zhang S, Zeng H, Xia C, Zuo T, et al. Cancer incidence and mortality in China, 2013. Cancer Lett. (2017) 401:63–71. doi: 10.1016/j.canlet.2017.04.024

PubMed Abstract | Crossref Full Text | Google Scholar

2. Miller KD, Nogueira L, Devasia T, Mariotto AB, Yabroff KR, Jemal A, et al. Cancer treatment and survivorship statistics, 2022. CA: A Cancer J For Clinicians. (2022) 72:409–36. doi: 10.3322/caac.21731

PubMed Abstract | Crossref Full Text | Google Scholar

3. Makama M, Drukker CA, Rutgers EJT, Slaets L, Cardoso F, Rookus MA, et al. An association study of established breast cancer reproductive and lifestyle risk factors with tumour subtype defined by the prognostic 70-gene expression signature (MammaPrint®). Eur J Cancer (Oxford England: 1990). (2017) 75:5–13. doi: 10.1016/j.ejca.2016.12.024

PubMed Abstract | Crossref Full Text | Google Scholar

4. Jin Z-Q, Lin M-Y, Hao W-Q, Jiang H-T, Zhang L, Hu W-H, et al. Diagnostic evaluation of ductal carcinoma in situ of the breast: ultrasonographic, mammographic and histopathologic correlations. Ultrasound In Med Biol. (2015) 41:47–55. doi: 10.1016/j.ultrasmedbio.2014.09.023

PubMed Abstract | Crossref Full Text | Google Scholar

5. Su X, Lin Q, Cui C, Xu W, Wei Z, Fei J, et al. Non-calcified ductal carcinoma in situ of the breast: comparison of diagnostic accuracy of digital breast tomosynthesis, digital mammography, and ultrasonography. Breast Cancer (Tokyo Japan). (2017) 24:562–70. doi: 10.1007/s12282-016-0739-7

PubMed Abstract | Crossref Full Text | Google Scholar

6. Schneble EJ, Graham LJ, Shupe MP, Flynt FL, Banks KP, Kirkpatrick AD, et al. Future directions for the early detection of recurrent breast cancer. J Cancer. (2014) 5:291–300. doi: 10.7150/jca.8017

PubMed Abstract | Crossref Full Text | Google Scholar

7. Niell BL, Freer PE, Weinfurtner RJ, Arleo EK, and Drukteinis JS. Screening for breast cancer. Radiologic Clinics. (2017) 55:1145–62. doi: 10.1016/j.rcl.2017.06.004

PubMed Abstract | Crossref Full Text | Google Scholar

8. Osako T, Takahashi K, Iwase T, Iijima K, Miyagi Y, Nishimura S, et al. Diagnostic ultrasonography and mammography for invasive and noninvasive breast cancer in women aged 30 to 39 years. Breast Cancer (Tokyo Japan). (2007) 14:229–33. doi: 10.2325/jbcs.891

PubMed Abstract | Crossref Full Text | Google Scholar

9. Spick C, Schernthaner M, Pinker K, Kapetas P, Bernathova M, Polanec SH, et al. MR-guided vacuum-assisted breast biopsy of MRI-only lesions: a single center experience. Eur Radiol. (2016) 26:3908–16. doi: 10.1007/s00330-016-4267-9

PubMed Abstract | Crossref Full Text | Google Scholar

10. Aswathy M and Jagannath M. Detection of breast cancer on digital histopathology images: Present status and future possibilities. Inf Med Unlocked. (2017) 8:74–9. doi: 10.1016/j.imu.2016.11.001

Crossref Full Text | Google Scholar

11. Liberman L and Menell JH. Breast imaging reporting and data system (BI-RADS). Radiologic Clinics. (2002) 40:409–30. doi: 10.1016/S0033-8389(01)00017-3

PubMed Abstract | Crossref Full Text | Google Scholar

12. Afrifa S, Varadarajan V, Appiahene P, and Zhang T. A novel artificial intelligence techniques for women breast cancer classification using ultrasound images. CEOG. (2023) 50:271–84. doi: 10.31083/j.ceog5012271

Crossref Full Text | Google Scholar

13. Kaddes M, Ayid YM, Elshewey AM, and Fouad Y. Breast cancer classification based on hybrid CNN with LSTM model. Sci Rep. (2025) 15:4409. doi: 10.1038/s41598-025-88459-6

PubMed Abstract | Crossref Full Text | Google Scholar

14. Chung M, Calabrese E, Mongan J, Ray KM, Hayward JH, Kelil T, et al. Deep learning to simulate contrast-enhanced breast MRI of invasive breast cancer. Radiology. (2023) 306:e213199. doi: 10.1148/radiol.213199

PubMed Abstract | Crossref Full Text | Google Scholar

15. Afrifa S, Varadarajan V, Appiahene P, Zhang T, and Domfeh EA. Ensemble machine learning techniques for accurate and efficient detection of botnet attacks in connected computers. Eng. (2023) 4:650–64. doi: 10.3390/eng4010039

Crossref Full Text | Google Scholar

16. Wang S-R, Cao C-L, Du T-T, Wang J-L, Li J, Li W-X, et al. Machine learning model for predicting axillary lymph node metastasis in clinically node positive breast cancer based on peritumoral ultrasound radiomics and SHAP feature analysis. J Ultrasound Med. (2024) 43:1611–25. doi: 10.1002/jum.16483

PubMed Abstract | Crossref Full Text | Google Scholar

17. Kavitha T, Mathai PP, Karthikeyan C, Ashok M, Kohar R, Avanija J, et al. Deep learning based capsule neural network model for breast cancer diagnosis using mammogram images. Interdiscip Sciences: Comput Life Sci. (2022) 14:113–29. doi: 10.1007/s12539-021-00467-y

PubMed Abstract | Crossref Full Text | Google Scholar

18. Huang J, Mei L, Long M, Liu Y, Sun W, Li X, et al. BM-net: CNN-based mobileNet-V3 and bilinear structure for breast cancer detection in whole slide images. Bioengineering. (2022) 9:261. doi: 10.3390/bioengineering9060261

PubMed Abstract | Crossref Full Text | Google Scholar

19. Das HS, Das A, Neog A, Mallik S, Bora K, and Zhao Z. Breast cancer detection: Shallow convolutional neural network against deep convolutional neural networks based approach. Front Genet. (2023) 13. doi: 10.3389/fgene.2022.1097207

PubMed Abstract | Crossref Full Text | Google Scholar

20. Aljuaid H, Alturki N, Alsubaie N, Cavallaro L, and Liotta A. Computer-aided diagnosis for breast cancer classification using deep neural networks and transfer learning. Comput Methods Programs Biomedicine. (2022) 223:106951. doi: 10.1016/j.cmpb.2022.106951

PubMed Abstract | Crossref Full Text | Google Scholar

21. Latha M, Kumar PS, Chandrika RR, Mahesh TR, Kumar VV, and Guluwadi S. Revolutionizing breast ultrasound diagnostics with EfficientNet-B7 and Explainable AI. BMC Med Imaging. (2024) 24:230. doi: 10.1186/s12880-024-01404-3

PubMed Abstract | Crossref Full Text | Google Scholar

22. Schnider E, Wolleb J, Huck A, Toranelli M, Rauter G, Müller-Gerbl M, et al. Improved distinct bone segmentation in upper-body CT through multi-resolution networks. Int J Comput Assisted Radiol Surg. (2023) 18:2091–9. doi: 10.1007/s11548-023-02957-4

PubMed Abstract | Crossref Full Text | Google Scholar

23. Yousefirizi F, Klyuzhin IS, O JH, Harsini S, Tie X, Shiri I, et al. TMTV-Net: fully automated total metabolic tumor volume segmentation in lymphoma PET/CT images — a multi-center generalizability analysis. Eur J Nuclear Med Mol Imaging. (2024) 51:1937–54. doi: 10.1007/s00259-024-06616-x

PubMed Abstract | Crossref Full Text | Google Scholar

24. Afrifa S, Appiahene P, Zhang T, and Varadarajan V. Machine learning algorithms for iron deficiency anemia detection in children using palm images. Int J Educ Manage Eng. (2024) 14(1):1–15. doi: 10.5815/ijeme.2024.01.01

Crossref Full Text | Google Scholar

25. Appiahene P, Arthur EJ, Korankye S, Afrifa S, Asare JW, and Donkoh ET. Detection of anemia using conjunctiva images: A smartphone application approach. Med Novel Technol Devices. (2023) 18:100237. doi: 10.1016/j.medntd.2023.100237

Crossref Full Text | Google Scholar

26. Appiahene P, Dogbe SSD, Kobina EEY, Dartey PS, Afrifa S, Donkoh ET, et al. Application of ensemble models approach in anemia detection using images of the palpable palm. Med Novel Technol Devices. (2023) 20:100269. doi: 10.1016/j.medntd.2023.100269

Crossref Full Text | Google Scholar

27. Adu WK, Appiahene P, and Afrifa S. VAR, ARIMAX and ARIMA models for nowcasting unemployment rate in Ghana using Google trends. J Electrical Syst Inf Technol. (2023) 10:12. doi: 10.1186/s43067-023-00078-1

Crossref Full Text | Google Scholar

28. Huang C, Sarabi M, and Ragab AE. MobileNet-V2 /IFHO model for Accurate Detection of early-stage diabetic retinopathy. Heliyon. (2024) 10:e37293. doi: 10.1016/j.heliyon.2024.e37293

PubMed Abstract | Crossref Full Text | Google Scholar

29. Mandiya RE, Kongo HM, Kasereka SK, Kyandoghere K, Tshakwanda PM, and Kasoro NM. Enhancing COVID-19 detection: an xception-based model with advanced transfer learning from X-ray thorax images. J Imaging. (2024) 10:63. doi: 10.3390/jimaging10030063

PubMed Abstract | Crossref Full Text | Google Scholar

30. Huang G, Liu Z, van der Maaten L, and Weinberger KQ. Densely connected convolutional networks. In: Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). Honolulu, HI, USA. Piscataway, NJ: IEEE Computer Society (2017). p. 2261–9. doi: 10.1109/CVPR.2017.243

Crossref Full Text | Google Scholar

31. He K, Zhang X, Ren S, and Sun J. Deep residual learning for image recognition. In: Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). Las Vegas, NV, USA. Piscataway, NJ: IEEE (2016). p. 770–8. doi: 10.1109/CVPR.2016.90

Crossref Full Text | Google Scholar

32. Latif S, Zou Z, Idrees Z, and Ahmad J. A novel attack detection scheme for the industrial internet of things using a lightweight random neural network. IEEE Access. (2020) 8:89337–50. doi: 10.1109/Access.6287639

Crossref Full Text | Google Scholar

33. Phan H, Liu Z, Huynh D, Savvides M, Cheng K-T, and Shen Z. Binarizing MobileNet via evolution-based searching. In: Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). (2020). p. 13417–26. doi: 10.1109/CVPR42600.2020.01343

Crossref Full Text | Google Scholar

34. Chen H-Y and Su C-Y. (2018). An enhanced hybrid MobileNet, in: 2018 9th International Conference on Awareness Science and Technology (iCAST), , Vol. 2018. pp. 308–12.

Google Scholar

35. Howard AG, Zhu M, Chen B, Kalenichenko D, Wang W, Weyand T, et al. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv preprint arXiv:170404861. (2017).

Google Scholar

36. Ansar W, Shahid AR, Raza B, and Dar AH. Breast cancer detection and localization using MobileNet based transfer learning for mammograms. In: Intelligent Computing Systems: Third International Symposium, ISICS 2020, Sharjah, United Arab Emirates, March 18–19, 2020, Proceedings 3. Cham, Switzerland: Springer International Publishing (2020). p. 11–21. doi: 10.1007/978-3-030-43364-2_2

Crossref Full Text | Google Scholar

37. Gómez-Flores W and de Albuquerque Pereira WC. A comparative study of pre-trained convolutional neural networks for semantic segmentation of breast tumors in ultrasound. Comput IN Biol AND Med. (2020) 126:104036. doi: 10.1016/j.compbiomed.2020.104036

PubMed Abstract | Crossref Full Text | Google Scholar

38. Badawy SM, Mohamed AE-NA, Hefnawy AA, Zidan HE, GadAllah MT, and El-Banby GM. Automatic semantic segmentation of breast tumors in ultrasound images based on combining fuzzy logic and deep learning-A feasibility study. PloS One. (2021) 16:e0251899. doi: 10.1371/journal.pone.0251899

PubMed Abstract | Crossref Full Text | Google Scholar

39. Chunhapran O and Yampaka T. Combination ultrasound and mammography for breast cancer classification using deep learning. In: 2021 18th International Joint Conference on Computer Science and Software Engineering (JCSSE). Lampang, Thailand. Piscataway, NJ: IEEE (2021). p. 1–6. doi: 10.1109/JCSSE53117.2021.9493840

Crossref Full Text | Google Scholar

40. Krishna ST and Kalluri HK. Deep learning and transfer learning approaches for image classification. Int J Recent Technol Eng (IJRTE). (2019) 7:427–32. doi: 10.35940/ijrte.F9811.038620

Crossref Full Text | Google Scholar

41. Gao Y and Mosalam KM. Deep transfer learning for imageing7-4 structural damage recognition. Computerion32 Civil Infrastructure Eng. (2018) 33:748–68. doi: 10.1111/mice.2018.33.issue-9

Crossref Full Text | Google Scholar

42. Feng X, Yao H, and Zhang S. An efficient way to refine DenseNet. Signal Image Video Process. (2019) 13:959–65. doi: 10.1007/s11760-019-01433-4

Crossref Full Text | Google Scholar

43. Kim SY, Choi Y, Kim EK, Han B-K, Yoon JH, Choi JS, et al. Deep learning-based computer-aided diagnosis in screening breast ultrasound to reduce false-positive diagnoses. Sci Rep. (2021) 11:395. doi: 10.1038/s41598-020-79880-0

PubMed Abstract | Crossref Full Text | Google Scholar

Keywords: artificial intelligence, breast cancer, diagnosis, mammography, ultrasound

Citation: Yin Y, Fang J, Zhang W and Shen X (2025) Optimizing breast cancer ultrasound diagnosis: a comparative study of AI model performance and image resolution. Front. Oncol. 15:1536365. doi: 10.3389/fonc.2025.1536365

Received: 28 November 2024; Accepted: 07 May 2025;
Published: 06 June 2025; Corrected: 17 June 2025.

Edited by:

Sharon R. Pine, University of Colorado Anschutz Medical Campus, United States

Reviewed by:

Jaber H. Jaradat, Mutah University, Jordan
Stephen Afrifa, Tianjin University, China

Copyright © 2025 Yin, Fang, Zhang and Shen. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Wei Zhang, emh3ZGV0b21AMTYzLmNvbQ==; Xinying Shen, c2hlbnhpbnlpbmcwNzU1QDEyNi5jb20=

These authors have contributed equally to this work and share first authorship

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.