Your new experience awaits. Try the new design now and help us make it even better

SYSTEMATIC REVIEW article

Front. Artif. Intell., 05 November 2025

Sec. Medicine and Public Health

Volume 8 - 2025 | https://doi.org/10.3389/frai.2025.1649746

Artificial intelligence for ovarian cancer diagnosis via ultrasound: a systematic review and quantitative assessment of model performance

Igor Garcia-Atutxa,&#x;Igor Garcia-Atutxa1,2Francisca Villanueva-Flores&#x;Francisca Villanueva-Flores2Ekaitz Dudagotia Barrio&#x;Ekaitz Dudagotia Barrio3Javier I. Sanchez-Villamil&#x;Javier I. Sanchez-Villamil2Jos Martínez-Ms&#x;José Martínez-Más4Andrs Bueno-Crespo
&#x;Andrés Bueno-Crespo1*
  • 1Escuela Politécnica Superior, Universidad Católica de Murcia (UCAM), Murcia, Spain
  • 2Centro de Investigación en Ciencia Aplicada y Tecnología Avanzada, (CICATA) Unidad Morelos del Instituto Politécnica Nacional (IPN), Xochitepec, Mexico
  • 3University of Murcia, Murcia, Spain
  • 4Facultad de Medicina, Universidad Católica de Murcia (UCAM), Murcia, Spain

Background: Early and accurate detection of ovarian cancer (OC) remains clinically challenging, prompting exploration of artificial intelligence (AI)-based ultrasound diagnostics. This systematic review and meta-analysis critically evaluate diagnostic accuracy, methodological rigor, and clinical applicability of AI models for ovarian mass classification using B-mode ultrasound.

Methods: A systematic literature search following PRISMA guidelines was conducted in PubMed, IEEE Xplore, and Scopus up to December 2024. Eligible studies included AI-based ovarian mass classification using B-mode ultrasound, reporting accuracy, sensitivity, specificity, and/or area under the ROC curve (AUC). Data extraction, quality assessment (PROBAST), and meta-analysis (random effects) were independently performed by two reviewers. Heterogeneity sources were explored.

Results: From 823 identified records, 44 studies met inclusion criteria, covering over 650,000 images. Pooled performance metrics indicated high accuracy (92.3%), sensitivity (91.6%), specificity (90.1%), and AUC (0.93). Automated segmentation significantly outperformed manual segmentation in accuracy and sensitivity, demonstrating standardization benefits and reduced observer variability. Dataset size minimally correlated with performance, highlighting methodological rigor as a primary determinant. No specific AI architecture consistently outperformed others. Substantial methodological heterogeneity and frequent risk-of-bias issues (limited validation, small datasets) currently limit clinical translation.

Conclusion: AI models show promising diagnostic performance for OC ultrasound imaging. However, addressing methodological challenges, including rigorous validation, standardized reporting (TRIPOD-AI, STARD-AI), and prospective multicenter studies, is essential for clinical integration. This review provides clear recommendations to enhance clinical translation of AI-based ultrasound diagnostics.

Graphical abstract
B-mode ultrasound images show ovarian tumors in original, manual, and automatic segmentation. A comparison between expert diagnosis and artificial intelligence highlights AI's accuracy, sensitivity, specificity, and AUC. Illustrations depict healthy, early, and advanced tumor stages, with AI symbols like a magnifying glass, network, and brain.

Graphical Abstract.

1 Introduction

Ovarian cancer (OC) is the most lethal gynecological neoplasm and the fifth leading cause of cancer-related mortality in women (Dalmartello et al., 2022; Siegel et al., 2021). The incidence of OC in the USA is estimated to be 10.2 cases, and the mortality rate is 6.0 deaths per 100,000 women per year, indicating approximately 1.1% lifetime risk for women (NIH, 2020). Despite advances in diagnosis and treatment, the mortality rate has not shown a significant decline over the past three decades, primarily due to the challenges in early detection and limited therapeutic efficacy in advanced-stage disease (Badgwell and Bast, 2007; Jacobs and Menon, 2004; Bast et al., 2007; Torre et al., 2018). Currently, 70% of OC cases are diagnosed at advanced stages, where five-year survival drastically drops to 20%–30%, in contrast to 80%–95% survival when detected at early stages (Bowtell et al., 2015). Therefore, improving early-stage detection methods is critically important to enhance patient outcomes.

Transvaginal ultrasound (TVS), computed tomography (CT), and magnetic resonance imaging (MRI) are the primary imaging modalities for OC detection. TVS is particularly advantageous as a non-invasive, cost-effective, accessible, and real-time imaging method, allowing assessment of ovarian masses’ size, shape, and internal structures (e.g., septa, solid tissue), aiding differentiation between benign and malignant tumors. However, the limited resolution of TVS may fail to detect small or early-stage tumors, and overlapping anatomical structures, such as bowel loops or normal ovarian tissue, complicate mass differentiation. Moreover, specific OC subtypes do not exhibit significant morphological changes in early phases, reducing sensitivity (Wu et al., 2018; Rosati et al., 2020).

Despite these limitations, TVS remains indispensable in the initial OC diagnosis due to accessibility and low cost. However, interpretation often varies significantly with radiologist experience, resulting in diagnostic inconsistencies and clinical errors (Bäumler et al., 2020; Nebgen et al., 2019). These issues underscore the critical need for standardized, objective, and automated diagnostic methods that enhance accuracy and reduce inter-observer variability.

Artificial intelligence (AI) integration has significantly advanced medical imaging diagnostics, improving tumor identification accuracy and consistency. Convolutional neural networks (CNNs), a sophisticated deep learning (DL) architecture, have demonstrated over 90% accuracy in extracting complex TVS image features and classifying ovarian malignancy in several studies (Akazawa and Hashimoto, 2021; Sone et al., 2021). Beyond accuracy, AI reduces human interpretation errors and enables the analysis of large datasets (Falana et al., 2023; Sahu and Shrivastava, 2023).

However, several challenges hinder the clinical translation of AI in OC detection. Crucially, existing DL models suffer from inadequate dataset representativeness, as most research uses datasets from single institutions, leading to population biases. AI model performance notably declines when tested on populations with different ethnic, geographic, or technological characteristics (Noseworthy et al., 2020). Additionally, heterogeneity in TVS image quality, formats, resolution, and acquisition protocols across institutions further impairs AI model reproducibility and generalizability (Raciti et al., 2023).

A significant barrier is the absence of robust prospective clinical validation. Most AI algorithms have been validated retrospectively, limiting insights into their real-time clinical applicability (Raciti et al., 2023). Furthermore, lack of standardized annotation and segmentation protocols significantly impacts AI model accuracy, with manual segmentation inconsistencies affecting performance by up to 20% (Heinlein et al., 2024; Ho et al., 2022). The absence of international consortia, standardized benchmarking, and sufficiently large, diverse, open-access TVS image databases also restrict the comparative evaluation and robust training of AI models. While initiatives like The Cancer Imaging Archive (TCIA) have addressed similar needs in other cancer areas, a parallel effort for OC is currently lacking (NIH, n.d.).

This systematic review and quantitative meta-analysis address these critical knowledge gaps by evaluating the diagnostic performance of AI models applied specifically to B-mode TVS images for early OC detection. Through a comprehensive comparison of accuracy, sensitivity, specificity, and area under the curve (AUC) across CNNs, classical machine learning algorithms, and transformer-based models, the study assesses how methodological factors, such as segmentation and dataset size, influence model performance. Clarifying these factors is expected to enhance clinical practice directly by guiding the development of robust, standardized AI tools capable of improving early OC diagnosis, thereby increasing patient survival rates and clinical outcomes.

2 Methodology

This study was designed and conducted following the PRISMA 2020 guidelines (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) (Liberati et al., 2009; Page et al., 2021) to ensure transparency, reproducibility, and comprehensiveness in the systematic review and meta-analysis (Supplementary Tables S1, S2). The methodological protocol included a predefined search strategy, explicit eligibility criteria, risk of bias assessment, and statistical analysis of the extracted data. Specifically, the systematic review was guided by the following research questions:

• What is the diagnostic accuracy (accuracy, sensitivity, specificity, and AUC) of AI-based models for ovarian mass classification using B-mode ultrasound?

• Which methodological factors, such as segmentation methodology (automatic vs. manual), dataset size, AI model architecture, and risk of bias significantly influence the diagnostic performance of AI models?

2.1 Search strategy

In January 2025, a comprehensive search was conducted in three high-impact scientific databases: PubMed, IEEE Xplore, and Scopus. The search strategy included the following terms combined using Boolean operators: (“machine learning” OR “artificial intelligence” OR “deep learning” OR “neural network”) AND (“ovarian cancer” OR “ovarian tumor”) AND “ultrasound.” No language or publication type restrictions were applied during the initial search.

2.2 Eligibility criteria

The following inclusion and exclusion criteria were explicitly defined to ensure transparency and reproducibility in the systematic selection of studies (see Table 1).

Table 1
www.frontiersin.org

Table 1. Inclusion and exclusion criteria for study selection.

Studies meeting all inclusion criteria and none of the exclusion criteria were eligible for inclusion in this systematic review and meta-analysis.

2.3 Study selection process

Two independent reviewers (IGA and FVF) initially evaluated the title and abstract of each article identified through the database search, applying the inclusion and exclusion criteria explicitly defined in Section 2.2. This preliminary assessment allowed for the exclusion of clearly irrelevant or ineligible studies. In cases of discrepancies during this initial stage, a third reviewer (EDB) was consulted to reach consensus. Subsequently, the full texts of the preselected articles were reviewed again by both reviewers (IGA and FVF) to confirm their definitive eligibility for inclusion in the quantitative analysis.

2.4 Data extraction

The following variables were extracted from each study: author, year, type of segmentation, model architecture, model name, image dataset size, type of ovarian masses, number of classes, and performance metrics (accuracy, sensitivity, specificity, and AUC). For studies reporting multiple models, the one with the best overall performance was selected to avoid data duplication. The information was systematized into a structured database for subsequent statistical analysis.

2.5 Risk of bias assessment

The methodological quality of the included studies was assessed using the PROBAST tool (Prediction model Risk of Bias Assessment Tool) (Wolff et al., 2019), which evaluates the risk of bias and applicability in studies that develop or validate prediction models. PROBAST comprises 20 items grouped into four domains: (i) participant selection, (ii) predictors, (iii) outcomes, and (iv) statistical analysis. Two reviewers (IGA and FVF) independently performed this evaluation. Discrepancies were resolved by consensus. A detailed assessment of the 20 PROBAST items for each study and domain-specific classifications is included as Supplementary Tables S3, S4.

2.6 Statistical analysis

2.6.1 Descriptive statistics and distribution assessment

Descriptive statistics were calculated for performance metrics, including accuracy, sensitivity, specificity, and AUC, and are reported as means, standard deviations, and ranges. The Shapiro–Wilk test was used to assess the normality of distributions, while Levene’s test was applied to evaluate the homogeneity of variances.

Because F1-score was rarely reported and often lacked the underlying confusion matrix, we did not meta-analyze F1. For interpretability, accuracy, sensitivity, specificity, and AUC remained our primary endpoints.

2.6.2 Comparison between segmentation methods (automatic vs. manual)

Given the presence of non-normal distributions and limited subgroup sizes, non-parametric tests were prioritized to enhance statistical validity. Specifically, a Mann–Whitney U test was used to compare accuracy between automatic and manual segmentation strategies.

2.6.3 Comparison across AI model architectures

Differences in accuracy across AI architecture categories (e.g., CNN, ML, ANN) were assessed using a Kruskal–Wallis H test. Additionally, performance variation between DL models (e.g., CNNs) and classical machine learning approaches was evaluated using one-way analysis of variance (ANOVA).

2.6.4 Correlation between dataset size and diagnostic performance

The relationship between dataset size and diagnostic performance was explored using Pearson’s correlation, excluding studies with more than 5,000 images to mitigate the influence of extreme values.

2.6.5 Meta-regression analysis of methodological factors

A meta-regression analysis was performed using ordinary least squares (OLS) modeling to investigate the combined influence of methodological variables on diagnostic performance. Accuracy was modeled as the dependent variable, and key predictors included dataset size, segmentation type (automatic vs. manual), model architecture (CNN vs. other), and risk of bias (high vs. low). The regression included 26 studies with complete data and demonstrated that segmentation type was a significant predictor of accuracy (β = 0.0656, p = 0.007), while the other covariates did not reach statistical significance. The model explained approximately 32% of the variance in accuracy (adjusted R2 = 0.32), supporting the relevance of segmentation quality as a determinant of AI model performance.

2.6.6 Subgroup and sensitivity analyses

Subgroup analyses were conducted based on risk of bias (assessed by PROBAST), and a sensitivity analysis was performed by excluding studies rated as high risk to determine the robustness of findings.

2.6.7 Software and reproducibility

All statistical analyses were conducted using Python (v3.12), leveraging the pandas, numpy, scipy, statsmodels, matplotlib, and seaborn libraries. Complete analysis code and data visualizations are available upon request.

All figures include concise alternative text in the captions, and a separate Supplementary material provides long textual descriptions.

3 Results

The systematic search in the PubMed, IEEE Xplore, and Scopus databases yielded 823 studies. After removing 58 duplicates, 765 titles and abstracts were screened. Of these, 686 were excluded for not meeting the inclusion criteria, resulting in 79 articles for full-text review. Finally, 44 studies were included in the quantitative analysis (Figure 1).

Figure 1
Flowchart detailing the study selection process for a systematic review. Identification phase: 823 records from PubMed (177), IEEE (52), and Scopus (594) are identified. Screening phase: 58 duplicates removed, 765 abstracts reviewed; 686 records excluded based on title and abstract. Full-text review: 79 articles, with 35 excluded for specific criteria, resulting in 44 studies included in the systematic review.

Figure 1. PRISMA flow diagram of our study. The figure illustrates the study selection process following PRISMA 2020 guidelines. A total of 823 records were identified through three databases (PubMed, IEEE Xplore, and Scopus). After removing 58 duplicates, 765 titles and abstracts were screened, excluding 686 studies that did not meet the eligibility criteria. Seventy-nine full-text articles were assessed for inclusion, of which 35 were excluded for using non-ultrasound imaging modalities (e.g., CT or MRI), focusing exclusively on segmentation methods, or lacking histopathological validation. Ultimately, 44 studies were included in the final quantitative analysis.

The studies cover a period up to December 2024. Collectively, they analyzed over 650,000 B-mode TVS images for ovarian mass classification using various AI models. Most studies (n = 27; 61.4%) used automatic segmentation, while the remainder (n = 17; 38.6%) employed manual segmentation. The predominant architectures were CNNs, followed by classical ML algorithms, conventional artificial neural networks (ANNs), and transformer-based architectures that have emerged in recent years (Table 2).

Table 2
www.frontiersin.org

Table 2. List of the 44 articles analyzed.

3.1 Overall performance of AI models

The analysis of the 44 included studies revealed the high average diagnostic performance of AI models applied to B-mode TVS images for OC detection. The mean accuracy was 92.3% ± 5.8, with mean sensitivity and specificity of 91.6% ± 7.2 and 90.1% ± 8.1, respectively. AUC values were reported in only 23 studies, with a mean of 0.93 ± 0.04, reflecting strong overall discriminative capacity. However, the partial availability of AUC reporting may indicate a potential reporting bias that limits the robustness of comparative analysis across all models. Beyond AUC’s limited reporting (23/44 studies), F1-score was scarcely available across the corpus. This pattern likely reflects historical reliance on accuracy/sensitivity/specificity in ultrasound AI, frequent absence of continuous model scores (hindering AUC), and the lack of confusion matrices or class-wise results needed for F1. In addition, F1 is sometimes reported as Dice in segmentation studies; because our review targets classification performance, segmentation-specific Dice metrics were not pooled, which might also contribute to the perceived under-reporting of F1.

Figure 2 provides a comparative overview of the four main performance metrics for the 10 top-performing models. This visualization highlights how specific models exhibit strong accuracy yet relatively lower specificity, an observation with important clinical implications when considering false-positive rates in diagnostic triage.

Figure 2
Bar chart comparing performance scores of various models with metrics: accuracy, sensitivity, specificity, and AUC. Models include MLP, DenseNet, ResNext50, LR, ResNet-50, YOLOv8, ResNet-34, EL, DenseNet161, and Pyramid Visual Transformer, on a scale from 0.800 to 1.000.

Figure 2. Top 10 AI models (comparison of accuracy, sensitivity, specificity, and AUC). This figure compares performance metrics (accuracy, sensitivity, specificity, and AUC) of the 10 highest-performing AI models identified in the review. Generally high accuracy is observed across models, but some exhibit trade-offs between sensitivity and specificity. This variability highlights the importance of selecting models not only with high overall accuracy but also clinically meaningful balance to minimize diagnostic errors, particularly false positives and false negatives.

Several models, including those based on OCD-FCNN, probabilistic neural networks (PNN), and ResNet-34, reported peak performance values exceeding 95%. However, many of these models were trained and tested on small or non-external datasets, lacked proper cross-validation, or relied exclusively on internal test sets. Such methodological limitations increase the likelihood of overfitting and restrict the generalizability of reported outcomes. None of the highest-performing models reported prospective validation or integration into clinical workflows, which remains essential for evaluating real-world applicability.

To assess whether methodological design influenced diagnostic performance, non-parametric tests were conducted using accuracy as the outcome variable. A Mann–Whitney U-test revealed a statistically significant difference in accuracy between models using automatic versus manual segmentation (U = 234.0, p = 0.007), favoring automatic methods. This finding suggests that automated segmentation enhances standardization and reduces variability across studies.

Conversely, a Kruskal–Wallis test comparing performance across AI architectures (e.g., CNN, ML, ANN) did not identify statistically significant differences (H = 6.53, p = 0.258), indicating that no specific architectural family demonstrated superior accuracy within the current dataset. Nevertheless, visual inspection using violin plots (Figure 3) showed a moderately higher central tendency and reduced the variance in accuracy among CNN-based models compared to classical machine learning (ML) approaches. While this pattern may reflect the architectural strengths of CNNs in capturing spatial hierarchies within medical images (Litjens et al., 2017), it should be interpreted cautiously. CNN-based models were more frequently applied in recent studies, which may also have benefited from advances in data augmentation, automatic segmentation, and preprocessing pipelines. Therefore, the observed trend could be confounded by methodological improvements rather than an inherent advantage of architecture.

Figure 3
Violin plot comparing accuracy of CNN and ML architectures. The CNN distribution is in blue, and ML is in orange. Both show similar median accuracies around 0.9.

Figure 3. Accuracy distribution by AI architecture: CNN vs. classical machine learning (ML). This figure illustrates that, although there is no statistically significant difference between architectures (p = 0.258), CNN-based models tend to display higher median accuracy and reduced variability compared to classical ML models. This result suggests a potential advantage of CNNs, likely due to their superior ability to capture complex features from medical images, though methodological advances in recent studies may also contribute to this observed trend.

It is also important to note that model performance was evaluated exclusively using accuracy, as this was the most consistently reported metric across studies. While this allowed for comparability, it may limit interpretability in class-imbalanced settings, where metrics such as AUC or F1-score are often more informative. Future studies should prioritize the reporting of multiple complementary metrics to capture diagnostic value more comprehensively.

These descriptive findings are further explored and formally tested in the meta-regression presented in Section 3.4.

In summary, while reported performance metrics are generally high, the absence of standardized validation protocols, partial reporting of key metrics, lack of weighted or stratified analyses, and underreporting of methodological variables (especially segmentation and validation strategies) limit the interpretability and clinical generalizability of the findings. Future studies should adopt harmonized reporting guidelines (e.g., TRIPOD-AI, PROBAST-AI), employ multicenter and external validation, and report performance metrics in clinically meaningful terms to support reliable and reproducible integration into diagnostic workflows.

3.2 Relationship between dataset size and performance

The relationship between the number of images used to train AI models and their diagnostic performance was evaluated using non-parametric correlation analysis. Although Pearson’s method was initially considered, the Shapiro–Wilk test confirmed that dataset size and performance metrics (accuracy, sensitivity, specificity) were not normally distributed (p < 0.001 for all), prompting the use of Spearman’s rank correlation.

Studies with more than 5,000 images were excluded from this analysis to reduce the risk of statistical distortion from highly imbalanced sample sizes. While large-scale datasets (up to 575,000 images) have become increasingly common in AI development, such volumes do not reflect typical clinical practice and may disproportionately drive correlation estimates. The 5,000-image threshold was selected to capture real-world data conditions better while preserving inter-study variability. Descriptive analysis of the full dataset showed that this threshold approximately corresponds to the 75th percentile of dataset sizes among included studies.

Spearman correlation coefficients between dataset size (≤5,000) and model performance metrics were weak and statistically non-significant. Specifically, the correlation with accuracy was ρ = 0.080 (p = 0.653), with a 95% confidence interval of −0.27 to 0.41 and an R2 of 0.006, suggesting that less than 1% of the variation in accuracy could be explained by dataset size. For sensitivity, the correlation was ρ = 0.246 (p = 0.154; 95% CI: −0.09 to 0.54; R2 = 0.061), and for specificity, ρ = 0.183 (p = 0.350; 95% CI: −0.20 to 0.52; R2 = 0.034). Table 3 summarizes these results, including the correlation coefficients, confidence intervals, and the proportion of explained variance.

Table 3
www.frontiersin.org

Table 3. Correlation between dataset size and performance metrics.

While these findings suggest that increasing dataset size within the studied range does not systematically improve model performance, this interpretation should be cautiously made. The exclusion of large datasets may limit the generalizability of these findings, and potential interaction effects, such as those involving the segmentation method, risk of bias, or model architecture, were not examined in this univariate analysis. These results are, however, consistent with the multivariable meta-regression analysis presented in Section 4, in which dataset size did not emerge as a significant independent predictor of accuracy.

One plausible explanation lies in the widespread adoption of data augmentation strategies. Techniques such as image rotation, scaling, contrast adjustment, and noise addition simulate data variability and may reduce the dependency on raw volume. However, excessive use of augmentation may also lead to redundancy or learning saturation, where additional data no longer meaningfully improves generalization.

This interpretation is in line with prior literature. For instance, Roberts et al. (2021) found that dataset size was not consistently associated with performance in a comprehensive medical imaging AI studies review. Instead, methodological rigor, validation strategy, and data diversity were identified as stronger predictors of performance. Furthermore, the risk of performance overestimation due to augmented or homogeneous datasets remains a critical concern in model evaluation.

Figure 4 presents scatterplots of accuracy, sensitivity, and specificity versus dataset size (≤5,000), each overlaid with a non-parametric LOWESS regression line and 95% confidence bands. While substantial scatters are observed across all metrics, the absence of clear or consistent directional trends underscores the importance of factors beyond sample size, such as annotation quality and experimental design in developing reliable diagnostic models.

Figure 4
Three scatter plots labeled A, B, and C show accuracy, sensitivity, and specificity versus dataset size, respectively, for sizes up to 5000. Orange dots represent data points, with red lines indicating trends. Plot A shows accuracy stabilizing after increasing. Plot B shows sensitivity slightly fluctuating but generally increasing. Plot C shows specificity peaking and then declining.

Figure 4. Non-parametric association between dataset size and model performance metrics (≤5,000 images). Scatter plots depict the relationship between dataset size and three key performance metrics: (A) accuracy, (B) sensitivity, and (C) specificity across studies with ≤5,000 training images. The red lines represent LOWESS (locally weighted scatterplot smoothing) regression fits with 95% confidence bands. Although minor local variations are observed, no clear or systematic trend indicates a significant improvement in these performance metrics with increasing dataset size within the clinically relevant range analyzed. This finding suggests that methodological factors other than dataset size may have a greater influence on diagnostic model accuracy.

3.3 Comparison between automatic and manual segmentation

To evaluate the impact of segmentation strategy on the diagnostic performance of AI models, performance metrics were compared between studies that implemented automatic segmentation (n = 27) and those that used manual segmentation (n = 17).

Models using automatic segmentation achieved a significantly higher average accuracy (94.2% ± 4.3) than manual segmentation (88.2% ± 6.6, p = 0.012). Sensitivity also favored automatic segmentation (93.7% ± 5.6 vs. 88.6% ± 8.3, p = 0.042). Although specificity was higher in the automatic group (92.5% ± 6.0 vs. 87.3% ± 9.5), the difference was not statistically significant (p = 0.084). AUC values were nearly identical between both groups (p = 0.839). Confidence intervals for these comparisons were not reported but are recommended for future studies to enhance the interpretability and reproducibility of statistical estimates.

Levene’s test revealed a significantly more significant variance in specificity within the manual segmentation group (p = 0.045), indicating less consistency. This is consistent with previous literature findings, highlighting manual segmentation’s susceptibility to inter- and intra-observer variability, particularly when standardized annotation protocols or multiple expert raters are not employed (Taha and Hanbury, 2015; Menze et al., 2015).

Although the primary studies reported heterogeneous segmentation details that precluded a stratified meta-analysis by architecture, a brief practical comparison is informative for clinical implementation. U-Net remains the canonical encoder–decoder with skip connections that performs well when lesion boundaries are reasonably defined, and training data are limited. AdaResU-Net augments U-Net with residual blocks and adaptive mechanisms that enlarge the effective receptive field and stabilize gradient flow, improving boundary delineation in speckle-rich ultrasound and in the presence of heterogeneous echotexture. In practice, U-Net offers simplicity and fast deployment; AdaResU-Net can yield crisper contours and fewer leakage errors near cyst walls at the cost of extra parameters. These architectural tradeoffs are likely to contribute to the higher and less variable accuracy we observed with automated vs. manual segmentation. Future primary studies should report standardized segmentation metrics (e.g., Dice, surface distance) alongside classification endpoints to enable formal architecture-level synthesis.

Figure 5 presents comparative boxplots of accuracy, sensitivity, specificity, and AUC by segmentation type. The distributions reveal higher mean values for automatic segmentation across most metrics, lower dispersion, and fewer outliers, especially for specificity and sensitivity. This visual trend suggests increased consistency, which may be attributed to the standardization benefits of automated pipelines.

Figure 5
Box plots comparing automatic and manual methods across four metrics: (A) Accuracy, (B) Sensitivity, (C) Specificity, and (D) AUC. Automatic methods show higher median values in all metrics, with narrower interquartile ranges compared to manual methods. Outliers are present for the sensitivity and specificity metrics.

Figure 5. Diagnostic performance metrics stratified by segmentation type (automatic vs. manual). Boxplots illustrate the distribution of (A) accuracy, (B) sensitivity, (C) specificity, and (D) AUC across studies using either automatic or manual segmentation. Automatic segmentation models demonstrate higher mean values and reduced dispersion for most metrics, particularly accuracy and specificity. The manual segmentation group shows greater variability and outliers, suggesting less consistency. These visual differences align with the hypothesis that automated segmentation enhances reproducibility and standardization; however, causal interpretation should be made cautiously due to potential confounding factors.

Table 4 provides a detailed summary of methodological characteristics and performance metrics stratified by segmentation type. Studies using automatic segmentation not only performed better on average but also used considerably larger datasets (mean = 22,941 vs. 1,121 images) and demonstrated lower standard deviation across metrics such as accuracy (4% vs. 7%) and specificity (6% vs. 10%).

Table 4
www.frontiersin.org

Table 4. Dataset size and performance by segmentation type.

However, this interpretation should be approached with caution. Although these group-level comparisons suggest superior performance with automatic segmentation, the analysis did not control for potential confounders such as dataset size, model architecture, training methodology, or publication year. Importantly, these variables may co-vary with segmentation strategy, particularly since automatic methods are more prevalent in recent, technically advanced studies.

As shown in the meta-regression (Section 3.4), the segmentation strategy remained a significant independent predictor of accuracy even after adjusting for these covariates. Nevertheless, the unadjusted differences observed here might still reflect broader methodological convergence rather than a causal advantage of automatic segmentation per se.

Future studies should incorporate multivariable models, harmonized annotation protocols, and prospective designs to clarify the segmentation method’s isolated effect on AI model performance. Moreover, reporting standards such as STARD-AI and TRIPOD-AI should be adopted to ensure replicability and transparency in performance evaluation across studies (Collins et al., 2021; Sounderajah et al., 2021).

3.4 Meta-regression analysis of methodological factors

To further explore the drivers of diagnostic performance, a meta-regression was performed using accuracy as the dependent variable and four predictors: dataset size, segmentation strategy (automatic vs. manual), model architecture (CNN vs. other), and risk of bias (high vs. low). The regression model included 26 studies with complete data across all variables (Figure 6).

Figure 6
Bar chart showing the effect on accuracy by different predictors. Segmentation_binary has the highest positive effect, followed by Bias_HighRisk and Architecture_CNN. The size of the dataset shows no effect.

Figure 6. Meta-regression coefficients with 95% confidence intervals. This figure presents the coefficients resulting from the meta-regression analysis evaluating the relative influence of various methodological factors on diagnostic accuracy. Only automatic segmentation demonstrated a statistically significant effect on accuracy (p = 0.007), whereas other factors such as dataset size, architecture type (CNN vs. others), and risk of bias did not reach statistical significance. This outcome emphasizes the critical importance of automated segmentation quality as a key determinant of AI model performance.

The overall model was statistically significant (F = 3.98, p = 0.015), with an adjusted R2 of 0.32, indicating that the included predictors could explain approximately 32% of the variance in reported accuracy.

Among the covariates, segmentation strategy emerged as a significant predictor: studies using automatic segmentation reported on average, a 6.6 percentage point higher accuracy compared to those using manual segmentation (β = 0.0656, p = 0.007). This aligns with previous findings suggesting that automated preprocessing may reduce inter-observer variability and improve reproducibility.

Other predictors, such as dataset size, CNN architecture, and risk of bias, were not statistically significant at the conventional threshold (p > 0.050). However, the effect of high risk of bias approached significance (β = 0.0427, p = 0.096), suggesting a possible inflation of performance estimates in studies with methodological limitations.

Notably, dataset size was not a significant predictor (p = 0.323), corroborating earlier findings that performance does not linearly scale with sample size within the studied range, possibly due to saturation effects or compensatory use of data augmentation techniques.

These results reinforce the critical role of segmentation quality in shaping model performance and highlight the need for more standardized methodologies and transparent reporting in AI-based diagnostic research.

3.5 Temporal analysis

The progression of AI architectures in the included studies reflects a clear methodological shift over time. Temporal analysis of architectural usage revealed a transition from traditional ML techniques and ANNs to DL approaches, particularly CNNs. Between 2014 and 2018, studies primarily employed ML methods such as support vector machines (SVM), random forests (RF), and logistic regression (LR), representing approximately 85% of the methodologies used during this period. ANN-type architecture was also present, constituting roughly 15% of studies, while no CNN-based models were recorded before 2020.

Figure 7 presents a stacked bar chart showing the number of studies using each architecture per year. CNNs emerged in 2020, accounting for 20% of the methodologies that year, and showed a marked increase in 2021, becoming the predominant architecture (65% of studies) in 2022. This trend intensified through 2023 (72% of studies) and 2024 (78% of studies), with CNNs accounting for more than half of the models evaluated annually. Transformer-based architectures, specifically Swin Transformer and Pyramid Vision Transformer, appeared exclusively in 2024, accounting for approximately 10% of the methodologies that year, indicating the beginning of a new phase of exploration focused on models with advanced contextual attention mechanisms and long-range feature integration.

Figure 7
Stacked bar chart showing the number of AI studies published from 2014 to 2024, with categories for AI architectures: ANN, CNN, FDA, ML, RNN, and TBN. Notable increase in 2024.

Figure 7. This figure illustrates the temporal evolution in the use of different AI architectures from 2014 to 2024. A clear shift is observed from classical machine learning techniques (ML and ANN) to deep learning models (CNN), particularly from 2021 onwards, with transformer-based models appearing more recently (2024). This evolution reflects a continuous methodological transition toward increasingly sophisticated architectures capable of directly and deeply extracting information from images. However, this shift may also coincide with general methodological improvements over time.

This shift mirrors broader trends observed across diagnostic imaging AI, where deep architectures have largely replaced classical ML techniques due to their ability to learn hierarchical features directly from raw images without manual feature engineering (Litjens et al., 2017; Esteva et al., 2019). However, this evolution may also explain some of the performance differences observed in earlier sections. For example, the predominance of CNNs in recent years may co-occur with advances in preprocessing, data augmentation, and training infrastructure, confounding the interpretation of architecture-based performance gains.

Notably, this trend may influence perceived model superiority, as CNN-based studies often reflect newer methodological standards, including automatic segmentation and more rigorous evaluation protocols. These temporal patterns underscore the importance of accounting for publication year and technological maturity when comparing performance across architectures or studies.

3.6 Heterogeneity and risk of bias analysis

To evaluate how methodological quality influences the reported performance of AI models, the 44 included studies were classified according to their overall risk of bias using the PROBAST tool. In total, 26 studies were classified as low risk of bias, 12 as unclear risk, and six as high risk (Figure 8). However, not all studies reported all performance metrics, and the number of studies included varied substantially by metric. This heterogeneity in reporting introduces selection bias and impairs comparability.

Figure 8
Chart A displays the number of studies across different PROBAST domains with risk levels: low (green), unclear (yellow), and high (red). Chart B shows mean values for accuracy, sensitivity, specificity, and AUC with the same risk levels. Both charts highlight the distribution and influence of risk levels in study assessments.

Figure 8. Distribution of risk of bias across PROBAST domains and associated model performance metrics. (A) Distribution of the 44 included studies across the four PROBAST domains (participants, predictors, outcomes, and analysis), categorized by overall risk of bias (low, unclear, and high). (B) Mean values of accuracy, sensitivity, specificity, and AUC, stratified by overall risk of bias. Studies classified as high risk consistently report higher mean values for accuracy, sensitivity, and specificity but exhibit greater variability and frequently lack AUC reporting, indicating potential methodological overfitting. Studies with low risk exhibit more consistent and reliable performance metrics.

Studies classified as high risk of bias showed markedly elevated performance metrics, with a mean accuracy of 96.1% ± 7.6, a sensitivity of 95.4% ± 7.9, and a specificity of 93.9% ± 11.7. However, none of these studies reported areas under the curve (AUC) values, precluding the complete evaluation of discriminative performance. Moreover, the elevated standard deviation in specificity suggests potential overfitting, likely arising from methodologically weak practices such as internal validation without cross-validation, small sample sizes, and operator-dependent manual segmentation (Wolff et al., 2019; Collins et al., 2021). These practices have been consistently linked to inflated model performance in machine learning for medical imaging (Kelly et al., 2019).

In contrast, low-risk studies showed more conservative but consistent performance metrics, with lower dispersion and complete reporting of AUC. Specifically, they reported an accuracy of 90.4% ± 5.2 (n = 20), a sensitivity of 90.7% ± 7.0 (n = 25), a specificity of 89.1% ± 8.2 (n = 22), and an AUC of 93.5% ± 3.9 (n = 20). While these results appear numerically inferior, the reduced variability and broader metric completeness suggest higher methodological reliability and clinical applicability.

The unclear-risk group, often the result of poor reporting rather than clearly defined methodological shortcomings, yielded intermediate metrics (accuracy: 93.6% ± 5.0; sensitivity: 91.9% ± 7.7; specificity: 91.1% ± 5.4; AUC: 92.2% ± 4.6). However, the small and variable sample sizes for each metric (e.g., n = 6 for specificity) compromise interpretability and hinder statistical power.

A sensitivity analysis was conducted excluding high-risk studies, which resulted in a notable reduction in extreme values (e.g., 100% accuracy) and decreased overall dispersion, particularly in specificity. However, this analysis was limited by the lack of formal statistical testing (e.g., ANOVA or Kruskal–Wallis) to assess whether differences between groups were statistically significant. Furthermore, no regression adjustment was made for potential confounders such as dataset size or model complexity. These omissions limit the strength of causal inferences between risk of bias and reported model performance.

The primary sources of methodological heterogeneity across studies were identified as follows: (i) reliance on manual, operator-dependent image segmentation; (ii) absence of cross-validation or external validation; (iii) small sample sizes (<200 cases); and (iv) lack of standardized metric reporting formats. These deficiencies were most prevalent in high- and unclear-risk studies, consistent with prior evidence from systematic reviews of machine learning in healthcare (Sounderajah et al., 2021; Liu et al., 2019).

Finally, this section would benefit from including confidence intervals and formal effect size estimates to contextualize differences across bias strata. Without these, claims about “superior” or “more stable” performance remain largely descriptive and potentially misleading.

4 Discussion

The analyzed studies demonstrate generally high diagnostic performance of AI models for classifying ovarian masses using B-mode TVS. Most studies achieved AUC values ranging from 0.85 to 0.95, with sensitivity and specificity typically above 80% (Acharya et al., 2018; Sadeghi et al., 2024). However, substantial heterogeneity across studies suggests caution when interpreting pooled performance metrics due to sample size, class balance, and data quality variations.

Automatic segmentation demonstrated statistically significant accuracy and sensitivity superior to manual segmentation, likely due to reduced observer variability (p = 0.007 and p = 0.042, respectively). Although manual segmentation provides potentially optimal delineation by experts, it is prone to operator-dependent biases. Automatic segmentation offers reproducibility and scalability but introduces errors if segmentation quality is suboptimal (Chiappa et al., 2021). Given these findings, future research should explicitly compare different segmentation algorithms and validate them against expert delineation standards to enhance clinical applicability.

Contrary to expectations, dataset size alone showed no clear correlation with diagnostic performance within datasets ≤5,000 images. Small datasets sometimes reported exceptional accuracy, likely due to extensive data augmentation strategies, while larger datasets faced increased variability and complexity, offsetting potential accuracy gains (Acharya et al., 2018; Liu et al., 2019). Thus, methodological rigor, dataset diversity, and validation strategy appear more influential than dataset volume alone.

Comparison among model architectures (classical ML, CNNs, and Transformers) revealed no consistent superiority. Traditional ML models and simple neural networks occasionally matched the performance of sophisticated CNNs and Transformers (Acharya et al., 2018). Although CNN-based models showed visually higher median accuracy and reduced variance, no statistically significant differences were observed across architectures (p = 0.258). Superior results often correlated with modern methodological standards rather than architectural innovation alone. Future research should directly compare these architectures under standardized experimental conditions to clarify intrinsic performance differences clearly.

The risk of bias significantly affected the result’s validity. Studies with high or unclear bias frequently reported inflated performance metrics with notable dispersion (e.g., specificity SD = 11.7%) due to methodological weaknesses like internal-only validation, small sample sizes, and lack of external validation (Sounderajah et al., 2021; Roberts et al., 2021). Such biases undermine clinical generalizability, emphasizing the need for rigorous validation standards (e.g., TRIPOD-AI, STARD-AI) and multicenter, prospective validation.

Temporal analysis showed methodological evolution influencing perceived performance gains. Despite less sophisticated techniques, earlier studies occasionally reported superior outcomes due to less rigorous validation methods, whereas recent studies employed stricter evaluation, tempering observed performance improvements (Sounderajah et al., 2021; Roberts et al., 2021). Therefore, performance comparisons across time should account for these evolving methodological contexts.

Recent advances emphasize the critical importance of precise methodological design and targeted biological understanding. For instance, Liu et al. (2024) highlighted how epitope-specific targeting of HER2 significantly influences therapeutic outcomes in solid tumors, illustrating the necessity of methodological precision in clinical efficacy. Similarly, Ma et al. (2023) underscored the potential benefits of integrating detailed molecular insights into advanced diagnostics by showing how the glycolytic enzyme ENO1 modulates choline phospholipid metabolism and tumor proliferation. Such multidisciplinary integration may substantially improve cancer characterization and patient outcomes.

AI currently demonstrates potential as an adjunctive diagnostic tool rather than a standalone solution. Real-world applicability requires rigorous external validations, standardized metric reporting (e.g., sensitivity, specificity, predictive values), and integration within clinical workflows. Moreover, evaluating the clinical impact of AI, particularly for distinguishing borderline ovarian lesions, remains essential due to their clinical complexity and diagnostic challenges (Roberts et al., 2021).

Given the limited reporting of comprehensive metrics like AUC and F1-score, future studies should ensure the consistent inclusion of these metrics to improve clinical interpretability and decision-making.

In conclusion, while AI models for ovarian mass classification via TVS demonstrate promising diagnostic accuracy, significant methodological limitations currently restrict their clinical translation. Future research must prioritize external validation, robust methodological standards, multidisciplinary integration, and transparent, standardized reporting. Prospective, multicenter studies remain crucial to fully validate these models’ clinical utility, generalizability, and real-world applicability.

5 Conclusion

AI models applied to B-mode TVS images demonstrate strong diagnostic performance for classifying ovarian masses, achieving high sensitivity, specificity, and overall discriminative ability (AUC). Automated segmentation significantly outperformed manual methods in accuracy and sensitivity, likely due to enhanced standardization and reduced inter-observer variability.

Nevertheless, these findings must be interpreted cautiously due to considerable methodological heterogeneity, variations in dataset size and quality, and significant risks of bias identified among several studies. Additionally, no consistent superiority emerged among different AI architectures (CNNs, classical ML, or Transformers), suggesting that methodological rigor, validation procedures, and data standardization may be more influential determinants of performance than the specific model architecture itself.

Future research should prioritize prospective, multicenter external validation under realistic clinical conditions for reliable clinical translation. Moreover, rigorous adherence to standardized reporting guidelines (e.g., TRIPOD-AI, STARD-AI), comprehensive metric reporting, including sensitivity, specificity, AUC, and F1-score, and explicit evaluation of clinical utility, especially in distinguishing borderline ovarian lesions, are essential.

In summary, while AI holds significant promise for OC diagnosis using TVS, overcoming current methodological limitations through robust validation and standardized methodological practices is imperative for successful integration into clinical practice.

Data availability statement

The original contributions presented in the study are included in the article/Supplementary material, further inquiries can be directed to the corresponding author.

Author contributions

IG-A: Formal analysis, Data curation, Writing – original draft, Conceptualization, Investigation, Writing – review & editing, Methodology. FV-F: Formal analysis, Writing – review & editing, Methodology, Funding acquisition, Writing – original draft. EB: Methodology, Writing – review & editing, Investigation, Conceptualization, Formal analysis. JS-V: Investigation, Writing – review & editing, Methodology, Conceptualization. JM-M: Conceptualization, Methodology, Investigation, Writing – review & editing. AB-C: Writing – review & editing, Methodology, Conceptualization, Investigation, Funding acquisition.

Funding

The author(s) declare that financial support was received for the research and/or publication of this article. This work was supported by the Program for Mobility, Cooperation, and Internationalization ‘‘Jiménez de la Espada’’ under Grant 22466/EE/24 of Andrés Bueno-Crespo funded by the Seneca Foundation – Agency for Science and Technology in the Region of Murcia.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The authors declare that Gen AI was used in the creation of this manuscript. During the preparation of this manuscript, the authors used ChatGPT to enhance language and readability. After employing this tool, the authors reviewed and edited the content as necessary and assume full responsibility for the content of the publication.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/frai.2025.1649746/full#supplementary-material

References

Acharya, U. R., Akter, A., Chowriappa, P., Dua, S., Raghavendra, U., Koh, J. E. W., et al. (2018). Use of nonlinear features for automated characterization of suspicious ovarian tumors using ultrasound images in fuzzy forest framework. Int. J. Fuzzy Syst. 20, 1385–1402. doi: 10.1007/s40815-018-0456-9

Crossref Full Text | Google Scholar

Acharya, U. R., Mookiah, M. R. K., Vinitha Sree, S., Yanti, R., Martis, R. J., Saba, L., et al. (2014). Evolutionary algorithm-based classifier parameter tuning for automatic ovarian cancer tissue characterization and classification. Ultraschall Med. 35, 237–245. doi: 10.1055/s-0032-1330336

PubMed Abstract | Crossref Full Text | Google Scholar

Acharya, U. R., Sree, S. V., Kulshreshtha, S., Molinari, F., Koh, J. E. W., Saba, L., et al. (2014). GyneScan: an improved online paradigm for screening of ovarian Cancer via tissue characterization. Technol. Cancer Res. Treat. 13, 529–539. doi: 10.7785/tcrtexpress.2013.600273

PubMed Abstract | Crossref Full Text | Google Scholar

Akazawa, M., and Hashimoto, K. (2021). Artificial intelligence in gynecologic cancers: current status and future challenges – a systematic review. Artif. Intell. Med. 120:102164. doi: 10.1016/j.artmed.2021.102164

PubMed Abstract | Crossref Full Text | Google Scholar

Al-karawi, D., Al-Assam, H., Du, H., Sayasneh, A., Landolfo, C., Timmerman, D., et al. (2021). An evaluation of the effectiveness of image-based texture features extracted from static B-mode ultrasound images in distinguishing between benign and malignant ovarian masses. Ultrason. Imaging 43, 124–138. doi: 10.1177/0161734621998091

PubMed Abstract | Crossref Full Text | Google Scholar

Alwan, O. F., Kadhim, Q. K., Issa, R. B., and Ahmed, S. T. (2023). Early detection and segmentation of ovarian tumor using convolutional neural network with ultrasound imaging. Rev. Intell. Artif. 37, 1503–1509. doi: 10.18280/ria.370614

Crossref Full Text | Google Scholar

Aramendía-Vidaurreta, V., Cabeza, R., Villanueva, A., Navallas, J., and Alcázar, J. L. (2015). Ultrasound image discrimination between benign and malignant adnexal masses based on a neural network approach. Ultrasound Med. Biol. 42, 742–752. doi: 10.1016/j.ultrasmedbio.2015.11.014

Crossref Full Text | Google Scholar

Badgwell, D., and Bast, R. C. (2007). Early detection of ovarian cancer. Dis. Markers 23, 397–410. doi: 10.1155/2007/309382

PubMed Abstract | Crossref Full Text | Google Scholar

Bast, R. C., Brewer, M., Zou, C., Hernandez, M. A., Daley, M., Ozols, R., et al. (2007). Prevention and early detection of ovarian cancer: mission impossible? Recent Results Cancer Res. 174, 91–100. doi: 10.1007/978-3-540-37696-5_9

Crossref Full Text | Google Scholar

Bäumler, M., Gallant, D., Druckmann, R., and Kuhn, W. (2020). Ultrasound screening of ovarian cancer. Horm. Mol. Biol. Clin. Invest. 41:20190022. doi: 10.1515/hmbci-2019-0022

PubMed Abstract | Crossref Full Text | Google Scholar

Bowtell, D., Böhm, S., Ahmed, A., Aspuria, P., Bast, R., and Beral, V. (2015). Rethinking ovarian cancer part II: reducing mortality from high-grade serous ovarian cancer. Nat. Rev. Cancer 15, 668–679. doi: 10.1038/nrc4019

PubMed Abstract | Crossref Full Text | Google Scholar

Chen, H., Yang, B.-W., Qian, L., Meng, Y.-S., Bai, X.-H., Hong, X.-W., et al. (2022). Deep learning prediction of ovarian malignancy at US compared with O-RADS and expert assessment. Radiology 304, 106–113. doi: 10.1148/radiol.211367

PubMed Abstract | Crossref Full Text | Google Scholar

Chiappa, V., Bogani, G., Interlenghi, M., Salvatore, C., Bertolina, F., Sarpietro, G., et al. (2021). The adoption of radiomics and machine learning improves the diagnostic processes of women with ovarian masses (the AROMA pilot study). J. Ultrasound 24, 429–437. doi: 10.1007/s40477-020-00503-5

PubMed Abstract | Crossref Full Text | Google Scholar

Christiansen, F., Epstein, E., Smedberg, E., Åkerlund, M., Smith, K., and Epstein, E. (2021). Ultrasound image analysis using deep neural networks for discriminating between benign and malignant ovarian tumors: comparison with expert subjective assessment. Ultrasound Obstet. Gynecol. 57, 155–163. doi: 10.1002/uog.23530

PubMed Abstract | Crossref Full Text | Google Scholar

Collins, G. S., Dhiman, P., Andaur Navarro, C. L., Ma, J., Hooft, L., Reitsma, J. B., et al. (2021). Protocol for development of a reporting guideline (TRIPOD-AI) and risk of bias tool (PROBAST-AI) for diagnostic and prognostic prediction model studies based on artificial intelligence. BMJ Open 11:e048008. doi: 10.1136/bmjopen-2020-048008

PubMed Abstract | Crossref Full Text | Google Scholar

Dai, W.-L., Wu, Y.-N., Ling, Y.-T., Zhao, J., Zhang, S., Gu, Z.-W., et al. (2024). Development and validation of a deep learning pipeline to diagnose ovarian masses using ultrasound screening: a retrospective multicenter study. EClinicalMedicine 78:102923. doi: 10.1016/j.eclinm.2024.102923

PubMed Abstract | Crossref Full Text | Google Scholar

Dalmartello, M., La Vecchia, C., Bertuccio, P., Boffetta, P., Levi, F., Negri, E., et al. (2022). European cancer mortality predictions for the year 2022 with focus on ovarian cancer. Ann. Oncol. Off. J. Eur. Soc. Med. Oncol. 33, 330–339. doi: 10.1016/j.annonc.2021.12.007

PubMed Abstract | Crossref Full Text | Google Scholar

Du, Y., Guo, W., Xiao, Y., Chen, H., Yao, J., and Wu, J. (2024). Ultrasound-based deep learning radiomics model for differentiating benign, borderline, and malignant ovarian tumours: a multi-class classification exploratory study. BMC Med. Imaging 24:89. doi: 10.1186/s12880-024-01251-2

PubMed Abstract | Crossref Full Text | Google Scholar

Du, Y., Xiao, Y., Guo, W., Yao, J., Lan, T., Li, S., et al. (2024). Development and validation of an ultrasound-based deep learning radiomics nomogram for predicting the malignant risk of ovarian tumours. Biomed. Eng. Online 23:41. doi: 10.1186/s12938-024-01234-y

PubMed Abstract | Crossref Full Text | Google Scholar

Esteva, A., Robicquet, A., Ramsundar, B., Kuleshov, V., DePristo, M., Chou, K., et al. (2019). A guide to deep learning in healthcare. Nat. Med. 25, 24–29. doi: 10.1038/s41591-018-0316-z

PubMed Abstract | Crossref Full Text | Google Scholar

Falana, W. O., Serener, A., and Serte, S. (2023). Deep learning for comparative study of ovarian cancer detection on histopathological images. 2023 7th Int. Symp. Multidiscip. Stud. Innov. Technol., 1–6. doi: 10.1109/ISMSIT58785.2023.10304913

Crossref Full Text | Google Scholar

Fan, J., Liu, J., Chen, Q., Wang, W., and Wu, Y. (2023). Accurate ovarian cyst classification with a lightweight deep learning model for ultrasound images. IEEE Access 11, 110681–110691. doi: 10.1109/ACCESS.2023.3321408

Crossref Full Text | Google Scholar

Gao, Y., Zeng, S., Xu, X., Li, H., Yao, S., Song, K., et al. (2022). Deep learning-enabled pelvic ultrasound images for accurate diagnosis of ovarian cancer in China: a retrospective, multicentre, diagnostic study. Lancet Digit Health 4, e179–e187. doi: 10.1016/S2589-7500(21)00278-8

PubMed Abstract | Crossref Full Text | Google Scholar

Giourga, M., Petropoulos, I., Stavros, S., Potiris, A., Gerede, A., Sapantzoglou, I., et al. (2024). Enhancing ovarian tumor diagnosis: performance of convolutional neural networks in classifying ovarian masses using ultrasound images. J. Clin. Med. 13:4123. doi: 10.3390/jcm13144123

PubMed Abstract | Crossref Full Text | Google Scholar

He, X., Bai, X.-H., Chen, H., and Feng, W.-W. (2024). Machine learning models in evaluating the malignancy risk of ovarian tumors: a comparative study. J. Ovarian Res. 17:219. doi: 10.1186/s13048-024-01544-8

PubMed Abstract | Crossref Full Text | Google Scholar

Heinlein, L., Maron, R. C., Hekler, A., Haggenmüller, S., Wies, C., Utikal, J. S., et al. (2024). Clinical melanoma diagnosis with artificial intelligence: insights from a prospective multicenter study. arXiv. [Preprint]. doi: 10.48550/arXiv.2401.14193

Crossref Full Text | Google Scholar

Ho, D. J., Chui, M. H., Vanderbilt, C. M., Jung, J., Robson, M. E., Park, C.-S., et al. (2022). Deep interactive learning-based ovarian cancer segmentation of H&E-stained whole slide images to study morphological patterns of BRCA mutation. arXiv. [Preprint]. doi: 10.48550/arXiv.2203.15015

Crossref Full Text | Google Scholar

Hussein, I. J., Burhanuddin, M. A., Mohammed, M. A., Benameur, N., Maashi, M. S., and Maashi, M. S. (2022). Fully-automatic identification of gynaecological abnormality using a new adaptive frequency filter and histogram of oriented gradients (HOG). Expert. Syst. 39:e12789. doi: 10.1111/exsy.12789

Crossref Full Text | Google Scholar

Hussein, I. J., Burhanuddin, M., Mohammed, M., Elhoseny, M., Garcia-Zapirain, B., Maashi, M., et al. (2020). Fully automatic segmentation of gynaecological abnormality using a new Viola–Jones model. Comput. Mater. Contin. 66, 3161–3182. doi: 10.32604/cmc.2021.012691

Crossref Full Text | Google Scholar

Jacobs, L., and Menon, U. (2004). Progress and challenges in screening for early detection of ovarian cancer. Mol. Cell Proteomics 3, 355–366. doi: 10.1074/mcp.R400006-MCP200

Crossref Full Text | Google Scholar

Jeevitha, S., and Priya, N. (2022). Identifying and classifying an ovarian cyst using SCBOD (size and count-based ovarian detection) algorithm in ultrasound image. Int. J. Electr. Comp. Eng. Syst. 13, 799–806. doi: 10.32985/ijeces.13.9.8

Crossref Full Text | Google Scholar

Kelly, C. J., Karthikesalingam, A., Suleyman, M., Corrado, G., and King, D. (2019). Key challenges for delivering clinical impact with artificial intelligence. BMC Med. 17:195. doi: 10.1186/s12916-019-1426-2

PubMed Abstract | Crossref Full Text | Google Scholar

Kiruthika, V., Sathiya, S., Ramya, M. M., and Sakthidasan Sankaran, K. (2023). An intelligent machine learning approach for ovarian detection and classification system using ultrasonogram images. Eng. Sci. 23:879. doi: 10.1201/9781003224068-2

PubMed Abstract | Crossref Full Text | Google Scholar

Kongara, S. R., Prakasha, S., Brindha, A., Pathak, S. K., Miya, J., Taqui, S. N., et al. (2024). Performance evaluation of optimized convolutional neural network mechanism in the detection and classification of ovarian cancer. Multimed. Tools Appl. 83, 71311–71334. doi: 10.1007/s11042-024-18115-0

Crossref Full Text | Google Scholar

Li, J., Chen, Y., Zhang, M., Zhang, P., He, K., Yan, F., et al. (2022). A deep learning model system for diagnosis and management of adnexal masses. Cancer 14:5291. doi: 10.3390/cancers14215291

PubMed Abstract | Crossref Full Text | Google Scholar

Li, L., He, L., Guo, W., Ma, J., Sun, G., and Ma, H. (2024). PMFFNet: a hybrid network based on feature pyramid for ovarian tumor segmentation. PLoS One 19:e0299360. doi: 10.1371/journal.pone.0299360

PubMed Abstract | Crossref Full Text | Google Scholar

Liberati, A., Altman, D. G., Tetzlaff, J., Mulrow, C., Gøtzsche, P. C., Ioannidis, J. P. A., et al. (2009). The PRISMA statement for reporting systematic reviews and meta-analyses of studies that evaluate health care interventions: explanation and elaboration. J. Clin. Epidemiol. 62, e1–e34. doi: 10.1016/j.jclinepi.2009.06.006

PubMed Abstract | Crossref Full Text | Google Scholar

Litjens, G., Kooi, T., Bejnordi, B. E., Setio, A. A. A., Ciompi, F., Ghafoorian, M., et al. (2017). A survey on deep learning in medical image analysis. Med. Image Anal. 42, 60–88. doi: 10.1016/j.media.2017.07.005

PubMed Abstract | Crossref Full Text | Google Scholar

Liu, L., Cai, W., Tian, H., Wu, B., Zhang, J., Wang, T., et al. (2024). Ultrasound image-based nomogram combining clinical, radiomics, and deep transfer learning features for automatic classification of ovarian masses according to O-RADS. Front. Oncol. 14:1377489. doi: 10.3389/fonc.2024.1377489

PubMed Abstract | Crossref Full Text | Google Scholar

Liu, L., Cai, W., Zhou, C., Tian, H., Wu, B., Zhang, J., et al. (2024). Ultrasound radiomics-based artificial intelligence model to assist in the differential diagnosis of ovarian endometrioma and ovarian dermoid cyst. Front. Med. 11:11. doi: 10.3389/fmed.2024.1362588

PubMed Abstract | Crossref Full Text | Google Scholar

Liu, X., Faes, L., Kale, A. U., Wagner, S. K., Fu, D. J., Bruynseels, A., et al. (2019). A comparison of deep learning performance against health-care professionals in detecting diseases from medical imaging: a systematic review and meta-analysis. Lancet Digit Health 1, e271–e297. doi: 10.1016/S2589-7500(19)30123-2

PubMed Abstract | Crossref Full Text | Google Scholar

Liu, X., Song, Y., Cheng, P., Liang, B., and Xing, D. (2024). Targeting HER2 in solid tumors: unveiling the structure and novel epitopes. Cancer Treat. Rev. 130:102826. doi: 10.1016/j.ctrv.2024.102826

PubMed Abstract | Crossref Full Text | Google Scholar

Ma, Q., Jiang, H., Ma, L., Zhao, G., Xu, Q., Guo, D., et al. (2023). The moonlighting function of glycolytic enzyme enolase-1 promotes choline phospholipid metabolism and tumor cell proliferation. Proc. Natl. Acad. Sci. USA 120:e2209435120. doi: 10.1073/pnas.2209435120

PubMed Abstract | Crossref Full Text | Google Scholar

Martínez-Más, J., Bueno-Crespo, A., Khazendar, S., Remezal-Solano, M., Martínez-Cendán, J.-P., Jassim, S., et al. (2019). Evaluation of machine learning methods with Fourier transform features for classifying ovarian tumors based on ultrasound images. PLoS One 14:e0219388. doi: 10.1371/journal.pone.0219388

PubMed Abstract | Crossref Full Text | Google Scholar

Meijing, W., Guangxia, C., Shuchang, L., Lijiang, C., Zongmei, T., Min, Y., et al. (2023). Deep convolutional neural networks for multiple histologic types of ovarian tumors classification in ultrasound images. Front. Oncol. 13:1154200. doi: 10.3389/fonc.2023.1154200

Crossref Full Text | Google Scholar

Menze, B. H., Jakab, A., Bauer, S., Kalpathy-Cramer, J., Farahani, K., Kirby, J., et al. (2015). The multimodal brain tumor image segmentation benchmark (BRATS). IEEE Trans. Med. Imaging 34, 1993–2024. doi: 10.1109/TMI.2014.2377694

PubMed Abstract | Crossref Full Text | Google Scholar

Miao, K., Lv, Q., Zhang, L., Zhao, N., and Dong, X. (2024). Discriminative diagnosis of ovarian endometriosis cysts and benign mucinous cystadenomas based on the ConvNeXt algorithm. Eur. J. Obstet. Gynecol. Reprod. Biol. 298, 135–139. doi: 10.1016/j.ejogrb.2024.05.010

PubMed Abstract | Crossref Full Text | Google Scholar

Miao, K., Zhao, N., Lv, Q., He, X., Xu, M., Dong, X., et al. (2023). Prediction of benign and malignant ovarian tumors using Resnet34 on ultrasound images. J. Obstet. Gynaecol. Res. 49, 2910–2917. doi: 10.1111/jog.15788

PubMed Abstract | Crossref Full Text | Google Scholar

Moro, F., Vagni, M., Tran, H. E., Bernardini, F., Mascilini, F., Ciccarone, F., et al. (2024). Radiomics analysis of ultrasound images to discriminate between benign and malignant adnexal masses with solid ultrasound morphology. Ultrasound Obstet. Gynecol. Off. J. Int. Soc. 65, 353–363. doi: 10.1002/uog.27680

Crossref Full Text | Google Scholar

Narmatha, C., Manimegalai, P., Krishnadass, J., Valsalan, P., Manimurugan, S., and Mustafa, M. (2023). Ovarian cysts classification using novel deep reinforcement learning with Harris hawks optimization method. J. Supercomput. 79, 1374–1397. doi: 10.1007/s11227-022-04709-8

Crossref Full Text | Google Scholar

Nebgen, D. R., Lu, K. H., and Bast, R. C. (2019). Novel approaches to ovarian cancer screening. Curr. Oncol. Rep. 21:75. doi: 10.1007/s11912-019-0816-0

PubMed Abstract | Crossref Full Text | Google Scholar

NIH. Cancer stat facts: ovarian cancer. (2020). Available online at: https://seer.cancer.gov/statfacts/html/ovary.html (Accessed June 10, 2025).

Google Scholar

NIH. The cancer imaging archive. (n.d.). Available online at: https://www.cancerimagingarchive.net/ (Accessed June 10, 2025).

Google Scholar

Noseworthy, P. A., Attia, Z. I., Brewer, L. C., Hayes, S. N., Yao, X., Kapa, S., et al. (2020). Assessing and mitigating bias in medical artificial intelligence. Circ. Arrhythm. Electrophysiol. 13:e007988. doi: 10.1161/CIRCEP.119.007988

PubMed Abstract | Crossref Full Text | Google Scholar

Page, M. J., McKenzie, J. E., Bossuyt, P. M., Boutron, I., Hoffmann, T. C., Mulrow, C. D., et al. (2021). The PRISMA 2020 statement: an updated guideline for reporting systematic reviews. BMJ 372:n71. doi: 10.1136/bmj.n71

Crossref Full Text | Google Scholar

Patil, S. D., Deore, P. J., and Patil, V. B. (2024). An intelligent computer aided diagnosis system for classification of ovarian masses using machine learning approach. Int. Res. J. Multidiscip. Technov. 6, 45–57. doi: 10.54392/irjmt2434

Crossref Full Text | Google Scholar

Pham, T.-L., and Le, V.-H. (2024). Ovarian tumors detection and classification from ultrasound images based on YOLOv8. J. Adv. Inf. Technol. 15, 264–275. doi: 10.12720/jait.15.2.264-275

Crossref Full Text | Google Scholar

Raciti, P., Sue, J., Retamero, J. A., Ceballos, R., Godrich, R., Kunz, J. D., et al. (2023). Clinical validation of artificial intelligence–augmented pathology diagnosis demonstrates significant gains in diagnostic accuracy in prostate Cancer detection. Arch. Pathol. Lab Med. 147, 1178–1185. doi: 10.5858/arpa.2022-0066-OA

PubMed Abstract | Crossref Full Text | Google Scholar

Ravishankar, T. N., Makarand Jadhav, H., Satheesh Kumar, N., Ambala, S., and Pillai, N. M. (2023). A deep learning approach for ovarian cysts detection and classification (OCD-FCNN) using fuzzy convolutional neural network. Meas. Sens. 27:100797. doi: 10.1016/j.measen.2023.100797

Crossref Full Text | Google Scholar

Roberts, M., Driggs, D., Thorpe, M., Gilbey, J., Yeung, M., Ursprung, S., et al. (2021). Common pitfalls and recommendations for using machine learning to detect and prognosticate for COVID-19 using chest radiographs and CT scans. Nat. Mach. Intell. 3, 199–217. doi: 10.1038/s42256-021-00307-0

Crossref Full Text | Google Scholar

Rosati, A., Gueli Alletti, S., Capozzi, V. A., Mirandola, M., Vargiu, V., Fedele, C., et al. (2020). Role of ultrasound in the detection of recurrent ovarian cancer: a review of the literature. Gland Surg. 9, 1092–1101. doi: 10.21037/gs-20-357

PubMed Abstract | Crossref Full Text | Google Scholar

Sadeghi, M. H., Sina, S., Omidi, H., Farshchitabrizi, A. H., and Alavi, M. (2024). Deep learning in ovarian cancer diagnosis: a comprehensive review of various imaging modalities. Pol. J. Radiol. 89, 30–48. doi: 10.5114/pjr.2024.134817

PubMed Abstract | Crossref Full Text | Google Scholar

Sahu, S. A., and Shrivastava, D. (2023). A comprehensive review of screening methods for ovarian masses: towards earlier detection. Cureus 15:e48534. doi: 10.7759/cureus.48534

PubMed Abstract | Crossref Full Text | Google Scholar

Sha, M. (2024). Segmentation of ovarian cyst in ultrasound images using AdaResU-net with optimization algorithm and deep learning model. Sci. Rep. 14:18868. doi: 10.1038/s41598-024-69427-y

PubMed Abstract | Crossref Full Text | Google Scholar

Shih-Tien, H., Yu-Jie, S., Chian-Huei, H., Ming-Jer, C., Chien-Hsing, L., and Chih-En, K. (2022). Automatic ovarian tumors recognition system based on ensemble convolutional neural network with ultrasound imaging. BMC Med. Inform. Decis. Mak. 22:298. doi: 10.1186/s12911-022-02047-6

Crossref Full Text | Google Scholar

Siegel, R. L., Miller, K. D., Fuchs, H. E., and Jemal, A. (2021). Cancer statistics, 2021. CA Cancer J. Clin. 71, 7–33. doi: 10.3322/caac.21654

PubMed Abstract | Crossref Full Text | Google Scholar

Sone, K., Toyohara, Y., Taguchi, A., Miyamoto, Y., Tanikawa, M., Uchino-Mori, M., et al. (2021). Application of artificial intelligence in gynecologic malignancies: a review. J. Obstet. Gynaecol. Res. 47, 2577–2585. doi: 10.1111/jog.14818

PubMed Abstract | Crossref Full Text | Google Scholar

Sounderajah, V., Ashrafian, H., Golub, R. M., Shetty, S., De Fauw, J., Hooft, L., et al. (2021). Developing a reporting guideline for artificial intelligence-centred diagnostic test accuracy studies: the STARD-AI protocol. BMJ Open 11:e047709. doi: 10.1136/bmjopen-2020-047709

PubMed Abstract | Crossref Full Text | Google Scholar

Ștefan, P.-A., Lupean, R.-A., Mihu, C. M., Lebovici, A., Oancea, M. D., Hu, L., et al. (2021). Ultrasonography in the diagnosis of adnexal lesions: the role of texture analysis. Diagnostics 11:812. doi: 10.3390/diagnostics11050812

PubMed Abstract | Crossref Full Text | Google Scholar

Taha, A. A., and Hanbury, A. (2015). Metrics for evaluating 3D medical image segmentation: analysis, selection, and tool. BMC Med. Imaging 15:29. doi: 10.1186/s12880-015-0068-x

PubMed Abstract | Crossref Full Text | Google Scholar

Tang, Z.-P., Ma, Z., He, Y., Liu, R.-C., Jin, B.-B., Wen, D.-Y., et al. (2022). Ultrasound-based radiomics for predicting different pathological subtypes of epithelial ovarian cancer before surgery. BMC Med. Imaging 22:147. doi: 10.1186/s12880-022-00879-2

PubMed Abstract | Crossref Full Text | Google Scholar

Torre, L. A., Trabert, B., DeSantis, C. E., Miller, K. D., Samimi, G., Runowicz, C. D., et al. (2018). Ovarian cancer statistics, 2018. CA Cancer J. Clin. 68, 284–296. doi: 10.3322/caac.21456

PubMed Abstract | Crossref Full Text | Google Scholar

Wang, H., Liu, C., Zhao, Z., Zhang, C., Wang, X., Li, H., et al. (2021). Application of deep convolutional neural networks for discriminating benign, borderline, and malignant serous ovarian tumors from ultrasound images. Front. Oncol. 11:770683. doi: 10.3389/fonc.2021.770683

PubMed Abstract | Crossref Full Text | Google Scholar

Wang, Z., Luo, S., Chen, J., Jiao, Y., Cui, C., Shi, S., et al. (2024). Multi-modality deep learning model reaches high prediction accuracy in the diagnosis of ovarian cancer. iScience 27:109403. doi: 10.1016/j.isci.2024.109403

PubMed Abstract | Crossref Full Text | Google Scholar

Wolff, R. F., Moons, K. G. M., Riley, R. D., Whiting, P. F., Westwood, M., Collins, G. S., et al. (2019). PROBAST: a tool to assess the risk of Bias and applicability of prediction model studies. Ann. Intern. Med. 170, 51–58. doi: 10.7326/M18-1376

PubMed Abstract | Crossref Full Text | Google Scholar

Wu, C., Wang, Y., and Wang, F. Deep learning for ovarian tumor classification with ultrasound images. Pacific RIM Conference on Multimedia (2018). p. 395–406.

Google Scholar

Xi, M., Zheng, R., Wang, M., Shi, X., Chen, C., Qian, J., et al. (2023). Ultrasonographic diagnosis of ovarian tumors through the deep convolutional neural network. Ginekol. Pol. 95, 181–189. doi: 10.5603/gpl.94956

PubMed Abstract | Crossref Full Text | Google Scholar

Xiang, H., Xiao, Y., Li, F., Li, C., Liu, L., Deng, T., et al. (2024). Development and validation of an interpretable model integrating multimodal information for improving ovarian cancer diagnosis. Nat. Commun. 15:2681. doi: 10.1038/s41467-024-46700-2

PubMed Abstract | Crossref Full Text | Google Scholar

Xie, W., Lin, W., Li, P., Lai, H., Wang, Z., Liu, P., et al. (2024). Developing a deep learning model for predicting ovarian cancer in ovarian-adnexal reporting and data system ultrasound (O-RADS US) category 4 lesions: a multicenter study. J. Cancer Res. Clin. Oncol. 150:346. doi: 10.1007/s00432-024-05872-6

PubMed Abstract | Crossref Full Text | Google Scholar

Yuyeon, J., Taewan, K., Mi-Ryung, H., Sejin, K., Geunyoung, K., Seungchul, L., et al. (2022). Ovarian tumor diagnosis using deep convolutional neural networks and a denoising convolutional autoencoder. Sci. Rep. 12:17024. doi: 10.1038/s41598-022-20653-2

Crossref Full Text | Google Scholar

Keywords: systematic review, meta-analysis, artificial intelligence, ultrasound, ovarian cancer, early detection, deep learning

Citation: Garcia-Atutxa I, Villanueva-Flores F, Barrio ED, Sanchez-Villamil JI, Martínez-Más J and Bueno-Crespo A (2025) Artificial intelligence for ovarian cancer diagnosis via ultrasound: a systematic review and quantitative assessment of model performance. Front. Artif. Intell. 8:1649746. doi: 10.3389/frai.2025.1649746

Received: 18 June 2025; Accepted: 20 October 2025;
Published: 05 November 2025.

Edited by:

Tim Hulsen, Rotterdam University of Applied Sciences, Netherlands

Reviewed by:

Wenjing Zhu, University of Health and Rehabilitation Sciences (Qingdao Municipal Hospital), China
Uğurcan Zorlu, Yuksek Ihtisas Training and Research Hospital, Türkiye

Copyright © 2025 Garcia-Atutxa, Villanueva-Flores, Barrio, Sanchez-Villamil, Martínez-Más and Bueno-Crespo. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Andrés Bueno-Crespo, YWJ1ZW5vQHVjYW0uZWR1

ORCID: Igor Garcia-Atutxa, orcid.org/0009-0002-1551-2685
Francisca Villanueva-Flores, orcid.org/0000-0001-6092-4211
Ekaitz Dudagotia Barrio, orcid.org/0000-0003-0990-3081
Javier I. Sanchez-Villamil, orcid.org/0000-0003-4928-5994
José Martínez-Más, orcid.org/0000-0002-8506-9933
Andrés Bueno-Crespo, orcid.org/0000-0003-1734-6852

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.