Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Oncol., 03 October 2025

Sec. Genitourinary Oncology

Volume 15 - 2025 | https://doi.org/10.3389/fonc.2025.1661695

This article is part of the Research TopicLeveraging Artificial Intelligence for Biomarker Discovery in Prostate CancerView all 3 articles

Interpretable machine learning models based on multi-dimensional fusion data for predicting positive surgical margins in robot-assisted radical prostatectomy: a retrospective study

  • 1Department of Urology, The First Affiliated Hospital of Chongqing Medical University, Chongqing, China
  • 2Department of Urology, The Second People’s Hospital of Neijiang, Neijiang, Sichuan, China
  • 3Department of Urology, Guangan People's Hospital, Guangan, Sichuan, China
  • 4Department of Radiology, The Second People’s Hospital of Neijiang, Neijiang, Sichuan, China
  • 5Department of Respiratory and Critical Care Medicine, The First People’s Hospital of Neijiang, Neijiang, Sichuan, China
  • 6Department of Medical Informatics Library, Chongqing Medical University, Chongqing, China
  • 7Department of Urology, The University-Town Hospital of Chongqing Medical University, Chongqing, China

Objective: This study aimed to develop and validate interpretable machine learning (ML) models based on multi-dimensional fusion data for predicting positive surgical margins (PSM) in robot-assisted radical prostatectomy (RARP).

Methods: Patients who underwent RARP at our institution between January 2016 and July 2025 were enrolled. Demographic, clinical, biopsy pathology data, and MRI-derived anatomical features (measured using ITK-SNAP on axial, sagittal, and coronal planes) were collected. Feature selection was performed using intraobserver and interobserver correlation coefficients (ICCs), low-variance filtering, univariable logistic regression, Spearman’s correlation analysis, the least absolute shrinkage and selection operator (LASSO) algorithm, and the Boruta algorithm. Six ML models were constructed, with performance evaluated using area under the curve (AUC), calibration curves, and decision curve analyses (DCA) to identify the optimal model. Five-fold and ten-fold cross-validation were used to assess the optimal model’s generalizability, and its interpretability was evaluated via Shapley Additive exPlanations (SHAP) analysis.

Results: A total of 347 patients were included, comprising a training set (n=193, January 2016–December 2024), validation set (n=84, January 2016–December 2024), and test set (n=70, January 2025–July 2025). From 164 initial features, 7 key features were retained through a four-step screening. The Random Forest (RF) model outperformed other models, achieving AUCs of 0.99 (95% CI: 0.97–1.00) in the training set, 0.88 (95% CI: 0.80–0.95) in the validation set, and 0.97 (95% CI: 0.94–1.00) in the test set. Calibration curve and decision curve analyses confirmed its strong clinical utility. Five-fold cross-validation for the RF model showed fold-specific AUCs of 0.82–0.92, with a mean AUC of 0.87 (95% CI: 0.84–0.90). Ten-fold cross-validation showed fold-specific AUCs of 0.80–0.99, with a mean AUC of 0.88 (95% CI: 0.83–0.93). SHAP analysis revealed five novel spatial anatomical features (such as Sagittal plane-posterior spatial anatomical structure index, Coronal plane-Left anatomical structure interval) were negatively associated with PSM risk, while the number of positive biopsy cores and clinical tumor stage were positively associations.

Conclusions: Multi-dimensional fusion data combined with ML models improves PSM prediction accuracy in RARP. The RF model, with excellent performance and interpretability, shows promise for preoperative PSM risk stratification, facilitates optimized clinical decision-making, and supports personalized treatment discussions during preoperative planning, but requires prospective and external validation before clinical implementation.

Introduction

Prostate cancer (PCa) is one of the most common malignancies in men worldwide and ranks fifth among cancer-related deaths in males (1, 2). The “Cancer Statistics 2024” report estimates 299,010 new PCa cases and 35,250 related deaths in 2024 (2). Robot-assisted radical prostatectomy (RARP) is the primary surgical treatment for localized PCa and has become the gold standard for radical prostatectomy (RP) (3, 4), significantly improving overall and tumor-specific survival rates (5). By 2013, up to 80% of RPs in the United States were RARP procedures (6).

Positive surgical margin (PSM) in the prostate specimen following RP is a well-established predictor of biochemical recurrence (BCR) (7, 8). The incidence of PSM is influenced by multiple factors, including preoperative prostate-specific antigen (PSA) levels, clinical tumor stage (cT stage), Gleason score/International Society of Urological Pathology (ISUP) grade group, pathological extension of the primary tumor, and others (915). Patients with PSM face higher risks of BCR, disease progression, additional treatments, and psychological distress, which negatively impact quality of life (1618). Given the diversity of RARP patients, preoperative prediction of surgical complexity and prognostic factors is critical for ensuring safety, optimizing scheduling, enhancing care, and reducing costs (19, 20).

Due to the prostate’s deep location within the pelvic cavity, RARP presents challenges such as limited surgical spatial related to prostate size and pelvic anatomy (21, 22). Recent studies have proposed pelvic measurement indicators to characterize pelvic anatomy (21, 2326) and demonstrated that artificial intelligence (AI) models based on pelvic-prostate spatial features can predict RP surgical difficulty (22, 2729). However, the stability of AI models depends on the quantity and quality of the training set, and existing models lack integration of radiomics, clinical, and biopsy pathology features, limiting their generalizability.

This study aimed to establish and validate a comprehensive machine learning (ML) algorithm integrating multi-dimensional fusion data (radiomics, prostate/pelvic measurements, clinical, and biopsy pathology features) for preoperative PSM prediction in RARP.

Materials and methods

Study cohorts

This retrospective single-center study was conducted at the Department of Urology, The First Affiliated Hospital of Chongqing Medical University, with collaborative support from co-authors at other institutions for data analysis and imaging feature quantification. This study was approved by the Institutional Review Board (IRB) of our hospital (Approval No. K2023-599) (Supplementary Material 1). As a retrospective study, informed consent from patients was waived. All study protocols were in accordance with the Declaration of Helsinki (30). Clinical data (demographic and laboratory variables), mpMRI data (anatomical features and relevant parameters), and biopsy pathological data were anonymized prior to analysis. PSM was defined as tumor cells at the inked surgical margin, regardless of anatomical location.

Training and validation sets

Patients who underwent RARP between January 2016 and December 2024 were enrolled. Exclusion criteria: (1) Missing or poor-quality mpMRI (n=279); (2) Incomplete biopsy pathology data (n=271); (3) Prior PCa treatment (androgen deprivation therapy, radiotherapy and others; n=200); (4) Non-puncture biopsy pathology diagnosis (such as transurethral resection of the prostate (TURP) and light laser vaporization, which may cause edema of the surrounding tissues) (n=70). (5) mpMRI performed after biopsy (n=42); (6) Missing laboratory data (n=17); (7) Other treated malignancies (n=13); (8) Distant metastases (n=4); (9) mpMRI performed more than 5 months before RARP (n=2); (10) Prostatic leiomyosarcoma (n=1); (11) TURP within 1 year (n=1).

Test set

Patients who underwent RARP between January 2025 and July 2025 were enrolled. Exclusion criteria: (1) Missing or poor-quality mpMRI (n=54); (2) Incomplete biopsy pathology data (n=58); (3) Prior PCa treatment (n=45); (4) Non-puncture biopsy pathology diagnosis (n=14); (5) Other treated malignancies (n=2); (6) Distant metastases (n=2); (7) mpMRI performed more than 5 months before RARP (n=2); (8) prostatic leiomyosarcoma (n=0); (9) TURP within 1 year (n=0). The patient screening flowchart is shown in Figure 1.

Figure 1
Flowchart showing the selection criteria and allocation of RARP patients from two periods: 1,177 patients from January 1, 2016, to December 31, 2024, and 247 patients from January 1, 2025, to July 31, 2025. Exclusion criteria include poor-quality mpMRI, prior PCa treatment, and other factors. Enrolled patients total 277 from the first group and 70 from the second. The first group is split into training (193) and validation (84) sets. The second group forms the test set.

Figure 1. Screening flowchart of patients in this study (January 1, 2016, to July 31, 2025). RARP, Robot-Assisted Radical Prostatectomy; mpMRI, Multiparametric Magnetic Resonance Imaging; PCa, Prostate Cancer.

Clinical, mpMRI, and biopsy pathology data collection

Clinical data

A clinician blinded to mpMRI and pathology data extracted 49 features from electronic records, including demographics (age, BMI), lifestyle factors (smoking status, alcohol consumption), comorbidities (hypertension, diabetes), surgical details, laboratory tests (complete blood count, biochemical function, coagulation), and PCa markers (total prostate-specific antigen (tPSA), free prostate-specific antigen (fPSA), fPSA/PSA ratio).

mpMRI data

Imaging was performed using a 3.0 T MR scanner (GE Discovery MR750W, General Healthcare, Milwaukee, USA) with T2-weighted sequences (TR = 3,500 ms, TE = 85 ms, slice thickness=3 mm) and diffusion-weighted imaging (b-values=0, 1,000 s/mm²). Measurements were manually performed by two radiologists (Readers A and B, >8 years of PCa diagnosis experience) using ITK-SNAP (http://www.itksnap.org/). They were blinded to clinicopathological data and assessed Prostate Imaging-Reporting and Data System (PI-RADS) scores, clinical tumour stage, and measured pelvic, prostate, and tumor features. Intraobserver and interobserver correlation coefficients (ICCs) for MRI features ranged from 0.70 to 0.99, indicating good agreement (Supplementary Table 1). Controversial cases were re-evaluated by a senior radiologist (>15 years of experience).

Radiomics features (10 items): PI-RADS score, seminal vesicle invasion (SVI), lymph-node invasion (LNI), lympho-vascular invasion (LVI), perineural invasion, and others.

Anatomical measurements (97 items, Supplementary Table 2, Supplementary Figure 1):

● Axial plane (26 items): thickness of right obturator internus muscle (TROIM), thickness of left obturator internus muscle (TLOIM), distance of outer of the levator ani muscle (DOLAM), and others.

● Sagittal plane (31 items): prostatic urethral length (PUL), membranous urethral length (MUL), membranous urethral angle (MUA), and others.

● Coronal plane (12 items): right anal sphincter thickness (RST), left anal sphincter thickness (LST), thickness of right levator ani muscle (TRLAM), and others.

● Calculated values (28 items): thickness of levator ani muscle (TLAM), prostate-muscle index (PMI), roundness ratio (RR), and others.

Biopsy pathology evaluation

All patients underwent transrectal ultrasound-guided prostate biopsy by a single urologist. Biopsy pathology was reviewed by a senior pathologist (>10 years of PCa experience), blinded to MRI and postoperative pathology. Tumor classification was based on the 2016 WHO criteria, with grading via Gleason score and cancer group grades (31, 32). Eight features were recorded, including biopsy method, number of positive biopsy cores (PBC), Percentage of PBC, and others.

Feature extraction and selection

Features with missing rates <10% were included. For imputation, continuous variables were filled with median values, and categorical variables with mode values to ensure comparability.

A four-step selection process was used: (1) Removal of low-variance features (baseline analysis). (2) Initial screening using univariable logistic regression. (3) Remove redundancy using Spearman’s rank correlation analysis (r ≥ 0.7). (4) Select optimal subsets via LASSO and Boruta algorithms, with the final features being the intersection of both.

Hyperparameter tuning

To optimize each algorithm’s performance, we conducted hyperparameter tuning (33). This modeling process utilized a Bayesian hyperparameter search method (34), which systematically evaluated a comprehensive set of hyperparameter values to identify configurations maximizing efficiency and accuracy. Through this detailed and iterative exploration of the hyperparameter space, we were able to fine-tune the models effectively. This meticulous adjustment ensures that our models are precisely calibrated, significantly enhancing their ability to analyze and predict outcomes accurately with the dataset at hand.

ML model construction, validation, and testing

Six ML models for predicting PSM were built: Logistic Regression (LR), Support Vector Machine (SVM), K-nearest Neighbor (KNN), Decision Tree (DT), Random Forest (RF), and Extreme Gradient Boosting (XGBoost). The receiver operating characteristic (ROC) curve analysis, area under the ROC curve (AUC), accuracy (ACC), sensitivity (SEN), specificity (SPE), positive predictive value (PPV), negative predictive value (NPV), and F1 score were calculated to evaluate model performance. To compare the predictive performance and clinical utility of the constructed ML models, the DeLong test, calibration curve analysis with Brier score loss, and decision curve analysis were conducted. A lower Brier score indicated superior model calibration.

Cross-validation of the target model

To further validate model robustness, five-fold and ten-fold cross-validation were performed for the optimal model. In N-fold cross-validation, the dataset is divided into N equal folds; the model is trained on N-1 folds and validated on the remaining fold in each iteration. This process is repeated N times, with the final performance metric derived by averaging results to ensure a robust assessment (35).

Machine learning model interpretation

The optimal model was interpreted using SHAP (Shapley Additive exPlanations) analysis (36, 37). Based on cooperative game theory, SHAP quantifies each feature’s contribution to model predictions by evaluating its marginal impact across all feature combinations, ensuring a balanced representation of feature importance. It provides interpretability at two scales: (1) Local interpretability: clarifies individual predictions by quantifying feature contributions; (2) Global interpretability: synthesizes features’ relative impacts across the entire dataset. The workflow is illustrated in Figure 2.

Figure 2
Flowchart showcasing a machine learning process for medical analysis. It includes four main sections: “All feature” with demographic, laboratory, and pathology data; “Features selection” using statistical analyses and algorithms like LASSO and Boruta; “Machine learning models” establishment and comparison through ROC and calibration curves; and “Model interpretation” with SHAP methodology, illustrating feature importance using bar charts and plots. Various graphs and images detail each section.

Figure 2. The overall workflow of this study. LASSO, Least Absolute Shrinkage and Selection Operator; SHAP, Shapley Additive exPlanations; ROC, Receiver Operating Characteristic Curve.

Statistical analysis

SPSS 25.0 (SPSS, Armonk, NY, USA), R software (version 4.3.1; https://www.r-project.org/), and Python (version 3.8.0; https://www.python.org/) were used for statistical analysis. Continuous variables were presented as medians with interquartile ranges (IQRs) and compared using Mann–Whitney U tests. Categorical data were presented as counts (percentages) and compared using chi-square, Fisher’s exact test, or Yates’ continuity correction. Accuracy, sensitivity, specificity, PPV, and NPV based on the optimal cutoff (Youden index) were calculated, with 95% confidence intervals (CIs) estimated using 1,000 bootstraps. A two-tailed P-value < 0.05 was considered statistically significant.

Results

Clinical characteristics

A total of 347 patients (median age: 70 years, IQR: 65.00-74.00 years) were included, with 238 (68.6%) negative surgical margins (NSM) and 109 (31.4%) PSM. No significant differences in clinical, mpMRI, or biopsy pathology features were observed between the training and validation sets (all P > 0.05; Table 1).

Table 1
www.frontiersin.org

Table 1. Baseline characteristics (training vs. validation sets).

Features selection

Based on previous studies (915, 2128) and the authors’ interests. From 164 initial features, 7 key features were retained through four-step screening (Figure 3): (1) 8 low-variance features were excluded (Supplementary Table 3). (2) 119 features unrelated to PSM were excluded via univariable logistic regression analysis (Supplementary Table 4). (3) 16 redundant features (r ≥ 0.7) were removed via Spearman’s rank correlation analysis (Supplementary Figure 2). (4) The LASSO logistic algorithm and the Boruta algorithm retained 10 and 9 features (Supplementary Figure 3), with final features as their intersection. The 7 features (Supplementary Figure 3) included:

▪ Number of positive biopsy cores (Number of PBC).

▪ Clinical tumor stage (cT stage).

▪ Sagittal plane-posterior spatial anatomical structure index (S-PSAI).

▪ Sagittal plane-total spatial anatomical structure index (S-TSAI).

▪ Coronal plane-left anatomical structure interval (C-LAI).

▪ Coronal plane-right anatomical structure interval (C-RAI).

▪ Axial plane-inferior margin of symphysis pubis-bilateral ischial spinous angle (A-SP-BIS Angle).

Figure 3
Flowchart illustrating feature selection for analysis. Initial 164 features undergo low variance filtering, univariate logistic regression reducing to 155, and Spearman's rank correlation down to 37 features. Those with correlation greater than 0.7 are narrowed to 21. Further selection through LASSO logistic (10 features) and Boruta algorithm (9 features), intersecting at 7 enrolled features. Final selected features include cT stage, Number of PBC, S-PSAI, S-TSAI, C-LAI, C-RAI, and A-SP-BIS Angle.

Figure 3. Screening flowchart of key features for model establishment in this study. cT stage, Clinical primary tumor stage. Number of PBC, Number of positive biopsy cores. S-PSAI, Sagittal plane-Posterior spatial anatomical structure index. S-TSAI, Sagittal plane-Total spatial anatomical structure index. C-LAI, Coronal plane-Left anatomical structure interval. C-RAI, Coronal plane-Right anatomical structure interval. A-SP-BIS Angle, Axial plane-Inferior margin of symphysis pubis - bilateral ischial spinous angle. LASSO, the least absolute shrinkage and selection operator.

No significant differences in these 7 features were observed across the training, validation, and test sets (all P > 0.05; Table 2).

Table 2
www.frontiersin.org

Table 2. Key features comparison in the training, validation, and test sets.

ML model establishment

Table 3 summarizes model parameters, adjustment ranges, and optimal values via Bayesian optimization. Among the six models, KNN and XGB showed high training-set AUCs of 1.00 (95% CI: 1.00-1.00), suggesting overfitting. The RF model achieved optimal balanced performance: (1) Training set: AUC of 0.99 (95% CI: 0.97–1.00), accuracy of 0.94. (2) Validation set: AUC of 0.88 (95% CI: 0.80–0.95), accuracy of 0.83. (3) Test set: AUC of 0.97 (95% CI: 0.94–1.00), accuracy of 0.93. The DT, SVM, and LR models ranked as the second, third, and fourth predicting models in the training set. The DT, LR, and SVM models ranked as the second, third, and fourth predicting models in the validation set. The LR, SVM, and DT models ranked as the second, third, and fourth predicting models in the test set. Model performance metrics are detailed in Table 4 and Figures 4A–F.

Table 3
www.frontiersin.org

Table 3. Model parameters screening via Bayesian analysis.

Table 4
www.frontiersin.org

Table 4. Predictive performance of six ML models.

Figure 4
ROC and radar charts for machine learning models across training, validation, and test sets. Charts A-C show ROC curves with AUC values for LR, SVM, KNN, DT, RF, and XGB models. Radar charts D-F compare metrics like accuracy, AUC, sensitivity, specificity, PPV, and NPV for the same models across the datasets. Each chart highlights the performance differences among the models in varying conditions.

Figure 4. Predictive performance of six ML models. This figure presents the Receiver Operating Characteristic (ROC) curve analysis of the established models in the training (A), validation (B), and test (C) sets, as well as the radar plots of the models’ prediction metrics in the training (D), validation (E), and test (F) sets. LR, Logistic Regression; SVM, Support Vector Machine; KNN, K-Nearest Neighbors; DT, Decision Tree; RF, Random Forest; XGB, Extreme Gradient Boosting; AUC, Area Under the ROC Curve; PPV, Positive Predictive Value; NPV, Negative Predictive Value.

Comparison of ML models

DeLong tests confirmed the RF model outperformed LR, SVM, and DT in the training set (all P < 0.05) but not KNN/XGB (overfitting models). In validation and test sets, RF outperformed all five other models (positive Z-scores). In the validation set, RF showed non-significant differences vs. LR/XGB (P > 0.05) but superiority vs. SVM/KNN/DT (P < 0.05). In the test set, RF showed non-significant differences vs. LR/SVM/XGB (P > 0.05) but superiority vs. KNN/DT (P < 0.05) (Table 5).

Table 5
www.frontiersin.org

Table 5. Results of DeLong’s test analysis comparing AUCs of the RF model with those of other ML models.

Also, RF had the lowest Brier scores (except overfitted KNN/XGB models) and well-matched calibration curves (Figures 5A–C), with the highest net benefit in most threshold probabilities at decision curve analysis (Figures 5D–F).

Figure 5
Graphs depicting model performance across training, validation, and test sets. Panels A-C show calibration curves with true vs. predicted probabilities for various models: LR, SVM, KNN, DT, RF, XGB. Brier scores vary across sets. Panels D-F display decision curve analyses, illustrating net benefit against high-risk thresholds for the same models. Different colored lines represent different models, with “Treat All” and “Treat None” as baselines.

Figure 5. Clinical utility evaluation via calibration and decision curves. Calibration curve analysis of the six models in the training (A), validation (B) and test (C) sets. Decision curve analysis of the six models in the training (D), validation (E) and test (F) sets. "Treat All" and "Treat None" are baselines. Different colors represent models. Brier score, measure of calibration (lower = better).

Robustness checks

Five-fold cross-validation for RF showed fold-specific AUCs of 0.82–0.92, with a mean AUC of 0.87 (95% CI: 0.84–0.90). Ten-fold cross-validation showed fold-specific AUCs of 0.80–0.99, with a mean AUC of 0.88 (95% CI: 0.83–0.93), indicating stable performance (Figure 6).

Figure 6
ROC curves with cross-validation for two datasets. Panel A shows five curves with AUC values ranging from 0.82 to 0.92 and a mean ROC of 0.87. Panel B displays ten curves with AUC values ranging from 0.80 to 0.99 and a mean ROC of 0.88. Both panels include a dashed line representing chance.

Figure 6. Five-fold and ten-fold cross-validation of the RF model. (A) Five-fold cross-validation ROC curves of the RF model (AUC values range from 0.82 to 0.92, mean AUC = 0.87, 95% CI: 0.84–0.90); (B) Ten-fold cross-validation ROC curves of the RF model (AUC values range from 0.80 to 0.99, mean AUC = 0.88, 95% CI: 0.83–0.93). The dashed line represents chance (AUC = 0.5). RF, Random Forest; AUC, Area Under the ROC Curve; CI, Confidence Interval.

SHAP interpretation of the RF model

Feature importance rankings were consistent across the training and validation datasets: S-PSAI > C-LAI > S-TSAI > A-SP-BIS Angle > C-RAI > Number of PBC > cT stage (Figures 7A, B). In the test set, the ranking in the test set was: S-SPAI > S-TSAI > C-LAI > A-SP-BIS Angle > C-RAI > Number of PBC > cT stage (Figures 7A-C). Five spatial features (S-PSAI, C-LAI, S-TSAI, A-SP-BIS Angle, C-RAI) were negatively associated with PSM risk, while Number of PBC and cT stage were positively associated (Figures 7D–F). The SHAP decision plot illustrates the influences of all contributing features on the final predicted probability (Figures 7G-I). SHAP dependence plots further clarified feature relationships (Figure 8). Representative cases (NSM vs. PSM) illustrated feature contributions of each of the 7 key features within the RF model (Figure 9).

Figure 7
Grouped bar and violin plots showing SHAP values for three sets: Training, Validation, and Test. Panels A, B, and C display bar plots of mean SHAP values, indicating feature impact on model output. Panels D, E, and F present violin plots of SHAP values with color gradients representing feature value by impact. Panels G, H, and I show parallel plots of model output values and feature impact. Features include S-PSAI, C-LAI, S-TSAI, A-SP-BIS Angle, C-RAI, Number of PBC, and cT stage.

Figure 7. The SHAP analysis of the RF model. Bar plot of mean SHAP values (feature importance) in the training, validation and test sets (A–C); Bee-swarm plot of SHAP values (color gradients represent feature value by impact) in the training, validation and test sets (D–F); Parallel plot of model output values and feature impact in the training, validation and test sets (G–I); SHAP, Shapley Additive exPlanations; RF, Random Forest; S-PSAI, Sagittal plane-posterior spatial anatomical structure index; C-LAI, Coronal plane-left anatomical structure interval; S-TSAI, Sagittal plane-total spatial anatomical structure index; A-SP-BIS Angle, Axial plane-inferior margin of symphysis pubis-bilateral ischial spinous angle; C-RAI, Coronal plane-right anatomical structure interval; Number of PBC, Number of positive biopsy cores; cT stage, Clinical primary tumor Stage.

Figure 8
Scatter plots show SHAP values versus various parameters for training, validation, and test sets. Each row corresponds to a different dataset: training (n=193), validation (n=84), and test (n=70). Parameters include S-PISA, C-LAI, S-TSIA, A-Sp-PBS Angle, C-RAI, and Number of PBC. Each plot highlights the relationship between a parameter and SHAP values, with a red dashed line indicating zero influence.

Figure 8. SHAP dependence plots for seven features. SHAP dependence plot is a visualization tool in SHAP tools used to analyze the influence of a single feature on model prediction and its interaction with other features, which reveals the direct impact (positive or negative) and nonlinear relationship of features on the prediction results by presenting the relationship between feature values and the corresponding SHAP values.

Figure 9
Three sets of bar charts show feature importance for predicting surgical margin status in patients. Each set contains two charts comparing positive and negative margins. The charts are divided into training, validation, and test sets, highlighting features such as PSA, BMI, and stages with corresponding weights. Blue bars indicate negative margins, while pink bars indicate positive margins, with values showing the impact of each feature.

Figure 9. Representative NSM (A, C, E) and PSM (B, D, F) cases predicted by the RF model. Representative NSM (negative surgical margin) case in the training (A), validation (C) and test (E) sets; Representative PSM (positive surgical margin) case in the training (B), validation (D) and test (F) sets. The distinct contributions of each feature within the RF model for individual predictions are illustrated using the SHAP waterfall plot. RF, random forest. SHAP, SHapley Additive exPlanations.

Discussion

To our knowledge, this is the first study to investigate ML models that integrate clinical, mpMRI, and biopsy pathology data for predicting PSM before RARP. The RF model exhibited excellent performance across the training, validation, and test sets, with its robustness validated via cross-validation. SHAP analysis identified the feature importance rankings, thereby improving model transparency. This innovative approach will improve preoperative surgical risk stratification, optimize clinical decision-making processes, and establish a framework for automated robotic surgery case screening, ultimately advancing the precision and individualization of RARP therapeutic strategies.

Previous studies have identified predictors of PSM, including surgical experience, body mass index (BMI), tPSA, extracapsular extension (ECE), neurovascular bundle (NVB) invasion, cT stage, percentage of positive biopsy cores, number of positive biopsy cores, Gleason score (GS), pathological stage (pT stage), time interval between prostate biopsy and surgery, among others (914), but these studies focused on single feature types. Multiparametric MRI (mpMRI), a standard PCa imaging tool (3841), provides critical anatomical insights, with prior studies linking prostate and pelvic dimensions (e.g., prostate volume (PV), pelvic dimension index (PDI)/PV ratio, prostate-muscle index (PMI), apical depth (AD), symphysis angle (SA), transverse diameter of the pelvic entrance and intertuberous distance (ITD), among others) to PSM risk (2124, 26, 28, 42, 43). However, focusing exclusively on a single category of features while neglecting their holistic nature when evaluating PSM in RARP offers a limited perspective. This study addressed this limitation by integrating 164 features across multiple domains and screening 7 key predictors through a rigorous multi-step selection process, thereby ensuring the scientific validity and rigor of the selected features.

Optimal hyperparameter tuning is critical for ML performance (44). Data-efficient optimization algorithms, such as Bayesian optimization (44), were employed to automate this process, screening both the parameter adjustment range and optimal parameters. Based on evaluation metrics, KNN and XGB approached or reached a value of 1 for accuracy, AUC, and other metrics, indicating overfitting. The KNN and XGB regression methods were susceptible to overfitting and fit discontinuity, which remain significant challenges in the field (45). In contrast, the RF model obtained suitable AUCs of 0.99 (95% CI: 0.97-1.00), 0.88 (95% CI: 0.80-0.95), and 0.97 (95% CI: 0.94-1.00) in the training, validation, and test sets. The RF outperformed LR, SVM, and DT, indicating superior generalization. Its high accuracy (0.94, 0.83, 0.93), specificity (0.96, 0.86, 0.93), and sensitivity (0.91, 0.78, 0.92) across datasets confirm its predictive reliability. Overall, these results confirm that the RF model is the optimal classifier, consistent with previous studies (46, 47).

Notably, the DeLong test confirmed that in the training set, the AUCs of the RF model were superior to those of LR, SVM, and DT, but lower than those of KNN and XGB. Given that KNN and XGB are overfitting models, these comparisons lack practical significance. The AUCs of the RF model were comparable to those of LR and XGB in the validation set, but superior to those of SVM, KNN, and DT. The AUCs of the RF model were comparable to those of LR, SVM, and XGB in the test set, but superior to those of KNN and DT. The DeLong test confirmed that the overall advantage of the RF model holds, but this advantage has practical value for high-variance models (KNN/DT/SVM models). For LR and XGB models in the validation and test sets, the RF model does not exhibit a significant advantage, which may be attributed to the insufficient sample size of the current validation and test sets. In addition, the RF model exhibited the optimal calibration (lower Brier score, well-aligned calibration curves) and the highest net benefit across most threshold probabilities (decision curve analysis). In conclusion, the RF model demonstrated excellent performance in terms of sensitivity, specificity, accuracy, ROC, and F1 score across the three sets, affirming its predictive reliability and clinical decision-support value. Five-fold and ten-fold cross-validation confirmed the stable performance of the RF model. These results suggest that the RF model could facilitate the identification of surgical difficulty, guide personalized surgical planning, and optimize resource allocation (such as assigning experienced surgeons), thereby reducing the risk of biochemical recurrence after surgery.

ML models are often criticized as “black boxes” (48, 49), which limits their clinical acceptance, particularly in critical applications such as healthcare, where transparency and reliability in clinical decision-making tools are crucial (50, 51). To address this challenge, researchers have focused on developing methods to improve the interpretability of these models, such as SHapley Additive exPlanations (SHAP) analysis, which assigns contribution values to individual features in the dataset to indicate the extent of each feature’s influence on predicted outcomes. This holistic approach enables researchers to identify which features most significantly impact outcomes and whether their influence is positive or negative, thereby promoting the acceptance of ML-based diagnostic or predictive tools in clinical settings (47, 5256). To our knowledge, this is the first study to investigate ML models based on multi-dimensional fusion data that use SHAP methods for PSM prediction. The contribution relationships of the 7 selected features were successfully visualized using SHAP bar plots, bee-swarm plots, and decision plots. As a result, our study identified that 5 newly discovered spatial features were negatively associated with PSM, with S-PSAI being the most influential. In preoperative PSM prediction, the RF model assigns the highest importance to this feature. Specifically, lower values of S-PSAI, S-TSAI, C-LAI, and C-RAI indicate limited surgical space, which increases surgical difficulty and the risk of tumor residue, prompting surgeons to adjust dissection techniques (such as expanding the resection range or performing more meticulous operations) or assign experienced surgeons. A narrow A-SP-BIS Angle could inform surgical planning (such as adjusting port placement to improve access). However, the number of PBC and cT stage are positively correlated with PSM, consistent with previous studies (11, 12, 15). Surgeons can use feature contributions to prioritize intraoperative vigilance in high-risk regions. SHAP visualizations enhance transparency, thereby improving trust in model-derived decisions and patient understanding and compliance.

This study has several key strengths: (1) Robust data quality: Strict inclusion and exclusion criteria were applied, and enrolled cases underwent rigorous screening; (2) Comprehensive feature integration: Integration of radiomics, prostate and pelvic measurements, clinical, and biopsy pathological features, comprehensively covering factors influencing PSM; (3) Scientific rigor in feature screening: A four-step screening process (low-variance elimination, univariate regression, Spearman correlation-based redundancy removal, and intersection of LASSO and Boruta algorithms) was applied to 164 initial features; (4) Model diversity and validation: Six ML algorithms were used to develop models, and stable predictive models were identified through comparison, with robustness confirmed via sensitivity analysis and N-fold (5-fold and 10-fold) cross-validation; (5) Model interpretability: The SHAP method was employed to clarify the model’s decision-making process, providing valuable insights into its predictive mechanism.

Several limitations of this study should be acknowledged. (1) Retrospective bias: The retrospective design introduced inevitable selection bias, highlighting the need for prospective studies with predefined criteria. (2) Selection bias: The high exclusion rate (primarily due to missing data) may introduce selection bias, as patients with complete data may differ from those excluded. (3) Limited generalizability: Although the dataset is larger than those in previous studies, the single-center Chinese cohort limits global applicability, necessitating external and multi-ethnic validation. (4) Automation limitations: Automatic mpMRI feature recognition is lacking and will be addressed in future work. (5) Restricted surgical scope: Patients undergoing laparoscopic RP were excluded because the surgical assistant’s experience level may impact outcomes, and thus the generalizability of the RF model to RP patients requires further investigation. (6) Data limitations: Genomic data (e.g., PTEN deletion) were not included; future iterations will integrate genomic data and long-term functional outcomes to provide a more comprehensive risk assessment.

In conclusion, ML models based on multi-dimensional fusion data improve PSM prediction in RARP. The RF model, with robust performance and SHAP-based interpretability, enhances preoperative risk stratification, optimizes decision-making, and supports personalized treatment, thereby improving patient treatment compliance and potentially enhancing patient outcomes. Prospective and external validation are required prior to clinical implementation.

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Ethics statement

The studies involving humans were approved by Clinical Research Ethics Committee of the First Affiliated Hospital of Chongqing Medical University. The studies were conducted in accordance with the local legislation and institutional requirements. The ethics committee/institutional review board waived the requirement of written informed consent for participation from the participants or the participants’ legal guardians/next of kin because As a retrospective study, informed consent from patients was waived. All study protocols were in accordance with the Declaration of Helsinki.

Author contributions

ZL: Conceptualization, Data curation, Formal Analysis, Investigation, Methodology, Project administration, Resources, Validation, Visualization, Writing – original draft. WZ: Data curation, Formal Analysis, Investigation, Writing – original draft. PD: Data curation, Formal Analysis, Software, Writing – review & editing. JL: Data curation, Formal Analysis, Writing – review & editing. LL: Data curation, Formal Analysis, Writing – review & editing. YL: Investigation, Methodology, Writing – review & editing. SS: Investigation, Methodology, Writing – review & editing. SJS: Investigation, Methodology, Writing – review & editing. YW: Data curation, Methodology, Writing – review & editing. LHL: Data curation, Methodology, Writing – review & editing. YZ: Data curation, Methodology, Writing – review & editing. SQ: Data curation, Methodology, Writing – review & editing. LJ: Data curation, Methodology, Writing – review & editing. KH: Data curation, Methodology, Writing – review & editing. JZ: Data curation, Formal Analysis, Investigation, Methodology, Project administration, Writing – review & editing. JH: Data curation, Formal Analysis, Funding acquisition, Investigation, Methodology, Project administration, Writing – review & editing. DW: Conceptualization, Funding acquisition, Investigation, Project administration, Resources, Writing – review & editing.

Funding

The author(s) declare financial support was received for the research and/or publication of this article. This work was funded by the Chongqing Technology Innovation and Application Development Special Project (Grant Number: CSTB2023TIAD-KPX0053 and CSTB2024TIAD-GPX0025).

Acknowledgments

The authors thank the editors and reviewers for their valuable time and feedback.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declare that no Generative AI was used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fonc.2025.1661695/full#supplementary-material

Supplementary Figure 1 | Definitions and images of parameters in MRI measurements. Note: MRI measurement abbreviations, names, and definitions were detailed in Supplementary Table 2.

Supplementary Figure 2 | Feature elimination via Spearman’s rank correlation analysis. Note: A-RSAI, Axial plane-Right spatial anatomical structure index. A-RAI, Axial plane-Right anatomical structure interval. A-SP-BIS Angle, Axial plane-Inferior margin of symphysis pubis - bilateral ischial spinous angle. A-ASP, Axial plane-The angle of the symphysis pubis. C-TSAI, Coronal plane-Total spatial anatomical structure index. C-RSAI, Coronal plane-Right spatial anatomical structure index. C-LSAI, Coronal plane-Left spatial anatomical structure index. C-TAI, Coronal plane-Total anatomical structure interval. C-RAI, Coronal plane-Right anatomical structure interval. C-LAI, Coronal plane-Left anatomical structure interval. S-TSAI, Sagittal plane-Total spatial anatomical structure index. S-PSAI, Sagittal plane-Posterior spatial anatomical structure index. S-ASAI, Sagittal plane-Anterior spatial anatomical structure index. S-TAI, Sagittal plane-Total anatomical structure interval. S-PAI, Sagittal plane-Posterior anatomical structure interval. S-AAI, Sagittal plane-Anterior anatomical structure interval. S-SP-PA-CA Angle, Sagittal plane-Symphysis pubis-prostate apical-coccyx apical angle. S-SP-PA-S5 Angle, Sagittal plane-Symphysis pubis-prostate apical-S5 angle. S-SP-PA-S1 Angle, Sagittal plane-Symphysis pubis-prostate apical-S1 angle. S-MTSP-IMSPA Angle, Sagittal plane-Angle between the medial tangent of the symphysis pubis, and the line connecting the inferior margin of the symphysis pubis and the prostate apical. S-LASP-API Angle, Sagittal plane-Angle between the long axis diameter of the symphysis pubis and anteroposterior diameter of the pelvic outlet angle. PSAD, Total prostate-specific antigen density. AFSI, Anterior fibromuscular stroma invasion. UI, Urethral invasion. A-CCL-PZ, Axial plane-Cumulative contact length-prostate peripheral zone. A-LLD, Axial plane-Largest lesion diameter. cT stage, Clinical primary tumor stage. GGG, Gleason grade group. GS, Gleason score. PGG, Primary Gleason grade. tPSA, Total prostate-specific antigen. fPSA, Free prostate-specific antigen. ALT, Alanine aminotransferase. AST, Aspartate aminotransferas. LND, Lymph node dissection.

Supplementary Figure 3 | Feature screening via LASSO and Boruta algorithms. Note: LND, Lymph node dissection. A-ASP, Axial plane-The angle of the symphysis pubis. A-RAI, Axial plane-Right anatomical structure interval. PGG, Primary Gleason grade. S-LASP-API Angle, Sagittal plane-Angle between the long axis diameter of the symphysis pubis and anteroposterior diameter of the pelvic outlet angle. fPSA, Free prostate-specific antigen. AFSI, Anterior fibromuscular stroma invasion. S-SP-PA-S5 Angle, Sagittal plane-Symphysis pubis-prostate apical-S5 angle. AST, Aspartate aminotransferas. S-MTSP-IMSPA Angle, Sagittal plane-Angle between the medial tangent of the symphysis pubis, and the line connecting the inferior margin of the symphysis pubis and the prostate apical. UI, Urethral invasion. cT stage, Clinical primary tumor stage. PSAD, Total prostate-specific antigen density. Number of PBC, Number of positive biopsy cores. A-LLD, Axial plane-Largest lesion diameter. C-RAI, Coronal plane-Right anatomical structure interval. C-LAI, Coronal plane-Left anatomical structure interval. S-TSAI, Sagittal plane-Total spatial anatomical structure index. A-SP-BIS Angle, Axial plane-Inferior margin of symphysis pubis - bilateral ischial spinous angle. S-PSAI, Sagittal plane-Posterior spatial anatomical structure index. LASSO, the least absolute shrinkage and selection operator.

Supplementary Figure 4 | Definitions and images of five key features in MRI measurements. Note: (A) Axial plane of MRI on T2 showing the maximum area of the prostate. A-SP-BIS-Angle was defined as the angle between the lines connecting the inferior margin of the symphysis pubis to the medial aspects of the bilateral ischial spinous processes. (B) Sagittal plane of MRI on T2 showing the maximum area of the prostate. Prostate maximum anteroposterior diameter(PAD) was defined as The maximum anteroposterior diameter at the largest area of the prostate(bc). Prostate maximum craniocaudal diameter(PCD) was defined as The maximum craniocaudal diameter at the largest area of the prostate(ef). Anterior anatomical structure interval diameter(S-AAI) was defined as the horizontal distance between the posterior margin of the symphysis pubis and the anterior margin of the prostate, which was measured at the level of the maximum anteroposterior diameter with the largest area of the prostate (ab). Posterior anatomical structure interval diameter(S-PAI) was defined as the horizontal distance between the posterior margin of the prostate and the anterior margin of the rectum, which was measured at the level of the maximum anteroposterior diameter with the largest area of the prostate (cd). Posterior spatial anatomical structure index(S-PSAI), Calculated based on the formula: (S-RAI) / (S-PAD); Total spatial anatomical structure index(S-TSAI), Calculated based on the formula: ('S-AAI'+'S-PAI') / (S-PAD). (C) Coronal plane of MRI on T2 showing the maximum area of the prostate. Prostate transverse diameter(C-PTD) was defined as the maximum transverse diameter at the largest area of the prostate(bc). Prostate craniocaudal diameter(C-PCD) was defined as the maximum craniocaudal diameter at the largest area of the prostate and the inner edge of the left obturator internus muscle(ef). Left anatomical structure interval(C-LAI) was defined as the horizontal distance between the left edge at the maximum transverse diameter of the prostate and the inner edge of the left obturator internus muscle(cd). Right anatomical structure interval(C-RAI) was defined as the horizontal distance between the horizontal distance between the right edge at the maximum transverse diameter of the prostate and the inner edge of the right obturator internus muscle(ab). P, prostate. SP, symphysis pubis. OIM, obturator internus muscle. Fe, femoral bone. R, rectum. Bl, bladder.

References

1. Bray F, Laversanne M, Sung H, Ferlay J, Siegel RL, Soerjomataram I, et al. Global cancer statistics 2022: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA Cancer J Clin. (2024) 74:229–63. doi: 10.3322/caac.21834

PubMed Abstract | Crossref Full Text | Google Scholar

2. Siegel RL, Giaquinto AN, and Jemal A. Cancer statistics, 2024. CA Cancer J Clin. (2024) 74:12–49. doi: 10.3322/caac.21820

PubMed Abstract | Crossref Full Text | Google Scholar

3. Hu JC, Gu X, Lipsitz SR, Barry MJ, D'Amico AV, Weinberg AC, et al. Comparative effectiveness of minimally invasive vs open radical prostatectomy. Jama. (2009) 302:1557–64. doi: 10.1001/jama.2009.1451

PubMed Abstract | Crossref Full Text | Google Scholar

4. Patel VR, Sivaraman A, Coelho RF, Chauhan S, Palmer KJ, Orvieto MA, et al. Pentafecta: a new concept for reporting outcomes of robot-assisted laparoscopic radical prostatectomy. Eur Urol. (2011) 59:702–07. doi: 10.1016/j.eururo.2011.01.032

PubMed Abstract | Crossref Full Text | Google Scholar

5. Mottet N, van den Bergh R, Briers E, Van den Broeck T, Cumberbatch MG, De Santis M, et al. EAU-EANM-ESTRO-ESUR-SIOG guidelines on prostate cancer-2020 update. Part 1: screening, diagnosis, and local treatment with curative intent. Eur Urol. (2021) 79:243–62. doi: 10.1016/j.eururo.2020.09.042

PubMed Abstract | Crossref Full Text | Google Scholar

6. Fantus RJ, Cohen A, Riedinger CB, Kuchta K, Wang CH, Yao K, et al. Facility-level analysis of robot utilization across disciplines in the National Cancer Database. J Robot Surg. (2019) 13:293–99. doi: 10.1007/s11701-018-0855-9

PubMed Abstract | Crossref Full Text | Google Scholar

7. Stephenson AJ, Kattan MW, Eastham JA, Dotan ZA, Bianco FJJ, Lilja H, et al. Defining biochemical recurrence of prostate cancer after radical prostatectomy: a proposal for a standardized definition. J Clin Oncol. (2006) 24:3973–78. doi: 10.1200/JCO.2005.04.0756

PubMed Abstract | Crossref Full Text | Google Scholar

8. Zhang L, Wu B, Zha Z, Zhao H, Jiang Y, and Yuan J. Positive surgical margin is associated with biochemical recurrence risk following radical prostatectomy: a meta-analysis from high-quality retrospective cohort studies. World J Surg Oncol. (2018) 16:124. doi: 10.1186/s12957-018-1433-3

PubMed Abstract | Crossref Full Text | Google Scholar

9. Udo K, Cronin AM, Carlino LJ, Savage CJ, Maschino AC, Al-Ahmadie HA, et al. Prognostic impact of subclassification of radical prostatectomy positive margins by linear extent and Gleason grade. J Urol. (2013) 189:1302–07. doi: 10.1016/j.juro.2012.10.004

PubMed Abstract | Crossref Full Text | Google Scholar

10. Patel VR, Coelho RF, Rocco B, Orvieto M, Sivaraman A, Palmer KJ, et al. Positive surgical margins after robotic assisted radical prostatectomy: a multi-institutional study. J Urol. (2011) 186:511–16. doi: 10.1016/j.juro.2011.03.112

PubMed Abstract | Crossref Full Text | Google Scholar

11. Ficarra V, Novara G, Secco S, D'Elia C, Boscolo-Berto R, Gardiman M, et al. Predictors of positive surgical margins after laparoscopic robot assisted radical prostatectomy. J Urol. (2009) 182:2682–88. doi: 10.1016/j.juro.2009.08.037

PubMed Abstract | Crossref Full Text | Google Scholar

12. Yang R, Cao K, Han T, Zhang Y, Zhang G, Xu L, et al. Perineural invasion status, Gleason score and number of positive cores in biopsy pathology are predictors of positive surgical margin following laparoscopic radical prostatectomy. Asian J Androl. (2017) 19:468–72. doi: 10.4103/1008-682X.173444

PubMed Abstract | Crossref Full Text | Google Scholar

13. Turan T, Boylu U, Başataç C, and Gümüş E. Predictive preoperative factors for positive surgical margins in robotic radical prostatectomy in low-risk prostate cancer. Turk J Urol. (2013) 39:69–73. doi: 10.5152/tud.2013.020

PubMed Abstract | Crossref Full Text | Google Scholar

14. Yossepowitch O, Briganti A, Eastham JA, Epstein J, Graefen M, Montironi R, et al. Positive surgical margins after radical prostatectomy: a systematic review and contemporary update. Eur Urol. (2014) 65:303–13. doi: 10.1016/j.eururo.2013.07.039

PubMed Abstract | Crossref Full Text | Google Scholar

15. Lee RS, Ma R, Pham S, Maya-Silva J, Nguyen JH, Aron M, et al. Machine learning to delineate surgeon and clinical factors that anticipate positive surgical margins after robot-assisted radical prostatectomy. J Endourol. (2022) 36:1192–98. doi: 10.1089/end.2021.0890

PubMed Abstract | Crossref Full Text | Google Scholar

16. Kanehira M, Takata R, Ishii S, Ito A, Ikarashi D, Matsuura T, et al. Predictive factors for short-term biochemical recurrence-free survival after robot-assisted laparoscopic radical prostatectomy in high-risk prostate cancer patients. Int J Clin Oncol. (2019) 24:1099–104. doi: 10.1007/s10147-019-01445-7

PubMed Abstract | Crossref Full Text | Google Scholar

17. Rosas-Nava JE, Herranz-Amo F, Paños-Fagundo EV, Lledó-García E, Verdú-Tartajo F, and Hernández-Fernández C. Value of positive resection margins in patients with pT2 prostate cancer. Implications for adjuvant treatment. Actas Urol Esp. (2011) 35:272–76. doi: 10.1016/j.acuro.2011.01.004

PubMed Abstract | Crossref Full Text | Google Scholar

18. Hong YM, Hu JC, Paciorek AT, Knight SJ, and Carroll PR. Impact of radical prostatectomy positive surgical margins on fear of cancer recurrence: results from CaPSURE. Urol Oncol. (2010) 28:268–73. doi: 10.1016/j.urolonc.2008.07.004

PubMed Abstract | Crossref Full Text | Google Scholar

19. Wu S, Lin SX, Wirth GJ, Lu M, Lu J, Subtelny AO, et al. Impact of multifocality and multilocation of positive surgical margin after radical prostatectomy on predicting oncological outcome. Clin Genitourin Cancer. (2019) 17:e44–52. doi: 10.1016/j.clgc.2018.08.007

PubMed Abstract | Crossref Full Text | Google Scholar

20. Lu J, Wirth GJ, Wu S, Chen J, Dahl DM, Olumi AF, et al. A close surgical margin after radical prostatectomy is an independent predictor of recurrence. J Urol. (2012) 188:91–7. doi: 10.1016/j.juro.2012.02.2565

PubMed Abstract | Crossref Full Text | Google Scholar

21. Chen J, Chu T, Ghodoussipour S, Bowman S, Patel H, King K, et al. Effect of surgeon experience and bony pelvic dimensions on surgical performance and patient outcomes in robot-assisted radical prostatectomy. BJU Int. (2019) 124:828–35. doi: 10.1111/bju.14857

PubMed Abstract | Crossref Full Text | Google Scholar

22. Lu H, Yu C, Yu X, Yang D, Yu S, Xia L, et al. Effects of bony pelvic and prostate dimensions on surgical difficulty of robot-assisted radical prostatectomy: an original study and meta-analysis. Ann Surg Oncol. (2024) 31:8405–20. doi: 10.1245/s10434-024-15769-w

PubMed Abstract | Crossref Full Text | Google Scholar

23. Mason BM, Hakimi AA, Faleck D, Chernyak V, Rozenblitt A, and Ghavamian R. The role of preoperative endo-rectal coil magnetic resonance imaging in predicting surgical difficulty for robotic prostatectomy. Urology. (2010) 76:1130–35. doi: 10.1016/j.urology.2010.05.037

PubMed Abstract | Crossref Full Text | Google Scholar

24. Matikainen MP, von Bodman CJ, Secin FP, Yunis LH, Vora K, Guillonneau B, et al. The depth of the prostatic apex is an independent predictor of positive apical margins at radical prostatectomy. BJU Int. (2010) 106:622–26. doi: 10.1111/j.1464-410X.2009.09184.x

PubMed Abstract | Crossref Full Text | Google Scholar

25. Yao A, Iwamoto H, Masago T, Morizane S, Honda M, Sejima T, et al. Anatomical dimensions using preoperative magnetic resonance imaging: impact on the learning curve of robot-assisted laparoscopic prostatectomy. Int J Urol. (2015) 22:74–9. doi: 10.1111/iju.12602

PubMed Abstract | Crossref Full Text | Google Scholar

26. Kimura N, Yamada Y, Takeshima Y, Otsuka M, Akamatsu N, Hakozaki Y, et al. The 'prostate-muscle index': a simple pelvic cavity measurement predicting estimated blood loss and console time in robot-assisted radical prostatectomy. Sci Rep. (2022) 12:11945. doi: 10.1038/s41598-022-16202-6

PubMed Abstract | Crossref Full Text | Google Scholar

27. Mei H, Wang Z, Zheng Q, Jiao P, Wu J, Liu X, et al. Deep learning for predicting difficulty in radical prostatectomy: A novel evaluation scheme. Urology. (2025) 198:1–07. doi: 10.1016/j.urology.2025.01.028

PubMed Abstract | Crossref Full Text | Google Scholar

28. Tezcan S, Ozturk E, Savran B, Ciledag N, Ulu Ozturk F, Keten T, et al. Value of the newly developed pelvic dimension index/prostate volume ratio in predicting positive surgical margin in prostate cancer. Int Urol Nephrol. (2023) 55:3111–17. doi: 10.1007/s11255-023-03750-7

PubMed Abstract | Crossref Full Text | Google Scholar

29. Wen J, Liu W, Zhang Y, and Shen X. MRI-based radiomics for prediction of extraprostatic extension of prostate cancer: a systematic review and meta-analysis. Radiol Med. (2024) 129:702–11. doi: 10.1007/s11547-024-01810-1

PubMed Abstract | Crossref Full Text | Google Scholar

30. Goodyear MDE, Krleza-Jeric K, and Lemmens T. The declaration of helsinki. Bmj. (2007) 335:624–25. doi: 10.1136/bmj.39339.610000.BE

PubMed Abstract | Crossref Full Text | Google Scholar

31. Epstein JI, Egevad L, Amin MB, Delahunt B, Srigley JR, and Humphrey PA. The 2014 international society of urological pathology (ISUP) consensus conference on gleason grading of prostatic carcinoma: definition of grading patterns and proposal for a new grading system. Am J Surg Pathol. (2016) 40:244–52. doi: 10.1097/PAS.0000000000000530

PubMed Abstract | Crossref Full Text | Google Scholar

32. Humphrey PA, Moch H, Cubilla AL, Ulbright TM, and Reuter VE. The 2016 WHO classification of tumours of the urinary system and male genital organs-part B: prostate and bladder tumours. Eur Urol. (2016) 70:106–19. doi: 10.1016/j.eururo.2016.02.028

PubMed Abstract | Crossref Full Text | Google Scholar

33. Sejuti ZA and Islam MS. A hybrid CNN-KNN approach for identification of COVID-19 with 5-fold cross validation. Sens Int. (2023) 4:100229. doi: 10.1016/j.sintl.2023.100229

PubMed Abstract | Crossref Full Text | Google Scholar

34. Lu Q, Polyzos KD, Li B, and Giannakis GB. Surrogate modeling for bayesian optimization beyond a single gaussian process. IEEE Trans Pattern Anal Mach Intell. (2023) 45:11283–96. doi: 10.1109/TPAMI.2023.3264741

PubMed Abstract | Crossref Full Text | Google Scholar

35. Lin Y, Lin P, Lin C, Wu T, Liu L, Su C, et al. Adding nutritional status to the original BCLC stage improves mortality prediction for hepatocellular carcinoma patients in HBV-endemic regions. Am J Cancer Res. (2023) 13:3618–28.

PubMed Abstract | Google Scholar

36. Ali S, Akhlaq F, Imran AS, Kastrati Z, Daudpota SM, and Moosa M. The enlightening role of explainable artificial intelligence in medical & healthcare domains: A systematic literature review. Comput Biol Med. (2023) 166:107555. doi: 10.1016/j.compbiomed.2023.107555

PubMed Abstract | Crossref Full Text | Google Scholar

37. Loh HW, Ooi CP, Seoni S, Barua PD, Molinari F, and Acharya UR. Application of explainable artificial intelligence for healthcare: A systematic review of the last decade (2011-2022). Comput Methods Programs BioMed. (2022) 226:107161. doi: 10.1016/j.cmpb.2022.107161

PubMed Abstract | Crossref Full Text | Google Scholar

38. Baxter MT, Conlin CC, Bagrodia A, Barrett T, Bartsch H, Brau A, et al. Advanced restriction imaging and reconstruction technology for prostate magnetic resonance imaging (ART-pro): A study protocol for a multicenter, multinational trial evaluating biparametric magnetic resonance imaging and advanced, quantitative diffusion magnetic resonance imaging for the detection of prostate cancer. Eur Urol Open Sci. (2025) 71:132–43. doi: 10.1016/j.euros.2024.12.003

PubMed Abstract | Crossref Full Text | Google Scholar

39. Wang C, Shen D, Yuan L, Dong Q, Xu S, Liu Y, et al. Creating a novel multiparametric magnetic resonance imaging-based biopsy strategy for reducing unnecessary prostate biopsies: a retrospective cohort study. Quant Imaging Med Surg. (2024) 14:2021–33. doi: 10.21037/qims-23-875

PubMed Abstract | Crossref Full Text | Google Scholar

40. Patel P, Wang S, and Siddiqui MM. The use of multiparametric magnetic resonance imaging (mpMRI) in the detection, evaluation, and surveillance of clinically significant prostate cancer (csPCa). Curr Urol Rep. (2019) 20:60. doi: 10.1007/s11934-019-0926-0

PubMed Abstract | Crossref Full Text | Google Scholar

41. Álvarez-Maestro M, Gómez Rivas J, Quesada Olarte J, Carrión DM, Trelles Guzman C, Ballesteros C, et al. Magnetic resonance as imaging diagnostic tool in prostate cancer: New evidences-The EAU Section of Uro-Technology position. Actas Urol Esp (Engl Ed). (2020) 44:148–55. doi: 10.1016/j.acuro.2019.08.003

PubMed Abstract | Crossref Full Text | Google Scholar

42. von Bodman C, Matikainen MP, Yunis LH, Laudone V, Scardino PT, Akin O, et al. Ethnic variation in pelvimetric measures and its impact on positive surgical margins at radical prostatectomy. Urology. (2010) 76:1092–96. doi: 10.1016/j.urology.2010.02.020

PubMed Abstract | Crossref Full Text | Google Scholar

43. Neill MG, Lockwood GA, McCluskey SA, and Fleshner NE. Preoperative evaluation of the "hostile pelvis" in radical prostatectomy with computed tomographic pelvimetry. BJU Int. (2007) 99:534–38. doi: 10.1111/j.1464-410X.2006.06640.x

PubMed Abstract | Crossref Full Text | Google Scholar

44. Berkenkamp F, Krause A, and Schoellig AP. Bayesian optimization with safety constraints: safe and automatic parameter tuning in robotics. Mach Learn. (2023) 112:3713–47. doi: 10.1007/s10994-021-06019-1

PubMed Abstract | Crossref Full Text | Google Scholar

45. Srisuradetchai P and Suksrikran K. Random kernel k-nearest neighbors regression. Front Big Data. (2024) 7:1402384. doi: 10.3389/fdata.2024.1402384

PubMed Abstract | Crossref Full Text | Google Scholar

46. Li L, Gu L, Kang B, Yang J, Wu Y, Liu H, et al. Evaluation of the efficiency of MRI-based radiomics classifiers in the diagnosis of prostate lesions. Front Oncol. (2022) 12:934108. doi: 10.3389/fonc.2022.934108

PubMed Abstract | Crossref Full Text | Google Scholar

47. Zhou W, Liu Z, Zhang J, Su S, Luo Y, Jiang L, et al. Interpretable multiparametric MRI radiomics-based machine learning model for preoperative differentiation between benign and Malignant prostate masses: a diagnostic, multicenter study. Front Oncol. (2025) 15:1541618. doi: 10.3389/fonc.2025.1541618

PubMed Abstract | Crossref Full Text | Google Scholar

48. Rasheed K, Qayyum A, Ghaly M, Al-Fuqaha A, Razi A, and Qadir J. Explainable, trustworthy, and ethical machine learning for healthcare: A survey. Comput Biol Med. (2022) 149:106043. doi: 10.1016/j.compbiomed.2022.106043

PubMed Abstract | Crossref Full Text | Google Scholar

49. Karim MR, Islam T, Shajalal M, Beyan O, Lange C, Cochez M, et al. Explainable AI for bioinformatics: methods, tools and applications. Brief Bioinform. (2023) 24. doi: 10.1093/bib/bbad236

PubMed Abstract | Crossref Full Text | Google Scholar

50. Kerr WT and McFarlane KN. Machine learning and artificial intelligence applications to epilepsy: a review for the practicing epileptologist. Curr Neurol Neurosci Rep. (2023) 23:869–79. doi: 10.1007/s11910-023-01318-7

PubMed Abstract | Crossref Full Text | Google Scholar

51. Luo J, Pan M, Mo K, Mao Y, and Zou D. Emerging role of artificial intelligence in diagnosis, classification and clinical management of glioma. Semin Cancer Biol. (2023) 91:110–23. doi: 10.1016/j.semcancer.2023.03.006

PubMed Abstract | Crossref Full Text | Google Scholar

52. Ma M, Liu R, Wen C, Xu W, Xu Z, Wang S, et al. Predicting the molecular subtype of breast cancer and identifying interpretable imaging features using machine learning algorithms. Eur Radiol. (2022) 32:1652–62. doi: 10.1007/s00330-021-08271-4

PubMed Abstract | Crossref Full Text | Google Scholar

53. Yun K, He T, Zhen S, Quan M, Yang X, Man D, et al. Development and validation of explainable machine-learning models for carotid atherosclerosis early screening. J Transl Med. (2023) 21:353. doi: 10.1186/s12967-023-04093-8

PubMed Abstract | Crossref Full Text | Google Scholar

54. Liu Z, Luo C, Chen X, Feng Y, Feng J, Zhang R, et al. Noninvasive prediction of perineural invasion in intrahepatic cholangiocarcinoma by clinicoradiological features and computed tomography radiomics based on interpretable machine learning: a multicenter cohort study. Int J Surg. (2024) 110:1039–51. doi: 10.1097/JS9.0000000000000881

PubMed Abstract | Crossref Full Text | Google Scholar

55. Sun J, Sun C, Tang Y, Liu T, and Lu C. Application of SHAP for explainable machine learning on age-based subgrouping mammography questionnaire data for positive mammography prediction and risk factor identification. Healthcare (Basel). (2023) 11:2000. doi: 10.3390/healthcare11142000

PubMed Abstract | Crossref Full Text | Google Scholar

56. Moulaei K, Afshari L, Moulaei R, Sabet B, Mousavi SM, and Afrash MR. Explainable artificial intelligence for stroke prediction through comparison of deep learning and machine learning models. Sci Rep. (2024) 14:31392. doi: 10.1038/s41598-024-82931-5

PubMed Abstract | Crossref Full Text | Google Scholar

Keywords: prostate cancer, robot-assisted radical prostatectomy (RARP), multi-dimensional fusion data, multiparametric magnetic resonance imaging (mpMRI), machine learning, interpretation

Citation: Liu Z, Zhou W, Dong P, Liu J, Luo L, Luo Y, Su S, Sankoh SJ, Wang Y, Liu L, Zhang Y, Qiu S, Jiang L, Han K, Zhang J, He J and Wang D (2025) Interpretable machine learning models based on multi-dimensional fusion data for predicting positive surgical margins in robot-assisted radical prostatectomy: a retrospective study. Front. Oncol. 15:1661695. doi: 10.3389/fonc.2025.1661695

Received: 08 July 2025; Accepted: 17 September 2025;
Published: 03 October 2025.

Edited by:

Neil Mendhiratta, George Washington University, United States

Reviewed by:

Bernhard Ralla, Charité University Medicine Berlin, Germany
Karan Jatwani, George Washington University Hospital, United States

Copyright © 2025 Liu, Zhou, Dong, Liu, Luo, Luo, Su, Sankoh, Wang, Liu, Zhang, Qiu, Jiang, Han, Zhang, He and Wang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Delin Wang, V2FuZ2RlbGluX2NxbXVAMTYzLmNvbQ==; Jiang He, Nzg3NjQ0Nzc0QHFxLmNvbQ==; Jindong Zhang, MjA0OTY5QGhvc3BpdGFsLmNxbXUuZWR1LmNu

These authors have contributed equally to this work

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.