Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Urol., 19 January 2026

Sec. Male Urology

Volume 5 - 2025 | https://doi.org/10.3389/fruro.2025.1726795

This article is part of the Research TopicArtificial Intelligence and Medical Image ProcessingView all 10 articles

Lesion detection using artificial intelligence models in MR images of prostate cancer and prostatitis patients and comparison of model performance

Muhammed Kaya*&#x;Muhammed Kaya1*†Osman DurdagOsman Durdag2Merve SolakMerve Solak1Ahmet CoskuncayAhmet Coskuncay3Gulen Burakgazi*Gulen Burakgazi1*
  • 1Faculty of Medicine, Recep Tayyip Erdoğan University, Rize, Türkiye
  • 2Department of Software Engineering, Ataturk University, Erzurum, Türkiye
  • 3Department of Computer Engineering, Ataturk University, Erzurum, Türkiye

Aim: The diagnosis of prostate cancer and prostatitis becomes challenging when using biparametric Magnetic Resonance (MR) images. This research investigates deep learning models to assess their capability for improving diagnostic accuracy and assisting radiologists.

Methods: This retrospective study analyzed 153 patients who received histopathological diagnoses of prostate cancer or prostatitis between January 2017 and December 2023. Patients were categorized according to PI-RADS scores, and both T2A and ADC-DWI (Apparent Diffusion Coefficient–Diffusion-Weighted Imaging) sequences were examined. Expert radiologists labeled the images prior to lesion detection with the Faster R-CNN (Faster Region-based Convolutional Neural Network) model. Nine different classification models were trained using normal and augmented datasets to evaluate their performance. Model reliability was further assessed through cross-validation and statistical significance testing.

Results: The Faster R-CNN model achieved 96% accuracy (95% CI: 93.2–98.8%) for P5 and 99% accuracy (95% CI: 96.7–100%) for prostatitis in T2A sequences, and 90% accuracy (95% CI: 85.4–94.6%) for P5 and 97% accuracy (95% CI: 93.8–100%) for prostatitis in ADC-DWI sequences. However, the model failed to effectively detect P4 lesions (0% sensitivity in T2A and 30% in ADC-DWI). The model demonstrated comparable performance to expert radiologists, with no significant difference in overall P5 detection (p > 0.05), and Cohen’s kappa indicated substantial agreement (κ = 0.86). The classification models achieved up to 97% accuracy with InceptionV3 in T2A sequences and up to 99% accuracy with DenseNet201 in ADC-DWI sequences. To further evaluate discriminative performance, AUROC values were calculated for all classification models. In T2A sequences, AUROC scores were DenseNet201 (0.98), EfficientNetV2L (0.99), InceptionV3 (0.99), MobileNetV2 (0.92), NASNetLarge (0.83), ResNet50 (0.76), VGG16 (0.98), VGG19 (0.97), and Xception (0.96). In ADC-DWI sequences, AUROC values were DenseNet201 (0.99), EfficientNetV2L (0.96), InceptionV3 (0.99), MobileNetV2 (0.82), NASNetLarge (0.90), ResNet50 (0.64), VGG16 (0.96), VGG19 (0.86), and Xception (0.97), reinforcing the superior discriminative ability of DenseNet201 and InceptionV3 across modalities.

Conclusion: The deep learning models demonstrated promising diagnostic capabilities, comparable to radiologists, in distinguishing prostatitis and P5 prostate cancer lesions. Overall, the findings suggest that AI-based diagnostic tools hold potential as clinical decision support systems.

1 Introduction

Prostate cancer stands as the second most fatal type of cancer that affects men (1, 2). The clinical challenge of accurate prostate cancer diagnosis has driven the evolution of diagnostic approaches from traditional methods to advanced imaging and artificial intelligence-assisted techniques.

The diagnosis of prostate cancer relies on three main methods which include prostate-specific antigen (PSA) testing and digital rectal examination (DRE) and systematic prostate biopsy under transrectal ultrasound (TRUS) guidance (36). The glycoprotein PSA functions as a biomarker but shows poor cancer specificity because it originates from prostate epithelial cells. Serum PSA levels increase in patients with BPH and prostatitis and after prostate biopsy procedures because this test lacks sufficient cancer specificity (68). The diagnostic value of PSA measurements is restricted since histopathological confirmation of PCa through TRUS-guided biopsy occurs in less than half of patients (810). The anatomical coverage limitations and random systematic biopsy approach of TRUS produce high false-negative rates (11, 12).

The development of MRI technology has revolutionized the process of diagnosing prostate cancer. The diagnostic capabilities of MRI significantly improved through the addition of functional imaging sequences including diffusion-weighted imaging (DWI) and dynamic contrast-enhanced MRI (DCE-MRI) in the late 1990s, leading to enhanced PCa detection (13, 14). The combination of these techniques led to the development of multiparametric MRI (mpMRI) in the 2000s, enabling detailed anatomical and functional examination of prostate tissue using T2-weighted imaging, DWI, DCE-MRI, and, when necessary, magnetic resonance spectroscopy (MRS) sequences (15).

Since the 2010s, mpMRI has become the standard imaging method for the diagnosis, local staging, and treatment planning of PCa (16). Structured reporting systems, particularly the Prostate Imaging Reporting and Data System (PI-RADS), were developed to improve both diagnostic precision and reporting uniformity in mpMRI assessments (1719). Currently, mpMRI serves as the primary non-invasive imaging technique for both selecting biopsy locations and guiding biopsy targets (19).

Medical imaging has experienced a transformative shift through artificial intelligence (AI) which enhances both diagnostic precision and operational speed. AI applications have shown significant advancement in prostate cancer diagnosis through automated image analysis and pattern recognition and clinical decision support systems (20). Research shows that deep learning algorithms specifically convolutional neural networks (CNNs) achieve superior results in complex imaging data analysis by matching or surpassing the diagnostic abilities of trained radiologists (20).

The current developments in AI-assisted prostate MRI analysis include automated lesion detection, segmentation, and characterization. The systems use different deep learning architectures such as U-Net, Mask R-CNN, and Faster R-CNN for different analytical tasks (21). The implementation of these deep learning-based algorithms has demonstrated high accuracy in performing lesion segmentation, PI-RADS scoring, and malignancy prediction (2123). These systems demonstrate potential to decrease inter-reader variability while enhancing diagnostic accuracy (21). The technology demonstrates potential for differentiating prostate cancer from benign conditions which represents a major diagnostic obstacle (24).

The literature contains multiple essential gaps despite recent progress. The majority of research studies detect cancer but fail to distinguish between cancerous and benign conditions. The majority of AI models were developed and validated using limited datasets from individual institutions which creates doubts about their ability to generalize. The current evidence base lacks sufficient data to compare AI diagnostic accuracy with that of experienced radiologists in actual clinical settings (21, 24).

In this study, deep learning-based artificial intelligence models were used to automatically detect prostate lesions and perform zonal localization (transitional zone [TZ] and peripheral zone [PZ]) with high accuracy using biparametric MR (bpMR) images from patients with histopathologically confirmed PCa or prostatitis. The research focuses on developing precise methods to distinguish prostate cancer from prostatitis because this distinction determines important clinical treatment choices. The diagnostic performance of the developed models was compared with visual evaluations by experienced radiologists to scientifically demonstrate the contribution of AI-supported analysis compared to classical interpretation methods based on human observation. The research uses advanced deep learning architectures together with thorough validation methods to establish strong evidence about AI utility in prostate imaging. In this context, the feasibility of AI as a decision support tool in prostate imaging was evaluated, and its potential for clinical integration was explored.

2 Methods

2.1 Study design and sample size calculation

The research followed STARD (Standards for Reporting of Diagnostic Accuracy Studies) guidelines in its retrospective cohort study design. The G*Power 3.1 software performed power analysis to establish the smallest necessary sample size. The literature shows AI-based prostate cancer detection achieves sensitivity rates between 85-90% so we used 90% sensitivity and 85% specificity with a 60% disease prevalence in our biopsy-confirmed population at an alpha error of 0.05 and power of 0.80 to calculate a minimum sample size of 138 patients. The research included 153 patients which exceeded the calculated minimum sample size for sufficient statistical power.

2.2 Study population and patient selection criteria

The research examined patients with confirmed prostate cancer or prostatitis diagnoses at our hospital from January 2017 to December 2023. The study included participants who met the following criteria: (1) aged 45–85 years; (2) histopathological confirmation of prostate cancer or prostatitis through systematic 12-core TRUS-guided biopsy or MRI-targeted biopsy performed within 6 weeks of MRI; (3) complete biparametric MRI sequences (T2-weighted and DWI/ADC); (4) no prior prostate surgery, radiation therapy, or hormonal treatment; (5) PSA levels between 2.5–50 ng/mL at the time of MRI. The exclusion criteria consisted of: (1) MRI sequences that were either incomplete or motion-degraded with artifact score >2 on a 5-point scale; (2) hip prosthesis or other metallic implants that caused significant susceptibility artifacts; (3) the time interval between MRI and biopsy was more than 6 weeks; (4) other malignancies were present; (5) insufficient histopathological sampling (<10 cores for systematic biopsy). The study included 153 patient images from a total of 180 patients because 23 patients failed exclusion criteria and 4 patients had lost or corrupted data before model training.

2.2.1 Important clarification

PI-RADS scores were used for risk stratification and initial categorization, not as definitive diagnoses. All cases were confirmed through histopathological examination. For analysis purposes, histopathologically confirmed prostate cancer lesions smaller than 1.5 cm were labeled as P4, whereas those larger than 1.5 cm were labeled as cancer. Lesions histopathologically confirmed as prostatitis were labeled as prostatitis. In the T2 sequence, lesions diagnosed as prostatitis and categorized under the P3 label comprise 277 training images and 71 test images. Among the 1,906 training images labeled as prostate cancer, only 65 were assigned to the P4 category, while the remaining 1,841 were classified as P5; similarly, of the 488 test images, 13 were labeled as P4 and 475 as P5. In the ADC–DWI sequence, prostatitis cases corresponding to the P3 label include 267 training and 69 test images. Within this modality, of the 1,965 training images associated with prostate cancer, only 95 were categorized as P4 and 1,870 as P5; likewise, among the 489 test images, 19 were assigned to the P4 class and 470 to the P5 class However, due to the insufficient number of P4-labeled samples—resulting in limited statistical power and suboptimal model performance—and considering that the P4 label inherently denotes cancer, the P4 and P5 categories were consolidated and treated as a single cancer class during the classification stages of the study. This approach ensures that our model training was based on ground truth pathology results rather than imaging scores alone.

2.3 Imaging protocol and data acquisition

All patients underwent biparametric MR imaging using a standardized protocol on a 3.0 Tesla MR device (GE Discovery MR750w, GE Healthcare, Waukesha, Wisconsin, USA) with a 32-channel phased-array body coil. Standard imaging protocols included T2-weighted turbo spin-echo scans with TR = 4068 ms, TE = 130 ms, field of view=180×180 mm, matrix=320×320, and slice thickness of 3.5 mm with no gap, and diffusion-weighted echo-planar imaging scans with TR = 4600 ms, TE = 90 ms, field of view=240×240 mm, matrix=128×128, and slice thickness of 4 mm at b50, b800, and b1400 values. ADC maps were automatically generated using monoexponential fitting.

2.4 Image annotation and inter-rater reliability

The entire imaging dataset underwent evaluation by two specialized radiologists with 3 and 5 years of prostate MRI experience, working independently and blinded to histopathological results. Final verification was performed by a senior radiologist who has practiced for 15 years to resolve discrepancies. The reliability between raters was evaluated through Cohen’s kappa coefficient for p4, p5(cancer) and prostatitis (κ=0.82, indicating substantial agreement) and intraclass correlation coefficient (ICC) for lesion measurements (ICC = 0.89, 95% CI: 0.84-0.93, indicating excellent reliability). The data collection tool reliability and validity was confirmed through MakeSense platform labeling and XML format coordinate recording with experienced radiologist verification (Figure 1). All annotations were performed according to PI-RADS v2.1 guidelines (25).

Figure 1
Images labeled A, B, and C depict different aspects of medical imaging and data annotation. A shows an MRI with color-coded regions marked. B displays a label creation interface for tagging objects, with labels such as “pz” and “tz.” C features a blurred image with a red rectangular region labeled by coordinates, titled “Apparent Diffusion Coefficient.

Figure 1. Data labeling methodology and coordinate system. (A) The axial T2 MR image displays multiple labels for peripheral zone (pz - purple), transitional zone (tz - pink) and cancer (cancer - orange). (B) The MakeSense platform interface and label creation panel. (C) The ADC image shows a coordinate system which displays the xmin, ymin (top left corner) and xmax, ymax (bottom right corner) coordinates of the label. The image coordinate system begins at the top left corner (0,0) and extends to the right and downwards.

2.5 Image processing and data augmentation

Standardization of T2 images involved transforming various resolution levels into 512×512 pixels using bilinear interpolation, followed by ADC image standardization to 256×256 pixels. Images were normalized to zero mean and unit variance. The OpenCV library’s convertScaleAbs function was applied to data augmentation protocols for image processing through two sets of parameters (alpha=2, beta=0 and alpha=1.5, beta=25) to enhance contrast and brightness variations. The dataset received additional augmentation through random rotation (± 15 degrees) and horizontal flipping (probability=0.5). The images generated through the data augmentation process were reviewed by expert radiologists, who confirmed that no distortions or alterations occurred in the morphology of the prostate tissue. The summary of the dataset obtained through data augmentation is presented in Figure 2.

Figure 2
Bar chart titled “Summary of Data” showing five categories: T2, Augmented T2, ADC, Augmented ADC, with series p3, p4, cancer, tz, and pz. Augmented T2 and ADC have higher values, especially in tz and p3.

Figure 2. A chart summarizing the dataset composed of T2 and ADC images, along with the corresponding augmented dataset.

2.6 Deep learning model architecture

2.6.1 Faster R-CNN architecture details

The Faster R-CNN model architecture consisted of a ResNet-50 backbone pre-trained on ImageNet, a Region Proposal Network (RPN) with 3×3 convolutional layers, and a Fast R-CNN detector head. The RPN used three anchor scales (128², 256², 512²) and three aspect ratios (1:2, 1:1, 2:1) at each spatial position. Non-maximum suppression with IoU threshold of 0.7 was applied to reduce redundant detections. The model received integration through Facebook AI Research’s detectron2 library version 0.6 during the model training protocol. The training hyperparameters consisted of a learning rate set to 0.001 with step decay at epochs 60 and 80, momentum set to 0.9, weight decay set to 0.0001, batch size set to 4, and the number of epochs was set to 90 with early stopping patience of 10 epochs based on validation loss. Training was performed on the Google Colab platform using NVIDIA Tesla P100 GPU with 16GB memory.

2.6.2 Classification model architecture

The classification models employed transfer learning with ImageNet pre-trained weights for nine architectures: InceptionV3, Xception, DenseNet201, EfficientNetV2L, VGG16, VGG19, NASNetLarge, MobileNetV2, and ResNet50. All models were modified by removing the original classification head and adding new final layers with a Global Average Pooling + Dropout (rate=0.5) + Dense (256 units, ReLU activation) + Dropout (rate=0.3) + Dense (2 units) + Softmax structure. Fine-tuning was performed on the last 30% of layers while keeping earlier layers frozen for the first 20 epochs, then unfreezing all layers for continued training.

2.7 Model training and validation strategy

A stratified splitting approach ensured balanced representation of both classes across datasets. The training-testing ratio used for Faster R-CNN was 80% for training and 20% for testing, while the classification models received 70% for training and 15% for both validation and testing. The model stability and generalizability was assessed by performing 5-fold stratified cross-validation. Each fold had the same class distribution as the full dataset. The performance metrics were averaged across all folds with standard deviations reported.

2.8 Statistical analysis

Python programming language version 3.9 with scikit-learn 1.0.2, TensorFlow 2.8.0, and statsmodels 0.13.2 libraries served for statistical analysis to calculate model performance through accuracy, precision, recall (sensitivity), specificity, and F1-score values. The evaluation metrics included area under the receiver operating characteristic curve (AUC-ROC), area under the precision-recall curve (AUC-PR), and Matthews correlation coefficient (MCC).

The chi-square test evaluated categorical data while the Student’s t-test examined continuous variables for normally distributed data, with Mann-Whitney U test used for non-parametric data. A P-value of <0.05 established statistical significance in all comparative analyses, with Bonferroni correction applied for multiple comparisons. Cohen’s kappa coefficient assessed radiologist-artificial intelligence model agreement. The analysis excluded images with missing labels, and the researchers checked data consistency through automated scripts and manual verification. The study evaluated T2 and ADC-DWI sequence performance separately while comparing outcomes between normal and enhanced data sets. The confusion matrix enabled model classification success visualization, and ROC analysis found the optimal threshold values using Youden’s index. Bootstrap resampling (n=1000) was used to calculate 95% confidence intervals for all performance metrics.

3 Results

3.1 Patient characteristics and dataset description

In our study, a total of 153 patients were analyzed, and 27 of the 180 patients initially evaluated were excluded from the study due to various exclusion criteria. The patients’ biparametric MR images were obtained using a 3.0 Tesla MR device, and T2-weighted sequences and diffusion-weighted imaging techniques were applied. A total of 2,742 images from T2 sequences and 2,790 images from ADC-DWI sequences were used for the Faster R-CNN model. Two different datasets were prepared for training the classification models, one with normal images and one with augmented data, with cancer and prostatitis-diagnosed images grouped separately (Table 1).

Table 1
www.frontiersin.org

Table 1. Patient demographics and dataset characteristics.

3.2 Faster R-CNN model performance

The performance evaluation of the Faster R-CNN model resulted in 96% accuracy (95% CI: 93.2-98.8%), 96% precision, and 98% F1-score for P5 lesions in T2-weighted images. In the prostatitis category, which includes prostatitis lesions, the model achieved 99% accuracy (95% CI: 96.7-100%) and 91% F1 score. The model failed to detect P4 lesions in T2 sequences because it misclassified all P4 cases as prostatitis or P5 (sensitivity = 0%). In ADC-DWI sequences, the model achieved 90% accuracy (95% CI: 85.4-94.6%) and 94% F1 score for P5 lesions, and 97% accuracy (95% CI: 93.8-100%) and 88% F1 score for prostatitis lesions. In the P4 category, low performance was again observed, with only 30% sensitivity and 43% F1 score values calculated (p<0.001 compared to prostatitis and P5 categories) (Table 2).

Table 2
www.frontiersin.org

Table 2. Faster R-CNN model performance metrics with statistical significance.

3.3 Comparison with radiologist performance

The artificial intelligence model achieved high agreement rates when compared to an experienced radiologist (Cohen’s κ = 0.86, 95% CI: 0.79-0.93, p<0.001). The artificial intelligence model correctly identified zonal anatomical localization in T2-weighted images at a 98.1% rate except for one image that misclassified a peripheral zone lesion as a transitional zone. The PI-RADS scoring results showed high agreement between the radiologist and artificial intelligence model with rates between 92% and 96%. The artificial intelligence model achieved 100% accuracy in cancer case classification which exceeded the radiologist’s 97.9% accuracy rate, though this difference was not statistically significant (p=0.314, McNemar’s test). The prostatitis lesions in ADC-DWI sequences received 100% agreement, but no significant statistical difference emerged between the two evaluators in the overall assessment (Table 3).

Table 3
www.frontiersin.org

Table 3. Comparison of radiologist versus artificial intelligence with statistical analysis.

PIRADS 5 peripheral zone lesion and AI-compatible classification example shown in Figure 3. The radiologist and AI model agreed that this lesion should be classified as PIRADS 5. Histopathology confirmed the diagnosis as prostate adenocarcinoma. As another example, a bilateral peripheral zone prostatitis lesion and the corresponding AI-consistent prostatitis diagnosis are presented in Figure 4. Both the radiologist and the AI model concurred in the classification of the lesion as prostatitis. Histopathology confirmed the diagnosis as chronic prostatitis. As another example, AI-based probability estimations for a patient with a PIRADS 5 lesion are presented in Figure 5. Both the radiologist and the AI model assigned a PIRADS 5 score based on the T2WI and DWI characteristics. The histopathological diagnosis confirmed the presence of prostate adenocarcinoma.

Figure 3
Three medical images labeled A, B, and C display MRI scans with an orange arrow indicating specific areas. Image A shows a cross-sectional view of pelvic anatomy. Image B highlights a bright spot on a grayscale image. Image C reveals another highlighted area within a darker, blurry scan.

Figure 3. PIRADS 5 peripheral zone lesion and AI-compatible classification example. (A) A right peripheral zone contains an abnormal hypointense lesion which appears in an axial T2-weighted image (arrow). (B) Hyperintense signal with restriction on DWI b1400. (C) The ADC map shows a hypointense area with a low ADC value (0.56 × 10−3 mm2/s) (arrows). The radiologist and AI model agreed that this lesion should be classified as PIRADS 5. Histopathology: prostate adenocarcinoma.

Figure 4
Three medical images labeled A, B, and C, displaying MRI scans. Image A shows a cross-sectional view with clear anatomical structures. Images B and C are closer views of specific areas, with bright spots indicated by orange arrows. Each image highlights different details of the scan.

Figure 4. The Prostatitis bilateral peripheral zone lesion and AI-compatible prostatitis diagnosis. (A) The bilateral peripheral zone displays multiple hypointense lesions in axial T2-weighted imaging (arrows). (B) The DWI b1400 images show minimal diffusion restriction. (C) The ADC map shows high ADC values (0.96 × 10−3 mm2/s) in the areas marked by arrows. The radiologist and artificial intelligence model both agreed on prostatitis classification. Histopathology: chronic prostatitis.

Figure 5
MRI scans of the pelvic region in four panels labeled A to D. Panel A shows a cross-section with an arrow pointing to the right. Panel B highlights a bright area on the right with an arrow. Panel C displays a white area in the center with an arrow. Panel D indicates three regions with percentages: tz 98%, pz 82%, and cancer 94%.

Figure 5. AI-based probability estimates in a patient with a PIRADS 5 lesion. (A) Axial T2-weighted image shows an irregularly marginated hypointense lesion larger than 15 mm located in the left peripheral zone at the prostatic apex (arrow). (B) The lesion appears hyperintense on high b-value diffusion-weighted imaging (b1400) (arrow). (C) The ADC map reveals a hypointense area with a low mean ADC value of 0.68 × 10−3 mm2/s, consistent with restricted diffusion. (D) Both the radiologist and the AI model assigned a PIRADS 5 score based on T2WI and DWI features. Histopathological diagnosis confirmed prostate adenocarcinoma.

3.4 Confusion matrix analysis

Confusion matrix analyses have revealed the classification performance of the model in detail. In T2-weighted images, the model correctly classified 64 of the 72 test images, successfully identifying 5 prostatitis lesions. However, no correct predictions were made in the P4 class. In ADC-DWI images, out of 114 test images, 89 P5 lesions, 3 P4 lesions, and 11 prostatitis lesions were correctly classified (overall accuracy: 90.4%, 95% CI: 84.7-96.1%) (Figure 6).

Figure 6
Two confusion matrices labeled A and B show predicted versus actual classifications for prostatitis categories. Matrix A has high correct predictions for class p5. Matrix B demonstrates similar trends, with slight variations in misclassification numbers. Color intensity reflects prediction accuracy.

Figure 6. Faster R-CNN model confusion matrices. (A) T2-weighted images: The model correctly identified 64/67 P5 lesions together with 5/6 prostatitis lesions yet it failed to detect any P4 lesions (0/2). (B) ADC-DWI images: The model reached 89/90 P5 accuracy and 3/10 P4 accuracy and 11/14 prostatitis accuracy. The actual case numbers are represented by numbers while correct predictions appear on the diagonal.

3.5 Classification model performance and overfitting analysis

The evaluation of nine classification models showed that data augmentation led to better performance results for all models. We conducted extra validation through cross-validation and learning curve analysis to address potential overfitting issues. The InceptionV3, Xception, DenseNet201, and EfficientNetV2L models achieved 100% values in all metrics after data augmentation in T2 sequences. Although the cross-validation metrics for the other models did not reach perfect performance, they still demonstrated improved results following the data augmentation process. The lowest performance in the normal data set was recorded by the ResNet50 model with 82.5% accuracy. In the ADC-DWI sequences, the DenseNet201 model showed the highest performance with 99.1% accuracy even in the normal data set and achieved excellent results after data augmentation. The learning curves showed that the model converged well without any significant difference between training and validation performance after epoch 60 (Table 4). In addition to these metrics, AUROC values were calculated to further assess the discriminative capability of the classification models. For the T2 sequence, AUROC scores were as follows: DenseNet201 (0.98), EfficientNetV2L (0.99), InceptionV3 (0.99), MobileNetV2 (0.92), NASNetLarge (0.83), ResNet50 (0.76), VGG16 (0.98), VGG19 (0.97), and Xception (0.96). For the ADC-DWI sequence, AUROC values were DenseNet201 (0.99), EfficientNetV2L (0.96), InceptionV3 (0.99), MobileNetV2 (0.82), NASNetLarge (0.90), ResNet50 (0.64), VGG16 (0.96), VGG19 (0.86), and Xception (0.97), further confirming the superior discrimination performance of DenseNet201 and InceptionV3 across both modalities.

Table 4
www.frontiersin.org

Table 4. Classification model performance with cross-validation results.

4 Discussion

In this study, the diagnostic performance of Faster R-CNN and nine different deep learning classification models was evaluated using biparametric MR images of 153 patients with histopathologically confirmed PCa or prostatitis. In our study, the Faster R-CNN model achieved 96% accuracy and 98% F1-score for P5 lesions in T2 images, and 90% accuracy and 94% F1-score in ADC-DWI sequences. Among the classification models, InceptionV3, Xception, DenseNet201, and EfficientNetV2L demonstrated excellent performance after data augmentation, achieving 100% values in all metrics. No significant difference was found between the experienced radiologist and the artificial intelligence model in terms of PI-RADS scoring, and both approaches showed high diagnostic success.

4.1 Clinical context and diagnostic challenges

Prostate cancer can remain silent for a long time or behave very aggressively. The process of determining tumour behavior stands equally important to the process of making a diagnosis. Cancer seen in the prostate gland can also be very similar to other benign prostate diseases (prostatitis, etc.). For this reason, the diagnosis stage is very critical as it changes the treatment plan. In the PI-RADS v2.1 assessment, a probability estimate on a 5-point scale is made for each lesion in the prostate gland based on the combination of T2A, DWI, and DCE (Dynamic Contrast-Enhanced) MRI findings (25). The diagnostic precision advancement capabilities of artificial intelligence systems represent an essential aspect in this particular scenario.

4.2 Technical considerations and imaging protocols

High-resolution T2A imaging serves as a vital sequence for evaluating anatomy and detecting tumors while determining tumor localization and local staging. The diagnostic precision of T2A images stands at a relatively low level on its own. The T2A images present low specificity because hemorrhage alongside chronic prostatitis and scarring and atrophy and radiotherapy and hormone therapy generate hypointense signals that make this method less specific. T2A images require functional sequence integration for better PCa diagnosis because these images exhibit high sensitivity yet low specificity (26). Our research employed a biparametric MR approach because of this. A.E.S Bosaily and colleagues performed an observational study showing that dynamic contrast-enhanced imaging provided no benefit to prostate cancer diagnosis accuracy over T2 and diffusion sequences, so there was no need for prostate MR contrast injection (27).

4.3 Comparison with existing literature

Our research findings need to be placed in the context of the wider field of AI applications in medical imaging. Other fields in the literature have utilized Faster R-CNN in numerous studies. The system shows high precision in detecting lung nodules and shows promise for clinical application in diagnosing lung diseases (28). Faster R-CNN reaches 100% accuracy in meningioma tumour classification and 87.5% accuracy in glioma classification while achieving a 94.6% confidence rate in meningioma tumour segmentation compared to MRI (29). The sensitivity of Faster R-CNN reaches 95.64% when detecting lung nodules at a rate of 1.72 false positives per scan (30). Pre-processing of colonoscopy images alongside Faster R-CNN detection enables significant improvement in colon polyp detection rates with an average sensitivity of 91.43% (31).

Specifically in prostate cancer detection, our performance metrics align with or exceed several recent studies. Aldoj et al. (2020) reported that, for the detection of clinically significant prostate cancer defined by biopsy results, their 3D convolutional neural network (CNN) model, trained on multiparametric MRI data from 200 patients, achieved a receiver operating characteristic area under the curve (AUC) ranging from 0.89 (with 88.6% sensitivity and 90.0% specificity) to 0.91 (with 81.2% sensitivity and 90.5% specificity), with an average AUC of 0.897 for the ADC, DWI, and K-trans input combination. Other sequence combinations yielded lower performance. The classification performance was comparable to that of experienced radiologists using PI-RADS, and lesion size or largest diameter did not affect the model’s performance (32). The study by Schelb et al. (2019) showed that a deep learning model (U-Net), trained on T2-weighted and diffusion MRI, had sensitivity and specificity values similar to clinical PI-RADS assessment for the detection of clinically significant prostate cancer. In the external test set, U-Net achieved sensitivities of 96% and 92% at probability thresholds of ≥0.22 and ≥0.33, with corresponding specificities of 31% and 47%, respectively. These results were not statistically different from those obtained by clinical PI-RADS assessment, highlighting that the deep learning approach performed similarly to expert human readers in this context (33).

The DWI and ADC values play essential roles in both lesion characterization and tumour aggressiveness assessment. Prostate cancer replaces tubules by destroying glandular tissue. The ADC values in tumour tissue remain lower than those in the healthy peripheral zones because tumour tissue contains denser cells. The meta-analysis conducted by Lian-Ming Wu and colleagues demonstrated that T2A combined with DWI produced better results than T2WI alone in prostate carcinoma detection with sensitivity at 0.76 and specificity at 0.82 (34). The accuracy range of our study (90-96%) exceeds the reported values in the literature, although direct comparison is limited by differences in patient populations and reference standards.

The research conducted by Mukesh Soni and colleagues demonstrated that SEMRCNN outperformed V-Net and ResNet50-U-Net for prostate cancer detection from MP-MRI with a Dice (similarity) coefficient of 0.654 and a sensitivity of 0.695 (35). The Faster R-CNN model demonstrated superior sensitivity at 0.99 for P5 lesions in ADC-DWI images but this comparison requires careful interpretation because of our limited sample size and different evaluation metrics. Hamide Nematollahi and colleagues conducted a systematic review to determine how different artificial intelligence systems performed in cancer detection tasks. The authors demonstrated that supervised machine learning techniques including deep learning and random forest and logistic regression proved effective for prostate cancer detection through multiparametric MRI (36).

Artificial intelligence models have brought significant progress to prostate cancer detection and segmentation procedures. The non-local Mask R-CNN model achieved 89.5% detection and segmentation accuracy for intraprostatic lesions during histopathological training (37). A. Saha and colleagues demonstrated that a 3D CAD system achieved 83.69% detection sensitivity and AUROC of 0.882 in biparametric MR imaging (38). The AUROC equivalent of approximately 0.94 based on F1-scores indicates competitive performance but we recognize the requirement for direct AUROC calculation in future research. A study conducted in 2024 worked with 400 patients and 11,000 images. The combination of Faster R-CNN and ResNet models produced a modified ResNet which reached 97.4% accuracy in detecting prostate cancer from MR images (39). The accuracy rates between our model (96% for P5 in T2) and their model are similar but their larger dataset provides better evidence for generalizability.

4.4 Critical analysis of model performance

Our assessment shows both positive aspects and major weaknesses in our results. Our study results match the outcomes found in previous research conducted in similar contexts. The model shows significant issues in P4 lesion classification because it fails to detect any P4 lesions in T2 sequences and only detects 30% of P4 lesions in ADC-DWI sequences. This represents a major clinical limitation, as P4 lesions carry a 50-70% probability of clinically significant cancer according to PI-RADS v2.1 guidelines (25). The main cause of this outcome stems from both the limited P4 case number (only 2 cases in T2 test set and 10 in ADC-DWI) and the challenging process of measuring these lesions, which requires size-based distinction from P5 lesions. Therefore, it was deemed more appropriate to evaluate P4 and P5 lesions together under the ‘cancer’ label in classification models. When the test results were evaluated, it was observed that the Faster R-CNN model was more effective in analyzing ADC-DWI images and made more predictions based on these images.

4.5 Clinical integration considerations

The integration of AI systems into clinical workflows requires careful consideration of multiple factors. Our proposed system would function as a decision support tool rather than a replacement for radiologist interpretation. The workflow would involve: (1) initial AI analysis providing probability scores and anatomical localization, (2) radiologist review with AI assistance highlighting areas of concern, and (3) final interpretation combining AI outputs with clinical context and radiologist expertise. The implementation of this approach faces several challenges because it requires high-quality detailed spatial annotations of training data and methods to handle intra-lesion heterogeneity and weak labeling limitations and sufficient computational resources for full-resolution radiology image analysis (40).

4.6 Study limitations

The research contains multiple essential restrictions. The study’s 153 patient sample remains small compared to current multicenter research which reduces the ability to apply our results to other populations. Due to the small number of P4 samples, the model was unable to effectively identify this class, which represents a significant clinical gap as these intermediate-risk lesions require careful evaluation. The model’s performance on an independent dataset remains unverified which creates doubts about its ability to function across different institutions and imaging protocols. The 100% accuracy in some augmented datasets raises concerns about overfitting which needs further investigation because cross-validation results showed 96.8% accuracy. The study did not conduct a prospective clinical trial to evaluate the model’s performance in real-world settings. The study did not evaluate inter-scanner variability because all imaging data was collected from a single 3.0T scanner. The study failed to evaluate how variations in image quality and motion artifacts affect the model’s performance.

4.7 Study strengths and contributions

Nevertheless, our study has important strengths. Using the Faster R-CNN architecture, a comprehensive analysis was performed in which zonal anatomical structures were evaluated separately in both DWI-ADC and T2 sequences. The results were directly compared with the evaluations of an expert radiologist with 15 years of experience using nine different deep learning classification models. The sequences that did not need contrast agent achieved high performance levels which provided practical benefits such as shorter scan times and gadolinium risk elimination and cost-effectiveness. Our systematic comparison of multiple architectures provides insights into the relative performance of different deep learning approaches for this specific clinical task.

4.8 Future directions

P5 lesions differ from P4 lesions in that they tend to exhibit more locally aggressive behavior. Consequently, the development of models capable of accurately detecting P4 lesions remains a key priority for future research, as such advancements would facilitate the creation of systems that yield more consistent and clinically meaningful predictions. Moreover, the generalizability of current findings requires multi-institutional validation studies to confirm performance across diverse populations, imaging equipment, and imaging protocols.

The clinical utility and impact on patient outcomes of AI-assisted interpretation workflows should be established through prospective clinical trials comparing these workflows with standard diagnostic approaches. Additionally, the clinical value of such systems can be further enhanced by integrating them into electronic health records and clinical decision support platforms. Finally, diagnostic performance may be improved by exploring the synergistic use of radiomic features and deep learning–based methods (41).

5 Conclusion

Artificial intelligence models have demonstrated promising diagnostic capabilities in distinguishing prostate cancer from prostatitis using biparametric MR images and may provide valuable support to radiologists as a supplementary tool. The Faster R-CNN model achieved competitive performance compared to similar models in the literature, with the classification models InceptionV3, DenseNet201, and EfficientNetV2L showing excellent performance in our limited dataset. The results show that the experienced radiologist and the artificial intelligence model have similar performance in detecting P5 and prostatitis lesions, which suggests potential clinical utility, but the failure to detect P4 lesions is a critical limitation that must be addressed before clinical deployment.

These findings suggest that artificial intelligence models have potential for integration as clinical decision support systems in prostate cancer and prostatitis diagnosis, but substantial validation work remains necessary. The current evidence supports continued research and development rather than immediate clinical implementation. Future work should focus on addressing the identified limitations, particularly P4 lesion detection and external validation, before these systems can be considered ready for routine clinical use. The path forward requires collaborative efforts between AI researchers, radiologists, and clinicians to ensure that these technologies truly enhance patient care while maintaining diagnostic safety and reliability.

Representative case examples

Case 1: PI-RADS 5 Lesion (Figure 3). The right peripheral zone lesion of the 66-year-old patient received a PI-RADS 5 score from both the radiologist and artificial intelligence system. The T2-weighted images revealed an irregular hypointense lesion with significant DWI b1400 restriction and an ADC map diffusion coefficient of 0.56 × 10−3 mm2/s. The histopathological examination confirmed prostate adenocarcinoma as the diagnosis (Gleason score 4 + 3 = 7).

Case 2: Prostatitis Lesion (Figure 4). The 49-year-old patient received a PI-RADS 3 score from both evaluators for bilateral peripheral zone patchy lesions. The lesions displayed patchy hypointense T2-weighted imaging characteristics and mild DWI restriction and an ADC map diffusion coefficient of 0.96 × 10−3 mm2/s. The histopathological result was reported as chronic prostatitis.

Case 3: AI Probability Estimation (Figure 5). This case demonstrates the AI model’s capability to provide probability estimates for different tissue types. The model correctly identified the cancer with 94% confidence, while also detecting the transitional zone (98%) and peripheral zone (82%) anatomical structures.

Data availability statement

The original contributions presented in the study are included in the article/supplementary material. Further inquiries can be directed to the corresponding authors.

Ethics statement

The study involving humans was approved by RECEP TAYYIP ERDOGAN UNIVERSITY EDUCATION AND RESEARCH HOSPITAL and was conducted with institutional ethical committee approval (approval number: 2025/44; approval date: 30 January 2025). Personal data were anonymised and coded for analysis to protect patient privacy. Data security was ensured on encrypted computers and in databases accessible only to the research team. The study was conducted in accordance with the principles of the Helsinki Declaration. Patient written consent for publication was obtained.

Author contributions

MK: Writing – original draft, Visualization, Validation, Formal Analysis, Writing – review & editing, Project administration. OD: Writing – review & editing, Software. MS: Resources, Writing – review & editing. AC: Conceptualization, Writing – review & editing. GB: Writing – review & editing.

Funding

The author(s) declared that financial support was not received for this work and/or its publication.

Conflict of interest

The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declared that generative AI was used in the creation of this manuscript. Generative artificial intelligence was used.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

1. Siegel DA, O'Neil ME, Richards TB, Dowling NF, and Weir HK. Prostate cancer incidence and survival, by stage and race/ethnicity — United states, 2001–2017. Morbidity Mortality Weekly Rep. (2020) 69:1473–1480. doi: 10.15585/mmwr.mm6941a1

PubMed Abstract | Crossref Full Text | Google Scholar

2. Siegel C. Re: molecular imaging of prostate cancer. J Urol. (2016) 196:415. doi: 10.1016/j.juro.2016.05.026

PubMed Abstract | Crossref Full Text | Google Scholar

3. Men S, Cakar B, Conkbayir I, and Hekimoglu B. Detection of prostatic carcinoma: the role of TRUS, TRUS guided biopsy, digital rectal examination, PSA and PSA density. J Exp Clin Cancer Research: CR. (2001) 20:473–80. PMID: 11876539

Google Scholar

4. Krishna S. V, Rani SR, and Kiran PU. Comparing the effectiveness of digital rectal examination, prostate specific antigen, and TRUS guided prostate biopsy, in the detection of prostate cancer. Int J Clin Biochem Res. (2025) 11:248–253. doi: 10.18231/j.ijcbr.2024.037

Crossref Full Text | Google Scholar

5. Enzmann T and Sondermann M. Der transrektale ultraschall (TRUS) in der urologischen praxis. J für Urologie und Urogynäkologie/Österreich. (2021) 28:16–20. doi: 10.1007/s41972-021-00128-8

Crossref Full Text | Google Scholar

6. Scattoni V, Roscigno M, Raber M, Montorsi F, Da Pozzo L, Guazzoni G, et al. Multiple vesico-urethral biopsies following radical prostatectomy: the predictive rroles of TRUS, DRE, PSA and the pathological stage. Eur Urol. (2003) 44:407–14. doi: 10.1016/S0302-2838(03)00320-8

PubMed Abstract | Crossref Full Text | Google Scholar

7. Bansal R. High levels of the PSA (Prostate specific antigen) in patients with benign prostatic hyperplasia and prostate cancer. Arch Cancer Res. (2021) 8:14. doi: 10.36648/2254-6081.8.4.14

Crossref Full Text | Google Scholar

8. Van Iersel MP, Witjes WP, Debruyne FM, Karthaus HF, and De la Rosette JJ. Prostate-specific antigen density: correlation with histological diagnosis of prostate cancer, benign prostatic hyperplasia and prostatitis. Br J Urol. (1995) 76:47–53. doi: 10.1111/J.1464-410X.1995.TB07830.X

PubMed Abstract | Crossref Full Text | Google Scholar

9. Chou A, Darst BF, Wilkens L. R, Le Marchand L, Lilja H, Conti DV, et al. Association of prostate-specific antigen levels with prostate cancer risk in a multiethnic population: stability over time and comparison with polygenic risk score. Cancer Epidemiology Biomarkers Prevention: A Publ Am Assoc Cancer Research Cosponsored by Am Soc Prev Oncol. (2022) 31:2199–2207. doi: 10.1158/1055-9965.EPI-22-0443

PubMed Abstract | Crossref Full Text | Google Scholar

10. Jehle DVK, Nguyen N, Garza MA, Kim DK, Paul KK, Bilby NJ, et al. PSA levels and mortality in prostate cancer patients. Clin Genitourin Cancer. (2024) 22:102162. doi: 10.1016/j.clgc.2024.102162

PubMed Abstract | Crossref Full Text | Google Scholar

11. Feleppa EJ, Kalisz A, Sokil-Melgar JB, Lizzi F. L, Liu T, Rosado A. L, et al. Typing of prostate tissue by ultrasonic spectrum analysis. IEEE Trans Ultrasonics Ferroelectrics Frequency Control. (1996) 43:609–19. doi: 10.1109/58.503779

Crossref Full Text | Google Scholar

12. Fascelli M, Sussman R, Frye TP, George AK, Abboud S, Chelluri R, et al. How reliable is a negative MRI/TRUS fusion biopsy? The predictive value of targeted biopsy for prostate cancer. J Clin Oncol. (2015) 33:51–1. doi: 10.1200/JCO.2015.33.7_SUPPL.51

Crossref Full Text | Google Scholar

13. Itatani R, Namimoto T, Atsuji S, Katahira K, Morishita S, Kitani K, et al. Negative predictive value of multiparametric MRI for prostate cancer detection: outcome of 5-year follow-up in men with negative findings on initial MRI studies. Eur J Radiol. (2014) 83:1740–5. doi: 10.1016/j.ejrad.2014.06.026

PubMed Abstract | Crossref Full Text | Google Scholar

14. Yuan Q, Costa DN, Senegas J, Xi Y, Wiethoff AJ, Rofsky NM, et al. Quantitative diffusion-weighted imaging and dynamic contrast-enhanced characterization of the index lesion with multiparametric MRI in prostate cancer patients. J Magnetic Resonance Imaging. (2017) 45:908–916. doi: 10.1002/jmri.25391

PubMed Abstract | Crossref Full Text | Google Scholar

15. Shen Y and Wu S. Value of dynamic contrast-enhanced MRI and DWI in diagnosis of prostate cancer and its pathological comparison. J Men's Health. (2023) 19:119–126. doi: 10.22514/jomh.2023.138

Crossref Full Text | Google Scholar

16. De Visschere P, Lumen N, Ost P, Decaestecker K, Pattyn E, Villeirs G, et al. Dynamic contrast-enhanced imaging has limited added value over T2-weighted imaging and diffusion-weighted imaging when using PI-radsv2 for diagnosis of clinically significant prostate cancer in patients with elevated PSA. Clin Radiol. (2017) 72:23–32. doi: 10.1016/j.crad.2016.09.011

PubMed Abstract | Crossref Full Text | Google Scholar

17. Sackett J, Shih JH, Reese SE, Brender JR, Harmon SA, Barrett T, et al. Quality of prostate MRI: is the PI-RADS standard sufficient? Acad Radiol. (2020) 28:199–207. doi: 10.1016/j.acra.2020.01.031

PubMed Abstract | Crossref Full Text | Google Scholar

18. Pickersgill NA, Vetter JM, Andriole GL, Shetty AS, Fowler KJ, Mintz AJ, et al. Accuracy and variability of prostate multiparametric magnetic resonance imaging interpretation using the prostate imaging reporting and data system: A blinded comparison of radiologists. Eur Urol Focus. (2020) 6:267–272. doi: 10.1016/j.euf.2018.10.008

PubMed Abstract | Crossref Full Text | Google Scholar

19. Park YW, Kang KA, Kim CK, and Park SY. Prostate imaging-reporting and data system version 2 has improved biopsy tumor grade accuracy: A single, tertiary institutional experience. Abdominal Radiol. (2023) 48:2370–2378. doi: 10.1007/s00261-023-03917-x

PubMed Abstract | Crossref Full Text | Google Scholar

20. Li W, Hu R, Zhang Q, Yu Z, Deng L, Zhu X, et al. Artificial intelligence in prostate cancer. Chin Med J. (2025) 138:1769–82. doi: 10.1097/CM9.0000000000003689

PubMed Abstract | Crossref Full Text | Google Scholar

21. Marletta S, Eccher A, Martelli FM, Santonicco N, Girolami I, Scarpa A, et al. Artificial intelligence-based algorithms for the diagnosis of prostate cancer: A systematic review. Am J Clin Pathol. (2024) 161:526–534. doi: 10.1093/ajcp/aqad182

PubMed Abstract | Crossref Full Text | Google Scholar

22. Rabaan AA, Bakhrebah MA, AlSaihati H, Alhumaid S, Alsubki RA, Turkistani SA, et al. Artificial intelligence for clinical diagnosis and treatment of prostate cancer. Cancers. (2022) 14:5595. doi: 10.3390/cancers14225595

PubMed Abstract | Crossref Full Text | Google Scholar

23. Liu G, Pan S, Zhao R, Zhou H, Chen J, Zhou X, et al. The added value of AI-based computer-aided diagnosis in classification of cancer at prostate MRI. Eur Radiol. (2023) 33:5118–30. doi: 10.1007/s00330-023-09433-2

PubMed Abstract | Crossref Full Text | Google Scholar

24. Goldenberg S, Nir G, and Salcudean SE. A new era: artificial intelligence and machine learning in prostate cancer. Nat Rev Urol. (2019) 16:391–403. doi: 10.1038/s41585-019-0193-3

PubMed Abstract | Crossref Full Text | Google Scholar

25. Turkbey B and Choyke PL. Prostate magnetic resonance imaging: lesion detection and local staging. Annu Rev Med. (2019) 70:451–9. doi: 10.1146/annurev-med-053117-123215

PubMed Abstract | Crossref Full Text | Google Scholar

26. Toivonen J, Perez I, Movahedi P, Merisaari H, Pesola M, Taimen P, et al. Radiomics and machine learning of multisequence multiparametric prostate MRI: towards improved non-invasive prostate cancer characterization. PloS One. (2019) 14:e0217702. doi: 10.1371/journal.pone.0217702

PubMed Abstract | Crossref Full Text | Google Scholar

27. Bosaily AE-S, Frangou E, Ahmed HU, Emberton M, Punwani S, Kaplan R, et al. Additional value of dynamic contrast-enhanced sequences in multiparametric prostate magnetic resonance imaging: data from the PROMIS study. Eur Urology. (2020) 78:503–11. doi: 10.1016/j.eururo.2020.03.002

PubMed Abstract | Crossref Full Text | Google Scholar

28. Su Y, Li D, and Chen X. Lung nodule detection based on faster R-CNN framework. Comput Methods Programs Biomedicine. (2021) 200:105866. doi: 10.1016/j.cmpb.2020.105866

PubMed Abstract | Crossref Full Text | Google Scholar

29. Kaldera HNTK, Gunasekara SR, and Dissanayake MB. (2019). Brain tumor classification and segmentation using faster R-CNN, in: 2019 Advances in Science and Engineering Technology International Conferences (ASET), IEEE:Piscataway, NJ, USA. pp. 1–6. doi: 10.1109/ICASET.2019.8714263

Crossref Full Text | Google Scholar

30. Nguyen CC, Tran GS, Nguyen VT, Burie JC, and Nghiem TP. Pulmonary nodule detection based on faster R-CNN with adaptive anchor box. IEEE Access. (2021) 9:154740–51. doi: 10.1109/ACCESS.2021.3128942

Crossref Full Text | Google Scholar

31. Qian Z, Lv Y, Lv D, Gu H, Wang K, Zhang W, et al. A new approach to polyp detection by pre-processing of images and enhanced faster R-CNN. IEEE Sensors J. (2020) 11:1–1. doi: 10.1109/JSEN.2020.3036005

Crossref Full Text | Google Scholar

32. Aldoj N, Lukas S, Dewey M, and Penzkofer T. Semi-automatic classification of prostate cancer on multi-parametric MR imaging using a multi-channel 3D convolutional neural network. Eur Radiol. (2020) 30:1243–53. doi: 10.1007/s00330-019-06417-z

PubMed Abstract | Crossref Full Text | Google Scholar

33. Schelb P, Kohl S, Radtke JP, Wiesenfarth M, Kickingereder P, Bickelhaupt S, et al. Classification of cancer at prostate MRI: deep learning versus clinical PI-RADS assessment. Radiology. (2019) 293:607–17. doi: 10.1148/radiol.2019190938

PubMed Abstract | Crossref Full Text | Google Scholar

34. Wu L. M, Xu J. R, Ye Y. Q, Lu Q, and Hu J. N. The clinical value of diffusion-weighted imaging in combination with T2-weighted imaging in diagnosing prostate carcinoma: A systematic review and meta-analysis. Am J Roentgenology. (2012) 199:103–10. doi: 10.2214/AJR.11.7634

PubMed Abstract | Crossref Full Text | Google Scholar

35. Soni M, Khan IR, Babu KS, Nasrullah S, Madduri A, Rahin S. A, et al. Light weighted healthcare CNN model to detect prostate cancer on multiparametric MRI. Comput Intell Neurosci. (2022) 2022:5497120. doi: 10.1155/2022/5497120

PubMed Abstract | Crossref Full Text | Google Scholar

36. Nematollahi H, Moslehi M, Aminolroayaei F, Maleki M, and Shahbazi-Gahrouei D. Diagnostic performance evaluation of multiparametric magnetic resonance imaging in the detection of prostate cancer with supervised machine learning methods. Diagnostics. (2023) 13:806. doi: 10.3390/diagnostics13040806

PubMed Abstract | Crossref Full Text | Google Scholar

37. Dai Z, Jambor I, Taimen P, Pantelic M, Elshaikh M, Dabaja A, et al. Prostate cancer detection and segmentation on MRI using non-local mask R-CNN with histopathological ground truth. Med Physics. (2023) 50:7748–63. doi: 10.1002/mp.16557

PubMed Abstract | Crossref Full Text | Google Scholar

38. Saha A, Hosseinzadeh M, and Huisman H. End-to-End Prostate Cancer Detection in bpMRI via 3D CNNs: Effects of Attention Mechanisms, Clinical Priori and Decoupled False Positive Reduction. Med Image Analysis. (2021) 73:102155. doi: 10.1016/j.media.2021.102155

PubMed Abstract | Crossref Full Text | Google Scholar

39. Talaat FM, El-Sappagh S, Alnowaiser K, and Hassan E. Improved prostate cancer diagnosis using a modified resNet50-based deep learning architecture. BMC Med Inf Decision Making. (2024) 24:119. doi: 10.1186/s12911-024-02419-0

PubMed Abstract | Crossref Full Text | Google Scholar

40. Harmon SA, Tuncer S, Sanford T, Choyke PL, and Türkbey B. Artificial intelligence at the intersection of pathology and radiology in prostate cancer. Diagn Interv Radiol. (2019) 25:183–8. doi: 10.5152/dir.2019.19125

PubMed Abstract | Crossref Full Text | Google Scholar

41. Castillo JM, Arif M, Starmans MPA, Niessen WJ, Bangma CH, Schoots IG, et al. Classification of clinically significant prostate cancer on multi-parametric MRI: A validation study comparing deep learning and radiomics. Cancers. (2022) 14:12. doi: 10.3390/cancers14010012

PubMed Abstract | Crossref Full Text | Google Scholar

Keywords: artificial intelligence, faster R-CNN, magnetic resonance imaging, prostatic neoplasms, prostatitis

Citation: Kaya M, Durdag O, Solak M, Coskuncay A and Burakgazi G (2026) Lesion detection using artificial intelligence models in MR images of prostate cancer and prostatitis patients and comparison of model performance. Front. Urol. 5:1726795. doi: 10.3389/fruro.2025.1726795

Received: 16 October 2025; Accepted: 18 December 2025; Revised: 05 December 2025;
Published: 19 January 2026.

Edited by:

Nasser Kashou, Kash Global Tech, United States

Reviewed by:

Zheng Yuan, China Academy of Chinese Medical Sciences, China
Xiao Jian Tan, Tunku Abdul Rahman University College, Malaysia

Copyright © 2026 Kaya, Durdag, Solak, Coskuncay and Burakgazi. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Muhammed Kaya, ZHJtdWhhbW1lZGtheWExOTAzQGdtYWlsLmNvbQ==; Gulen Burakgazi, YnVyYWtnYXppZ3VsZW5AZ21haWwuY29t

Present address: Muhammed Kaya, Department of Radiology, Kelkit State Hospital, Gümüşhane, Turkey

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.