ORIGINAL RESEARCH article
Front. Oncol.
Sec. Genitourinary Oncology
Robust Prostate Cancer Risk Stratification from Unregistered mpMRI via Learned Cross-Modal Correspondence
Provisionally accepted- 1Department of Radiology, The First Hospital of Jiaxing or The Affiliated Hospital of Jiaxing University, Jiaxing, China
- 2Perception Vision Medical Technologies Co Ltd, Guangzhou, China, Guangzhou, China
- 3Zhejiang Chinese Medical University, Hangzhou, China
Select one of your emails
You have multiple emails registered with Frontiers:
Notify me on publication
Please enter your email address:
If you already have an account, please login
You don't have a Frontiers account ? You can register here
Background and Objective: Accurate prostate cancer risk stratification benefits from the fusion of T2-weighted (T2WI) and Apparent Diffusion Coefficient (ADC) MRI. However, patient motion and imaging distortions frequently cause spatial misalignments between these sequences. While radiologists compensate for this via subjective cognitive fusion, the process introduces inter-reader variability and can be particularly challenging in equivocal cases. Conventional fusion models are even more vulnerable, as they require perfect image registration, making them brittle in real-world clinical scenarios. We aimed to develop and validate a deep learning framework that overcomes these limitations by robustly fusing unregistered mpMRI data. Methods: We retrospectively analyzed a cohort of 300 consecutive men (mean age, 71.5 ± 7.6 years) who underwent pre-biopsy prostate mpMRI at our institution between January 2021 and May 2023. All included patients had pathologically confirmed prostate cancer, with high-risk prostate cancer, as defined by NCCN guidelines, present in 184 of 300 cases (61.3%). The dataset was partitioned chronologically into a development cohort (n=250) for 5-fold cross-validation and a temporal test cohort (n=50) for independent evaluation. We developed Cross-Modal Optimal Transport Fusion (CMOT-Fusion), a deep learning framework that learns to identify and match diagnostically relevant regions between misaligned T2WI and ADC images. This approach enables robust multimodal fusion without requiring an explicit image registration step. Results: For discriminating NCCN high-risk versus low/intermediate-risk disease among pathologically confirmed prostate cancer cases, CMOT-Fusion achieved a mean Area Under the Curve (AUC) of 0.849 ± 0.034 in 5-fold cross-validation, outperforming single-modality baselines and conventional fusion methods. On an independent test set, the model's performance remained robust, with an ensemble AUC of 0.824 (95% CI: 0.694–0.930. Robust Risk Stratification from Unregistered mpMRI reference based on routine radiology suspicion scoring, PI-RADS achieved an AUC of 0.839 (95% CI: 0.726–0.930) on the same test cohort. Conclusion: Our results demonstrate that learning a direct correspondence between unregistered mpMRI sequences significantly improves prostate cancer risk stratification. The proposed CMOT-Fusion framework offers a robust solution to the common clinical problem of inter-sequence misalignment, potentially enhancing diagnostic reliability and streamlining clinical workflows by removing the need for a separate image registration step.
Keywords: deep learning, Feature fusion, image registration, Multiparametric MRI, Optimal transport, prostate cancer
Received: 06 Sep 2025; Accepted: 19 Dec 2025.
Copyright: © 2025 Gong, Mao, Gong, Lu, Ding and Zhu. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
* Correspondence: Xiang Zhu
Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.
