Your new experience awaits. Try the new design now and help us make it even better

SYSTEMATIC REVIEW article

Front. Oncol., 20 August 2025

Sec. Gastrointestinal Cancers: Hepato Pancreatic Biliary Cancers

Volume 15 - 2025 | https://doi.org/10.3389/fonc.2025.1597969

This article is part of the Research TopicDiagnosis and Management of Pancreatic CancerView all 3 articles

Diagnosis methods for pancreatic cancer with the technique of deep learning: a review and a meta-analysis

Yuanbo Bi&#x;Yuanbo Bi1†Dongrui Li&#x;Dongrui Li1†Ruochen PangRuochen Pang2Chengxv DuChengxv Du1Da LiDa Li1Xiaoyv ZhaoXiaoyv Zhao3Haitao Lv*Haitao Lv1*
  • 1Department of Hepatobiliary Surgery, The Second Hospital of Hebei Medical University, Shijiazhuang, Hebei, China
  • 2Department of Spine Surgery, The Third Hospital of Hebei Medical University, Shijiazhuang, Hebei, China
  • 3Department of Oncology, Hebei Medical University, Shijiazhuang, Hebei, China

Background: Early diagnosis can significantly improve survival rate of Pancreatic ductal adenocarcinoma (PDAC), but due to the insidious and non-specific early symptoms, most patients are not suitable for surgery when diagnosed. Traditional imaging techniques and an increasing number of non-imaging diagnostic methods have been used for the early diagnosis of pancreatic cancer (PC) through deep learning (DL).

Objective: This review summarizes diagnosis methods for pancreatic cancer with the technique of deep learning and looks forward to the future development directions of deep learning for early diagnosis of pancreatic cancer.

Methods: This study follows the PRISMA-ScR (Preferred Reporting Items for Systematic Reviews and Meta-Analyses Extension for Scoping Reviews) guidelines, retrieving studies on deep learning for early pancreatic cancer diagnosis from PubMed, Embase, Web of Science, IEEE, and Cochrane Library over the past 5 years. Inclusion criteria were studies involving PDAC patients, using deep learning algorithms for diagnosis evaluation, using histopathological results as the reference standard, and having sufficient data. Two reviewers independently screened and extracted data. Quality was assessed using QUADAS-2, with StataMP 17 for meta-analysis.

Results: In this study, 422 articles were retrieved, and 7 were finally included for meta-analysis. The analysis showed that the accuracy of deep learning in the early diagnosis of pancreatic cancer was 80%-98.9%, and the combined sensitivity, specificity and AUC were 0.92 (95% CI: 0.85-0.96), 0.92 (95% CI: 0.85-0.96), and 0.97 (95% CI: 0.95-0.98). The positive and negative likelihood ratio were 11.52 (95% CI, 6.15-21.55) and 0.09 (95% CI, 0.04-0.17). Endoscopic ultrasound (EUS) and Contrast-Enhanced Computed Tomography (CE-CT) were the main diagnostic methods. Non-imaging diagnostic methods such as deep learning urine markers, disease trajectory also performed good diagnostic potential.

Conclusions: Artificial intelligence (AI) technology holds promise for clinical guidance in pancreatic cancer risk prediction and diagnosis. Future research may focus on leveraging diverse data sources like genomics and biomarkers through deep learning; utilizing multi - center or international samples; tackling the challenge of early diagnosis for small pancreatic cancers; enhancing the explainability of AI models and multi-modal approaches.

Introduction

Background

Pancreatic Ductal Adenocarcinoma (PDAC) is a highly aggressive and lethal malignant tumor of the pancreas. Currently, it is the fourth leading cause of cancer death and is projected to become the second by 2030 (1, 2). The high mortality rate of PDAC is mainly due to the challenge of its diagnosis (3). Early diagnosis can significantly improve the 5-year survival rate (4). But due to the similar attenuation of early PDAC to healthy pancreas, approximately 40% of pancreatic tumors less than 2cm are missed on abdominal computed tomography (CT) (58). However, PDAC is asymptomatic until it progresses to an advanced stage. At the time of diagnosis, only about 20% of cases are suitable for surgical resection (9).

It is clear that if cancer can be detected early and appropriate surgical and systemic treatments provided, the survival rate of PC could be significantly improved (10). Imaging techniques play a crucial role in the diagnosis of PDAC. Current clinical imaging modalities include endoscopic ultrasound (EUS), CT, and positron emission tomography-computed tomography (PET/CT), each with its own advantages and disadvantages in clinical application (1113). In the traditional process of medical image analysis, experienced radiologists are required. With artificial intelligence technology, radiologists can be liberated from tedious and repetitive tasks to handle those that require more creativity. The gradual development of deep learning and radiomics provides a new perspective for the use of the above as the diagnosis and screening of early PC (14).

Deep learning (DL) has recently made progress on many problems, especially in medical imaging diagnosis, and in some cases has even surpassed human performance (13). Deep learning constitutes a specialized branch of machine learning that employs multi-layered (deep) neural networks to automatically extract hierarchical features from high-dimensional data, yielding breakthrough performance in tasks such as image recognition, speech processing, and natural language understanding. Machine learning (ML), in turn, is a sub-discipline of AI; it leverages statistical techniques to enable computer systems to learn patterns from data and make predictions or decisions without explicitly programming every rule (6, 7). AI is the broad discipline devoted to enabling machines to exhibit intelligent behavior. Its canonical formulation, first articulated by John McCarthy, defines AI as “the science and engineering of making intelligent machines, especially intelligent computer programs that perform tasks which, if carried out by humans, would require human intelligence” (15). Consequently, AI, ML, and DL exhibit a nested relationship: AI ⊃ ML ⊃ DL. In recent years, a large number of studies related to deep learning have been published, but some of them seem to blur the definitions of deep learning and machine learning. Deep learning is a subset of ML algorithms. The conceptual difference between the two is easy to overlook (16).

With the increase in research, various new diagnostic markers for PC diagnosis using deep learning have emerged one after another (17). Based on the aforementioned, this study aims to assess the clinical relevance and translational potential of deep learning algorithms in the early detection of pancreatic cancer. We systematically review applications across imaging modalities, urinary biomarkers, and disease-trajectory models, with particular emphasis on diagnostic methods, neural-network architectures, and downstream clinical utility. In addition, we performed a comprehensive meta-analysis of the diagnostic accuracy reported by the selected studies.

Methods

Overview

In this review, PRISMA-ScR (Preferred Reporting Items for Systematic Reviews and Meta-Analyses Extension for Scoping Reviews) guidelines were followed to ensure the transparency and reliability of this study.

Search strategy

Search sources

We searched five databases (PubMed, Embase, Web of science (WOS), IEEE, and Cochrane Library) for studies on DL for early diagnosis of PDAC published between December 31, 2019 and December 31, 2024 (Figure 1).

Figure 1
Bar and line chart titled “Publication” showing data from 2020 to 2024. Orange bars represent “systematic review,” gray bars indicate “meta-analysis,” and  a blue line depicts “all literature.” Both bars show increases in 2023 and 2024. The blue  line rises steadily across the years. Systematic review and meta-analysis

Figure 1. “Systematic review and meta-analysis“ based on the axis labels on the left, and “All literature” based on the axis labels on the right (classified by year).

Search terms

To develop the search strategy, 2 experts in the fields of artificial intelligence and hepatobiliary surgery were consulted, and previous relevant reviews were consulted to identify subject headings and free words. Search terms were selected according to the target intervention (i.e., deep learning), the target disease (i.e., PC), and the purpose of the intervention (i.e., early diagnosis). Details of the exact search terms used to search each database are provided in Supplementary File 1.

Study eligibility criteria

Following the PICOS principle, studies meeting the following inclusion criteria were admitted: (1) included patients with a diagnosis of PDAC; (2) develop or use clinical predictive models or deep learning algorithms to evaluate PC diagnosis instead of segmentation; (3) used histopathologic results as the reference standard (4) had sufficient data to reconstruct the 2 × 2 contingency table to be included.

The following studies were excluded: (1) reviews, case reports, letters, commentaries, errata, meta-analyses and studies published only as conference abstracts; (2) Studies were also excluded if they lacked internal or external validation; (3) The number of the same diagnostic method was less than 3; (4) The sample size was less than 50; (5) The sample was not on a per-patient basis.

Study selection

The eligibility of articles was determined by two reviewers (Yuanbo Bi and Dongrui Li), who independently screened the titles and abstracts of the search results. In case of disagreement, the third reviewer (Ruochen Pang) will intervene and decide together. To measure inter-reviewer agreement, we calculated Cohen kappa, which was 0.892 for title and abstract screening and 0.810 for full-text screening, with almost perfect agreement (13, 18).

Data extraction

To accurately extract data from the included studies from the selected studies, a data extraction table was created using Microsoft Excel, and the data extraction fields are described in Supplementary File 3 and filled in using the included studies. Two reviewers independently performed the process, and any disagreements between the two reviewers were resolved through discussion by the intervention of a third reviewer.

Data synthesis

To estimate the accuracy of deep learning algorithms, we conducted a meta-analysis of studies that provided sufficient data to construct contingency tables. StataMP 17 (64-bit) was used for forest plot, funnel plot, SROC curve, heterogeneity test and meta-analysis. Studies evaluating deep learning in robotic programs were not included in the analysis. Microsoft Excel was used to manage the comprehensive data. Endnote X9 was used for literature management.

Quality assessment

The methodological quality of eligible articles was determined by two authors (Yuanbo Bi and Dongrui Li) using the QUADAS-2. QUADAS-2 tool is recommended for use in systematic reviews of diagnostic accuracy based on sources of bias and variation and comprises four domains: patient selection, index test, reference standard, and flow and timing. Each domain is assessed in terms of risk of bias, and the first three domains are also assessed in terms of concerns regarding applicability. Each component contains several questions used to help judge the risk of bias (low, high or unclear) (13).

Results

Search results

We initially identified 422 articles using five publicly available online databases: PubMed (n=79), Embase (n=161), Web of science (n=127), IEEE (n=51), Cochrane Library (n=4) +Additional records identified through other sources (n=0) entered into PRISMA-ScR flow chart. We excluded 144 replicates. Among the remaining studies, 150 articles unrelated to the research topic were excluded according to the abstract, and 17 articles were included in the systematic review after the full-text evaluation of the remaining 128 articles. A total of 7 articles met our inclusion criteria and were included in the meta-analysis (Figure 2).

Figure 2
Flowchart illustrating a systematic review process. The identification phase lists 422 records from databases like PubMed, Embase, and others. After removing duplicates, 278 records remain. Screening excludes 150 records, leaving 128 full-text articles for eligibility. Of these, 111 articles are excluded for reasons like no deep learning or lack of data, resulting in 17 qualitative and 7 quantitative synthesis studies.

Figure 2. Preferred reporting items for systematic review and meta-analysis of diagnostic test accuracy studies (PRISMA) flow diagram for study selection.

Study selection and characteristics

The results of 17 studies (out of 278) that met the inclusion criteria are summarized in Supplementary File 2. These studies were published between 2019 and 2024. The techniques applied included CE-CT (23.53%), EUS (29.41%), with one study using Contrast-enhanced harmonic endoscopic ultrasound (CH-EUS), pathological examination (11.76%), disease trajectories (11.76%), urine biomarkers (11.76%), RNA (5.88%), and DNA (5.88%). Among them, 7 studies met both the inclusion and exclusion criteria and were included in the meta-analysis (Table 1).

Table 1
www.frontiersin.org

Table 1. Publications reporting on deep learning use in early diagnosis of pancreatic cancer.

Systemic review

In this section, we systematically review the 17 included studies according to their diagnostic modalities, critically appraise the advantages and limitations of each method, and synthesize the corresponding findings.

CE-CT

In recent years, research on the early diagnosis of PC using deep learning-based CE-CT has gained significant attention. CE-CT remains a primary imaging modality for detecting (PC). However, the complex anatomical structure of the pancreas and the generally low contrast of these images present substantial challenges for accurate segmentation of pancreatic CT images. Precise pancreatic segmentation is critical in clinical practice, particularly for diagnosing and treating PC (19). Notably, CE-CT exhibits low sensitivity for small tumors, with approximately 40% of tumors smaller than 2 cm being undetected (5, 20).

The accuracy of deep learning-based CE-CT for the early diagnosis of PC ranges from 80% to 98.90%, with sensitivity ranging from 0.79 to 0.973, specificity from 0.76 to 1, and AUC from 0.85 to 0.999. One study had a sample size of less than 100, resulting in statistical results that deviated significantly from those of the other three studies. Nevertheless, due to its minimal impact on the overall findings during subsequent subgroup analysis, this study was retained (21).

Chen et al. reported a sensitivity of 74.7% (95% CI: 64.5, 83.3) for malignant tumors smaller than 2 cm (5). Liu et al. provided sensitivities of 92.1%, 92.1%, and 63.1% for malignant tumors smaller than 2 cm in two internal validations and one external validation, respectively (20). The early diagnosis of PC smaller than 2 cm remains unclear, necessitating further research to confirm these findings, which may represent a promising new research direction.

EUS

Endoscopic ultrasound (EUS) is a non-invasive and highly precise technique for detecting pancreatic diseases and has become a widely used tool for diagnosing digestive system disorders (2224). EUS-guided fine-needle aspiration biopsy (EUS-FNA/B) is considered the first-line method for the pathological diagnosis of PC due to its high accuracy (25). Reports indicate that the diagnostic sensitivity of EUS-FNA/B for pancreatic ductal adenocarcinoma (PDAC) is 85%–92%, with specificity ranging from 96% to 98%. Furthermore, it has been proven to be a feasible and safe technique with a complication rate of less than 1% (26).

In this study, the number of studies utilizing deep learning-based EUS for the early diagnosis of PC was the largest (n = 5), with accuracy ranging from 86.18% to 94.17%, sensitivity from 0.831 to 0.94, specificity from 0.822 to 1, and AUC from 0.9221 to 0.96 (3, 4, 6, 25, 27).

Among these, a prospective study on the early diagnosis of PC using deep learning CH-EUS demonstrated that the accuracy of CH-EUS NASTER-guided EUS-FNA was 93.8%, compared to 91.3% in the control group. Although the diagnostic rate of the CH-EUS MASTER group appeared higher, no statistically significant difference was observed between the two groups (p > 0.05). The AUC values for the CH-EUS MASTER group and the control group were 0.955 and 0.933, respectively (25). CH-EUS MASTER may serve as a promising real-time objective system for differentiating benign and malignant pancreatic masses. However, due to the small sample size (n = 39) and partial data loss, it was excluded from the subsequent meta-analysis.

Based on interval estimation, deep learning-based EUS image data appears slightly superior to CE-CT in terms of accuracy, sensitivity, specificity, and AUC, demonstrating greater stability. Nevertheless, additional studies are required to confirm these findings.

PET/CT

Although PET/CT holds significant value for the early diagnosis and differentiation of tumors, as well as for determining metastasis, its high cost and prolonged exposure time limit its application in the early diagnosis and screening of PC at present. Most current studies on deep learning-based PET/CT focus on tumor segmentation (2830), rather than diagnosis. Consequently, no PET/CT studies were included in the meta-analysis.

Pathological examination

Deep learning of PC pathological tissues may have far-reaching significance for improving the accuracy of pancreatic cancer pathological diagnosis and targeted therapy (31). In routine pathological smears, stromal components account for 90% of the specimen and interact dynamically with the tumor, which poses a clinical challenge (32, 33). By leveraging deep learning and image analysis, information clues from stromal interactions can be extracted for the identification of new cancer biomarkers. This approach holds potential for enhancing diagnostic accuracy and providing deeper insights into the biology of pancreatic cancer (34).

Pancreatic cancer cells typically accumulate a large number of lipid droplets (LDs), which regulate lipid storage (35). To facilitate rapid diagnosis, Hong et al. proposes a deep convolutional neural network based automatic identification system for pancreatic cancer cells, which uses optical diffraction tomography to quantitatively image LDs of unstained cytology samples (35). The accuracy rate is 97.06% (± 1.021), and the AUC is 99.8%. This reflects the application prospects of deep learning in quantifying cell lipid droplet content in pancreatic cancer pathological examinations.

Fu et al. developed a deep learning model for the classification of pancreatic cancer pathological tissues. The automatic patch-level approach achieved a classification accuracy of 95.3%, and the whole-slide images-level (WSIs-level) approach achieved 100% (36). Fassler et al. quantitatively assessed T cells in the tumor microenvironment (TME) from multiplex immunohistochemistry (mIHC) WSIs of pancreatic ductal adenocarcinoma using the ColorAE: U-Net deep learning tool suite. They analyzed the spatial distribution of (CD3, CD4, CD8), B cells (CD20), macrophages (CD16), and tumor cells (K17) (33). This might promote the wider adoption of mIHC to support precision medicine, especially in the field of immunotherapy for pancreatic cancer.

Urine biomarkers

A study proposed the utilization of four urinary proteome biomarkers (creatinine, LYVE1, REG1B, and TFF1) to successfully develop a novel and highly efficient 1D CNN-LSTM model for the early diagnosis of PDAC. The proposed model is composed of one-dimensional convolutional neural networks (1D-CNNs) and long short-term memory (LSTM). Successful experiments and evaluations were conducted on 590 urine samples from the public dataset (183 healthy pancreatic samples, 208 samples with benign hepatobiliary diseases, and 199 PDAC samples). The study demonstrated that the 1-D CNN + LSTM model achieved an accuracy of 97% and an area under the curve (AUC) of 98% (37). This research shows a high accuracy and AUC, highlighting the great potential of urine markers for the early diagnosis of pancreatic cancer.

Linh, et al. introduce a label-free surface-enhanced Raman scattering sensor based on a three-dimensional plasmonic coral nanoarchitecture (3D-PCN), and successfully distinguished prostate cancer from pancreatic cancer by identifying urine with the aid of deep learning (38). However, there are few studies on the early diagnosis of pancreatic cancer using urine markers with deep learning, and more related research could be conducted to supplement this area. Acer et al. utilized non-invasive urine biomarkers and CA19-9, combined with seven machine learning models, to conduct early detection of PDAC in data from 590 participants, and found that the ensemble learning model performed best, with the GBC model achieving an accuracy of 92.99% (AUC = 0.9761) (39).

Disease trajectories

Deep learning models based on disease trajectories are used to integrate longitudinal clinical data from electronic medical records to infer the risk of PC, which may also be a breakthrough for the early diagnosis of pancreatic cancer (40).

Placido et al. predicted PC through deep learning of large sample disease trajectories, using (Electronic Health Records)EHR disease codes, and the best (Danish National Patient Registry) DNPR model’s performance was an area under the receiver operating characteristic (AUROC) curve of 0.88 (41). Park et al. used deep learning based on time series laboratory test results from EHR for the early detection of pancreatic cancer. The deep learning model demonstrated better performance on early detection (AUROC 0.671, CI 95% 0.667 – 0.675, p < 0.001) at 12 months prior to diagnosis compared to a logistic regression, xgboost, and a feedforward neural network baseline (42).

The predictive performance at the level demonstrated by deep-learning Disease trajectories may be useful for the initial design of real-world clinical predictive surveillance programs. This approach potentially provides a scalable workflow for community-level early cancer detection, shifting the focus from late-stage treatment to early-stage intervention.

RNA

Hong et al. proposed a novel deep learning framework, named HATZFS, for identifying driver biomarkers of pancreatic cancer using RNA differential expression data. This model integrates HDQN, GAT, and ZFS to construct a pancreatic cancer RNA transcriptional regulatory network involving lncRNA, miRNA, and mRNA relationships. To validate the effectiveness of this model, comprehensive experiments were conducted on a benchmark dataset containing 14 databases and compared with eight other algorithms (43). Overall, HATZFS not only provides the importance of all RNA molecules in the pancreatic cancer RNA regulatory network but also determines the set of driver nodes in this network as driver biomarkers.

DNA

Tumor type guides clinical treatment decisions for cancer. Histological diagnosis remains challenging, and genomic alterations have high diagnostic value for tumor types (36, 44).

Darmofal et al. developed Genome-Derived-Diagnosis Ensemble (GDD-ENS), a hyperparameter ensemble using deep neural networks to classify tumor types, based on genomic features from a dataset of 39,787 solid tumors sequenced with a clinical targeted cancer gene panel. GDD-ENS achieved an accuracy of 93% in high-confidence predictions for 38 cancer types (45).

Methodological quality

The methodological quality of the studies according to QUADAS-2 assessment is illustrated in Figure 3. Three studies showed low risk of bias and applicability issues in all domains, while others showed unclear risks in some domains. No study had high risk of bias or applicability issues.

Figure 3
Table showing the risk of bias and applicability concerns for several studies. Green circles with plus signs indicate low risk or concern, while yellow circles with question marks indicate unclear status. A red circle with a minus sign represents high risk. Each study is evaluated on patient selection, index test, reference standard, and flow and timing under both categories.

Figure 3. QUADAS-2 assessment of the studies.

Meta−analysis results

Seven studies provided enough data to construct 8 contingency tables, with a pooled sensitivity of 0.92 (95% CI: 0.85-0.96), specificity of 0.92 (95% CI: 0.85-0.96), and AUC of 0.97 (95% CI, 0.95.-0.98) (Table 2). The combined positive likelihood ratio and negative likelihood ratio were 11.52 (95% CI, 6.15-21.55) and 0.09 (95% CI, 0.04-0.17), respectively (Figures 46). The likelihood ratio scatter plot distribution is relatively scattered, and there may be large inter-study heterogeneity, which needs further analysis of the source of heterogeneity (Figure 7).

Table 2
www.frontiersin.org

Table 2. Summary table of the results of the meta-analysis.

Figure 4
Forest plots comparing sensitivity and specificity for seven studies. The left plot shows sensitivity values ranging from 0.79 to 0.99, with a combined sensitivity of 0.92. The right plot shows specificity values from 0.76 to 0.99, with a combined specificity of 0.92. The confidence intervals and combined estimates are indicated for each study. Statistical heterogeneity is noted with I-squared values of 87.53 for sensitivity and 84.74 for specificity.

Figure 4. Forest plot of sensitivity and specificity of deep learning (DL) in identifying pancreatic tumors.

Figure 5
SROC plot showing sensitivity versus specificity with prediction and confidence contours. Observed data points are marked as circles, and the summary operating point is represented by a red diamond with sensitivity and specificity both at 0.92. The SROC curve has an AUC of 0.97. Dashed lines indicate the 95% confidence contour, while dotted lines show the 95% prediction contour.

Figure 5. Summary receiver operating characteristic (SROC) curves for the diagnosis of pancreatic tumors using DL. Each circle indicates an individual study, red diamond represents summary sensitivity and specificity.

Figure 6
Forest plot displaying diagnostic likelihood ratios (DLR) with 95% confidence intervals for different studies. On the left, DLR positive values range from 2.2 to 611.7. On the right, DLR negative values range from 0.0 to 0.5. Individual study results are plotted as squares with horizontal lines for confidence intervals. The summary measure for combined studies is shown as a diamond. The combined DLR positive is 11.52, and DLR negative is 0.09. Statistical metrics, Q and I2, indicate heterogeneity.

Figure 6. Forest plot for likelihood ratio after combination (LR+, LR-).

Figure 7
Scatter plot titled “PLR vs NLR” showcasing data points labeled as 'PLR' across 'NLR' values. A red line indicates fitted values, showing a downward trend. The x-axis is labeled 0 to 0.25, and the y-axis is labeled 0 to 100.

Figure 7. Likelihood ratio scatter plot is relatively scattered, and there may be large inter-study heterogeneity. PLR is the positive likelihood ratio and NLR is the negative likelihood ratio.

Inter-study heterogeneity was quantified using the Mantel–Haenszel Q test and the I² statistic. Mantel-Haenszel Q test: The Q value is 43.72, the degrees of freedom are 7, and the p-value is 0.000 < 0.01, indicating significant heterogeneity among the studies. I² value: I² = 84.0% (95% CI 67.8 - 90.2%). Since the I² value is greater than 50%, it indicates that a significant proportion of the variation is caused by the heterogeneity among the studies.

Given that fewer than ten studies were available, we performed stratified analyses to identify potential sources of this heterogeneity. Studies were stratified by sample size (≤ 1–000 vs > 1 000), imaging modality (CE-CT vs EUS), deep-learning architecture (CNN vs non-CNN), and validation strategy (internal vs external).

In the CE-CT subgroup, heterogeneity remained pronounced (Q = 41.61, df = 4, p < 0.001; I² = 90.4%, 95% CI 79.9–94.2%). By contrast, the EUS subgroup exhibited negligible heterogeneity (Q = 1.53, df = 2, p = 0.464; I² = 0.0%, 95% CI 0.0–72.9%). The markedly lower heterogeneity in the EUS subset may reflect greater consistency in image acquisition, reader expertise, lesion size, or histopathological reference standards; however, the wide confidence interval (0.0–72.9%) indicates that imprecision due to the small number of studies cannot be excluded. None of the remaining stratifications materially reduced the observed heterogeneity (Supplementary File 4).

The Fagan chart analysis, which assessed a 50% predicted probability to simulate a clinical situation, resulted in a 92% posterior probability of a positive test result, compared with a negative likelihood ratio of 0.09 and a negative posterior probability of 8% (Figure 8).

Figure 8
Fagan nomogram showing the relationship between pre-test probability, likelihood ratio, and post-test probability. With a pre-test probability of 50%, a positive likelihood ratio of 12 leads to a post-test probability of 92%. A negative likelihood ratio of 0.09 results in a post-test probability of 8%.

Figure 8. Fagan nomogram of the accuracy of DL in the diagnosis of pancreatic tumors.

A Deeks funnel plot asymmetry test with a p value of 0.90 implies that the probability of obtaining the current or more extreme data, given the null hypothesis (i.e., no publication bias), is 90%. The funnel plot has good symmetry, and the distribution of the study results is in line with the expectation of random error, with no obvious signs of bias (Figure 9). The Egger regression yielded an intercept of 1.318 (p = 0.447; F-test p = 0.4465), indicating that the intercept did not significantly deviate from zero (Supplementary File 5). Consequently, there was no statistical evidence of small-study effects or publication bias.

Figure 9
Funnel plot showing Deeks' funnel plot asymmetry test with a p-value of 0.90. The x-axis represents the diagnostic odds ratio on a logarithmic scale, and the y-axis shows one over the square root of effective sample size. Data points, numbered one to seven, represent individual studies. A dashed blue line indicates the regression line.

Figure 9. Deek funnel plot showing publication bias.

Discussion

We conducted a systematic review and meta-analysis on the research on deep learning for early diagnosis of pancreatic cancer over the past five years. Through the analysis of indicators such as accuracy, sensitivity, specificity, and AUC, we aimed to demonstrate the efficacy and clinical applicability of deep learning models in the early diagnosis of pancreatic cancer by learning from CE-CT, EUS images, as well as urine biomarkers, disease trajectories, etc. According to our meta-analysis, the accuracy rate of deep learning models for early diagnosis of pancreatic cancer has been at a relatively high level in the past five years, and the publication bias is relatively small. However, the heterogeneity among the studies is relatively strong. In this regard, we provided further explanations in the Limitations section.

Among the included studies, convolutional neural networks (CNN) were the most commonly used algorithm because it can easily handle multi-dimensional data by using a large number of parameters. At present, imaging indicators such as CE-CT and EUS seem to be more popular data for deep learning, but genomic data, RNA, urine biomarkers, and disease trajectories have also become new objects for deep learning, and perhaps they can become one of the main directions of future research for diagnosing and predicting pancreatic cancer using artificial intelligence algorithms.

In recent years, multi-modal approaches has also become a research hotspot, but its application in the early diagnosis of pancreatic cancer remains to be explored. Multi-modal approaches refer to the simultaneous generation of multiple omics data from the same sample, thereby providing a comprehensive view of biological system processes. Therefore, multi-omics data offer valuable and holistic insights into the complex interactions between biological mechanisms and biomarkers (46, 47). Multi-modal approaches have the potential to integrate different data sources (such as imaging data, genomics, and clinical records), thereby enriching the feature space and improving diagnostic accuracy (48). However, this integration is challenged by data heterogeneity, which arises from differences in data acquisition protocols, patient populations, and measurement scales. Optimization can be achieved from two perspectives. First, from the perspective of data sources, establishing standardized public databases can reduce heterogeneity caused by single data sources. Providing open-source code can help researchers better reproduce study results and reduce heterogeneity caused by different parameter settings. Second, from the perspective of model stability and generalizability, it is encouraged to conduct multicenter studies to train models with large amounts of data, thereby improving the stability and generalizability of models under different devices and operating habits (49, 50).

Deep learning algorithms typically contain a large number of parameters, which need to be learned through data to establish a complex mapping relationship between input features and output results. The more parameters, the stronger the model’s expressive power, but it is also prone to overfitting (i.e., performing well on the training set but having poor generalization ability on external validation sets) (51, 52). Large sample sizes can provide more diverse data distributions, helping deep learning models obtain high-robust features and improve generalization ability (53). Therefore, large sample size data is needed for model training. However, in our meta-analysis, only one study used more than 1000 samples, and other studies were limited by the size of the data. Therefore, in future research, multicenter large sample sizes or even multiple countries may become the main trend.

An interesting finding of the review is that the number of published studies on deep learning models for early diagnosis of pancreatic cancer has been on the rise since 2020. In our meta-analysis, the number of included studies has been on the rise in all years except for 2022 (Figure 1). The reason for the relatively small number of relevant studies in 2022 may be affected by the COVID-19 pandemic. Currently, deep learning for early diagnosis of pancreatic cancer is in a research boom, and this field requires more high-quality research.

Many studies describe deep learning models as “black boxes” models, and we cannot accurately explain their specific working principles (2, 6, 7, 45). Currently, there are studies in the medical field that have explained machine learning models, such as Castagno et al. using the KernelSHAP tool to evaluate the importance of features for model prediction and selecting the top 5 features as core variables to build a new model (54). Du et al. used the SHAP algorithm to explain the model for differentiating hepatocellular carcinoma using gray-scale ultrasound (55).

Furthermore, enhancing model interpretability is crucial for clinical acceptance, as it allows clinicians to understand and trust the decision-making processes of deep-learning algorithms (56). Techniques such as feature visualization, saliency maps, and model-agnostic interpretability methods can provide insights into how models derive their predictions, thereby bridging the gap between complex algorithms and clinical practice (57). As the complexity of models increases, such as deep learning models, the difficulty of model interpretability has significantly risen. However, to enhance the reliability of deep learning models and promote their clinical application in early diagnosis of pancreatic cancer, the interpretability of deep learning models may become the focus of research in the coming years (54, 55, 58).

Limitations

Lv et al. observed that 10 included studies exhibited high heterogeneity; however, even after conducting subgroup analyses and sensitivity analyses, no source of heterogeneity was identified (18). In our Meta-analysis, we encountered the same predicament. Heterogeneity could not be reduced by subgroup analysis.

We analyzed this. Firstly, the number of included studies was limited. Although grouping could be done based on the above criteria, this might be one of the reasons for the high heterogeneity. Secondly, since the deep learning models currently applied in the studies are mostly modified versions of large models such as CNN, we cannot achieve complete uniformity in training the models, which might also be the main reason for the heterogeneity. Additionally, various parameter settings, image quality, and inspection equipment could be sources of heterogeneity, but further exploration is needed.

Conclusion

We believe that this review will be helpful for the scientific community to better understand the application of deep learning technology in risk prediction and diagnosis of pancreatic cancer. Recent published studies have shown high accuracy rates. However, due to limitations in dataset size and diversity, insufficient model interpretability, difficulties in clinical validation and implementation, ethical and legal concerns, as well as the acceptance and trust of clinicians and patients, no deep learning models have been widely applied in clinical practice.

To promote the future clinical application of deep learning models, it is imperative to address several key issues: exploring multi-modal approaches, resolving data heterogeneity problems, enhancing the interpretability of DL models, designing large-sample prospective studies, and conducting external validations.

We also believe that there are still cutting-edge issues that need to be addressed in the risk prediction and diagnosis of pancreatic cancer using artificial intelligence technology, which are specifically reflected in the following four aspects: (1) Genomics, RNA, Urine biomarkers, Disease trajectories and other data seem to have also become new objects for deep learning, and perhaps can become one of the main directions for future research; (2) Multi-center large-sample or even multi-national samples may become the main theme; (3) Early diagnosis of pancreatic cancer smaller than 2 cm remains a challenge; (4) The interpretability of deep learning models and the integration of multi-modal approaches may become the focus of research in the coming years.

Data availability statement

The original contributions presented in the study are included in the article/Supplementary Material. Further inquiries can be directed to the corresponding author.

Author contributions

YB: Data curation, Software, Formal analysis, Methodology, Conceptualization, Writing – review & editing, Investigation, Writing – original draft, Project administration. DRL: Methodology, Conceptualization, Data curation, Validation, Supervision, Formal analysis, Writing – review & editing, Writing – original draft, Project administration. RP: Writing – original draft, Investigation, Software, Validation, Data curation. CD: Validation, Data curation, Writing – original draft, Investigation, Software. DL: Software, Data curation, Conceptualization, Validation, Writing – review & editing, Methodology. XZ: Writing – review & editing, Conceptualization, Resources, Project administration, Formal analysis, Methodology. HL: Writing – original draft, Methodology, Project administration, Conceptualization, Writing – review & editing, Funding acquisition, Validation, Resources, Supervision, Formal analysis.

Funding

The author(s) declare that no financial support was received for the research and/or publication of this article.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declare that no Generative AI was used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fonc.2025.1597969/full#supplementary-material

References

1. Shafeeq Ahmed MH. Convolutional neural network-based deep learning model for the detection of pancreatic ductal adenocarcinoma from computed tomography. HPB. (2024) p:S836. doi: 10.1016/j.hpb.2024.07.267

Crossref Full Text | Google Scholar

2. Lanhong Y, Zheyuan Z, Elif K, Cemal Y, Temel T, Ulas B, et al. A review of deep learning and radiomics approaches for pancreatic cancer diagnosis from medical imaging. Curr Opin Gastroenterol. (2023) 39:436–47. doi: 10.1097/MOG.0000000000000966

PubMed Abstract | Crossref Full Text | Google Scholar

3. Gu J, Pan J, Hu J, Dai L, Zhang K, Wang B, et al. Prospective assessment of pancreatic ductal adenocarcinoma diagnosis from endoscopic ultrasonography images with the assistance of deep learning. Cancer. (2023) 129:2214–23. doi: 10.1002/cncr.34772

PubMed Abstract | Crossref Full Text | Google Scholar

4. Hu SS, Duan BW, Xu L, Huang DP, Liu XG, Gou SH, et al. Enhancing physician support in pancreatic cancer diagnosis: New M-F-RCNN artificial intelligence model using endoscopic ultrasound. Endoscopy Int Open. (2024) 12:E1277–84. doi: 10.1055/a-2422-9214

PubMed Abstract | Crossref Full Text | Google Scholar

5. Chen PT, Wu T, Wang P, Chang D, Liu KL, Wu MS, et al. Pancreatic cancer detection on CT scans with deep learning: A nationwide population-based study. Radiology. (2023) 306:172–82. doi: 10.1148/radiol.220152

PubMed Abstract | Crossref Full Text | Google Scholar

6. Kuwahara T, Hara K, Mizuno N, Haba S, Okuno N, Kuraishi Y, et al. Artificial intelligence using deep learning analysis of endoscopic ultrasonography images for the differential diagnosis of pancreatic masses. Endoscopy. (2023) 55:140–9. doi: 10.1055/a-1873-7920

PubMed Abstract | Crossref Full Text | Google Scholar

7. Aisen AM and Rodrigues PS. Deep learning to detect pancreatic cancer at CT: artificial intelligence living up to its hype. Radiology. (2023) 306:183–5. doi: 10.1148/radiol.222126

PubMed Abstract | Crossref Full Text | Google Scholar

8. Hoogenboom SA, Ravi KK, Engels MM, Irmakci I, Keles E, Bolan CW, et al. Missed diagnosis of pancreatic ductal adenocarcinoma detection using deep convolutional neural network. Gastroenterology. (2021) 160:S–18. doi: 10.1016/S0016-5085(21)00794-0

Crossref Full Text | Google Scholar

9. Viriyasaranon T, Chun JW, Koh YH, Cho JH, Jung MK, Kim SH, et al. Annotation-efficient deep learning model for pancreatic cancer diagnosis and classification using CT images: A retrospective diagnostic study. Cancers. (2023) 15:3392. doi: 10.3390/cancers15133392

PubMed Abstract | Crossref Full Text | Google Scholar

10. Liao WC. Early detection of pancreatic cancer: opportunities provided by cancer-induced paraneoplastic phenomena and artificial intelligence. J Cancer Res Pract. (2023) 10:129–33. doi: 10.4103/ejcrp.eJCRP-D-23-00002

Crossref Full Text | Google Scholar

11. Kurnaz E, Ceylan R, Bozkurt MA, Cebeci H, and Koplay M. A novel deep learning model for pancreas segmentation: pascal U-net. Inteligencia Artificial-Iberoamerican J Artif Intell. (2024) 27:22–36. doi: 10.4114/intartif.vol27iss74pp22-36

Crossref Full Text | Google Scholar

12. Jain S, Sikka G, and Dhir R. A systematic literature review on pancreas segmentation from traditional to non-supervised techniques in abdominal medical images. Artif Intell Rev. (2024) 57:317. doi: 10.1007/s10462-024-10966-1

Crossref Full Text | Google Scholar

13. Castellana R, Fanni SC, Roncella C, Romei C, Natrella M, and Neri E. Radiomics and deep learning models for CT pre-operative lymph node staging in pancreatic ductal adenocarcinoma: A systematic review and meta-analysis. Eur J Radiol. (2024) 176:111510. doi: 10.1016/j.ejrad.2024.111510

PubMed Abstract | Crossref Full Text | Google Scholar

14. Mandal S, Balraj K, Kodamana H, Arora C, Clark JM, Kwon DS, et al. Weakly supervised large-scale pancreatic cancer detection using multi-instance learning. Front Oncol. (2024) 14. doi: 10.3389/fonc.2024.1362850

PubMed Abstract | Crossref Full Text | Google Scholar

15. Jakhar D and Kaur I. Artificial intelligence, machine learning and deep learning: definitions and differences. Clin Exp Dermatol. (2020) 45:131–2. doi: 10.1111/ced.14029

PubMed Abstract | Crossref Full Text | Google Scholar

16. Huang B, Huang H, Zhang S, Zhang D, Shi Q, Liu J, et al. Artificial intelligence in pancreatic cancer. Theranostics. (2022) 12:6931–54. doi: 10.7150/thno.77949

PubMed Abstract | Crossref Full Text | Google Scholar

17. Sriraman H, Badarudeen S, Vats S, and Balasubramanian P. A systematic review of real-time deep learning methods for image-based cancer diagnostics. J Multidiscip Healthcare. (2024) 17:4411–25. doi: 10.2147/JMDH.S446745

PubMed Abstract | Crossref Full Text | Google Scholar

18. Lv B, Wang K, Wei N, Yu F, Tao T, and Shi Y. Diagnostic value of deep learning-assisted endoscopic ultrasound for pancreatic tumors: a systematic review and meta-analysis. Front Oncol. (2023) 13. doi: 10.3389/fonc.2023.1191008

PubMed Abstract | Crossref Full Text | Google Scholar

19. Zhang CH, Achuthan A, and Himel GMS. State-of-the-art and challenges in pancreatic CT segmentation: A systematic review of U-net and its variants. IEEE Access. (2024) 12:78726–42. doi: 10.1109/ACCESS.2024.3392595

Crossref Full Text | Google Scholar

20. Liu KL, Wu T, Chen PT, Tsai YM, Roth H, Wu MS, et al. Deep learning to distinguish pancreatic cancer tissue from non-cancerous pancreatic tissue: a retrospective study with cross-racial external validation. Lancet Digit Health. (2020) 2:e303–13. doi: 10.1016/S2589-7500(20)30078-9

PubMed Abstract | Crossref Full Text | Google Scholar

21. Xiang F, Meng QT, Deng JJ, Wang J, Liang XY, Liu XY, et al. A deep learning model based on contrast-enhanced computed tomography for differential diagnosis of gallbladder carcinoma. Hepatobiliary Pancreatic Dis Int. (2024) 23:376–84. doi: 10.1016/j.hbpd.2023.04.001

PubMed Abstract | Crossref Full Text | Google Scholar

22. Rousta F, Esteki A, Shalbaf A, Sadeghi A, Moghadam PK, and Voshagh A. Application of artificial intelligence in pancreas endoscopic ultrasound imaging- A systematic review. Comput Methods Programs BioMed. (2024) 250:108205. doi: 10.1016/j.cmpb.2024.108205

PubMed Abstract | Crossref Full Text | Google Scholar

23. Huang J, Fan X, and Liu W. Applications and prospects of artificial intelligence-assisted endoscopic ultrasound in digestive system diseases. Diagnostics. (2023) 13:2815. doi: 10.3390/diagnostics13172815

PubMed Abstract | Crossref Full Text | Google Scholar

24. Kuwahara T, Hara K, Mizuno N, Haba S, Okuno N, Koda H, et al. Current status of artificial intelligence analysis for endoscopic ultrasonography. Digestive Endoscopy. (2021) 33:298–305. doi: 10.1111/den.13880

PubMed Abstract | Crossref Full Text | Google Scholar

25. Tang AL, Tian L, Gao K, Liu R, Hu S, Liu JZ, et al. Contrast-enhanced harmonic endoscopic ultrasound (CH-EUS) MASTER: A novel deep learning-based system in pancreatic mass diagnosis. Cancer Med. (2023) 12:7962–73. doi: 10.1002/cam4.5578

PubMed Abstract | Crossref Full Text | Google Scholar

26. Qin X, Zhang M, Zhou C, Ran T, Pan Y, Deng Y, et al. A deep learning model using hyperspectral image for EUS-FNA cytology diagnosis in pancreatic ductal adenocarcinoma. Cancer Med. (2023) 12:17005–17. doi: 10.1002/cam4.6335

PubMed Abstract | Crossref Full Text | Google Scholar

27. Naito Y, Tsuneki M, Fukushima N, Koga Y, Higashi M, Notohara K, et al. A deep learning model to detect pancreatic ductal adenocarcinoma on endoscopic ultrasound-guided fine-needle biopsy. Sci Rep. (2021) 11:8454. doi: 10.1038/s41598-021-87748-0

PubMed Abstract | Crossref Full Text | Google Scholar

28. Shao M, Cheng C, Hu CY, Zheng J, Zhang B, Wang T, et al. Semisupervised 3D segmentation of pancreatic tumors in positron emission tomography/computed tomography images using a mutual information minimization and cross-fusion strategy. Quantitative Imaging Med Surg. (2024) 14:1747–65. doi: 10.21037/qims-23-1153

PubMed Abstract | Crossref Full Text | Google Scholar

29. Kaur H, Saini SK, Thakur N, and Juneja M. Survey of denoising, segmentation and classification of pancreatic cancer imaging. Curr Med Imaging. (2024) 20:e150523216892. doi: 10.2174/1573405620666230515090523

PubMed Abstract | Crossref Full Text | Google Scholar

30. Jiang X, Hu Z, Wang S, and Zhang Y. Deep learning for medical image-based cancer diagnosis. Cancers. (2023) 15:3608. doi: 10.3390/cancers15143608

PubMed Abstract | Crossref Full Text | Google Scholar

31. Watson MD, Baimas-George MR, Murphy KJ, Pickens RC, Iannitti DA, Martinie JB, et al. Pure and hybrid deep learning models can predict pathologic tumor response to neoadjuvant therapy in pancreatic adenocarcinoma: A pilot study. Am Surg. (2021) 87:1901–9. doi: 10.1177/0003134820982557

PubMed Abstract | Crossref Full Text | Google Scholar

32. Kim J, Bae S, Yoon SM, and Jeong S. Roadmap for providing and leveraging annotated data by cytologists in the PDAC domain as open data: support for AI-based pathology image analysis development and data utilization strategies. Front Oncol. (2024) 14. doi: 10.3389/fonc.2024.1346237

PubMed Abstract | Crossref Full Text | Google Scholar

33. Fassler DJ, Abousamra S, Gupta R, Chen C, Zhao M, Paredes D, et al. Deep learning-based image analysis methods for brightfield-acquired multiplex immunohistochemistry images. Diagn Pathol. (2020) 15:100. doi: 10.1186/s13000-020-01003-0

PubMed Abstract | Crossref Full Text | Google Scholar

34. Khellaf A, Lando L, and Trinh QH. Classical computer vision and modern deep-learning of pancreatic stroma histology features to diagnose cancer. Ann Oncol. (2023) 34:S1537. doi: 10.1016/j.annonc.2023.10.294

Crossref Full Text | Google Scholar

35. Hong SJ, Hou JU, Chung BMJ, Kang SH, Shim BS, Lee SL, et al. Convolutional neural network model for automatic recognition and classification of pancreatic cancer cell based on analysis of lipid droplet on unlabeled sample by 3D optical diffraction tomography. Comput Methods Programs Biomedicine. (2024) 246:108041. doi: 10.1016/j.cmpb.2024.108041

PubMed Abstract | Crossref Full Text | Google Scholar

36. Fu H, Mi WM, Pan BJ, Guo YC, Li J, Xu RY, et al. Automatic pancreatic ductal adenocarcinoma detection in whole slide images using deep convolutional neural networks. Front Oncol. (2021) 11. doi: 10.3389/fonc.2021.665929

PubMed Abstract | Crossref Full Text | Google Scholar

37. Karar ME, El-Fishawy N, and Radad M. Automated classification of urine biomarkers to diagnose pancreatic cancer using 1-D convolutional neural networks. J Biol Eng. (2023) 17:28. doi: 10.1186/s13036-023-00340-0

PubMed Abstract | Crossref Full Text | Google Scholar

38. Linh VTN, Mun J, Kim Y, Han IW, Kim H, Park SG, et al. 3D plasmonic coral nanoarchitecture paper for label-free human urine sensing and deep learning-assisted cancer screening. Biosensors Bioelectronics. (2023) 224:115076. doi: 10.1016/j.bios.2023.115076

PubMed Abstract | Crossref Full Text | Google Scholar

39. Acer I, Bulucu FO, Içer S, and Latifoglu F. Early diagnosis of pancreatic cancer by machine learning methods using urine biomarker combinations. Turkish J Electrical Eng Comput Sci. (2023) 31:112–25. doi: 10.55730/1300-0632.3974

Crossref Full Text | Google Scholar

40. Placido D, Yuan B, Hjaltelin JX, Zheng C, Haue AD, Chmura PJ, et al. Pancreatic cancer risk predicted from disease trajectories using deep learning. Nat Med. (2021) 29:1113–22. doi: 10.1101/2021.06.27.449937

Crossref Full Text | Google Scholar

41. Placido D, Yuan B, Hjaltelin JX, Zheng CL, Haue AD, Chmura PJ, et al. A deep learning algorithm to predict risk of pancreatic cancer from disease trajectories. Nat Med. (2023) 29:1113–+. doi: 10.1038/s41591-023-02332-5

PubMed Abstract | Crossref Full Text | Google Scholar

42. Park J, Artin MG, Lee KE, Pumpalova YS, Ingram MA, May BL, et al. Deep learning on time series laboratory test results from electronic health records for early detection of pancreatic cancer. J BioMed Inform. (2022) 131:104095. doi: 10.1016/j.jbi.2022.104095

PubMed Abstract | Crossref Full Text | Google Scholar

43. Hong J, Hou WJ, Sheng N, Zuo CM, and Wang Y. HATZFS predicts pancreatic cancer driver biomarkers by hierarchical reinforcement learning and zero-forcing set. Expert Syst Appl. (2025) 260:125435. doi: 10.1016/j.eswa.2024.125435

Crossref Full Text | Google Scholar

44. Shen H, Yang M, Liu J, Chen K, and Li X. Development of a deep learning model for cancer diagnosis by inspecting cell-free DNA end-motifs. NPJ Precis Oncol. (2024) 8:160. doi: 10.1038/s41698-024-00635-5

PubMed Abstract | Crossref Full Text | Google Scholar

45. Darmofal M, Suman S, Atwal G, Toomey M, Chen JF, Chang JC, et al. Deep-learning model for tumor-type prediction using targeted clinical genomic sequencing data. Cancer Discov. (2024) 14:1064–81. doi: 10.1158/2159-8290.CD-23-0996

PubMed Abstract | Crossref Full Text | Google Scholar

46. Ramesh D, Manickavel P, Ghosh S, and Bhat M. The integration of multi-omics with artificial intelligence in hepatology: A comprehensive review of personalized medicine, biomarker identification, and drug discovery. J Clin Exp Hepatol. (2025) 15:102611. doi: 10.1016/j.jceh.2025.102611

PubMed Abstract | Crossref Full Text | Google Scholar

47. Abdullakutty F, Akbari Y, Al-Maadeed S, Bouridane A, Talaat IM, and Hamoudi R. Histopathology in focus: a review on explainable multi-modal approaches for breast cancer diagnosis. Front Med (Lausanne). (2024) 11:1450103. doi: 10.3389/fmed.2024.1450103

PubMed Abstract | Crossref Full Text | Google Scholar

48. Wang X, Figueredo G, Li R, Zhang WE, Chen W, and Chen X. A survey of deep-learning-based radiology report generation using multimodal inputs. Med Image Anal. (2025) 103:103627. doi: 10.1016/j.media.2025.103627

PubMed Abstract | Crossref Full Text | Google Scholar

49. Berloco F, Zaccaria GM, Altini N, Colucci S, and Bevilacqua V. A multimodal framework for assessing the link between pathomics, transcriptomics, and pancreatic cancer mutations. Comput Med Imaging Graph. (2025) 123:102526. doi: 10.1016/j.compmedimag.2025.102526

PubMed Abstract | Crossref Full Text | Google Scholar

50. Xu X, Li J, Zhu Z, Zhao L, Wang H, Song C, et al. A comprehensive review on synergy of multi-modal data and AI technologies in medical diagnosis. Bioengineering (Basel). (2024) 11:219. doi: 10.3390/bioengineering11030219

PubMed Abstract | Crossref Full Text | Google Scholar

51. Rai HM, Yoo J, and Razaque A. Comparative analysis of machine learning and deep learning models for improved cancer detection: A comprehensive review of recent advancements in diagnostic techniques. Expert Syst Appl. (2024) 255:124838. doi: 10.1016/j.eswa.2024.124838

Crossref Full Text | Google Scholar

52. Ma Y, Jia F, Liu K, and Liu Y. Current status of the application of artificial intelligence in the diagnosis and treatment of pancreatic cancer. J Clin Hepatol. (2024) 40:2121–6. doi: 10.12449/JCH241032

Crossref Full Text | Google Scholar

53. Patel H, Zanos T, and Hewitt DB. Deep learning applications in pancreatic cancer. Cancers. (2024) 16:436. doi: 10.3390/cancers16020436

PubMed Abstract | Crossref Full Text | Google Scholar

54. Castagno S, Birch M, van der Schaar M, and McCaskie A. Predicting rapid progression in knee osteoarthritis: a novel and interpretable automated machine learning approach, with. specific Focus young patients early disease. Ann Rheum Dis. (2025) 84:124–35. doi: 10.1136/ard-2024-225872

PubMed Abstract | Crossref Full Text | Google Scholar

55. Du Z, Fan F, Ma J, Liu J, Yan X, Chen X, et al. Development and validation of an ultrasound-based interpretable machine learning model for the classification of ≤3 cm hepatocellular carcinoma: a multicentre retrospective diagnostic study. EClinicalMedicine. (2025) 81:103098. doi: 10.1016/j.eclinm.2025.103098

PubMed Abstract | Crossref Full Text | Google Scholar

56. Zhang X, Shen YY, Su GH, Guo Y, Zheng RC, Du SY, et al. A dynamic contrast-enhanced MRI-based vision transformer model for distinguishing HER2-zero, -low, and -positive expression in breast cancer and exploring model interpretability. Adv Sci (Weinh). (2025) p:e03925. doi: 10.1002/advs.202503925

PubMed Abstract | Crossref Full Text | Google Scholar

57. Harris CE, Liu L, Almeida L, Kassick C, and Makrogiannis S. Artificial intelligence in pediatric osteopenia diagnosis: evaluating deep network classification and model interpretability using wrist X-rays. Bone Rep. (2025) 25:101845. doi: 10.1016/j.bonr.2025.101845

PubMed Abstract | Crossref Full Text | Google Scholar

58. Hanassab S, Nelson SM, Akbarov A, Yeung AC, Hramyka A, Alhamwi T, et al. Explainable artificial intelligence to identify follicles that optimize clinical outcomes during assisted conception. Nat Commun. (2025) 16:296. doi: 10.1038/s41467-024-55301-y

PubMed Abstract | Crossref Full Text | Google Scholar

Keywords: pancreatic cancer (PC), deep learning, diagnosis methods, research trends, meta-analysis

Citation: Bi Y, Li D, Pang R, Du C, Li D, Zhao X and Lv H (2025) Diagnosis methods for pancreatic cancer with the technique of deep learning: a review and a meta-analysis. Front. Oncol. 15:1597969. doi: 10.3389/fonc.2025.1597969

Received: 22 March 2025; Accepted: 30 July 2025;
Published: 20 August 2025.

Edited by:

Ravi Kumar Sharma, Chandigarh University, India

Reviewed by:

Vinod Kumar Yata, Malla Reddy University, India
Xubo Song, Oregon Health and Science University, United States
Archana Machireddy, Oregon Health and Science University, United States

Copyright © 2025 Bi, Li, Pang, Du, Li, Zhao and Lv. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Haitao Lv, Z2FuZGFubHZoYWl0YW9AMTYzLmNvbQ==

These authors have contributed equally to this work and share first authorship

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.