Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Oncol., 08 October 2025

Sec. Cancer Imaging and Image-directed Interventions

Volume 15 - 2025 | https://doi.org/10.3389/fonc.2025.1678525

Development of an AI-driven digital assistance system for real-time safety evaluation and quality control in laparoscopic liver surgery

Zi-Yang Peng,&#x;Zi-Yang Peng1,2†Zhi-Bo Wang&#x;Zhi-Bo Wang2†Yan YanYan Yan3Hao-Qian PengHao-Qian Peng3Yong-Tai MaYong-Tai Ma3Yu-Tong LiYu-Tong Li3Yao-Xing RenYao-Xing Ren2Jun-Xi XiangJun-Xi Xiang4Kun GuoKun Guo4Gang WangGang Wang5Jian-Feng DuanJian-Feng Duan6Xiao-Wen LiXiao-Wen Li7Yu GuanYu Guan8Xue-Min LiuXue-Min Liu4Rong-Qian Wu,Rong-Qian Wu2,4Yi Lyu,*Yi Lyu2,4*Li Yu*Li Yu4*
  • 1Department of Otorhinolaryngology-Head and Neck Surgery, The First Affiliated Hospital of Xi’an Jiaotong University, Xi’an, Shaanxi, China
  • 2School of Future Technology, National Local Joint Engineering Research Center for Precision Surgery & Regenerative Medicine, Xi’an Jiaotong University, Xi’an, Shaanxi, China
  • 3Health Science Center, Xi’an Jiaotong University, Xi’an, Shaanxi, China
  • 4Department of Hepatobiliary Surgery, The First Affiliated Hospital of Xi’an Jiaotong University, Xi’an, Shaanxi, China
  • 5Department of Hepatobiliary and Pancreatic Surgery II, Baoji Central Hospital, Baoji, Shaanxi, China
  • 6Department of Hepatobiliary Surgery, Hanzhong 3201 Hospital, Hanzhong, Shaanxi, China
  • 7Department of Geriatric General Surgery, The Second Affiliated Hospital of Xi’an Jiaotong University, Xi’an, Shaanxi, China
  • 8Department of Anesthesiology and Perioperative Medicine, The First Affiliated Hospital of Xi’an Jiaotong University, Xi’an, Shaanxi, China

Background: By performing AI-driven workflow analysis, intelligent surgical systems can provide real-time intraoperative quality control and alerts. We have upgraded an Intelligent Surgical Assistant (ISA) through integrating a redesigned hierarchical recognition algorithm, an expanded surgical dataset, and an optimized real-time intraoperative feedback framework.

Objective: We aimed to assess the accuracy of the ISA in real-time instrument tracking, organ segmentation, and phase classification during laparoscopic hemi-hepatectomy.

Methods: In this retrospective multi-center analysis, a total of 142861 annotated frames were collected from 403 laparoscopic hemi-hepatectomy videos across 4 centers to build a comprehensive database of surgical video annotations. Each frame was labeled for surgical phase, organs, and instruments. The algorithm in the ISA was retrained using a hybrid deep learning framework integrating instrument tracking, organ segmentation, and phase classification. We then established a scoring system for surgical image recognition and evaluated the algorithm’s recognition accuracy and inter-operator consistency across different surgical teams.

Results: The upgraded ISA achieved an accuracy of 89% in real-time recognition of instruments and organs. The programmatic phase classification for laparoscopic hemi-hepatectomy reached an average accuracy of 91% (p<0.001), enabling a correct recognition of surgical events. The inter-operator variability in recognition was reduced to 14.3%, highlighting the potential of AI-assisted quality control to standardize intraoperative alerts. Overall, the ISA demonstrated high precision and consistency in phase recognition and operative field evaluation across all phases (accuracy >87%, specificity ~90% in each phase). Notably, critical phases (Phase 1 and Phase 5) were identified with an exceptional accuracy area under the curve (AUC 0.96 in Phase 1; AUC 0.87 in Phase 5), indicating that key surgical procedures could be phased with very low false-alarm rates.

Conclusions: The optimized ISA provides a highly accurate real-time interpretation of surgical phases and a strong potential to standardize surgical procedures, thus guaranteeing the outcomes and safety of laparoscopic hemi-hepatectomy.

1 Introduction

Since the first laparoscopic cholecystectomy accomplished by Philippe Mouret in 1987, minimally invasive techniques have flourished, allowing an array of surgical procedures from simple elective to complex comprehensive, such as tumor and organ removal (13). Surgical robots have further polished these procedures. The da Vinci system, FDA-approved in 2000, offers 3D high-definition vision, wristed instruments, and tremor filtering, and can markedly increase the precision of surgical procedures (4). Clinical studies have confirmed that, compared with open surgery, minimally invasive and robotic surgery boasts a lower perioperative complication rate (from 15.2% down to 9.8%), a shorter hospital stay (on average by 2.3 days), and a milder postoperative pain (VAS reduced by 1.7 points) (5). These advantages have been translated into a lower morbidity and a faster recovery (6, 7). Meta-analyses show that laparoscopic approaches can be safely applied in liver surgery, even among patients with malignant diseases, offering similar oncological outcomes with less blood loss and shorter hospitalization, compared to open surgery (8, 9).

However, traditional laparoscopy is still challenged by a “fulcrum effect” (opposite motion beneath trocar fulcrums) (10), limited tactile feedback, and a two-dimensional operative view. It is particularly difficulty to overcome these challenges when dissecting dense adhesions or structures with complex anatomies (1114). Cognitive errors and fatigue of surgeons may affect the surgical outcomes (15). Therefore, artificial intelligence systems may be integrated with laparoscopy to bring more surgical benefits (16).

While AI models have demonstrated significant success in discrete tasks such as tool tracking or anatomical segmentation, a key challenge remains: integrating these functions into a single, cohesive system that maintains both high accuracy across multiple tasks and the real-time inference speed required for clinical utility. Many existing systems excel at one task but often struggle to perform comprehensive, multi-faceted analysis without sacrificing speed (17, 18). To address this gap, we developed and validated an Intelligent Surgical Assistant (ISA) for laparoscopic hemi-hepatectomy. Our system is specifically designed to perform simultaneous instrument tracking, organ segmentation, and surgical phase classification, all while operating at a clinically viable frame rate. By providing this holistic, real-time analysis, the ISA aims to deliver timely and relevant feedback to surgeons, enhancing intraoperative quality control and safety (19).

This ISA has been trained to distinguish six stages on the phase-labeled videos of laparoscopic hemi-hepatectomy: Phase 1 (intraoperative ultrasound), Phase 2 (first hepatic hilum dissection), Phase 3 (second hilum dissection), Phase 4 (exposure of the middle hepatic vein), Phase 5 (post-resection hemostasis on liver cut surface), and Phase 0 (non-critical steps). In clinical settings, ISA processes the incoming laparoscopic video frame-by-frame at ≥30 FPS, meanwhile labeling the current phase and offering a clarity score in real time. The surgeon can thus verify if a critical phase has been satisfactorily completed. For example, a high clarity score in “Phase 5” indicates a clear surgical field, in which liver transection is complete and hemostasis is successful. Conversely, a low clarity score indicates the presence of smoke or bleeding, warning that the surgeon should pause to clear them. Using ISA, a surgeon can recognize the phase and check the field more precisely, thereby ensuring the safety of all procedures.

2 Methods

2.1 Study design and ethics

This observational study involved no additional interventions beyond standard laparoscopic hemi-hepatectomy. All patient data were de-identified before analysis in compliance with local data privacy regulations and the Declaration of Helsinki. Data use was approved by the Clinical Research Ethics Committee of Xi’an Jiaotong University, Approval Date: July 15, 2023, Approval No. XJTU1AF2023LSK-429. All patients (aged >18 years) had shown consent to our videoing surgical procedures for research purposes. All participating surgeons provided informed consent for the retrospective use of their surgical videos in workflow evaluation and frame recognition. Video data were retrospectively collected between Aug 30, 2023, and Aug 7, 2024.

2.2 Dataset construction (laparoscopic hemi-hepatectomy)

This retrospective study included a cohort of 403 patients who underwent laparoscopic hemi-hepatectomy between August 2023 and August 2024, from which 403 surgical videos were obtained from four participating centers. The inclusion criteria were: (1) adult patients (age > 18 years); (2) undergoing elective laparoscopic hemi-hepatectomy; (3) availability of complete, high-quality surgical video recordings; and (4) provision of informed consent for the research use of video data. Exclusion criteria were: (1) emergency surgeries; (2) procedures converted to open surgery due to non-oncological reasons (e.g., equipment failure); (3) patients with prior major upper abdominal surgery; and (4) videos with significant portions obscured by technical issues or poor quality. The baseline demographic and clinical characteristics of the patient cohort are summarized in Table 1.

Table 1
www.frontiersin.org

Table 1. Baseline characteristics of the patient cohort (N = 403).

From these videos, the internal deep learning cohort was constructed through a rigorous, two-step quality control process. First, an initial frame selection was conducted by junior surgeons (PZY, YY, PHQ, MYT, LYT). They were tasked with selecting representative frames for each key surgical phase, based on predefined visual criteria designed to ensure anatomical clarity, as detailed in our scoring system (now Table 2). For instance, frames selected for Phase 2 (‘First hepatic hilum dissection’) were required to show a clear exposure of the hilar structures. Each frame selected in this initial step also had to meet a confidence level exceeding 50% for clear identification.

Table 2
www.frontiersin.org

Table 2. Scoring criteria for surgical phases.

These initially selected frames then underwent a second-step review and supervision by senior surgeons (XJX, GK, LXM, LY). The senior surgeons’ review protocol was twofold. First, they qualitatively validated that each frame was a high-quality, representative example of its designated phase, rejecting images with visual obstructions such as excessive smoke, blood, or off-target camera angles. Second, they applied a much stricter quantitative threshold, excluding any frame with a final confidence level below 90%. During this stage, images exhibiting similar surgical features were also reduced to minimize redundancy. This stringent quality control process resulted in the exclusion of 5,934 frames, with the final 136,927 high-confidence frames being retained for annotation.

Each of the retained frames was annotated with segmentation masks in distinct colors using LabelMe software, highlighting critical areas such as liver parenchyma (primary target), biliary structures, major blood vessels, surgical instruments, and background structures (Figure 1B). The inter-operator consistency of the review process was assessed using Fleiss’ Kappa. To quantify this, we calculated the coefficient on a randomly selected subset of 10% of the annotated frames, yielding a Kappa value of 0.88 (p < 0.001), indicating almost perfect agreement.

Figure 1
Flowchart depicting an ISA model development process for surgical classification. Top left shows dataset construction involving 403 videos and 142,861 lesion images with annotations. Top right illustrates model development with expert and AI data identifying surgical instruments and organs. Bottom left presents classification performance across five phases of surgical procedures. Bottom right highlights detection results, noting improved operation quality control, advanced algorithm design, quick critical site identification, safer surgeries, and faster recovery, with an AUC of 0.959.

Figure 1. Integrated dataset and ISA model overview. (A) The dataset construction pipeline. The dataset was sourced from 403 surgical videos, and the process included frame extraction, annotation by junior surgeons, and final review by experts. (B) The development framework of the Intelligent Surgical Assistant (ISA) model with separate branches for identifying surgical instruments and anatomical structures. Example outputs are shown as expert-annotated masks (Expert data), compared to the model’s predicted segmentation (AI data) for both tasks. (C) The classification performance across five surgical phases: Phase 1 (Endoscopic ultrasound examination), Phase 2 (First hepatic hilum anatomy), Phase 3 (Second hepatic portal anatomy), Phase 4 (Anatomy of the hepatic portal vein), and Phase 5 (Observation of hemostasis at the resection site after removal of the diseased liver). (D) The significance of the detection results: this advanced algorithmic system enables whole-process quality control, accurately identifies critical sites (arrow), and contributes to safer surgical procedures and faster postoperative recovery. The model achieves an AUC of 0.959, demonstrating a robust discriminatory capability.

Each annotated frame was then labeled according to the surgical phase, with the dataset stratified to ensure representation of the five primary surgical stages (Phases 1-5). To enhance the model’s generalizability, the study incorporated 10-fold cross-validation (Figure 2). This technique divided the dataset into 10 subsets, with each subset serving as the test set once, while the remaining nine subsets were used for training. This approach facilitated multiple rounds of training and testing, ensuring robust performance across diverse datasets.

Figure 2
Diagram illustrating a neural network architecture for image analysis. Part A shows the “Split the Branch” process using Deeplabv3 with Xception65, transforming input images into segmented outputs. Part B displays the “Feature Extraction Branch” using ResNet-50, followed by the “Predicting Branch” for stage and image quality using dropout layers.

Figure 2. Dual-branch network architecture for joint prediction of stages and quality. (A) The top branch (Segmentation) uses DeepLabv3+ with Xception65 backbone to generate 5-channel outputs highlighting tools and anatomy. (B) The bottom branch (Feature Extraction) passes the image through ResNet-50 convolutional layers (Conv1–Conv5) to capture contextual information. These outputs are concatenated into a 1920×1080×5 tensor, which is then processed by two parallel dense “Predicting Branches” ending in softmax layers. By fusing segmentation cues with deep semantic features, the network robustly predicts the visual quality of each surgical phase.

2.3 Training ISA using AI models

We designed a deep learning model of hybrid segmentation and multi-task joint learning (Figure 3). Input frames (1920×1080 resolution) were first processed by a “Split-and-Branch” semantic segmentation module to identify and mask the key anatomical structures. The resultant mask was fused with the original image and fed into a pre-trained ResNet-50 backbone to extract deep features (2048-dimensional from the Conv5_x stage). The ResNet’s layers (Conv1 to Conv5_x) were run to extract features in a hierarchical sequence: upper layers to capture low-level edges and textures, while deeper layers to model complex organs and instruments. The segmentation mask emphasizes salient regions, allowing ResNet to selectively catch instrument shapes or liver tissue textures.

Figure 3
Flowchart of patient cohort data analysis for laparoscopic hepatectomy. It shows the division of patients into internal and external cohorts. Internal cohort data is labeled, and clips are categorized into phases with training and test sets. Cross-validation is performed, and the International Surgical Association (ISA) performance is compared between senior and junior surgeons.

Figure 3. Flowchart of ISA development and evaluation. ISA refers to the Artificial Intelligence model constructed in this study for intraoperative phase recognition during laparoscopic hepatectomy. The flowchart outlines the entire pipeline including patient enrollment, video acquisition, phase-wise annotation, data screening, and model validation. Patients were divided into internal and external cohorts. Only clips with phase transition (PT) confidence above 90% were retained for analysis. A 10-fold cross-validation strategy was applied on the internal dataset for performance evaluation.

The model’s architecture was intentionally designed for computational efficiency to ensure its utility in a real-time clinical setting. This efficiency is primarily achieved through a shared feature extractor and lightweight prediction branches. By using a single ResNet-50 backbone to generate shared features for all downstream tasks, we effectively avoid the redundant computations that would arise from running multiple independent models.

From these shared backbone features, two lightweight fully-connected branches simultaneously (1) determine the surgical phase (Phases 0–5) (Figure 1C) and (2) score the image clarity. We trained the model in PyTorch on an NVIDIA Tesla V100 GPU, achieving an average inference latency of approximately 52 ms per frame (corresponding to 19.2 FPS as reported in Table 3), which was sufficient for generating intraoperative real-time feedback. The phase classification branch output a probability distribution over the 6 phases, while the quality branch predicted a scalar clarity score reflecting visibility (smoking, bleeding, etc.). We employed joint loss optimization in both tasks, so that the shared features could benefit both phase recognition and clarity assessment.

Table 3
www.frontiersin.org

Table 3. Performance of the model.

We designed a deep learning model of hybrid segmentation and multi-task joint learning (Figure 3). Input frames (1920×1080 resolution) were first processed by a “Split-and-Branch” semantic segmentation module to identify and mask the key anatomical structures. The resultant mask was fused with the original image and fed into a pre-trained ResNet-50 backbone to extract deep features (2048-dimensional from the Conv5_x stage). The ResNet’s layers (Conv1 to Conv5_x) were run to extract features in a hierarchical sequence: upper layers to capture low-level edges and textures, while deeper layers to model complex organs and instruments. The segmentation mask emphasizes salient regions, allowing ResNet to selectively catch instrument shapes or liver tissue textures.

From these shared backbone features, two lightweight fully-connected branches simultaneously (1) determine the surgical phase (Phases 0–5) (Figure 1C) and (2) score the image clarity. We trained the model in PyTorch on an NVIDIA Tesla V100 GPU, achieving a millisecond inference (average latency ~52 ms per frame), which was sufficient for generating intraoperative real-time feedback. The phase classification branch output a probability distribution over the 6 phases, while the quality branch predicted a scalar clarity score reflecting visibility (smoking, bleeding, etc.). We employed joint loss optimization in both tasks, so that the shared features could benefit both phase recognition and clarity assessment.

2.4 Evaluation metrics

We evaluate model performance using standard object detection metrics. AP50, AP75, and AP50:95 represent the Average Precision (AP) under Intersection-over-Union (IoU) thresholds of 0.5, 0.75, and the average from 0.5 to 0.95 (step = 0.05), respectively. AP is computed as the area under the Precision-Recall (PR) curve:

AP=01p(r)dr

where p(r)denotes precision at recall r. In addition, APM and APL measure detection accuracy for medium-sized and large-sized objects, respectively, following the COCO evaluation protocol. Frame rate (FPS) reflects inference speed and computational efficiency.

2.5 Statistical analysis

The performance of the ISA was evaluated according to accuracy, precision, recall, and F1-score. Phase recognition results were also summarized in a confusion matrix. To ensure performance was superior to random assignment, the statistical significance of the phase classification results was validated using a chi-square test (p < 0.05).

2.6 Clarity scoring system for five key intraoperative phases of laparoscopic liver resection

Table 2 defines a three-point system (0–2) for scoring the clarity in each of the five phases of laparoscopic liver resection, including: Phase 1 (intraoperative liver ultrasound), Phase 2 (first porta hepatis dissection), Phase 3 (second porta hepatis dissection), Phase 4 (middle hepatic vein exposure), and Phase 5 (electrocoagulation of the liver section surface). Phase 0 was set as a non-critical background phase. For each phase, a score of 2 represented a full visualization of anatomical structures and an optimal position of the camera, 1 indicated partial anatomical exposure or a suboptimal position, and 0 denoted indistinguishable anatomical structures or an obstructed vision (e.g., smoke, blood, off-target lens). The clinical validity of this scoring system is rooted in its development by senior hepatobiliary surgeons and its direct correlation with intraoperative safety. The criteria for each score were established through expert consensus based on extensive surgical experience. A high clarity score (Score 2) represents an optimal surgical field, which is a prerequisite for the safe identification of critical anatomical structures and the prevention of iatrogenic injury. Conversely, a low score (0 or 1) signifies a compromised view due to factors like bleeding, smoke, or suboptimal exposure. Such situations are clinically significant as they substantially increase the risk of complications. Therefore, this scoring system serves as a clinically relevant and valid proxy for quantifying the quality and safety of the operative field in real-time.

3 Results

3.1 Phase classification

The ISA demonstrated an average accuracy of 91% in classifying the five key phases (p<0.001). Figure 4 presents phase recognition results for a representative test case, with color-coded ribbons comparing the model’s predictions to the ground-truth annotations. Performance details are shown in Table 4 (confusion matrix). The ISA matched most of the frames with the phase correctly, with an accuracy of 89.0% in Phase 1 and 90.5% in Phase 5, indicating its high reliability in phase segmentation. Misclassifications were rare (<8% for any off-diagonal), primarily in frames at the transition between two phases. Overall, the model could clearly distinguish between major procedural steps of laparoscopic hemi-hepatectomy, with a recall >82% in each phase.

Figure 4
Bar diagrams comparing ground truth and estimated phases. The ground truth bar shows large segments of red, blue, gray, and purple with a long green section. The estimated phase (original) displays more fragmented colors with many thin lines, while the estimated phase (postprocessing) closely resembles the ground truth with fewer divisions.

Figure 4. Phase recognition results for one laparoscopic hemi-hepatectomy video. For a representative case from the test set, the three color-coded ribbons illustrate surgical-phase predictions versus ground truth along the temporal axis. The top ribbon shows the ground-truth labels; the middle ribbon presents the primary model predictions; the bottom ribbon displays the refined predictions after post-processing. The laparoscopic cholecystectomy (LC) procedure was temporally divided into five phases: (1) Endoscopic ultrasound examination, (2) First hepatic hilum anatomy, (3) Second hepatic portal anatomy, (4) Anatomy of the hepatic portal vein, and (5) Observation of hemostasis at the section after isolated diseased liver. In the top ribbon of Figure 5, these phases are sequentially encoded with five distinct colors.

Table 4
www.frontiersin.org

Table 4. Confusion matrix of phase prediction (percent).

While overall misclassifications were rare (<8% for any off-diagonal), a closer analysis of the confusion matrix (Table 4) reveals that the most frequent misclassification occurred between Phase 3 and Phase 0 (11.4%), suggesting that the final moments of the second hilum dissection can be visually similar to non-critical operative steps. Similarly, some confusion was observed between Phase 2 and Phase 0 (8.2%). These specific transition errors highlight key areas for future model refinement.

3.2 Spatial focus of the model during key surgical phases

To investigate the model’s visual attention during critical operative tasks, we analyzed its multi-level feature extraction across two representative stages: first hepatic hilum occlusion and hepatic pedicle dissection (Figure 5). The network gradually constructed semantic representations by extracting local textures and anatomical boundaries from raw laparoscopic images. During the hilum occlusion phase, the activation maps concentrated around the portal vein and the site of vascular clamping, successfully capturing the convergence zone of the hepatic triad. In the pedicle dissection phase, the model’s focus shifted toward the hepatic artery and bile duct trajectories, aligning well with the surgeon’s operative field. The final output heatmaps exhibited high-intensity responses localized precisely over the regions of surgical manipulation, reflecting accurate anatomical comprehension by the network. These attention distributions were tightly aligned with intraoperative targets, suggesting effective feature learning in anatomically complex environments.

Figure 5
A 4x4 grid of images shows various medical and thermal imaging scenes. Panels A, H, I, and P depict surgical procedures with tubes and tissue. Panels B, C, D, E, F, G, J, K, L, M, N, and O display thermal or heat map images with varying color intensities indicating different temperature levels. Each panel serves to compare the visual and thermal perspectives in medical imagery.

Figure 5. Heatmap visualizations of neural network activations. Heatmap visualizations of neural network activation are shown to illustrate the model’s response during phase recognition in liver surgery. (A, I) depict the input surgical images. (B–G, J–O) display the corresponding feature maps extracted by the backbone network, capturing key visual cues relevant to the identification of the first hepatic portal occlusion phase and the hepatic pedicle dissection phase, respectively. (H, P) present the final output images, highlighting the network’s interpretive focus for accurate surgical phase classification.

3.3 Image clarity evaluation

Table 5 summarizes the precision, recall, F1-score, specificity, and overall accuracy of the ISA in judging the image clarity in each phase. The ISA achieved the highest AUC (0.96) in Phase 1 (Figure 1D), indicating its strongest ability to discriminate Phase 1. By contrast, the lowest AUC (0.87) and lowest accuracy (0.85) were observed in Phase 3, indicating its relatively weaker performance in recognizing the procedures in Phase 3.

Table 5
www.frontiersin.org

Table 5. Classification performance metrics by phase.

3.4 Performances of the ISA across multiple cohorts

As shown by the results from the validation cohorts (Figure 6A), the ISA achieved the highest AUC (0.9598) in Phase 1, followed by Phase 5 (0.93), Phase 4 (0.92), Phase 2 (0.9137). In Phase 0 and Phase 3, the AUCs were slightly lower (0.8839, 0.8776, respectively), reflecting the relatively weaker yet still reliable discriminatory ability of the ISA.

Figure 6
Panel A shows ROC curves for various phases, indicating the true positive rate against the false positive rate. Panel B and C display precision-recall curves for different models. Legends differentiate by AUC values and IoU thresholds.

Figure 6. Evaluation of model classification performance. (A) ROC curves for the classification of all surgical phases (One-vs-Rest). The high AUC values for each phase demonstrate the model’s robust discriminatory power. (B) PR curves comparing our model’s performance against several baseline methods, showing that our model consistently achieves higher precision across varying recall levels. (C) PR curves for our model evaluated at increasingly stringent IoU thresholds (from 0.5 to 0.95), demonstrating robust performance even at high IoU values and reflecting its strong spatial localization capabilities.

Notably, the ISA achieved AUC values above 0.87 across all phases of the procedure, indicating a robust and consistent discriminative capability for key frame identification in surgical videos irrespective of stage. Moreover, the performances in recognizing static background and dynamic key phases showed a notable disparity, suggesting that a higher degree of visual complexity (i.e., richer visual information) could enable a more accurate recognition, as shown by that in active surgical scenes, the ISA showed a stronger performance in reading frames. Overall, the consistently high AUC values at each phase demonstrated the model’s stable discriminatory ability.

3.5 Performances of the model

As shown in Table 3, our method achieves the best performance across most evaluation metrics. Specifically, it obtains the highest values in AP50 (95.2%), AP75 (65.4%), and AP50:95 (62.1%), outperforming the second-best method SurgeNet by 2–4 percentage points. It also ranks first in APL (64.8%), indicating better performance in detecting large anatomical structures. While SurgeNet achieves a slightly higher APM (55.6%), our method remains competitive at 54.3%, demonstrating consistent performance across different object scales.

Regarding efficiency, our model achieves a frame rate of 19.2 FPS, significantly faster than other methods such as TransUNet (15.3 FPS) and DeepLabv3+ (10.1 FPS). This indicates that our method is not only accurate but also practical for real-time applications in clinical scenarios.

The PR performance of our model is detailed in Figures 6B, C. Figure 6B compares our model to several baselines, demonstrating that our method consistently maintains higher precision across varying recall levels. Furthermore, as shown in Figure 6C, our model maintains robust performance even under stricter IoU thresholds (e.g., 0.85 and 0.9), reflecting its strong spatial localization capabilities.

In summary, the results validate the effectiveness and robustness of our method in terms of both detection accuracy and inference speed, highlighting its potential for real-world medical image analysis tasks.

3.6 Temporal phase prediction across the full surgical timeline

We further evaluated the model’s temporal prediction performance over the entire course of laparoscopic liver resection, dividing the procedure into five sequential phases and comparing model outputs to expert-annotated ground truth. Without post-processing, the model was able to reproduce the general phase order, though minor misclassifications occurred at transitional boundaries—particularly between the second hepatic portal anatomy and portal vein dissection stages. After applying temporal smoothing and transition constraints, the predicted phase sequence exhibited improved continuity, reduced fragmentation, and better alignment with surgical annotations. In low-motion frames such as post-resection hemostasis observation, the model maintained stable predictions, indicating reliable temporal awareness and rhythmic phase modeling even in visually ambiguous intervals.

3.7 Performance on the independent external validation cohort

To rigorously assess the model’s generalizability, we evaluated its performance on a completely independent external validation cohort, which consisted of 122 surgical videos from a center whose data was not used for training. The ISA was applied to this unseen dataset without any retraining or fine-tuning.

On this external cohort, the model demonstrated strong and consistent performance, achieving an average phase recognition accuracy of 89.5%, which is comparable to the 91% accuracy observed in the internal cross-validation. Key performance metrics, including precision, recall, and F1-score, also remained robust, confirming that the model did not overfit to the training data and can generalize effectively to different surgical teams and environments. The detailed performance on the external cohort is summarized in Table 6.

Table 6
www.frontiersin.org

Table 6. Key performance metrics on the external validation cohort (N=122).

4 Discussion

Our development and multi-center validation of the ISA system directly addresses several key challenges recently highlighted in the surgical AI literature. While many studies have focused on single-task excellence, our approach emphasizes a multi-task framework that maintains real-time performance, a critical requirement for clinical adoption (20). Furthermore, by creating a large, multi-center dataset and rigorously validating our model on an independent external cohort, we contribute to solving the issues of data scarcity and model generalizability that are frequently cited as major hurdles in the field (21). Our work therefore represents a significant step toward translating AI from a research concept into a clinically valuable tool, as envisioned by recent reviews (24).

Compared with previous single-task AI systems used in laparoscopic surgery, such as those focused solely on tool tracking or static segmentation, the ISA achieves a comprehensive integration of intraoperative visual information (21). The average classification accuracy exceeded 89% across key phases, with AUCs consistently above 0.87. Notably, our approach outperformed SurgeNet, TransUNet, and EndoViT in both segmentation accuracy (AP50: 95.2%) and frame rate (19.2 FPS), providing not only precision but also practical operability in real surgical environments. These metrics collectively support the reliability of ISA as a real-time clinical decision support tool (2224).

From an oncological perspective, achieving precise anatomical exposure and reliable intraoperative phase control is essential in liver cancer resection. The ISA’s ability to evaluate phase-specific image clarity and detect critical procedural transitions (e.g., hilum dissection and hemostasis) may directly contribute to complete tumor excision and reduced intraoperative complications. By alerting the surgeon in real time when visual clarity is compromised, the system is designed to mitigate the risks of transecting tissue with inadequate visualization. Whether this function ultimately translates into a reduced incidence of residual tumor warrants investigation in future prospective studies. Although our study did not evaluate long-term oncologic outcomes, the integration of ISA into hepatobiliary workflows may ultimately translate into reduced margin positivity and improved surgical radicality, warranting future investigation.

Despite the promising results, this study has several limitations. The primary limitation is that our validation is confined to technical metrics of accuracy and speed, rather than clinical endpoints. While our system’s ability to accurately identify surgical phases and assess image clarity suggests a strong potential for improving safety, we did not measure its direct impact on outcomes such as operative time, blood loss, or complication rates (24). Therefore, the clinical benefits of the ISA remain a well-founded hypothesis that requires rigorous validation in future prospective, randomized controlled trials. Second, although the dataset is relatively large and multi-institutional, it may not fully capture the heterogeneity of all intraoperative environments, especially in complex tumor resections involving vascular invasion or cirrhotic livers. Third, the current ISA system relies exclusively on endoscopic video input; incorporation of multimodal data such as intraoperative ultrasound or fluorescence imaging may further enhance decision-making accuracy (25, 26).

In conclusion, the proposed ISA demonstrates high accuracy, robustness, and real-time responsiveness in phase-specific analysis during laparoscopic liver surgery. Preliminary feedback from participating surgeons suggests that the system enhances intraoperative decision-making, particularly by clarifying critical transitions such as hemostasis and hepatic hilum dissection. This study exemplifies how AI can bridge the gap between real-time endoscopic imaging and surgical decision-making, supporting procedural consistency and situational awareness.

However, it is important to acknowledge the model’s potential limitations and “failure modes,” particularly in challenging clinical scenarios. As the system relies on visual input, its performance could be compromised by severe intraoperative bleeding that completely obscures the camera, extensive adhesions from reoperations that alter typical anatomy, or rare anatomical variations not well-represented in the training data. Addressing these challenges will be a key direction for future model improvements and is essential for ensuring the system’s reliability in the full spectrum of surgical situations.

Despite these considerations, with continued optimization and integration into clinical workflows, the ISA holds strong potential to improve intraoperative safety and to standardize surgical procedures—particularly in oncologic contexts where precision and margin control are critical. Future prospective trials are warranted to evaluate its clinical impact on operative time, complication rates, and long-term oncologic outcomes. Ultimately, the intelligent vision systems demonstrated by ISA could serve as a foundational component for future integrated platforms that provide intelligent intraoperative navigation and quality control in minimally invasive oncologic surgery.

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Ethics statement

The studies involving humans were approved by Ethics Committee of The First Affiliated Hospital of Xi’an Jiaotong University. The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their written informed consent to participate in this study. Written informed consent was obtained from the individual(s) for the publication of any potentially identifiable images or data included in this article.

Author contributions

ZP: Data curation, Formal Analysis, Writing – original draft, Writing – review & editing. ZW: Conceptualization, Methodology, Project administration, Writing – review & editing. YY: Formal Analysis, Investigation, Resources, Writing – review & editing. HP: Software, Supervision, Validation, Writing – review & editing. YM: Conceptualization, Software, Visualization, Writing – review & editing. YL: Conceptualization, Data curation, Writing – review & editing. YR: Conceptualization, Software, Writing – review & editing. JX: Funding acquisition, Methodology, Project administration, Writing – review & editing. KG: Investigation, Methodology, Project administration, Writing – review & editing. GW: Project administration, Supervision, Writing – review & editing. JD: Project administration, Resources, Supervision, Writing – review & editing. X-WL: Investigation, Project administration, Supervision, Writing – review & editing. YG: Data curation, Formal Analysis, Project administration, Writing – review & editing. X-ML: Funding acquisition, Project administration, Resources, Writing – review & editing. RW: Data curation, Funding acquisition, Project administration, Resources, Supervision, Writing – review & editing. YL: Formal Analysis, Funding acquisition, Supervision, Writing – review & editing. LY: Funding acquisition, Investigation, Project administration, Supervision, Writing – review & editing.

Funding

The author(s) declare financial support was received for the research and/or publication of this article. This experiment was supported by Major Research Plan of the National Natural Science Foundation of China, grant No. 92048202 (Referred to Yi Lyu); National Natural Science Foundation of China, grant No. 82471190 (Referred to Xue-Min Liu); School Enterprise Joint Project of Xi’an Jiaotong University, grant No. HX202439 and HX202197 (Referred to Rong-Qian Wu); Noncommunicable Chronic Diseases-National Science and Technology Major Project of China, grant No. 2023ZD0502004 (Referred to Jun-Xi Xiang); Free Exploration and Innovation Project of the Basic Scientific Research Fund of Xi’an Jiaotong University, grant No. xzy022023069 and No. xzy022024026 (Referred to Zi-Yang Peng and Zhi-Bo Wang).

Acknowledgments

We do appreciate the National Local Joint Engineering Research Center for Precision Surgery & Regenerative Medicine, Xi’an Jiaotong University, for its invaluable support and resources. Authors acknowledge Professor Yi Lyu for providing a platform to realize our research dreams, and authors appreciate the support from the School of Future Technology, Xi’an Jiaotong University.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declare that no Generative AI was used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fonc.2025.1678525/full#supplementary-material

References

1. Dezzani EO. Minimally invasive surgery: an overview. Minerva Surg. (2023) 78:616–25. doi: 10.23736/S2724-5691.23.10126-2

PubMed Abstract | Crossref Full Text | Google Scholar

2. Jastaniah A and Grushka J. The role of minimally invasive surgeries in trauma. Surg Clin North Am. (2024) 104:437–49. doi: 10.1016/j.suc.2023.10.003

PubMed Abstract | Crossref Full Text | Google Scholar

3. Saeidi H, Opfermann JD, Kam M, Wei S, Leonard S, Hsieh MH, et al. Autonomous robotic laparoscopic surgery for intestinal anastomosis. Sci Robot. (2022) 7:eabj2908. doi: 10.1126/scirobotics.abj2908

PubMed Abstract | Crossref Full Text | Google Scholar

4. Marino F, Moretto S, Rossi F, Pio Bizzarri F, Gandi C, Filomena GB, et al. Robot-assisted radical prostatectomy with the Hugo RAS and da Vinci surgical robotic systems: A systematic review and meta-analysis of comparative studies. Eur Urol Focus. (2024) 24:S2405-4569. doi: 10.1016/j.euf.2024.10.005

PubMed Abstract | Crossref Full Text | Google Scholar

5. Veldkamp R, Kuhry E, Hop WC, Jeekel J, Kazemier G, Bonjer HJ, et al. Laparoscopic surgery versus open surgery for colon cancer: short-term outcomes of a randomised trial. Lancet Oncol. (2005) 6:477–84. doi: 10.1016/S1470-2045(05)70221-7

PubMed Abstract | Crossref Full Text | Google Scholar

6. Cianchi F. Robotics in general surgery: a promising evolution. Minerva Surg. (2021) 76:103–4. doi: 10.23736/S2724-5691.21.08764-2

PubMed Abstract | Crossref Full Text | Google Scholar

7. Troisi RI, Patriti A, Montalti R, and Casciola L. Robot assistance in liver surgery: a real advantage over a fully laparoscopic approach? Results of a comparative bi-institutional analysis. Int J Med Robot. (2013) 9:160–6. doi: 10.1002/rcs.1495

PubMed Abstract | Crossref Full Text | Google Scholar

8. Ficarra V, Novara G, Rosen RC, Artibani W, Carroll PR, Costello A, et al. Systematic review and meta-analysis of studies reporting urinary continence recovery after robot-assisted radical prostatectomy. Eur Urol. (2012) 62:405–17. doi: 10.1016/j.eururo.2012.05.045

PubMed Abstract | Crossref Full Text | Google Scholar

9. Chen IY, Joshi S, and Ghassemi M. Treating health disparities with artificial intelligence. Nat Med. (2020) 26:16–7. doi: 10.1038/s41591-019-0649-2

PubMed Abstract | Crossref Full Text | Google Scholar

10. Hanna GB, Shimi SM, and Cuschieri A. Task performance in endoscopic surgery is influenced by location of the image display. Ann Surg. (1998) 227:481–4. doi: 10.1097/00000658-199804000-00005

PubMed Abstract | Crossref Full Text | Google Scholar

11. Kitaguchi D, Takeshita N, Matsuzaki H, Takano H, Owada Y, Enomoto T, et al. Real-time automatic surgical phase recognition in laparoscopic sigmoidectomy using the convolutional neural network-based deep learning approach. Surg Endosc. (2020) 34:4924–31. doi: 10.1007/s00464-019-07281-0

PubMed Abstract | Crossref Full Text | Google Scholar

12. Han HS, Shehta A, Ahn S, Yoon YS, Cho JY, and Choi Y. Laparoscopic versus open liver resection for hepatocellular carcinoma: Case-matched study with propensity score matching. J Hepatol. (2015) 63:643–50. doi: 10.1016/j.jhep.2015.04.005

PubMed Abstract | Crossref Full Text | Google Scholar

13. Yoon YI, Kim KH, Kang SH, Kim WJ, Shin MH, Lee SK, et al. Pure laparoscopic versus open right hepatectomy for hepatocellular carcinoma in patients with cirrhosis: A propensity score matched analysis. Ann Surg. (2017) 265:856–63. doi: 10.1097/SLA.0000000000002072

PubMed Abstract | Crossref Full Text | Google Scholar

14. McHugh DJ, Gleeson JP, and Feldman DR. Testicular cancer in 2023: Current status and recent progress. CA Cancer J Clin. (2024) 74:167–86. doi: 10.3322/caac.21819

PubMed Abstract | Crossref Full Text | Google Scholar

15. Halls MC, Cipriani F, Berardi G, Barkhatov L, Lainas P, Alzoubi M, et al. Conversion for unfavorable intraoperative events results in significantly worse outcomes during laparoscopic liver resection: lessons learned from a multicenter review of 2861 cases. Ann Surg. (2018) 268:1051–7. doi: 10.1097/SLA.0000000000002332

PubMed Abstract | Crossref Full Text | Google Scholar

16. Lam K, Chen J, Wang Z, Iqbal FM, Darzi A, Lo B, et al. Machine learning for technical skill assessment in surgery: a systematic review. NPJ Digit Med. (2022) 5:24. doi: 10.1038/s41746-022-00566-0

PubMed Abstract | Crossref Full Text | Google Scholar

17. Mascagni P, Vardazaryan A, Alapatt D, Urade T, Emre T, Fiorillo C, et al. Artificial intelligence for surgical safety: automatic assessment of the critical view of safety in laparoscopic cholecystectomy using deep learning. Ann Surg. (2022) 275:955–61. doi: 10.1097/SLA.0000000000004351

PubMed Abstract | Crossref Full Text | Google Scholar

18. Luongo F, Hakim R, Nguyen JH, Anandkumar A, and Hung AJ. Deep learning-based computer vision to recognize and classify suturing gestures in robot-assisted surgery. Surgery. (2021) 169:1240–4. doi: 10.1016/j.surg.2020.08.016

PubMed Abstract | Crossref Full Text | Google Scholar

19. Chen Z, An J, Wu S, Cheng K, You J, Liu J, et al. Surgesture: a novel instrument based on surgical actions for objective skill assessment. Surg Endosc. (2022) 36:6113–21. doi: 10.1007/s00464-022-09108-x

PubMed Abstract | Crossref Full Text | Google Scholar

20. Kenig N, Monton Echeverria J, and Muntaner Vives A. Artificial intelligence in surgery: A systematic review of use and validation. J Clin Med. (2024) 13:7108. doi: 10.3390/jcm13237108

PubMed Abstract | Crossref Full Text | Google Scholar

21. Tasci B, Dogan S, and Tuncer T. Artificial intelligence in gastrointestinal surgery: A systematic review. World J Gastrointest Surg. (2025) 17:109463. doi: 10.4240/wjgs.v17.i8.109463

PubMed Abstract | Crossref Full Text | Google Scholar

22. Madani A, Namazi B, Altieri MS, Hashimoto DA, Rivera AM, Pucher PH, et al. Artificial intelligence for intraoperative guidance: using semantic segmentation to identify surgical anatomy during laparoscopic cholecystectomy. Ann Surg. (2022) 276:363–9. doi: 10.1097/SLA.0000000000004594

PubMed Abstract | Crossref Full Text | Google Scholar

23. Anteby R, Horesh N, Soffer S, Zager Y, Barash Y, Amiel I, et al. Deep learning visual analysis in laparoscopic surgery: a systematic review and diagnostic test accuracy meta-analysis. Surg Endosc. (2021) 35:1521–33. doi: 10.1007/s00464-020-08168-1

PubMed Abstract | Crossref Full Text | Google Scholar

24. Mascagni P, Alapatt D, Sestini L, Altieri MS, Madani A, Watanabe Y, et al. Computer vision in surgery: from potential to clinical value. NPJ Digit Med. (2022) 5:163. doi: 10.1038/s41746-022-00707-5

PubMed Abstract | Crossref Full Text | Google Scholar

25. Tel A, Murta F, Sembronio S, Costa F, and Robiony M. Virtual planning and navigation for targeted excision of intraorbital space-occupying lesions: proposal of a computer-guided protocol. Int J Oral Maxillofac Surg. (2022) 51:269–78. doi: 10.1016/j.ijom.2021.07.013

PubMed Abstract | Crossref Full Text | Google Scholar

26. Sun P, Zhao Y, Men J, Ma ZR, Jiang HZ, Liu CY, et al. Application of virtual and augmented reality technology in hip surgery: systematic review. J Med Internet Res. (2023) 25:e37599. doi: 10.2196/37599

PubMed Abstract | Crossref Full Text | Google Scholar

27. Jaspers TJM, de Jong RLPD, Al Khalil Y, Mishra A, Mongan J, Huang CW, et al. Exploring the effect of dataset diversity in self-supervised learning for surgical computer vision[C]//MICCAI Workshop on Data Engineering in Medical Imaging. Cham: Springer Nature Switzerland (2024) p. 43–53.

Google Scholar

28. Chen J, Lu Y, Yu Q, et al. arXiv preprint arXiv:2102.04306, 2021.

Google Scholar

29. Zhang J, Li X, Wang Y, Li K, Wang Z, Liu W, et al. Eatformer: Improving vision transformer inspired by evolutionary algorithm. Int J Comput Vision. (2024) 132:3509–36. doi: 10.1007/s11263-024-02034-6

Crossref Full Text | Google Scholar

30. Batić D, Holm F, Özsoy E, Kader A, Daum S, Nolte A, et al. EndoViT: pretraining vision transformers on a large collection of endoscopic images. Int J Comput Assisted Radiol Surg. (2024) 19:1085–91. doi: 10.1007/s11548-024-03091-5

PubMed Abstract | Crossref Full Text | Google Scholar

31. Chen LC, Zhu Y, Papandreou G, Schroff F, and Adam H. (2018). Encoder-decoder with atrous separable convolution for semantic image segmentation, in: Proceedings of the European conference on computer vision (ECCV), . pp. 801–18.

Google Scholar

Keywords: digital surgery, AI assistance, intraoperative quality control, surgical decision support, real-time safety evaluation

Citation: Peng Z-Y, Wang Z-B, Yan Y, Peng H-Q, Ma Y-T, Li Y-T, Ren Y-X, Xiang J-X, Guo K, Wang G, Duan J-F, Li X-W, Guan Y, Liu X-M, Wu R-Q, Lyu Y and Yu L (2025) Development of an AI-driven digital assistance system for real-time safety evaluation and quality control in laparoscopic liver surgery. Front. Oncol. 15:1678525. doi: 10.3389/fonc.2025.1678525

Received: 04 August 2025; Accepted: 23 September 2025;
Published: 08 October 2025.

Edited by:

Chen Liu, Army Medical University, China

Reviewed by:

Jun Liu, Dongguan Hospital of Guangzhou University of Chinese Medicine, China
Cao Yinghao, National University of Singapore, Singapore

Copyright © 2025 Peng, Wang, Yan, Peng, Ma, Li, Ren, Xiang, Guo, Wang, Duan, Li, Guan, Liu, Wu, Lyu and Yu. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Yi Lyu, bHV5aTE2OUAxMjYuY29t; Li Yu, bGl5dS04MjAyMTlAMTYzLmNvbQ==

†These authors share first authorship

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.