Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Surg., 02 February 2026

Sec. Surgical Oncology

Volume 13 - 2026 | https://doi.org/10.3389/fsurg.2026.1711392

This article is part of the Research TopicIntegrating Real-time Diagnostics in Laparoscopic Surgery for Enhanced Precision in Tumor ResectionView all articles

Construction and validation of a high-precision annotated dataset for developing intelligent critical vein recognition models in laparoscopic pancreatic surgery


Hu Zhou,,&#x;Hu Zhou1,2,†Lu Ping,&#x;Lu Ping1,†Ruohan CuiRuohan Cui2Junyi GaoJunyi Gao1Xianlin HanXianlin Han1Wenming Wu
Wenming Wu1*Surong Hua

Surong Hua3*
  • 1Department of General Surgery, State Key Laboratory of Complex, Severe, and Rare Diseases, Peking Union Medical College Hospital, Chinese Academy of Medical Sciences & Peking Union Medical College, Beijing, China
  • 2Chinese Academy of Medical Sciences & Peking Union Medical College, Beijing, China
  • 3Beijing United Family Hospital, Beijing, China

Background: Laparoscopic operation holds multiple advantages as a minimal invasive method of surgical treatment. Vascular-related manipulations, including identification and dissection of vascular structures and control of bleeding, are highly experience-based and demand a tortuous learning curve. With the rapid development of artificial intelligence (AI) in the entire diagnosis and treatment process of diseases, data-driven AI models have shown promising potentials in both education and real-time aiding in surgery. However, there is no dedicated dataset existing for developing vascular identification models in laparoscopic settings.

Methods: Videos from 23 laparoscopic distal pancreatectomy (LDP) and laparoscopic pancreaticoduodenectomy (LPD) performed at Peking Union Medical College Hospital (PUMCH) between January 2021 and June 2022 were collected. Senior surgeons systematically reviewed surgical videos to visually identify critical venous vasculature, precisely annotating frame-accurate start and end timestamps on the video timeline. Frames were extracted from these video segments at a fixed temporal interval of one frame per second, then stored to compile the source image dataset. The contours of superior mesenteric vein (SMV), portal vein (PV), splenic vein (SV) were labelled and reviewed according to standard procedure. A High-Resolution Network (HRNet) was combined with a fully convolutional network (FCN) output head to construct a preliminary segmentation model for initial validation and comparison.

Results: A dataset comprises 19,003 annotated frames and is publicly available. The baseline model achieved a recall of 79.6%, precision of 95.8%, and Dice coefficient of 0.69 on the testing set.

Conclusion: This study constructed and released the first large-scale, expert-annotated dataset of key venous structures from pancreatic surgery (PS) videos and established benchmark performance for intraoperative vein segmentation using open-source models. This resource provides a foundation for advancing AI-assisted vascular identification in laparoscopic surgery.

Introduction

The minimally invasive approach of laparoscopic surgery offers clear advantages over open procedures in the management of pancreatic diseases, including reduced surgical trauma and attenuated systemic reaction (1). However, the pancreas is surrounded by a dense network of arteries and veins; thus, tumor invasion into single or multiple blood vessels makes vascular manipulation a primary source of surgical risk (2, 3). While arterial anatomy, such as the gastroduodenal artery, is routinely emphasized (4), venous structures also require equal attention as they serve both as high-risk zones of bleeding and as indispensable anatomical landmarks. During both laparoscopic pancreaticoduodenectomy (LPD) and distal pancreatectomy (LDP), the superior mesenteric vein (SMV), portal vein (PV), and splenic vein (SV) require meticulous handling due to their thin, fragile walls and critical role as anatomical landmarks (5). Their exposure demands precision, as inadvertent injury to these structures may precipitate uncontrollable massive hemorrhage. The infiltration of the SMV-PV axis is one of the key indicators for assessing the surgical resectability of pancreatic cancer (6, 7), as well as tumors in the distal common bile duct, the duodenum and the area surrounding the ampulla of Vater. For resectable and borderline resectable pancreatic cancers, venous involvement often necessitates combined vein resection and reconstruction, substantially increasing surgical complexity (810). Improper vascular handling may result in severe postoperative complications, including pancreatic fistula, hemorrhage, gastroparesis, and portal vein thrombosis (1113), underscoring the need for precise vascular management.

In laparoscopic pancreatic surgery (PS), visual identification of vessels is the sole means of anatomical recognition due to the lack of tactile feedback (14, 15). However, visual clarity is frequently compromised by intraoperative bleeding, smoke from electrocautery and cutting, and rapid camera movements, hindering accurate anatomical recognition (16, 17). Additionally, less-experienced surgeons may face elevated stress and reduced performance under such conditions (18, 19), further jeopardizing procedural safety and patient outcomes. These challenges underscore the urgent need for intraoperative assistance systems to enhance anatomical visualization and surgical precision.

Data-driven deep learning (DL) image recognition technology holds promise for mitigating challenges arising from the limitations of human vision and surgical inexperience. In diagnosis, DL systems such as the PANDA model have achieved expert-level performance in detecting subclinical pancreatic lesions via non-contrast abdominal CT scans (20). This model leverages a prospective cohort study of 3,000 subjects with standardized CT imaging, reflecting both significant scientific rigor and clinical value. For preoperative resectability assessment of PS, numerous researchers utilize imaging data for semantic segmentation and three-dimensional reconstruction to simulate anatomical relationships between pancreatic lesions and surrounding structures (2123). This enables precise identification and localization of critical anatomy during surgery, providing visual decision support to enhance operative precision. However, as far as the surgical operation itself is concerned, there is an extreme lack of studies on intraoperative anatomical identification in PS. Beyond our team's publication of the first study on intraoperative vascular identification in pancreatic procedures (24), no other reports exist. This stands in remarkable contrast to the explosive growth in laparoscopic cholecystectomy (LC) research, where systematic breakthroughs have been achieved across multiple dimensions: anatomical structure recognition (25), surgical phase identification (26, 27), and critical view of safety (CVS) evaluation (28, 29). These advances are largely attributable to the release of Cholec80 (30), the largest sample size of annotated intraoperative LC image dataset at that time. Therefore, to accelerate intelligent research on PS, the creation of high-quality annotated intraoperative datasets has emerged as a critical bottleneck requiring urgent resolution.

To address this gap, we present VIP20K, the first expert-annotated dataset (20,000 images for Vein Identification in Pancreatic surgery) focused on key venous structures in laparoscopic PS. VIP20K concentrates on the need for real-time intraoperative identification of high-risk vessels such as the SMV-PV axis and the SV, and provides a foundation for advancing intraoperative vascular recognition. We further establish baseline semantic segmentation performance using models constructed entirely from publicly available components. This dataset is expected to facilitate the development of advanced algorithms for real-time, precise venous identification, ultimately reducing operative risk, lowering surgeon cognitive burden, and improving clinical outcomes in pancreatic surgery.

Materials and methods

Our dataset construction followed a five-stage workflow (Figure 1): data collection using standard laparoscopy equipment at Peking Union Medical College Hospital (PUMCH), data anonymization, target video segments selection, frames extraction, and key veins annotation and expert review. This methodical framework facilitates the creation of a standardized laparoscopic PS video repository with precise annotations and rigorous quality control.

Figure 1
Medical imagery workflow diagram showing four main processes: data collection in an operating room with two surgeons and surgery videos; data anonymization removing personal information from images; annotation and expert review with visual annotations on surgery images and an approval interface; frames extraction capturing images per second from videos. Target video segments focus on key veins in surgical footage. Labels clarify each step.

Figure 1. Overview of data collection process and methodology in the research.

Surgical procedure

All operations were carried out in accordance with established minimally invasive techniques. The key steps are described as follows. In LPD, the gastrocolic ligament was divided, and the hepatic flexure was mobilized, followed by Kocher maneuver for duodeno-pancreatic mobilization. Vascular dissection of the hepatoduodenal ligament and gastroduodenal artery preceded inferior-border dissection of the pancreas, during which the SMV and PV became clearly visible. A tunnel was created between the pancreatic neck and SMV/PV before pancreatic-neck transection and completion of resection and reconstruction (31). In LDP, access to the lesser sac was achieved by dividing the gastrocolic ligament, followed by mobilization of the transverse colon and splenic flexure to expose the pancreatic body–tail. A tunnel was created between the pancreatic neck and the SMV and PV, and then the splenic vein and artery were dissected, and the pancreas was transected at the planned line using a linear stapler (32).

Laparoscopic pancreatic surgery video acquisition

We retrospectively collected intraoperative videos from laparoscopic PS performed at PUMCH between January 2021 and June 2022. All procedures were performed by senior surgeons (each with >100 documented Whipple procedures) utilizing KANGJI® laparoscopic endoscopes capable of 4K imaging (3,840 × 2,160 pixels). Real-time recording was achieved via KJ.FPAK-02 recording hosts for intraoperative monitoring and archival. Video capture commenced upon placing laparoscopy into abdominal cavity through trocar and continued uninterrupted until removal of trocar symbolizing the end of the surgery, ensuring comprehensive procedural documentation. Video inclusion criteria: (a) Complete procedural record of LPD or LDP, (b) Continuous visualization of the SMV-PV axis for ≥2 min. Video exclusion criteria: (a) Intraoperative conversion to open surgery, (b) Prior abdominal or pelvic surgical interventions, (c) Confirmed vascular invasion or borderline resectable pancreatic cancer. After screening according to the standards, 23 surgeries were included, including 15 cases of LDP and 8 cases of LPD. The demographics and tumor details were illustrated in Table 1. All patient inclusion required both clinical indications for surgery and written informed consent. This study received ethical approval from the Ethics Committee of PUMCH (Approval NO. I-25PJ0451 for collection of surgical videos in the institution).

Table 1
www.frontiersin.org

Table 1. Demographics and tumor characteristics of the patients.

Data anonymization

Our core team members conducted comprehensive reviews of all surgical videos to identify and remove personally identifiable information of both physicians and patients. The processed videos were systematically renamed according to standardized protocols, and sidebars containing temporal data, device specifications, physician names, and patient medical record numbers were excised. Additionally, we deleted operationally irrelevant footage recorded when the laparoscope was temporarily withdrawn from the abdominal cavity for lens cleaning. These measures collectively ensured full data anonymization in accordance with general data protection regulation (GDPR) requirements.

Target video segments selection and frames extraction

Senior surgeons systematically reviewed surgical videos to visually identify critical venous vasculature, precisely annotating frame-accurate start and end timestamps on the video timeline. Surgical scenes such as obvious bloodstains covering, smoke, and shaky camera movements were all included to enhance the richness of the video scenes. We then employed video processing algorithms to automatically extract all segments containing target anatomical structures based on this timestamp metadata, thus establishing an initial surgical video library for subsequent analysis and reference. Subsequently, frames were systematically extracted from these video segments at a fixed temporal interval of one frame per second, then stored as JPEG files with a resolution of 1,920 × 1,080 pixels to compile the source image dataset.

Annotation and expert review

Following image dataset generation, four uniformly trained junior surgeons utilized our team's dedicated annotation software (supporting cross-device pixel-level annotation on tablets/desktops/laptops) to delineate contours of the SMV, PV, and SV based on standardized protocols. Key annotation criteria included (Figure 2): (a) Instrument obstruction handling: if the target vein is partially or completely obscured by surgical instruments, retraction cords, catheters, or other objects, trace along the edge of the obstruction during annotation, bypassing the obstructed area to reflect the actual visible portion of the vein; if the obstruction is a suture, disregard the suture and directly trace the contour of the vein beneath it. (b) Minimum annotation size requirement: target veins with area of minimum bounding rectangle box smaller than 64 × 64 pixels in the image are deemed ineligible for annotation and should be excluded. (c) Exclusion criteria for image quality: images exhibiting motion blur, fog interference, severe glare, or other artifacts affecting interpretation must be skipped during marking; images where the target vein is not within the field of view are also excluded. (d) Handling fascia and adipose tissue: direct contouring for thin fascial/adipose coverage over veins, whereas thick coverage required outlining along its periphery and bypassing the obscured area. (e) Blood interference processing: direct annotation for thin blood layers which permits venous wall visualization, exclusion for thick blood obscuring veins or obvious bloodstain at vessel margins. (f) Temporal consistency: for continuous frame sequences, annotations must maintain strict consistency across frames to prevent contour discrepancies of the same structure in different frames. All annotations underwent independent binary assessment (Yes/No) by two senior surgeons who completed over 100 Whipple procedures evaluating anatomical accuracy and protocol compliance. Annotations receiving unanimous approval were accepted; conflicting or rejected cases required re-annotation under senior guidance until consensus was achieved, developing a high-precision annotated image dataset.

Figure 2
Examples of the critical vein annotations for different intraoperative scenarios. Images A to H show the annotated images for various scenarios, highlighted by blue outlines marking vein areas of interest.

Figure 2. Examples of the critical vein annotations for different intraoperative scenarios. The blue polygon delineates the contour of the target vein. (A) Trace along the edge of the instruments during annotation, and bypass the obstructed area. (B) Disregard the suture and directly trace the contour of the vein beneath it. (C) Minimum annotation size requirement, exclude targets with minimum bounding rectangle ≤64 × 64 pixels. (D) Exclude images where the target veins are not clearly visible. (E) Outline the contours of the veins covered by thin fascial/adipose coverage. (F) Direct contour delineation of venous structures through thin blood layers, skip the instrument and mark each section separately. (G) Bypass superficial surgical instruments and thick fascial/adipose tissues that has not yet been stripped. (H) Annotations must exclude thick tissues and blood residues at marginal regions.

Technical validation

In this semantic segmentation task targeting surgical scenarios, we employed High-Resolution Network (HRNet) as the backbone network, combined with a fully convolutional network (FCN) output head to construct the segmentation model (Figure 3). The adopted HRNet-W18-C variant is distinguished by its four-stage progressive multi-resolution architecture, which maintains high-resolution representations throughout the network (33). In the initial stage of the network, only the high-resolution branch was included. The original input image was convolved to a resolution of 1/4 scale and then input into HRNet. In this stage, the number of channels was 18, and residual convolution was used to extract basic spatial features. Subsequent stages progressively integrated low-resolution branches at 1/2, 1/4, and 1/8 resolutions, with channel counts doubling to 36, 72, and 144, respectively. Each branch utilized cross-resolution exchange units to enable feature interaction. High-resolution branches incorporated upsampled semantic information from low-resolution branches to enhance contextual awareness. Low-resolution branches integrated downsampled detail features from high-resolution branches to improve spatial sensitivity. This bidirectional information exchange, iteratively reinforced across residual units, established a hierarchical feature enhancement mechanism. By preserving high-resolution features and enabling dynamic multi-scale fusion, HRNet mitigated spatial information loss caused by repeated downsampling, particularly advantageous for precise localization of fine structures in surgical imaging such as vessels.

Figure 3
Diagram illustrating a 4-stage convolutional neural network (CNN) process. It starts with an input image, proceeds through stages involving feature maps of various sizes, and involves operations like copy, propagation, preprocessing, upsampling, and stride convolution. The output channels are concatenated and passed through a fully convolutional network (FCN) producing a segmented output. A legend explains the color coding and symbols for different operations.

Figure 3. Overview of HRNet-FCN segmentation model.

To adapt HRNet for semantic segmentation, we replaced its classification output head with an FCN output layer. The core principle of FCN lies in fully convolutional operations and end-to-end pixel-wise prediction (34). The FCN head employs deconvolution to refine segmentation boundaries, recovering spatial details lost during downsampling. The architecture we developed integrated high-resolution feature preservation with dynamic cross-scale fusion, leveraging FCN's pixel-level prediction capabilities to enable precise vein segmentation in complex surgical scenes.

Corresponding strategies were also taken in training configuration and optimization. The backbone weights of HRNet had been initialized through ImageNet pre-training to accelerate convergence. In data preprocessing, input images were resized to 960 × 540 while maintaining the original aspect ratio. From the resized images, 512 × 512 patches were randomly cropped and horizontally flipped with a 50% probability to increase spatial diversity. RGB channels were standardized with a mean of 123.68, 116.28, and 103.53, respectively, and a standard deviation of 57.12. No additional augmentation techniques were applied to preserve anatomical authenticity. The model was trained across 8 NVIDIA Tesla V100 GPUs with Python 3.6 and PyTorch 0.4.1. Training hyperparameters were as follows: a per-GPU batch size of 32, Stochastic Gradient Descent (SGD) Optimizer with momentum 0.9 and weight decay factor 0.0005, learning rate initialized at 0.01and decayed exponentially per epoch with a decay exponent of 0.9, and iterations for 100,000 epochs. To address severe pixel imbalance between the target veins and background, a weighted cross-entropy loss was employed:

L=1ni=1ncCwcyc,itruelog(yc,ipre)
wc=totalpixelsnpixelsinclassc=nnc

where n was total pixels per image, i was pixel index, c presented a class, yc,itrue was ground truth label, yc,ipre was predicted probability, wc was weight of class c, and nc was the pixel count of class c. In postprocessing, morphological refinement was employed via cv2.morphologyEx to eliminate edge artifacts and connect fragmented regions.

For comprehensive evaluation of the segmentation model, we employed three key metrics to rigorously quantify performance: recall (sensitivity), precision, and Dice coefficient. Recall quantifies a model's detection completeness by measuring the ratio of true positives to all actual positives, emphasizing minimization of false negatives. Recall evaluates the proportion of among. Precision assesses prediction reliability through the fraction of true positives among all predicted positives, thereby reducing false alarms. Dice coefficient quantifies spatial overlap between predictions and ground truth, offering robustness against class imbalance.

Dice=2*area(Prediction)area(GroundTruth)area(Prediction)+area(GroundTruth)

Results

Dataset architecture

The complete dataset is publicly available on the Kaggle platform and can be accessed via https://www.kaggle.com/datasets/prohuge/vip20k. The root directory of the dataset, named VIP20K, comprises two subdirectories: LPD-PUMCH and LDP-PUMCH, which respectively store imaging data for two distinct pancreatic surgical procedures, laparoscopic pancreaticoduodenectomy and laparoscopic distal pancreatectomy. This segregated storage design facilitates cross-technique comparative studies, where LPD-PUMCH contains data from 8 LPD procedures (Case 01–08) and LDP-PUMCH includes 15 procedures (Case 01–15) (Table 2). Each case data is organized into a three-tiered hierarchical structure:

Case-level: Folders named by case identifiers (e.g., LPD-PUMCH/Case 01/) store comprehensive data for individual surgeries.

Segment-level: Subfolders are created under the case folder based on the surgical procedure sequence (e.g., LPD-PUMCH/Case 01/segment_01/), with each segment representing a continuous operational unit.

Data-level: Segment folders contain three subdirectories:

3.1 Image/: Raw surgical scene images (JPEG format)

3.2 Mask/: Pixel-wise annotation masks (PNG format)

3.3 Merge/: Visual composite images (JPEG format) generated by overlaying masks on raw images.


Table 2
www.frontiersin.org

Table 2. Size of the dataset.

The file naming convention ensures strict spatial consistency across modalities, for example, within the directory LPD-PUMCH/Case 01/segment_01, the files Image/00050.jpg, Mask/00050_mask.png, and Merge/00050.jpg represent the same frame. The resolution is uniformly set to 1,920 × 1,080 pixels to eliminate scale discrepancies. This dataset is organized using a standardized hierarchical structure, aiming to provide high-quality anatomical annotation resources for computer vision-assisted surgery research.

Segmentation validation

In the validation of the HRNet-FCN model on the dataset, a 4:1 ratio was applied to partition the data into training and testing sets. To enhance the model's generalization capability across pancreatic surgical scenarios, data from two distinct procedures LDP and LPD were incorporated into the training phase. For the LDP cohort, 15 cases were included, with the training set comprising 6,417 frames from 14 patients and the testing set containing 1,849 frames from a single patient (Case 15). Similarly, in the LPD cohort with 8 cases, the training set consisted of 8,708 frames from 6 patients, while the testing set included 2,029 frames from 2 patients (Case 07 and Case 08). As summarized in Table 3, the model achieved a recall of 79.6%, precision of 95.8%, and Dice coefficient of 0.69 on the testing set, demonstrating robust practical utility. While false alarms were infrequent, false negatives or called miss rate particularly in veins detection warranted attention (Figure 4). Veins with a relatively large exposed area and clear surfaces were rarely missed, however, recognition failures occurred in scenes involving veins with small exposed area, blood occlusion, multiple segmented targets, or blurry scenes. These observations highlighted key areas for future optimization in venous segmentation under complex intraoperative conditions. The established performance metrics provide a comparative baseline for subsequent research in anatomical landmark identification during PS.

Table 3
www.frontiersin.org

Table 3. Validation results of dataset.

Figure 4
A series of nine surgical images labeled from A to I, showing organs with superimposed green and blue outlines. The green indicates ground truth, and the blue indicates predictions. Surgical tools are visible in each image.

Figure 4. Examples of inference results from the semantic segmentation model. IoU: Intersection over Union, indicates the degree of overlap between the prediction bounding box and the ground truth bounding box. (A) Correct recognition of the splenic vein (SV) during laparoscopic distal pancreatectomy (LDP) surgery. (B) Correct recognition of the portal vein (PV) and superior mesenteric vein (SMV) during laparoscopic pancreaticoduodenectomy (LPD). (C) Failure to identify a visible SV during LDP. (D) Omission of a small-target SV with misidentification of other structures as the target vein in LDP. (E) Missed detection of the SV surrounded by blood during LDP. (F) Incomplete segmentation of the SMV-PV axis, with two regions unannotated in multi-area labeling, during LPD. (G) PV omission due to heavy surface blood coverage during LPD. (H) Failure to detect a PV segment submerged in pooled blood during LPD. (I) Missed SMV-PV identification in an obscured surgical view with smoke and blood interference during LPD.

Discussion

This work introduced the first publicly available dataset specifically designed for intraoperative vascular recognition in PS, addressing a critical gap in surgical data science. Our dataset captures dynamic and diverse forms of the SMV, PV, and SV under realistic surgical conditions, including blood occlusion, instrument obstruction, and tissue deformation, which enabled model training for high-risk scenarios where vascular injury may lead to severe complications. By open-sourcing this resource, we encourage the community to engage in safety-critical computer vision applications.

The core value of this dataset stems from its complex pancreatic surgical scenes and annotation difficulty, necessitating close collaboration between surgeons and AI engineers to overcome intraoperative labeling challenges. Specifically, vessel visibility exhibits dynamic fluctuations due to interference factors such as bleeding, smoke interference, instrument movement, and changes in the surgical fields, requiring meticulous frame-by-frame review. Accurate delineation of vascular boundaries demands specialized surgical expertise to distinguish true vessel walls from artifacts, necessitating standardized training protocols for all annotators to ensure annotation consistency. The entire workflow, from video acquisition, annotation, to quality review, and dataset finalization, involves over 1,000 h of work, highlighting the tremendous effort required to develop a high-precision annotated surgical vision dataset.

The HRNet-FCN architecture demonstrates exceptional performance in vessel segmentation by leveraging its intrinsic high-resolution feature preservation mechanism. Through parallel multi-scale branches, HRNet maintains spatial accuracy at 1/4 input resolution, significantly mitigating information loss typically observed in encoder-decoder architectures. The model's robustness is further evidenced by its Dice coefficient of 0.69 in laparoscopic PS, enabling precise anatomical landmark identification while reducing risk of critical vein injury. We assembled a straightforward baseline model from open-source components; its performance on the proposed dataset can serve as a reference for other researchers to innovate in architecture design and algorithmic development, thereby advancing intraoperative vein recognition accuracy and ultimately enhancing clinical benefit for surgeons. This initiative aims to enhance intraoperative vein recognition accuracy, ultimately advancing clinical benefits for surgeons and patients through improved surgical safety and decision support.

Limitations remain that must guide future work in the research. The current dataset exhibits several limitations. First, the limited sample size of 23 patients, while consistent with the pilot nature of this work, may constrain the generalizability of the model to the broader population of patients undergoing PS. Second, all data were obtained from a single medical center. Although the use of standardized instruments and procedures ensures internal consistency, it may limit the model's applicability across diverse clinical settings and potentially introduce systematic biases. Third, while the dataset is the largest publicly available resource (20,000 frames) for identifying important veins in PS, the incidence of clinically pivotal scenarios, including massive hemorrhage and vascular anomalies, remains low. Thus, the model's robustness in identifying the rare but high-stakes events is constrained. Furthermore, reliance solely on intraoperative video neglects the guidance value of preoperative planning. This creates a fragmented decision-making workflow, requiring surgeons to manually cross-reference preoperative CT/MR vascular reconstructions across isolated systems, which compromises operational efficiency and spatial accuracy. Regarding model performance, the current framework demonstrates suboptimal detection sensitivity for veins with limited exposure and in blood-occluded scenarios. Enhancing recognition under these challenging conditions necessitates innovations in both deep learning architectures and multimodal fusion strategies.

In the future, there are some directions that can be explored. To enhance model generalizability, a cross-institutional federated learning system will integrate multi-center data using encrypted parameter exchange and feature alignment techniques, ensuring privacy-compliant collaboration across hospitals. Generative adversarial networks (GANs) will synthesize high-fidelity intraoperative hemorrhage and vascular anomaly samples to address class imbalance in rare scenarios. For operating-room deployment, neural architecture search (NAS) and quantization-aware training (QAT) will optimize models for embedded systems. A transformer-based cross-modal registration mechanism will align intraoperative video with preoperative CT/MR vascular models at a precision of millimeter level, adapting dynamically to tissue deformation. This enables an augmented reality navigation platform that overlays real-time segmentation and surgical pathways onto the surgical field, providing millimeter level guidance accuracy.

Conclusion

This study constructed and released the first large-scale annotated dataset of major veins sourced from PS videos. We validate its usability through baseline evaluations using open-source models, establishing benchmark metrics for intraoperative vein segmentation. This resource will empower practitioners to innovate algorithms and models, advancing the precision and efficiency of venous structure identification during surgery, ultimately translating into direct clinical benefits for patients and surgeons.

Data availability statement

The datasets presented in this study can be found in online repositories. The names of the repository/repositories and accession number(s) can be found below: www.kaggle.com/datasets/prohuge/vip20k.

Ethics statement

The studies involving humans were approved by Institutional Review Board of Peking Union Medical College Hospital. The studies were conducted in accordance with the local legislation and institutional requirements. Written informed consent for participation was not required from the participants or the participants’ legal guardians/next of kin in accordance with the national legislation and institutional requirements.

Author contributions

HZ: Data curation, Formal analysis, Investigation, Software, Validation, Visualization, Writing – original draft. LP: Data curation, Investigation, Validation, Writing – original draft. RC: Data curation, Investigation, Validation, Writing – review & editing. JG: Data curation, Resources, Writing – review & editing. XH: Data curation, Resources, Writing – review & editing. WW: Conceptualization, Funding acquisition, Project administration, Resources, Supervision, Writing – review & editing. SH: Conceptualization, Methodology, Project administration, Resources, Writing – review & editing.

Funding

The author(s) declared that financial support was received for this work and/or its publication. The research was supported by the Independent Research Fund of the State Key Laboratory of Complex, Severe, and Rare Diseases (grant no. 2025-I-ZD-002).

Conflict of interest

The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declared that generative AI was used in the creation of this manuscript. This manuscript used AI tools for language editing, with all generated content reviewed and approved by the authors. The use of AI is transparently disclosed in accordance with journal guidelines. No patient data were processed using AI, ensuring full compliance with privacy and ethical standards.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

1. Tan-Tam C, Chung SW. Minireview on laparoscopic hepatobiliary and pancreatic surgery. World J Gastrointest Endosc. (2014) 6(3):60–7. doi: 10.4253/wjge.v6.i3.60

PubMed Abstract | Crossref Full Text | Google Scholar

2. Filipoiu FM, Badea GT, Enyedi M, Oprea Ș, Filipoiu ZF, Mutu DEG. Mesopancreas-anatomical insights and its implications for diagnosis and clinical and surgical practice. Diagnostics (Basel). (2025) 15(7):20–1. doi: 10.3390/diagnostics15070914

Crossref Full Text | Google Scholar

3. Tseng JF, Raut CP, Lee JE, Pisters PWT, Vauthey JN, Abdalla EK, et al. Pancreaticoduodenectomy with vascular resection: margin status and survival duration. J Gastrointest Surg. (2004) 8(8):935–49. discussion 949–950. doi: 10.1016/j.gassur.2004.09.046

PubMed Abstract | Crossref Full Text | Google Scholar

4. Yamaguchi T, Hasegawa K, Sauvain MO, Passoni S, Kazami Y, Kokudo T, et al. An aberrant right hepatic artery arising from the gastroduodenal artery: a pitfall encountered during pancreaticoduodenectomy. Surg Today. (2021) 51(10):1577–82. doi: 10.1007/s00595-021-02242-4

PubMed Abstract | Crossref Full Text | Google Scholar

5. Ball CG, Dixon E, Vollmer CM, Howard TJ. The view from 10,000 procedures: technical tips and wisdom from master pancreatic surgeons to avoid hemorrhage during pancreaticoduodenectomy. BMC Surg. (2015) 15:122. doi: 10.1186/s12893-015-0109-y

PubMed Abstract | Crossref Full Text | Google Scholar

6. Xu D, Wu P, Zhang K, Cai B, Yin J, Shi G, et al. The short-term outcomes of distal pancreatectomy with portal vein/superior mesenteric vein resection. Langenbecks Arch Surg. (2022) 407(5):2161–8. doi: 10.1007/s00423-021-02382-8

PubMed Abstract | Crossref Full Text | Google Scholar

7. Lopez NE, Prendergast C, Lowy AM. Borderline resectable pancreatic cancer: definitions and management. World J Gastroenterol. (2014) 20(31):10740–51. doi: 10.3748/wjg.v20.i31.10740

PubMed Abstract | Crossref Full Text | Google Scholar

8. Javed AA, Bleich K, Bagante F, He J, Weiss MJ, Wolfgang CL, et al. Pancreaticoduodenectomy with venous resection and reconstruction: current surgical techniques and associated postoperative imaging findings. Abdom Radiol (NY). (2018) 43(5):1193–203. doi: 10.1007/s00261-017-1290-5

PubMed Abstract | Crossref Full Text | Google Scholar

9. Ramacciato G, Mercantini P, Petrucciani N, Giaccaglia V, Nigri G, Ravaioli M, et al. Does portal-superior mesenteric vein invasion still indicate irresectability for pancreatic carcinoma? Ann Surg Oncol. (2009) 16(4):817–25. doi: 10.1245/s10434-008-0281-8

PubMed Abstract | Crossref Full Text | Google Scholar

10. Giovinazzo F, Turri G, Katz MH, Heaton N, Ahmed I. Meta-analysis of benefits of portal-superior mesenteric vein resection in pancreatic resection for ductal adenocarcinoma. Br J Surg. (2016) 103(3):179–91. doi: 10.1002/bjs.9969

PubMed Abstract | Crossref Full Text | Google Scholar

11. Hackert T, Klaiber U, Hinz U, Strunk S, Loos M, Strobel O, et al. Portal vein resection in pancreatic cancer surgery: risk of thrombosis and radicality determine survival. Ann Surg. (2023) 277(6):e1291–8. doi: 10.1097/SLA.0000000000005444

PubMed Abstract | Crossref Full Text | Google Scholar

12. Lu JW, Ding HF, Wu XN, Liu XM, Wang B, Wu Z, et al. Intra-abdominal hemorrhage following 739 consecutive pancreaticoduodenectomy: risk factors and treatments. J Gastroenterol Hepatol. (2019) 34(6):1100–7. doi: 10.1111/jgh.14560

PubMed Abstract | Crossref Full Text | Google Scholar

13. Kokkinakis S, Kritsotakis EI, Maliotis N, Karageorgiou I, Chrysos E, Lasithiotakis K. Complications of modern pancreaticoduodenectomy: a systematic review and meta-analysis. Hepatobiliary Pancreat Dis Int. (2022) 21(6):527–37. doi: 10.1016/j.hbpd.2022.04.006

PubMed Abstract | Crossref Full Text | Google Scholar

14. Othman W, Vandyck KE, Abril C, Barajas-Gamboa JS, Pantoja JP, Kroh M, et al. Stiffness assessment and lump detection in minimally invasive surgery using in-house developed smart laparoscopic forceps. IEEE J Transl Eng Health Med. (2022) 10:2500410. doi: 10.1109/JTEHM.2022.3180937

PubMed Abstract | Crossref Full Text | Google Scholar

15. Al-Taan OS, Stephenson JA, Briggs C, Pollard C, Metcalfe MS, Dennison AR. Laparoscopic pancreatic surgery: a review of present results and future prospects. HPB (Oxford). (2010) 12(4):239–43. doi: 10.1111/j.1477-2574.2010.00168.x

PubMed Abstract | Crossref Full Text | Google Scholar

16. Lin X, Fan Q, Li R, Chen R, Yang Z, Li Y. Enhancing laparoscopic visibility: efficient surgical smoke clearance innovatively using nebulization technology. Biomed Eng Online. (2025) 24(1):65. doi: 10.1186/s12938-025-01395-4

PubMed Abstract | Crossref Full Text | Google Scholar

17. Nwoye CI, Mutter D, Marescaux J, Padoy N. Weakly supervised convolutional LSTM approach for tool tracking in laparoscopic videos. Int J Comput Assist Radiol Surg. (2019) 14(6):1059–67. doi: 10.1007/s11548-019-01958-6

PubMed Abstract | Crossref Full Text | Google Scholar

18. Stoop TF, Ateeb Z, Ghorbani P, Scholten L, Arnelo U, Besselink MG, et al. Surgical outcomes after total pancreatectomy: a high-volume center experience. Ann Surg Oncol. (2021) 28(3):1543–51. doi: 10.1245/s10434-020-08957-x

PubMed Abstract | Crossref Full Text | Google Scholar

19. Wang M, Peng B, Liu J, Yin X, Tan Z, Liu R, et al. Practice patterns and perioperative outcomes of laparoscopic pancreaticoduodenectomy in China: a retrospective multicenter analysis of 1029 patients. Ann Surg. (2021) 273(1):145–53. doi: 10.1097/SLA.0000000000003190

PubMed Abstract | Crossref Full Text | Google Scholar

20. Cao K, Xia Y, Yao J, Han X, Lambert L, Zhang T, et al. Large-scale pancreatic cancer detection via non-contrast CT and deep learning. Nat Med. (2023) 29(12):3033–43. doi: 10.1038/s41591-023-02640-w

PubMed Abstract | Crossref Full Text | Google Scholar

21. Miyamoto R, Takahashi A, Ogasawara A, Ogura T, Kitamura K, Ishida H, et al. Three-dimensional simulation of the pancreatic parenchyma, pancreatic duct and vascular arrangement in pancreatic surgery using a deep learning algorithm. PLoS One. (2022) 17(10):e0276600. doi: 10.1371/journal.pone.0276600

PubMed Abstract | Crossref Full Text | Google Scholar

22. Akbari H, Kosugi Y, Khorgami Z. Image-guided preparation of the calot’s triangle in laparoscopic cholecystectomy. Annu Int Conf IEEE Eng Med Biol Soc. (2009) 2009:5649–52. doi: 10.1109/IEMBS.2009.5333766

PubMed Abstract | Crossref Full Text | Google Scholar

23. Lin C, Gao J, Zheng H, Zhao J, Yang H, Lin G, et al. Three-dimensional visualization technology used in pancreatic surgery: a valuable tool for surgical trainees. J Gastrointest Surg. (2020) 24(4):866–73. doi: 10.1007/s11605-019-04214-z

PubMed Abstract | Crossref Full Text | Google Scholar

24. Shi J, Cui R, Wang Z, Yan Q, Ping L, Zhou H, et al. Deep learning HRNet FCN for blood vessel identification in laparoscopic pancreatic surgery. NPJ Digit Med. (2025) 8(1):235. doi: 10.1038/s41746-025-01663-6

PubMed Abstract | Crossref Full Text | Google Scholar

25. Silva B, Oliveira B, Morais P, Buschle LR, Correia-Pinto J, Lima E, et al. Analysis of current deep learning networks for semantic segmentation of anatomical structures in laparoscopic surgery. Annu Int Conf IEEE Eng Med Biol Soc. (2022) 2022:3502–5. doi: 10.1109/EMBC48229.2022.9871583

PubMed Abstract | Crossref Full Text | Google Scholar

26. Guédon ACP, Meij SEP, Osman KNMMH, Kloosterman HA, van Stralen KJ, Grimbergen MCM, et al. Deep learning for surgical phase recognition using endoscopic videos. Surg Endosc. (2021) 35(11):6150–7. doi: 10.1007/s00464-020-08110-5

PubMed Abstract | Crossref Full Text | Google Scholar

27. Shinozuka K, Turuda S, Fujinaga A, Nakanuma H, Kawamura M, Matsunobu Y, et al. Artificial intelligence software available for medical devices: surgical phase recognition in laparoscopic cholecystectomy. Surg Endosc. (2022) 36(10):7444–52. doi: 10.1007/s00464-022-09160-7

PubMed Abstract | Crossref Full Text | Google Scholar

28. Mascagni P, Vardazaryan A, Alapatt D, Urade T, Emre T, Fiorillo C, et al. Artificial intelligence for surgical safety: automatic assessment of the critical view of safety in laparoscopic cholecystectomy using deep learning. Ann Surg. (2022) 275(5):955–61. doi: 10.1097/SLA.0000000000004351

PubMed Abstract | Crossref Full Text | Google Scholar

29. Wu S, Tang M, Liu J, Qin D, Wang Y, Zhai S, et al. Impact of an AI-based laparoscopic cholecystectomy coaching program on the surgical performance: a randomized controlled trial. Int J Surg. (2024) 110(12):7816–23. doi: 10.1097/JS9.0000000000001798

PubMed Abstract | Crossref Full Text | Google Scholar

30. Twinanda AP, Shehata S, Mutter D, Marescaux J, de Mathelin M, Padoy N. Endonet: a deep architecture for recognition tasks on laparoscopic videos. IEEE Trans Med Imaging. (2017) 36(1):86–97. doi: 10.1109/TMI.2016.2593957

PubMed Abstract | Crossref Full Text | Google Scholar

31. Tuan LQA, Hai PM. Laparoscopic pancreaticoduodenectomy. In: Lomanto D, Tzu-Liang Chen W, Fuentes MB, editors. Mastering Endo-Laparoscopic and Thoracoscopic Surgery: ELSA Manual. Singapore: Springer (2022). p. 357–66.

Google Scholar

32. Hai PM, Tuan LQA. Laparoscopic distal pancreatectomy. In: Lomanto D, Tzu-Liang Chen W, Fuentes MB, editors. Mastering Endo-Laparoscopic and Thoracoscopic Surgery: ELSA Manual. Singapore: Springer (2022). p. 349–55.

Google Scholar

33. Wang J, Sun K, Cheng T, Jiang B, Deng C, Zhao Y, et al. Deep high-resolution representation learning for visual recognition. IEEE Trans Pattern Anal Mach Intell. (2021) 43(10):3349–64. doi: 10.1109/TPAMI.2020.2983686

PubMed Abstract | Crossref Full Text | Google Scholar

34. Long J, Shelhamer E, Darrell T. Fully convolutional networks for semantic segmentation. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2015). p. 3431–40

Google Scholar

Keywords: benchmark metrics, dataset, deep learning, laparoscopic pancreatic surgery, semantic segmentation, vein recognition

Citation: Zhou H, Ping L, Cui R, Gao J, Han X, Wu W and Hua S (2026) Construction and validation of a high-precision annotated dataset for developing intelligent critical vein recognition models in laparoscopic pancreatic surgery. Front. Surg. 13:1711392. doi: 10.3389/fsurg.2026.1711392

Received: 23 September 2025; Revised: 9 December 2025;
Accepted: 14 January 2026;
Published: 2 February 2026.

Edited by:

Lei Wang, Peking University, China

Reviewed by:

Cheng-Peng Li, Peking University, China
Diletta Corallino, Sapienza University of Rome, Italy

Copyright: © 2026 Zhou, Ping, Cui, Gao, Han, Wu and Hua. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Wenming Wu, ZG9jdG9yd3V1QDEyNi5jb20=; Surong Hua, aHVhc3Vyb25nQHRzaW5naHVhLm9yZy5jbg==

These authors have contributed equally to this work and share first authorship

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.