- 1Laboratory for Pathology Dynamics, Department of Biomedical Engineering, Georgia Institute of Technology and Emory University, Atlanta, GA, United States
- 2Center for Machine Learning, Georgia Institute of Technology, Atlanta, GA, United States
- 3School of Computer Science, Georgia Institute of Technology, Atlanta, GA, United States
- 4Department of Orthopedic Surgery, Emory University, Atlanta, GA, United States
- 5Department of Orthopedic Surgery, University of Minnesota, Minneapolis, MN, United States
Introduction: Accurate segmentation of cervical spinous processes on lateral X-rays is essential for reliable anatomical landmarking, surgical planning, and longitudinal assessment of spinal deformity. However, no publicly available dataset provides pixel-level annotations of these structures, and manual delineation remains time-consuming and operator dependent. To address this gap, we curated an expert-labeled dataset of 500 cervical spine radiographs and developed CervSpineNet, a hybrid deep learning framework for automated spinous process segmentation.
Methods: CervSpineNet integrates a transformer-based encoder to capture global anatomical context with a lightweight convolutional decoder to refine local boundaries. Training used a compound loss function that combines Dice, Focal Tversky, Hausdorff distance transform, and Structural Similarity (SSIM) terms to jointly optimize region overlap, class balance, structural fidelity, and boundary accuracy. The model was trained and evaluated on three dataset variants: original images, contrast-enhanced images using CLAHE, and augmented images. Performance was benchmarked against four baselines: U-Net, DeepLabV3+, the Segment Anything Model (SAM), and a text-guided SegFormer.
Results: Across all experimental settings, CervSpineNet consistently outperformed competing methods, achieving mean Dice coefficients above 0.93, IoU values above 0.87, and SSIM above 0.98, with substantially lower HD95 distances. The model demonstrated strong agreement with ground truth, with global MAE ≈ 0.005, and maintained efficient inference times of 5–10 seconds per image. With a compact footprint of approximately 345 MB, CervSpineNet runs on standard clinical hardware and reduces manual annotation time by about 96%.
Discussion: These results indicate that combining transformer-driven global context with convolutional boundary refinement enables robust and reproducible spinous process segmentation on lateral cervical radiographs. By pairing an expert-annotated dataset with a high-performing, computationally efficient model, this work provides a scalable foundation for AI-assisted cervical spine analysis, supporting rapid segmentation for surgical evaluation, deformity monitoring, and large-scale retrospective studies in both research and clinical practice.
1 Introduction
The cervical spine—the upper segment of the vertebral column comprising seven articulating vertebrae (C1–C7)—plays a critical biomechanical and protective role. The craniocervical junction, formed by the atlas (C1) and axis (C2), enables head rotation and flexion–extension, while the subaxial spine (C3–C7) provides load-bearing stability and supports the weight of the head (Kaiser et al., 2025). Each vertebra features a posterior bony prominence called the spinous process, whose morphology varies substantially across levels and between individuals (Cramer, 2014). These spinous processes serve as essential anatomic landmarks for vertebral identification, muscle attachment, and radiologic orientation. Their visibility and geometry make them key indicators in diagnostic imaging, surgical navigation, and postoperative evaluation.
Accurate assessment of the cervical spinous processes is clinically significant in several contexts. Avulsion injuries such as clay-shoveler’s fractures typically involve the lower cervical or upper thoracic processes and are common in activities requiring rapid spinal rotation (Posthuma de Boer et al., 2016). Similarly, congenital variants—such as hypoplasia, dysplasia, or persistent apophyses—may mimic fractures and lead to misinterpretation if not correctly recognized (Riew et al., 2010; Farooqi et al., 2017; Gould et al., 2020). Understanding these variants is vital for distinguishing developmental anomalies from acute pathology and for avoiding unnecessary interventions (Chen et al., 2006). Beyond diagnosis, spinous process delineation also assists in evaluating outcomes following a spinal surgery called anterior cervical discectomy and fusion (ACDF), which is a widely performed procedure for degenerative conditions of the cervical spine. Postoperative evaluation of fusion status, indicating successful bony union and mechanical stability between adjacent vertebrae, represents a critical clinical outcome. However, studies show that inter-observer variability in assessing postoperative fusion from dynamic radiographs can be substantial, often requiring supplemental computed tomography (CT) scans or quantitative methods (Ariyaratne et al., 2024). With ACDF case volumes projected to rise markedly through 2040 (Neifert et al., 2020), automated segmentation tools could streamline assessment and reduce dependence on manual measurements.
Semantic segmentation—the pixel-level delineation of structures within images—is central to computational imaging and biomedical analysis (Hollon et al., 2020; Ouyang et al., 2020). Manual segmentation remains the reference standard but is labor-intensive, time-consuming, and prone to inter-observer variability (Ma et al., 2024). Automated methods can dramatically reduce workload, improve consistency, and enable high-throughput image analysis (Wang et al., 2019). However, manual labeling of X-rays is particularly challenging due to overlapping anatomy, limited soft-tissue contrast, and the absence of public datasets for small bony structures such as spinous processes (Wang et al., 2019; Wang S. et al., 2021; Galbusera and Cina, 2024). While recent advances in deep learning have achieved outstanding results for CT and MRI segmentation (Ahmad et al., 2023; Muthukrishnan et al., 2024; Son et al., 2024; Yang et al., 2024), segmentation in plain radiographs remains relatively underexplored.
To date, there is no publicly available dataset containing cervical spine X-rays with corresponding pixel-level annotations of the spinous processes. Prior work suggests that certain vertebral levels, notably C1/C2 and C6/C7, are particularly difficult to identify using machine learning models (Shim et al., 2022; van Santbrink et al., 2025). Preliminary zero-shot experiments using transformer-based encoders such as SAM (Bucher et al., 2019) and conventional segmentation architectures like DeepLabV3+ produced unsatisfactory results. These findings underscore the need for a domain-specific dataset and a dedicated segmentation framework optimized for the cervical spine.
Recent literature highlights that effective medical image segmentation models must balance local boundary precision with global contextual awareness. Conventional convolutional neural networks (CNNs) excel at fine detail but are limited in modeling long-range dependencies, whereas transformer architectures capture global context at the expense of spatial granularity. Hybrid models that integrate these complementary strengths have demonstrated improved performance across biomedical imaging tasks. Approaches such as MedSAM and TransUNet combine transformer-based encoders with U-Net–style decoders, yielding sharper boundaries and stronger structural consistency across diverse medical modalities (Chen J. et al., 2024; Ma et al., 2024). Similarly, hybrid convolution–transformer systems in musculoskeletal imaging achieve strong Dice and IoU performance and near-expert agreement (Xu et al., 2022; Ham et al., 2023; Mostafa et al., 2024; Al-Antari et al., 2025; Bao et al., 2025). These developments support hybrid encoder–decoder architectures as a robust design paradigm for accurate and data-efficient medical image segmentation.
Building on this foundation, the present study introduces CervSpineNet, a hybrid deep learning framework for automated segmentation of cervical spinous processes in lateral X-rays. The primary contributions are threefold:
1. A new expert-labeled dataset of cervical spine radiographs with pixel-level binary masks of the spinous processes.
2. A hybrid transformer–CNN architecture that integrates a ViT-B encoder for global context modeling with a traditional convolutional decoder with added layers for edge refinement.
3. A compound loss function that jointly optimizes region overlap, edge sharpness, class balance, and structural similarity.
CervSpineNet consistently achieved mean Dice coefficients exceeding 0.93 across experiments and demonstrated excellent agreement with ground-truth masks. Beyond accuracy, the framework reduces manual annotation time by approximately 96% and is efficient enough to run on standard hospital hardware. Together, the proposed dataset and model provide a scalable foundation for automated, high-fidelity cervical spine analysis, advancing both the methodological and translational frontiers of biomedical image segmentation.
2 Methodology
This section outlines the methodological framework used to develop and evaluate the proposed CervSpineNet model, from dataset curation and preprocessing to model training, benchmarking, and performance assessment. The overall workflow is illustrated in Figure 1.
Figure 1. Schematic overview of the CervSpineNet pipeline for automated cervical spinous process segmentation. The workflow includes dataset curation from the publicly available Cervical Spine X-ray Atlas (CSXA), manual annotation of spinous processes, and annotation-to-mask generation for binary ground-truth creation. Preprocessing steps include Contrast Limited Adaptive Histogram Equalization (CLAHE) and data augmentation to enhance variability and image contrast. The processed datasets are used for model training and inference, with results evaluated through quantitative metrics and visual comparison to expert-annotated ground truth.
2.1 Data acquisition
Prior work shows CNN-based segmentation is effective for vertebral bodies in radiographs and MRI (Chen Y. et al., 2024; Wang H. et al., 2025). However, no public dataset provides pixel-level masks of the cervical spinous processes, so we curated a task-specific corpus of lateral cervical spine X-rays and their corresponding binary masks.
We randomly sampled 500 PNG images from the Cervical Spine X-ray Atlas (CSXA) (Ran et al., 2024) (4,963 PNGs with JSON annotations; one image per patient) and manually annotated the spinous processes on a tablet with a stylus. Annotations were converted to binary masks using an OpenCV color-segmentation pipeline: images were converted to HSV; red hue bands (0 °–10 °, 160 °–180 °) corresponding to tracings were thresholded; masks from both bands were combined; contours were extracted and filled to produce clean binary ground-truth masks. All 500 image–mask pairs were created with consistent dimensions. Trained annotators labeled the images; an expert spine surgeon (Dr. Sangwook T. Yoon) provided final review and approval. The annotation workflow is shown in Figure 2.
Figure 2. Representative workflow for creating binary masks of the cervical spinous processes from lateral X-ray images. (A) Original cervical spine radiograph from the Cervical Spine X-ray Atlas (CSXA) dataset. (B) Manual annotation of individual spinous processes using a tablet and stylus. (C) Automated mask conversion using a color-based segmentation algorithm in HSV space, where annotated red regions are thresholded and converted to binary masks. The resulting ground-truth masks serve as pixel-level labels for model training and evaluation.
2.2 Data bifurcation, pre-processing, and augmentation
To ensure fair evaluation and minimize overfitting (Muraina, 2022), we used an 80/20 split, allocating 400 images for training and 100 for testing (Bichri et al., 2024); the test set remained untouched throughout all experiments (Lones, 2021). For the training images, we generated three distinct dataset variants, each used to train a separate model configuration:
1. Original dataset: the 400 unaltered radiographs and their corresponding masks.
2. CLAHE dataset: the 400 original radiographs processed with Contrast Limited Adaptive Histogram Equalization (CLAHE) to compensate for non-uniform intensity distributions common in radiographs. Specifically, CLAHE operates by dividing the image into small contextual regions, equalizing each region’s histogram, and applying a clip limit to avoid noise amplification before recombining the tiles via bilinear interpolation (Zuiderveld, 1994). We used the standard OpenCV implementation.
3. Augmented dataset: the 400 original + 400 augmented image–mask pairs (800 total). To increase the effective training set size and improve model generalization (Goceri, 2023), augmentations including affine transformations with rotations of ±10 ° and ±45 ° and translations of ±10 pixels along x and y axes were applied to the original dataset. The larger ±45 ° rotations were included intentionally, as several expert-provided sample radiographs contained substantial initial misalignment. Mask transformations used nearest-neighbor interpolation to preserve labels.
All three training variants (original, CLAHE, and augmented) were derived from the same underlying set of 400 cervical radiographs. These variants reflect different preprocessing pipelines applied to identical base images rather than independent datasets, thus enabling us to evaluate how raw, intensity-normalized, and synthetically diversified training distributions affected model performance.
2.3 Experimental setup
We trained and evaluated each model on the three dataset variants: Original (400 images); CLAHE (400 images); Augmented (800 images). The test set (n = 100) remained constant across experiments. We compared GPU vs. CPU training/inference for efficiency. All experiments used a single system with PyTorch and an NVIDIA H100 Tensor Core GPU; the uniform environment ensured consistent timing and memory profiles. The H100’s tensor cores and memory bandwidth supported large-batch training and stable convergence.
A test across five different images from the testing set show that GPU inference averaged 5–8 s/image; CPU inference on an Intel® Xeon® Gold 6,136 (2 × 12 cores) averaged 9–10 s/image. The proposed CervSpineNet is computationally lightweight (∼345 MB) and typically utilizes ∼8–9 GPU cores during inference, enabling deployment without specialized hardware.
2.4 Proposed hybrid segmentation approach: CervSpineNet architecture
Figure 3 shows the overall architecture and I/O. The encoder is transformer-based; the decoder is a lightweight U-Net–style module with residual and squeeze-and-excitation (SE) blocks and bilinear up-sampling.
Figure 3. Schematic representation of the CervSpineNet hybrid architecture for automated segmentation of cervical spinous processes. The framework performs training and inference using paired cervical spine X-ray images and binary masks. CervSpineNet integrates a Vision Transformer (ViT-B) encoder for global contextual feature extraction with a U-Net–style decoder for spatial detail recovery. Each stage of the decoder includes residual blocks (RB) to maintain gradient flow, squeeze-and-excitation blocks (SEB) for adaptive channel attention, and bilinear upsampling (BU) layers for smooth boundary reconstruction. Preprocessing involves image resizing and binarization, while the output mask is generated via sigmoid activation and optimized using the composite loss function combining Dice, Focal Tversky, Hausdorff Distance Transform, and SSIM terms.
2.4.1 Inputs and notations
We denote the input image by
2.4.2 Encoder: ViT-B
Vision transformers (ViT) capture global context and long-range dependencies valuable for medical segmentation (Zhang et al., 2024), but benefit from local refinements from downstream decoders (Khan and Khan, 2025). As a result, we base our encoder on the ViT-B backbone of the Segment Anything Model (SAM). Specifically, we load the ViT-B variant from the original SAM repository through the sam_model_registry interface using the official PyTorch implementation of SAM. The prompt encoder and mask decoder heads are eliminated from this model, leaving only the image encoder.
Given an input radiograph/x-ray image that is resized to
Our unique U-Net-style decoder receives these feature maps and uses multi-scale up sampling to return the image to its original resolution.
The encoder is based on the SAM ViT-B image encoder, which we use as a generic feature extractor for cervical spine radiographs. Each input radiograph is first resized to 1,024 × 1,024 and linearly projected into a sequence of non-overlapping 16 × 16 patches. These patches are embedded into a high-dimensional token space, augmented with learned 2D positional encodings, and passed through a stack of multi-head self-attention and feedforward transformer blocks. Across these layers, the encoder aggregates information over the entire field-of-view, so that each token encodes both local appearance (e.g., vertebral boundaries) and long-range context (e.g., overall spinal alignment). Finally, the token sequence is reshaped back into a 2D feature map of size 64 × 64 with 256 channels, which serves as the latent representation fed into our U-Net–style decoder for pixel-wise mask prediction.
The encoder weights are initialized from the SA-1B pre-trained SAM ViT-B checkpoint and then fine-tuned end-to-end together with our decoder on the curated cervical spine dataset using the compound loss described in Section 2.4.4. In contrast, we also train a complete SAM baseline (Section 2.5.3) that keeps the original image encoder, prompt encoder, and mask decoder. In this baseline, the entire SAM model is optimized using the same training set, and centroids obtained from the ground-truth masks are utilized as point prompts.
To qualitatively consider if the ViT-B encoder in CervSpineNet uses global contextual information, we generated three types of complementary attention maps for the trained hybrid model. During a forward pass, we cached the self-attention inputs to all transformer blocks and the final encoder feature map. First, we computed an attention-rollout map by creating the self-attention matrix for each block at the final token resolution, adding an identity term, normalizing rows, and multiplying the matrices across blocks; the vector of token importances that resulted was reshaped to the encoder grid (≈64 × 64). Second, we computed a query-centric attention map from the last block by selecting a query token centered on the mid-cervical spinous process and visualizing its attention weights to all other tokens. Third, we produced a Grad-CAM map on the output of the encoder by back-propagating the mean foreground prediction to the encoder feature map and aggregating gradient-weighted activations. All maps were upsampled to image resolution and overlaid with the radiograph for visual interpretation.
2.4.3 Decoder
U-Net decoders progressively restore spatial resolution and boundary fidelity (Yuan and Cheng, 2024). Our decoder stages include Conv + ReLU → Residual Block → SE Block → Bilinear Upsample → 1 × 1 Conv + Sigmoid. SE blocks re-weight channels to enhance task-relevant features (Wang J. et al., 2021); residual blocks maintain gradient flow and stabilize training (Ashkani Chenarlogh et al., 2022).
2.4.3.1 Convolution stage
Equation 2 denotes the convolution stage where
2.4.3.2 Residual block
The residual block depicted by Equation 3 contains two stacked Conv + ReLU layers with a skip connection.
2.4.3.3 SE block—squeeze (global average pooling)
Equation 4 gives the squeeze formula for the SE block where
2.4.3.4 SE block—excitation (2-layer MLP)
Equation 5 gives the excitation formula for the SE block where
2.4.3.5 Channel re-weighting
Channel re-weighing is represented by Equation 6, where
2.4.3.6 Bilinear upsampling
The output feature map F⊖ after sampling is given by Equation 7.
Our decoder is U-Net–inspired but intentionally departs from the textbook U-Net design. Because the encoder is a SAM ViT-B transformer that produces a single low-resolution feature map (256 channels at 64 × 64), the decoder operates on this representation only and does not use multi-scale encoder–decoder skip connections or a symmetric convolutional contracting path. Instead, it forms a purely expanding path with four stages of bilinear up-sampling followed by 3 × 3 convolutions, residual blocks, and SE blocks that refine features at each scale before the final 1 × 1 convolution and sigmoid.
2.4.4 Loss function
Composite losses can improve robustness by balancing region overlap, boundary accuracy, and imbalance (Taghanaki et al., 2019; Mu et al., 2022). To jointly optimize region-wise overlap, class imbalance, boundary accuracy, and structural fidelity, we selected the four loss elements: Dice, Focal Tversky (FT), Hausdorff Distance Transform (HD95), and Structural Similarity Index (SSIM). The Hausdorff Distance Transform term specifically penalizes boundary misalignment, the SSIM term promotes structurally coherent, non-blurry masks that respect the overall spine morphology, Focal Tversky down-weights easy background pixels and concentrates learning on imbalanced, thin vertebral structures, and Dice loss offers a powerful region-overlap term.
Let
The Dice Loss is given by Equation 8:
For the Focal Tversky loss, we first compute the True Positive (TP), False Positive (FP) and False Negative (FN) as given in Equation 9:
Thus, the Tversky Index, represented by Equation 10, is given by
And the Focal Tversky Loss, given by Equation 11, is computed as:
Where,
For the Hausdorff Distance Transform Loss, depicted by Equation 12, we form binarized masks,
Finally, for the SSIM loss given by Equation 13, we compute the image-wise average SSIM following the standard definition (Equation 18), and define the loss as:
After ablations with the composition of loss functions (Section 2.7.2), the best results were obtained with a weighted sum of Dice, Focal Tversky, Hausdorff Distance Transform, and SSIM (Equation 14):
Based on our pilot experiments, we empirically selected the weights (0.5, 0.3, 0.1, 0.1). We prioritize Dice as the primary region-overlap term, give Focal Tversky a moderate weight to address class imbalance, and use smaller regularization weights for the boundary-focused HDT and SSIM terms to sharpen contours without destabilizing optimization.
2.5 Baselines implemented: performance comparison
To avoid model-class bias and follow current benchmarking guidance (Antonelli et al., 2022; Tejani et al., 2024), we compared CervSpineNet to strong and architecturally diverse baselines (Wolfrath et al., 2024): DeepLabV3+, U-Net, the full Segment Anything Model (SAM), and a text-guided SegFormer. All baselines and CervSpineNet were trained and evaluated on the same training/test split with identical pre-processing and data augmentations. U-Net, DeepLabV3+, and text-guided SegFormer were trained on images resized to 512 × 512, which is standard for CNN-based medical segmentation and keeps GPU memory requirements manageable. The SAM ViT-B and the proposed CervSpineNet were trained on 1,024 × 1,024 inputs to match the native input resolution of the SAM image encoder. All predictions were resampled back to the original image resolution for evaluation. The batch size for every training experiment was kept 1, and the models were trained on 50 epochs, selecting the checkpoint with the best testing metrics as the metrics saturated around 40–45 for all experiments. This ensured that the performance differences mainly reflect architectural choices rather than differences in data or optimization.
2.5.1 DeepLabV3 +
DeepLabV3+ performs strongly across medical segmentation tasks (Hou et al., 2024; Ketenci Çay et al., 2025). We implemented DeepLabV3+ with a ResNet-50 backbone pretrained on ImageNet. The single-channel radiographs were resized to 512 × 512 pixels and duplicated to three channels. A 1-channel logit map that has been upsampled to the input resolution is produced by the model. An AdamW optimizer and BCE Loss were used to refine DeepLabV3+ on our cervical dataset.
2.5.2 U-net
U-Net remains a robust baseline across modalities (Ronneberger et al., 2015; Du et al., 2020; Azad et al., 2022). We created a 4-level U-Net from scratch using 64 initial filters as a convolutional baseline. The inputs were 512 × 512. We used the AdamW optimizer with BCE Loss.
2.5.3 Segment anything model (SAM)
SAM is trained on >1B masks (Kirillov et al., 2023) and, with task-specific prompting, can be adapted to medical images (Mazurowski et al., 2023). For SAM, we used the official ViT-B variant from the Meta Segment-Anything repository, loaded via the sam_model_registry. We kept the full architecture, namely image encoder, prompt encoder, mask decoder, and resized X-rays to 1,024 × 1,024 as recommended. For each training image, we derived oracle point prompts by placing a single positive point at the centroid of each ground-truth spinous process mask. We passed these points through the prompt encoder and supervised the resulting masks with binary cross-entropy + Dice loss. We then fine-tuned all SAM parameters with the AdamW optimizer for 50 epochs. At inference time, we used one positive centroid point per spinous process and placed a threshold for the output probabilities at 0.5. This setup follows recent SAM adaptations to medical imaging and represents a reasonable, strong SAM baseline for our task.
2.5.4 Vision-language model—text guided SegFormer
We used MedCLIP (domain-adapted CLIP) (Wang et al., 2022) to encode text prompts describing the spinal processes; embeddings were L2-normalized and pooled to a fixed vector
2.6 Evaluation metrics
Following common practice (Li et al., 2025; Zhang et al., 2025), we report Dice, Intersection-over-Union (IoU), Structural Similarity (SSIM), Hausdorff Distance (HD95), and Volumetric Similarity (VS) on the held-out test set. Equations 15-17 display the Dice, IoU and VS terms as implemented in this study.
Here, TP/FP/TN/FN are computed from thresholded (0.5) binary predictions and ground-truth masks for each pixel.
• True Positive (TP) – prediction = 1, ground truth = 1 (pixels correctly classified as foreground)
• False Positive (FP) – prediction = 1, ground truth = 0 (pixels incorrectly classified as foreground)
• True Negative – prediction = 0, ground truth = 0 (pixels correctly classified as background)
• False Negative (FN) – prediction = 0, ground truth = 1 (missed foreground pixels)
SSIM captures luminance, contrast, and structure similarity. We used the standard SSIM metric (Bakurov et al., 2022), defined for a predicted mask
where,
HD95, as given by Equation 20, is the 95th percentile surface-to-surface distance between predicted and ground-truth objects (Ramachandran et al., 2023). Let
Thus, for all distances,
2.7 Ablation studies
Ablation experiments are essential in medical image segmentation, particularly for hybrid architectures, as they clarify the relative contributions of individual components to overall performance (Gao et al., 2021; Wang X. et al., 2025). To systematically characterize the behavior of CervSpineNet, we performed two sets of ablations: (i) architectural ablations isolating encoder and decoder contributions, and (ii) loss-function ablations assessing how each loss term influences error modes and structural fidelity.
2.7.1 Architectural ablation experiments
To quantify how much performance arises from the encoder versus the decoder, we evaluated three architectural variants under identical training conditions, preprocessing, composite loss, and data splits.
i. Pure CNN Encoder–Decoder Baseline
We first constructed a fully convolutional baseline in which both the encoder and decoder are simple CNNs. The encoder comprises four convolution–BatchNorm–ReLU blocks, each followed by 2 × 2 max pooling, reducing the spatial resolution from 1,024 × 1,024 to 64 × 64 while expanding feature channels from 3 to 256. The decoder is deliberately minimal, consisting of stacked 3 × 3 convolutions with ReLU activations and bilinear upsampling, terminating in a 1 × 1 sigmoid layer for mask generation. No residual connections, attention modules, or skip connections are used. This configuration provides a reference model that isolates the performance of a straightforward fully convolutional architecture.
ii. CNN Encoder + Full Decoder
Next, we evaluated a model with the same CNN encoder as above but augmented with our full decoder. Each decoding stage includes a 3 × 3 convolution, a residual block to enhance representational depth without compromising gradient flow, and a squeeze-and-excitation (SE) block for adaptive channel reweighting based on global context. Bilinear upsampling is applied at each stage. Because the encoder and training protocol remain constant, differences in performance directly reflect the contribution of residual pathways and channel attention to anatomical structure reconstruction.
iii. ViT-B Encoder (SAM) + Simple Decoder
Finally, we substituted the CNN encoder with the Vision Transformer–B (ViT-B) image encoder from Segment Anything, while retaining the simple decoder of the first experiment. The ViT-B encoder processes 1,024 × 1,024 inputs to produce 256-channel 64 × 64 feature maps enriched with long-range dependencies via self-attention. All transformer layers are unfrozen during training. Using the same composite loss, optimizer (AdamW), and data splits allows this ablation to cleanly assess the effect of replacing local convolutional features with global transformer-based representations, independent of decoder capacity.
2.7.2 Loss ablation experiments
We also performed a dedicated loss-function ablation using the full CervSpineNet architecture (ViT-B encoder + residual/SE decoder). The optimizer, learning rate schedule, and data partitions were held constant across conditions. Loss terms were added progressively to address complementary error modes.
Beginning with a Dice-only baseline emphasizing volumetric overlap, we incorporated the Focal Tversky (FT) loss to penalize false negatives—critical for foreground–background imbalance in postoperative cervical spine X-rays. We then added the SSIM term to encourage structural and textural fidelity, particularly along vertebral margins and disc spaces. Finally, we introduced a differentiable Hausdorff distance surrogate (HD95) to explicitly penalize boundary deviations and outlier predictions.
This yielded five configurations—Dice; Dice + FT; Dice + SSIM; Dice + FT + SSIM; and Dice + FT + SSIM + HD95—each trained for 50 epochs and evaluated on an identical test set using Dice, IoU, SSIM, HD95, and volumetric similarity metrics.
3 Results
This section presents the outcomes of the experiments designed to evaluate the proposed CervSpineNet model and its comparative baselines. Quantitative results are reported across the original, CLAHE-enhanced, and augmented datasets, followed by statistical significance testing, qualitative visual assessments, and an analysis of time efficiency between automated and manual segmentation.
3.1 Quantitative comparison
Mean performance metrics (Dice, IoU, SSIM, HD95, VS) on the testing data were computed for all models across the three dataset variants: the original set (Table 1), CLAHE-enhanced images (Table 2), and the augmented dataset (Table 3).
Repeated trials with identical hyperparameters produced standard deviations in the range 0.001–0.005 for Dice, IoU, SSIM, and VS and ∼0.3–0.6 for HD95, confirming the stability and reproducibility of model performance across the testing data. Figure 4 shows the distribution of Dice coefficients for all models evaluated on the 100 test images under the three preprocessing conditions: (A) original data, (B) CLAHE-enhanced data, and (C) augmented data. Across all three dataset variants, CervSpineNet exhibited the narrowest spread and highest median Dice scores, indicating robustness to preprocessing variations and the radiographic variability that these variants are designed to reflect.
Figure 4. Distribution of Dice coefficients across test datasets for different preprocessing conditions. Boxplots show per-image Dice scores (n = 100) for all segmentation models evaluated on (A) the original dataset, (B) the CLAHE-enhanced dataset, and (C) the augmented dataset. CervSpineNet demonstrates the highest median Dice and the smallest interquartile range across all three data variants, indicating superior segmentation accuracy and stability relative to baseline models.
CervSpineNet achieved the highest or near-highest mean values for nearly every metric, with particularly strong performance in SSIM and HD95, which reflect structural accuracy and boundary precision, respectively. To determine whether these improvements were statistically significant, formal non-parametric tests were conducted as described in the following section.
A summary of the mean test-set performance for all four architectural configurations adopted in this study is shown in Table 4.
Table 4. Architectural Ablation Experiments and the metrics yielded: This table follows a similar structure as Tables 1–3 and demonstrates a comparison between different ablation experiments done with the hybrid model architecture utilizing the same metrics as used previously.
Finally, the results of the loss ablation experiment are demonstrated in Table 5. Different sets of loss functions used, and their metrics scores are depicted in this table. The experiments demonstrated in Tables 4, 5 also yielded minute variations in the range of 0.002–0.004 for Dice, IoU, SSIM, and VS and variations of ∼0.45–0.70 for HD95.
Table 5. Loss Ablation Experiments and the metrics yielded: different combinations of loss functions and their results.
3.2 Statistical significance analysis of model performances
Prior studies highlight the robustness of the Wilcoxon Signed-Rank test and the Friedman test, which detects overall performance differences across multiple models evaluated on the same samples, for evaluating segmentation algorithms (Mostafa et al., 2024). We compared five segmentation models (SAM, U-Net, text-guided SegFormer, DeepLabV3+, and CervSpineNet) using five metrics (Dice, IoU, SSIM, HD95, and VS) across the same per-image testing scores.
The Friedman test assessed overall model differences for each metric, followed by pairwise Wilcoxon signed-rank tests with Bonferroni correction (p < 0.05). Across all metrics, significant overall differences were observed (Friedman p < 0.001).
CervSpineNet significantly outperformed SAM and U-Net (corrected p < 0.001) in every metric and dataset variant. It also exceeded DeepLabV3+ in SSIM, HD95, and Volumetric Similarity while maintaining comparable Dice and IoU (corrected p ≈ 0.05). Against the text-guided SegFormer, CervSpineNet achieved statistically higher performance in all metrics except VS, where the difference was not significant. These results confirm that the proposed hybrid model, CervSpineNet, provides robust and consistent segmentation.
3.3 Qualitative inference visualization and error analysis
Visual inspection of unseen test radiographs further corroborated the quantitative findings (Figure 5). Panels (A), (B), and (C) show representative segmentation results from the original, CLAHE-enhanced, and augmented datasets, respectively, comparing CervSpineNet with all baseline models. Across all preprocessing conditions, CervSpineNet produced the most accurate and anatomically consistent delineations, with smoother contours, sharper boundaries, and markedly fewer false or incomplete predictions than either CNN-only or transformer-only architectures. These qualitative outcomes parallel the quantitative improvements reported in Section 3.1, illustrating how the hybrid architecture captures global contextual features while preserving fine boundary details.
Figure 5. Qualitative comparison of model predictions on representative test radiographs. Representative examples (A–C) show three randomly selected test radiographs and their corresponding segmentation results for each baseline model and the proposed CervSpineNet. As evidenced by the red markings, CervSpineNet demonstrates superior boundary sharpness and anatomical conformity of the spinous processes compared with U-Net, DeepLabV3+, SAM, and the text-guided SegFormer. The red circles display the visual prediction errors made by the baselines when compared to the ground-truth mask.
To examine segmentation fidelity in more detail, Figure 6 visualizes per-pixel discrepancies between CervSpineNet predictions and expert ground-truth masks. In these heatmaps, uncolored regions indicate perfect agreement, while increasing color intensity reflects greater deviation from the annotated boundaries. False positives represent regions erroneously predicted as foreground outside the annotated structure, false negatives correspond to missed spinous process pixels within the ground-truth region, and thin color bands along the contours indicate minor boundary mismatches or thickness differences.
Figure 6. Qualitative error analysis and per-image Dice, HD95, and mean absolute error (MAE) for high-quality CervSpineNet predictions. Representative examples show original cervical spine X-rays, expert ground-truth (GT) masks, CervSpineNet-predicted masks, and corresponding error heatmaps. Colored overlays highlight mis-segmented regions, where greater intensity denotes larger deviation from the GT boundaries. Quantitative values include Dice, HD95, Global MAE (overall pixel-wise deviation) and Foreground-only MAE (error within annotated regions). Lower values indicate higher structural fidelity and more accurate boundary reconstruction.
A complementary quantitative error assessment was performed using Dice, HD95, and Mean Absolute Error (MAE) calculated on continuous (non-binarized) probability maps to preserve boundary sensitivity. Figure 6 demonstrate a sample of radiographs with good model predictions backed by the metrics used for comparison. On the other hand, Figure 7 depicts some edge cases where the model failed to perform well due to a variety of reasons often resulting in over/under prediction of structures or significant disagreement from ground truth masks.
Figure 7. Qualitative error analysis and per-image Dice, HD95, and mean absolute error (MAE) for low-quality CervSpineNet predictions. It is like Figure 6 in terms of the results displayed but this figure demonstrates the failure modes of the hybrid model and includes some edge case radiographs that causes the model to predict masks poorly and yield below-par metric results.
A pixel-by-pixel error heatmap between the ground-truth spinous process mask and the model prediction is superimposed on the lateral cervical X-ray in Figures 6, 7. Areas where prediction and ground truth agree appear like the original image, whereas colored bands around the spinous processes mark disagreement: cooler colors indicate small or no error, and warmer colors (green–yellow–red) highlight increasing mismatch, with the brightest regions corresponding to the largest segmentation errors. In these figures, the Global MAE captures the average pixel-wise deviation across the entire image, reflecting overall segmentation fidelity, whereas the Foreground-only MAE quantifies deviation exclusively within annotated spinous process regions, emphasizing local structural precision. Both metrics were computed at the native image resolution, where lower MAE values indicate stronger correspondence with expert annotations. CervSpineNet consistently achieved the lowest Global and Foreground-only MAE scores, confirming high boundary accuracy and strong structural alignment across the test set.
Figure 6 shows three representative test radiographs with high-quality CervSpineNet segmentations. In these typical cases, spinous process masks from C1–C7 closely follow the cortical margins with minimal over- or under-segmentation. These examples support the good overall performance observed in the aggregate metrics by achieving high quantitative agreement with the reference annotations (Dice ≈ 0.93, HD95 ≈ 3, Global MAE ≈ 0.005, and Foreground-only MAE ≈ 0.05).
Figure 7 illustrates three characteristic failure modes encountered by CervSpineNet. In Case 1, adjacent spinous processes are extremely close and partially overlapping, causing the model to merge two levels (C3–C4) into a single elongated mask and consequently overestimate the superior spinous process. In Case 2, a fractured spinous process produces an irregular and discontinuous contour. While the expert ground truth precisely outlines the fracture margins, the model predicts a smoother, unbroken process, resulting in reduced Dice scores and a prominent error band in the heatmap. In Case 3, the C7 spinous process is scarcely visible due to low contrast and soft-tissue overlap at the cervicothoracic junction; although the model accurately segments C1–C6, it fails to detect C7 entirely, yielding a level-specific false negative. Together, these examples indicate that severe pathology, low bone–soft-tissue contrast, and tightly packed vertebral levels are the conditions under which CervSpineNet is most vulnerable. Future work will prioritize augmenting the training set with such challenging cases and exploring level-aware constraints or post-processing strategies to mitigate mask merging and missed levels.
Figure 8 presents attention visualizations for three representative held-out test cases—two typical high-quality segmentations and one challenging example with substantial shoulder overlap. In the well-segmented cases, both attention-rollout and Grad-CAM overlays highlight a continuous chain of vertebral bodies and spinous processes spanning the upper through lower cervical levels. This pattern indicates that the transformer encoder integrates information across the full cervical column rather than relying solely on local image patches. Similarly, the query-centric attention map, when the query token is placed on a mid-cervical spinous process, assigns high attention weights to both adjacent and more distant vertebral levels, reinforcing evidence of long-range contextual reasoning. In contrast, the difficult case shows more spatially diffuse attention that partially shifts toward the overlapping shoulder and soft-tissue regions, corresponding to the observed segmentation errors. Collectively, these qualitative results support the conclusion that CervSpineNet’s transformer encoder leverages global anatomical context along the cervical spine.
Figure 8. Transformer Attention Maps for ViT-B. Attention maps generated for three test cases; two with good (1, 2) and one with bad metrics and predicted masks (3). (a) is the original radiograph, (b) is the hybrid model predicted mask, (c) is the query centric map, (d) represents the attention-rollout map, and (e) depicts the Grad-CAM map.
3.4 Comparison of manual and automated segmentation efficiency
We conducted a systematic analysis to compare the time required for manual versus automated segmentation of cervical spinous processes. For manual annotation, three annotators—two trained annotators and one expert spine surgeon—outlined the spinous processes on a predefined set of radiographs using a tablet interface. The total manual annotation time was determined by averaging the duration recorded across all three annotators. This process was divided into two stages: (1) annotation, measuring the average time required to trace the spinous processes, and (2) mask generation, recording the time needed to transform these annotations into binary masks using the color-based segmentation algorithm described in Section 2.1.
Figure 9 compares the time required for manual versus automated segmentation. Figure 9A summarizes the average annotation time per image for all manual annotators— two annotators and 1 expert—alongside the automated CervSpineNet model. Manual annotation times ranged from 46 to 98 s, with an overall average of 72 s, whereas CervSpineNet produced segmentations in only ∼7.5 s per image. Figure 9B shows total processing time, combining annotation and mask-generation stages. Manual processing required approximately 3–4 min per image (mean = 210 s). In contrast, automated segmentation using CervSpineNet took only 5–10 s (mean = 7.5 s), corresponding to an overall ∼96.43% reduction in total processing time.
Figure 9. (A) Horizontal bar chart showing average annotation time per image for each manual annotator (Expert, Annotator 1, Annotator 2, and Annotator Group Average) compared with automated segmentation using CervSpineNet. Manual annotation required 46–98 s per image on average, whereas automated inference required only ∼7.5 s. (B) Vertical bar chart illustrating total segmentation time (annotation + mask generation) for manual versus automated approaches. Manual processing required approximately 3–4 min per image (mean = 210 s), while CervSpineNet produced binary masks in 5–10 s (mean = 7.5 s), representing a ∼96.43% reduction in total processing time. Error bars indicate standard deviation.
Moreover, for inter-rater variability and to compare the model performance to human variability, we selected three of the test set images that the expert and trained annotators had already annotated and compared those segmentations to the masks predicted by the hybrid model using standard overlap and boundary metrics: Dice coefficient, Intersection-over-Union (IoU), 95th-percentile Hausdorff distance (HD95, in pixels), structural similarity index (SSIM), and volumetric similarity (VS).
Human–human agreement was high: expert vs. human 1 and expert vs. human two reached mean Dice scores of 0.91 (IoU ≈ 0.83–0.84, HD95 ≈ 6.3 pixels, SSIM ≈ 0.97, VS ≈ 0.96–0.98), while human 1 vs. human 2 yielded a similar Dice of 0.92 (IoU 0.85, HD95 6.24 pixels, SSIM 0.97, VS 0.98). On the same images, the proposed model reached a Dice of 0.92 vs. the expert (0.924), with higher IoU (0.86), lower HD95 (5.71 pixels), SSIM of 0.98, and VS of 0.99. Model vs. R1 and model vs. R2 comparisons were also in the same range: Dice ≈ 0.92, IoU ≈ 0.86, HD95 ≈ 6.8 and 5.5 pixels, SSIM ≈ 0.98. These results collectively indicate that model segmentations are as consistent with the expert as those of the additional human raters, and considering boundary accuracy, HD95 is slightly better than the average human–human variability on this sample, though we acknowledge this was based on three test cases.
4 Discussion
This section interprets the results of the study in the context of existing literature, emphasizing the methodological advances, clinical relevance, and potential impact of the developed CervSpineNet model. The discussion also outlines the limitations of the current work and identifies future directions for improving performance and generalizability.
4.1 Overview and model interpretation
This study introduces an automated segmentation framework for delineating cervical spinous processes from lateral spine radiographs. Because no public dataset existed for this task and manual annotation is labor-intensive, both a curated dataset and a novel hybrid model, CervSpineNet, were developed. Early zero-shot testing with standard segmentation models produced inconsistent and inaccurate delineations, confirming the need for a dedicated dataset capable of capturing the subtle morphology of spinous processes.
The CervSpineNet combines a Vision Transformer (ViT-B) encoder for global contextual modeling with a U-Net–style convolutional decoder for fine-grained boundary reconstruction (Chen J. et al., 2024; Al-Antari et al., 2025). This hybrid design maintains full spatial resolution by remaining fully convolutional, ensuring that predicted masks match input dimensions—a practical advantage for radiological workflows. Integrated residual and squeeze-and-excitation blocks selectively enhance task-relevant features and stabilize training. Collectively, these design elements allow the model to generate binary masks that closely align with ground truth and exhibit lower structural error, as confirmed across all dataset variants.
Although CervSpineNet follows the overall concept of a transformer–CNN hybrid, its architecture is specifically tailored to the demands of cervical spine X-ray segmentation. We decouple the SAM ViT-B image encoder and fine-tune it on 1,024 × 1,024 radiographs, leveraging its large-scale pretraining to capture long-range anatomical context along the full C1–C7 column. On top of these rich features, we employ a compact U-Net style decoder with residual and squeeze-and-excitation blocks, sharpening local boundaries and preserving thin, curved spinous process shapes. Training is driven by a compound loss that combines Dice overlap, Focal Tversky, Hausdorff distance–transform and SSIM terms to jointly encourage correct vertebral coverage, suppression of false positives and accurate contour geometry. In our experiments (Section 3.1, Tables 1–3), this SAM-based hybrid consistently outperforms standard U-Net, DeepLabV3+, a fine-tuned full SAM model and a text-guided SegFormer, indicating that the proposed combination of encoder, decoder and loss function confers a tangible advantage for this specific small-structure segmentation task.
Overall, the architectural ablations demonstrate that both encoder choice and decoder design substantially influence segmentation performance, albeit in complementary ways. Enhancing the decoder with residual and squeeze-and-excitation blocks consistently improves overlap, volumetric, and boundary metrics, indicating that richer decoding capacity is critical for refining fine spinal structures and suppressing noise. In contrast, replacing a purely convolutional encoder with a ViT-B transformer encoder yields the largest performance gains—particularly for boundary accuracy—highlighting the importance of long-range contextual reasoning and global shape modeling in cervical spine anatomy. Together, these results validate the design of CervSpineNet: the combination of a transformer-based encoder with an enriched decoder provides measurable advantages, while simpler components remain competitive options for lower-complexity or resource-constrained settings.
Loss ablations similarly show that region-based and structure-aware supervision outperforms Dice alone. Adding either Focal Tversky or SSIM to Dice produces consistent improvements in Dice, IoU, and SSIM, underscoring the value of class-imbalance handling and structural consistency for stable mask learning. The Dice + SSIM configuration is particularly effective for overlap and volumetric similarity, reflecting the benefit of enforcing local textural coherence in postoperative spine X-rays. These trends support our use of a compound loss: Dice and FT promote reliable region overlap and recall, SSIM preserves fine anatomical detail, and the HD95 surrogate sharpens boundaries by explicitly penalizing outlier deviations.
4.2 Performance evaluation and clinical relevance
Across the original, CLAHE-enhanced, and augmented datasets, CervSpineNet consistently outperformed CNN-based and transformer-based baselines on Dice, IoU, SSIM, HD95, and volumetric similarity metrics. Statistical testing confirmed that these improvements were significant, highlighting the robustness and reproducibility of the approach. Notably, the model accurately segmented difficult lower cervical levels (C6–C7), which are often challenging even for experts, underscoring its ability to generalize to low-contrast or ambiguous regions.
To the best of our knowledge, this work provides the first public dataset and segmentation framework dedicated to 2D X-ray–based cervical spinous process delineation. Accurate and efficient segmentation of these structures has broad clinical value. In Anterior Cervical Discectomy and Fusion (ACDF), precise localization of spinous processes supports surgical planning and fusion evaluation (Neifert et al., 2020; Martin et al., 2025). Beyond ACDF, automatic spinous process segmentation could aid in the assessment of vertebral alignment after trauma, the monitoring of spinal deformities such as scoliosis or kyphosis, and image-guided navigation during minimally invasive procedures.
From a workflow standpoint, CervSpineNet offers substantial efficiency gains. It reduces average manual segmentation time from approximately 3–4 min per image to 5–10 s, representing a ∼96% time reduction. The lightweight architecture (∼345 MB) and CPU-compatible inference make the model suitable for clinical integration without specialized hardware. Collectively, these features position CervSpineNet as a practical, reproducible tool for advancing AI-assisted cervical spine imaging and analysis.
4.3 Limitations and future work
Despite strong performance, several limitations warrant consideration. First, training the hybrid architecture is more computationally demanding than lighter CNN baselines due to the transformer encoder and multi-stage decoding. Model performance may also be sensitive to threshold selection, class imbalance, and annotation noise, and the fixed-size image resizing may introduce subtle aspect-ratio bias. Although CervSpineNet performs well on internal data, external validation across different scanners, institutions, and acquisition protocols remains necessary to assess generalization and potential domain shift. Additionally, while we report strong segmentation metrics (Dice, IoU, HD95, etc.), we did not directly evaluate downstream clinical indices—such as spinous-process–derived angular measures or deformity parameters. Establishing links between CervSpineNet’s segmentations and clinically meaningful quantitative metrics will be essential for demonstrating translational impact.
Future optimization efforts should investigate encoder freezing, mixed-precision training, gradient checkpointing, or knowledge distillation to reduce computational burden. Boundary accuracy may be further improved through boundary-aware decoders, attention-gated skip connections, or uncertainty-based quality control mechanisms. Expanding the dataset to incorporate diverse imaging protocols, patient populations, and post-operative presentations will enhance robustness. Extending the framework to multi-view radiographs or 3D modalities (CT, MRI) may also enrich anatomical representation and support more complex clinical tasks.
Prospective clinical validation and workflow-integrated deployment studies will be critical for determining real-world utility in surgical planning, postoperative monitoring, and routine spine care. Beyond surgical settings, automated spinous-process segmentation may facilitate longitudinal deformity surveillance, trauma assessment, and AI-assisted spine navigation. These directions position CervSpineNet as a promising foundation for scalable, clinically relevant spine-imaging solutions.
5 Conclusion
This study presents CervSpineNet, a hybrid transformer–CNN framework for automated segmentation of cervical spinous processes on lateral X-ray images, along with the first curated dataset developed specifically for this task. By coupling a ViT-B encoder that captures global anatomical context with a lightweight convolutional decoder optimized for fine structural reconstruction, CervSpineNet achieves strong performance, with mean Dice scores exceeding 0.93 and SSIM values above 0.98 across multiple dataset variants. Statistical comparisons demonstrate significant improvements over established baselines, and the model reduces manual annotation time by approximately 96%, producing accurate binary masks within 5–10 s on standard clinical hardware.
With its high accuracy, compact computational footprint, and openly accessible dataset, CervSpineNet offers a practical and reproducible foundation for future clinical integration and methodological research. The framework has potential applications in spine imaging, surgical planning, postoperative monitoring, and quantitative radiology, and it provides a scalable platform for advancing automated analysis of cervical spine anatomy.
Data availability statement
The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.
Ethics statement
Ethical approval was not required for the study involving humans in accordance with the local legislation and institutional requirements. Written informed consent to participate in this study was not required from the participants or the participants’ legal guardians/next of kin in accordance with the national legislation and the institutional requirements.
Author contributions
JS: Validation, Formal Analysis, Writing – review and editing, Software, Data curation, Writing – original draft, Conceptualization, Visualization, Methodology. LM: Visualization, Writing – original draft, Validation, Formal Analysis, Writing – review and editing, Data curation, Software, Methodology. AN: Writing – original draft, Data curation, Writing – review and editing, Validation, Methodology, Formal Analysis, Visualization, Software. AM: Methodology, Formal Analysis, Software, Writing – review and editing, Data curation, Conceptualization. JB: Writing – review and editing, Visualization, Formal Analysis, Writing – original draft, Methodology, Data curation. IP: Writing – review and editing, Data curation, Methodology. SY: Data curation, Validation, Conceptualization, Writing – review and editing. ChM: Data curation, Resources, Writing – review and editing, Conceptualization. CaM: Funding acquisition, Writing – review and editing, Conceptualization, Writing – original draft, Investigation, Resources, Project administration, Formal Analysis, Validation, Supervision, Visualization, Methodology.
Funding
The author(s) declared that financial support was received for this work and/or its publication. The work was supported by National Institutes of Health grant R35GM152245, the National Science Foundation CAREER award 1944247, and the Chan Zuckerberg Initiative award 253558 to CaM. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
Conflict of interest
The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Generative AI statement
The author(s) declared that generative AI was not used in the creation of this manuscript.
Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
Ahmad, N., Strand, R., Sparresäter, B., Tarai, S., Lundström, E., Bergström, G., et al. (2023). Automatic segmentation of large-scale CT image datasets for detailed body composition analysis. BMC Bioinforma. 24, 346. doi:10.1186/s12859-023-05462-2
Al-Antari, M. A., Al-Tam, R. M., Al-Hejri, A. M., Al-Huda, Z., Lee, S., Yıldırım, Ö., et al. (2025). A hybrid segmentation and classification CAD framework for automated myocardial infarction prediction from MRI images. Sci. Rep. 15, 14196. doi:10.1038/s41598-025-98893-1
Antonelli, M., Reinke, A., Bakas, S., Farahani, K., Kopp-Schneider, A., Landman, B. A., et al. (2022). The medical segmentation decathlon. Nat. Commun. 13, 4128. doi:10.1038/s41467-022-30695-9
Ariyaratne, S., Jenko, N., Iyengar, K. P., Davies, M., Azzopardi, C., Hughes, S., et al. (2024). Anatomy and pathologies of the spinous process. Diseases 12, 302. doi:10.3390/diseases12120302
Ashkani Chenarlogh, V., Shabanzadeh, A., Ghelich Oghli, M., Sirjani, N., Farzin Moghadam, S., Akhavan, A., et al. (2022). Clinical target segmentation using a novel deep neural network: double attention Res-U-Net. Sci. Rep. 12, 6717. doi:10.1038/s41598-022-10429-z
Azad, R., Khodapanah Aghdam, E., Rauland, A., Jia, Y., Haddadi Avval, A., Bozorgpour, A., et al. (2022). Medical image segmentation review: the success of U-Net. Available online at: https://ui.adsabs.harvard.edu/abs/2022arXiv221114830A (Accessed November 01, 2022).
Bakurov, I., Buzzelli, M., Schettini, R., Castelli, M., and Vanneschi, L. (2022). Structural similarity index (SSIM) revisited: a data-driven approach. Expert Syst. Appl. 189, 116087. doi:10.1016/j.eswa.2021.116087
Bao, J., Tan, Z., Sun, Y., Xu, X., Liu, H., Cui, W., et al. (2025). Deep ensemble learning-driven fully automated multi-structure segmentation for precision craniomaxillofacial surgery. Front. Bioeng. Biotechnol. 13, 1580502. doi:10.3389/fbioe.2025.1580502
Bichri, H., Chergui, A., and Hain, M. (2024). Investigating the impact of train/test split ratio on the performance of pre-trained models with custom datasets. Int. J. Adv. Comput. Sci. Appl. 15. doi:10.14569/ijacsa.2024.0150235
Bucher, M., Vu, T.-H., Cord, M., and Pérez, P. (2019). Zero-shot semantic segmentation. Available online at: https://ui.adsabs.harvard.edu/abs/2019arXiv190600817B (Accessed June 01, 2019).
Chen, J. J., Branstetter, B. F. T., and Welch, W. C. (2006). Multiple posterior vertebral fusion abnormalities: a case report and review of the literature. AJR Am. J. Roentgenol. 186, 1256–1259. doi:10.2214/AJR.04.1874
Chen, J., Mei, J., Li, X., Lu, Y., Yu, Q., Wei, Q., et al. (2024a). TransUNet: rethinking the U-Net architecture design for medical image segmentation through the lens of transformers. Med. Image Anal. 97, 103280. doi:10.1016/j.media.2024.103280
Chen, Y., Mo, Y., Readie, A., Ligozio, G., Mandal, I., Jabbar, F., et al. (2024b). VertXNet: an ensemble method for vertebral body segmentation and identification from cervical and lumbar spinal X-rays. Sci. Rep. 14, 3341. doi:10.1038/s41598-023-49923-3
Cramer, G. D. (2014). “General characteristics of the spine,” in Clinical anatomy of the spine, spinal cord, and Ans (Elsevier), 15–64.
Du, G., Cao, X., Liang, J., Chen, X., and Zhan, Y. (2020). Medical image segmentation based on U-Net: a review. J. Imaging Sci. Technol. 64, 020508. doi:10.2352/j.imagingsci.technol.2020.64.2.020508
Farooqi, R. R., Mehmood, M., and Kotwal, H. A. (2017). Hyperplasia of lamina and spinous process of C5 vertebrae and associated hemivertebra at C4 level. J. Orthop. Case Rep. 7, 79–81. doi:10.13107/jocr.2250-0685.698
Galbusera, F., and Cina, A. (2024). Image annotation and curation in radiology: an overview for machine learning practitioners. Eur. Radiol. Exp. 8, 11. doi:10.1186/s41747-023-00408-y
Gao, Y., Zhou, M., and Metaxas, D. N. (2021). “UTNet: a hybrid transformer architecture for medical image segmentation,” in Medical image computing and computer assisted intervention – MICCAI 2021 (Cham: Springer International Publishing), 61–71.
Goceri, E. (2023). Medical image data augmentation: techniques, comparisons and interpretations. Artif. Intell. Rev. 56, 1–45. doi:10.1007/s10462-023-10453-z
Gould, H., Sohail, O. A., and Haines, C. M. (2020). Anterior cervical discectomy and fusion: techniques, complications, and future directives. Semin. Spine Surg. 32, 100772. doi:10.1016/j.semss.2019.100772
Ham, D.-W., Choi, Y.-S., Yoo, Y., Park, S.-M., and Song, K.-S. (2023). Measurement of interspinous motion in dynamic cervical radiographs using a deep learning-based segmentation model. J. Neurosurg. Spine 39, 329–334. doi:10.3171/2023.5.SPINE23293
Hollon, T. C., Pandian, B., Adapa, A. R., Urias, E., Save, A. V., Khalsa, S. S. S., et al. (2020). Near real-time intraoperative brain tumor diagnosis using stimulated Raman histology and deep neural networks. Nat. Med. 26, 52–58. doi:10.1038/s41591-019-0715-9
Hou, X., Chen, P., and Gu, H. (2024). LM-DeeplabV3+: a lightweight image segmentation algorithm based on multi-scale feature interaction. Appl. Sci. 14, 1558. doi:10.3390/app14041558
Kaiser, J. T., Reddy, V., Launico, M. V., and Lugo-Pico, J. G. (2025). “Anatomy, head and neck: cervical vertebrae,” in StatPearls (Treasure Island (FL): StatPearls Publishing LLC).
Ketenci Çay, F., Yeşil, Ç., Çay, O., Yılmaz, B. G., Özçini, F. H., and İlgüy, D. (2025). DeepLabv3 + method for detecting and segmenting apical lesions on panoramic radiography. Clin. Oral Investig. 29, 101. doi:10.1007/s00784-025-06156-0
Khan, A. R., and Khan, A. (2025). Multi-axis vision transformer for medical image segmentation. Eng. Appl. Artif. Intell. 158, 111251. doi:10.1016/j.engappai.2025.111251
Kirillov, A., Mintun, E., Ravi, N., Mao, H., Rolland, C., Gustafson, L., et al. (2023). Segment anything. Available online at: https://ui.adsabs.harvard.edu/abs/2023arXiv230402643K (Accessed April 01, 2023).
Li, J., Ma, X., and Shi, Y. (2025). Multi-label conditioned diffusion for cardiac MR image augmentation and segmentation. Bioeng. (Basel) 12, 812. doi:10.3390/bioengineering12080812
Lones, M. A. (2021). How to avoid machine learning pitfalls: a guide for academic researchers. arXiv [cs.LG]. doi:10.48550/arXiv.2108.02497
Ma, J., He, Y., Li, F., Han, L., You, C., and Wang, B. (2024). Segment anything in medical images. Nat. Commun. 15, 654. doi:10.1038/s41467-024-44824-z
Martin, C. T., Yoon, S. T., Alluri, R. K., Benzel, E. C., Bono, C. M., Cho, S. K., et al. (2025). How reliable is the assessment of fusion status following ACDF using dynamic flexion-extension radiographs? Glob. Spine J. 15, 2450–2457. doi:10.1177/21925682241303107
Mazurowski, M. A., Dong, H., Gu, H., Yang, J., Konz, N., and Zhang, Y. (2023). Segment anything model for medical image analysis: an experimental study. Med. Image Anal. 89, 102918. doi:10.1016/j.media.2023.102918
Mostafa, R. R., Khedr, A. M., Aghbari, Z. A., Afyouni, I., Kamel, I., and Ahmed, N. (2024). Medical image segmentation approach based on hybrid adaptive differential evolution and crayfish optimizer. Comput. Biol. Med. 180, 109011. doi:10.1016/j.compbiomed.2024.109011
Mu, Y., Sun, J., and He, J. (2022). “The combined focal cross entropy and dice loss function for segmentation of protein secondary structures from cryo-EM 3D density maps,” in 2022 IEEE International Conference on Bioinformatics and Biomedicine (BIBM) (IEEE).
Muraina, I. (2022). “Ideal dataset splitting ratios in machine learning algorithms: general concerns for data scientists and data analysts,” in 7th International Mardin Artuklu Scientific Researches Conference. (Mardin, Turkey).
Muthukrishnan, V., Jaipurkar, S., and Damodaran, N. (2024). Continuum topological derivative - a novel application tool for segmentation of CT and MRI images. Neuroimage Rep. 4, 100215. doi:10.1016/j.ynirp.2024.100215
Neifert, S. N., Martini, M. L., Yuk, F., Mcneill, I. T., Caridi, J. M., Steinberger, J., et al. (2020). Predicting trends in cervical spinal surgery in the United States from 2020 to 2040. World Neurosurg. 141, e175–e181. doi:10.1016/j.wneu.2020.05.055
Ouyang, D., He, B., Ghorbani, A., Yuan, N., Ebinger, J., Langlotz, C. P., et al. (2020). Video-based AI for beat-to-beat assessment of cardiac function. Nature 580, 252–256. doi:10.1038/s41586-020-2145-8
Posthuma De Boer, J., Van Wulfften Palthe, A. F. Y., Stadhouder, A., and Bloemers, F. W. (2016). The clay shoveler’s fracture: a case report and review of the literature. J. Emerg. Med. 51, 292–297. doi:10.1016/j.jemermed.2016.03.020
Ramachandran, P., Eswarlal, T., Lehman, M., and Colbert, Z. (2023). Assessment of optimizers and their performance in autosegmenting lung tumors. J. Med. Phys. 48, 129–135. doi:10.4103/jmp.jmp_54_23
Ran, Y., Qin, W., Qin, C., Li, X., Liu, Y., Xu, L., et al. (2024). A high-quality dataset featuring classified and annotated cervical spine X-ray atlas. Sci. Data 11, 625. doi:10.1038/s41597-024-03383-0
Riew, K. D., Ecker, E., and Dettori, J. R. (2010). Anterior cervical discectomy and fusion for the management of axial neck pain in the absence of radiculopathy or myelopathy. Evid. Based Spine Care J. 1, 45–50. doi:10.1055/s-0030-1267067
Ronneberger, O., Fischer, P., and Brox, T. (2015). U-Net: Convolutional networks for biomedical image segmentation. Springer International Publishing, 234–241.
Shim, J.-H., Kim, W. S., Kim, K. G., Yee, G. T., Kim, Y. J., and Jeong, T. S. (2022). Evaluation of U-Net models in automated cervical spine and cranial bone segmentation using X-ray images for traumatic atlanto-occipital dislocation diagnosis. Sci. Rep. 12, 21438. doi:10.1038/s41598-022-23863-w
Son, W. J., Ahn, S. J., Lee, J. Y., and Lee, H. (2024). Automated brain segmentation on computed tomographic images using perceptual loss based convolutional neural networks. Investig. Magn. Reson. Imaging 28, 193. doi:10.13104/imri.2024.0023
Taghanaki, S. A., Zheng, Y., Kevin Zhou, S., Georgescu, B., Sharma, P., Xu, D., et al. (2019). Combo loss: handling input and output imbalance in multi-organ segmentation. Comput. Med. Imaging Graph. 75, 24–33. doi:10.1016/j.compmedimag.2019.04.005
Tejani, A. S., Klontzas, M. E., Gatti, A. A., Mongan, J. T., Moy, L., Park, S. H., et al. (2024). Checklist for artificial intelligence in medical imaging (CLAIM): 2024 update. Radiol. Artif. Intell. 6, e240300. doi:10.1148/ryai.240300
Van Santbrink, E., Schuermans, V., Cerfonteijn, E., Breeuwer, M., Smeets, A., Van Santbrink, H., et al. (2025). AI-assisted image recognition of cervical spine vertebrae in dynamic X-ray recordings. Bioeng. (Basel) 12, 679. doi:10.3390/bioengineering12070679
Wang, G., Zuluaga, M. A., Li, W., Pratt, R., Patel, P. A., Aertsen, M., et al. (2019). DeepIGeoS: a deep interactive geodesic framework for medical image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 41, 1559–1572. doi:10.1109/TPAMI.2018.2840695
Wang, J., Lv, P., Wang, H., and Shi, C. (2021a). SAR-U-Net: squeeze-and-excitation block and atrous spatial pyramid pooling based residual U-Net for automatic liver segmentation in computed tomography. Comput. Methods Programs Biomed. 208, 106268. doi:10.1016/j.cmpb.2021.106268
Wang, S., Li, C., Wang, R., Liu, Z., Wang, M., Tan, H., et al. (2021b). Annotation-efficient deep learning for automatic medical image segmentation. Nat. Commun. 12, 5915. doi:10.1038/s41467-021-26216-9
Wang, Z., Wu, Z., Agarwal, D., and Sun, J. (2022). MedCLIP: contrastive learning from unpaired medical images and text. Proc. Conf. Empir. Methods Nat. Lang. Process 2022, 3876–3887. doi:10.18653/v1/2022.emnlp-main.256
Wang, H., Lu, J., Yang, S., Xiao, Y., He, L., Dou, Z., et al. (2025a). Cervical vertebral body segmentation in X-ray and magnetic resonance imaging based on YOLO-UNet: automatic segmentation approach and available tool. Digit. Health 11, 20552076251347695. doi:10.1177/20552076251347695
Wang, X., Zhu, Y., Cui, Y., Huang, X., Guo, D., Mu, P., et al. (2025b). Lightweight multi-stage aggregation transformer for robust medical image segmentation. Med. Image Anal. 103, 103569. doi:10.1016/j.media.2025.103569
Wolfrath, N., Wolfrath, J., Hu, H., Banerjee, A., and Kothari, A. N. (2024). Stronger baseline models -- A key requirement for aligning machine learning research with clinical utility. Available online at: https://ui.adsabs.harvard.edu/abs/2024arXiv240912116W (Accessed September 01, 2024).
Xu, Y., He, X., Xu, G., Qi, G., Yu, K., Yin, L., et al. (2022). A medical image segmentation method based on multi-dimensional statistical features. Front. Neurosci. 16, 1009581. doi:10.3389/fnins.2022.1009581
Yang, T., Lu, X., Yang, L., Yang, M., Chen, J., and Zhao, H. (2024). Application of MRI image segmentation algorithm for brain tumors based on improved YOLO. Front. Neurosci. 18, 1510175. doi:10.3389/fnins.2024.1510175
Yuan, Y., and Cheng, Y. (2024). Medical image segmentation with UNet-based multi-scale context fusion. Sci. Rep. 14, 15687. doi:10.1038/s41598-024-66585-x
Zhang, J., Li, F., Zhang, X., Wang, H., and Hei, X. (2024). Automatic medical image segmentation with vision transformer. Appl. Sci. (Basel) 14, 2741. doi:10.3390/app14072741
Zhang, X., Zhang, Y., Li, Z., Song, Y., Chen, S., Mao, Z., et al. (2025). A real-time cell image segmentation method based on multi-scale feature fusion. Bioengineering 12, 843. doi:10.3390/bioengineering12080843
Keywords: artificial intelligence, automated musculoskeletal landmark detection, cervical spine segmentation, cervical spinous process dataset, deep learning in radiology, hybrid transformer–CNN architecture, machine learning, radiology workflow automation
Citation: Sawant JS, Moukheiber L, Nair A, Mahajan A, Byun J, Pichaimani I, Yoon ST, Martin CT and Mitchell CS (2026) CervSpineNet: a hybrid deep learning-based approach for the segmentation of cervical spinous processes. Front. Bioeng. Biotechnol. 13:1733689. doi: 10.3389/fbioe.2025.1733689
Received: 27 October 2025; Accepted: 18 December 2025;
Published: 19 January 2026.
Edited by:
Christopher G. Provatidis, National Technical University of Athens, GreeceReviewed by:
Zifei Liang, New York University, United StatesMoxin Zhao, Chinese Academy of Sciences (CAS), China
Copyright © 2026 Sawant, Moukheiber, Nair, Mahajan, Byun, Pichaimani, Yoon, Martin and Mitchell. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Cassie S. Mitchell, Y2Fzc2llLm1pdGNoZWxsQGJtZS5nYXRlY2guZWR1
Lama Moukheiber1,2