Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Med., 23 July 2025

Sec. Precision Medicine

Volume 12 - 2025 | https://doi.org/10.3389/fmed.2025.1582439

gamUnet: designing global attention-based CNN architectures for enhanced oral cancer detection and segmentation


Jinyang Zhang&#x;Jinyang Zhang1Hongxin Ding,&#x;Hongxin Ding1,2Runchuan ZhuRunchuan Zhu1Weibin Liao,Weibin Liao1,2Junfeng Zhao,Junfeng Zhao1,2Min GaoMin Gao3Xiaoyun Zhang
Xiaoyun Zhang3*
  • 1School of Computer Science, Peking University, Beijing, China
  • 2Key Laboratory of High Confidence Software Technologies, Ministry of Education, Peking University, Beijing, China
  • 3School and Hospital of Stomatology, Peking University, Beijing, China

Introduction: Oral squamous cell carcinoma (OSCC) is a significant global health burden, where timely and accurate diagnosis is essential for improved patient outcomes. Conventional diagnosis relies on manual evaluation of hematoxylin and eosin (H&E)-stained slides, a time-consuming process requiring specialized expertise and prone to variability. While deep learning methods, especially convolutional neural networks (CNNs), have advanced automated analysis of histopathological images for cancerous tissues in various body parts, OSCC presents unique challenges. Its infiltrative growth patterns and poorly defined boundaries, coupled with the complex architecture of the oral cavity, make accurate segmentation particularly difficult. Traditional CNNs which sturggle to capture critical global contextual information often fail to distinguish the complex tissue structures in OSCC images.

Methods: To address these challenges, we propose a novel architecture called gamUnet, which integrates the Global Attention Mechanism (GAM) to enhance the model's ability to capture global cross-modal information. This allows the model to focus on key diagnostic regions while retaining detailed spatial information. Additionally, we introduce an extended model, gamResNet, to further improve OSCC detection performance. Both architectures show significant improvements in handling the unique challenges of oral cancer images.

Results: Extensive experiments on public datasets show that our GAM-enhanced architecture significantly outperforms conventional models, achieving superior accuracy, robustness, and efficiency in OSCC diagnosis.

Discussion: Our approach provides an effective tool for clinicians in diagnosing OSCC, reducing diagnostic variability, and ultimately contributing to improved patient care and treatment planning.

1 Introduction

Oral squamous cell carcinoma (OSCC) is an aggressive malignancy characterized by high global incidence and mortality (1, 2). OSCC arises from abnormal cellular proliferation within the oral epithelium and represents a significant global health burden, accounting for over 90% of oral malignancies (35). Annually, more than 657,000 new OSCC cases are diagnosed, resulting in ~330,000 deaths (6). Timely and accurate diagnosis is critical, as survival rates drastically decrease from 80% in early-stage cases to 20%–30% in advanced stages (7). Currently, OSCC diagnosis primarily relies on microscopic examination of hematoxylin and eosin (H&E)-stained tissue slides (8). However, this manual process is highly labor-intensive, time-consuming and subject to inter-observer variability, which underscores the necessity for automated solutions to improve diagnostic accuracy and consistency.

Deep learning, particularly convolutional neural networks (CNNs), has emerged as a powerful tool for medical image analysis. For instance, U-Net (9) and its variants have achieved notable success in segmenting medical images across modalities such as CT, MRI, and histopathology (1013). Similarly, ResNet (14) is widely applied for classification tasks, effectively identifying diseases including colorectal cancer, brain tumors, and glioma subtypes (1517). Building on these successes, CNNs hold promise for OSCC image analysis, potentially enhancing diagnostic efficiency. Previous studies on OSCC segmentation have predominantly utilized variants of U-Net or other CNN-based models, achieving promising results through enhanced architectures, multi-stage approaches, and feature extraction techniques (1821). For OSCC classification, architectures and transfer learning strategies have been employed (2227), reporting high accuracy but often facing limited sensitivity, robustness, and generalizability due to small datasets or inconsistent preprocessing. Overall, despite these advances, existing methods for OSCC still encounter significant challenges in effectively balancing accuracy, computational efficiency, and robustness in clinical settings.

The complexity of oral cavity tissue structures introduces unique challenges to OSCC imaging analysis, differentiating it from other cancers characterized by more homogeneous tissues (28). OSCC typically exhibits invasive growth patterns with ill-defined boundaries (29, 30), frequently infiltrating adjacent complex tissues including mucous membranes, bone, and soft tissues (3133). This invasive behavior complicates tumor boundary delineation, as critical pathological features may span multiple image regions, making it particularly challenging to delineate the exact borders of the tumor. Traditional CNN approaches, like U-Net, ResNet, and exisiting OSCC methods building upon them, relying heavily on localized convolutional operations, often fall short in effectively capturing the global context and long-range dependencies critical for accurately segmenting OSCC. As a result, current automated solutions for OSCC segmentation and classification remain suboptimal

A possible solution is utilizing the attention mechanism. Recently, research has explored incorporating attention mechanisms to enhance CNN performance in image tasks. SENet (34) used channel attention but lacked spatial awareness. CBAM (35) and BAM (36) incorporated spatial attention but couldn't fully capture interactions across all dimensions, and TAM (37) improved this by addressing two dimensions at a time. GAM (38) further advanced this by simultaneously capturing cross-dimensional interactions, making it ideal for handling complex medical images.

However, the application of attention mechanisms, including GAM, to OSCC H&E-stained histopathological image segmentation and classification remains underexplored. Existing research primarily focuses on refining CNN architectures or applying transfer learning without fully addressing the inherent challenges in OSCC imaging, as analyzed aforementioned.

To address these challenges, we propose a novel approach designed specifically for OSCC. We introduce gamUnet, an innovative model integrating the Global Attention Mechanism into the U-Net architecture to enhance the capture of global context and cross-dimensional interactions, crucial for delineating complex OSCC tissue structures. Furthermore, we propose gamResNet, extending our approach by incorporating residual networks enhanced by GAM to improve OSCC detection. Our proposed models effectively address existing limitations, significantly improving the segmentation and classification accuracy and robustness for OSCC.

Our key contributions are as follows:

• We propose gamUnet and gamResNet, two novel architectures integrating GAM for improved OSCC segmentation and detection.

• We demonstrate the advantage of global attention mechanism in capturing complex tissue structures and delineating tumor regions, providing insight into the benefits of GAM in medical image analysis.

• Our experimental results validate that our models significantly outperform traditional CNN approaches, offering a promising tool for clinical application in OSCC diagnosis.

2 Related work

2.1 Deep learning for OSCC analysis

Past studies have examined various deep learning architectures for OSCC segmentation. Martino et al. (18), for instance, introduced the Oral Cancer Annotated (ORCA) dataset and evaluated multiple CNN-based architectures on the dataset, including SegNet, U-Net, and U-Net variants with VGG-16 and ResNet50 encoders, which showed promising results. Dos et al. (19) incorporated color space features into a U-Net-based model for better tumor region identification and background removal, and achieved an impressive accuracy of 97.6% on the Oral Cavity-Derived Cancer(OCDC) dataset. Pennisi et al. (20) designed a Multi-encoder U-Net where input images were divided into tiles and processed by separate encoders before merging, allowing multi-region feature fusion. Musulin et al. (21) proposed a two-stage system for diagnosing OSCC, demonstrating effectiveness in multiclass grading and segmentation. Beyond segmentation, classification of OSCC using deep learning has also seen notable developments. Wang et al. (1) introduced a semi-supervised boundary-aware U-Net with transformation consistency and contrastive learning, outperforming fully-supervised models on the OCDC dataset in low-label regimes. Ünsal et al. (39) applied a U2-Net encoder-decoder architecture on private OSCC datasets and reported robust performance with a Dice coefficient of 0.86. Shah et al. (40) proposed OCANet, which incorporates both local and global attention branches to enrich contextual understanding, achieving 86.1% Dice and 77.1% mIoU on ORCA. A range of CNN architectures has been applied to H&E-stained histopathological images for oral cancer detection (2227), leveraging transfer learning and hybrid models to improve accuracy.

Despite these advancements, OSCC segmentation remains challenging due to the inherent complex anatomical structure of the oral cavity, invasive tumor growth, and ill-defined tumor margins (2830). While existing OSCC-specific methods have demonstrated promising results, they are often not explicitly designed to address these unique challenges, lacking mechanisms to effectively capture global contextual information, resulting in suboptimal performance. In contrast, our work specifically addresses this challenge utilizing the Global Attention Mechanism.

2.2 Attention mechanism in CNNs

In recent years, attention mechanism has garnered significant interest in enhancing CNN-based models to improve the performance of image classification and segmentation tasks, by allowing them to focus on relevant features within an image. Early examples include SENet (34), which pioneered this field by using the channel attention to highlight significant features while suppressing irrelevant ones. However, its primary focus on channel attention restricted its capacity to address spatial information effectively. More recent approaches, such as CBAM (35) and BAM (36), have introduced spatial attention alongside channel attention. Their designed Convolutional Block Attention Module (CBAM) and Bottleneck Attention Module (BAM) have improved model performance in a variety of applications. However, these models still fall short in capturing interactions between channel, spatial width, and spatial height, which are critical for understanding complex medical images. Recognizing the importance of cross-dimension interactions, Misra et al. (37) introduced the triplet attention module (TAM), which considered the relationships between channel, spatial width, and spatial height. Although TAM improved efficiency, it still applied attention to only two dimensions at a time, limiting its ability to fully capture interactions across all three dimensions. Addressing this, Liu et al. (38) introduced the Global Attention Mechanism (GAM) to enhance feature extraction by capturing cross-dimensional interactions across the channel, spatial width, and spatial height simultaneously, making it particularly suited for complex medical imaging tasks that demand comprehensive feature representation. Recently, attention modules have been directly embedded into OSCC segmentation architectures. OCANet (40), for instance, utilizes a dual-branch attention mechanism to integrate local fine-grained detail with global structural awareness.

Still, the potential of attention mechanisms in enhancing OSCC imaging analysis remains underexplored. Many of these models either employ decoupled attention paths or require extensive network modifications, leading to increased computational cost. In contrast, our proposed architecture utilizing GAM offers a lightweight yet expressive approach that integrates seamlessly with CNNs, enabling effective cross-dimensional feature fusion for OSCC segmentation and classification.

2.3 Comparative summary of related methods

To provide a clearer understanding of the landscape of OSCC analysis and attention mechanisms in convolutional networks, we summarize representative studies in Table 1. The table categorizes prior works based on dataset, model architecture, core strengths, and limitations. Notably, many OSCC-specific segmentation models leverage U-Net variants, often falling short in capturing comprehensive global context, which is critical for handling the anatomical complexity of OSCC. In parallel, the adoption of advanced attention modules like GAM remains limited in OSCC-focused studies. By proposing a new model architecture integrating GAM, we aim to address this gap and enhance both segmentation accuracy and contextual awareness with minimal architectural overhead.

Table 1
www.frontiersin.org

Table 1. Comparison of related OSCC segmentation and attention-based methods.

3 Methods

In this study, we introduce a novel model architecture gamUnet, and an extended model gamResNet, which integrate a Global Attention Mechanism into U-Net and ResNet architectures, respectively. These models specifically target the challenges posed by tumor infiltration and poorly defined lesion boundaries in OSCC histopathological image analysis. Unlike traditional CNN architectures, which excel at capturing local features but fail in modeling global context, the proposed GAM significantly enhances global contextual awareness and captures long-range dependencies.

3.1 Base architecture

We adopt U-Net as the base architecture for image segmentation and ResNet-18 for image classification, both of which have demonstrated strong performance in medical image analysis. U-Net is selected as the base architecture for segmentation due to its robust performance in medical imaging segmentation. The U-Net architecture employs an encoder-decoder structure, where the encoder progressively reduces the spatial resolution of the input image, capturing abstract features, while the decoder upscales these features to create high-resolution segmentation masks. Skip connections between corresponding layers of the encoder and decoder preserve fine-grained spatial details during the upsampling process. For classification, ResNet-18 (14) is adopted. Despite having fewer layers compared to deeper ResNet variants, ResNet-18 provides a balanced trade-off between computational efficiency and predictive performance, essential for deployment in clinical settings. ResNet-18, reduces the spatial resolution of the input through convolution and pooling layers, capturing high-level features. A key feature of ResNet-18 is the use of residual (skip) connections, which allow for the efficient training of deeper networks by bypassing certain layers, thereby addressing the vanishing gradient problem. In classification tasks, ResNet-18 outputs class probabilities after reducing the feature map through a fully connected layer.

3.2 Global attention mechanism

While both U-Net and ResNet-18 excel in capturing local features, they struggle with modeling global dependencies, a crucial aspect in tasks involving complex medical images in OSCC. To address their limitations in global context modeling, we integrate GAM into both architectures. GAM computes attention across channel and spatial dimensions, significantly enhancing feature representation.

Given a feature map FRC×H×W, GAM calculates channel attention Mc and spatial attention Ms as:

Mc(F)=σ(MLP(AvgPool(F))+MLP(MaxPool(F)))    (1)
Ms(F)=σ(Conv7×7([AvgPool(F);MaxPool(F)]))    (2)

Here, σ denotes the sigmoid activation, AvgPool and MaxPool represent global average and max pooling operations, respectively, and MLP denotes a multilayer perceptron for channel-wise computations. For an input feature map F1, GAM computes the intermediate feature map F2 and the final output F3 as follows:

F2=Mc(F1)F1    (3)
F3=Ms(F2)F2    (4)

where ⊗ represents element-wise multiplication.

The channel attention module captures relationships between channels, enhancing important features, while the spatial attention module emphasizes crucial spatial regions. GAM enables the model to focus on regions that are critical for diagnosis, thus enhancing performance, especially in tasks with complex tissue structures like OSCC.

3.3 Model variants

We propose two key model variants: gamUnet for segmentation tasks and gamResNet for classification tasks. These models incorporate the GAM module at different points in their architecture to enhance their ability to capture global context while preserving local feature details.

3.3.1 gamUnet

For segmentation, the GAM module is incorporated into each convolutional block of the U-Net encoder (Figure 1). Each encoder block comprises the following steps: First perform convolution Conv3 × 3 to obtain preliminary feature maps. Then, apply GAM to these feature maps to compute attention weights. Finally, multiply the attention-weighted features to refine and amplify significant regions. Refined features are passed through skip connections to the decoder, preserving both global context and local spatial details for accurate segmentation.

Figure 1
Flowchart of a neural network architecture featuring a Global Attention Module (GAM). The process includes convolutional layers, max pooling, up-convolution, and copy-and-crop operations. The input, a histological image, progresses through various stages, with channel and spatial attention mechanisms refining output features. Detailed block and arrow symbols represent each operation.

Figure 1. gamUnet.

3.3.2 gamResNet

For gamResNet, GAM modules replace the second convolution within each residual block of ResNet-18 (Figure 2). Specifically: First perform initial convolution Conv3 × 3 to extract preliminary features. Then replace second convolution with GAM, refining features by computing global attention weights. Finally, utilize the residual connection to combine GAM-enhanced features with the initial block input.

Figure 2
Flowchart of a convolutional neural network architecture for cancer detection. It starts with input features processed through GAM with channel and spatial attention, followed by multiple 3 Ă— 3 convolution and residual modules. After three stacked residual modules, the output undergoes average pooling and a fully connected layer to predict cancer. Arrows indicate skip connections, with dashed lines for downsampling. An inset shows a microscopic tissue image.

Figure 2. gamResNet.

Through these explicit technical refinements, our gamUnet and gamResNet architectures robustly address the complex and heterogeneous nature of OSCC histopathological images for enhanced segmentation and classification, significantly surpassing traditional CNN methods by effectively capturing critical local and global contextual information.

4 Experiments

In this section, we first describe in detail the datasets, training procedures, experimental setup and evaluation metrics, providing a clear explanation of how our models are trained and evaluated, including critical steps involved in data preparation and model training, including image pre-processing, feature extraction, and model optimization. We then present the main experimental results, and the ablation studies conducted to rigorously assess the effectiveness of our proposed gamUnet and gamResNet architectures for OSCC segmentation and classification tasks.

4.1 Datasets

4.1.1 Segmentation datasets

Table 2 presents the statistics of datasets utilized for segmentation task, Figure 3 presents the image examples.

Table 2
www.frontiersin.org

Table 2. Dataset statistics for segmentation.

Figure 3
Two sets of microscopic images showing tissue samples. The first set, labeled “OCDC Images,” includes two rectangular images with varying cellular structures and textures. The second set, labeled “ORCA Images,” contains two circular images showing denser, differently stained tissues.

Figure 3. Image examples of OCDC and ORCA datasets.

4.1.1.1 OCDC dataset

The OCDC dataset consists of 15 H&E-stained whole slide images (WSIs) derived from human patients diagnosed with OSCC. A total of 1,020 annotated image patches, each measuring 640 × 640 pixels, were extracted from these WSIs, with tumor regions manually marked by a pathologist. This dataset is specifically designed for the automated detection and segmentation of OSCC, featuring a training set of 840 image patches and a test set of 180 patches, all digitized at a 20 × magnification.

4.1.1.2 ORCA dataset

The ORCA dataset includes 200 H&E-stained WSIs sourced from the Cancer Genome Atlas (TCGA) and manually annotated by expert pathologists. Each WSI is composed of one or two cores with a size of 4,500 × 4,500 pixels, containing ground-truth data for tumor pixels. The dataset is split into two subsets: a validation set and a test set, each consisting of 100 core images. In this study, the validation set was utilized for training purposes. Because the original core images only the middle round part is a valid image, so we cut the original image into 1,200*1,200 patches in preprocessing. Then, in order to filter out the low-quality images located at the edges, we removed the patches whose corresponding mask black part of the larger than 80% of the patches, resulting in 728 training images and 796 validation images. The ORCA dataset presents a higher level of complexity due to its larger size and greater complexity, making it suitable for evaluating advanced segmentation models.

4.1.2 Classification datasets

Table 3 presents the statistics of the datasets utilized for classification, Figure 4 presents the image examples.

Table 3
www.frontiersin.org

Table 3. Dataset statistics for classification.

Figure 4
Histological images display skin tissue sections. The left image shows dense purple staining with undulating epidermal layers and distinct cellular structures. The right image reveals hair follicles in close detail with concentric layers, highlighted in vivid magenta and purple hues.

Figure 4. Image examples of the classification dataset. (Left) Normal epithelium. (Right) OSCC.

This dataset (41) comprises a total of 1,224 histopathological images, divided into two distinct sets with varying resolutions. The first set contains 89 images of normal oral epithelium and 439 images of OSCC, all captured at 100 × magnification. The second set includes 201 images of normal oral epithelium and 495 OSCC images, captured at 400 × magnification. All images were acquired using a Leica ICC50 HD microscope from H&E stained tissue slides. These slides were collected, prepared, and cataloged by medical professionals from 230 patients.

4.2 Evaluation metrics

We adopt standard evaluation metrics for both segmentation and classification tasks. Accuracy reflects the overall proportion of correctly classified instances and is defined as:

Accuracy=TP+TNTP+TN+FP+FN

Sensitivity (also referred to as recall in classification) quantifies the ability to correctly identify positive samples:

Sensitivity/recall=TPTP+FN

Specificity measures the proportion of actual negatives correctly identified:

Specificity=TNTN+FP

Precision evaluates the proportion of predicted positives that are truly positive:

Precision=TPTP+FP

The F1 score, commonly used in classification, is the harmonic mean of precision and sensitivity:

F1 score=2×precision×sensitivityprecision+sensitivity

For segmentation tasks, we also report the Dice coefficient, which shares the same formula as the F1 score:

Dice=2TP2TP+FP+FN

The Jaccard Index or Intersection over Union (IoU) assesses the overlap between predicted and ground truth regions:

IoU=TPTP+FP+FN

For classification, we further report the Area Under the Receiver Operating Characteristic Curve (AUROC), which reflects the model's ability to distinguish between classes by summarizing the trade-off between sensitivity and false positive rate across thresholds.

4.3 Baselines

Below are brief descriptions of the baseline algorithms. For image segmentation:

U-Net (9): an established encoder-decoder CNN architecture widely employed in medical image segmentation, characterized by its symmetrical U-shaped structure.

Attention U-Net (42): an extension of U-Net that incorporates attention mechanisms to focus on relevant regions and emphasize important features.

Depthwise U-Net (43): a variant of U-Net that uses depthwise separable convolutions to reduce the number of parameters and computational complexity.

Res U-Net (44): a U-Net model enhanced with residual connections, which help to alleviate the vanishing gradient problem and improve feature propagation through the network.

Shuffle U-Net (45): a lightweight U-Net variant that utilizes shuffle operations to improve feature mixing and maintain high segmentation accuracy with reduced computational cost.

For image classification:

ResNet (ResNet-18 and ResNet-50) (14): residual networks that use skip connections to enable deeper architectures with improved gradient flow.

MobileNet-V3-Small (46): a compact and efficient CNN model designed for mobile and edge devices, utilizing efficient layers like depthwise convolutions and squeeze-and-excitation blocks.

MobileNet-V2 (47): a lightweight CNN model featuring inverted residuals and linear bottlenecks, optimized for performance on mobile and low-resource devices.

MobileNet-V1 (48): the original MobileNet architecture that employs depthwise separable convolutions to reduce model size and complexity.

EfficientNet (EfficientNet-B0 and EfficientNet-B1) (49): utilizes compound scaling methods that systematically balance depth, width, and resolution, achieving superior accuracy with fewer parameters.

4.4 Training and implementation details

We train all models on both medical image segmentation datasets and medical image classification datasets, including the OCDC dataset, the ORCA dataset and 100X and 400X classification datasets. Prior to training, the original images from both datasets were resized to 256 × 256 pixels and augmented using random rotations, flips, and scaling to enhance model robustness. Each image undergoes the process shown in Figure 5.

Figure 5
Flowchart depicting image processing steps. A circular input image is magnified into patches, undergoing random rotation, flip, and cropping. The processes lead to a final square image.

Figure 5. Input image preprocess workflow.

Models were trained using binary cross-entropy loss and optimized using Adam optimizer. The training was conducted with a batch size of 32 and a learning rate of 0.01, using early stopping to prevent overfitting. Each model's performance was evaluated using key metrics introduced above. All experiments were implemented using PyTorch and conducted on an NVIDIA RTX 3090 GPU with 24GB memory. The classification models were trained for ~5.5 h, while the segmentation models required about 4 h on the OCDC dataset and 1.5 h on the smaller ORCA dataset.

5 Results

5.1 Experimental results for OSCC segmentation

The segmentation performance on both the OCDC and ORCA datasets, as shown in Tables 4, 5, highlights the effectiveness of our proposed model architecture gamUnet, outperforming traditional architectures such as plain U-Net, Attention-Unet, Depthwise-Unet, Res-Unet, and Shuffle-Unet. Given the class imbalance in the datasets, our primary evaluation metrics are the F1/Dice score and the Intersection-over-Union (IOU, also known as the Jaccard Index), which effectively measure segmentation accuracy under imbalanced scenarios. On the OCDC dataset, gamUnet achieves the highest F1/Dice score (0.9096) and IOU (0.8315), indicating enhanced ability to precisely delineate tumor regions while minimizing false predictions. While all models experience a decrease in performance on the more complex ORCA dataset, gamUnet maintained robust results with the highest F1/Dice score (0.7047) and IOU (0.5631), demonstrating its resilience against complex anatomical structures and tumor morphology. The superior performance of gamUnet highlights its capacity for capturing global context through the integrated GAM, effectively balancing local details with global semantic information crucial for OSCC segmentation tasks.

Table 4
www.frontiersin.org

Table 4. Segmentation performance comparison on OCDC dataset.

Table 5
www.frontiersin.org

Table 5. Segmentation performance comparison on OCRA dataset.

Figure 6 presents qualitative and visualized segmentation results on the OCDC dataset. Compared to baselines, gamUnet yields segmentation masks that more closely align with the ground truth, effectively capturing tumor boundaries. In contrast, other models exhibit either over-segmentation or structural inaccuracies, particularly in complex regions. These results further validate the advantage of incorporating global attention for precise and robust OSCC segmentation.

Figure 6
Two rows of images comparing segmentation results. The first column shows the original histological images. The second column shows ground truth (GT) segmentations. The remaining columns depict results from different segmentation methods: Ours, U-Net, AttU-Net, DepU-Net, ResU-Net, and ShufU-Net. Black and white contrast highlights differences in segmentation accuracy among methods.

Figure 6. Segmentation performance evaluation of various models for oral cancer detection using the OCDC dataset. The classes are defined as: white for carcinoma pixels and black for non-carcinoma pixels. The bar chart displays the average performance of each model on the test set.

5.2 Experimental results for OSCC classification

Tables 6, 7 detail the classification performance evaluated at 100X and 400X magnifications on the oral cancer classification datasets. Considering the dataset imbalance, AUROC and F1 Score are the key evaluation metrics. On the 100X dataset, our proposed gamResNet attained the highest AUROC (0.991), accuracy (0.902), and F1 Score (0.945), outperforming traditional baselines such as ResNet-50, MobileNet-V2, and EfficientNet-B1. This underscores gamResNet's superior capacity to accurately distinguish between positive and negative samples. Additionally, mobilenet-v1 and efficientnet-b0 exhibit competitive F1 scores (0.897 and 0.898, respectively), although they fall behind in terms of AUROC and accuracy, indicating a trade-off in model performance across different metrics.

Table 6
www.frontiersin.org

Table 6. Classification performance comparison on OCDC 100X dataset.

Table 7
www.frontiersin.org

Table 7. Classification performance comparison on OCDC 400X dataset.

On the 400X dataset, similar trends are observed. GamResNet again achieves the highest AUROC (0.924), accuracy (0.899), and F1 score (0.933), coupled with notably high sensitivity (0.991), which is essential for clinical cancer detection applications. Although efficientnet-b0 performs well with an AUROC of 0.875 and an F1 score of 0.863, it is still notably outperformed by gamResNet. The baseline ResNet-50 model achieves the highest precision (0.905) on the 400X dataset, indicating its capability to minimize false positives, but it falls behind in sensitivity and F1 score.

Collectively, these results demonstrate that integrating the GAM mechanism significantly enhances model robustness and generalization across varied magnification levels and dataset complexities, confirming the clinical utility of gamResNet for reliable OSCC classification.

5.3 Ablation studies

To investigate the contribution of the GAM in our architectures, we conducted ablation studies on both gamUnet and gamResNet.

For gamUnet, we selectively integrated the GAM modules into the encoder, decoder, or both components of the U-Net architecture. All experiments were conducted under identical training settings on the OCDC dataset. As shown in Table 8, the variant with GAM integrated solely in the encoder (gamUnet) achieved the best performance across key segmentation metrics, including accuracy (0.9294), F1/Dice score (0.9168), and Jaccard Index (0.8467). These results highlight the benefit of enhancing global contextual representation at the encoding stage. Conversely, applying GAM to both encoder and decoder improved specificity, suggesting enhanced ability in distinguishing background regions, albeit with a slight drop in overall accuracy. These findings support our design choice of encoder-focused GAM integration.

Table 8
www.frontiersin.org

Table 8. Segmentation performance comparison of gamUnet variants on OCDC.

For gamResNet, we explored how different placements of the GAM module within the residual blocks affect classification performance. The original design of gamResNet replaces the second convolution layer of each residual block with a GAM module. We compared this configuration against two variants: (1) replacing the first convolution instead of the second, and (2) replacing both convolutions with GAM. Results on OCDC 100X dataset are summarized in Table 9. The original gamResNet achieved the best trade-off across all metrics, with the highest AUROC, accuracy, and F1 score, while maintaining a balanced sensitivity and specificity. Replacing only the first convolution slightly reduced AUROC (0.895), indicating suboptimal background discrimination. Substituting both convolutions with GAM led to near-perfect sensitivity but drastically reduced specificity (0.125), suggesting severe over-sensitivity to positive class predictions. These results confirm the efficacy of selectively applying GAM to the second convolutional layer, balancing attention to relevant features without compromising classification reliability.

Table 9
www.frontiersin.org

Table 9. Classification performance comparison of gamResNet variants on the OCDC 100X dataset.

Collectively, our ablation studies across segmentation and classification tasks demonstrate that a targeted integration of GAM as designed in our gamUnet and gamResNet strikes the best balance between leveraging global attention and preserving task-specific structural inductive biases.

5.4 Statistical significance

To evaluate the statistical significance and robustness of our experimental results, we conducted four independent runs for each model configuration and report both the mean and standard deviation in the main results. The relatively small standard deviations across most metrics indicate high stability and reproducibility of our results, which also enhances tthe credibility of the comparisons. Additionally, for the OCDC segmentation task, we performed statistical analysis by calculating 95% confidence intervals and p-values for key metrics, including F1/Dice and IOU. As summarized in Table 10, our proposed model gamUnet achieves significantly higher F1/Dice and IOU scores with narrow confidence intervals and p-values well below 0.05. These findings confirm that the observed performance gains are not only consistent but also statistically significant, reinforcing the effectiveness and reliability of our approach in handling complex and imbalanced OSCC image analysis tasks.

Table 10
www.frontiersin.org

Table 10. Statistical significance analysis on OCDC segmentation results.

5.5 Model efficiency and computational cost

Regarding the computational overhead introduced by the GAM module, we provide a detailed comparison of model complexity in terms of parameter count across both classification and segmentation tasks.

For segmentation, Table 11 shows that our proposed gamUnet contains around 28.9 million parameters. Although this is higher than vanilla Unet (6.8M) and lightweight variants such as Shuffle Unet (0.61M), it remains substantially more efficient than heavy models that similarly incorporates attention mechanisms like Attention Unet (34.9M), while achieving better performance.

Table 11
www.frontiersin.org

Table 11. Parameter counts for segmentation models.

For classification, Table 12 shows that our classification model gamResNet contains ~8.4 million parameters, significantly fewer than standard ResNet-18 (11.2M) and ResNet-50 (25M), while still outperforming them. Compared to lightweight models such as MobileNet or EfficientNet-B0, gamResNet maintains a reasonable parameter size and achieves superior diagnostic accuracy, striking a strong balance between efficiency and effectiveness.

Table 12
www.frontiersin.org

Table 12. Parameter counts for classification models.

Importantly, our design selectively integrates GAM in a minimal yet effective way—encoder-only for gamUnet and second-conv-only for gamResNet—thereby avoiding unnecessary overhead while preserving the benefits of global attention. This makes our models well-suited for clinical deployment, where both accuracy and resource efficiency are critical.

6 Discussion

The experimental results demonstrate that our gamUnet and gamResNet consistently outperformed baselines across key metrics in both segmentation and classification of OSCC histopathological images. The ablation studies also confirms the effectiveness of our designed model architecture. In classification, the gamResNet consistently achieved the highest AUROC and F1 scores across both magnifications. Notably, the enhanced model's ability to maintain high sensitivity across different magnifications underscores its versatility in handling datasets with varying levels of detail, which is highly important given the variability in histopathological images, where different magnification levels can emphasize different cellular structures. By incorporating GAM, our models effectively address the inherent limitations of conventional CNNs, which struggle with capturing global context and long-range dependencies. This enhancement is particularly crucial for tasks on OSCC histopathological images, where tumors could be ill-defined and essential diagnostic features may span across multiple regions of an image.

Overall, our results validate the effectiveness of the proposed model architecture in enhancing the diagnostic accuracy and reliability of deep learning models in OSCC analysis. This approach not only offers a more efficient alternative to manual examination but also has the potential to streamline diagnostic workflows, supporting pathologists in making timely and informed decisions. Future work may involve exploring the application of GAM in other cancer types and expanding the dataset to further validate model robustness. Additionally, integrating multi-scale attention mechanisms could further refine the model's ability to adapt to different magnifications and improve clinical applicability.

7 Conclusion

This study presents a deep learning-based approach to enhance the accuracy and efficiency of OSCC diagnosis and segmentation using histopathological images. By designing novel model architectures gamUnet and gamResnet that integrate Global Attention Mechanism to enable the model to capture global information, our method significantly improves the segmentation and classification performance on OSCC H&E-stained images, where ill-defined tumors and cell infiltration pose great challenges for other models. Experimental results on three datasets demonstrate that our gamUnet and gamResNet outperform traditional architectures, effectively distinguishing tumor regions within the complex tissue structures characteristic of OSCC. The ability of our method to provide reliable and consistent results highlights their potential as a diagnostic tool, supporting pathologists in delivering timely and accurate assessments. This work contributes to the advancement of automated pathology tools, aiming to improve patient outcomes in oral cancer care and can be further expanded to other cancer types and more diverse datasets.

Data availability statement

The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.

Ethics statement

Ethical approval was not required for the study involving humans in accordance with the local legislation and institutional requirements. Written informed consent to participate in this study was not required from the participants or the participants' legal guardians/next of kin in accordance with the national legislation and the institutional requirements.

Author contributions

JZhan: Investigation, Methodology, Writing – original draft. HD: Data curation, Methodology, Writing – original draft. RZ: Data curation, Methodology, Resources, Writing – review & editing. WL: Data curation, Methodology, Resources, Writing – review & editing. JZhao: Conceptualization, Investigation, Writing – review & editing. MG: Conceptualization, Writing – review & editing. XZ: Conceptualization, Investigation, Writing – review & editing.

Funding

The author(s) declare that financial support was received for the research and/or publication of this article. This work was supported by the National Natural Science Foundation of China (No. U23A20468) and Capital's Funds for the Health Improvement and Research (No. 2024-2G-4106).

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declare that no Gen AI was used in the creation of this manuscript.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

1. Wang R, Lyu C, Yu L. A transformation uncertainty and multi-scale contrastive learning-based semi-supervised segmentation method for oral cavity-derived cancer. Front Oncol. (2025) 15:1577198. doi: 10.3389/fonc.2025.1577198

PubMed Abstract | Crossref Full Text | Google Scholar

2. Warnakulasuriya S, Greenspan JS. Epidemiology of oral and oropharyngeal cancers. In:Warnakulasuriya S, Greenspan JS, , editors. Textbook of Oral Cancer: Prevention, Diagnosis and Management. Cham: Springer (2020). p. 5–21. doi: 10.1007/978-3-030-32316-5_2

Crossref Full Text | Google Scholar

3. Tan Y, Wang Z, Xu M, Li B, Huang Z, Qin S, et al. Oral squamous cell carcinomas: state of the field and emerging directions. Int J Oral Sci. (2023) 15:44. doi: 10.1038/s41368-023-00249-w

PubMed Abstract | Crossref Full Text | Google Scholar

4. Badwelan M, Muaddi H, Ahmed A, Lee KT, Tran SD. Oral squamous cell carcinoma and concomitant primary tumors, what do we know? A review of the literature. Curr Oncol. (2023) 30:3721–34. doi: 10.3390/curroncol30040283

PubMed Abstract | Crossref Full Text | Google Scholar

5. Coletta RD, Yeudall WA, Salo T. Grand challenges in oral cancers. Front Oral Health. (2020) 1:3. doi: 10.3389/froh.2020.00003

PubMed Abstract | Crossref Full Text | Google Scholar

6. Perdomo S, Roa GM, Brennan P, Forman D, Sierra MS. Head and neck cancer burden and preventive measures in Central and South America. Cancer Epidemiol. (2016) 44:S43–52. doi: 10.1016/j.canep.2016.03.012

PubMed Abstract | Crossref Full Text | Google Scholar

7. Speight PM, Khurram SA, Kujan O. Oral potentially malignant disorders: risk of progression to malignancy. Oral Surg Oral Med Oral Pathol Oral Radiol. (2018) 125:612–27. doi: 10.1016/j.oooo.2017.12.011

PubMed Abstract | Crossref Full Text | Google Scholar

8. Chakraborty D, Natarajan C, Mukherjee A. Advances in oral cancer detection. Adv Clin Chem. (2019) 91:181–200. doi: 10.1016/bs.acc.2019.03.006

PubMed Abstract | Crossref Full Text | Google Scholar

9. Ronneberger O, Fischer P, Brox T. U-net: Convolutional networks for biomedical image segmentation. In: Medical image Computing and Computer-Assisted Intervention-MICCAI 2015: 18th International Conference, Munich, Germany, October 5-9, 2015, Proceedings, Part III 18. Cham: Springer (2015). p. 234–41. doi: 10.1007/978-3-319-24574-4_28

Crossref Full Text | Google Scholar

10. Albishri A, Shah SJH, Lee Y. TLU-Net: transfer learning framework using U-Net convolutional neural networks for CT-based lungs and COVID-19 segmentation. In: 2022 IEEE International Conference on Bioinformatics and Biomedicine (BIBM). Las Vegas, NV: IEEE. (2022). p. 2238–46. doi: 10.1109/BIBM55620.2022.9994991

Crossref Full Text | Google Scholar

11. Albishri AA, Shah SJH, Kang SS, Lee Y. AM-UNet: automated mini 3D end-to-end U-net based network for brain claustrum segmentation. Multimed Tools Appl. (2022) 81:36171–94. doi: 10.1007/s11042-021-11568-7

PubMed Abstract | Crossref Full Text | Google Scholar

12. Chandrashekar G, AlQarni S, Bumann EE, Lee Y. Collaborative deep learning model for tooth segmentation and identification using panoramic radiographs. Comput Biol Med. (2022) 148:105829. doi: 10.1016/j.compbiomed.2022.105829

PubMed Abstract | Crossref Full Text | Google Scholar

13. Fraz MM, Khurram SA, Graham S, Shaban M, Hassan M, Loya A, et al. FABnet: feature attention-based network for simultaneous segmentation of microvessels and nerves in routine histology images of oral cancer. Neural Comput Appl. (2020) 32:9915–28. doi: 10.1007/s00521-019-04516-y

Crossref Full Text | Google Scholar

14. He K, Zhang X, Ren S, Sun J. Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Las Vegas, NV: IEEE (2016). p. 770–8. doi: 10.1109/CVPR.2016.90

Crossref Full Text | Google Scholar

15. Sarwinda D, Paradisa RH, Bustamam A, Anggia P. Deep learning in image classification using residual network (ResNet) variants for detection of colorectal cancer. Procedia Comput Sci. (2021) 179:423–31. doi: 10.1016/j.procs.2021.01.025

Crossref Full Text | Google Scholar

16. Ghosal P, Nandanwar L, Kanchan S, Bhadra A, Chakraborty J, Nandi D. Brain tumor classification using ResNet-101 based squeeze and excitation deep neural network. In: 2019 Second International Conference on Advanced Computational and Communication Paradigms (ICACCP). Gangtok: IEEE. (2019). p. 1–6. doi: 10.1109/ICACCP.2019.8882973

Crossref Full Text | Google Scholar

17. Lu Z, Bai Y, Chen Y, Su C, Lu S, Zhan T, et al. The classification of gliomas based on a pyramid dilated convolution resnet model. Pattern Recognit Lett. (2020) 133:173–9. doi: 10.1016/j.patrec.2020.03.007

Crossref Full Text | Google Scholar

18. Martino F, Bloisi DD, Pennisi A, Fawakherji M, Ilardi G, Russo D, et al. Deep learning-based pixel-wise lesion segmentation on oral squamous cell carcinoma images. Appl Sci. (2020) 10:8285. doi: 10.3390/app10228285

Crossref Full Text | Google Scholar

19. dos Santos DF, de Faria PR, Travencolo BA, do Nascimento MZ. Automated detection of tumor regions from oral histological whole slide images using fully convolutional neural networks. Biomed Signal Process Control. (2021) 69:102921. doi: 10.1016/j.bspc.2021.102921

Crossref Full Text | Google Scholar

20. Pennisi A, Bloisi DD, Nardi D, Varricchio S, Merolla F. Multi-encoder U-Net for Oral Squamous Cell Carcinoma Image Segmentation. In: 2022 IEEE International Symposium on Medical Measurements and Applications (MeMeA). Messina: IEEE (2022). p. 1–6. doi: 10.1109/MeMeA54994.2022.9856482

Crossref Full Text | Google Scholar

21. Musulin J, Štifanić D, Zulijani A, Ćabov T, Dekanić A, Car Z. An enhanced histopathology analysis: an AI-based system for multiclass grading of oral squamous cell carcinoma and segmenting of epithelial and stromal tissue. Cancers. (2021) 13:1784. doi: 10.3390/cancers13081784

PubMed Abstract | Crossref Full Text | Google Scholar

22. Alhazmi A, Alhazmi Y, Makrami A, Masmali A, Salawi N, Masmali K, et al. Application of artificial intelligence and machine learning for prediction of oral cancer risk. J Oral Pathol Med. (2021) 50:444–50. doi: 10.1111/jop.13157

PubMed Abstract | Crossref Full Text | Google Scholar

23. Bansal K, Bathla R, Kumar Y. Deep transfer learning techniques with hybrid optimization in early prediction and diagnosis of different types of oral cancer. Soft Comput. (2022) 26:11153–84. doi: 10.1007/s00500-022-07246-x

Crossref Full Text | Google Scholar

24. Shavlokhova V, Sandhu S, Flechtenmacher C, Koveshazi I, Neumeier F. Padrón-Laso V, et al. Deep learning on oral squamous cell carcinoma ex vivo fluorescent confocal microscopy data: a feasibility study. J Clin Med. (2021) 10:5326. doi: 10.3390/jcm10225326

PubMed Abstract | Crossref Full Text | Google Scholar

25. Aubreville M, Knipfer C, Oetter N, Jaremenko C, Rodner E, Denzler J, et al. Automatic classification of cancerous tissue in laserendomicroscopy images of the oral cavity using deep learning. Sci Rep. (2017) 7:11979. doi: 10.1038/s41598-017-12320-8

PubMed Abstract | Crossref Full Text | Google Scholar

26. Tanriver G, Soluk Tekkesin M, Ergen O. Automated detection and classification of oral lesions using deep learning to detect oral potentially malignant disorders. Cancers. (2021) 13:2766. doi: 10.3390/cancers13112766

PubMed Abstract | Crossref Full Text | Google Scholar

27. Welikala RA, Remagnino P, Lim JH, Chan CS, Rajendran S, Kallarakkal TG, et al. Automated detection and classification of oral lesions using deep learning for early detection of oral cancer. IEEE Access. (2020) 8:132677–93. doi: 10.1109/ACCESS.2020.3010180

Crossref Full Text | Google Scholar

28. Dhanuthai K, Rojanawatsirivej S, Thosaporn W, Kintarak S, Subarnbhesaj A, Darling M, et al. Oral cancer: a multicenter study. Med Oral Patol Oral Cir Bucal. (2018) 23:e23. doi: 10.4317/medoral.21999

PubMed Abstract | Crossref Full Text | Google Scholar

29. Farag AF, Abou-Alnour DA, Abu-Taleb NS. Oral carcinoma cuniculatum, an unacquainted variant of oral squamous cell carcinoma: a systematic review. Imaging Sci Dent. (2018) 48:233–44. doi: 10.5624/isd.2018.48.4.233

PubMed Abstract | Crossref Full Text | Google Scholar

30. Woolgar JA, Triantafyllou A. Pitfalls and procedures in the histopathological diagnosis of oral and oropharyngeal squamous cell carcinoma and a review of the role of pathology in prognosis. Oral Oncol. (2009) 45:361–85. doi: 10.1016/j.oraloncology.2008.07.016

PubMed Abstract | Crossref Full Text | Google Scholar

31. Rivera C, Venegas B. Histological and molecular aspects of oral squamous cell carcinoma. Oncol Lett. (2014) 8:7–11. doi: 10.3892/ol.2014.2103

PubMed Abstract | Crossref Full Text | Google Scholar

32. Shimizu S, Hiratsuka H, Koike K, Tsuchihashi K, Sonoda T, Ogi K, et al. Tumor-infiltrating CD8+ T-cell density is an independent prognostic marker for oral squamous cell carcinoma. Cancer Med. (2019) 8:80–93. doi: 10.1002/cam4.1889

PubMed Abstract | Crossref Full Text | Google Scholar

33. Campisi G, Calvino F, Carinci F, Matranga D, Carella M, Mazzotta M, et al. Peri-tumoral inflammatory cell infiltration in OSCC: a reliable marker of local recurrence and prognosis? An investigation using artificial neural networks. Int J Immunopathol Pharmacol. (2011) 24(2_suppl):113–20. doi: 10.1177/03946320110240S220

PubMed Abstract | Crossref Full Text | Google Scholar

34. Hu J, Shen L, Sun G. Squeeze-and-excitation networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Salt Lake City, UT: IEEE (2018). p. 7132–41. doi: 10.1109/CVPR.2018.00745

Crossref Full Text | Google Scholar

35. Woo S, Park J, Lee JY, Kweon IS. CBAM: convolutional block attention module. In: Proceedings of the European Conference on Computer Vision (ECCV). Cham: IEEE (2018). p. 3–19. doi: 10.1007/978-3-030-01234-2_1

Crossref Full Text | Google Scholar

36. Park J. Bam: bottleneck attention module. arXiv. (2018) [Preprint]. arXiv:1807.06514. doi: 10.48550/arXiv.1807.06514

Crossref Full Text | Google Scholar

37. Misra D, Nalamada T, Arasanipalai AU, Hou Q. Rotate to attend: convolutional triplet attention module. In: Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision. Waikoloa, HI: IEEE (2021). p. 3139–48. doi: 10.1109/WACV48630.2021.00318

Crossref Full Text | Google Scholar

38. Liu Y, Shao Z, Hoffmann N. Global attention mechanism: retain information to enhance channel-spatial interactions. arXiv. (2021) [Preprint]. arXiv:2112.05561. doi: 10.48550/arXiv.2112.05561

Crossref Full Text | Google Scholar

39. Ünsal G, Sevim S, Akkaya N, Aktaş V, Özcan I, Ünsal RBK, et al. Deep learning-based automatic segmentation of oral squamous cell carcinoma in histopathological images: a comprehensive evaluation and performance analysis. J Stomatol. (2025) 78:127–31. doi: 10.5114/jos.2025.151576

Crossref Full Text | Google Scholar

40. Shah SJH, Albishri A, Wang R, Lee Y. Integrating local and global attention mechanisms for enhanced oral cancer detection and explainability. Comput Biol Med. (2025) 189:109841. doi: 10.1016/j.compbiomed.2025.109841

PubMed Abstract | Crossref Full Text | Google Scholar

41. Rahman TY. A histopathological image repository of normal epithelium of oral cavity and oral squamous cell carcinoma. Mendeley Data. (2019) 1. doi: 10.17632/ftmp4cvtmb.1

Crossref Full Text | Google Scholar

42. Oktay O. Attention u-net: learning where to look for the pancreas. arXiv. (2018) [Preprint]. arXiv:1804.03999. doi: 10.48550/arXiv.1804.03999

Crossref Full Text | Google Scholar

43. Chollet F. Xception: deep learning with depthwise separable convolutions. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Honolulu, HI: IEEE (2017). p. 1251–8. doi: 10.1109/CVPR.2017.195

Crossref Full Text | Google Scholar

44. Diakogiannis FI, Waldner F, Caccetta P, Wu C. ResUNet-a: a deep learning framework for semantic segmentation of remotely sensed data. ISPRS J Photogramm Remote Sens. (2020) 162:94–114. doi: 10.1016/j.isprsjprs.2020.01.013

Crossref Full Text | Google Scholar

45. Chatterjee S, Sciarra A, Dünnwald M, Mushunuri RV, Podishetti R, Rao RN, et al. ShuffleUNet: Super resolution of diffusion-weighted MRIs using deep learning. In: 2021 29th European Signal Processing Conference (EUSIPCO). Dublin: IEEE. (2021). p. 940–4. doi: 10.23919/EUSIPCO54536.2021.9615963

Crossref Full Text | Google Scholar

46. Howard A, Sandler M, Chu G, Chen LC, Chen B, Tan M, et al. Searching for mobilenetv3. In: Proceedings of the IEEE/CVF International Conference on Computer Vision. Seoul: IEEE (2019). p. 1314–24. doi: 10.1109/ICCV.2019.00140

Crossref Full Text | Google Scholar

47. Sandler M, Howard A, Zhu M, Zhmoginov A, Chen LC. Mobilenetv2: inverted residuals and linear bottlenecks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Salt Lake City, UT: IEEE (2018). p. 4510–20. doi: 10.1109/CVPR.2018.00474

Crossref Full Text | Google Scholar

48. Howard AG. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv. (2017) [Preprint] arXiv:1704.04861. doi: 10.48550/arXiv.1704.04861

Crossref Full Text | Google Scholar

49. Tan M, Le Q. Efficientnet: rethinking model scaling for convolutional neural networks. In: International Conference on Machine Learning. PMLR. (2019). p. 6105–14.

Google Scholar

Keywords: oral squamous cell carcinoma (OSCC), segmentation, image processing, image classification, convolutional neural networks, deep learning–artificial intelligence, artificial intelligence

Citation: Zhang J, Ding H, Zhu R, Liao W, Zhao J, Gao M and Zhang X (2025) gamUnet: designing global attention-based CNN architectures for enhanced oral cancer detection and segmentation. Front. Med. 12:1582439. doi: 10.3389/fmed.2025.1582439

Received: 24 February 2025; Accepted: 27 June 2025;
Published: 23 July 2025.

Edited by:

Devesh U. Kapoor, Gujarat Technological University, India

Reviewed by:

Harsh Bansal, Amazon, United States
Vikram Yadav, Amazon.com Inc, United States

Copyright © 2025 Zhang, Ding, Zhu, Liao, Zhao, Gao and Zhang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Xiaoyun Zhang, eGlhb3l1bnpoYW5ncGt1QDEyNi5jb20=

These authors have contributed equally to this work and share first authorship

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.