ORIGINAL RESEARCH article
Front. Med.
Sec. Obstetrics and Gynecology
Uterine Cancer Classification from CT Images Using Convolutional Feature Extraction and Transformer-Based Self-Attention
Provisionally accepted- 1Yarmouk University, Irbid, Jordan
- 2Jordan University of Science and Technology, Irbid, Jordan
- 3Weill Cornell Medicine - Qatar, Doha, Qatar
Select one of your emails
You have multiple emails registered with Frontiers:
Notify me on publication
Please enter your email address:
If you already have an account, please login
You don't have a Frontiers account ? You can register here
Background: Accurate and early diagnosis of uterine cancer from computed tomography images remains a challenging task due to the complexity of anatomical structures and the subtle visual differences between normal, benign, and malignant uterine tissues. Traditional diagnostic approaches and conventional deep learning models often fail to effectively capture both local and global image characteristics. Objective: This study aims to develop and validate a novel hybrid deep learning framework that integrates convolutional feature extraction with transformer-based global attention mechanisms to improve the accuracy and robustness of uterine cancer classification from computed tomography images. Methods: In the proposed framework, DenseNet121 is employed as a convolutional neural network feature extractor, while a transformer encoder is utilized to model long-range contextual dependencies through multi-head self-attention. DenseNet121 captures discriminative local features from computed tomography images, which are subsequently processed by the transformer to enhance global feature representation. The performance of the proposed model is evaluated using the KAUH uterine cancer computed tomography dataset, which includes three classes: normal, benign, and malignant. The proposed approach is compared with several state-of-the-art deep learning models, including VGG16, VGG19, MobileNetV2, ResNet50, and DenseNet121. Results: Experimental results demonstrate that the proposed hybrid model outperforms the comparative models. It achieves an accuracy of 87.44%, sensitivity of 87.13%, specificity of 95.20%, an F1 score of 87.17%, and an area under the receiver operating characteristic curve of 99.41%. Conclusion: The results confirm the effectiveness of integrating convolutional neural networks with transformer-based self-attention mechanisms for significantly improving uterine cancer classification from computed tomography images. The proposed system shows strong potential as a computer-aided decision-support tool for radiologists to assist in the detection of uterine cancer and may be extended to various real-world clinical applications.
Keywords: Classification, CT images, deep learning, diagnosis, Real Dataset, Uterine Cancer
Received: 05 Jan 2026; Accepted: 05 Feb 2026.
Copyright: © 2026 Alshdaifat, Sindiani, Alhatamleh, Malkawi, Madain, Almahmoud, Al-Smadi, Al-Mnayyis, Amin and Abd-alrazaq. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
* Correspondence: Alaa Abd-alrazaq
Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.
