ORIGINAL RESEARCH article
Front. Med.
Sec. Ophthalmology
Volume 12 - 2025 | doi: 10.3389/fmed.2025.1660920
This article is part of the Research TopicInnovative Advancements in Eye Image Processing for Improved Ophthalmic DiagnosisView all 4 articles
HyReti-Net: Hybrid Retinal Diseases Classification and Diagnosis Network Using Optical Coherence Tomography
Provisionally accepted- 1Aier Eye Medical Center of Anhui Medical University, Anhui, China
- 2Shenyang Aier Excellence Eye Hospital, Shenyang, China
- 3University of Southern California Translational Biotechnology Department, Los Angeles, United States
- 4Shenyang Aerospace University, Shenyang, China
Select one of your emails
You have multiple emails registered with Frontiers:
Notify me on publication
Please enter your email address:
If you already have an account, please login
You don't have a Frontiers account ? You can register here
Background: With optical coherence tomography (OCT), doctors are able to see cross-sections of the retinal layers and diagnose retinal diseases. Computer-aided diagnosis algorithms such as convolutional neural networks (CNNs) and vision Transformers (ViTs) enhance diagnostic efficiency by automatically analyzing these OCT images. However, CNNs are less effective in extracting global features and ViTs lack the local inductive bias and typically require large amounts of training data. Methods: In this paper, we presented a hybrid retinal diseases classification and diagnosis network named HyReti-Net which incorporated two branches. One branch extracted local features by leveraging the spatial hierarchy learning capabilities of ResNet-50, while the other branch was established based on Swin Transformer to consider the global information. In addition, we proposed a feature fusion module (FFM) consisting of a concatenation and residual block and the improved channel attention block to retain local and global features more effectively. The multi-level features fusion mechanism was used to further enhance the ability of global feature extraction. Results: Evaluation and comparison were used to show the advantage of the proposed architecture. Five metrics were applied to compare the performance of existing methods. Moreover, ablation studies were carried out to evaluate their effects on the foundational model. For each public dataset, heatmaps were also generated to enhance the interpretability of OCT image classification. The results underscored the effectiveness and advantage of the proposed method which achieved the highest classification accuracy. Conclusion: In this article, a hybrid multi-scale network model integrating dual-branches and a features fusion module was proposed to diagnose retinal diseases. The performance of the proposed method produced promising classification results. On the OCT-2014, OCT-2017 and OCT-C8, experimental results indicated that HyReti-Net achieved better performance than the state-of-the-art networks. This study can provide a reference for clinical diagnosis of ophthalmologists through artificial intelligence technology.
Keywords: vision Transformer, Convolutional Neural Network, Feature fusion, Retinal Diseases, Classification
Received: 07 Jul 2025; Accepted: 15 Sep 2025.
Copyright: © 2025 Yang, Hsu, Wang, Wu, Lu, Ding, Zhao, Tang, Lu and Ma. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
* Correspondence:
Jikun Yang, jkyang66@163.com
Liwei Ma, 18900913588@163.com
Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.