Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Med.

Sec. Ophthalmology

Volume 12 - 2025 | doi: 10.3389/fmed.2025.1660920

This article is part of the Research TopicInnovative Advancements in Eye Image Processing for Improved Ophthalmic DiagnosisView all 4 articles

HyReti-Net: Hybrid Retinal Diseases Classification and Diagnosis Network Using Optical Coherence Tomography

Provisionally accepted
Jikun  YangJikun Yang1,2*Chaoliang  HsuChaoliang Hsu3Jing  WangJing Wang2Bin  WuBin Wu2Yuanyuan  LuYuanyuan Lu2Yuxi  DingYuxi Ding2Zhenbo  ZhaoZhenbo Zhao1,2Kaili  TangKaili Tang1,2Feng  LuFeng Lu4Liwei  MaLiwei Ma1,2*
  • 1Aier Eye Medical Center of Anhui Medical University, Anhui, China
  • 2Shenyang Aier Excellence Eye Hospital, Shenyang, China
  • 3University of Southern California Translational Biotechnology Department, Los Angeles, United States
  • 4Shenyang Aerospace University, Shenyang, China

The final, formatted version of the article will be published soon.

Background: With optical coherence tomography (OCT), doctors are able to see cross-sections of the retinal layers and diagnose retinal diseases. Computer-aided diagnosis algorithms such as convolutional neural networks (CNNs) and vision Transformers (ViTs) enhance diagnostic efficiency by automatically analyzing these OCT images. However, CNNs are less effective in extracting global features and ViTs lack the local inductive bias and typically require large amounts of training data. Methods: In this paper, we presented a hybrid retinal diseases classification and diagnosis network named HyReti-Net which incorporated two branches. One branch extracted local features by leveraging the spatial hierarchy learning capabilities of ResNet-50, while the other branch was established based on Swin Transformer to consider the global information. In addition, we proposed a feature fusion module (FFM) consisting of a concatenation and residual block and the improved channel attention block to retain local and global features more effectively. The multi-level features fusion mechanism was used to further enhance the ability of global feature extraction. Results: Evaluation and comparison were used to show the advantage of the proposed architecture. Five metrics were applied to compare the performance of existing methods. Moreover, ablation studies were carried out to evaluate their effects on the foundational model. For each public dataset, heatmaps were also generated to enhance the interpretability of OCT image classification. The results underscored the effectiveness and advantage of the proposed method which achieved the highest classification accuracy. Conclusion: In this article, a hybrid multi-scale network model integrating dual-branches and a features fusion module was proposed to diagnose retinal diseases. The performance of the proposed method produced promising classification results. On the OCT-2014, OCT-2017 and OCT-C8, experimental results indicated that HyReti-Net achieved better performance than the state-of-the-art networks. This study can provide a reference for clinical diagnosis of ophthalmologists through artificial intelligence technology.

Keywords: vision Transformer, Convolutional Neural Network, Feature fusion, Retinal Diseases, Classification

Received: 07 Jul 2025; Accepted: 15 Sep 2025.

Copyright: © 2025 Yang, Hsu, Wang, Wu, Lu, Ding, Zhao, Tang, Lu and Ma. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

* Correspondence:
Jikun Yang, jkyang66@163.com
Liwei Ma, 18900913588@163.com

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.