ORIGINAL RESEARCH article

Front. Med., 02 April 2025

Sec. Ophthalmology

Volume 12 - 2025 | https://doi.org/10.3389/fmed.2025.1551315

This article is part of the Research TopicEfficient Artificial Intelligence in Ophthalmic Imaging – Volume IIView all 5 articles

A lightweight multi-deep learning framework for accurate diabetic retinopathy detection and multi-level severity identification

  • 1Department of Artificial Intelligence and Robotics, Sejong University, Seoul, Republic of Korea
  • 2Department of Scientific Computing, Pukyong National University, Busan, Republic of Korea
  • 3Department of Mathematics, Institute of Mathematical Science, Pusan National University, Busan, Republic of Korea
  • 4Finance Fishery Manufacture Industrial Mathematics Center on Big Data, Pusan National University, Busan, Republic of Korea

Accurate and timely detection of diabetic retinopathy (DR) is crucial for managing its progression and improving patient outcomes. However, developing algorithms to analyze complex fundus images continues to be a major challenge. This work presents a lightweight deep-learning network developed for DR detection. The proposed framework consists of two stages. In the first step, the developed model is used to assess the presence of DR [i.e., healthy (no DR) or diseased (DR)]. The next step involves the use of transfer learning for further subclassification of DR severity (i.e., mild, moderate, severe DR, and proliferative DR). The designed model is reused for transfer learning, as correlated images facilitate further classification of DR severity. The online dataset is used to validate the proposed framework, and results show that the proposed model is lightweight and has comparatively low learnable parameters compared to others. The proposed two-stage framework enhances the classification performance, achieving a 99.06% classification rate for DR detection and an accuracy of 90.75% for DR severity identification for APTOS 2019 dataset.

1 Introduction

Diabetes is a growing global health concern, with approximately 830 million people affected worldwide in 2022, according to the World Health Organization (1). The prevalence of diabetes is expected to continue rising, resulting in an increasing number of people developing related complications, such as diabetic retinopathy (DR). Maintaining optimal visual health is essential for ensuring the well-being of individuals worldwide. Visual impairment and blindness pose significant threats to global health, driven by various factors including chronic diseases and unequal access to healthcare (2). As the number of people with DR grows from an estimated 103 million in 2020 to around 160 million by 2045, so does the number of individuals experiencing vision problems, posing a substantial challenge to public health and the economy (3). DR is a significant contributor to visual impairment and blindness worldwide. According to a study published by Al-Ghamdi (4), in Saudi Arabia, individuals over the age of 50 experience higher rates of blindness and visual impairment. For instance, among those aged 60 years and above, about 20% have blindness, while as many as 66.2% have visual impairment. Being a chronic and insidious disease, DR generally affects persons with diabetes, often developing insidiously without overt early symptoms. If left unattended, DR can result in irreversible blindness.

DR can be effectively managed through regular screening, which enables the condition to be detected early and treated. Advanced techniques promote the diagnosing and severity identification of tiny lesions at higher resolution imaging of the fundus. There are primarily two forms in which DR manifests itself: proliferative DR (PDR) and non-proliferative DR (NPDR) (5). NPDR can be classified into four different severity levels, ranging from mild to severe (6). Typical symptoms in DR include microaneurysms, retinal hemorrhage, and exudates. In the early development of DR, there are usually only slight manifestations of microaneurysms. With its progress, along with significant intraretinal hemorrhages and venous beading or microvascular abnormalities, complications can become more severe (7). The formation of new blood vessels defines PDR, often accompanied by vitreous or retina hemorrhage (8). Fundus imaging plays an imperative role in the diagnosis of DR (9), enabling ophthalmologists to analyze lesions, decide on the severity, and hence recommend treatment.

Grading DR from a fundus image is difficult even for expert ophthalmologists because diagnosis involves several lesions and often overlapped borders of those lesions, making perfect diagnosis time-consuming and unreliable. Hence, computer-aided systems are needed to grade DR. Various computer-aided systems have been presented in the literature to grade the DR severity (1014). Such solutions intend to assist ophthalmologists in correctly assessing fundus images to detect lesions.

Various image processing methods are used to automatically detect hemorrhages, one common retinal manifestation related to diabetic patients (10). Other works have focused on developing strategies for detecting microaneurysms (11), one of the primary indicators of DR. Dimensionality reduction using techniques such as principal component analysis has been used to enhance microaneurysm detection. Fuzzy C-means clustering has also been utilized to diagnose DR and maculopathy automatically (15). Other researchers have also tried curvelet-based edge enhancement and wideband bandpass filters to improve the contrast between lesions and retina background (16).

Several classical machine learning methods have been employed by researchers using decision trees, support vector machines (17), random forests (18), logistic regression (19, 20), and Gaussian naive Bayes (19) for DR grading. These algorithms extract features from images with the aid of image processing techniques. For instance, Lachure et al. (17) used erosion, dilatation, opening, and closure for feature detection by applying morphology-based image processing methods to differentiate microaneurysms from exudates. Recently, studies of deep features for local and global feature extractions from fundus images have been conducted (18). These features are then utilized to train classifiers and predict the final DR classification. Most of the classical machine-learning methods depend heavily on feature extraction, which is often a complicated and time-consuming process. This may not capture all the intricate characteristics necessary for accurate classification.

In contrast, deep learning techniques have emerged as a powerful tool in imaging applications. Deep learning models have demonstrated remarkable success in identifying DR by automatically extracting complex features using convolutional layers (21, 22). This approach eliminates manual feature extraction, allowing for more efficient and accurate classification (1214). Zhou et al. (23) introduced a multitasking deep neural network architecture for DR classification. By leveraging a multitasking approach, they forecasted labels through classification and regression, achieving an 84% kappa score. The interconnectedness of DR phases contributed to the achieved performance. In addition to deep learning models, transfer learning has also been explored for detection. Researchers have leveraged pre-trained models and fine-tuned them for DR detection, achieving promising results (24, 25). Furthermore, various feature selection approaches, including wrapper-based (26, 27) and filter-based methods (28), have been investigated for DR detection. These techniques aim to identify the relevant features, reducing the data dimensionality and improving model performance. Deep neural networks dominate image processing and computer vision applications. However, traditional deep learning models may not effectively capture complex, irregular objects. Despite advancements, multi-class categorization remains an area for improvement. On the other hand, transfer learning models are effective, but can be even more beneficial when fine-tuned on weights frozen with images that are highly correlated with the target task. This study aimed to enhance classification performance while minimizing layers and parameters.

This research presents a novel deep-learning architecture and framework for detecting DR and identifying its severity. The approach involves a two-stage process. Initially, a custom-designed deep learning lightweight classification model is developed to distinguish between healthy (no DR) and DR-affected (mild DR, moderate DR, severe DR, and PDR) fundus images. Upon identifying the presence of DR, the pre-trained model is reused using transfer learning to subclassify the severity of DR into specific categories, including mild, moderate, severe DR, and PDR. Leveraging the initial model’s frozen weights on related images facilitates the severity identification process. The proposed deep learning model and framework are validated using a publicly available online dataset. The proposed framework is compared to existing state-of-the-art models to check its simplicity and efficiency.

2 Materials and methods

This section provides a complete description of the methodologies and materials in this work, particularly the developed deep learning model and its framework. Additionally, it describes the data augmentation techniques used to tackle the dataset’s imbalanced issues.

2.1 Proposed DR detection multi-level severity identification framework

This research presents a deep learning-based approach for detecting and identifying the severity of DR. This approach is divided into two phases. Phase I involves designing a deep learning model to classify fundus images as healthy (no DR) and DR-affected (mild DR, moderate DR, severe DR, and PDR). Stage II utilizes the trained model from Stage I to further categorize DR retinas into mild, moderate, severe DR, and PDR. The overall framework is illustrated in Figure 1.

Figure 1
www.frontiersin.org

Figure 1. DR detection and severity identification framework using deep learning approach.

2.2 Proposed deep learning model and transfer learning model for DR detection and severity identification

This section explores two deep learning approaches for DR detection and severity identification: (i) isolated deep learning networks and (ii) transfer learning.

An isolated model is trained from scratch without prior knowledge (29), whereas transfer learning leverages pre-acquired knowledge from other models (30). The transfer learning approach involves training a base model on a base dataset and then reusing the learned features to train a new model for a specific task (3133). This study employs a developed deep learning model approach for transfer learning, where a pre-trained model is reutilized by modifying its architecture to accommodate the specific task of DR severity identification.

Two deep learning models are examined: the isolated deep learning model and the transfer learning-based deep learning model. The isolated deep learning model is designed to categorize fundus images into healthy (no DR) and DR-affected (mild DR, moderate DR, severe DR, and PDR) classes, while the transfer learning-based models are used to subclassify DR-affected retinas into mild, moderate, severe DR, and PDR.

The isolated model consists of five primary layers: input, convolution, pooling, fully connected, and classification. The features are extracted in convolution and pooling layers, while the fully connected and classification layers facilitate prediction. This study develops and evaluates various isolated models with different architectures (36-layer, 33-layer with one parallel branch, 37-layer with one parallel branch, 41-layer with one parallel branch, and 42-layer with two parallel branch) to determine the optimal model for DR detection. Figure 2 illustrates the detailed architecture of the optimal model (37-layer with one parallel branch) used for further analysis of DR detection.

Figure 2
www.frontiersin.org

Figure 2. Proposed deep learning model architecture for DR detection.

The optimal isolated deep learning model is reused for DR severity identification. The last three layers of the pre-trained model are replaced and retrained to accommodate the multi-class classification task. Figure 3 presents the detailed architecture of the transfer learning model used for DR severity identification.

Figure 3
www.frontiersin.org

Figure 3. Proposed transfer learning deep learning model architecture for DR severity identification.

2.3 Dataset and preprocessing

The study employed the FairVision dataset (34), the Asia Pacific Tele-Ophthalmology Society (APTOS) 2019 dataset (https://www.kaggle.com/c/aptos2019-blindness-detection/overview, accessed on 20 November 2024), and the DDR dataset (https://www.kaggle.com/datasets/mariaherrerot/ddrdataset, accessed on 28 February 2025). The dataset comprises fundus photographs captured from patients, representing a diverse range of imaging conditions. The FairVision, APTOS 2019, and DDR datasets included 6,000, 3,662, and 12,522 fundus images, respectively. Figure 4 offers more information regarding the distribution of fundus images.

Figure 4
www.frontiersin.org

Figure 4. Overview of the various datasets for DR detection and severity identification.

The developed model required uniform input images, which were achieved by cropping all images to a fixed size of 227 × 227 pixels. Normalization was performed using a zero-center approach during the preprocessing stage.

A significant challenge in deep learning is the scarcity of data. Increasing the dataset size enables models to learn more robust features and reduces overfitting risks. Our datasets were initially imbalanced, with varying image counts across classes. We employed data augmentation techniques to address this, including flipping, rotation, scaling, and translation. These augmentations enhanced the dataset’s size and image quality; details are presented in Table 1.

Table 1
www.frontiersin.org

Table 1. Details about the images used for DR detection and severity identification.

3 Results

This research used MATLAB 2023a for all simulations and analyses on a personal computer, which has the following specs: Core i7, 12th Generation, 32 GB RAM, NVIDIA GeForce RTX 3050, 1 TB SSD, and 64-bit Windows 11 operating system. Using hold-out validation, we divided the dataset randomly into 80/20 ratios for model training and testing. The images used for model testing were not part of the training set. The initial parameters were as follows: 100 epochs, 0.9 momentum, 128 mini batch-size, and 0.001 learning rate. The stochastic gradient descent with momentum (SGDM) solver is used for training and testing.

The APTOS 2019 dataset was initially utilized to evaluate the proposed model’s performance. A preliminary analysis was conducted to identify the optimal layer configuration for the lightweight deep learning model. The results of this analysis are summarized in Table 2. Based on the findings, the 37-layer model was selected, as it demonstrated superior accuracy.

Table 2
www.frontiersin.org

Table 2. Results of the ablation study performed for the selection of layers using APTOS 2019 dataset.

Several state-of-the-art models, including VGG-16, MobileNet-v2, Efficient-b0, Inception-v3, ResNET-50, and GoogLeNet, were also trained and tested to compare comprehensively using APTOS 2019 dataset. The results of this evaluation are summarized in Table 3.

Table 3
www.frontiersin.org

Table 3. Performance of various deep learning models against APTOS 2019 dataset.

After comprehensively analyzing Table 3, it can be observed that the Efficient-b0 showed the highest classification rate of 82.4% (kappa value = 0.73), with almost 162 min 6 s training time, and also has 4 million learnable parameters. The second best is the Inception-v3 model with 81.86% classification accuracy (kappa value = 0.72), 110 min 17 s training time, and 21.8 million learnable parameters. ResNET-50 showed the lowest performance, with only 75.3% accuracy (kappa value = 0.62) and 23.5 million learnable parameters. Notably, our proposed deep learning model demonstrated a reasonable classification rate of 80.08% (kappa value = 0.70), with a significantly shorter training time of just 14 min 28 s. As shown in Table 3, the model correctly classifies 587 out of 733 fundus images. Moreover, the developed network has the fewest learnable parameters compared to the other.

Therefore, the proposed model was employed to investigate DR detection and severity identification using the APTOS 2019 dataset, as discussed in section 2. The process involves initially distinguishing between healthy and DR-infected fundus images and detecting the DR severity in the second stage. Data augmentation was applied to address dataset imbalance, as illustrated in Table 1. The outcomes of this analysis for the APTOS 2019 dataset are presented in Figures 57.

Figure 5
www.frontiersin.org

Figure 5. Performance of the developed deep learning model and DR detection and severity identification framework for the APTOS 2019 dataset.

Figure 6
www.frontiersin.org

Figure 6. Learning curves DR detection for the APTOS 2019 dataset.

Figure 7
www.frontiersin.org

Figure 7. Learning curves DR severity identification for the APTOS 2019 dataset.

Figure 5 demonstrates that the presented methodology effectively enhances the accuracy of DR detection. Specifically, the model achieved a high accuracy rate of 99.06% (kappa value = 0.98), correctly classifying 792 out of 800 DR images and misclassifying only seven healthy (no DR) fundus images.

In the case of DR severity level identification, some misclassification occurred, with 40 out of 200 fundus images for moderate DR class. However, these numbers are relatively low compared to the true positives. The model performed strongly in identifying mild and severe DR classes, with 95.5% recall for both classes (see Figure 5). Overall, the approach accomplished a high classification rate of 90.75% (kappa value = 0.88) for DR severity identification.

Figures 6, 7 illustrate both trained models’ learning curves, showing that they become stable after approximately 40 epochs. To further validate the generalizability of the proposed approach, we evaluated its performance on two additional datasets: FairVision and DDR (dataset details are provided in Figure 4 and Table 1). The results obtained from these datasets using 5-fold cross-validation are presented in Figures 8, 9.

Figure 8
www.frontiersin.org

Figure 8. Performance of the developed deep learning model and DR detection framework for the FairVision dataset.

Figure 9
www.frontiersin.org

Figure 9. Performance of the developed deep learning model and DR detection and severity identification framework for the DDR dataset.

The FairVision dataset offered two classes for DR detection: non-vision-threatening DR and vision-threatening DR. However, the dataset was highly imbalanced, comprising 5,435 images of non-vision-threatening DR and 565 images of vision-threatening DR. To address this imbalance, we applied an augmentation approach, resulting in 6,000 images for each class, as shown in Table 1. Our model achieved a high classification rate of 98.56% (kappa value = 0.97) to distinguish between non-vision-threatening DR and vision-threatening DR class, as shown in Figure 8. Similarly, the DDR dataset is also highly imbalanced, comprising 6,266, 630, 4,477, 236, and 913 images of none, mild, moderate, severe, and PDR. We applied an augmentation approach to address this imbalance, resulting in 6,000 images for the no DR class and 1,500 images for the remaining DR classes (1,500 × 4 = 6,000). The model again achieved a high classification rate of 95.43% (kappa value = 0.91) for detecting DR and 92.63% (kappa value = 0.90) for DR severity identification, as shown in Figure 9. The results of these well-known mainstream public datasets further validated the performance of the proposed model and multistage approach. Furthermore, a comparison of the proposed approach with other studies is presented in Table 4.

Table 4
www.frontiersin.org

Table 4. Comparative analysis of classification performance with state-of-the-art models on the APTOS 2019 dataset.

4 Discussion

DR is one of the major vision-threatening conditions faced by the healthcare department, which is quite challenging to diagnose and manage. Grading the severity of DR is vital for timely management and prevention of vision loss. This study presents a novel deep-learning network and framework for determining the detection and severity of DR, which would help the healthcare professional community manage patients effectively and efficiently.

An ablation study was conducted, where various deep-learning models were designed and tested for DR detection and severity identification. The 37-layer deep-learning model with a single parallel branch was selected due to its high classification accuracy and low training time (Table 2). Subsequently, various pre-trained models were trained and tested alongside the proposed 37-layer model, with detailed results presented in Table 3. After thorough analysis, it was observed that the developed deep-learning model achieved reasonable accuracy with reduced training time and minimal learnable parameters (1.3 million), making it lightweight. Onward, the model was initially trained to detect DR using augmented fundus images (APTOS 2019), and the results are presented in Figure 5. The model demonstrated a high performance for DR detection (99.06%). Furthermore, in the case of DR severity detection, the results achieved a high severity level detection accuracy of 90.75%. To validate the proposed model and approach further, it was evaluated using the FairVision and DDR datasets. The results demonstrated excellent performance, with a classification accuracy of 98.56% on the FairVision dataset for distinguishing between non-vision-threatening and vision-threatening DR (see Figure 8). On the DDR dataset, our model achieved an accuracy of 95.43% for DR detection and 92.63% for severity level identification (see Figure 9).

APTOS 2019 is a benchmark dataset used for the performance evaluation of many different models in detecting DR severity. Table 4 summarizes the performance of state-of-the-art models using this dataset. Recently, Alam et al. (35) investigated the use of the swapping assignments across multiple views technique for DR grading in conjunction with contrasting cluster assignments in retinal fundus images. The model achieved an accuracy of 87% with fewer parameters and layers. In another recent study (36), the authors introduced a novel transformer-based model called TMILv4 with an accuracy of 85.6%. However, its sensitivity was relatively low, at 73.7%. Similarly, Shaik and Cherukuri (37) combined a convolutional autoencoder with a neural support vector machine. They reported an accuracy of 84.31%, although the sensitivity score was as low as 66.16%. Contrarily, Kassani et al. (38) used the modified Xception network; it had higher sensitivity but lower accuracy at 83.09%. In contrast, the proposed model beats those state-of-the-art deep learning and transformer-based models, as represented in Table 4.

5 Conclusion

This study presents a novel, lightweight, deep-learning DR detection and severity identification framework. A 37-layer deep-learning model is designed to detect the DR using fundus images. The model is validated using a publicly available dataset and augmented using image augmentation techniques to balance classes. Comprehensive experiments are conducted to evaluate the proposed model’s performance using various metrics. The results demonstrate exceptional accuracy, with the proposed model achieving 99.06% accuracy for DR detection. Notably, the model has significantly fewer learnable parameters than existing models. The trained model is reutilized to evaluate its performance in detecting DR severity levels using transfer learning, yielding a high classification accuracy of 90.75%. Furthermore, our proposed model achieved high accuracy rates of 98.56% for FairVision, and 95.43% in detecting DR and 92.63% its severity for DDR dataset. The outstanding performance surpasses previous deep learning algorithms, providing superior DR detection and severity identification results. This work serves as a reference for classifying fundus images using deep learning techniques, offering a lightweight, efficient, and robust model.

Data availability statement

The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.

Ethics statement

Ethical approval was not required for the study involving humans in accordance with the local legislation and institutional requirements. Written informed consent to participate in this study was not required from the participants or the participants’ legal guardians/next of kin in accordance with the national legislation and the institutional requirements.

Author contributions

AZ: Conceptualization, Methodology, Software, Writing – original draft. KK: Conceptualization, Formal analysis, Methodology, Writing – original draft. MA: Conceptualization, Formal analysis, Methodology, Software, Writing – original draft. JB: Formal analysis, Investigation, Validation, Writing – review & editing. S-HK: Conceptualization, Funding acquisition, Resources, Supervision, Validation, Writing – review & editing.

Funding

The author(s) declare that financial support was received for the research and/or publication of this article. This work was partially supported by the MSIT (Ministry of Science and ICT), Korea, under the ICAN (ICT Challenge and Advanced Network of HRD) program (IITP-2024-RS-2022-00156345) under the supervision of the IITP (Institute of Information & Communications Technology Planning & Evaluation). This work was partially supported by the National Research Foundation of Korea (NRF) grant funded by the Korea Government (MSIT) (Nos. RS-2023-00219051 and RS-2023-00209107).

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The authors declare that no Gen AI was used in the creation of this manuscript.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

1. World Health Organization. (2024). Diabetes. Available online at: https://www.who.int/news-room/fact-sheets/detail/diabetes. (Accessed December 1, 2024)

Google Scholar

2. Zheng, DD, Christ, SL, Lam, BL, Feaster, DJ, McCollister, K, and Lee, DJ. Patterns of chronic conditions and their association with visual impairment and health care use. JAMA Ophthalmol. (2020) 138:387–94. doi: 10.1001/jamaophthalmol.2020.0052

PubMed Abstract | Crossref Full Text | Google Scholar

3. Teo, ZL, Tham, Y-C, Yu, M, Chee, ML, Rim, TH, Cheung, N, et al. Global prevalence of diabetic retinopathy and projection of burden through 2045: systematic review and meta-analysis. Ophthalmology. (2021) 128:1580–91. doi: 10.1016/j.ophtha.2021.04.027

PubMed Abstract | Crossref Full Text | Google Scholar

4. Al-Ghamdi, AS. Adults visual impairment and blindness—an overview of prevalence and causes in Saudi Arabia. Saudi J Ophthalmol. (2019) 33:374–81. doi: 10.1016/j.sjopt.2019.10.001

PubMed Abstract | Crossref Full Text | Google Scholar

5. Fong, DS, Aiello, LP, Ferris, FL III, and Klein, R. Diabetic retinopathy. Diabetes Care. (2004) 27:2540–53. doi: 10.2337/diacare.27.10.2540

PubMed Abstract | Crossref Full Text | Google Scholar

6. Zulaikha, BS. Multi-level severity classification for diabetic retinopathy based on hybrid optimization enabled deep learning. Biomed Signal Process Control. (2023) 84:104736. doi: 10.1016/j.bspc.2023.104736

PubMed Abstract | Crossref Full Text | Google Scholar

7. Manjiri, P, Ramesh, M, Yogesh, R, Manoj, S, and Neha, D (2014). Automated localization of optic disk, detection of microaneurysms and extraction of blood vessels to bypass angiography. Proceedings of the 3rd International Conference on Frontiers of Intelligent Computing: Theory and Applications (FICTA)

Google Scholar

8. Shukla, UV, and Tripathy, K. Diabetic retinopathy In: StatPearls. Treasure Island, FL: StatPearls Publishing (2023)

Google Scholar

9. Prathibha, S, and Siddappaji,. Advancing diabetic retinopathy diagnosis with fundus imaging: a comprehensive survey of computer-aided detection, grading and classification methods. Glob Transit. (2024) 6:93–112. doi: 10.1016/j.glt.2024.04.001

Crossref Full Text | Google Scholar

10. Mumtaz, R, Hussain, M, Sarwar, S, Khan, K, Mumtaz, S, and Mumtaz, M. Automatic detection of retinal hemorrhages by exploiting image processing techniques for screening retinal diseases in diabetic patients. Int J Diabetes Dev Ctries. (2018) 38:80–7. doi: 10.1007/s13410-017-0561-6

Crossref Full Text | Google Scholar

11. Usman Akram, M, Khalid, S, Tariq, A, Khan, SA, and Azam, F. Detection and classification of retinal lesions for grading of diabetic retinopathy. Comput Biol Med. (2014) 45:161–71. doi: 10.1016/j.compbiomed.2013.11.014

PubMed Abstract | Crossref Full Text | Google Scholar

12. Alyoubi, WL, Shalash, WM, and Abulkhair, MF. Diabetic retinopathy detection through deep learning techniques: a review. Inform Med Unlocked. (2020) 20:100377. doi: 10.1016/j.imu.2020.100377

Crossref Full Text | Google Scholar

13. Das, D, Biswas, SK, and Bandyopadhyay, S. A critical review on diagnosis of diabetic retinopathy using machine learning and deep learning. Multimed Tools Appl. (2022) 81:25613–55. doi: 10.1007/s11042-022-12642-4

PubMed Abstract | Crossref Full Text | Google Scholar

14. Nielsen, KB, Lautrup, ML, Andersen, JKH, Savarimuthu, TR, and Grauslund, J. Deep learning-based algorithms in screening of diabetic retinopathy: a systematic review of diagnostic performance. Ophthalmol Retina. (2019) 3:294–304. doi: 10.1016/j.oret.2018.10.014

Crossref Full Text | Google Scholar

15. Rahim, SS, Palade, V, Shuttleworth, J, and Jayne, C. Automatic screening and classification of diabetic retinopathy and maculopathy using fuzzy image processing. Brain Inform. (2016) 3:249–67. doi: 10.1007/s40708-016-0045-3

PubMed Abstract | Crossref Full Text | Google Scholar

16. Kar, SS, and Maity, SP. Automatic detection of retinal lesions for screening of diabetic retinopathy. IEEE Trans Biomed Eng. (2018) 65:608–18. doi: 10.1109/TBME.2017.2707578

Crossref Full Text | Google Scholar

17. Lachure, J, Deorankar, AV, Lachure, S, Gupta, S, and Jadhav, R. (2015). Diabetic retinopathy using morphological operations and machine learning. 2015 IEEE International Advance Computing Conference (IACC). 12–13

Google Scholar

18. Gayathri, S, Gopi, VP, and Palanisamy, P. Diabetic retinopathy classification based on multipath CNN and machine learning classifiers. Phys Eng Sci Med. (2021) 44:639–53. doi: 10.1007/s13246-021-01012-3

PubMed Abstract | Crossref Full Text | Google Scholar

19. Asha, PR, and Karpagavalli, S (2015). Diabetic retinal exudates detection using machine learning techniques. 2015 International Conference on Advanced Computing and Communication Systems. 5–7.

Google Scholar

20. Wang, J-Z, Lu, N-H, Du, W-C, Liu, K-Y, Hsu, S-Y, Wang, C-Y, et al. Classification of color fundus photographs using fusion extracted features and customized CNN models. Healthcare. (2023) 11:2228. doi: 10.3390/healthcare11152228

PubMed Abstract | Crossref Full Text | Google Scholar

21. Bhimavarapu, U, and Battineni, G. Deep learning for the detection and classification of diabetic retinopathy with an improved activation function. Healthcare. (2023) 11:97. doi: 10.3390/healthcare11010097

PubMed Abstract | Crossref Full Text | Google Scholar

22. Kashyap, R, Nair, R, Gangadharan, SMP, Botto-Tobar, M, Farooq, S, and Rizwan, A. Glaucoma detection and classification using improved U-Net deep learning model. Healthcare. (2022) 10:2497. doi: 10.3390/healthcare10122497

PubMed Abstract | Crossref Full Text | Google Scholar

23. Zhou, K, Gu, Z, Liu, W, Luo, W, Cheng, J, Gao, S, et al. (2018). Multi-cell multi-task convolutional neural networks for diabetic retinopathy grading. 2018 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC)

Google Scholar

24. Bodapati, JD, Shaik, NS, and Naralasetti, V. Deep convolution feature aggregation: an application to diabetic retinopathy severity level prediction. SIViP. (2021) 15:923–30. doi: 10.1007/s11760-020-01816-y

Crossref Full Text | Google Scholar

25. Bilal, A, Zhu, L, Deng, A, Lu, H, and Wu, N. AI-based automatic detection and classification of diabetic retinopathy using U-Net and deep learning. Symmetry. (2022) 14:1427. doi: 10.3390/sym14071427

Crossref Full Text | Google Scholar

26. Bilal, A, Sun, G, Mazhar, S, and Imran, A. (2022). Improved grey wolf optimization-based feature selection and classification using CNN for diabetic retinopathy detection. Evolutionary Computing and Mobile Sustainable Networks

Google Scholar

27. Canayaz, M. Classification of diabetic retinopathy with feature selection over deep features using nature-inspired wrapper methods. Appl Soft Comput. (2022) 128:109462. doi: 10.1016/j.asoc.2022.109462

Crossref Full Text | Google Scholar

28. Mohan, NJ, Murugan, R, Goel, T, Mirjalili, S, Singh, YK, Deb, D, et al. Optimal hybrid feature selection technique for diabetic retinopathy grading using fundus images. Sādhanā. (2023) 48:102. doi: 10.1007/s12046-023-02175-3

Crossref Full Text | Google Scholar

29. Chollet, F. Deep learning with Python. New York, NY: Simon and Schuster (2017).

Google Scholar

30. Tan, C, Sun, F, Kong, T, Zhang, W, Yang, C, and Liu, C. (2018). A survey on deep transfer learning. International Conference on Artificial Neural Networks

Google Scholar

31. Akram, MW, Li, G, Jin, Y, Chen, X, Zhu, C, and Ahmad, A. Automatic detection of photovoltaic module defects in infrared images with isolated and develop-model transfer deep learning. Sol Energy. (2020) 198:175–86. doi: 10.1016/j.solener.2020.01.055

Crossref Full Text | Google Scholar

32. Alanazi, MF, Ali, MU, Hussain, SJ, Zafar, A, Mohatram, M, Irfan, M, et al. Brain tumor/mass classification framework using magnetic-resonance-imaging-based isolated and developed transfer deep-learning model. Sensors. (2022) 22:372. doi: 10.3390/s22010372

PubMed Abstract | Crossref Full Text | Google Scholar

33. Ahmed, W, Hanif, A, Kallu, KD, Kouzani, AZ, Ali, MU, and Zafar, A. Photovoltaic panels classification using isolated and transfer learned deep neural models using infrared thermographic images. Sensors. (2021) 21:5668. doi: 10.3390/s21165668

PubMed Abstract | Crossref Full Text | Google Scholar

34. Luo, Y, Khan, MO, Tian, Y, Shi, M, Dou, Z, Elze, T, et al. (2023). FairVision: equitable deep learning for eye disease screening via fair identity scaling. arXiv. Available online at: https://arxiv.org/abs/2310.02492. [Epub ahead of preprint]

Google Scholar

35. Alam, MNU, Bahadur, EH, Masum, AKM, Noori, FM, and Uddin, MZ. SwAV-driven diagnostics: new perspectives on grading diabetic retinopathy from retinal photography. Front Robot AI. (2024) 11:11. doi: 10.3389/frobt.2024.1445565

PubMed Abstract | Crossref Full Text | Google Scholar

36. Yang, Y, Cai, Z, Qiu, S, and Xu, P. A novel transformer model with multiple instance learning for diabetic retinopathy classification. IEEE Access. (2024) 12:6768–76. doi: 10.1109/ACCESS.2024.3351473

Crossref Full Text | Google Scholar

37. Shaik, NS, and Cherukuri, TK. Lesion-aware attention with neural support vector machine for retinopathy diagnosis. Mach Vis Appl. (2021) 32:126. doi: 10.1007/s00138-021-01253-y

Crossref Full Text | Google Scholar

38. Kassani, SH, Kassani, PH, Khazaeinezhad, R, Wesolowski, MJ, Schneider, KA, and Deters, R. (2019). Diabetic retinopathy classification using a modified Xception architecture. 2019 IEEE International Symposium on Signal Processing and Information Technology (ISSPIT).

Google Scholar

39. Kumar, G, Chatterjee, S, and Chattopadhyay, C. DRISTI: a hybrid deep neural network for diabetic retinopathy diagnosis. SIViP. (2021) 15:1679–86. doi: 10.1007/s11760-021-01904-7

PubMed Abstract | Crossref Full Text | Google Scholar

40. Dondeti, V, Bodapati, JD, Shareef, SN, and Veeranjaneyulu, N. Deep convolution features in non-linear embedding space for fundus image classification. Rev Intell Artif. (2020) 34:307–13. doi: 10.18280/ria.340308

Crossref Full Text | Google Scholar

41. Bodapati, JD, Naralasetti, V, Shareef, SN, Hakak, S, Bilal, M, Maddikunta, PKR, et al. Blended multi-modal deep ConvNet features for diabetic retinopathy severity prediction. Electronics. (2020) 9:914. doi: 10.3390/electronics9060914

Crossref Full Text | Google Scholar

42. Gangwar, AK, and Ravi, V. (2021). Diabetic retinopathy detection using transfer learning and deep learning. Evolution in Computational Intelligence Singapore: Springer

Google Scholar

43. Bodapati, JD, Shaik, NS, and Naralasetti, V. Composite deep neural network with gated-attention mechanism for diabetic retinopathy severity classification. J Ambient Intell Humaniz Comput. (2021) 12:9825–39. doi: 10.1007/s12652-020-02727-z

Crossref Full Text | Google Scholar

Keywords: diabetic retinopathy, fundus imaging, deep learning model, lightweight model, transfer learning

Citation: Zafar A, Kim KS, Ali MU, Byun JH and Kim S-H (2025) A lightweight multi-deep learning framework for accurate diabetic retinopathy detection and multi-level severity identification. Front. Med. 12:1551315. doi: 10.3389/fmed.2025.1551315

Received: 25 December 2024; Accepted: 10 March 2025;
Published: 02 April 2025.

Edited by:

Yukun Zhou, University College London, United Kingdom

Reviewed by:

Min Shi, Harvard University, United States
Lie Ju, Monash University, Australia

Copyright © 2025 Zafar, Kim, Ali, Byun and Kim. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Seong-Han Kim, c2hraW04QHNlam9uZy5hYy5rcg==

These authors have contributed equally to this work and share first authorship

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.