<?xml version="1.0" encoding="utf-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" article-type="research-article" dtd-version="2.3" xml:lang="EN">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Med.</journal-id>
<journal-title>Frontiers in Medicine</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Med.</abbrev-journal-title>
<issn pub-type="epub">2296-858X</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fmed.2023.1150295</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Medicine</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>DME-DeepLabV3+: a lightweight model for diabetic macular edema extraction based on DeepLabV3+ architecture</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name>
<surname>Bai</surname>
<given-names>Yun</given-names>
</name>
<xref rid="aff1" ref-type="aff"><sup>1</sup></xref>
<xref rid="fn0001" ref-type="author-notes"><sup>&#x2020;</sup></xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Li</surname>
<given-names>Jing</given-names>
</name>
<xref rid="aff1" ref-type="aff"><sup>1</sup></xref>
<xref rid="fn0001" ref-type="author-notes"><sup>&#x2020;</sup></xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Shi</surname>
<given-names>Lianjun</given-names>
</name>
<xref rid="aff2" ref-type="aff"><sup>2</sup></xref>
<xref rid="fn0001" ref-type="author-notes"><sup>&#x2020;</sup></xref>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Jiang</surname>
<given-names>Qin</given-names>
</name>
<xref rid="aff2" ref-type="aff"><sup>2</sup></xref>
<xref rid="c003" ref-type="corresp"><sup>&#x002A;</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/1960293/overview"/>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Yan</surname>
<given-names>Biao</given-names>
</name>
<xref rid="aff3" ref-type="aff"><sup>3</sup></xref>
<xref rid="c001" ref-type="corresp"><sup>&#x002A;</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/1099889/overview"/>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Wang</surname>
<given-names>Zhenhua</given-names>
</name>
<xref rid="aff1" ref-type="aff"><sup>1</sup></xref>
<xref rid="c002" ref-type="corresp"><sup>&#x002A;</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/1520996/overview"/>
</contrib>
</contrib-group>
<aff id="aff1"><sup>1</sup><institution>College of Information Science, Shanghai Ocean University</institution>, <addr-line>Shanghai</addr-line>, <country>China</country></aff>
<aff id="aff2"><sup>2</sup><institution>The Affiliated Eye Hospital, Nanjing Medical University</institution>, <addr-line>Nanjing</addr-line>, <country>China</country></aff>
<aff id="aff3"><sup>3</sup><institution>Eye Institute, Eye and ENT Hospital, Shanghai Medical College, Fudan University</institution>, <addr-line>Shanghai</addr-line>, <country>China</country></aff>
<author-notes>
<fn fn-type="edited-by" id="fn0002"><p>Edited by: Darren Shu Jeng Ting, University of Nottingham, United Kingdom</p></fn>
<fn fn-type="edited-by" id="fn0003"><p>Reviewed by: Zizhong Hu, Nanjing Medical University, China; Rizwan Ali Naqvi, Sejong University, Republic of Korea</p></fn>
<corresp id="c001">&#x002A;Correspondence: Biao Yan, <email>yanbiao1982@hotmail.com</email></corresp>
<corresp id="c002">Zhenhua Wang, <email>zh-wang@shou.edu.cn</email></corresp>
<corresp id="c003">Qin Jiang, <email>jiangqin710@126.com</email></corresp>
<fn fn-type="equal" id="fn0001"><p><sup>&#x2020;</sup>These authors have contributed equally to this work</p></fn>
</author-notes>
<pub-date pub-type="epub">
<day>08</day>
<month>09</month>
<year>2023</year>
</pub-date>
<pub-date pub-type="collection">
<year>2023</year>
</pub-date>
<volume>10</volume>
<elocation-id>1150295</elocation-id>
<history>
<date date-type="received">
<day>24</day>
<month>01</month>
<year>2023</year>
</date>
<date date-type="accepted">
<day>25</day>
<month>08</month>
<year>2023</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x00A9; 2023 Bai, Li, Shi, Jiang, Yan and Wang.</copyright-statement>
<copyright-year>2023</copyright-year>
<copyright-holder>Bai, Li, Shi, Jiang, Yan and Wang</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/">
<p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p>
</license>
</permissions>
<abstract>
<sec id="sec1001">
<title>Introduction</title>
<p>Diabetic macular edema (DME) is a major cause of vision impairment in the patients with diabetes. Optical Coherence Tomography (OCT) is an important ophthalmic imaging method, which can enable early detection of DME. However, it is difficult to achieve high-efficiency and high-precision extraction of DME in OCT images because the sources of OCT images are diverse and the quality of OCT images is not stable. Thus, it is still required to design a model to improve the accuracy of DME extraction in OCT images.</p>
</sec>
<sec id="sec2001">
<title>Methods</title>
<p>A lightweight model (DME-DeepLabV3+) was proposed for DME extraction using a DeepLabV3+ architecture. In this model, MobileNetV2 model was used as the backbone for extracting low-level features of DME. The improved ASPP with sawtooth wave-like dilation rate was used for extracting high-level features of DME. Then, the decoder was used to fuse and refine low-level and high-level features of DME. Finally, 1711 OCT images were collected from the Kermany dataset and the Affiliated Eye Hospital. 1369, 171, and 171 OCT images were randomly selected for training, validation, and testing, respectively.</p>
</sec>
<sec id="sec3001">
<title>Conclusion</title>
<p>In ablation experiment, the proposed DME-DeepLabV3+ model was compared against DeepLabV3+ model with different setting to evaluate the effects of MobileNetV2 and improved ASPP on DME extraction. DME-DeepLabV3+ had better extraction performance, especially in small-scale macular edema regions. The extraction results of DME-DeepLabV3+ were close to ground truth. In comparative experiment, the proposed DME-DeepLabV3+ model was compared against other models, including FCN, UNet, PSPNet, ICNet, and DANet, to evaluate DME extraction performance. DME-DeepLabV3+ model had better DME extraction performance than other models as shown by greater pixel accuracy (PA), mean pixel accuracy (MPA), precision (Pre), recall (Re), F1-score (F1), and mean Intersection over Union (MIoU), which were 98.71%, 95.23%, 91.19%, 91.12%, 91.15%, and 91.18%, respectively.</p>
</sec>
<sec id="sec4001">
<title>Discussion</title>
<p>DME-DeepLabV3+ model is suitable for DME extraction in OCT images and can assist the ophthalmologists in the management of ocular diseases.</p>
</sec>
</abstract>
<kwd-group>
<kwd>diabetic macular edema</kwd>
<kwd>optical coherence tomography</kwd>
<kwd>deep learning</kwd>
<kwd>DeepLabV3+</kwd>
<kwd>extraction model</kwd>
</kwd-group>
<counts>
<fig-count count="8"/>
<table-count count="4"/>
<equation-count count="7"/>
<ref-count count="35"/>
<page-count count="11"/>
<word-count count="5620"/>
</counts>
<custom-meta-wrap>
<custom-meta>
<meta-name>section-at-acceptance</meta-name>
<meta-value>Ophthalmology</meta-value>
</custom-meta>
</custom-meta-wrap>
</article-meta>
</front>
<body>
<sec sec-type="intro" id="sec1">
<label>1.</label>
<title>Introduction</title>
<p>Diabetic macular edema (DME) is the major cause of vision loss in the patients with diabetic retinopathy. Increasing prevalence of DME is tightly correlated with the global epidemic of diabetes mellitus (<xref ref-type="bibr" rid="ref1">1</xref>, <xref ref-type="bibr" rid="ref2">2</xref>). DME is usually caused by the rupture of retinal barrier and increased permeability of retinal vessels, which is characterized by the leakage of fluid and other plasma components. The effusion can accumulate in the macula, resulting in edema (<xref ref-type="bibr" rid="ref3">3</xref>, <xref ref-type="bibr" rid="ref4">4</xref>). In the clinical work, the presence and severity of retinopathy are required to be determined according to the size of edema area.</p>
<p>Optical Coherence Tomography (OCT) is a non-contact, non-invasive, and highly sensitive ophthalmic imaging method, which can enable early detection of diabetic macular edema by observing the transverse section of macular degeneration (<xref ref-type="bibr" rid="ref5">5</xref>). Normal OCT image is shown in <xref rid="fig1" ref-type="fig">Figure 1A</xref> and OCT image with DME is shown in <xref rid="fig1" ref-type="fig">Figure 1B</xref>. DMEs accumulated in typical relative positions within the main retinal layers. Based on OCT patterns of DME, DME can be classified into three different patterns, including diffuse retinal thickening (DRT), cystoid macular edema (CME), and serous retinal detachment (SRD). CME normally starts to manifest symptoms in the inner retina, while SRD and DRT typically appear in the outer retina. In the severe advanced stages of DR, CMEs can also proliferate from the inner to the outer retina and merge with DRT (<xref ref-type="bibr" rid="ref6">6</xref>). Thus, rapid and accurate detection of all types of edemas is of great significance for evaluating the progression of diabetic retinopathy. In the clinical work, DME is usually segmented by the well-trained experts (<xref ref-type="bibr" rid="ref7">7</xref>). However, manual extraction of DME edemas is time-consuming and labor-intensive. Moreover, there is inevitable variability in the extraction results by different experts. With increased prevalence of diabetes, an increasing number of patients require disease management based on OCT images in the clinical practices. Thus, it is highly required to design an automatic method for rapid and accurate detection of DME in OCT images.</p>
<fig position="float" id="fig1">
<label>Figure 1</label>
<caption>
<p>Optical coherence tomography images in diabetic patients and healthy controls. <bold>(A)</bold> Normal OCT image; <bold>(B)</bold> OCT image with DME.</p>
</caption>
<graphic xlink:href="fmed-10-1150295-g001.tif"/>
</fig>
<p>Image extraction is processed and analyzed according to the features, including image color, spatial structure, and texture information (<xref ref-type="bibr" rid="ref8">8</xref>). Image extraction models can divide an image into several specific regions, such as threshold-based extraction model (<xref ref-type="bibr" rid="ref9">9</xref>, <xref ref-type="bibr" rid="ref10">10</xref>), region-based extraction model (<xref ref-type="bibr" rid="ref11">11</xref>, <xref ref-type="bibr" rid="ref12">12</xref>), and edge detection-based extraction model (<xref ref-type="bibr" rid="ref13">13</xref>). With the development of deep learning, several models have been developed to extract DME, such as fully convolutional network (FCN), U-Net, and PSPNet. Based on these deep learning models, several scholars have also developed the improved models for DME extraction. <xref rid="tab1" ref-type="table">Table 1</xref> showed the strengths and weaknesses of different models for DME extraction.</p>
<table-wrap position="float" id="tab1">
<label>Table 1</label>
<caption>
<p>Strengths and weaknesses of different models for DME extraction.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top">Models</th>
<th align="left" valign="top">Strengths</th>
<th align="left" valign="top">Weaknesses</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="middle">FCN (<xref ref-type="bibr" rid="ref14">14</xref>)</td>
<td align="left" valign="middle">End-to-end pixel-level classification without inputting size constraints</td>
<td align="left" valign="middle">Ignore target boundary details and lack spatial consistency</td>
</tr>
<tr>
<td align="left" valign="middle">U-Net (<xref ref-type="bibr" rid="ref15">15</xref>)</td>
<td align="left" valign="middle">Good extraction performance on small objects</td>
<td align="left" valign="middle">Down-sampling operators cause spatial information loss during encoding</td>
</tr>
<tr>
<td align="left" valign="middle">PSPNet (<xref ref-type="bibr" rid="ref16">16</xref>)</td>
<td align="left" valign="middle">Aggregate contextual information from different regions and improve the ability of obtaining global information</td>
<td align="left" valign="middle">No effective fusion of shallow features and missing target boundary details</td>
</tr>
<tr>
<td align="left" valign="middle">FCN&#x2009;+&#x2009;Sobel operator + Dijkstra algorithm (<xref ref-type="bibr" rid="ref17">17</xref>)</td>
<td align="left" valign="middle">Achieve better results in DICE index</td>
<td align="left" valign="middle" rowspan="2">Divide OCT extraction tasks into two stages, coarse and fine extraction, which makes OCT extraction cumbersome</td>
</tr>
<tr>
<td align="left" valign="middle">FCN&#x2009;+&#x2009;multiphase level set (<xref ref-type="bibr" rid="ref18">18</xref>)</td>
<td align="left" valign="middle">Avoid overlapping phenomenon of boundary and reduce the need for large training datasets</td>
</tr>
<tr>
<td align="left" valign="middle">U-Net&#x2009;+&#x2009;Bayesian deep learning (<xref ref-type="bibr" rid="ref19">19</xref>)</td>
<td align="left" valign="middle">Improve the accuracy of OCT image extraction with better versatility and interpretability</td>
<td align="left" valign="middle">Poor extraction performance for small-area objects</td>
</tr>
<tr>
<td align="left" valign="middle">PSPNet + dual attention mechanism (<xref ref-type="bibr" rid="ref20">20</xref>)</td>
<td align="left" valign="middle">Aggregate context information of different regions</td>
<td align="left" valign="middle">Insensitive to the information of fluid accumulation regions in DME</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>The sources of OCT images are diverse and the quality of OCT images is not always stable. Moreover, the size and distribution of DMEs are not uniform and the borders of DMEs are blurred. Thus, it is still required to design a novel model to improve the accuracy of DME extraction in OCT images. In this study, we proposed a lightweight automatic model (DME-DeepLabV3+) based on the DeepLabV3+ architecture. The major contributions of the proposed DME-DeepLabV3+ are shown below:</p>
<list>
<list-item>
<p>Taking MobileNetV2 as the backbone, the ability of DME-DeepLabV3+ is improved in extracting the low-level features of DME.</p>
</list-item>
<list-item>
<p>Improving ASPP by the sawtooth wave-like dilation rate, DME-DeepLabV3+ avoids grid effects, learns more local information, and extracts high-level features of DME better.</p>
</list-item>
<list-item>
<p>Based on the decoder, DME-DeepLabV3+ fuses the low-level and high-level features of DME, and refines the results of DME extraction.</p>
</list-item>
</list>
</sec>
<sec sec-type="materials|methods" id="sec2">
<label>2.</label>
<title>Materials and methods</title>
<p>The flowchart of the proposed model, DME-DeepLabV3+, was shown in <xref rid="fig2" ref-type="fig">Figure 2</xref>.</p>
<fig position="float" id="fig2">
<label>Figure 2</label>
<caption>
<p>Flowchart of DME extraction by DME-DeepLabV3+ model.</p>
</caption>
<graphic xlink:href="fmed-10-1150295-g002.tif"/>
</fig>
<p>Low-level features of DME extraction by MobileNetV2; High-level features of DME extraction by the improved ASPP; Fusion and refinement of low-level and high-level features of DME by the decoder.</p>
<sec id="sec3">
<label>2.1.</label>
<title>Low-level features of DME extraction by MobileNetV2</title>
<p>DeepLabV3+ is a deep learning model for image extraction with deep convolutional nets, which takes Xception as the backbone network (<xref ref-type="bibr" rid="ref21">21</xref>). Xception uses numerous parameters, complicated operations, and high computer performance requirements (<xref ref-type="bibr" rid="ref22">22</xref>), which leads to several challenges for DME extraction, such as fault-extraction and over-extraction problems. MobileNetV2 is a lightweight network, which shows a great advantage to solve the fault-extraction and over-extraction problems (<xref ref-type="bibr" rid="ref23">23</xref>). In DME-DeepLabV3+ model, we used MobileNetV2 as the backbone to simplify model structure, which could improve the extraction efficiency and reduce the problems of fault-extraction and over-extraction.</p>
<p>MobileNetV2 used depthwise separable convolution to reduce the number of parameters and complex operations. Depthwise separable convolution consisted of DepthWise (DW) and PointWise (PW), whereas DW performed convolution operations on each channel of the input layer and PW fused the features and obtained the feature information with stronger expressive ability. MobileNetV2 used Inverted Residual to improve the memory efficiency.</p>
<p>In Inverted Residual, the dimension of DME features was increased by 1&#x2009;&#x00D7;&#x2009;1 convolution. Next, DME features were extracted by 3&#x2009;&#x00D7;&#x2009;3 DW convolution, and the dimension of DME features was reduced by 1&#x2009;&#x00D7;&#x2009;1 convolution (<xref rid="fig3" ref-type="fig">Figure 3</xref>). When the stride was 1, DME output features were consistent with the input features and shortcuts were used to add the elements of DME input and output. When the stride was 2, no shortcut was required. At the same time, a linear bottleneck neural network was used in the last 1&#x2009;&#x00D7;&#x2009;1 convolutional layer of Inverted Residual, which could reduce the loss of low-dimensional feature of DME information.</p>
<fig position="float" id="fig3">
<label>Figure 3</label>
<caption>
<p>Structure of Inverted Residual. <bold>(A)</bold> Stride&#x2009;=&#x2009;1; <bold>(B)</bold> Stride&#x2009;=&#x2009;2.</p>
</caption>
<graphic xlink:href="fmed-10-1150295-g003.tif"/>
</fig>
<p>Compared with DeepLabV3+ with Xception as the backbone network, DME-DeepLabV3+ with MobileNetV2 as the backbone network not only improved the accuracy but also improved the efficiency in DME extraction.</p>
</sec>
<sec id="sec4">
<label>2.2.</label>
<title>High-level features of DME extraction by the improved ASPP</title>
<p>ASPP consists of atrous convolution with different dilation rates, which strikes the best trade-off between multi-scale feature extraction and context assimilation, especially for small objects (<xref ref-type="bibr" rid="ref24">24</xref>). DME has multi-scale features, especially with several small areas of edema. ASPP was then used to extract high-level features of DME. However, the dilation rate in ASPP had a grid effect, which not only lost the semantic information but also ignored the consistency of local information in edema regions (<xref ref-type="bibr" rid="ref25">25</xref>). Here, we replaced the original dilation rate with the sawtooth wave-like dilation rate to improve ASPP for extracting the high-level features of DME. A sawtooth wave-like dilation rate was formed by the repeated combination of two sets of the same &#x201C;rising edge&#x201D; type dilation rate.</p>
<p><xref rid="fig4" ref-type="fig">Figures 4</xref>, <xref rid="fig5" ref-type="fig">5</xref> show the illustration of the atrous convolution principle of DeepLabV3+ and DME-DeepLabV3+, respectively. <xref rid="fig4" ref-type="fig">Figures 4A</xref>,<xref rid="fig4" ref-type="fig">B</xref> show RF (receptive field) and the number of calculation times of DeepLabV3+. <xref rid="fig5" ref-type="fig">Figures 5A</xref>,<xref rid="fig5" ref-type="fig">B</xref> show RF and the number of calculation times of DME-DeepLabV3+. The results show that there was about 73% of information loss due to the grid effect in DeepLabV3+ model. In DME-DeepLabV3+ model, each pixel was effectively used and involved in further computations. Compared with DeepLabV3+ model, increased dilation rate in DME-DeepLabV3+ model can avoid the grid effects and learn more local information.</p>
<fig position="float" id="fig4">
<label>Figure 4</label>
<caption>
<p>Illustration of atrous convolution principle of DeepLabV3&#x2009;+ with &#x2009;dilation rate = [1, 6, 12, 18] and RF&#x2009; =&#x2009; 75&#x2009; &#x00D7;&#x2009; 75. <bold>(A)</bold> Effective pixels in RF, which were marked in blue; <bold>(B)</bold> The number of calculation times of each pixel.</p>
</caption>
<graphic xlink:href="fmed-10-1150295-g004.tif"/>
</fig>
<fig position="float" id="fig5">
<label>Figure 5</label>
<caption>
<p>Illustration of atrous convolution principle of DME-DeepLabV3+ with dilation rate = [1, 3, 9, 1, 3, 9] and RF&#x2009; =&#x2009; 53&#x2009; &#x00D7;&#x2009; 53. <bold>(A)</bold> Effective pixels in RF, which were marked in blue; <bold>(B)</bold> The number of calculation times of each pixel.</p>
</caption>
<graphic xlink:href="fmed-10-1150295-g005.tif"/>
</fig>
</sec>
<sec id="sec5">
<label>2.3.</label>
<title>Fusion and refinement of low-level and high-level features by the decoder</title>
<p>Low-level and high-level features of DME were extracted by MobileNetV2 and the improved ASPP, respectively. All features of DME were fused and refined by the decoder. The decoder is mainly composed of ordinary convolution and fusion layers. It fuses the features extracted from the encoder, uses the up-sampling to restore the feature dimension, and outputs the prediction results of the same size with less information loss as possible (<xref ref-type="bibr" rid="ref26">26</xref>). In the decoder, low-level features with fewer channels were obtained by 1&#x2009;&#x00D7;&#x2009;1 convolution. Bilinear up-sampling of high-level features were conducted by a factor of 4. The concatenation features were obtained by concatenating the low-level features and high-level features and a feature concatenation was refined by a few 3&#x2009;&#x00D7;&#x2009;3 convolutions. Finally, the results of DME extraction were output following another bilinear up-sampling by a factor of 4.</p>
</sec>
<sec id="sec6">
<label>2.4.</label>
<title>Ethical statement</title>
<p>The design and conduct of this study adhere to the intent and principles of the Declaration of Helsinki. The protocols were also reviewed and approved by the ethical committee of Eye Hospital (Nanjing medical university). Informed consents were obtained from all participants.</p>
</sec>
<sec id="sec7">
<label>2.5.</label>
<title>Datasets</title>
<p>The datasets contained 1711 OCT images, including 416 images (512&#x2009;&#x00D7;&#x2009;512 pixels) selected from the Kermany dataset (<xref ref-type="bibr" rid="ref27">27</xref>) and 1,295 OCT datasets (938&#x2009;&#x00D7;&#x2009;420 pixels) collected from the Affiliated Eye Hospital, Nanjing Medical University. All patients were required to undergo OCT scanning by a spectral domain OCT (RTVue, Optovue Inc., United States). These OCT images were centered on the macula with an axial resolution of 10&#x2009;&#x03BC;m and a 24-bit depth and were acquired in 2&#x2009;s, covering 4&#x2009;&#x00D7;&#x2009;4&#x2009;mm area. Inclusion criteria were as follows: the presence of macular edema in at least one eye and clear optical media allowing OCT imaging with good quality. Subsets of 1,369, 171, and 171 OCT images were randomly selected for training, validation, and testing, respectively. Each OCT image was individually labeled by three experienced clinicians who had more than 10-year clinical working experience. The annotation results were binarized by MATLAB software, where the background was labeled as 0 and the DME labeled as 1. Due to the limited human energy, some artificial deviations were inevitable. For these images, a senior expert was consulted and thorough rounds of discussion and adjudication were conducted to ensure the accuracy of the labeling. The original OCT images and ground truth are shown in <xref rid="fig6" ref-type="fig">Figure 6</xref>.</p>
<fig position="float" id="fig6">
<label>Figure 6</label>
<caption>
<p>Original OCT images and DME labeling by three experienced clinicians.</p>
</caption>
<graphic xlink:href="fmed-10-1150295-g006.tif"/>
</fig>
</sec>
</sec>
<sec id="sec8">
<label>3.</label>
<title>Implementation</title>
<p>The hardware configurations used for this study are shown below: Windows 10, NVIDIA GeForce RTX 3060. The software environment is the deep-learning framework PyTorch 1.10.0, CUDA 11.3, and the programming language Python 3.9.</p>
</sec>
<sec id="sec9">
<label>4.</label>
<title>Evaluation metrics</title>
<p>Seven metrics were calculated to estimate the extraction performance of DME-DeepLabV3+, including pixel accuracy (PA), mean pixel accuracy (MPA), precision (Pre), recall (Re), F1-score (F1), mean intersection over union (MIoU), and frames per second (FPS).</p>
<disp-formula id="EQ1">
<label>(1)</label>
<mml:math id="M1">
<mml:mrow>
<mml:mi mathvariant="normal">PA</mml:mi>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:msubsup>
<mml:mstyle displaystyle="true">
<mml:mo>&#x2211;</mml:mo>
</mml:mstyle>
<mml:mrow>
<mml:mi mathvariant="normal">i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
<mml:mi mathvariant="normal">k</mml:mi>
</mml:msubsup>
<mml:msub>
<mml:mi mathvariant="normal">p</mml:mi>
<mml:mrow>
<mml:mi mathvariant="normal">ii</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:msubsup>
<mml:mstyle displaystyle="true">
<mml:mo>&#x2211;</mml:mo>
</mml:mstyle>
<mml:mrow>
<mml:mi mathvariant="normal">i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
<mml:mi mathvariant="normal">k</mml:mi>
</mml:msubsup>
<mml:msubsup>
<mml:mstyle displaystyle="true">
<mml:mo>&#x2211;</mml:mo>
</mml:mstyle>
<mml:mrow>
<mml:mi mathvariant="normal">j</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
<mml:mi mathvariant="normal">k</mml:mi>
</mml:msubsup>
<mml:msub>
<mml:mi mathvariant="normal">p</mml:mi>
<mml:mrow>
<mml:mi mathvariant="normal">ij</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
</disp-formula>
<disp-formula id="EQ2">
<label>(2)</label>
<mml:math id="M2">
<mml:mrow>
<mml:mi mathvariant="normal">MPA</mml:mi>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mrow>
<mml:mi mathvariant="normal">k</mml:mi>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:mfrac>
<mml:munderover>
<mml:mstyle displaystyle="true">
<mml:mo>&#x2211;</mml:mo>
</mml:mstyle>
<mml:mrow>
<mml:mi mathvariant="normal">i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
<mml:mi mathvariant="normal">k</mml:mi>
</mml:munderover>
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="normal">p</mml:mi>
<mml:mrow>
<mml:mi mathvariant="normal">ii</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:msubsup>
<mml:mstyle displaystyle="true">
<mml:mo>&#x2211;</mml:mo>
</mml:mstyle>
<mml:mrow>
<mml:mi mathvariant="normal">j</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
<mml:mi mathvariant="normal">k</mml:mi>
</mml:msubsup>
<mml:msub>
<mml:mi mathvariant="normal">p</mml:mi>
<mml:mrow>
<mml:mi mathvariant="normal">ij</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
</disp-formula>
<disp-formula id="EQ3">
<label>(3)</label>
<mml:math id="M3">
<mml:mrow>
<mml:mi mathvariant="normal">Pre</mml:mi>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="normal">TP</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">TP</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="normal">FP</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
</disp-formula>
<disp-formula id="EQ4">
<label>(4)</label>
<mml:math id="M4">
<mml:mrow>
<mml:mi>Re</mml:mi>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="normal">TP</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">TP</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="normal">FN</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
</disp-formula>
<disp-formula id="EQ5">
<label>(5)</label>
<mml:math id="M5">
<mml:mrow>
<mml:mi mathvariant="normal">F</mml:mi>
<mml:mn>1</mml:mn>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mn>2</mml:mn>
<mml:mo>&#x00D7;</mml:mo>
<mml:mi mathvariant="normal">Pre</mml:mi>
<mml:mo>&#x00D7;</mml:mo>
<mml:mi>Re</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">Pre</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi>Re</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
</disp-formula>
<disp-formula id="EQ6">
<label>(6)</label>
<mml:math id="M6">
<mml:mrow>
<mml:mi mathvariant="normal">MIoU</mml:mi>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mrow>
<mml:mi mathvariant="normal">k</mml:mi>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:mfrac>
<mml:munderover>
<mml:mstyle displaystyle="true">
<mml:mo>&#x2211;</mml:mo>
</mml:mstyle>
<mml:mrow>
<mml:mi mathvariant="normal">i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
<mml:mi mathvariant="normal">k</mml:mi>
</mml:munderover>
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="normal">p</mml:mi>
<mml:mrow>
<mml:mi mathvariant="normal">ii</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:msubsup>
<mml:mstyle displaystyle="true">
<mml:mo>&#x2211;</mml:mo>
</mml:mstyle>
<mml:mrow>
<mml:mi mathvariant="normal">j</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
<mml:mi mathvariant="normal">k</mml:mi>
</mml:msubsup>
<mml:msub>
<mml:mi mathvariant="normal">p</mml:mi>
<mml:mrow>
<mml:mi mathvariant="normal">ij</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>+</mml:mo>
<mml:msubsup>
<mml:mstyle displaystyle="true">
<mml:mo>&#x2211;</mml:mo>
</mml:mstyle>
<mml:mrow>
<mml:mi mathvariant="normal">j</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
<mml:mi mathvariant="normal">k</mml:mi>
</mml:msubsup>
<mml:msub>
<mml:mi mathvariant="normal">p</mml:mi>
<mml:mrow>
<mml:mi mathvariant="normal">ji</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mi mathvariant="normal">p</mml:mi>
<mml:mrow>
<mml:mi mathvariant="normal">ii</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
</disp-formula>
<disp-formula id="EQ7">
<label>(7)</label>
<mml:math id="M7">
<mml:mrow>
<mml:mi mathvariant="normal">FPS</mml:mi>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="normal">frameNum</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">elapsedTime</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
</disp-formula>
<p>TP, FP, and FN denote the true positive region, false positive region, and false negative region, respectively. <inline-formula>
<mml:math id="M8">
<mml:mrow>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> is the number of edema area pixels which was correctly classified as edema areas; <inline-formula>
<mml:math id="M9">
<mml:mrow>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> is the number of background area pixels which are misclassified as edema areas; <inline-formula>
<mml:math id="M10">
<mml:mrow>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> is the number of edema area pixels which are incorrectly classified as the background; k is the labeling results of different classes, where k&#x2009;=&#x2009;0 expressed as background class and k&#x2009;=&#x2009;1 as DME class; <inline-formula>
<mml:math id="M11">
<mml:mrow>
<mml:mi>f</mml:mi>
<mml:mi>r</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>m</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>N</mml:mi>
<mml:mi>u</mml:mi>
<mml:mi>m</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> is the number of OCT images that are input to the model when performing inference; <inline-formula>
<mml:math id="M12">
<mml:mrow>
<mml:mi>e</mml:mi>
<mml:mi>l</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>p</mml:mi>
<mml:mi>s</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>d</mml:mi>
<mml:mi>T</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>m</mml:mi>
<mml:mi>e</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> is the time consumed by the model when performing inference. PA is the overall pixel accuracy. MPA is the average pixel accuracy of DME and background. Pre and Re are the proportion of real DME regions in the samples predicted as DME and the proportion of correct predictions in all DME, respectively. F1-score (F1) is a balanced metric and determined by precision and recall. MIoU is a metric to measure the similarity of ground truth and prediction. FPS is the number of OCT images inferred per second.</p>
</sec>
<sec sec-type="results" id="sec10">
<label>5.</label>
<title>Results</title>
<p>To evaluate the performance of DME extraction of DME-DeepLabV3+ model, two comparative experiments were performed. In experiment 1, DME extraction performance of DME-DeepLabV3+ model was evaluated by comparing against DeepLabV3+ model under different settings. In experiment 2, DME extraction performance of DME-DeepLabV3+ model was evaluated by comparing against other end-to-end models, including FCN, UNet, PSPNet, ICNet, and DANet.</p>
<sec id="sec11">
<label>5.1.</label>
<title>Experiment 1 (ablation experiment)</title>
<p>To evaluate the effects of MobileNetV2 and the improved ASPP on DME extraction performance, the proposed DME-DeepLabV3+ model was compared against DeepLabV3+ model with different settings, including DeepLabV3+, DeepLabV3+ with MobileNetV2 (MobileNetV2-DeepLabV3+), DeepLabV3+ with the improved ASPP (Improved ASPP-DeepLabV3+). <xref rid="fig7" ref-type="fig">Figure 7</xref> showed the DME extraction results by DeepLabV3+ model with different settings, where red, blue, and white DME regions represented true positive (TP) regions, false positive (FP) regions and false negative (FN) regions, respectively. DeepLabV3+ model led to some missed and false extraction of DME. MobileNetV2-DeepLabV3+ and improved ASPP-DeepLabV3+ reduced the missed and false extraction of DME. However, the missed extraction still existed in small edematous regions as shown in OCT2. DME-DeepLabV3+ had better extraction performance, especially in small-scale macular edema regions. The extraction results of DME-DeepLabV3+ were close to the ground truth.</p>
<fig position="float" id="fig7">
<label>Figure 7</label>
<caption>
<p>DME extraction results by DeepLabV3+ model with different settings Red, blue, and white DME regions represented TP regions, FP regions and FN regions.</p>
</caption>
<graphic xlink:href="fmed-10-1150295-g007.tif"/>
</fig>
<p><xref rid="tab2" ref-type="table">Table 2</xref> showed the results of evaluation metrics for DeepLabV3+ under different settings. Compared with DeepLabV3+ model, the MobileNetV2-DeepLabV3+ enhanced the scores of PA, MPA, Pre, Re, F1, MIoU, and FPS of DME extraction results, which were 98.69(0.28&#x2191;), 94.95(1.50&#x2191;), 91.02(0.57&#x2191;), 91.09(3.31&#x2191;), 91.06(1.97&#x2191;), 91.05(1.78&#x2191;), and 9.24(3.40&#x2191;), respectively. FPS increased by about 58%. The improved ASPP-DeepLabV3+ enhanced the scores of PA, MPA, Re, F1, and MIoU of DME extraction results, which were 98.47(0.06&#x2191;), 93.90(0.45&#x2191;), 88.49(0.71&#x2191;), 89.45(0.36&#x2191;), and 89.61(0.34&#x2191;), respectively. DME-DeepLabV3+ enhanced the scores of PA, MPA, Pre, Re, F1, and MIoU of DME extraction results, which were 98.71(0.30&#x2191;), 95.23(1.78&#x2191;), 91.19(0.74&#x2191;), 91.12(3.34&#x2191;), 91.15(2.06&#x2191;), and 91.18(1.91&#x2191;), respectively. FPS was 9.03, which was lower than that of MobileNetV2-DeepLabV3+ (0.21&#x2193;).</p>
<table-wrap position="float" id="tab2">
<label>Table 2</label>
<caption>
<p>Evaluation metrics of DME extraction by DeepLabV3+ model with different settings.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top" rowspan="2">Models</th>
<th align="center" valign="top" colspan="7">Evaluation metrics</th>
</tr>
<tr>
<th align="center" valign="top">PA(%)</th>
<th align="center" valign="top">MPA(%)</th>
<th align="center" valign="top">Pre(%)</th>
<th align="center" valign="top">Re(%)</th>
<th align="center" valign="top">F1(%)</th>
<th align="center" valign="top">MIoU(%)</th>
<th align="center" valign="top">FPS(it/s)</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="middle">DeepLabV3+</td>
<td align="char" valign="middle" char="&#x00B1;">98.41 &#x00B1; 0.07</td>
<td align="char" valign="middle" char="&#x00B1;">93.45 &#x00B1; 0.36</td>
<td align="char" valign="middle" char="&#x00B1;">90.45 &#x00B1; 0.85</td>
<td align="char" valign="middle" char="&#x00B1;">87.78 &#x00B1; 0.62</td>
<td align="char" valign="middle" char="&#x00B1;">89.09 &#x00B1; 0.16</td>
<td align="char" valign="middle" char="&#x00B1;">89.27 &#x00B1; 0.20</td>
<td align="char" valign="middle" char="&#x00B1;">5.84 &#x00B1; 0.87</td>
</tr>
<tr>
<td align="left" valign="middle">MobileNetV2-DeepLabV3+</td>
<td align="char" valign="middle" char="&#x00B1;">98.69 &#x00B1; 0.01</td>
<td align="char" valign="middle" char="&#x00B1;">94.95 &#x00B1; 0.14</td>
<td align="char" valign="middle" char="&#x00B1;">91.02 &#x00B1; 0.58</td>
<td align="char" valign="middle" char="&#x00B1;">91.09 &#x00B1; 0.43</td>
<td align="char" valign="middle" char="&#x00B1;">91.06 &#x00B1; 0.08</td>
<td align="char" valign="middle" char="&#x00B1;">91.05 &#x00B1; 0.05</td>
<td align="char" valign="middle" char="&#x00B1;">9.24 &#x00B1; 0.33</td>
</tr>
<tr>
<td align="left" valign="middle">Improved ASPP-DeepLabV3+</td>
<td align="char" valign="middle" char="&#x00B1;">98.47 &#x00B1; 0.03</td>
<td align="char" valign="middle" char="&#x00B1;">93.90 &#x00B1; 0.40</td>
<td align="char" valign="middle" char="&#x00B1;">90.43 &#x00B1; 0.30</td>
<td align="char" valign="middle" char="&#x00B1;">88.49 &#x00B1; 0.68</td>
<td align="char" valign="middle" char="&#x00B1;">89.45 &#x00B1; 0.42</td>
<td align="char" valign="middle" char="&#x00B1;">89.61 &#x00B1; 0.25</td>
<td align="char" valign="middle" char="&#x00B1;">5.59 &#x00B1; 0.13</td>
</tr>
<tr>
<td align="left" valign="middle">DME-DeepLabV3+</td>
<td align="char" valign="middle" char="&#x00B1;">98.71 &#x00B1; 0.02</td>
<td align="char" valign="middle" char="&#x00B1;">95.23 &#x00B1; 0.26</td>
<td align="char" valign="middle" char="&#x00B1;">91.19 &#x00B1; 0.44</td>
<td align="char" valign="middle" char="&#x00B1;">91.12 &#x00B1; 0.48</td>
<td align="char" valign="middle" char="&#x00B1;">91.15 &#x00B1; 0.02</td>
<td align="char" valign="middle" char="&#x00B1;">91.18 &#x00B1; 0.09</td>
<td align="char" valign="middle" char="&#x00B1;">9.03 &#x00B1; 0.43</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="sec12">
<label>5.2.</label>
<title>Experiment 2 (comparative experiment)</title>
<p>We evaluate DME extraction performance of DME-DeepLabV3+ model by comparing against other models, including FCN, UNet, PSPNet, ICNet, and DANet. <xref rid="fig8" ref-type="fig">Figure 8</xref> showed DME extraction results by different models, where red, blue, and white DME regions represented true positive (TP) regions, false positive (FP) regions, and false negative (FN) regions, respectively. As shown in <xref rid="fig8" ref-type="fig">Figure 8</xref>, DME extraction results of DME-DeepLabV3+ were close to the ground truth. A part of the background was extracted falsely by FCN, U-Net, or DANet models. Compared with FCN and U-Net, PSPNet and ICNet reduced the fault-extraction, but the small-scale macular edema was over-extracted. <xref rid="tab3" ref-type="table">Table 3</xref> showed the parameter configurations of different models and <xref rid="tab4" ref-type="table">Table 4</xref> showed the results of DME evaluation metrics. Compared with FCN, U-Net, PSPNet, and DANet models, DME-DeepLabV3+ achieved higher scores of PA, MPA, and FPS. As for Pre, Re, F1, and MIoU, DME-DeepLabV3+ substantially exceeded other models. Compared with ICNet, DME-DeepLabV3+ achieved a better trade-off in the accuracy and efficiency for DME extraction.</p>
<fig position="float" id="fig8">
<label>Figure 8</label>
<caption>
<p>DME extraction results by different models Red, blue, and white DME regions represented TP regions, FP regions and FN regions.</p>
</caption>
<graphic xlink:href="fmed-10-1150295-g008.tif"/>
</fig>
<table-wrap position="float" id="tab3">
<label>Table 3</label>
<caption>
<p>Parameter configurations of different DME extraction models.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top">Models</th>
<th align="left" valign="top">Backbone</th>
<th align="center" valign="top">Learning rate</th>
<th align="center" valign="top">Total epochs</th>
<th align="center" valign="top">Batch size</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="middle">FCN</td>
<td align="left" valign="middle">ResNet50</td>
<td align="char" valign="middle" char=".">0.01</td>
<td align="center" valign="middle">200</td>
<td align="center" valign="middle">1</td>
</tr>
<tr>
<td align="left" valign="middle">U-Net</td>
<td align="left" valign="middle">VGG16</td>
<td align="char" valign="middle" char=".">0.01</td>
<td align="center" valign="middle">200</td>
<td align="center" valign="middle">2</td>
</tr>
<tr>
<td align="left" valign="middle">PSPNet</td>
<td align="left" valign="middle">ResNet50</td>
<td align="char" valign="middle" char=".">0.01</td>
<td align="center" valign="middle">200</td>
<td align="center" valign="middle">2</td>
</tr>
<tr>
<td align="left" valign="middle">ICNet</td>
<td align="left" valign="middle">ResNet50</td>
<td align="char" valign="middle" char=".">0.01</td>
<td align="center" valign="middle">200</td>
<td align="center" valign="middle">2</td>
</tr>
<tr>
<td align="left" valign="middle">DANet</td>
<td align="left" valign="middle">ResNet101</td>
<td align="char" valign="middle" char=".">0.0001</td>
<td align="center" valign="middle">200</td>
<td align="center" valign="middle">2</td>
</tr>
<tr>
<td align="left" valign="middle">DME-DeepLabV3+</td>
<td align="left" valign="middle">MobileNetV2</td>
<td align="char" valign="middle" char=".">0.01</td>
<td align="center" valign="middle">200</td>
<td align="center" valign="middle">2</td>
</tr>
</tbody>
</table>
</table-wrap>
<table-wrap position="float" id="tab4">
<label>Table 4</label>
<caption>
<p>Evaluation metrics of DME extraction by different models.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top" rowspan="2">Models</th>
<th align="center" valign="top" colspan="7">Evaluation metrics</th>
</tr>
<tr>
<th align="center" valign="top">PA(%)</th>
<th align="center" valign="top">MPA(%)</th>
<th align="center" valign="top">Pre(%)</th>
<th align="center" valign="top">Re(%)</th>
<th align="center" valign="top">F1(%)</th>
<th align="center" valign="top">MIoU(%)</th>
<th align="center" valign="top">FPS(it/s)</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="middle">FCN</td>
<td align="char" valign="middle" char="&#x00B1;">98.27 &#x00B1; 0.11</td>
<td align="char" valign="middle" char="&#x00B1;">89.88 &#x00B1; 1.50</td>
<td align="char" valign="middle" char="&#x00B1;">81.76 &#x00B1; 1.20</td>
<td align="char" valign="middle" char="&#x00B1;">79.44 &#x00B1; 1.40</td>
<td align="char" valign="middle" char="&#x00B1;">80.58 &#x00B1; 0.85</td>
<td align="char" valign="middle" char="&#x00B1;">82.66 &#x00B1; 0.48</td>
<td align="char" valign="middle" char="&#x00B1;">4.08 &#x00B1; 0.20</td>
</tr>
<tr>
<td align="left" valign="middle">U-Net</td>
<td align="char" valign="middle" char="&#x00B1;">98.61 &#x00B1; 0.01</td>
<td align="char" valign="middle" char="&#x00B1;">90.73 &#x00B1; 1.05</td>
<td align="char" valign="middle" char="&#x00B1;">86.31 &#x00B1; 0.41</td>
<td align="char" valign="middle" char="&#x00B1;">81.15 &#x00B1; 0.65</td>
<td align="char" valign="middle" char="&#x00B1;">83.58 &#x00B1; 0.15</td>
<td align="char" valign="middle" char="&#x00B1;">85.32 &#x00B1; 0.23</td>
<td align="char" valign="middle" char="&#x00B1;">3.43 &#x00B1; 0.24</td>
</tr>
<tr>
<td align="left" valign="middle">PSPNet</td>
<td align="char" valign="middle" char="&#x00B1;">98.69 &#x00B1; 0.04</td>
<td align="char" valign="middle" char="&#x00B1;">92.52 &#x00B1; 0.41</td>
<td align="char" valign="middle" char="&#x00B1;">84.61 &#x00B1; 1.27</td>
<td align="char" valign="middle" char="&#x00B1;">85.76 &#x00B1; 0.90</td>
<td align="char" valign="middle" char="&#x00B1;">85.17 &#x00B1; 0.24</td>
<td align="char" valign="middle" char="&#x00B1;">86.41 &#x00B1; 0.21</td>
<td align="char" valign="middle" char="&#x00B1;">7.96 &#x00B1; 0.48</td>
</tr>
<tr>
<td align="left" valign="middle">ICNet</td>
<td align="char" valign="middle" char="&#x00B1;">98.07 &#x00B1; 0.02</td>
<td align="char" valign="middle" char="&#x00B1;">90.94 &#x00B1; 0.20</td>
<td align="char" valign="middle" char="&#x00B1;">90.73 &#x00B1; 0.78</td>
<td align="char" valign="middle" char="&#x00B1;">82.57 &#x00B1; 0.93</td>
<td align="char" valign="middle" char="&#x00B1;">86.45 &#x00B1; 0.29</td>
<td align="char" valign="middle" char="&#x00B1;">86.86 &#x00B1; 0.13</td>
<td align="char" valign="middle" char="&#x00B1;">15.86 &#x00B1; 0.30</td>
</tr>
<tr>
<td align="left" valign="middle">DANet</td>
<td align="char" valign="middle" char="&#x00B1;">98.06 &#x00B1; 0.01</td>
<td align="char" valign="middle" char="&#x00B1;">92.15 &#x00B1; 0.16</td>
<td align="char" valign="middle" char="&#x00B1;">87.44 &#x00B1; 0.77</td>
<td align="char" valign="middle" char="&#x00B1;">85.68 &#x00B1; 1.16</td>
<td align="char" valign="middle" char="&#x00B1;">86.54 &#x00B1; 0.21</td>
<td align="char" valign="middle" char="&#x00B1;">87.11 &#x00B1; 0.08</td>
<td align="char" valign="middle" char="&#x00B1;">6.12 &#x00B1; 0.50</td>
</tr>
<tr>
<td align="left" valign="middle">DME-DeepLabV3+</td>
<td align="char" valign="middle" char="&#x00B1;">98.71 &#x00B1; 0.02</td>
<td align="char" valign="middle" char="&#x00B1;">95.23 &#x00B1; 0.26</td>
<td align="char" valign="middle" char="&#x00B1;">91.19 &#x00B1; 0.44</td>
<td align="char" valign="middle" char="&#x00B1;">91.12 &#x00B1; 0.48</td>
<td align="char" valign="middle" char="&#x00B1;">91.15 &#x00B1; 0.02</td>
<td align="char" valign="middle" char="&#x00B1;">91.18 &#x00B1; 0.09</td>
<td align="char" valign="middle" char="&#x00B1;">9.03 &#x00B1; 0.43</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
</sec>
<sec id="sec13">
<label>6.</label>
<title>Conclusion and discussion</title>
<p>With the increased incidence of diabetes, DME has become a major cause of visual impairment in diabetic patients (<xref ref-type="bibr" rid="ref28">28</xref>). DME occurs as a result of the disruption of blood-retinal barrier and consequent increase in vascular permeability (<xref ref-type="bibr" rid="ref29">29</xref>). OCT allows longitudinal, functional, and microstructural analysis of human macula (<xref ref-type="bibr" rid="ref30">30</xref>). However, manual labeling DME is time-consuming and labor-intensive (<xref ref-type="bibr" rid="ref31">31</xref>). Automatic extraction of DME based on machine learning can help physicians assess disease severity, determine treatment options, and improve life quality of patients (<xref ref-type="bibr" rid="ref32">32</xref>). Thus, it is urgent to develop an efficient model for DME detection. In this study, we proposed a lightweight model based on DeepLabV3+, termed DME-DeepLabV3+, to extract DME in OCT images. MobileNetV2 architecture was used as the backbone to extract the low-level features of DME and reduce the model complexity to enhance DME detection accuracy. With the help of improved ASPP structure, DME-DeepLabV3+ avoided the grid effects and learned more local information. Finally, the decoder was used to fuse the low-level and high-level features of DME and refined the results of DME extraction.</p>
<p>OCT image modality has been widely used for detecting DME due to its non-invasive and high-resolution features. Considering the clinical characteristics that are present in OCT images such as thickness, reflectivity or intraretinal fluid accumulation, DMEs have been categorized into three different types: SRD, DRT, and CME. Traditional DME detection is based on the low-level hand-crafted features, which require significant domain knowledge and are sensitive to the variations of lesions. Given great variability of morphology, shape, and relative ME position, it is difficult to detect all three ME types simultaneously. Our proposed model can achieve automatic and simultaneous detection of all three types of ME (SRD, DRT, and CME) in the ophthalmological field. However, the accuracy of DRT detection is still not good as SRD or CME detection. DRT is characterized by a sponge-like retinal swelling of the macula with reduced intraretinal reflectivity. In addition, DRT is characterized by uniform thickening of inner retinal layers but without macroscopic optical empty spaces. Thus, further improvement of our proposed model is still required for enhancing the accuracy of the automatic detection of DRT edemas.</p>
<p>In clinical practice, layer segmentation and fluid area segmentation can provide qualitative information and visualization of retinal structure, which is important for DME assessment and monitoring. Although commercial OCT devices with on-board proprietary segmentation software are available, the definition of retinal boundaries varies between the manufacturers, making the quantitative retinal thickness difficult. In addition, proprietary software is difficult to be used for image analysis from other OCT devices, which poses a great challenge for effective diagnosis of DME (<xref ref-type="bibr" rid="ref33">33</xref>). Although automated methods for layer segmentation have been proposed, most of them usually ignore the priority of mutually exclusive relationships between different layers, which can also affect the accuracy of DME assessment (<xref ref-type="bibr" rid="ref34">34</xref>). In future study, we will improve our model to consider both layer segmentation and fluid area segmentation for better monitoring the progression of DME in retinal diseases.</p>
<p>Both microaneurysm (MA) formation and DME lesions are the important signs of DR. Early and accurate detection of DME and MAs can reduce the risk of DR. Due to the small size of MA lesions and low contrast between MA lesion and retinal background, automated MA detection is still challenging. Many imaging modalities have been used to detect MAs, including color fundus images, optical coherence tomography angiography (OCTA), and fluorescein fundus angiography (FFA). However, MAs are situated on the capillaries, which are not often visible in color fundus images. Although FFA can capture the small changes of retinal vessels, FFA is an invasive method compared with other imaging modalities. OCTA can provide the detailed visualization of vascular perfusion and allow for the examination of retinal vasculature in 3D (<xref ref-type="bibr" rid="ref35">35</xref>). In future study, we would also improve our model by considering the segmentation of FFA for better monitoring the progression of DME in retinal diseases. We would design modules with better feature extraction capabilities, such as embedding attention mechanism to the model, strengthening key information, suppressing useless information, and better capturing contextual information, to improve the generalization of the model for the diagnosis of retinal diseases.</p>
</sec>
<sec sec-type="data-availability" id="sec14">
<title>Data availability statement</title>
<p>The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.</p>
</sec>
<sec id="sec15">
<title>Author contributions</title>
<p>BY, ZW, and QJ were responsible for the conceptualization and data collection. BY, ZW, and YB were responsible for the experiment design and manuscript writing. JL and LS conducted the data collection and data entry. BY and ZW were responsible for overall supervision and manuscript revision. All authors contributed to the article and approved the submitted version.</p>
</sec>
<sec sec-type="funding-information" id="sec16">
<title>Funding</title>
<p>This research was generously supported by the grants from the National Natural Science Foundation of China (Grant Nos. 82171074 and 82070983).</p>
</sec>
<sec sec-type="COI-statement" id="sec17">
<title>Conflict of interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
<p>The reviewer ZH declared a shared parent affiliation with the authors LS, QJ, and BY to the handling editor at the time of review.</p>
</sec>
<sec id="sec100" sec-type="disclaimer">
<title>Publisher&#x2019;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
</body>
<back>
<ref-list>
<title>References</title>
<ref id="ref1"><label>1.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Schmidt-Erfurth</surname><given-names>U</given-names></name> <name><surname>Garcia-Arumi</surname><given-names>J</given-names></name> <name><surname>Bandello</surname><given-names>F</given-names></name> <name><surname>Berg</surname><given-names>K</given-names></name> <name><surname>Chakravarthy</surname><given-names>U</given-names></name> <name><surname>Gerendas</surname><given-names>BS</given-names></name> <etal/></person-group>. <article-title>Guidelines for the Management of Diabetic Macular Edema by the European Society of Retina Specialists (Euretina)</article-title>. <source>Ophthalmologica</source>. (<year>2017</year>) <volume>237</volume>:<fpage>185</fpage>&#x2013;<lpage>222</lpage>. doi: <pub-id pub-id-type="doi">10.1159/000458539</pub-id>, PMID: <pub-id pub-id-type="pmid">28423385</pub-id></citation></ref>
<ref id="ref2"><label>2.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Steinmetz</surname><given-names>JD</given-names></name> <name><surname>Bourne</surname><given-names>RR</given-names></name> <name><surname>Briant</surname><given-names>PS</given-names></name> <name><surname>Flaxman</surname><given-names>SR</given-names></name> <name><surname>Taylor</surname><given-names>HR</given-names></name> <name><surname>Jonas</surname><given-names>JB</given-names></name> <etal/></person-group>. <article-title>Causes of blindness and vision impairment in 2020 and trends over 30 years, and prevalence of avoidable blindness in relation to vision 2020: the right to sight: an analysis for the global burden of disease study</article-title>. <source>Lancet Glob Health</source>. (<year>2021</year>) <volume>9</volume>:<fpage>e144</fpage>&#x2013;<lpage>60</lpage>. doi: <pub-id pub-id-type="doi">10.1016/S2214-109X(20)30489-7</pub-id>, PMID: <pub-id pub-id-type="pmid">33275949</pub-id></citation></ref>
<ref id="ref3"><label>3.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Tan</surname><given-names>GS</given-names></name> <name><surname>Cheung</surname><given-names>N</given-names></name> <name><surname>Simo</surname><given-names>R</given-names></name> <name><surname>Cheung</surname><given-names>GC</given-names></name> <name><surname>Wong</surname><given-names>TY</given-names></name></person-group>. <article-title>Diabetic Macular Oedema</article-title>. <source>Lancet Diabetes Endocrinol</source>. (<year>2017</year>) <volume>5</volume>:<fpage>143</fpage>&#x2013;<lpage>55</lpage>. doi: <pub-id pub-id-type="doi">10.1016/S2213-8587(16)30052-3</pub-id></citation></ref>
<ref id="ref4"><label>4.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Das</surname><given-names>A</given-names></name> <name><surname>McGuire</surname><given-names>PG</given-names></name> <name><surname>Rangasamy</surname><given-names>S</given-names></name></person-group>. <article-title>Diabetic macular edema: pathophysiology and novel therapeutic targets</article-title>. <source>Ophthalmology</source>. (<year>2015</year>) <volume>122</volume>:<fpage>1375</fpage>&#x2013;<lpage>94</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.ophtha.2015.03.024</pub-id>, PMID: <pub-id pub-id-type="pmid">25935789</pub-id></citation></ref>
<ref id="ref5"><label>5.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Van Melkebeke</surname><given-names>L</given-names></name> <name><surname>Barbosa-Breda</surname><given-names>J</given-names></name> <name><surname>Huygens</surname><given-names>M</given-names></name> <name><surname>Stalmans</surname><given-names>I</given-names></name></person-group>. <article-title>Optical coherence tomography angiography in Glaucoma: a review</article-title>. <source>Ophthalmic Res</source>. (<year>2018</year>) <volume>60</volume>:<fpage>139</fpage>&#x2013;<lpage>51</lpage>. doi: <pub-id pub-id-type="doi">10.1159/000488495</pub-id>, PMID: <pub-id pub-id-type="pmid">29794471</pub-id></citation></ref>
<ref id="ref6"><label>6.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wu</surname><given-names>Q</given-names></name> <name><surname>Zhang</surname><given-names>B</given-names></name> <name><surname>Hu</surname><given-names>Y</given-names></name> <name><surname>Liu</surname><given-names>B</given-names></name> <name><surname>Cao</surname><given-names>D</given-names></name> <name><surname>Yang</surname><given-names>D</given-names></name> <etal/></person-group>. <article-title>Detection of morphologic patterns of diabetic macular edema using a deep learning approach based on optical coherence tomography images</article-title>. <source>Retina</source>. <volume>41</volume>:<fpage>1110</fpage>&#x2013;<lpage>7</lpage>. doi: <pub-id pub-id-type="doi">10.1097/iae.0000000000002992</pub-id>, PMID: <pub-id pub-id-type="pmid">33031250</pub-id></citation></ref>
<ref id="ref7"><label>7.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname><given-names>Z</given-names></name> <name><surname>Zhong</surname><given-names>Y</given-names></name> <name><surname>Yao</surname><given-names>M</given-names></name> <name><surname>Ma</surname><given-names>Y</given-names></name> <name><surname>Zhang</surname><given-names>W</given-names></name> <name><surname>Li</surname><given-names>C</given-names></name> <etal/></person-group>. <article-title>Automated segmentation of macular edema for the diagnosis of ocular disease using deep learning method</article-title>. <source>Sci Rep</source>. (<year>2021</year>) <volume>11</volume>:<fpage>1</fpage>&#x2013;<lpage>12</lpage>. doi: <pub-id pub-id-type="doi">10.1038/s41598-021-92458-8</pub-id>, PMID: <pub-id pub-id-type="pmid">34183684</pub-id></citation></ref>
<ref id="ref8"><label>8.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Liu</surname><given-names>X</given-names></name> <name><surname>Song</surname><given-names>L</given-names></name> <name><surname>Liu</surname><given-names>S</given-names></name> <name><surname>Zhang</surname><given-names>Y</given-names></name></person-group>. <article-title>A review of deep-learning-based medical image segmentation methods</article-title>. <source>Sustainability</source>. (<year>2021</year>) <volume>13</volume>:<fpage>1224</fpage>. doi: <pub-id pub-id-type="doi">10.3390/su13031224</pub-id></citation></ref>
<ref id="ref9"><label>9.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Makandar</surname><given-names>A</given-names></name> <name><surname>Halalli</surname><given-names>B</given-names></name></person-group>. <article-title>Threshold based segmentation technique for mass detection in mammography</article-title>. <source>J Comput</source>. (<year>2016</year>) <volume>11</volume>:<fpage>472</fpage>&#x2013;<lpage>8</lpage>. doi: <pub-id pub-id-type="doi">10.17706/jcp.11.6.472-478</pub-id></citation></ref>
<ref id="ref10"><label>10.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zebari</surname><given-names>DA</given-names></name> <name><surname>Zeebaree</surname><given-names>DQ</given-names></name> <name><surname>Abdulazeez</surname><given-names>AM</given-names></name> <name><surname>Haron</surname><given-names>H</given-names></name> <name><surname>Hamed</surname><given-names>HNA</given-names></name></person-group>. <article-title>Improved threshold based and trainable fully automated segmentation for breast Cancer boundary and pectoral muscle in mammogram images</article-title>. <source>IEEE Access</source>. (<year>2020</year>) <volume>8</volume>:<fpage>203097</fpage>&#x2013;<lpage>116</lpage>. doi: <pub-id pub-id-type="doi">10.1109/ACCESS.2020.3036072</pub-id></citation></ref>
<ref id="ref11"><label>11.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Biratu</surname><given-names>ES</given-names></name> <name><surname>Schwenker</surname><given-names>F</given-names></name> <name><surname>Debelee</surname><given-names>TG</given-names></name> <name><surname>Kebede</surname><given-names>SR</given-names></name> <name><surname>Negera</surname><given-names>WG</given-names></name> <name><surname>Molla</surname><given-names>HT</given-names></name></person-group>. <article-title>Enhanced region growing for brain tumor Mr image segmentation</article-title>. <source>J Imaging</source>. (<year>2021</year>) <volume>7</volume>:<fpage>22</fpage>. doi: <pub-id pub-id-type="doi">10.3390/jimaging7020022</pub-id>, PMID: <pub-id pub-id-type="pmid">34460621</pub-id></citation></ref>
<ref id="ref12"><label>12.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Liu</surname><given-names>J</given-names></name> <name><surname>Yan</surname><given-names>S</given-names></name> <name><surname>Lu</surname><given-names>N</given-names></name> <name><surname>Yang</surname><given-names>D</given-names></name> <name><surname>Fan</surname><given-names>C</given-names></name> <name><surname>Lv</surname><given-names>H</given-names></name> <etal/></person-group>. <article-title>Automatic segmentation of foveal avascular zone based on adaptive watershed algorithm in retinal optical coherence tomography angiography images</article-title>. <source>J Innov Opt Health Sci</source>. (<year>2022</year>) <volume>15</volume>:<fpage>2242001</fpage>. doi: <pub-id pub-id-type="doi">10.1142/S1793545822420019</pub-id></citation></ref>
<ref id="ref13"><label>13.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chatterjee</surname><given-names>S</given-names></name> <name><surname>Suman</surname><given-names>A</given-names></name> <name><surname>Gaurav</surname><given-names>R</given-names></name> <name><surname>Banerjee</surname><given-names>S</given-names></name> <name><surname>Singh</surname><given-names>AK</given-names></name> <name><surname>Ghosh</surname><given-names>BK</given-names></name> <etal/></person-group>. <article-title>Retinal blood vessel segmentation using edge detection method</article-title>. <source>J Phys Conf Ser</source>. (<year>2021</year>) <volume>1717</volume>:<fpage>012008</fpage>. doi: <pub-id pub-id-type="doi">10.1088/1742-6596/1717/1/012008</pub-id></citation></ref>
<ref id="ref14"><label>14.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Long</surname><given-names>J</given-names></name> <name><surname>Shelhamer</surname><given-names>E</given-names></name> <name><surname>Darrell</surname><given-names>T</given-names></name></person-group>. <article-title>Fully convolutional networks for semantic segmentation</article-title>. <source>IEEE Trans Pattern Anal Mach Intell</source>. (<year>2017</year>) <volume>39</volume>:<fpage>640</fpage>&#x2013;<lpage>51</lpage>. doi: <pub-id pub-id-type="doi">10.1109/cvpr.2015.7298965</pub-id></citation></ref>
<ref id="ref15"><label>15.</label><citation citation-type="other"><person-group person-group-type="author"><name><surname>Ronneberger</surname><given-names>O</given-names></name> <name><surname>Fischer</surname><given-names>P</given-names></name> <name><surname>Brox</surname><given-names>T</given-names></name></person-group>. <italic>U-net: convolutional networks for biomedical image segmentation</italic>. International conference on medical image computing and computer-assisted intervention. Berlin: Springer (<year>2015</year>).</citation></ref>
<ref id="ref16"><label>16.</label><citation citation-type="other"><person-group person-group-type="author"><name><surname>Zhao</surname><given-names>H</given-names></name> <name><surname>Shi</surname><given-names>J</given-names></name> <name><surname>Qi</surname><given-names>X</given-names></name> <name><surname>Wang</surname><given-names>X</given-names></name> <name><surname>Jia</surname><given-names>J</given-names></name></person-group>. <italic>Pyramid scene parsing network. Proceedings of the IEEE conference on computer vision and pattern recognition</italic>. (<year>2017</year>).</citation></ref>
<ref id="ref17"><label>17.</label><citation citation-type="book"><person-group person-group-type="author"><name><surname>Ben-Cohen</surname><given-names>A</given-names></name> <name><surname>Mark</surname><given-names>D</given-names></name> <name><surname>Kovler</surname><given-names>I</given-names></name> <name><surname>Zur</surname><given-names>D</given-names></name> <name><surname>Barak</surname><given-names>A</given-names></name> <name><surname>Iglicki</surname><given-names>M</given-names></name> <etal/></person-group>. <source>Retinal layers segmentation using fully convolutional network in Oct images</source>. <publisher-loc>Israel</publisher-loc>: <publisher-name>RSIP Vision</publisher-name>, <fpage>1</fpage>&#x2013;<lpage>8</lpage>. (<year>2017</year>).</citation></ref>
<ref id="ref18"><label>18.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ruan</surname><given-names>Y</given-names></name> <name><surname>Xue</surname><given-names>J</given-names></name> <name><surname>Li</surname><given-names>T</given-names></name> <name><surname>Liu</surname><given-names>D</given-names></name> <name><surname>Lu</surname><given-names>H</given-names></name> <name><surname>Chen</surname><given-names>M</given-names></name> <etal/></person-group>. <article-title>Multi-phase level set algorithm based on fully convolutional networks (Fcn-Mls) for retinal layer segmentation in Sd-Oct images with central serous Chorioretinopathy (Csc)</article-title>. <source>Biomed Opt Express</source>. (<year>2019</year>) <volume>10</volume>:<fpage>3987</fpage>&#x2013;<lpage>4002</lpage>. doi: <pub-id pub-id-type="doi">10.1364/BOE.10.003987</pub-id>, PMID: <pub-id pub-id-type="pmid">31452990</pub-id></citation></ref>
<ref id="ref19"><label>19.</label><citation citation-type="other"><person-group person-group-type="author"><name><surname>Orlando</surname><given-names>JI</given-names></name> <name><surname>Seeb&#x00F6;ck</surname><given-names>P</given-names></name> <name><surname>Bogunovi&#x0107;</surname><given-names>H</given-names></name> <name><surname>Klimscha</surname><given-names>S</given-names></name> <name><surname>Grechenig</surname><given-names>C</given-names></name> <name><surname>Waldstein</surname><given-names>S</given-names></name> <etal/></person-group>., <italic>U2-net: a Bayesian U-net model with epistemic uncertainty feedback for photoreceptor layer segmentation in pathological Oct scans</italic>. 2019 IEEE 16th International Symposium on Biomedical Imaging (ISBI 2019) (<year>2019</year>).</citation></ref>
<ref id="ref20"><label>20.</label><citation citation-type="other"><person-group person-group-type="author"><name><surname>Wu</surname><given-names>J</given-names></name> <name><surname>Chen</surname><given-names>J</given-names></name> <name><surname>Xiao</surname><given-names>Z</given-names></name> <name><surname>Geng</surname><given-names>L</given-names></name></person-group>. <italic>Automatic layering of retinal Oct images with dual attention mechanism</italic>. 2021 3rd international conference on intelligent medicine and image processing (<year>2021</year>).</citation></ref>
<ref id="ref21"><label>21.</label><citation citation-type="other"><person-group person-group-type="author"><name><surname>Chen</surname><given-names>L-C</given-names></name> <name><surname>Zhu</surname><given-names>Y</given-names></name> <name><surname>Papandreou</surname><given-names>G</given-names></name> <name><surname>Schroff</surname><given-names>F</given-names></name> <name><surname>Adam</surname><given-names>H</given-names></name></person-group>. <italic>Encoder-decoder with Atrous separable convolution for semantic image segmentation</italic>. Proceedings of the European conference on computer vision (ECCV) (<year>2018</year>).</citation></ref>
<ref id="ref22"><label>22.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname><given-names>R</given-names></name> <name><surname>Du</surname><given-names>L</given-names></name> <name><surname>Xiao</surname><given-names>Q</given-names></name> <name><surname>Liu</surname><given-names>J</given-names></name></person-group>. <article-title>Comparison of backbones for semantic segmentation network</article-title>. <source>J Phys Conf Ser</source>. (<year>2020</year>) <volume>1544</volume>:<fpage>012196</fpage>. doi: <pub-id pub-id-type="doi">10.1088/1742-6596/1544/1/012196</pub-id></citation></ref>
<ref id="ref23"><label>23.</label><citation citation-type="other"><person-group person-group-type="author"><name><surname>Sandler</surname><given-names>M</given-names></name> <name><surname>Howard</surname><given-names>A</given-names></name> <name><surname>Zhu</surname><given-names>M</given-names></name> <name><surname>Zhmoginov</surname><given-names>A</given-names></name> <name><surname>Chen</surname><given-names>L-C</given-names></name></person-group>. <italic>Mobilenetv2: Inverted Residuals and Linear Bottlenecks</italic>. Proceedings of the IEEE conference on computer vision and pattern recognition (<year>2018</year>).</citation></ref>
<ref id="ref24"><label>24.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chen</surname><given-names>L-C</given-names></name> <name><surname>Papandreou</surname><given-names>G</given-names></name> <name><surname>Kokkinos</surname><given-names>I</given-names></name> <name><surname>Murphy</surname><given-names>K</given-names></name> <name><surname>Yuille</surname><given-names>AL</given-names></name></person-group>. <article-title>Deeplab: semantic image segmentation with deep convolutional nets, Atrous convolution, and fully connected Crfs</article-title>. <source>IEEE Trans Pattern Anal Mach Intell</source>. (<year>2017</year>) <volume>40</volume>:<fpage>834</fpage>&#x2013;<lpage>48</lpage>. doi: <pub-id pub-id-type="doi">10.1109/tpami.2017.2699184</pub-id>, PMID: <pub-id pub-id-type="pmid">28463186</pub-id></citation></ref>
<ref id="ref25"><label>25.</label><citation citation-type="other"><person-group person-group-type="author"><name><surname>Wang</surname><given-names>P</given-names></name> <name><surname>Chen</surname><given-names>P</given-names></name> <name><surname>Yuan</surname><given-names>Y</given-names></name> <name><surname>Liu</surname><given-names>D</given-names></name> <name><surname>Huang</surname><given-names>Z</given-names></name> <name><surname>Hou</surname><given-names>X</given-names></name> <etal/></person-group>. <italic>Understanding convolution for semantic segmentation</italic>. 2018 IEEE winter conference on applications of computer vision (WACV) (<year>2018</year>).</citation></ref>
<ref id="ref26"><label>26.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Badrinarayanan</surname><given-names>V</given-names></name> <name><surname>Kendall</surname><given-names>A</given-names></name> <name><surname>Cipolla</surname><given-names>R</given-names></name></person-group>. <article-title>Segnet: a deep convolutional encoder-decoder architecture for image segmentation</article-title>. <source>IEEE Trans Pattern Anal Mach Intell</source>. (<year>2017</year>) <volume>39</volume>:<fpage>2481</fpage>&#x2013;<lpage>95</lpage>. doi: <pub-id pub-id-type="doi">10.1109/tpami.2016.2644615</pub-id>, PMID: <pub-id pub-id-type="pmid">28060704</pub-id></citation></ref>
<ref id="ref27"><label>27.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kermany</surname><given-names>DS</given-names></name> <name><surname>Goldbaum</surname><given-names>M</given-names></name> <name><surname>Cai</surname><given-names>W</given-names></name> <name><surname>Valentim</surname><given-names>CC</given-names></name> <name><surname>Liang</surname><given-names>H</given-names></name> <name><surname>Baxter</surname><given-names>SL</given-names></name> <etal/></person-group>. <article-title>Identifying medical diagnoses and treatable diseases by image-based deep learning</article-title>. <source>Cells</source>. (<year>2018</year>) <volume>172</volume>:<fpage>1122</fpage>&#x2013;<lpage>1131.e9</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.cell.2018.02.010</pub-id>, PMID: <pub-id pub-id-type="pmid">29474911</pub-id></citation></ref>
<ref id="ref28"><label>28.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kim</surname><given-names>EJ</given-names></name> <name><surname>Lin</surname><given-names>WV</given-names></name> <name><surname>Rodriguez</surname><given-names>SM</given-names></name> <name><surname>Chen</surname><given-names>A</given-names></name> <name><surname>Loya</surname><given-names>A</given-names></name> <name><surname>Weng</surname><given-names>CY</given-names></name></person-group>. <article-title>Treatment of diabetic macular edema</article-title>. <source>Curr Diab Rep</source>. (<year>2019</year>) <volume>19</volume>:<fpage>1</fpage>&#x2013;<lpage>10</lpage>. doi: <pub-id pub-id-type="doi">10.1007/s11892-019-1188-4</pub-id></citation></ref>
<ref id="ref29"><label>29.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Noma</surname><given-names>H</given-names></name> <name><surname>Yasuda</surname><given-names>K</given-names></name> <name><surname>Shimura</surname><given-names>M</given-names></name></person-group>. <article-title>Involvement of cytokines in the pathogenesis of diabetic macular edema</article-title>. <source>Int J Mol Sci</source>. (<year>2021</year>) <volume>22</volume>:<fpage>3427</fpage>. doi: <pub-id pub-id-type="doi">10.3390/ijms22073427</pub-id>, PMID: <pub-id pub-id-type="pmid">33810434</pub-id></citation></ref>
<ref id="ref30"><label>30.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Daruich</surname><given-names>A</given-names></name> <name><surname>Matet</surname><given-names>A</given-names></name> <name><surname>Moulin</surname><given-names>A</given-names></name> <name><surname>Kowalczuk</surname><given-names>L</given-names></name> <name><surname>Nicolas</surname><given-names>M</given-names></name> <name><surname>Sellam</surname><given-names>A</given-names></name> <etal/></person-group>. <article-title>Mechanisms of macular edema: beyond the surface</article-title>. <source>Prog Retin Eye Res</source>. (<year>2018</year>) <volume>63</volume>:<fpage>20</fpage>&#x2013;<lpage>68</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.preteyeres.2017.10.006</pub-id>, PMID: <pub-id pub-id-type="pmid">29126927</pub-id></citation></ref>
<ref id="ref31"><label>31.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Pekala</surname><given-names>M</given-names></name> <name><surname>Joshi</surname><given-names>N</given-names></name> <name><surname>Liu</surname><given-names>TA</given-names></name> <name><surname>Bressler</surname><given-names>NM</given-names></name> <name><surname>DeBuc</surname><given-names>DC</given-names></name> <name><surname>Burlina</surname><given-names>P</given-names></name></person-group>. <article-title>Deep learning based retinal Oct segmentation</article-title>. <source>Comput Biol Med</source>. (<year>2019</year>) <volume>114</volume>:<fpage>103445</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.compbiomed.2019.103445</pub-id></citation></ref>
<ref id="ref32"><label>32.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>de Moura</surname><given-names>J</given-names></name> <name><surname>Samagaio</surname><given-names>G</given-names></name> <name><surname>Novo</surname><given-names>J</given-names></name> <name><surname>Almuina</surname><given-names>P</given-names></name> <name><surname>Fern&#x00E1;ndez</surname><given-names>MI</given-names></name> <name><surname>Ortega</surname><given-names>M</given-names></name></person-group>. <article-title>Joint diabetic macular edema segmentation and characterization in Oct images</article-title>. <source>J Digit Imaging</source>. (<year>2020</year>) <volume>33</volume>:<fpage>1335</fpage>&#x2013;<lpage>51</lpage>. doi: <pub-id pub-id-type="doi">10.1007/s10278-020-00360-y</pub-id>, PMID: <pub-id pub-id-type="pmid">32562127</pub-id></citation></ref>
<ref id="ref33"><label>33.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Alex</surname><given-names>V</given-names></name> <name><surname>Motevasseli</surname><given-names>T</given-names></name> <name><surname>Freeman</surname><given-names>WR</given-names></name> <name><surname>Jayamon</surname><given-names>JA</given-names></name> <name><surname>Bartsch</surname><given-names>D-UG</given-names></name> <name><surname>Borooah</surname><given-names>S</given-names></name></person-group>. <article-title>Assessing the validity of a cross-platform retinal image segmentation tool in Normal and diseased retina</article-title>. <source>Sci Rep</source>. (<year>2021</year>) <volume>11</volume>:<fpage>21784</fpage>. doi: <pub-id pub-id-type="doi">10.21203/rs.3.rs-396609/v1</pub-id></citation></ref>
<ref id="ref34"><label>34.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wei</surname><given-names>H</given-names></name> <name><surname>Peng</surname><given-names>P</given-names></name></person-group>. <article-title>The segmentation of retinal layer and fluid in Sd-Oct images using Mutex Dice loss based fully convolutional networks</article-title>. <source>IEEE Access.</source> (<year>2020</year>) <volume>8</volume>:<fpage>60929</fpage>&#x2013;<lpage>39</lpage>. doi: <pub-id pub-id-type="doi">10.1109/ACCESS.2020.2983818</pub-id></citation></ref>
<ref id="ref35"><label>35.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hervella</surname><given-names>&#x00C1;S</given-names></name> <name><surname>Rouco</surname><given-names>J</given-names></name> <name><surname>Novo</surname><given-names>J</given-names></name> <name><surname>Ortega</surname><given-names>M</given-names></name></person-group>. <article-title>Retinal microaneurysms detection using adversarial pre-training with unlabeled multimodal images</article-title>. <source>Inf Fusion</source>. (<year>2022</year>) <volume>79</volume>:<fpage>146</fpage>&#x2013;<lpage>61</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.inffus.2021.10.003</pub-id></citation></ref>
</ref-list>
</back>
</article>