<?xml version="1.0" encoding="UTF-8" standalone="no"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Med.</journal-id>
<journal-title>Frontiers in Medicine</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Med.</abbrev-journal-title>
<issn pub-type="epub">2296-858X</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fmed.2021.782664</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Medicine</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>U-shaped GAN for Semi-Supervised Learning and Unsupervised Domain Adaptation in High Resolution Chest Radiograph Segmentation</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name><surname>Wang</surname> <given-names>Hongyu</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/1484456/overview"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Gu</surname> <given-names>Hong</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/703448/overview"/>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name><surname>Qin</surname> <given-names>Pan</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>&#x0002A;</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/635358/overview"/>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name><surname>Wang</surname> <given-names>Jia</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<xref ref-type="corresp" rid="c002"><sup>&#x0002A;</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/703145/overview"/>
</contrib>
</contrib-group>
<aff id="aff1"><sup>1</sup><institution>Faculty of Electronic Information and Electrical Engineering, Dalian University of Technology</institution>, <addr-line>Dalian</addr-line>, <country>China</country></aff>
<aff id="aff2"><sup>2</sup><institution>Department of Surgery, The Second Hospital of Dalian Medical University</institution>, <addr-line>Dalian</addr-line>, <country>China</country></aff>
<author-notes>
<fn fn-type="edited-by"><p>Edited by: Kuanquan Wang, Harbin Institute of Technology, China</p></fn>
<fn fn-type="edited-by"><p>Reviewed by: Nurlan Dauletbayev, McGill University, Canada; Zhixin Li, Guangxi Normal University, China</p></fn>
<corresp id="c001">&#x0002A;Correspondence: Pan Qin <email>qp112cn&#x00040;dlut.edu.cn</email></corresp>
<corresp id="c002">Jia Wang <email>wangjia77&#x00040;hotmail.com</email></corresp>
<fn fn-type="other" id="fn001"><p>This article was submitted to Precision Medicine, a section of the journal Frontiers in Medicine</p></fn></author-notes>
<pub-date pub-type="epub">
<day>13</day>
<month>01</month>
<year>2022</year>
</pub-date>
<pub-date pub-type="collection">
<year>2021</year>
</pub-date>
<volume>8</volume>
<elocation-id>782664</elocation-id>
<history>
<date date-type="received">
<day>24</day>
<month>09</month>
<year>2021</year>
</date>
<date date-type="accepted">
<day>14</day>
<month>12</month>
<year>2021</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x000A9; 2022 Wang, Gu, Qin and Wang.</copyright-statement>
<copyright-year>2022</copyright-year>
<copyright-holder>Wang, Gu, Qin and Wang</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p></license> </permissions>
<abstract><p>Deep learning has achieved considerable success in medical image segmentation. However, applying deep learning in clinical environments often involves two problems: (1) scarcity of annotated data as data annotation is time-consuming and (2) varying attributes of different datasets due to domain shift. To address these problems, we propose an improved generative adversarial network (GAN) segmentation model, called U-shaped GAN, for limited-annotated chest radiograph datasets. The semi-supervised learning approach and unsupervised domain adaptation (UDA) approach are modeled into a unified framework for effective segmentation. We improve GAN by replacing the traditional discriminator with a U-shaped net, which predicts each pixel a label. The proposed U-shaped net is designed with high resolution radiographs (1,024 &#x000D7; 1,024) for effective segmentation while taking computational burden into account. The pointwise convolution is applied to U-shaped GAN for dimensionality reduction, which decreases the number of feature maps while retaining their salient features. Moreover, we design the U-shaped net with a pretrained ResNet-50 as an encoder to reduce the computational burden of training the encoder from scratch. A semi-supervised learning approach is proposed learning from limited annotated data while exploiting additional unannotated data with a pixel-level loss. U-shaped GAN is extended to UDA by taking the source and target domain data as the annotated data and the unannotated data in the semi-supervised learning approach, respectively. Compared to the previous models dealing with the aforementioned problems separately, U-shaped GAN is compatible with varying data distributions of multiple medical centers, with efficient training and optimizing performance. U-shaped GAN can be generalized to chest radiograph segmentation for clinical deployment. We evaluate U-shaped GAN with two chest radiograph datasets. U-shaped GAN is shown to significantly outperform the state-of-the-art models.</p></abstract>
<kwd-group>
<kwd>semi-supervised learning</kwd>
<kwd>unsupervised domain adaptation</kwd>
<kwd>generative adversarial network</kwd>
<kwd>medical image segmentation</kwd>
<kwd>chest radiograph</kwd>
</kwd-group>
<contract-num rid="cn001">61633006</contract-num>
<contract-num rid="cn001">81872247</contract-num>
<contract-num rid="cn002">DUT21YG118</contract-num>
<contract-sponsor id="cn001">National Natural Science Foundation of China<named-content content-type="fundref-id">10.13039/501100001809</named-content></contract-sponsor>
<contract-sponsor id="cn002">Fundamental Research Funds for the Central Universities<named-content content-type="fundref-id">10.13039/501100012226</named-content></contract-sponsor>
<counts>
<fig-count count="7"/>
<table-count count="6"/>
<equation-count count="10"/>
<ref-count count="39"/>
<page-count count="12"/>
<word-count count="7898"/>
</counts>
</article-meta>
</front>
<body>
<sec sec-type="intro" id="s1">
<title>1. Introduction</title>
<p>Recently, deep learning models have gained increasing popularity in medical segmentation. However, deep learning models with supervision require substantial pixel-level annotated data to achieve sufficient accuracy and prevent over-fitting (<xref ref-type="bibr" rid="B1">1</xref>&#x02013;<xref ref-type="bibr" rid="B4">4</xref>). Pixel-level annotation is expensive, especially with medical images, because it is time-consuming and requires highly skilled experts (<xref ref-type="bibr" rid="B3">3</xref>, <xref ref-type="bibr" rid="B5">5</xref>). Therefore, medical image datasets are usually small, which cannot meet the requirement of deep learning, due to a lack of annotations (<xref ref-type="bibr" rid="B6">6</xref>, <xref ref-type="bibr" rid="B7">7</xref>). Even if a model is well-trained on a certain medical dataset, its accuracy decreases when it is applied to unseen domains (<xref ref-type="bibr" rid="B8">8</xref>, <xref ref-type="bibr" rid="B9">9</xref>). The deep learning models suffer an accuracy drop between two domains due to domain shift (<xref ref-type="bibr" rid="B8">8</xref>). These problems limit the application of deep learning models in clinical environments.</p>
<p>An alternative to supervised learning is semi-supervised learning, which involves using unannotated data to identify specific hidden features of the dataset to facilitate prediction. Semi-supervised deep learning using generative adversarial networks (GANs) (<xref ref-type="bibr" rid="B10">10</xref>) has been highly successful (<xref ref-type="bibr" rid="B1">1</xref>, <xref ref-type="bibr" rid="B2">2</xref>), especially with nonmedical images (<xref ref-type="bibr" rid="B3">3</xref>). A semi-supervised GAN model was proposed to distinguish between predicted probability maps and the ground truth (<xref ref-type="bibr" rid="B2">2</xref>). In Souly et al. (<xref ref-type="bibr" rid="B1">1</xref>), a GAN was used to generate fake images close to real images as learned by the segmentation network. In medical segmentation, a few semi-supervised models have been developed using the model in Hung et al. (<xref ref-type="bibr" rid="B2">2</xref>) to improve the prediction accuracy with specially designed loss functions for particular image types, such as computed tomography of the liver (<xref ref-type="bibr" rid="B3">3</xref>), retinal optical coherence tomography (<xref ref-type="bibr" rid="B11">11</xref>), and pediatric MRI (<xref ref-type="bibr" rid="B12">12</xref>).</p>
<p>Domain adaptation (DA) suffers domain shift by transferring knowledge from the source domain to the target domain (<xref ref-type="bibr" rid="B13">13</xref>&#x02013;<xref ref-type="bibr" rid="B16">16</xref>). A popular solution is transfer learning, which fine-tunes parts of a pre-trained model with annotated target domain data (<xref ref-type="bibr" rid="B13">13</xref>). However, transfer learning approaches rely on additional annotated data on the target domain, which is expensive or sometimes impractical. Instead, unsupervised domain adaptation (UDA) is more appealing to generalize models in clinical practice. UDA using GAN is becoming increasingly popular in the medical field (<xref ref-type="bibr" rid="B14">14</xref>&#x02013;<xref ref-type="bibr" rid="B16">16</xref>). GAN-based UDA models use generators to transform the target data to the source domain, discriminate the source data from the target data, and improve segmentation accuracy with a specific training method (<xref ref-type="bibr" rid="B16">16</xref>), net structure (<xref ref-type="bibr" rid="B14">14</xref>), or training loss (<xref ref-type="bibr" rid="B15">15</xref>).</p>
<p>Deep learning algorithms require large amounts of data, which cannot be collected from a single medical center. Therefore, data from multiple collection centers, comprising large medical centers and small clinics, are required (<xref ref-type="bibr" rid="B17">17</xref>&#x02013;<xref ref-type="bibr" rid="B19">19</xref>). The large medical centers provide partly annotated datasets for semi-supervised learning, while the clinics provide unannotated data. The annotated and unannotated data may come from either the same or different domains in the dataset collected from multiple centers. A single model that can deal with the semi-supervised and UDA approach at the same time is urgently needed.</p>
<p>To tackle the aforementioned problems, we propose an improved GAN model, called U-shaped GAN, for medical image segmentation. U-shaped GAN is improved by replacing the traditional discriminator with a U-shaped net to assign each pixel a label. Training the segmentation model with images of high resolution is effective; however, it increases the computational burden (<xref ref-type="bibr" rid="B20">20</xref>, <xref ref-type="bibr" rid="B21">21</xref>). U-shaped GAN is designed with high resolution radiographs for effective segmentation while considering computational burden. The pointwise convolution is applied to U-shaped GAN for dimensionality reduction, which decreases the number of feature maps while retaining their salient features. Moreover, the U-shaped net takes a pretrained ResNet-50 as an encoder to reduce the computational burden of training from scratch. A pixel-level semi-supervised loss is proposed to leverage the unannotated data to assist the annotated data for semi-supervised learning. U-shaped GAN is extended to UDA with minimal modification. The semi-supervised learning approach and UDA approach are merged into a single model to handle datasets from multiple medical centers conveniently and efficiently. We evaluate U-shaped GAN on lung segmentation for radiographs.</p>
<p>To conclude the introduction, we outline the major contributions of this work as follows:</p>
<list list-type="simple">
<list-item><p>(1) U-shaped GAN is proposed for high resolution medical image segmentation while taking computational burden into account.</p></list-item>
<list-item><p>(2) A semi-supervised learning approach is proposed to overcome the lack of annotated data. We employ a pixel-level semi-supervised loss that leverages the unannotated data to assist the annotated data for segmentation.</p></list-item>
<list-item><p>(3) U-shaped GAN is extended to UDA with minimal modification to transfer knowledge among different domains without additional annotated data on the target domain.</p></list-item>
<list-item><p>(4) In our framework, the semi-supervised learning approach and UDA approach are merged into a single model to handle datasets from multiple medical centers conveniently and efficiently.</p></list-item>
</list>
</sec>
<sec sec-type="methods" id="s2">
<title>2. Methods</title>
<sec>
<title>2.1. Background</title>
<p>In recent years, GAN has garnered considerable attention because of its superior performance in terms of generating images (<xref ref-type="bibr" rid="B2">2</xref>). GAN consists of a generator network <italic>G</italic> and discriminator net <italic>D</italic>. <italic>G</italic> generates fake images close to real data from a noise distribution deceiving the discriminator, while <italic>D</italic> distinguishes the real images from fake ones. <italic>G</italic> and <italic>D</italic> can be considered as two competitors in a min-max game with the following formulation:</p>
<disp-formula id="E1"><label>(1)</label><mml:math id="M1"><mml:mtable columnalign="right"><mml:mtr><mml:mtd><mml:mtable columnalign="right"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mo class="qopname">min</mml:mo></mml:mrow><mml:mrow><mml:mi>G</mml:mi></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mo class="qopname">max</mml:mo></mml:mrow><mml:mrow><mml:mi>D</mml:mi></mml:mrow></mml:msub><mml:mi>V</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>D</mml:mi><mml:mo>,</mml:mo><mml:mi>G</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:mi>x</mml:mi><mml:mo>&#x0007E;</mml:mo><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mi>d</mml:mi><mml:mi>a</mml:mi><mml:mi>t</mml:mi><mml:mi>a</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">[</mml:mo><mml:mrow><mml:mo class="qopname">log</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>D</mml:mi><mml:msub><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mi>r</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>l</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">]</mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mtext>&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;</mml:mtext><mml:mo>&#x0002B;</mml:mo><mml:msub><mml:mrow><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:mi>z</mml:mi><mml:mo>&#x0007E;</mml:mo><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mi>n</mml:mi><mml:mi>o</mml:mi><mml:mi>i</mml:mi><mml:mi>s</mml:mi><mml:mi>e</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>z</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">[</mml:mo><mml:mrow><mml:mo class="qopname">log</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mn>1</mml:mn><mml:mo>-</mml:mo><mml:mi>D</mml:mi><mml:msub><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>G</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>z</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mi>r</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>l</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">]</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where <italic>E</italic> is the expectation of a random variable, <italic>p</italic><sub><italic>data</italic></sub>(<italic>x</italic>) is the real data distribution, and <italic>p</italic><sub><italic>noise</italic></sub>(<italic>z</italic>) is a noise distribution. <italic>D</italic>(&#x0002A;)<sub><italic>real</italic></sub> stands for the possibility that the sample is from the real data. <italic>G</italic> transforms the noise variable <italic>z</italic> from the distribution <italic>p</italic><sub><italic>noise</italic></sub>(<italic>z</italic>) into <italic>G</italic>(<italic>z</italic>). The min-max game provides a useful feature representation for auxiliary supervised discrimination tasks (<xref ref-type="bibr" rid="B22">22</xref>).</p>
</sec>
<sec>
<title>2.2. Proposed Model</title>
<p>The goal of this study is to develop a unified framework for semi-supervised learning and UDA. Analyzing the influence factors in semi-supervised learning and UDA on chest radiographs, we propose a similar solution for the semi-supervised learning approach and UDA approach. A pixel-level semi-supervised loss is proposed to leverage the unannotated data to assist the annotated data for segmentation.</p>
<sec>
<title>2.2.1. Semi-Supervised Learning</title>
<p>We propose a U-shaped GAN for semi-supervised lung segmentation from chest radiograph datasets. U-shaped GAN is based on the following hypothesis: the features in an ideal representation correspond to the underlying causes of the data. If label <italic>y</italic> is among the salient causes of data <italic>x</italic>, a suitable representation for the probability distribution <italic>p</italic>(<italic>x</italic>) may also be a suitable representation for computing the distribution of conditional probability <italic>p</italic>(<italic>y</italic>|<italic>x</italic>) (<xref ref-type="bibr" rid="B23">23</xref>). The marginal probability <italic>p</italic>(<italic>x</italic>) is related to the conditional probability <italic>p</italic>(<italic>y</italic>|<italic>x</italic>) through the Bayes rule:</p>
<disp-formula id="E3"><label>(2)</label><mml:math id="M3"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mi>p</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>y</mml:mi><mml:mo stretchy="false">|</mml:mo><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi>p</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi><mml:mo stretchy="false">|</mml:mo><mml:mi>y</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mi>p</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mi>p</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:mfrac></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>

<p>Under this hypothesis, we use the unannotated and annotated data to find a representation for the radiographs. A particular semi-supervised loss, which can be divided into a supervised loss and an unsupervised loss, is proposed. The supervised loss using the annotated data is employed for segmentation prediction, and the unsupervised loss using the unannotated data is utilized for a better representation of the whole dataset, as shown in <xref ref-type="fig" rid="F1">Figure 1A</xref>. The unannotated data generalize the model as a regularizer. In U-shaped GAN, we employ a generator to generate realistic data segmented by a multiclass classifier (our discriminator) from the noise input, which in addition to classifying the pixels into lungs, determines whether a given pixel belongs to the real or generated data. The generator converges the realistic data to the real distribution <italic>p</italic>(<italic>x</italic>) of the partly annotated real data. This enables the discriminator to learn better features to represent the radiographs and to filter irrelevant individual features. Moreover, we employ the annotated data to find the relations among those features and the segmentation task. We modify a GAN by replacing the original discriminator with a U-shaped net (<xref ref-type="bibr" rid="B24">24</xref>) for assigning a label to each pixel. Training convolutional neural networks (CNNs) with a high resolution is effective for lung segmentation (<xref ref-type="bibr" rid="B20">20</xref>, <xref ref-type="bibr" rid="B21">21</xref>). U-shaped GAN is designed for high-resolution chest radiographs. Unlike the segmentation GAN (<xref ref-type="bibr" rid="B1">1</xref>), we use the semantic classes and the background as an additional class for segmentation as the background contains several unannotated organs that reflect the imaging condition.</p>
<fig id="F1" position="float">
<label>Figure 1</label>
<caption><p><bold>(A)</bold> Overview of the proposed semi-supervised segmentation approach. We utilize an unsupervised loss to obtain a better representation for the radiographs and a supervised loss to relate those features to the segmentation task. <bold>(B)</bold> Overview of the proposed unsupervised domain adaptation (UDA) segmentation approach. We leverage U-shaped GAN to find a better representation with basic features supporting radiograph imaging with an unsupervised loss. The distributions of the source and target data are aligned in this representation.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fmed-08-782664-g0001.tif"/>
</fig>
<p>We improve GAN by replacing the traditional discriminator with a U-shaped net, which, instead of predicting each image a label, assigns to each pixel a label. The proposed end-to-end deep learning model is illustrated in <xref ref-type="fig" rid="F2">Figure 2</xref>. The discriminator acts as a segmentation network to assign one of the following labels to each pixel: lung class, background, or fake data. The annotated data is used to train the discriminator <italic>D</italic> to minimize the loss function <italic>L</italic><sub><italic>l</italic></sub>:</p>
<disp-formula id="E4"><label>(3)</label><mml:math id="M4"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mtable style="text-align:axis;" equalrows="false" equalcolumns="false" class="array"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:mi>x</mml:mi><mml:mo>&#x0007E;</mml:mo><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mi>d</mml:mi><mml:mi>a</mml:mi><mml:mi>t</mml:mi><mml:msub><mml:mrow><mml:mi>a</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi><mml:mo>,</mml:mo><mml:mi>y</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">[</mml:mo><mml:mrow><mml:mo class="qopname">log</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>D</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>y</mml:mi><mml:mo stretchy="false">|</mml:mo><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">]</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where <italic>p</italic><sub><italic>data</italic><sub><italic>l</italic></sub></sub>(<italic>x, y</italic>) is the joint distribution of the pixel-level labels <italic>y</italic> and pixel values <italic>x</italic> of the annotated data; the discriminator <italic>D</italic> predicts the possibility <italic>D</italic>(<italic>y</italic>|<italic>x</italic>) of pixel <italic>x</italic> belonging to label <italic>y</italic>.</p>
<fig id="F2" position="float">
<label>Figure 2</label>
<caption><p>Schematic description of U-shaped GAN. The generator transforms noise into fake images. The fake and real data are used to train the discriminator for pixel-level predictions as lungs <italic>C</italic><sub><italic>lung</italic></sub>, background <italic>C</italic><sub><italic>back</italic></sub>, and fake <italic>C</italic><sub><italic>fake</italic></sub>.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fmed-08-782664-g0002.tif"/>
</fig>
<p>In semi-supervised learning, where the labels are partly available among the training images, it is convenient to leverage the unannotated data for estimating a representation with useful features for segmentation. The true labels <italic>y</italic> of the pixels of the unannotated data are set as real data. The loss function for training the discriminator <italic>D</italic> with the unannotated data is defined as follows:</p>
<disp-formula id="E5"><label>(4)</label><mml:math id="M5"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mtable columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mrow><mml:mi>u</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:mi>x</mml:mi><mml:mo>&#x0007E;</mml:mo><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mi>d</mml:mi><mml:mi>a</mml:mi><mml:mi>t</mml:mi><mml:msub><mml:mrow><mml:mi>a</mml:mi></mml:mrow><mml:mrow><mml:mi>u</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">[</mml:mo><mml:mrow><mml:mo class="qopname">log</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>D</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>y</mml:mi><mml:mo stretchy="false">|</mml:mo><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">]</mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mtext>&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;</mml:mtext><mml:mo>=</mml:mo><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:mi>x</mml:mi><mml:mo>&#x0007E;</mml:mo><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mi>d</mml:mi><mml:mi>a</mml:mi><mml:mi>t</mml:mi><mml:msub><mml:mrow><mml:mi>a</mml:mi></mml:mrow><mml:mrow><mml:mi>u</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">[</mml:mo><mml:mrow><mml:mo class="qopname">log</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mn>1</mml:mn><mml:mo>-</mml:mo><mml:mi>D</mml:mi><mml:msub><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mi>f</mml:mi><mml:mi>a</mml:mi><mml:mi>k</mml:mi><mml:mi>e</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">]</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where <italic>p</italic><sub><italic>data</italic><sub><italic>u</italic></sub></sub>(<italic>x</italic>) is the distribution of pixels of chest radiographs without annotation; <italic>D</italic>(<italic>x</italic>)<sub><italic>fake</italic></sub> is the possibility of the pixel belonging to the fake data. In U-shaped GAN, there is no output designed with the label &#x0201C;real data.&#x0201D; We used the 1 &#x02212; <italic>D</italic>(<italic>x</italic>)<sub><italic>fake</italic></sub> instead.</p>
<p>The generator <italic>G</italic> maps a random noise <italic>z</italic> to a sample <italic>G</italic>(<italic>z</italic>) that is close to chest radiographs, while the discriminator <italic>D</italic> is trained to label the generated sample <italic>G</italic>(<italic>z</italic>) as fake. The true labels <italic>y</italic> of the pixels of the generated data are set as fake data. The loss function for the discriminator <italic>D</italic> with the generated data is given as</p>
<disp-formula id="E6"><label>(5)</label><mml:math id="M6"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mtable style="text-align:axis;" equalrows="false" equalcolumns="false" class="array"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:mi>z</mml:mi><mml:mo>&#x0007E;</mml:mo><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mi>n</mml:mi><mml:mi>o</mml:mi><mml:mi>i</mml:mi><mml:mi>s</mml:mi><mml:mi>e</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>z</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">[</mml:mo><mml:mrow><mml:mo class="qopname">log</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>D</mml:mi><mml:msub><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>G</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>z</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mi>f</mml:mi><mml:mi>a</mml:mi><mml:mi>k</mml:mi><mml:mi>e</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">]</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where <italic>D</italic>(<italic>G</italic>(<italic>z</italic>))<sub><italic>fake</italic></sub> is the possibility of the generated pixel belonging to the fake data.</p>
<p>We minimize the pixel-level discriminator loss <italic>L</italic><sub><italic>D</italic></sub> with respect to three types of input data, as follows:</p>
<disp-formula id="E7"><label>(6)</label><mml:math id="M7"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mtable style="text-align:axis;" equalrows="false" equalcolumns="false" class="array"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mrow><mml:mi>D</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:mi>x</mml:mi><mml:mo>&#x0007E;</mml:mo><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mi>d</mml:mi><mml:mi>a</mml:mi><mml:mi>t</mml:mi><mml:msub><mml:mrow><mml:mi>a</mml:mi></mml:mrow><mml:mrow><mml:mi>u</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">[</mml:mo><mml:mrow><mml:mo class="qopname">log</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mn>1</mml:mn><mml:mo>-</mml:mo><mml:mi>D</mml:mi><mml:msub><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mi>f</mml:mi><mml:mi>a</mml:mi><mml:mi>k</mml:mi><mml:mi>e</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">]</mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mtext>&#x000A0;&#x000A0;&#x000A0;&#x000A0;</mml:mtext><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:mi>z</mml:mi><mml:mo>&#x0007E;</mml:mo><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mi>n</mml:mi><mml:mi>o</mml:mi><mml:mi>i</mml:mi><mml:mi>s</mml:mi><mml:mi>e</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>z</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">[</mml:mo><mml:mrow><mml:mo class="qopname">log</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>D</mml:mi><mml:msub><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>G</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>z</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mi>f</mml:mi><mml:mi>a</mml:mi><mml:mi>k</mml:mi><mml:mi>e</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">]</mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mtext>&#x000A0;&#x000A0;&#x000A0;&#x000A0;</mml:mtext><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:mi>x</mml:mi><mml:mo>&#x0007E;</mml:mo><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mi>d</mml:mi><mml:mi>a</mml:mi><mml:mi>t</mml:mi><mml:msub><mml:mrow><mml:mi>a</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi><mml:mo>,</mml:mo><mml:mi>y</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">[</mml:mo><mml:mrow><mml:mo class="qopname">log</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>D</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>y</mml:mi><mml:mo stretchy="false">|</mml:mo><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">]</mml:mo></mml:mrow><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>The first and second terms of <italic>L</italic><sub><italic>D</italic></sub> are devised for unannotated data as an unsupervised loss to increase the ability of the discriminator to identify the real radiographs from fake ones and to find salient features in the chest radiograph. The third term is devised for annotated data as a supervised loss training <italic>D</italic> to find correct relations among these features and the segmentation task. We use a soft maximum over the outputs. <italic>D</italic>(<italic>y</italic>|<italic>x</italic>) for the annotated data is a component of the 1&#x02212;<italic>D</italic>(<italic>x</italic>)<sub><italic>fake</italic></sub>. Increasing the probability <italic>D</italic>(<italic>y</italic>|<italic>x</italic>) will decrease the probability of <italic>D</italic>(<italic>x</italic>)<sub><italic>fake</italic></sub>. The third term has the same effect as the first term and acts as an unsupervised loss to increase the ability of the discriminating and salient features finding. We minimize the generator loss <italic>L</italic><sub><italic>G</italic></sub> to train the generator <italic>G</italic> as follows:</p>
<disp-formula id="E8"><label>(7)</label><mml:math id="M8"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mtable style="text-align:axis;" equalrows="false" equalcolumns="false" class="array"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:mi>z</mml:mi><mml:mo>&#x0007E;</mml:mo><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mi>n</mml:mi><mml:mi>o</mml:mi><mml:mi>i</mml:mi><mml:mi>s</mml:mi><mml:mi>e</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>z</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">[</mml:mo><mml:mrow><mml:mo class="qopname">log</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>D</mml:mi><mml:msub><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>G</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>z</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mi>f</mml:mi><mml:mi>a</mml:mi><mml:mi>k</mml:mi><mml:mi>e</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">]</mml:mo></mml:mrow><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>Because all of the annotated and unannotated data contribute to the discriminating ability of U-shaped GAN, <italic>G</italic> generates a distribution <italic>p</italic><sub><italic>g</italic></sub>(<italic>G</italic>(<italic>z</italic>)) converging to the real distribution <italic>p</italic>(<italic>x</italic>) of the whole dataset consisting of annotated and unannotated radiographs.</p>
</sec>
<sec>
<title>2.2.2. UDA Approach</title>
<p>The chest radiographs from various sources acquired by the same imaging modality differ in three aspects: image quality, image appearance, and spatial configuration (<xref ref-type="bibr" rid="B8">8</xref>). The features relevant to the basic imaging causes among the radiographs are similar. Based on this property, we aim to develop a UDA approach to find these similar features, filter features of individual domains, and align the source and target domains in a representation, which is similar to our semi-supervised learning approach. U-shaped GAN is extended to UDA with minimal modification and uses nearly the same training process in the semi-supervised learning and UDA approaches. We use a U-shaped GAN to search for the aforementioned features with the annotated source domain and unannotated target domain, as shown in <xref ref-type="fig" rid="F1">Figure 1B</xref>. The target data serves as the unannotated data in the semi-supervised learning approach to generalize the model trained on the source dataset. The source data serves as the annotated data to ensure the accuracy of the segmentation task.</p>
<p>U-shaped GAN is extended to UDA, where only the source data are annotated, and the labels of target data are not available. The source and target domain data are used as the annotated and unannotated data in <italic>L</italic><sub><italic>D</italic></sub>, respectively, to train the discriminator. The generator loss function is the same as <italic>L</italic><sub><italic>G</italic></sub> in the semi-supervised approach. The generator <italic>G</italic> produces the fake data matching the aligned representation of the source and target data in the min-max game of <italic>G</italic> and <italic>D</italic>. The fake data are close to the real radiographs in both domains, and the discriminator learns better features related to the basic imaging causes of radiographs. U-shaped GAN then finds a suitable representation for the radiographs of the target and source data. The distributions of source and target data are aligned in this representation. <italic>L</italic><sub><italic>D</italic></sub> and <italic>L</italic><sub><italic>G</italic></sub> encourage domain-invariant detectors to emerge in U-shaped GAN. The annotated source data also guarantees the correct segmentation prediction.</p>
</sec>
</sec>
<sec>
<title>2.3. Network Architecture</title>
<p>U-shaped GAN is proposed to label each input image pixel <italic>y</italic> as a lung, background, or fake pixel. The U-shaped net is incorporated into the structure of GAN, serving as the discriminator, to label each pixel. A schematic description of U-shaped GAN is shown in <xref ref-type="fig" rid="F2">Figure 2</xref>. Training CNNs at a high resolution is effective for lung segmentation predictions; however, it increases the computational burden (<xref ref-type="bibr" rid="B20">20</xref>, <xref ref-type="bibr" rid="B21">21</xref>). Therefore, we use a pointwise convolution layer followed by two 4 &#x000D7; 4 fractionally strided convolution layers to form a new generator block (<xref ref-type="fig" rid="F3">Figure 3</xref>), instead of the fractionally strided convolution layers used in deep convolutional GANs (<xref ref-type="bibr" rid="B25">25</xref>), to achieve a high resolution of 1, 024 &#x000D7; 1, 024. The pointwise convolution layer reduces feature dimensions while retaining the salient features and leaves the 4 &#x000D7; 4 layers with fewer parameters. This approach significantly reduces the computational complexity. The model parameters decrease from 1.34 &#x000D7; 10<sup>8</sup> to 1.35 &#x000D7; 10<sup>7</sup> and the floating-point operations (FLOPs) decrease from 5.84 &#x000D7; 10<sup>8</sup> to 1.53 &#x000D7; 10<sup>8</sup>. In addition, we improve U-shaped GAN with a modified U-shaped network as the discriminator, as shown in <xref ref-type="fig" rid="F4">Figure 4</xref>. The U-shaped net consists of a feature encoder and decoder modules (<xref ref-type="bibr" rid="B24">24</xref>). We replace the encoder with a pretrained ResNet-50 (<xref ref-type="bibr" rid="B26">26</xref>), which further reduces the computational burden of training from scratch. Moreover, ResNet-50 solves the degradation problem by adding identity connections to the convolution network (<xref ref-type="bibr" rid="B26">26</xref>). The feature decoder module restores the high-level semantic features extracted from the feature encoder module. The modified decoder module comprises four building blocks, as shown in <xref ref-type="fig" rid="F4">Figure 4</xref>. The fundamental building block mainly comprises a 3 &#x000D7; 3 convolution layer followed by a 4 &#x000D7; 4 fractionally strided convolution layer. A pointwise convolution layer is used to connect them to reduce relevant parameters. Skip connections take information directly from the encoder to the decoder layers and recover the information loss due to consecutive pooling and striding convolutional operations (<xref ref-type="bibr" rid="B24">24</xref>). We use instance normalization (<xref ref-type="bibr" rid="B27">27</xref>) followed by LeakyReLU activation functions (<xref ref-type="bibr" rid="B28">28</xref>) between each layer.</p>
<fig id="F3" position="float">
<label>Figure 3</label>
<caption><p>The overall structure of the proposed generator. The generator block consists of a pointwise convolution layer and two 4 &#x000D7; 4 fractionally strided convolution layers. Conv3&#x000D7;3 denotes the 3 &#x000D7; 3 convolution layer. Conv1&#x000D7;1 represents the 1 &#x000D7; 1 convolution layer.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fmed-08-782664-g0003.tif"/>
</fig>
<fig id="F4" position="float">
<label>Figure 4</label>
<caption><p>The overall structure of the discriminator. ResNet block stands for the block used in ResNet-50. Conv3&#x000D7;3 indicates the 3 &#x000D7; 3 convolution layer. Conv1&#x000D7;1 indicates the 1 &#x000D7; 1 convolution layer. Decoder block mainly comprises a 3 &#x000D7; 3 convolution layer and 4 &#x000D7; 4 fractionally strided convolution layer. A pointwise convolution layer is used to connect them.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fmed-08-782664-g0004.tif"/>
</fig>
</sec>
</sec>
<sec id="s3">
<title>3. Experiments</title>
<sec>
<title>3.1. Datasets</title>
<p>In our experiments, we utilize the Japanese Society of Radiological Technology (JSRT) (<xref ref-type="bibr" rid="B29">29</xref>) and Montgomery County (MC) datasets (<xref ref-type="bibr" rid="B30">30</xref>, <xref ref-type="bibr" rid="B31">31</xref>). The JSRT dataset contains 247 posterior-anterior (PA) chest radiographs, of which 154 contain lung nodules and 93 have no nodules (<xref ref-type="bibr" rid="B29">29</xref>). The ground truth lung masks can be obtained in the Segmentation in Chest Radiographs dataset (<xref ref-type="bibr" rid="B32">32</xref>). The MC dataset contains PA chest radiographs collected from the National Library of Medicine, National Institutes of Health, Bethesda, MD, USA. It consists of 80 normal and 58 abnormal cases with manifestations of tuberculosis (<xref ref-type="bibr" rid="B30">30</xref>, <xref ref-type="bibr" rid="B31">31</xref>). The ground truth lung masks are also contained in the MC dataset.</p>
</sec>
<sec>
<title>3.2. Metrics</title>
<p>Several algorithms with different evaluation metrics are available in the literature. We used two commonly used methods, the Jaccard index and Dice score metrics, to compare U-shaped GAN with the state-of-the-art models.</p>
<p>(1) The Jaccard index statistic is used for gauging the similarity and diversity of sample sets. It shows the agreement between the ground truth B and the predicted set of pixels A and is given as:</p>
<disp-formula id="E9"><label>(8)</label><mml:math id="M9"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mi>J</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>A</mml:mi><mml:mo>,</mml:mo><mml:mi>B</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mo stretchy="false">|</mml:mo><mml:mi>A</mml:mi><mml:mo>&#x02229;</mml:mo><mml:mi>B</mml:mi><mml:mo stretchy="false">|</mml:mo></mml:mrow><mml:mrow><mml:mo stretchy="false">|</mml:mo><mml:mi>A</mml:mi><mml:mo>&#x0222A;</mml:mo><mml:mi>B</mml:mi><mml:mo stretchy="false">|</mml:mo></mml:mrow></mml:mfrac><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>(2) The Dice score measures the overlap between the ground truth B and the predicted set of pixels A as follows:</p>
<disp-formula id="E10"><label>(9)</label><mml:math id="M10"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mi>D</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>A</mml:mi><mml:mo>,</mml:mo><mml:mi>B</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mn>2</mml:mn><mml:mo>&#x000D7;</mml:mo><mml:mo stretchy="false">|</mml:mo><mml:mi>A</mml:mi><mml:mo>&#x02229;</mml:mo><mml:mi>B</mml:mi><mml:mo stretchy="false">|</mml:mo></mml:mrow><mml:mrow><mml:mo stretchy="false">|</mml:mo><mml:mi>A</mml:mi><mml:mo stretchy="false">|</mml:mo><mml:mo>&#x0002B;</mml:mo><mml:mo stretchy="false">|</mml:mo><mml:mi>B</mml:mi><mml:mo stretchy="false">|</mml:mo></mml:mrow></mml:mfrac><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
</sec>
<sec>
<title>3.3. Implementation Details</title>
<p>U-shaped GAN is implemented in Python using the PyTorch framework. The gray chest radiographs are resized to 1, 024 &#x000D7; 1, 024 and converted to RGB images compatible with the pre-trained ResNet-50 before placement in the network. The weights of ResNet-50 in our discriminator are pre-trained on the ImageNet dataset (<xref ref-type="bibr" rid="B33">33</xref>). The default 5-fold cross-validation is considered to train the semi-supervised model. In the semi-supervised approach, U-shaped GAN is trained with the JSRT or MC dataset with a portion of pixel-wise annotated data and the remainder without pixel-level annotations with 500 epochs. The Adam optimizer (<xref ref-type="bibr" rid="B34">34</xref>) is used to train 500 epochs of the generator and discriminator with initial learning rates of 0.001 and 0.0001, respectively, and multiplied by 0.1 after 200 epochs. For semi-supervised learning, we randomly select 12.5, 25, and 50% of the radiographs from the training set as the annotated set, with the remainder forming the unannotated set. Moreover, we train our model with 100% of the training dataset in the supervised approach. An ablation study is performed to discover the performance contribution from the modified architecture of U-shaped GAN and the proposed pixel-level semi-supervised loss. As in Li et al. (<xref ref-type="bibr" rid="B35">35</xref>), we train U-shaped GAN <italic>via</italic> the semi-supervised approach and supervised approach with 35 annotated radiographs on the JSRT dataset and with 24 annotated radiographs on the MC dataset. The supervised approach is conducted solely with the same annotated set and segmentation network. Moreover, we explore the effect of U-shaped GAN with the original GAN loss, called the original approach, by adding classification layers to U-shaped GAN paralleling with the decoder of our discriminator. The classification layers are identical to those in ResNet-50. In the original approach, the classification layers discriminate the real radiographs from the fake ones, and our discriminator just works as a segmentation network to predict the probability of belonging to the lungs of each pixel in the original approach. For UDA, we first employ the MC and JSRT datasets as the source and target domains, respectively, and then swap their roles. We randomly split each dataset into 7:1:2 for training, validation, and test sets. We train U-shaped GAN similarly to the semi-supervised approach using the source and target data as the annotated and unannotated data, respectively.</p>
</sec>
</sec>
<sec sec-type="results" id="s4">
<title>4. Results</title>
<sec>
<title>4.1. Semi-Supervised Segmentation</title>
<p>U-shaped GAN is trained on the MC and JSRT datasets independently. The comparison with the state-of-the-art semi-supervised CNN (<xref ref-type="bibr" rid="B35">35</xref>) is shown in <xref ref-type="table" rid="T1">Table 1</xref>. As few semi-supervised models on chest radiographs are available, we also compare U-shaped GAN with 1) human observation (<xref ref-type="bibr" rid="B32">32</xref>); 2) traditional methods (<xref ref-type="bibr" rid="B30">30</xref>, <xref ref-type="bibr" rid="B36">36</xref>); and 3) supervised CNNs (<xref ref-type="bibr" rid="B21">21</xref>, <xref ref-type="bibr" rid="B35">35</xref>, <xref ref-type="bibr" rid="B37">37</xref>, <xref ref-type="bibr" rid="B38">38</xref>). The comparison is shown in <xref ref-type="table" rid="T2">Table 2</xref>.</p>
<table-wrap position="float" id="T1">
<label>Table 1</label>
<caption><p>Comparison of U-shaped generative adversarial network (GAN) with the state-of-the-art semi-supervised model.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="left"><bold>Dataset</bold></th>
<th valign="top" align="left"><bold>Model</bold></th>
<th valign="top" align="left"><bold>Main method</bold></th>
<th valign="top" align="left"><bold>Dice</bold></th>
<th valign="top" align="left"><bold>IoU</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">JSRT</td>
<td valign="top" align="left">Li (<xref ref-type="bibr" rid="B35">35</xref>)</td>
<td valign="top" align="left">Semi-supervised CNN(35)</td>
<td valign="top" align="left">0.967</td>
<td valign="top" align="left">-</td>
</tr>
<tr>
<td/>
<td valign="top" align="left">U-shaped GAN</td>
<td valign="top" align="left">Semi-supervised GAN(35)</td>
<td valign="top" align="left"><bold>0.971</bold></td>
<td valign="top" align="left"><bold>0.944</bold></td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p><italic>The numbers after the main method are the numbers of the annotated radiographs used in the semi-supervised approach. Numbers in bold indicate the best result among the models</italic>.</p>
</table-wrap-foot>
</table-wrap>
<table-wrap position="float" id="T2">
<label>Table 2</label>
<caption><p>Comparison of U-shaped GAN with other lung segmentation methods for chest radiograph datasets.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="left"><bold>Dataset</bold></th>
<th valign="top" align="left"><bold>Model</bold></th>
<th valign="top" align="center"><bold>Main method</bold></th>
<th valign="top" align="center"><bold>Dice</bold></th>
<th valign="top" align="center"><bold>IoU</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">JSRT</td>
<td valign="top" align="left">Human (<xref ref-type="bibr" rid="B32">32</xref>)</td>
<td valign="top" align="center">Human observation</td>
<td valign="top" align="center">-</td>
<td valign="top" align="center">0.946 &#x000B1; 0.018</td>
</tr>
<tr>
<td/>
<td valign="top" align="left">Candemir (<xref ref-type="bibr" rid="B30">30</xref>)</td>
<td valign="top" align="center">Traditional method</td>
<td valign="top" align="center">0.967 &#x000B1; 0.008</td>
<td valign="top" align="center">0.954 &#x000B1; 0.015</td>
</tr>
<tr>
<td/>
<td valign="top" align="left">U-net (<xref ref-type="bibr" rid="B35">35</xref>)</td>
<td valign="top" align="center">Supervised CNN</td>
<td valign="top" align="center">0.946</td>
<td valign="top" align="center">-</td>
</tr>
<tr>
<td/>
<td valign="top" align="left">InvertedNet (<xref ref-type="bibr" rid="B21">21</xref>)</td>
<td valign="top" align="center">Supervised CNN</td>
<td valign="top" align="center">0.974</td>
<td valign="top" align="center">0.950</td>
</tr>
<tr>
<td/>
<td valign="top" align="left">Li (<xref ref-type="bibr" rid="B35">35</xref>)</td>
<td valign="top" align="center">Supervised CNN</td>
<td valign="top" align="center">0.967</td>
<td valign="top" align="center">-</td>
</tr>
<tr>
<td/>
<td valign="top" align="left">U-shaped GAN</td>
<td valign="top" align="center">Supervised GAN</td>
<td valign="top" align="center"><bold>0.979</bold> <bold>&#x000B1;</bold> <bold>0.001</bold></td>
<td valign="top" align="center"><bold>0.958</bold> &#x000B1; <bold>0.003</bold></td>
</tr>
<tr>
<td/>
<td valign="top" align="left">U-shaped GAN</td>
<td valign="top" align="center">Semi-supervised GAN(25%)</td>
<td valign="top" align="center">0.975 &#x000B1; 0.001</td>
<td valign="top" align="center">0.951 &#x000B1; 0.002</td>
</tr>
<tr>
<td valign="top" align="left">MC</td>
<td valign="top" align="left">Bosdelekidis (<xref ref-type="bibr" rid="B36">36</xref>)</td>
<td valign="top" align="center">Traditional method</td>
<td valign="top" align="center">0.923</td>
<td valign="top" align="center">0.862</td>
</tr>
<tr>
<td/>
<td valign="top" align="left">Candemir (<xref ref-type="bibr" rid="B30">30</xref>)</td>
<td valign="top" align="center">Traditional method</td>
<td valign="top" align="center">0.960 &#x000B1; 0.018</td>
<td valign="top" align="center">0.941 &#x000B1; 0.034</td>
</tr>
<tr>
<td/>
<td valign="top" align="left">U-net (<xref ref-type="bibr" rid="B37">37</xref>)</td>
<td valign="top" align="center">Supervised CNN</td>
<td valign="top" align="center">-</td>
<td valign="top" align="center">0.942 &#x000B1; 0.046</td>
</tr>
<tr>
<td/>
<td valign="top" align="left">Souza (<xref ref-type="bibr" rid="B38">38</xref>)</td>
<td valign="top" align="center">Supervised CNN</td>
<td valign="top" align="center">0.936</td>
<td valign="top" align="center">0.881</td>
</tr>
<tr>
<td/>
<td valign="top" align="left">U-shaped GAN</td>
<td valign="top" align="center">Supervised GAN</td>
<td valign="top" align="center"><bold>0.976</bold> <bold>&#x000B1;</bold> <bold>0.006</bold></td>
<td valign="top" align="center"><bold>0.955</bold> <bold>&#x000B1;</bold> <bold>0.010</bold></td>
</tr>
<tr>
<td/>
<td valign="top" align="left">U-shaped GAN</td>
<td valign="top" align="center">Semi-supervised GAN(25%)</td>
<td valign="top" align="center">0.968 &#x000B1; 0.011</td>
<td valign="top" align="center">0.940 &#x000B1; 0.019</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p><italic>The number after the main method is the proportion of annotated radiographs from the training set in the semi-supervised approach. Numbers in bold indicate the best result among the models</italic>.</p>
</table-wrap-foot>
</table-wrap>
<p>U-shaped GAN trained with 100% annotated data achieves a performance increase of 0.4&#x02013;10.8% over the state-of-the-art traditional models and supervised CNNs on both the JSRT and MC datasets. The results validate the effectiveness of the design of the segmentation network.</p>
<p>Our semi-supervised model (Dice &#x0003D; 0.975, IoU &#x0003D; 0.951) trained with 25% annotated data outperforms the state-of-the-art supervised models and human observation on the JSRT dataset. Our semi-supervised model (Dice &#x0003D; 0.968, IoU &#x0003D; 0.940) trained with 25% annotated data outperforms the most state-of-the-art supervised models but performs slightly worse than the U-net model (IoU &#x0003D; 0.942) on the MC dataset. Our proposed semi-supervised model achieves outstanding performance with limited annotated datasets. Moreover, U-shaped GAN outperforms the state-of-the-art semi-supervised model (<xref ref-type="bibr" rid="B35">35</xref>) in both supervised and semi-supervised settings by 1.2 and 0.8%, respectively. <xref ref-type="fig" rid="F5">Figure 5</xref> shows a few examples of semi-supervised results with U-shaped GAN. The ground truth contour of the lungs is shown in green, and the segmentation result of the algorithm is in red.</p>
<fig id="F5" position="float">
<label>Figure 5</label>
<caption><p>Semi-supervised results (25% annotated data) with U-shaped GAN. The radiographs from the Japanese Society of Radiological Technology (JSRT) dataset and Montgomery County (MC) dataset appear at the top and bottom, respectively. Green and red contours indicate the true ground and automatic segmentation results, respectively.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fmed-08-782664-g0005.tif"/>
</fig>
<p>We evaluate our approach with 12.5, 25, 50, and 100% annotated radiographs (remaining portions consist of unannotated radiographs). The annotated radiographs in the data splits are randomly sampled from the whole dataset. Notably, the approach works well even with 12.5% annotated data, as shown in <xref ref-type="table" rid="T3">Table 3</xref>. For the details, readers are referred to <xref ref-type="supplementary-material" rid="SM1">Supplementary Figures 1</xref>, <xref ref-type="supplementary-material" rid="SM1">2</xref>.</p>
<table-wrap position="float" id="T3">
<label>Table 3</label>
<caption><p>Comparison of the results of U-shaped GAN trained with different proportions of annotated data.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="left"><bold>Dataset</bold></th>
<th valign="top" align="center"><bold>Annotated data (%)</bold></th>
<th valign="top" align="center"><bold>Dice</bold></th>
<th valign="top" align="center"><bold>IoU</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">JSRT</td>
<td valign="top" align="center">100</td>
<td valign="top" align="center">0.979 &#x000B1; 0.001</td>
<td valign="top" align="center">0.958 &#x000B1; 0.003</td>
</tr>
<tr>
<td/>
<td valign="top" align="center">50</td>
<td valign="top" align="center">0.977 &#x000B1; 0.002</td>
<td valign="top" align="center">0.956 &#x000B1; 0.003</td>
</tr>
<tr>
<td/>
<td valign="top" align="center">25</td>
<td valign="top" align="center">0.975 &#x000B1; 0.001</td>
<td valign="top" align="center">0.951 &#x000B1; 0.002</td>
</tr>
<tr>
<td/>
<td valign="top" align="center">12.5</td>
<td valign="top" align="center">0.964 &#x000B1; 0.007</td>
<td valign="top" align="center">0.934 &#x000B1; 0.008</td>
</tr>
<tr>
<td valign="top" align="left">MC</td>
<td valign="top" align="center">100</td>
<td valign="top" align="center">0.976 &#x000B1; 0.006</td>
<td valign="top" align="center">0.955 &#x000B1; 0.010</td>
</tr>
<tr>
<td/>
<td valign="top" align="center">50</td>
<td valign="top" align="center">0.973 &#x000B1; 0.007</td>
<td valign="top" align="center">0.949 &#x000B1; 0.013</td>
</tr>
<tr>
<td/>
<td valign="top" align="center">25</td>
<td valign="top" align="center">0.968 &#x000B1; 0.011</td>
<td valign="top" align="center">0.940 &#x000B1; 0.019</td>
</tr>
<tr>
<td/>
<td valign="top" align="center">12.5</td>
<td valign="top" align="center">0.958 &#x000B1; 0.021</td>
<td valign="top" align="center">0.922 &#x000B1; 0.034</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>We apply U-shaped GAN on a pneumothorax segmentation dataset (<xref ref-type="bibr" rid="B39">39</xref>) with a semi-supervised approach. This dataset contains 2,669 radiographs with annotated pneumothorax lesion areas. U-shaped GAN shows promising performance on the pneumothorax segmentation in the semi-supervised approach. Most of the results predict rough areas of the pneumothorax lesion correctly, which provides credible help to the radiologist to find the lesion rapidly. For the details, readers are referred to <xref ref-type="supplementary-material" rid="SM1">Supplementary Figure 3</xref>.</p>
</sec>
<sec>
<title>4.2. Unsupervised Domain Adaptation</title>
<p>We use the MC and JSRT datasets as the source and target domains, respectively, and then swap their roles for UDA. The performances of our UDA model on the target domains are compared under various settings: 1) the model being trained on source data and tested on the target domain with no DA (T-noDA); 2) UDA model testing on the source domain (S-test); 3) human observation (<xref ref-type="bibr" rid="B32">32</xref>); 4) UDA models with CNNs (<xref ref-type="bibr" rid="B15">15</xref>, <xref ref-type="bibr" rid="B37">37</xref>).</p>
<p>As shown in <xref ref-type="table" rid="T4">Table 4</xref>, when directly applying the learned source domain model to target data, the model performance significantly degrades, indicating that domain shift would severely impede the generalization performance of CNNs. However, remarkable improvements are achieved by applying the unsupervised loss on the target images. Compared to the T-noDA results, the segmentation predictions increase by 3.3% and 5.3% on the JSRT and MC datasets, respectively, with our UDA approach.</p>
<table-wrap position="float" id="T4">
<label>Table 4</label>
<caption><p>Comparison of segmentation results among different unsupervised domain adaptation (UDA) methods.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="left"><bold>Dataset</bold></th>
<th valign="top" align="left"><bold>Model</bold></th>
<th valign="top" align="left"><bold>Main method</bold></th>
<th valign="top" align="left"><bold>Dice</bold></th>
<th valign="top" align="left"><bold>IoU</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">MC &#x02192; JSRT</td>
<td valign="top" align="left">Human observation (<xref ref-type="bibr" rid="B32">32</xref>)</td>
<td valign="top" align="left">Human observation</td>
<td valign="top" align="left">-</td>
<td valign="top" align="left">0.946</td>
</tr>
<tr>
<td/>
<td valign="top" align="left">T-noDAg</td>
<td valign="top" align="left">-</td>
<td valign="top" align="left">0.934</td>
<td valign="top" align="left">0.895</td>
</tr>
<tr>
<td/>
<td valign="top" align="left">S-test</td>
<td valign="top" align="left">-</td>
<td valign="top" align="left">0.981</td>
<td valign="top" align="left">0.963</td>
</tr>
<tr>
<td/>
<td valign="top" align="left">MUNIT (<xref ref-type="bibr" rid="B37">37</xref>)</td>
<td valign="top" align="left">UDA with MUNIT</td>
<td valign="top" align="left">-</td>
<td valign="top" align="left">0.882</td>
</tr>
<tr>
<td/>
<td valign="top" align="left">CyUDA (<xref ref-type="bibr" rid="B15">15</xref>)</td>
<td valign="top" align="left">UDA with CycleGAN</td>
<td valign="top" align="left">0.928</td>
<td valign="top" align="left">-</td>
</tr>
<tr>
<td/>
<td valign="top" align="left">SeUDA (<xref ref-type="bibr" rid="B15">15</xref>)</td>
<td valign="top" align="left">UDA with CycleGAN</td>
<td valign="top" align="left">0.945</td>
<td valign="top" align="left">-</td>
</tr>
<tr>
<td/>
<td valign="top" align="left">Our method</td>
<td valign="top" align="left">UDA with GAN</td>
<td valign="top" align="left"><bold>0.965</bold></td>
<td valign="top" align="left"><bold>0.932</bold></td>
</tr>
<tr>
<td valign="top" align="left">JSRT &#x02192; MC</td>
<td valign="top" align="left">T-noDAg</td>
<td valign="top" align="left">-</td>
<td valign="top" align="left">0.918</td>
<td valign="top" align="left">0.880</td>
</tr>
<tr>
<td/>
<td valign="top" align="left">S-test</td>
<td valign="top" align="left">-</td>
<td valign="top" align="left">0.980</td>
<td valign="top" align="left">0.961</td>
</tr>
<tr>
<td/>
<td valign="top" align="left">Our method</td>
<td valign="top" align="left">UDA with GAN</td>
<td valign="top" align="left"><bold>0.967</bold></td>
<td valign="top" align="left"><bold>0.936</bold></td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p><italic>Numbers in bold indicate the best result among the models</italic>.</p>
</table-wrap-foot>
</table-wrap>
<p>Experimental results demonstrate a significant enhancement in performance compared to other models. Compared with other UDA models based on CNNs (<xref ref-type="bibr" rid="B15">15</xref>, <xref ref-type="bibr" rid="B37">37</xref>), U-shaped GAN achieves a significant improvement over MUNIT (5.7%), CyUDA (4.0%), and SeUDA (2.1%). Moreover, U-shaped GAN is even comparable to human observation. <xref ref-type="fig" rid="F6">Figure 6</xref> shows a few examples of UDA results with U-shaped GAN. The ground truth contour of the lungs is shown in green, and the segmentation result of the algorithm is in red.</p>
<fig id="F6" position="float">
<label>Figure 6</label>
<caption><p>UDA results with U-shaped GAN. The radiographs from the JSRT dataset with MC &#x02192; JSRT UDA results are on the top; the radiographs from the MC dataset with JSRT &#x02192; MC UDA results are on the bottom. Green and red contours indicate the ground truth and automatic segmentation results, respectively.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fmed-08-782664-g0006.tif"/>
</fig>
<p>Compared to models trained with 100% annotated data on source domains, the UDA models obtain similar accuracy on the source domains (<italic>S</italic>-test). Therefore, U-shaped GAN is able to improve results on the target domains while maintaining segmentation performance on the source domains. For the details, readers are referred to <xref ref-type="supplementary-material" rid="SM1">Supplementary Figures 4</xref>, <xref ref-type="supplementary-material" rid="SM1">5</xref>.</p>
</sec>
<sec>
<title>4.3. Ablation Study</title>
<p>For the ablation study, we study the effectiveness of our modified segmentation network, the architecture of U-shaped GAN, and the proposed pixel-level semi-supervised loss. To demonstrate the advantage of U-shaped GAN, we report the scores of U-net (the original U-shaped net), U-shaped GAN trained with the supervised approach, U-shaped GAN trained with the original approach, and U-shaped GAN trained with our pixel-level loss.</p>
<p>First, the results of the comparison of our modified segmentation network and the original U-shaped net are shown in <xref ref-type="table" rid="T5">Table 5</xref>. The segmentation network of U-shaped GAN is designed with high resolution radiographs following the main idea of the U-shaped net (<xref ref-type="bibr" rid="B24">24</xref>). It is shown that our modified segmentation network improves the prediction by 3.28% on the JSRT dataset and 2.02% on the MC dataset when trained with the whole annotated datasets.</p>
<table-wrap position="float" id="T5">
<label>Table 5</label>
<caption><p>Comparison of U-shaped GAN and U-net trained with the whole dataset.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="left"><bold>Dataset</bold></th>
<th valign="top" align="left"><bold>Model</bold></th>
<th valign="top" align="left"><bold>Main method</bold></th>
<th valign="top" align="left"><bold>Dice</bold></th>
<th valign="top" align="left"><bold>IoU</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">JSRT</td>
<td valign="top" align="left">U-net (<xref ref-type="bibr" rid="B35">35</xref>)</td>
<td valign="top" align="left">Supervised approach</td>
<td valign="top" align="left">0.946</td>
<td valign="top" align="left">-</td>
</tr>
<tr>
<td/>
<td valign="top" align="left">U-shaped GAN</td>
<td valign="top" align="left">Supervised approach</td>
<td valign="top" align="left"><bold>0.977</bold></td>
<td valign="top" align="left"><bold>0.955</bold></td>
</tr>
<tr>
<td valign="top" align="left">MC</td>
<td valign="top" align="left">U-net (<xref ref-type="bibr" rid="B37">37</xref>)</td>
<td valign="top" align="left">Supervised approach</td>
<td valign="top" align="left">-</td>
<td valign="top" align="left">0.942</td>
</tr>
<tr>
<td/>
<td valign="top" align="left">U-shaped GAN</td>
<td valign="top" align="left">Supervised approach</td>
<td valign="top" align="left"><bold>0.980</bold></td>
<td valign="top" align="left"><bold>0.961</bold></td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p><italic>Numbers in bold indicate the best result among the models</italic>.</p>
</table-wrap-foot>
</table-wrap>
<p>Second, the effectiveness of the architecture of U-shaped GAN is investigated. The U-shaped net is incorporated into the structure of GAN leveraging unannotated data to assist the segmentation task. By adding an original GAN loss to the supervised approach, the Dice scores increase from 0.968 to 0.970 on the JSRT dataset and from 0.966 to 0.971 on the MC dataset. The architecture of U-shaped GAN is successful in leveraging unannotated data to find a representation for the whole dataset, shown in <xref ref-type="table" rid="T6">Table 6</xref>.</p>
<table-wrap position="float" id="T6">
<label>Table 6</label>
<caption><p>Comparison of U-shaped GAN with different training approaches.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="left"><bold>Dataset</bold></th>
<th valign="top" align="left"><bold>Model</bold></th>
<th valign="top" align="left"><bold>Main method</bold></th>
<th valign="top" align="left"><bold>Dice</bold></th>
<th valign="top" align="left"><bold>IoU</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">JSRT</td>
<td valign="top" align="left">U-shaped GAN</td>
<td valign="top" align="left">Supervised approach(35)</td>
<td valign="top" align="left">0.968</td>
<td valign="top" align="left">0.939</td>
</tr>
<tr>
<td/>
<td valign="top" align="left">U-shaped GAN</td>
<td valign="top" align="left">Original approach(35)</td>
<td valign="top" align="left">0.970</td>
<td valign="top" align="left">0.941</td>
</tr>
<tr>
<td/>
<td valign="top" align="left">U-shaped GAN</td>
<td valign="top" align="left">Semi-supervised approach(35)</td>
<td valign="top" align="left"><bold>0.971</bold></td>
<td valign="top" align="left"><bold>0.944</bold></td>
</tr>
<tr>
<td valign="top" align="left">MC</td>
<td valign="top" align="left">U-shaped GAN</td>
<td valign="top" align="left">Supervised approach(24)</td>
<td valign="top" align="left">0.966</td>
<td valign="top" align="left">0.936</td>
</tr>
<tr>
<td/>
<td valign="top" align="left">U-shaped GAN</td>
<td valign="top" align="left">Original approach(24)</td>
<td valign="top" align="left">0.971</td>
<td valign="top" align="left">0.945</td>
</tr>
<tr>
<td/>
<td valign="top" align="left">U-shaped GAN</td>
<td valign="top" align="left">Semi-supervised approach(24)</td>
<td valign="top" align="left"><bold>0.973</bold></td>
<td valign="top" align="left"><bold>0.948</bold></td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p><italic>The number after the main method is the number of the annotated radiographs used in training. Numbers in bold indicate the best result among the models</italic>.</p>
</table-wrap-foot>
</table-wrap>
<p>Third, we show the comparison between results gained by the GAN original loss and our pixel-level GAN loss with semi-supervised training. The pixel-level GAN loss increases the capacity of U-shaped GAN in finding the representation of the whole dataset. The segmentation results improve from Dice &#x0003D; 0.970, IoU &#x0003D; 0.941 to Dice &#x0003D; 0.971, IoU &#x0003D; 0.944 on the JSRT dataset and Dice &#x0003D; 0.971, IoU &#x0003D; 0.945 to Dice &#x0003D; 0.973, IoU &#x0003D; 0.948 on the MC dataset, shown in <xref ref-type="table" rid="T6">Table 6</xref>.</p>
<p>Some results of U-shaped GAN with different training approaches on confusing samples are shown in <xref ref-type="fig" rid="F7">Figure 7</xref>. U-shaped GAN seems to be resistant to interference by irrelevant features, such as other organs and lesion areas, by using the GAN architecture and to increase the resistant capability by the pixel-level loss.</p>
<fig id="F7" position="float">
<label>Figure 7</label>
<caption><p>Comparison of the results of U-shaped GAN with different training approaches on confusing samples. The radiographs in the first row are results of the supervised approach. The second row is the results of the original approach. The third row is the results of the semi-supervised approach. Green and red contours indicate the ground truth and automatic segmentation results, respectively.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fmed-08-782664-g0007.tif"/>
</fig>
</sec>
</sec>
<sec sec-type="discussion" id="s5">
<title>5. Discussion</title>
<p>In this study, we propose U-shaped GAN to address the scarcity of annotated data and domain shift. U-shaped GAN for radiographs shows strong performance in semi-supervised learning and UDA approaches. To handle datasets from multiple medical centers conveniently and efficiently, we combine semi-supervised learning and UDA in radiograph segmentation into a single method. U-shaped GAN functions similarly with the annotated and unannotated data in the semi-supervised and UDA approaches. The effectiveness of the model is demonstrated through extensive experiments.</p>
<p>Training models with high image resolution is effective; however, it would increase the computational burden (<xref ref-type="bibr" rid="B20">20</xref>, <xref ref-type="bibr" rid="B21">21</xref>). The previous models concentrated on the 256 &#x000D7; 256 or 512 &#x000D7; 512 image resolutions (<xref ref-type="bibr" rid="B15">15</xref>, <xref ref-type="bibr" rid="B21">21</xref>, <xref ref-type="bibr" rid="B35">35</xref>, <xref ref-type="bibr" rid="B37">37</xref>, <xref ref-type="bibr" rid="B38">38</xref>). We propose U-shaped GAN for high resolution radiographs (1, 024 &#x000D7; 1, 024). We use pointwise convolution for dimensionality reduction, decreasing the number of feature maps while retaining their salient features. Moreover, we design the U-shaped net with a pretrained ResNet-50 as encoder, which further reduces the computational burden of the training encoder from scratch. U-shaped GAN trained with the whole annotated data achieves a more accurate performance than the state-of-the-art supervised models as well as the original U-shaped net. This prediction result shows that U-shaped GAN is effective for segmentation prediction.</p>
<p>The previous studies often analyzed semi-supervised learning and UDA problems separately (<xref ref-type="bibr" rid="B3">3</xref>, <xref ref-type="bibr" rid="B11">11</xref>, <xref ref-type="bibr" rid="B12">12</xref>, <xref ref-type="bibr" rid="B14">14</xref>&#x02013;<xref ref-type="bibr" rid="B16">16</xref>). In semi-supervised learning, the previous GANs were usually proposed to distinguish between segmentation probability maps and the ground truth (<xref ref-type="bibr" rid="B3">3</xref>, <xref ref-type="bibr" rid="B11">11</xref>, <xref ref-type="bibr" rid="B12">12</xref>). The generators produced the segmentation probability maps as the segmentation networks (<xref ref-type="bibr" rid="B3">3</xref>, <xref ref-type="bibr" rid="B11">11</xref>, <xref ref-type="bibr" rid="B12">12</xref>). In UDA, the previous GANs were usually proposed to distinguish between source data and target data (<xref ref-type="bibr" rid="B14">14</xref>&#x02013;<xref ref-type="bibr" rid="B16">16</xref>). The generators transferred the target domain to the source domain and extra networks were designed for the segmentation (<xref ref-type="bibr" rid="B14">14</xref>&#x02013;<xref ref-type="bibr" rid="B16">16</xref>). The annotated and unannotated data may come from either the same or different domains in the dataset collected from multiple centers. Dealing with the two problems separately increases the model complexity. Moreover, separating the dataset to train two segmentation networks decreases the utilization efficiency of collected data. Therefore, we propose a single model to deal with the semi-supervised learning and UDA approaches at the same time. We attribute the model&#x00027;s remarkable generalization capabilities to the effective use of the unannotated data. We use adversarial learning to achieve a representation for lung segmentation in chest radiographs. In U-shaped GAN, we employ a generator to generate realistic data, which, together with the real data (most of them are unannotated data), force the discriminator to find the most salient features. Our discriminator, which in addition to classifying the pixels into lungs, determines whether a given pixel belongs to the real or generated data.</p>
<p>U-shaped GAN exploits more widely available unannotated data to complement small annotated data with a semi-supervised loss. U-shaped GAN achieves greater performance than the state-of-the-art semi-supervised model. Moreover, it is comparable to the supervised models with 25% annotated data. U-shaped GAN works well even with 12.5% annotated data with Dice scores of 0.964 and 0.958 on JSRT and MC datasets, respectively. Unlike the previous semi-supervised study using the generators to produce the segmentation probability maps as the segmentation networks (<xref ref-type="bibr" rid="B3">3</xref>, <xref ref-type="bibr" rid="B11">11</xref>, <xref ref-type="bibr" rid="B12">12</xref>, <xref ref-type="bibr" rid="B25">25</xref>), U-shaped GAN uses the generator to generate realistic data. The realistic data, together with the real data, force our discriminator to learn a better representation for the radiographs. Compared with the result achieved by trained in the supervised approach, U-shaped GAN achieves increased performance in the semi-supervised learning approach with adversarial learning. Instead of discriminating real or fake labels on image-level (<xref ref-type="bibr" rid="B3">3</xref>, <xref ref-type="bibr" rid="B11">11</xref>, <xref ref-type="bibr" rid="B12">12</xref>, <xref ref-type="bibr" rid="B25">25</xref>), a pixel-level loss is proposed to extract more information from the radiographs. The segmentation accuracy is improved when using the proposed loss. For the data from one domain, U-shaped GAN effectively leverages the unannotated data to achieve high segmentation accuracy and reduces the cost of medical image annotation.</p>
<p>U-shaped GAN is extended to UDA to reduce domain shift without the extra expense of annotation on the target domain. Instead of transferring the target domain to the source domain (<xref ref-type="bibr" rid="B14">14</xref>&#x02013;<xref ref-type="bibr" rid="B16">16</xref>, <xref ref-type="bibr" rid="B37">37</xref>), our generator generates realistic data. Discriminating the realistic data from the real ones, our discriminator learns a better representation. U-shaped GAN is better than the state-of-the-art UDA models and comparable to human observation. It achieves high accuracy on the target domain while maintaining the accuracy on the source domain (S-test). Thus, the model can be trained with data collected from multiple medical centers. Regardless of whether the unannotated data come from single or multiple domains, the prediction accuracies on their corresponding domains are increased, and the accuracies on other domains are maintained. Because the same networks are used in the two approaches, the datasets are sufficient to train U-shaped GAN and generalize the model among various domains, making it suitable for clinical applications in a multiple center system. In addition, using the same architecture at multiple medical centers reduces the model complexity.</p>
<p>Results of our evaluation are promising, but U-shaped GAN has only been fully tested with lung segmentation. In the future, we will extend the model to detect a wider range of lung diseases by collecting additional chest radiographs of different diseases from multiple medical centers.</p>
</sec>
<sec sec-type="conclusions" id="s6">
<title>6. Conclusion</title>
<p>In this study, we propose U-shaped GAN to overcome the crucial problems caused by scarce labeled data and inevitable domain shift. The GAN-based model is designed at a high resolution (1, 024 &#x000D7; 1, 024) for effective segmentation. The semi-supervised learning approach and UDA approach are modeled into a unified framework for effective radiograph segmentation. We leverage unannotated and annotated data with a pixel-level semi-supervised loss. U-shaped GAN is compatible with varying data distributions of multiple medical centers, with efficient training and optimizing performance. Our experiment results demonstrate that U-shaped GAN achieved more accurate lung segmentation performance as compared with the state-of-the-art models. U-shaped GAN is more appealing to the model development and clinical application by eliminating the need to use two different models to deal with the aforementioned problems.</p>
</sec>
<sec sec-type="data-availability" id="s7">
<title>Data Availability Statement</title>
<p>The original contributions presented in the study are included in the article/<xref ref-type="supplementary-material" rid="SM1">Supplementary Material</xref>, further inquiries can be directed to the corresponding author/s.</p>
</sec>
<sec id="s8">
<title>Author Contributions</title>
<p>HG and HW conceived the idea for this study. HW worked on the end-to-end implementation of the study. JW provided relevant insights on the clinical impact of the research work and handled the redaction of the paper. PQ managed the project and provided the funding for the research. All authors contributed to the article and approved the submitted version.</p>
</sec>
<sec sec-type="funding-information" id="s9">
<title>Funding</title>
<p>This work was supported by the National Natural Science Foundation of China (grant nos. 61633006 and 81872247) and the Fundamental Research Funds for the Central Universities, China (grant no. DUT21YG118).</p>
</sec>
<sec sec-type="COI-statement" id="conf1">
<title>Conflict of Interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="disclaimer" id="s10">
<title>Publisher&#x00027;s Note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec> </body>
<back><sec sec-type="supplementary-material" id="s11">
<title>Supplementary Material</title>
<p>The Supplementary Material for this article can be found online at: <ext-link ext-link-type="uri" xlink:href="https://www.frontiersin.org/articles/10.3389/fmed.2021.782664/full#supplementary-material">https://www.frontiersin.org/articles/10.3389/fmed.2021.782664/full#supplementary-material</ext-link></p>
<supplementary-material xlink:href="Image_1.PDF" id="SM1" mimetype="application/pdf" xmlns:xlink="http://www.w3.org/1999/xlink"/>
</sec>
<ref-list>
<title>References</title>
<ref id="B1">
<label>1.</label>
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Souly</surname> <given-names>N</given-names></name> <name><surname>Spampinato</surname> <given-names>C</given-names></name> <name><surname>Shah</surname> <given-names>M</given-names></name></person-group>. <article-title>Semi supervised semantic segmentation using generative adversarial network</article-title>. In: <source>Proceedings of the IEEE International Conference on Computer Vision</source>. <publisher-loc>Venice</publisher-loc>: <publisher-name>IEEE</publisher-name> (<year>2017</year>). p. <fpage>5688</fpage>&#x02013;<lpage>96</lpage>. <pub-id pub-id-type="doi">10.1109/ICCV.2017.606</pub-id><pub-id pub-id-type="pmid">32658792</pub-id></citation></ref>
<ref id="B2">
<label>2.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hung</surname> <given-names>WC</given-names></name> <name><surname>Tsai</surname> <given-names>YH</given-names></name> <name><surname>Liou</surname> <given-names>YT</given-names></name> <name><surname>Lin</surname> <given-names>YY</given-names></name> <name><surname>Yang</surname> <given-names>MH</given-names></name></person-group>. <article-title>Adversarial learning for semi-supervised semantic segmentation</article-title>. <source>arXiv preprint</source> arXiv:180207934 (<year>2018</year>).</citation>
</ref>
<ref id="B3">
<label>3.</label>
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Zheng</surname> <given-names>H</given-names></name> <name><surname>Lin</surname> <given-names>L</given-names></name> <name><surname>Hu</surname> <given-names>H</given-names></name> <name><surname>Zhang</surname> <given-names>Q</given-names></name> <name><surname>Chen</surname> <given-names>Q</given-names></name> <name><surname>Iwamoto</surname> <given-names>Y</given-names></name> <etal/></person-group>. <article-title>Semi-supervised segmentation of liver using adversarial learning with deep atlas prior</article-title>. In: <source>International Conference on Medical Image Computing and Computer-Assisted Intervention</source>. <publisher-loc>Cham</publisher-loc>: <publisher-name>Springer</publisher-name> (<year>2019</year>). p. <fpage>148</fpage>&#x02013;<lpage>56</lpage>.</citation>
</ref>
<ref id="B4">
<label>4.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>H</given-names></name> <name><surname>Gu</surname> <given-names>H</given-names></name> <name><surname>Qin</surname> <given-names>P</given-names></name> <name><surname>Wang</surname> <given-names>J</given-names></name></person-group>. <article-title>CheXLocNet: automatic localization of pneumothorax in chest radiographs using deep convolutional neural networks</article-title>. <source>PLoS ONE</source>. (<year>2020</year>) <volume>15</volume>:<fpage>e0242013</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pone.0242013</pub-id><pub-id pub-id-type="pmid">33166371</pub-id></citation></ref>
<ref id="B5">
<label>5.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Van Eycke</surname> <given-names>YR</given-names></name> <name><surname>Foucart</surname> <given-names>A</given-names></name> <name><surname>Decaestecker</surname> <given-names>C</given-names></name></person-group>. <article-title>Strategies to reduce the expert supervision required for deep Learning-Based segmentation of histopathological images</article-title>. <source>Front Med</source>. (<year>2019</year>) <volume>6</volume>:<fpage>222</fpage>. <pub-id pub-id-type="doi">10.3389/fmed.2019.00222</pub-id><pub-id pub-id-type="pmid">31681779</pub-id></citation></ref>
<ref id="B6">
<label>6.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Nawrocki</surname> <given-names>T</given-names></name> <name><surname>Maldjian</surname> <given-names>PD</given-names></name> <name><surname>Slasky</surname> <given-names>SE</given-names></name> <name><surname>Contractor</surname> <given-names>SG</given-names></name></person-group>. <article-title>Artificial intelligence and radiology: have rumors of the radiologist&#x00027;s demise been greatly exaggerated?</article-title> <source>Acad Radiol</source>. (<year>2018</year>) <volume>25</volume>:<fpage>967</fpage>&#x02013;<lpage>72</lpage>. <pub-id pub-id-type="doi">10.1016/j.acra.2017.12.027</pub-id></citation>
</ref>
<ref id="B7">
<label>7.</label>
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Jin</surname> <given-names>D</given-names></name> <name><surname>Harrison</surname> <given-names>AP</given-names></name> <name><surname>Zhang</surname> <given-names>L</given-names></name> <name><surname>Yan</surname> <given-names>K</given-names></name> <name><surname>Wang</surname> <given-names>Y</given-names></name> <name><surname>Cai</surname> <given-names>J</given-names></name> <etal/></person-group>. <article-title>Chapter 14 - artificial intelligence in radiology</article-title>. In: Xing L, Giger ML, Min JK, editors. <source>Artificial Intelligence in Radiology</source>. <publisher-loc>Cambridge, MA</publisher-loc>: <publisher-name>Academic Press</publisher-name> (<year>2021</year>). p. <fpage>265</fpage>&#x02013;<lpage>89</lpage>.</citation>
</ref>
<ref id="B8">
<label>8.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>L</given-names></name> <name><surname>Wang</surname> <given-names>X</given-names></name> <name><surname>Yang</surname> <given-names>D</given-names></name> <name><surname>Sanford</surname> <given-names>T</given-names></name> <name><surname>Harmon</surname> <given-names>S</given-names></name> <name><surname>Turkbey</surname> <given-names>B</given-names></name> <etal/></person-group>. <article-title>Generalizing deep learning for medical image segmentation to unseen domains via deep stacked transformation</article-title>. <source>IEEE Trans Med Imaging</source>. (<year>2020</year>) <volume>39</volume>:<fpage>2531</fpage>&#x02013;<lpage>40</lpage>. <pub-id pub-id-type="doi">10.1109/TMI.2020.2973595</pub-id><pub-id pub-id-type="pmid">32070947</pub-id></citation></ref>
<ref id="B9">
<label>9.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Briganti</surname> <given-names>G</given-names></name> <name><surname>Le Moine</surname> <given-names>O</given-names></name></person-group>. <article-title>Artificial intelligence in medicine: today and tomorrow</article-title>. <source>Front Med</source>. (<year>2020</year>) <volume>7</volume>:<fpage>27</fpage>. <pub-id pub-id-type="doi">10.3389/fmed.2020.00027</pub-id><pub-id pub-id-type="pmid">32118012</pub-id></citation></ref>
<ref id="B10">
<label>10.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Goodfellow</surname> <given-names>I</given-names></name> <name><surname>Pouget-Abadie</surname> <given-names>J</given-names></name> <name><surname>Mirza</surname> <given-names>M</given-names></name> <name><surname>Xu</surname> <given-names>B</given-names></name> <name><surname>Warde-Farley</surname> <given-names>D</given-names></name> <name><surname>Ozair</surname> <given-names>S</given-names></name> <etal/></person-group>. <article-title>Generative adversarial nets</article-title>. In: <source>Advances in Neural Information Processing Systems.</source> Montr&#x000E9;al (<year>2014</year>). p. <fpage>2672</fpage>&#x02013;<lpage>80</lpage>.</citation>
</ref>
<ref id="B11">
<label>11.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Liu</surname> <given-names>X</given-names></name> <name><surname>Cao</surname> <given-names>J</given-names></name> <name><surname>Fu</surname> <given-names>T</given-names></name> <name><surname>Pan</surname> <given-names>Z</given-names></name> <name><surname>Hu</surname> <given-names>W</given-names></name> <name><surname>Zhang</surname> <given-names>K</given-names></name> <etal/></person-group>. <article-title>Semi-supervised automatic segmentation of layer and fluid region in retinal optical coherence tomography images using adversarial learning</article-title>. <source>IEEE Access</source>. (<year>2018</year>) <volume>7</volume>:<fpage>3046</fpage>&#x02013;<lpage>3061</lpage>. <pub-id pub-id-type="doi">10.1109/ACCESS.2018.2889321</pub-id><pub-id pub-id-type="pmid">27295638</pub-id></citation></ref>
<ref id="B12">
<label>12.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Decourt</surname> <given-names>C</given-names></name> <name><surname>Duong</surname> <given-names>L</given-names></name></person-group>. <article-title>Semi-supervised generative adversarial networks for the segmentation of the left ventricle in pediatric MRI</article-title>. <source>Comput Biol Med</source>. (<year>2020</year>) <volume>123</volume>:<fpage>103884</fpage>. <pub-id pub-id-type="doi">10.1016/j.compbiomed.2020.103884</pub-id><pub-id pub-id-type="pmid">32658792</pub-id></citation></ref>
<ref id="B13">
<label>13.</label>
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Ghafoorian</surname> <given-names>M</given-names></name> <name><surname>Mehrtash</surname> <given-names>A</given-names></name> <name><surname>Kapur</surname> <given-names>T</given-names></name> <name><surname>Karssemeijer</surname> <given-names>N</given-names></name> <name><surname>Marchiori</surname> <given-names>E</given-names></name> <name><surname>Pesteie</surname> <given-names>M</given-names></name> <etal/></person-group>. <article-title>Transfer learning for domain adaptation in mri: application in brain lesion segmentation</article-title>. In: <source>International Conference on Medical Image Computing and Computer-Assisted Intervention</source>. <publisher-loc>Cham</publisher-loc>: <publisher-name>Springer</publisher-name> (<year>2017</year>). p. <fpage>516</fpage>&#x02013;<lpage>24</lpage>.</citation>
</ref>
<ref id="B14">
<label>14.</label>
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Kamnitsas</surname> <given-names>K</given-names></name> <name><surname>Baumgartner</surname> <given-names>C</given-names></name> <name><surname>Ledig</surname> <given-names>C</given-names></name> <name><surname>Newcombe</surname> <given-names>V</given-names></name> <name><surname>Simpson</surname> <given-names>J</given-names></name> <name><surname>Kane</surname> <given-names>A</given-names></name> <etal/></person-group>. <article-title>Unsupervised domain adaptation in brain lesion segmentation with adversarial networks</article-title>. In: <source>International Conference on Information Processing in Medical Imaging</source>. <publisher-loc>Cham</publisher-loc>: <publisher-name>Springer</publisher-name> (<year>2017</year>). p. <fpage>597</fpage>&#x02013;<lpage>609</lpage>.</citation>
</ref>
<ref id="B15">
<label>15.</label>
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Chen</surname> <given-names>C</given-names></name> <name><surname>Dou</surname> <given-names>Q</given-names></name> <name><surname>Chen</surname> <given-names>H</given-names></name> <name><surname>Heng</surname> <given-names>PA</given-names></name></person-group>. <article-title>Semantic-aware generative adversarial nets for unsupervised domain adaptation in chest x-ray segmentation</article-title>. In: <source>International Workshop on Machine Learning in Medical Imaging</source>. <publisher-loc>Cham</publisher-loc>: <publisher-name>Springer</publisher-name> (<year>2018</year>). p. <fpage>143</fpage>&#x02013;<lpage>51</lpage>.</citation>
</ref>
<ref id="B16">
<label>16.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Shen</surname> <given-names>R</given-names></name> <name><surname>Yao</surname> <given-names>J</given-names></name> <name><surname>Yan</surname> <given-names>K</given-names></name> <name><surname>Tian</surname> <given-names>K</given-names></name> <name><surname>Jiang</surname> <given-names>C</given-names></name> <name><surname>Zhou</surname> <given-names>K</given-names></name></person-group>. <article-title>Unsupervised domain adaptation with adversarial learning for mass detection in mammogram</article-title>. <source>Neurocomputing</source>. (<year>2020</year>) <volume>393</volume>:<fpage>27</fpage>&#x02013;<lpage>37</lpage>. <pub-id pub-id-type="doi">10.1016/j.neucom.2020.01.099</pub-id></citation>
</ref>
<ref id="B17">
<label>17.</label>
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>X</given-names></name> <name><surname>Peng</surname> <given-names>Y</given-names></name> <name><surname>Lu</surname> <given-names>L</given-names></name> <name><surname>Lu</surname> <given-names>Z</given-names></name> <name><surname>Bagheri</surname> <given-names>M</given-names></name> <name><surname>Summers</surname> <given-names>RM</given-names></name></person-group>. <article-title>Chestx-ray8: hospital-scale chest x-ray database and benchmarks on weakly-supervised classification and localization of common thorax diseases</article-title>. In: <source>Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition</source>. <publisher-loc>Honolulu, HI</publisher-loc>: <publisher-name>IEEE</publisher-name> (<year>2017</year>). p. <fpage>2097</fpage>&#x02013;<lpage>106</lpage>.</citation>
</ref>
<ref id="B18">
<label>18.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>McBee</surname> <given-names>MP</given-names></name> <name><surname>Awan</surname> <given-names>OA</given-names></name> <name><surname>Colucci</surname> <given-names>AT</given-names></name> <name><surname>Ghobadi</surname> <given-names>CW</given-names></name> <name><surname>Kadom</surname> <given-names>N</given-names></name> <name><surname>Kansagra</surname> <given-names>AP</given-names></name> <etal/></person-group>. <article-title>Deep learning in radiology</article-title>. <source>Acad Radiol</source>. (<year>2018</year>) <volume>25</volume>:<fpage>1472</fpage>&#x02013;<lpage>80</lpage>. <pub-id pub-id-type="doi">10.1016/j.acra.2018.02.018</pub-id><pub-id pub-id-type="pmid">29606338</pub-id></citation></ref>
<ref id="B19">
<label>19.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Cohen</surname> <given-names>JP</given-names></name> <name><surname>Morrison</surname> <given-names>P</given-names></name> <name><surname>Dao</surname> <given-names>L</given-names></name> <name><surname>Roth</surname> <given-names>K</given-names></name> <name><surname>Duong</surname> <given-names>TQ</given-names></name> <name><surname>Ghassemi</surname> <given-names>M</given-names></name></person-group>. <article-title>Covid-19 image data collection: prospective predictions are the future</article-title>. <source>arXiv preprint</source> arXiv:200611988. (<year>2020</year>).</citation>
</ref>
<ref id="B20">
<label>20.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Eslami</surname> <given-names>M</given-names></name> <name><surname>Tabarestani</surname> <given-names>S</given-names></name> <name><surname>Albarqouni</surname> <given-names>S</given-names></name> <name><surname>Adeli</surname> <given-names>E</given-names></name> <name><surname>Navab</surname> <given-names>N</given-names></name> <name><surname>Adjouadi</surname> <given-names>M</given-names></name></person-group>. <article-title>Image-to-images translation for multi-task organ segmentation and bone suppression in Chest X-Ray radiography</article-title>. <source>IEEE Trans Med Imaging</source>. (<year>2020</year>) <volume>39</volume>:<fpage>2553</fpage>&#x02013;<lpage>65</lpage>. <pub-id pub-id-type="doi">10.1109/TMI.2020.2974159</pub-id><pub-id pub-id-type="pmid">32078541</pub-id></citation></ref>
<ref id="B21">
<label>21.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Novikov</surname> <given-names>AA</given-names></name> <name><surname>Lenis</surname> <given-names>D</given-names></name> <name><surname>Major</surname> <given-names>D</given-names></name> <name><surname>Hladvka</surname> <given-names>J</given-names></name> <name><surname>Wimmer</surname> <given-names>M</given-names></name> <name><surname>Bhler</surname> <given-names>K</given-names></name></person-group>. <article-title>Fully convolutional architectures for multiclass segmentation in chest radiographs</article-title>. <source>IEEE Trans Med Imaging</source>. (<year>2018</year>) <volume>37</volume>:<fpage>1865</fpage>&#x02013;<lpage>76</lpage>. <pub-id pub-id-type="doi">10.1109/TMI.2018.2806086</pub-id><pub-id pub-id-type="pmid">29994439</pub-id></citation></ref>
<ref id="B22">
<label>22.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Donahue</surname> <given-names>J</given-names></name> <name><surname>Kr&#x000E4;henb&#x000FC;hl</surname> <given-names>P</given-names></name> <name><surname>Darrell</surname> <given-names>T</given-names></name></person-group>. <article-title>Adversarial feature learning</article-title>. <source>arXiv preprint</source> arXiv:160509782. (<year>2016</year>).</citation>
</ref>
<ref id="B23">
<label>23.</label>
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Goodfellow</surname> <given-names>I</given-names></name> <name><surname>Bengio</surname> <given-names>Y</given-names></name> <name><surname>Courville</surname> <given-names>A</given-names></name> <name><surname>Bengio</surname> <given-names>Y</given-names></name></person-group>. <source>Deep Learning</source>. <publisher-loc>Cambridge, MA</publisher-loc>: <publisher-name>MIT press</publisher-name> (<year>2016</year>).</citation>
</ref>
<ref id="B24">
<label>24.</label>
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Ronneberger</surname> <given-names>O</given-names></name> <name><surname>Fischer</surname> <given-names>P</given-names></name> <name><surname>Brox</surname> <given-names>T</given-names></name></person-group>. <article-title>U-net: convolutional networks for biomedical image segmentation</article-title>. In: <source>International Conference on Medical Image Computing and Computer-Assisted Intervention</source>. <publisher-loc>Springer</publisher-loc> (<year>2015</year>). p. <fpage>234</fpage>&#x02013;<lpage>41</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-319-24574-4_28</pub-id></citation>
</ref>
<ref id="B25">
<label>25.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Radford</surname> <given-names>A</given-names></name> <name><surname>Metz</surname> <given-names>L</given-names></name> <name><surname>Chintala</surname> <given-names>S</given-names></name></person-group>. <article-title>Unsupervised representation learning with deep convolutional generative adversarial networks</article-title>. <source>arXiv preprint</source> arXiv:151106434. (<year>2015</year>).<pub-id pub-id-type="pmid">33873122</pub-id></citation></ref>
<ref id="B26">
<label>26.</label>
<citation citation-type="book"><person-group person-group-type="author"><name><surname>He</surname> <given-names>K</given-names></name> <name><surname>Zhang</surname> <given-names>X</given-names></name> <name><surname>Ren</surname> <given-names>S</given-names></name> <name><surname>Sun</surname> <given-names>J</given-names></name></person-group>. <article-title>Deep residual learning for image recognition</article-title>. In: <source>Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition</source>. <publisher-loc>Las Vegas, NV</publisher-loc>: <publisher-name>IEEE</publisher-name> (<year>2016</year>). p. <fpage>770</fpage>&#x02013;<lpage>8</lpage>.<pub-id pub-id-type="pmid">32166560</pub-id></citation></ref>
<ref id="B27">
<label>27.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ulyanov</surname> <given-names>D</given-names></name> <name><surname>Vedaldi</surname> <given-names>A</given-names></name> <name><surname>Lempitsky</surname> <given-names>V</given-names></name></person-group>. <article-title>Instance normalization: The missing ingredient for fast stylization</article-title>. <source>arXiv preprint</source> arXiv:160708022. (<year>2016</year>).</citation>
</ref>
<ref id="B28">
<label>28.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Maas</surname> <given-names>AL</given-names></name> <name><surname>Hannun</surname> <given-names>AY</given-names></name> <name><surname>Ng</surname> <given-names>AY</given-names></name></person-group>. <article-title>Rectifier nonlinearities improve neural network acoustic models</article-title>. In: <source>Proceedings of the 30th International Conference on Machine Learning, ICML 2013, Vol. 30</source>. Atlanta, GA (<year>2013</year>). p. <fpage>3</fpage>.</citation>
</ref>
<ref id="B29">
<label>29.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Shiraishi</surname> <given-names>J</given-names></name> <name><surname>Katsuragawa</surname> <given-names>S</given-names></name> <name><surname>Ikezoe</surname> <given-names>J</given-names></name> <name><surname>Matsumoto</surname> <given-names>T</given-names></name> <name><surname>Kobayashi</surname> <given-names>T</given-names></name> <name><surname>Komatsu</surname> <given-names>Ki</given-names></name> <etal/></person-group>. <article-title>Development of a digital image database for chest radiographs with and without a lung nodule: receiver operating characteristic analysis of radiologists&#x00027; detection of pulmonary nodules</article-title>. <source>Am J Roentgenol</source>. (<year>2000</year>) <volume>174</volume>:<fpage>71</fpage>&#x02013;<lpage>4</lpage>. <pub-id pub-id-type="doi">10.2214/ajr.174.1.1740071</pub-id><pub-id pub-id-type="pmid">10628457</pub-id></citation></ref>
<ref id="B30">
<label>30.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Candemir</surname> <given-names>S</given-names></name> <name><surname>Jaeger</surname> <given-names>S</given-names></name> <name><surname>Palaniappan</surname> <given-names>K</given-names></name> <name><surname>Musco</surname> <given-names>JP</given-names></name> <name><surname>Singh</surname> <given-names>RK</given-names></name> <name><surname>Xue</surname> <given-names>Z</given-names></name> <etal/></person-group>. <article-title>Lung segmentation in chest radiographs using anatomical atlases with nonrigid registration</article-title>. <source>IEEE Trans Med Imaging</source>. (<year>2013</year>) <volume>33</volume>:<fpage>577</fpage>&#x02013;<lpage>90</lpage>. <pub-id pub-id-type="doi">10.1109/TMI.2013.2290491</pub-id><pub-id pub-id-type="pmid">24239990</pub-id></citation></ref>
<ref id="B31">
<label>31.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jaeger</surname> <given-names>S</given-names></name> <name><surname>Karargyris</surname> <given-names>A</given-names></name> <name><surname>Candemir</surname> <given-names>S</given-names></name> <name><surname>Folio</surname> <given-names>L</given-names></name> <name><surname>Siegelman</surname> <given-names>J</given-names></name> <name><surname>Callaghan</surname> <given-names>F</given-names></name> <etal/></person-group>. <article-title>Automatic tuberculosis screening using chest radiographs</article-title>. <source>IEEE Trans Med Imaging</source>. (<year>2013</year>) <volume>33</volume>:<fpage>233</fpage>&#x02013;<lpage>45</lpage>. <pub-id pub-id-type="doi">10.1109/TMI.2013.2284099</pub-id><pub-id pub-id-type="pmid">29959539</pub-id></citation></ref>
<ref id="B32">
<label>32.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Van Ginneken</surname> <given-names>B</given-names></name> <name><surname>Stegmann</surname> <given-names>MB</given-names></name> <name><surname>Loog</surname> <given-names>M</given-names></name></person-group>. <article-title>Segmentation of anatomical structures in chest radiographs using supervised methods: a comparative study on a public database</article-title>. <source>Med Image Anal</source>. (<year>2006</year>) <volume>10</volume>:<fpage>19</fpage>&#x02013;<lpage>40</lpage>. <pub-id pub-id-type="doi">10.1016/j.media.2005.02.002</pub-id><pub-id pub-id-type="pmid">15919232</pub-id></citation></ref>
<ref id="B33">
<label>33.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Russakovsky</surname> <given-names>O</given-names></name> <name><surname>Deng</surname> <given-names>J</given-names></name> <name><surname>Su</surname> <given-names>H</given-names></name> <name><surname>Krause</surname> <given-names>J</given-names></name> <name><surname>Satheesh</surname> <given-names>S</given-names></name> <name><surname>Ma</surname> <given-names>S</given-names></name> <etal/></person-group>. <article-title>Imagenet large scale visual recognition challenge</article-title>. <source>Int J Comput Vis</source>. (<year>2015</year>) <volume>115</volume>:<fpage>211</fpage>&#x02013;<lpage>52</lpage>. <pub-id pub-id-type="doi">10.1007/s11263-015-0816-y</pub-id></citation>
</ref>
<ref id="B34">
<label>34.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kingma</surname> <given-names>DP</given-names></name> <name><surname>Ba</surname> <given-names>J</given-names></name></person-group>. <article-title>Adam: a method for stochastic optimization</article-title>. <source>arXiv preprint</source> arXiv:14126980. (<year>2014</year>).</citation>
</ref>
<ref id="B35">
<label>35.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>D</given-names></name> <name><surname>Yang</surname> <given-names>J</given-names></name> <name><surname>Kreis</surname> <given-names>K</given-names></name> <name><surname>Torralba</surname> <given-names>A</given-names></name> <name><surname>Fidler</surname> <given-names>S</given-names></name></person-group>. <article-title>Semantic segmentation with generative models: semi-supervised learning and strong out-of-domain generalization</article-title>. In: <source>Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition.</source> (<year>2021</year>) p. <fpage>8300</fpage>&#x02013;<lpage>11</lpage>.</citation>
</ref>
<ref id="B36">
<label>36.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bosdelekidis</surname> <given-names>V</given-names></name> <name><surname>Ioakeimidis</surname> <given-names>NS</given-names></name></person-group>. <article-title>Lung field segmentation in chest X-rays: a deformation-tolerant procedure based on the approximation of rib cage seed points</article-title>. <source>Appl Sci</source>. (<year>2020</year>) <volume>10</volume>:<fpage>6264</fpage>. <pub-id pub-id-type="doi">10.3390/app10186264</pub-id></citation>
</ref>
<ref id="B37">
<label>37.</label>
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Oliveira</surname> <given-names>H</given-names></name> <name><surname>dos Santos</surname> <given-names>J</given-names></name></person-group>. <article-title>Deep transfer learning for segmentation of anatomical structures in chest radiographs</article-title>. In: <source>2018 31st SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI)</source>. <publisher-loc>Parana</publisher-loc>: <publisher-name>IEEE</publisher-name> (<year>2018</year>). p. <fpage>204</fpage>&#x02013;<lpage>11</lpage>.</citation>
</ref>
<ref id="B38">
<label>38.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Souza</surname> <given-names>JC</given-names></name> <name><surname>Diniz</surname> <given-names>JOB</given-names></name> <name><surname>Ferreira</surname> <given-names>JL</given-names></name> <name><surname>da Silva</surname> <given-names>GLF</given-names></name> <name><surname>Silva</surname> <given-names>AC</given-names></name> <name><surname>de Paiva</surname> <given-names>AC</given-names></name></person-group>. <article-title>An automatic method for lung segmentation and reconstruction in chest X-ray using deep neural networks</article-title>. <source>Comput Methods Programs Biomed</source>. (<year>2019</year>) <volume>177</volume>:<fpage>285</fpage>&#x02013;<lpage>96</lpage>. <pub-id pub-id-type="doi">10.1016/j.cmpb.2019.06.005</pub-id><pub-id pub-id-type="pmid">31319957</pub-id></citation></ref>
<ref id="B39">
<label>39.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Filice</surname> <given-names>RW</given-names></name> <name><surname>Stein</surname> <given-names>A</given-names></name> <name><surname>Wu</surname> <given-names>CC</given-names></name> <name><surname>Arteaga</surname> <given-names>VA</given-names></name> <name><surname>Borstelmann</surname> <given-names>S</given-names></name> <name><surname>Gaddikeri</surname> <given-names>R</given-names></name> <etal/></person-group>. <article-title>Crowdsourcing pneumothorax annotations using machine learning annotations on the NIH chest X-ray dataset</article-title>. <source>J Digit Imaging</source>. (<year>2020</year>) <volume>33</volume>:<fpage>490</fpage>&#x02013;<lpage>6</lpage>. <pub-id pub-id-type="doi">10.1007/s10278-019-00299-9</pub-id><pub-id pub-id-type="pmid">31768897</pub-id></citation></ref>
</ref-list>
 
</back>
</article>