<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article article-type="research-article" dtd-version="2.3" xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Phys.</journal-id>
<journal-title>Frontiers in Physics</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Phys.</abbrev-journal-title>
<issn pub-type="epub">2296-424X</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">965095</article-id>
<article-id pub-id-type="doi">10.3389/fphy.2022.965095</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Physics</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Three-Dimensional Virtual Optical Clearing With Cycle-Consistent Generative Adversarial Network</article-title>
<alt-title alt-title-type="left-running-head">Chen et al.</alt-title>
<alt-title alt-title-type="right-running-head">Three-Dimensional Virtual Optical Clearing</alt-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name>
<surname>Chen</surname>
<given-names>Jiajia</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<xref ref-type="aff" rid="aff3">
<sup>3</sup>
</xref>
<xref ref-type="fn" rid="fn1">
<sup>&#x2020;</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/1765989/overview"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Du</surname>
<given-names>Zhenhong</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<xref ref-type="fn" rid="fn1">
<sup>&#x2020;</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/1872949/overview"/>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Si</surname>
<given-names>Ke</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<xref ref-type="aff" rid="aff3">
<sup>3</sup>
</xref>
<xref ref-type="aff" rid="aff4">
<sup>4</sup>
</xref>
<xref ref-type="corresp" rid="c001">&#x2a;</xref>
<uri xlink:href="https://loop.frontiersin.org/people/1185488/overview"/>
</contrib>
</contrib-group>
<aff id="aff1">
<sup>1</sup>
<institution>State Key Laboratory of Modern Optical Instrumentation</institution>, <institution>Department of Psychiatry</institution>, <institution>First Affiliated Hospital</institution>, <institution>Zhejiang University School of Medicine</institution>, <addr-line>Hangzhou</addr-line>, <country>China</country>
</aff>
<aff id="aff2">
<sup>2</sup>
<institution>College of Optical Science and Engineering</institution>, <institution>Zhejiang University</institution>, <addr-line>Hangzhou</addr-line>, <country>China</country>
</aff>
<aff id="aff3">
<sup>3</sup>
<institution>Intelligent Optics and Photonics Research Center</institution>, <institution>Jiaxing Research Institute</institution>, <institution>Zhejiang University</institution>, <addr-line>Jiaxing</addr-line>, <country>China</country>
</aff>
<aff id="aff4">
<sup>4</sup>
<institution>MOE Frontier Science Center for Brain Science and Brain-Machine Integration</institution>, <institution>NHC and CAMS Key Laboratory of Medical Neurobiology</institution>, <institution>School of Brain Science and Brain Medicine</institution>, <institution>Zhejiang University</institution>, <addr-line>Hangzhou</addr-line>, <country>China</country>
</aff>
<author-notes>
<fn fn-type="edited-by">
<p>
<bold>Edited by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/851359/overview">Liwei Liu</ext-link>, Shenzhen University, China</p>
</fn>
<fn fn-type="edited-by">
<p>
<bold>Reviewed by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/998698/overview">Ming Lei</ext-link>, Xi&#x2019;an Jiaotong University, China</p>
<p>
<ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/888006/overview">Sihua Yang</ext-link>, South China Normal University, China</p>
</fn>
<corresp id="c001">&#x2a;Correspondence: Ke Si, <email>kesi@zju.edu.cn</email>
</corresp>
<fn fn-type="equal" id="fn1">
<label>
<sup>&#x2020;</sup>
</label>
<p>These authors have contributed equally to this work</p>
</fn>
<fn fn-type="other">
<p>This article was submitted to Optics and Photonics, a section of the journal Frontiers in Physics</p>
</fn>
</author-notes>
<pub-date pub-type="epub">
<day>19</day>
<month>07</month>
<year>2022</year>
</pub-date>
<pub-date pub-type="collection">
<year>2022</year>
</pub-date>
<volume>10</volume>
<elocation-id>965095</elocation-id>
<history>
<date date-type="received">
<day>09</day>
<month>06</month>
<year>2022</year>
</date>
<date date-type="accepted">
<day>22</day>
<month>06</month>
<year>2022</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#xa9; 2022 Chen, Du and Si.</copyright-statement>
<copyright-year>2022</copyright-year>
<copyright-holder>Chen, Du and Si</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/">
<p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p>
</license>
</permissions>
<abstract>
<p>High-throughput deep tissue imaging and chemical tissue clearing protocols have brought out great promotion in biological research. However, due to uneven transparency introduced by tissue anisotropy in imperfectly cleared tissues, fluorescence imaging based on direct chemical tissue clearing still encounters great challenges, such as image blurring, low contrast, artifacts and so on. Here we reported a three-dimensional virtual optical clearing method based on unsupervised cycle-consistent generative adversarial network, termed 3D-VoCycleGAN, to digitally improve image quality and tissue transparency of biological samples. We demonstrated the good image deblurring and denoising capability of our method on imperfectly cleared mouse brain and kidney tissues. With 3D-VoCycleGAN prediction, the signal-to-background ratio (SBR) of images in imperfectly cleared brain tissue areas also showed above 40% improvement. Compared to other deconvolution methods, our method could evidently eliminate the tissue opaqueness and restore the image quality of the larger 3D images deep inside the imperfect cleared biological tissues with higher efficiency. And after virtually cleared, the transparency and clearing depth of mouse kidney tissues were increased by up to 30%. To our knowledge, it is the first interdisciplinary application of the CycleGAN deep learning model in the 3D fluorescence imaging and tissue clearing fields, promoting the development of high-throughput volumetric fluorescence imaging and deep learning techniques.</p>
</abstract>
<kwd-group>
<kwd>optical clearing</kwd>
<kwd>deep learning</kwd>
<kwd>deep tissue imaging</kwd>
<kwd>light-sheet</kwd>
<kwd>image processing</kwd>
</kwd-group>
</article-meta>
</front>
<body>
<sec id="s1">
<title>Introduction</title>
<p>Fluorescence microscopy has been playing an increasingly indispensable role in depiction of biological microstructures and functions. Up to now, confocal microscopy is still the most extensive and successful commercial fluorescence imaging system [<xref ref-type="bibr" rid="B1">1</xref>]. Nevertheless, the tissue anisotropy, the signal attenuation or absorption, the optical aberration of imaging system will all cause severe image blurring and degradation in the practical imaging process, limiting the further development of biological research at micro-scale [<xref ref-type="bibr" rid="B2">2</xref>]. On the one hand, the low fluorescence image quality greatly decreases the resolving power and further analysis accuracy of imaging systems for microstructural information. On the other hand, the reduction of fluorescence signal deep inside biological tissues will influence the imaging depth and imaging speed of thick tissues, restricting the experimental research efficiency for large-scale biological tissues. So far, many researchers have made great efforts to improve the imaging efficiency and image quality from different aspects, including physical, chemical, and digital ways [<xref ref-type="bibr" rid="B3">3</xref>&#x2013;<xref ref-type="bibr" rid="B5">5</xref>].</p>
<p>To acquire detailed 3D information physically, a series of advanced optical imaging techniques have been developed for deep tissue imaging in recent years, such as two-photon excitation microscopy (TPEM) [<xref ref-type="bibr" rid="B6">6</xref>], fluorescence micro-optical sectioning tomography (fMOST) [<xref ref-type="bibr" rid="B7">7</xref>], and light-sheet fluorescence microscopy (LSFM) [<xref ref-type="bibr" rid="B8">8</xref>]. Compared with the confocal microscopy, the two-photon absorption effect provides lower background signal level, phototoxicity and photobleaching for biological imaging. Besides, longer wavelength of laser used in TPEF could realize larger penetration depth for fluorescence excitation and detection, improving the 3D imaging capability of fluorescence microscopy. fMOST broke the 3D imaging limitations for brainwide mapping neurite level by using a continuous tissue sectioning microtome and synchronous wide-field detection. And the synchronous tissue sectioning and imaging idea could be also introduced into various conventional imaging systems for high-speed 3D imaging and reconstruction, including serial two-photon tomography [<xref ref-type="bibr" rid="B9">9</xref>], automatic serial sectioning polarization sensitive optical coherence tomography [<xref ref-type="bibr" rid="B10">10</xref>] and so on. Further, as a rapid, high-resolution imaging technique, LSFM has played an important part in large-scale mesoscopic biological research due to its large field of view and good optical sectioning capability. Especially for millimeter-level biological tissue imaging, LSFM has shown unprecedented imaging speed and throughput, which is at least dozens of times higher than some conventional fluorescence microscopies [<xref ref-type="bibr" rid="B11">11</xref>].</p>
<p>Although these microscopic imaging systems have made significant progress in depicting biological microstructures and functions, it is not sufficient for us to improve the image quality and imaging depth only by physical means. It is because the strong scattering and attenuation effect introduced by the biological tissue anisotropy will directly cause severe image degradation and noise, which could not easily be overcome or bypassed by upgrading the optical system. Hence, the chemical tissue clearing techniques were proposed to improve the tissue homogeneity and ensure refractive matching between tissues and surrounding buffers. Especially as a powerful combination with LSFM imaging techniques, various tissue clearing protocols have been developed and modified for larger imaging depth and better imaging quality in 3D tissue imaging [<xref ref-type="bibr" rid="B12">12</xref>&#x2013;<xref ref-type="bibr" rid="B14">14</xref>]. For example, CUBIC-series allows whole-brain even whole-body clearing and enables single-cell-resolution visualization and quantification of nucleus and neural activities in centimeter-scale brains [<xref ref-type="bibr" rid="B15">15</xref>]. And the previously unknown details and anatomical connections such as non-dividing stem cells near perisinusoidal areas under the fluorescence microscope could also be revealed by using DISCO-series protocols [<xref ref-type="bibr" rid="B16">16</xref>]. Besides, an ultrafast optical clearing method (FOCM) was also proposed to clarify 300-um-thick mouse tissue slices in 2&#xa0;min with low morphological deformation, fluorescent toxicity and easy operation [<xref ref-type="bibr" rid="B17">17</xref>]. Nevertheless, in spite of the great clearing effect on tissues of rodent animals, these clearing protocols have not perfectly resolved the compactness and refractoriness of brain tissues, especially the white matter. So far, as the most important partner of high-throughput imaging systems (e.g., LSFM), the development of chemical tissue clearing techniques is still booming for expanding the 3D tissue imaging depth and image quality.</p>
<p>Except for improvements of imaging systems and tissue preparation techniques, another popular approach is image deconvolution. An image restoration process, namely deconvolution, is established for enhancing the tissue details and image quality by typically modeling the image acquisition and degradation process as the summation of image noise and convolution between sample and systematic point spread function [<xref ref-type="bibr" rid="B18">18</xref>]. Many classic deconvolution methods, such as Richardson-Lucy deconvolution and Huygens deconvolution have shown great image enhancement performance for different requirements, including resolution improvement, image deblurring and noise suppression [<xref ref-type="bibr" rid="B19">19</xref>, <xref ref-type="bibr" rid="B20">20</xref>]. However, inevitably, tissue anisotropy and scattering generally lead to deficiency of some 3D information, getting in the way of accurate acquiring or estimation of systematic point spread function, which is very important for the deconvolution process. Particularly, the fast large-scale tissue imaging process is generally accompanied by unforeseeable uncertainty of point spread function distortion and image degradation. As a kind of emerging state-of-the-art technique, deep learning has gradually shown powerful efficiency and wide feasibility, especially in image super-resolution, image restoration and aberration correction [<xref ref-type="bibr" rid="B21">21</xref>&#x2013;<xref ref-type="bibr" rid="B23">23</xref>]. However, almost current deep learning-based image processing methods need a large number of exquisitely prepared paired datasets. Due to the hardware and experimental limitation, it is difficult even impossible to acquire enough high-quality ground-truth in paired datasets for some deep learning models, such as convolutional neural networks and U-Net [<xref ref-type="bibr" rid="B24">24</xref>&#x2013;<xref ref-type="bibr" rid="B26">26</xref>]. In recent years, a series of unsupervised deep learning models are proposed to realize feature transformation between two types of data with unpaired datasets [<xref ref-type="bibr" rid="B27">27</xref>&#x2013;<xref ref-type="bibr" rid="B29">29</xref>]. For example, CycleGAN model has been widely applied in two-dimensional (2D) medical image processing, which realized good efficiency comparable to supervised deep learning models [<xref ref-type="bibr" rid="B30">30</xref>&#x2013;<xref ref-type="bibr" rid="B32">32</xref>]. Nevertheless, CycleGAN was mainly used in non-fluorescent imaging and 2D image processing. The wide application and successful verification in 3D high-throughput fluorescence microscopy has hardly ever been reported before.</p>
<p>Here we report a three-dimensional virtual optical clearing method based on cycle-consistent generative adversarial network, termed 3D-VoCycleGAN, to improve the transparency of imperfect cleared biological tissues and image quality of LSFM images. First of all, we selected the blurred 3D image volumes and clear 3D image volumes from the raw 3D image data with varying transparency and contrast as datasets for further network training. Then we built a CycleGAN deep learning model with two 3D ResUNet-based generators and two 3D PatchGAN-based discriminators to realize fast prediction and transformation from blurred image volumes to clear image volumes. By testing the method on the 3D image data acquired from a custom-built LSFM system, we verified evident improvements with homogeneous digital tissue clearing and good image contrast on imperfect cleared mouse brain tissues and mouse kidney tissues. Besides, compared to other deconvolution methods, our virtual clearing method showed good image restoration and transparency enhancement effect with evident speed advantage, especially for 3D images deep inside biological tissues. To our knowledge, it is the first time that CycleGAN model has been used for enhancing the clearing effect of chemical tissue clearing, and restoring the 3D blurred LSFM images. Our virtual optical clearing method could effectively remedy the insufficiency of chemical tissue clearing and deep tissue imaging techniques, illustrating promising potential in future 3D histology and volumetric fluorescence imaging.</p>
</sec>
<sec sec-type="methods" id="s2">
<title>Methods</title>
<sec id="s2-1">
<title>The Main Framework of Virtual Optical Clearing</title>
<p>The tissue spatial anisotropy and imperfect tissue clearing effect generally leads to heterogeneous image contrast or image blurring, limiting further image biological structure identification and analysis. We proposed a CycleGAN-based approach to restore image quality with various tissue transparency and contrast in 3D LSFM imaging. The main framework of our virtual optical clearing method is shown in <xref ref-type="fig" rid="F1">Figure 1A</xref>. According to different types of tissue properties and clearing effect, we selected several representative 3D image volumes with specific structural information from the raw LSFM data to generate datasets for network training. In the tissues, 3D volumes with high image contrast, low noise and good transparency were regarded as well-cleared tissue areas. And the cropped 3D image data from the well-cleared tissue areas was performed 3D deconvolution to further suppress background signal out of focus and improve image contrast for generating the target domain image data. At the same time, 3D image volumes with severe blurring and background noise in imperfectly cleared area was selected as the source domain image data. It is noticed that unlike the supervised deep learning-based methods, we need not realize accurate data pre-alignment between the two domains in our method. Hence, various types of biological samples could be used to build datasets for our experiments, including mouse brain tissues and kidney tissues acquired by using our custom-built LSFM system.</p>
<fig id="F1" position="float">
<label>FIGURE 1</label>
<caption>
<p>Framework of virtual optical clearing method and 3D-VoCycleGAN architecture. <bold>(A)</bold> The framework and workflow of our virtual optical clearing method. The 3D-VoCycleGAN consists of two generators and two discriminations <bold>(B)</bold> The network structure of generators. 3D ResUNet was used to build the two generators in our network <bold>(C)</bold> The network structure of discriminators. 3D PatchGAN was used to build the two discriminators in our network.</p>
</caption>
<graphic xlink:href="fphy-10-965095-g001.tif"/>
</fig>
<p>In the specific network training process, we defined the source domain and the target domain as domain X and domain Y, respectively. Taking the GPU memory requirements of network training into account, 200 groups of unpaired 3D image volumes with 256 &#xd7; 256&#xd7;16 pixel<sup>3</sup> size were randomly cropped from the domain X and domain Y to generate the training datasets. In order to improve the training efficiency and avoid overfitting, image volumes with insufficient foreground were automatically discarded before training [<xref ref-type="bibr" rid="B33">33</xref>]. And these image volumes were normalized and then fed into the network for training. To ensure higher generalization capability of our network, during each training process, we performed data augmentation on the input image volumes by introducing a series of random changes, such as flipping, rotation, and so on. As the loss function iteratively minimized, the non-transparent 3D tissue image identified and learned the clearer structural features from the cleared 3D tissue image. After the deep learning network was well optimized, the original large LSFM image data with varying spatial transparency and contrast was sent to the network for implementing fast image prediction and image quality restoration.</p>
</sec>
<sec id="s2-2">
<title>3D-VoCycleGAN Architecture</title>
<p>In our 3D-VoCycleGAN, we built the two generators by using ResUNet, which has been proved to have great biological feature extraction capability [<xref ref-type="bibr" rid="B34">34</xref>]. The ResUNet used in our generators is a kind of encoder-decoder cascade structure, which is composed of four downsampling blocks and upsampling blocks (<xref ref-type="fig" rid="F1">Figure 1B</xref>). In the encoder path, each downsampling block contained a max pooling and two convolution layers, each layer of which comprised 3 &#xd7; 3&#xd7;3 kernel followed by a batch-normalization layer and a ReLU activation function. And the encoder encoded the input image stack into multiple feature representations at different levels through max pooling. Then the decoder consisting of four upsampling blocks symmetrical to the encoder was used to decode data information back to the original dimension. Each upsampling block consisted of trilinear interpolation, skip connection, and convolution layers. The skip connection concatenated the high-level features and spatial information between encoder and decoder, thus retaining more details and capturing finer information. Finally, we established a residual connection between the input and the output of the decoder to avoid the gradient vanishing problem and improve the performance of the network [<xref ref-type="bibr" rid="B35">35</xref>].</p>
<p>For building the discriminators, we modified the five-layer conventional structure of PatchGAN [<xref ref-type="bibr" rid="B31">31</xref>] into a 3D form (<xref ref-type="fig" rid="F1">Figure 1C</xref>). In the PatchGAN, conventional operation with a stride of 2 and padding of 1 was used in the first three conventional layers, each layer of which was followed by an instance normalization layer, and a LeakyReLU activation function. And the channel numbers were doubled as the image resolution was halved in the first three conventional each time. And the channel numbers of the five convolutional layers were 64, 128, 256, 512 and 1, respectively. Considering that the size of the input stack was 256 &#xd7; 256&#xd7;16, we set the kernel size of the last two convolution layers to 3 &#xd7; 3&#xd7;3, making the size of the output patch 32 &#xd7; 32&#xd7;2. Finally, the last convolutional layer reduced the channel numbers of the feature map to 1, and a sigmoid activation function was used to normalize the output value into [0, 1].</p>
<p>Our 3D-VoCycleGAN was implemented based on the Pytorch deep learning framework. The learning rate of the Adam optimizer was set to 0.0002 in the first 100 epochs and linearly decayed to 0 in the next 100 epochs. The batch size was set to 1. The overall training and prediction processes based on our method were implemented in a Dell 7,920 workstation equipped with RTX 3090 GPU (24&#xa0;GB memory).</p>
</sec>
<sec id="s2-3">
<title>Loss Function</title>
<p>The deep learning model of 3D-VoCycleGAN contained two generators <inline-formula id="inf1">
<mml:math id="m1">
<mml:mrow>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mrow>
<mml:mi>X</mml:mi>
<mml:mi>Y</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> (from domain X to Y), <inline-formula id="inf2">
<mml:math id="m2">
<mml:mrow>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mrow>
<mml:mi>Y</mml:mi>
<mml:mi>X</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> (from domain Y to X) and two discriminators <inline-formula id="inf3">
<mml:math id="m3">
<mml:mrow>
<mml:msub>
<mml:mi>D</mml:mi>
<mml:mi>X</mml:mi>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mtext>&#x2009;</mml:mtext>
<mml:msub>
<mml:mi>D</mml:mi>
<mml:mi>Y</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>. The two generators <inline-formula id="inf4">
<mml:math id="m4">
<mml:mrow>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mrow>
<mml:mi>X</mml:mi>
<mml:mi>Y</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf5">
<mml:math id="m5">
<mml:mrow>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mrow>
<mml:mi>Y</mml:mi>
<mml:mi>X</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> aimed to realize transformation from 3D image volumes of domain X to 3D image volumes of domain Y and inverse image transformation from domain Y to domain X, respectively. In each domain, an image volume could be the output of the generator or come from the original training data. Hence, two discriminators were used to judge which situation the image volumes belong to. In a word, the goal of the generator was synthesizing images and fooling the discriminator, while the discriminator tried to accurately winnow truth from falsehood. The standard GAN loss [<xref ref-type="bibr" rid="B36">36</xref>] was given as follows:<disp-formula id="e1">
<mml:math id="m6">
<mml:mrow>
<mml:msub>
<mml:mi>l</mml:mi>
<mml:mrow>
<mml:mi>G</mml:mi>
<mml:mi>A</mml:mi>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mrow>
<mml:mi>X</mml:mi>
<mml:mi>Y</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mi mathvariant="normal">&#x45;</mml:mi>
<mml:mrow>
<mml:mi>y</mml:mi>
<mml:mo>&#x223c;</mml:mo>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mrow>
<mml:mi mathvariant="normal">d</mml:mi>
<mml:mi mathvariant="normal">a</mml:mi>
<mml:mi mathvariant="normal">t</mml:mi>
<mml:mi mathvariant="normal">a</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mrow>
<mml:mi>log</mml:mi>
<mml:mo>&#x2061;</mml:mo>
<mml:msub>
<mml:mi>D</mml:mi>
<mml:mi>Y</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>]</mml:mo>
</mml:mrow>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mtext mathvariant="normal">&#x45;</mml:mtext>
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>&#x223c;</mml:mo>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mrow>
<mml:mi mathvariant="normal">d</mml:mi>
<mml:mi mathvariant="normal">a</mml:mi>
<mml:mi mathvariant="normal">t</mml:mi>
<mml:mi mathvariant="normal">a</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mrow>
<mml:mo>[</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>log</mml:mi>
<mml:mo>&#x2061;</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
</mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mi>D</mml:mi>
<mml:mi>Y</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mrow>
<mml:mi>X</mml:mi>
<mml:mi>Y</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>]</mml:mo>
</mml:mrow>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(1)</label>
</disp-formula>
<disp-formula id="e2">
<mml:math id="m7">
<mml:mrow>
<mml:msub>
<mml:mi>l</mml:mi>
<mml:mrow>
<mml:mi>G</mml:mi>
<mml:mi>A</mml:mi>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mrow>
<mml:mi>Y</mml:mi>
<mml:mi>X</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mtext mathvariant="normal">&#x45;</mml:mtext>
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>&#x223c;</mml:mo>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mrow>
<mml:mi mathvariant="normal">d</mml:mi>
<mml:mi mathvariant="normal">a</mml:mi>
<mml:mi mathvariant="normal">t</mml:mi>
<mml:mi mathvariant="normal">a</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mrow>
<mml:mi>log</mml:mi>
<mml:mo>&#x2061;</mml:mo>
<mml:msub>
<mml:mi>D</mml:mi>
<mml:mi>X</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>]</mml:mo>
</mml:mrow>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mtext mathvariant="normal">&#x45;</mml:mtext>
<mml:mrow>
<mml:mi>y</mml:mi>
<mml:mo>&#x223c;</mml:mo>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mrow>
<mml:mi mathvariant="normal">d</mml:mi>
<mml:mi mathvariant="normal">a</mml:mi>
<mml:mi mathvariant="normal">t</mml:mi>
<mml:mi mathvariant="normal">a</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mrow>
<mml:mo>[</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>log</mml:mi>
<mml:mo>&#x2061;</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
</mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mi>D</mml:mi>
<mml:mi>X</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mrow>
<mml:mi>Y</mml:mi>
<mml:mi>X</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>]</mml:mo>
</mml:mrow>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(2)</label>
</disp-formula>where <italic>x</italic> denotes the blurred stack in domain X, <italic>y</italic> denotes the clear stack in domain Y. <inline-formula id="inf6">
<mml:math id="m8">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="double-struck">E</mml:mi>
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>&#x223c;</mml:mo>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mrow>
<mml:mi mathvariant="normal">d</mml:mi>
<mml:mi mathvariant="normal">a</mml:mi>
<mml:mi mathvariant="normal">t</mml:mi>
<mml:mi mathvariant="normal">a</mml:mi>
<mml:mtext>&#xa0;</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf7">
<mml:math id="m9">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="double-struck">E</mml:mi>
<mml:mrow>
<mml:mi>y</mml:mi>
<mml:mo>&#x223c;</mml:mo>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mrow>
<mml:mi mathvariant="normal">d</mml:mi>
<mml:mi mathvariant="normal">a</mml:mi>
<mml:mi mathvariant="normal">t</mml:mi>
<mml:mi mathvariant="normal">a</mml:mi>
<mml:mtext>&#xa0;</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> were the expectation operators. However, standard GAN loss usually suffers from mode collapse and vanishing gradients, which severely degrades network performance and stability. Here we used LS-GAN loss [<xref ref-type="bibr" rid="B37">37</xref>] as objective function:<disp-formula id="e3">
<mml:math id="m10">
<mml:mrow>
<mml:msub>
<mml:mi>l</mml:mi>
<mml:mrow>
<mml:mi>G</mml:mi>
<mml:mi>A</mml:mi>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mrow>
<mml:mi>X</mml:mi>
<mml:mi>Y</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:mfrac>
<mml:msub>
<mml:mtext mathvariant="normal">&#x45;</mml:mtext>
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>&#x223c;</mml:mo>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mrow>
<mml:mi>d</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>a</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
</mml:mrow>
<mml:msub>
<mml:mi>D</mml:mi>
<mml:mi>Y</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
</mml:mrow>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mrow>
<mml:mi>X</mml:mi>
<mml:mi>Y</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
</mml:mrow>
<mml:mi>x</mml:mi>
<mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
<mml:msup>
<mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mn>2</mml:mn>
</mml:msup>
</mml:mrow>
<mml:mo>]</mml:mo>
</mml:mrow>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(3)</label>
</disp-formula>
<disp-formula id="e4">
<mml:math id="m11">
<mml:mrow>
<mml:msub>
<mml:mi>l</mml:mi>
<mml:mrow>
<mml:mi>G</mml:mi>
<mml:mi>A</mml:mi>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mi>D</mml:mi>
<mml:mi>Y</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:mfrac>
<mml:msub>
<mml:mtext mathvariant="normal">&#x45;</mml:mtext>
<mml:mrow>
<mml:mi>y</mml:mi>
<mml:mo>&#x223c;</mml:mo>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mrow>
<mml:mi mathvariant="normal">d</mml:mi>
<mml:mi mathvariant="normal">a</mml:mi>
<mml:mi mathvariant="normal">t</mml:mi>
<mml:mi mathvariant="normal">a</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
</mml:mrow>
<mml:msub>
<mml:mi>D</mml:mi>
<mml:mi>Y</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
</mml:mrow>
<mml:mi>y</mml:mi>
<mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
<mml:msup>
<mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mn>2</mml:mn>
</mml:msup>
</mml:mrow>
<mml:mo>]</mml:mo>
</mml:mrow>
<mml:mo>&#x2b;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:mfrac>
<mml:msub>
<mml:mtext mathvariant="normal">&#x45;</mml:mtext>
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>&#x223c;</mml:mo>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mrow>
<mml:mi>d</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>a</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
</mml:mrow>
<mml:msub>
<mml:mi>D</mml:mi>
<mml:mi>Y</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
</mml:mrow>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mrow>
<mml:mi>X</mml:mi>
<mml:mi>Y</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
</mml:mrow>
<mml:mi>x</mml:mi>
<mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mn>2</mml:mn>
</mml:msup>
</mml:mrow>
<mml:mo>]</mml:mo>
</mml:mrow>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(4)</label>
</disp-formula>
<disp-formula id="e5">
<mml:math id="m12">
<mml:mrow>
<mml:msub>
<mml:mi>l</mml:mi>
<mml:mrow>
<mml:mi>G</mml:mi>
<mml:mi>A</mml:mi>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mrow>
<mml:mi>Y</mml:mi>
<mml:mi>X</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:mfrac>
<mml:msub>
<mml:mtext mathvariant="normal">&#x45;</mml:mtext>
<mml:mrow>
<mml:mi>y</mml:mi>
<mml:mo>&#x223c;</mml:mo>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mrow>
<mml:mi>d</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>a</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
</mml:mrow>
<mml:msub>
<mml:mi>D</mml:mi>
<mml:mi>X</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
</mml:mrow>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mrow>
<mml:mi>Y</mml:mi>
<mml:mi>X</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
</mml:mrow>
<mml:mi>y</mml:mi>
<mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
<mml:msup>
<mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mn>2</mml:mn>
</mml:msup>
</mml:mrow>
<mml:mo>]</mml:mo>
</mml:mrow>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(5)</label>
</disp-formula>
<disp-formula id="e6">
<mml:math id="m13">
<mml:mrow>
<mml:msub>
<mml:mi>l</mml:mi>
<mml:mrow>
<mml:mi>G</mml:mi>
<mml:mi>A</mml:mi>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mi>D</mml:mi>
<mml:mi>X</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:mfrac>
<mml:msub>
<mml:mtext mathvariant="normal">&#x45;</mml:mtext>
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>&#x223c;</mml:mo>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mrow>
<mml:mi mathvariant="normal">d</mml:mi>
<mml:mi mathvariant="normal">a</mml:mi>
<mml:mi mathvariant="normal">t</mml:mi>
<mml:mi mathvariant="normal">a</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
</mml:mrow>
<mml:msub>
<mml:mi>D</mml:mi>
<mml:mi>X</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
</mml:mrow>
<mml:mi>x</mml:mi>
<mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
<mml:msup>
<mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mn>2</mml:mn>
</mml:msup>
</mml:mrow>
<mml:mo>]</mml:mo>
</mml:mrow>
<mml:mo>&#x2b;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:mfrac>
<mml:msub>
<mml:mtext mathvariant="normal">&#x45;</mml:mtext>
<mml:mrow>
<mml:mi>y</mml:mi>
<mml:mo>&#x223c;</mml:mo>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mrow>
<mml:mi>d</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>a</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
</mml:mrow>
<mml:msub>
<mml:mi>D</mml:mi>
<mml:mi>X</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
</mml:mrow>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mrow>
<mml:mi>Y</mml:mi>
<mml:mi>X</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
</mml:mrow>
<mml:mi>y</mml:mi>
<mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mn>2</mml:mn>
</mml:msup>
</mml:mrow>
<mml:mo>]</mml:mo>
</mml:mrow>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(6)</label>
</disp-formula>
</p>
<p>The original cycle consistency loss calculated the L<sub>1</sub> loss between the original image and reconstructed image. We incorporated a 3D multi-scale structural similarity index metric (MS-SSIM) to construct a detail preserving transformation [<xref ref-type="bibr" rid="B38">38</xref>]. The cycle consistency loss was defined as follows:<disp-formula id="e7">
<mml:math id="m14">
<mml:mrow>
<mml:mtable columnalign="left">
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:msub>
<mml:mi>l</mml:mi>
<mml:mrow>
<mml:mi mathvariant="normal">c</mml:mi>
<mml:mi mathvariant="normal">y</mml:mi>
<mml:mi mathvariant="normal">c</mml:mi>
<mml:mi>l</mml:mi>
<mml:mi mathvariant="normal">e</mml:mi>
<mml:mtext>&#xa0;</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mrow>
<mml:mi>X</mml:mi>
<mml:mi>Y</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mrow>
<mml:mi>Y</mml:mi>
<mml:mi>X</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mtext mathvariant="normal">&#x45;</mml:mtext>
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>&#x223c;</mml:mo>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mrow>
<mml:mi mathvariant="normal">d</mml:mi>
<mml:mi mathvariant="normal">a</mml:mi>
<mml:mi mathvariant="normal">t</mml:mi>
<mml:mi mathvariant="normal">a</mml:mi>
<mml:mtext>&#xa0;</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mo>&#x2016;</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mrow>
<mml:mi>Y</mml:mi>
<mml:mi>X</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mrow>
<mml:mi>X</mml:mi>
<mml:mi>Y</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mo>&#x2016;</mml:mo>
</mml:mrow>
<mml:mn>1</mml:mn>
</mml:msub>
</mml:mrow>
<mml:mo>]</mml:mo>
</mml:mrow>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mtext mathvariant="normal">&#x45;</mml:mtext>
<mml:mrow>
<mml:mi>y</mml:mi>
<mml:mo>&#x223c;</mml:mo>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mrow>
<mml:mi mathvariant="normal">d</mml:mi>
<mml:mi mathvariant="normal">a</mml:mi>
<mml:mi mathvariant="normal">t</mml:mi>
<mml:mi mathvariant="normal">a</mml:mi>
<mml:mtext>&#xa0;</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mo>&#x2016;</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mrow>
<mml:mi>X</mml:mi>
<mml:mi>Y</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mrow>
<mml:mi>Y</mml:mi>
<mml:mi>X</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mo>&#x2016;</mml:mo>
</mml:mrow>
<mml:mn>1</mml:mn>
</mml:msub>
</mml:mrow>
<mml:mo>]</mml:mo>
</mml:mrow>
<mml:mo>&#x2b;</mml:mo>
<mml:mi>&#x3b7;</mml:mi>
<mml:mo>&#x22c5;</mml:mo>
<mml:msub>
<mml:mi>l</mml:mi>
<mml:mrow>
<mml:mi>M</mml:mi>
<mml:mi>S</mml:mi>
<mml:mo>&#x2010;</mml:mo>
<mml:mi>S</mml:mi>
<mml:mi>S</mml:mi>
<mml:mi>I</mml:mi>
<mml:mi>M</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mtd>
</mml:mtr>
</mml:mtable>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(7)</label>
</disp-formula>
<disp-formula id="e8">
<mml:math id="m15">
<mml:mrow>
<mml:mtable columnalign="left">
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:msub>
<mml:mi>l</mml:mi>
<mml:mrow>
<mml:mi>M</mml:mi>
<mml:mi>S</mml:mi>
<mml:mo>&#x2010;</mml:mo>
<mml:mi>S</mml:mi>
<mml:mi>S</mml:mi>
<mml:mi>I</mml:mi>
<mml:mi>M</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mrow>
<mml:mi>X</mml:mi>
<mml:mi>Y</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mrow>
<mml:mi>Y</mml:mi>
<mml:mi>X</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mi mathvariant="normal">M</mml:mi>
<mml:mi mathvariant="normal">S</mml:mi>
<mml:mo>&#x2010;</mml:mo>
<mml:mi mathvariant="normal">S</mml:mi>
<mml:mi mathvariant="normal">S</mml:mi>
<mml:mi mathvariant="normal">I</mml:mi>
<mml:mi mathvariant="normal">M</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mrow>
<mml:mi>Y</mml:mi>
<mml:mi>X</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mrow>
<mml:mi>X</mml:mi>
<mml:mi>Y</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mi mathvariant="normal">M</mml:mi>
<mml:mi mathvariant="normal">S</mml:mi>
<mml:mo>&#x2010;</mml:mo>
<mml:mi mathvariant="normal">S</mml:mi>
<mml:mi mathvariant="normal">S</mml:mi>
<mml:mi mathvariant="normal">I</mml:mi>
<mml:mi mathvariant="normal">M</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>y</mml:mi>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mrow>
<mml:mi>X</mml:mi>
<mml:mi>Y</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mrow>
<mml:mi>Y</mml:mi>
<mml:mi>X</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mtd>
</mml:mtr>
</mml:mtable>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(8)</label>
</disp-formula>where <inline-formula id="inf8">
<mml:math id="m16">
<mml:mrow>
<mml:mi>&#x3b7;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> represents the weight factor for the <inline-formula id="inf9">
<mml:math id="m17">
<mml:mrow>
<mml:msub>
<mml:mi>l</mml:mi>
<mml:mrow>
<mml:mi>M</mml:mi>
<mml:mi>S</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>S</mml:mi>
<mml:mi>S</mml:mi>
<mml:mi>I</mml:mi>
<mml:mi>M</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> term. And it was set to 0.2 in all experiments. An additional identity loss was used to constrain the generator to build an identity mapping when well-cleared volumes of the target domain were provided.<disp-formula id="e9">
<mml:math id="m18">
<mml:mrow>
<mml:msub>
<mml:mi>l</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>d</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>n</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>y</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mrow>
<mml:mi>X</mml:mi>
<mml:mi>Y</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mrow>
<mml:mi>Y</mml:mi>
<mml:mi>X</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mtext mathvariant="normal">&#x45;</mml:mtext>
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>&#x223c;</mml:mo>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mrow>
<mml:mi mathvariant="normal">d</mml:mi>
<mml:mi mathvariant="normal">a</mml:mi>
<mml:mi mathvariant="normal">t</mml:mi>
<mml:mi mathvariant="normal">a</mml:mi>
<mml:mtext>&#xa0;</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mo>&#x2016;</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mrow>
<mml:mi>Y</mml:mi>
<mml:mi>X</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mo>&#x2016;</mml:mo>
</mml:mrow>
<mml:mn>1</mml:mn>
</mml:msub>
</mml:mrow>
<mml:mo>]</mml:mo>
</mml:mrow>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mtext mathvariant="normal">&#x45;</mml:mtext>
<mml:mrow>
<mml:mi>y</mml:mi>
<mml:mo>&#x223c;</mml:mo>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mrow>
<mml:mi mathvariant="normal">d</mml:mi>
<mml:mi mathvariant="normal">a</mml:mi>
<mml:mi mathvariant="normal">t</mml:mi>
<mml:mi mathvariant="normal">a</mml:mi>
<mml:mtext>&#xa0;</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mo>&#x2016;</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mrow>
<mml:mi>X</mml:mi>
<mml:mi>Y</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mo>&#x2016;</mml:mo>
</mml:mrow>
<mml:mn>1</mml:mn>
</mml:msub>
</mml:mrow>
<mml:mo>]</mml:mo>
</mml:mrow>
<mml:mtext>&#xa0;</mml:mtext>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(9)</label>
</disp-formula>
</p>
<p>Saliency constraint has been proved to be effective for content preservation on 2D microscopic images [<xref ref-type="bibr" rid="B39">39</xref>]. And we added it to our 3D-VoCycleGAN for maintaining fine structures and information. By appropriately setting the threshold value, we could separate the foreground with detailed structure from the 3D samples. Therefore, the two generators focused on the essential structures of foreground at the early training stage and avoided appearance of artifacts in background with the aid of saliency constraint. The saliency constraint loss can be written as:<disp-formula id="e10">
<mml:math id="m19">
<mml:mrow>
<mml:mtable columnalign="left">
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:mtext>&#x2003;</mml:mtext>
<mml:msub>
<mml:mi>l</mml:mi>
<mml:mrow>
<mml:mi>s</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>l</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>n</mml:mi>
<mml:mi>c</mml:mi>
<mml:mi>y</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mrow>
<mml:mi>X</mml:mi>
<mml:mi>Y</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mrow>
<mml:mi>Y</mml:mi>
<mml:mi>X</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mtext mathvariant="normal">&#x45;</mml:mtext>
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>&#x223c;</mml:mo>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mrow>
<mml:mi mathvariant="normal">d</mml:mi>
<mml:mi mathvariant="normal">a</mml:mi>
<mml:mi mathvariant="normal">t</mml:mi>
<mml:mi mathvariant="normal">a</mml:mi>
<mml:mtext>&#xa0;</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mo>&#x2016;</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="normal">&#x54;</mml:mi>
<mml:mi>&#x3b1;</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mi mathvariant="normal">&#x54;</mml:mi>
<mml:mi>&#x3b2;</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mrow>
<mml:mi>X</mml:mi>
<mml:mi>Y</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>&#x2016;</mml:mo>
</mml:mrow>
<mml:mn>1</mml:mn>
</mml:msub>
</mml:mrow>
<mml:mo>]</mml:mo>
</mml:mrow>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mtext mathvariant="normal">&#x45;</mml:mtext>
<mml:mrow>
<mml:mi>y</mml:mi>
<mml:mo>&#x223c;</mml:mo>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mrow>
<mml:mi mathvariant="normal">d</mml:mi>
<mml:mi mathvariant="normal">a</mml:mi>
<mml:mi mathvariant="normal">t</mml:mi>
<mml:mi mathvariant="normal">a</mml:mi>
<mml:mtext>&#xa0;</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mo>&#x2016;</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="normal">&#x54;</mml:mi>
<mml:mi>&#x3b2;</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mi mathvariant="normal">&#x54;</mml:mi>
<mml:mi>&#x3b1;</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mrow>
<mml:mi>Y</mml:mi>
<mml:mi>X</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>&#x2016;</mml:mo>
</mml:mrow>
<mml:mn>1</mml:mn>
</mml:msub>
</mml:mrow>
<mml:mo>]</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mtd>
</mml:mtr>
</mml:mtable>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(10)</label>
</disp-formula>where <inline-formula id="inf10">
<mml:math id="m20">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="normal">&#x54;</mml:mi>
<mml:mi>&#x3b1;</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mi mathvariant="normal">s</mml:mi>
<mml:mi mathvariant="normal">i</mml:mi>
<mml:mi mathvariant="normal">g</mml:mi>
<mml:mi mathvariant="normal">m</mml:mi>
<mml:mi mathvariant="normal">o</mml:mi>
<mml:mi mathvariant="normal">i</mml:mi>
<mml:mi mathvariant="normal">d</mml:mi>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mrow>
<mml:mn>100</mml:mn>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>&#x3b1;</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>]</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf11">
<mml:math id="m21">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="normal">&#x54;</mml:mi>
<mml:mi>&#x3b2;</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mi mathvariant="normal">s</mml:mi>
<mml:mi mathvariant="normal">i</mml:mi>
<mml:mi mathvariant="normal">g</mml:mi>
<mml:mi mathvariant="normal">m</mml:mi>
<mml:mi mathvariant="normal">o</mml:mi>
<mml:mi mathvariant="normal">i</mml:mi>
<mml:mi mathvariant="normal">d</mml:mi>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mrow>
<mml:mn>100</mml:mn>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>&#x3b2;</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>]</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> are binary functions parameterized by threshold &#x3b1; and &#x3b2;.</p>
<p>The total loss for generators could be expressed by:<disp-formula id="e11">
<mml:math id="m22">
<mml:mrow>
<mml:mtable columnalign="left">
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:msub>
<mml:mi>L</mml:mi>
<mml:mrow>
<mml:mi mathvariant="normal">G</mml:mi>
<mml:mi mathvariant="normal">e</mml:mi>
<mml:mi mathvariant="normal">n</mml:mi>
<mml:mi mathvariant="normal">e</mml:mi>
<mml:mi mathvariant="normal">r</mml:mi>
<mml:mi mathvariant="normal">a</mml:mi>
<mml:mi mathvariant="normal">t</mml:mi>
<mml:mi mathvariant="normal">o</mml:mi>
<mml:mi mathvariant="normal">r</mml:mi>
<mml:mtext>&#xa0;</mml:mtext>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mi>l</mml:mi>
<mml:mrow>
<mml:mi>G</mml:mi>
<mml:mi>A</mml:mi>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mrow>
<mml:mi>X</mml:mi>
<mml:mi>Y</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mi>l</mml:mi>
<mml:mrow>
<mml:mi>G</mml:mi>
<mml:mi>A</mml:mi>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mrow>
<mml:mi>Y</mml:mi>
<mml:mi>X</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x2b;</mml:mo>
<mml:mi>&#x3bb;</mml:mi>
<mml:mo>&#x22c5;</mml:mo>
<mml:msub>
<mml:mi>l</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>d</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>n</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>y</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mrow>
<mml:mi>X</mml:mi>
<mml:mi>Y</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mrow>
<mml:mi>Y</mml:mi>
<mml:mi>X</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x2b;</mml:mo>
<mml:mi>&#x3c1;</mml:mi>
<mml:mo>&#x22c5;</mml:mo>
<mml:msub>
<mml:mi>l</mml:mi>
<mml:mrow>
<mml:mi>s</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>l</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>n</mml:mi>
<mml:mi>c</mml:mi>
<mml:mi>y</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mrow>
<mml:mi>X</mml:mi>
<mml:mi>Y</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mrow>
<mml:mi>Y</mml:mi>
<mml:mi>X</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x2b;</mml:mo>
<mml:mi>&#x3bc;</mml:mi>
<mml:mo>&#x22c5;</mml:mo>
<mml:msub>
<mml:mi>l</mml:mi>
<mml:mrow>
<mml:mi mathvariant="normal">c</mml:mi>
<mml:mi mathvariant="normal">y</mml:mi>
<mml:mi mathvariant="normal">c</mml:mi>
<mml:mi>l</mml:mi>
<mml:mi mathvariant="normal">e</mml:mi>
<mml:mtext>&#xa0;</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mrow>
<mml:mi>X</mml:mi>
<mml:mi>Y</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mrow>
<mml:mi>Y</mml:mi>
<mml:mi>X</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mtd>
</mml:mtr>
</mml:mtable>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(11)</label>
</disp-formula>where the parameters <inline-formula id="inf12">
<mml:math id="m23">
<mml:mrow>
<mml:mi>&#x3bb;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf13">
<mml:math id="m24">
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf14">
<mml:math id="m25">
<mml:mrow>
<mml:mi>&#x3c1;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> denotes the weight factor for the identity loss, saliency constraint loss and cycle loss, respectively. We assigned <inline-formula id="inf15">
<mml:math id="m26">
<mml:mrow>
<mml:mi>&#x3bb;</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn mathvariant="italic">5</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf16">
<mml:math id="m27">
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn mathvariant="italic">10</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, respectively. And we set <inline-formula id="inf17">
<mml:math id="m28">
<mml:mrow>
<mml:mi>&#x3c1;</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>10</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:msup>
<mml:mi>e</mml:mi>
<mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>0.1</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mi mathvariant="normal">n</mml:mi>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula>, where n denotes training epoch number, making the network pay more attention to the transformation of important details and ignoring the interference from the background at the early stage of training.</p>
</sec>
<sec id="s2-4">
<title>Sample Preparation and 3D Image Data Acquisition</title>
<p>The sample used in this study contained mouse brain and mouse kidney tissues. For verifying the performance of our method on tissue slices, we prepared Thy1-GFP mouse brains and mouse kidney tissues, which were sectioned into 300-&#x3bc;m and 200-&#x3bc;m thick slices, respectively. The mouse kidney tissue slices were stained with DRAQ5 for labelling the cell nucleus before cleared by FOCM reagents. The FOCM reagents were prepared as 30% (wt/vol) urea (Vetec), 20% (wt/vol) D-sorbitol (Vetec), and 5% (wt/vol) glycerol dissolved in DMSO. When preparing the reagent, urea and D-sorbitol were dissolved in DMSO and stirred at room temperature overnight. After complete dissolution, glycerol was added and stirred further. The reagents should be stored at room temperature and shaken gently before using. Before imaging experiments, the well-stained mouse brain and kidney tissue slices were incubated in FOCM reagents for several minutes.</p>
<p>For verifying the performance of our method on deep tissues with millimeter-thickness, we prepared stereoscopic mouse brain and kidney tissue blocks labelled with Alexa Fluor 647 anti-mouse CD31 antibody (CD31-AF647, BioLegend) by caudal vein injection. The Alexa Fluor 647 anti-mouse CD31 antibody (20&#xa0;mg) was then diluted in sterile saline (total volume of 150&#xa0;ml). After the injection, mice were placed in a warm cage for 30&#xa0;min prior to perfusion. Then mice were rapidly anesthetized with chloral hydrate [5%, wt/vol, 0.1 ml/10&#xa0;g, intraperitoneal (i.p.)] and transcardially perfused with ice-cold 0.01&#xa0;M phosphate buffered saline (PBS, Coolaber) and paraformaldehyde (PFA, 4% in PBS wt/vol, Saiguo Biotechnology Co., Ltd). Mouse brains and kidneys were collected and incubated in the same PFA solution at 4&#xb0;C for 24&#x2013;48&#xa0;h for uniform fixation. After fixation, mouse brains and kidneys were washed in 0.01&#xa0;M PBS at room temperature (20&#x2013;25&#xb0;C) for 6&#x2013;12&#xa0;h. The mouse brains and kidneys were clarified by the CUBIC-L/R&#x2b; protocol [xxx]. CUBIC-L [10&#xa0;wt% of N-butyldiethanolamine (Vetec) and 10&#xa0;wt% of Triton X-100 (Sigma) in water] and CUBIC-R&#x2b; [45&#xa0;wt% of antipyrine (Vetec) and 30&#xa0;wt% of nicotinamide (Vetec) in water, buffered with 0.5% (v/w) N-butyldiethanolamine (pH &#x223c; 10)] was prepared for tissue clearing. Mouse organs and tissues were incubated in CUBIC-L for 7 d at 37&#xb0;C with gentle shaking followed by PBS washing at room temperature. After PBS clearing, the mouse brains and kidneys was incubated and stored in CUBIC-R&#x2b; at room temperature.</p>
<p>For generating 3D datasets and verifying the application of our method, we acquired the experimental data of mouse organs and tissues via a custom-built LSFM system. For exciting fluorescence signal of mouse tissues labeled with Thy1-GFP, DRAQ5, and CD31-AF647, two semiconductor lasers (OBIS 488LS/637LX nm, Coherent) were aligned and expanded by a pair of achromatic lenses with 30 and 250&#xa0;mm focal length, respectively. Like the classical selective plane illumination microscopy, we used a cylindrical lens with 100&#xa0;mm focal length and a low-NA objective (Olympus &#xd7;4/NA 0.1/WD 18.5&#xa0;mm) with long working distance to generate a thin illumination sheet. And a mechanical slit (VA100&#xa0;C/M, Thorlabs) was set to 1&#xa0;mm for controlling the thickness of the illumination sheet. The fluorescence signal was collected by a tube lens (ITL200, Thorlabs), detection objective (Nikon &#xd7;4/NA 0.2/WD 20&#xa0;mm), a multi-channel emission filter (&#x23;87-247, Edmund Optics) and a sCMOS camera (ORCA-Flash 4.0 V3, Hamamatsu). And the biological samples were loaded on a 3D motorized stage (KMT50SE/M-3D, Thorlabs) and performed scanned imaging. In this system, we used the external trigger mode with a synchronous signal to ensure high-frame image acquisition. During the LSFM imaging process, each frame was captured with a constant acquisition interval of 2&#xa0;&#x3bc;m and total exposure time of 10&#xa0;ms. The corresponding camera acquisition speed could be up to 100 fps. Due to the rapid imaging advantage, the image acquisition process of a whole mouse brain tissue slice could be finished in 4&#xa0;min. All images were transformed and stored as 16-bit Multi-TIFF format for post-processing.</p>
</sec>
</sec>
<sec sec-type="results" id="s3">
<title>Results</title>
<sec id="s3-1">
<title>Image Enhancement Performance of 3D-VoCycleGAN for Mouse Brain Slices</title>
<p>Due to spatial anisotropy and structural complexity of biological tissues, chemical tissue clearing still could not realize perfect tissue transparency and image contrast. As for a pre-cleared tissue, the clearing extent and spatial image degradation is random across the whole 3D volume. Here we first demonstrated our 3D-VoCycleGAN on Thy1-GFP mouse brain slices to digitally improve the optical clearing extent and image quality. Although the whole image showed the distribution of fluorescence signal in the brain slice, some details such as nerve fiber were still blurred as shown in <xref ref-type="fig" rid="F2">Figure 2A</xref>. By using our 3D-VoCycleGAN, the image background was greatly suppressed. Besides, the blurred nerve fiber in the original image became more distinguishable after enhancement. Further, we cropped a region-of-interest (ROI) from the brain slice to compare the virtual optical clearing efficiency and image enhancing performance with other deconvolution methods. Here we performed sparse deconvolution [<xref ref-type="bibr" rid="B40">40</xref>], Richardson-Lucy deconvolution and 3D-VoCycleGAN on a 500 &#xd7; 500&#xd7;32 pixel<sup>3</sup> image volume. According to the results in <xref ref-type="fig" rid="F2">Figure 2B</xref>, three methods all realized image quality improvement to different extents. Then we plotted the profiles of two lines across the nerve fiber to quantitatively evaluate their differences of performance. Although Richardson-Lucy also enhanced the image resolution to some extent, the image background noise suppression capability was not as good as the other methods. Our 3D-VoCycleGAN and sparse deconvolution showed better image contrast and SBR improvements, which were both increased by above 40% (<xref ref-type="fig" rid="F2">Figure 2C</xref>). It is worth noting that although sparse deconvolution showed very powerful image quality enhancing capability comparable to our 3D-VoCycleGAN, the processing time consumption for the image volume with same size is larger than our method (<xref ref-type="fig" rid="F2">Figure 2D</xref>). It usually takes a considerable amount of time to finish the image processing by using Richardson-Lucy deconvolution. And sparse deconvolution could shorten the processing time with the aid of GPU acceleration. Nevertheless, due to GPU memory limitation, sparse deconvolution could not realize image processing of large 3D data in one time. Although sparse deconvolution could process the 3D image data volume by volume after splitting the large 3D image data into several image stacks, the final stitched 3D image will show uneven brightness and background since these sub-volumes exist sparsity differences. As for this image volume, the processing time consumption of 3D-VoCycleGAN, sparse deconvolution, and Richardson-Lucy deconvolution were 0.8, 48 and 45s, respectively. Thereinto, we performed 50 iterations of the Richardson-Lucy deconvolution by using the DeconvolutionLab2 plugin in ImageJ/Fiji. Then, with 3D-VoCycleGAN, image processing of the whole brain slice with 3,100 &#xd7; 3,500&#xd7;180 pixel<sup>3</sup> volume could be finished in only 141s. Hence, our 3D-VoCycleGAN could realize great image quality enhancement with short time consumption, showing the great image processing efficiency in brain slice imaging.</p>
<fig id="F2" position="float">
<label>FIGURE 2</label>
<caption>
<p>The image enhancement performance of 3D-VoCycleGAN. <bold>(A)</bold> The comparison results before and after using 3D-VoCyleGAN <bold>(B)</bold> The image quality enhancement results with Richardson-Lucy deconvolution, sparse deconvolution, and 3D-VoCycleGAN method. The size of cropped image volume was limited by the GPU memory requirements of Richardson-Lucy deconvolution and sparse deconvolution. <bold>(C)</bold> The quantitative evaluation of the image quality enhancement performance. The plotting curves of two line profiles showed the signal level and SBR of nerve fiber in the brain slice <bold>(D)</bold> The time consumption of different image processing methods for the image volume. RL: Richardson-Lucy deconvolution, Sparse: sparse deconvolution, 3D-VoC: 3D-VoCycleGAN.</p>
</caption>
<graphic xlink:href="fphy-10-965095-g002.tif"/>
</fig>
</sec>
<sec id="s3-2">
<title>Information Restoration of Images Deep Inside Kidney Tissue Slices</title>
<p>Tissue slice imaging with chemical optical clearing protocols are widely used in biological research. Although many chemical tissue clearing methods have promoted biological structure and function research, tissue scattering and refractive index mismatching between multiple media in the imaging system usually influence the image quality and microstructure analysis under the slice surface. Our 3D-VoCycleGAN could contribute to the information restoration of images deep inside tissue slices. As shown in <xref ref-type="fig" rid="F3">Figure 3A</xref>, labelled cell nucleus of the mouse kidney tissue slice showed the distribution of glomeruli. However, the tissue transparency was not enough to distinguish the detailed nucleus or glomeruli. Especially, due to the uneven thickness and structural anisotropy in axial direction, the clearing reagents could not perfectly make the kidney tissue slice transparent. For example, although the kidney tissue slice could be optical cleared by the clearing reagents, the images deep inside the tissue slice were still blurred due to inevitable light scattering or attenuation. The fluorescence signal was nearly overwhelmed in the strong background signal except the area from the tissue surface to 115-&#x3bc;m depth (<xref ref-type="fig" rid="F3">Figure 3B</xref>). By using our 3D-VoCycleGAN, the fluorescence signal in deep tissue could be quickly recovered. The axial images in <xref ref-type="fig" rid="F3">Figures 3B,C</xref> showed the profile of glomeruli and tubules across about 150-&#x3bc;m depth, which improved the clearing depth of mouse kidney tissues by up to 30%. Besides, we could also compare the 2D images in different depths. When the imaging depth was 100&#xa0;&#x3bc;m, the original image could show distinguished cell nucleus with faint noise. Our virtual optical clearing method could improve the image contrast with details maintained as shown in <xref ref-type="fig" rid="F3">Figure 3D</xref>. When the imaging depth was 140&#xa0;&#x3bc;m, the morphology of cell nucleus and glomeruli were severely blurred (<xref ref-type="fig" rid="F3">Figure 3E</xref>). With 3D-VoCycleGAN, the final image quality was evidently improved, where some information of nucleus and glomeruli was recovered. Hence, to some extent, we could realize structural information restoration of imperfectly cleared tissue slices in deep depth by virtual optical clearing technique.</p>
<fig id="F3" position="float">
<label>FIGURE 3</label>
<caption>
<p>The image information restoration effect in a mouse kidney tissue slice. <bold>(A)</bold> Fluorescence image of a mouse kidney slice labelled with cell nucleus <bold>(B,C)</bold> The comparison of clearing extent between original and fluorescence image processed by 3D-VoCycleGAN. <bold>(D,E)</bold> The image improvements in different imaging depths. The image of z &#x3d; 100&#xa0;&#x3bc;m was given in <bold>(D)</bold> and image of z &#x3d; 140&#xa0;&#x3bc;m was given in <bold>(E)</bold>.</p>
</caption>
<graphic xlink:href="fphy-10-965095-g003.tif"/>
</fig>
</sec>
<sec id="s3-3">
<title>Virtual Optical Clearing of Large-Scale 3D Tissues</title>
<p>Except for tissue slice imaging, high-throughput imaging for large-scale 3D tissues also plays an increasingly important part in biological research, especially in digital organ mapping and brain network reconstruction. Here we demonstrated the 3D image enhancement capability for stereoscopic mouse tissue with above 1 &#xd7; 1 &#xd7; 1&#xa0;mm<sup>3</sup> volume by using our virtual clearing method. To further depict the glomeruli distribution or morphology of kidney tissue, we imaged and reconstructed a mouse kidney tissue block with a custom-bulit LSFM system. As shown in <xref ref-type="fig" rid="F4">Figure 4A</xref>, the original 3D image volume has evident space-variant opaqueness and structure blurring. In particular, the structural contours of glomeruli were gradually blurred as the axial depth increases, which was marked by the white arrow in <xref ref-type="fig" rid="F4">Figure 4A</xref>. The connections between glomeruli and arteries were also very important for supporting some kidney functions. However, in the imperfectly cleared kidney tissue, a branch of the kidney artery was also overwhelmed in the background noise (yellow arrow in <xref ref-type="fig" rid="F4">Figure 4A</xref>). By using our virtual optical clearing method, the glomeruli and artery branch were all evidently recovered and distinguished (<xref ref-type="fig" rid="F4">Figure 4B</xref>). And the anisotropic tissue transparency was eliminated after virtual optical clearing. From the depth color-coded z-projections, we could see the whole image contrast and fluorescence signal intensity in various depths were enhanced (<xref ref-type="fig" rid="F4">Figures 4C,D</xref>).</p>
<fig id="F4" position="float">
<label>FIGURE 4</label>
<caption>
<p>The virtual optical clearing effect for millimeter-thickness mouse tissues. <bold>(A)</bold> 3D fluorescence image of imperfectly cleared mouse kidney tissue. The two dashed boxes represents 2D sections from two imaging depths <bold>(B)</bold> 3D fluorescence image of virtually optically cleared mouse kidney tissue by using 3D-VoCycleGAN. The white and yellow arrows represents the glomerulus and artery branch, respectively. <bold>(C,D)</bold> The depth color-coded image of image volume in <bold>(A)</bold> and <bold>(B)</bold>, respectively. <bold>(E)</bold> The comparison results between original image and virtually cleared image for shallow depth shown in <bold>(A)</bold>. The magnifications of insets were &#xd7;2. <bold>(F)</bold> The comparison results between original image and virtually cleared image for deep depth shown in <bold>(A)</bold>. The magnifications of insets were &#xd7;1.5. The blue and green arrows in <bold>(E)</bold> and <bold>(F)</bold> represented the structure details of glomeruli and artery branch, respectively. <bold>(G)</bold> The quantitative evaluation of tissue structural details. The plotting curves of four groups of line profiles represented the glomeruli and artery branch <bold>(H)</bold> The comparison results between original and virtually cleared mouse brain vessels by using depth color-coded z-projections. <bold>(I)</bold> The original 2D image showing blurred brain vessels <bold>(J)</bold> The 2D image corresponding to image <bold>(I)</bold> after virtual optical clearing. The blue and green arrows represented the brain vessel details. The magnifications of insets were &#xd7;3.</p>
</caption>
<graphic xlink:href="fphy-10-965095-g004.tif"/>
</fig>
<p>In order to quantify the virtual optical clearing effect, we selected two 2D images from two different z-depths. As shown in <xref ref-type="fig" rid="F4">Figure 4E</xref>, the original 2D image had low image quality, including strong noise, blurring, and structure missing (two blue arrows). After virtual optical clearing, the details and contours of glomeruli were enhanced greatly, which might be meaningful to the morphological analysis of kidney microstructures. Although the original 2D image in <xref ref-type="fig" rid="F4">Figure 4F</xref> had lower noise and blurring extent than the original image in <xref ref-type="fig" rid="F4">Figure 4E</xref> because of shallow imaging depth, the vascular walls of two artery branches were still difficult to be distinguished directly (two green arrows). Our virtual optical clearing successfully recovered the details of two vascular walls. Further, we evaluated the fluorescence signal intensity and noise level in the two depths by plotting four lines. As shown in <xref ref-type="fig" rid="F4">Figure 4G</xref>, images processed by our method had more distinguishable details about glomeruli and artery, especially the contour of glomeruli and vascular walls. Besides, our method improved the SBR of images by above 25%. Meanwhile, we tested our method on 3D images of brain vessels, which showed complex structures and dense distribution in the mouse brain. For deep brain tissue 3D imaging, tissue anisotropy and imperfect tissue clearing will introduce strong scattering and attenuation of fluorescence signal, resulting in vessel blurring and artifacts as shown in <xref ref-type="fig" rid="F4">Figure 4H</xref>. According to the comparison results of depth color-coded z-projections, we could see the definition and sharpness of vessels were greatly improved in various depths by using our virtual optical clearing method. As shown in <xref ref-type="fig" rid="F4">Figures 4I,J</xref>, the vessel artifacts were suppressed after virtual clearing. Similar to the results of kidney tissue, the original vessel contours existed severe blurring. Our virtual optical clearing method could effectively restore the information and improve the image quality, indicating its powerful image enhancement capability in 3D fluorescence imaging.</p>
</sec>
</sec>
<sec sec-type="discussion" id="s4">
<title>Discussion</title>
<p>High-throughput 3D fluorescence imaging and tissue clearing techniques are playing an increasingly important role in biological research. However, due to the tissue anisotropy and structural complexity, chemical tissue clearing techniques sometimes could not imperfectly clear the whole tissue, resulting in image quality degradation, such as image blurring, background noise, artifacts and so on. A series of methods including physical, chemical and digital ways have been proposed to improve the fluorescence image quality in recent years. Here, we presented an unsupervised deep learning-based image processing method, called 3D-VoCycleGAN to realize further virtual optical clearing of imperfectly cleared tissues. By making full use of the tissue anisotropy and space-variant clearing extent, we built a virtual optical clearing method to enhance the clearing effect of chemical tissue clearing techniques. In our virtual optical clearing, we established a CycleGAN architecture which consists of two pairs of 3D image generators and discriminators to realize the transformation and evaluation from low-quality LSFM images to high-quality LSFM images. Needless of accurate data pre-alignment between the source domain and target domain, our method need not accurate data pre-alignment between the source domain and target domain, which greatly improves the image processing efficiency. Compared with other image enhancement methods, our method showed more powerful image enhancement effect and faster processing speed. With the 3D-VoCycleGAN, we could restore more detailed mouse tissue structural information with high SBR and image contrast, such as distinguished nerve fibers, somas, glomeruli, and vessels.</p>
<p>Particularly, it is the first time that the CycleGAN deep learning model has been used for enhancing the clearing effect of chemical tissue clearing, and restoring the 3D blurred LSFM images. Furthermore, except for LSFM, our 3D-VoCycleGAN could also be used to process 3D images captured by other 3D fluorescence imaging systems such as confocal, two-photon microscopy and so on. From the aspect of deep learning network architecture, the 3D operations such as 3D max pooling and 3D convolution in our 3D-VoCycleGAN are not only limited to specific fluorescence imaging systems. From the aspect of dataset preparation, the sparse deconvolution method which we used to pre-process the 3D cleared tissue images was proposed to improve the image quality of structured illumination microscopy and proved to be available in various 3D fluorescence imaging systems such as confocal, two-photon, expansion microscopy and so on [<xref ref-type="bibr" rid="B40">40</xref>]. Hence, the whole image processing flow of our 3D-VoCycleGAN could be transferred to various 3D fluorescence imaging systems by using related datasets. In summary, our study promoted the combination and application of digital image processing, chemical tissue clearing and 3D fluorescence imaging techniques, showing the promising development of interdisciplinary technology in future high-throughput 3D biomedical imaging and biological research.</p>
</sec>
</body>
<back>
<sec id="s5" sec-type="data-availability">
<title>Data Availability Statement</title>
<p>The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding author.</p>
</sec>
<sec id="s6">
<title>Ethics Statement</title>
<p>All mouse experiments followed the guidelines for the Care and Use of Laboratory Animals of Zhejiang University approved by the Committee of Laboratory Animal Center of Zhejiang University.</p>
</sec>
<sec id="s7">
<title>Author Contributions</title>
<p>JC and ZD conceived the idea. KS supervised the project. JC and ZD performed the experiments and image processing. JC, ZD, and KS wrote the paper. All the authors contributed to the discussion on the results for this manuscript.</p>
</sec>
<sec id="s8">
<title>Funding</title>
<p>This work was supported in part by the Key R&#x26;D Program of Zhejiang Province (2021C03001), National Natural Science Foundation of China (61735016), CAMS Innovation Fund for Medical Sciences (2019-I2M-5-057), Fundamental Research Funds for the Central Universities, Alibaba Cloud.</p>
</sec>
<sec sec-type="COI-statement" id="s9">
<title>Conflict of Interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="disclaimer" id="s10">
<title>Publisher&#x2019;s Note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<ref-list>
<title>References</title>
<ref id="B1">
<label>1.</label>
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Wilson</surname>
<given-names>T</given-names>
</name>
<name>
<surname>Sheppard</surname>
<given-names>C</given-names>
</name>
</person-group>. <source>Theory and Practice of Scanning Optical Microscopy</source>. <publisher-loc>London</publisher-loc>: <publisher-name>Academic Press</publisher-name> (<year>1984</year>). </citation>
</ref>
<ref id="B2">
<label>2.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Feuchtinger</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Walch</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Dobosz</surname>
<given-names>M</given-names>
</name>
</person-group>. <article-title>Deep Tissue Imaging: A Review from a Preclinical Cancer Research Perspective</article-title>. <source>Histochem Cel Biol</source> (<year>2016</year>) <volume>146</volume>:<fpage>781</fpage>&#x2013;<lpage>806</lpage>. <pub-id pub-id-type="doi">10.1007/s00418-016-1495-7</pub-id> </citation>
</ref>
<ref id="B3">
<label>3.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ntziachristos</surname>
<given-names>V</given-names>
</name>
</person-group>. <article-title>Going Deeper Than Microscopy: The Optical Imaging Frontier in Biology</article-title>. <source>Nat Methods</source> (<year>2010</year>) <volume>7</volume>:<fpage>603</fpage>&#x2013;<lpage>14</lpage>. <pub-id pub-id-type="doi">10.1038/nmeth.1483</pub-id> </citation>
</ref>
<ref id="B4">
<label>4.</label>
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Yu</surname>
<given-names>H</given-names>
</name>
<name>
<surname>Rahim</surname>
<given-names>NAA</given-names>
</name>
</person-group>. <source>Imaging in Cellular and Tissue Engineering</source>. <edition>1st ed.</edition> <publisher-loc>Boca Raton</publisher-loc>: <publisher-name>CRC Press</publisher-name> (<year>2013</year>). </citation>
</ref>
<ref id="B5">
<label>5.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Yoon</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Kim</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Jang</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Choi</surname>
<given-names>Y</given-names>
</name>
<name>
<surname>Choi</surname>
<given-names>W</given-names>
</name>
<name>
<surname>Kang</surname>
<given-names>S</given-names>
</name>
<etal/>
</person-group> <article-title>Deep Optical Imaging within Complex Scattering media</article-title>. <source>Nat Rev Phys</source> (<year>2020</year>) <volume>2</volume>:<fpage>141</fpage>&#x2013;<lpage>58</lpage>. <pub-id pub-id-type="doi">10.1038/s42254-019-0143-2</pub-id> </citation>
</ref>
<ref id="B6">
<label>6.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Helmchen</surname>
<given-names>F</given-names>
</name>
<name>
<surname>Denk</surname>
<given-names>W</given-names>
</name>
</person-group>. <article-title>Deep Tissue Two-Photon Microscopy</article-title>. <source>Nat Methods</source> (<year>2005</year>) <volume>2</volume>:<fpage>932</fpage>&#x2013;<lpage>40</lpage>. <pub-id pub-id-type="doi">10.1038/nmeth818</pub-id> </citation>
</ref>
<ref id="B7">
<label>7.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Li</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Gong</surname>
<given-names>H</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>B</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>Q</given-names>
</name>
<name>
<surname>Yan</surname>
<given-names>C</given-names>
</name>
<name>
<surname>Wu</surname>
<given-names>J</given-names>
</name>
<etal/>
</person-group> <article-title>Micro-Optical Sectioning Tomography to Obtain a High-Resolution Atlas of the Mouse Brain</article-title>. <source>Science</source> (<year>2010</year>) <volume>330</volume>:<fpage>1404</fpage>&#x2013;<lpage>8</lpage>. <pub-id pub-id-type="doi">10.1126/science.1191776</pub-id> </citation>
</ref>
<ref id="B8">
<label>8.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Dodt</surname>
<given-names>H-U</given-names>
</name>
<name>
<surname>Leischner</surname>
<given-names>U</given-names>
</name>
<name>
<surname>Schierloh</surname>
<given-names>A</given-names>
</name>
<name>
<surname>J&#xe4;hrling</surname>
<given-names>N</given-names>
</name>
<name>
<surname>Mauch</surname>
<given-names>CP</given-names>
</name>
<name>
<surname>Deininger</surname>
<given-names>K</given-names>
</name>
<etal/>
</person-group> <article-title>Ultramicroscopy: Three-Dimensional Visualization of Neuronal Networks in the Whole Mouse Brain</article-title>. <source>Nat Methods</source> (<year>2007</year>) <volume>4</volume>:<fpage>331</fpage>&#x2013;<lpage>6</lpage>. <pub-id pub-id-type="doi">10.1038/nmeth1036</pub-id> </citation>
</ref>
<ref id="B9">
<label>9.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Economo</surname>
<given-names>MN</given-names>
</name>
<name>
<surname>Clack</surname>
<given-names>NG</given-names>
</name>
<name>
<surname>Lavis</surname>
<given-names>LD</given-names>
</name>
<name>
<surname>Gerfen</surname>
<given-names>CR</given-names>
</name>
<name>
<surname>Svoboda</surname>
<given-names>K</given-names>
</name>
<name>
<surname>Myers</surname>
<given-names>EW</given-names>
</name>
<etal/>
</person-group> <article-title>A Platform for Brain-wide Imaging and Reconstruction of Individual Neurons</article-title>. <source>eLife</source> (<year>2016</year>) <volume>5</volume>:<fpage>e10566</fpage>. <pub-id pub-id-type="doi">10.7554/eLife.10566</pub-id> </citation>
</ref>
<ref id="B10">
<label>10.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>H</given-names>
</name>
<name>
<surname>Magnain</surname>
<given-names>C</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>R</given-names>
</name>
<name>
<surname>Dubb</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Varjabedian</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Tirrell</surname>
<given-names>LS</given-names>
</name>
<etal/>
</person-group> <article-title>as-PSOCT: Volumetric Microscopic Imaging of Human Brain Architecture and Connectivity</article-title>. <source>NeuroImage</source> (<year>2018</year>) <volume>165</volume>:<fpage>56</fpage>&#x2013;<lpage>68</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuroimage.2017.10.012</pub-id> </citation>
</ref>
<ref id="B11">
<label>11.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Stelzer</surname>
<given-names>EHK</given-names>
</name>
<name>
<surname>Strobl</surname>
<given-names>F</given-names>
</name>
<name>
<surname>Chang</surname>
<given-names>B-J</given-names>
</name>
<name>
<surname>Preusser</surname>
<given-names>F</given-names>
</name>
<name>
<surname>Preibisch</surname>
<given-names>S</given-names>
</name>
<name>
<surname>McDole</surname>
<given-names>K</given-names>
</name>
<etal/>
</person-group> <article-title>Light Sheet Fluorescence Microscopy</article-title>. <source>Nat Rev Methods Primers</source> (<year>2021</year>) <volume>1</volume>:<fpage>73</fpage>. <pub-id pub-id-type="doi">10.1038/s43586-021-00069-4</pub-id> </citation>
</ref>
<ref id="B12">
<label>12.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Richardson</surname>
<given-names>DS</given-names>
</name>
<name>
<surname>Guan</surname>
<given-names>W</given-names>
</name>
<name>
<surname>Matsumoto</surname>
<given-names>K</given-names>
</name>
<name>
<surname>Pan</surname>
<given-names>C</given-names>
</name>
<name>
<surname>Chung</surname>
<given-names>K</given-names>
</name>
<name>
<surname>Ert&#xfc;rk</surname>
<given-names>A</given-names>
</name>
<etal/>
</person-group> <article-title>Tissue Clearing</article-title>. <source>Nat Rev Methods Primers</source> (<year>2021</year>) <volume>1</volume>:<fpage>84</fpage>. <pub-id pub-id-type="doi">10.1038/s43586-021-00080-9</pub-id> </citation>
</ref>
<ref id="B13">
<label>13.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Brenna</surname>
<given-names>C</given-names>
</name>
<name>
<surname>Simioni</surname>
<given-names>C</given-names>
</name>
<name>
<surname>Varano</surname>
<given-names>G</given-names>
</name>
<name>
<surname>Conti</surname>
<given-names>I</given-names>
</name>
<name>
<surname>Costanzi</surname>
<given-names>E</given-names>
</name>
<name>
<surname>Melloni</surname>
<given-names>M</given-names>
</name>
<etal/>
</person-group> <article-title>Optical Tissue Clearing Associated with 3D Imaging: Application in Preclinical and Clinical Studies</article-title>. <source>Histochem Cel Biol</source> (<year>2022</year>) <volume>157</volume>:<fpage>497</fpage>&#x2013;<lpage>511</lpage>. <pub-id pub-id-type="doi">10.1007/s00418-022-02081-5</pub-id> </citation>
</ref>
<ref id="B14">
<label>14.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Yu</surname>
<given-names>T</given-names>
</name>
<name>
<surname>Zhu</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>D</given-names>
</name>
<name>
<surname>Zhu</surname>
<given-names>D</given-names>
</name>
</person-group>. <article-title>Physical and Chemical Mechanisms of Tissue Optical Clearing</article-title>. <source>iScience</source> (<year>2021</year>) <volume>24</volume>:<fpage>102178</fpage>. <pub-id pub-id-type="doi">10.1016/j.isci.2021.102178</pub-id> </citation>
</ref>
<ref id="B15">
<label>15.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Matsumoto</surname>
<given-names>K</given-names>
</name>
<name>
<surname>Mitani</surname>
<given-names>TT</given-names>
</name>
<name>
<surname>Horiguchi</surname>
<given-names>SA</given-names>
</name>
<name>
<surname>Kaneshiro</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Murakami</surname>
<given-names>TC</given-names>
</name>
<name>
<surname>Mano</surname>
<given-names>T</given-names>
</name>
<etal/>
</person-group> <article-title>Advanced CUBIC Tissue Clearing for Whole-Organ Cell Profiling</article-title>. <source>Nat Protoc</source> (<year>2019</year>) <volume>14</volume>:<fpage>3506</fpage>&#x2013;<lpage>37</lpage>. <pub-id pub-id-type="doi">10.1038/s41596-019-0240-9</pub-id> </citation>
</ref>
<ref id="B16">
<label>16.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Molbay</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Kolabas</surname>
<given-names>ZI</given-names>
</name>
<name>
<surname>Todorov</surname>
<given-names>MI</given-names>
</name>
<name>
<surname>Ohn</surname>
<given-names>T-L</given-names>
</name>
<name>
<surname>Ert&#xfc;rk</surname>
<given-names>A</given-names>
</name>
</person-group>. <article-title>A Guidebook for DISCO Tissue Clearing</article-title>. <source>Mol Syst Biol</source> (<year>2021</year>) <volume>17</volume>:<fpage>e9807</fpage>. <pub-id pub-id-type="doi">10.15252/msb.20209807</pub-id> </citation>
</ref>
<ref id="B17">
<label>17.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhu</surname>
<given-names>X</given-names>
</name>
<name>
<surname>Huang</surname>
<given-names>L</given-names>
</name>
<name>
<surname>Zheng</surname>
<given-names>Y</given-names>
</name>
<name>
<surname>Song</surname>
<given-names>Y</given-names>
</name>
<name>
<surname>Xu</surname>
<given-names>Q</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>J</given-names>
</name>
<etal/>
</person-group> <article-title>Ultrafast Optical Clearing Method for Three-Dimensional Imaging with Cellular Resolution</article-title>. <source>Proc Natl Acad Sci U.S.A</source> (<year>2019</year>) <volume>116</volume>:<fpage>11480</fpage>&#x2013;<lpage>9</lpage>. <pub-id pub-id-type="doi">10.1073/pnas.1819583116</pub-id> </citation>
</ref>
<ref id="B18">
<label>18.</label>
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Cannell</surname>
<given-names>MB</given-names>
</name>
<name>
<surname>McMorland</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Soeller</surname>
<given-names>C</given-names>
</name>
</person-group>. <article-title>Image Enhancement by Deconvolution</article-title>. In: <person-group person-group-type="editor">
<name>
<surname>Pawley</surname>
<given-names>JB</given-names>
</name>
</person-group>, editor. <source>Handbook of Biological Confocal Microscopy</source>. <publisher-loc>Boston, MA</publisher-loc>: <publisher-name>Springer US</publisher-name> (<year>2006</year>). p. <fpage>488</fpage>&#x2013;<lpage>500</lpage>. <pub-id pub-id-type="doi">10.1007/978-0-387-45524-2_25</pub-id> </citation>
</ref>
<ref id="B19">
<label>19.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Richardson</surname>
<given-names>WH</given-names>
</name>
</person-group>. <article-title>Bayesian-Based Iterative Method of Image Restoration&#x2a;</article-title>. <source>J Opt Soc Am</source> (<year>1972</year>) <volume>62</volume>:<fpage>55</fpage>&#x2013;<lpage>9</lpage>. <pub-id pub-id-type="doi">10.1364/josa.62.000055</pub-id> </citation>
</ref>
<ref id="B20">
<label>20.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Lucy</surname>
<given-names>LB</given-names>
</name>
</person-group>. <article-title>An Iterative Technique for the Rectification of Observed Distributions</article-title>. <source>Astronomical J</source> (<year>1974</year>) <volume>79</volume>:<fpage>745</fpage>. <pub-id pub-id-type="doi">10.1086/111605</pub-id> </citation>
</ref>
<ref id="B21">
<label>21.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Belthangady</surname>
<given-names>C</given-names>
</name>
<name>
<surname>Royer</surname>
<given-names>LA</given-names>
</name>
</person-group>. <article-title>Applications, Promises, and Pitfalls of Deep Learning for Fluorescence Image Reconstruction</article-title>. <source>Nat Methods</source> (<year>2019</year>) <volume>16</volume>:<fpage>1215</fpage>&#x2013;<lpage>25</lpage>. <pub-id pub-id-type="doi">10.1038/s41592-019-0458-z</pub-id> </citation>
</ref>
<ref id="B22">
<label>22.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>H</given-names>
</name>
<name>
<surname>Rivenson</surname>
<given-names>Y</given-names>
</name>
<name>
<surname>Jin</surname>
<given-names>Y</given-names>
</name>
<name>
<surname>Wei</surname>
<given-names>Z</given-names>
</name>
<name>
<surname>Gao</surname>
<given-names>R</given-names>
</name>
<name>
<surname>G&#xfc;nayd&#x131;n</surname>
<given-names>H</given-names>
</name>
<etal/>
</person-group> <article-title>Deep Learning Enables Cross-Modality Super-resolution in Fluorescence Microscopy</article-title>. <source>Nat Methods</source> (<year>2019</year>) <volume>16</volume>:<fpage>103</fpage>&#x2013;<lpage>10</lpage>. <pub-id pub-id-type="doi">10.1038/s41592-018-0239-0</pub-id> </citation>
</ref>
<ref id="B23">
<label>23.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhang</surname>
<given-names>B</given-names>
</name>
<name>
<surname>Zhu</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Si</surname>
<given-names>K</given-names>
</name>
<name>
<surname>Gong</surname>
<given-names>W</given-names>
</name>
</person-group>. <article-title>Deep Learning Assisted Zonal Adaptive Aberration Correction</article-title>. <source>Front Phys</source> (<year>2021</year>) <volume>8</volume>. <pub-id pub-id-type="doi">10.3389/fphy.2020.621966</pub-id> </citation>
</ref>
<ref id="B24">
<label>24.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Schmidhuber</surname>
<given-names>J</given-names>
</name>
</person-group>. <article-title>Deep Learning in Neural Networks: An Overview</article-title>. <source>Neural Networks</source> (<year>2015</year>) <volume>61</volume>:<fpage>85</fpage>&#x2013;<lpage>117</lpage>. <pub-id pub-id-type="doi">10.1016/j.neunet.2014.09.003</pub-id> </citation>
</ref>
<ref id="B25">
<label>25.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Jin</surname>
<given-names>KH</given-names>
</name>
<name>
<surname>McCann</surname>
<given-names>MT</given-names>
</name>
<name>
<surname>Froustey</surname>
<given-names>E</given-names>
</name>
<name>
<surname>Unser</surname>
<given-names>M</given-names>
</name>
</person-group>. <article-title>Deep Convolutional Neural Network for Inverse Problems in Imaging</article-title>. <source>IEEE Trans Image Process</source> (<year>2017</year>) <volume>26</volume>:<fpage>4509</fpage>&#x2013;<lpage>22</lpage>. <pub-id pub-id-type="doi">10.1109/tip.2017.2713099</pub-id> </citation>
</ref>
<ref id="B26">
<label>26.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Salim</surname>
<given-names>UT</given-names>
</name>
<name>
<surname>Ali</surname>
<given-names>F</given-names>
</name>
<name>
<surname>Dawwd</surname>
<given-names>SA</given-names>
</name>
<name>
<surname>System</surname>
<given-names>D</given-names>
</name>
</person-group>. <article-title>U-net Convolutional Networks Performance Based on Software-Hardware Cooperation Parameters: A Review</article-title>. <source>Int J Comput Digital Syst</source> (<year>2021</year>) <volume>11</volume>. <pub-id pub-id-type="doi">10.12785/ijcds/110180</pub-id> </citation>
</ref>
<ref id="B27">
<label>27.</label>
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Karhunen</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Raiko</surname>
<given-names>T</given-names>
</name>
<name>
<surname>Cho</surname>
<given-names>K</given-names>
</name>
</person-group>. <article-title>Unsupervised Deep Learning</article-title>. In: <person-group person-group-type="editor">
<name>
<surname>Bingham</surname>
<given-names>E</given-names>
</name>
<name>
<surname>Kaski</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Laaksonen</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Lampinen</surname>
<given-names>J</given-names>
</name>
</person-group>, editors. <source>Advances in Independent Component Analysis and Learning Machines</source>. <publisher-name>Academic Press</publisher-name> (<year>2015</year>). p. <fpage>125</fpage>&#x2013;<lpage>42</lpage>. <pub-id pub-id-type="doi">10.1016/b978-0-12-802806-3.00007-5</pub-id> </citation>
</ref>
<ref id="B28">
<label>28.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Li</surname>
<given-names>Z</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>T</given-names>
</name>
<name>
<surname>Wan</surname>
<given-names>P</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>D</given-names>
</name>
</person-group>. <article-title>SEGAN: Structure-Enhanced Generative Adversarial Network for Compressed Sensing MRI Reconstruction</article-title>. <source>Proceedings of the AAAI Conference on Artificial Intelligence</source> (<year>2019</year>) <volume>33</volume>:<fpage>1012</fpage>&#x2013;<lpage>9</lpage>. <pub-id pub-id-type="doi">10.1609/aaai.v33i01.33011012</pub-id> </citation>
</ref>
<ref id="B29">
<label>29.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Lv</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>C</given-names>
</name>
<name>
<surname>Yang</surname>
<given-names>G</given-names>
</name>
</person-group>. <article-title>PIC-GAN: A Parallel Imaging Coupled Generative Adversarial Network for Accelerated Multi-Channel MRI Reconstruction</article-title>. <source>Diagnostics</source> (<year>2021</year>) <volume>11</volume>:<fpage>61</fpage>. <pub-id pub-id-type="doi">10.3390/diagnostics11010061</pub-id> </citation>
</ref>
<ref id="B30">
<label>30.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Quan</surname>
<given-names>TM</given-names>
</name>
<name>
<surname>Nguyen-Duc</surname>
<given-names>T</given-names>
</name>
<name>
<surname>Jeong</surname>
<given-names>W-K</given-names>
</name>
</person-group>. <article-title>Compressed Sensing MRI Reconstruction Using a Generative Adversarial Network with a Cyclic Loss</article-title>. <source>IEEE Trans Med Imaging</source> (<year>2018</year>) <volume>37</volume>:<fpage>1488</fpage>&#x2013;<lpage>97</lpage>. <pub-id pub-id-type="doi">10.1109/tmi.2018.2820120</pub-id> </citation>
</ref>
<ref id="B31">
<label>31.</label>
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Zhu</surname>
<given-names>J-Y</given-names>
</name>
<name>
<surname>Park</surname>
<given-names>T</given-names>
</name>
<name>
<surname>Isola</surname>
<given-names>P</given-names>
</name>
<name>
<surname>Efros</surname>
<given-names>AA</given-names>
</name>
</person-group>. <article-title>Unpaired Image-To-Image Translation Using Cycle-Consistent Adversarial Networks</article-title>. In: <conf-name>IEEE International Conference on Computer Vision (ICCV)</conf-name>; <conf-date>22-29 October, 2017</conf-date>; <conf-loc>Venice, Italy</conf-loc> (<year>2017</year>). p. <fpage>2242</fpage>&#x2013;<lpage>51</lpage>. <pub-id pub-id-type="doi">10.1109/ICCV.2017.244</pub-id> </citation>
</ref>
<ref id="B32">
<label>32.</label>
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Li</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Huang</surname>
<given-names>H</given-names>
</name>
<name>
<surname>Ma</surname>
<given-names>L</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>W</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>T</given-names>
</name>
<name>
<surname>Jiang</surname>
<given-names>Y</given-names>
</name>
</person-group>. <article-title>Unsupervised Image-To-Image Translation with Stacked Cycle-Consistent Adversarial Networks</article-title>. In: <conf-name>Proceedings of the European conference on computer vision (ECCV)</conf-name>; <conf-date>8-14 September, 2018</conf-date>; <conf-loc>Munich, Germany</conf-loc> (<year>2018</year>). p. <fpage>184</fpage>&#x2013;<lpage>99</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-030-01240-3_12</pub-id> </citation>
</ref>
<ref id="B33">
<label>33.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Chen</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Sasaki</surname>
<given-names>H</given-names>
</name>
<name>
<surname>Lai</surname>
<given-names>H</given-names>
</name>
<name>
<surname>Su</surname>
<given-names>Y</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Wu</surname>
<given-names>Y</given-names>
</name>
<etal/>
</person-group> <article-title>Three-dimensional Residual Channel Attention Networks Denoise and Sharpen Fluorescence Microscopy Image Volumes</article-title>. <source>Nat Methods</source> (<year>2021</year>) <volume>18</volume>:<fpage>678</fpage>&#x2013;<lpage>87</lpage>. <pub-id pub-id-type="doi">10.1038/s41592-021-01155-x</pub-id> </citation>
</ref>
<ref id="B34">
<label>34.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Weigert</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Schmidt</surname>
<given-names>U</given-names>
</name>
<name>
<surname>Boothe</surname>
<given-names>T</given-names>
</name>
<name>
<surname>M&#xfc;ller</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Dibrov</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Jain</surname>
<given-names>A</given-names>
</name>
<etal/>
</person-group> <article-title>Content-aware Image Restoration: Pushing the Limits of Fluorescence Microscopy</article-title>. <source>Nat Methods</source> (<year>2018</year>) <volume>15</volume>:<fpage>1090</fpage>&#x2013;<lpage>7</lpage>. <pub-id pub-id-type="doi">10.1038/s41592-018-0216-7</pub-id> </citation>
</ref>
<ref id="B35">
<label>35.</label>
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>He</surname>
<given-names>K</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>X</given-names>
</name>
<name>
<surname>Ren</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Sun</surname>
<given-names>J</given-names>
</name>
</person-group>. <article-title>Deep Residual Learning for Image Recognition</article-title>. In: <conf-name>Proceedings of the IEEE conference on computer vision and pattern recognition</conf-name>; <conf-date>27-30 June 2016</conf-date>; <conf-loc>Las Vegas, NV, USA</conf-loc> (<year>2016</year>). p. <fpage>770</fpage>&#x2013;<lpage>8</lpage>. <pub-id pub-id-type="doi">10.1109/cvpr.2016.90</pub-id> </citation>
</ref>
<ref id="B36">
<label>36.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Goodfellow</surname>
<given-names>I</given-names>
</name>
<name>
<surname>Pouget-Abadie</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Mirza</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Xu</surname>
<given-names>B</given-names>
</name>
<name>
<surname>Warde-Farley</surname>
<given-names>D</given-names>
</name>
<name>
<surname>Ozair</surname>
<given-names>S</given-names>
</name>
<etal/>
</person-group> <article-title>Generative Adversarial Nets</article-title>. <source>Adv Neural Inf Process Syst</source> (<year>2014</year>) <volume>2</volume>:<fpage>2672</fpage>&#x2013;<lpage>80</lpage>. <pub-id pub-id-type="doi">10.5555/2969033.2969125</pub-id> </citation>
</ref>
<ref id="B37">
<label>37.</label>
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Mao</surname>
<given-names>X</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>Q</given-names>
</name>
<name>
<surname>Xie</surname>
<given-names>H</given-names>
</name>
<name>
<surname>Lau</surname>
<given-names>RY</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>Z</given-names>
</name>
<name>
<surname>Paul Smolley</surname>
<given-names>S</given-names>
</name>
</person-group>. <article-title>Least Squares Generative Adversarial Networks</article-title>. In: <conf-name>Proceedings of the IEEE international conference on computer vision</conf-name>; <conf-date>22-29 October, 2017</conf-date>; <conf-loc>Venice, Italy</conf-loc> (<year>2017</year>). <fpage>2794</fpage>&#x2013;<lpage>802</lpage>. <pub-id pub-id-type="doi">10.1109/iccv.2017.304</pub-id> </citation>
</ref>
<ref id="B38">
<label>38.</label>
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>Z</given-names>
</name>
<name>
<surname>Simoncelli</surname>
<given-names>EP</given-names>
</name>
<name>
<surname>Bovik</surname>
<given-names>AC</given-names>
</name>
</person-group>. <article-title>Multiscale Structural Similarity for Image Quality Assessment</article-title>. In: <conf-name>The Thrity-Seventh Asilomar Conference on Signals, Systems &#x26; Computers</conf-name>; <conf-date>November 09-12, 2003</conf-date>; <conf-loc>Pacific Grove,CA</conf-loc> (<year>2003</year>). p. <fpage>1398</fpage>&#x2013;<lpage>402</lpage>. <pub-id pub-id-type="doi">10.1109/ACSSC.2003.1292216</pub-id> </citation>
</ref>
<ref id="B39">
<label>39.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Li</surname>
<given-names>X</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>G</given-names>
</name>
<name>
<surname>Qiao</surname>
<given-names>H</given-names>
</name>
<name>
<surname>Bao</surname>
<given-names>F</given-names>
</name>
<name>
<surname>Deng</surname>
<given-names>Y</given-names>
</name>
<name>
<surname>Wu</surname>
<given-names>J</given-names>
</name>
<etal/>
</person-group> <article-title>Unsupervised Content-Preserving Transformation for Optical Microscopy</article-title>. <source>Light Sci Appl</source> (<year>2021</year>) <volume>10</volume>:<fpage>44</fpage>. <pub-id pub-id-type="doi">10.1038/s41377-021-00484-y</pub-id> </citation>
</ref>
<ref id="B40">
<label>40.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhao</surname>
<given-names>W</given-names>
</name>
<name>
<surname>Zhao</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>L</given-names>
</name>
<name>
<surname>Huang</surname>
<given-names>X</given-names>
</name>
<name>
<surname>Xing</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>Y</given-names>
</name>
<etal/>
</person-group> <article-title>Sparse Deconvolution Improves the Resolution of Live-Cell Super-resolution Fluorescence Microscopy</article-title>. <source>Nat Biotechnol</source> (<year>2021</year>) <volume>40</volume>:<fpage>606</fpage>&#x2013;<lpage>17</lpage>. <pub-id pub-id-type="doi">10.1038/s41587-021-01092-2</pub-id> </citation>
</ref>
</ref-list>
</back>
</article>