<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article article-type="research-article" dtd-version="2.3" xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Genet.</journal-id>
<journal-title>Frontiers in Genetics</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Genet.</abbrev-journal-title>
<issn pub-type="epub">1664-8021</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">913372</article-id>
<article-id pub-id-type="doi">10.3389/fgene.2022.913372</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Genetics</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>CIEGAN: A Deep Learning Tool for Cell Image Enhancement</article-title>
<alt-title alt-title-type="left-running-head">Sun et al.</alt-title>
<alt-title alt-title-type="right-running-head">Cell Image Enhancement</alt-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name>
<surname>Sun</surname>
<given-names>Qiushi</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="fn" rid="fn1">
<sup>&#x2020;</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/1878145/overview"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Yang</surname>
<given-names>Xiaochun</given-names>
</name>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<xref ref-type="fn" rid="fn1">
<sup>&#x2020;</sup>
</xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Guo</surname>
<given-names>Jingtao</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Zhao</surname>
<given-names>Yang</given-names>
</name>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<xref ref-type="corresp" rid="c001">&#x2a;</xref>
<uri xlink:href="https://loop.frontiersin.org/people/1448980/overview"/>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Liu</surname>
<given-names>Yi</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="corresp" rid="c001">&#x2a;</xref>
<uri xlink:href="https://loop.frontiersin.org/people/1110703/overview"/>
</contrib>
</contrib-group>
<aff id="aff1">
<sup>1</sup>
<institution>Beijing Key Lab of Traffic Data Analysis and Mining</institution>, <institution>School of Computer and Information Technology</institution>, <institution>Beijing Jiaotong University</institution>, <addr-line>Beijing</addr-line>, <country>China</country>
</aff>
<aff id="aff2">
<sup>2</sup>
<institution>State Key Laboratory of Natural and Biomimetic Drugs</institution>, <institution>MOE Key Laboratory of Cell Proliferation and Differentiation</institution>, <institution>Beijing Key Laboratory of Cardiometabolic Molecular Medicine</institution>, <institution>Institute of Molecular Medicine</institution>, <institution>College of Future Technology</institution>, <institution>Peking University</institution>, <addr-line>Beijing</addr-line>, <country>China</country>
</aff>
<author-notes>
<fn fn-type="edited-by">
<p>
<bold>Edited by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/674022/overview">Pu-Feng Du</ext-link>, Tianjin University, China</p>
</fn>
<fn fn-type="edited-by">
<p>
<bold>Reviewed by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1665937/overview">Feng Yang</ext-link>, National Institutes of Health (NIH), United States</p>
<p>
<ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1702617/overview">Chunyu Jin</ext-link>, University of California, San Diego, United States</p>
</fn>
<corresp id="c001">&#x2a;Correspondence: Yang Zhao, <email>yangzhao@pku.edu.cn</email>; Yi Liu, <email>yiliu@bjtu.edu.cn</email>
</corresp>
<fn fn-type="equal" id="fn1">
<label>
<sup>&#x2020;</sup>
</label>
<p>These authors have contributed equally to this work and share first authorship</p>
</fn>
<fn fn-type="other">
<p>This article was submitted to Computational Genomics, a section of the journal Frontiers in Genetics</p>
</fn>
</author-notes>
<pub-date pub-type="epub">
<day>04</day>
<month>07</month>
<year>2022</year>
</pub-date>
<pub-date pub-type="collection">
<year>2022</year>
</pub-date>
<volume>13</volume>
<elocation-id>913372</elocation-id>
<history>
<date date-type="received">
<day>05</day>
<month>04</month>
<year>2022</year>
</date>
<date date-type="accepted">
<day>25</day>
<month>05</month>
<year>2022</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#xa9; 2022 Sun, Yang, Guo, Zhao and Liu.</copyright-statement>
<copyright-year>2022</copyright-year>
<copyright-holder>Sun, Yang, Guo, Zhao and Liu</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/">
<p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p>
</license>
</permissions>
<abstract>
<p>Long-term live-cell imaging technology has emerged in the study of cell culture and development, and it is expected to elucidate the differentiation or reprogramming morphology of cells and the dynamic process of interaction between cells. There are some advantages to this technique: it is noninvasive, high-throughput, low-cost, and it can help researchers explore phenomena that are otherwise difficult to observe. Many challenges arise in the real-time process, for example, low-quality micrographs are often obtained due to unavoidable human factors or technical factors in the long-term experimental period. Moreover, some core dynamics in the developmental process are rare and fleeting in imaging observation and difficult to recapture again. Therefore, this study proposes a deep learning method for microscope cell image enhancement to reconstruct sharp images. We combine generative adversarial nets and various loss functions to make blurry images sharp again, which is much more convenient for researchers to carry out further analysis. This technology can not only make up the blurry images of critical moments of the development process through image enhancement but also allows long-term live-cell imaging to find a balance between imaging speed and image quality. Furthermore, the scalability of this technology makes the methods perform well in fluorescence image enhancement. Finally, the method is tested in long-term live-cell imaging of human-induced pluripotent stem cell-derived cardiomyocyte differentiation experiments, and it can greatly improve the image space resolution ratio.</p>
</abstract>
<kwd-group>
<kwd>cell image</kwd>
<kwd>image enhancement</kwd>
<kwd>long-term imaging</kwd>
<kwd>deep learning</kwd>
<kwd>generative adversarial network</kwd>
</kwd-group>
<contract-num rid="cn001">31771475</contract-num>
<contract-sponsor id="cn001">National Natural Science Foundation of China<named-content content-type="fundref-id">10.13039/501100001809</named-content>
</contract-sponsor>
</article-meta>
</front>
<body>
<sec id="s1">
<title>Introduction</title>
<p>Microscopic imaging and fluorescence imaging technology have brought great convenience to biological research, and allow researchers to visually observe subcellular structures and the interaction between cells. The emergence of long-term live-cell imaging technology has made it possible to observe the cultivation and growth process of cells, which is expected to explain more biological phenomena over time. In particular, the dynamics of changes in cellular and subcellular structures and protein subcellular localization, and the dynamic process of cell differentiation and reprogramming were studied. It is crucial to decipher the mechanism behind the dynamic heterogeneous cellular responses.</p>
<p>Many studies require long-term imaging of living cells, so brightfield imaging should be carried out for further analysis to keep cells alive. The brightfield imaging process is simple without a fluorescent staining operation and the noise introduced into the experimental system is quite low. The non-intrusive experimental method shows great advantages: 1) no complex experimental operations, 2) does not introduce noise into the experimental system, and 3) does not interfere and destroy the cells themselves, while the phototoxicity can be reduced to a minimum. Since long-term live-cell imaging has such advantages, there have been many studies.</p>
<p>
<xref ref-type="bibr" rid="B36">Smith et al. (2010)</xref> used high-resolution time-lapse imaging to track the reprogramming process from single mouse embryonic fibroblasts (MEFs) to induced pluripotent stem (iPS) cell colonies over 2&#xa0;weeks. <xref ref-type="bibr" rid="B33">Schroeder (2011)</xref> conducted continuous long-term single-cell tracking observations of mammalian stem cells and found a set of technical solutions for long-term imaging and tracking. <xref ref-type="bibr" rid="B20">McQuate et al. (2017)</xref> established a pipeline for long-term live-cell imaging of infected cells and subsequent image analysis methods for <italic>Salmonella</italic> effector proteins SseG and SteA. <xref ref-type="bibr" rid="B5">Chen et al. (2010)</xref> developed a machine learning-based classification, segmentation, and statistical modeling system based on a time-lapse brightfield imaging analysis system to guide iPSC colony selection, counting, and classification automatically. In their research, AlexNet and hidden Markov model (HMM) technology were used. <xref ref-type="bibr" rid="B3">Buggenthin et al. (2017)</xref> used long-term time-lapse microscopy data and single-cell tracking annotation to prospectively predict differentiation outcomes in differentiating primary hematopoietic progenitors. They propose a convolutional neural network (CNN) combined with a recurrent neural network (RNN) architecture to process images from brightfield microscopy and cell motion. They predicted primary murine hematopoietic stem and progenitor cells (HSPCs) differentiating into either the granulocytic/monocytic (GM) or the megakaryocytic/erythroid (MegE) lineage. <xref ref-type="bibr" rid="B42">Wang et al. (2020)</xref> developed a live-cell imaging platform that tracks cell state changes by incorporating endogenous fluorescent labels. It can minimize the perturbation to cell physiology when processing live-cell imaging. In the field of cell differentiation and reprogramming, continuous long-term single-cell observation provides an insight into the mechanisms of cell fate. Even in the field of education, the low-cost long-term live-cell imaging platform also has high application prospects (<xref ref-type="bibr" rid="B41">Walzik et al., 2015</xref>).</p>
<p>A summary of the general processing pipeline of long-term live-cell imaging research is shown in <xref ref-type="fig" rid="F1">Figure 1</xref>. Once the research question has been set up, appropriate microscopy strategies must be tailored according to the experimental system to be used at the beginning of the study. In addition, it is necessary to balance the trade-offs between the image space resolution ratio, experimental throughput, and imaging speed (<xref ref-type="bibr" rid="B47">Weigert et al., 2018</xref>), limited by imaging technology and cost.</p>
<fig id="F1" position="float">
<label>FIGURE 1</label>
<caption>
<p>General processing pipeline for long-term live-cell research. The image preprocessing stage (blue part in this figure) is very important in the entire research pipeline and directly determines the accuracy and results of the further analysis.</p>
</caption>
<graphic xlink:href="fgene-13-913372-g001.tif"/>
</fig>
<p>However, many difficulties and challenges arise in actual long-term imaging experiments. This is the congenital deficiency of long-term live-cell imaging. It takes considerable effort to maintain regular cell culture conditions while performing long-term high-resolution imaging (<xref ref-type="bibr" rid="B35">Skylaki et al., 2016</xref>). For example, the obtained photos may not be as sharp and distinguishable as traditional fluorescent label imaging because the noninvasive label-free observation method has no conspicuous calibrations. At the same time, it is necessary to reduce the phototoxicity to a range that can be tolerated in an experimental system while exposing the live cells to the transmitted light in long-term incubation. Thus, it reduces the signal-to-noise ratio of image acquisition because the light intensity is limited. Moreover, a large number of cells will aggregate into clusters as a result of cell growth and culture in long-term live-cell culturing. The sudden growth of cells in a mass can cause loss of the focal surface. Dead cells will become pollutants, float up and block the view. Furthermore, artificial placement errors will be introduced in the long-term experimental system. For example, the medium was changed every certain period of time to maintain regular cell survival or differentiation. Nevertheless, thermal expansion and contraction of the culture chamber are caused by temperature changes during the movement of the culture chamber in and out of the thermostatic incubator. The quality and clarity of the medium will lead to a decrease in the quality of the acquired images. Some of the conditions that cause blurring in long-term live-cell imaging are shown in <xref ref-type="fig" rid="F2">Figure 2</xref>. Most importantly, it does not leave enough time for the researcher to take another image because many cellular dynamic response processes are rare and occur quickly. On the other hand, it always takes several days or weeks to reproduce the entire biological experiment again, which wastes considerable time. Therefore, the industry urgently needs a tool that can efficiently improve the quality of once-taken bad images and reconstruct high-quality microscopic images of cells.</p>
<fig id="F2" position="float">
<label>FIGURE 2</label>
<caption>
<p>Challenges in long-term live-cell imaging. Changes in individual steps or components may influence the next series of steps and cause a reduction in image quality. Trade-offs must be made between time and image quality in almost every long-term experiment.</p>
</caption>
<graphic xlink:href="fgene-13-913372-g002.tif"/>
</fig>
<p>Image processing methods such as image inpainting or image completion can be used to restore imperfect cell images. The rapid progress of deep learning technology (<xref ref-type="bibr" rid="B13">Hinton et al., 2006</xref>; <xref ref-type="bibr" rid="B14">Hinton and Salakhutdinov, 2006</xref>) and deep convolutional neural networks (CNNs) has led to many new applications in computer vision and image processes. The emergence of generative adversarial networks (GANs) (<xref ref-type="bibr" rid="B12">Goodfellow et al., 2014</xref>) has brought almost a leap in image generation, inpainting, repair, and completion. A conditional generative adversarial net (CGAN) (<xref ref-type="bibr" rid="B21">Mirza and Osindero, 2014</xref>) can generate custom outputs by adding class information to the model. The best of these methods in image processing is deep convolutional generative adversarial networks (DCGANs) proposed by <xref ref-type="bibr" rid="B30">Radford et al. (2016</xref>), which replace fully connected layers in the original GANs with the convolutional layers in both the generator and the discriminator. Recently, many excellent image repair methods based on DCGAN structures have been proposed for real-world photo restoration, such as those by <xref ref-type="bibr" rid="B25">Pathak et al. (2016</xref>), <xref ref-type="bibr" rid="B15">Iizuka et al. (2017</xref>), and <xref ref-type="bibr" rid="B50">Yu et al. (2018</xref>). These methods work very well on landscape, architecture, or portrait retouching.</p>
<p>Recently, image-to-image translation tasks have been proposed to address image style transfer, which aims to translate an input image from a source domain to a target domain. The &#x201c;pix2pix&#x201d; proposed by <xref ref-type="bibr" rid="B16">Isola et al. (2017</xref>) is an image translation method based on conditional adversarial networks, which has shown the super ability of street scene restoration in the real world. &#x201c;Pix2pix&#x201d; uses input&#x2013;output image pairs as training data, and pixel-wise reconstruction loss coupled with adversarial loss is used to optimize the model building process.</p>
<p>On the other hand, the single-image super-resolution (SISR) method has emerged to recover a high-resolution (HR) image from a single low-resolution (LR) image. Wei et al. proposed the &#x201c;artificial neural network accelerated-photoactivated localization microscopy&#x201d; (ANNA-PALM) method for reconstructing high-quality cell super-resolution views from sparse, rapidly acquired, single-molecule localization data and widefield images (<xref ref-type="bibr" rid="B24">Ouyang et al., 2018</xref>). Based on the &#x201c;pix2pix&#x201d; architecture, this method greatly facilitates studies of rare events, cellular heterogeneity, or stochastic structures. The super-resolution generative adversarial network (SRGAN) proposed by <xref ref-type="bibr" rid="B19">Ledig et al. (2017</xref>) is one of the milestones in single image super-resolution, and it significantly improves the overall visual quality of reconstruction over traditional methods. The SRGAN innovatively uses content loss coupled with adversarial loss instead of PSNR-oriented loss as the objective function. There are many variants of SRGAN methods, such as the enhanced SRGAN (ESRGAN) proposed by <xref ref-type="bibr" rid="B45">Wang et al. (2018</xref>) and the practical restoration application ESRGAN (Real-ESRGAN) proposed by <xref ref-type="bibr" rid="B44">Wang et al. (2021b</xref>). In an ESRGAN, a residual-in-residual dense block (RRDB) was introduced to the model as the basic network building unit, which combines a multilevel residual network and dense connections. The RRDB can further improve the recovered textures by adopting a deeper and more complex structure than the original residual block in the SRGAN. The Real-ESRGAN uses the U-net discriminator with spectral normalization as a modification to the ESRGAN to increase the discriminator capability and stabilize the training dynamics. Therefore, it is better at restoring most real-world images than previous works, especially low-quality web images or videos with compression degradations.</p>
<p>While the aforementioned methods perform well on macroscopic photographs such as street views, these methods do not perform well enough in the reconstruction of biological images, which require very precise fine structure recovery. Therefore, inspired by the methods in the field of image completion and image super-resolution (<xref ref-type="bibr" rid="B45">Wang et al., 2018</xref>, <xref ref-type="bibr" rid="B43">2021a</xref>, <xref ref-type="bibr" rid="B44">2021b</xref>; <xref ref-type="bibr" rid="B29">Rad et al., 2019</xref>; <xref ref-type="bibr" rid="B28">Qiao et al., 2021</xref>), we propose a cell image-enhanced generative adversarial network (referred to as CIEGAN) for image enhancement to address the challenges mentioned previously. In addition to using adversarial loss, the CIEGAN introduced perceptual losses comprising feature reconstruction loss and style reconstruction loss (<xref ref-type="bibr" rid="B11">Gatys et al., 2015</xref>; <xref ref-type="bibr" rid="B17">Johnson et al., 2016</xref>), which greatly improves the image restoration efficiency of the model. Coupled with image reconstruction loss and the total variation regulator, our method can solve various blurry problems of biological cell images. This method is very convenient and especially optimized for long-term live-cell imaging. Moreover, it can increase the imaging speed because there is no need to take more Z-axes layers for focus finding. Researchers can have more time to scan more conditions or increase the frequency of image acquisition. Furthermore, it can handle the force majeure during cell culture: cell clumping, cell bulging or blurring caused by floating dead cells, etc., even if the blur was caused by the beating of the differentiated mature cardiomyocytes. Nevertheless, the processing is fast, of low cost, and can easily be extended to other photos. It is convenient for researchers to obtain the differentiation or development trajectories of cell lines from the image stream and conduct research such as differentiation trajectory tracking, subtype search, or protein subcellular localizations (<xref ref-type="bibr" rid="B2">Aggarwal et al., 2021</xref>).</p>
<p>We applied the CIEGAN to long-term live-cell imaging of a human-induced pluripotent stem cell (hiPSC)-derived cardiomyocyte (hiPSC-CM) differentiation system, which greatly enhanced the quality of brightfield cell images. Through the comparison of results, it is found that the CIEGAN based on generative adversarial networks is better than the traditional image enhancement algorithm and can use the original blurred images to reconstruct sharper images. The information entropy of the enhanced image is increased and its resolution ratio is also significantly improved. At the same time, we also found that it is quite suitable for the enhancement of fluorescence images.</p>
</sec>
<sec sec-type="materials|methods" id="s2">
<title>Materials and Methods</title>
<p>This section describes the experimental steps and methods of the hiPSC-CM differentiation system. In addition, microscopy techniques and strategies have been used in the image data acquisition of live cells in a long-term culture. Notably, there are many challenges in the acquisition of microscopic images in long-term live-cell culture systems, and in some cases, image quality is sacrificed to balance the pros and cons. Here, the main technology and workflow of a cell image enhancement GAN are shown in detail and explained how a CIEGAN improves the sharpness of microscopic cell images. Finally, the deployment and training process of the model are also described.</p>
<sec id="s2-1">
<title>Human-Induced Pluripotent Stem Cell Culture and Differentiation</title>
<p>Our experimental system is the differentiation induction process of human pluripotent stem cells into cardiomyocytes. The main differentiation process is shown in <xref ref-type="fig" rid="F3">Figure 3</xref>, and images were captured and saved throughout the process.</p>
<fig id="F3" position="float">
<label>FIGURE 3</label>
<caption>
<p>HiPSC-CM experimental system. Stage 0 is the hiPSC seeding and growth stage, and the differentiation process starts from day 0. After stage III, the cells were fixed and stained for readout and further analysis.</p>
</caption>
<graphic xlink:href="fgene-13-913372-g003.tif"/>
</fig>
<p>First, the iPSC-18 cell line was chosen for induction experiments. iPSC-18 cells (Y00300, Takara) were routinely cultured in a PGM1 medium (CELLAPY) on growth factor&#x2013;reduced Matrigel (corning)-coated 6-well plates. iPSC-18 cells were passaged every 4&#xa0;days using EDTA (Gibco). hiPSCs were split into a CDM medium (Cauliscell Inc.) at a ratio of 1:10 before differentiation in 24-well or 96-well plates. When they attained &#x223c;80&#x2013;90% confluence, the medium was changed to a RPMI 1640 medium (RPMI, Gibco), 1x B27 without insulin (Gibco), and 100 U penicillin (Gibco), or RPMI&#x2b;B27 minus, for shorting. During the first 48&#xa0;h, hiPSCs were treated with CHIR99021 (CHIR, a WNT activator). From 48 to 72&#xa0;h (day 3), the medium was changed to RPMI&#x2b;B27 minus. During days 4&#x2013;5, RPMI&#x2b;B27 minus medium was supplemented with IWR1 (a WNT inhibitor). On day 6, IWR1 was withdrawn instead of the RPMI&#x2b;B27 minus medium. After day 7 through to the end of the differentiation process (up to 14&#xa0;days), the RPMI 1640 medium (RPMI, Gibco), 1 x B27 (Gibco), and 100&#xa0;U penicillin (Gibco) were used and refreshed every 3&#xa0;days.</p>
</sec>
<sec id="s2-2">
<title>Immunofluorescence Staining</title>
<p>After the final stage of induction (stage III: day 8&#x2013;day 12), the cells were fixed in 4% paraformaldehyde (DING GUO) for 20&#xa0;min at room temperature, permeabilized, and blocked in 3% normal donkey serum (Jackson) and 0.1% Triton X-100 for 45&#xa0;min at room temperature. Then, the cells were incubated with a cTnT antibody (Thermo, MA5-12960, use 1:300) overnight at 4&#xb0;C in PBS plus 0.1% Triton X-100 and 3% normal donkey serum (Jackson). The cells were washed with PBS and then incubated with secondary antibodies for 1&#xa0;h at 37&#xb0;C in a dark environment in PBS and 1% bovine serum albumin (BSA). Nuclei were stained with Hoechst 33342 (Yeasen) for 5&#xa0;min at room temperature.</p>
</sec>
<sec id="s2-3">
<title>Microscopic Image Acquisition</title>
<p>The irresistible degradation of image quality introduced in long-term live-cell experiments has been described previously. Errors are introduced during the culture medium changing operation on average every 24&#x2013;48&#xa0;h according to the experimental steps of the iPSC-CM differentiation system. An elaborate microscopy strategy must be carefully designed to reduce these errors as much as possible. A good microscopy strategy can maximize the image quality and speed up photographing so that higher throughput experimental image data can be obtained within one culture cycle. It is essential to find a balance between the imaging resolution ratio, experimental throughput, and imaging speed (<xref ref-type="fig" rid="F2">Figure 2</xref>).</p>
<p>Because the imaging field of view of the microscope is limited to the optical device itself and light path design, multiple scanning imaging is required to expand the field of observation view, and the whole picture is stitched after the image acquisition. Therefore, the larger the observation field to be photographed, the more time-consuming it will be. It will cost more time to scan more culture chambers in a parallel multi-condition comparison observation of cell differentiation or reprogramming studies. On the other hand, it will further reduce the imaging speed of the system to perform the Z-axis layer imaging if the three-dimensional structure needs to be observed. Therefore, it is necessary to accelerate the imaging speed of each imaging experimental cycle to ensure the acceptable frequency of observation. If a higher imaging speed is required, a narrower imaging breadth is to be obtained, and vice versa. You cannot have your cake and eat it, too.</p>
<p>In addition, there are many options for different focusing strategies in an experiment. If the fixed focus or the one-time autofocus has been chosen, the out-of-focus caused by various emergencies in the long-term live-cell imaging process cannot be handled. For example, culture chamber expansion and contraction are caused by temperature or dead cell contamination. In particular, cells are raised into multiple layers because of cell growth. At this time, each layer of cells has its own focus surface due to the overlapping of multiple cell layers. Only multiple Z-axis microscopic imaging can obtain each sharp view of the overlapping cells. Notably, different cell types have different clonal heights, and the optimal focus surface may span more than 50&#x2013;60&#xa0;&#x3bc;M or even 100&#xa0;&#x3bc;M in our iPSC-CM differentiation system.</p>
<p>On the other hand, if the autofocus every-time mode has been chosen, it is almost impossible to carry on the experiment because the focusing process will take plenty of time. Each well requires a 5 &#xd7; 5 pattern mosaic tile stitching to obtain a square field of view of approximately 6.3&#xa0;mm&#x2a;6.3&#xa0;mm according to the 96-well plates in this research. If only the center of each well is used as the focus reference point, then it will take approximately 48&#xa0;min for 96 focus points to perform high-speed hardware autofocus. Moreover, if autofocus needs to be performed on each tile, the total time spent focusing will be 25 times larger, which is overtime to an incredible 20&#xa0;h. Unfortunately, more focal points per well are required for the 24-well plates because of the larger culture area of each well. There is a way to combine a one-time autofocus strategy and multiple Z-axis imaging, then select the sharpest layer for use after imaging experiments (more images are shown in Supplementary Materials), but it also comes at the expense of time. These limits necessitate trade-offs between the imaging resolution ratio, experimental throughput, and imaging speed (<xref ref-type="fig" rid="F2">Figure 2</xref>).</p>
<p>Here, the &#x201c;Celldiscoverer 7,&#x201d; a long-term live-cell culture instrument manufactured by Carl Zeiss, is used. It has an internal incubator to ensure regular cell growth, and the cell culture environment is kept stable at 37&#xa0;&#xb0;C with 5% CO<sub>2</sub>. The ORCA-Flash 4.0 V3 digital CMOS camera is used as HD picture acquisition equipment. The effective resolution of the camera is 2,048&#x2a;2,048 pixels. The objective is a ZEISS Plan-Apochromat &#xd7;5 objective. The objective can easily handle thin and thick vessel bottoms made of glass or plastic, which is essential to the hiPSC-CM differentiation system because our cells can only grow on plastic. With a &#xd7;2 tube lens, it achieves 10x/0.35 magnification and spatial resolution. Finally, the resolution ratio of all the photos is 0.65&#xa0;&#x3bc;M per pixel.</p>
<p>For culture chambers, 96-well and 24-well plates produced by Falcon are used. A 2,048&#x2a;2,048-pixel photo can cover a square of approximately 1.33&#x2a;1.33&#xa0;mm because of the resolution ratio of 0.65&#xa0;&#x3bc;M per pixel. Therefore, the scanning imaging method was adopted for image acquisition, and the whole images were stitched after the imaging experiments. The larger the observation field is, the more mosaic tiles will be needed.</p>
<p>Multiple Z-axis layers are photographed to study the multiple Z-axis layer aggregation of the cells in the iPSC-CM differentiation process and to find the cause of blurring, and more importantly, to obtain the training data for the model. Eleven, seven, and five layers at 1.5&#xa0;&#x3bc;M, 6&#xa0;&#x3bc;M, and 18&#xa0;&#x3bc;m intervals with total vertical distances of 15&#xa0;&#x3bc;M, 36&#xa0;&#x3bc;M, and 72&#xa0;&#x3bc;M, respectively, were obtained for study (the images are shown in the Supplementary Materials).</p>
<p>Finally, the microscopic images were acquired by Carl Zeiss ZEN software version V2.5, and the images were saved in the CZI format or PNG format. A real-time microscopic image processing framework has been compiled for the long-term live-cell imaging system. It can automatically acquire and perform image preprocessing correspondingly, including image stitching and image segmentation. The segmented images will be sent to the CIEGAN for image enhancement for further analysis.</p>
</sec>
<sec id="s2-4">
<title>Cell Image Enhancement Generative Adversarial Networks</title>
<p>The deep convolutional generative adversarial network structure (<xref ref-type="bibr" rid="B12">Goodfellow et al., 2014</xref>; <xref ref-type="bibr" rid="B30">Radford et al., 2016</xref>) is adopted as the main body of the model to reconstruct high-quality and high-resolution images from low-quality microscopic cell images.</p>
<p>The GAN architecture in our model comprises a pair of generators and discriminators. Typically, the generator is trained to generate fake samples from random noise vector <italic>z</italic>. However, in our model, we take the blurred original image as the input <italic>z</italic> to enhance it. The input image flows through a pair of our carefully designed encoder-decoder-like structures in the generator. The latent implicit representation of the input image is obtained from the encoder module. The output image is precisely reconstructed using the information provided by the latent representation. On the other hand, the discriminator is trained to distinguish between the real cell images and the generated fake images. This framework can be represented as a two-player min-max game between generator <inline-formula id="inf1">
<mml:math id="m1">
<mml:mi>G</mml:mi>
</mml:math>
</inline-formula> and discriminator <inline-formula id="inf2">
<mml:math id="m2">
<mml:mi>D</mml:mi>
</mml:math>
</inline-formula> with value function <inline-formula id="inf3">
<mml:math id="m3">
<mml:mrow>
<mml:mi>V</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>D</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>G</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>:<disp-formula id="e1">
<mml:math id="m4">
<mml:mrow>
<mml:munder>
<mml:mrow>
<mml:mi>min</mml:mi>
</mml:mrow>
<mml:mi>G</mml:mi>
</mml:munder>
<mml:munder>
<mml:mrow>
<mml:mi>max</mml:mi>
</mml:mrow>
<mml:mi>D</mml:mi>
</mml:munder>
<mml:mi>V</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>D</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>G</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mi>&#x395;</mml:mi>
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>&#x223c;</mml:mo>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mrow>
<mml:mi>d</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>a</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>x</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mrow>
<mml:mi>log</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>D</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>x</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>]</mml:mo>
</mml:mrow>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mi>&#x395;</mml:mi>
<mml:mrow>
<mml:mi>z</mml:mi>
<mml:mo>&#x223c;</mml:mo>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mi>z</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>z</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mrow>
<mml:mi>log</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>D</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>G</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>z</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>]</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>.</mml:mo>
</mml:math>
<label>(1)</label>
</disp-formula>In <xref ref-type="disp-formula" rid="e1">Eq. 1</xref>, <italic>x</italic> represents the real-world high-resolution cell image examples. Discriminator <inline-formula id="inf4">
<mml:math id="m5">
<mml:mi>D</mml:mi>
</mml:math>
</inline-formula> was trained to maximize the probability of assigning the correct label to both generated enhanced samples from G and the real-world cell image examples. At the same time, the generator <inline-formula id="inf5">
<mml:math id="m6">
<mml:mi>G</mml:mi>
</mml:math>
</inline-formula> was trained to minimize <inline-formula id="inf6">
<mml:math id="m7">
<mml:mrow>
<mml:mi>log</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>D</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>G</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>z</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> simultaneously, that is, let the generated fake samples deceive the discriminator <inline-formula id="inf7">
<mml:math id="m8">
<mml:mi>D</mml:mi>
</mml:math>
</inline-formula> to the maximum extent.</p>
<p>In the GAN structure, only the strongest generator survives in the game, which is very suitable for image restoration tasks. The adversarial loss ensures a high degree of realism of the image, making the image more natural and realistic. The following description will use <inline-formula id="inf8">
<mml:math id="m9">
<mml:mrow>
<mml:mover accent="true">
<mml:mi>x</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
</mml:mrow>
</mml:math>
</inline-formula> to represent <inline-formula id="inf9">
<mml:math id="m10">
<mml:mrow>
<mml:mrow>
<mml:mover accent="true">
<mml:mi>x</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>G</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>z</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>, the generated image samples for brevity. The adversarial loss of the discriminator is formulated as <xref ref-type="disp-formula" rid="e2">Eq. 2</xref>:<disp-formula id="e2">
<mml:math id="m11">
<mml:mrow>
<mml:msub>
<mml:mi>L</mml:mi>
<mml:mrow>
<mml:mi>a</mml:mi>
<mml:mi>d</mml:mi>
<mml:msub>
<mml:mi>v</mml:mi>
<mml:mi>D</mml:mi>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>log</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>D</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>x</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x2b;</mml:mo>
<mml:mi>log</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>D</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mover accent="true">
<mml:mi>x</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>.</mml:mo>
</mml:math>
<label>(2)</label>
</disp-formula>
</p>
<p>The two parts are the true labels for the ground truth samples and the false labels for the generated samples. The optimization objective of the adversarial loss of the discriminator is formulated as <xref ref-type="disp-formula" rid="e3">Eq. 3</xref>:<disp-formula id="e3">
<mml:math id="m12">
<mml:mrow>
<mml:munder>
<mml:mrow>
<mml:mi>max</mml:mi>
</mml:mrow>
<mml:mi>D</mml:mi>
</mml:munder>
<mml:msub>
<mml:mi>L</mml:mi>
<mml:mrow>
<mml:mi>a</mml:mi>
<mml:mi>d</mml:mi>
<mml:msub>
<mml:mi>v</mml:mi>
<mml:mi>D</mml:mi>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mi>&#x395;</mml:mi>
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>&#x223c;</mml:mo>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mrow>
<mml:mi>d</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>a</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mrow>
<mml:mi>log</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>D</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>x</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x2b;</mml:mo>
<mml:mi>log</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>D</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mover accent="true">
<mml:mi>x</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>]</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>.</mml:mo>
</mml:math>
<label>(3)</label>
</disp-formula>
</p>
<p>Similarly, the adversarial loss of the generator and its optimization objective are formulated as <xref ref-type="disp-formula" rid="e4">Eqs. 4</xref>, <xref ref-type="disp-formula" rid="e5">5</xref>:<disp-formula id="e4">
<mml:math id="m13">
<mml:mrow>
<mml:msub>
<mml:mi>L</mml:mi>
<mml:mrow>
<mml:mi>a</mml:mi>
<mml:mi>d</mml:mi>
<mml:msub>
<mml:mi>v</mml:mi>
<mml:mi>G</mml:mi>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>log</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>D</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mover accent="true">
<mml:mi>x</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>,</mml:mo>
</mml:math>
<label>(4)</label>
</disp-formula>
<disp-formula id="e5">
<mml:math id="m14">
<mml:mrow>
<mml:munder>
<mml:mrow>
<mml:mi>min</mml:mi>
</mml:mrow>
<mml:mi>G</mml:mi>
</mml:munder>
<mml:msub>
<mml:mi>L</mml:mi>
<mml:mrow>
<mml:mi>a</mml:mi>
<mml:mi>d</mml:mi>
<mml:msub>
<mml:mi>v</mml:mi>
<mml:mi>G</mml:mi>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mi>&#x395;</mml:mi>
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>&#x223c;</mml:mo>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mrow>
<mml:mi>d</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>a</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mrow>
<mml:mi>log</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>D</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mover accent="true">
<mml:mi>x</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>]</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>.</mml:mo>
</mml:math>
<label>(5)</label>
</disp-formula>
</p>
<p>In the GAN structure, the latent representation can capture valuable information in the input images, and the rest of the details and textures are handed over to network parameters for completion and reconstruction. However, it is not enough to determine the network parameters precisely in the generator only by the adversarial loss of the GAN, and more penalties are required to generate more accurate images and perform more refined image restoration.</p>
<p>Inspired by image style transfer (<xref ref-type="bibr" rid="B16">Isola et al., 2017</xref>), single-image super-resolution (SISR) methods (<xref ref-type="bibr" rid="B49">Yang et al., 2019</xref>; <xref ref-type="bibr" rid="B22">Ooi and Ibrahim, 2021</xref>), and high photorealistic image synthesis (<xref ref-type="bibr" rid="B45">Wang et al., 2018</xref>; <xref ref-type="bibr" rid="B44">2021b</xref>), a series of image reconstruction losses are introduced to the model, such as pixel-wise loss and perceptual loss.</p>
<p>Specifically, using only reconstruction loss can reconstruct sharp images, but the generalization abilities are poor because of its pixel-wise properties. Therefore, images generated by reconstruction loss only may have excellent results superficially but suffer overfitting problems: just a single pixel translation may lead to model failure. Therefore, combining with the perceptual loss is a wise choice. The perceptual loss enables the contents and styles of the image reappearance. The reconstruction loss, also known as pixel-wise loss, is denoted as <xref ref-type="disp-formula" rid="e6">Eq. 6</xref>:<disp-formula id="e6">
<mml:math id="m15">
<mml:mrow>
<mml:msub>
<mml:mi>L</mml:mi>
<mml:mrow>
<mml:mi>r</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>c</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mrow>
<mml:mi>B</mml:mi>
<mml:mi>C</mml:mi>
<mml:mi>H</mml:mi>
<mml:mi>W</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:msubsup>
<mml:mrow>
<mml:mrow>
<mml:mo>&#x2016;</mml:mo>
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mrow>
<mml:mover accent="true">
<mml:mi>x</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
</mml:mrow>
</mml:mrow>
<mml:mo>&#x2016;</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mn>2</mml:mn>
<mml:mn>2</mml:mn>
</mml:msubsup>
</mml:mrow>
<mml:mo>.</mml:mo>
</mml:math>
<label>(6)</label>
</disp-formula>
</p>
<p>In <xref ref-type="disp-formula" rid="e6">Eq. 6</xref>, B, C, H, and W represent the training batch size, the number of channels of the image or feature map, and the height and width of the feature map, respectively.</p>
<p>The perceptual loss comprises two parts: the feature loss part and the style loss part. The feature perceptual loss is formulated as <xref ref-type="disp-formula" rid="e7">Eq. 7</xref> (<xref ref-type="bibr" rid="B11">Gatys et al., 2015</xref>; <xref ref-type="bibr" rid="B17">Johnson et al., 2016</xref>):<disp-formula id="e7">
<mml:math id="m16">
<mml:mrow>
<mml:msub>
<mml:mi>L</mml:mi>
<mml:mrow>
<mml:mi>f</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mstyle displaystyle="true">
<mml:msubsup>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>N</mml:mi>
</mml:msubsup>
<mml:mrow>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mrow>
<mml:mi>B</mml:mi>
<mml:mi>C</mml:mi>
<mml:mi>H</mml:mi>
<mml:mi>W</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:msubsup>
<mml:mrow>
<mml:mrow>
<mml:mo>&#x2016;</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="normal">&#x3a6;</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>x</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mi mathvariant="normal">&#x3a6;</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mover accent="true">
<mml:mi>x</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>&#x2016;</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mn>2</mml:mn>
<mml:mn>2</mml:mn>
</mml:msubsup>
</mml:mrow>
</mml:mstyle>
</mml:mrow>
<mml:mo>.</mml:mo>
</mml:math>
<label>(7)</label>
</disp-formula>In <xref ref-type="disp-formula" rid="e7">Eq. 7</xref>, <inline-formula id="inf10">
<mml:math id="m17">
<mml:mrow>
<mml:msub>
<mml:mi>&#x3a6;</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> is the <italic>i</italic>-th layer of a pre-trained VGG-16 or VGG-19 network (<xref ref-type="bibr" rid="B34">Simonyan and Zisserman, 2015</xref>), and <inline-formula id="inf11">
<mml:math id="m18">
<mml:mrow>
<mml:msub>
<mml:mi>&#x3a6;</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>x</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> is the feature map of input image <italic>x</italic>. In the actual data flow, the shape of the feature map is the same as mentioned previously: <inline-formula id="inf12">
<mml:math id="m19">
<mml:mrow>
<mml:mi>B</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mi>C</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mi>H</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mi>W</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>. <italic>N</italic> is the total number of VGG network layers. Here, we use the VGG-19 network, which is pre-trained on the ImageNet dataset (<xref ref-type="bibr" rid="B9">Deng et al., 2009</xref>). The style perceptual loss is formulated as <xref ref-type="disp-formula" rid="e8">Eq. 8</xref>:<disp-formula id="e8">
<mml:math id="m20">
<mml:mrow>
<mml:msub>
<mml:mi>L</mml:mi>
<mml:mrow>
<mml:mi>s</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>y</mml:mi>
<mml:mi>l</mml:mi>
<mml:mi>e</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mstyle displaystyle="true">
<mml:msubsup>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>N</mml:mi>
</mml:msubsup>
<mml:mrow>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mrow>
<mml:mi>B</mml:mi>
<mml:mi>C</mml:mi>
<mml:mi>H</mml:mi>
<mml:mi>W</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:msubsup>
<mml:mrow>
<mml:mrow>
<mml:mo>&#x2016;</mml:mo>
<mml:mrow>
<mml:mi>G</mml:mi>
<mml:mi>r</mml:mi>
<mml:mi>a</mml:mi>
<mml:msub>
<mml:mi>m</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>x</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>G</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>r</mml:mi>
<mml:msub>
<mml:mi>m</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mover accent="true">
<mml:mi>x</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>&#x2016;</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mi>F</mml:mi>
<mml:mn>2</mml:mn>
</mml:msubsup>
</mml:mrow>
</mml:mstyle>
</mml:mrow>
<mml:mo>.</mml:mo>
</mml:math>
<label>(8)</label>
</disp-formula>In <xref ref-type="disp-formula" rid="e8">Eq. 8</xref>, the Gram matrix can be calculated using the following formula: <inline-formula id="inf13">
<mml:math id="m21">
<mml:mrow>
<mml:mi>G</mml:mi>
<mml:mi>r</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>m</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>A</mml:mi>
<mml:msup>
<mml:mi>A</mml:mi>
<mml:mi>T</mml:mi>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula>, where <italic>A</italic> represents a matrix. Here, the use of the squared Frobenius norm instead of the squared Euclidean distance was used before.</p>
<p>Nevertheless, a total variation regularization is imported to the model to remove noise and mosaics from images and further reduce the spikey artifacts of the generated images. The total variation regulator is formulated as <xref ref-type="disp-formula" rid="e9">Eq. 9</xref>:<disp-formula id="e9">
<mml:math id="m22">
<mml:mrow>
<mml:msub>
<mml:mi>L</mml:mi>
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mi>v</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mstyle displaystyle="true">
<mml:munderover>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>j</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>H</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>W</mml:mi>
</mml:mrow>
</mml:munderover>
<mml:mrow>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mrow>
<mml:mi>B</mml:mi>
<mml:mi>C</mml:mi>
<mml:mi>H</mml:mi>
<mml:mi>W</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msubsup>
<mml:mrow>
<mml:mrow>
<mml:mo>&#x2016;</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mi>x</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mi>x</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>&#x2016;</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mn>2</mml:mn>
<mml:mn>2</mml:mn>
</mml:msubsup>
<mml:mo>&#x2b;</mml:mo>
<mml:msubsup>
<mml:mrow>
<mml:mrow>
<mml:mo>&#x2016;</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mi>x</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>j</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mi>x</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>&#x2016;</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mn>2</mml:mn>
<mml:mn>2</mml:mn>
</mml:msubsup>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mstyle>
</mml:mrow>
<mml:mo>.</mml:mo>
</mml:math>
<label>(9)</label>
</disp-formula>
</p>
<p>In <xref ref-type="disp-formula" rid="e9">Eq. 9</xref>, <inline-formula id="inf14">
<mml:math id="m23">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mi>x</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> represents a pixel from the generated enhanced image. Here, the rows and columns are calculated separately by the difference between adjacent pixels.</p>
<p>Finally, the loss of the CIEGAN is divided into two parts: the discriminator loss <inline-formula id="inf15">
<mml:math id="m24">
<mml:mrow>
<mml:msub>
<mml:mi>L</mml:mi>
<mml:mi>D</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> and the generator loss <inline-formula id="inf16">
<mml:math id="m25">
<mml:mrow>
<mml:msub>
<mml:mi>L</mml:mi>
<mml:mi>G</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, shown as <xref ref-type="disp-formula" rid="e10">Eqs. 10</xref>, <xref ref-type="disp-formula" rid="e11">11</xref>, respectively:<disp-formula id="e10">
<mml:math id="m26">
<mml:mrow>
<mml:msub>
<mml:mi>L</mml:mi>
<mml:mi>D</mml:mi>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mi>L</mml:mi>
<mml:mrow>
<mml:mi>a</mml:mi>
<mml:mi>d</mml:mi>
<mml:msub>
<mml:mi>v</mml:mi>
<mml:mi>D</mml:mi>
</mml:msub>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>,</mml:mo>
</mml:math>
<label>(10)</label>
</disp-formula>
<disp-formula id="e11">
<mml:math id="m27">
<mml:mrow>
<mml:msub>
<mml:mi>L</mml:mi>
<mml:mi>G</mml:mi>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mi>&#x3bb;</mml:mi>
<mml:mrow>
<mml:mi>r</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>c</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mi>L</mml:mi>
<mml:mrow>
<mml:mi>r</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>c</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mi>&#x3bb;</mml:mi>
<mml:mrow>
<mml:mi>f</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mi>L</mml:mi>
<mml:mrow>
<mml:mi>f</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mi>&#x3bb;</mml:mi>
<mml:mrow>
<mml:mi>s</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>y</mml:mi>
<mml:mi>l</mml:mi>
<mml:mi>e</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mi>L</mml:mi>
<mml:mrow>
<mml:mi>s</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>y</mml:mi>
<mml:mi>l</mml:mi>
<mml:mi>e</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mi>&#x3bb;</mml:mi>
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mi>v</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mi>L</mml:mi>
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mi>v</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mi>&#x3bb;</mml:mi>
<mml:mrow>
<mml:mi>a</mml:mi>
<mml:mi>d</mml:mi>
<mml:msub>
<mml:mi>v</mml:mi>
<mml:mi>G</mml:mi>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mi>L</mml:mi>
<mml:mrow>
<mml:mi>a</mml:mi>
<mml:mi>d</mml:mi>
<mml:msub>
<mml:mi>v</mml:mi>
<mml:mi>G</mml:mi>
</mml:msub>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>.</mml:mo>
</mml:math>
<label>(11)</label>
</disp-formula>The corresponding coefficients <inline-formula id="inf17">
<mml:math id="m28">
<mml:mi>&#x3bb;</mml:mi>
</mml:math>
</inline-formula> are added in front of different losses in the generator loss <inline-formula id="inf18">
<mml:math id="m29">
<mml:mrow>
<mml:msub>
<mml:mi>L</mml:mi>
<mml:mi>G</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> to balance the weights of different losses.</p>
<p>These losses and regularizations are merged together to reconstruct high-quality images and are referred to as the combined loss shown in <xref ref-type="fig" rid="F4">Figure 4</xref>. The main structure of the CIEGAN model and the training and testing processes are depicted.</p>
<fig id="F4" position="float">
<label>FIGURE 4</label>
<caption>
<p>Network architecture and the main pipeline of the CIEGAN. The training process of the model is shown in the blue box, and the testing process of the model is shown in the green box in the lower-left corner, where the generator is shared. Once the model is trained, it can predict a sharp image from the original imperfect cell image.</p>
</caption>
<graphic xlink:href="fgene-13-913372-g004.tif"/>
</fig>
</sec>
<sec id="s2-5">
<title>Model Building and Training</title>
<p>CIEGAN model coding mainly uses the TensorFlow deep learning framework (<xref ref-type="bibr" rid="B1">Abadi et al., 2016</xref>) and TF-slim library to build our generative adversarial nets. A cloud computing environment is used for model training and testing. The main hardware configuration list is an Intel Xeon Cascade Lake (2.5&#xa0;GHz) 8-core processor, 32&#xa0;GB of memory, and the CUDA computational acceleration unit is an NVIDIA T4 (with 15&#xa0;GB of video memory).</p>
<p>Then, 128&#x2a;128-pixel images, 256&#x2a;256-pixel images, and 512&#x2a;512-pixel images are successively tested on the CIEGAN model. According to the memory size of the CUDA unit, the network was trained using a batch size of 32, 20, and 4 images for the 128&#x2a;128-pixel, 256&#x2a;256-pixel, and 512&#x2a;512-pixel inputs, respectively. Finally, the combination of 256&#x2a;256-pixel image size and a batch size of 20 is chosen for the final training process according to the results.</p>
<p>The datasets used in training and testing come from two sources: 1) the original data are obtained from multiple Z-axis layers with an out-of-focus and sharp focus for each field of view. 2) Additional data were generated with a Gaussian blur from the original high-definition image to simulate the out-of-focus effect. In this way, more samples can be generated. Finally, a pair of blurry and high-resolution images of the same field of view are input into the model for training.</p>
<p>To ensure a stable and efficient training process and make the generator and discriminator converge, a multistep training strategy is adopted. 1) First, the generator is trained so that it can output primary-quality images. 2) Then, the discriminator is trained to identify fake images generated by the generator. 3) Finally, fire up the game and start the game process between the generator and the discriminator.</p>
<p>Initially, approximately 4,200 brightfield live-cell images (256&#x2a;256 pixels) of the iPSC-CM differentiation process were used in the model test. The CIEGAN model only takes a few hours to achieve decent results on an NVIDIA T4. The time spent on training varies depending on the size of the training set in other applications. However, the use of the CIEGAN model is very fast; it can run 128 images (256&#x2a;256 pixels) at a time of only a few minutes on VIDIA T4. Most of this is the load time of the model checkpoints. Once the model is loaded, its prediction timeliness is comparable to real-time processing. Nevertheless, our program provides automatic segmentation and assembly to handle larger input images. Finally, the performance of the CIEGAN should be similar to the same level of the CUDA computational acceleration unit. For example, it should have approximately the same time cost as this article on an NVIDIA 1080Ti GPU (12&#xa0;GB).</p>
</sec>
</sec>
<sec sec-type="results" id="s3">
<title>Results</title>
<p>In this section, the CIEGAN is applied to long-term live-cell imaging of iPSC-CM differentiation. It significantly facilitates the research works by enhancing time-lapse microscopy images and carrying out the next analysis work. The CIEGAN successively enhanced the brightfield image of induced cardiomyocytes and obtained many good results. Then, the method is extended to the enhancement of fluorescence images, and the results are promising. Finally, several other similar methods are compared and public databases are used to explore the practicality and scalability of these methods.</p>
<sec id="s3-1">
<title>Brightfield Image Enhancement</title>
<p>First, the brightfield images in the hiPSC-CM differentiation process are enhanced for qualitative testing and the results are shown in <xref ref-type="fig" rid="F5">Figure 5</xref>. A variety of cell morphologies are selected to test the robustness of this method. The thickness of the observed cells varies from flat monolayer to three-dimensional structures.</p>
<fig id="F5" position="float">
<label>FIGURE 5</label>
<caption>
<p>Brightfield cell image enhancement results of iPSC-CM differentiation experiments. The results for different cell morphologies are shown in subfigures <bold>(A)</bold>, <bold>(B)</bold>, <bold>(C)</bold>, and <bold>(D)</bold>. TE stands for the traditional enhancement method, CIEGAN, CIEGAN plus (CIEGANP) is our method, GT is the ground truth, and EGT is the enhanced ground truth. <bold>(E)</bold> Boxplot comparison results of resolution, MS-SSIM, mutual information (MI) entropy, PSNR, and NRMSE (<italic>n</italic> &#x3d; 1128).</p>
</caption>
<graphic xlink:href="fgene-13-913372-g005.tif"/>
</fig>
<p>In the results, the traditional enhancement method (TE) is used to enhance the blurred input images for comparison. The TE method adopts the combination of the unsharp masking (<xref ref-type="bibr" rid="B27">Polesel et al., 2000</xref>; <xref ref-type="bibr" rid="B8">Deng, 2010</xref>) and contrast limited adaptive histogram equalization (CLAHE) (<xref ref-type="bibr" rid="B26">Pisano et al., 1998</xref>; <xref ref-type="bibr" rid="B31">Reza, 2004</xref>). Unsharp masking and CLAHE are classical tools for sharpness enhancement that can adaptively adjust and enhance the sharpness and contrast of the image, respectively. Here, the TE columns in <xref ref-type="fig" rid="F5">Figure 5</xref> show the results of the enhancement of the input cell images. Although the brightness and contrast of the images have been significantly improved through the enhancement of the traditional method, the blurring problem of the image has not been fundamentally solved. However, our CIEGAN method outperforms the traditional methods and benefit from the adversarial process and perceptual loss. The results of the CIEGAN are very close to the ground truth (GT), especially on the reconstruction of the fine structure of cells.</p>
<p>On the other hand, inspired by the Real-ESRGAN (Real Enhanced Super-Resolution Generative Adversarial Network) (<xref ref-type="bibr" rid="B44">Wang et al., 2021b</xref>), an improved CIEGAN was trained by using an enhanced version of the training sets, which we call CIEGAN plus (CIEGANP). The difference between them is the training inputs; the CIEGANP model is trained with enhanced ground truth images. Here, enhanced ground truth (EGT) is the result of image enhancement using unsharp masking and CLAHE methods for the GT. Interestingly, the image enhanced by the EGT method highlights the dead cells in the image (black dots) because the size of the dead cells or impurities is much smaller than the cells. <xref ref-type="fig" rid="F5">Figure 5</xref> (A&#x2013;D) shows that the result of the CIEGANP is better than that of the CIEGAN in brightness and sharpness. Moreover, the results of the CIEGANP have less pepper noise or spikey artifacts than the enhanced ground truth (EGT) due to the introduced variation regulator.</p>
<p>To further evaluate the method, we performed a package of quantitative evaluations between traditional methods. The normalized root mean square error (NRMSE), peak signal-to-noise ratio (PSNR), and multi-scale structural similarity index (MS-SSIM) (<xref ref-type="bibr" rid="B46">Wang et al., 2003</xref>) are used in the image similarity assessment between the generated image and ground truth. The NRMSE reflects the pixel difference between the two images, and the smaller the value is, the better. The PSNR is the ratio of the maximum possible power of a signal to the power of corrupting noise that affects representation fidelity, which can objectively measure the image quality; the larger the value is, the better. The MS-SSIM is an improved version of the SSIM that is also used to measure image quality; the closer the value is to 1, the better. In addition, mutual information (MI) is used to measure the similarity of two images. The mutual information <inline-formula id="inf19">
<mml:math id="m30">
<mml:mi>I</mml:mi>
</mml:math>
</inline-formula> between two pictures is formulated as <xref ref-type="disp-formula" rid="e12">Eq. 12</xref>:<disp-formula id="e12">
<mml:math id="m31">
<mml:mrow>
<mml:mi>I</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>X</mml:mi>
<mml:mo>;</mml:mo>
<mml:mi>Y</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mstyle displaystyle="true">
<mml:munder>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>y</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mi>Y</mml:mi>
</mml:mrow>
</mml:munder>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:munder>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mi>X</mml:mi>
</mml:mrow>
</mml:munder>
<mml:mrow>
<mml:mi>p</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mi>log</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mfrac>
<mml:mrow>
<mml:mi>p</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mrow>
<mml:mi>p</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>x</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mi>p</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>y</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mstyle>
</mml:mrow>
</mml:mstyle>
</mml:mrow>
<mml:mo>.</mml:mo>
</mml:math>
<label>(12)</label>
</disp-formula>
</p>
<p>Mutual information in <xref ref-type="disp-formula" rid="e12">Eq. 12</xref> describes the reciprocity between objects in two images, and the larger the value is, the higher the similarity between the two images will be.</p>
<p>We also used some no-reference methods for a single image quality evaluation in addition to the full reference method. The information gain and the estimation of the resolution ratio are used for evaluation after image enhancement. The information entropy <inline-formula id="inf20">
<mml:math id="m32">
<mml:mi>H</mml:mi>
</mml:math>
</inline-formula> can be expressed as <xref ref-type="disp-formula" rid="e13">Eq. 13</xref>:<disp-formula id="e13">
<mml:math id="m33">
<mml:mrow>
<mml:mi>H</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>x</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>E</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>I</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>x</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#x2212;</mml:mo>
<mml:mstyle displaystyle="true">
<mml:munder>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mi>X</mml:mi>
</mml:mrow>
</mml:munder>
<mml:mrow>
<mml:mi>p</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>x</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mi>log</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>p</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>x</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mstyle>
</mml:mrow>
<mml:mo>.</mml:mo>
</mml:math>
<label>(13)</label>
</disp-formula>In <xref ref-type="disp-formula" rid="e13">Eq. 13</xref>, <italic>x</italic> represents the gray value of a pixel in the image, and <inline-formula id="inf21">
<mml:math id="m34">
<mml:mrow>
<mml:mi>p</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>x</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> is the probability of occurrence of its pixel gray value. For a normal 8-bit depth grayscale image, <italic>X</italic> contains the grayscale from 0&#x2013;255 in this image. The information entropy is a nonnegative value, that is, it is used to describe the uncertainty of the pixels in the picture. The larger the information entropy is, the greater will be the amount of information contained in the picture.</p>
<p>On the other hand, resolution ratio estimation is widely used in biological image evaluation because it can indicate the actual resolution per pixel (<xref ref-type="bibr" rid="B28">Qiao et al., 2021</xref>). The resolution ratio calculation is performed by a decorrelation analysis, where the cross-correlation coefficient is expressed as <xref ref-type="disp-formula" rid="e14">Eq. 14</xref> (<xref ref-type="bibr" rid="B10">Descloux et al., 2019</xref>):<disp-formula id="e14">
<mml:math id="m35">
<mml:mrow>
<mml:mi>d</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>r</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mrow>
<mml:mo>&#x222b;</mml:mo>
<mml:mrow>
<mml:mi>R</mml:mi>
<mml:mi>e</mml:mi>
<mml:mrow>
<mml:mo>{</mml:mo>
<mml:mrow>
<mml:mi>I</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>k</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:msub>
<mml:mi>I</mml:mi>
<mml:mi>n</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>k</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mi>M</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>r</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>}</mml:mo>
</mml:mrow>
<mml:mi>d</mml:mi>
<mml:msub>
<mml:mi>k</mml:mi>
<mml:mi>x</mml:mi>
</mml:msub>
<mml:mi>d</mml:mi>
<mml:msub>
<mml:mi>k</mml:mi>
<mml:mi>y</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mrow>
</mml:mstyle>
</mml:mrow>
<mml:mrow>
<mml:msqrt>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mrow>
<mml:mo>&#x222b;</mml:mo>
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mrow>
<mml:mo>&#x7c;</mml:mo>
<mml:mrow>
<mml:mi>I</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>k</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>&#x7c;</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mn>2</mml:mn>
</mml:msup>
<mml:mi>d</mml:mi>
<mml:msub>
<mml:mi>k</mml:mi>
<mml:mi>x</mml:mi>
</mml:msub>
<mml:mi>d</mml:mi>
<mml:msub>
<mml:mi>k</mml:mi>
<mml:mi>y</mml:mi>
</mml:msub>
<mml:mstyle displaystyle="true">
<mml:mrow>
<mml:mo>&#x222b;</mml:mo>
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mrow>
<mml:mo>&#x7c;</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>I</mml:mi>
<mml:mi>n</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>k</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mi>M</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>r</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>&#x7c;</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mn>2</mml:mn>
</mml:msup>
<mml:mi>d</mml:mi>
<mml:msub>
<mml:mi>k</mml:mi>
<mml:mi>x</mml:mi>
</mml:msub>
<mml:mi>d</mml:mi>
<mml:msub>
<mml:mi>k</mml:mi>
<mml:mi>y</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mrow>
</mml:mstyle>
</mml:mrow>
</mml:mrow>
</mml:mstyle>
</mml:mrow>
</mml:msqrt>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
<mml:mo>.</mml:mo>
</mml:math>
<label>(14)</label>
</disp-formula>In <xref ref-type="disp-formula" rid="e14">Eq. 14</xref>, <italic>k</italic> is the Fourier space coordinates, and <italic>I</italic> is the Fourier transform function. <inline-formula id="inf22">
<mml:math id="m36">
<mml:mrow>
<mml:mi>I</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>k</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> represents the Fourier transform of the input image, and <inline-formula id="inf23">
<mml:math id="m37">
<mml:mrow>
<mml:msub>
<mml:mi>I</mml:mi>
<mml:mi>n</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>k</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> represents the normalization of <inline-formula id="inf24">
<mml:math id="m38">
<mml:mrow>
<mml:mi>I</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>k</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>. <inline-formula id="inf25">
<mml:math id="m39">
<mml:mrow>
<mml:mi>M</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>r</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> is the circular mask of the radius <inline-formula id="inf26">
<mml:math id="m40">
<mml:mi>r</mml:mi>
</mml:math>
</inline-formula> (<xref ref-type="bibr" rid="B10">Descloux et al., 2019</xref>). The input image is passed through a series of high-pass filters and found the local maximum of the highest frequency; the normalized frequencies were denoted as <inline-formula id="inf27">
<mml:math id="m41">
<mml:mrow>
<mml:msub>
<mml:mi>k</mml:mi>
<mml:mi>c</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>. The resolution ratio is <inline-formula id="inf28">
<mml:math id="m42">
<mml:mrow>
<mml:mi>r</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>s</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>l</mml:mi>
<mml:mi>u</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>n</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mn>2</mml:mn>
<mml:mi>p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mi>k</mml:mi>
<mml:mi>c</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
</inline-formula>, where <italic>p</italic> is the pixel size in photo acquisition.</p>
<p>The resolution ratio can measure the recognizability of structures in biological images; the smaller the resolution ratio value is, the greater the accuracy will be. The physical resolution in the iPSC-CM experiments is 0.65&#xa0;&#x3bc;M per pixel, while the actual resolution of the ground truth images obtained may be poorer: 1.465&#xa0;&#x3bc;M on average. The input images have a resolution of 4.305&#xa0;&#x3bc;M on average due to out-of-focus, and our method can enhance this to 2.488&#xa0;&#x3bc;M by the CIEGAN on average and 1.416&#xa0;&#x3bc;M by the CIEGANP on average. Here, the traditional enhancement (TE) method only has a resolution of 3.546&#xa0;&#x3bc;M on average. The results of the boxplot comparison are shown in <xref ref-type="fig" rid="F5">Figure 5E</xref>.</p>
<p>An expert questionnaire is conducted to investigate whether the resulting pictures generated by this model are suitable for scientific research purposes. The results also show that the CIEGANP generally performs better than the other methods and has appropriate contrast and brightness.</p>
</sec>
<sec id="s3-2">
<title>DAPI Image Enhancement</title>
<p>The assessment of cardiomyocyte quality is required after the third stage of differentiation (<xref ref-type="fig" rid="F3">Figure 3</xref>) in the iPSC-CM differentiation process. Therefore, fluorescent staining experiments were performed. Here, we stained two types of cell markers: cardiomyocyte-specific cTnT antibody and the nucleus-specific Hoechst 33342, which are used to assess the differentiation ratio and the cardiomyocyte quality. The enhancement results of the cell image stained with Hoechst 33342 are shown in <xref ref-type="fig" rid="F6">Figure 6</xref>.</p>
<fig id="F6" position="float">
<label>FIGURE 6</label>
<caption>
<p>Enhancement results of Hoechst 33342 fluorescence microscopy images of the iPSC-CM differentiation experiment. TE stands for the traditional enhancement method, CIEGAN and CIEGAN plus (CIEGANP) are our methods, GT is the ground truth, and EGT is the enhanced ground truth.</p>
</caption>
<graphic xlink:href="fgene-13-913372-g006.tif"/>
</fig>
<p>The CIEGAN algorithm has significantly enhanced the sharpness and contrast of the blurred cell staining images. It even obtained a higher signal-to-noise ratio than the ground truth. However, the performance in brightness is not perfect, which is a common problem of this model. The reason for this is that there are many black images with nothing in the training set. For this reason, the model trained with sharped ground-truth images as the CIEGAN plus has been introduced, and it significantly improves the sharpness and brightness of generating biological images while increasing the signal-to-noise ratio.</p>
</sec>
<sec id="s3-3">
<title>CTnT Image Enhancement</title>
<p>The enhanced cTnT fluorescent stained images are shown in <xref ref-type="fig" rid="F7">Figure 7</xref>. The CIEGAN model achieves excellent generalization performance for different types of fluorescent images. The phototoxicity can be ignored in this experiment because the fluorescent staining method kills cells. Longer and stronger exposures can be used for imaging. However, photobleaching cannot be ignored because of the poor stability of dyes under strong light irradiation. It is not possible to use a strong intensity of light for a long time during the exposure process. Therefore, the balance between exposure time, light intensity, and image clarity also needs to be considered when taking photographs of fluorescence microscopy images. Nevertheless, it is sometimes impossible to obtain fluorescent photographs again because of severe photobleaching. In this case, the CIEGANP can not only deal with various out-of-focus images but can also enhance images with a low signal-to-noise ratio due to photobleaching.</p>
<fig id="F7" position="float">
<label>FIGURE 7</label>
<caption>
<p>Results of cTnT fluorescence microscopy images of the iPSC-CM differentiation experiment. TE stands for the traditional enhancement method, CIEGAN, CIEGAN plus (CIEGANP) is our method, GT is the ground truth, and EGT is the enhanced ground truth.</p>
</caption>
<graphic xlink:href="fgene-13-913372-g007.tif"/>
</fig>
</sec>
<sec id="s3-4">
<title>Comparison With Other Methods</title>
<p>The comparison with other methods is also carried out here. The &#x201c;pix2pix&#x201d; proposed by <xref ref-type="bibr" rid="B16">Isola et al. (2017</xref>) and the Real-ESRGAN proposed by <xref ref-type="bibr" rid="B44">Wang et al. (2021b</xref>) are used for comparison. The comparison results are shown in <xref ref-type="fig" rid="F8">Figure 8</xref>.</p>
<fig id="F8" position="float">
<label>FIGURE 8</label>
<caption>
<p>Comparison results of brightfield image enhancement of the iPSC-CM differentiation experiment. TE stands for the traditional enhancement method, CIEGAN, CIEGAN plus (CIEGANP) is our method, GT is the ground truth, and EGT is the enhanced ground truth. &#x201c;pix2pix&#x201d; is the method proposed by <xref ref-type="bibr" rid="B16">Isola et al. (2017</xref>) and RESRGAN is the real-ESRGAN method proposed by <xref ref-type="bibr" rid="B44">Wang et al. (2021b</xref>).</p>
</caption>
<graphic xlink:href="fgene-13-913372-g008.tif"/>
</fig>
<p>First, the &#x201c;pix2pix&#x201d; model is trained on the dataset of brightfield images in the hiPSC-CM experiments and then achieves convergence. The results of &#x201c;pix2pix&#x201d; show its excellent performance, but there are a small number of artifacts and compression blur. On the other hand, the Real-ESRGAN shows its remarkable performance in real-world photos. Here, the Real-ESRGAN can increase the input biological image from 256&#x2a;256 pixels to 1,024&#x2a;1,024 pixels, which is 16 times the quantity of pixels. Because it was designed to perform super-resolution enhancement instead of dealing with blur degradations in biological micrographs, the high-definition pictures generated by the Real-ESRGAN are biologically distorted.</p>
</sec>
</sec>
<sec sec-type="discussion" id="s4">
<title>Discussion</title>
<p>To overcome challenges in long-term live-cell imaging, this study proposes a cell image-enhanced generative adversarial network (CIEGAN). This method can resolve various blurred degradations in biological brightfield cell images and significantly improve the image space resolution ratio. It can maximize the effectiveness of the information mining of the biological image. Needless to say, it is very convenient to make the blurred images sharp again with a few steps. Moreover, it accelerated the imaging speed because there is no need to take multiple Z-axis layers to prevent out-of-focus problems. It creates more time for experimental throughput, so researchers can investigate more conditions or increase the frequency of image acquisition. Most importantly, many cellular dynamic response processes are rare and quick and do not give us a second chance to recapture the study of cell differentiation and reprogramming. Here, the CIEGAN can give researchers a second chance to reproduce sharp biological images in a short time. Furthermore, it can handle imaging mishaps during cell culture: cell clumping bulging, blurring caused by floating dead cells or the poor clarity of the medium, out-of-focus problems caused by thermal expansion and contraction of the culture chamber, and even the blur caused by the beating of differentiated mature cardiomyocytes, etc. Nevertheless, the image enhancement process is fast, of low cost, and can easily be extended to other applications. It is convenient for researchers to reproduce the developmental trajectories of cell lines from long-term time-lapse unstable image streams.</p>
<p>On the other hand, the blurred cTnT staining results of myocardial cells could be enhanced by the aforementioned method. It is necessary to photograph Z-stacks to ensure full-field vision, as monolayer cardiomyocytes are still stereoscopic (<xref ref-type="bibr" rid="B6">Christiansen et al., 2018</xref>). The CIEGAN can obtain clear cTnT staining images from single-layer imaging and reduce the requirements and complexity of microscopic photography. In addition, sharp images have more cell features, such as the sarcomere structure of cardiomyocytes, which can indicate the state of maturity of the cardiomyocytes (<xref ref-type="bibr" rid="B40">Veerman et al., 2015</xref>). This method can not only be applied to cardiomyocytes but also to enhance the image of other cells, such as neurons, hepatocytes, adipocytes, etc., which will obtain more valuable information for biological-image study for further application.</p>
<p>Notably, once the deep learning model was trained, the model performed well on the same cell type of microscopic images. It is best to retrain the model to generalize other types of cells. The performance of the model is positively correlated with the sharpness of the input training examples. Therefore, researchers cannot expect this model to perform well on poor training data sets. This model also has some common limitations similar to other deep learning models. Because image transformation with deep learning models is not perfect in any way, real-world situations tend to be more complex (<xref ref-type="bibr" rid="B4">Cai et al., 2019</xref>; <xref ref-type="bibr" rid="B49">Yang et al., 2019</xref>; <xref ref-type="bibr" rid="B28">Qiao et al., 2021</xref>). The deep learning model cannot predict new or unseen fine structures limited by the image morphology and granularity of the training set, which is also a great challenge faced by the industry. Therefore, improving the quality of the first-hand images obtained by the microscope is a fundamental and indispensable part of biological studies. On the other hand, the method proposed in this study can improve the image quality in long-term living cell images to its best. It is very helpful in saving time, especially in long-term live-cell imaging with long experimental periods. Because it is impossible to repeat photograph processing due to the rare phenomenon of photobleaching, another time-consuming biological experiment must be restarted.</p>
<p>In further research, the CIEGAN will be improved by introducing more advanced generator structures or more penalty functions. U-net is becoming widely used in deep learning processing schemes for biological image processing (<xref ref-type="bibr" rid="B23">Ounkomol et al., 2018</xref>; <xref ref-type="bibr" rid="B47">Weigert et al., 2018</xref>; <xref ref-type="bibr" rid="B18">Kandel et al., 2020</xref>; <xref ref-type="bibr" rid="B7">Dance, 2021</xref>; <xref ref-type="bibr" rid="B48">Wieslander et al., 2021</xref>). It is widely implemented in image segmentation and classification thanks to its structure of directly copying the feature maps of convolutional layers to deconvolutional layers (<xref ref-type="bibr" rid="B32">Ronneberger et al., 2015</xref>). We can try to introduce this network mechanism into our model and in addition, the concept of the network structure of GoogLeNet (<xref ref-type="bibr" rid="B38">Szegedy et al., 2015</xref>, <xref ref-type="bibr" rid="B39">2016</xref>, <xref ref-type="bibr" rid="B37">2017</xref>). The method in this study mainly uses multiple image difference losses as the training criteria for the GAN generator, and more losses could be tried in the next step.</p>
<p>The CIEGAN method has high scalability and broad application prospects in image enhancement scenarios, which can help biologists observe and investigate image phenomena in the process of cell differentiation and reprogramming more intuitively and deeply. In turn, more efficient experimental models can be designed, and even effective potential treatments for related diseases can be found. We will continue to refine the application of the CIEGAN method to more image enhancement scenarios.</p>
</sec>
</body>
<back>
<sec id="s5" sec-type="data-availability">
<title>Data Availability Statement</title>
<p>The original contributions presented in the study are included in the article/<xref ref-type="sec" rid="s10">Supplementary Material</xref>; further inquiries can be directed to the corresponding authors.</p>
</sec>
<sec id="s6">
<title>Author Contributions</title>
<p>All authors participated in the conception and experimental work of the study. YL provided method guidance and gave guidance for this manuscript writing. YZ raised scientific questions and guided the biological experiments. XY performed all the biological experiments: cell culture and differentiation induction of hiPSCs. In addition, XY put forward many suggestions for the improvement of this tool, and feedback on its use. JG modified some deep learning algorithms and made many valuable suggestions for model training and testing. The long-term living-cell microscopic imaging experimental design and image acquisition and preprocessing tasks were mainly completed by QS. Finally, QS carried out the construction of the mainframe and was a major contributor to algorithm coding and manuscript writing. All authors read and approved the final manuscript.</p>
</sec>
<sec id="s7">
<title>Funding</title>
<p>This work was supported in part by the National Natural Science Foundation of China (No. 31771475).</p>
</sec>
<sec sec-type="COI-statement" id="s8">
<title>Conflict of Interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="disclaimer" id="s9">
<title>Publisher&#x2019;s Note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors, and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<ack>
<p>The authors would like to thank Prof. Jue Zhang for some advice. Special thanks are extended to Prof. YZ (State Key Laboratory of Natural and Biomimetic Drugs, MOE Key Laboratory of Cell Proliferation and Differentiation, Beijing Key Laboratory of Cardiometabolic Molecular Medicine, Institute of Molecular Medicine, College of Future Technology, Peking University, Beijing, China.) for providing the experimental environment and materials, and great support for the research. Finally, thanks to everyone who contributed to this article for their efforts, especially for overcoming the inconvenience of their work during the COVID-19 pandemic.</p>
</ack>
<sec id="s10">
<title>Supplementary Material</title>
<p>The Supplementary Material for this article can be found online at: <ext-link ext-link-type="uri" xlink:href="https://www.frontiersin.org/articles/10.3389/fgene.2022.913372/full#supplementary-material">https://www.frontiersin.org/articles/10.3389/fgene.2022.913372/full&#x23;supplementary-material</ext-link>
</p>
<supplementary-material xlink:href="DataSheet2.PDF" id="SM1" mimetype="application/PDF" xmlns:xlink="http://www.w3.org/1999/xlink"/>
<supplementary-material xlink:href="DataSheet1.PDF" id="SM2" mimetype="application/PDF" xmlns:xlink="http://www.w3.org/1999/xlink"/>
</sec>
<ref-list>
<title>Reference</title>
<ref id="B1">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Abadi</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Barham</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Davis</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Dean</surname>
<given-names>J.</given-names>
</name>
<etal/>
</person-group> (<year>2016</year>). &#x201c;<article-title>TensorFlow: A System for Large-Scale Machine Learning</article-title>,&#x201d; in <conf-name>USENIX Symposium on Operating Systems Design and Implementation</conf-name>. Editors <person-group person-group-type="editor">
<name>
<surname>Keeton</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Roscoe</surname>
<given-names>T.</given-names>
</name>
</person-group>, <fpage>265</fpage>&#x2013;<lpage>283</lpage>. <comment>Available at: <ext-link ext-link-type="uri" xlink:href="https://www.usenix.org/conference/osdi16/technical-sessions/presentation/abadi">https://www.usenix.org/conference/osdi16/technical-sessions/presentation/abadi</ext-link>.</comment> </citation>
</ref>
<ref id="B2">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Aggarwal</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Gupta</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Ahuja</surname>
<given-names>R.</given-names>
</name>
</person-group> (<year>2021</year>). &#x201c;<article-title>A Review on Protein Subcellular Localization Prediction Using Microscopic Images</article-title>,&#x201d; in <conf-name>2021 6th International Conference on Signal Processing, Computing and Control (ISPCC)</conf-name> (<publisher-loc>Washington, D.C., United States</publisher-loc>: <publisher-name>IEEE</publisher-name>), <fpage>72</fpage>&#x2013;<lpage>77</lpage>. <pub-id pub-id-type="doi">10.1109/ispcc53510.2021.9609437</pub-id> </citation>
</ref>
<ref id="B3">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Buggenthin</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Buettner</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Hoppe</surname>
<given-names>P. S.</given-names>
</name>
<name>
<surname>Endele</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Kroiss</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Strasser</surname>
<given-names>M.</given-names>
</name>
<etal/>
</person-group> (<year>2017</year>). <article-title>Prospective Identification of Hematopoietic Lineage Choice by Deep Learning</article-title>. <source>Nat. Methods</source> <volume>14</volume>, <fpage>403</fpage>&#x2013;<lpage>406</lpage>. <pub-id pub-id-type="doi">10.1038/nmeth.4182</pub-id> </citation>
</ref>
<ref id="B4">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Cai</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Zeng</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Yong</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Cao</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>L.</given-names>
</name>
</person-group> (<year>2019</year>). &#x201c;<article-title>Toward Real-World Single Image Super-resolution: A New Benchmark and a New Model</article-title>,&#x201d; in <conf-name>Proceedings of the IEEE/CVF International Conference on Computer Vision</conf-name>, <fpage>3086</fpage>&#x2013;<lpage>3095</lpage>. <pub-id pub-id-type="doi">10.1109/iccv.2019.00318</pub-id> </citation>
</ref>
<ref id="B5">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Chen</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Han</surname>
<given-names>Q.</given-names>
</name>
<name>
<surname>Qin</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Xu</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>Y.</given-names>
</name>
<etal/>
</person-group> (<year>2010</year>). <article-title>Towards an Optimized Culture Medium for the Generation of Mouse Induced Pluripotent Stem Cells</article-title>. <source>J. Biol. Chem.</source> <volume>285</volume>, <fpage>31066</fpage>&#x2013;<lpage>31072</lpage>. <pub-id pub-id-type="doi">10.1074/jbc.m110.139436</pub-id> </citation>
</ref>
<ref id="B6">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Christiansen</surname>
<given-names>E. M.</given-names>
</name>
<name>
<surname>Yang</surname>
<given-names>S. J.</given-names>
</name>
<name>
<surname>Ando</surname>
<given-names>D. M.</given-names>
</name>
<name>
<surname>Javaherian</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Skibinski</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Lipnick</surname>
<given-names>S.</given-names>
</name>
<etal/>
</person-group> (<year>2018</year>). <article-title>In Silico labeling: Predicting Fluorescent Labels in Unlabeled Images</article-title>. <source>Cell.</source> <volume>173</volume>, <fpage>792</fpage>&#x2013;<lpage>803</lpage>. <pub-id pub-id-type="doi">10.1016/j.cell.2018.03.040</pub-id> </citation>
</ref>
<ref id="B7">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Dance</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>AI Spots Cell Structures that Humans Can&#x27;t</article-title>. <source>Nature</source> <volume>592</volume>, <fpage>154</fpage>&#x2013;<lpage>155</lpage>. <pub-id pub-id-type="doi">10.1038/d41586-021-00812-7</pub-id> </citation>
</ref>
<ref id="B8">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Deng</surname>
<given-names>G.</given-names>
</name>
</person-group> (<year>2010</year>). <article-title>A Generalized Unsharp Masking Algorithm</article-title>. <source>IEEE Trans. Image Process</source> <volume>20</volume>, <fpage>1249</fpage>&#x2013;<lpage>1261</lpage>. <pub-id pub-id-type="doi">10.1109/TIP.2010.2092441</pub-id> </citation>
</ref>
<ref id="B9">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Deng</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Dong</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Socher</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>L.-J.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Fei-Fei</surname>
<given-names>L.</given-names>
</name>
</person-group> (<year>2009</year>). &#x201c;<article-title>Imagenet: A Large-Scale Hierarchical Image Database</article-title>,&#x201d; in <conf-name>2009 IEEE conference on computer vision and pattern recognition</conf-name> (<publisher-loc>Washington, D.C., United States</publisher-loc>: <publisher-name>IEEE</publisher-name>), <fpage>248</fpage>&#x2013;<lpage>255</lpage>. <pub-id pub-id-type="doi">10.1109/cvpr.2009.5206848</pub-id> </citation>
</ref>
<ref id="B10">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Descloux</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Gru&#xdf;mayer</surname>
<given-names>K. S.</given-names>
</name>
<name>
<surname>Radenovic</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>Parameter-free Image Resolution Estimation Based on Decorrelation Analysis</article-title>. <source>Nat. Methods</source> <volume>16</volume>, <fpage>918</fpage>&#x2013;<lpage>924</lpage>. <pub-id pub-id-type="doi">10.1038/s41592-019-0515-7</pub-id> </citation>
</ref>
<ref id="B11">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Gatys</surname>
<given-names>L. A.</given-names>
</name>
<name>
<surname>Ecker</surname>
<given-names>A. S.</given-names>
</name>
<name>
<surname>Bethge</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2015</year>). <source>A Neural Algorithm of Artistic Style. <italic>CoRR</italic> abs/1508.06576</source>. <comment>Available at: <ext-link ext-link-type="uri" xlink:href="http://arxiv.org/abs/1508.06576">http://arxiv.org/abs/1508.06576</ext-link>.</comment> </citation>
</ref>
<ref id="B12">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Goodfellow</surname>
<given-names>I. J.</given-names>
</name>
<name>
<surname>Pouget-Abadie</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Mirza</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Xu</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Warde-Farley</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Ozair</surname>
<given-names>S.</given-names>
</name>
<etal/>
</person-group> (<year>2014</year>). &#x201c;<article-title>Generative Adversarial Nets</article-title>,&#x201d; in <source>Advances in Neural Information Processing Systems</source>. Editors <person-group person-group-type="editor">
<name>
<surname>Ghahramani</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Welling</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Cortes</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Lawrence</surname>
<given-names>N. D.</given-names>
</name>
<name>
<surname>Weinberger</surname>
<given-names>K. Q.</given-names>
</name>
</person-group>, <fpage>2672</fpage>&#x2013;<lpage>2680</lpage>. </citation>
</ref>
<ref id="B13">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Hinton</surname>
<given-names>G. E.</given-names>
</name>
<name>
<surname>Osindero</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Teh</surname>
<given-names>Y.-W.</given-names>
</name>
</person-group> (<year>2006</year>). <article-title>A Fast Learning Algorithm for Deep Belief Nets</article-title>. <source>Neural Comput.</source> <volume>18</volume>, <fpage>1527</fpage>&#x2013;<lpage>1554</lpage>. <pub-id pub-id-type="doi">10.1162/neco.2006.18.7.1527</pub-id> </citation>
</ref>
<ref id="B14">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Hinton</surname>
<given-names>G. E.</given-names>
</name>
<name>
<surname>Salakhutdinov</surname>
<given-names>R. R.</given-names>
</name>
</person-group> (<year>2006</year>). <article-title>Reducing the Dimensionality of Data with Neural Networks</article-title>. <source>science</source> <volume>313</volume>, <fpage>504</fpage>&#x2013;<lpage>507</lpage>. <pub-id pub-id-type="doi">10.1126/science.1127647</pub-id> </citation>
</ref>
<ref id="B15">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Iizuka</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Simo-Serra</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>Ishikawa</surname>
<given-names>H.</given-names>
</name>
</person-group> (<year>2017</year>). <article-title>Globally and Locally Consistent Image Completion</article-title>. <source>ACM Trans. Graph.</source> <volume>36</volume>, <fpage>1</fpage>&#x2013;<lpage>14</lpage>. <pub-id pub-id-type="doi">10.1145/3072959.3073659</pub-id> </citation>
</ref>
<ref id="B16">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Isola</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Zhu</surname>
<given-names>J.-Y.</given-names>
</name>
<name>
<surname>Zhou</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Efros</surname>
<given-names>A. A.</given-names>
</name>
</person-group> (<year>2017</year>). &#x201c;<article-title>Image-to-Image Translation with Conditional Adversarial Networks</article-title>,&#x201d; in <conf-name>IEEE Conference on Computer Vision and Pattern Recognition</conf-name> (<publisher-loc>Washington, D.C., United States</publisher-loc>: <publisher-name>IEEE Computer Society</publisher-name>), <fpage>5967</fpage>&#x2013;<lpage>5976</lpage>. <pub-id pub-id-type="doi">10.1109/CVPR.2017.632</pub-id> </citation>
</ref>
<ref id="B17">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Johnson</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Alahi</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Fei-Fei</surname>
<given-names>L.</given-names>
</name>
</person-group> (<year>2016</year>). &#x201c;<article-title>Perceptual Losses for Real-Time Style Transfer and Super-resolution</article-title>,&#x201d; in <source>
<italic>Computer Vision</italic> Lecture Notes in Computer Science.</source> Editors <person-group person-group-type="editor">
<name>
<surname>Leibe</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Matas</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Sebe</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Welling</surname>
<given-names>M.</given-names>
</name>
</person-group> (<publisher-loc>Berlin, Germany</publisher-loc>: <publisher-name>Springer</publisher-name>), <fpage>694</fpage>&#x2013;<lpage>711</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-319-46475-6_43</pub-id> </citation>
</ref>
<ref id="B18">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kandel</surname>
<given-names>M. E.</given-names>
</name>
<name>
<surname>He</surname>
<given-names>Y. R.</given-names>
</name>
<name>
<surname>Lee</surname>
<given-names>Y. J.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>T. H.</given-names>
</name>
<name>
<surname>Sullivan</surname>
<given-names>K. M.</given-names>
</name>
<name>
<surname>Aydin</surname>
<given-names>O.</given-names>
</name>
<etal/>
</person-group> (<year>2020</year>). <article-title>Phase Imaging with Computational Specificity (PICS) for Measuring Dry Mass Changes in Sub-cellular Compartments</article-title>. <source>Nat. Commun.</source> <volume>11</volume>, <fpage>6256</fpage>&#x2013;<lpage>6310</lpage>. <pub-id pub-id-type="doi">10.1038/s41467-020-20062-x</pub-id> </citation>
</ref>
<ref id="B19">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Ledig</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Theis</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Husz&#xe1;r</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Caballero</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Cunningham</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Acosta</surname>
<given-names>A.</given-names>
</name>
<etal/>
</person-group> (<year>2017</year>). &#x201c;<article-title>Photo-realistic Single Image Super-resolution Using a Generative Adversarial Network</article-title>,&#x201d; in <conf-name>Proceedings of the IEEE conference on computer vision and pattern recognition</conf-name>, <fpage>4681</fpage>&#x2013;<lpage>4690</lpage>. <pub-id pub-id-type="doi">10.1109/cvpr.2017.19</pub-id> </citation>
</ref>
<ref id="B20">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>McQuate</surname>
<given-names>S. E.</given-names>
</name>
<name>
<surname>Young</surname>
<given-names>A. M.</given-names>
</name>
<name>
<surname>Silva-Herzog</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>Bunker</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>Hernandez</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>de Chaumont</surname>
<given-names>F.</given-names>
</name>
<etal/>
</person-group> (<year>2017</year>). <article-title>Long-term Live-Cell Imaging Reveals New Roles forSalmonellaeffector Proteins SseG and SteA</article-title>. <source>Cell. Microbiol.</source> <volume>19</volume>, <fpage>e12641</fpage>. <pub-id pub-id-type="doi">10.1111/cmi.12641</pub-id> </citation>
</ref>
<ref id="B21">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Mirza</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Osindero</surname>
<given-names>S.</given-names>
</name>
</person-group> (<year>2014</year>). <article-title>Conditional Generative Adversarial Nets</article-title>. <source>Comput. Sci</source>, <fpage>2672</fpage>&#x2013;<lpage>2680</lpage>. </citation>
</ref>
<ref id="B22">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ooi</surname>
<given-names>Y. K.</given-names>
</name>
<name>
<surname>Ibrahim</surname>
<given-names>H.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>Deep Learning Algorithms for Single Image Super-resolution: a Systematic Review</article-title>. <source>Electronics</source> <volume>10</volume>, <fpage>867</fpage>. <pub-id pub-id-type="doi">10.3390/electronics10070867</pub-id> </citation>
</ref>
<ref id="B23">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ounkomol</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Seshamani</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Maleckar</surname>
<given-names>M. M.</given-names>
</name>
<name>
<surname>Collman</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Johnson</surname>
<given-names>G. R.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>Label-free Prediction of Three-Dimensional Fluorescence Images from Transmitted-Light Microscopy</article-title>. <source>Nat. Methods</source> <volume>15</volume>, <fpage>917</fpage>&#x2013;<lpage>920</lpage>. <pub-id pub-id-type="doi">10.1038/s41592-018-0111-2</pub-id> </citation>
</ref>
<ref id="B24">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ouyang</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Aristov</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Lelek</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Hao</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Zimmer</surname>
<given-names>C.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>Deep Learning Massively Accelerates Super-resolution Localization Microscopy</article-title>. <source>Nat. Biotechnol.</source> <volume>36</volume>, <fpage>460</fpage>&#x2013;<lpage>468</lpage>. <pub-id pub-id-type="doi">10.1038/nbt.4106</pub-id> </citation>
</ref>
<ref id="B25">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Pathak</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Kr&#xe4;henb&#xfc;hl</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Donahue</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Darrell</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Efros</surname>
<given-names>A. A.</given-names>
</name>
</person-group> (<year>2016</year>). &#x201c;<article-title>Context Encoders: Feature Learning by Inpainting</article-title>,&#x201d; in <conf-name>IEEE Conference on Computer Vision and Pattern Recognition</conf-name> (<publisher-loc>Washington, D.C., United States</publisher-loc>: <publisher-name>IEEE Computer Society</publisher-name>), <fpage>2536</fpage>&#x2013;<lpage>2544</lpage>. <pub-id pub-id-type="doi">10.1109/CVPR.2016.278</pub-id> </citation>
</ref>
<ref id="B26">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Pisano</surname>
<given-names>E. D.</given-names>
</name>
<name>
<surname>Zong</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Hemminger</surname>
<given-names>B. M.</given-names>
</name>
<name>
<surname>DeLuca</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Johnston</surname>
<given-names>R. E.</given-names>
</name>
<name>
<surname>Muller</surname>
<given-names>K.</given-names>
</name>
<etal/>
</person-group> (<year>1998</year>). <article-title>Contrast Limited Adaptive Histogram Equalization Image Processing to Improve the Detection of Simulated Spiculations in Dense Mammograms</article-title>. <source>J. Digit. Imaging</source> <volume>11</volume>, <fpage>193</fpage>&#x2013;<lpage>200</lpage>. <pub-id pub-id-type="doi">10.1007/bf03178082</pub-id> </citation>
</ref>
<ref id="B27">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Polesel</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Ramponi</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Mathews</surname>
<given-names>V. J.</given-names>
</name>
</person-group> (<year>2000</year>). <article-title>Image Enhancement via Adaptive Unsharp Masking</article-title>. <source>IEEE Trans. Image Process.</source> <volume>9</volume>, <fpage>505</fpage>&#x2013;<lpage>510</lpage>. <pub-id pub-id-type="doi">10.1109/83.826787</pub-id> </citation>
</ref>
<ref id="B28">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Qiao</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Guo</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Jiang</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Dai</surname>
<given-names>Q.</given-names>
</name>
<etal/>
</person-group> (<year>2021</year>). <article-title>Evaluation and Development of Deep Neural Networks for Image Super-resolution in Optical Microscopy</article-title>. <source>Nat. Methods</source> <volume>18</volume>, <fpage>194</fpage>&#x2013;<lpage>202</lpage>. <pub-id pub-id-type="doi">10.1038/s41592-020-01048-5</pub-id> </citation>
</ref>
<ref id="B29">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Rad</surname>
<given-names>M. S.</given-names>
</name>
<name>
<surname>Bozorgtabar</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Marti</surname>
<given-names>U.-V.</given-names>
</name>
<name>
<surname>Basler</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Ekenel</surname>
<given-names>H. K.</given-names>
</name>
<name>
<surname>Thiran</surname>
<given-names>J.-P.</given-names>
</name>
</person-group> (<year>2019</year>). &#x201c;<article-title>Srobb: Targeted Perceptual Loss for Single Image Super-resolution</article-title>,&#x201d; in <conf-name>Proceedings of the IEEE/CVF International Conference on Computer Vision</conf-name>, <fpage>2710</fpage>&#x2013;<lpage>2719</lpage>. <pub-id pub-id-type="doi">10.1109/iccv.2019.00280</pub-id> </citation>
</ref>
<ref id="B30">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Radford</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Metz</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Chintala</surname>
<given-names>S.</given-names>
</name>
</person-group> (<year>2016</year>). &#x201c;<article-title>Unsupervised Representation Learning with Deep Convolutional Generative Adversarial Networks</article-title>,&#x201d; in <conf-name>International Conference on Learning Representations</conf-name>. Editors <person-group person-group-type="editor">
<name>
<surname>Bengio</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>LeCun</surname>
<given-names>Y.</given-names>
</name>
</person-group>. <comment>Available at: <ext-link ext-link-type="uri" xlink:href="http://arxiv.org/abs/1511.06434">http://arxiv.org/abs/1511.06434</ext-link>.</comment> </citation>
</ref>
<ref id="B31">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Reza</surname>
<given-names>A. M.</given-names>
</name>
</person-group> (<year>2004</year>). <article-title>Realization of the Contrast Limited Adaptive Histogram Equalization (CLAHE) for Real-Time Image Enhancement</article-title>. <source>J. VLSI Signal Processing-Systems Signal, Image, Video Technol.</source> <volume>38</volume>, <fpage>35</fpage>&#x2013;<lpage>44</lpage>. <pub-id pub-id-type="doi">10.1023/b:vlsi.0000028532.53893.82</pub-id> </citation>
</ref>
<ref id="B32">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Ronneberger</surname>
<given-names>O.</given-names>
</name>
<name>
<surname>Fischer</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Brox</surname>
<given-names>T.</given-names>
</name>
</person-group> (<year>2015</year>). &#x201c;<article-title>U-net: Convolutional Networks for Biomedical Image Segmentation</article-title>,&#x201d; in <conf-name>International Conference on Medical image computing and computer-assisted intervention</conf-name> (<publisher-loc>Berlin, Germany</publisher-loc>: <publisher-name>Springer</publisher-name>), <fpage>234</fpage>&#x2013;<lpage>241</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-319-24574-4_28</pub-id> </citation>
</ref>
<ref id="B33">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Schroeder</surname>
<given-names>T.</given-names>
</name>
</person-group> (<year>2011</year>). <article-title>Long-term Single-Cell Imaging of Mammalian Stem Cells</article-title>. <source>Nat. Methods</source> <volume>8</volume>, <fpage>S30</fpage>&#x2013;<lpage>S35</lpage>. <pub-id pub-id-type="doi">10.1038/nmeth.1577</pub-id> </citation>
</ref>
<ref id="B34">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Simonyan</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Zisserman</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2015</year>). &#x201c;<article-title>Very Deep Convolutional Networks for Large-Scale Image Recognition</article-title>,&#x201d; in <conf-name>International Conference on Learning Representations</conf-name>. Editors <person-group person-group-type="editor">
<name>
<surname>Bengio</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>LeCun</surname>
<given-names>Y.</given-names>
</name>
</person-group>. <comment>Available at: <ext-link ext-link-type="uri" xlink:href="http://arxiv.org/abs/1409.1556">http://arxiv.org/abs/1409.1556</ext-link>.</comment> </citation>
</ref>
<ref id="B35">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Skylaki</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Hilsenbeck</surname>
<given-names>O.</given-names>
</name>
<name>
<surname>Schroeder</surname>
<given-names>T.</given-names>
</name>
</person-group> (<year>2016</year>). <article-title>Challenges in Long-Term Imaging and Quantification of Single-Cell Dynamics</article-title>. <source>Nat. Biotechnol.</source> <volume>34</volume>, <fpage>1137</fpage>&#x2013;<lpage>1144</lpage>. <pub-id pub-id-type="doi">10.1038/nbt.3713</pub-id> </citation>
</ref>
<ref id="B36">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Smith</surname>
<given-names>Z. D.</given-names>
</name>
<name>
<surname>Nachman</surname>
<given-names>I.</given-names>
</name>
<name>
<surname>Regev</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Meissner</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2010</year>). <article-title>Dynamic Single-Cell Imaging of Direct Reprogramming Reveals an Early Specifying Event</article-title>. <source>Nat. Biotechnol.</source> <volume>28</volume>, <fpage>521</fpage>&#x2013;<lpage>526</lpage>. <pub-id pub-id-type="doi">10.1038/nbt.1632</pub-id> </citation>
</ref>
<ref id="B37">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Szegedy</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Ioffe</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Vanhoucke</surname>
<given-names>V.</given-names>
</name>
<name>
<surname>Alemi</surname>
<given-names>A. A.</given-names>
</name>
</person-group> (<year>2017</year>). &#x201c;<article-title>Inception-v4, Inception-Resnet and the Impact of Residual Connections on Learning</article-title>,&#x201d; in <conf-name>Thirty-first AAAI conference on artificial intelligence</conf-name>. </citation>
</ref>
<ref id="B38">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Szegedy</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Jia</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Sermanet</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Reed</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Anguelov</surname>
<given-names>D.</given-names>
</name>
<etal/>
</person-group> (<year>2015</year>). &#x201c;<article-title>Going Deeper with Convolutions</article-title>,&#x201d; in <conf-name>Proceedings of the IEEE conference on computer vision and pattern recognition</conf-name>, <fpage>1</fpage>&#x2013;<lpage>9</lpage>. <pub-id pub-id-type="doi">10.1109/cvpr.2015.7298594</pub-id> </citation>
</ref>
<ref id="B39">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Szegedy</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Vanhoucke</surname>
<given-names>V.</given-names>
</name>
<name>
<surname>Ioffe</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Shlens</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Wojna</surname>
<given-names>Z.</given-names>
</name>
</person-group> (<year>2016</year>). &#x201c;<article-title>Rethinking the Inception Architecture for Computer Vision</article-title>,&#x201d; in <conf-name>IEEE Conference on Computer Vision and Pattern Recognition</conf-name> (<publisher-loc>Washington, D.C., United States</publisher-loc>: <publisher-name>IEEE Computer Society</publisher-name>), <fpage>2818</fpage>&#x2013;<lpage>2826</lpage>. <pub-id pub-id-type="doi">10.1109/CVPR.2016.308</pub-id> </citation>
</ref>
<ref id="B40">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Veerman</surname>
<given-names>C. C.</given-names>
</name>
<name>
<surname>Kosmidis</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Mummery</surname>
<given-names>C. L.</given-names>
</name>
<name>
<surname>Casini</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Verkerk</surname>
<given-names>A. O.</given-names>
</name>
<name>
<surname>Bellin</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2015</year>). <article-title>Immaturity of Human Stem-Cell-Derived Cardiomyocytes in Culture: Fatal Flaw or Soluble Problem?</article-title> <source>Stem cells Dev.</source> <volume>24</volume>, <fpage>1035</fpage>&#x2013;<lpage>1052</lpage>. <pub-id pub-id-type="doi">10.1089/scd.2014.0533</pub-id> </citation>
</ref>
<ref id="B41">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Walzik</surname>
<given-names>M. P.</given-names>
</name>
<name>
<surname>Vollmar</surname>
<given-names>V.</given-names>
</name>
<name>
<surname>Lachnit</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Dietz</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Haug</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Bachmann</surname>
<given-names>H.</given-names>
</name>
<etal/>
</person-group> (<year>2015</year>). <article-title>A Portable Low-Cost Long-Term Live-Cell Imaging Platform for Biomedical Research and Education</article-title>. <source>Biosens. Bioelectron.</source> <volume>64</volume>, <fpage>639</fpage>&#x2013;<lpage>649</lpage>. <pub-id pub-id-type="doi">10.1016/j.bios.2014.09.061</pub-id> </citation>
</ref>
<ref id="B42">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Douglas</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Kumari</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Enuameh</surname>
<given-names>M. S.</given-names>
</name>
<name>
<surname>Dai</surname>
<given-names>Y.</given-names>
</name>
<etal/>
</person-group> (<year>2020</year>). <article-title>Live-cell Imaging and Analysis Reveal Cell Phenotypic Transition Dynamics Inherently Missing in Snapshot Data</article-title>. <source>Sci. Adv.</source> <volume>6</volume>, <fpage>eaba9319</fpage>. <pub-id pub-id-type="doi">10.1126/sciadv.aba9319</pub-id> </citation>
</ref>
<ref id="B43">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Shan</surname>
<given-names>Y.</given-names>
</name>
</person-group> (<year>2021a</year>). &#x201c;<article-title>Towards Real-World Blind Face Restoration with Generative Facial Prior</article-title>,&#x201d; in <conf-name>Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition</conf-name>, <fpage>9168</fpage>&#x2013;<lpage>9178</lpage>. <pub-id pub-id-type="doi">10.1109/cvpr46437.2021.00905</pub-id> </citation>
</ref>
<ref id="B44">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Xie</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Dong</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Shan</surname>
<given-names>Y.</given-names>
</name>
</person-group> (<year>2021b</year>). &#x201c;<article-title>Real-esrgan: Training Real-World Blind Super-resolution with Pure Synthetic Data</article-title>,&#x201d; in <conf-name>Proceedings of the IEEE/CVF International Conference on Computer Vision</conf-name>, <fpage>1905</fpage>&#x2013;<lpage>1914</lpage>. <pub-id pub-id-type="doi">10.1109/iccvw54120.2021.00217</pub-id> </citation>
</ref>
<ref id="B45">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Yu</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Wu</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Gu</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Dong</surname>
<given-names>C.</given-names>
</name>
<etal/>
</person-group> (<year>2018</year>). &#x201c;<article-title>Esrgan: Enhanced Super-resolution Generative Adversarial Networks</article-title>,&#x201d; in <conf-name>Proceedings of the European conference on computer vision (ECCV) workshops</conf-name>. </citation>
</ref>
<ref id="B46">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Simoncelli</surname>
<given-names>E. P.</given-names>
</name>
<name>
<surname>Bovik</surname>
<given-names>A. C.</given-names>
</name>
</person-group> (<year>2003</year>). &#x201c;<article-title>Multiscale Structural Similarity for Image Quality Assessment</article-title>,&#x201d; in <conf-name>The Thrity-Seventh Asilomar Conference on Signals, Systems &#x26; Computers</conf-name> (<publisher-loc>Washington, D.C., United States</publisher-loc>: <publisher-name>IEEE</publisher-name>), <fpage>1398</fpage>&#x2013;<lpage>1402</lpage>. </citation>
</ref>
<ref id="B47">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Weigert</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Schmidt</surname>
<given-names>U.</given-names>
</name>
<name>
<surname>Boothe</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>M&#xfc;ller</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Dibrov</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Jain</surname>
<given-names>A.</given-names>
</name>
<etal/>
</person-group> (<year>2018</year>). <article-title>Content-aware Image Restoration: Pushing the Limits of Fluorescence Microscopy</article-title>. <source>Nat. Methods</source> <volume>15</volume>, <fpage>1090</fpage>&#x2013;<lpage>1097</lpage>. <pub-id pub-id-type="doi">10.1038/s41592-018-0216-7</pub-id> </citation>
</ref>
<ref id="B48">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wieslander</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Gupta</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Bergman</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>Hallstr&#xf6;m</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>Harrison</surname>
<given-names>P. J.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>Learning to See Colours: Generating Biologically Relevant Fluorescent Labels from Bright-Field Images</article-title>. <source>Cold Spring Harb. Lab.</source>. </citation>
</ref>
<ref id="B49">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Yang</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Tian</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Xue</surname>
<given-names>J.-H.</given-names>
</name>
<name>
<surname>Liao</surname>
<given-names>Q.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>Deep Learning for Single Image Super-resolution: A Brief Review</article-title>. <source>IEEE Trans. Multimed.</source> <volume>21</volume>, <fpage>3106</fpage>&#x2013;<lpage>3121</lpage>. <pub-id pub-id-type="doi">10.1109/tmm.2019.2919431</pub-id> </citation>
</ref>
<ref id="B50">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Yu</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Lin</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Yang</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Shen</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Lu</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Huang</surname>
<given-names>T. S.</given-names>
</name>
</person-group> (<year>2018</year>). &#x201c;<article-title>Generative Image Inpainting with Contextual Attention</article-title>,&#x201d; in <conf-name>IEEE Conference on Computer Vision and Pattern Recognition</conf-name> (<publisher-loc>Washington, D.C., United States</publisher-loc>: <publisher-name>Computer Vision Foundation/IEEE Computer Society</publisher-name>), <fpage>5505</fpage>&#x2013;<lpage>5514</lpage>. <pub-id pub-id-type="doi">10.1109/CVPR.2018.00577</pub-id> </citation>
</ref>
</ref-list>
</back>
</article>