<?xml version="1.0" encoding="UTF-8" standalone="no"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" article-type="review-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Neuroinform.</journal-id>
<journal-title>Frontiers in Neuroinformatics</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Neuroinform.</abbrev-journal-title>
<issn pub-type="epub">1662-5196</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fninf.2022.933230</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Neuroscience</subject>
<subj-group>
<subject>Review</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Review of Generative Adversarial Networks in mono- and cross-modal biomedical image registration</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name><surname>Han</surname> <given-names>Tingting</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/1695161/overview"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Wu</surname> <given-names>Jun</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/1682218/overview"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Luo</surname> <given-names>Wenting</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
</contrib>
<contrib contrib-type="author">
<name><surname>Wang</surname> <given-names>Huiming</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
</contrib>
<contrib contrib-type="author">
<name><surname>Jin</surname> <given-names>Zhe</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/1961454/overview"/>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name><surname>Qu</surname> <given-names>Lei</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
<xref ref-type="aff" rid="aff4"><sup>4</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>&#x0002A;</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/1453371/overview"/>
</contrib>
</contrib-group>
<aff id="aff1"><sup>1</sup><institution>Ministry of Education Key Laboratory of Intelligent Computing and Signal Processing, Information Materials and Intelligent Sensing Laboratory of Anhui Province, Anhui University</institution>, <addr-line>Hefei</addr-line>, <country>China</country></aff>
<aff id="aff2"><sup>2</sup><institution>School of Artificial Intelligence, Anhui University</institution>, <addr-line>Hefei</addr-line>, <country>China</country></aff>
<aff id="aff3"><sup>3</sup><institution>Institute of Artificial Intelligence, Hefei Comprehensive National Science Center</institution>, <addr-line>Hefei</addr-line>, <country>China</country></aff>
<aff id="aff4"><sup>4</sup><institution>SEU-ALLEN Joint Center, Institute for Brain and Intelligence, Southeast University</institution>, <addr-line>Nanjing</addr-line>, <country>China</country></aff>
<author-notes>
<fn fn-type="edited-by"><p>Edited by: Zhi Zhou, Microsoft, United States</p></fn>
<fn fn-type="edited-by"><p>Reviewed by: Fazhi He, Wuhan University, China; Dong Huang, Fourth Military Medical University, China</p></fn>
<corresp id="c001">&#x0002A;Correspondence: Lei Qu <email>qulei&#x00040;ahu.edu.cn</email></corresp>
</author-notes>
<pub-date pub-type="epub">
<day>22</day>
<month>11</month>
<year>2022</year>
</pub-date>
<pub-date pub-type="collection">
<year>2022</year>
</pub-date>
<volume>16</volume>
<elocation-id>933230</elocation-id>
<history>
<date date-type="received">
<day>30</day>
<month>04</month>
<year>2022</year>
</date>
<date date-type="accepted">
<day>13</day>
<month>10</month>
<year>2022</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x000A9; 2022 Han, Wu, Luo, Wang, Jin and Qu.</copyright-statement>
<copyright-year>2022</copyright-year>
<copyright-holder>Han, Wu, Luo, Wang, Jin and Qu</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p></license>
</permissions>
<abstract>
<p>Biomedical image registration refers to aligning corresponding anatomical structures among different images, which is critical to many tasks, such as brain atlas building, tumor growth monitoring, and image fusion-based medical diagnosis. However, high-throughput biomedical image registration remains challenging due to inherent variations in the intensity, texture, and anatomy resulting from different imaging modalities, different sample preparation methods, or different developmental stages of the imaged subject. Recently, Generative Adversarial Networks (GAN) have attracted increasing interest in both mono- and cross-modal biomedical image registrations due to their special ability to eliminate the modal variance and their adversarial training strategy. This paper provides a comprehensive survey of the GAN-based mono- and cross-modal biomedical image registration methods. According to the different implementation strategies, we organize the GAN-based mono- and cross-modal biomedical image registration methods into four categories: modality translation, symmetric learning, adversarial strategies, and joint training. The key concepts, the main contributions, and the advantages and disadvantages of the different strategies are summarized and discussed. Finally, we analyze the statistics of all the cited works from different points of view and reveal future trends for GAN-based biomedical image registration studies.</p></abstract>
<kwd-group>
<kwd>cross-modal</kwd>
<kwd>biomedical image registration</kwd>
<kwd>Generative Adversarial Networks</kwd>
<kwd>image translation</kwd>
<kwd>adversarial training</kwd>
</kwd-group>
<contract-sponsor id="cn001">National Natural Science Foundation of China<named-content content-type="fundref-id">10.13039/501100001809</named-content></contract-sponsor>
<counts>
<fig-count count="7"/>
<table-count count="8"/>
<equation-count count="28"/>
<ref-count count="89"/>
<page-count count="19"/>
<word-count count="10778"/>
</counts>
</article-meta>
</front>
<body>
<sec sec-type="intro" id="s1">
<title>Introduction</title>
<p>The goal of biomedical image registration (BIR) is to estimate a linear or non-linear spatial transformation by geometrically aligning the corresponding anatomical structures between images. The images can be acquired across time, modalities, subjects, or species. By aligning the corresponding structures or mapping the images onto a canonical coordinate space, the registration allows quantitative comparison across the subjects imaged under different conditions. It enables the analysis of their distinct aspects in pathology or neurobiology in a coordinated manner (Oliveira and Tavares, <xref ref-type="bibr" rid="B60">2014</xref>). In addition, image registration is also fundamental to image-guided intervention and radiotherapy.</p>
<p>In recent years, there has been a steady emergence of high-resolution and high-throughput biomedical imaging techniques (Li and Gong, <xref ref-type="bibr" rid="B42">2012</xref>; Chen et al., <xref ref-type="bibr" rid="B13">2021</xref>). Some commonly used macroscale imaging techniques include magnetic resonance imaging (MRI), computed tomography (CT), positron emission tomography (PET), and single photon emission computed tomography (SPECT) (Gering et al., <xref ref-type="bibr" rid="B20">2001</xref>; Staring et al., <xref ref-type="bibr" rid="B70">2009</xref>). However, mesoscale and microscale imaging techniques, such as serial two-photon tomography (STPT) (Ragan et al., <xref ref-type="bibr" rid="B64">2012</xref>), fluorescence micro-optical sectioning tomography (FMOST) (Gong et al., <xref ref-type="bibr" rid="B21">2016</xref>), volumetric imaging with synchronous on-the-fly scan and readout (VISOR) (Xu et al., <xref ref-type="bibr" rid="B80">2021</xref>), and the electron microscope (EM) (Ruska, <xref ref-type="bibr" rid="B66">1987</xref>), play pivotal roles in various neuroscience studies. The resulting exploration of the number, resolution, dimensionality, and modalities of biomedical images not only provides researchers with unprecedented opportunities to study tissue functions, diagnose diseases, etc. but also poses enormous challenges to image registration techniques.</p>
<p>A large number of image registration methods, ranging from the traditional iterative method to the one-shot end-to-end method (Klein S. et al., <xref ref-type="bibr" rid="B40">2009</xref>; Qu et al., <xref ref-type="bibr" rid="B62">2021</xref>), from the fully supervised strategy to the unsupervised strategy (Balakrishnan et al., <xref ref-type="bibr" rid="B6">2019</xref>; He et al., <xref ref-type="bibr" rid="B26">2021</xref>), have been developed to take full advantage of the rapidly accumulating biomedical images with different geometric and modalities. According to the different acquisition techniques of the images, these methods can also be classified into two main categories: mono-modal (intra-modal) registration and cross-modal (or inter-model) registration. Generally, the images of different modalities often vary substantially in their voxel intensity, image texture, and anatomical structures (e.g., due to uneven brain shrinkage resulting from different sample preparation methods). Therefore, cross-modal registration is even more challenging to achieve.</p>
<p>To the best of our knowledge, few studies focus on cross-modal medical image registration. Among the available reviews (Jiang S. et al., <xref ref-type="bibr" rid="B34">2021</xref>), traditional feature-based cross-modal alignment methods have been reviewed in detail. Most of these traditional registration methods are based on iterative training, which is time consuming. The supervised alignment methods are limited by insufficient labels among the learning-based methods. However, unsupervised methods are proposed with various loss functions due to the absence of ground truth and supervision.</p>
<p>Additionally, these unsupervised methods do not perform as well as unimodal on cross-modal images due to too much variation between cross-modal appearances. Nevertheless, efforts are being directed toward removing the modal differences between cross-modalities. Among these various deep-learning-based methods, the Generative Adversarial Networks (GAN) (Goodfellow et al., <xref ref-type="bibr" rid="B22">2014</xref>) have received increasing attention from researchers for their unique network structures and training strategies. In addition, the GAN-based methods have shown extraordinary potential in dealing with cross-modal registration. In particular, the conditional GAN (CGAN) can realize transformation between different styles of images, which provides a new solution to the difficult cross-modal registration method, which has been plagued by the different modality characteristics for a long time.</p>
<p>Since it was proposed, the GAN has been widely used for biomedical image analysis, such as classification, detection, and registration. Its outstanding performance in image synthesis, style translation (Kim et al., <xref ref-type="bibr" rid="B37">2017</xref>; Jing et al., <xref ref-type="bibr" rid="B35">2020</xref>), and the adversarial training strategy has attracted attention in many areas (Li et al., <xref ref-type="bibr" rid="B45">2021</xref>). GAN has been applied to image registration tasks since 2018 (Yan et al., <xref ref-type="bibr" rid="B82">2018</xref>). However, compared with other deep-learning-based registration methods, the GAN-based methods are still in their infancy, and their potential needs further exploration. To the best of our knowledge, there has not yet been a specific review on GAN in biomedical image registration. Therefore, we hereby try to provide an up-to-date and comprehensive review of existing GAN applications in biomedical image registration.</p>
<p>In the survey, we focus on both the GAN-based mono- and cross-modal biomedical image registrations but may highlight more on the contribution of the GAN-based cross-modal image registration. Cross-modal biomedical image registration is still facing many challenges compared with mature mono-modal biomedical image registration.</p>
<p>This review is structured as follows: Section Common GAN structures briefly introduces the basic theory of common GAN related to image registrations; Section Strategies of GAN based biomedical image registration provides a comprehensive analysis of four GAN-based registration strategies; in Section Statistics, we analyze the ratio distribution of some important characteristics of these studies, and in Section Future perspectives, we discuss some open issues and future research perspectives.</p></sec>
<sec id="s2">
<title>Common GAN structures</title>
<p>This section gives a brief introduction to the GAN structures used for image generation. The structures considered are often used directly or indirectly in the cross-modal biomedical image registration model. We summarize this literature, which considers GAN structures, in <xref ref-type="table" rid="T2">Tables 2</xref>&#x02013;<xref ref-type="table" rid="T5">5</xref>. The section emphasizes the overall architecture, data flow, and objective function of GAN. The differences between the various methods are also presented.</p>
<sec>
<title>Original GAN</title>
<p>The framework of the original GAN is shown in <xref ref-type="fig" rid="F1">Figure 1</xref>. The original GAN consists of two networks, the generator (G) and the discriminator (D), both of which are fully connected. The input to the generator is a random noise vector from the noise distribution &#x0007E;<italic>p</italic>(z) (random noise is Gaussian noise or uniform noise). The generator can learn a mapping from the low-dimensional noise vector space to the high-dimensional data space. The input of the discriminator is the real data &#x0007E;<italic>P</italic><sub><italic>r</italic></sub><italic>(x)</italic> and the synthetic data &#x0007E;<italic>P</italic><sub><italic>g</italic></sub><italic>(x)</italic> by the generator. If the input to the discriminator is real data x, the purpose of the discriminator is to represent the probability that <italic>x</italic> comes from &#x0007E;<italic>P</italic><sub><italic>r</italic></sub><italic>(x)</italic> rather than &#x0007E;<italic>P</italic><sub><italic>g</italic></sub><italic>(x)</italic>, and the discriminator should classify it as real data and return a value close to 1.</p>
<fig id="F1" position="float">
<label>Figure 1</label>
<caption><p>The architecture of the original GAN.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fninf-16-933230-g0001.tif"/>
</fig>
<p>Conversely, if the input is synthetic data, the discriminator should classify it as false data and return a value close to 0. The false signal output from the discriminator is back propagated to the generator to update the network parameters. This framework is trained in an adversarial strategy corresponding to a two-player minimax game. The minimax GAN loss is equivalent to the game&#x00027;s rules, while the generator and the discriminator are equivalent to the two players. The goal of the generator is to minimize the loss by generating synthetic images that look as similar to the real images as possible to fool the discriminator.</p>
<p>In contrast, the discriminator maximizes the loss to maximize the probability of assigning the correct label to both the training examples and the samples from the generator. The training improves the performance of both the generator and the discriminator networks. Their loss functions can be formulated as follows:</p>
<disp-formula id="E1"><mml:math id="M1"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mrow><mml:mi>D</mml:mi></mml:mrow></mml:msub></mml:mtd><mml:mtd><mml:mo>=</mml:mo></mml:mtd><mml:mtd><mml:mstyle displaystyle="true"><mml:munder class="msub"><mml:mrow><mml:mi>m</mml:mi><mml:mi>a</mml:mi><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>D</mml:mi></mml:mrow></mml:munder></mml:mstyle><mml:msub><mml:mrow><mml:mo>&#x1D53C;</mml:mo></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>x</mml:mtext></mml:mstyle><mml:mo>&#x0007E;</mml:mo><mml:msub><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>p</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>d</mml:mtext></mml:mstyle><mml:mi>a</mml:mi><mml:mi>t</mml:mi><mml:mstyle mathvariant="bold"><mml:mtext>a</mml:mtext></mml:mstyle><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>x</mml:mtext></mml:mstyle></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msub></mml:mrow></mml:msub><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mi>l</mml:mi><mml:mi>o</mml:mi><mml:mi>g</mml:mi><mml:mi>D</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mo>&#x0002B;</mml:mo><mml:msub><mml:mrow><mml:mo>&#x1D53C;</mml:mo></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>z</mml:mtext></mml:mstyle><mml:mo>&#x0007E;</mml:mo><mml:msub><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>p</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>z</mml:mtext></mml:mstyle></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>z</mml:mtext></mml:mstyle></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mo class="qopname">log</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mn>1</mml:mn><mml:mo>-</mml:mo><mml:mi>D</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>G</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>Z</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="E2"><label>(1)</label><mml:math id="M2"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi></mml:mrow></mml:msub></mml:mtd><mml:mtd><mml:mo>=</mml:mo></mml:mtd><mml:mtd><mml:mi>m</mml:mi><mml:mi>i</mml:mi><mml:msub><mml:mrow><mml:mi>n</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mo>&#x1D53C;</mml:mo></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>z</mml:mtext></mml:mstyle><mml:mo>&#x0007E;</mml:mo><mml:msub><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>p</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>z</mml:mtext></mml:mstyle></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>z</mml:mtext></mml:mstyle></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mo class="qopname">log</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mn>1</mml:mn><mml:mo>-</mml:mo><mml:mi>D</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>G</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>Z</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where <italic>L</italic><sub><italic>D</italic></sub> and <italic>L</italic><sub><italic>G</italic></sub> are the loss functions of <italic>D</italic> and <italic>G</italic>, respectively, and <italic>D</italic> is the binary classifier; it is expected that the data distribution generated by G(z) is close to the real data when the model is optimized.</p>
</sec>
<sec>
<title>DCGAN</title>
<p>Compared with the original GAN, the deep convolutional generative adversarial networks (DCGAN) (Radford et al., <xref ref-type="bibr" rid="B63">2015</xref>) add specific architectural constraints to GAN by replacing all the full-connected neural networks with CNN, which results in stable training. <xref ref-type="fig" rid="F2">Figure 2A</xref> illustrates the structure of DCGAN, in which there are three important changes in the convolutional neural network (CNN) architecture. Firstly, the pooling layers in the discriminator and the generator are replaced by the stridden convolution and the fractionally strung convolutions, respectively, which allow the generator to learn the specific spatial upsampling from the input noise distribution to the output image. Secondly, batch normalization (Ioffe and Szegedy, <xref ref-type="bibr" rid="B30">2015</xref>) is utilized to regulate poor initialization to prevent the generator from collapsing from all samples to a single point. Thirdly, the LeakyReLU (Maas et al., <xref ref-type="bibr" rid="B54">2013</xref>) activation is adopted to replace the maxout activation in all layers of the discriminator, promoting the output of higher-resolution images.</p>
<fig id="F2" position="float">
<label>Figure 2</label>
<caption><p>The architecture of the variant GANs. <bold>(A)</bold> The architecture of DCGAN, <bold>(B)</bold> the architecture of CGAN, <bold>(C)</bold> the architecture of Pix2Pix, <bold>(D)</bold> the architecture of CyclGAN, and <bold>(E)</bold> the architecture of UNIT.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fninf-16-933230-g0002.tif"/>
</fig>
</sec>
<sec>
<title>CGAN</title>
<p>The structure of CGAN (Mirza and Osindero, <xref ref-type="bibr" rid="B58">2014</xref>) is illustrated in <xref ref-type="fig" rid="F2">Figure 2B</xref>. The CGAN performs the conditioning for the output mode by feeding the auxiliary information related to the desired properties <italic>y</italic> and noise vector <italic>z</italic> to the generator and the discriminator. The objective function of the CGAN can be formulated as follows:</p>
<disp-formula id="E3"><mml:math id="M3"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mrow><mml:mi>D</mml:mi></mml:mrow></mml:msub></mml:mtd><mml:mtd><mml:mo>=</mml:mo></mml:mtd><mml:mtd><mml:mstyle displaystyle="true"><mml:munder class="msub"><mml:mrow><mml:mi>m</mml:mi><mml:mi>a</mml:mi><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>D</mml:mi></mml:mrow></mml:munder></mml:mstyle><mml:msub><mml:mrow><mml:mo>&#x1D53C;</mml:mo></mml:mrow><mml:mrow><mml:mi>x</mml:mi><mml:mo>&#x0007E;</mml:mo><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mi>d</mml:mi><mml:mi>a</mml:mi><mml:mi>t</mml:mi><mml:mi>a</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mi>l</mml:mi><mml:mi>o</mml:mi><mml:mi>g</mml:mi><mml:mi>D</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi><mml:mo>|</mml:mo><mml:mi>y</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mo>&#x0002B;</mml:mo><mml:msub><mml:mrow><mml:mo>&#x1D53C;</mml:mo></mml:mrow><mml:mrow><mml:mi>z</mml:mi><mml:mo>&#x0007E;</mml:mo><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mi>z</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>z</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mi>l</mml:mi><mml:mi>o</mml:mi><mml:mi>g</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mn>1</mml:mn><mml:mo>-</mml:mo><mml:mi>D</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>G</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>z</mml:mi><mml:mo>|</mml:mo><mml:mi>y</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="E4"><label>(2)</label><mml:math id="M4"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi></mml:mrow></mml:msub></mml:mtd><mml:mtd><mml:mo>=</mml:mo></mml:mtd><mml:mtd><mml:mi>m</mml:mi><mml:mi>i</mml:mi><mml:msub><mml:mrow><mml:mi>n</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mo>&#x1D53C;</mml:mo></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>z</mml:mtext></mml:mstyle><mml:mo>&#x0007E;</mml:mo><mml:msub><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>p</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>z</mml:mtext></mml:mstyle></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>z</mml:mtext></mml:mstyle></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mtext>&#x000A0;</mml:mtext><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mo class="qopname">log</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mn>1</mml:mn><mml:mo>-</mml:mo><mml:mi>D</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>G</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>z</mml:mi><mml:mo>|</mml:mo><mml:mi>y</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where <italic>y</italic> is the auxiliary information, which could be a class label, an image, or even the data from different modes. For instance, Pix2Pix (Isola et al., <xref ref-type="bibr" rid="B31">2017</xref>) translates the label image or edge image to an image with another style. InfoGAN (Chen et al., <xref ref-type="bibr" rid="B14">2016</xref>) is regarded as a special CGAN whose condition is a conditional constraint on the random noise <italic>z</italic> for guiding the thickness, slope, and other features of the generated image.</p>
</sec>
<sec>
<title>Pix2Pix</title>
<p>Pix2Pix (Isola et al., <xref ref-type="bibr" rid="B31">2017</xref>) is the first GAN framework for image-to-image translation, which can learn a mapping that transforms an image from one modality to another based on paired-wise images. The structure of Pix2Pix is depicted in <xref ref-type="fig" rid="F2">Figure 2C</xref>. A paired-wise image means that the internal structures in the image pair are accurately aligned, while their texture, brightness, and other modality-related features differ. The objective loss combines CGAN with the <italic>L1</italic> loss so that the generator is also asked to generate images as close as possible to the ground truth:</p>
<disp-formula id="E5"><label>(3)</label><mml:math id="M5"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>G</mml:mi><mml:mo>,</mml:mo><mml:mi>D</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:mi>x</mml:mi><mml:mo>,</mml:mo><mml:mi>y</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mi>l</mml:mi><mml:mi>o</mml:mi><mml:mi>g</mml:mi><mml:mi>D</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi><mml:mo>,</mml:mo><mml:mi>y</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mo>&#x0002B;</mml:mo><mml:msub><mml:mrow><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:mi>x</mml:mi><mml:mo>,</mml:mo><mml:mi>z</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mo class="qopname">log</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mn>1</mml:mn><mml:mo>-</mml:mo><mml:mi>D</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi><mml:mo>,</mml:mo><mml:mi>G</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi><mml:mo>,</mml:mo><mml:mi>z</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="E6"><label>(4)</label><mml:math id="M6"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mrow><mml:mi>L</mml:mi><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>G</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:mi>x</mml:mi><mml:mo>,</mml:mo><mml:mi>y</mml:mi><mml:mo>,</mml:mo><mml:mi>z</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mo>|</mml:mo><mml:mo>|</mml:mo><mml:mi>y</mml:mi><mml:mo>-</mml:mo><mml:mi>G</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi><mml:mo>,</mml:mo><mml:mi>z</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>|</mml:mo><mml:msub><mml:mrow><mml:mo>|</mml:mo></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="E7"><label>(5)</label><mml:math id="M7"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msup><mml:mrow><mml:mi>G</mml:mi></mml:mrow><mml:mrow><mml:mo>*</mml:mo></mml:mrow></mml:msup><mml:mo>=</mml:mo><mml:mi>a</mml:mi><mml:mi>r</mml:mi><mml:mi>g</mml:mi><mml:mtable style="text-align:axis;" equalrows="false" columnlines="none none none none none none none none none" equalcolumns="false" class="array"><mml:mtr><mml:mtd><mml:mo class="qopname">min</mml:mo></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mi>G</mml:mi></mml:mtd></mml:mtr><mml:mtr></mml:mtr></mml:mtable><mml:mtable style="text-align:axis;" equalrows="false" columnlines="none none none none none none none none none" equalcolumns="false" class="array"><mml:mtr><mml:mtd><mml:mo class="qopname">max</mml:mo></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mi>D</mml:mi></mml:mtd></mml:mtr><mml:mtr></mml:mtr></mml:mtable><mml:msub><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>G</mml:mi><mml:mo>,</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:mi>D</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x0002B;</mml:mo><mml:mi>&#x003BB;</mml:mi><mml:msub><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mrow><mml:mi>L</mml:mi><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>G</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where <italic>x</italic> and <italic>y</italic> represent the images from the source and the target domain, respectively; the <italic>L1</italic> loss is a pixel-level metric between the target domain image and the generated image to impose a constraint on <italic>G</italic>, which could recover the low-frequency part of the image; the adversarial loss could recovery the high-frequency part of the image, and &#x003BB; is the adjustable parameter.</p>
</sec>
<sec>
<title>Cycle-GAN</title>
<p>CycleGAN (Zhu et al., <xref ref-type="bibr" rid="B89">2017</xref>) contains two generators and two discriminators, which are self-bounded by an inverse loop to transform the image between the two domains. The structure of Cycle-GAN is illustrated in <xref ref-type="fig" rid="F2">Figure 2D</xref>. One generator, <italic>G</italic>, translates the source domain image <italic>X</italic> to the target domain image <italic>Y</italic>. Another generator, <italic>F</italic>, learns the inverse mapping of <italic>G</italic>, which brings <italic>G(X)</italic> back to its original image <italic>X</italic>., i.e., <italic>x</italic>&#x02192;<italic>G(x)</italic>&#x02192;<italic>F[G(x)]</italic> &#x02248; <italic>x</italic>. Similarly, for <italic>y</italic> from the domain <italic>Y, F</italic> and <italic>G</italic> also satisfy the cycle-consistent, i.e., <italic>y</italic>&#x02192;<italic>F(y)</italic>&#x02192;<italic>G[F(y)]</italic> &#x02248; <italic>y</italic>. The cycle-consistent loss <italic>L</italic><sub><italic>cycle</italic></sub> measures the reconstructed image and the real image by pixel-level loss calculation to constrain the training of <italic>G</italic> and <italic>F</italic>, ensuring the consistency of its morphological structure in the transformation process. Two discriminators distinguish between the reconstructed image and the real image. The adversarial loss and the cycle-consistent loss are as follows:</p>
<disp-formula id="E8"><mml:math id="M8"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>G</mml:mi><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mrow><mml:mi>Y</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd><mml:mtd><mml:mo>=</mml:mo></mml:mtd><mml:mtd><mml:mi>E</mml:mi><mml:msub><mml:mrow><mml:mtext>&#x000A0;</mml:mtext></mml:mrow><mml:mrow><mml:mi>y</mml:mi><mml:mo>&#x0007E;</mml:mo><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mi>Y</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mi>l</mml:mi><mml:mi>o</mml:mi><mml:mi>g</mml:mi><mml:msub><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mrow><mml:mi>Y</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="E9"><label>(6)</label><mml:math id="M9"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mo>&#x0002B;</mml:mo><mml:mi>E</mml:mi><mml:msub><mml:mrow><mml:mtext>&#x000A0;</mml:mtext></mml:mrow><mml:mrow><mml:mi>x</mml:mi><mml:mo>&#x0007E;</mml:mo><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mi>X</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mi>l</mml:mi><mml:mi>o</mml:mi><mml:mi>g</mml:mi><mml:msub><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mrow><mml:mi>X</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="E10"><mml:math id="M10"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>y</mml:mi><mml:mi>c</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>G</mml:mi><mml:mo>,</mml:mo><mml:mi>F</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mtext>&#x000A0;</mml:mtext><mml:mo>=</mml:mo><mml:mo>&#x1D53C;</mml:mo><mml:mi>x</mml:mi><mml:msub><mml:mrow><mml:mtext>&#x000A0;</mml:mtext></mml:mrow><mml:mrow><mml:mi>x</mml:mi><mml:mo>&#x0007E;</mml:mo><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mi>X</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mo>|</mml:mo><mml:mo>|</mml:mo><mml:mi>F</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>G</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>-</mml:mo><mml:mi>x</mml:mi><mml:msub><mml:mrow><mml:mo>|</mml:mo><mml:mo>|</mml:mo></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="E11"><label>(7)</label><mml:math id="M11"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mo>&#x0002B;</mml:mo><mml:mo>&#x1D53C;</mml:mo><mml:msub><mml:mrow><mml:mtext>&#x000A0;</mml:mtext></mml:mrow><mml:mrow><mml:mi>y</mml:mi><mml:mo>&#x0007E;</mml:mo><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mi>Y</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mo>|</mml:mo><mml:mo>|</mml:mo><mml:mi>G</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>F</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>-</mml:mo><mml:mi>y</mml:mi><mml:msub><mml:mrow><mml:mo>|</mml:mo><mml:mo>|</mml:mo></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="E12"><label>(8)</label><mml:math id="M12"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mi>L</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>G</mml:mi><mml:mo>,</mml:mo><mml:mi>F</mml:mi><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mrow><mml:mi>X</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mrow><mml:mi>Y</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mtext>&#x000A0;&#x000A0;</mml:mtext><mml:msub><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>G</mml:mi><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mrow><mml:mi>Y</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x0002B;</mml:mo><mml:msub><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>F</mml:mi><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mrow><mml:mi>X</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x0002B;</mml:mo><mml:mi>&#x003BB;</mml:mi><mml:mtext>&#x000A0;</mml:mtext><mml:msub><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>y</mml:mi><mml:mi>c</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>G</mml:mi><mml:mo>,</mml:mo><mml:mi>F</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>The training procedure uses the least squares and replays buffer for training stability. UNET (Ronneberger et al., <xref ref-type="bibr" rid="B65">2015</xref>) and PatchGAN (Li and Wand, <xref ref-type="bibr" rid="B43">2016</xref>; Isola et al., <xref ref-type="bibr" rid="B31">2017</xref>) are used to build the generator and the discriminator.</p>
</sec>
<sec>
<title>UNIT</title>
<p>UNIT (Liu et al., <xref ref-type="bibr" rid="B49">2017</xref>) can also perform unpaired image-to-image transformation by combining two variational autoencoder generative adversarial networks (VAEGAN) (Xian et al., <xref ref-type="bibr" rid="B79">2019</xref>), with each responsible for one modality but sharing the same latent space. The UNIT structure is illustrated in <xref ref-type="fig" rid="F2">Figure 2E</xref>, which consists of six subnetworks: two domain image encoders, E1 and E2, two domain image generators, G1 and G2, and two adversarial domain discriminators, D1 and D2. The encoder&#x02013;generator pair {E1, G1} constitutes a VAE for the X1 domain, called VAE1. For an input image x1&#x02208;X1, the VAE1 first maps x1 to a code in a latent space Z via the encoder E1 and then decodes a random-perturbed version of the code to reconstruct the input image via the generator G1. The image x<sub>2</sub> in X<sub>2</sub> can be translated to an image in X<sub>1</sub> by applying G<sub>1</sub>(Z<sub>2</sub>). For real images sampled from the X<sub>1</sub> domain, D1 should output true, whereas, for images generated by G<sub>1</sub>(Z<sub>2</sub>), it should output false. The cycle-consistency constraint exists in x<sub>1</sub> = F<sub>2 &#x02212; 1</sub>[F<sub>1 &#x02212; 2</sub>(x1)], where F<sub>1 &#x02212; 2</sub> = G<sub>2</sub>[E<sub>1</sub>(X<sub>1</sub>)].</p></sec>
</sec>
<sec id="s3">
<title>Strategies for GAN-based biomedical image registration</title>
<p>Cross-modal biomedical image registration using GAN has given rise to an increasing number of registration algorithms to solve the current problems mentioned in the introduction section. Based on the different strategies, the algorithms are divided into four categories: modality translation, symmetric learning, adversarial strategies, and joint training. A category overview of the biomedical image registration methods using GAN is shown in <xref ref-type="table" rid="T1">Table 1</xref>. In the table, we describe the key ideas of the four categories, respectively, and summarize the different implementation methods for each strategy. In the subsequent subsections, we review all the relevant works as classified in <xref ref-type="table" rid="T1">Table 1</xref>.</p>
<table-wrap position="float" id="T1">
<label>Table 1</label>
<caption><p>A category overview of biomedical image registration methods using GAN.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="left"><bold>Category</bold></th>
<th valign="top" align="left"><bold>Key idea</bold></th>
<th valign="top" align="left"><bold>Method</bold></th>
<th valign="top" align="left"><bold>Publication</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Modality independent</td>
<td valign="top" align="left">Translate two different modal-image to the</td>
<td valign="top" align="left">Modality translation</td>
<td valign="top" align="left">Tanner et al., <xref ref-type="bibr" rid="B71">2018</xref>; Wei et al., <xref ref-type="bibr" rid="B77">2019</xref></td>
</tr>
<tr>
<td/>
<td valign="top" align="left">same domain, then perform mono-modal</td>
<td/>
<td valign="top" align="left">Wei et al., <xref ref-type="bibr" rid="B76">2020</xref>; Xu et al., <xref ref-type="bibr" rid="B81">2020</xref></td>
</tr>
<tr>
<td/>
<td valign="top" align="left">registration</td>
<td/>
<td valign="top" align="left">Zhang et al., <xref ref-type="bibr" rid="B84">2020</xref>; Zhou et al., <xref ref-type="bibr" rid="B88">2021</xref></td>
</tr>
<tr>
<td/>
<td/>
<td/>
<td valign="top" align="left">Lu et al., <xref ref-type="bibr" rid="B50">2021</xref></td>
</tr>
<tr>
<td/>
<td/>
<td valign="top" align="left">Latent representation</td>
<td valign="top" align="left">Mahapatra and Ge, <xref ref-type="bibr" rid="B55">2020</xref>; Yang et al., <xref ref-type="bibr" rid="B83">2020</xref></td>
</tr>
<tr style="border-bottom: thin solid #000000;">
<td/>
<td/>
<td valign="top" align="left">Image decomposition</td>
<td valign="top" align="left">Qin et al., <xref ref-type="bibr" rid="B61">2019</xref>; Wu and Zhou, <xref ref-type="bibr" rid="B78">2021</xref></td>
</tr> <tr>
<td valign="top" align="left">Symmetric Learning</td>
<td valign="top" align="left">The accuracy of bidirectional registration is optimized by making transformation inverse consistency</td>
<td valign="top" align="left">GAN</td>
<td valign="top" align="left">Zheng et al., <xref ref-type="bibr" rid="B87">2021</xref></td>
</tr>
<tr style="border-bottom: thin solid #000000;">
<td/>
<td/>
<td valign="top" align="left">Cyclic-consistency</td>
<td valign="top" align="left">Lu et al., <xref ref-type="bibr" rid="B52">2019</xref>, <xref ref-type="bibr" rid="B51">2020</xref></td>
</tr> <tr>
<td valign="top" align="left">Adversarial learning</td>
<td valign="top" align="left">Adopt the way of adversarial training to perform image registration. The generator is regarded as registration, and similarity loss is instead of the discriminator</td>
<td valign="top" align="left">Semi-supervised</td>
<td valign="top" align="left">Hu et al., <xref ref-type="bibr" rid="B29">2018</xref>; Elmahdy et al., <xref ref-type="bibr" rid="B16">2019</xref>; Li and Ogino, <xref ref-type="bibr" rid="B46">2019</xref>; Luo et al., <xref ref-type="bibr" rid="B53">2021</xref></td>
</tr>
<tr>
<td/>
<td/>
<td valign="top" align="left">Knowledge distillation</td>
<td valign="top" align="left">Tran et al., <xref ref-type="bibr" rid="B72">2022</xref></td>
</tr>
<tr>
<td/>
<td/>
<td valign="top" align="left">Attention mechanisms</td>
<td valign="top" align="left">Li M. et al., <xref ref-type="bibr" rid="B44">2021</xref></td>
</tr>
<tr style="border-bottom: thin solid #000000;">
<td/>
<td/>
<td valign="top" align="left">Adversarial training</td>
<td valign="top" align="left">Fan et al., <xref ref-type="bibr" rid="B18">2018</xref>, <xref ref-type="bibr" rid="B17">2019</xref>; Yan et al., <xref ref-type="bibr" rid="B82">2018</xref></td>
</tr> <tr>
<td valign="top" align="left">Joint learning</td>
<td valign="top" align="left">Segmentation, synthesis, and registration network jointly train to improve the performance of each other</td>
<td valign="top" align="left">Multitask</td>
<td valign="top" align="left">Mahapatra et al., <xref ref-type="bibr" rid="B56">2018</xref>; Liu et al., <xref ref-type="bibr" rid="B48">2020</xref>; Zhou et al., <xref ref-type="bibr" rid="B88">2021</xref></td>
</tr>
</tbody>
</table>
</table-wrap>
<p>To present a comprehensive overview of all the relevant works on GANs in biomedical image registration, we searched science datasets, including Google Scholar, SpringerLink, and PubMed, for all relevant published articles from 2018 to 2021. The keywords included medical image registration/matching alignment, GAN (Generative Adversarial Networks), multimodal (cross-modality) medical image registration, GAN, adversarial medical image registration, segmentation, and registration. About 300 papers are indexed, including 36 papers that completely match our criteria. There are two requirements for our inclusion in the articles. The first is that the topic of the papers is image registration, and the second is that the method is of GAN based on and used to implement the registration strategy. To verify the comprehensiveness of the search, we also searched them separately in the International Conference on Medical Image Computing and Computer-Assisted Intervention (MICCAI), the IEEE International Symposium on Biomedical Imaging (ISBI), and SPIE Medical Imaging to compare with the already searched papers. During the literature review process, we try to integrate all relevant papers to reach a reasonable conclusion; however, because this topic is still in its infancy, the number of published papers is minimal. Therefore, we are unable to conduct an experimental review on this topic because most of the searched articles have no open-source code, and the data are private.</p>
<sec>
<title>Modality-Independent based strategy</title>
<p>Biomedical image registration algorithms of modality-independent based strategies mainly focus on cross-modal images. The key idea of the strategy is to eliminate the variance between modalities so that cross-modality registration can be performed on modality-independent data. A modality-independent strategy can be implemented by translating cross-modality, image disentangling, and latent representation methods. This strategy can avoid the design of cross-modal similarity loss. It only uses robust mono-modal similarity loss to guide the optimization of the model. In <xref ref-type="table" rid="T2">Table 2</xref>, we provide an overview of the important elements of all the reviewed papers. Among these papers, 12 directly use Cycle-GAN as the baseline model, and seven are applied to the registration tasks of MRI-CT, with the organs covered by the brain, liver, retina, and heart.</p>
<table-wrap position="float" id="T2">
<label>Table 2</label>
<caption><p>Overview of the modality-independent based strategy.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="left"><bold>Publications</bold></th>
<th valign="top" align="center"><bold>Organ</bold></th>
<th valign="top" align="center"><bold>Method</bold></th>
<th valign="top" align="center"><bold>Modality</bold></th>
<th valign="top" align="center"><bold>Evaluation metrics</bold></th>
<th valign="top" align="center"><bold>Loss</bold></th>
<th valign="top" align="center"><bold>Dataset</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Han et al. (<xref ref-type="bibr" rid="B25">2021</xref>)</td>
<td valign="top" align="center">Brain</td>
<td valign="top" align="center">Cycle-GAN</td>
<td valign="top" align="center">MRI-CT</td>
<td valign="top" align="center">M1, 2</td>
<td valign="top" align="center">L3</td>
<td valign="top" align="center">&#x02013;</td>
</tr>
<tr>
<td valign="top" align="left">Fu et al. (<xref ref-type="bibr" rid="B19">2020</xref>)</td>
<td valign="top" align="center">Head and neck</td>
<td valign="top" align="center">Cycle-GAN</td>
<td valign="top" align="center">MRI-CT</td>
<td valign="top" align="center">M2</td>
<td/>
<td valign="top" align="center">&#x02013;</td>
</tr>
<tr>
<td valign="top" align="left">Lu et al. (<xref ref-type="bibr" rid="B51">2020</xref>)</td>
<td valign="top" align="center">Heart</td>
<td valign="top" align="center">Cycle-GAN</td>
<td valign="top" align="center">CT-TEE</td>
<td valign="top" align="center">M1, 3, 4</td>
<td valign="top" align="center">L1</td>
<td valign="top" align="center">&#x02013;</td>
</tr>
<tr>
<td valign="top" align="left">Wei et al. (<xref ref-type="bibr" rid="B76">2020</xref>)</td>
<td valign="top" align="center">Liver</td>
<td valign="top" align="center">Cycle-GAN</td>
<td valign="top" align="center">MRI-CT</td>
<td valign="top" align="center">M1, 2</td>
<td valign="top" align="center">L3</td>
<td valign="top" align="center">&#x02013;</td>
</tr>
<tr>
<td valign="top" align="left">Wei et al. (<xref ref-type="bibr" rid="B77">2019</xref>)</td>
<td valign="top" align="center">Liver</td>
<td valign="top" align="center">Cycle-GAN</td>
<td valign="top" align="center">MRI-CT</td>
<td valign="top" align="center">M1, 2</td>
<td valign="top" align="center">L3</td>
<td valign="top" align="center">&#x02013;</td>
</tr>
<tr>
<td valign="top" align="left">Arar et al. (<xref ref-type="bibr" rid="B2">2020</xref>)</td>
<td valign="top" align="center">/</td>
<td valign="top" align="center">Cycle-GAN</td>
<td valign="top" align="center">/</td>
<td valign="top" align="center">M5</td>
<td valign="top" align="center">L1, 2</td>
<td valign="top" align="center">&#x02013;</td>
</tr>
<tr>
<td valign="top" align="left">Zhang et al. (<xref ref-type="bibr" rid="B84">2020</xref>)</td>
<td valign="top" align="center">Brain</td>
<td valign="top" align="center">Pix2Pix</td>
<td valign="top" align="center">T1&#x02013;T2</td>
<td/>
<td valign="top" align="center">L1, 2, 5</td>
<td valign="top" align="center">D3</td>
</tr>
<tr>
<td valign="top" align="left">Tanner et al. (<xref ref-type="bibr" rid="B71">2018</xref>)</td>
<td valign="top" align="center">Retina and heart</td>
<td valign="top" align="center">Cycle-GAN</td>
<td valign="top" align="center">MRI-CT</td>
<td valign="top" align="center">M1</td>
<td valign="top" align="center">L1, 3</td>
<td valign="top" align="center">&#x02013;</td>
</tr>
<tr>
<td valign="top" align="left">Zhou et al. (<xref ref-type="bibr" rid="B88">2021</xref>)</td>
<td valign="top" align="center">Liver</td>
<td valign="top" align="center">Cycle-GAN</td>
<td valign="top" align="center">MRI-CBCT</td>
<td valign="top" align="center">M1, 10</td>
<td valign="top" align="center">L3, 4, 5, 6, 8</td>
<td valign="top" align="center">D8,9</td>
</tr>
<tr>
<td valign="top" align="left">Yang et al. (<xref ref-type="bibr" rid="B83">2020</xref>)</td>
<td valign="top" align="center">Brain</td>
<td valign="top" align="center">VAE&#x0002B;GAN</td>
<td valign="top" align="center">MRI-CT</td>
<td valign="top" align="center">M7, 9</td>
<td valign="top" align="center">L2, 3</td>
<td valign="top" align="center">&#x02013;</td>
</tr>
<tr>
<td valign="top" align="left">Lu et al. (<xref ref-type="bibr" rid="B50">2021</xref>)</td>
<td valign="top" align="center">/</td>
<td valign="top" align="center">Cycle-GAN, Pix2Pix, Drit, StarGAN-v2</td>
<td valign="top" align="center">/</td>
<td valign="top" align="center">M5</td>
<td valign="top" align="center">/</td>
<td valign="top" align="center">&#x02013;</td>
</tr>
<tr>
<td valign="top" align="left">Xu et al. (<xref ref-type="bibr" rid="B81">2020</xref>)</td>
<td valign="top" align="center">Kidney</td>
<td valign="top" align="center">Cycle-GAN</td>
<td valign="top" align="center">CT-MR</td>
<td/>
<td/>
<td/>
</tr>
<tr>
<td valign="top" align="left">Qin et al. (<xref ref-type="bibr" rid="B61">2019</xref>)</td>
<td valign="top" align="center">Lung and brain</td>
<td valign="top" align="center">CycleGAN</td>
<td valign="top" align="center">T1&#x02013;T2</td>
<td valign="top" align="center">M1, 3, 13</td>
<td valign="top" align="center">L1, 2, 5, 10</td>
<td valign="top" align="center">D24</td>
</tr>
<tr>
<td valign="top" align="left">Wu and Zhou (<xref ref-type="bibr" rid="B78">2021</xref>)</td>
<td valign="top" align="center">Brain</td>
<td valign="top" align="center">CyclegGAN</td>
<td valign="top" align="center">T1&#x02013;T2</td>
<td valign="top" align="center">M14</td>
<td valign="top" align="center">L1, 3, 9</td>
<td valign="top" align="center">D25</td>
</tr>
<tr>
<td valign="top" align="left">Mahapatra and Ge (<xref ref-type="bibr" rid="B55">2020</xref>)</td>
<td valign="top" align="center">Lung, brain, retinal</td>
<td valign="top" align="center">GAN</td>
<td valign="top" align="center">X-rays- X-rays/T1-T2</td>
<td valign="top" align="center">M1, 3</td>
<td valign="top" align="center">L1, 5, 10</td>
<td valign="top" align="center">D23</td>
</tr>
<tr>
<td valign="top" align="left">Lin et al. (<xref ref-type="bibr" rid="B47">2021</xref>)</td>
<td valign="top" align="center">Brain</td>
<td valign="top" align="center">RevGAN</td>
<td valign="top" align="center">MRI-PET</td>
<td valign="top" align="center">ACC</td>
<td valign="top" align="center">L1</td>
<td valign="top" align="center">&#x02013;</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p>Some brief descriptions of the loss, dataset, and evaluation metrics can be found in <xref ref-type="table" rid="T6">Tables 6</xref>&#x02013;<xref ref-type="table" rid="T8">8</xref>. The symbol&#x02013;in the last column means that the data used in the paper is a private dataset.</p>
</table-wrap-foot>
</table-wrap>
<sec>
<title>Modality translation</title>
<p>To register MRI to CT, Tanner et al. (<xref ref-type="bibr" rid="B71">2018</xref>) make the first attempt at modality translation using Cycle-GAN and subsequently perform the mono-modality registration on two images in the same domain. The Patch-GAN uses N &#x000D7; N patches instead of a single value as the output of the discriminator for spatial corresponded-preservation. This pioneering work assessed the feasibility of this strategy. However, the mono-registration significantly relies on the quality of synthetic images. In a subsequent study, to constrain the geometric changes during modality translation, Wei et al. (<xref ref-type="bibr" rid="B76">2020</xref>) designed the mutual information (MI) loss to regularize the anatomy and between the classic mono-modal registration method ANTS (Avants et al., <xref ref-type="bibr" rid="B3">2008</xref>, <xref ref-type="bibr" rid="B4">2009</xref>) as the registration network. Xu et al. (<xref ref-type="bibr" rid="B81">2020</xref>) combined the deformation field from uni- and multimodal stream networks by dual stream fusion network for cross-modality registration. The uni-modal stream model preserves the anatomy during modality translation using the Cycle-GAN by combining several losses, including the modality independent neighborhood descriptor (MIND) (Heinrich et al., <xref ref-type="bibr" rid="B28">2012</xref>), the correlation coefficient loss (CC), and the L2 loss. The basic structures of these methods are shown in <xref ref-type="fig" rid="F3">Figure 3D</xref>. To further solve the uni-modal mismatch problem caused by the unrealistic soft-tissue details generated by the modality translation, the multimodal stream network is proposed on the UNET-based cross-modal network to learn the original information from both the fixed and moving images. The dual stream fusion network is responsible for fusing the deformation fields of the uni-modal and multimodal streams. The two registration streams&#x00027; complementary functions preserve the edge details of images. However, the multimodal stream also learns some redundant features, which is not beneficial to the alignment. Unlike the aforementioned methods, Zhou et al. (<xref ref-type="bibr" rid="B88">2021</xref>) translate the CBCT and MRI to the CT modality by Cycle-GAN. The UNET-based segmentation network is trained to get the segmentation map of the synthetic CT image for guiding the robust point matching (RPM) registration. The systems combine the synthesis and segmentation networks to implement cross-modality image segmentation. Arar et al. (<xref ref-type="bibr" rid="B2">2020</xref>) assume that a spatial translation network (STN) (Jaderberg et al., <xref ref-type="bibr" rid="B32">2015</xref>) registration network (R) and the CGAN-based translation net (T) are commutative, i.e., T&#x000B0;R=R&#x000B0;T. Based on this assumption, the optimized L1-reconstruction loss <italic>L</italic><sub><italic>recon</italic></sub>(<italic>T, R</italic>) &#x0003D; ||<italic>O</italic><sub><italic>RT</italic></sub>&#x02212;<italic>I</italic><sub><italic>target</italic></sub>||<sub>1</sub>&#x0002B;||<italic>O</italic><sub><italic>TR</italic></sub>&#x02212;<italic>I</italic><sub><italic>target</italic></sub>||<sub>1</sub>, which encourages T to be geometrically preserved. Benefited from the anatomy-consistency constraints, the registration accuracy can be improved. However, the training of GAN may suffer from non-convergence, which may pose certain additional difficulties to the training of the registration network. Lu et al. (<xref ref-type="bibr" rid="B50">2021</xref>) assessed what role image translation plays in the cross-modal registration based on the performance of Cycle-GAN, which also shows the instability of this approach and the overdependence on the data.</p>
<fig id="F3" position="float">
<label>Figure 3</label>
<caption><p>Overall structures of existing cross-modal image registration methods. <bold>(A)</bold> The overall structure of the latent representation method. <bold>(B)</bold> The overall structure of joint learning-based strategy. <bold>(C)</bold> The overall structure of adversarial learning-based strategy. <bold>(D)</bold> The basic structures of modality translation.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fninf-16-933230-g0003.tif"/>
</fig>
</sec>
<sec>
<title>Image disentangling</title>
<p>Qin et al. (<xref ref-type="bibr" rid="B61">2019</xref>) try to learn a registration function in modal-independent latent shape space in an unsupervised manner. The proposed framework consists of three parts: a disentangling image network <italic>via</italic> unpaired modality translation, a deformable registration network in the disentangled latent space, and a GAN modal to learn a similarity metric in the image space implicitly. Several losses are used in the network to train the three parts, including the self-reconstruction loss, the latent reconstruction loss, the cross-cycle consistency, and the adversarial loss with similarity metrics defined in latent space. The work is capable of translating cross-modal images by image disentangling to obtain shape latent representation related to the image alignment. This method can alleviate unrealistic image generation from the Cycle-GAN-based approaches. However, the deformation field generated by latent shape representation introduces unsmooth edges. Wu and Zhou (<xref ref-type="bibr" rid="B78">2021</xref>) propose a fully unsupervised registration network through image disentangling. The proposed registration framework consists of two parts: one registration network aligns the image from x to y, and the other aligns the image from y to x. Each part consists of two subnetworks: an unsupervised deformable registration network and a disentangling representation network <italic>via</italic> unpaired image-to-image translation. Unlike Qin et al. (<xref ref-type="bibr" rid="B61">2019</xref>), the representation disentangling model aims to drive a deformable registration network for learning the mapping between the two modalities.</p></sec>
<sec>
<title>Latent representation</title>
<p>Mahapatra and Ge (<xref ref-type="bibr" rid="B55">2020</xref>) use a convolutional autoencoder (CAE) network to learn latent space representation for different modalities of images. The generator is fed into latent features from CAE to generate the deformation fields. The intensity and shape constraints are achieved by content loss, including the normal mutual information (NMI), the structural similarity index measure (SSIM) (Wang et al., <xref ref-type="bibr" rid="B73">1987</xref>, <xref ref-type="bibr" rid="B75">2004</xref>), and the visual graphics generator (VGG) (Simonyan and Zisserman, <xref ref-type="bibr" rid="B69">2014</xref>) with L2 loss. The cycle consistency loss and the adversarial loss are used to constrain the deformation field consistency, which is calculated as follows:</p>
<disp-formula id="E13"><label>(9)</label><mml:math id="M13"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mi>L</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>G</mml:mi><mml:mo>,</mml:mo><mml:mi>F</mml:mi><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mrow><mml:msup><mml:mrow><mml:mi>I</mml:mi></mml:mrow><mml:mrow><mml:mi>F</mml:mi><mml:mi>l</mml:mi><mml:mi>t</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mrow><mml:msup><mml:mrow><mml:mi>I</mml:mi></mml:mrow><mml:mrow><mml:mi>R</mml:mi><mml:mi>e</mml:mi><mml:mi>f</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow></mml:msub><mml:mo>&#x0002B;</mml:mo><mml:msub><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi><mml:mi>t</mml:mi><mml:mi>e</mml:mi><mml:mi>n</mml:mi><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mo>&#x0002B;</mml:mo><mml:mi>&#x003BB;</mml:mi><mml:msub><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>y</mml:mi><mml:mi>c</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>G</mml:mi><mml:mo>,</mml:mo><mml:mi>F</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="E14"><mml:math id="M14"><mml:mtable columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi><mml:mi>t</mml:mi><mml:mi>e</mml:mi><mml:mi>n</mml:mi><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msup><mml:mrow><mml:mi>I</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi><mml:mi>r</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi><mml:mi>s</mml:mi></mml:mrow></mml:msup><mml:mo>,</mml:mo><mml:msup><mml:mrow><mml:mi>I</mml:mi></mml:mrow><mml:mrow><mml:mi>R</mml:mi><mml:mi>e</mml:mi><mml:mi>f</mml:mi></mml:mrow></mml:msup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mtext>&#x02003;</mml:mtext><mml:mo>=</mml:mo><mml:mi>N</mml:mi><mml:mi>M</mml:mi><mml:mi>I</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msup><mml:mrow><mml:mi>I</mml:mi></mml:mrow><mml:mrow><mml:mi>R</mml:mi><mml:mi>e</mml:mi><mml:mi>f</mml:mi></mml:mrow></mml:msup><mml:mo>,</mml:mo><mml:msup><mml:mrow><mml:mi>I</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi><mml:mi>r</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi><mml:mi>s</mml:mi></mml:mrow></mml:msup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x0002B;</mml:mo><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mn>1</mml:mn><mml:mo>-</mml:mo><mml:mi>S</mml:mi><mml:mi>S</mml:mi><mml:mi>I</mml:mi><mml:mi>M</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msup><mml:mrow><mml:mi>I</mml:mi></mml:mrow><mml:mrow><mml:mi>R</mml:mi><mml:mi>e</mml:mi><mml:mi>f</mml:mi></mml:mrow></mml:msup><mml:mo>,</mml:mo><mml:msup><mml:mrow><mml:mi>I</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi><mml:mi>r</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi><mml:mi>s</mml:mi></mml:mrow></mml:msup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mtext>&#x02003;&#x000A0;</mml:mtext><mml:mo>&#x0002B;</mml:mo><mml:mi>V</mml:mi><mml:mi>G</mml:mi><mml:mi>G</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msup><mml:mrow><mml:mi>I</mml:mi></mml:mrow><mml:mrow><mml:mi>R</mml:mi><mml:mi>e</mml:mi><mml:mi>f</mml:mi></mml:mrow></mml:msup><mml:mo>,</mml:mo><mml:msup><mml:mrow><mml:mi>I</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi><mml:mi>r</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi><mml:mi>s</mml:mi></mml:mrow></mml:msup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="E15"><label>(10)</label><mml:math id="M15"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>y</mml:mi><mml:mi>c</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>G</mml:mi><mml:mo>,</mml:mo><mml:mi>F</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:mi>x</mml:mi></mml:mrow></mml:msub><mml:mo>|</mml:mo><mml:mo>|</mml:mo><mml:mi>F</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>G</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>-</mml:mo><mml:mi>x</mml:mi><mml:mo>|</mml:mo><mml:msub><mml:mrow><mml:mo>|</mml:mo></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>&#x0002B;</mml:mo><mml:msub><mml:mrow><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:mi>y</mml:mi></mml:mrow></mml:msub><mml:mo>|</mml:mo><mml:mo>|</mml:mo><mml:mi>G</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>F</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>-</mml:mo><mml:mi>y</mml:mi><mml:mo>|</mml:mo><mml:msub><mml:mrow><mml:mo>|</mml:mo></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="E16"><mml:math id="M16"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow></mml:msub></mml:mtd><mml:mtd><mml:mo>=</mml:mo></mml:mtd><mml:mtd><mml:mtext>&#x000A0;&#x000A0;</mml:mtext><mml:msub><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>y</mml:mi><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>G</mml:mi><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mrow><mml:msup><mml:mrow><mml:mi>I</mml:mi></mml:mrow><mml:mrow><mml:mi>R</mml:mi><mml:mi>e</mml:mi><mml:mi>f</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x0002B;</mml:mo><mml:msub><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>y</mml:mi><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>N</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>F</mml:mi><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mrow><mml:msup><mml:mrow><mml:mi>I</mml:mi></mml:mrow><mml:mrow><mml:mi>F</mml:mi><mml:mi>l</mml:mi><mml:mi>t</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="E17"><label>(11)</label><mml:math id="M17"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mo>&#x0002B;</mml:mo><mml:mi>l</mml:mi><mml:mi>o</mml:mi><mml:mi>g</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mn>1</mml:mn><mml:mo>-</mml:mo><mml:mi>M</mml:mi><mml:mi>S</mml:mi><mml:msub><mml:mrow><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:mi>N</mml:mi><mml:mi>o</mml:mi><mml:mi>r</mml:mi><mml:mi>m</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msup><mml:mrow><mml:mi>I</mml:mi></mml:mrow><mml:mrow><mml:mi>D</mml:mi><mml:mi>e</mml:mi><mml:mi>f</mml:mi><mml:mo>-</mml:mo><mml:mi>A</mml:mi><mml:mi>P</mml:mi><mml:mi>P</mml:mi></mml:mrow></mml:msup><mml:mo>,</mml:mo><mml:msup><mml:mrow><mml:mi>I</mml:mi></mml:mrow><mml:mrow><mml:mi>D</mml:mi><mml:mi>e</mml:mi><mml:mi>f</mml:mi><mml:mo>-</mml:mo><mml:mi>R</mml:mi><mml:mi>e</mml:mi><mml:mi>c</mml:mi><mml:mi>v</mml:mi></mml:mrow></mml:msup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where (<italic>G, F)</italic> represents the two generators, <inline-formula><mml:math id="M18"><mml:msub><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mrow><mml:msup><mml:mrow><mml:mi>I</mml:mi></mml:mrow><mml:mrow><mml:mi>F</mml:mi><mml:mi>l</mml:mi><mml:mi>t</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:msub></mml:math></inline-formula> and <inline-formula><mml:math id="M19"><mml:msub><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mrow><mml:msup><mml:mrow><mml:mi>I</mml:mi></mml:mrow><mml:mrow><mml:mi>R</mml:mi><mml:mi>e</mml:mi><mml:mi>f</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:msub></mml:math></inline-formula> represent <italic>I</italic><sup><italic>Flt</italic></sup> and <italic>I</italic><sup><italic>Ref</italic></sup> as the real data of the discriminator, <italic>x</italic> and <italic>y</italic> represent the original images of the two modalities, and <italic>MSE</italic><sub><italic>Norm</italic></sub> is the MSE normalized to [0, 1]. Yang et al. (<xref ref-type="bibr" rid="B83">2020</xref>) transform image modality through a conditional auto-encoder generative adversarial network (CAE-GAN), which redesigns VAE (Kingma and Welling, <xref ref-type="bibr" rid="B38">2013</xref>) and GAN to form the symmetric UNET. The registration network uses a traditional nonparametric deformable method based on local phase differences at multiple scales. The overall structure of the latent representation method is shown in <xref ref-type="fig" rid="F3">Figure 3A</xref>.</p>
</sec>
</sec>
<sec>
<title>Symmetric learning-based strategy</title>
<p><xref ref-type="table" rid="T3">Table 3</xref> lists two papers about using the symmetric learning-based GAN methods. Both of them perform image registration for CT-MRI. One is used on the brain, and another on the heart.</p>
<table-wrap position="float" id="T3">
<label>Table 3</label>
<caption><p>Overview of symmetric learning-based methods.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="left"><bold>Publication</bold></th>
<th valign="top" align="center"><bold>Organ</bold></th>
<th valign="top" align="center"><bold>Method</bold></th>
<th valign="top" align="center"><bold>Modality</bold></th>
<th valign="top" align="center"><bold>Evaluation metrics</bold></th>
<th valign="top" align="center"><bold>Loss</bold></th>
<th valign="top" align="center"><bold>Dataset</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Zheng et al. (<xref ref-type="bibr" rid="B87">2021</xref>)</td>
<td valign="top" align="center">brain</td>
<td valign="top" align="center">GAN</td>
<td valign="top" align="center">CT-MRI</td>
<td valign="top" align="center">M1</td>
<td valign="top" align="center">L1, 2</td>
<td valign="top" align="center">D3, 4, 5, 6, 7</td>
</tr>
<tr>
<td valign="top" align="left">Lu et al. (<xref ref-type="bibr" rid="B52">2019</xref>)</td>
<td valign="top" align="center">heart</td>
<td valign="top" align="center">Cycle-GAN</td>
<td valign="top" align="center">MRI-CT</td>
<td valign="top" align="center">M1</td>
<td valign="top" align="center">L3, 7</td>
<td valign="top" align="center">D2</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p>Some brief descriptions of the loss, dataset, and evaluation metrics can be found in <xref ref-type="table" rid="T6">Tables 6</xref>&#x02013;<xref ref-type="table" rid="T8">8</xref>.</p>
</table-wrap-foot>
</table-wrap>
<sec>
<title>Cyclic-consistency</title>
<p>From the perspective of cyclic learning, symmetric learning can assist and supervise each other. The CIRNet (Lu et al., <xref ref-type="bibr" rid="B52">2019</xref>) uses two cascaded networks with identical structures as the symmetric registration networks. The two cascaded networks share the weights. The L2 loss is used as L<italic>cyc</italic> to enforce image A translating through two deformation fields &#x003D5;1 and &#x003D5;2, and A(&#x003D5;1, &#x003D5;2) = A. L<italic>cyc</italic> is defined as follows:</p>
<disp-formula id="E18"><label>(12)</label><mml:math id="M20"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>y</mml:mi><mml:mi>c</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>A</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>&#x003C6;</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003C6;</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:mi>A</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:mfrac><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:mfrac><mml:mstyle displaystyle="true"><mml:munder class="msub"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mo>&#x02208;</mml:mo><mml:mi>&#x003A9;</mml:mi></mml:mrow></mml:munder></mml:mstyle><mml:msup><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>A</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>&#x003C6;</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003C6;</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:mi>A</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where <italic>N</italic> represents the number of all the voxels, and &#x003A9; refers to all the voxels in the image. The total loss of the two registration networks is represented by Equations (13, 14), respectively:</p>
<disp-formula id="E19"><mml:math id="M21"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mrow><mml:mi>R</mml:mi><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>A</mml:mi><mml:mo>,</mml:mo><mml:mi>B</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd><mml:mtd><mml:mo>=</mml:mo></mml:mtd><mml:mtd><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mrow><mml:mi>N</mml:mi><mml:mi>C</mml:mi><mml:mi>C</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>A</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>&#x003C6;</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:mi>B</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x0002B;</mml:mo><mml:mi>&#x003B1;</mml:mi><mml:msub><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi><mml:mi>e</mml:mi><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>&#x003C6;</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="E20"><label>(13)</label><mml:math id="M22"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mo>&#x0002B;</mml:mo><mml:mi>&#x003B2;</mml:mi><mml:msub><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>y</mml:mi><mml:mi>c</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>A</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>&#x003C6;</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003C6;</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:mi>A</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="E21"><mml:math id="M23"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mn>2</mml:mn><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>A</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>&#x003C6;</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:mi>A</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd><mml:mtd><mml:mo>=</mml:mo></mml:mtd><mml:mtd><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mrow><mml:mi>N</mml:mi><mml:mi>C</mml:mi><mml:mi>C</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>A</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>&#x003C6;</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003C6;</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:mi>A</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x0002B;</mml:mo><mml:mi>&#x003B1;</mml:mi><mml:msub><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi><mml:mi>e</mml:mi><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>&#x003C6;</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="E22"><label>(14)</label><mml:math id="M24"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mo>&#x0002B;</mml:mo><mml:mi>&#x003B2;</mml:mi><mml:msub><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>y</mml:mi><mml:mi>c</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>A</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>&#x003C6;</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003C6;</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:mi>A</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="E23"><label>(15)</label><mml:math id="M25"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi><mml:mi>e</mml:mi><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>&#x003C6;</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mstyle displaystyle="true"><mml:munder class="msub"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mo>&#x02208;</mml:mo><mml:mi>&#x003A9;</mml:mi></mml:mrow></mml:munder></mml:mstyle><mml:msup><mml:mrow><mml:mo>&#x02207;</mml:mo><mml:mi>&#x003C6;</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where <italic>A</italic>(&#x003C6;<sub>1</sub>) denotes a warped image produced by the module R1, <italic>A</italic>(&#x003C6;<sub>1</sub>, &#x003C6;<sub>2</sub>) denotes a warped image produced by the module R2, and <italic>L</italic><sub><italic>reg</italic></sub>(&#x003C6;<sub>1</sub>) is the smooth regularization<italic>L</italic><sub><italic>reg</italic></sub>.</p>
<p>SymReg-GAN (Zheng et al., <xref ref-type="bibr" rid="B87">2021</xref>) proposes a GAN-based symmetric registration to resolve the inverse-consistent translation between cross-modal images. A generator performs the modality translation, consisting of an affine-translation regressor and a non-linear-deformation regressor. The discriminator distinguishes between the translation estimation and the ground truth. The SymReg-GAN is trained by jointly utilizing labeled and unlabeled images. It encourages symmetry in registration by enforcing a condition that in the cycle composed of the transformation from one image to the other, its reverse transformation should bring the original image back. The total loss combines the symmetry loss, registration loss, and supervision loss into one. This method takes full advantage of both labeled and unlabeled data and resolves the limitation of iterative optimization by non-learning techniques. However, the spatial transformation and the modality transformation may not be the same, and even if the spatial transformation is symmetric, the transformation error may still be cyclic.</p>
</sec>
</sec>
<sec>
<title>Adversarial learning-based strategy</title>
<p>Biomedical image registration algorithms of adversarial learning-based strategies utilize adversarial loss to drive the learning of registration networks such as GAN. Adversarial loss consists of two parts: the training aim of the generator is to generate an image that makes the discriminator consider it real, and the optimization objective of the discriminator is to distinguish between an image generated by the generator and a real image in the dataset as accurately as possible. Based on this strategy, several methods, as shown in <xref ref-type="table" rid="T1">Table 1</xref>, including semi-supervised, knowledge distillation, attention mechanisms, and adversarial training, are implemented to improve registration performance. The similarity loss is instead by learning a discriminator network. Although GAN can be trained unsupervised, paired training data may be more helpful for model convergence for the cross-modal registration modal. An overview of the crucial elements of all the reviewed papers is shown in <xref ref-type="table" rid="T4">Table 4</xref>. Five papers are for mono-modal image registration and four for cross-modal registration. The overall structure of this strategy is shown in <xref ref-type="fig" rid="F3">Figure 3C</xref>.</p>
<table-wrap position="float" id="T4">
<label>Table 4</label>
<caption><p>Publications of adversarial strategy based.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="left"><bold>Publications</bold></th>
<th valign="top" align="center"><bold>Organ</bold></th>
<th valign="top" align="center"><bold>Method</bold></th>
<th valign="top" align="center"><bold>Modality</bold></th>
<th valign="top" align="center"><bold>Evaluation metrics</bold></th>
<th valign="top" align="center"><bold>Loss</bold></th>
<th valign="top" align="left"><bold>Dataset</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Tran et al. (<xref ref-type="bibr" rid="B72">2022</xref>)</td>
<td valign="top" align="center">Liver, brain</td>
<td valign="top" align="center">GAN</td>
<td valign="top" align="center">CT-CT/MRI-MRI</td>
<td valign="top" align="center">M1, 11</td>
<td valign="top" align="center">L1, 2</td>
<td valign="top" align="left">D8, 9, 10, 11, 12, 13, 14, 15, 16, 17</td>
</tr>
<tr>
<td valign="top" align="left">Li and Ogino (<xref ref-type="bibr" rid="B46">2019</xref>)</td>
<td valign="top" align="center">Liver</td>
<td valign="top" align="center">GAN</td>
<td valign="top" align="center">/</td>
<td valign="top" align="center">M1, 2</td>
<td valign="top" align="center">L1, 2, 5</td>
<td/>
</tr>
<tr>
<td valign="top" align="left">Hu et al. (<xref ref-type="bibr" rid="B29">2018</xref>)</td>
<td valign="top" align="center">Prostate</td>
<td valign="top" align="center">GAN</td>
<td valign="top" align="center">MRI-TRUS</td>
<td/>
<td/>
<td/>
</tr>
<tr>
<td valign="top" align="left">Bessadok et al. (<xref ref-type="bibr" rid="B9">2021</xref>)</td>
<td valign="top" align="center">Brain</td>
<td valign="top" align="center">GAN</td>
<td valign="top" align="center">MRI-MRI</td>
<td valign="top" align="center">M5, 12</td>
<td valign="top" align="center">L1</td>
<td valign="top" align="left">D15</td>
</tr>
<tr>
<td valign="top" align="left">Li M. et al. (<xref ref-type="bibr" rid="B44">2021</xref>)</td>
<td valign="top" align="center">Brain</td>
<td valign="top" align="center">GAN</td>
<td valign="top" align="center">MRI-PET</td>
<td valign="top" align="center">M1</td>
<td valign="top" align="center">L2, 5</td>
<td valign="top" align="left">D29</td>
</tr>
<tr>
<td valign="top" align="left">Fan et al. (<xref ref-type="bibr" rid="B18">2018</xref>)</td>
<td valign="top" align="center">Brain</td>
<td valign="top" align="center">GAN</td>
<td valign="top" align="center">MRI-MRI</td>
<td valign="top" align="center">M1, 4</td>
<td valign="top" align="center">L1</td>
<td valign="top" align="left">D18</td>
</tr>
<tr>
<td valign="top" align="left">Fan et al. (<xref ref-type="bibr" rid="B17">2019</xref>)</td>
<td valign="top" align="center">Brain</td>
<td valign="top" align="center">GAN</td>
<td valign="top" align="center">MRI-MRI</td>
<td valign="top" align="center">M1, 4</td>
<td valign="top" align="center">L1</td>
<td valign="top" align="left">D18</td>
</tr>
<tr>
<td valign="top" align="left">Yan et al. (<xref ref-type="bibr" rid="B82">2018</xref>)</td>
<td valign="top" align="center">Rectum</td>
<td valign="top" align="center">GAN</td>
<td valign="top" align="center">MRI-TRUS</td>
<td valign="top" align="center">M1, 2</td>
<td valign="top" align="center">L1</td>
<td/>
</tr>
<tr>
<td valign="top" align="left">Tran et al. (<xref ref-type="bibr" rid="B72">2022</xref>)</td>
<td valign="top" align="center">Prostate cancer</td>
<td valign="top" align="center">GAN</td>
<td valign="top" align="center">MRI-TRUS</td>
<td valign="top" align="center">M2</td>
<td valign="top" align="center">L1</td>
<td/>
</tr>
<tr>
<td valign="top" align="left">Luo et al. (<xref ref-type="bibr" rid="B53">2021</xref>)</td>
<td valign="top" align="center">Lung</td>
<td valign="top" align="center">GAN</td>
<td valign="top" align="center">X-rays-X-rays</td>
<td valign="top" align="center">M1, 3, 4, 11</td>
<td valign="top" align="center">L6, 7, 9</td>
<td valign="top" align="left">D20</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p>A brief description of the loss, dataset, and evaluation metrics can be found in <xref ref-type="table" rid="T6">Tables 6</xref>&#x02013;<xref ref-type="table" rid="T8">8</xref>. The symbol&#x02013;in the last column means that the data used in the paper is a private dataset.</p>
</table-wrap-foot>
</table-wrap>
<sec>
<title>Adversarial training</title>
<p>For image registration tasks, the effect of adversarial loss is to make the warped image closer to the target image (Yan et al., <xref ref-type="bibr" rid="B82">2018</xref>). The role of the generator is to generate a deformation field, and the task of the discriminator is to identify the alignment image. For more stable loss, Wasserstein GAN (WGAN) is adopted. Since a discriminator can be considered a registration image quality assessor, the quality of a warped image can be improved with cross-modal similarity metrics. However, the training of GAN may suffer from non-convergence, which may pose additional difficulties in training the registration network. Compared to Fan et al. (<xref ref-type="bibr" rid="B18">2018</xref>, <xref ref-type="bibr" rid="B17">2019</xref>) and Yan et al. (<xref ref-type="bibr" rid="B82">2018</xref>) select more reasonable reference data for training the discriminator for better model convergence.</p></sec>
<sec>
<title>Semi-supervised</title>
<p>Hu et al. (<xref ref-type="bibr" rid="B29">2018</xref>) use a biomechanical simulation deformation field to regularize the deformation fields formed by the alignment network. The generator is fed into simulated motion data to form a translation. The discriminator tries to distinguish the dense displacement field from ground truth deformation. Another similarity loss metric warps moving labels and fixed labels in a weakly supervised manner. Li and Ogino (<xref ref-type="bibr" rid="B46">2019</xref>) propose a general end-to-end registration network in which a CNN similar to UNET is trained to generate the deformation field. For better guiding, the anatomical shape alignment and masks of moving and fixed objects are also fed into the registration network. The input of the discriminator net is a positive and negative alignment pair, consisting of masks and images of the fixed and warped images for guiding a finer anatomy alignment. In addition, an encoder extracts anatomical shape differences as another registration loss. The studies by Elmahdy et al. (<xref ref-type="bibr" rid="B16">2019</xref>) and Luo et al. (<xref ref-type="bibr" rid="B53">2021</xref>) are similar to that of Li and Ogino (<xref ref-type="bibr" rid="B46">2019</xref>), in which the adversarial learning-based registration network joint segmentation and registration with segmentation label information as the input of the generator and the discriminator. The dice similarity coefficient (DSC) and the normalized cross-correlation (NCC) are added to the generator to avoid slow convergence and suboptimal registration.</p></sec>
<sec>
<title>Knowledge distillation</title>
<p>Knowledge distillation is a process of transferring knowledge from a cumbersome pre-trained model (i.e., the teacher network) to a light-weighted one (i.e., the student network). Tran et al. (<xref ref-type="bibr" rid="B72">2022</xref>) used knowledge distillation by adversarial learning to streamline the expensive and effective teacher registration model to a light-weighted student registration model. In their proposed method, the teacher network is the recursive cascaded network (RCN) (Zhao et al., <xref ref-type="bibr" rid="B85">2019a</xref>) for extracting meaningful deformations, and the student network is a CNN-based registration network. When training the registration network, the teacher network and the student network are optimized by <italic>L</italic><sub><italic>adv</italic></sub> &#x0003D; &#x003A5;<italic>l</italic><sub><italic>rec</italic></sub>&#x0002B;(1&#x02212;&#x003A5;)<italic>l</italic><sub><italic>dis</italic></sub>, where <italic>l</italic><sub><italic>rec</italic></sub> represents the reconstructed loss and the discriminator loss, and <italic>l</italic><sub><italic>rec</italic></sub> and <italic>l</italic><sub><italic>dis</italic></sub> are expressed as follows:</p>
<disp-formula id="E24"><label>(16)</label><mml:math id="M26"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>l</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi><mml:mi>e</mml:mi><mml:mi>c</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msubsup><mml:mrow><mml:mi>I</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow><mml:mrow><mml:mi>h</mml:mi></mml:mrow></mml:msubsup><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi>I</mml:mi></mml:mrow><mml:mrow><mml:mi>f</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mn>1</mml:mn><mml:mo>-</mml:mo><mml:mi>C</mml:mi><mml:mi>o</mml:mi><mml:mi>r</mml:mi><mml:mi>r</mml:mi><mml:mi>C</mml:mi><mml:mi>o</mml:mi><mml:mi>e</mml:mi><mml:mi>f</mml:mi><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msubsup><mml:mrow><mml:mi>I</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow><mml:mrow><mml:mi>h</mml:mi></mml:mrow></mml:msubsup><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi>I</mml:mi></mml:mrow><mml:mrow><mml:mi>f</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="E25"><label>(17)</label><mml:math id="M27"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>l</mml:mi></mml:mrow><mml:mrow><mml:mi>d</mml:mi><mml:mi>i</mml:mi><mml:mi>s</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mo>|</mml:mo><mml:mo>|</mml:mo><mml:msub><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mo class="MathClass-ord">&#x02205;</mml:mo></mml:mrow><mml:mrow><mml:mi>s</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mo class="MathClass-ord">&#x02205;</mml:mo></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo stretchy="false">)</mml:mo><mml:mo>|</mml:mo><mml:msubsup><mml:mrow><mml:mo>|</mml:mo></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup><mml:mo>&#x0002B;</mml:mo><mml:mi>&#x003BB;</mml:mi><mml:msup><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mo>|</mml:mo><mml:mo>|</mml:mo><mml:msub><mml:mrow><mml:mo>&#x02207;</mml:mo></mml:mrow><mml:mrow><mml:mover accent="true"><mml:mrow><mml:msub><mml:mrow><mml:mo class="MathClass-ord">&#x02205;</mml:mo></mml:mrow><mml:mrow><mml:mi>s</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mover accent="true"><mml:mrow><mml:msub><mml:mrow><mml:mo class="MathClass-ord">&#x02205;</mml:mo></mml:mrow><mml:mrow><mml:mi>s</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>|</mml:mo><mml:msub><mml:mrow><mml:mo>|</mml:mo></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="E26"><label>(18)</label><mml:math id="M28"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mover accent="true"><mml:mrow><mml:msub><mml:mrow><mml:mo class="MathClass-ord">&#x02205;</mml:mo></mml:mrow><mml:mrow><mml:mi>s</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>^</mml:mo></mml:mover><mml:mo>=</mml:mo><mml:mi>&#x003B2;</mml:mi><mml:msub><mml:mrow><mml:mo class="MathClass-ord">&#x02205;</mml:mo></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mo>&#x0002B;</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mn>1</mml:mn><mml:mo>-</mml:mo><mml:mi>&#x003B2;</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:msub><mml:mrow><mml:mo class="MathClass-ord">&#x02205;</mml:mo></mml:mrow><mml:mrow><mml:mi>s</mml:mi></mml:mrow></mml:msub></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where <inline-formula><mml:math id="M29"><mml:msubsup><mml:mrow><mml:mi>I</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow><mml:mrow><mml:mi>h</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> denotes a warped image by the student network, <italic>CorrCoef</italic>[<italic>I</italic><sub>1</sub>&#x02212;<italic>I</italic><sub>2</sub>] is the correlation between images I1 and I2, &#x02205;<sub><italic>s</italic></sub> and &#x02205;<sub><italic>t</italic></sub> denote the deformation of the teacher and the student networks, respectively, and<inline-formula><mml:math id="M30"><mml:mtext>&#x000A0;</mml:mtext><mml:mover accent="true"><mml:mrow><mml:msub><mml:mrow><mml:mo class="MathClass-ord">&#x02205;</mml:mo></mml:mrow><mml:mrow><mml:mi>s</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:math></inline-formula> denotes the joint deformation. Applying knowledge distillation by means of adversarial learning provides a new and efficient way to reduce computational costs and achieve competitive accuracy.</p></sec>
<sec>
<title>Attention mechanisms</title>
<p>To reduce feature loss of the upsampling process in a registration network, Li M. et al. (<xref ref-type="bibr" rid="B44">2021</xref>) proposed a GAN-based registration network combining UNET with dual attention mechanisms. The dual attention mechanisms consist of the channel attention mechanism and the location attention mechanism. Meanwhile, the residual structure is also introduced into the upsampling process for improving feature restoration.</p>
</sec>
</sec>
<sec>
<title>Joint learning-based strategy</title>
<p>Joint learning of segmentation, registration, and synthesis networks can improve their performance for each other. An overview of the essential elements of all the reviewed papers is shown in <xref ref-type="table" rid="T5">Table 5</xref>. Two of which are for mono-model registration methods. The overall structure is shown in <xref ref-type="fig" rid="F3">Figure 3B</xref>.</p>
<table-wrap position="float" id="T5">
<label>Table 5</label>
<caption><p>Joint learning-based methods and publications.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="left"><bold>Publication</bold></th>
<th valign="top" align="center"><bold>Organ</bold></th>
<th valign="top" align="center"><bold>Method</bold></th>
<th valign="top" align="center"><bold>Modality</bold></th>
<th valign="top" align="center"><bold>Evaluation metrics</bold></th>
<th valign="top" align="center"><bold>Loss</bold></th>
<th valign="top" align="center"><bold>Dataset</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Liu et al. (<xref ref-type="bibr" rid="B48">2020</xref>)</td>
<td valign="top" align="center">Liver tumor</td>
<td valign="top" align="center">CGAN</td>
<td valign="top" align="center">CT-CT</td>
<td valign="top" align="center">M1, 3, 4</td>
<td valign="top" align="center">L1,2</td>
<td valign="top" align="center">&#x02013;</td>
</tr>
<tr>
<td valign="top" align="left">Zhou et al. (<xref ref-type="bibr" rid="B88">2021</xref>)</td>
<td valign="top" align="center">Liver tumor</td>
<td valign="top" align="center">CycleGAN</td>
<td valign="top" align="center">MRI-CBCT</td>
<td valign="top" align="center">M1, 10</td>
<td valign="top" align="center">L3, 4, 5, 6, 8</td>
<td valign="top" align="center">D8</td>
</tr>
<tr>
<td valign="top" align="left">Mahapatra et al. (<xref ref-type="bibr" rid="B56">2018</xref>)</td>
<td valign="top" align="center">Lung</td>
<td valign="top" align="center">CycleGAN</td>
<td valign="top" align="center">X-rays- X-rays</td>
<td valign="top" align="center">M1, 10</td>
<td valign="top" align="center">L1, 3, 9</td>
<td valign="top" align="center">D23</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p>A brief description of the loss, dataset, and evaluation metrics can be found in <xref ref-type="table" rid="T6">Tables 6</xref>&#x02013;<xref ref-type="table" rid="T8">8</xref>. The symbol&#x02013;in the last column means that the data used in the paper is a private dataset.</p>
</table-wrap-foot>
</table-wrap>
<table-wrap position="float" id="T6">
<label>Table 6</label>
<caption><p>A brief summary of different losses used in the reviewed publications in <xref ref-type="table" rid="T2">Tables 2</xref>&#x02013;<xref ref-type="table" rid="T5">5</xref>.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="left"><bold>Abbre</bold></th>
<th valign="top" align="left"><bold>Loss</bold></th>
<th valign="top" align="left"><bold>Name</bold></th>
<th valign="top" align="left"><bold>Remark</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">L1</td>
<td valign="top" align="left">L<sub>adv</sub></td>
<td valign="top" align="left">23 adversarial learning loss</td>
<td valign="top" align="left">The discriminator introduces the adversarial loss to distinguish synthetic data from real data, consisting of cross-entropy loss and least squares loss</td>
</tr>
<tr>
<td valign="top" align="left">L2</td>
<td valign="top" align="left">L<sub>pix</sub></td>
<td valign="top" align="left">Pix-level supervision loss</td>
<td valign="top" align="left">Pix-level loss evaluates the different intensity values, consisting of L1, L2, and Frobenius norm</td>
</tr>
<tr>
<td valign="top" align="left">L3</td>
<td valign="top" align="left">L<sub>cyc</sub></td>
<td valign="top" align="left">Cycle consistency loss</td>
<td valign="top" align="left">Element-wise loss measures the self-similarity of the image of cycled translation to the source domain image when training with an unpaired image from two domains</td>
</tr>
<tr>
<td valign="top" align="left">L4</td>
<td valign="top" align="left">L<sub>mind</sub></td>
<td valign="top" align="left">Modality-independent neighborhood descriptor</td>
<td valign="top" align="left">The pixel-level similarity metric is used to measure the structural similarity between two different modal images</td>
</tr>
<tr>
<td valign="top" align="left">L5</td>
<td valign="top" align="left">L<sub>cc</sub></td>
<td valign="top" align="left">Correlation coefficient loss</td>
<td valign="top" align="left">Structural similarity metrics between two different modal images</td>
</tr>
<tr>
<td valign="top" align="left">L6</td>
<td valign="top" align="left">L<sub>seg</sub></td>
<td valign="top" align="left">Segmentation loss</td>
<td valign="top" align="left">Measuring the difference between a segmented prediction label and a ground truth label</td>
</tr>
<tr>
<td valign="top" align="left">L7</td>
<td valign="top" align="left">L<sub>Ncc</sub></td>
<td valign="top" align="left">Normalized cross-correlation</td>
<td valign="top" align="left">Used for mono alignment tasks to measure the level of alignment of the warped image to the fixed image</td>
</tr>
<tr>
<td valign="top" align="left">L8</td>
<td valign="top" align="left">L<sub>idt</sub></td>
<td valign="top" align="left">Identity loss</td>
<td valign="top" align="left">Identity loss regularizes the generators to be near an identity mapping when real samples of the target domain are provided</td>
</tr>
<tr>
<td valign="top" align="left">L9</td>
<td valign="top" align="left">DM</td>
<td valign="top" align="left">Hamming distance</td>
<td valign="top" align="left">Pix-level similarity metrics for image feature focus on a hash value</td>
</tr>
<tr>
<td valign="top" align="left">L10</td>
<td valign="top" align="left">L<sub>lat</sub></td>
<td valign="top" align="left">Latent reconstruction loss</td>
<td valign="top" align="left">Similarity measure of latent spatial features</td>
</tr>
<tr>
<td valign="top" align="left">L11</td>
<td valign="top" align="left">L<sub>Leastsquares</sub></td>
<td valign="top" align="left">Least squares loss</td>
<td valign="top" align="left">Used in generators and discriminators as adversarial loss</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p>The L in the first column represents the loss.</p>
</table-wrap-foot>
</table-wrap>
<table-wrap position="float" id="T7">
<label>Table 7</label>
<caption><p>A brief summary of different metrics, which are all in respect to the ground truth.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="left"><bold>Abbr</bold></th>
<th valign="top" align="left"><bold>Metrics</bold></th>
<th valign="top" align="left"><bold>Remarks</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">M1</td>
<td valign="top" align="left">DICE, Median DSC</td>
<td valign="top" align="left">The dice coefficient calculates the degree of overlap between the aligned image and the ground truth</td>
</tr>
<tr>
<td valign="top" align="left">M2</td>
<td valign="top" align="left">TRE (Targeted registration error)</td>
<td valign="top" align="left">TRE represents the distance sum of the corresponding landmark point between the target image and the aligned image</td>
</tr>
<tr>
<td valign="top" align="left">M3</td>
<td valign="top" align="left">HD (Hausdorff distance)</td>
<td valign="top" align="left">HD computes the distance between the contour of the predicted segmentation region and the ground truth to measure shape similarity</td>
</tr>
<tr>
<td valign="top" align="left">M4</td>
<td valign="top" align="left">ASD, ASSD (Average symmetric surface distance)</td>
<td valign="top" align="left">Average symmetric surface distance metrics for measuring image alignment</td>
</tr>
<tr>
<td valign="top" align="left">M5</td>
<td valign="top" align="left">AE (Average euclidean)</td>
<td valign="top" align="left">Pixel-level description of the distance between the aligned image and the fixed image to measure the similarity</td>
</tr>
<tr>
<td valign="top" align="left">M6</td>
<td valign="top" align="left">RMSE (Root mean square error)</td>
<td valign="top" align="left">Used in alignment to describe the deviation of pixels between the aligned image and the real image</td>
</tr>
<tr>
<td valign="top" align="left">M7</td>
<td valign="top" align="left">MI (Mutual information)</td>
<td valign="top" align="left">A metric commonly used for cross-modal registration</td>
</tr>
<tr>
<td valign="top" align="left">M8</td>
<td valign="top" align="left">NMI (Normal mutual information)</td>
<td valign="top" align="left">Used to measure intensity consistency of images</td>
</tr>
<tr>
<td valign="top" align="left">M9</td>
<td valign="top" align="left">SLPD (Sum of local phase differences)</td>
<td valign="top" align="left">Measure the similarity by the sum of local phase differences</td>
</tr>
<tr>
<td valign="top" align="left">M10</td>
<td valign="top" align="left">Mean &#x000B1; Std</td>
<td valign="top" align="left">Means and standard deviations between two images</td>
</tr>
<tr>
<td valign="top" align="left">M11</td>
<td valign="top" align="left">Jcd (Jaccard)</td>
<td valign="top" align="left">Used in alignment tasks to describe the dissimilarity between images</td>
</tr>
<tr>
<td valign="top" align="left">M12</td>
<td valign="top" align="left">PCC (Pearson correlation coefficient)</td>
<td valign="top" align="left">Like M5, but is more suitable in higher dimensions</td>
</tr>
<tr>
<td valign="top" align="left">M13</td>
<td valign="top" align="left">MCD (Mean contour distance)</td>
<td valign="top" align="left">Measure the similarity of images by Mean Contour Distance</td>
</tr>
<tr>
<td valign="top" align="left">M15</td>
<td valign="top" align="left">MNCC (Mean normalized cross correlation)</td>
<td valign="top" align="left">Mean Normalized correlation between two images</td>
</tr>
<tr>
<td valign="top" align="left">M16</td>
<td valign="top" align="left">SSIM</td>
<td valign="top" align="left">Metrics the structural similarity with respect to a given ground truth</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p>The M in the first column indicates metrics.</p>
</table-wrap-foot>
</table-wrap>
<table-wrap position="float" id="T8">
<label>Table 8</label>
<caption><p>Common datasets used in the reviewed literature.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="left"><bold>Abbre</bold></th>
<th valign="top" align="left"><bold>Dataset</bold></th>
<th valign="top" align="left"><bold>Anatomy</bold></th>
<th valign="top" align="left"><bold>Purpose</bold></th>
<th valign="top" align="left"><bold>Modality</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">D1</td>
<td valign="top" align="left">Abdomen (ABD)</td>
<td valign="top" align="left">Kidney</td>
<td valign="top" align="left">Healthy abdominal organ segmentation</td>
<td valign="top" align="left">CT, MR</td>
</tr>
<tr>
<td valign="top" align="left">D2 (Bernard et al., <xref ref-type="bibr" rid="B8">2018</xref>)</td>
<td valign="top" align="left">ACDC</td>
<td valign="top" align="left">Heart</td>
<td valign="top" align="left">Heart segmentation</td>
<td valign="top" align="left">MRI</td>
</tr>
<tr>
<td valign="top" align="left">D3 (Bakas et al., <xref ref-type="bibr" rid="B5">2018</xref>)</td>
<td valign="top" align="left">(BraTS) 2018</td>
<td valign="top" align="left">Brain</td>
<td valign="top" align="left">Brain tumor segmentation</td>
<td valign="top" align="left">MRI</td>
</tr>
<tr>
<td valign="top" align="left">D4 (Gousias et al., <xref ref-type="bibr" rid="B23">2012</xref>)</td>
<td valign="top" align="left">ALBERTs</td>
<td valign="top" align="left">Newborn brain</td>
<td valign="top" align="left">Manual segmentation of labeled atlases</td>
<td valign="top" align="left">MRI</td>
</tr>
<tr>
<td valign="top" align="left">D5</td>
<td valign="top" align="left">LPBA40</td>
<td valign="top" align="left">Brain</td>
<td valign="top" align="left">Medical image registration for Continuous Registration Challenge</td>
<td valign="top" align="left">MRI</td>
</tr>
<tr>
<td valign="top" align="left">D6</td>
<td valign="top" align="left">IBSR18</td>
<td valign="top" align="left">Brain</td>
<td valign="top" align="left">Medical image registration for Continuous Registration Challenge</td>
<td valign="top" align="left">T1-weighted</td>
</tr>
<tr>
<td valign="top" align="left">D7</td>
<td valign="top" align="left">CUMC12 and MGH10</td>
<td valign="top" align="left">Brain</td>
<td valign="top" align="left">Medical image registration for Continuous Registration Challenge</td>
<td valign="top" align="left">MRI</td>
</tr>
<tr>
<td valign="top" align="left">D8 (Bilic et al., <xref ref-type="bibr" rid="B10">2019</xref>)</td>
<td valign="top" align="left">LiTS</td>
<td valign="top" align="left">Liver</td>
<td valign="top" align="left">Liver segmentation</td>
<td valign="top" align="left">CT</td>
</tr>
<tr>
<td valign="top" align="left">D9 (Kavur et al., <xref ref-type="bibr" rid="B36">2021</xref>)</td>
<td valign="top" align="left">CHAOS</td>
<td valign="top" align="left">Liver</td>
<td valign="top" align="left">Liver segmentation</td>
<td valign="top" align="left">CT</td>
</tr>
<tr>
<td valign="top" align="left">D10 (Antonelli et al., <xref ref-type="bibr" rid="B1">2022</xref>)</td>
<td valign="top" align="left">MSD</td>
<td valign="top" align="left">Liver</td>
<td valign="top" align="left">Liver tumor segmentation</td>
<td valign="top" align="left">CT</td>
</tr>
<tr>
<td valign="top" align="left">D11 (Zhao et al., <xref ref-type="bibr" rid="B86">2019b</xref>)</td>
<td valign="top" align="left">BFH</td>
<td valign="top" align="left">Liver</td>
<td valign="top" align="left">Liver tumor segmentation</td>
<td valign="top" align="left">CT</td>
</tr>
<tr>
<td valign="top" align="left">D12 (Heimann et al., <xref ref-type="bibr" rid="B27">2009</xref>)</td>
<td valign="top" align="left">SLIVER</td>
<td valign="top" align="left">Liver</td>
<td valign="top" align="left">Liver segmentation</td>
<td valign="top" align="left">CT</td>
</tr>
<tr>
<td valign="top" align="left">D13</td>
<td valign="top" align="left">LSPIG</td>
<td valign="top" align="left">Liver</td>
<td valign="top" align="left">Liver Segmentation of Pigs</td>
<td valign="top" align="left">CT</td>
</tr>
<tr>
<td valign="top" align="left">D14 (Mueller et al., <xref ref-type="bibr" rid="B59">2005</xref>)</td>
<td valign="top" align="left">ADNI</td>
<td valign="top" align="left">Brain</td>
<td valign="top" align="left">Brain MRI</td>
<td valign="top" align="left">MRI</td>
</tr>
<tr>
<td valign="top" align="left">D15 (Di Martino et al., <xref ref-type="bibr" rid="B15">2014</xref>)</td>
<td valign="top" align="left">ABIDE</td>
<td valign="top" align="left">Brain</td>
<td valign="top" align="left">Toward a large-scale evaluation of the intrinsic brain architecture in autism</td>
<td valign="top" align="left">R-FMRI</td>
</tr>
<tr>
<td valign="top" align="left">D16 (Bellec et al., <xref ref-type="bibr" rid="B7">2017</xref>)</td>
<td valign="top" align="left">ADHD</td>
<td valign="top" align="left">Brain</td>
<td valign="top" align="left">Brain MRI</td>
<td valign="top" align="left">MRI</td>
</tr>
<tr>
<td valign="top" align="left">D17 (Shattuck et al., <xref ref-type="bibr" rid="B67">2008</xref>)</td>
<td valign="top" align="left">LPBA</td>
<td valign="top" align="left">Brain</td>
<td valign="top" align="left">Brain MRI</td>
<td valign="top" align="left">MRI</td>
</tr>
<tr>
<td valign="top" align="left">D18 (Klein A. et al., <xref ref-type="bibr" rid="B39">2009</xref>)</td>
<td valign="top" align="left">&#x02022;LPBA40 &#x02022;IBSR18 &#x02022;CUMC12 &#x02022;MGH10</td>
<td valign="top" align="left">Brain</td>
<td valign="top" align="left">Medical image registration for Continuous Registration Challenge</td>
<td valign="top" align="left">MRI</td>
</tr>
<tr>
<td valign="top" align="left">D20 (Shiraishi et al., <xref ref-type="bibr" rid="B68">2000</xref>)</td>
<td valign="top" align="left">JSRT</td>
<td valign="top" align="left">Chest radiographs</td>
<td valign="top" align="left">Radiologists&#x00027; detection of pulmonary nodules</td>
<td valign="top" align="left">x-ray</td>
</tr>
<tr>
<td valign="top" align="left">D2 (Candemir et al., <xref ref-type="bibr" rid="B12">2013</xref>)</td>
<td valign="top" align="left">MONT</td>
<td valign="top" align="left">Chest radiographs</td>
<td valign="top" align="left">Lung segmentation</td>
<td valign="top" align="left">x-ray</td>
</tr>
<tr>
<td valign="top" align="left">D22 (Jaeger et al., <xref ref-type="bibr" rid="B33">2013</xref>)</td>
<td valign="top" align="left">SHEN</td>
<td valign="top" align="left">Chest radiographs</td>
<td valign="top" align="left">Automatic tuberculosis screening</td>
<td valign="top" align="left">x-ray</td>
</tr>
<tr>
<td valign="top" align="left">D23 (Wang et al., <xref ref-type="bibr" rid="B74">2017</xref>)</td>
<td valign="top" align="left">NIH ChestXray14</td>
<td valign="top" align="left">Chest</td>
<td valign="top" align="left">Classification studies</td>
<td valign="top" align="left">x-ray</td>
</tr>
<tr>
<td valign="top" align="left">D24 (Menze et al., <xref ref-type="bibr" rid="B57">2014</xref>)</td>
<td valign="top" align="left">BraTS&#x00027;2017</td>
<td valign="top" align="left">Brain</td>
<td valign="top" align="left">The Brain Tumor Segmentation</td>
<td valign="top" align="left">MRI</td>
</tr>
<tr>
<td valign="top" align="left">D25</td>
<td valign="top" align="left">IXI</td>
<td valign="top" align="left">Brain</td>
<td valign="top" align="left">Analysis of brain development</td>
<td valign="top" align="left">MR</td>
</tr>
<tr>
<td valign="top" align="left">D27</td>
<td valign="top" align="left">&#x02022;RIRE &#x02022;Atlas</td>
<td valign="top" align="left">Brain</td>
<td valign="top" align="left">image registration evaluation</td>
<td valign="top" align="left">CT, MR</td>
</tr>
<tr>
<td valign="top" align="left">D28 (LaMontagne et al., <xref ref-type="bibr" rid="B41">2019</xref>)</td>
<td valign="top" align="left">OASIS-3</td>
<td valign="top" align="left">Brain</td>
<td valign="top" align="left">Cognitive Dataset for Normal Aging and Alzheimer&#x00027;s Disease</td>
<td valign="top" align="left">MR</td>
</tr>
</tbody>
</table>
</table-wrap>
<sec>
<title>Multitask</title>
<p>Existing experiments show that a segmentation map can help registration by joint learning. However, in real registration tasks, segmentation labels may not be available. Liu et al. (<xref ref-type="bibr" rid="B48">2020</xref>) propose a joint system of segmentation, registration, and synthesis <italic>via</italic> multi-task learning. The objectives of the CGAN-based synthesis model and the registration model are optimized <italic>via</italic> a joint loss. The segmentation network is trained in a supervised manner. The segmentation module estimates the segmentation map for the moving, fixed, and synthesized images. During the training procedure, a dice loss is optimized between the segmentation maps of the warped moving image and the fixed image. The result proves that the segmentation task can improve registration accuracy. Zhou et al. (<xref ref-type="bibr" rid="B88">2021</xref>) take advantage of each other through the joint Cycle-GAN and UNET-based segmentation network to solve the missing label problem <italic>via</italic> Cycle-GAN&#x00027;s translating of the two modalities to the third one with a large number of available labels. Thus, the synthesis network improves the segmentation accuracy and further improves the accuracy of the RPM registration. Mahapatra et al. (<xref ref-type="bibr" rid="B56">2018</xref>) have trained the generation network to complete the alignment of the reference and moving images by combining the segmentation map that is used directly as the input to the generator, with no need to train an additional segmentation network. Segmentation and alignment are mutually driven. The ways of joining image registration task and image segmentation task may improve the accuracy by sharing the result of learning, which can expand the goal of the registration research.</p>
</sec>
</sec>
</sec>
<sec id="s4">
<title>Statistics</title>
<p>It is essential to conduct relevant analyses from a global perspective after a detailed study of each category of biomedical image registration strategies. In the past 4 years, more than half of the reviewed works have used the modality-independent-based strategy to solve cross-modal biomedical registration&#x02014;the methods of Adversarial Learning Based Strategy account for 32%. From 2020 to 2021, the number of articles published on the modality-independent-based strategy was higher than others, peaking in 2020. However, there is a drop-down trend in 2021. As noted, no paper on the adversarial learning-based strategy was published in 2020. In the other years, the works on the adversarial learning-based strategy are published in a balanced proportion, with the detailed percentages shown in <xref ref-type="fig" rid="F4">Figures 4</xref>, <xref ref-type="fig" rid="F5">5</xref>. In addition to analyzing the trends in the published number of papers and the popularity of the four strategy categories, we also analyzed the percentage distribution of the other characteristics, which is shown in <xref ref-type="fig" rid="F6">Figure 6</xref>. A total of 75% of the works aim to solve the problem of the cross-modal domain of biomedical image registration, among which 46 and 42% adopt direct or indirect Cycle-GAN and GAN are part of the important structure of the registration framework. Cycle-GAN is utilized only for the cross-modal domain of biomedical image registration, whereas GAN is utilized for both cross-modal and uni-modal image registration. In the cross-modal bioimage registration, 33% of the works perform image registration between CT and MRI. The number of articles using MRI accounts for 97%. Regarding the region of interest (ROI), the brain and liver are the most studied sites. The brain is the top registration target in all works. The reason for the wide adoption of the brain consists of its clinical importance, availability in public datasets, and relative simplicity of registration.</p>
<fig id="F4" position="float">
<label>Figure 4</label>
<caption><p>Proportional distribution pie chart of the number of publications on different implementation strategies.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fninf-16-933230-g0004.tif"/>
</fig>
<fig id="F5" position="float">
<label>Figure 5</label>
<caption><p>Contrast bar graph of the number of publications on the four strategies for the GAN-based biomedical image registration over the last 5 years from 2018&#x02013;2022.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fninf-16-933230-g0005.tif"/>
</fig>
<fig id="F6" position="float">
<label>Figure 6</label>
<caption><p>Percentage pie chart of various attributes of the GAN-based biomedical image registration methods.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fninf-16-933230-g0006.tif"/>
</fig>
<p>In addition, the metrics used in the cross-modal registration methods are shown in <xref ref-type="fig" rid="F7">Figure 7</xref>. As seen in the figure, dice and TRE are the top two most frequently used metrics. The Dice coefficient calculates the degree of overlap between the aligned image and the ground truth, and the confusion matrix formula is as follows:</p>
<disp-formula id="E27"><label>(19)</label><mml:math id="M31"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mi>D</mml:mi><mml:mi>i</mml:mi><mml:mi>c</mml:mi><mml:mi>e</mml:mi><mml:mo>=</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:mfrac><mml:mrow><mml:msup><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mo>*</mml:mo></mml:mrow></mml:msup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>r</mml:mi><mml:mi>e</mml:mi><mml:mi>f</mml:mi><mml:mo>&#x02229;</mml:mo><mml:mi>w</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mi>r</mml:mi><mml:mi>e</mml:mi><mml:mi>f</mml:mi><mml:mo>&#x0222A;</mml:mo><mml:mi>w</mml:mi></mml:mrow></mml:mfrac></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where &#x0201C;ref&#x0201D; refers to the reference image, and &#x0201C;w&#x0201D; represents a warped image. Obviously, when the two images overlap exactly, the Dice coefficient is 1. TRE represents the distance sum of the corresponding landmark between the target image and the aligned image and is expressed as follows:</p>
<disp-formula id="E28"><label>(20)</label><mml:math id="M32"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mi>T</mml:mi><mml:mi>R</mml:mi><mml:mi>E</mml:mi><mml:mo>=</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:mfrac><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>n</mml:mi></mml:mrow></mml:mfrac><mml:mstyle displaystyle="true"><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>n</mml:mi></mml:mrow></mml:munderover></mml:mstyle><mml:mo>|</mml:mo><mml:msubsup><mml:mrow><mml:mi>d</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>-</mml:mo><mml:msubsup><mml:mrow><mml:mi>d</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>w</mml:mi></mml:mrow></mml:msubsup><mml:mo>|</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where <italic>n</italic> is the number of landmarks, <italic>r</italic> is the reference image, <italic>w</italic> is the aligned image, <italic>i</italic> is the <italic>i</italic>-th corresponding point, and <italic>d</italic> indicates the Euclidean distance.</p>
<fig id="F7" position="float">
<label>Figure 7</label>
<caption><p>Performance metrics statistics of existing registration methods.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fninf-16-933230-g0007.tif"/>
</fig>
</sec>
<sec id="s5">
<title>Future perspectives</title>
<sec>
<title>Exploring in-between representatives of two modalities</title>
<p>Many existing modality-translation-based methods for cross-modality biomedical image registration rely on synthetic images to train the mono-modal registration network because it is difficult to develop cross-modality similarity measures. Although such a training scheme does not need to perform cross-modal similarity metrics to improve the image synthesis performance, it is still necessary to design various losses to constrain other feature changes. Additionally, is the synthetic modality useful for improving registration performance? As far as we know, this intensity information does not play a key role in improving image performance. Some shape features, such as edges and corners, are essential for image registration. An in-between representation was found (Lu et al., <xref ref-type="bibr" rid="B50">2021</xref>), i.e., COMIR, which maps the modalities to their established &#x0201C;common ground.&#x0201D; An in-between representative with characteristics relevant to the accurate alignment would be good. In the future, more workers are expected to be carried out in this direction to find the in-between representatives.</p>
</sec>
<sec>
<title>Exploring quality assessment guided modal translation network</title>
<p>The image quality generated by the mode translation network directly affects the accuracy of the registration algorithm. Therefore, an important research direction is how to effectively and reasonably evaluate the quality of images generated by the GAN network. Additionally, an effective generated image quality evaluation method can be used to constrain the mode translation network&#x00027;s training process and improve the modal translation&#x00027;s effectiveness. There have recently been quality evaluation methods for images generated by GAN (Gu et al., <xref ref-type="bibr" rid="B24">2020</xref>), but there is still a lack of quality evaluation methods for synthetic biomedical images.</p>
</sec>
<sec>
<title>Designing large-scale biomedical image generation GAN network</title>
<p>The size of images existing image generation networks can generate is minimal (Brock et al., <xref ref-type="bibr" rid="B11">2019</xref>), but biomedical images are generally of high resolution, especially biological images used in neurological research. The training process of the existing GAN network is difficult to converge, especially with the increase in image size. The dimension of data space will dramatically increase. This challenge is difficult with current hardware levels and GAN-based image-synthesized methods. Therefore, designing an image synthesis network capable of synthesizing large-scale biomedical images is also a future direction.</p>
</sec>
<sec>
<title>Designing prior knowledge-guided registration methods</title>
<p>Traditional image registration models often use some standard landmarks like points and lines as guidance to optimize the model. Several recent studies have shown that a segmentation mask can be utilized in the discriminator (Luo et al., <xref ref-type="bibr" rid="B53">2021</xref>) or generator (Mahapatra et al., <xref ref-type="bibr" rid="B56">2018</xref>) for guiding the edge alignment. However, their works simply use only a segmentation mask as the edge space correspondence guidance. More space correspondence features are expected to be explored and verified.</p>
</sec>
</sec>
<sec sec-type="conclusions" id="s6">
<title>Conclusions</title>
<p>This paper provides a comprehensive survey of cross-modal and mono-modal biomedical image registration approaches based on GAN. The commonly used GAN structures are summarized, followed by the analyses of the biomedical image registration studies of the modality-independent based strategy, the symmetric learning-based strategy, the adversarial learning-based strategy, and the joint learning-based strategy from different implementation methods and perspectives. In addition, we have conducted a statistical analysis of the existing literature in various aspects and have drawn the corresponding conclusions. Finally, we outline four interesting research directions for future studies.</p></sec>
<sec id="s7">
<title>Author contributions</title>
<p>TH, JW, and LQ contributed to the conception and design of this paper. TH completed the literature selection and the writing of the first draft. JW and ZJ improved the writing of the paper. LQ provided constructive comments and determined the final draft of the paper. All authors contributed to the drafting of the manuscript. All authors contributed to the article and approved the submitted version.</p></sec>
<sec sec-type="funding-information" id="s8">
<title>Funding</title>
<p>This research was funded by the National Natural Science Foundation of China (61871411, 62271003, and 62201008), the Sci-Tech Innovation 2030 Agenda (2022ZD0205200 and 2022ZD0205204), the University Synergy Innovation Program of Anhui Province (GXXT-2021-001), and the Natural Science Foundation of the Education Department of Anhui Province (KJ2021A0017).</p>
</sec>
<sec sec-type="COI-statement" id="conf1">
<title>Conflict of interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p></sec>
<sec sec-type="disclaimer" id="s9">
<title>Publisher&#x00027;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p></sec>
</body>
<back>
<ack><p>The authors acknowledge the high-performance computing platform of Anhui University for providing computing resources.</p>
</ack>
<sec sec-type="supplementary-material" id="s10">
<title>Supplementary material</title>
<p>The Supplementary Material for this article can be found online at: <ext-link ext-link-type="uri" xlink:href="https://www.frontiersin.org/articles/10.3389/fninf.2022.933230/full#supplementary-material">https://www.frontiersin.org/articles/10.3389/fninf.2022.933230/full#supplementary-material</ext-link></p>
<supplementary-material xlink:href="Data_Sheet_1.pdf" id="SM1" mimetype="application/pdf" xmlns:xlink="http://www.w3.org/1999/xlink"/>
</sec>
<ref-list>
<title>References</title>
<ref id="B1">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Antonelli</surname> <given-names>M.</given-names></name> <name><surname>Reinke</surname> <given-names>A.</given-names></name> <name><surname>Bakas</surname> <given-names>S.</given-names></name> <name><surname>Farahani</surname> <given-names>K.</given-names></name> <name><surname>Kopp-Schneider</surname> <given-names>A.</given-names></name> <name><surname>Landman</surname> <given-names>B. A.</given-names></name> <etal/></person-group>. (<year>2022</year>). <article-title>The medical segmentation decathlon</article-title>. <source>Nat. Commun.</source> <volume>13</volume>, <fpage>4128</fpage>. <pub-id pub-id-type="doi">10.1038/s41467-022-30695-9</pub-id><pub-id pub-id-type="pmid">35840566</pub-id></citation></ref>
<ref id="B2">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Arar</surname> <given-names>M.</given-names></name> <name><surname>Ginger</surname> <given-names>Y.</given-names></name> <name><surname>Danon</surname> <given-names>D.</given-names></name> <name><surname>Bermano</surname> <given-names>A. H.</given-names></name> <name><surname>Cohen-Or</surname> <given-names>D.</given-names></name></person-group> (<year>2020</year>). <article-title>&#x0201C;Unsupervised multi-modal image registration via geometry preserving image-to-image translation,&#x0201D;</article-title> in <source>IEEE Conference on Computer Vision and Pattern Recognition (CVPR)</source> (<publisher-loc>Seattle, WA</publisher-loc>), <fpage>13410</fpage>&#x02013;<lpage>13419</lpage>. <pub-id pub-id-type="doi">10.1109/CVPR42600.2020.01342</pub-id></citation>
</ref>
<ref id="B3">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Avants</surname> <given-names>B. B.</given-names></name> <name><surname>Epstein</surname> <given-names>C. L.</given-names></name> <name><surname>Grossman</surname> <given-names>M.</given-names></name> <name><surname>Gee</surname> <given-names>J. C.</given-names></name></person-group> (<year>2008</year>). <article-title>Symmetric diffeomorphic image registration with cross-correlation: evaluating automated labeling of elderly and neurodegenerative brain</article-title>. <source>Med. Image Anal.</source> <volume>12</volume>, <fpage>26</fpage>&#x02013;<lpage>41</lpage>. <pub-id pub-id-type="doi">10.1016/j.media.2007.06.004</pub-id><pub-id pub-id-type="pmid">17659998</pub-id></citation></ref>
<ref id="B4">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Avants</surname> <given-names>B. B.</given-names></name> <name><surname>Tustison</surname> <given-names>N.</given-names></name> <name><surname>Song</surname> <given-names>G.</given-names></name></person-group> (<year>2009</year>). <article-title>Advanced normalization tools (ANTS)</article-title>. <source>Insight J.</source> <volume>2</volume>, <fpage>1</fpage>&#x02013;<lpage>35</lpage>. <pub-id pub-id-type="doi">10.54294/uvnhin</pub-id></citation>
</ref>
<ref id="B5">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bakas</surname> <given-names>S.</given-names></name> <name><surname>Reyes</surname> <given-names>M.</given-names></name> <name><surname>Jakab</surname> <given-names>A.</given-names></name> <name><surname>Bauer</surname> <given-names>S.</given-names></name> <name><surname>Rempfler</surname> <given-names>M.</given-names></name> <name><surname>Crimi</surname> <given-names>A.</given-names></name> <etal/></person-group>. (<year>2018</year>). <source>Identifying the Best Machine Learning Algorithms for Brain Tumor Segmentation, Progression Assessment, and Overall Survival Prediction in the BRATS Challenge.</source></citation>
</ref>
<ref id="B6">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Balakrishnan</surname> <given-names>G.</given-names></name> <name><surname>Zhao</surname> <given-names>A.</given-names></name> <name><surname>Sabuncu</surname> <given-names>M. R.</given-names></name> <name><surname>Guttag</surname> <given-names>J.</given-names></name> <name><surname>Dalca</surname> <given-names>A. V.</given-names></name></person-group> (<year>2019</year>). <article-title>VoxelMorph: a learning framework for deformable medical image registration</article-title>. <source>IEEE Trans. Med. Imaging</source> <volume>38</volume>, <fpage>1788</fpage>&#x02013;<lpage>1800</lpage>. <pub-id pub-id-type="doi">10.1109/TMI.2019.2897538</pub-id><pub-id pub-id-type="pmid">30716034</pub-id></citation></ref>
<ref id="B7">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bellec</surname> <given-names>P.</given-names></name> <name><surname>Chu</surname> <given-names>C.</given-names></name> <name><surname>Chouinard-Decorte</surname> <given-names>F.</given-names></name> <name><surname>Benhajali</surname> <given-names>Y.</given-names></name> <name><surname>Margulies</surname> <given-names>D. S.</given-names></name> <name><surname>Craddock</surname> <given-names>R. C.</given-names></name></person-group> (<year>2017</year>). <article-title>The neuro bureau ADHD-200 preprocessed repository</article-title>. <source>Neuroimage</source> <volume>144</volume>, <fpage>275</fpage>&#x02013;<lpage>286</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuroimage.2016.06.034</pub-id><pub-id pub-id-type="pmid">27423255</pub-id></citation></ref>
<ref id="B8">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bernard</surname> <given-names>O.</given-names></name> <name><surname>Lalande</surname> <given-names>A.</given-names></name> <name><surname>Zotti</surname> <given-names>C.</given-names></name> <name><surname>Cervenansky</surname> <given-names>F.</given-names></name> <name><surname>Yang</surname> <given-names>X.</given-names></name> <name><surname>Heng</surname> <given-names>P.-A.</given-names></name> <etal/></person-group>. (<year>2018</year>). <article-title>Deep learning techniques for automatic MRI cardiac multi-structures segmentation and diagnosis: is the problem solved?</article-title> <source>IEEE Trans. Med. Imaging</source> <volume>37</volume>, <fpage>2514</fpage>&#x02013;<lpage>2525</lpage>. <pub-id pub-id-type="doi">10.1109/TMI.2018.2837502</pub-id><pub-id pub-id-type="pmid">29994302</pub-id></citation></ref>
<ref id="B9">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bessadok</surname> <given-names>A.</given-names></name> <name><surname>Mahjoub</surname> <given-names>M. A.</given-names></name> <name><surname>Rekik</surname> <given-names>I.</given-names></name></person-group> (<year>2021</year>). <article-title>Brain graph synthesis by dual adversarial domain alignment and target graph prediction from a source graph</article-title>. <source>Med. Image Anal.</source> <volume>68</volume>, <fpage>101902</fpage>&#x02013;<lpage>101902</lpage>. <pub-id pub-id-type="doi">10.1016/j.media.2020.101902</pub-id><pub-id pub-id-type="pmid">33338871</pub-id></citation></ref>
<ref id="B10">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bilic</surname> <given-names>P.</given-names></name> <name><surname>Christ</surname> <given-names>P. F.</given-names></name> <name><surname>Vorontsov</surname> <given-names>E.</given-names></name> <name><surname>Chlebus</surname> <given-names>G.</given-names></name> <name><surname>Chen</surname> <given-names>H.</given-names></name> <name><surname>Dou</surname> <given-names>Q.</given-names></name> <etal/></person-group>. (<year>2019</year>). <source>The Liver Tumor Segmentation Benchmark (Lits).</source></citation>
</ref>
<ref id="B11">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Brock</surname> <given-names>A.</given-names></name> <name><surname>Donahue</surname> <given-names>J.</given-names></name> <name><surname>Simonyan</surname> <given-names>K.</given-names></name></person-group> (<year>2019</year>). <article-title>&#x0201C;Large scale GAN training for high fidelity natural image synthesis,&#x0201D;</article-title> in <source>International Conference on Learning Representations(ICLR)</source> (<publisher-loc>New Orleans, LA</publisher-loc>).</citation>
</ref>
<ref id="B12">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Candemir</surname> <given-names>S.</given-names></name> <name><surname>Jaeger</surname> <given-names>S.</given-names></name> <name><surname>Palaniappan</surname> <given-names>K.</given-names></name> <name><surname>Musco</surname> <given-names>J. P.</given-names></name> <name><surname>Singh</surname> <given-names>R. K.</given-names></name> <name><surname>Xue</surname> <given-names>Z.</given-names></name> <etal/></person-group>. (<year>2013</year>). <article-title>Lung segmentation in chest radiographs using anatomical atlases with nonrigid registration</article-title>. <source>IEEE Trans. Med. Imaging</source> <volume>33</volume>, <fpage>577</fpage>&#x02013;<lpage>590</lpage>. <pub-id pub-id-type="doi">10.1109/TMI.2013.2290491</pub-id><pub-id pub-id-type="pmid">24239990</pub-id></citation></ref>
<ref id="B13">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chen</surname> <given-names>W.</given-names></name> <name><surname>Liu</surname> <given-names>M.</given-names></name> <name><surname>Du</surname> <given-names>H.</given-names></name> <name><surname>Radojevi&#x00107;</surname> <given-names>M.</given-names></name> <name><surname>Wang</surname> <given-names>Y.</given-names></name> <name><surname>Meijering</surname> <given-names>E.</given-names></name></person-group> (<year>2021</year>). <article-title>Deep-Learning-based automated neuron reconstruction from 3D microscopy images using synthetic training images</article-title>. <source>IEEE Trans. Med. Imaging</source> <volume>41</volume>, <fpage>1031</fpage>&#x02013;<lpage>1042</lpage>. <pub-id pub-id-type="doi">10.1109/TMI.2021.3130934</pub-id><pub-id pub-id-type="pmid">34847022</pub-id></citation></ref>
<ref id="B14">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chen</surname> <given-names>X.</given-names></name> <name><surname>Duan</surname> <given-names>Y.</given-names></name> <name><surname>Houthooft</surname> <given-names>R.</given-names></name> <name><surname>Schulman</surname> <given-names>J.</given-names></name> <name><surname>Sutskever</surname> <given-names>I.</given-names></name> <name><surname>Abbeel</surname> <given-names>P.</given-names></name></person-group> (<year>2016</year>). <article-title>Infogan: interpretable representation learning by information maximizing generative adversarial nets</article-title>. <source>Adv. Neural Inf. Process. Syst.</source> <volume>29</volume>, <fpage>2180</fpage>&#x02013;<lpage>2188</lpage>.</citation>
</ref>
<ref id="B15">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Di Martino</surname> <given-names>A.</given-names></name> <name><surname>Yan</surname> <given-names>C.-G.</given-names></name> <name><surname>Li</surname> <given-names>Q.</given-names></name> <name><surname>Denio</surname> <given-names>E.</given-names></name> <name><surname>Castellanos</surname> <given-names>F. X.</given-names></name> <name><surname>Alaerts</surname> <given-names>K.</given-names></name> <etal/></person-group>. (<year>2014</year>). <article-title>The autism brain imaging data exchange: towards a large-scale evaluation of the intrinsic brain architecture in autism</article-title>. <source>Mol. Psychiatry</source> <volume>19</volume>, <fpage>659</fpage>&#x02013;<lpage>667</lpage>. <pub-id pub-id-type="doi">10.1038/mp.2013.78</pub-id><pub-id pub-id-type="pmid">23774715</pub-id></citation></ref>
<ref id="B16">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Elmahdy</surname> <given-names>M. S.</given-names></name> <name><surname>Wolterink</surname> <given-names>J. M.</given-names></name> <name><surname>Sokooti</surname> <given-names>H.</given-names></name> <name><surname>I&#x00161;gum</surname> <given-names>I.</given-names></name> <name><surname>Staring</surname> <given-names>M.</given-names></name></person-group> (<year>2019</year>). <article-title>&#x0201C;Adversarial optimization for joint registration and segmentation in prostate CT radiotherapy,&#x0201D;</article-title> in <source>Medical Image Computing and Computer-Assisted Intervention (MICCAI)</source> (<publisher-loc>Shanghai</publisher-loc>), <fpage>366</fpage>&#x02013;<lpage>374</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-030-32226-7_41</pub-id></citation>
</ref>
<ref id="B17">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Fan</surname> <given-names>J.</given-names></name> <name><surname>Cao</surname> <given-names>X.</given-names></name> <name><surname>Wang</surname> <given-names>Q.</given-names></name> <name><surname>Yap</surname> <given-names>P. T.</given-names></name> <name><surname>Shen</surname> <given-names>D.</given-names></name></person-group> (<year>2019</year>). <article-title>Adversarial learning for mono- or multi-modal registration</article-title>. <source>Med. Image Anal.</source> <volume>58</volume>, <fpage>101545</fpage>. <pub-id pub-id-type="doi">10.1016/j.media.2019.101545</pub-id><pub-id pub-id-type="pmid">31557633</pub-id></citation></ref>
<ref id="B18">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Fan</surname> <given-names>J.</given-names></name> <name><surname>Cao</surname> <given-names>X.</given-names></name> <name><surname>Xue</surname> <given-names>Z.</given-names></name> <name><surname>Yap</surname> <given-names>P. T.</given-names></name> <name><surname>Shen</surname> <given-names>D.</given-names></name></person-group> (<year>2018</year>). <article-title>Adversarial similarity network for evaluating image alignment in deep learning based registration</article-title>. <source>Med. Image Comput. Comput. Assist. Interv.</source> <volume>11070</volume>, <fpage>739</fpage>&#x02013;<lpage>746</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-030-00928-1_83</pub-id><pub-id pub-id-type="pmid">30627709</pub-id></citation></ref>
<ref id="B19">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Fu</surname> <given-names>Y.</given-names></name> <name><surname>Lei</surname> <given-names>Y.</given-names></name> <name><surname>Zhou</surname> <given-names>J.</given-names></name> <name><surname>Wang</surname> <given-names>T.</given-names></name> <name><surname>David</surname> <given-names>S. Y.</given-names></name> <name><surname>Beitler</surname> <given-names>J. J.</given-names></name> <etal/></person-group>. (<year>2020</year>). <article-title>Synthetic CT-aided MRI-CT image registration for head and neck radiotherapy</article-title>. <source>Int. Soc. Opt. Photon.</source> 11317. <pub-id pub-id-type="doi">10.1117/12.2549092</pub-id></citation>
</ref>
<ref id="B20">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gering</surname> <given-names>D. T.</given-names></name> <name><surname>Nabavi</surname> <given-names>A.</given-names></name> <name><surname>Kikinis</surname> <given-names>R.</given-names></name> <name><surname>Hata</surname> <given-names>N.</given-names></name> <name><surname>O&#x00027;Donnell</surname> <given-names>L. J.</given-names></name> <name><surname>Grimson</surname> <given-names>W. E. L.</given-names></name> <etal/></person-group>. (<year>2001</year>). <article-title>An integrated visualization system for surgical planning and guidance using image fusion and an open MR</article-title>. <source>J. Mag. Reson. Imaging</source> <volume>13</volume>, <fpage>967</fpage>&#x02013;<lpage>975</lpage>. <pub-id pub-id-type="doi">10.1002/jmri.1139</pub-id><pub-id pub-id-type="pmid">11382961</pub-id></citation></ref>
<ref id="B21">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gong</surname> <given-names>H.</given-names></name> <name><surname>Xu</surname> <given-names>D.</given-names></name> <name><surname>Yuan</surname> <given-names>J.</given-names></name> <name><surname>Li</surname> <given-names>X.</given-names></name> <name><surname>Guo</surname> <given-names>C.</given-names></name> <name><surname>Peng</surname> <given-names>J.</given-names></name> <etal/></person-group>. (<year>2016</year>). <article-title>High-throughput dual-colour precision imaging for brain-wide connectome with cytoarchitectonic landmarks at the cellular level</article-title>. <source>Nat. Commun.</source> <volume>7</volume>, <fpage>12142</fpage>. <pub-id pub-id-type="doi">10.1038/ncomms12142</pub-id><pub-id pub-id-type="pmid">27374071</pub-id></citation></ref>
<ref id="B22">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Goodfellow</surname> <given-names>I.</given-names></name> <name><surname>Pouget-Abadie</surname> <given-names>J.</given-names></name> <name><surname>Mirza</surname> <given-names>M.</given-names></name> <name><surname>Xu</surname> <given-names>B.</given-names></name> <name><surname>Warde-Farley</surname> <given-names>D.</given-names></name> <name><surname>Ozair</surname> <given-names>S.</given-names></name> <etal/></person-group>. (<year>2014</year>). <article-title>Generative adversarial nets</article-title>. <source>Adv. Neural Inf. Process. Syst.</source> <volume>27</volume>, <fpage>139</fpage>&#x02013;<lpage>144</lpage>. <pub-id pub-id-type="doi">10.1145/3422622</pub-id></citation>
</ref>
<ref id="B23">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gousias</surname> <given-names>I. S.</given-names></name> <name><surname>Edwards</surname> <given-names>A. D.</given-names></name> <name><surname>Rutherford</surname> <given-names>M. A.</given-names></name> <name><surname>Counsell</surname> <given-names>S. J.</given-names></name> <name><surname>Hajnal</surname> <given-names>J. V.</given-names></name> <name><surname>Rueckert</surname> <given-names>D.</given-names></name> <etal/></person-group>. (<year>2012</year>). <article-title>Magnetic resonance imaging of the newborn brain: manual segmentation of labelled atlases in term-born and preterm infants</article-title>. <source>Neuroimage</source> <volume>62</volume>, <fpage>1499</fpage>&#x02013;<lpage>1509</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuroimage.2012.05.083</pub-id><pub-id pub-id-type="pmid">22713673</pub-id></citation></ref>
<ref id="B24">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gu</surname> <given-names>S.</given-names></name> <name><surname>Bao</surname> <given-names>J.</given-names></name> <name><surname>Chen</surname> <given-names>D.</given-names></name> <name><surname>Wen</surname> <given-names>F.</given-names></name></person-group> (<year>2020</year>). <article-title>&#x0201C;Giqa: generated image quality assessment,&#x0201D;</article-title> in inj <italic>European Conference on Computer Vision, ECCV</italic> (Glasgow), <fpage>369</fpage>&#x02013;<lpage>385</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-030-58621-8_22</pub-id><pub-id pub-id-type="pmid">12758132</pub-id></citation></ref>
<ref id="B25">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Han</surname> <given-names>R.</given-names></name> <name><surname>Jones</surname> <given-names>C. K.</given-names></name> <name><surname>Ketcha</surname> <given-names>M. D.</given-names></name> <name><surname>Wu</surname> <given-names>P.</given-names></name> <name><surname>Vagdargi</surname> <given-names>P.</given-names></name> <name><surname>Uneri</surname> <given-names>A.</given-names></name> <etal/></person-group>. (<year>2021</year>). <article-title>Deformable MR-CT image registration using an unsupervised end-to-end synthesis and registration network for endoscopic neurosurgery</article-title>. <source>Int. Soc. Opt. Photon.</source> 11598. <pub-id pub-id-type="doi">10.1117/12.2581567</pub-id></citation>
</ref>
<ref id="B26">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>He</surname> <given-names>Y.</given-names></name> <name><surname>Li</surname> <given-names>T.</given-names></name> <name><surname>Ge</surname> <given-names>R.</given-names></name> <name><surname>Yang</surname> <given-names>J.</given-names></name> <name><surname>Kong</surname> <given-names>Y.</given-names></name> <name><surname>Zhu</surname> <given-names>J.</given-names></name> <etal/></person-group>. (<year>2021</year>). <article-title>Few-shot learning for deformable medical image registration with perception-correspondence decoupling and reverse teaching</article-title>. <source>IEEE J. Biomed. Health Inform.</source> <volume>26</volume>, <fpage>1177</fpage>&#x02013;<lpage>1187</lpage>. <pub-id pub-id-type="doi">10.1109/JBHI.2021.3095409</pub-id><pub-id pub-id-type="pmid">34232899</pub-id></citation></ref>
<ref id="B27">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Heimann</surname> <given-names>T.</given-names></name> <name><surname>Van Ginneken</surname> <given-names>B.</given-names></name> <name><surname>Styner</surname> <given-names>M. A.</given-names></name> <name><surname>Arzhaeva</surname> <given-names>Y.</given-names></name> <name><surname>Aurich</surname> <given-names>V.</given-names></name> <name><surname>Bauer</surname> <given-names>C.</given-names></name> <etal/></person-group>. (<year>2009</year>). <article-title>Comparison and evaluation of methods for liver segmentation from CT datasets</article-title>. <source>IEEE Trans. Med. Imaging</source> <volume>28</volume>, <fpage>1251</fpage>&#x02013;<lpage>1265</lpage>. <pub-id pub-id-type="doi">10.1109/TMI.2009.2013851</pub-id><pub-id pub-id-type="pmid">19211338</pub-id></citation></ref>
<ref id="B28">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Heinrich</surname> <given-names>M. P.</given-names></name> <name><surname>Jenkinson</surname> <given-names>M.</given-names></name> <name><surname>Bhushan</surname> <given-names>M.</given-names></name> <name><surname>Matin</surname> <given-names>T.</given-names></name> <name><surname>Gleeson</surname> <given-names>F. V.</given-names></name> <name><surname>Brady</surname> <given-names>M.</given-names></name> <etal/></person-group>. (<year>2012</year>). <article-title>MIND: Modality independent neighbourhood descriptor for multi-modal deformable registration</article-title>. <source>Med. Image Anal.</source> <volume>16</volume>, <fpage>1423</fpage>&#x02013;<lpage>1435</lpage>. <pub-id pub-id-type="doi">10.1016/j.media.2012.05.008</pub-id><pub-id pub-id-type="pmid">22722056</pub-id></citation></ref>
<ref id="B29">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Hu</surname> <given-names>Y.</given-names></name> <name><surname>Gibson</surname> <given-names>E.</given-names></name> <name><surname>Ghavami</surname> <given-names>N.</given-names></name> <name><surname>Bonmati</surname> <given-names>E.</given-names></name> <name><surname>Moore</surname> <given-names>C. M.</given-names></name> <name><surname>Emberton</surname> <given-names>M.</given-names></name> <etal/></person-group>. (<year>2018</year>). <article-title>&#x0201C;Adversarial deformation regularization for training image registration neural networks,&#x0201D;</article-title> in <source>International Conference on Medical Image Computing and Computer-Assisted Intervention</source> (<publisher-loc>Granada</publisher-loc>), <fpage>774</fpage>&#x02013;<lpage>782</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-030-00928-1_87</pub-id></citation>
</ref>
<ref id="B30">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ioffe</surname> <given-names>S.</given-names></name> <name><surname>Szegedy</surname> <given-names>C.</given-names></name></person-group> (<year>2015</year>). <article-title>Batch normalization: accelerating deep network training by reducing internal covariate shift</article-title>. <source>PMLR</source>. <volume>27</volume>, <fpage>448</fpage>&#x02013;<lpage>456</lpage>.</citation>
</ref>
<ref id="B31">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Isola</surname> <given-names>P.</given-names></name> <name><surname>Zhu</surname> <given-names>J.-Y.</given-names></name> <name><surname>Zhou</surname> <given-names>T.</given-names></name> <name><surname>Efros</surname> <given-names>A. A.</given-names></name></person-group> (<year>2017</year>). <article-title>&#x0201C;Image-to-Image translation with conditional adversarial networks,&#x0201D;</article-title> in <source>IEEE Conference on Computer Vision and Pattern Recognition (CVPR)</source> (<publisher-loc>Hawaii, HI</publisher-loc>), <fpage>1125</fpage>&#x02013;<lpage>1134</lpage>. <pub-id pub-id-type="doi">10.1109/CVPR.2017.632</pub-id><pub-id pub-id-type="pmid">34940729</pub-id></citation></ref>
<ref id="B32">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jaderberg</surname> <given-names>M.</given-names></name> <name><surname>Simonyan</surname> <given-names>K.</given-names></name> <name><surname>Zisserman</surname> <given-names>A.</given-names></name></person-group> (<year>2015</year>). <article-title>Spatial transformer networks</article-title>. <source>Adv. Neural Inform. Proc. Syst.</source> 28, 607.</citation>
</ref>
<ref id="B33">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jaeger</surname> <given-names>S.</given-names></name> <name><surname>Karargyris</surname> <given-names>A.</given-names></name> <name><surname>Candemir</surname> <given-names>S.</given-names></name> <name><surname>Folio</surname> <given-names>L.</given-names></name> <name><surname>Siegelman</surname> <given-names>J.</given-names></name> <name><surname>Callaghan</surname> <given-names>F.</given-names></name> <etal/></person-group>. (<year>2013</year>). <article-title>Automatic tuberculosis screening using chest radiographs</article-title>. <source>IEEE Trans. Med. Imaging</source> <volume>33</volume>, <fpage>233</fpage>&#x02013;<lpage>245</lpage>. <pub-id pub-id-type="doi">10.1109/TMI.2013.2284099</pub-id><pub-id pub-id-type="pmid">29959539</pub-id></citation></ref>
<ref id="B34">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Jiang</surname> <given-names>S.</given-names></name> <name><surname>Wang</surname> <given-names>C.</given-names></name> <name><surname>Huang</surname> <given-names>C.</given-names></name></person-group> (<year>2021</year>). <article-title>&#x0201C;Image registration improved by generative adversarial networks,&#x0201D;</article-title> in <source>International Conference on Multimedia Modeling</source> (<publisher-loc>Prague</publisher-loc>), <fpage>26</fpage>&#x02013;<lpage>35</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-030-67835-7_3</pub-id></citation>
</ref>
<ref id="B35">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jing</surname> <given-names>Y.</given-names></name> <name><surname>Yang</surname> <given-names>Y.</given-names></name> <name><surname>Feng</surname> <given-names>Z.</given-names></name> <name><surname>Ye</surname> <given-names>J.</given-names></name> <name><surname>Yu</surname> <given-names>Y.</given-names></name> <name><surname>Song</surname> <given-names>M.</given-names></name></person-group> (<year>2020</year>). <article-title>Neural style transfer: a review</article-title>. <source>IEEE Trans. Vis. Comput. Graph.</source> <volume>26</volume>, <fpage>3365</fpage>&#x02013;<lpage>3385</lpage>. <pub-id pub-id-type="doi">10.1109/TVCG.2019.2921336</pub-id><pub-id pub-id-type="pmid">31180860</pub-id></citation></ref>
<ref id="B36">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kavur</surname> <given-names>A. E.</given-names></name> <name><surname>Gezer</surname> <given-names>N. S.</given-names></name> <name><surname>Bari&#x0015F;</surname> <given-names>M.</given-names></name> <name><surname>Aslan</surname> <given-names>S.</given-names></name> <name><surname>Conze</surname> <given-names>P.-H.</given-names></name> <name><surname>Groza</surname> <given-names>V.</given-names></name> <etal/></person-group>. (<year>2021</year>). <article-title>CHAOS challenge-combined (CT-MR) healthy abdominal organ segmentation</article-title>. <source>Med. Image Anal.</source> <volume>69</volume>, <fpage>101950</fpage>. <pub-id pub-id-type="doi">10.1016/j.media.2020.101950</pub-id><pub-id pub-id-type="pmid">33421920</pub-id></citation></ref>
<ref id="B37">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Kim</surname> <given-names>T.</given-names></name> <name><surname>Cha</surname> <given-names>M.</given-names></name> <name><surname>Kim</surname> <given-names>H.</given-names></name> <name><surname>Lee</surname> <given-names>J. K.</given-names></name> <name><surname>Kim</surname> <given-names>J.</given-names></name></person-group> (<year>2017</year>). <article-title>&#x0201C;Learning to discover cross-domain relations with generative adversarial networks,&#x0201D;</article-title> in <source>International Conference on Machine Learning</source> (<publisher-loc>Sydney</publisher-loc>) <fpage>1857</fpage>&#x02013;<lpage>1865</lpage>.</citation>
</ref>
<ref id="B38">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kingma</surname> <given-names>D. P.</given-names></name> <name><surname>Welling</surname> <given-names>M.</given-names></name></person-group> (<year>2013</year>). <article-title>Auto-encoding variational bayes</article-title>. <source>arXiv preprint</source>. arXiv:1312.6114.<pub-id pub-id-type="pmid">32176273</pub-id></citation></ref>
<ref id="B39">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Klein</surname> <given-names>A.</given-names></name> <name><surname>Andersson</surname> <given-names>J.</given-names></name> <name><surname>Ardekani</surname> <given-names>B. A.</given-names></name> <name><surname>Ashburner</surname> <given-names>J.</given-names></name> <name><surname>Avants</surname> <given-names>B.</given-names></name> <name><surname>Chiang</surname> <given-names>M.-C.</given-names></name> <etal/></person-group>. (<year>2009</year>). <article-title>Evaluation of 14 nonlinear deformation algorithms applied to human brain MRI registration</article-title>. <source>Neuroimage</source> <volume>46</volume>, <fpage>786</fpage>&#x02013;<lpage>802</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuroimage.2008.12.037</pub-id><pub-id pub-id-type="pmid">19195496</pub-id></citation></ref>
<ref id="B40">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Klein</surname> <given-names>S.</given-names></name> <name><surname>Staring</surname> <given-names>M.</given-names></name> <name><surname>Murphy</surname> <given-names>K.</given-names></name> <name><surname>Viergever</surname> <given-names>M. A.</given-names></name> <name><surname>Pluim</surname> <given-names>J. P. W.</given-names></name></person-group> (<year>2009</year>). <article-title>Elastix: a toolbox for intensity-based medical image registration</article-title>. <source>IEEE Trans. Med. Imaging</source> <volume>29</volume>, <fpage>196</fpage>&#x02013;<lpage>205</lpage>. <pub-id pub-id-type="doi">10.1109/TMI.2009.2035616</pub-id><pub-id pub-id-type="pmid">19923044</pub-id></citation></ref>
<ref id="B41">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>LaMontagne</surname> <given-names>P. J.</given-names></name> <name><surname>Benzinger</surname> <given-names>T. L. S.</given-names></name> <name><surname>Morris</surname> <given-names>J. C.</given-names></name> <name><surname>Keefe</surname> <given-names>S.</given-names></name> <name><surname>Hornbeck</surname> <given-names>R.</given-names></name> <name><surname>Xiong</surname> <given-names>C.</given-names></name> <etal/></person-group>. (<year>2019</year>). <article-title>OASIS-3: longitudinal neuroimaging, clinical, and cognitive dataset for normal aging and Alzheimer disease</article-title>. <source>Alzheimers Dementia</source>. <pub-id pub-id-type="doi">10.1101/2019.12.13.19014902</pub-id><pub-id pub-id-type="pmid">35187166</pub-id></citation></ref>
<ref id="B42">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>A. A.</given-names></name> <name><surname>Gong</surname> <given-names>H.</given-names></name></person-group> (<year>2012</year>). <article-title>Progress on whole brain imaging methods at the level of optical microscopy</article-title>. <source>Prog. Biochem. Biophys.</source> 39,498&#x02013;504. <pub-id pub-id-type="doi">10.3724/SP.J.1206.2012.00237</pub-id><pub-id pub-id-type="pmid">35777803</pub-id></citation></ref>
<ref id="B43">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>C.</given-names></name> <name><surname>Wand</surname> <given-names>M.</given-names></name></person-group> (<year>2016</year>). <article-title>&#x0201C;Precomputed real-time texture synthesis with markovian generative adversarial networks,&#x0201D;</article-title> in <source>European Conference on Computer Vision</source> (<publisher-loc>Amsterdam</publisher-loc>), <fpage>702</fpage>&#x02013;<lpage>716</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-319-46487-9_43</pub-id></citation>
</ref>
<ref id="B44">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>M.</given-names></name> <name><surname>Wang</surname> <given-names>Y.</given-names></name> <name><surname>Zhang</surname> <given-names>F.</given-names></name> <name><surname>Li</surname> <given-names>G.</given-names></name> <name><surname>Hu</surname> <given-names>S.</given-names></name> <name><surname>Wu</surname> <given-names>L.</given-names></name></person-group> (<year>2021</year>). <article-title>&#x0201C;Deformable medical image registration based on unsupervised generative adversarial network integrating dual attention mechanisms,&#x0201D;</article-title> in <source>International Congress on Image and Signal Processing, BioMedical Engineering and Informatics (CISP-BMEI)</source> (<publisher-loc>Shenyang</publisher-loc>) <fpage>1</fpage>&#x02013;<lpage>6</lpage>. <pub-id pub-id-type="doi">10.1109/CISP-BMEI53629.2021.9624229</pub-id><pub-id pub-id-type="pmid">27295638</pub-id></citation></ref>
<ref id="B45">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>X.</given-names></name> <name><surname>Jiang</surname> <given-names>Y.</given-names></name> <name><surname>Rodriguez-Andina</surname> <given-names>J. J.</given-names></name> <name><surname>Luo</surname> <given-names>H.</given-names></name> <name><surname>Yin</surname> <given-names>S.</given-names></name> <name><surname>Kaynak</surname> <given-names>O.</given-names></name></person-group> (<year>2021</year>). <article-title>When medical images meet generative adversarial network: recent development and research opportunities</article-title>. <source>Discover Art. Intell.</source> <volume>1</volume>, <fpage>1</fpage>&#x02013;<lpage>20</lpage>. <pub-id pub-id-type="doi">10.1007/s44163-021-00006-0</pub-id></citation>
</ref>
<ref id="B46">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>Z.</given-names></name> <name><surname>Ogino</surname> <given-names>M.</given-names></name></person-group> (<year>2019</year>). <article-title>&#x0201C;Adversarial learning for deformable image registration: application to 3d ultrasound image fusion,&#x0201D;</article-title> in <source>Smart Ultrasound Imaging and Perinatal, Preterm and Paediatric Image Analysis</source> (<publisher-loc>Cham</publisher-loc>: <publisher-name>Springer</publisher-name>), <fpage>56</fpage>&#x02013;<lpage>64</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-030-32875-7_7</pub-id></citation>
</ref>
<ref id="B47">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lin</surname> <given-names>W.</given-names></name> <name><surname>Lin</surname> <given-names>W.</given-names></name> <name><surname>Chen</surname> <given-names>G.</given-names></name> <name><surname>Zhang</surname> <given-names>H.</given-names></name> <name><surname>Gao</surname> <given-names>Q.</given-names></name> <name><surname>Huang</surname> <given-names>Y.</given-names></name> <etal/></person-group>. (<year>2021</year>). <article-title>Bidirectional mapping of brain MRI and PET with 3D reversible GAN for the diagnosis of Alzheimer&#x00027;s disease</article-title>. <source>Front. Neurosci.</source> <volume>15</volume>, <fpage>357</fpage>. <pub-id pub-id-type="doi">10.3389/fnins.2021.646013</pub-id><pub-id pub-id-type="pmid">33935634</pub-id></citation></ref>
<ref id="B48">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Liu</surname> <given-names>F.</given-names></name> <name><surname>Cai</surname> <given-names>J.</given-names></name> <name><surname>Huo</surname> <given-names>Y.</given-names></name> <name><surname>Cheng</surname> <given-names>C.-T.</given-names></name> <name><surname>Raju</surname> <given-names>A.</given-names></name> <name><surname>Jin</surname> <given-names>D.</given-names></name> <etal/></person-group>. (<year>2020</year>). <article-title>&#x0201C;Jssr: a joint synthesis, segmentation, and registration system for 3d multi-modal image alignment of large-scale pathological ct scans,&#x0201D;</article-title> in <source>European Conference on Computer Vision</source> (<publisher-loc>Glasgow</publisher-loc>) 257&#x02013;274. <pub-id pub-id-type="doi">10.1007/978-3-030-58601-0_16</pub-id></citation>
</ref>
<ref id="B49">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Liu</surname> <given-names>M.-Y.</given-names></name> <name><surname>Breuel</surname> <given-names>T.</given-names></name> <name><surname>Kautz</surname> <given-names>J.</given-names></name></person-group> (<year>2017</year>). <article-title>Unsupervised image-to-image translation networks</article-title>. <source>Adv. Neural Inf. Process. Syst.</source> <volume>30</volume>, <fpage>721</fpage>&#x02013;<lpage>730</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-319-70139-4</pub-id></citation>
</ref>
<ref id="B50">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lu</surname> <given-names>J.</given-names></name> <name><surname>&#x000D6;fverstedt</surname> <given-names>J.</given-names></name> <name><surname>Lindblad</surname> <given-names>J.</given-names></name> <name><surname>Sladoje</surname> <given-names>N.</given-names></name></person-group> (<year>2021</year>). <source>Is Image-to-Image Translation the Panacea for Multimodal Image Registration? A Comparative Study.</source></citation>
</ref>
<ref id="B51">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lu</surname> <given-names>Y.</given-names></name> <name><surname>Li</surname> <given-names>B.</given-names></name> <name><surname>Liu</surname> <given-names>N.</given-names></name> <name><surname>Chen</surname> <given-names>J. W.</given-names></name> <name><surname>Xiao</surname> <given-names>L.</given-names></name> <name><surname>Gou</surname> <given-names>S.</given-names></name> <etal/></person-group>. (<year>2020</year>). <article-title>CT-TEE image registration for surgical navigation of congenital heart disease based on a cycle adversarial network</article-title>. <source>Comput. Math. Methods Med.</source> <volume>2020</volume>, <fpage>4942121</fpage>. <pub-id pub-id-type="doi">10.1155/2020/4942121</pub-id><pub-id pub-id-type="pmid">32802148</pub-id></citation></ref>
<ref id="B52">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Lu</surname> <given-names>Z.</given-names></name> <name><surname>Yang</surname> <given-names>G.</given-names></name> <name><surname>Hua</surname> <given-names>T.</given-names></name> <name><surname>Hu</surname> <given-names>L.</given-names></name> <name><surname>Kong</surname> <given-names>Y.</given-names></name> <name><surname>Tang</surname> <given-names>L.</given-names></name> <etal/></person-group>. (<year>2019</year>). <article-title>&#x0201C;Unsupervised three-dimensional image registration using a cycle convolutional neural network,&#x0201D;</article-title> in <source>IEEE International Conference on Image Processing (ICIP)</source> (<publisher-loc>Taipei</publisher-loc>), <fpage>2174</fpage>&#x02013;<lpage>2178</lpage>. <pub-id pub-id-type="doi">10.1109/ICIP.2019.8803163</pub-id><pub-id pub-id-type="pmid">34726267</pub-id></citation></ref>
<ref id="B53">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Luo</surname> <given-names>Y.</given-names></name> <name><surname>Cao</surname> <given-names>W.</given-names></name> <name><surname>He</surname> <given-names>Z.</given-names></name> <name><surname>Zou</surname> <given-names>W.</given-names></name> <name><surname>He</surname> <given-names>Z.</given-names></name></person-group> (<year>2021</year>). <article-title>Deformable adversarial registration network with multiple loss constraints</article-title>. <source>Comput. Med. Imaging Graph.</source> <volume>91</volume>, <fpage>101931</fpage>. <pub-id pub-id-type="doi">10.1016/j.compmedimag.2021.101931</pub-id><pub-id pub-id-type="pmid">34090262</pub-id></citation></ref>
<ref id="B54">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Maas</surname> <given-names>A. L.</given-names></name> <name><surname>Hannun</surname> <given-names>A. Y.</given-names></name> <name><surname>Ng</surname> <given-names>A. Y.</given-names></name></person-group> (<year>2013</year>). <article-title>&#x0201C;Rectifier nonlinearities improve neural network acoustic models,&#x0201D;</article-title> in <source>Proceeding ICML</source> (<publisher-loc>Shenzhen</publisher-loc>).</citation>
</ref>
<ref id="B55">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Mahapatra</surname> <given-names>D.</given-names></name> <name><surname>Ge</surname> <given-names>Z.</given-names></name></person-group> (<year>2020</year>). <article-title>Training data independent image registration using generative adversarial networks and domain adaptation</article-title>. <source>Patt. Recognit.</source> <volume>100</volume>, <fpage>107109</fpage>. <pub-id pub-id-type="doi">10.1016/j.patcog.2019.107109</pub-id></citation>
</ref>
<ref id="B56">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Mahapatra</surname> <given-names>D.</given-names></name> <name><surname>Ge</surname> <given-names>Z.</given-names></name> <name><surname>Sedai</surname> <given-names>S.</given-names></name> <name><surname>Chakravorty</surname> <given-names>R.</given-names></name></person-group> (<year>2018</year>). <article-title>&#x0201C;Joint registration and segmentation of Xray images using generative adversarial networks,&#x0201D;</article-title> in <source>Machine Learning in Medical Imaging</source> (<publisher-loc>Granada</publisher-loc>), <fpage>73</fpage>&#x02013;<lpage>80</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-030-00919-9_9</pub-id></citation>
</ref>
<ref id="B57">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Menze</surname> <given-names>B. H.</given-names></name> <name><surname>Jakab</surname> <given-names>A.</given-names></name> <name><surname>Bauer</surname> <given-names>S.</given-names></name> <name><surname>Kalpathy-Cramer</surname> <given-names>J.</given-names></name> <name><surname>Farahani</surname> <given-names>K.</given-names></name> <name><surname>Kirby</surname> <given-names>J.</given-names></name> <etal/></person-group>. (<year>2014</year>). <article-title>The multimodal brain tumor image segmentation benchmark (BRATS)</article-title>. <source>IEEE Trans. Med. Imaging</source> <volume>34</volume>, <fpage>1993</fpage>&#x02013;<lpage>2024</lpage>. <pub-id pub-id-type="doi">10.1109/TMI.2014.2377694</pub-id><pub-id pub-id-type="pmid">25494501</pub-id></citation></ref>
<ref id="B58">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Mirza</surname> <given-names>M.</given-names></name> <name><surname>Osindero</surname> <given-names>S.</given-names></name></person-group> (<year>2014</year>). <source>Conditional Generative Adversarial Nets</source>.</citation>
</ref>
<ref id="B59">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Mueller</surname> <given-names>S. G.</given-names></name> <name><surname>Weiner</surname> <given-names>M. W.</given-names></name> <name><surname>Thal</surname> <given-names>L. J.</given-names></name> <name><surname>Petersen</surname> <given-names>R. C.</given-names></name> <name><surname>Jack</surname> <given-names>C. R.</given-names></name> <name><surname>Jagust</surname> <given-names>W.</given-names></name> <etal/></person-group>. (<year>2005</year>). <article-title>Ways toward an early diagnosis in Alzheimer&#x00027;s disease: the Alzheimer&#x00027;s disease neuroimaging initiative (ADNI)</article-title>. <source>Alzheimers Dementia</source> <volume>1</volume>, <fpage>55</fpage>&#x02013;<lpage>66</lpage>. <pub-id pub-id-type="doi">10.1016/j.jalz.2005.06.003</pub-id><pub-id pub-id-type="pmid">17476317</pub-id></citation></ref>
<ref id="B60">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Oliveira</surname> <given-names>F. P.</given-names></name> <name><surname>Tavares</surname> <given-names>J. M.</given-names></name></person-group> (<year>2014</year>). <article-title>Medical image registration: a review</article-title>. <source>Comput. Methods Biomech. Biomed. Engin.</source> <volume>17</volume>, <fpage>73</fpage>&#x02013;<lpage>93</lpage>. <pub-id pub-id-type="doi">10.1080/10255842.2012.670855</pub-id><pub-id pub-id-type="pmid">22435355</pub-id></citation></ref>
<ref id="B61">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Qin</surname> <given-names>C.</given-names></name> <name><surname>Shi</surname> <given-names>B.</given-names></name> <name><surname>Liao</surname> <given-names>R.</given-names></name> <name><surname>Mansi</surname> <given-names>T.</given-names></name> <name><surname>Rueckert</surname> <given-names>D.</given-names></name> <name><surname>Kamen</surname> <given-names>A.</given-names></name></person-group> (<year>2019</year>). <article-title>&#x0201C;Unsupervised deformable registration for multi-modal images via disentangled representations,&#x0201D;</article-title> in <source>International Conference on Information Processing in Medical Imaging</source> (<publisher-loc>Hong Kong</publisher-loc>), <fpage>249</fpage>&#x02013;<lpage>261</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-030-20351-1_19</pub-id></citation>
</ref>
<ref id="B62">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Qu</surname> <given-names>L.</given-names></name> <name><surname>Li</surname> <given-names>Y.</given-names></name> <name><surname>Xie</surname> <given-names>P.</given-names></name> <name><surname>Liu</surname> <given-names>L.</given-names></name> <name><surname>Wang</surname> <given-names>Y.</given-names></name> <etal/></person-group>. (<year>2021</year>). <source>Cross-Modality Coherent Registration of Whole Mouse Brains</source>. <pub-id pub-id-type="doi">10.1038/s41592-021-01334-w</pub-id><pub-id pub-id-type="pmid">34887551</pub-id></citation></ref>
<ref id="B63">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Radford</surname> <given-names>A.</given-names></name> <name><surname>Metz</surname> <given-names>L.</given-names></name> <name><surname>Chintala</surname> <given-names>S.</given-names></name></person-group> (<year>2015</year>). <article-title>&#x0201C;Unsupervised representation learning with deep convolutional generative adversarial networks,&#x0201D;</article-title> in <source>International Conference on Learning Representations(ICLR)</source> (<publisher-loc>San Diego, CA</publisher-loc>).<pub-id pub-id-type="pmid">33873122</pub-id></citation></ref>
<ref id="B64">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ragan</surname> <given-names>T.</given-names></name> <name><surname>Kadiri</surname> <given-names>L. R.</given-names></name> <name><surname>Venkataraju</surname> <given-names>K. U.</given-names></name> <name><surname>Bahlmann</surname> <given-names>K.</given-names></name> <name><surname>Sutin</surname> <given-names>J.</given-names></name> <name><surname>Taranda</surname> <given-names>J.</given-names></name> <etal/></person-group>. (<year>2012</year>). <article-title>Serial two-photon tomography for automated <italic>ex vivo</italic> mouse brain imaging</article-title>. <source>Nat. Methods</source> <volume>9</volume>, <fpage>255</fpage>&#x02013;<lpage>258</lpage>. <pub-id pub-id-type="doi">10.1038/nmeth.1854</pub-id><pub-id pub-id-type="pmid">22245809</pub-id></citation></ref>
<ref id="B65">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Ronneberger</surname> <given-names>O.</given-names></name> <name><surname>Fischer</surname> <given-names>P.</given-names></name> <name><surname>Brox</surname> <given-names>T.</given-names></name></person-group> (<year>2015</year>). <article-title>&#x0201C;U-Net: convolutional networks for biomedical image segmentation,&#x0201D;</article-title> in <source>Medical Image Computing and Computer-Assisted Intervention (MICCAI)</source> (<publisher-loc>Munich</publisher-loc>), <fpage>234</fpage>&#x02013;<lpage>241</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-319-24574-4_28</pub-id></citation>
</ref>
<ref id="B66">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ruska</surname> <given-names>E.</given-names></name></person-group> (<year>1987</year>). <article-title>The development of the electron microscope and of electron microscopy</article-title>. <source>Rev. Mod. Phys.</source> <volume>59</volume>, <fpage>627</fpage>. <pub-id pub-id-type="doi">10.1103/RevModPhys.59.627</pub-id><pub-id pub-id-type="pmid">25832233</pub-id></citation></ref>
<ref id="B67">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Shattuck</surname> <given-names>D. W.</given-names></name> <name><surname>Mirza</surname> <given-names>M.</given-names></name> <name><surname>Adisetiyo</surname> <given-names>V.</given-names></name> <name><surname>Hojatkashani</surname> <given-names>C.</given-names></name> <name><surname>Salamon</surname> <given-names>G.</given-names></name> <name><surname>Narr</surname> <given-names>K. L.</given-names></name> <etal/></person-group>. (<year>2008</year>). <article-title>Construction of a 3D probabilistic atlas of human cortical structures</article-title>. <source>Neuroimage</source> <volume>39</volume>, <fpage>1064</fpage>&#x02013;<lpage>1080</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuroimage.2007.09.031</pub-id><pub-id pub-id-type="pmid">18037310</pub-id></citation></ref>
<ref id="B68">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Shiraishi</surname> <given-names>J.</given-names></name> <name><surname>Katsuragawa</surname> <given-names>S.</given-names></name> <name><surname>Ikezoe</surname> <given-names>J.</given-names></name> <name><surname>Matsumoto</surname> <given-names>T.</given-names></name> <name><surname>Kobayashi</surname> <given-names>T.</given-names></name> <name><surname>Komatsu</surname> <given-names>K.-I.</given-names></name> <etal/></person-group>. (<year>2000</year>). <article-title>Development of a digital image database for chest radiographs with and without a lung nodule: receiver operating characteristic analysis of radiologists&#x00027; detection of pulmonary nodules</article-title>. <source>Am. J. Roentgenol.</source> <volume>174</volume>, <fpage>71</fpage>&#x02013;<lpage>74</lpage>. <pub-id pub-id-type="doi">10.2214/ajr.174.1.1740071</pub-id><pub-id pub-id-type="pmid">10628457</pub-id></citation></ref>
<ref id="B69">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Simonyan</surname> <given-names>K.</given-names></name> <name><surname>Zisserman</surname> <given-names>A.</given-names></name></person-group> (<year>2014</year>). <article-title>Very deep convolutional networks for large-scale image recognition</article-title>. <source>arXiv preprint</source> arXiv:1409.1556.</citation>
</ref>
<ref id="B70">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Staring</surname> <given-names>M.</given-names></name> <name><surname>van der Heide</surname> <given-names>U. A.</given-names></name> <name><surname>Klein</surname> <given-names>S.</given-names></name> <name><surname>Viergever</surname> <given-names>M. A.</given-names></name> <name><surname>Pluim</surname> <given-names>J. P.</given-names></name></person-group> (<year>2009</year>). <article-title>Registration of cervical MRI using multifeature mutual information</article-title>. <source>IEEE Trans. Med. Imaging</source> <volume>28</volume>, <fpage>1412</fpage>&#x02013;<lpage>1421</lpage>. <pub-id pub-id-type="doi">10.1109/TMI.2009.2016560</pub-id><pub-id pub-id-type="pmid">19278929</pub-id></citation></ref>
<ref id="B71">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Tanner</surname> <given-names>C.</given-names></name> <name><surname>Ozdemir</surname> <given-names>F.</given-names></name> <name><surname>Profanter</surname> <given-names>R.</given-names></name> <name><surname>Vishnevsky</surname> <given-names>V.</given-names></name> <name><surname>Konukoglu</surname> <given-names>E.</given-names></name> <name><surname>Goksel</surname> <given-names>O.</given-names></name></person-group> (<year>2018</year>). <source>Generative Adversarial Networks for MR-CT Deformable Image Registration.</source><pub-id pub-id-type="pmid">36305424</pub-id></citation></ref>
<ref id="B72">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Tran</surname> <given-names>M. Q.</given-names></name> <name><surname>Do</surname> <given-names>T.</given-names></name> <name><surname>Tran</surname> <given-names>H.</given-names></name> <name><surname>Tjiputra</surname> <given-names>E.</given-names></name> <name><surname>Tran</surname> <given-names>Q. D.</given-names></name> <name><surname>Nguyen</surname> <given-names>A.</given-names></name></person-group> (<year>2022</year>). <article-title>Light-weight deformable registration using adversarial learning with distilling knowledge</article-title>. <source>IEEE Trans. Med. Imaging</source> <volume>41</volume>, <fpage>1443</fpage>&#x02013;<lpage>1453</lpage>. <pub-id pub-id-type="doi">10.1109/TMI.2022.3141013</pub-id><pub-id pub-id-type="pmid">34990354</pub-id></citation></ref>
<ref id="B73">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>L. T.</given-names></name> <name><surname>Hoover</surname> <given-names>N. E.</given-names></name> <name><surname>Porter</surname> <given-names>E. H.</given-names></name> <name><surname>Zasio</surname> <given-names>J. J.</given-names></name></person-group> (<year>1987</year>). <article-title>&#x0201C;SSIM: a software levelized compiled-code simulator,&#x0201D;</article-title> in <source>Proceedings of the 24th ACM/IEEE Design Automation Conference</source> (<publisher-loc>Miami Beach, FL</publisher-loc>), <fpage>2</fpage>-<lpage>8</lpage>. <pub-id pub-id-type="doi">10.1145/37888.37889</pub-id></citation>
</ref>
<ref id="B74">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>X.</given-names></name> <name><surname>Peng</surname> <given-names>Y.</given-names></name> <name><surname>Lu</surname> <given-names>L.</given-names></name> <name><surname>Lu</surname> <given-names>Z.</given-names></name> <name><surname>Bagheri</surname> <given-names>M.</given-names></name> <name><surname>Summers</surname> <given-names>R. M.</given-names></name></person-group> (<year>2017</year>). <article-title>&#x0201C;Chestx-ray8: hospital-scale chest x-ray database and benchmarks on weakly-supervised classification and localization of common thorax diseases,&#x0201D;</article-title> in <source>Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition</source> (<publisher-loc>Miami Beach, FL</publisher-loc>) <fpage>2097</fpage>&#x02013;<lpage>2106</lpage>. <pub-id pub-id-type="doi">10.1109/CVPR.2017.369</pub-id></citation>
</ref>
<ref id="B75">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>Z.</given-names></name> <name><surname>Bovik</surname> <given-names>A. C.</given-names></name> <name><surname>Sheikh</surname> <given-names>H. R.</given-names></name> <name><surname>Simoncelli</surname> <given-names>E. P.</given-names></name></person-group> (<year>2004</year>). <article-title>&#x0201C;Image quality assessment: from error visibility to structural similarity,&#x0201D;</article-title> in <source>IEEE transactions on image processing</source>. <volume>13</volume>, <fpage>600</fpage>&#x02013;<lpage>612</lpage>.<pub-id pub-id-type="pmid">15376593</pub-id></citation></ref>
<ref id="B76">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wei</surname> <given-names>D.</given-names></name> <name><surname>Ahmad</surname> <given-names>S.</given-names></name> <name><surname>Huo</surname> <given-names>J.</given-names></name> <name><surname>Huang</surname> <given-names>P.</given-names></name> <name><surname>Yap</surname> <given-names>P. T.</given-names></name> <name><surname>Xue</surname> <given-names>Z.</given-names></name> <etal/></person-group>. (<year>2020</year>). <article-title>SLIR: synthesis, localization, inpainting, and registration for image-guided thermal ablation of liver tumors</article-title>. <source>Med. Image Anal.</source> <volume>65</volume>, <fpage>101763</fpage>. <pub-id pub-id-type="doi">10.1016/j.media.2020.101763</pub-id><pub-id pub-id-type="pmid">32623279</pub-id></citation></ref>
<ref id="B77">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Wei</surname> <given-names>D.</given-names></name> <name><surname>Ahmad</surname> <given-names>S.</given-names></name> <name><surname>Huo</surname> <given-names>J.</given-names></name> <name><surname>Peng</surname> <given-names>W.</given-names></name> <name><surname>Ge</surname> <given-names>Y.</given-names></name> <name><surname>Xue</surname> <given-names>Z.</given-names></name> <etal/></person-group>. (<year>2019</year>). <article-title>&#x0201C;Synthesis and inpainting-based MR-CT registration for image-guided thermal ablation of liver tumors,&#x0201D;</article-title> in <source>International Conference on Medical Image Computing and Computer-Assisted Intervention</source> (<publisher-loc>Shanghai</publisher-loc>), <fpage>512</fpage>&#x02013;<lpage>520</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-030-32254-0_57</pub-id></citation>
</ref>
<ref id="B78">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wu</surname> <given-names>J.</given-names></name> <name><surname>Zhou</surname> <given-names>S.</given-names></name></person-group> (<year>2021</year>). <article-title>A disentangled representations based unsupervised deformable framework for cross-modality image registration</article-title>. <source>Annu. Int. Conf. IEEE Eng. Med. Biol. Soc</source>. <volume>2021</volume>, <fpage>3531</fpage>&#x02013;<lpage>3534</lpage>. <pub-id pub-id-type="doi">10.1109/EMBC46164.2021.9630778</pub-id><pub-id pub-id-type="pmid">34892001</pub-id></citation></ref>
<ref id="B79">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Xian</surname> <given-names>Y.</given-names></name> <name><surname>Sharma</surname> <given-names>S.</given-names></name> <name><surname>Schiele</surname> <given-names>B.</given-names></name> <name><surname>Akata</surname> <given-names>Z.</given-names></name></person-group> (<year>2019</year>). <article-title>&#x0201C;f-vaegan-d2: a feature generating framework for any-shot learning,&#x0201D;</article-title> in <source>Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition</source> (<publisher-loc>Oxford</publisher-loc>), <fpage>10275</fpage>&#x02013;<lpage>10284</lpage>. <pub-id pub-id-type="doi">10.1109/CVPR.2019.01052</pub-id></citation>
</ref>
<ref id="B80">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Xu</surname> <given-names>F.</given-names></name> <name><surname>Shen</surname> <given-names>Y.</given-names></name> <name><surname>Ding</surname> <given-names>L.</given-names></name> <name><surname>Yang</surname> <given-names>C.-Y.</given-names></name> <name><surname>Tan</surname> <given-names>H.</given-names></name> <name><surname>Wang</surname> <given-names>H.</given-names></name> <etal/></person-group>. (<year>2021</year>). <article-title>High-throughput mapping of a whole rhesus monkey brain at micrometer resolution</article-title>. <source>Nat. Biotechnol.</source> <volume>39</volume>, <fpage>1521</fpage>&#x02013;<lpage>1528</lpage>. <pub-id pub-id-type="doi">10.1038/s41587-021-00986-5</pub-id><pub-id pub-id-type="pmid">34312500</pub-id></citation></ref>
<ref id="B81">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Xu</surname> <given-names>Z.</given-names></name> <name><surname>Luo</surname> <given-names>J.</given-names></name> <name><surname>Yan</surname> <given-names>J.</given-names></name> <name><surname>Pulya</surname> <given-names>R.</given-names></name> <name><surname>Li</surname> <given-names>X.</given-names></name> <name><surname>Wells</surname> <given-names>W.</given-names></name> <etal/></person-group>. (<year>2020</year>). <article-title>&#x0201C;Adversarial uni-and multi-modal stream networks for multimodal image registration,&#x0201D;</article-title> in <source>International Conference on Medical Image Computing and Computer-Assisted Intervention</source>, <fpage>222</fpage>&#x02013;<lpage>232</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-030-59716-0_22</pub-id><pub-id pub-id-type="pmid">33283210</pub-id></citation></ref>
<ref id="B82">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Yan</surname> <given-names>P.</given-names></name> <name><surname>Xu</surname> <given-names>S.</given-names></name> <name><surname>Rastinehad</surname> <given-names>A. R.</given-names></name> <name><surname>Wood</surname> <given-names>B. J.</given-names></name></person-group> (<year>2018</year>). <article-title>&#x0201C;Adversarial image registration with application for MR and TRUS image fusion,&#x0201D;</article-title> in <source>International Workshop on Machine Learning in Medical Imaging</source>, <fpage>197</fpage>&#x02013;<lpage>204</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-030-00919-9_23</pub-id></citation>
</ref>
<ref id="B83">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Yang</surname> <given-names>H.</given-names></name> <name><surname>Qian</surname> <given-names>P.</given-names></name> <name><surname>Fan</surname> <given-names>C.</given-names></name></person-group> (<year>2020</year>). <article-title>An indirect multimodal image registration and completion method guided by image synthesis</article-title>. <source>Comput. Math. Methods Med.</source> <volume>2020</volume>, <fpage>2684851</fpage>. <pub-id pub-id-type="doi">10.1155/2020/2684851</pub-id><pub-id pub-id-type="pmid">32670390</pub-id></citation></ref>
<ref id="B84">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>X.</given-names></name> <name><surname>Jian</surname> <given-names>W.</given-names></name> <name><surname>Chen</surname> <given-names>Y.</given-names></name> <name><surname>Yang</surname> <given-names>S.</given-names></name></person-group> (<year>2020</year>). <source>Deform-GAN: An Unsupervised Learning Model for Deformable Registration.</source></citation>
</ref>
<ref id="B85">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Zhao</surname> <given-names>S.</given-names></name> <name><surname>Dong</surname> <given-names>Y.</given-names></name> <name><surname>Chang</surname> <given-names>E. I.</given-names></name> <name><surname>Xu</surname> <given-names>Y.</given-names></name></person-group> (<year>2019a</year>). <article-title>&#x0201C;Recursive cascaded networks for unsupervised medical image registration,&#x0201D;</article-title> in <source>Proceedings of the IEEE/CVF International Conference on Computer Vision</source> (<publisher-loc>Seoul</publisher-loc>) <fpage>10600</fpage>&#x02013;<lpage>10610</lpage>. <pub-id pub-id-type="doi">10.1109/ICCV.2019.01070</pub-id></citation>
</ref>
<ref id="B86">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhao</surname> <given-names>S.</given-names></name> <name><surname>Lau</surname> <given-names>T.</given-names></name> <name><surname>Luo</surname> <given-names>J.</given-names></name> <name><surname>Eric</surname> <given-names>I.</given-names></name> <name><surname>Chang</surname> <given-names>C.</given-names></name> <name><surname>Xu</surname> <given-names>Y.</given-names></name></person-group> (<year>2019b</year>). <article-title>Unsupervised 3D end-to-end medical image registration with volume tweening network</article-title>. <source>IEEE J. Biomed. Health Inform.</source> <volume>24</volume>, <fpage>1394</fpage>&#x02013;<lpage>1404</lpage>. <pub-id pub-id-type="doi">10.1109/JBHI.2019.2951024</pub-id><pub-id pub-id-type="pmid">31689224</pub-id></citation></ref>
<ref id="B87">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zheng</surname> <given-names>Y.</given-names></name> <name><surname>Sui</surname> <given-names>X.</given-names></name> <name><surname>Jiang</surname> <given-names>Y.</given-names></name> <name><surname>Che</surname> <given-names>T.</given-names></name> <name><surname>Zhang</surname> <given-names>S.</given-names></name> <name><surname>Yang</surname> <given-names>J.</given-names></name> <etal/></person-group>. (<year>2021</year>). <article-title>SymReg-GAN: symmetric image registration with generative adversarial networks</article-title>. <source>IEEE Trans. Patt. Anal. Mach. Intell</source>. <volume>44</volume>, <fpage>5631</fpage>&#x02013;<lpage>5646</lpage>. <pub-id pub-id-type="doi">10.1109/TPAMI.2021.3083543</pub-id><pub-id pub-id-type="pmid">34033536</pub-id></citation></ref>
<ref id="B88">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhou</surname> <given-names>B.</given-names></name> <name><surname>Augenfeld</surname> <given-names>Z.</given-names></name> <name><surname>Chapiro</surname> <given-names>J.</given-names></name> <name><surname>Zhou</surname> <given-names>S. K.</given-names></name> <name><surname>Liu</surname> <given-names>C.</given-names></name> <name><surname>Duncan</surname> <given-names>J. S.</given-names></name></person-group> (<year>2021</year>). <article-title>Anatomy-guided multimodal registration by learning segmentation without ground truth: application to intraprocedural CBCT/MR liver segmentation and registration</article-title>. <source>Med. Image Anal.</source> <volume>71</volume>, <fpage>102041</fpage>. <pub-id pub-id-type="doi">10.1016/j.media.2021.102041</pub-id><pub-id pub-id-type="pmid">33823397</pub-id></citation></ref>
<ref id="B89">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Zhu</surname> <given-names>J.-Y.</given-names></name> <name><surname>Park</surname> <given-names>T.</given-names></name> <name><surname>Isola</surname> <given-names>P.</given-names></name> <name><surname>Efros</surname> <given-names>A. A.</given-names></name></person-group> (<year>2017</year>). <article-title>&#x0201C;Unpaired image-to-image translation using cycle-consistent adversarial networks,&#x0201D;</article-title> in <source>Proceedings of the IEEE International Conference on Computer Vision</source> (<publisher-loc>Venice</publisher-loc>), <fpage>2223</fpage>&#x02013;<lpage>2232</lpage> <pub-id pub-id-type="doi">10.1109/ICCV.2017.244</pub-id></citation>
</ref>
</ref-list> 
</back>
</article>