<?xml version="1.0" encoding="UTF-8" standalone="no"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Neurosci.</journal-id>
<journal-title>Frontiers in Neuroscience</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Neurosci.</abbrev-journal-title>
<issn pub-type="epub">1662-453X</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fnins.2020.615435</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Neuroscience</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>LatLRR-FCNs: Latent Low-Rank Representation With Fully Convolutional Networks for Medical Image Fusion</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name><surname>Xu</surname> <given-names>Zhengyuan</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<xref ref-type="author-notes" rid="fn002"><sup>&#x02020;</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/1103759/overview"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Xiang</surname> <given-names>Wentao</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="author-notes" rid="fn002"><sup>&#x02020;</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/1142421/overview"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Zhu</surname> <given-names>Songsheng</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
</contrib>
<contrib contrib-type="author">
<name><surname>Zeng</surname> <given-names>Rui</given-names></name>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
</contrib>
<contrib contrib-type="author">
<name><surname>Marquez-Chin</surname> <given-names>Cesar</given-names></name>
<xref ref-type="aff" rid="aff4"><sup>4</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/846172/overview"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Chen</surname> <given-names>Zhen</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
</contrib>
<contrib contrib-type="author">
<name><surname>Chen</surname> <given-names>Xianqing</given-names></name>
<xref ref-type="aff" rid="aff5"><sup>5</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/713236/overview"/>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name><surname>Liu</surname> <given-names>Bin</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="corresp" rid="c002"><sup>&#x0002A;</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/1142448/overview"/>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name><surname>Li</surname> <given-names>Jianqing</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>&#x0002A;</sup></xref>
</contrib>
</contrib-group>
<aff id="aff1"><sup>1</sup><institution>The Key Laboratory of Clinical and Medical Engineering, School of Biomedical Engineering and Informatics, Nanjing Medical University</institution>, <addr-line>Nanjing</addr-line>, <country>China</country></aff>
<aff id="aff2"><sup>2</sup><institution>The Department of Medical Engineering, Wannan Medical College</institution>, <addr-line>Wuhu</addr-line>, <country>China</country></aff>
<aff id="aff3"><sup>3</sup><institution>The Brain and Mind Centre, The University of Sydney</institution>, <addr-line>Sydney, NSW</addr-line>, <country>Australia</country></aff>
<aff id="aff4"><sup>4</sup><institution>The KITE Research Institute, Toronto Rehabilitation Institute-University Health Network</institution>, <addr-line>Toronto, ON</addr-line>, <country>Canada</country></aff>
<aff id="aff5"><sup>5</sup><institution>The Department of Electrical Engineering, College of Engineering, Zhejiang Normal University</institution>, <addr-line>Jinhua</addr-line>, <country>China</country></aff>
<author-notes>
<fn fn-type="edited-by"><p>Edited by: Yu-Dong Zhang, University of Leicester, United Kingdom</p></fn>
<fn fn-type="edited-by"><p>Reviewed by: Siyuan Lu, University of Leicester, United Kingdom; David Haynor, University of Washington, United States</p></fn>
<corresp id="c001">&#x0002A;Correspondence: Jianqing Li <email>jqli&#x00040;njmu.edu.cn</email></corresp>
<corresp id="c002">Bin Liu <email>liubin&#x00040;njmu.edu.cn</email></corresp>
<fn fn-type="other" id="fn001"><p>This article was submitted to Brain Imaging Methods, a section of the journal Frontiers in Neuroscience</p></fn>
<fn fn-type="other" id="fn002"><p>&#x02020;These authors have contributed equally to this work</p></fn></author-notes>
<pub-date pub-type="epub">
<day>13</day>
<month>01</month>
<year>2021</year>
</pub-date>
<pub-date pub-type="collection">
<year>2020</year>
</pub-date>
<volume>14</volume>
<elocation-id>615435</elocation-id>
<history>
<date date-type="received">
<day>09</day>
<month>10</month>
<year>2020</year>
</date>
<date date-type="accepted">
<day>07</day>
<month>12</month>
<year>2020</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x000A9; 2021 Xu, Xiang, Zhu, Zeng, Marquez-Chin, Chen, Chen, Liu and Li.</copyright-statement>
<copyright-year>2021</copyright-year>
<copyright-holder>Xu, Xiang, Zhu, Zeng, Marquez-Chin, Chen, Chen, Liu and Li</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p></license> </permissions>
<abstract><p>Medical image fusion, which aims to derive complementary information from multi-modality medical images, plays an important role in many clinical applications, such as medical diagnostics and treatment. We propose the LatLRR-FCNs, which is a hybrid medical image fusion framework consisting of the latent low-rank representation (LatLRR) and the fully convolutional networks (FCNs). Specifically, the LatLRR module is used to decompose the multi-modality medical images into low-rank and saliency components, which can provide fine-grained details and preserve energies, respectively. The FCN module aims to preserve both global and local information by generating the weighting maps for each modality image. The final weighting map is obtained using the weighted local energy and the weighted sum of the eight-neighborhood-based modified Laplacian method. The fused low-rank component is generated by combining the low-rank components of each modality image according to the guidance provided by the final weighting map within pyramid-based fusion. A simple sum strategy is used for the saliency components. The usefulness and efficiency of the proposed framework are thoroughly evaluated on four medical image fusion tasks, including computed tomography (CT) and magnetic resonance (MR), T1- and T2-weighted MR, positron emission tomography and MR, and single-photon emission CT and MR. The results demonstrate that by leveraging the LatLRR for image detail extraction and the FCNs for global and local information description, we can achieve performance superior to the state-of-the-art methods in terms of both objective assessment and visual quality in some cases. Furthermore, our method has a competitive performance in terms of computational costs compared to other baselines.</p></abstract>
<kwd-group>
<kwd>multi-modality medical image</kwd>
<kwd>latent low-rank representation</kwd>
<kwd>fully convolutional networks</kwd>
<kwd>medical image fusion</kwd>
<kwd>Laplacian pyramid</kwd>
</kwd-group>
<contract-num rid="cn001">2017YFB1303200</contract-num>
<contract-num rid="cn002">61571113</contract-num>
<contract-sponsor id="cn001">National Key Research and Development Program of China<named-content content-type="fundref-id">10.13039/501100012166</named-content></contract-sponsor>
<contract-sponsor id="cn002">National Natural Science Foundation of China<named-content content-type="fundref-id">10.13039/501100001809</named-content></contract-sponsor>
<counts>
<fig-count count="10"/>
<table-count count="3"/>
<equation-count count="20"/>
<ref-count count="49"/>
<page-count count="17"/>
<word-count count="7951"/>
</counts>
</article-meta>
</front>
<body>
<sec sec-type="intro" id="s1">
<title>1. Introduction</title>
<p>Medical image fusion is a key technology that has been used extensively in clinical diagnosis and treatment planning (James and Dasarathy, <xref ref-type="bibr" rid="B16">2014</xref>). Modern medical imaging techniques mainly include computed tomography (CT), magnetic resonance (MR), single-photon emission computed tomography (SPECT), and positron emission tomography (PET) (Walrand et al., <xref ref-type="bibr" rid="B38">2017</xref>). CT has a high spatial and density resolution for dense structures (e.g., bones, implants), while MR has a high resolution for soft tissue (Wang et al., <xref ref-type="bibr" rid="B42">2016</xref>) (e.g., muscle, tendon, and fascia). PET is an advanced nuclear medical examination technique that allows visualization of biomolecular metabolism, receptors, and neurotransmitter activity <italic>in vivo</italic>. SPECT is often applied to quantify images of the physiological and pathological changes of organs or tissues <italic>in vitro</italic>. Evaluating different perspectives of these imaging techniques reveals that they do, to an extent, complement each other (Walrand et al., <xref ref-type="bibr" rid="B38">2017</xref>). Thus, medical image fusion can be utilized to combine different medical images and generate a new fusing image, providing the clinical information from each original image (Du et al., <xref ref-type="bibr" rid="B8">2016</xref>; Huang et al., <xref ref-type="bibr" rid="B15">2020</xref>).</p>
<p>To date, many medical image fusion studies have been reported (Toet, <xref ref-type="bibr" rid="B36">1989</xref>; Li et al., <xref ref-type="bibr" rid="B20">1995</xref>, <xref ref-type="bibr" rid="B22">2013</xref>; Petrovic and Xydeas, <xref ref-type="bibr" rid="B34">2004</xref>; Lewis et al., <xref ref-type="bibr" rid="B18">2007</xref>; Zhang and Guo, <xref ref-type="bibr" rid="B46">2009</xref>; Bhatnagar et al., <xref ref-type="bibr" rid="B5">2015</xref>; Wang Q. et al., <xref ref-type="bibr" rid="B41">2015</xref>; Geng et al., <xref ref-type="bibr" rid="B11">2017</xref>; Zhao and Lu, <xref ref-type="bibr" rid="B48">2017</xref>; Li H. et al., <xref ref-type="bibr" rid="B19">2018</xref>; Manchanda and Sharma, <xref ref-type="bibr" rid="B32">2018</xref>). Among them, multiscale transform (MST)-based methods are commonly used. The key point of MST-based fusion techniques is to decompose the original images into a multiscale transform domain (Li et al., <xref ref-type="bibr" rid="B20">1995</xref>). Some fusion rule strategies can be utilized to merge the transformed coefficients, and the merged coefficients are employed to reconstruct the composite image. Note that the current literature indicates that the non-subsampled shearlet transform (NSST) and non-subsampled contourlet transform (NSCT) achieve the optimum performance in terms of image representation among MST-based methods (Anitha et al., <xref ref-type="bibr" rid="B1">2015</xref>; Li Y. et al., <xref ref-type="bibr" rid="B23">2018</xref>; Yin et al., <xref ref-type="bibr" rid="B45">2018</xref>; Zhu et al., <xref ref-type="bibr" rid="B49">2019</xref>). Zhu et al. used NSCT to decompose medical image pairs into low-pass and high-pass sub-bands, where a phase congruency rule was applied to fuse the high-pass sub-bands and a local Laplacian energy-based fusion rule was utilized for the low-pass sub-bands (Zhu et al., <xref ref-type="bibr" rid="B49">2019</xref>). Later, Yin et al. introduced a novel framework in which the high-frequency coefficients were fused by a parameter-adaptive pulse coupled neural network (PA-PCNN), and the weighted local energy and the weighted sum of eight-neighborhood-based modified Laplacian were utilized to fuse low-frequency bands in the NSST domain (Yin et al., <xref ref-type="bibr" rid="B45">2018</xref>). However, due to the nature of the transformation, MST-based (including NCST-based and NSST-based) fusion methods may not express and extract certain significant structures of source images properly without being sensitive to misregistration.</p>
<p>To address the misregistration problem in MST-based methods, sparse representation (SR) has emerged as another popular and powerful theory in the medical image fusion field (Liu and Wang, <xref ref-type="bibr" rid="B28">2014</xref>; Liu et al., <xref ref-type="bibr" rid="B26">2016</xref>, <xref ref-type="bibr" rid="B27">2019</xref>; Fei et al., <xref ref-type="bibr" rid="B10">2017</xref>). A typical SR-based medical image fusion method includes three basic steps: (1) a given dictionary is used to find the sparsest representation of source images; (2) some fusion rules are used to integrate the sparse representation coefficients; and (3) the integrated sparse representation coefficients and given dictionary are utilized to construct the fused image. For example, Liu and Wang (<xref ref-type="bibr" rid="B28">2014</xref>) proposed a novel adaptive sparse representation model for medical image fusion, where a set of more compact sub-dictionaries was learned to replace the single redundant dictionary in the traditional SR approach and achieved better results. Although the SR-based and extended methods are robust in terms of noise and misregistration to some extent, they cannot capture global information and suffer from significant energy loss.</p>
<p>In the field of medical image fusion, a key issue is to calculate a weight map since it reflects pixel activity information from different modality images, determining the quality of the final fused image. The weight map is calculated by two steps: activity level measurement and weight assignment. However, these two steps suffer from the robustness problem because traditional methods cannot deal with noise and misregistration well, as indicated in Liu et al. (<xref ref-type="bibr" rid="B25">2017</xref>). To improve the robustness of activity level measurement and weight assignment, Liu et al. (<xref ref-type="bibr" rid="B25">2017</xref>) introduced a deep learning fusion method with a simple multi-layer convolutional neural network (CNN) using the decision map and the medical image under the pyramid-based image fusion framework to reconstruct the fused medical image. While such a method achieves some success in specific medical image fusion tasks, this work may fail in multi-modal image fusion because the simple use of the CNN cannot extract fine-grained details efficiently.</p>
<p>To address the aforementioned challenges, we propose a novel hybrid medical image fusion framework with two principal elements (e.g., LatLRR and FCNs), inspired by Liu and Wang (<xref ref-type="bibr" rid="B28">2014</xref>) and Liu et al. (<xref ref-type="bibr" rid="B25">2017</xref>). The main contributions of this paper are as follows:</p>
<list list-type="bullet">
<list-item><p>The latent low-rank representation (LatLRR) is applied to decompose the medical image into low-rank (for extraction of details) and saliency components (for the preservation of energies).</p></list-item>
<list-item><p>In the context of the low-rank component, to avoid the fixed-length feature vector from the final full connection layer and the information loss in the traditional CNN, three different FCNs (due to the nature of an input image of arbitrary size) are applied to produce a correspondingly-sized feature map with an efficient deconvolution layer (Guo et al., <xref ref-type="bibr" rid="B12">2018</xref>), where a prediction is generated for each pixel and the spatial information in the original input image is retained. A sum strategy is used to fuse the saliency parts for energy preservation.</p></list-item>
<list-item><p>To the best of our knowledge, this fusion strategy in combination with LatLRR and FCNs is the first to be applied in the medical image domain.</p></list-item>
</list>
<p>The remainder of this paper is structured as follows. In section 2, the proposed fusion strategy is described in detail. Section 3 gives the experimental configurations. Section 4 illustrates a comparative study between the proposed frameworks and five representative medical image fusion methods in terms of visual quality and quantitative and computational cost assessments. The conclusion is drawn in section 5.</p>
</sec>
<sec sec-type="methods" id="s2">
<title>2. Methodology</title>
<p>As shown in <xref ref-type="fig" rid="F1">Figure 1</xref>, each proposed framework is fed with a pair of pre-registered multi-modality medical source images, and outputs the fused medical image via the following four steps:</p>
<list list-type="bullet">
<list-item><p>We use the <bold>LatLRR</bold> theory to decompose the two medical source images into low-rank and saliency components (see section 2.1).</p></list-item>
<list-item><p>To capture the detailed information of each source, a novel <bold>fusion framework of the low-rank components</bold> for each paired source based on FCNs, score maps, weight maps, and pyramid fusion is described (see section 2.2).</p></list-item>
<list-item><p>To retain the energies of each source, a simple sum strategy is used to <bold>fuse the saliency components</bold> and <bold>reconstruct the fused image</bold> (see section 2.3).</p></list-item>
</list>
<fig id="F1" position="float">
<label>Figure 1</label>
<caption><p>Schematic diagram of the proposed end-to-end frameworks (LatLRR-FCNs). The proposed LatLRR-FCNs enable the fusion image to extract details and preserve energies from paired sources. It is composed of four parts: <bold>(a)</bold> LatLRR decomposition, <bold>(b)</bold> fusion of low-rank components, <bold>(c)</bold> fusion of saliency components, and <bold>(d)</bold> reconstruction of fused image. <italic>Img</italic><sub>1</sub> and <italic>Img</italic><sub>2</sub> are the source medical images, <italic>Lr</italic><sub>1</sub> and <italic>Lr</italic><sub>2</sub> are the low-rank components of <italic>Img</italic><sub>1</sub> and <italic>Img</italic><sub>2</sub>, <italic>Ls</italic><sub>1</sub> and <italic>Ls</italic><sub>2</sub> are the saliency components of <italic>Img</italic><sub>1</sub> and <italic>Img</italic><sub>2</sub>, <inline-formula><mml:math id="M2"><mml:msup><mml:mrow><mml:msub><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>:</mml:mo><mml:mi>C</mml:mi></mml:mrow></mml:msup></mml:math></inline-formula> and <inline-formula><mml:math id="M3"><mml:msup><mml:mrow><mml:msub><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>:</mml:mo><mml:mi>C</mml:mi></mml:mrow></mml:msup></mml:math></inline-formula> are the score maps, <italic>W</italic><sub>1</sub> and <italic>W</italic><sub>1</sub> are the initial weight maps of <italic>Lr</italic><sub>1</sub> and <italic>Lr</italic><sub>2</sub>, and the final fused weight map is <italic>F</italic><sub><italic>w</italic></sub>. <italic>F</italic><sub><italic>lr</italic></sub> is the fused low-rank component, <italic>F</italic><sub><italic>ls</italic></sub> is the fused saliency component, and the final fused image is <italic>F</italic>.</p></caption>
<graphic xlink:href="fnins-14-615435-g0001.tif"/>
</fig>
<sec>
<title>2.1. LatLRR Decomposition</title>
<p>The LatLRR theory was first proposed by Liu and Yan (<xref ref-type="bibr" rid="B24">2011</xref>), integrating subspace segmentation and feature extraction simultaneously, to extract the global and local structure from raw data in the context of natural images. It can be summarized into the following problem (Li and Wu, <xref ref-type="bibr" rid="B21">2018</xref>):</p>
<disp-formula id="E1"><label>(1)</label><mml:math id="M1"><mml:mtable columnalign='left'><mml:mtr><mml:mtd><mml:munder><mml:mrow><mml:mi>min</mml:mi></mml:mrow><mml:mrow><mml:mi>X</mml:mi><mml:mo>,</mml:mo><mml:mi>Y</mml:mi><mml:mo>,</mml:mo><mml:mi>Z</mml:mi></mml:mrow></mml:munder><mml:msub><mml:mrow><mml:mo>&#x02016;</mml:mo><mml:mi>X</mml:mi><mml:mo>&#x02016;</mml:mo></mml:mrow><mml:mo>*</mml:mo></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mrow><mml:mo>&#x02016;</mml:mo><mml:mi>Y</mml:mi><mml:mo>&#x02016;</mml:mo></mml:mrow><mml:mo>*</mml:mo></mml:msub><mml:mo>+</mml:mo><mml:mi>&#x003BB;</mml:mi><mml:msub><mml:mrow><mml:mo>&#x02016;</mml:mo><mml:mi>Z</mml:mi><mml:mo>&#x02016;</mml:mo></mml:mrow><mml:mn>1</mml:mn></mml:msub></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mi>s</mml:mi><mml:mo>.</mml:mo><mml:mi>t</mml:mi><mml:mo>.</mml:mo><mml:mi>I</mml:mi><mml:mi>m</mml:mi><mml:mi>g</mml:mi><mml:mo>=</mml:mo><mml:mi>I</mml:mi><mml:mi>m</mml:mi><mml:mi>g</mml:mi><mml:mi>X</mml:mi><mml:mo>+</mml:mo><mml:mi>I</mml:mi><mml:mi>m</mml:mi><mml:mi>g</mml:mi><mml:mi>Y</mml:mi><mml:mo>+</mml:mo><mml:mi>Z</mml:mi></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where &#x02016;&#x02016;<sub>&#x0002A;</sub> denotes the nuclear norm, &#x02016;&#x02016;<sub>1</sub> denotes the <italic>l</italic><sub>1</sub>-norm, and &#x003BB; &#x0003E; 0 is the balance coefficient. <italic>Img</italic> is the observed data matrix, and <italic>X</italic> and <italic>Y</italic> denote the low-rank and saliency coefficients, respectively. Note that <xref ref-type="fig" rid="F2">Figure 2</xref> explains the subject to Equation (1), where <italic>ImgX</italic>, <italic>ImgY</italic>, and <italic>Z</italic> are the low-rank, saliency, and noise components of <italic>Img</italic>, respectively.</p>
<fig id="F2" position="float">
<label>Figure 2</label>
<caption><p>The LatLRR decomposing operation. <italic>Img</italic> is the observed image. <italic>ImgX</italic> and <italic>ImgY</italic> are the low-rank and saliency components of <italic>Img</italic>, respectively. <italic>Z</italic> denotes the noisy component.</p></caption>
<graphic xlink:href="fnins-14-615435-g0002.tif"/>
</fig>
<p>In this paper, <bold>the LatLRR decomposition</bold> of Equation (1) can be solved by the inexact augmented Lagrangian multiplier method (Wang et al., <xref ref-type="bibr" rid="B40">2013</xref>), where it extracts the low-rank and saliency components (e.g., <italic>ImgX</italic><sub><italic>j</italic></sub> and <italic>ImgY</italic><sub><italic>j</italic></sub>) from medical image <italic>Img</italic><sub><italic>j</italic></sub> with <italic>j</italic> &#x0003D; 1, 2 (here, we consider two medical images, as shown in <xref ref-type="fig" rid="F1">Figure 1a</xref>).</p>
</sec>
<sec>
<title>2.2. Fusion of Low-Rank Components</title>
<p>The <bold>fusion of low-rank component</bold> details can be seen in <xref ref-type="fig" rid="F1">Figure 1b</xref>, including the FCN model for producing score maps (section 2.2.1), zero-phase component analysis (ZCA) (Kessy et al., <xref ref-type="bibr" rid="B17">2018</xref>), and <italic>l</italic><sub>1</sub>-norm operations (section 2.2.2) for whiting the score maps and generating the weight maps, respectively, weighted local energy (WLE) and weighted sum of eight-neighborhood-based modified Laplacian (WSEML) (Yin et al., <xref ref-type="bibr" rid="B45">2018</xref>) operations (section 2.2.3) for obtaining the fused weight map, and pyramid fusion strategy (section 2.2.4) for reconstructing the fused low-rank component.</p>
<sec>
<title>2.2.1. FCN Model</title>
<p>The fully convolutional networks (FCNs), demonstrated in many studies (Long et al., <xref ref-type="bibr" rid="B29">2015</xref>; Wang L. et al., <xref ref-type="bibr" rid="B39">2015</xref>; Chen et al., <xref ref-type="bibr" rid="B6">2017</xref>; Guo et al., <xref ref-type="bibr" rid="B12">2018</xref>), achieved significant performance in image semantic segmentation. In the FCN architecture, after multiple convolutions and pooling processes, the obtained image size will be progressively smaller with a lower resolution, resulting in a heatmap (coarse output). To keep the output the same size as the input, a skip architecture is used for upsampling. In this work, three different scenarios are tested, as shown in <xref ref-type="fig" rid="F3">Figure 3</xref>. For each scenario, there are 38 layers of FCNs before upsampling, including 16 convolutional layers (blue color block in <xref ref-type="fig" rid="F3">Figure 3A</xref>), 15 rule layers, five pooling layers (green color block in <xref ref-type="fig" rid="F3">Figure 3A</xref>), and two dropout layers. In <xref ref-type="fig" rid="F3">Figure 3A</xref>, the FCN-32s is a single-stream net in which up-samples stride 32 predictions back to pixels in a single step, but the upsampling output is very coarse. To obtain the refined outputs of FCN-16s, the final layer and the pool4 layer are used to combine the predictions in <xref ref-type="fig" rid="F3">Figure 3B</xref> at stride 16. In <xref ref-type="fig" rid="F3">Figure 3C</xref>, to obtain the outputs of FCN-8s with greater precision, the pool3 layer, the pool4 layer, and the final layer are utilized to combine the predictions at stride 8. As shown in <xref ref-type="fig" rid="F1">Figure 1b</xref>, the three trained FCNs (FCN-32s, FCN-16s, and FCN-8s) are utilized to classify a pair of the low-rank components of medical source images <italic>Lr</italic><sub>1</sub> &#x0003D; <italic>ImgX</italic><sub>1</sub> and <italic>Lr</italic><sub>2</sub> &#x0003D; <italic>ImgX</italic><sub>2</sub> pixel by pixel, producing the corresponding score maps <inline-formula><mml:math id="M4"><mml:msup><mml:mrow><mml:msub><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>:</mml:mo><mml:mi>C</mml:mi></mml:mrow></mml:msup><mml:mo>,</mml:mo><mml:mi>C</mml:mi><mml:mo>=</mml:mo><mml:mn>21</mml:mn><mml:mo>,</mml:mo><mml:mi>j</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mn>2</mml:mn></mml:math></inline-formula> (the choice of <italic>C</italic> &#x0003D; 21 can be seen in section 3.4).</p>
<fig id="F3" position="float">
<label>Figure 3</label>
<caption><p>The skip architecture for upsampling for three scenarios: <bold>(A)</bold> FNC-32s, <bold>(B)</bold> FNC-16s, and <bold>(C)</bold> FNC-8s.</p></caption>
<graphic xlink:href="fnins-14-615435-g0003.tif"/>
</fig>
</sec>
<sec>
<title>2.2.2. ZCA and <italic>l</italic><sub>1</sub>-Norm Operations</title>
<p>The details for ZCA and <italic>l</italic><sub>1</sub>-norm operations are depicted in <xref ref-type="fig" rid="F4">Figure 4</xref>. To project the original redundancy score maps into a sparse subspace, we used ZCA to whiten those score maps <inline-formula><mml:math id="M5"><mml:msup><mml:mrow><mml:msub><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>:</mml:mo><mml:mi>C</mml:mi></mml:mrow></mml:msup></mml:math></inline-formula> and to obtain the score maps <inline-formula><mml:math id="M6"><mml:msup><mml:mrow><mml:msub><mml:mrow><mml:mi>&#x0015C;</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>:</mml:mo><mml:mi>C</mml:mi></mml:mrow></mml:msup></mml:math></inline-formula>. Among the ZCA, the covariance matrix <inline-formula><mml:math id="M7"><mml:mi>C</mml:mi><mml:msup><mml:mrow><mml:msub><mml:mrow><mml:mi>o</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup></mml:math></inline-formula> is decomposed as follows:</p>
<disp-formula id="E2"><label>(2)</label><mml:math id="M8"><mml:mtable class="eqnarray" columnalign="right center left"><mml:mtr><mml:mtd><mml:mi>C</mml:mi><mml:msup><mml:mrow><mml:msub><mml:mrow><mml:mi>o</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup><mml:mo>=</mml:mo><mml:msup><mml:mrow><mml:msub><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup><mml:mo>&#x000D7;</mml:mo><mml:msup><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msup><mml:mrow><mml:msub><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mi>T</mml:mi></mml:mrow></mml:msup><mml:mo>,</mml:mo><mml:mtext>&#x00A0;</mml:mtext><mml:mi>C</mml:mi><mml:msup><mml:mrow><mml:msub><mml:mrow><mml:mi>o</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup><mml:mo>=</mml:mo><mml:mi>U</mml:mi><mml:mi>&#x003A3;</mml:mi><mml:msup><mml:mrow><mml:mi>V</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi></mml:mrow></mml:msup></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where <italic>i</italic> &#x0003D; 1, 2, &#x022EF;&#x02009;, <italic>C</italic>; <italic>j</italic> &#x0003D; 1, 2, and <italic>i</italic> denote the <italic>i</italic>&#x02212;<italic>th</italic> channel score map. Note that <italic>U</italic>, &#x003A3; and <italic>V</italic> define the left singular, singular values, and right singular matrixes, respectively (Chen et al., <xref ref-type="bibr" rid="B7">2018</xref>). An alternative solution named <inline-formula><mml:math id="M9"><mml:msup><mml:mrow><mml:msub><mml:mrow><mml:mi>&#x0015C;</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup></mml:math></inline-formula> is given as follows:</p>
<disp-formula id="E3"><label>(3)</label><mml:math id="M10"><mml:mtable class="eqnarray" columnalign="right center left"><mml:mtr><mml:mtd><mml:msup><mml:mrow><mml:msub><mml:mrow><mml:mi>&#x0015C;</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup><mml:mo>=</mml:mo><mml:msup><mml:mrow><mml:msub><mml:mrow><mml:mi>K</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup><mml:mo>&#x000D7;</mml:mo><mml:msup><mml:mrow><mml:msub><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup><mml:mo>,</mml:mo><mml:mtext>&#x00A0;</mml:mtext><mml:msup><mml:mrow><mml:msub><mml:mrow><mml:mi>K</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup><mml:mo>=</mml:mo><mml:mi>U</mml:mi><mml:msup><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>&#x003A3;</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mi>&#x003B7;</mml:mi><mml:mi>I</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mo>-</mml:mo><mml:mfrac><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:mfrac></mml:mrow></mml:msup><mml:msup><mml:mrow><mml:mi>U</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi></mml:mrow></mml:msup></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where &#x003B7; is a small value avoiding bad matrix inversion and <italic>I</italic> is the identity matrix. Then, the local <italic>l</italic><sub>1</sub>-norm and average operations are used to calculate the initial weight map <italic>W</italic><sub><italic>j</italic></sub>:</p>
<disp-formula id="E4"><label>(4)</label><mml:math id="M11"><mml:mtable class="eqnarray" columnalign="right center left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>W</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:msubsup><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>x</mml:mi><mml:mo>=</mml:mo><mml:mi>u</mml:mi><mml:mo>-</mml:mo><mml:mi>k</mml:mi></mml:mrow><mml:mrow><mml:mi>u</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mi>k</mml:mi></mml:mrow></mml:msubsup><mml:msubsup><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>y</mml:mi><mml:mo>=</mml:mo><mml:mi>v</mml:mi><mml:mo>-</mml:mo><mml:mi>k</mml:mi></mml:mrow><mml:mrow><mml:mi>v</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mi>k</mml:mi></mml:mrow></mml:msubsup><mml:msub><mml:mrow><mml:mo>|</mml:mo><mml:mo>|</mml:mo><mml:msup><mml:mrow><mml:msub><mml:mrow><mml:mi>&#x0015C;</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi><mml:mo>,</mml:mo><mml:mi>y</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>|</mml:mo><mml:mo>|</mml:mo></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mn>2</mml:mn><mml:mi>k</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x000D7;</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mn>2</mml:mn><mml:mi>k</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:mfrac></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where <italic>k</italic> &#x0003D; 2 and the average <italic>l</italic><sub>1</sub>-norm is calculated by a window centered at <inline-formula><mml:math id="M12"><mml:msup><mml:mrow><mml:msub><mml:mrow><mml:mi>&#x0015C;</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mi>v</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula>.</p>
<fig id="F4" position="float">
<label>Figure 4</label>
<caption><p>The ZCA and <italic>l</italic><sub>1</sub>-norm operations for whiting the score maps and generating the weight maps, respectively.</p></caption>
<graphic xlink:href="fnins-14-615435-g0004.tif"/>
</fig>
</sec>
<sec>
<title>2.2.3. WLE and WSEML Operations</title>
<p>Once the initial weight maps <italic>W</italic><sub>1</sub> and <italic>W</italic><sub>2</sub> is calculated, the WLE and WSEML are applied to acquire the final fused weight map <italic>F</italic><sub><italic>w</italic></sub>, which is described in <xref ref-type="fig" rid="F1">Figure 1b</xref> with the orange block.</p>
<p>First, the WLE of each <italic>W</italic><sub><italic>j</italic></sub> (i.e., &#x003A6;<sub><italic>j</italic></sub>) is calculated as follows:</p>
<disp-formula id="E5"><label>(5)</label><mml:math id="M13"><mml:mtable class="eqnarray" columnalign="right center left"><mml:mtr><mml:mtd><mml:mtable class="aligned"><mml:mtr><mml:mtd columnalign="right"><mml:msub><mml:mrow><mml:mi>&#x003A6;</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mi>v</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mstyle displaystyle="true"><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>p</mml:mi><mml:mo>=</mml:mo><mml:mo>-</mml:mo><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:munderover></mml:mstyle><mml:mstyle displaystyle="true"><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>q</mml:mi><mml:mo>=</mml:mo><mml:mo>-</mml:mo><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:munderover></mml:mstyle><mml:mrow><mml:mo>{</mml:mo><mml:mrow><mml:mi>&#x003A9;</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>p</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mi>r</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>q</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mi>r</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:msub><mml:mrow><mml:mi>W</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub><mml:msup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>u</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mi>p</mml:mi><mml:mo>,</mml:mo><mml:mi>v</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mi>q</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:mrow><mml:mo>}</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where <italic>j</italic> &#x02208; {1, 2} and &#x003A9; denote a (2<italic>r</italic>&#x0002B;1) &#x000D7; (2<italic>r</italic>&#x0002B;1) weighting matrix. The value of each element in &#x003A9; is 2<sup>2<italic>r</italic>&#x02212;<italic>d</italic></sup> with radius <italic>r</italic>, <italic>d</italic> denotes the element of a four-neighborhood distance to the center. If <italic>r</italic> is 1, &#x003A9; is equal to</p>
<disp-formula id="E6"><mml:math id="M14"><mml:mfrac><mml:mrow><mml:mtext class="textrm" mathvariant="normal">1</mml:mtext></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">16</mml:mtext></mml:mrow></mml:mfrac><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mtable style="text-align:axis;" equalrows="false" columnlines="none none" equalcolumns="false" class="array"><mml:mtr><mml:mtd><mml:mtext class="textrm" mathvariant="normal">1</mml:mtext></mml:mtd><mml:mtd><mml:mtext class="textrm" mathvariant="normal">2</mml:mtext></mml:mtd><mml:mtd><mml:mtext class="textrm" mathvariant="normal">1</mml:mtext></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mtext class="textrm" mathvariant="normal">2</mml:mtext></mml:mtd><mml:mtd><mml:mtext class="textrm" mathvariant="normal">4</mml:mtext></mml:mtd><mml:mtd><mml:mtext class="textrm" mathvariant="normal">2</mml:mtext></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mtext class="textrm" mathvariant="normal">1</mml:mtext></mml:mtd><mml:mtd><mml:mtext class="textrm" mathvariant="normal">2</mml:mtext></mml:mtd><mml:mtd><mml:mtext class="textrm" mathvariant="normal">1</mml:mtext></mml:mtd></mml:mtr></mml:mtable></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:math></disp-formula>
<p>Second, the WSEML of each <italic>W</italic><sub><italic>j</italic></sub> (i.e., &#x003A8;<sub><italic>j</italic></sub>) is given as follows:</p>
<disp-formula id="E7"><label>(6)</label><mml:math id="M15"><mml:mtable columnalign='left'><mml:mtr><mml:mtd><mml:msub><mml:mi>&#x003A8;</mml:mi><mml:mi>j</mml:mi></mml:msub><mml:mo stretchy='false'>(</mml:mo><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mi>v</mml:mi><mml:mo stretchy='false'>)</mml:mo><mml:mo>=</mml:mo><mml:mstyle displaystyle='true'><mml:munderover><mml:mo>&#x02211;</mml:mo><mml:mrow><mml:mi>p</mml:mi><mml:mo>=</mml:mo><mml:mo>&#x02212;</mml:mo><mml:mi>r</mml:mi></mml:mrow><mml:mi>r</mml:mi></mml:munderover><mml:mrow><mml:mstyle displaystyle='true'><mml:munderover><mml:mo>&#x02211;</mml:mo><mml:mrow><mml:mi>q</mml:mi><mml:mo>=</mml:mo><mml:mo>&#x02212;</mml:mo><mml:mi>r</mml:mi></mml:mrow><mml:mi>r</mml:mi></mml:munderover><mml:mrow><mml:mrow><mml:mo>{</mml:mo><mml:mrow><mml:mi>&#x003A9;</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:mi>p</mml:mi><mml:mo>+</mml:mo><mml:mi>r</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>q</mml:mi><mml:mo>+</mml:mo><mml:mi>r</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:mrow></mml:mrow></mml:mstyle></mml:mrow></mml:mstyle></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mtext>&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;</mml:mtext><mml:mrow><mml:mrow><mml:mo>&#x000D7;</mml:mo><mml:mi>E</mml:mi><mml:mi>M</mml:mi><mml:msub><mml:mi>L</mml:mi><mml:mi>j</mml:mi></mml:msub><mml:mo stretchy='false'>(</mml:mo><mml:mi>u</mml:mi><mml:mo>+</mml:mo><mml:mi>p</mml:mi><mml:mo>,</mml:mo><mml:mi>v</mml:mi><mml:mo>+</mml:mo><mml:mi>q</mml:mi><mml:mo stretchy='false'>)</mml:mo></mml:mrow><mml:mo>}</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where <italic>EML</italic> is expressed as follows:</p>
<disp-formula id="E8"><label>(7)</label><mml:math id="M16"><mml:mtable class="eqnarray" columnalign="right center left"><mml:mtr><mml:mtd><mml:mtable style="text-align:axis;" equalrows="false" columnlines="none" equalcolumns="false" class="array"><mml:mtr><mml:mtd><mml:mi>E</mml:mi><mml:mi>M</mml:mi><mml:msub><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mi>v</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mo>|</mml:mo><mml:mn>2</mml:mn><mml:msub><mml:mrow><mml:mi>W</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mi>v</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>W</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>v</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>W</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>v</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>|</mml:mo></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mo>&#x0002B;</mml:mo><mml:mo>|</mml:mo><mml:mn>2</mml:mn><mml:msub><mml:mrow><mml:mi>W</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mi>v</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>W</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mi>v</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>W</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mi>v</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>|</mml:mo></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mo>&#x0002B;</mml:mo><mml:mfrac><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:msqrt><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msqrt></mml:mrow></mml:mfrac><mml:mo>|</mml:mo><mml:mn>2</mml:mn><mml:msub><mml:mrow><mml:mi>W</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mi>v</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>W</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>v</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>W</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>v</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>|</mml:mo></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mo>&#x0002B;</mml:mo><mml:mfrac><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:msqrt><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msqrt></mml:mrow></mml:mfrac><mml:mo>|</mml:mo><mml:mn>2</mml:mn><mml:msub><mml:mrow><mml:mi>W</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mi>v</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>W</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>v</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>W</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>v</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>|</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>Finally, the fused weight map <italic>F</italic><sub><italic>w</italic></sub> is calculated by the following rule:</p>
<disp-formula id="E9"><label>(8)</label><mml:math id="M17"><mml:mtable class="eqnarray" columnalign="right center left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>F</mml:mi></mml:mrow><mml:mrow><mml:mi>w</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mi>v</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mrow><mml:mo>{</mml:mo><mml:mtable style="text-align:axis;" columnalign="left" equalrows="false" columnlines="none none" equalcolumns="false" class="array"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>W</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mi>v</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:mtd><mml:mtd><mml:mtext class="textrm" mathvariant="normal">if</mml:mtext><mml:mtext>&#x00A0;</mml:mtext><mml:msub><mml:mrow><mml:mi>&#x003A6;</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mi>v</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x000B7;</mml:mo><mml:msub><mml:mrow><mml:mtext>&#x003A8;</mml:mtext></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mi>v</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd><mml:mtd><mml:mo>&#x02265;</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003A6;</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mi>v</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x000B7;</mml:mo><mml:msub><mml:mrow><mml:mtext>&#x003A8;</mml:mtext></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mi>v</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>W</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mi>v</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:mtd><mml:mtd><mml:mtext class="textrm" mathvariant="normal">otherwise</mml:mtext></mml:mtd></mml:mtr><mml:mtr></mml:mtr></mml:mtable></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
</sec>
<sec>
<title>2.2.4. Pyramid Fusion Strategy</title>
<p>As shown in <xref ref-type="fig" rid="F1">Figure 1b</xref>, the fused weight map <italic>F</italic><sub><italic>w</italic></sub> is decomposed into a Gaussian pyramid <italic>G</italic>{<italic>S</italic>}<sup><italic>l</italic></sup> (green color arrow). The low-rank components <italic>Lr</italic><sub>1</sub> and <italic>Lr</italic><sub>2</sub> are decomposed into a Laplacian pyramid (dark blue color arrow) <italic>L</italic>{<italic>C</italic>}<sup><italic>l</italic></sup> and <italic>L</italic>{<italic>M</italic>}<sup><italic>l</italic></sup>, respectively. Note that <italic>l</italic> denotes the <italic>l</italic>-th decomposition level, which is calculated by the following:</p>
<disp-formula id="E10"><label>(9)</label><mml:math id="M18"><mml:mtable class="eqnarray" columnalign="right center left"><mml:mtr><mml:mtd><mml:mi>l</mml:mi><mml:mo>=</mml:mo><mml:mo>&#x0230A;</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mo class="qopname">log</mml:mo></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>X</mml:mi><mml:mo>,</mml:mo><mml:mi>Y</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo>&#x0230B;</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where &#x0230A;&#x000B7;&#x0230B; is the flooring operation and the spatial size of the low-rank component is <italic>X</italic> &#x000D7; <italic>Y</italic>.</p>
<p>Next, those coefficients about <italic>L</italic>{<italic>F</italic>} are calculated at each decomposition level <italic>l</italic>:</p>
<disp-formula id="E11"><label>(10)</label><mml:math id="M19"><mml:mtable columnalign='left'><mml:mtr><mml:mtd><mml:mi>L</mml:mi><mml:msup><mml:mrow><mml:mo>&#x0007B;</mml:mo><mml:mi>F</mml:mi><mml:mo>&#x0007D;</mml:mo></mml:mrow><mml:mi>l</mml:mi></mml:msup><mml:mo stretchy='false'>(</mml:mo><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:mi>v</mml:mi><mml:mo stretchy='false'>)</mml:mo></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mo>=</mml:mo><mml:mrow><mml:mo>{</mml:mo><mml:mrow><mml:mtable columnalign='left'><mml:mtr columnalign='left'><mml:mtd columnalign='left'><mml:mrow><mml:mo>&#x0007B;</mml:mo><mml:mi>G</mml:mi><mml:msup><mml:mrow><mml:mo>&#x0007B;</mml:mo><mml:mi>S</mml:mi><mml:mo>&#x0007D;</mml:mo></mml:mrow><mml:mi>l</mml:mi></mml:msup><mml:mo stretchy='false'>(</mml:mo><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:mi>v</mml:mi><mml:mo stretchy='false'>)</mml:mo><mml:mo>&#x000B7;</mml:mo><mml:mi>L</mml:mi><mml:msup><mml:mrow><mml:mo>&#x0007B;</mml:mo><mml:mi>C</mml:mi><mml:mo>&#x0007D;</mml:mo></mml:mrow><mml:mi>l</mml:mi></mml:msup><mml:mo stretchy='false'>(</mml:mo><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:mi>v</mml:mi><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr columnalign='left'><mml:mtd columnalign='left'><mml:mrow><mml:mo>+</mml:mo><mml:mo stretchy='false'>(</mml:mo><mml:mn>1</mml:mn><mml:mo>&#x02212;</mml:mo><mml:mi>G</mml:mi><mml:msup><mml:mrow><mml:mo>&#x0007B;</mml:mo><mml:mi>S</mml:mi><mml:mo>&#x0007D;</mml:mo></mml:mrow><mml:mi>l</mml:mi></mml:msup><mml:mo stretchy='false'>(</mml:mo><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mi>v</mml:mi><mml:mo stretchy='false'>)</mml:mo><mml:mo stretchy='false'>)</mml:mo><mml:mi>L</mml:mi><mml:msup><mml:mrow><mml:mo>&#x0007B;</mml:mo><mml:mi>M</mml:mi><mml:mo>&#x0007D;</mml:mo></mml:mrow><mml:mi>l</mml:mi></mml:msup><mml:mo stretchy='false'>(</mml:mo><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:mi>v</mml:mi><mml:mo stretchy='false'>)</mml:mo><mml:mo>&#x0007D;</mml:mo><mml:mo>,</mml:mo><mml:mtext>&#x02003;</mml:mtext><mml:mi>i</mml:mi><mml:mi>f</mml:mi><mml:mo>&#x0007B;</mml:mo><mml:msup><mml:mi>Q</mml:mi><mml:mi>l</mml:mi></mml:msup><mml:mo stretchy='false'>(</mml:mo><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:mi>v</mml:mi><mml:mo stretchy='false'>)</mml:mo><mml:mo>&#x02265;</mml:mo><mml:mi>&#x003C4;</mml:mi><mml:mo>&#x0007D;</mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr columnalign='left'><mml:mtd columnalign='left'><mml:mrow><mml:mi>L</mml:mi><mml:msup><mml:mrow><mml:mo>&#x0007B;</mml:mo><mml:mi>C</mml:mi><mml:mo>&#x0007D;</mml:mo></mml:mrow><mml:mi>l</mml:mi></mml:msup><mml:mo stretchy='false'>(</mml:mo><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:mi>v</mml:mi><mml:mo stretchy='false'>)</mml:mo><mml:mo>,</mml:mo><mml:mtext>&#x02003;</mml:mtext><mml:mi>i</mml:mi><mml:mi>f</mml:mi><mml:mo>&#x0007B;</mml:mo><mml:msup><mml:mi>Q</mml:mi><mml:mi>l</mml:mi></mml:msup><mml:mo stretchy='false'>(</mml:mo><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:mi>v</mml:mi><mml:mo stretchy='false'>)</mml:mo><mml:mo>&#x0003C;</mml:mo><mml:mi>&#x003C4;</mml:mi><mml:mtext>&#x000A0;</mml:mtext><mml:mo>&#x00026;</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:msubsup><mml:mi>E</mml:mi><mml:mi>C</mml:mi><mml:mi>l</mml:mi></mml:msubsup><mml:mo stretchy='false'>(</mml:mo><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:mi>v</mml:mi><mml:mo stretchy='false'>)</mml:mo><mml:mo>&#x02265;</mml:mo><mml:msubsup><mml:mi>E</mml:mi><mml:mi>M</mml:mi><mml:mi>l</mml:mi></mml:msubsup><mml:mo stretchy='false'>(</mml:mo><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:mi>v</mml:mi><mml:mo stretchy='false'>)</mml:mo><mml:mo>&#x0007D;</mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr columnalign='left'><mml:mtd columnalign='left'><mml:mrow><mml:mi>L</mml:mi><mml:msup><mml:mrow><mml:mo>&#x0007B;</mml:mo><mml:mi>M</mml:mi><mml:mo>&#x0007D;</mml:mo></mml:mrow><mml:mi>l</mml:mi></mml:msup><mml:mo stretchy='false'>(</mml:mo><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:mi>v</mml:mi><mml:mo stretchy='false'>)</mml:mo><mml:mo>,</mml:mo><mml:mtext>&#x02003;</mml:mtext><mml:mi>i</mml:mi><mml:mi>f</mml:mi><mml:mo>&#x0007B;</mml:mo><mml:msup><mml:mi>Q</mml:mi><mml:mi>l</mml:mi></mml:msup><mml:mo stretchy='false'>(</mml:mo><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:mi>v</mml:mi><mml:mo stretchy='false'>)</mml:mo><mml:mo>&#x0003C;</mml:mo><mml:mi>&#x003C4;</mml:mi><mml:mtext>&#x000A0;</mml:mtext><mml:mo>&#x00026;</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:msubsup><mml:mi>E</mml:mi><mml:mi>C</mml:mi><mml:mi>l</mml:mi></mml:msubsup><mml:mo stretchy='false'>(</mml:mo><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:mi>v</mml:mi><mml:mo stretchy='false'>)</mml:mo><mml:mo>&#x02265;</mml:mo><mml:msubsup><mml:mi>E</mml:mi><mml:mi>M</mml:mi><mml:mi>l</mml:mi></mml:msubsup><mml:mo stretchy='false'>(</mml:mo><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:mi>v</mml:mi><mml:mo stretchy='false'>)</mml:mo><mml:mo>&#x0007D;</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:mrow></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where the threshold &#x003C4; determines the corresponding fusion mode. <italic>Q</italic><sup><italic>l</italic></sup>(<italic>x, y</italic>) is given as follows:</p>
<disp-formula id="E12"><label>(11)</label><mml:math id="M20"><mml:mtable class="eqnarray" columnalign="right center left"><mml:mtr><mml:mtd><mml:msup><mml:mrow><mml:mi>Q</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi></mml:mrow></mml:msup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mi>v</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mn>2</mml:mn><mml:mstyle displaystyle="true"><mml:munder class="msub"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>p</mml:mi></mml:mrow></mml:munder></mml:mstyle><mml:mstyle displaystyle="true"><mml:munder class="msub"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>q</mml:mi></mml:mrow></mml:munder></mml:mstyle><mml:mi>L</mml:mi><mml:msup><mml:mrow><mml:mrow><mml:mo>{</mml:mo><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mo>}</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mi>l</mml:mi></mml:mrow></mml:msup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mi>p</mml:mi><mml:mo>,</mml:mo><mml:mi>v</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mi>q</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mi>L</mml:mi><mml:msup><mml:mrow><mml:mrow><mml:mo>{</mml:mo><mml:mrow><mml:mi>M</mml:mi></mml:mrow><mml:mo>}</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mi>l</mml:mi></mml:mrow></mml:msup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mi>p</mml:mi><mml:mo>,</mml:mo><mml:mi>v</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mi>q</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:msubsup><mml:mrow><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi></mml:mrow></mml:msubsup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mi>v</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x0002B;</mml:mo><mml:msubsup><mml:mrow><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:mi>M</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi></mml:mrow></mml:msubsup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mi>v</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:mfrac></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where <inline-formula><mml:math id="M21"><mml:msubsup><mml:mrow><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi></mml:mrow></mml:msubsup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mi>v</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula> and <inline-formula><mml:math id="M22"><mml:msubsup><mml:mrow><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:mi>M</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi></mml:mrow></mml:msubsup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mi>v</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula> are the local energy maps of <italic>L</italic>{<italic>C</italic>}<sup><italic>l</italic></sup> and <italic>L</italic>{<italic>M</italic>}<sup><italic>l</italic></sup>, respectively. <inline-formula><mml:math id="M23"><mml:msubsup><mml:mrow><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi></mml:mrow></mml:msubsup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mi>v</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula> and <inline-formula><mml:math id="M24"><mml:msubsup><mml:mrow><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:mi>M</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi></mml:mrow></mml:msubsup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mi>v</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula> are defined as follows:</p>
<disp-formula id="E13"><label>(12)</label><mml:math id="M25"><mml:mtable class="eqnarray" columnalign="right center left"><mml:mtr><mml:mtd><mml:mtable style="text-align:axis;" equalrows="false" columnlines="" equalcolumns="false" class="array"><mml:mtr><mml:mtd><mml:msubsup><mml:mrow><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi></mml:mrow></mml:msubsup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mi>v</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mstyle displaystyle="true"><mml:munder class="msub"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>p</mml:mi></mml:mrow></mml:munder></mml:mstyle><mml:mstyle displaystyle="true"><mml:munder class="msub"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>q</mml:mi></mml:mrow></mml:munder></mml:mstyle><mml:mi>L</mml:mi><mml:msup><mml:mrow><mml:mrow><mml:mo>{</mml:mo><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mo>}</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mi>l</mml:mi></mml:mrow></mml:msup><mml:msup><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mi>p</mml:mi><mml:mo>,</mml:mo><mml:mi>v</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mi>q</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:msubsup><mml:mrow><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:mi>M</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi></mml:mrow></mml:msubsup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mi>v</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mstyle displaystyle="true"><mml:munder class="msub"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>p</mml:mi></mml:mrow></mml:munder></mml:mstyle><mml:mstyle displaystyle="true"><mml:munder class="msub"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>q</mml:mi></mml:mrow></mml:munder></mml:mstyle><mml:mi>L</mml:mi><mml:msup><mml:mrow><mml:mrow><mml:mo>{</mml:mo><mml:mrow><mml:mi>M</mml:mi></mml:mrow><mml:mo>}</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mi>l</mml:mi></mml:mrow></mml:msup><mml:msup><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mi>p</mml:mi><mml:mo>,</mml:mo><mml:mi>v</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mi>q</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:mtd></mml:mtr></mml:mtable></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>Finally, the Laplacian pyramid reconstruction method (Mertens et al., <xref ref-type="bibr" rid="B33">2009</xref>) (bottle green color arrow in <xref ref-type="fig" rid="F1">Figure 1b</xref>) is used to reconstruct the fused low-rank components <italic>F</italic><sub><italic>lr</italic></sub> from <italic>L</italic>{<italic>F</italic>}<sup><italic>l</italic></sup>, as indicated in Equation (10).</p>
</sec>
</sec>
<sec>
<title>2.3. The Flowchart of the Proposed LatLRR-FCNs</title>
<p>The FCN architectures (FCN-32s or FCN-16s or FCN-8s) are inserted to produce two score maps with the focus property after the LatLRR decomposition once a pair of low-rank components for two images are calculated (hereafter, we named the proposed LatRR-FCNs: including proposed LatRR-FCN-32s, LatRR-FCN-16s and LatRR-FCN-8s, respectively). Algorithm 1 provides a pseudo-code of the proposed LatRR-FCN-32s, LatRR-FCN-16s and LatRR-FCN-8s networks. Then, ZCA and <italic>l</italic><sub>1</sub>-norm are utilized to white the score maps and obtain the initial weight maps for the low-rank components of paired source images [see <bold>Part 2-(ii)</bold> in Algorithm 1]. The WLE and WSEML techniques are used to fuse the two initial weight maps [see <bold>Part 2-(iii)</bold>]. The fused weight map and a pair of low-rank components under the pyramid-based image fusion framework (Mertens et al., <xref ref-type="bibr" rid="B33">2009</xref>) are used to reconstruct the fused low-rank components&#x00027; image <italic>F</italic><sub><italic>lr</italic></sub> [see <bold>Part 2-(iv)</bold>]. We sum the saliency components to obtain the fused saliency components&#x00027; image <italic>F</italic><sub><italic>ls</italic></sub> (see <bold>Part 3</bold>). Finally, the fused image <italic>F</italic> is obtained by combining <italic>F</italic><sub><italic>lr</italic></sub> and <italic>F</italic><sub><italic>ls</italic></sub> (see <bold>Part 4</bold>).</p>
<table-wrap position="float">
<label>Algorithm 1</label>
<caption><p>LatRR-FCN-32s, LatRR-FCN-16s, and LatRR-FCN-8s networks.</p></caption>
<table frame="hsides" rules="groups">
<tbody>
<tr>
<td valign="top" align="left"><inline-graphic xlink:href="fnins-14-615435-i0001.tif"/></td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
</sec>
<sec id="s3">
<title>3. Experimental Configurations</title>
<sec>
<title>3.1. FCN Training Sets</title>
<p>Currently, transfer learning (Bar et al., <xref ref-type="bibr" rid="B3">2015</xref>; Liu et al., <xref ref-type="bibr" rid="B25">2017</xref>; Razzak et al., <xref ref-type="bibr" rid="B35">2018</xref>; Lu et al., <xref ref-type="bibr" rid="B30">2019</xref>, <xref ref-type="bibr" rid="B31">2020</xref>) has become an active topic in the field of medical image analysis. In this study, we directly adopted a transfer learning strategy, and we trained the FCNs (FCN-32s, FCN-16s, and FCN-8s) on the PASCAL VOC 2012 dataset (Everingham et al., <xref ref-type="bibr" rid="B9">2012</xref>) and the semantic boundary dataset (SBD) (Hariharan et al., <xref ref-type="bibr" rid="B14">2011</xref>). The PASCAL VOC 2012 dataset contains 20 foreground object classes and 1 background class. The original dataset contains 1,464 (train), 1,449 (val), and 1,456 (test) pixel-level annotated images. The dataset is augmented with the SBD by extra annotations (Mertens et al., <xref ref-type="bibr" rid="B33">2009</xref>), resulting in 10,582 training images.</p>
</sec>
<sec>
<title>3.2. Source Medical Image Testing Sets</title>
<p>In our experiments, we used 40 pairs of multi-modal medical images (each medical image fusion problem contains 10 image pairs) to demonstrate the usefulness and efficiency of the proposed methods. Most of the test images were gathered from the Whole Brain Atlas databases (Vidoni, <xref ref-type="bibr" rid="B37">2012</xref>) and have been widely adopted in previous related publications (Liu and Wang, <xref ref-type="bibr" rid="B28">2014</xref>; Liu et al., <xref ref-type="bibr" rid="B25">2017</xref>, <xref ref-type="bibr" rid="B27">2019</xref>; Yin et al., <xref ref-type="bibr" rid="B45">2018</xref>; Zhu et al., <xref ref-type="bibr" rid="B49">2019</xref>). Each pair of images was geometrically aligned, and all the test images were normalized to 256 &#x000D7; 256.</p>
</sec>
<sec>
<title>3.3. State-of-the-Art Methods</title>
<p>Five superior medical image fusion methods were collected for comparison against our proposed methods. These included the adaptive sparse representation (ASR) method (Liu and Wang, <xref ref-type="bibr" rid="B28">2014</xref>) (<ext-link ext-link-type="uri" xlink:href="https://github.com/yuliu316316/MST-SR-Fusion-Toolbox">https://github.com/yuliu316316/MST-SR-Fusion-Toolbox</ext-link>), the convolutional neural network (CNN)-based (LP-CNN) method (Liu et al., <xref ref-type="bibr" rid="B25">2017</xref>) (<ext-link ext-link-type="uri" xlink:href="https://github.com/yuliu316316/CNN-Fusion">https://github.com/yuliu316316/CNN-Fusion</ext-link>), the phase congruency and local Laplacian energy-based NSCT (NSCT-PC-LLE) method (Zhu et al., <xref ref-type="bibr" rid="B49">2019</xref>) (<ext-link ext-link-type="uri" xlink:href="https://github.com/zhiqinzhu123/Source-code-of-medical-image-fusion-in-NSCT-domain">https://github.com/zhiqinzhu123/Source-code-of-medical-image-fusion-in-NSCT-domain</ext-link>), the parameter-adaptive pulse coupled-neural network (NSST-PAPCNN) in the NSST domain method (Yin et al., <xref ref-type="bibr" rid="B45">2018</xref>) (<ext-link ext-link-type="uri" xlink:href="https://github.com/yuliu316316/NSST-PAPCNN-Fusion">https://github.com/yuliu316316/NSST-PAPCNN-Fusion</ext-link>), and the convolutional sparsity-based morphological component analysis (CSMCA) method (Liu et al., <xref ref-type="bibr" rid="B27">2019</xref>) (<ext-link ext-link-type="uri" xlink:href="https://github.com/yuliu316316/CSMCA-Fusion">https://github.com/yuliu316316/CSMCA-Fusion</ext-link>). Among them, the NSCT-PC-LLE, NSST-PAPCNN, and CSMCA methods were proposed in last year.</p>
</sec>
<sec>
<title>3.4. Parameter Choices</title>
<p>The parameters of all compared methods were set to the default values. The key parameters for our proposed algorithms were given in <xref ref-type="table" rid="T1">Table 1</xref>. According to this table, the parameter &#x003BB; in LatLRR decomposition was 0.8 (Li and Wu, <xref ref-type="bibr" rid="B21">2018</xref>), and the threshold &#x003C4; in Equation (10) was set to 0.8 (Liu et al., <xref ref-type="bibr" rid="B25">2017</xref>). The PASCAL VOC 2012 dataset contained 20 foreground object classes and one background class, so that the <italic>C</italic> in <inline-formula><mml:math id="M27"><mml:msup><mml:mrow><mml:msub><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>:</mml:mo><mml:mi>C</mml:mi></mml:mrow></mml:msup></mml:math></inline-formula> was equal to 21. Note that we adopted a transfer learning strategy directly to train the FCN-VGG16 (Long et al., <xref ref-type="bibr" rid="B29">2015</xref>) by MacInnes, and the trained models were obtained after 50 epochs using the training data. The choice of epoch was dependent on <xref ref-type="fig" rid="F5">Figure 5</xref>. When the epoch was lower than 50, the accuracy of the training and validation sets increased with the values of epochs. However, the accuracy of the validation set leveled off when the epoch was higher than 50, although the accuracy of the training set still increased regardless of the scenario (FCN-32s, FCN-16s, and FCN-8s). In terms of the loss function, the values for all FCN architectures decreased with the epoch in the case of the training set, but the loss of the scenarios tended to converge at the 50 epochs. Therefore, to balance the computational complexity and accuracy, the epochs for FCN models in this paper were chosen as 50.</p>
<table-wrap position="float" id="T1">
<label>Table 1</label>
<caption><p>The key parameters used in our algorithms.</p></caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th valign="top" align="left"><bold>LatLRR decomposition</bold></th>
<th valign="top" align="center"><bold>Threshold</bold></th>
<th valign="top" align="center"><bold>Classes</bold></th>
<th valign="top" align="center"><bold>No. of epochs</bold></th>
</tr>
<tr>
<th valign="top" align="left"><bold>&#x003BB;</bold></th>
<th valign="top" align="center"><bold>&#x003C4;</bold></th>
<th valign="top" align="center"><bold>C</bold></th>
<th valign="top" align="center"><bold><italic>epoch</italic></bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">0.8</td>
<td valign="top" align="center">0.8</td>
<td valign="top" align="center">21</td>
<td valign="top" align="center">50</td>
</tr>
</tbody>
</table>
</table-wrap>
<fig id="F5" position="float">
<label>Figure 5</label>
<caption><p>The process of training FCN models. <bold>(A)</bold> FNC-32s model, <bold>(B)</bold> FNC-16s model, and <bold>(C)</bold> FNC-8s model.</p></caption>
<graphic xlink:href="fnins-14-615435-g0005.tif"/>
</fig>
</sec>
<sec>
<title>3.5. Experimental Environment</title>
<p>All the experiments were implemented in MATLAB R2019a on a WIN64 Intel(R) Core (TM)i7-8750H CPU&#x00040;2.20GHz 8GB RAM. The training models of the proposed method were trained in MATLAB R2019a&#x0002B;VS2017&#x0002B; MatConvNet 1.0-beta25.</p>
</sec>
<sec>
<title>3.6. Objective Evaluation Metrices</title>
<p>In this study, five common representative quantitative metrics, e.g., <italic>EN</italic> (Liu and Wang, <xref ref-type="bibr" rid="B28">2014</xref>), <italic>Q</italic><sub><italic>MI</italic></sub> (Bhatnagar et al., <xref ref-type="bibr" rid="B4">2013</xref>), <italic>Q</italic><sup><italic>AB</italic>/<italic>F</italic></sup> (Xydeas et al., <xref ref-type="bibr" rid="B44">2000</xref>), <italic>SCD</italic> (Aslantas and Bendes, <xref ref-type="bibr" rid="B2">2015</xref>), and <italic>VIFF</italic>(Han et al., <xref ref-type="bibr" rid="B13">2013</xref>) (for all metrics, a larger value indicates a better performance), were used to evaluate the quality of fused images. The metrics were briefly described as follows:</p>
<p><bold>(i) Entropy (EN)</bold> Liu and Wang (<xref ref-type="bibr" rid="B28">2014</xref>), Wang et al. (<xref ref-type="bibr" rid="B43">2017</xref>), Zhang et al. (<xref ref-type="bibr" rid="B47">2017</xref>): Entropy measures the amount of information in the fused image.</p>
<p><bold>(ii) Mutual information (MI) of two images <italic>Q</italic><sub><italic>MI</italic></sub></bold> Bhatnagar et al. (<xref ref-type="bibr" rid="B4">2013</xref>): MI is a quantitative assessment of the information shared by two images. Mathematically, MI can be expressed with joint entropy <italic>H</italic>(<italic>C, D</italic>), marginal entropy <italic>H</italic>(<italic>C</italic>), and <italic>H</italic>(<italic>D</italic>) of two variables <italic>C</italic> and <italic>D</italic> as follows:</p>
<disp-formula id="E14"><label>(13)</label><mml:math id="M28"><mml:mtable class="eqnarray" columnalign="right center left"><mml:mtr><mml:mtd><mml:mi>M</mml:mi><mml:mi>I</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>C</mml:mi><mml:mo>,</mml:mo><mml:mi>D</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mi>H</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x0002B;</mml:mo><mml:mi>H</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>-</mml:mo><mml:mi>H</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>C</mml:mi><mml:mo>,</mml:mo><mml:mi>D</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where <inline-formula><mml:math id="M29"><mml:mi>H</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mo>-</mml:mo><mml:munder><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>c</mml:mi></mml:mrow></mml:munder><mml:mi>p</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>c</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:msub><mml:mrow><mml:mo class="qopname">log</mml:mo></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mi>p</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>c</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:mi>H</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mo>-</mml:mo><mml:munder><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>d</mml:mi></mml:mrow></mml:munder><mml:mi>p</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>d</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:msub><mml:mrow><mml:mo class="qopname">log</mml:mo></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mi>p</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>d</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:mi>H</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>C</mml:mi><mml:mo>,</mml:mo><mml:mi>D</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mo>-</mml:mo><mml:munder><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mo>,</mml:mo><mml:mi>d</mml:mi></mml:mrow></mml:munder><mml:mi>p</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>c</mml:mi><mml:mo>,</mml:mo><mml:mi>d</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:msub><mml:mrow><mml:mo class="qopname">log</mml:mo></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mi>p</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>c</mml:mi><mml:mo>,</mml:mo><mml:mi>d</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula>. <italic>p</italic>(<italic>c</italic>) and <italic>p</italic>(<italic>d</italic>) denote the marginal probability distributions of <italic>C</italic> and <italic>D</italic>, respectively. <italic>p</italic>(<italic>c, d</italic>) denotes the joint probability distribution of <italic>C</italic> and <italic>D</italic>. Therefore, the quality of the fused image with respect to input images <italic>Img</italic><sub>1</sub> and <italic>Img</italic><sub>2</sub> can be defined as:</p>
<disp-formula id="E15"><label>(14)</label><mml:math id="M30"><mml:mtable class="eqnarray" columnalign="right center left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>Q</mml:mi></mml:mrow><mml:mrow><mml:mi>M</mml:mi><mml:mi>I</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>2</mml:mn><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mfrac><mml:mrow><mml:mi>M</mml:mi><mml:mi>I</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>I</mml:mi><mml:mi>m</mml:mi><mml:msub><mml:mrow><mml:mi>g</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:mi>F</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mi>H</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>I</mml:mi><mml:mi>m</mml:mi><mml:msub><mml:mrow><mml:mi>g</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x0002B;</mml:mo><mml:mi>H</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>F</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:mfrac><mml:mo>&#x0002B;</mml:mo><mml:mfrac><mml:mrow><mml:mi>M</mml:mi><mml:mi>I</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>I</mml:mi><mml:mi>m</mml:mi><mml:msub><mml:mrow><mml:mi>g</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:mi>F</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mi>H</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>I</mml:mi><mml:mi>m</mml:mi><mml:msub><mml:mrow><mml:mi>g</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x0002B;</mml:mo><mml:mi>H</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>F</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:mfrac></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p><bold>(iii) Edge-based similarity measure <italic>Q</italic><sup><italic>AB</italic>/<italic>F</italic></sup></bold>: The authors in Xydeas et al. (<xref ref-type="bibr" rid="B44">2000</xref>) proposed a metric <italic>Q</italic><sup><italic>AB</italic>/<italic>F</italic></sup> to produce the similarity between the edges that transform in the fusion process. This metric is defined as follows:</p>
<disp-formula id="E16"><label>(15)</label><mml:math id="M31"><mml:mtable class="eqnarray" columnalign="right center left"><mml:mtr><mml:mtd><mml:msup><mml:mrow><mml:mi>Q</mml:mi></mml:mrow><mml:mrow><mml:mi>A</mml:mi><mml:mi>B</mml:mi><mml:mo>/</mml:mo><mml:mi>F</mml:mi></mml:mrow></mml:msup><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mstyle displaystyle="true"><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>u</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:munderover></mml:mstyle><mml:mstyle displaystyle="true"><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>v</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>M</mml:mi></mml:mrow></mml:munderover></mml:mstyle><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msup><mml:mrow><mml:mi>Q</mml:mi></mml:mrow><mml:mrow><mml:mi>A</mml:mi><mml:mi>F</mml:mi></mml:mrow></mml:msup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mi>v</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:msup><mml:mrow><mml:mi>w</mml:mi></mml:mrow><mml:mrow><mml:mi>A</mml:mi></mml:mrow></mml:msup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mi>v</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x0002B;</mml:mo><mml:msup><mml:mrow><mml:mi>Q</mml:mi></mml:mrow><mml:mrow><mml:mi>B</mml:mi><mml:mi>F</mml:mi></mml:mrow></mml:msup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mi>v</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:msup><mml:mrow><mml:mi>w</mml:mi></mml:mrow><mml:mrow><mml:mi>B</mml:mi></mml:mrow></mml:msup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mi>v</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mstyle displaystyle="true"><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>u</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:munderover></mml:mstyle><mml:mstyle displaystyle="true"><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>v</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>M</mml:mi></mml:mrow></mml:munderover></mml:mstyle><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msup><mml:mrow><mml:mi>w</mml:mi></mml:mrow><mml:mrow><mml:mi>A</mml:mi></mml:mrow></mml:msup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mi>v</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x0002B;</mml:mo><mml:msup><mml:mrow><mml:mi>w</mml:mi></mml:mrow><mml:mrow><mml:mi>B</mml:mi></mml:mrow></mml:msup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mi>v</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:mfrac></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where <italic>A, B</italic>, and <italic>F</italic> represent the two input images (<italic>Img</italic><sub>1</sub> and <italic>Img</italic><sub>2</sub>) and fused images. The size of each image is <italic>N</italic> &#x000D7; <italic>M</italic>, <italic>Q</italic><sup><italic>AF</italic></sup>(<italic>u, v</italic>) and <italic>Q</italic><sup><italic>BF</italic></sup>(<italic>u, v</italic>) are defined as follows:</p>
<disp-formula id="E17"><label>(16)</label><mml:math id="M32"><mml:mtable class="eqnarray" columnalign="right center left"><mml:mtr><mml:mtd><mml:mtable style="text-align:axis;" equalrows="false" columnlines="" equalcolumns="false" class="array"><mml:mtr><mml:mtd><mml:msup><mml:mrow><mml:mi>Q</mml:mi></mml:mrow><mml:mrow><mml:mi>A</mml:mi><mml:mi>F</mml:mi></mml:mrow></mml:msup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mi>v</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:msubsup><mml:mrow><mml:mi>Q</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow><mml:mrow><mml:mi>A</mml:mi><mml:mi>F</mml:mi></mml:mrow></mml:msubsup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mi>v</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:msubsup><mml:mrow><mml:mi>Q</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B1;</mml:mi></mml:mrow><mml:mrow><mml:mi>A</mml:mi><mml:mi>F</mml:mi></mml:mrow></mml:msubsup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mi>v</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:msup><mml:mrow><mml:mi>Q</mml:mi></mml:mrow><mml:mrow><mml:mi>B</mml:mi><mml:mi>F</mml:mi></mml:mrow></mml:msup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mi>v</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:msubsup><mml:mrow><mml:mi>Q</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow><mml:mrow><mml:mi>B</mml:mi><mml:mi>F</mml:mi></mml:mrow></mml:msubsup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mi>v</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:msubsup><mml:mrow><mml:mi>Q</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B1;</mml:mi></mml:mrow><mml:mrow><mml:mi>B</mml:mi><mml:mi>F</mml:mi></mml:mrow></mml:msubsup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mi>v</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where <inline-formula><mml:math id="M33"><mml:msubsup><mml:mrow><mml:mi>Q</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x0002A;</mml:mo><mml:mi>F</mml:mi></mml:mrow></mml:msubsup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mi>v</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula> and <inline-formula><mml:math id="M34"><mml:msubsup><mml:mrow><mml:mi>Q</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B1;</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x0002A;</mml:mo><mml:mi>F</mml:mi></mml:mrow></mml:msubsup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mi>v</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula> are the edge strength and orient preservation values at location (<italic>u, v</italic>) in images <italic>A</italic> and <italic>B</italic>, respectively. The dynamic range for <italic>Q</italic><sup><italic>AB</italic>/<italic>F</italic></sup> is equal to [0, 1], where a larger value for <italic>Q</italic><sup><italic>AB</italic>/<italic>F</italic></sup> indicates a better fusion result. For more details of this metric, please refer to Xydeas et al. (<xref ref-type="bibr" rid="B44">2000</xref>).</p>
<p><bold>(iv) The sum of the correlations of differences (SCD)</bold> Aslantas and Bendes (<xref ref-type="bibr" rid="B2">2015</xref>) is a quality metric formulated as follows:</p>
<disp-formula id="E18"><label>(17)</label><mml:math id="M35"><mml:mtable class="eqnarray" columnalign="right center left"><mml:mtr><mml:mtd><mml:mi>S</mml:mi><mml:mi>C</mml:mi><mml:mi>D</mml:mi><mml:mo>=</mml:mo><mml:mi>r</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:mi>I</mml:mi><mml:mi>m</mml:mi><mml:msub><mml:mrow><mml:mi>g</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x0002B;</mml:mo><mml:mi>r</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:mi>I</mml:mi><mml:mi>m</mml:mi><mml:msub><mml:mrow><mml:mi>g</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where <italic>D</italic><sub>1</sub> &#x0003D; <italic>F</italic> &#x02212; <italic>Img</italic><sub>2</sub>, <italic>D</italic><sub>2</sub> &#x0003D; <italic>F</italic> &#x02212; <italic>Img</italic><sub>1</sub>, <italic>F</italic> is the fused image, and <italic>Img</italic><sub>1</sub> and <italic>Img</italic><sub>2</sub> are the input images. The <italic>r</italic>(.) function calculates the correlation between <italic>S</italic><sub><italic>k</italic></sub> and <italic>D</italic><sub><italic>k</italic></sub>, given as:</p>
<disp-formula id="E19"><label>(18)</label><mml:math id="M36"><mml:mtable class="eqnarray" columnalign="right center left"><mml:mtr><mml:mtd><mml:mi>r</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:mi>I</mml:mi><mml:mi>m</mml:mi><mml:msub><mml:mrow><mml:mi>g</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mstyle displaystyle="true"><mml:munder><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>u</mml:mi></mml:mrow></mml:munder></mml:mstyle><mml:mstyle displaystyle="true"><mml:munder><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>v</mml:mi></mml:mrow></mml:munder></mml:mstyle><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mi>v</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mo>&#x00304;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>I</mml:mi><mml:mi>m</mml:mi><mml:msub><mml:mrow><mml:mi>g</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mi>v</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>I</mml:mi><mml:mi>m</mml:mi><mml:mi>g</mml:mi></mml:mrow><mml:mo>&#x00304;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:msqrt><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mstyle displaystyle="true"><mml:munder><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>u</mml:mi></mml:mrow></mml:munder></mml:mstyle><mml:mstyle displaystyle="true"><mml:munder><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>v</mml:mi></mml:mrow></mml:munder></mml:mstyle><mml:mi>j</mml:mi><mml:msup><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mi>v</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mo>&#x00304;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mstyle displaystyle="true"><mml:munder><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>u</mml:mi></mml:mrow></mml:munder></mml:mstyle><mml:mstyle displaystyle="true"><mml:munder><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>v</mml:mi></mml:mrow></mml:munder></mml:mstyle><mml:msup><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>I</mml:mi><mml:mi>m</mml:mi><mml:msub><mml:mrow><mml:mi>g</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi><mml:mo>,</mml:mo><mml:mi>v</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>I</mml:mi><mml:mi>m</mml:mi><mml:mi>g</mml:mi></mml:mrow><mml:mo>&#x00304;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:mrow></mml:msqrt></mml:mrow></mml:mfrac></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where <italic>k</italic> &#x0003D; 1, 2, <inline-formula><mml:math id="M38"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mo>&#x00304;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> and <inline-formula><mml:math id="M39"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>I</mml:mi><mml:mi>m</mml:mi><mml:mi>g</mml:mi></mml:mrow><mml:mo>&#x00304;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> are the average of the pixel values of <italic>D</italic><sub><italic>k</italic></sub> and <italic>Img</italic><sub><italic>k</italic></sub>, respectively.</p>
<p><bold>(v) The human visual perception-based metric visual information fidelity fusion (VIFF)</bold> Han et al. (<xref ref-type="bibr" rid="B13">2013</xref>): To obtain the <italic>VIFF</italic>, four steps are needed. First, the source and fused images are filtered and then divided into blocks. Second, visual information is evaluated with and without distortion information in each block. Third, the <italic>VIFF</italic> of each sub-band is calculated. Finally, the overall quality measure is determined by weighting the <italic>VIFF</italic> of each sub-band.</p>
</sec>
<sec>
<title>3.7. Color Space Fusion</title>
<p>In our proposed methods, the YUV color space was used to solve the grayscale and RGB color image (PET, SPECT) fusion issues. First, the RGB color image was converted into a YUV color space, resulting in three channel components of Y, U, and V. Then, the grayscale image and the Y channel were fused by using the proposed fusion methods, as described in section 2. Finally, the fused Y-channel component, the U-channel component, and the V-channel component were inversely transformed by YUV space, obtaining the fused color image.</p>
</sec>
</sec>
<sec id="s4">
<title>4. Results and Discussion</title>
<p>This section is devoted to showing that the proposed LatRR-FCNs can improve the information details and energy preservation in terms of <bold>visual quality assessment</bold> (section 4.1), <bold>quantitative assessment</bold> (section 4.2) and <bold>computational cost assessment</bold> (section 4.3), compared with five recently proposed methods: ASR (Liu and Wang, <xref ref-type="bibr" rid="B28">2014</xref>), LP-CNN (Liu et al., <xref ref-type="bibr" rid="B25">2017</xref>), NSCT-PC-LLE (Zhu et al., <xref ref-type="bibr" rid="B49">2019</xref>), NSST-PAPCNN (Yin et al., <xref ref-type="bibr" rid="B45">2018</xref>), and CSMCA (Liu et al., <xref ref-type="bibr" rid="B27">2019</xref>). In this study, the usefulness and efficiency of each method are investigated with four sets of medical image fusion studies, including CT and MR, MR-T1 and MR-T2, PET and MR, and SPECT and MR.</p>
<sec>
<title>4.1. Visual Quality</title>
<p>The fusion examples of CT and MR images are given in <xref ref-type="fig" rid="F6">Figure 6</xref>. Furthermore, one representative region of each result is enlarged for better comparison. The ASR and CSMCA methods reveal a significant energy loss in both the CT and MR images (resulting in an intensity and contrast decrease in the fused images), especially for the bone and lesion regions in the <xref ref-type="fig" rid="F6">Figures 6a3&#x02013;c3,a7&#x02013;c7</xref>. The fusion results of the NSCT-PC-LLE, LP-CNN, NSST-PAPCNN, and the proposed methods have better information preservation for the CT and MR modalities. However, the NSCT-PC-LLE, LP-CNN, and NSST-PAPCNN methods cannot extract the detailed information well in the MR image, which can be seen in the <xref ref-type="fig" rid="F6">Figures 6a4&#x02013;c4,a5&#x02013;c5,a6&#x02013;c6</xref> and the corresponding highlighted close-ups. Furthermore, the ASR method fails to extract the structural and edge details from the CT modality (see <xref ref-type="fig" rid="F6">Figures 6a4&#x02013;c4</xref>). The NSCT-PC-LLE and NSST-PAPCNN methods outperform the ASR method, even though some structural details cannot be extracted (see the <xref ref-type="fig" rid="F6">Figures 6a3&#x02013;c3,a4&#x02013;c4,a6&#x02013;c6</xref>). The proposed frameworks and LP-CNN method can effectively extract the structure and edge details from both CT and MR modalities (see <xref ref-type="fig" rid="F6">Figures 6a8&#x02013;a10,b8&#x02013;b10,c8&#x02013;c10,a5&#x02013;c5</xref>, respectively). The proposed methods perform well on the preservation of detailed and structural information for all three examples.</p>
<fig id="F6" position="float">
<label>Figure 6</label>
<caption><p>Three fusion examples of CT and MR images. One close-up is provided in each set for better comparison. The original images: <bold>(a1&#x02013;c1)</bold> CT and <bold>(a2&#x02013;c2)</bold> MR. The fusion results <bold>(a3&#x02013;c3)</bold> ASR, <bold>(a4&#x02013;c4)</bold> NSCT-PC-LLE, <bold>(a5&#x02013;c5)</bold> LP-CNN, <bold>(a6&#x02013;c6)</bold> NSST-PAPCNN, <bold>(a7&#x02013;c7)</bold> CSMCA, <bold>(a8&#x02013;c8)</bold> LatLRR-FCN-32s, <bold>(a9&#x02013;c9)</bold> LatLRR-FCN-16s, and <bold>(a10&#x02013;c10)</bold> LatLRR-FCN-8s.</p></caption>
<graphic xlink:href="fnins-14-615435-g0006.tif"/>
</fig>
<p><xref ref-type="fig" rid="F7">Figure 7</xref> gives three fusion examples of MR-T1 and MR-T2 images. The ASR and CSMCA methods suffer from low intensity and contrast caused by the loss of energy (see the <xref ref-type="fig" rid="F7">Figures 7a3&#x02013;c3,a7&#x02013;c7</xref> with the close-up). In addition, the NSCT-PC-LLE, LP-CNN, and NSST-PAPCNN methods cannot preserve the detailed information (see the close-ups in <xref ref-type="fig" rid="F7">Figures 7a4&#x02013;c4,a5&#x02013;c5,a6&#x02013;c6</xref>, respectively). Furthermore, the ASR and NSCT-PC-LLE methods exhibit lower ability in structure and edge detail extraction within the MR-T1 modality, explained by the close-up in <xref ref-type="fig" rid="F7">Figures 7a4&#x02013;c4,a5&#x02013;c5</xref>. Finally, compared to the other tested methods, our proposed LatLRR-FCN-based methods achieve the best performance, as shown with the close-ups in <xref ref-type="fig" rid="F7">Figures 7a8&#x02013;c8,a9&#x02013;c9,a10&#x02013;c10</xref>, respectively.</p>
<fig id="F7" position="float">
<label>Figure 7</label>
<caption><p>Three fusion examples of MR-T1 and MR-T2 images. One close-up is provided in each set for better comparison. The original images: <bold>(a1&#x02013;c1)</bold> MR-T1 and <bold>(a2&#x02013;c2)</bold> MR-T2. The fusion results are as follows: <bold>(a3&#x02013;c3)</bold> ASR, <bold>(a4&#x02013;c4)</bold> NSCT-PC-LLE, <bold>(a5&#x02013;c5)</bold> LP-CNN, <bold>(a6&#x02013;c6)</bold> NSST-PAPCNN, <bold>(a7&#x02013;c7)</bold> CSMCA, <bold>(A8&#x02013;C8)</bold> LatLRR-FCN-32s, <bold>(a9&#x02013;c9)</bold> LatLRR-FCN-16s, and <bold>(a10&#x02013;c10)</bold> LatLRR-FCN-8s.</p></caption>
<graphic xlink:href="fnins-14-615435-g0007.tif"/>
</fig>
<p><xref ref-type="fig" rid="F8">Figure 8</xref> shows the three fusion examples of MR and PET images. The ASR and CSMCA methods lose a significant amount of energy in both the MR and PET modalities, as viewed in the <xref ref-type="fig" rid="F8">Figures 8a3&#x02013;c3,a7&#x02013;c7</xref> and the corresponding close-ups. Note that the NSCT-PC-LLE and LP-CNN methods are subjected to a severe color distortion (see the close-ups in <xref ref-type="fig" rid="F8">Figures 8a4&#x02013;c4,a5&#x02013;c5</xref>). Furthermore, the color distortion existed more or less in the fusion results of the NSST-PAPCNN method (see <xref ref-type="fig" rid="F8">Figures 8a7&#x02013;c7</xref> and the close-ups). Overall, the color preservation of our proposed algorithms (see <xref ref-type="fig" rid="F8">Figures 8a8&#x02013;c8,a9&#x02013;c9,a10&#x02013;c10</xref> together with their close-ups) are also significantly higher than the other methods.</p>
<fig id="F8" position="float">
<label>Figure 8</label>
<caption><p>Three fusion examples of MR and PET images. One close-up is provided in each set for better comparison. The original images: <bold>(a1&#x02013;c1)</bold> MR and <bold>(a2&#x02013;c2)</bold> PET. The fusion results: <bold>(a3&#x02013;c3)</bold> ASR, <bold>(a4&#x02013;c4)</bold> NSCT-PC-LLE, <bold>(a5&#x02013;c5)</bold> LP-CNN, <bold>(a6&#x02013;c6)</bold> NSST-PAPCNN, <bold>(a7&#x02013;c7)</bold> CSMCA, <bold>(a8&#x02013;c8)</bold> LatLRR-FCN-32s, <bold>(a9&#x02013;c9)</bold> LatLRR-FCN-16s, and <bold>(a10&#x02013;c10)</bold> LatLRR-FCN-8s.</p></caption>
<graphic xlink:href="fnins-14-615435-g0008.tif"/>
</fig>
<p>The fusion examples of three sets of MR and SPECT images are shown in <xref ref-type="fig" rid="F9">Figure 9</xref>. The ASR and CSMCA methods still lose much energy in both the MR and PET modalities (see <xref ref-type="fig" rid="F9">Figures 9a3&#x02013;c3,a7&#x02013;c7</xref>). Moreover, color distortion exists in the NSCT-PC-LLE and LP-CNN methods (see the close-up in <xref ref-type="fig" rid="F9">Figures 9a4&#x02013;c4,a5&#x02013;c5</xref>). Furthermore, in the results of the NSST-PAPCNN method, color distortion also exists (in <xref ref-type="fig" rid="F9">Figures 9a6&#x02013;c6</xref>, especially the close-up). The visual quality of color preservation of our proposed methods significantly outperforms the others.</p>
<fig id="F9" position="float">
<label>Figure 9</label>
<caption><p>Three fusion examples of MR and SPECT images. One close-up is provided in each set for better comparison. The original images: <bold>(a1&#x02013;c1)</bold> MR and <bold>(a2&#x02013;c2)</bold> SPECT. The fusion results: <bold>(a3&#x02013;c3)</bold> ASR, <bold>(a4&#x02013;c4)</bold> NSCT-PC-LLE, <bold>(a5&#x02013;c5)</bold> LP-CNN, <bold>(a6&#x02013;c6)</bold> NSST-PAPCNN, <bold>(a7&#x02013;c7)</bold> CSMCA, <bold>(a8&#x02013;c8)</bold> LatLRR-FCN-32s, <bold>(a9&#x02013;c9)</bold> LatLRR-FCN-16s, and <bold>(a10&#x02013;c10)</bold> LatLRR-FCN-8s.</p></caption>
<graphic xlink:href="fnins-14-615435-g0009.tif"/>
</fig>
</sec>
<sec>
<title>4.2. Quantitative Assessment</title>
<p>Here, five common quantitative metrics as described in section 3.6 are employed to appraise the fusion performance. The average score of each method in each fusion problem is reported in <xref ref-type="table" rid="T2">Table 2</xref>. The top three values of all the fusion methods are shown in bold, and their rank is indicated by a superscript. For CT and MRI fusion, the proposed methods achieve the best performance in terms of <italic>EN</italic> (i.e., the values of <italic>EN</italic> for LatLRR-FCN-8s, LatLRR-FCN-16s, and LatLRR-FCN-32s are equal to <bold>5.4531<sup>1</sup></bold>, <bold>5.4037<sup>2</sup></bold> and <bold>5.3915<sup>3</sup></bold>, respectively), <italic>Q</italic><sub><italic>MI</italic></sub> (the value of <italic>Q</italic><sub><italic>MI</italic></sub> for LatLRR-FCN-32s, <bold>1.2072<sup>1</sup></bold>, higher than that of LatLRR-FCN-16s, <bold>1.2065<sup>2</sup></bold>, and LatLRR-FCN-8s, <bold>1.2043<sup>3</sup></bold>), <italic>SCD</italic>, and <italic>VIFF</italic> metrics. Note that in the context of the <italic>Q</italic><sup><italic>AB</italic>/<italic>F</italic></sup> metric, our proposed LatLRR-FCN-32s (<italic>Q</italic><sup><italic>AB</italic>/<italic>F</italic></sup> &#x0003D; <bold>0.5541<sup>2</sup></bold>) is slightly lower than the LP-CNN method (<italic>Q</italic><sup><italic>AB</italic>/<italic>F</italic></sup> &#x0003D; <bold>0.5688<sup>1</sup></bold>) but slightly superior to the CSMCA method (<italic>Q</italic><sup><italic>AB</italic>/<italic>F</italic></sup> &#x0003D; <bold>0.5535<sup>3</sup></bold>). In the case of MR-T1 and MR-T2 fusion, the proposed methods show the best values in three of the five metrics with <italic>EN</italic>, <italic>SCD</italic> and <italic>VIFF</italic>. Among them, an increase improvement in the proposed LatLRR-FCN-32s for <italic>SCD</italic> about <bold>10.23%</bold> [(1.7483 &#x02212; 1.5694)/1.7483 &#x0003D; 0.1023] is reported in <xref ref-type="table" rid="T2">Table 2</xref>, compared to the best performance among the other five methods, i.e., NSST-PAPCNN algorithm. For MRI and PET fusion, overall, the proposed LatRR-FCNs obtain the best results in all five objective metrics except that the NSST-PAPCNN method achieves the rank first in the metric <italic>Q</italic><sup><italic>AB</italic>/<italic>F</italic></sup> (i.e., <italic>Q</italic><sup><italic>AB</italic>/<italic>F</italic></sup> &#x0003D; <bold>1.0946<sup>1</sup></bold>) with a slight improvement <bold>1.16%</bold> (e.g., 0.0116 &#x0003D; (1.0946 &#x02212; 1.0819)/1.0946) compared with our proposed LatLRR-FCN-8s. Finally, our proposed LatRR-FCNs outperform the other fusion methods in the aspect of <italic>EN</italic>, <italic>SCD</italic>, and <italic>VIFF</italic> metrics for MRI and SPECT fusion, especially for the <italic>SCD</italic> metric of LatRR-FCN-8s with a significant improvement in <bold>17.22%</bold> (0.1722 &#x0003D; (1.8265&#x02212;1.5119)/1.8265) compared to that of the NSST-PAPCNN approach.</p>
<table-wrap position="float" id="T2">
<label>Table 2</label>
<caption><p>Five metrics of different methods for the four medical image fusion problems (a higher value for each metric indicates a better performance).</p></caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th valign="top" align="left"><bold>CT and MR</bold></th>
<th valign="top" align="center"><bold><italic>EN</italic></bold></th>
<th valign="top" align="center"><bold><italic>Q</italic><sub><italic>MI</italic></sub></bold></th>
<th valign="top" align="center"><bold><italic>Q</italic><sup><italic>AB</italic>/<italic>F</italic></sup></bold></th>
<th valign="top" align="center"><bold><italic>SCD</italic></bold></th>
<th valign="top" align="center"><bold><italic>VIFF</italic></bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">ASR</td>
<td valign="top" align="center">4.7319</td>
<td valign="top" align="center">1.1068</td>
<td valign="top" align="center">0.5378</td>
<td valign="top" align="center">1.5400</td>
<td valign="top" align="center">0.3588</td>
</tr>
<tr>
<td valign="top" align="left">NSCT-PC-LLE</td>
<td valign="top" align="center">5.1619</td>
<td valign="top" align="center">1.1930</td>
<td valign="top" align="center">0.5206</td>
<td valign="top" align="center">1.5786</td>
<td valign="top" align="center">0.4570</td>
</tr>
<tr>
<td valign="top" align="left">LP-CNN</td>
<td valign="top" align="center">4.9503</td>
<td valign="top" align="center">1.1674</td>
<td valign="top" align="center"><bold>0.5688<sup>1</sup></bold></td>
<td valign="top" align="center">1.5715</td>
<td valign="top" align="center">0.4476</td>
</tr>
<tr>
<td valign="top" align="left">NSST-PAPCNN</td>
<td valign="top" align="center">5.2323</td>
<td valign="top" align="center">1.1687</td>
<td valign="top" align="center">0.5383</td>
<td valign="top" align="center">1.6259</td>
<td valign="top" align="center">0.4703</td>
</tr>
<tr>
<td valign="top" align="left">CSMCA</td>
<td valign="top" align="center">4.8532</td>
<td valign="top" align="center">1.1120</td>
<td valign="top" align="center"><bold>0.5535<sup>3</sup></bold></td>
<td valign="top" align="center">1.5462</td>
<td valign="top" align="center">0.4135</td>
</tr>
<tr>
<td valign="top" align="left">LatLRR-FCN-32s</td>
<td valign="top" align="center"><bold>5.3915<sup>3</sup></bold></td>
<td valign="top" align="center"><bold>1.2072<sup>1</sup></bold></td>
<td valign="top" align="center"><bold>0.5541<sup>2</sup></bold></td>
<td valign="top" align="center"><bold>1.6946<sup>3</sup></bold></td>
<td valign="top" align="center"><bold>0.4998<sup>3</sup></bold></td>
</tr>
<tr>
<td valign="top" align="left">LatLRR-FCN-16s</td>
<td valign="top" align="center"><bold>5.4037<sup>2</sup></bold></td>
<td valign="top" align="center"><bold>1.2065<sup>2</sup></bold></td>
<td valign="top" align="center">0.5531</td>
<td valign="top" align="center"><bold>1.6986<sup>2</sup></bold></td>
<td valign="top" align="center"><bold>0.5003<sup>2</sup></bold></td>
</tr>
<tr>
<td valign="top" align="left">LatLRR-FCN-8s</td>
<td valign="top" align="center"><bold>5.4531<sup>1</sup></bold></td>
<td valign="top" align="center"><bold>1.2043<sup>3</sup></bold></td>
<td valign="top" align="center">0.5517</td>
<td valign="top" align="center"><bold>1.7034<sup>1</sup></bold></td>
<td valign="top" align="center"><bold>0.5007<sup>1</sup></bold></td>
</tr>
<tr style="border-top: thin solid #000000;border-bottom: thin solid #000000;">
<td valign="top" align="left"><bold>MR-T1 and MR-T2</bold></td>
<td valign="top" align="center"><italic><bold>EN</bold></italic></td>
<td valign="top" align="center"><bold><italic>Q</italic><sub><italic>MI</italic></sub></bold></td>
<td valign="top" align="center"><bold><italic>Q</italic><sup><italic>AB</italic>/<italic>F</italic></sup></bold></td>
<td valign="top" align="center"><italic><bold>SCD</bold></italic></td>
<td valign="top" align="center"><italic><bold>VIFF</bold></italic></td>
</tr>
<tr>
<td valign="top" align="left">ASR</td>
<td valign="top" align="center">4.0307</td>
<td valign="top" align="center">0.9965</td>
<td valign="top" align="center">0.5916</td>
<td valign="top" align="center">1.4834</td>
<td valign="top" align="center">0.4553</td>
</tr>
<tr>
<td valign="top" align="left">NSCT-PC-LLE</td>
<td valign="top" align="center">4.5478</td>
<td valign="top" align="center"><bold>1.1500<sup>1</sup></bold></td>
<td valign="top" align="center"><bold>0.6432<sup>3</sup></bold></td>
<td valign="top" align="center">1.4883</td>
<td valign="top" align="center">0.5934</td>
</tr>
<tr>
<td valign="top" align="left">LP-CNN</td>
<td valign="top" align="center">4.6327</td>
<td valign="top" align="center"><bold>1.1294<sup>3</sup></bold></td>
<td valign="top" align="center"><bold>0.6670<sup>1</sup></bold></td>
<td valign="top" align="center">1.5333</td>
<td valign="top" align="center">0.6006</td>
</tr>
<tr>
<td valign="top" align="left">NSST-PAPCNN</td>
<td valign="top" align="center">4.6654</td>
<td valign="top" align="center"><bold>1.1305<sup>2</sup></bold></td>
<td valign="top" align="center">0.6035</td>
<td valign="top" align="center">1.5694</td>
<td valign="top" align="center">0.6053</td>
</tr>
<tr>
<td valign="top" align="left">CSMCA</td>
<td valign="top" align="center">4.1134</td>
<td valign="top" align="center">1.0401</td>
<td valign="top" align="center"><bold>0.6606<sup>2</sup></bold></td>
<td valign="top" align="center">1.4677</td>
<td valign="top" align="center">0.5421</td>
</tr>
<tr>
<td valign="top" align="left">LatLRR-FCN-32s</td>
<td valign="top" align="center"><bold>5.0259<sup>3</sup></bold></td>
<td valign="top" align="center">1.0940</td>
<td valign="top" align="center">0.6380</td>
<td valign="top" align="center"><bold>1.7483<sup>1</sup></bold></td>
<td valign="top" align="center"><bold>0.6335<sup>3</sup></bold></td>
</tr>
<tr>
<td valign="top" align="left">LatLRR-FCN-16s</td>
<td valign="top" align="center"><bold>5.0524<sup>2</sup></bold></td>
<td valign="top" align="center">1.0953</td>
<td valign="top" align="center">0.6384</td>
<td valign="top" align="center"><bold>1.7475<sup>2</sup></bold></td>
<td valign="top" align="center"><bold>0.6346<sup>2</sup></bold></td>
</tr>
<tr>
<td valign="top" align="left">LatLRR-FCN-8s</td>
<td valign="top" align="center"><bold>5.0949<sup>1</sup></bold></td>
<td valign="top" align="center">1.0974</td>
<td valign="top" align="center">0.6388</td>
<td valign="top" align="center"><bold>1.7457<sup>3</sup></bold></td>
<td valign="top" align="center"><bold>0.6358<sup>1</sup></bold></td>
</tr>
<tr style="border-top: thin solid #000000;border-bottom: thin solid #000000;">
<td valign="top" align="left"><bold>PET and MR</bold></td>
<td valign="top" align="center"><italic><bold>EN</bold></italic></td>
<td valign="top" align="center"><bold><italic>Q</italic><sub><italic>MI</italic></sub></bold></td>
<td valign="top" align="center"><bold><italic>Q</italic><sup><italic>AB</italic>/<italic>F</italic></sup></bold></td>
<td valign="top" align="center"><italic><bold>SCD</bold></italic></td>
<td valign="top" align="center"><italic><bold>VIFF</bold></italic></td>
</tr>
<tr>
<td valign="top" align="left">ASR</td>
<td valign="top" align="center">3.9649</td>
<td valign="top" align="center">0.9899</td>
<td valign="top" align="center">0.6153</td>
<td valign="top" align="center">1.6446</td>
<td valign="top" align="center">0.4206</td>
</tr>
<tr>
<td valign="top" align="left">NSCT-PC-LLE</td>
<td valign="top" align="center">4.4502</td>
<td valign="top" align="center">1.0739</td>
<td valign="top" align="center">0.6547</td>
<td valign="top" align="center">1.7013</td>
<td valign="top" align="center">0.5789</td>
</tr>
<tr>
<td valign="top" align="left">LP-CNN</td>
<td valign="top" align="center">4.5261</td>
<td valign="top" align="center">1.0523</td>
<td valign="top" align="center">0.6540</td>
<td valign="top" align="center">1.7121</td>
<td valign="top" align="center">0.5828</td>
</tr>
<tr>
<td valign="top" align="left">NSST-PAPCNN</td>
<td valign="top" align="center">4.4790</td>
<td valign="top" align="center"><bold>1.0946<sup>1</sup></bold></td>
<td valign="top" align="center">0.6486</td>
<td valign="top" align="center">1.7398</td>
<td valign="top" align="center">0.5997</td>
</tr>
<tr>
<td valign="top" align="left">CSMCA</td>
<td valign="top" align="center">4.0233</td>
<td valign="top" align="center">1.0078</td>
<td valign="top" align="center">0.6573</td>
<td valign="top" align="center">1.6725</td>
<td valign="top" align="center">0.5015</td>
</tr>
<tr>
<td valign="top" align="left">LatLRR-FCN-32s</td>
<td valign="top" align="center"><bold>4.5648<sup>1</sup></bold></td>
<td valign="top" align="center">1.0808</td>
<td valign="top" align="center"><bold>0.6588<sup>1</sup></bold></td>
<td valign="top" align="center"><bold>1.8861<sup>1</sup></bold></td>
<td valign="top" align="center"><bold>0.6699<sup>1</sup></bold></td>
</tr>
<tr>
<td valign="top" align="left">LatLRR-FCN-16s</td>
<td valign="top" align="center"><bold>4.5556<sup>3</sup></bold></td>
<td valign="top" align="center"><bold>1.0816<sup>3</sup></bold></td>
<td valign="top" align="center"><bold>0.6587<sup>2</sup></bold></td>
<td valign="top" align="center"><bold>1.8851<sup>2</sup></bold></td>
<td valign="top" align="center"><bold>0.6693<sup>2</sup></bold></td>
</tr>
<tr>
<td valign="top" align="left">LatLRR-FCN-8s</td>
<td valign="top" align="center"><bold>4.5573<sup>2</sup></bold></td>
<td valign="top" align="center"><bold>1.0819<sup>2</sup></bold></td>
<td valign="top" align="center"><bold>0.6584<sup>3</sup></bold></td>
<td valign="top" align="center"><bold>1.8834<sup>3</sup></bold></td>
<td valign="top" align="center"><bold>0.6689<sup>3</sup></bold></td>
</tr>
<tr style="border-top: thin solid #000000;border-bottom: thin solid #000000;">
<td valign="top" align="left"><bold>SPECT and MR</bold></td>
<td valign="top" align="center"><italic><bold>EN</bold></italic></td>
<td valign="top" align="center"><bold><italic>Q</italic><sub><italic>MI</italic></sub></bold></td>
<td valign="top" align="center"><bold><italic>Q</italic><sup><italic>AB</italic>/<italic>F</italic></sup></bold></td>
<td valign="top" align="center"><italic><bold>SCD</bold></italic></td>
<td valign="top" align="center"><italic><bold>VIFF</bold></italic></td>
</tr>
<tr>
<td valign="top" align="left">ASR</td>
<td valign="top" align="center">4.5089</td>
<td valign="top" align="center">1.1563</td>
<td valign="top" align="center">0.5548</td>
<td valign="top" align="center">1.3995</td>
<td valign="top" align="center">0.4692</td>
</tr>
<tr>
<td valign="top" align="left">NSCT-PC-LLE</td>
<td valign="top" align="center">4.9146</td>
<td valign="top" align="center"><bold>1.3703<sup>2</sup></bold></td>
<td valign="top" align="center"><bold>0.6405<sup>2</sup></bold></td>
<td valign="top" align="center">1.3839</td>
<td valign="top" align="center">0.5629</td>
</tr>
<tr>
<td valign="top" align="left">LP-CNN</td>
<td valign="top" align="center">5.4404</td>
<td valign="top" align="center"><bold>1.3146<sup>3</sup></bold></td>
<td valign="top" align="center"><bold>0.6496<sup>1</sup></bold></td>
<td valign="top" align="center">1.4958</td>
<td valign="top" align="center">0.6024</td>
</tr>
<tr>
<td valign="top" align="left">NSST-PAPCNN</td>
<td valign="top" align="center">4.9242</td>
<td valign="top" align="center"><bold>1.3912<sup>1</sup></bold></td>
<td valign="top" align="center"><bold>0.6355<sup>3</sup></bold></td>
<td valign="top" align="center">1.5119</td>
<td valign="top" align="center">0.5814</td>
</tr>
<tr>
<td valign="top" align="left">CSMCA</td>
<td valign="top" align="center">4.6187</td>
<td valign="top" align="center">1.1977</td>
<td valign="top" align="center">0.6233</td>
<td valign="top" align="center">1.4073</td>
<td valign="top" align="center">0.5339</td>
</tr>
<tr>
<td valign="top" align="left">LatLRR-FCN-32s</td>
<td valign="top" align="center"><bold>5.6361<sup>1</sup></bold></td>
<td valign="top" align="center">1.2880</td>
<td valign="top" align="center">0.6205</td>
<td valign="top" align="center"><bold>1.8224<sup>3</sup></bold></td>
<td valign="top" align="center"><bold>0.7075<sup>3</sup></bold></td>
</tr>
<tr>
<td valign="top" align="left">LatLRR-FCN-16s</td>
<td valign="top" align="center"><bold>5.6270<sup>3</sup></bold></td>
<td valign="top" align="center">1.2899</td>
<td valign="top" align="center">0.6207</td>
<td valign="top" align="center"><bold>1.8244<sup>2</sup></bold></td>
<td valign="top" align="center"><bold>0.7078<sup>2</sup></bold></td>
</tr>
<tr>
<td valign="top" align="left">LatLRR-FCN-8s</td>
<td valign="top" align="center"><bold>5.6341<sup>2</sup></bold></td>
<td valign="top" align="center">1.2916</td>
<td valign="top" align="center">0.6210</td>
<td valign="top" align="center"><bold>1.8265<sup>1</sup></bold></td>
<td valign="top" align="center"><bold>0.7081<sup>1</sup></bold></td>
</tr>
</tbody>
</table>
</table-wrap>
<p>As also shown in <xref ref-type="table" rid="T2">Table 2</xref>, for different metrics, it can be concluded as follows. (1) For the <italic>EN</italic> metric, our proposed techniques have the optimal energy preservation in four medical image fusion problems. (2) The <italic>Q</italic><sub><italic>MI</italic></sub> metric shows that our proposed LatLRR-FCN-8s and LatLRR-FCN-16s architectures obtain the best performance in detail information extraction than others in the context of CT and MR image fusion and PET and MR image fusion problems. (3) In terms of the <italic>Q</italic><sup><italic>AB</italic>/<italic>F</italic></sup> metric, our proposed frameworks are also close to the other comparison algorithms in edge and direction retention. (4) For the <italic>SCD</italic> metric, our proposed methods have a higher cross-correlation between the fused image and the input image than the others in all four medical image fusion problems. (5) For the <italic>VIFF</italic> metric, compared to the other methods, our proposed approaches are more consistent with the visual mechanism of human eyes in four medical image fusion problems.</p>
<p>Moreover, <xref ref-type="fig" rid="F10">Figure 10</xref> shows the objective performance of different methods in each fusion problem. The ten scores of each method in each fusion problem are connected for each metric. Obviously, the proposed three methods show the optimal performance among them. More specifically, the proposed LatLRR-FCNs are the best three ranks on the metrics of <italic>EN</italic>, <italic>SCD</italic>, and <italic>VIFF</italic> for all four problems, which is also concluded in <xref ref-type="table" rid="T2">Table 2</xref>.</p>
<fig id="F10" position="float">
<label>Figure 10</label>
<caption><p>Objective performance of different fusion methods in each fusion problem. <bold>(A)</bold> CT and MR, <bold>(B)</bold> MR-T1 and MR-T2, <bold>(C)</bold> PET and MR, and <bold>(D)</bold> SPECT and MR.</p></caption>
<graphic xlink:href="fnins-14-615435-g0010.tif"/>
</fig>
</sec>
<sec>
<title>4.3. Computational Cost Assessment</title>
<p>The average computational costs of different methods are shown in <xref ref-type="table" rid="T3">Table 3</xref>, including gray-level and color images. Although the performances of LP-CNN, NSCT-PC-LLE, and NSST-PAPCNN are better than the proposed methods, the proposed methods achieve a better performance in terms of both visual perception and objective assessment. However, the processing cost of ASR and CSMCA is 6 times and 10 times higher than our proposed methods. In total, the experimental results show that the proposed methods can achieve competitive performance in terms of computational costs in practice.</p>
<table-wrap position="float" id="T3">
<label>Table 3</label>
<caption><p>Computation cost of different methods.</p></caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th valign="top" align="left"><bold>Methods</bold></th>
<th valign="top" align="center"><bold>Gray-level image fusion time (s)</bold></th>
<th valign="top" align="center"><bold>Color image fusion time (s)</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">ASR</td>
<td valign="top" align="center">128.2032</td>
<td valign="top" align="center">163.4961</td>
</tr>
<tr>
<td valign="top" align="left">NSCT-PC-LLE</td>
<td valign="top" align="center">2.1638</td>
<td valign="top" align="center">4.2115</td>
</tr>
<tr>
<td valign="top" align="left">LP-CNN</td>
<td valign="top" align="center">11.7130</td>
<td valign="top" align="center">11.9583</td>
</tr>
<tr>
<td valign="top" align="left">NSST-PAPCNN</td>
<td valign="top" align="center">4.2226</td>
<td valign="top" align="center">4.8991</td>
</tr>
<tr>
<td valign="top" align="left">CSMCA</td>
<td valign="top" align="center">233.4387</td>
<td valign="top" align="center">252.1574</td>
</tr>
<tr>
<td valign="top" align="left">LatLRR-FCN-32s</td>
<td valign="top" align="center">21.5983</td>
<td valign="top" align="center">19.5076</td>
</tr>
<tr>
<td valign="top" align="left">LatLRR-FCN-16s</td>
<td valign="top" align="center">21.7452</td>
<td valign="top" align="center">19.7250</td>
</tr>
<tr>
<td valign="top" align="left">LatLRR-FCN-8s</td>
<td valign="top" align="center">20.2011</td>
<td valign="top" align="center">18.7253</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
</sec>
<sec sec-type="conclusions" id="s5">
<title>5. Conclusion</title>
<p>In this paper, three LatRR-FCNs have been proposed to improve energy conservation and detail extraction during medical image fusion. Based on LatLRR, the LatRR-FCNs decompose the medical image into low-rank and saliency components, which can enhance the extraction of detail in the SR-based methods. Then, three different fully convolutional networks (FCN-32s, FCN-16s, and FCN-8s), ZCA, <italic>l</italic><sub>1</sub>-norm, WLE, and WSEML operations together with a pyramid-based fusion method are applied to fuse the low-rank components, which can simultaneously enhance the energy preservation and detail extraction. We sum the saliency components to obtain the fused saliency components. Finally, the fused image is obtained by combining the fused low-rank components and fused saliency components. The proposed frameworks were evaluated in the context of four kinds of medical image fusion problems, including CT and MR, MR-T1 and MR-T2, PET and MR, and SPECT and MR. The results of our experiments demonstrated that the proposed frameworks can achieve optimal performance in both visual quality and objective assessment.</p>
</sec>
<sec sec-type="data-availability-statement" id="s6">
<title>Data Availability Statement</title>
<p>Publicly available datasets were analyzed in this study. This data can be found here: <ext-link ext-link-type="uri" xlink:href="http://www.med.harvard.edu/aanlib/">http://www.med.harvard.edu/aanlib/</ext-link>.</p>
</sec>
<sec id="s7">
<title>Author Contributions</title>
<p>ZX and JL: conceptualization. ZX and WX: methodology and writing&#x02013;review and editing. ZX: software, visualization, and writing&#x02013;original draft preparation. SZ, CM-C, and ZC: validation. RZ and XC: formal analysis. JL: resources and supervision. BL: data curation and project administration. All authors have read and agreed to the published version of the manuscript.</p>
</sec>
<sec sec-type="COI-statement" id="conf1">
<title>Conflict of Interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest. The handling editor declared a past co-authorship with one of the authors XC.</p>
</sec>
</body>
<back>
<ack><p>The authors would like to thank the editor and reviewers for the valuable comments and suggestions. We would like to thank Yu Liu for providing the experimental codes for their methods.</p></ack>
<ref-list>
<title>References</title>
<ref id="B1">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Anitha</surname> <given-names>S.</given-names></name> <name><surname>Subhashini</surname> <given-names>T.</given-names></name> <name><surname>Kamaraju</surname> <given-names>M.</given-names></name></person-group> (<year>2015</year>). <article-title>A novel multimodal medical image fusion approach based on phase congruency and directive contrast in nsct domain</article-title>. <source>Int. J. Comput. Appl</source>. <volume>129</volume>, <fpage>30</fpage>&#x02013;<lpage>35</lpage>. <pub-id pub-id-type="doi">10.5120/ijca2015907014</pub-id></citation>
</ref>
<ref id="B2">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Aslantas</surname> <given-names>V.</given-names></name> <name><surname>Bendes</surname> <given-names>E.</given-names></name></person-group> (<year>2015</year>). <article-title>A new image quality metric for image fusion: the sum of the correlations of differences</article-title>. <source>AEU Int. J. Electron. Commun</source>. <volume>69</volume>, <fpage>1890</fpage>&#x02013;<lpage>1896</lpage>. <pub-id pub-id-type="doi">10.1016/j.aeue.2015.09.004</pub-id></citation>
</ref>
<ref id="B3">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Bar</surname> <given-names>Y.</given-names></name> <name><surname>Diamant</surname> <given-names>I.</given-names></name> <name><surname>Wolf</surname> <given-names>L.</given-names></name> <name><surname>Lieberman</surname> <given-names>S.</given-names></name> <name><surname>Konen</surname> <given-names>E.</given-names></name> <name><surname>Greenspan</surname> <given-names>H.</given-names></name></person-group> (<year>2015</year>). <article-title>&#x0201C;Chest pathology detection using deep learning with non-medical training,&#x0201D;</article-title> in <source>2015 IEEE 12th International Symposium on Biomedical Imaging (ISBI)</source> (<publisher-loc>Brooklyn, NY</publisher-loc>: <publisher-name>IEEE</publisher-name>), <fpage>294</fpage>&#x02013;<lpage>297</lpage>. <pub-id pub-id-type="doi">10.1109/ISBI.2015.7163871</pub-id></citation></ref>
<ref id="B4">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bhatnagar</surname> <given-names>G.</given-names></name> <name><surname>Wu</surname> <given-names>Q. J.</given-names></name> <name><surname>Liu</surname> <given-names>Z.</given-names></name></person-group> (<year>2013</year>). <article-title>Directive contrast based multimodal medical image fusion in nsct domain</article-title>. <source>IEEE Trans. Multimedia</source> <volume>15</volume>, <fpage>1014</fpage>&#x02013;<lpage>1024</lpage>. <pub-id pub-id-type="doi">10.1109/TMM.2013.2244870</pub-id></citation></ref>
<ref id="B5">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bhatnagar</surname> <given-names>G.</given-names></name> <name><surname>Wu</surname> <given-names>Q. J.</given-names></name> <name><surname>Liu</surname> <given-names>Z.</given-names></name></person-group> (<year>2015</year>). <article-title>A new contrast based multimodal medical image fusion framework</article-title>. <source>Neurocomputing</source> <volume>157</volume>, <fpage>143</fpage>&#x02013;<lpage>152</lpage>. <pub-id pub-id-type="doi">10.1016/j.neucom.2015.01.025</pub-id></citation></ref>
<ref id="B6">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Chen</surname> <given-names>Q.</given-names></name> <name><surname>Xu</surname> <given-names>J.</given-names></name> <name><surname>Koltun</surname> <given-names>V.</given-names></name></person-group> (<year>2017</year>). <article-title>&#x0201C;Fast image processing with fully-convolutional networks,&#x0201D;</article-title> in <source>Proceedings of the IEEE International Conference on Computer Vision</source> (<publisher-loc>Venice</publisher-loc>), <fpage>2497</fpage>&#x02013;<lpage>2506</lpage>. <pub-id pub-id-type="doi">10.1109/ICCV.2017.273</pub-id></citation>
</ref>
<ref id="B7">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chen</surname> <given-names>Y.</given-names></name> <name><surname>Yang</surname> <given-names>M.</given-names></name> <name><surname>Chen</surname> <given-names>X.</given-names></name> <name><surname>Liu</surname> <given-names>B.</given-names></name> <name><surname>Wang</surname> <given-names>H.</given-names></name> <name><surname>Wang</surname> <given-names>S.</given-names></name></person-group> (<year>2018</year>). <article-title>Sensorineural hearing loss detection via discrete wavelet transform and principal component analysis combined with generalized eigenvalue proximal support vector machine and tikhonov regularization</article-title>. <source>Multimedia Tools Appl</source>. <volume>77</volume>, <fpage>3775</fpage>&#x02013;<lpage>3793</lpage>. <pub-id pub-id-type="doi">10.1007/s11042-016-4087-6</pub-id></citation></ref>
<ref id="B8">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Du</surname> <given-names>J.</given-names></name> <name><surname>Li</surname> <given-names>W.</given-names></name> <name><surname>Lu</surname> <given-names>K.</given-names></name> <name><surname>Xiao</surname> <given-names>B.</given-names></name></person-group> (<year>2016</year>). <article-title>An overview of multi-modal medical image fusion</article-title>. <source>Neurocomputing</source> <volume>215</volume>, <fpage>3</fpage>&#x02013;<lpage>20</lpage>. <pub-id pub-id-type="doi">10.1016/j.neucom.2015.07.160</pub-id></citation></ref>
<ref id="B9">
<citation citation-type="web"><person-group person-group-type="author"><name><surname>Everingham</surname> <given-names>M.</given-names></name> <name><surname>Van Gool</surname> <given-names>L.</given-names></name> <name><surname>Williams</surname> <given-names>C. K. I.</given-names></name> <name><surname>Winn</surname> <given-names>J.</given-names></name> <name><surname>Zisserman</surname> <given-names>A.</given-names></name></person-group> (<year>2012</year>). <source>The PASCAL Visual Object Classes Challenge 2012 (VOC2012) Results</source>. Available online at: <ext-link ext-link-type="uri" xlink:href="http://www.pascal-network.org/challenges/VOC/voc2012/workshop/index.html">http://www.pascal-network.org/challenges/VOC/voc2012/workshop/index.html</ext-link></citation></ref>
<ref id="B10">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Fei</surname> <given-names>Y.</given-names></name> <name><surname>Wei</surname> <given-names>G.</given-names></name> <name><surname>Zongxi</surname> <given-names>S.</given-names></name></person-group> (<year>2017</year>). <article-title>Medical image fusion based on feature extraction and sparse representation</article-title>. <source>Int. J. Biomed. Imaging</source> <volume>2017</volume>, <fpage>1</fpage>&#x02013;<lpage>13</lpage>. <pub-id pub-id-type="doi">10.1155/2017/3020461</pub-id><pub-id pub-id-type="pmid">28321246</pub-id></citation></ref>
<ref id="B11">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Geng</surname> <given-names>P.</given-names></name> <name><surname>Sun</surname> <given-names>X.</given-names></name> <name><surname>Liu</surname> <given-names>J.</given-names></name></person-group> (<year>2017</year>). <article-title>Adopting quaternion wavelet transform to fuse multi-modal medical images</article-title>. <source>J. Med. Biol. Eng</source>. <volume>37</volume>, <fpage>230</fpage>&#x02013;<lpage>239</lpage>. <pub-id pub-id-type="doi">10.1007/s40846-016-0200-6</pub-id><pub-id pub-id-type="pmid">29755307</pub-id></citation></ref>
<ref id="B12">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Guo</surname> <given-names>X.</given-names></name> <name><surname>Nie</surname> <given-names>R.</given-names></name> <name><surname>Cao</surname> <given-names>J.</given-names></name> <name><surname>Zhou</surname> <given-names>D.</given-names></name> <name><surname>Qian</surname> <given-names>W.</given-names></name></person-group> (<year>2018</year>). <article-title>Fully convolutional network-based multifocus image fusion</article-title>. <source>Neural Comput</source>. <volume>30</volume>, <fpage>1775</fpage>&#x02013;<lpage>1800</lpage>. <pub-id pub-id-type="doi">10.1162/neco_a_01098</pub-id><pub-id pub-id-type="pmid">29894654</pub-id></citation></ref>
<ref id="B13">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Han</surname> <given-names>Y.</given-names></name> <name><surname>Cai</surname> <given-names>Y.</given-names></name> <name><surname>Cao</surname> <given-names>Y.</given-names></name> <name><surname>Xu</surname> <given-names>X.</given-names></name></person-group> (<year>2013</year>). <article-title>A new image fusion performance metric based on visual information fidelity</article-title>. <source>Inform. Fusion</source> <volume>14</volume>, <fpage>127</fpage>&#x02013;<lpage>135</lpage>. <pub-id pub-id-type="doi">10.1016/j.inffus.2011.08.002</pub-id></citation></ref>
<ref id="B14">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Hariharan</surname> <given-names>B.</given-names></name> <name><surname>Arbel&#x000E1;ez</surname> <given-names>P.</given-names></name> <name><surname>Bourdev</surname> <given-names>L.</given-names></name> <name><surname>Maji</surname> <given-names>S.</given-names></name> <name><surname>Malik</surname> <given-names>J.</given-names></name></person-group> (<year>2011</year>). <article-title>&#x0201C;Semantic contours from inverse detectors,&#x0201D;</article-title> in <source>2011 International Conference on Computer Vision</source> (<publisher-loc>Barcelona</publisher-loc>: <publisher-name>IEEE</publisher-name>), <fpage>991</fpage>&#x02013;<lpage>998</lpage>. <pub-id pub-id-type="doi">10.1109/ICCV.2011.6126343</pub-id></citation></ref>
<ref id="B15">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Huang</surname> <given-names>B.</given-names></name> <name><surname>Yang</surname> <given-names>F.</given-names></name> <name><surname>Yin</surname> <given-names>M.</given-names></name> <name><surname>Mo</surname> <given-names>X.</given-names></name> <name><surname>Zhong</surname> <given-names>C.</given-names></name></person-group> (<year>2020</year>). <article-title>A review of multimodal medical image fusion techniques</article-title>. <source>Comput. Math. Methods Med</source>. <volume>2020</volume>, <fpage>1</fpage>&#x02013;<lpage>16</lpage>. <pub-id pub-id-type="doi">10.1155/2020/8279342</pub-id><pub-id pub-id-type="pmid">32377226</pub-id></citation></ref>
<ref id="B16">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>James</surname> <given-names>A. P.</given-names></name> <name><surname>Dasarathy</surname> <given-names>B. V.</given-names></name></person-group> (<year>2014</year>). <article-title>Medical image fusion: a survey of the state of the art</article-title>. <source>Inform. Fusion</source> <volume>19</volume>:<fpage>4</fpage>&#x02013;<lpage>19</lpage>. <pub-id pub-id-type="doi">10.1016/j.inffus.2013.12.002</pub-id></citation></ref>
<ref id="B17">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kessy</surname> <given-names>A.</given-names></name> <name><surname>Lewin</surname> <given-names>A.</given-names></name> <name><surname>Strimmer</surname> <given-names>K.</given-names></name></person-group> (<year>2018</year>). <article-title>Optimal whitening and decorrelation</article-title>. <source>Am. Stat</source>. <volume>72</volume>, <fpage>309</fpage>&#x02013;<lpage>314</lpage>. <pub-id pub-id-type="doi">10.1080/00031305.2016.1277159</pub-id></citation></ref>
<ref id="B18">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lewis</surname> <given-names>J. J.</given-names></name> <name><surname>O&#x00027;Callaghan</surname> <given-names>R. J.</given-names></name> <name><surname>Nikolov</surname> <given-names>S. G.</given-names></name> <name><surname>Bull</surname> <given-names>D. R.</given-names></name> <name><surname>Canagarajah</surname> <given-names>N.</given-names></name></person-group> (<year>2007</year>). <article-title>Pixel-and region-based image fusion with complex wavelets</article-title>. <source>Inform. Fusion</source> <volume>8</volume>, <fpage>119</fpage>&#x02013;<lpage>130</lpage>. <pub-id pub-id-type="doi">10.1016/j.inffus.2005.09.006</pub-id></citation></ref>
<ref id="B19">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>H.</given-names></name> <name><surname>He</surname> <given-names>X.</given-names></name> <name><surname>Tao</surname> <given-names>D.</given-names></name> <name><surname>Tang</surname> <given-names>Y.</given-names></name> <name><surname>Wang</surname> <given-names>R.</given-names></name></person-group> (<year>2018</year>). <article-title>Joint medical image fusion, denoising and enhancement via discriminative low-rank sparse dictionaries learning</article-title>. <source>Pattern Recogn</source>. <volume>79</volume>, <fpage>130</fpage>&#x02013;<lpage>146</lpage>. <pub-id pub-id-type="doi">10.1016/j.patcog.2018.02.005</pub-id></citation></ref>
<ref id="B20">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>H.</given-names></name> <name><surname>Manjunath</surname> <given-names>B.</given-names></name> <name><surname>Mitra</surname> <given-names>S. K.</given-names></name></person-group> (<year>1995</year>). <article-title>Multisensor image fusion using the wavelet transform</article-title>. <source>Graph. Models Image Process</source>. <volume>57</volume>, <fpage>235</fpage>&#x02013;<lpage>245</lpage>. <pub-id pub-id-type="doi">10.1006/gmip.1995.1022</pub-id></citation></ref>
<ref id="B21">
<citation citation-type="web"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>H.</given-names></name> <name><surname>Wu</surname> <given-names>X.-J.</given-names></name></person-group> (<year>2018</year>). <source>Infrared and visible image fusion using latent low-rank representation. <italic>arXiv [Preprint]. arXiv:1804.08992</italic></source>. Available online at: <ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/1804.08992">https://arxiv.org/abs/1804.08992</ext-link></citation></ref>
<ref id="B22">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>S.</given-names></name> <name><surname>Kang</surname> <given-names>X.</given-names></name> <name><surname>Hu</surname> <given-names>J.</given-names></name></person-group> (<year>2013</year>). <article-title>Image fusion with guided filtering</article-title>. <source>IEEE Trans. Image Process</source>. <volume>22</volume>, <fpage>2864</fpage>&#x02013;<lpage>2875</lpage>. <pub-id pub-id-type="doi">10.1109/TIP.2013.2244222</pub-id><pub-id pub-id-type="pmid">23372084</pub-id></citation></ref>
<ref id="B23">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>Y.</given-names></name> <name><surname>Sun</surname> <given-names>Y.</given-names></name> <name><surname>Huang</surname> <given-names>X.</given-names></name> <name><surname>Qi</surname> <given-names>G.</given-names></name> <name><surname>Zheng</surname> <given-names>M.</given-names></name> <name><surname>Zhu</surname> <given-names>Z.</given-names></name></person-group> (<year>2018</year>). <article-title>An image fusion method based on sparse representation and sum modified-laplacian in nsct domain</article-title>. <source>Entropy</source> <volume>20</volume>:<fpage>522</fpage>. <pub-id pub-id-type="doi">10.3390/e20070522</pub-id><pub-id pub-id-type="pmid">33265611</pub-id></citation></ref>
<ref id="B24">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Liu</surname> <given-names>G.</given-names></name> <name><surname>Yan</surname> <given-names>S.</given-names></name></person-group> (<year>2011</year>). <article-title>&#x0201C;Latent low-rank representation for subspace segmentation and feature extraction,&#x0201D;</article-title> in <source>2011 International Conference on Computer Vision</source> (<publisher-loc>Barcelona</publisher-loc>: <publisher-name>IEEE</publisher-name>), <fpage>1615</fpage>&#x02013;<lpage>1622</lpage>. <pub-id pub-id-type="doi">10.1109/ICCV.2011.6126422</pub-id></citation></ref>
<ref id="B25">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Liu</surname> <given-names>Y.</given-names></name> <name><surname>Chen</surname> <given-names>X.</given-names></name> <name><surname>Cheng</surname> <given-names>J.</given-names></name> <name><surname>Peng</surname> <given-names>H.</given-names></name></person-group> (<year>2017</year>). <article-title>&#x0201C;A medical image fusion method based on convolutional neural networks,&#x0201D;</article-title> in <source>2017 20th International Conference on Information Fusion (Fusion)</source> (<publisher-loc>Xi&#x00027;an</publisher-loc>: <publisher-name>IEEE</publisher-name>), <fpage>1</fpage>&#x02013;<lpage>7</lpage>. <pub-id pub-id-type="doi">10.23919/ICIF.2017.8009769</pub-id></citation></ref>
<ref id="B26">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Liu</surname> <given-names>Y.</given-names></name> <name><surname>Chen</surname> <given-names>X.</given-names></name> <name><surname>Ward</surname> <given-names>R. K.</given-names></name> <name><surname>Wang</surname> <given-names>Z. J.</given-names></name></person-group> (<year>2016</year>). <article-title>Image fusion with convolutional sparse representation</article-title>. <source>IEEE Signal Process. Lett</source>. <volume>23</volume>, <fpage>1882</fpage>&#x02013;<lpage>1886</lpage>. <pub-id pub-id-type="doi">10.1109/LSP.2016.2618776</pub-id></citation></ref>
<ref id="B27">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Liu</surname> <given-names>Y.</given-names></name> <name><surname>Chen</surname> <given-names>X.</given-names></name> <name><surname>Ward</surname> <given-names>R. K.</given-names></name> <name><surname>Wang</surname> <given-names>Z. J.</given-names></name></person-group> (<year>2019</year>). <article-title>Medical image fusion via convolutional sparsity based morphological component analysis</article-title>. <source>IEEE Signal Process. Lett</source>. <volume>26</volume>, <fpage>485</fpage>&#x02013;<lpage>489</lpage>. <pub-id pub-id-type="doi">10.1109/LSP.2019.2895749</pub-id></citation></ref>
<ref id="B28">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Liu</surname> <given-names>Y.</given-names></name> <name><surname>Wang</surname> <given-names>Z.</given-names></name></person-group> (<year>2014</year>). <article-title>Simultaneous image fusion and denoising with adaptive sparse representation</article-title>. <source>IET Image Process</source>. <volume>9</volume>, <fpage>347</fpage>&#x02013;<lpage>357</lpage>. <pub-id pub-id-type="doi">10.1049/iet-ipr.2014.0311</pub-id></citation></ref>
<ref id="B29">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Long</surname> <given-names>J.</given-names></name> <name><surname>Shelhamer</surname> <given-names>E.</given-names></name> <name><surname>Darrell</surname> <given-names>T.</given-names></name></person-group> (<year>2015</year>). <article-title>&#x0201C;Fully convolutional networks for semantic segmentation,&#x0201D;</article-title> in <source>Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition</source> (<publisher-loc>Boston, MA</publisher-loc>), <fpage>3431</fpage>&#x02013;<lpage>3440</lpage>. <pub-id pub-id-type="doi">10.1109/CVPR.2015.7298965</pub-id><pub-id pub-id-type="pmid">27244717</pub-id></citation></ref>
<ref id="B30">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lu</surname> <given-names>S.</given-names></name> <name><surname>Lu</surname> <given-names>Z.</given-names></name> <name><surname>Zhang</surname> <given-names>Y.-D.</given-names></name></person-group> (<year>2019</year>). <article-title>Pathological brain detection based on alexnet and transfer learning</article-title>. <source>J. Comput. Sci</source>. <volume>30</volume>, <fpage>41</fpage>&#x02013;<lpage>47</lpage>. <pub-id pub-id-type="doi">10.1016/j.jocs.2018.11.008</pub-id></citation></ref>
<ref id="B31">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lu</surname> <given-names>S.</given-names></name> <name><surname>Wang</surname> <given-names>S.-H.</given-names></name> <name><surname>Zhang</surname> <given-names>Y.-D.</given-names></name></person-group> (<year>2020</year>). <article-title>Detection of abnormal brain in MRI via improved Alexnet and ELM optimized by chaotic bat algorithm</article-title>. <source>Neural Comput. Appl</source>. <volume>32</volume>, <fpage>1</fpage>&#x02013;<lpage>13</lpage>. <pub-id pub-id-type="doi">10.1007/s00521-020-05082-4</pub-id></citation></ref>
<ref id="B32">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Manchanda</surname> <given-names>M.</given-names></name> <name><surname>Sharma</surname> <given-names>R.</given-names></name></person-group> (<year>2018</year>). <article-title>An improved multimodal medical image fusion algorithm based on fuzzy transform</article-title>. <source>J. Visual Commun. Image Represent</source>. <volume>51</volume>, <fpage>76</fpage>&#x02013;<lpage>94</lpage>. <pub-id pub-id-type="doi">10.1016/j.jvcir.2017.12.011</pub-id></citation></ref>
<ref id="B33">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Mertens</surname> <given-names>T.</given-names></name> <name><surname>Kautz</surname> <given-names>J.</given-names></name> <name><surname>Van Reeth</surname> <given-names>F.</given-names></name></person-group> (<year>2009</year>). <article-title>&#x0201C;Exposure fusion: a simple and practical alternative to high dynamic range photography,&#x0201D;</article-title> in <source>Computer Graphics Forum</source>, <volume>Vol. 28</volume>, eds R. Scopigno and E. Gr&#x000F6;ller (<publisher-name>Wiley Online Library</publisher-name>), <fpage>161</fpage>&#x02013;<lpage>171</lpage>. <pub-id pub-id-type="doi">10.1111/j.1467-8659.2008.01171.x</pub-id></citation></ref>
<ref id="B34">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Petrovic</surname> <given-names>V. S.</given-names></name> <name><surname>Xydeas</surname> <given-names>C. S.</given-names></name></person-group> (<year>2004</year>). <article-title>Gradient-based multiresolution image fusion</article-title>. <source>IEEE Trans. Image Process</source>. <volume>13</volume>, <fpage>228</fpage>&#x02013;<lpage>237</lpage>. <pub-id pub-id-type="doi">10.1109/TIP.2004.823821</pub-id><pub-id pub-id-type="pmid">15376943</pub-id></citation></ref>
<ref id="B35">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Razzak</surname> <given-names>M. I.</given-names></name> <name><surname>Naz</surname> <given-names>S.</given-names></name> <name><surname>Zaib</surname> <given-names>A.</given-names></name></person-group> (<year>2018</year>). <article-title>&#x0201C;Deep learning for medical image processing: overview, challenges and the future,&#x0201D;</article-title> in <source>Classification in BioApps</source>, eds N. Dey, A. Ashour, and S. Borra (<publisher-loc>Cham</publisher-loc>: <publisher-name>Springer</publisher-name>), <fpage>323</fpage>&#x02013;<lpage>350</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-319-65981-7_12</pub-id></citation></ref>
<ref id="B36">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Toet</surname> <given-names>A.</given-names></name></person-group> (<year>1989</year>). <article-title>A morphological pyramidal image decomposition</article-title>. <source>Pattern Recogn. Lett</source>. <volume>9</volume>, <fpage>255</fpage>&#x02013;<lpage>261</lpage>. <pub-id pub-id-type="doi">10.1016/0167-8655(89)90004-4</pub-id></citation></ref>
<ref id="B37">
<citation citation-type="web"><person-group person-group-type="author"><name><surname>Vidoni</surname> <given-names>E. D.</given-names></name></person-group> (<year>2012</year>). <article-title>The whole brain atlas: <ext-link ext-link-type="uri" xlink:href="http://www.med.harvard.edu/aanlib/">www.med.harvard.edu/aanlib/</ext-link></article-title>. <source>J. Neurol. Phys. Therapy</source> <volume>36</volume>:<fpage>108</fpage>. <pub-id pub-id-type="doi">10.1097/NPT.0b013e3182563795</pub-id></citation></ref>
<ref id="B38">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Walrand</surname> <given-names>S.</given-names></name> <name><surname>Hesse</surname> <given-names>M.</given-names></name> <name><surname>Jamar</surname> <given-names>F.</given-names></name></person-group> (<year>2017</year>). <article-title>&#x0201C;SPECT/CT, PET/CT and PET/MR principles,&#x0201D;</article-title> in <source>Diagnostic and Therapeutic Nuclear Medicine for Neuroendocrine Tumors</source>, eds K. Pacak and D. Ta&#x000EF;eb (<publisher-loc>Cham</publisher-loc>: <publisher-name>Springer</publisher-name>), <fpage>163</fpage>&#x02013;<lpage>200</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-319-46038-3_8</pub-id><pub-id pub-id-type="pmid">26576737</pub-id></citation></ref>
<ref id="B39">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>L.</given-names></name> <name><surname>Ouyang</surname> <given-names>W.</given-names></name> <name><surname>Wang</surname> <given-names>X.</given-names></name> <name><surname>Lu</surname> <given-names>H.</given-names></name></person-group> (<year>2015</year>). <article-title>&#x0201C;Visual tracking with fully convolutional networks,&#x0201D;</article-title> in <source>Proceedings of the IEEE International Conference on Computer Vision</source> (<publisher-loc>Santiago</publisher-loc>), <fpage>3119</fpage>&#x02013;<lpage>3127</lpage>. <pub-id pub-id-type="doi">10.1109/ICCV.2015.357</pub-id><pub-id pub-id-type="pmid">33031042</pub-id></citation></ref>
<ref id="B40">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>P.</given-names></name> <name><surname>Zhang</surname> <given-names>C.</given-names></name> <name><surname>Cai</surname> <given-names>S.</given-names></name> <name><surname>Li</surname> <given-names>L.</given-names></name></person-group> (<year>2013</year>). <article-title>Accelerated matrix recovery via random projection based on inexact augmented lagrange multiplier method</article-title>. <source>Trans. Tianjin Univ</source>. <volume>19</volume>, <fpage>293</fpage>&#x02013;<lpage>299</lpage>. <pub-id pub-id-type="doi">10.1007/s12209-013-2135-0</pub-id></citation></ref>
<ref id="B41">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>Q.</given-names></name> <name><surname>Li</surname> <given-names>S.</given-names></name> <name><surname>Qin</surname> <given-names>H.</given-names></name> <name><surname>Hao</surname> <given-names>A.</given-names></name></person-group> (<year>2015</year>). <article-title>Robust multi-modal medical image fusion via anisotropic heat diffusion guided low-rank structural analysis</article-title>. <source>Inform. Fusion</source> <volume>26</volume>, <fpage>103</fpage>&#x02013;<lpage>121</lpage>. <pub-id pub-id-type="doi">10.1016/j.inffus.2015.01.001</pub-id></citation></ref>
<ref id="B42">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>S.</given-names></name> <name><surname>Yang</surname> <given-names>M.</given-names></name> <name><surname>Du</surname> <given-names>S.</given-names></name> <name><surname>Yang</surname> <given-names>J.</given-names></name> <name><surname>Liu</surname> <given-names>B.</given-names></name> <name><surname>Gorriz</surname> <given-names>J. M.</given-names></name> <etal/></person-group>. (<year>2016</year>). <article-title>Wavelet entropy and directed acyclic graph support vector machine for detection of patients with unilateral hearing loss in mri scanning</article-title>. <source>Front. Comput. Neurosci</source>. <volume>10</volume>:<fpage>106</fpage>. <pub-id pub-id-type="doi">10.3389/fncom.2016.00106</pub-id></citation>
</ref>
<ref id="B43">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>S.</given-names></name> <name><surname>Yang</surname> <given-names>M.</given-names></name> <name><surname>Li</surname> <given-names>J.</given-names></name> <name><surname>Wu</surname> <given-names>X.</given-names></name> <name><surname>Wang</surname> <given-names>H.</given-names></name> <name><surname>Liu</surname> <given-names>B.</given-names></name> <etal/></person-group>. (<year>2017</year>). <article-title>Texture analysis method based on fractional fourier entropy and fitness-scaling adaptive genetic algorithm for detecting left-sided and right-sided sensorineural hearing loss</article-title>. <source>Fundament. Inform</source>. <volume>151</volume>, <fpage>505</fpage>&#x02013;<lpage>521</lpage>. <pub-id pub-id-type="doi">10.3233/FI-2017-1507</pub-id></citation></ref>
<ref id="B44">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Xydeas</surname> <given-names>C.</given-names></name> <name><surname>Petrovic</surname> <given-names>V.</given-names></name></person-group> (<year>2000</year>). <article-title>Objective image fusion performance measure</article-title>. <source>Electron. Lett</source>. <volume>36</volume>, <fpage>308</fpage>&#x02013;<lpage>309</lpage>. <pub-id pub-id-type="doi">10.1049/el:20000267</pub-id></citation></ref>
<ref id="B45">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Yin</surname> <given-names>M.</given-names></name> <name><surname>Liu</surname> <given-names>X.</given-names></name> <name><surname>Liu</surname> <given-names>Y.</given-names></name> <name><surname>Chen</surname> <given-names>X.</given-names></name></person-group> (<year>2018</year>). <article-title>Medical image fusion with parameter-adaptive pulse coupled neural network in nonsubsampled shearlet transform domain</article-title>. <source>IEEE Trans. Instrument. Meas</source>. <volume>68</volume>, <fpage>49</fpage>&#x02013;<lpage>64</lpage>. <pub-id pub-id-type="doi">10.1109/TIM.2018.2838778</pub-id></citation></ref>
<ref id="B46">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>Q.</given-names></name> <name><surname>Guo</surname> <given-names>B. L.</given-names></name></person-group> (<year>2009</year>). <article-title>Multifocus image fusion using the nonsubsampled contourlet transform</article-title>. <source>Signal Process</source>. <volume>89</volume>, <fpage>1334</fpage>&#x02013;<lpage>1346</lpage>. <pub-id pub-id-type="doi">10.1016/j.sigpro.2009.01.012</pub-id></citation></ref>
<ref id="B47">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>Y.</given-names></name> <name><surname>Ranjan Nayak</surname> <given-names>D.</given-names></name> <name><surname>Yang</surname> <given-names>M.</given-names></name> <name><surname>Yuan</surname> <given-names>T.-F.</given-names></name> <name><surname>Liu</surname> <given-names>B.</given-names></name> <name><surname>Lu</surname> <given-names>H.</given-names></name> <etal/></person-group>. (<year>2017</year>). <article-title>Detection of unilateral hearing loss by stationary wavelet entropy</article-title>. <source>CNS Neurol. Disord. Drug Targets</source> <volume>16</volume>, <fpage>122</fpage>&#x02013;<lpage>128</lpage>. <pub-id pub-id-type="doi">10.2174/1871527315666161026115046</pub-id><pub-id pub-id-type="pmid">27784224</pub-id></citation></ref>
<ref id="B48">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhao</surname> <given-names>W.</given-names></name> <name><surname>Lu</surname> <given-names>H.</given-names></name></person-group> (<year>2017</year>). <article-title>Medical image fusion and denoising with alternating sequential filter and adaptive fractional order total variation</article-title>. <source>IEEE Trans. Instrument. Meas</source>. <volume>66</volume>, <fpage>2283</fpage>&#x02013;<lpage>2294</lpage>. <pub-id pub-id-type="doi">10.1109/TIM.2017.2700198</pub-id></citation></ref>
<ref id="B49">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhu</surname> <given-names>Z.</given-names></name> <name><surname>Zheng</surname> <given-names>M.</given-names></name> <name><surname>Qi</surname> <given-names>G.</given-names></name> <name><surname>Wang</surname> <given-names>D.</given-names></name> <name><surname>Xiang</surname> <given-names>Y.</given-names></name></person-group> (<year>2019</year>). <article-title>A phase congruency and local laplacian energy based multi-modality medical image fusion method in nsct domain</article-title>. <source>IEEE Access</source> <volume>7</volume>, <fpage>20811</fpage>&#x02013;<lpage>20824</lpage>. <pub-id pub-id-type="doi">10.1109/ACCESS.2019.2898111</pub-id></citation></ref>
</ref-list>
<fn-group>
<fn fn-type="financial-disclosure"><p><bold>Funding.</bold> This document is the results of the research project funded by the National Key Research and Development Programme of China (2017YFB1303200), NSFC (61571113, 62001240), Leading-edge Technology and Basic Research Programme of Jiangsu (BK20192004D), Jiangsu Graduate Scientific Research Innovation Programme (KYCX20_1391), and Zhejiang Provincial Natural Science Foundation of China under Grant (LY17F010003).</p></fn>
</fn-group>
</back>
</article>