<?xml version="1.0" encoding="UTF-8" standalone="no"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Neurosci.</journal-id>
<journal-title>Frontiers in Neuroscience</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Neurosci.</abbrev-journal-title>
<issn pub-type="epub">1662-453X</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fnins.2023.1202143</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Neuroscience</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Linear fine-tuning: a linear transformation based transfer strategy for deep MRI reconstruction</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name><surname>Bi</surname> <given-names>Wanqing</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/2271458/overview"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Xv</surname> <given-names>Jianan</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/2292343/overview"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Song</surname> <given-names>Mengdie</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
</contrib>
<contrib contrib-type="author">
<name><surname>Hao</surname> <given-names>Xiaohan</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/1045930/overview"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Gao</surname> <given-names>Dayong</given-names></name>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name><surname>Qi</surname> <given-names>Fulang</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>&#x0002A;</sup></xref>
</contrib>
</contrib-group>
<aff id="aff1"><sup>1</sup><institution>The Centers for Biomedical Engineering, University of Science and Technology of China, Hefei</institution>, <addr-line>Anhui</addr-line>, <country>China</country></aff>
<aff id="aff2"><sup>2</sup><institution>Fuqing Medical Co., Ltd., Hefei</institution>, <addr-line>Anhui</addr-line>, <country>China</country></aff>
<aff id="aff3"><sup>3</sup><institution>Department of Mechanical Engineering, University of Washington</institution>, <addr-line>Seattle, WA</addr-line>, <country>United States</country></aff>
<author-notes>
<fn fn-type="edited-by"><p>Edited by: Fabio Baselice, University of Naples Parthenope, Italy</p></fn>
<fn fn-type="edited-by"><p>Reviewed by: Sairam Geethanath, Icahn School of Medicine at Mount Sinai, United States; Hossam El-Din Moustafa, Mansoura University, Egypt</p></fn>
<corresp id="c001">&#x0002A;Correspondence: Fulang Qi <email>qaler&#x00040;mail.ustc.edu.cn</email></corresp>
</author-notes>
<pub-date pub-type="epub">
<day>20</day>
<month>06</month>
<year>2023</year>
</pub-date>
<pub-date pub-type="collection">
<year>2023</year>
</pub-date>
<volume>17</volume>
<elocation-id>1202143</elocation-id>
<history>
<date date-type="received">
<day>07</day>
<month>04</month>
<year>2023</year>
</date>
<date date-type="accepted">
<day>05</day>
<month>06</month>
<year>2023</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x000A9; 2023 Bi, Xv, Song, Hao, Gao and Qi.</copyright-statement>
<copyright-year>2023</copyright-year>
<copyright-holder>Bi, Xv, Song, Hao, Gao and Qi</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p></license> </permissions>
<abstract>
<sec>
<title>Introduction</title>
<p>Fine-tuning (FT) is a generally adopted transfer learning method for deep learning-based magnetic resonance imaging (MRI) reconstruction. In this approach, the reconstruction model is initialized with pre-trained weights derived from a source domain with ample data and subsequently updated with limited data from the target domain. However, the direct full-weight update strategy can pose the risk of &#x00022;catastrophic forgetting&#x00022; and overfitting, hindering its effectiveness. The goal of this study is to develop a zero-weight update transfer strategy to preserve pre-trained generic knowledge and reduce overfitting.</p></sec>
<sec>
<title>Methods</title>
<p>Based on the commonality between the source and target domains, we assume a linear transformation relationship of the optimal model weights from the source domain to the target domain. Accordingly, we propose a novel transfer strategy, linear fine-tuning (LFT), which introduces scaling and shifting (SS) factors into the pre-trained model. In contrast to FT, LFT only updates SS factors in the transfer phase, while the pre-trained weights remain fixed.</p></sec>
<sec>
<title>Results</title>
<p>To evaluate the proposed LFT, we designed three different transfer scenarios and conducted a comparative analysis of FT, LFT, and other methods at various sampling rates and data volumes. In the transfer scenario between different contrasts, LFT outperforms typical transfer strategies at various sampling rates and considerably reduces artifacts on reconstructed images. In transfer scenarios between different slice directions or anatomical structures, LFT surpasses the FT method, particularly when the target domain contains a decreasing number of training images, with a maximum improvement of up to 2.06 dB (5.89%) in peak signal-to-noise ratio.</p></sec>
<sec>
<title>Discussion</title>
<p>The LFT strategy shows great potential to address the issues of &#x00022;catastrophic forgetting&#x00022; and overfitting in transfer scenarios for MRI reconstruction, while reducing the reliance on the amount of data in the target domain. Linear fine-tuning is expected to shorten the development cycle of reconstruction models for adapting complicated clinical scenarios, thereby enhancing the clinical applicability of deep MRI reconstruction.</p></sec></abstract>
<kwd-group>
<kwd>magnetic resonance imaging reconstruction</kwd>
<kwd>deep learning</kwd>
<kwd>transfer learning</kwd>
<kwd>fine-tuning</kwd>
<kwd>transfer strategy</kwd>
</kwd-group>
<counts>
<fig-count count="8"/>
<table-count count="4"/>
<equation-count count="10"/>
<ref-count count="43"/>
<page-count count="12"/>
<word-count count="8248"/>
</counts>
<custom-meta-wrap>
<custom-meta>
<meta-name>section-at-acceptance</meta-name>
<meta-value>Brain Imaging Methods</meta-value>
</custom-meta>
</custom-meta-wrap>
</article-meta>
</front>
<body>
<sec sec-type="intro" id="s1">
<title>1. Introduction</title>
<p>Magnetic resonance imaging (MRI) includes diverse sequences that provide distinct types of anatomical and pathological information, catering to a wide range of clinical needs (Yousaf et al., <xref ref-type="bibr" rid="B42">2018</xref>). For instance, T1-weighted spin-echo sequences can be employed to measure the cross-sectional area of visceral and subcutaneous fat in the abdomen (Lancaster et al., <xref ref-type="bibr" rid="B22">1991</xref>), as T1-weighted images provide the most anatomically-relevant details. Additionally, T2-weighted images highlight lesions and can be used to identify myocardial edema (Eitel and Friedrich, <xref ref-type="bibr" rid="B9">2011</xref>), determine the area at risk in non-reperfused infarction (Aletras et al., <xref ref-type="bibr" rid="B1">2006</xref>), etc. As a downside, standard scan sequences used in clinical routines generally require a long imaging time, which can aggravate patient discomfort and introduce severe motion artifacts. In an effort to speed up the acquisition of MRI, reconstruction techniques (Roy and Kailath, <xref ref-type="bibr" rid="B33">1989</xref>; Pruessmann et al., <xref ref-type="bibr" rid="B30">1999</xref>; Griswold et al., <xref ref-type="bibr" rid="B13">2002</xref>; Block et al., <xref ref-type="bibr" rid="B6">2007</xref>; Hamilton et al., <xref ref-type="bibr" rid="B15">2017</xref>) based on under-sampled k-space data have been developed, albeit at the cost of increased hardware requirements or lengthened reconstruction time.</p>
<p>Recently, deep learning-based technologies have gained much attention owing to the powerful image representation capabilities and fast image generation speed (Lee et al., <xref ref-type="bibr" rid="B23">2018</xref>; Quan et al., <xref ref-type="bibr" rid="B31">2018</xref>; Hosseini et al., <xref ref-type="bibr" rid="B16">2020</xref>; Cole et al., <xref ref-type="bibr" rid="B7">2021</xref>). Numerous neural network frameworks have been proposed for MRI reconstruction, establishing an end-to-end non-linear mapping from under-sampled data to fully-sampled images. Wang et al. (<xref ref-type="bibr" rid="B38">2016</xref>) introduced a convolutional neural network (CNN) that employed a large number of MR images as training datasets to recover the delicate structures and details of test data. Additionally, Quan et al. (<xref ref-type="bibr" rid="B31">2018</xref>) designed a variant of a fully-residual convolutional auto-encoder and generative adversarial networks (GAN) called RefineGAN, which incorporated a cyclic data consistency loss and a chained architecture to enhance the reconstruction quality. Despite being a powerful tool for MRI reconstruction, deep learning still suffers from a notable inherent drawback: limited generalizability (Gavrikov and Keuper, <xref ref-type="bibr" rid="B12">2022</xref>). It was found that deep learning algorithms are sensitive to shifts in the distribution of input data (Knoll et al., <xref ref-type="bibr" rid="B21">2019</xref>; Antun et al., <xref ref-type="bibr" rid="B4">2020</xref>). Knoll et al. (<xref ref-type="bibr" rid="B21">2019</xref>) summarized the deviations between training and testing led to substantial decreases in reconstruction image quality. Antun et al. (<xref ref-type="bibr" rid="B4">2020</xref>) demonstrated that even some certain tiny, almost imperceptible perturbations may result in severe artifacts in the reconstruction. Moreover, even for one specific anatomical part, the reconstructed MR images exhibit obvious stylistic variations when different sequences or different scan parameters are performed. Therefore, adapting a reconstruction model to diverse clinical scenarios is challenging.</p>
<p>To enhance the generalization ability of the neural networks, a straightforward approach is to collect ample data encompassing various clinical scenarios. However, medical imaging domains generally face a dearth of data, which is both time-consuming and expensive to acquire. Alternatively, transfer learning strategies (Tajbakhsh et al., <xref ref-type="bibr" rid="B37">2016</xref>) have been developed, focusing on how to transfer knowledge from the source domain to the target domain (Romero et al., <xref ref-type="bibr" rid="B32">2020</xref>). Since an explicit paradigm for designing transfer strategies is not available, multiple factors such as the task attribute, the total amount of data, and the amount of labeled data in the source and target domains should be considered comprehensively. Kamphenkel et al. (<xref ref-type="bibr" rid="B19">2018</xref>) proposed an unsupervised domain adaptation method for breast cancer classification to transform the target data to the source domain without any label information of target data. Faced with the inability to obtain paired images of the same subject, Zhang et al. (<xref ref-type="bibr" rid="B43">2022</xref>) designed a novel unsupervised domain adaptation approach for tackling the distribution discrepancy across domains in medical image segmentation.</p>
<p>Regarding the MRI reconstruction scenario involved in this study, it is characterized by the following aspects. Firstly, the tasks in both the source and target domains are identical, i.e., reconstructing under-sampled aliased images into aliasing-free images. Secondly, the source domain is composed of a large-scale public dataset, while the target domain has a relatively small data volume. Thirdly, network training usually takes full-sampling data as the reference without requiring additional labeling. Hence, in this case, fine-tuning (FT) (Pan and Yang, <xref ref-type="bibr" rid="B29">2010</xref>; Dar et al., <xref ref-type="bibr" rid="B8">2020</xref>; Fr&#x000E9;gier and Gouray, <xref ref-type="bibr" rid="B11">2021</xref>) is the most commonly used transfer learning method that updates pre-trained models to target domains. The FT strategy leverages pre-trained weights as initialization and relearns all the weights of convolutional layers in the transfer phase, which has been investigated in many studies. Dar et al. (<xref ref-type="bibr" rid="B8">2020</xref>) pre-trained networks on large public MR datasets and then fine-tuned them using only tens of brain MR images from the target domain, achieving performance comparable to networks trained directly on thousands of target images. Arshad et al. (<xref ref-type="bibr" rid="B5">2021</xref>) obtained satisfactory reconstruction of MR images based on a pre-trained U-Net through end-to-end FT under various magnetic field strengths, anatomies, and acceleration factors. Lv et al. (<xref ref-type="bibr" rid="B25">2021a</xref>) focused on the generalization ability of the multi-channel MRI reconstruction network and highlighted the critical role of FT in adapting to a particular target application using only a few training cases.</p>
<p>All of the aforementioned research has employed FT to achieve reliable reconstructions of under-sampled images in different transfer scenarios, demonstrating the potential of knowledge transfer for reconstruction. However, two major limitations of FT hinder its effectiveness. Firstly, updating all the pre-trained weights implies that the target model may forget the pre-trained knowledge when adapting to a new task, known as &#x0201C;catastrophic forgetting&#x0201D; (Lopez-Paz and Ranzato, <xref ref-type="bibr" rid="B24">2017</xref>). This can be improved as distinct MR images exhibit many similar features despite variations in contrast, slicing direction, and anatomy. Additionally, networks mostly contain a mass of weights, and the full-weight adjustment on a small target dataset may result in overfitting (Sun et al., <xref ref-type="bibr" rid="B36">2019</xref>). Recently, layer-wise FT (Tajbakhsh et al., <xref ref-type="bibr" rid="B37">2016</xref>; Amiri et al., <xref ref-type="bibr" rid="B3">2020</xref>), a transfer strategy of FT, has been applied in the field of medical imaging for transfer learning. Layer-wise FT freezes the weights of specific layers and tunes the rest in the transfer phase. Although layer-wise FT has the potential to alleviate &#x0201C;catastrophic forgetting&#x0201D; by reducing the degree of weight updates, the determination of an optimal freezing strategy is scenario-dependent, making it difficult to design a general scheme for all transfer scenarios as in the case of FT. Therefore, layer-wise FT is rarely applied in the transfer learning of deep MRI reconstruction.</p>
<p>In this study, we propose a novel transfer strategy of FT, termed linear fine-tuning (LFT), which assists in effective transfer and is applicable to MRI reconstruction. The approach assumes a linear transformation relationship between the source and target models. Specifically, networks pre-trained on a large-scale source dataset can be transferred to the target domain by adjusting the scaling and shifting (<italic>SS</italic>) factors. These two factors are integrated into the <italic>SS</italic> block in a learnable fashion, which is merged with regular convolution to adjust for linear deviation. Unlike FT, LFT avoids changing the entire pre-trained weights and instead updates the <italic>SS</italic> factors continuously while freezing the pre-trained weights and biases. The proposed zero-weight strategy benefits learning specific knowledge while avoiding forgetting the generic knowledge by fixing the pre-trained weights. Additionally, LFT entails a considerably smaller number of weights to be adjusted, thereby reducing the risk of overfitting.</p>
<p>Our contributions can be summarized as follows.</p>
<list list-type="bullet">
<list-item><p>We hypothesize that advanced features of different MR images can be represented as different linear combinations of the same basic features.</p></list-item>
<list-item><p>Based on the proposed assumption, the LFT strategy introduces two coefficients, namely <italic>SS</italic> factors, in convolutional layers for the linear transformation of features. In the transfer phase, only the <italic>SS</italic> factors are updated, thereby mitigating the risk of &#x0201C;catastrophic forgetting&#x0201D; and overfitting.</p></list-item>
<list-item><p>We conducted extensive experiments on various MR datasets with different data volumes and sampling rates. The results suggest that the proposed LFT strategy generally outperforms other methods and achieves high-quality reconstruction, especially in the case of small samples.</p></list-item>
</list></sec>
<sec id="s2">
<title>2. Methods and materials</title>
<sec>
<title>2.1. Preliminary</title>
<sec>
<title>2.1.1. Definitions and notations</title>
<p>Since the key part of the LFT strategy implementation is the <italic>SS</italic> factors inserted into the convolutional layer, some concepts are defined to facilitate the explanation. In general, one of the fundamental components of a neural network is the convolutional layer, which acts as a feature extractor. The two-dimensional convolutional layer transforms an input tensor <italic>U</italic> of size <italic>M</italic> &#x000D7; <italic>H</italic> &#x000D7; <italic>W</italic> into an output tensor <italic>V</italic> of size <italic>N</italic> &#x000D7; <italic>H</italic> &#x000D7; <italic>W</italic> by applying <italic>N</italic> convolution filters <italic>F</italic><sup>(1)</sup>, <italic>F</italic><sup>(2)</sup>, &#x022EF;&#x02009;, <italic>F</italic><sup>(<italic>N</italic>)</sup>. The feature maps generated by each filter are defined as advanced features <italic>V</italic><sub><italic>a</italic></sub>, such that</p>
<disp-formula id="E1"><label>(1)</label><mml:math id="M1"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>V</mml:mi></mml:mrow><mml:mrow><mml:mi>a</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mi>U</mml:mi><mml:mo>*</mml:mo><mml:mi>F</mml:mi><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where <italic>F</italic> &#x02208; {<italic>F</italic><sup>(1)</sup>, <italic>F</italic><sup>(2)</sup>, &#x022EF;&#x02009;, <italic>F</italic><sup>(<italic>N</italic>)</sup>}, &#x0002A; represents the multichannel convolution operator, and <italic>V</italic><sub><italic>a</italic></sub> of each filter corresponds to one channel of <italic>V</italic>.</p>
<p>Each filter <italic>F</italic> contains <italic>M</italic> kernels, namely <italic>K</italic><sub>(1)</sub>, <italic>K</italic><sub>(2)</sub>, &#x022EF;&#x02009;, <italic>K</italic><sub>(<italic>M</italic>)</sub>, with weights <italic>W</italic>. The output of each kernel is defined as basic features <italic>V</italic><sub><italic>b</italic></sub>, via</p>
<disp-formula id="E2"><label>(2)</label><mml:math id="M2"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>V</mml:mi></mml:mrow><mml:mrow><mml:mi>b</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>m</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>U</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>m</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mo>&#x02299;</mml:mo><mml:msub><mml:mrow><mml:mi>W</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>m</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mo>&#x0002B;</mml:mo><mml:msub><mml:mrow><mml:mi>b</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>m</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where <italic>K</italic><sub>(<italic>m</italic>)</sub> &#x02208; {<italic>K</italic><sub>(1)</sub>, <italic>K</italic><sub>(2)</sub>, &#x022EF;&#x02009;, <italic>K</italic><sub>(<italic>M</italic>)</sub>}, with weight <italic>W</italic><sub>(<italic>m</italic>)</sub>. <italic>U</italic><sub>(<italic>m</italic>)</sub> indicates the channel of <italic>U</italic> corresponding to <italic>K</italic><sub>(<italic>m</italic>)</sub>. &#x02299; denotes the convolution operator and <italic>b</italic> is the bias of the convolutional layer.</p>
<p>Equations (1) and (2) are combined to obtain</p>
<disp-formula id="E3"><label>(3)</label><mml:math id="M3"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>V</mml:mi></mml:mrow><mml:mrow><mml:mi>a</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mstyle displaystyle="true"><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>m</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>M</mml:mi></mml:mrow></mml:munderover></mml:mstyle><mml:msub><mml:mrow><mml:mi>V</mml:mi></mml:mrow><mml:mrow><mml:mi>b</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>m</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mstyle displaystyle="true"><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>m</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>M</mml:mi></mml:mrow></mml:munderover></mml:mstyle><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>U</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>m</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mo>&#x02299;</mml:mo><mml:msub><mml:mrow><mml:mi>W</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>m</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mo>&#x0002B;</mml:mo><mml:msub><mml:mrow><mml:mi>b</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>m</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>Equation (3) indicates the convolution operation can be split into two parts: feature extraction and linear combination. Specifically, as illustrated in <xref ref-type="fig" rid="F1">Figure 1</xref>, an example of a filter <italic>F</italic> containing two kernels <italic>K</italic><sub>(1)</sub>, <italic>K</italic><sub>(2)</sub> is used. Each kernel slides over the corresponding channel of <italic>U</italic> and extracts basic features. Subsequently, the extracted features from different kernels are combined into advanced features, which are commonly described as a feature map. Advanced features typically contain special significance that reflects the function of the filter, especially in the deep layers of the network. <xref ref-type="fig" rid="F1">Figure 1</xref> shows two types of advanced features, including in-contour artifacts (top) and out-of-contour artifacts (bottom).</p>
<fig id="F1" position="float">
<label>Figure 1</label>
<caption><p>Exemplary basic and advanced features. <italic>V</italic><sub><italic>b</italic></sub> and <italic>V</italic><sub><italic>a</italic></sub> represent the outputs of each kernel and each filter, respectively. <italic>V</italic><sub><italic>a</italic></sub> shows two types of advanced features including in-contour artifacts <bold>(top)</bold> and out-of-contour artifacts <bold>(bottom)</bold>.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnins-17-1202143-g0001.tif"/>
</fig></sec>
<sec>
<title>2.1.2. Fine-tuning</title>
<p>The aim for transfer learning is to generalize high-performance networks trained in a related source domain to the target domain (Weiss et al., <xref ref-type="bibr" rid="B40">2016</xref>). Fine-tuning is a common transfer learning method for MRI reconstruction, which begins with pre-training a network in the source domain, usually with large-scale data. In the transfer phase, a new network is initialized with the pre-trained weights and continues to be trained for target task, with the updating process is shown in <xref ref-type="fig" rid="F2">Figure 2A</xref>.</p>
<fig id="F2" position="float">
<label>Figure 2</label>
<caption><p>Visualization of the difference between FT and LFT in the transfer phase. Taking a 3 &#x000D7; 3 filter as an example with four channels, the structure of <italic>SSConv</italic> is composed of <italic>SS</italic> block and regular convolution. The 3 &#x000D7; 3 and 1 &#x000D7; 1 squares represent the weight <italic>W</italic> and bias <italic>b</italic> of the kernel, respectively, and different colors in them symbolize different channels. <bold>(A)</bold> FT updates all the pre-trained <italic>W</italic> and <italic>b</italic>. <bold>(B)</bold> The propose LFT optimizes the scaling factor &#x003A6;<sub><italic>S</italic><sub><italic>W</italic></sub></sub> and shifting factor &#x003A6;<sub><italic>S</italic><sub><italic>b</italic></sub></sub> only.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnins-17-1202143-g0002.tif"/>
</fig>
<p>During pre-training, regular convolution <italic>Conv</italic> is applied to the input <italic>U</italic><sup><italic>s</italic></sup> in the source domain with a kernel <italic>K</italic><sub>(<italic>m</italic>)</sub> as</p>
<disp-formula id="E4"><label>(4)</label><mml:math id="M4"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msubsup><mml:mrow><mml:mi>V</mml:mi></mml:mrow><mml:mrow><mml:mi>b</mml:mi></mml:mrow><mml:mrow><mml:mi>s</mml:mi></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:msup><mml:mrow><mml:mi>U</mml:mi></mml:mrow><mml:mrow><mml:mi>s</mml:mi></mml:mrow></mml:msup><mml:mo>&#x02299;</mml:mo><mml:mi>W</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mi>b</mml:mi><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>Note that <italic>m</italic> is omitted for readability. During the transfer process, FT initializes the target network with the pre-trained <italic>W</italic> and <italic>b</italic>, giving</p>
<disp-formula id="E5"><label>(5)</label><mml:math id="M5"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msubsup><mml:mrow><mml:mi>V</mml:mi></mml:mrow><mml:mrow><mml:mi>b</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:msup><mml:mrow><mml:mi>U</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msup><mml:mo>&#x02299;</mml:mo><mml:mi>W</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mi>b</mml:mi><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>Afterwards, FT updates all the values of <italic>W</italic> and <italic>b</italic> utilizing the target data <italic>U</italic><sub><italic>t</italic></sub>, obtaining</p>
<disp-formula id="E6"><label>(6)</label><mml:math id="M6"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msubsup><mml:mrow><mml:mi>V</mml:mi></mml:mrow><mml:mrow><mml:mi>b</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:msup><mml:mrow><mml:mi>U</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msup><mml:mo>&#x02299;</mml:mo><mml:msup><mml:mrow><mml:mi>W</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x02032;</mml:mi></mml:mrow></mml:msup><mml:mo>&#x0002B;</mml:mo><mml:msup><mml:mrow><mml:mi>b</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x02032;</mml:mi></mml:mrow></mml:msup><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where <italic>W</italic>&#x02032; and <italic>b</italic>&#x02032; indicate the updated weights.</p>
<p>Fine-tuning is considered to shorten the convergence time and enhance the quality of the generated images by providing a strong initialization, particularly when the target dataset is limited (Jiang et al., <xref ref-type="bibr" rid="B18">2019</xref>). However, as stated in Section 1, such full-weight updating leads to &#x0201C;catastrophic forgetting&#x0201D; and overfitting.</p></sec></sec>
<sec>
<title>2.2. Linear fine-tuning</title>
<sec>
<title>2.2.1. Hypothesis</title>
<p>Despite the presence of distinct variations in advanced features, they share many of the basic features (Olah et al., <xref ref-type="bibr" rid="B28">2020</xref>). Hence, we hypothesize that for two similar types of images, their advanced features <italic>V</italic><sub><italic>a</italic></sub> can be derived from different linear combinations of the shared basic features <italic>V</italic><sub><italic>b</italic></sub> by means of the scaling factor &#x003A6;<sub><italic>S</italic><sub><italic>W</italic></sub></sub> and shifting factor &#x003A6;<sub><italic>S</italic><sub><italic>b</italic></sub></sub>. By substituting the inputs in Equation (3) with data <italic>U</italic><sup>1</sup> and <italic>U</italic><sup>2</sup> from these two datasets, and adding two factors, the hypothesis can be expressed as</p>
<disp-formula id="E7"><label>(7)</label><mml:math id="M7"><mml:mtable columnalign='left'><mml:mtr><mml:mtd><mml:msubsup><mml:mi>V</mml:mi><mml:mi>a</mml:mi><mml:mn>1</mml:mn></mml:msubsup><mml:mo>=</mml:mo><mml:mstyle displaystyle='true'><mml:munderover><mml:mo>&#x02211;</mml:mo><mml:mrow><mml:mi>m</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>M</mml:mi></mml:munderover><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:mo stretchy='false'>(</mml:mo><mml:msubsup><mml:mtext>&#x003A6;</mml:mtext><mml:mrow><mml:msub><mml:mi>S</mml:mi><mml:mi>W</mml:mi></mml:msub><mml:mo stretchy='false'>(</mml:mo><mml:mi>m</mml:mi><mml:mo stretchy='false'>)</mml:mo></mml:mrow><mml:mn>1</mml:mn></mml:msubsup><mml:mo>&#x000B7;</mml:mo><mml:mo stretchy='false'>(</mml:mo><mml:msubsup><mml:mi>U</mml:mi><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:mi>m</mml:mi><mml:mo stretchy='false'>)</mml:mo></mml:mrow><mml:mn>1</mml:mn></mml:msubsup><mml:mo>&#x02299;</mml:mo><mml:msub><mml:mi>W</mml:mi><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:mi>m</mml:mi><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:msub><mml:mo stretchy='false'>)</mml:mo><mml:mo>+</mml:mo><mml:mo stretchy='false'>(</mml:mo><mml:msubsup><mml:mtext>&#x003A6;</mml:mtext><mml:mrow><mml:msub><mml:mi>S</mml:mi><mml:mi>b</mml:mi></mml:msub><mml:mo stretchy='false'>(</mml:mo><mml:mi>m</mml:mi><mml:mo stretchy='false'>)</mml:mo></mml:mrow><mml:mn>1</mml:mn></mml:msubsup><mml:mo>+</mml:mo><mml:msub><mml:mi>b</mml:mi><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:mi>m</mml:mi><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:msub><mml:mo stretchy='false'>)</mml:mo><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:mstyle><mml:mo>,</mml:mo></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:msubsup><mml:mi>V</mml:mi><mml:mi>a</mml:mi><mml:mn>2</mml:mn></mml:msubsup><mml:mo>=</mml:mo><mml:mstyle displaystyle='true'><mml:munderover><mml:mo>&#x02211;</mml:mo><mml:mrow><mml:mi>m</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>M</mml:mi></mml:munderover><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:mo stretchy='false'>(</mml:mo><mml:msubsup><mml:mtext>&#x003A6;</mml:mtext><mml:mrow><mml:msub><mml:mi>S</mml:mi><mml:mi>W</mml:mi></mml:msub><mml:mo stretchy='false'>(</mml:mo><mml:mi>m</mml:mi><mml:mo stretchy='false'>)</mml:mo></mml:mrow><mml:mn>2</mml:mn></mml:msubsup><mml:mo>&#x000B7;</mml:mo><mml:mo stretchy='false'>(</mml:mo><mml:msubsup><mml:mi>U</mml:mi><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:mi>m</mml:mi><mml:mo stretchy='false'>)</mml:mo></mml:mrow><mml:mn>2</mml:mn></mml:msubsup><mml:mo>&#x02299;</mml:mo><mml:msub><mml:mi>W</mml:mi><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:mi>m</mml:mi><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:msub><mml:mo stretchy='false'>)</mml:mo><mml:mo>+</mml:mo><mml:mo stretchy='false'>(</mml:mo><mml:msubsup><mml:mtext>&#x003A6;</mml:mtext><mml:mrow><mml:msub><mml:mi>S</mml:mi><mml:mi>b</mml:mi></mml:msub><mml:mo stretchy='false'>(</mml:mo><mml:mi>m</mml:mi><mml:mo stretchy='false'>)</mml:mo></mml:mrow><mml:mn>2</mml:mn></mml:msubsup><mml:mo>+</mml:mo><mml:msub><mml:mi>b</mml:mi><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:mi>m</mml:mi><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:msub><mml:mo stretchy='false'>)</mml:mo><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:mstyle><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where &#x003A6;<sub><italic>S</italic><sub><italic>W</italic></sub>(<italic>m</italic>)</sub> and &#x003A6;<sub><italic>S</italic><sub><italic>b</italic></sub>(<italic>m</italic>)</sub> denote the two factors corresponding to <italic>K</italic><sub>(<italic>m</italic>)</sub>. Equation (7) indicates that <italic>V</italic><sub><italic>b</italic></sub> of <italic>U</italic><sup>1</sup> and <italic>U</italic><sup>2</sup> are extracted with the same kernel, and then their own <italic>V</italic><sub><italic>a</italic></sub> are formed by diverse linear combinations of <italic>V</italic><sub><italic>b</italic></sub>, which are adjusted by <italic>SS</italic> factors. The above equation can be rearranged into</p>
<disp-formula id="E8"><label>(8)</label><mml:math id="M8"><mml:mtable columnalign='left'><mml:mtr><mml:mtd><mml:msubsup><mml:mi>V</mml:mi><mml:mi>a</mml:mi><mml:mn>1</mml:mn></mml:msubsup><mml:mo>=</mml:mo><mml:mstyle displaystyle='true'><mml:munderover><mml:mo>&#x02211;</mml:mo><mml:mrow><mml:mi>m</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>M</mml:mi></mml:munderover><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:mo stretchy='false'>(</mml:mo><mml:msubsup><mml:mi>U</mml:mi><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:mi>m</mml:mi><mml:mo stretchy='false'>)</mml:mo></mml:mrow><mml:mn>1</mml:mn></mml:msubsup><mml:mo>&#x02299;</mml:mo><mml:mo stretchy='false'>(</mml:mo><mml:msubsup><mml:mtext>&#x003A6;</mml:mtext><mml:mrow><mml:msub><mml:mi>S</mml:mi><mml:mi>W</mml:mi></mml:msub><mml:mo stretchy='false'>(</mml:mo><mml:mi>m</mml:mi><mml:mo stretchy='false'>)</mml:mo></mml:mrow><mml:mn>1</mml:mn></mml:msubsup><mml:mo>&#x000B7;</mml:mo><mml:msub><mml:mi>W</mml:mi><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:mi>m</mml:mi><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:msub><mml:mo stretchy='false'>)</mml:mo><mml:mo>+</mml:mo><mml:mo stretchy='false'>(</mml:mo><mml:msubsup><mml:mtext>&#x003A6;</mml:mtext><mml:mrow><mml:msub><mml:mi>S</mml:mi><mml:mi>b</mml:mi></mml:msub><mml:mo stretchy='false'>(</mml:mo><mml:mi>m</mml:mi><mml:mo stretchy='false'>)</mml:mo></mml:mrow><mml:mn>1</mml:mn></mml:msubsup><mml:mo>+</mml:mo><mml:msub><mml:mi>b</mml:mi><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:mi>m</mml:mi><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:msub><mml:mo stretchy='false'>)</mml:mo><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:mstyle><mml:mo>,</mml:mo></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:msubsup><mml:mi>V</mml:mi><mml:mi>a</mml:mi><mml:mn>2</mml:mn></mml:msubsup><mml:mo>=</mml:mo><mml:mstyle displaystyle='true'><mml:munderover><mml:mo>&#x02211;</mml:mo><mml:mrow><mml:mi>m</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>M</mml:mi></mml:munderover><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:mo stretchy='false'>(</mml:mo><mml:msubsup><mml:mi>U</mml:mi><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:mi>m</mml:mi><mml:mo stretchy='false'>)</mml:mo></mml:mrow><mml:mn>2</mml:mn></mml:msubsup><mml:mo>&#x02299;</mml:mo><mml:mo stretchy='false'>(</mml:mo><mml:msubsup><mml:mtext>&#x003A6;</mml:mtext><mml:mrow><mml:msub><mml:mi>S</mml:mi><mml:mi>W</mml:mi></mml:msub><mml:mo stretchy='false'>(</mml:mo><mml:mi>m</mml:mi><mml:mo stretchy='false'>)</mml:mo></mml:mrow><mml:mn>2</mml:mn></mml:msubsup><mml:mo>&#x000B7;</mml:mo><mml:msub><mml:mi>W</mml:mi><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:mi>m</mml:mi><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:msub><mml:mo stretchy='false'>)</mml:mo><mml:mo>+</mml:mo><mml:mo stretchy='false'>(</mml:mo><mml:msubsup><mml:mtext>&#x003A6;</mml:mtext><mml:mrow><mml:msub><mml:mi>S</mml:mi><mml:mi>b</mml:mi></mml:msub><mml:mo stretchy='false'>(</mml:mo><mml:mi>m</mml:mi><mml:mo stretchy='false'>)</mml:mo></mml:mrow><mml:mn>2</mml:mn></mml:msubsup><mml:mo>+</mml:mo><mml:msub><mml:mi>b</mml:mi><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:mi>m</mml:mi><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:msub><mml:mo stretchy='false'>)</mml:mo><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:mstyle><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>Equation (8) indicates the effect of <italic>SS</italic> factors on the kernel. The &#x003A6;<sub><italic>S</italic><sub><italic>W</italic></sub></sub> uniformly scales all the weights <italic>W</italic> of the kernel, while &#x003A6;<sub><italic>S</italic><sub><italic>b</italic></sub></sub> shifts the outputs in addition to the original bias <italic>b</italic>. A linear transformation can be observed between the trained models for these two datasets.</p>
<p>Although the flexible sequences in MRI can result in images with varying advanced features, we argue that there is high repeatability of basic features due to the standardized views of medical images such as the limited texture variants or small patches (Alzubaidi et al., <xref ref-type="bibr" rid="B2">2021</xref>). According to the hypothesis above, we speculate that there is a linear correlation relationship between the trained models for two related MR datasets. If these two are treated as source and target datasets separately, it is feasible to adjust the source model pre-trained on the source dataset, by <italic>SS</italic> factors to obtain the target model.</p></sec>
<sec>
<title>2.2.2. Scaling and shifting factors</title>
<p>Based on the hypothesis and analysis, we propose the LFT strategy, which is implemented using the <italic>SS</italic> factors. The <italic>SS</italic> block, containing two factors, is integrated with regular convolution into <italic>SSConv</italic> in the transfer phase. The structure of <italic>SSConv</italic> is shown in <xref ref-type="fig" rid="F2">Figure 2B</xref>. Taking a 3 &#x000D7; 3 filter as an example, the lower part of <italic>SSConv</italic> shows the regular convolution <italic>Conv</italic>. The upper part is <italic>SS</italic> block, which contains a single scaling factor &#x003A6;<sub><italic>S</italic><sub><italic>W</italic></sub></sub> corresponding to <italic>W</italic> with 9 elements, and a shift factor &#x003A6;<sub><italic>S</italic><sub><italic>b</italic></sub></sub> corresponding to <italic>b</italic>.</p>
<p>In the following, the complete training process of LFT is described in detail. Firstly, in the pre-training phase, the <italic>W</italic> and <italic>b</italic> are initialized randomly, which is identical to classic pre-training. We train the <italic>W</italic> and <italic>b</italic> on large-scale datasets, as represented by Equation (4). Next in the transfer phase, the new target model is initialized with the pre-trained weights, and the <italic>SS</italic> block is inserted into the model. Here &#x003A6;<sub><italic>S</italic><sub><italic>W</italic></sub></sub> and &#x003A6;<sub><italic>S</italic><sub><italic>b</italic></sub></sub> are initialized as 1 and 0, respectively. Therefore, feeding input <italic>U</italic><sup><italic>t</italic></sup> in the target domain into <italic>SSConv</italic> can be expressed as</p>
<disp-formula id="E9"><label>(9)</label><mml:math id="M9"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msubsup><mml:mrow><mml:mi>V</mml:mi></mml:mrow><mml:mrow><mml:mi>b</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:msup><mml:mrow><mml:mi>U</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msup><mml:mo>&#x02299;</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mtext>&#x003A6;</mml:mtext></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:mi>W</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msub><mml:mo>&#x000B7;</mml:mo><mml:mi>W</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x0002B;</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mtext>&#x003A6;</mml:mtext></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:mi>b</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msub><mml:mo>&#x0002B;</mml:mo><mml:mi>b</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>Then LFT requires training the target model with target data, as shown in <xref ref-type="fig" rid="F2">Figure 2B</xref>. It is worth noting that <italic>W</italic> and <italic>b</italic> are frozen, and only &#x003A6;<sub><italic>S</italic><sub><italic>W</italic></sub></sub> and &#x003A6;<sub><italic>S</italic><sub><italic>b</italic></sub></sub> are optimized, giving</p>
<disp-formula id="E10"><label>(10)</label><mml:math id="M10"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msubsup><mml:mrow><mml:mi>V</mml:mi></mml:mrow><mml:mrow><mml:mi>b</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:msup><mml:mrow><mml:mi>U</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msup><mml:mo>&#x02299;</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msubsup><mml:mrow><mml:mtext>&#x003A6;</mml:mtext></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:mi>W</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>&#x02032;</mml:mi></mml:mrow></mml:msubsup><mml:mo>&#x000B7;</mml:mo><mml:mi>W</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x0002B;</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msubsup><mml:mrow><mml:mtext>&#x003A6;</mml:mtext></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:mi>b</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>&#x02032;</mml:mi></mml:mrow></mml:msubsup><mml:mo>&#x0002B;</mml:mo><mml:mi>b</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where <inline-formula><mml:math id="M11"><mml:msubsup><mml:mrow><mml:mtext>&#x003A6;</mml:mtext></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:mi>W</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>&#x02032;</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> and <inline-formula><mml:math id="M12"><mml:msubsup><mml:mrow><mml:mtext>&#x003A6;</mml:mtext></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:mi>b</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>&#x02032;</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> indicate the optimized weights.</p>
<p><xref ref-type="fig" rid="F2">Figure 2</xref> visualizes the difference between FT and LFT in the transfer phase. Fine-tuning updates the complete values of <italic>W</italic> and <italic>b</italic>, while LFT updates only the <italic>SS</italic> factors. It is obvious that LFT reduces the number of tuning weights, which contributes to lessening the risk of overfitting in the case of small samples sizes. Moreover, LFT prevents the problem of &#x0201C;catastrophic forgetting&#x0201D; by fixing the pre-trained weights, which benefits in learning specific knowledge of the target data without forgetting the generic knowledge learned from the source data.</p></sec></sec>
<sec>
<title>2.3. Materials</title>
<sec>
<title>2.3.1. Network</title>
<p>Generative adversarial networks (GANs) have shown strong performance in modeling the prior distributions of images (Shaul et al., <xref ref-type="bibr" rid="B34">2020</xref>; Lv et al., <xref ref-type="bibr" rid="B26">2021b</xref>). Therefore, GANs have been widely studied in MRI reconstruction (Shitrit and Riklin Raviv, <xref ref-type="bibr" rid="B35">2017</xref>; Yang et al., <xref ref-type="bibr" rid="B41">2017</xref>; Mardani et al., <xref ref-type="bibr" rid="B27">2018</xref>; Quan et al., <xref ref-type="bibr" rid="B31">2018</xref>), among which, RefineGAN (Quan et al., <xref ref-type="bibr" rid="B31">2018</xref>) gets a superior performance (Lv et al., <xref ref-type="bibr" rid="B26">2021b</xref>). In view of this, we constructed SSGAN as the reconstruction network by referring to RefineGAN&#x00027;s residual bottleneck block and double chain structure. To evaluate the proposed LFT strategy, the <italic>SS</italic> block is inserted into the network in the transfer phase. In addition, our network only contains basic building units, such as convolutional layers, non-linear activation functions and residual connections to facilitate the experiments on generalization and transferability. <xref ref-type="fig" rid="F3">Figure 3</xref> provides an overview of the SSGAN architecture with the inserted <italic>SS</italic> block. For more details, please refer to RefineGAN for more details.</p>
<fig id="F3" position="float">
<label>Figure 3</label>
<caption><p>Overview of the SSGAN architecture. The generator <italic>G</italic> of SSGAN is composed of two residual U-net with 2 encoder (pink box) and 2 decoder (blue box) blocks. The architecture of the discriminator <italic>D</italic> is the same as the encoding path of <italic>G</italic>. The inputs of <italic>G</italic> are the ZF image (i) and ZF image (k), which come from different collections.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnins-17-1202143-g0003.tif"/>
</fig></sec>
<sec>
<title>2.3.2. Datasets</title>
<p>The IXI dataset<xref ref-type="fn" rid="fn0001"><sup>1</sup></xref> is a large publicly available MR brain image dataset of healthy subjects, and the extensive data assists in improving the performance of the original model. For pre-training the source model, 58,000 sagittal T1-weighted brain images were selected from the IXI dataset. The transfer performance of the source model was explored on three different target datasets: the private sagittal T2-weighted brain dataset I, the private axial T1-weighted brain dataset II and the FastMRI<xref ref-type="fn" rid="fn0002"><sup>2</sup></xref> knee dataset. These datasets differ from the source data in distributions ranging from small to large, allowing us to study the effects of the distribution deviation on the proposed method. The two private datasets were derived from the study (Jiang et al., <xref ref-type="bibr" rid="B17">2022</xref>), and all the ethical and experimental procedures were approved by the First Affiliated Hospital of University of Science and Technology of China (in accordance with the Declaration of Helsinki), under Application No. 2021 KY205. Detailed acquisition parameters can be found in the <xref ref-type="supplementary-material" rid="SM1">Supplementary material</xref>. Slices (256 &#x000D7; 256 pixels) were extracted from raw data as the standard reference images, and retrospective under-sampling was performed with the mask to obtain the under-sampled data. The mask was constructed by random row (1D) sampling from a normal distribution, with denser sampling in the central part of the k-space.</p></sec>
<sec>
<title>2.3.3. Performance evaluation</title>
<p>The results were evaluated by the Peak Signal-to-Noise Ratio (PSNR) and the Structure SIMilarity (SSIM) (Wang et al., <xref ref-type="bibr" rid="B39">2004</xref>) as they are commonly used measures for evaluating image reconstruction quality. Both metrics rely on a pixel-wise comparison between the fully-sampled image and the reconstructed result. In addition, weighted peak signal-to-noise ratio (WPSNR) (Gupta and Aggarwal, <xref ref-type="bibr" rid="B14">2009</xref>; Erfurt et al., <xref ref-type="bibr" rid="B10">2019</xref>) is considered to be a metric more compatible with human visual perception, and its results are included in the <xref ref-type="supplementary-material" rid="SM1">Supplementary material</xref>.</p></sec></sec></sec>
<sec id="s3">
<title>3. Experiment and results</title>
<sec>
<title>3.1. Experimental design</title>
<p>As shown in <xref ref-type="fig" rid="F4">Figure 4</xref>, we utilized LFT to transfer the pre-trained SSGAN (gray part) to reconstruct the under-sampled MR data from three target datasets mentioned in II.D (color part). Hence, three transfer scenarios were designed: (1) sagittal brain MR images with different contrasts (T1-weighted and T2-weighted); (2) brain MR images with different slicing directions (axial and sagittal planes); (3) MR images with different anatomical structures (brain and knee).</p>
<fig id="F4" position="float">
<label>Figure 4</label>
<caption><p>Three transfer scenarios of the pre-trained SSGAN. The gray part shows SSGAN was pre-trained on the IXI dataset, then transferred to reconstruct images with different contrasts (blue part, scenario 1: T1-weighted to T2-weighted), slicing directions (purple part, scenario 2: axial planes to sagittal planes), and anatomical structures (green part, scenario 3: brain to knee).</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnins-17-1202143-g0004.tif"/>
</fig>
<p>The model transferred by LFT strategy was termed linear fine-tuning model (LFT model). For comparison, the following models were set up: (1) testing the target data directly with pre-trained model (PT model) to verify the necessity of transfer learning; (2) training the model directly (DT model) from scratch with the target data to prove the effectiveness of transfer learning; (3) transferring the PT model by FT, termed fine-tuning model (FT model).</p>
<p>Additionally, experiments were conducted at different sampling rates of 30, 40, and 50% to examine the robustness of the proposed method. We also investigated the effect of different sizes of training sets to validate the feasibility of LFT in the case of few samples.</p></sec>
<sec>
<title>3.2. Implementation details</title>
<p>All the models in this work were implemented using PyTorch framework on NVIDIA Geforce GTX 3090 with 24 GB memory. Adam optimizer (Kingma and Ba, <xref ref-type="bibr" rid="B20">2014</xref>) with an initial learning rate of 10<sup>&#x02212;4</sup> was applied for pre-training, and lowered the initial learning rate depending on the target datasets in the transfer phase. We stopped the network training based on the convergence criterion that the PSNR on the validation set does not increase within 15 epochs. The datasets used for the experiments were divided into training, validation and test datasets in the ratio of 16:5:4, but not all in the training set were used every time to examine the effect of the training set size.</p></sec>
<sec>
<title>3.3. Validation of the hypothesis</title>
<p>To verify the hypothesis in Section 2.2, the feature maps, i.e., advanced features extracted by filters of the same channel for four cases are visualized in <xref ref-type="fig" rid="F5">Figure 5</xref>, including: advanced features obtained by (a) feeding the source data into the PT model; (b) feeding the target data into the PT model; (c) feeding the target data into the FT model; (d) feeding the target data into the LFT model.</p>
<fig id="F5" position="float">
<label>Figure 5</label>
<caption><p>Comparison of advanced features extracted by filters of the same channel in four cases: advanced features obtained by <bold>(A)</bold> feeding the source data into the PT model; <bold>(B)</bold> feeding the target data into the PT model; <bold>(C)</bold> feeding the target data into the FT model; <bold>(D)</bold> feeding the target data into the LFT model. Source data is from IXI dataset, while target data are from private sagittal brain dataset I, private axial brain dataset II, and FastMRI knee dataset.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnins-17-1202143-g0005.tif"/>
</fig>
<p>As shown in column (a), the PT model extracted the out-of-contour artifacts (the first row) and in-contour artifacts (the last row) of source data, which presented the functions of these two pre-trained filters, respectively. The features in column (b) were the results of the PT model tested directly on the target data and were of different types than that in column (a) obviously. We attribute this to the fact that although the data in the source and target domains have similar basic features, the combination coefficients of advanced features are variable. Inappropriate coefficients prevented the PT model from extracting the expected advanced features of the target data. Column (c) shows the FT model restored the filters to extract the expected features, as in column (a). Although FT adjusted all the weights, it served to restore the original function of the filter, instead of relearning to extract new types. Column (d) displays the features of applying the LFT model to the target data. It is found that it also succeeded in extracting the desired advanced feature types. In particular, the LFT model adjusted the suitable coefficients for linear combinations of basic features by <italic>SS</italic> factors. This demonstrates that the hypothesis is reasonable and the LFT can achieve model transfer as FT with fewer tuning weights.</p></sec>
<sec>
<title>3.4. Results and analysis</title>
<p>In the pre-training phase, we trained three SSGANs with different under-sampling rates on the large-scale source IXI dataset. The networks reached the convergence stage within a few dozen iterations and achieved satisfactory reconstructions on the test data. <xref ref-type="table" rid="T1">Table 1</xref> presents the quantitative evaluation of both PSNR and SSIM, where the values of PSNR and SSIM are displayed in the form of mean &#x000B1; standard deviation. The zero-filled model (ZF model) indicates the zero-filled reconstruction of the under-sampled k-space measurement. Each network with the highest PSNR was saved as the basis for further transfer and named as PT model.</p>
<table-wrap position="float" id="T1">
<label>Table 1</label>
<caption><p>Performance evaluation of PT models on source dataset.</p></caption> 
<table frame="box" rules="all">
<thead>
<tr style="background-color:919498;color:ffffff">
<th valign="top" align="left"><bold>PSNR/SSIM</bold></th>
<th valign="top" align="center" colspan="2"><bold>30%</bold></th>
<th valign="top" align="center" colspan="2"><bold>40%</bold></th>
<th valign="top" align="center" colspan="2"><bold>50%</bold></th>
</tr>
<tr style="background-color:919498;color:ffffff">
<td valign="top" align="left"><bold>model</bold></td>
<td valign="top" align="center"><bold>PSNR</bold></td>
<td valign="top" align="center"><bold>SSIM</bold></td>
<td valign="top" align="center"><bold>PSNR</bold></td>
<td valign="top" align="center"><bold>SSIM</bold></td>
<td valign="top" align="center"><bold>PSNR</bold></td>
<td valign="top" align="center"><bold>SSIM</bold></td>
</tr> 
</thead>
<tbody>
<tr>
<td valign="top" align="left">ZF model<sup><italic>a</italic></sup></td>
<td valign="top" align="center">25.78 &#x000B1; 2.35</td>
<td valign="top" align="center">0.691 &#x000B1; 0.046</td>
<td valign="top" align="center">26.17 &#x000B1; 2.32</td>
<td valign="top" align="center">0.696 &#x000B1; 0.046</td>
<td valign="top" align="center">28.72 &#x000B1; 2.23</td>
<td valign="top" align="center">0.745 &#x000B1; 0.046</td>
</tr>
<tr>
<td valign="top" align="left">PT model<sup><italic>b</italic></sup></td>
<td valign="top" align="center">39.02 &#x000B1; 2.81</td>
<td valign="top" align="center">0.981 &#x000B1; 0.008</td>
<td valign="top" align="center">39.83 &#x000B1; 3.12</td>
<td valign="top" align="center">0.982 &#x000B1; 0.008</td>
<td valign="top" align="center">42.24 &#x000B1; 3.15</td>
<td valign="top" align="center">0.988 &#x000B1; 0.005</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p><sup><italic>a</italic></sup>Zero-filled model. <sup><italic>b</italic></sup>Pre-trained model.</p>
</table-wrap-foot>
</table-wrap>
<p>In transfer scenario 1, the target data are sagittal T2-weighted brain images, with different contrasts from the source data. There is not much deviation between their distributions. <xref ref-type="table" rid="T2">Table 2</xref> demonstrates the evaluation metrics of different reconstructions for the target data, with bolded indicating the best. The mean SSIM and PSNR values were both improved for two transfer models compared to the original PT and DT model. It indicates the necessity and effectiveness of transfer learning in the presence of contrast variation between the source and target data. In addition, the highest PSNR and SSIM were obtained by the LFT model at various under-sampling rates, reflecting the superiority and robustness of the proposed method. <xref ref-type="fig" rid="F6">Figure 6</xref> provides the reconstructions of each model. It can be seen that the reconstruction of the DT model had notable artifacts, indicating that training the network from scratch could not achieve a good result when there were only a few target samples. Instead, PT, FT, and LFT models basically succeeded in reconstructing, but the results of the LFT model achieved artifact minimization, which is more obvious at the red arrow indication in the error maps. Therefore, LFT model provided the most desirable reconstruction when transferring the network to data with different contrasts.</p>
<table-wrap position="float" id="T2">
<label>Table 2</label>
<caption><p>Performance evaluation of different reconstructions for the target data with different contrasts.</p></caption> 
<table frame="box" rules="all">
<thead>
<tr style="background-color:919498;color:ffffff">
<th valign="top" align="left"><bold>PSNR/SSIM</bold></th>
<th valign="top" align="center" colspan="2"><bold>30%</bold></th>
<th valign="top" align="center" colspan="2"><bold>40%</bold></th>
<th valign="top" align="center" colspan="2"><bold>50%</bold></th>
</tr>
<tr style="background-color:919498;color:ffffff">
<td valign="top" align="left"><bold>model</bold></td>
<td valign="top" align="center"><bold>PSNR</bold></td>
<td valign="top" align="center"><bold>SSIM</bold></td>
<td valign="top" align="center"><bold>PSNR</bold></td>
<td valign="top" align="center"><bold>SSIM</bold></td>
<td valign="top" align="center"><bold>PSNR</bold></td>
<td valign="top" align="center"><bold>SSIM</bold></td>
</tr> 
</thead>
<tbody>
<tr>
<td valign="top" align="left">ZF model<sup><italic>a</italic></sup></td>
<td valign="top" align="center">31.10 &#x000B1; 2.73</td>
<td valign="top" align="center">0.850 &#x000B1; 0.048</td>
<td valign="top" align="center">31.90 &#x000B1; 2.92</td>
<td valign="top" align="center">0.853 &#x000B1; 0.049</td>
<td valign="top" align="center">33.88 &#x000B1; 2.76</td>
<td valign="top" align="center">0.891 &#x000B1; 0.038</td>
</tr> <tr>
<td valign="top" align="left">PT model<sup><italic>b</italic></sup></td>
<td valign="top" align="center">40.45 &#x000B1; 2.39</td>
<td valign="top" align="center">0.973 &#x000B1; 0.007</td>
<td valign="top" align="center">41.28 &#x000B1; 2.56</td>
<td valign="top" align="center">0.978 &#x000B1; 0.008</td>
<td valign="top" align="center">44.52 &#x000B1; 2.61</td>
<td valign="top" align="center">0.980 &#x000B1; 0.003</td>
</tr> <tr>
<td valign="top" align="left">DT model<sup><italic>c</italic></sup></td>
<td valign="top" align="center">37.45 &#x000B1; 2.61</td>
<td valign="top" align="center">0.959 &#x000B1; 0.013</td>
<td valign="top" align="center">37.43 &#x000B1; 2.77</td>
<td valign="top" align="center">0.956 &#x000B1; 0.015</td>
<td valign="top" align="center">39.82 &#x000B1; 2.60</td>
<td valign="top" align="center">0.971 &#x000B1; 0.010</td>
</tr> <tr>
<td valign="top" align="left">FT model<sup><italic>d</italic></sup></td>
<td valign="top" align="center">40.72 &#x000B1; 2.54</td>
<td valign="top" align="center">0.981 &#x000B1; 0.008</td>
<td valign="top" align="center">42.35 &#x000B1; 2.73</td>
<td valign="top" align="center">0.984 &#x000B1; 0.008</td>
<td valign="top" align="center">44.78 &#x000B1; 2.69</td>
<td valign="top" align="center">0.991 &#x000B1; 0.004</td>
</tr>
<tr>
<td valign="top" align="left">LFT model<sup><italic>e</italic></sup></td>
<td valign="top" align="center"><bold>41.45</bold> <bold>&#x000B1;2.52</bold></td>
<td valign="top" align="center"><bold>0.984</bold> <bold>&#x000B1;0.008</bold></td>
<td valign="top" align="center"><bold>42.52</bold> <bold>&#x000B1;2.78</bold></td>
<td valign="top" align="center"><bold>0.985</bold> <bold>&#x000B1;0.008</bold></td>
<td valign="top" align="center"><bold>45.84</bold> <bold>&#x000B1;2.64</bold></td>
<td valign="top" align="center"><bold>0.992</bold> <bold>&#x000B1;0.003</bold></td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p><sup><italic>a</italic></sup>Zero-filled model. <sup><italic>b</italic></sup>Pre-trained model. <sup><italic>c</italic></sup>Directly trained model. <sup><italic>d</italic></sup>Fine-tuning model. <sup><italic>e</italic></sup>Linear fine-tuning model. The bold values indicate the best values of the evaluation metrics of different reconstructions for the target data.</p>
</table-wrap-foot>
</table-wrap>
<fig id="F6" position="float">
<label>Figure 6</label>
<caption><p>Typical reconstructions for sagittal T2-weighted brains from private dataset I by different methods. The last three models trained with 160 images at 50% sampling rate. From left to right are the results of: ground truth, zero-filled model, pre-trained model, directly trained model, fine-tuning model, and linear fine-tuning model, as well as their 10&#x000D7; magnified error maps.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnins-17-1202143-g0006.tif"/>
</fig>
<p>As for transfer scenario 2, the target data are axial T1-weighted brain images. They are sliced in different directions from the source data, and the variance in the data distribution leads to some respective features. <xref ref-type="table" rid="T3">Table 3</xref> shows the quantitative results of different reconstructions on the target data. The bolded rows mark the best results obtained by the LFT model, resolving the variation in different slicing directions. It is noteworthy that the LFT model demonstrated high reconstruction quality for small training sets, while the FT model showed a clear performance degradation as the sample size decreased. As shown in the results of the 100 training images, the LFT model reflects the most visible advantage over the FT model, with a 2.06 dB (5.89%) improvement in PSNR. <xref ref-type="fig" rid="F7">Figure 7</xref> presents intuitive reconstructions for the target data. A slight difference can be observed from the red arrow indication in the error map, which represents the LFT model outperforming the others in scenario 2.</p>
<table-wrap position="float" id="T3">
<label>Table 3</label>
<caption><p>Performance evaluation of different reconstructions for the target data with different slicing directions at 30% sampling rate.</p></caption> 
<table frame="box" rules="all">
<thead>
<tr style="background-color:919498;color:ffffff">
<th valign="top" align="left"><bold>PSNR/SSIM</bold></th>
<th valign="top" align="center" colspan="2"><bold>100 images</bold></th>
<th valign="top" align="center" colspan="2"><bold>200 images</bold></th>
<th valign="top" align="center" colspan="2"><bold>400 images</bold></th>
<th valign="top" align="center" colspan="2"><bold>800 images</bold></th>
</tr>
<tr style="background-color:919498;color:ffffff">
<td valign="top" align="left"><bold>model</bold></td>
<td valign="top" align="center"><bold>PSNR</bold></td>
<td valign="top" align="center"><bold>SSIM</bold></td>
<td valign="top" align="center"><bold>PSNR</bold></td>
<td valign="top" align="center"><bold>SSIM</bold></td>
<td valign="top" align="center"><bold>PSNR</bold></td>
<td valign="top" align="center"><bold>SSIM</bold></td>
<td valign="top" align="center"><bold>PSNR</bold></td>
<td valign="top" align="center"><bold>SSIM</bold></td>
</tr> 
</thead>
<tbody>
<tr>
<td valign="top" align="left">ZF model<sup><italic>a</italic></sup></td>
<td valign="top" align="center">23.88 &#x000B1; 3.47</td>
<td valign="top" align="center">0.806 &#x000B1; 0.040</td>
<td valign="top" align="center">23.88 &#x000B1; 3.47</td>
<td valign="top" align="center">0.806 &#x000B1; 0.040</td>
<td valign="top" align="center">23.88 &#x000B1; 3.47</td>
<td valign="top" align="center">0.806 &#x000B1; 0.040</td>
<td valign="top" align="center">23.88 &#x000B1; 3.47</td>
<td valign="top" align="center">0.806 &#x000B1; 0.040</td>
</tr> <tr>
<td valign="top" align="left">PT model<sup><italic>b</italic></sup></td>
<td valign="top" align="center">34.33 &#x000B1; 2.67</td>
<td valign="top" align="center">0.959 &#x000B1; 0.009</td>
<td valign="top" align="center">34.33 &#x000B1; 2.67</td>
<td valign="top" align="center">0.959 &#x000B1; 0.009</td>
<td valign="top" align="center">34.33 &#x000B1; 2.67</td>
<td valign="top" align="center">0.959 &#x000B1; 0.009</td>
<td valign="top" align="center">34.33 &#x000B1; 2.67</td>
<td valign="top" align="center">0.959 &#x000B1; 0.009</td>
</tr> <tr>
<td valign="top" align="left">DT model<sup><italic>c</italic></sup></td>
<td valign="top" align="center">31.46 &#x000B1; 2.84</td>
<td valign="top" align="center">0.925 &#x000B1; 0.016</td>
<td valign="top" align="center">32.62 &#x000B1; 2.87</td>
<td valign="top" align="center">0.941 &#x000B1; 0.014</td>
<td valign="top" align="center">33.92 &#x000B1; 2.77</td>
<td valign="top" align="center">0.958 &#x000B1; 0.011</td>
<td valign="top" align="center">34.86 &#x000B1; 2.71</td>
<td valign="top" align="center">0.964 &#x000B1; 0.011</td>
</tr> <tr>
<td valign="top" align="left">FT model<sup><italic>d</italic></sup></td>
<td valign="top" align="center">34.94 &#x000B1; 2.61</td>
<td valign="top" align="center">0.950 &#x000B1; 0.026</td>
<td valign="top" align="center">35.05 &#x000B1; 2.61</td>
<td valign="top" align="center">0.951 &#x000B1; 0.027</td>
<td valign="top" align="center">35.21 &#x000B1; 2.61</td>
<td valign="top" align="center">0.952 &#x000B1; 0.027</td>
<td valign="top" align="center">35.40 &#x000B1; 2.56</td>
<td valign="top" align="center">0.950 &#x000B1; 0.030</td>
</tr> <tr>
<td valign="top" align="left">RFT model<sup><italic>e</italic></sup></td>
<td valign="top" align="center">36.82 &#x000B1; 2.51</td>
<td valign="top" align="center">0.952 &#x000B1; 0.042</td>
<td valign="top" align="center">36.88 &#x000B1; 2.48</td>
<td valign="top" align="center">0.952 &#x000B1; 0.047</td>
<td valign="top" align="center">36.95 &#x000B1; 2.45</td>
<td valign="top" align="center">0.954 &#x000B1; 0.051</td>
<td valign="top" align="center">37.00 &#x000B1; 2.49</td>
<td valign="top" align="center">0.955 &#x000B1; 0.050</td>
</tr>
<tr>
<td valign="top" align="left">LFT model<sup><italic>f</italic></sup></td>
<td valign="top" align="center"><bold>37.00</bold> <bold>&#x000B1;1.89</bold></td>
<td valign="top" align="center"><bold>0.966</bold> <bold>&#x000B1;0.012</bold></td>
<td valign="top" align="center"><bold>37.03</bold> <bold>&#x000B1;1.90</bold></td>
<td valign="top" align="center"><bold>0.966</bold> <bold>&#x000B1;0.015</bold></td>
<td valign="top" align="center"><bold>37.13</bold> <bold>&#x000B1;1.97</bold></td>
<td valign="top" align="center"><bold>0.968</bold> <bold>&#x000B1;0.018</bold></td>
<td valign="top" align="center"><bold>37.14</bold> <bold>&#x000B1;1.93</bold></td>
<td valign="top" align="center"><bold>0.968</bold> <bold>&#x000B1;0.013</bold></td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p><sup><italic>a</italic></sup>Zero-filled model. <sup><italic>b</italic></sup>Pre-trained model. <sup><italic>c</italic></sup>Directly trained model. <sup><italic>d</italic></sup>Fine-tuning model. <sup><italic>e</italic></sup>Row fine-tuning model. <sup><italic>f</italic></sup>Linear fine-tuning model. The bold values indicate the best values of the evaluation metrics of different reconstructions for the target data.</p>
</table-wrap-foot>
</table-wrap>
<fig id="F7" position="float">
<label>Figure 7</label>
<caption><p>Typical reconstructions for axial T1-weighted brain images from private dataset II by different models (the last three networks trained with 200 images at 50% sampling rate). From left to right are the results of: ground truth, zero-filled model, pre-trained model, directly trained model, fine-tuning model, and linear fine-tuning model.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnins-17-1202143-g0007.tif"/>
</fig>
<p>Furthermore, we consider FT as a way for each element in a kernel to learn a transformation factor, and LFT as a way for all elements in a kernel to learn a transformation factor. Therefore, the method of learning one transformation factor for all elements in a single row within the kernel is set up for comparison, named row fine-tuning (RFT) model. The results of RFT were added to <xref ref-type="table" rid="T3">Table 3</xref> as a comparison, and we can observe that RFT improves upon FT by reducing adjustment weights. However, LFT, which minimizes the number of updated weights, yields the best performance. It implies that the performance of the transfer can be improved by reducing the number of weight updates, and LFT proves to be the optimal choice.</p>
<p>We transferred the model pre-trained on the brain data to reconstruct knee images in scenario 3, which belong to different anatomical structures and vary greatly from the source data. The reconstruction indices for the knee data at 50% sampling rate are presented in <xref ref-type="table" rid="T4">Table 4</xref>. The results indicate that LFT model obtained the optimal quality in most cases, but occasionally, the FT model performed better. We consider this phenomenon reasonable due to the large variation in diverse anatomical structures. Relatively few similar features limited the validity of LFT method. However, FT adjusted the model more adequately after providing more training data, resulting in better performance. It is also reflects by the decreasing gap between the metrics of FT and LFT models as the number of training sets increased. Besides, both FT and LFT models deteriorated the PT model due to overfitting when the dataset was extremely small. As the training data increased, LTF prioritized improving the situation. Typical reconstructions for knee images by different networks are shown in <xref ref-type="fig" rid="F8">Figure 8</xref>. Focusing on the error maps of the PT model here, there were more artifacts remaining in the background compared to the first two scenarios. This observation suggests that the PT model was more adaptable to the first two datasets as the data are more similar. In addition, the red arrow indicates that the LFT model had the least residual artifacts in the reconstruction images. Consequently, the LFT method is still the optimal transfer strategy when the source and limited target data vary widely.</p>
<table-wrap position="float" id="T4">
<label>Table 4</label>
<caption><p>Performance evaluation of different reconstructions for the target data with different anatomical structures at 50% sampling rate.</p></caption> 
<table frame="box" rules="all">
<thead>
<tr style="background-color:919498;color:ffffff">
<th valign="top" align="left"><bold>PSNR/SSIM</bold></th>
<th valign="top" align="center" colspan="2"><bold>100 images</bold></th>
<th valign="top" align="center" colspan="2"><bold>200 images</bold></th>
<th valign="top" align="center" colspan="2"><bold>400 images</bold></th>
<th valign="top" align="center" colspan="2"><bold>800 images</bold></th>
</tr>
<tr style="background-color:919498;color:ffffff">
<td valign="top" align="left"><bold>model</bold></td>
<td valign="top" align="center"><bold>PSNR</bold></td>
<td valign="top" align="center"><bold>SSIM</bold></td>
<td valign="top" align="center"><bold>PSNR</bold></td>
<td valign="top" align="center"><bold>SSIM</bold></td>
<td valign="top" align="center"><bold>PSNR</bold></td>
<td valign="top" align="center"><bold>SSIM</bold></td>
<td valign="top" align="center"><bold>PSNR</bold></td>
<td valign="top" align="center"><bold>SSIM</bold></td>
</tr> 
</thead>
<tbody>
<tr>
<td valign="top" align="left">ZF model<sup><italic>a</italic></sup></td>
<td valign="top" align="center">29.34 &#x000B1; 3.07</td>
<td valign="top" align="center">0.854 &#x000B1; 0.038</td>
<td valign="top" align="center">29.34 &#x000B1; 3.07</td>
<td valign="top" align="center">0.854 &#x000B1; 0.038</td>
<td valign="top" align="center">29.34 &#x000B1; 3.07</td>
<td valign="top" align="center">0.854 &#x000B1; 0.038</td>
<td valign="top" align="center">29.34 &#x000B1; 3.07</td>
<td valign="top" align="center">0.854 &#x000B1; 0.038</td>
</tr> <tr>
<td valign="top" align="left">PT model<sup><italic>b</italic></sup></td>
<td valign="top" align="center">35.35 &#x000B1; 2.86</td>
<td valign="top" align="center"><bold>0.934</bold> <bold>&#x000B1;0.025</bold></td>
<td valign="top" align="center">35.35 &#x000B1; 2.86</td>
<td valign="top" align="center">0.934 &#x000B1; 0.025</td>
<td valign="top" align="center">35.35 &#x000B1; 2.86</td>
<td valign="top" align="center">0.934 &#x000B1; 0.025</td>
<td valign="top" align="center">35.35 &#x000B1; 2.86</td>
<td valign="top" align="center">0.934 &#x000B1; 0.025</td>
</tr> <tr>
<td valign="top" align="left">DT model<sup><italic>c</italic></sup></td>
<td valign="top" align="center">32.51 &#x000B1; 2.86</td>
<td valign="top" align="center">0.907 &#x000B1; 0.024</td>
<td valign="top" align="center">33.92 &#x000B1; 2.75</td>
<td valign="top" align="center">0.922 &#x000B1; 0.021</td>
<td valign="top" align="center">34.76 &#x000B1; 2.64</td>
<td valign="top" align="center">0.930 &#x000B1; 0.020</td>
<td valign="top" align="center">35.14 &#x000B1; 2.59</td>
<td valign="top" align="center">0.934 &#x000B1; 0.018</td>
</tr> <tr>
<td valign="top" align="left">FT model<sup><italic>d</italic></sup></td>
<td valign="top" align="center">34.70 &#x000B1; 2.69</td>
<td valign="top" align="center">0.930 &#x000B1; 0.024</td>
<td valign="top" align="center">35.26 &#x000B1; 2.76</td>
<td valign="top" align="center">0.933 &#x000B1; 0.024</td>
<td valign="top" align="center">35.50 &#x000B1; 2.79</td>
<td valign="top" align="center"><bold>0.936</bold> <bold>&#x000B1;0.024</bold></td>
<td valign="top" align="center">35.86 &#x000B1; 2.72</td>
<td valign="top" align="center"><bold>0.936</bold> <bold>&#x000B1;0.024</bold></td>
</tr>
<tr>
<td valign="top" align="left">LFT model<sup><italic>e</italic></sup></td>
<td valign="top" align="center"><bold>35.64</bold> <bold>&#x000B1;2.33</bold></td>
<td valign="top" align="center">0.932 &#x000B1; 0.024</td>
<td valign="top" align="center"><bold>35.79</bold> <bold>&#x000B1;2.55</bold></td>
<td valign="top" align="center"><bold>0.934</bold> <bold>&#x000B1;0.024</bold></td>
<td valign="top" align="center"><bold>35.83</bold> <bold>&#x000B1;2.60</bold></td>
<td valign="top" align="center">0.934 &#x000B1; 0.024</td>
<td valign="top" align="center"><bold>35.96</bold> <bold>&#x000B1;2.66</bold></td>
<td valign="top" align="center">0.935 &#x000B1; 0.024</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p><sup><italic>a</italic></sup>Zero-filled model. <sup><italic>b</italic></sup>Pre-trained model. <sup><italic>c</italic></sup>Directly trained model. <sup><italic>d</italic></sup>Fine-tuning model. <sup><italic>e</italic></sup>Linear fine-tuning model. The bold values indicate the best values of the evaluation metrics of different reconstructions for the target data.</p>
</table-wrap-foot>
</table-wrap>
<fig id="F8" position="float">
<label>Figure 8</label>
<caption><p>Typical reconstructions for knee images from FastMRI dataset by different networks (the last three networks trained with 200 images at 50% sampling rate). From left to right are the results of: ground truth, zero-filled model, pre-trained model, directly trained model, fine-tuning model, and linear fine-tuning model.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnins-17-1202143-g0008.tif"/>
</fig></sec></sec>
<sec sec-type="discussion" id="s4">
<title>4. Discussion</title>
<p>This study centers on optimizing the FT strategy for MRI reconstruction. We conducted a comparative analysis of FT, LFT, and other transfer strategies in various transfer scenarios based on the reconstruction quality. Fine-tuning yields suboptimal reconstruction quality and appreciable residual artifacts, especially when the target domain training set contains fewer than 200 images in the anatomical structure transfer scenario, whereby the FT model performs worse than the PT model. We attribute the result to &#x0201C;catastrophic forgetting&#x0201D; and overfitting. Fine-tuning involves the update of all weights, which inevitably forgets the pre-trained knowledge. Moreover, as the target domain typically has limited data, updating a large number of weights is prone to overfitting (Sun et al., <xref ref-type="bibr" rid="B36">2019</xref>). Obviously, reducing the number of updated weights is crucial to improve the FT&#x00027;s performance. Several studies (Tajbakhsh et al., <xref ref-type="bibr" rid="B37">2016</xref>; Amiri et al., <xref ref-type="bibr" rid="B3">2020</xref>) have attempted to update only a subset of network weights, but the required level of tuning differs from one application to another. Hence, this approach is only applicable to certain transfer scenarios, limiting its scalability. The proposed LFT provides a new perspective of decoupling the pre-trained and to-be-updated weights by introducing learnable <italic>SS</italic> factors. While completely fixing the pre-trained weights, LFT can accomplish the transfer task by updating the weights fewer than those in the FT, thereby avoiding forgetting and overfitting. Consequently, LFT achieves more competitive results in various transfer scenarios.</p>
<p>Comprehensively analyzing the results of multiple transfer scenarios, it is found that with only 100 training images in the target domain, LFT outperforms FT with 400 training images. This implies that in the practical deployment, LFT can construct multiple reconstruction models at a lower cost and shorter development cycles to adapt to various complicated clinical scenarios. In addition, the <italic>SS</italic> block can be integrated into existing convolutional neural networks in a plug-and-play fashion, further enhancing the clinical applicability of LFT.</p>
<p>This study has some limitations. Firstly, we have exclusively assessed the proposed method using the linear sampling pattern of the Cartesian k-space trajectory, owing to its prevalence. Additional research is necessary to examine the feasibility of utilizing LFT in non-linear scenarios, such as radial sampling patterns. Secondly, deep MRI reconstruction is performed using retrospectively under-sampled data, which deviates from clinical routine, so prospective validation is required. Thirdly, LFT introduces a few additional parameters to the original model, causing an increase in the latency of inference. The comparison of inference time with and without additional parameters is shown in the <xref ref-type="supplementary-material" rid="SM1">Supplementary material</xref>. Despite a slight extension of reconstruction time, it does not considerably affect the overall MRI process in most clinical applications, as data acquisition and reconstruction can be executed asynchronously. Lastly, our source domain dataset only contains brain images, that is, the transfer effects were verified only from brain MR images to other scenarios. In future work, we intend to gather a more diverse range of image types to thoroughly evaluate the transferability of the LFT strategy across various scenarios.</p></sec>
<sec sec-type="conclusions" id="s5">
<title>5. Conclusion</title>
<p>To address the issues of &#x0201C;catastrophic forgetting&#x0201D; and overfitting in FT for MRI reconstruction, we have developed a novel transfer strategy, LFT, which is predicated on a linear transformation hypothesis. By focusing on optimizing the <italic>SS</italic> factors, as opposed to all weights, LFT achieves performance on par with FT, while requiring fewer training samples from the target domain. When applying deep learning for MRI reconstruction in diverse and complicated clinical scenarios, engineers only need to create a general pre-training model using MR images from various sequences and body parts, and subsequently gather a small quantity of images in the target scene. The LFT approach can then be employed to derive a customized reconstruction model with satisfactory performance. Additionally, as the LFT method can be seamlessly integrated with any reconstruction convolutional network, it does not limit the choice of architecture during the development phase. To conclude, LFT greatly enhances the viability of deep MRI reconstruction in scenarios with limited data.</p></sec>
<sec sec-type="data-availability" id="s6">
<title>Data availability statement</title>
<p>Publicly available datasets were analyzed in this study. This data can be found at: <ext-link ext-link-type="uri" xlink:href="http://brain-development.org/ixi-dataset/">http://brain-development.org/ixi-dataset/</ext-link> and <ext-link ext-link-type="uri" xlink:href="https://fastmri.org/dataset">https://fastmri.org/dataset</ext-link>.</p></sec>
<sec sec-type="ethics-statement" id="s7">
<title>Ethics statement</title>
<p>The studies involving human participants were reviewed and approved by the First Affiliated Hospital of University of Science and Technology of China. No. 2021 KY205. The patients/participants provided their written informed consent to participate in this study.</p></sec>
<sec sec-type="author-contributions" id="s8">
<title>Author contributions</title>
<p>WB: methodology, conducting the experiments, and writing&#x02014;original draft preparation. JX: implementing the algorithm. MS: data curation. XH: methodology, supervision, and writing&#x02014;review and editing. DG: guidance and supervision. FQ: methodology and writing&#x02014;review and editing. All authors contributed to the article and approved the submitted version.</p></sec>
</body>
<back>
<sec sec-type="funding-information" id="s9">
<title>Funding</title>
<p>This work was supported by a fund of the University Synergy Innovation Program of Anhui Province GXXT-2021-003 from Institute of Artificial Intelligence, Hefei Comprehensive National Science Center, and the National Science and Technology Innovation 2030 Major Program 2022ZD0204801.</p>
</sec>
<sec sec-type="COI-statement" id="conf1">
<title>Conflict of interest</title>
<p>XH was employed by Fuqing Medical Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="disclaimer" id="s10">
<title>Publisher&#x00027;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec><sec sec-type="supplementary-material" id="s11">
<title>Supplementary material</title>
<p>The Supplementary Material for this article can be found online at: <ext-link ext-link-type="uri" xlink:href="https://www.frontiersin.org/articles/10.3389/fnins.2023.1202143/full#supplementary-material">https://www.frontiersin.org/articles/10.3389/fnins.2023.1202143/full#supplementary-material</ext-link></p>
<supplementary-material xlink:href="Data_Sheet_1.PDF" id="SM1" mimetype="application/pdf" xmlns:xlink="http://www.w3.org/1999/xlink"/></sec>
<fn-group>
<fn id="fn0001"><p><sup>1</sup><ext-link ext-link-type="uri" xlink:href="http://brain-development.org/ixi-dataset/">http://brain-development.org/ixi-dataset/</ext-link></p></fn>
<fn id="fn0002"><p><sup>2</sup><ext-link ext-link-type="uri" xlink:href="https://fastmri.org/dataset/">https://fastmri.org/dataset/</ext-link></p></fn>
</fn-group>
<ref-list>
<title>References</title>
<ref id="B1">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Aletras</surname> <given-names>A. H.</given-names></name> <name><surname>Tilak</surname> <given-names>G. S.</given-names></name> <name><surname>Natanzon</surname> <given-names>A.</given-names></name> <name><surname>Hsu</surname> <given-names>L.-Y.</given-names></name> <name><surname>Gonzalez</surname> <given-names>F. M.</given-names></name> <name><surname>Hoyt</surname> <given-names>R. F.</given-names> <suffix>Jr.</suffix></name> <etal/></person-group>. (<year>2006</year>). <article-title>Retrospective determination of the area at risk for reperfused acute myocardial infarction with t2-weighted cardiac magnetic resonance imaging: histopathological and displacement encoding with stimulated echoes (dense) functional validations</article-title>. <source>Circulation</source> <volume>113</volume>, <fpage>1865</fpage>&#x02013;<lpage>1870</lpage>. <pub-id pub-id-type="doi">10.1161/CIRCULATIONAHA.105.576025</pub-id><pub-id pub-id-type="pmid">16606793</pub-id></citation></ref>
<ref id="B2">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Alzubaidi</surname> <given-names>L.</given-names></name> <name><surname>Al-Amidie</surname> <given-names>M.</given-names></name> <name><surname>Al-Asadi</surname> <given-names>A.</given-names></name> <name><surname>Humaidi</surname> <given-names>A. J.</given-names></name> <name><surname>Al-Shamma</surname> <given-names>O.</given-names></name> <name><surname>Fadhel</surname> <given-names>M. A.</given-names></name> <etal/></person-group>. (<year>2021</year>). <article-title>Novel transfer learning approach for medical imaging with limited labeled data</article-title>. <source>Cancers</source> <volume>13</volume>, <fpage>1590</fpage>. <pub-id pub-id-type="doi">10.3390/cancers13071590</pub-id><pub-id pub-id-type="pmid">33808207</pub-id></citation></ref>
<ref id="B3">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Amiri</surname> <given-names>M.</given-names></name> <name><surname>Brooks</surname> <given-names>R.</given-names></name> <name><surname>Rivaz</surname> <given-names>H.</given-names></name></person-group> (<year>2020</year>). <article-title>Fine-tuning U-Net for ultrasound image segmentation: different layers, different outcomes</article-title>. <source>IEEE Trans. Ultrasonics Ferroelectr. Frequency Control</source> <volume>67</volume>, <fpage>2510</fpage>&#x02013;<lpage>2518</lpage>. <pub-id pub-id-type="doi">10.1109/TUFFC.2020.3015081</pub-id><pub-id pub-id-type="pmid">32763853</pub-id></citation></ref>
<ref id="B4">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Antun</surname> <given-names>V.</given-names></name> <name><surname>Renna</surname> <given-names>F.</given-names></name> <name><surname>Poon</surname> <given-names>C.</given-names></name> <name><surname>Adcock</surname> <given-names>B.</given-names></name> <name><surname>Hansen</surname> <given-names>A. C.</given-names></name></person-group> (<year>2020</year>). <article-title>On instabilities of deep learning in image reconstruction and the potential costs of AI</article-title>. <source>Proc. Natl. Acad. Sci. U.S.A.</source> <volume>117</volume>, <fpage>30088</fpage>&#x02013;<lpage>30095</lpage>. <pub-id pub-id-type="doi">10.1073/pnas.1907377117</pub-id><pub-id pub-id-type="pmid">32393633</pub-id></citation></ref>
<ref id="B5">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Arshad</surname> <given-names>M.</given-names></name> <name><surname>Qureshi</surname> <given-names>M.</given-names></name> <name><surname>Inam</surname> <given-names>O.</given-names></name> <name><surname>Omer</surname> <given-names>H.</given-names></name></person-group> (<year>2021</year>). <article-title>Transfer learning in deep neural network based under-sampled MR image reconstruction</article-title>. <source>Magnet. Reson. Imaging</source> <volume>76</volume>, <fpage>96</fpage>&#x02013;<lpage>107</lpage>. <pub-id pub-id-type="doi">10.1016/j.mri.2020.09.018</pub-id><pub-id pub-id-type="pmid">32980504</pub-id></citation></ref>
<ref id="B6">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Block</surname> <given-names>K. T.</given-names></name> <name><surname>Uecker</surname> <given-names>M.</given-names></name> <name><surname>Frahm</surname> <given-names>J.</given-names></name></person-group> (<year>2007</year>). <article-title>Undersampled radial MRI with multiple coils. Iterative image reconstruction using a total variation constraint</article-title>. <source>Magnet. Reson. Med.</source> <volume>57</volume>, <fpage>1086</fpage>&#x02013;<lpage>1098</lpage>. <pub-id pub-id-type="doi">10.1002/mrm.21236</pub-id><pub-id pub-id-type="pmid">17534903</pub-id></citation></ref>
<ref id="B7">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Cole</surname> <given-names>E.</given-names></name> <name><surname>Cheng</surname> <given-names>J.</given-names></name> <name><surname>Pauly</surname> <given-names>J.</given-names></name> <name><surname>Vasanawala</surname> <given-names>S.</given-names></name></person-group> (<year>2021</year>). <article-title>Analysis of deep complex-valued convolutional neural networks for MRI reconstruction and phase-focused applications</article-title>. <source>Magnet. Reson. Med.</source> <volume>86</volume>, <fpage>1093</fpage>&#x02013;<lpage>1109</lpage>. <pub-id pub-id-type="doi">10.1002/mrm.28733</pub-id><pub-id pub-id-type="pmid">33724507</pub-id></citation></ref>
<ref id="B8">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Dar</surname> <given-names>S. U. H.</given-names></name> <name><surname>&#x000D6;zbey</surname> <given-names>M.</given-names></name> <name><surname>&#x000C7;atl&#x00131;</surname> <given-names>A. B.</given-names></name> <name><surname>&#x000C7;ukur</surname> <given-names>T.</given-names></name></person-group> (<year>2020</year>). <article-title>A transfer-learning approach for accelerated MRI using deep neural networks</article-title>. <source>Magnet. Reson. Med.</source> <volume>84</volume>, <fpage>663</fpage>&#x02013;<lpage>685</lpage>. <pub-id pub-id-type="doi">10.1002/mrm.28148</pub-id><pub-id pub-id-type="pmid">31898840</pub-id></citation></ref>
<ref id="B9">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Eitel</surname> <given-names>I.</given-names></name> <name><surname>Friedrich</surname> <given-names>M. G.</given-names></name></person-group> (<year>2011</year>). <article-title>T2-weighted cardiovascular magnetic resonance in acute cardiac disease</article-title>. <source>J. Cardiovasc. Magnet. Reson.</source> <volume>13</volume>, <fpage>1</fpage>&#x02013;<lpage>11</lpage>. <pub-id pub-id-type="doi">10.1186/1532-429X-13-13</pub-id><pub-id pub-id-type="pmid">21332972</pub-id></citation></ref>
<ref id="B10">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Erfurt</surname> <given-names>J.</given-names></name> <name><surname>Helmrich</surname> <given-names>C. R.</given-names></name> <name><surname>Bosse</surname> <given-names>S.</given-names></name> <name><surname>Schwarz</surname> <given-names>H.</given-names></name> <name><surname>Marpe</surname> <given-names>D.</given-names></name> <name><surname>Wiegand</surname> <given-names>T.</given-names></name></person-group> (<year>2019</year>). <article-title>&#x0201C;A study of the perceptually weighted peak signal-to-noise ratio (WPSNR) for image compression,&#x0201D;</article-title> in <source>2019 IEEE International Conference on Image Processing (ICIP)</source> (<publisher-loc>Taipei</publisher-loc>: <publisher-name>IEEE</publisher-name>), <fpage>2339</fpage>&#x02013;<lpage>2343</lpage>. <pub-id pub-id-type="doi">10.1109/ICIP.2019.8803307</pub-id></citation>
</ref>
<ref id="B11">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Fr&#x000E9;gier</surname> <given-names>Y.</given-names></name> <name><surname>Gouray</surname> <given-names>J.-B.</given-names></name></person-group> (<year>2021</year>). <article-title>&#x0201C;Mind2mind: transfer learning for GANs,&#x0201D;</article-title> in <source>Geometric Science of Information: 5th International Conference, GSI 2021</source> (<publisher-loc>Paris</publisher-loc>: <publisher-name>Springer</publisher-name>), <fpage>851</fpage>&#x02013;<lpage>859</lpage>.</citation>
</ref>
<ref id="B12">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Gavrikov</surname> <given-names>P.</given-names></name> <name><surname>Keuper</surname> <given-names>J.</given-names></name></person-group> (<year>2022</year>). <article-title>&#x0201C;CNN filter DB: an empirical investigation of trained convolutional filters,&#x0201D;</article-title> in <source>Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition</source> (<publisher-loc>New Orleans, LA</publisher-loc>), <fpage>19066</fpage>&#x02013;<lpage>19076</lpage>.</citation>
</ref>
<ref id="B13">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Griswold</surname> <given-names>M. A.</given-names></name> <name><surname>Jakob</surname> <given-names>P. M.</given-names></name> <name><surname>Heidemann</surname> <given-names>R. M.</given-names></name> <name><surname>Nittka</surname> <given-names>M.</given-names></name> <name><surname>Jellus</surname> <given-names>V.</given-names></name> <name><surname>Wang</surname> <given-names>J.</given-names></name> <etal/></person-group>. (<year>2002</year>). <article-title>Generalized autocalibrating partially parallel acquisitions (GRAPPA)</article-title>. <source>Magn. Reson. Med.</source> <volume>47</volume>, <fpage>1202</fpage>&#x02013;<lpage>1210</lpage>. <pub-id pub-id-type="doi">10.1002/mrm.10171</pub-id><pub-id pub-id-type="pmid">18219627</pub-id></citation></ref>
<ref id="B14">
<citation citation-type="web"><person-group person-group-type="author"><name><surname>Gupta</surname> <given-names>G.</given-names></name> <name><surname>Aggarwal</surname> <given-names>H.</given-names></name></person-group> (<year>2009</year>). <article-title>Digital image watermarking using two dimensional discrete wavelet transform, discrete cosine transform and fast fourier transform</article-title>. <source>Int. J. Recent Trends Eng.</source> <volume>1</volume>, <fpage>616</fpage>. Available online at: <ext-link ext-link-type="uri" xlink:href="https://www.semanticscholar.org/paper/Digital-image-Watermarking-using-Two-Dimensional-%2C-Gupta-Aggarwal/e154863ba286a4e7cdad655cc723d4413b822566">https://www.semanticscholar.org/paper/Digital-image-Watermarking-using-Two-Dimensional-%2C-Gupta-Aggarwal/e154863ba286a4e7cdad655cc723d4413b822566</ext-link></citation>
</ref>
<ref id="B15">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hamilton</surname> <given-names>J.</given-names></name> <name><surname>Franson</surname> <given-names>D.</given-names></name> <name><surname>Seiberlich</surname> <given-names>N.</given-names></name></person-group> (<year>2017</year>). <article-title>Recent advances in parallel imaging for MRI</article-title>. <source>Prog. Nuclear Magn. Reson. Spectrosc.</source> <volume>101</volume>, <fpage>71</fpage>&#x02013;<lpage>95</lpage>. <pub-id pub-id-type="doi">10.1016/j.pnmrs.2017.04.002</pub-id><pub-id pub-id-type="pmid">28844222</pub-id></citation></ref>
<ref id="B16">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hosseini</surname> <given-names>S. A. H.</given-names></name> <name><surname>Zhang</surname> <given-names>C.</given-names></name> <name><surname>Weing&#x000E4;rtner</surname> <given-names>S.</given-names></name> <name><surname>Moeller</surname> <given-names>S.</given-names></name> <name><surname>Stuber</surname> <given-names>M.</given-names></name> <name><surname>Ugurbil</surname> <given-names>K.</given-names></name> <etal/></person-group>. (<year>2020</year>). <article-title>Accelerated coronary MRI with sRAKI: a database-free self-consistent neural network k-space reconstruction for arbitrary undersampling</article-title>. <source>PLoS ONE</source> <volume>15</volume>, <fpage>e0229418</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pone.0229418</pub-id><pub-id pub-id-type="pmid">32084235</pub-id></citation></ref>
<ref id="B17">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jiang</surname> <given-names>J.</given-names></name> <name><surname>Qi</surname> <given-names>F.</given-names></name> <name><surname>Du</surname> <given-names>H.</given-names></name> <name><surname>Xu</surname> <given-names>J.</given-names></name> <name><surname>Zhou</surname> <given-names>Y.</given-names></name> <name><surname>Gao</surname> <given-names>D.</given-names></name> <etal/></person-group>. (<year>2022</year>). <article-title>Super-resolution reconstruction of 3t-like images from 0.35 t MRI using a hybrid attention residual network</article-title>. <source>IEEE Access</source> <volume>10</volume>, <fpage>32810</fpage>&#x02013;<lpage>32821</lpage>. <pub-id pub-id-type="doi">10.1109/ACCESS.2022.3155226</pub-id></citation>
</ref>
<ref id="B18">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jiang</surname> <given-names>M.</given-names></name> <name><surname>Yuan</surname> <given-names>Z.</given-names></name> <name><surname>Yang</surname> <given-names>X.</given-names></name> <name><surname>Zhang</surname> <given-names>J.</given-names></name> <name><surname>Gong</surname> <given-names>Y.</given-names></name> <name><surname>Xia</surname> <given-names>L.</given-names></name> <name><surname>Li</surname> <given-names>T.</given-names></name></person-group> (<year>2019</year>). <article-title>Accelerating CS-MRI reconstruction with fine-tuning wasserstein generative adversarial network</article-title>. <source>IEEE Access</source> <volume>7</volume>, <fpage>152347</fpage>&#x02013;<lpage>152357</lpage>. <pub-id pub-id-type="doi">10.1109/ACCESS.2019.2948220</pub-id></citation>
</ref>
<ref id="B19">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Kamphenkel</surname> <given-names>J.</given-names></name> <name><surname>J&#x000E4;ger</surname> <given-names>P. F.</given-names></name> <name><surname>Bickelhaupt</surname> <given-names>S.</given-names></name> <name><surname>Laun</surname> <given-names>F. B.</given-names></name> <name><surname>Lederer</surname> <given-names>W.</given-names></name> <name><surname>Daniel</surname> <given-names>H.</given-names></name> <etal/></person-group>. (<year>2018</year>). <article-title>&#x0201C;Domain adaptation for deviating acquisition protocols in CNN-based lesion classification on diffusion-weighted MR images,&#x0201D;</article-title> in <source>Image Analysis for Moving Organ, Breast, and Thoracic Images: Third International Workshop, RAMBO 2018, Fourth International Workshop, BIA 2018, and First International Workshop, TIA 2018</source> (<publisher-loc>Granada</publisher-loc>: <publisher-name>Springer</publisher-name>), pages <fpage>73</fpage>&#x02013;<lpage>80</lpage>.</citation>
</ref>
<ref id="B20">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kingma</surname> <given-names>D. P.</given-names></name> <name><surname>Ba</surname> <given-names>J.</given-names></name></person-group> (<year>2014</year>). <article-title>Adam: a method for stochastic optimization</article-title>. <source>arXiv preprint arXiv:1412.6980</source>.</citation>
</ref>
<ref id="B21">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Knoll</surname> <given-names>F.</given-names></name> <name><surname>Hammernik</surname> <given-names>K.</given-names></name> <name><surname>Kobler</surname> <given-names>E.</given-names></name> <name><surname>Pock</surname> <given-names>T.</given-names></name> <name><surname>Recht</surname> <given-names>M. P.</given-names></name> <name><surname>Sodickson</surname> <given-names>D. K.</given-names></name></person-group> (<year>2019</year>). <article-title>Assessment of the generalization of learned image reconstruction and the potential for transfer learning</article-title>. <source>Magn. Reson. Med.</source> <volume>81</volume>, <fpage>116</fpage>&#x02013;<lpage>128</lpage>. <pub-id pub-id-type="doi">10.1002/mrm.27355</pub-id><pub-id pub-id-type="pmid">29774597</pub-id></citation></ref>
<ref id="B22">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lancaster</surname> <given-names>J. L.</given-names></name> <name><surname>Ghiatas</surname> <given-names>A. A.</given-names></name> <name><surname>Alyassin</surname> <given-names>A.</given-names></name> <name><surname>Kilcoyne</surname> <given-names>R. F.</given-names></name> <name><surname>Bonora</surname> <given-names>E.</given-names></name> <name><surname>Defronzo</surname> <given-names>R. A.</given-names></name></person-group> (<year>1991</year>). <article-title>Measurement of abdominal fat with t1-weighted MR images</article-title>. <source>J. Magn. Reson. Imaging</source> <volume>1</volume>, <fpage>363</fpage>&#x02013;<lpage>369</lpage>. <pub-id pub-id-type="doi">10.1002/jmri.1880010315</pub-id><pub-id pub-id-type="pmid">1802150</pub-id></citation></ref>
<ref id="B23">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lee</surname> <given-names>D.</given-names></name> <name><surname>Yoo</surname> <given-names>J.</given-names></name> <name><surname>Tak</surname> <given-names>S.</given-names></name> <name><surname>Ye</surname> <given-names>J. C.</given-names></name></person-group> (<year>2018</year>). <article-title>Deep residual learning for accelerated MRI using magnitude and phase networks</article-title>. <source>IEEE Trans. Biomed. Eng.</source> <volume>65</volume>, <fpage>1985</fpage>&#x02013;<lpage>1995</lpage>. <pub-id pub-id-type="doi">10.1109/TBME.2018.2821699</pub-id><pub-id pub-id-type="pmid">29993390</pub-id></citation></ref>
<ref id="B24">
<citation citation-type="web"><person-group person-group-type="author"><name><surname>Lopez-Paz</surname> <given-names>D.</given-names></name> <name><surname>Ranzato</surname> <given-names>M.</given-names></name></person-group> (<year>2017</year>). <article-title>&#x0201C;Gradient episodic memory for continual learning,&#x0201D;</article-title> in <source>gradient episodic memory for continual learning, Vol. 30</source>, eds I. Guyon, U. Von Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett (Curran Associates). Available online at: <ext-link ext-link-type="uri" xlink:href="https://proceedings.neurips.cc/paper_files/paper/2017/file/f87522788a2be2d171666752f97ddebb-Paper.pdf">https://proceedings.neurips.cc/paper_files/paper/2017/file/f87522788a2be2d171666752f97ddebb-Paper.pdf</ext-link></citation>
</ref>
<ref id="B25">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lv</surname> <given-names>J.</given-names></name> <name><surname>Li</surname> <given-names>G.</given-names></name> <name><surname>Tong</surname> <given-names>X.</given-names></name> <name><surname>Chen</surname> <given-names>W.</given-names></name> <name><surname>Huang</surname> <given-names>J.</given-names></name> <name><surname>Wang</surname> <given-names>C.</given-names></name> <etal/></person-group>. (<year>2021a</year>). <article-title>Transfer learning enhanced generative adversarial networks for multi-channel MRI reconstruction</article-title>. <source>Comput. Biol. Med.</source> <volume>134</volume>, <fpage>104504</fpage>. <pub-id pub-id-type="doi">10.1016/j.compbiomed.2021.104504</pub-id><pub-id pub-id-type="pmid">34062366</pub-id></citation></ref>
<ref id="B26">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lv</surname> <given-names>J.</given-names></name> <name><surname>Zhu</surname> <given-names>J.</given-names></name> <name><surname>Yang</surname> <given-names>G.</given-names></name></person-group> (<year>2021b</year>). <article-title>Which GAN? A comparative study of generative adversarial network-based fast MRI reconstruction</article-title>. <source>Philos. Trans. R. Soc. A</source> <volume>379</volume>, <fpage>20200203</fpage>. <pub-id pub-id-type="doi">10.1098/rsta.2020.0203</pub-id><pub-id pub-id-type="pmid">33966462</pub-id></citation></ref>
<ref id="B27">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Mardani</surname> <given-names>M.</given-names></name> <name><surname>Gong</surname> <given-names>E.</given-names></name> <name><surname>Cheng</surname> <given-names>J. Y.</given-names></name> <name><surname>Vasanawala</surname> <given-names>S. S.</given-names></name> <name><surname>Zaharchuk</surname> <given-names>G.</given-names></name> <name><surname>Xing</surname> <given-names>L.</given-names></name> <etal/></person-group>. (<year>2018</year>). <article-title>Deep generative adversarial neural networks for compressive sensing MRI</article-title>. <source>IEEE Trans. Med. Imaging</source> <volume>38</volume>, <fpage>167</fpage>&#x02013;<lpage>179</lpage>. <pub-id pub-id-type="doi">10.1109/TMI.2018.2858752</pub-id><pub-id pub-id-type="pmid">30040634</pub-id></citation></ref>
<ref id="B28">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Olah</surname> <given-names>C.</given-names></name> <name><surname>Cammarata</surname> <given-names>N.</given-names></name> <name><surname>Schubert</surname> <given-names>L.</given-names></name> <name><surname>Goh</surname> <given-names>G.</given-names></name> <name><surname>Petrov</surname> <given-names>M.</given-names></name> <name><surname>Carter</surname> <given-names>S.</given-names></name></person-group> (<year>2020</year>). <article-title>An overview of early vision in inceptionv1</article-title>. <source>Distill</source> <volume>5</volume>, <fpage>e00024</fpage>-002. <pub-id pub-id-type="doi">10.23915/distill.00024.003</pub-id></citation>
</ref>
<ref id="B29">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Pan</surname> <given-names>S. J.</given-names></name> <name><surname>Yang</surname> <given-names>Q.</given-names></name></person-group> (<year>2010</year>). <article-title>A survey on transfer learning</article-title>. <source>IEEE Trans. Knowledge Data Eng.</source> <volume>22</volume>, <fpage>1345</fpage>&#x02013;<lpage>1359</lpage>. <pub-id pub-id-type="doi">10.1109/TKDE.2009.191</pub-id></citation>
</ref>
<ref id="B30">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Pruessmann</surname> <given-names>K. P.</given-names></name> <name><surname>Weiger</surname> <given-names>M.</given-names></name> <name><surname>Scheidegger</surname> <given-names>M. B.</given-names></name> <name><surname>Boesiger</surname> <given-names>P.</given-names></name></person-group> (<year>1999</year>). <article-title>Sense: sensitivity encoding for fast MRI</article-title>. <source>Magn. Reson. Med.</source> <volume>42</volume>, <fpage>952</fpage>&#x02013;<lpage>962</lpage>.<pub-id pub-id-type="pmid">10542355</pub-id></citation></ref>
<ref id="B31">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Quan</surname> <given-names>T. M.</given-names></name> <name><surname>Nguyen-Duc</surname> <given-names>T.</given-names></name> <name><surname>Jeong</surname> <given-names>W.-K.</given-names></name></person-group> (<year>2018</year>). <article-title>Compressed sensing MRI reconstruction using a generative adversarial network with a cyclic loss</article-title>. <source>IEEE Trans. Med. Imaging</source> <volume>37</volume>, <fpage>1488</fpage>&#x02013;<lpage>1497</lpage>. <pub-id pub-id-type="doi">10.1109/TMI.2018.2820120</pub-id><pub-id pub-id-type="pmid">29870376</pub-id></citation></ref>
<ref id="B32">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Romero</surname> <given-names>M.</given-names></name> <name><surname>Interian</surname> <given-names>Y.</given-names></name> <name><surname>Solberg</surname> <given-names>T.</given-names></name> <name><surname>Valdes</surname> <given-names>G.</given-names></name></person-group> (<year>2020</year>). <article-title>Targeted transfer learning to improve performance in small medical physics datasets</article-title>. <source>Med. Phys.</source> <volume>47</volume>, <fpage>6246</fpage>&#x02013;<lpage>6256</lpage>. <pub-id pub-id-type="doi">10.1002/mp.14507</pub-id><pub-id pub-id-type="pmid">33007112</pub-id></citation></ref>
<ref id="B33">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Roy</surname> <given-names>R.</given-names></name> <name><surname>Kailath</surname> <given-names>T.</given-names></name></person-group> (<year>1989</year>). <article-title>Esprit-estimation of signal parameters via rotational invariance techniques</article-title>. <source>IEEE Trans. Acoust. Speech Signal Process.</source> <volume>37</volume>, <fpage>984</fpage>&#x02013;<lpage>995</lpage>.<pub-id pub-id-type="pmid">26571506</pub-id></citation></ref>
<ref id="B34">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Shaul</surname> <given-names>R.</given-names></name> <name><surname>David</surname> <given-names>I.</given-names></name> <name><surname>Shitrit</surname> <given-names>O.</given-names></name> <name><surname>Raviv</surname> <given-names>T. R.</given-names></name></person-group> (<year>2020</year>). <article-title>Subsampled brain MRI reconstruction by generative adversarial neural networks</article-title>. <source>Med. Image Anal.</source> <volume>65</volume>, <fpage>101747</fpage>. <pub-id pub-id-type="doi">10.1016/j.media.2020.101747</pub-id><pub-id pub-id-type="pmid">32593933</pub-id></citation></ref>
<ref id="B35">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Shitrit</surname> <given-names>O.</given-names></name> <name><surname>Riklin Raviv</surname> <given-names>T.</given-names></name></person-group> (<year>2017</year>). <article-title>&#x0201C;Accelerated magnetic resonance imaging by adversarial neural network,&#x0201D;</article-title> in <source>Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support: Third International Workshop, DLMIA 2017, and 7th International Workshop, ML-CDS 2017</source> (<publisher-loc>Qu&#x000E9;bec City, QC</publisher-loc>: <publisher-name>Springer</publisher-name>), <fpage>30</fpage>&#x02013;<lpage>38</lpage>.</citation>
</ref>
<ref id="B36">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Sun</surname> <given-names>Q.</given-names></name> <name><surname>Liu</surname> <given-names>Y.</given-names></name> <name><surname>Chua</surname> <given-names>T.-S.</given-names></name> <name><surname>Schiele</surname> <given-names>B.</given-names></name></person-group> (<year>2019</year>). <article-title>&#x0201C;Meta-transfer learning for few-shot learning,&#x0201D;</article-title> in <source>Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition</source> (<publisher-loc>Long Beach, CA</publisher-loc>), <fpage>403</fpage>&#x02013;<lpage>412</lpage>.</citation>
</ref>
<ref id="B37">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Tajbakhsh</surname> <given-names>N.</given-names></name> <name><surname>Shin</surname> <given-names>J. Y.</given-names></name> <name><surname>Gurudu</surname> <given-names>S. R.</given-names></name> <name><surname>Hurst</surname> <given-names>R. T.</given-names></name> <name><surname>Kendall</surname> <given-names>C. B.</given-names></name> <name><surname>Gotway</surname> <given-names>M. B.</given-names></name> <etal/></person-group>. (<year>2016</year>). <article-title>Convolutional neural networks for medical image analysis: full training or fine tuning?</article-title> <source>IEEE Trans. Med. Imaging</source> <volume>35</volume>, <fpage>1299</fpage>&#x02013;<lpage>1312</lpage>. <pub-id pub-id-type="doi">10.1109/TMI.2016.2535302</pub-id><pub-id pub-id-type="pmid">26978662</pub-id></citation></ref>
<ref id="B38">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>S.</given-names></name> <name><surname>Su</surname> <given-names>Z.</given-names></name> <name><surname>Ying</surname> <given-names>L.</given-names></name> <name><surname>Peng</surname> <given-names>X.</given-names></name> <name><surname>Zhu</surname> <given-names>S.</given-names></name> <name><surname>Liang</surname> <given-names>F.</given-names></name> <etal/></person-group>. (<year>2016</year>). <article-title>&#x0201C;Accelerating magnetic resonance imaging via deep learning,&#x0201D;</article-title> in <source>2016 IEEE 13th International Symposium on Biomedical Imaging (ISBI)</source> (<publisher-loc>Prague</publisher-loc>: <publisher-name>IEEE</publisher-name>), <fpage>514</fpage>&#x02013;<lpage>517</lpage>.<pub-id pub-id-type="pmid">31709031</pub-id></citation></ref>
<ref id="B39">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>Z.</given-names></name> <name><surname>Bovik</surname> <given-names>A. C.</given-names></name> <name><surname>Sheikh</surname> <given-names>H. R.</given-names></name> <name><surname>Simoncelli</surname> <given-names>E. P.</given-names></name></person-group> (<year>2004</year>). <article-title>Image quality assessment: from error visibility to structural similarity</article-title>. <source>IEEE Trans. Image Process.</source> <volume>13</volume>, <fpage>600</fpage>&#x02013;<lpage>612</lpage>. <pub-id pub-id-type="doi">10.1109/TIP.2003.819861</pub-id><pub-id pub-id-type="pmid">15376593</pub-id></citation></ref>
<ref id="B40">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Weiss</surname> <given-names>K.</given-names></name> <name><surname>Khoshgoftaar</surname> <given-names>T. M.</given-names></name> <name><surname>Wang</surname> <given-names>D.</given-names></name></person-group> (<year>2016</year>). <article-title>A survey of transfer learning</article-title>. <source>J. Big Data</source> <volume>3</volume>, <fpage>1</fpage>&#x02013;<lpage>40</lpage>. <pub-id pub-id-type="doi">10.1186/s40537-016-0043-6</pub-id></citation>
</ref>
<ref id="B41">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Yang</surname> <given-names>G.</given-names></name> <name><surname>Yu</surname> <given-names>S.</given-names></name> <name><surname>Dong</surname> <given-names>H.</given-names></name> <name><surname>Slabaugh</surname> <given-names>G.</given-names></name> <name><surname>Dragotti</surname> <given-names>P. L.</given-names></name> <name><surname>Ye</surname> <given-names>X.</given-names></name> <etal/></person-group>. (<year>2017</year>). <article-title>Dagan: deep de-aliasing generative adversarial networks for fast compressed sensing MRI reconstruction</article-title>. <source>IEEE Trans. Med. Imaging</source> <volume>37</volume>, <fpage>1310</fpage>&#x02013;<lpage>1321</lpage>. <pub-id pub-id-type="doi">10.1109/TMI.2017.2785879</pub-id><pub-id pub-id-type="pmid">29870361</pub-id></citation></ref>
<ref id="B42">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Yousaf</surname> <given-names>T.</given-names></name> <name><surname>Dervenoulas</surname> <given-names>G.</given-names></name> <name><surname>Politis</surname> <given-names>M.</given-names></name></person-group> (<year>2018</year>). <article-title>Advances in MRI methodology</article-title>. <source>Int. Rev. Neurobiol.</source> <volume>141</volume>, <fpage>31</fpage>&#x02013;<lpage>76</lpage>. <pub-id pub-id-type="doi">10.1016/bs.irn.2018.08.008</pub-id><pub-id pub-id-type="pmid">30314602</pub-id></citation></ref>
<ref id="B43">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>Z.</given-names></name> <name><surname>Li</surname> <given-names>Y.</given-names></name> <name><surname>Shin</surname> <given-names>B.-S.</given-names></name></person-group> (<year>2022</year>). <article-title>C2-GAN: content-consistent generative adversarial networks for unsupervised domain adaptation in medical image segmentation</article-title>. <source>Med. Phys.</source> <volume>49</volume>, <fpage>6491</fpage>&#x02013;<lpage>6504</lpage>. <pub-id pub-id-type="doi">10.1002/mp.15944</pub-id><pub-id pub-id-type="pmid">35981348</pub-id></citation></ref>
</ref-list>


</back>
</article> 