<?xml version="1.0" encoding="utf-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Archiving and Interchange DTD v2.3 20070202//EN" "archivearticle.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" article-type="methods-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Med.</journal-id>
<journal-title>Frontiers in Medicine</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Med.</abbrev-journal-title>
<issn pub-type="epub">2296-858X</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fmed.2023.1188207</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Medicine</subject>
<subj-group>
<subject>Methods</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Pathologic complete response prediction in breast cancer lesion segmentation and neoadjuvant therapy</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" equal-contrib="yes">
<name>
<surname>Liu</surname>
<given-names>Yue</given-names>
</name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<xref ref-type="author-notes" rid="fn0001"><sup>&#x2020;</sup></xref>
</contrib>
<contrib contrib-type="author" equal-contrib="yes">
<name>
<surname>Chen</surname>
<given-names>Zhihong</given-names>
</name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
<xref ref-type="author-notes" rid="fn0001"><sup>&#x2020;</sup></xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Chen</surname>
<given-names>Junhao</given-names>
</name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Shi</surname>
<given-names>Zhenwei</given-names>
</name>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
<xref ref-type="aff" rid="aff4"><sup>4</sup></xref>
<xref ref-type="aff" rid="aff5"><sup>5</sup></xref>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Fang</surname>
<given-names>Gang</given-names>
</name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>&#x002A;</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/1582373/overview"/>
</contrib>
</contrib-group>
<aff id="aff1"><sup>1</sup><institution>Institute of Computing Science and Technology, Guangzhou University</institution>, <addr-line>Guangzhou</addr-line>, <country>China</country></aff>
<aff id="aff2"><sup>2</sup><institution>School of Information Engineering, Jiangxi College of Applied Technology</institution>, <addr-line>Ganzhou</addr-line>, <country>China</country></aff>
<aff id="aff3"><sup>3</sup><institution>Guangdong Provincial Key Laboratory of Artificial Intelligence in Medical Image Analysis and Application, Guangdong Provincial People's Hospital, Guangdong Academy of Medical Sciences</institution>, <addr-line>Guangzhou</addr-line>, <country>China</country></aff>
<aff id="aff4"><sup>4</sup><institution>Department of Radiology, Guangdong Provincial People&#x2019;s Hospital, Guangdong Academy of Medical Sciences</institution>, <addr-line>Guangzhou</addr-line>, <country>China</country></aff>
<aff id="aff5"><sup>5</sup><institution>Guangdong Cardiovascular Institute, Guangdong Provincial People's Hospital, Guangdong Academy of Medical Sciences</institution>, <addr-line>Guangzhou</addr-line>, <country>China</country></aff>
<author-notes>
<fn fn-type="edited-by" id="fn0002"><p>Edited by: Udhaya Kumar, Baylor College of Medicine, United States</p></fn>
<fn fn-type="edited-by" id="fn0003"><p>Reviewed by: Sriroop Reddy Ramireddy, The Apollo University, India; Alina Tudorica, Oregon Health and Science University, United States</p></fn>
<corresp id="c001">&#x002A;Correspondence: Gang Fang, <email>gangf@gzhu.edu.cn</email></corresp>
<fn fn-type="equal" id="fn0001"><p><sup>&#x2020;</sup>These authors have contributed equally to this work</p></fn>
</author-notes>
<pub-date pub-type="epub">
<day>04</day>
<month>12</month>
<year>2023</year>
</pub-date>
<pub-date pub-type="collection">
<year>2023</year>
</pub-date>
<volume>10</volume>
<elocation-id>1188207</elocation-id>
<history>
<date date-type="received">
<day>17</day>
<month>03</month>
<year>2023</year>
</date>
<date date-type="accepted">
<day>08</day>
<month>11</month>
<year>2023</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x00A9; 2023 Liu, Chen, Chen, Shi and Fang.</copyright-statement>
<copyright-year>2023</copyright-year>
<copyright-holder>Liu, Chen, Chen, Shi and Fang</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/">
<p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p>
</license>
</permissions>
<abstract>
<sec id="sec1">
<title>Objectives</title>
<p>Predicting whether axillary lymph nodes could achieve pathologic Complete Response (pCR) after breast cancer patients receive neoadjuvant chemotherapy helps make a quick follow-up treatment plan. This paper presents a novel method to achieve this prediction with the most effective medical imaging method, Dynamic Contrast-enhanced Magnetic Resonance Imaging (DCE-MRI).</p>
</sec>
<sec id="sec2">
<title>Methods</title>
<p>In order to get an accurate prediction, we first proposed a two-step lesion segmentation method to extract the breast cancer lesion region from DCE-MRI images. With the segmented breast cancer lesion region, we then used a multi-modal fusion model to predict the probability of axillary lymph nodes achieving pCR.</p>
</sec>
<sec id="sec3">
<title>Results</title>
<p>We collected 361 breast cancer samples from two hospitals to train and test the proposed segmentation model and the multi-modal fusion model. Both segmentation and prediction models obtained high accuracy.</p>
</sec>
<sec id="sec4">
<title>Conclusion</title>
<p>The results show that our method is effective in both the segmentation task and the pCR prediction task. It suggests that the presented methods, especially the multi-modal fusion model, can be used for the prediction of treatment response in breast cancer, given data from noninvasive methods only.</p>
</sec>
</abstract>
<kwd-group>
<kwd>PCR</kwd>
<kwd>DCE-MRI</kwd>
<kwd>neoadjuvant chemotherapy</kwd>
<kwd>multi-modal fusion model</kwd>
<kwd>breast cancer</kwd>
<kwd>lesion segmentation</kwd>
</kwd-group>
<counts>
<fig-count count="9"/>
<table-count count="7"/>
<equation-count count="4"/>
<ref-count count="40"/>
<page-count count="11"/>
<word-count count="6822"/>
</counts>
<custom-meta-wrap>
<custom-meta>
<meta-name>section-at-acceptance</meta-name>
<meta-value>Precision Medicine</meta-value>
</custom-meta>
</custom-meta-wrap>
</article-meta>
</front>
<body>
<sec sec-type="intro" id="sec5">
<label>1</label>
<title>Introduction</title>
<p>The incidence of breast cancer has been increasing in recent years, and breast cancer is one of the most common malignant cancers in women. In 2023, it is estimated that there will be 297,790 new cases of invasive breast cancer diagnosed and 43,170 women will die from breast cancer in the U.S. (<xref ref-type="bibr" rid="ref1">1</xref>). At present, neoadjuvant chemotherapy (NAC) plays an important role in breast cancer treatment, and research (<xref ref-type="bibr" rid="ref2 ref3 ref4 ref5">2&#x2013;5</xref>) shows that whether axillary lymph nodes achieve pathologic Complete Response (pCR) is an important prognostic predictor for breast cancer patients who receive NAC, and that pCR indicates a lower risk of local recurrence and a better long-term prognosis for patients. Therefore, it is of great importance if we can predict whether axillary lymph nodes will achieve pCR after patients receive NAC; this helps make a follow-up treatment plan and improve patients&#x2019; prognosis.</p>
<p>As a non-invasive method, imaging examination plays an important role in the clinical diagnosis and treatment of cancer. Specifically, in the evaluation of treatment response to NAC in breast cancer, Magnetic Resonance Imaging (MRI) is the most commonly used imaging evaluation method in clinical practice (<xref ref-type="bibr" rid="ref6">6</xref>). However, based on radiologists&#x2019; subjective evaluation of imaging features, MRI shows high sensitivities (83&#x2013;92%) and intermediate specificities (47&#x2013;63%) in preoperative diagnosis of axillary lymph nodes achieving pCR after NAC (<xref ref-type="bibr" rid="ref7">7</xref>). Nevertheless, recently, artificial intelligence has shown great promise in analyzing medical images, helping to identify image information beyond the ability of the naked eye, and providing objective quantitative evaluation to support clinical decision-making (<xref ref-type="bibr" rid="ref6">6</xref>). Specifically, deep neural networks, especially convolutional networks, attract great attention in the field of medical imaging analysis (<xref ref-type="bibr" rid="ref8">8</xref>). Thus, the objective of this paper is to utilize deep neural networks to process MRI images of breast cancer in order to predict whether axillary lymph node metastasis in breast cancer could achieve pCR after patients receive NAC.</p>
<p>Compared with other types of medical imaging technologies, such as Computed Tomography (CT) and Positron Emission Tomography (PET), MRI provides better imaging capability for soft tissues and is widely adopted in breast cancer diagnosis and treatment. In our work, we use Dynamic Contrast-enhanced Magnetic Resonance Imaging (DCE-MRI), which provides a high-quality image for soft tissues with better quality of blood flow around the lesion, which facilitates higher accuracy and earlier detection in breast cancer diagnosis. Despite the above, due to the nature of medical imaging technology, a common DCE-MRI image for breast cancer diagnosis (<xref ref-type="fig" rid="fig1">Figure 1A</xref>), contains a large amount of redundant information, so we need to extract only the lesion region of interest for further processing in order to achieve better performance. Image segmentation is widely used in medical imaging analysis, where pixels from specific regions are segmented from the background. With the prevalence of deep learning, models such as the Fully Convolutional Network (FCN) (<xref ref-type="bibr" rid="ref9">9</xref>) and UNet (<xref ref-type="bibr" rid="ref10">10</xref>) are applied in medical image segmentation and achieve great performance. It is also proven that neural networks are effective and efficient in breast tumor segmentation tasks (<xref ref-type="bibr" rid="ref11">11</xref>). In our work, considering the fact that breast cancer lesions are close to the chest wall and vary in size and distribution, we propose a two-step lesion segmentation method using nnUNet (<xref ref-type="bibr" rid="ref12">12</xref>) as shown in <xref ref-type="fig" rid="fig1">Figure 1</xref>: (1) segmentation of the mammary gland region; (2) segmentation of the breast cancer lesion region within the mammary gland region. When training medical image segmentation models, transferability should be taken into consideration because DCE-MRI images collected from different centers may vary in resolution, scanner used, protocol, and image quality. Hence, we apply a histogram matching method (<xref ref-type="bibr" rid="ref13">13</xref>) to augment the training samples in order to improve the model transferability.</p>
<fig position="float" id="fig1">
<label>Figure 1</label>
<caption>
<p>The proposed two-step breast cancer lesion segmentation method. <bold>(A)</bold> is a DCE-MRI image sample for breast cancer diagnosis, which contains irrelevant regions, such as the heart; <bold>(B)</bold> shows the segmented mammary gland region in red; <bold>(C)</bold> shows the breast cancer lesion region in red.</p>
</caption>
<graphic xlink:href="fmed-10-1188207-g001.tif"/>
</fig>
<p>Usually, after acquiring the model representation of breast cancer lesions, we can directly train a neural network to predict the probability of pCR after NAC. However, as pointed out by Ramos-Vara (<xref ref-type="bibr" rid="ref14">14</xref>), immunohistochemical detection can greatly help in the diagnosis of breast cancer, invasion and metastasis of tumors, and prognosis of breast cancer, so together with MRI data, we coordinate four common types of molecular typing data in breast cancer treatment to construct a multi-modal fusion model to predict whether axillary lymph nodes could achieve pCR. Our experiments show that the proposed multi-modal fusion model outperforms the predictive model with only MRI data.</p>
<p>In order to train and evaluate the proposed two-step lesion segmentation method and the multi-model fusion model, we collected 361 breast cancer samples from two hospitals: 246 samples from Guangdong Provincial People&#x2019;s Hospital using the Philips Achieva 1.5 T MRI system, and 115 samples from Henan Renmin Hospital using the Discovery MR750 3.0 T MRI system. Each sample comes with DCE-MRI imaging and molecular typing data, and each DCE-MRI image is labeled and verified by radiologists with more than 5&#x2009;years of breast cancer experience.</p>
<p>In this paper, we make the following three contributions: First, we propose a two-step lesion segmentation method to extract breast cancer lesion regions from DCE-MRI images. In the model training, considering the different sources of DCE-MRI images, we apply a simple histogram matching method to improve the model transferability. Second, we propose a multi-modal (i.e., segmented DCE-MRI image and molecular typing data) fusion model to predict the probability of axillary lymph nodes achieving pCR after patients receive neoadjuvant therapy. Finally, we evaluate our model through extensive ablation studies and experiments on a collected breast cancer dataset, and we show the promising performance of the proposed method.</p>
<p>The studies involving human participants were reviewed and approved by the Ethics Board of Guangdong Provincial People&#x2019;s Hospital and the Ethics Board of Henan Renmin Hospital. Written informed consent to participate in this study was provided by the participants.</p>
</sec>
<sec id="sec6">
<label>2</label>
<title>Related works</title>
<sec id="sec7">
<label>2.1</label>
<title>Convolutional neural network</title>
<p>Convolutional neural networks (CNN) have achieved great success in medical imaging analysis. CNN was first introduced to process medical images by Lo et al. (<xref ref-type="bibr" rid="ref15">15</xref>), and with the rapid development of CNN (<xref ref-type="bibr" rid="ref16">16</xref>, <xref ref-type="bibr" rid="ref17">17</xref>), it has been considered one of the most effective methods to process medical images. ResNet (<xref ref-type="bibr" rid="ref18">18</xref>), as one of the most classic CNNs, is widely adopted in all kinds of neural networks; with 152 layers of networks, it outperformed other models in the ImageNet Large Scale Visual Recognition Challenge (ILSVRC) 2015 by a large margin. Compared with AlexNet (<xref ref-type="bibr" rid="ref16">16</xref>) and VGGNet (<xref ref-type="bibr" rid="ref19">19</xref>), ResNet achieves smaller training errors and better testing accuracy, so in our proposed models, we chose ResNet as the backbone network.</p>
</sec>
<sec id="sec8">
<label>2.2</label>
<title>Medical image segmentation</title>
<p>With the rapid development and popularization of medical imaging devices and technologies, including MRI, CT, PET, etc., the amount of medical images produced by these devices is increasing; it is reported that medical images account for one-fifth of all images generated worldwide. Thus, it is urgent to process medical images effectively, and medical image segmentation is the first important step in image analysis. Among all medical imaging technologies, MRI is the most widely adopted one. With MRI, professionals can vary the image contrast to show different image intensities to reflect the difference between soft tissue, parenchyma, and fluid (<xref ref-type="bibr" rid="ref20">20</xref>, <xref ref-type="bibr" rid="ref21">21</xref>).</p>
<p>With the assistance of CNN, various medical image segmentation methods have been developed. In 2015, FCN (<xref ref-type="bibr" rid="ref9">9</xref>) was proposed to implement pixel-level classification to solve the semantic segmentation problem; it accepts images of arbitrary sizes. FCN was one of the first deep learning techniques that were applied to medical image processing, but the segmentation performance is not satisfactory. Based on FCN, Olaf et al. introduced U-Net (<xref ref-type="bibr" rid="ref10">10</xref>) for cell image segmentation; its surprising performance soon made it a standard backbone network. Later, 3D U-Net (<xref ref-type="bibr" rid="ref22">22</xref>), V-Net (<xref ref-type="bibr" rid="ref23">23</xref>), Res-UNet (<xref ref-type="bibr" rid="ref24">24</xref>), and other variants of U-Net (<xref ref-type="bibr" rid="ref25 ref26 ref27 ref28 ref29 ref30">25&#x2013;30</xref>) were proposed. Aside from FCN and U-Net, Recurrent Neural Networks (RNNs) are also utilized for medical image segmentation (<xref ref-type="bibr" rid="ref31">31</xref>, <xref ref-type="bibr" rid="ref32">32</xref>).</p>
</sec>
<sec id="sec9">
<label>2.3</label>
<title>Pathologic complete response prediction</title>
<p>Over the last decade, many methods have been developed in academia to predict pCR, including radiomics, machine learning, and deep learning. In radiomics, pre-designed features are extracted to build a predictive model, but these pre-designed features are complex (<xref ref-type="bibr" rid="ref33">33</xref>). Traditional machine learning methods such as SVM and AdaBoost also need well-designed features for prediction. In (<xref ref-type="bibr" rid="ref34">34</xref>), 13,950 imaging features are extracted from CT and MRI data for machine learning. Compared with traditional radiomics and machine learning methods, the predictive model based on deep learning provides an end-to-end training and inferring method that can be directly applied to medical images (<xref ref-type="bibr" rid="ref35">35</xref>). In our work, we use MRI images of breast cancer lesion regions, along with four types of molecular typing data commonly used in breast cancer treatment, to construct a multi-modal fusion model to predict whether or not the patient can achieve pCR.</p>
</sec>
</sec>
<sec sec-type="methods" id="sec10">
<label>3</label>
<title>Methods</title>
<p>In this section, we introduce the proposed method of processing MRI images of breast cancer with neural networks in order to predict whether axillary lymph node metastasis in breast cancer could achieve pCR after patients receive neoadjuvant therapy. We have divided this section into two parts: the first part gives details on how we extract the breast cancer lesion region from a common DCE-MRI breast cancer image; the second part introduces the multi-modal fusion model for pCR prediction.</p>
<sec id="sec11">
<label>3.1</label>
<title>Breast cancer lesion segmentation</title>
<p>As introduced in Section 1, and referring to <xref ref-type="fig" rid="fig1">Figure 1A</xref>, a common DCE-MRI image for breast cancer diagnosis contains a large amount of information that is irrelevant, so we needed to extract only the concerned lesion region for later processing in order to achieve better performance. Another notable reason for breast cancer lesion segmentation is that there is similar imaging intensity in the heart region in DCE-MRI images, as shown in <xref ref-type="fig" rid="fig2">Figure 2</xref>, so it is preferable to remove the heart region in order to reduce the probability of false positives. In the following, we elaborated on the proposed two-step lesion segmentation method to extract breast cancer lesion regions based on nnUNet and introduced a simple histogram matching method to augment the training samples in order to improve the model transferability between different centers.</p>
<fig position="float" id="fig2">
<label>Figure 2</label>
<caption>
<p>Two red rectangles show similar imaging intensity in the breast and heart regions.</p>
</caption>
<graphic xlink:href="fmed-10-1188207-g002.tif"/>
</fig>
<sec id="sec12">
<label>3.1.1</label>
<title>Backbone network for image segmentation</title>
<p>As the DCE-MRI data collected for training is three-dimensional, we chose nnUNet (<xref ref-type="bibr" rid="ref12">12</xref>) as the backbone network for the breast cancer lesion segmentation task. We specified the nnUNet as follows:</p>
<sec id="sec13">
<label>3.1.1.1</label>
<title>Pre-processing</title>
<p>Re-sampling and normalization were implemented at this stage. As the spatial resolution of each MRI image varied, which means one pixel of the image may represent a different size of physical space, the MRI image needs to be re-sampled according to the median of the spatial resolution of all data. Z-score normalization was done independently for each patient&#x2019;s imaging data.</p>
<p>Data augmentation was also implemented. Augmentation techniques include random rotation, random scaling, random elastic transformation, gamma correction, and inversion.</p>
</sec>
<sec id="sec14">
<label>3.1.1.2</label>
<title>Loss function</title>
<p>During training, we utilized the Cross-entropy loss and the Dice loss as follows:</p>
<disp-formula id="EQ1">
<label>(1)</label>
<mml:math id="M1">
<mml:msub>
<mml:mi>L</mml:mi>
<mml:mi mathvariant="italic">total</mml:mi>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mi>L</mml:mi>
<mml:mi mathvariant="italic">dice</mml:mi>
</mml:msub>
<mml:mo>+</mml:mo>
<mml:msub>
<mml:mi>L</mml:mi>
<mml:mrow>
<mml:mi>c</mml:mi>
<mml:mi>e</mml:mi>
</mml:mrow>
</mml:msub>
</mml:math>
</disp-formula>
<p>where the Cross-entropy loss <italic>L<sub>ce</sub></italic> is defined as follows:</p>
<disp-formula id="EQ2">
<label>(2)</label>
<mml:math id="M2">
<mml:msub>
<mml:mi>L</mml:mi>
<mml:mrow>
<mml:mi>c</mml:mi>
<mml:mi>e</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mo>&#x2212;</mml:mo>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mi>N</mml:mi>
</mml:mfrac>
<mml:munder>
<mml:mstyle displaystyle="true">
<mml:mo stretchy="true">&#x2211;</mml:mo>
</mml:mstyle>
<mml:mi>i</mml:mi>
</mml:munder>
<mml:munderover>
<mml:mstyle displaystyle="true">
<mml:mo stretchy="true">&#x2211;</mml:mo>
</mml:mstyle>
<mml:mi>k</mml:mi>
<mml:mi>K</mml:mi>
</mml:munderover>
<mml:msub>
<mml:mi>v</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>k</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>log</mml:mo>
<mml:mfenced open="(" close=")">
<mml:msub>
<mml:mi>u</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>k</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mfenced>
</mml:math>
</disp-formula>
<p>Dice loss was first introduced in (<xref ref-type="bibr" rid="ref23">23</xref>) to solve the imbalance between positive and negative samples. Dice loss is different from Cross-entropy loss: it helps minimize segmentation error and obtain more stable segmentation performance (<xref ref-type="bibr" rid="ref36">36</xref>). The Dice loss equation is as follows:</p>
<disp-formula id="EQ3">
<label>(3)</label>
<mml:math id="M3">
<mml:msub>
<mml:mi>L</mml:mi>
<mml:mi mathvariant="italic">dice</mml:mi>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mo>&#x2212;</mml:mo>
<mml:mfrac>
<mml:mn>2</mml:mn>
<mml:mfenced open="|" close="|">
<mml:mi>K</mml:mi>
</mml:mfenced>
</mml:mfrac>
<mml:munder>
<mml:mstyle displaystyle="true">
<mml:mo stretchy="true">&#x2211;</mml:mo>
</mml:mstyle>
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mi>K</mml:mi>
</mml:mrow>
</mml:munder>
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mstyle displaystyle="true">
<mml:mo stretchy="true">&#x2211;</mml:mo>
</mml:mstyle>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mi mathvariant="normal">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msubsup>
<mml:mi>u</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>k</mml:mi>
</mml:msubsup>
<mml:msubsup>
<mml:mi>v</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>k</mml:mi>
</mml:msubsup>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mstyle displaystyle="true">
<mml:mo stretchy="true">&#x2211;</mml:mo>
</mml:mstyle>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mi>I</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mspace width="0.25em"/>
<mml:msubsup>
<mml:mi>u</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>k</mml:mi>
</mml:msubsup>
<mml:mo>+</mml:mo>
<mml:msub>
<mml:mstyle displaystyle="true">
<mml:mo stretchy="true">&#x2211;</mml:mo>
</mml:mstyle>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mi mathvariant="normal">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msubsup>
<mml:mi>v</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>k</mml:mi>
</mml:msubsup>
</mml:mrow>
</mml:mfrac>
</mml:math>
</disp-formula>
<p>In <xref ref-type="disp-formula" rid="EQ2">Eqs. (2)</xref> and <xref ref-type="disp-formula" rid="EQ3">(3)</xref>, <italic>u</italic> is the model&#x2019;s predictive probability, <italic>v</italic> is the ground truth one-hot code, <italic>K</italic> is the number of classes, and <italic>I</italic> is the representation vector of the image. Empirically, Cross-entropy loss makes the model focus on the global representation, i.e., each pixel of the image, while Dice loss pays more attention to the positive region, so <italic>L<sub>total</sub></italic> takes advantage of both global and local information.</p>
</sec>
<sec id="sec15">
<label>3.1.1.3</label>
<title>Inference</title>
<p>Image segmentation inference was then performed on patches of MRI images. An image was divided into patches with an overlap of <italic>size/2</italic> pixels, where <italic>size</italic> is the size of the stride. Due to the lack of neighbor information, the segmentation accuracy of the edges of each patch will be relatively lower, so when fusing the segmentation result for pixels along the edges, we decreased the weight of edge pixels while increasing the weight of pixels close to the center.</p>
</sec>
<sec id="sec16">
<label>3.1.1.4</label>
<title>Post-processing</title>
<p>After obtaining a segmentation result, we found the largest connected contour and, in the meantime, neglected other smaller ones. This post-processing step can effectively reduce the occurrence of false positives.</p>
</sec>
</sec>
<sec id="sec17">
<label>3.1.2</label>
<title>Two-step lesion segmentation</title>
<p>In order to reduce the probability of false positives, we utilized a two-step lesion segmentation method to extract the breast cancer lesion region. As shown in <xref ref-type="fig" rid="fig1">Figure 1</xref>, given a DCE-MRI sample for breast cancer diagnosis, we first segmented the mammary gland region, based on which we then segmented the breast cancer lesion region.</p>
<sec id="sec18">
<label>3.1.2.1</label>
<title>Mammary gland segmentation</title>
<p>The pre-processing described in Section 3.1.1 was applied to the original DCE-MRI samples, and the backbone network, i.e., nnUNet, was utilized to implement the first step &#x201C;mammary gland segmentation&#x201D; task.</p>
</sec>
<sec id="sec19">
<label>3.1.2.2</label>
<title>Breast cancer lesion segmentation</title>
<p>After getting the result from the first segmentation step, we continued to segment the breast cancer lesion region. The mammary gland region was pre-processed only by Z-score normalization and is fed to the second segmentation step. We used the same backbone network to implement the &#x201C;breast cancer lesion segmentation&#x201D; task.</p>
<p>The details of the training are explained in Section 4, and the performance of our proposed two-step lesion segmentation method is also shown in the following section.</p>
</sec>
</sec>
<sec id="sec20">
<label>3.1.3</label>
<title>Domain adaptation</title>
<p>As our dataset was collected from two different centers, there will inevitably be a model transfer issue when training on samples from one center and testing on another. This is a common issue in medical image analysis because different medical imaging devices with different imaging protocols, methods, and different operators produce MRI images that vary in resolution, quality, etc.; therefore, many methods have been proposed to mitigate this issue (<xref ref-type="bibr" rid="ref13">13</xref>, <xref ref-type="bibr" rid="ref37">37</xref>, <xref ref-type="bibr" rid="ref38">38</xref>). With respect to the specific differences between DCE-MRI samples, we designed a domain adaptation method, i.e., simple histogram matching (<xref ref-type="bibr" rid="ref13">13</xref>), to improve the transferability of the model. Another advantage of histogram matching is that it only requires the gray-level distribution of the DCE-MRI images; thus, it does not reveal any personal information about the patient.</p>
<sec id="sec21">
<label>3.1.3.1</label>
<title>Histogram matching</title>
<p>At this point, we applied a simple histogram matching method (<xref ref-type="bibr" rid="ref13">13</xref>) to augment training samples in order to improve transferability. More specifically, we introduced the gray-level distribution to augment training samples, each of which was augmented by matching the gray-level histogram computed with samples from other centers. The histogram matching is implemented as follows:</p>
<disp-formula id="EQ4">
<label>(4)</label>
<mml:math id="M4">
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mi>L</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>M</mml:mi>
<mml:mo>&#x2217;</mml:mo>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:munderover>
<mml:mstyle displaystyle="true">
<mml:mo stretchy="true">&#x2211;</mml:mo>
</mml:mstyle>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
<mml:mi>k</mml:mi>
</mml:munderover>
<mml:msub>
<mml:mi>n</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
<mml:mo>&#x2217;</mml:mo>
<mml:mi>k</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo>,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>,</mml:mo>
<mml:mi>L</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:math>
</disp-formula>
<p>where <italic>L</italic> is the maximum gray-level value of the target histogram, <italic>M</italic> and <italic>N</italic> are the width and height of the image, and <italic>n<sub>j</sub></italic> is the gray-level value of pixel <italic>j</italic>.</p>
<p>As for the segmentation task, we first implemented the mammary gland segmentation without histogram matching. We then computed a gray-level histogram for each sample in the test dataset (in our experiment, samples from Henan Renmin Hospital are used as the test dataset) and then applied histogram matching to each sample in the training dataset (in our experiment, samples from Guangdong Provincial People&#x2019;s Hospital are used as the training dataset) with a randomly selected gray-level histogram from the test dataset. After the training dataset was augmented, it was fed to nnUNet for breast cancer lesion segmentation training.</p>
</sec>
</sec>
</sec>
<sec id="sec22">
<label>3.2</label>
<title>Pathologic complete response prediction</title>
<p>Among all treatments for breast cancer, NAC is emerging as a new and effective method. As introduced in Section 1, utilizing imaging examination as a non-invasive method, together with four types of molecular typing data commonly used in breast cancer treatment, we proposed a multi-modal fusion model to predict whether axillary lymph nodes could achieve pCR after patients receive NAC.</p>
<sec id="sec23">
<label>3.2.1</label>
<title>Multi-modal fusion</title>
<p>Although one can use the MRI data of the breast cancer lesion to directly predict the probability of pCR after neoadjuvant therapy, it has been proven that immunohistochemical detection can also help in breast cancer prognosis (<xref ref-type="bibr" rid="ref14">14</xref>). Thus, we propose to utilize common types of molecular typing data extracted by immunohistochemical detection of breast cancer. More specifically, we chose the following four common types of molecular typing data in breast cancer treatment: Human Epidermal Growth Factor Receptor 2 (HER2), Estrogen Receptor (ER), Progesterone Receptor (PR), and Ki-67.</p>
<p>HER2 protein is negative in normal breast tissue, and the amplification of HER2 is highly related to the growth, proliferation, transfer, and invasion of tumor cells; thus, it can be treated as one of the prognostic indicators of clinical treatment monitoring. ER and PR are nuclear hormone receptors; the expression of ER/PR indicates that tumor cells retain the characteristics of hormone-dependent growth and is significant in the prognosis judgment of breast cancer. Ki-67 is a monoclonal antibody; high expression of Ki-67 indicates a poor prognosis.</p>
<p>In our work, we used the above four types of molecular typing data, together with a DCE-MRI image of breast cancer lesions, to train the multi-modal (i.e., text and image) fusion model to predict pCR.</p>
</sec>
<sec id="sec24">
<label>3.2.2</label>
<title>Network structure</title>
<p>We used conventional ResNet (<xref ref-type="bibr" rid="ref18">18</xref>) as the backbone network to construct the prediction model as a common practice; more specifically, ResNet34 was selected, and we justify this choice in Section 4. The network structure is shown in <xref ref-type="fig" rid="fig3">Figure 3</xref>:</p>
<fig position="float" id="fig3">
<label>Figure 3</label>
<caption>
<p>The network structure of our multi-model fusion model for pCR prediction.</p>
</caption>
<graphic xlink:href="fmed-10-1188207-g003.tif"/>
</fig>
<p>As shown in <xref ref-type="fig" rid="fig3">Figure 3</xref>, non-imaging features and MRI data were processed by two separate network branches. The molecular typing data was processed by five Fully Connected (FC) layers, while the MRI image was processed by five convolutional network layers. It should be noted that the output of each FC layer in the non-imaging clinical data branch is fused with the output of each CNN layer in the MRI data branch by multiplication. This structure balances the weight of non-imaging data and MRI data to compute the model representation and makes the model utilize both molecular typing data and a DCE-MRI image of a breast cancer lesion to predict pCR. The proposed network is trained by the conventional Cross-entropy loss function.</p>
</sec>
</sec>
</sec>
<sec id="sec25">
<label>4</label>
<title>Experiments and analysis</title>
<sec id="sec26">
<label>4.1</label>
<title>Experiment setting</title>
<sec id="sec27">
<label>4.1.1</label>
<title>Dataset</title>
<p>In our work, we used DCE-MRI data for breast cancer lesion segmentation and pCR prediction. DCE-MRI can provide a high-quality image for soft tissues with better quality of blood flow around the lesion region, which facilitates higher accuracy and earlier detection in breast cancer diagnosis. Therefore, DCE-MRI is the most widely adopted imaging method in breast cancer diagnosis and treatment.</p>
<p>In order to train and test the proposed method, we collected 361 breast cancer samples from two hospitals: 246 samples from Guangdong Provincial People&#x2019;s Hospital and 115 samples from Henan Renmin Hospital. Each DCE-MRI image was labeled and verified by professionals. A labeled DCE-MRI sample is shown in <xref ref-type="fig" rid="fig4">Figure 4</xref>. We also collected the four types of molecular typing data commonly used in breast cancer treatment: HER2, ER, PR, and Ki-67, for each of the 361 samples.</p>
<fig position="float" id="fig4">
<label>Figure 4</label>
<caption>
<p>An example of the 361 labeled DCE-MRI samples in our dataset.</p>
</caption>
<graphic xlink:href="fmed-10-1188207-g004.tif"/>
</fig>
</sec>
<sec id="sec28">
<label>4.1.2</label>
<title>Network training setup</title>
<p>For breast cancer lesion segmentation, the network was trained by Adam optimizer (<xref ref-type="bibr" rid="ref39">39</xref>) with a learning rate of 3e-4 for 1,000 epochs. The learning rate was decayed by 5 if the decrease in the average training loss over 30 epochs was less than 5e-3. The model convergence criteria are: the decrease in the average training loss over 60 epochs must be less than 5e-3, or the learning rate must be less than 1e-6. For pCR prediction, after acquiring the segmentation result, the breast cancer lesion images were resampled to a size of 128&#x002A;128&#x002A;128. The initial learning rate was set to 1e-4, and the network was trained for 200 epochs.</p>
</sec>
</sec>
<sec id="sec29">
<label>4.2</label>
<title>Ablation studies</title>
<sec id="sec30">
<label>4.2.1</label>
<title>Effect of histogram matching</title>
<p>As introduced in Section 3.1.3, we used a simple histogram matching method to augment the training dataset in order to improve the transferability of the model. Here, we conducted an ablation study to show the effect of this domain adaptation method. As shown in <xref ref-type="fig" rid="fig5">Figure 5</xref>, for an MRI image (a) from the training dataset (i.e., samples from Guangdong Provincial People&#x2019;s Hospital), we randomly picked a target image (b) from Henan Renmin Hospital and applied gray-level histogram matching to the original training image so as to obtain an augmented sample (c).</p>
<fig position="float" id="fig5">
<label>Figure 5</label>
<caption>
<p>Example of gray-level histogram matching. <bold>(A)</bold> is a sample from Guangdong Provincial People&#x2019;s Hospital; <bold>(B)</bold> is a randomly picked sample from Henan Renmin Hospital; and <bold>(C)</bold> is the result of applying histogram matching to <bold>(A)</bold>.</p>
</caption>
<graphic xlink:href="fmed-10-1188207-g005.tif"/>
</fig>
<p>We show the results of the proposed segmentation model with and without domain adaptation in <xref ref-type="table" rid="tab1">Table 1</xref>. It is apparent that after applying the proposed domain adaptation method, i.e., histogram matching, the segmentation IoU increased by 7%. We also note that the training curve oscillates more than it does without histogram matching, as shown in <xref ref-type="fig" rid="fig6">Figure 6</xref>. This is because the gray-level distribution of the target dataset is introduced into the training samples.</p>
<table-wrap position="float" id="tab1">
<label>Table 1</label>
<caption>
<p>Performance of the breast cancer lesion region segmentation task with and without domain adaptation.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top">Task</th>
<th align="center" valign="top">Domain Adaptation</th>
<th align="center" valign="top">Dice</th>
<th align="center" valign="top">IoU</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top" rowspan="2">Breast cancer lesion region segmentation</td>
<td align="center" valign="top">No</td>
<td align="center" valign="top">0.78</td>
<td align="center" valign="top">0.65</td>
</tr>
<tr>
<td align="center" valign="top">Yes</td>
<td align="center" valign="top">0.83</td>
<td align="center" valign="top">0.72</td>
</tr>
</tbody>
</table>
</table-wrap>
<fig position="float" id="fig6">
<label>Figure 6</label>
<caption>
<p>Training curve of mammary gland segmentation.</p>
</caption>
<graphic xlink:href="fmed-10-1188207-g006.tif"/>
</fig>
</sec>
<sec id="sec31">
<label>4.2.2</label>
<title>ResNet depth</title>
<p>Normally, a deeper network implies stronger modeling ability; however, this is not always true in medical image processing models because of the higher risk of overfitting. We conducted an ablation study to show how the depth of the ResNet affects the model&#x2019;s performance. We trained the pCR prediction model with conventional ResNet18, ResNet34, and ResNet50, respectively, and the result is shown in <xref ref-type="table" rid="tab2">Table 2</xref>. We can see that the performance does not change much between models with different ResNet depths. Based on this, we chose ResNet43 in the experiments that follow.</p>
<table-wrap position="float" id="tab2">
<label>Table 2</label>
<caption>
<p>Result of pCR prediction with ResNet of different depths.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top">Task</th>
<th align="center" valign="top">ResNet depth</th>
<th align="center" valign="top">Accuracy</th>
<th align="center" valign="top">AUC</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top" rowspan="3">pCR prediction</td>
<td align="center" valign="top">ResNet18</td>
<td align="center" valign="top">0.70</td>
<td align="center" valign="top">0.68</td>
</tr>
<tr>
<td align="center" valign="top">ResNet34</td>
<td align="center" valign="top">0.72</td>
<td align="center" valign="top">0.69</td>
</tr>
<tr>
<td align="center" valign="top">ResNet50</td>
<td align="center" valign="top">0.71</td>
<td align="center" valign="top">0.69</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="sec32">
<label>4.2.3</label>
<title>Effect of the surrounding mammary gland</title>
<p>The performance of the prediction model is directly affected by the correlation between the input data and the prediction target. As for pCR prediction tasks, as pointed out in (<xref ref-type="bibr" rid="ref40">40</xref>), the mammary gland provides certain information when determining preoperative lymph node metastasis in breast cancer. Also, as pointed out by professionals, the MRI data of the mammary gland may contain abnormal information that may be related to patient prognosis. Thus, we conducted an ablation to test the influence of the surrounding mammary gland in predicting pCR. After acquiring the segmentation result of the breast cancer lesion, we expanded the periphery by using an expansion algorithm with kernels of three sizes, i.e., 5 pixels, 10 pixels, and 15 pixels. An example is shown in <xref ref-type="fig" rid="fig7">Figure 7</xref>. We used a circular expansion kernel in order to maintain the original shape of the segmented lesion. Then the expanded DCE-MRI data of the lesion region was used to train the proposed pCR prediction model, and the result is shown in <xref ref-type="table" rid="tab3">Table 3</xref>. It is quite obvious that the surrounding gland information does not help at all in pCR prediction, so we used the segmented lesion region directly for pCR prediction in the following experiments.</p>
<fig position="float" id="fig7">
<label>Figure 7</label>
<caption>
<p>Example of expanding a segmented lesion region by 5, 10 and 15 pixels, respectively, from left to right.</p>
</caption>
<graphic xlink:href="fmed-10-1188207-g007.tif"/>
</fig>
<table-wrap position="float" id="tab3">
<label>Table 3</label>
<caption>
<p>Result of pCR prediction with different expanding kernel sizes.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top">Task</th>
<th align="center" valign="top">Kernel size (pixel)</th>
<th align="center" valign="top">Accuracy</th>
<th align="center" valign="top">AUC</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top" rowspan="4">pCR prediction</td>
<td align="center" valign="top">-</td>
<td align="center" valign="top">0.72</td>
<td align="center" valign="top">0.69</td>
</tr>
<tr>
<td align="center" valign="top">5</td>
<td align="center" valign="top">0.42</td>
<td align="center" valign="top">0.60</td>
</tr>
<tr>
<td align="center" valign="top">10</td>
<td align="center" valign="top">0.57</td>
<td align="center" valign="top">0.77</td>
</tr>
<tr>
<td align="center" valign="top">15</td>
<td align="center" valign="top">0.42</td>
<td align="center" valign="top">0.76</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="sec33">
<label>4.2.4</label>
<title>Effects of multi-modal fusion</title>
<p>We also conducted an ablation study to show the effect of multi-modal fusion. We present the performance of pCR prediction with only DCE-MRI data of the segmented lesion and with both DCE-MRI data and four common types of molecular typing data (i.e., multi-modal fusion) in <xref ref-type="table" rid="tab4">Table 4</xref>. It is noted that the multi-modal fusion model provides a 13% increase in accuracy, which proves that the proposed model is effective.</p>
<table-wrap position="float" id="tab4">
<label>Table 4</label>
<caption>
<p>Result of pCR prediction with DCE-MRI data only and with multi-modal fusion.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top">Task</th>
<th align="left" valign="top">Model</th>
<th align="center" valign="top">Accuracy</th>
<th align="center" valign="top">AUC</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top" rowspan="2">pCR prediction</td>
<td align="left" valign="top">ResNet34</td>
<td align="center" valign="top">0.72</td>
<td align="center" valign="top">0.69</td>
</tr>
<tr>
<td align="left" valign="top">Multi-modal fusion</td>
<td align="center" valign="top">0.85</td>
<td align="center" valign="top">0.81</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
</sec>
<sec id="sec34">
<label>4.3</label>
<title>Experiment results</title>
<p>In this section, we present the results of the two-step lesion segmentation and pCR prediction. It should be noted that the experiments were conducted according to the method introduced in Section 3, and as explained in Section 4.2, the experiments were performed with domain adaptation, with ResNet34, without the surrounding mammary gland data of the lesion, and with multi-modal fusion.</p>
<sec id="sec35">
<label>4.3.1</label>
<title>Two-step lesion segmentation</title>
<sec id="sec36">
<label>4.3.1.1</label>
<title>Mammary gland segmentation.</title>
<p>The training curve is shown in <xref ref-type="fig" rid="fig6">Figure 6</xref> and examples of segmented mammary glands are shown in <xref ref-type="fig" rid="fig8">Figure 8</xref>.</p>
<fig position="float" id="fig8">
<label>Figure 8</label>
<caption>
<p>Examples of segmented mammary gland. <bold>(A)</bold> shows the mammary gland region in red, <bold>(B)</bold> shows a segmented mammary gland.</p>
</caption>
<graphic xlink:href="fmed-10-1188207-g008.tif"/>
</fig>
<p>The performance of the proposed method for mammary gland segmentation is shown in <xref ref-type="table" rid="tab5">Table 5</xref>. We achieved 93% IoU in the first segmentation task.</p>
<table-wrap position="float" id="tab5">
<label>Table 5</label>
<caption>
<p>Performance of the two-step lesion segmentation task.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top">Task</th>
<th align="center" valign="top">Dice</th>
<th align="center" valign="top">IoU</th>
<th align="center" valign="top">HD95</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top">Mammary gland segmentation</td>
<td align="center" valign="top">0.96&#x2009;&#x00B1;&#x2009;0.01</td>
<td align="center" valign="top">0.93&#x2009;&#x00B1;&#x2009;0.02</td>
<td align="center" valign="top">3.73&#x2009;&#x00B1;&#x2009;2.02</td>
</tr>
<tr>
<td align="left" valign="top">Breast cancer lesion segmentation</td>
<td align="center" valign="top">0.83</td>
<td align="center" valign="top">0.72</td>
<td align="center" valign="top">-</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="sec37">
<label>4.3.1.2</label>
<title>Breast cancer lesion segmentation.</title>
<p>Results from the first segmentation step, for example, <xref ref-type="fig" rid="fig8">Figure 8B</xref>, were used as input for the second segmentation step. The training curve is shown in <xref ref-type="fig" rid="fig9">Figure 9</xref>, and the performance of the proposed method for breast cancer lesion segmentation is shown in <xref ref-type="table" rid="tab5">Table 5</xref>.</p>
<fig position="float" id="fig9">
<label>Figure 9</label>
<caption>
<p>Training curve of breast cancer lesion segmentation.</p>
</caption>
<graphic xlink:href="fmed-10-1188207-g009.tif"/>
</fig>
</sec>
</sec>
<sec id="sec38">
<label>4.3.2</label>
<title>Pathologic complete response prediction</title>
<p>We used a multi-modal fusion model to predict whether axillary lymph nodes could achieve pCR after patients receive neoadjuvant therapy. DCE-MRI data of the segmented breast cancer lesion region and four types of molecular typing data commonly used in breast cancer treatment (i.e., HER2, ER, PR, and Ki-67) were utilized as input to the proposed multi-modal fusion model. The performance of pCR prediction by the proposed model is shown in <xref ref-type="table" rid="tab4">Table 4</xref>. The multi-modal fusion model achieved an accuracy of 85%, which is significantly high for pCR prediction of breast cancer with only non-invasive methods.</p>
<p>In addition, we performed McNemar&#x2019;s test of the two pCR prediction methods, one with DCE-MRI data only and the other with multi-modal fusion. We also performed a Chi-square goodness-of-fit test between each of the two methods and the ground truth. We randomly selected 200 samples to test each of the two methods, and the statistics of the pCR prediction results are shown in <xref ref-type="table" rid="tab6">Tables 6</xref>, <xref ref-type="table" rid="tab7">7</xref>, respectively. The McNemar&#x2019;s test showed that there exists a statistical difference between the two pCR prediction methods, while the Chi-square goodness-of-fit test revealed that the pCR prediction result by the multi-modal fusion method is more consistent with the ground truth distribution.</p>
<table-wrap position="float" id="tab6">
<label>Table 6</label>
<caption>
<p>Result of McNemar&#x2019;s test of two pCR prediction methods.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top" rowspan="2">Prediction with DCE-MRI</th>
<th align="center" valign="top" colspan="2">Prediction with Multi-modal</th>
<th align="center" valign="top" rowspan="2">Total</th>
<th align="center" valign="top" rowspan="2"><italic>&#x03C7;</italic><sup>2</sup></th>
<th align="center" valign="top" rowspan="2"><italic>p</italic></th>
</tr>
<tr>
<th align="center" valign="top">pCR</th>
<th align="center" valign="top">non-pCR</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="middle">pCR</td>
<td align="center" valign="top">63</td>
<td align="center" valign="top">1</td>
<td align="center" valign="top">64</td>
<td align="center" valign="top" rowspan="3">9.31</td>
<td align="center" valign="top" rowspan="3">0.0023</td>
</tr>
<tr>
<td align="left" valign="middle">non-pCR</td>
<td align="center" valign="top">12</td>
<td align="center" valign="top">124</td>
<td align="center" valign="top">136</td>
</tr>
<tr>
<td align="left" valign="middle">Total</td>
<td align="center" valign="top">75</td>
<td align="center" valign="top">125</td>
<td align="center" valign="top">200</td>
</tr>
</tbody>
</table>
</table-wrap>
<table-wrap position="float" id="tab7">
<label>Table 7</label>
<caption>
<p>Result of the Chi-square goodness-of-fit test between each of the two pCR prediction methods and the ground truth, respectively.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th/>
<th align="center" valign="top">pCR</th>
<th align="center" valign="top">non-pCR</th>
<th align="center" valign="top"><italic>&#x03C7;</italic><sup>2</sup></th>
<th align="center" valign="top"><italic>p</italic></th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="middle">Prediction with DCE-MRI</td>
<td align="center" valign="middle">64</td>
<td align="center" valign="middle">136</td>
<td align="center" valign="middle">7.43</td>
<td align="center" valign="middle">0.0064</td>
</tr>
<tr>
<td align="left" valign="middle">Prediction with Multi-modal</td>
<td align="center" valign="middle">75</td>
<td align="center" valign="middle">125</td>
<td align="center" valign="top">1.32</td>
<td align="center" valign="top">0.2509</td>
</tr>
<tr>
<td align="left" valign="middle">Ground truth</td>
<td align="center" valign="middle">83</td>
<td align="center" valign="middle">117</td>
<td/>
<td/>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
</sec>
</sec>
<sec sec-type="conclusions" id="sec39">
<label>5</label>
<title>Conclusion</title>
<p>In this paper, we presented a two-step lesion segmentation method to extract breast cancer lesion regions from DCE-MRI images, and in this process, we applied a simple histogram matching method to improve the transferability of the model. Then, we proposed a multi-modal (i.e., segmented DCE-MRI image and molecular typing data) fusion model to predict the probability of axillary lymph nodes achieving pCR after patients receive NAC. We collected 361 breast cancer samples from two hospitals to train and test the proposed segmentation method and the multi-modal fusion model. We demonstrated that our method achieves 93 and 72% IoU in mammary gland segmentation and breast cancer lesion segmentation tasks, respectively. We also showed that our multi-modal fusion model is effective and reaches 85% accuracy in pCR prediction using only data collected in a non-invasive manner. Although the IoU of breast cancer lesion segmentation is not very high (72%), it was used in the multi-modal fusion model and reached 85% accuracy in pCR prediction. This suggests that the presented method can be used for the prediction of treatment responses in breast cancer.</p>
<sec id="sec40">
<label>5.1</label>
<title>Limitations</title>
<p>The 361 breast cancer samples we collected for this study only include patients with solid tumors; therefore, this study focuses on lesion region segmentation of solid tumors and cannot be directly applied to other types of lesions, e.g., non-mass lesions or different breast parenchyma compositions. If the proposed method is to be used in other cases, it needs to be re-trained with enough specific data samples. Additionally, the proposed pCR prediction method requires a two-step process where we need to segment the breast cancer lesion from the DCE-MRI image, only then can we perform the final pCR prediction with the multi-model fusion model.</p>
</sec>
</sec>
<sec sec-type="data-availability" id="sec41">
<title>Data availability statement</title>
<p>The data analyzed in this study is subject to the following licenses/restrictions: The image and clinical medical data used to support the findings of this study are restricted by the Ethics Board of Guangdong Provincial People&#x2019;s Hospital in order to protect patient privacy. Data are available from Gang Fang, <email>gangf@gzhu.edu.cn</email> for researchers who meet the criteria for access to confidential data. Requests to access these datasets should be directed to Gang Fang, <email>gangf@gzhu.edu.cn</email>.</p>
</sec>
<sec sec-type="ethics-statement" id="sec42">
<title>Ethics statement</title>
<p>The studies involving human participants were reviewed and approved by the Ethics Board of Guangdong Provincial People&#x2019;s Hospital and Ethics Board of Henan Renmin Hospital. Written informed consent to participate in this study was provided by the participants.</p>
</sec>
<sec sec-type="author-contributions" id="sec43">
<title>Author contributions</title>
<p>GF designed and supervised the study. YL carried out the whole study. ZC assisted the whole study, conducted extra experiments. JC and ZS wrote the code for the study. All authors contributed to the article and approved the submitted version.</p>
</sec>
</body>
<back>
<ack>
<p>The study is supported by National Natural Science Foundation of China with grant number 61972107.</p>
</ack>
<sec sec-type="COI-statement" id="sec44">
<title>Conflict of interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec id="sec100" sec-type="disclaimer">
<title>Publisher&#x2019;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<ref-list>
<title>References</title>
<ref id="ref1"><label>1.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Siegel</surname><given-names>RL</given-names></name> <name><surname>Miller</surname><given-names>KD</given-names></name> <name><surname>Wagle</surname><given-names>NS</given-names></name> <name><surname>Jemal</surname><given-names>A</given-names></name></person-group>. <article-title>Cancer statistics, 2023</article-title>. <source>CA Cancer J Clin</source>. (<year>2023</year>) <volume>73</volume>:<fpage>17</fpage>&#x2013;<lpage>48</lpage>. doi: <pub-id pub-id-type="doi">10.3322/caac.21763</pub-id></citation></ref>
<ref id="ref2"><label>2.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Rouzier</surname><given-names>R</given-names></name> <name><surname>Extra</surname><given-names>JM</given-names></name> <name><surname>Klijanienko</surname><given-names>J</given-names></name> <name><surname>Falcou</surname><given-names>MC</given-names></name> <name><surname>Asselain</surname><given-names>B</given-names></name> <name><surname>Vincent-Salomon</surname><given-names>A</given-names></name> <etal/></person-group>. <article-title>Incidence and prognostic significance of complete axillary downstaging after primary chemotherapy in breast cancer patients with T1 to T3 tumors and cytologically proven axillary metastatic lymph nodes</article-title>. <source>J Clin Oncol</source>. (<year>2002</year>) <volume>20</volume>:<fpage>1304</fpage>&#x2013;<lpage>10</lpage>. doi: <pub-id pub-id-type="doi">10.1200/JCO.2002.20.5.1304</pub-id>, PMID: <pub-id pub-id-type="pmid">11870173</pub-id></citation></ref>
<ref id="ref3"><label>3.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Klauber-DeMore</surname><given-names>N</given-names></name> <name><surname>Ollila</surname><given-names>DW</given-names></name> <name><surname>Moore</surname><given-names>DT</given-names></name> <name><surname>Livasy</surname><given-names>C</given-names></name> <name><surname>Calvo</surname><given-names>BF</given-names></name> <name><surname>Kim</surname><given-names>HJ</given-names></name> <etal/></person-group>. <article-title>Size of residual lymph node metastasis after neoadjuvant chemotherapy in locally advanced breast cancer patients is prognostic</article-title>. <source>Ann Surg Oncol</source>. (<year>2006</year>) <volume>13</volume>:<fpage>685</fpage>&#x2013;<lpage>91</lpage>. doi: <pub-id pub-id-type="doi">10.1245/ASO.2006.03.010</pub-id>, PMID: <pub-id pub-id-type="pmid">16523367</pub-id></citation></ref>
<ref id="ref4"><label>4.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Mamounas</surname><given-names>EP</given-names></name> <name><surname>Anderson</surname><given-names>SJ</given-names></name> <name><surname>Dignam</surname><given-names>JJ</given-names></name> <name><surname>Bear</surname><given-names>HD</given-names></name> <name><surname>Julian</surname><given-names>TB</given-names></name> <name><surname>Geyer</surname><given-names>CE</given-names> <suffix>Jr</suffix></name> <etal/></person-group>. <article-title>Predictors of locoregional recurrence after neoadjuvant chemotherapy: results from combined analysis of National Surgical Adjuvant Breast and bowel project B-18 and B-27</article-title>. <source>J Clin Oncol</source>. (<year>2012</year>) <volume>30</volume>:<fpage>3960</fpage>&#x2013;<lpage>6</lpage>. doi: <pub-id pub-id-type="doi">10.1200/JCO.2011.40.8369</pub-id>, PMID: <pub-id pub-id-type="pmid">23032615</pub-id></citation></ref>
<ref id="ref5"><label>5.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Mougalian</surname><given-names>SS</given-names></name> <name><surname>Hernandez</surname><given-names>M</given-names></name> <name><surname>Lei</surname><given-names>X</given-names></name> <name><surname>Lynch</surname><given-names>S</given-names></name> <name><surname>Kuerer</surname><given-names>HM</given-names></name> <name><surname>Symmans</surname><given-names>WF</given-names></name> <etal/></person-group>. <article-title>Ten-year outcomes of patients with breast cancer with cytologically confirmed axillary lymph node metastases and pathologic complete response after primary systemic chemotherapy</article-title>. <source>JAMA Oncol</source>. (<year>2016</year>) <volume>2</volume>:<fpage>508</fpage>&#x2013;<lpage>16</lpage>. doi: <pub-id pub-id-type="doi">10.1001/jamaoncol.2015.4935</pub-id>, PMID: <pub-id pub-id-type="pmid">26720612</pub-id></citation></ref>
<ref id="ref6"><label>6.</label><citation citation-type="other"><person-group person-group-type="author"><collab id="coll1">Priscilla J. Slanetz</collab><name><surname>Moy</surname><given-names>L</given-names></name> <name><surname>Baron</surname><given-names>P</given-names></name> <name><surname>diFlorio</surname><given-names>RM</given-names></name> <name><surname>Green</surname><given-names>ED</given-names></name> <name><surname>Heller</surname><given-names>SL</given-names></name> <etal/></person-group>. <article-title>ACR appropriateness criteria((R)) monitoring response to neoadjuvant systemic therapy for breast cancer</article-title>. <source>J Am Coll Radiol</source>. (<year>2017</year>) <volume>14</volume>:<fpage>S462</fpage>&#x2013;<lpage>75</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.jacr.2017.08.037</pub-id></citation></ref>
<ref id="ref7"><label>7.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Mann</surname><given-names>RM</given-names></name> <name><surname>Cho</surname><given-names>N</given-names></name> <name><surname>Moy</surname><given-names>L</given-names></name></person-group>. <article-title>Breast MRI: state of the art</article-title>. <source>Radiology</source>. (<year>2019</year>) <volume>292</volume>:<fpage>520</fpage>&#x2013;<lpage>36</lpage>. doi: <pub-id pub-id-type="doi">10.1148/radiol.2019182947</pub-id>, PMID: <pub-id pub-id-type="pmid">31361209</pub-id></citation></ref>
<ref id="ref8"><label>8.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hosny</surname><given-names>A</given-names></name> <name><surname>Parmar</surname><given-names>C</given-names></name> <name><surname>Quackenbush</surname><given-names>J</given-names></name> <name><surname>Schwartz</surname><given-names>LH</given-names></name> <name><surname>Aerts</surname><given-names>H</given-names></name></person-group>. <article-title>Artificial intelligence in radiology</article-title>. <source>Nat Rev Cancer</source>. (<year>2018</year>) <volume>18</volume>:<fpage>500</fpage>&#x2013;<lpage>10</lpage>. doi: <pub-id pub-id-type="doi">10.1038/s41568-018-0016-5</pub-id>, PMID: <pub-id pub-id-type="pmid">29777175</pub-id></citation></ref>
<ref id="ref9"><label>9.</label><citation citation-type="other"><person-group person-group-type="author"><name><surname>Long</surname><given-names>Jonathan</given-names></name> <name><surname>Shelhamer</surname><given-names>Evan</given-names></name> <name><surname>Darrell</surname><given-names>Trevor</given-names></name></person-group>. (<year>2015</year>). &#x201C;Fully convolutional networks for semantic segmentation.&#x201D; in <italic>Proceedings of the IEEE conference on computer vision and pattern recognition</italic>.</citation></ref>
<ref id="ref10"><label>10.</label><citation citation-type="other"><person-group person-group-type="author"><name><surname>Olaf</surname><given-names>Ronneberger</given-names></name> <name><surname>Fischer</surname><given-names>Philipp</given-names></name> <name><surname>Brox</surname><given-names>Thomas</given-names></name></person-group>. (<year>2015</year>). &#x201C;U-net: convolutional networks for biomedical image segmentation.&#x201D; in <italic>International conference on medical image computing and computer-assisted intervention</italic>. Springer, Cham.</citation></ref>
<ref id="ref11"><label>11.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Singh</surname><given-names>VK</given-names></name> <name><surname>Rashwan</surname><given-names>HA</given-names></name> <name><surname>Romani</surname><given-names>S</given-names></name> <name><surname>Akram</surname><given-names>F</given-names></name> <name><surname>Pandey</surname><given-names>N</given-names></name> <name><surname>Sarker</surname><given-names>MMK</given-names></name> <etal/></person-group>. <article-title>Breast tumor segmentation and shape classification in mammograms using generative adversarial and convolutional neural network</article-title>. <source>Expert Syst Appl</source>. (<year>2020</year>) <volume>139</volume>:<fpage>112855</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.eswa.2019.112855</pub-id></citation></ref>
<ref id="ref12"><label>12.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Isensee</surname><given-names>F</given-names></name> <name><surname>Jaeger</surname><given-names>PF</given-names></name> <name><surname>Kohl</surname><given-names>SAA</given-names></name> <name><surname>Petersen</surname><given-names>J</given-names></name> <name><surname>Maier-Hein</surname><given-names>KH</given-names></name></person-group>. <article-title>nnU-net: a self-configuring method for deep learning-based biomedical image segmentation</article-title>. <source>Nat Methods</source>. (<year>2021</year>) <volume>18</volume>:<fpage>203</fpage>&#x2013;<lpage>11</lpage>. doi: <pub-id pub-id-type="doi">10.1038/s41592-020-01008-z</pub-id>, PMID: <pub-id pub-id-type="pmid">33288961</pub-id></citation></ref>
<ref id="ref13"><label>13.</label><citation citation-type="other"><person-group person-group-type="author"><name><surname>Ma</surname><given-names>Jun</given-names></name></person-group>. (<year>2020</year>). &#x201C;Histogram matching augmentation for domain adaptation with application to multi-centre, multi-vendor and multi-disease cardiac image segmentation.&#x201D; in <italic>International workshop on statistical atlases and computational models of the heart</italic>. Springer, Cham,</citation></ref>
<ref id="ref14"><label>14.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ramos-Vara</surname><given-names>JA</given-names></name></person-group>. <article-title>Technical aspects of immunohistochemistry</article-title>. <source>Vet Pathol</source>. (<year>2005</year>) <volume>42</volume>:<fpage>405</fpage>&#x2013;<lpage>26</lpage>. doi: <pub-id pub-id-type="doi">10.1354/vp.42-4-405</pub-id></citation></ref>
<ref id="ref15"><label>15.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lo</surname><given-names>SCB</given-names></name> <name><surname>Lou</surname><given-names>SLA</given-names></name> <name><surname>Jyh-Shyan Lin</surname></name> <name><surname>Freedman</surname><given-names>MT</given-names></name> <name><surname>Chien</surname><given-names>MV</given-names></name> <name><surname>Mun</surname><given-names>SK</given-names></name></person-group>. <article-title>Artificial convolution neural network techniques and applications for lung nodule detection</article-title>. <source>IEEE Trans Med Imaging</source>. (<year>1995</year>) <volume>14</volume>:<fpage>711</fpage>&#x2013;<lpage>8</lpage>. doi: <pub-id pub-id-type="doi">10.1109/42.476112</pub-id>, PMID: <pub-id pub-id-type="pmid">18215875</pub-id></citation></ref>
<ref id="ref16"><label>16.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Krizhevsky</surname><given-names>A</given-names></name> <name><surname>Sutskever</surname><given-names>I</given-names></name> <name><surname>Hinton</surname><given-names>GE</given-names></name></person-group>. <article-title>Imagenet classification with deep convolutional neural networks</article-title>. <source>Adv Neural Inf Proces Syst</source>. (<year>2012</year>) <volume>25</volume>:<fpage>1097</fpage>&#x2013;<lpage>105</lpage>.</citation></ref>
<ref id="ref17"><label>17.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Russakovsky</surname><given-names>O</given-names></name> <name><surname>Deng</surname><given-names>J</given-names></name> <name><surname>Su</surname><given-names>H</given-names></name> <name><surname>Krause</surname><given-names>J</given-names></name> <name><surname>Satheesh</surname><given-names>S</given-names></name> <name><surname>Ma</surname><given-names>S</given-names></name> <etal/></person-group>. <article-title>Imagenet large scale visual recognition challenge</article-title>. <source>Int J Comput Vis</source>. (<year>2015</year>) <volume>115</volume>:<fpage>211</fpage>&#x2013;<lpage>52</lpage>. doi: <pub-id pub-id-type="doi">10.1007/s11263-015-0816-y</pub-id></citation></ref>
<ref id="ref18"><label>18.</label><citation citation-type="other"><person-group person-group-type="author"><name><surname>He</surname><given-names>Kaiming</given-names></name> <name><surname>Zhang</surname><given-names>Xiangyu</given-names></name> <name><surname>Ren</surname><given-names>Shaoqing</given-names></name> <name><surname>Sun</surname><given-names>Jian</given-names></name></person-group> (<year>2016</year>). &#x201C;Deep residual learning for image recognition.&#x201D; in <italic>Proceedings of the IEEE conference on computer vision and pattern recognition</italic>.</citation></ref>
<ref id="ref19"><label>19.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Simonyan</surname><given-names>K</given-names></name> <name><surname>Zisserman</surname><given-names>A</given-names></name></person-group>. <article-title>Very deep convolutional networks for large-scale image recognition</article-title>. <source>arXiv</source>. (<year>2014</year>) arXiv:1409.1556</citation></ref>
<ref id="ref20"><label>20.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Shenton</surname><given-names>ME</given-names></name> <name><surname>Dickey</surname><given-names>CC</given-names></name> <name><surname>Frumin</surname><given-names>M</given-names></name> <name><surname>McCarley</surname><given-names>RW</given-names></name></person-group>. <article-title>A review of MRI findings in schizophrenia</article-title>. <source>Schizophr Res</source>. (<year>2001</year>) <volume>49</volume>:<fpage>1</fpage>&#x2013;<lpage>52</lpage>. doi: <pub-id pub-id-type="doi">10.1016/S0920-9964(01)00163-3</pub-id>, PMID: <pub-id pub-id-type="pmid">11343862</pub-id></citation></ref>
<ref id="ref21"><label>21.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Baur</surname><given-names>A</given-names></name> <name><surname>Reiser</surname><given-names>MF</given-names></name></person-group>. <article-title>Diffusion-weighted imaging of the musculoskeletal system in humans</article-title>. <source>Skelet Radiol</source>. (<year>2000</year>) <volume>29</volume>:<fpage>555</fpage>&#x2013;<lpage>62</lpage>. doi: <pub-id pub-id-type="doi">10.1007/s002560000243</pub-id></citation></ref>
<ref id="ref22"><label>22.</label><citation citation-type="other"><person-group person-group-type="author"><name><surname>&#x00C7;i&#x00E7;ek</surname><given-names>&#x00D6;zg&#x00FC;n</given-names></name> <name><surname>Abdulkadir</surname><given-names>Ahmed</given-names></name> <name><surname>Lienkamp</surname><given-names>Soeren S.</given-names></name> <name><surname>Brox</surname><given-names>Thomas</given-names></name> <name><surname>Ronneberger</surname><given-names>Olaf</given-names></name></person-group> (<year>2016</year>). &#x201C;3D U-net: learning dense volumetric segmentation from sparse annotation.&#x201D; in <italic>International conference on medical image computing and computer-assisted intervention</italic>. Springer, Cham.</citation></ref>
<ref id="ref23"><label>23.</label><citation citation-type="other"><person-group person-group-type="author"><name><surname>Milletari</surname><given-names>F</given-names></name> <name><surname>Navab</surname><given-names>N</given-names></name> <name><surname>Ahmadi</surname><given-names>S A</given-names></name></person-group>. (<year>2016</year>). &#x201C;V-net: fully convolutional neural networks for volumetric medical image segmentation.&#x201D; in <italic>The fourth international conference on 3D vision (3DV). IEEE</italic>. pp. 565&#x2013;571.</citation></ref>
<ref id="ref24"><label>24.</label><citation citation-type="other"><person-group person-group-type="author"><name><surname>Xiao</surname><given-names>Xiao</given-names></name> <name><surname>Lian</surname><given-names>Shen</given-names></name> <name><surname>Luo</surname><given-names>Zhiming</given-names></name> <name><surname>Li</surname><given-names>Shaozi</given-names></name></person-group> (<year>2018</year>). &#x201C;Weighted res-unet for high-quality retina vessel segmentation.&#x201D; in <italic>2018 9th international conference on information technology in medicine and education (ITME)</italic>. IEEE.</citation></ref>
<ref id="ref25"><label>25.</label><citation citation-type="book"><person-group person-group-type="author"><name><surname>Drozdzal</surname><given-names>Michal</given-names></name> <name><surname>Vorontsov</surname><given-names>Eugene</given-names></name> <name><surname>Chartrand</surname><given-names>Gabriel</given-names></name> <name><surname>Kadoury</surname><given-names>Samuel</given-names></name> <name><surname>Pal</surname><given-names>Chris</given-names></name></person-group> <source>The importance of skip connections in biomedical image segmentation. Deep learning and data labeling for medical applications</source>. <publisher-name>Springer</publisher-name>, <publisher-loc>Cham</publisher-loc>, (<year>2016</year>). <fpage>179</fpage>&#x2013;<lpage>187</lpage>.</citation></ref>
<ref id="ref26"><label>26.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname><given-names>C</given-names></name> <name><surname>Yan</surname><given-names>X</given-names></name> <name><surname>Smith</surname><given-names>M</given-names></name> <name><surname>Kochhar</surname><given-names>K</given-names></name> <name><surname>Rubin</surname><given-names>M</given-names></name> <name><surname>Warren</surname><given-names>SM</given-names></name> <etal/></person-group>. <article-title>A unified framework for automatic wound segmentation and analysis with deep convolutional neural networks</article-title>. <source>Annu Int Conf IEEE Eng Med Biol Soc</source>. (<year>2015</year>) <volume>2015</volume>:<fpage>2415</fpage>&#x2013;<lpage>8</lpage>. doi: <pub-id pub-id-type="doi">10.1109/EMBC.2015.7318881</pub-id>, PMID: <pub-id pub-id-type="pmid">26736781</pub-id></citation></ref>
<ref id="ref27"><label>27.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Brosch</surname><given-names>T</given-names></name> <name><surname>Tang</surname><given-names>LYW</given-names></name> <name><surname>Yoo</surname><given-names>Y</given-names></name> <name><surname>Li</surname><given-names>DKB</given-names></name> <name><surname>Traboulsee</surname><given-names>A</given-names></name> <name><surname>Tam</surname><given-names>R</given-names></name></person-group>. <article-title>Deep 3D convolutional encoder networks with shortcuts for multiscale feature integration applied to multiple sclerosis lesion segmentation</article-title>. <source>IEEE Trans Med Imaging</source>. (<year>2016</year>) <volume>35</volume>:<fpage>1229</fpage>&#x2013;<lpage>39</lpage>. doi: <pub-id pub-id-type="doi">10.1109/TMI.2016.2528821</pub-id></citation></ref>
<ref id="ref28"><label>28.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Iglovikov</surname><given-names>V</given-names></name> <name><surname>Shvets</surname><given-names>A</given-names></name></person-group>. <article-title>Ternausnet: U-net with vgg11 encoder pre-trained on imagenet for image segmentation</article-title>. <source>arXiv</source>. (<year>2018</year>) arXiv:1801.05746</citation></ref>
<ref id="ref29"><label>29.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ibtehaz</surname><given-names>N</given-names></name> <name><surname>Sohel Rahman</surname><given-names>M</given-names></name></person-group>. <article-title>MultiResUNet: rethinking the U-net architecture for multimodal biomedical image segmentation</article-title>. <source>Neural Netw</source>. (<year>2020</year>) <volume>121</volume>:<fpage>74</fpage>&#x2013;<lpage>87</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.neunet.2019.08.025</pub-id></citation></ref>
<ref id="ref30"><label>30.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chen</surname><given-names>J</given-names></name> <name><surname>Lu</surname><given-names>Y</given-names></name> <name><surname>Yu</surname><given-names>Q</given-names></name> <name><surname>Luo</surname><given-names>X</given-names></name> <name><surname>Adeli</surname><given-names>E</given-names></name> <name><surname>Wang</surname><given-names>Y</given-names></name> <etal/></person-group>. <article-title>Transunet: transformers make strong encoders for medical image segmentation</article-title>. <source>arXiv</source>. (<year>2021</year>) arXiv:2102.04306</citation></ref>
<ref id="ref31"><label>31.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Stollenga</surname><given-names>MF</given-names></name> <name><surname>Byron</surname><given-names>W</given-names></name> <name><surname>Liwichi</surname><given-names>M</given-names></name> <name><surname>Byeon</surname><given-names>W</given-names></name> <name><surname>Liwicki</surname><given-names>M</given-names></name> <name><surname>Schmidhuber</surname><given-names>J</given-names></name></person-group>. <article-title>Parallel multi-dimensional LSTM, with application to fast biomedical volumetric image segmentation</article-title>. <source>Comp Sci</source>. (<year>2015</year>) [Preprint]</citation></ref>
<ref id="ref32"><label>32.</label><citation citation-type="other"><person-group person-group-type="author"><name><surname>Andermatt</surname><given-names>S</given-names></name> <name><surname>Pezold</surname><given-names>S</given-names></name> <name><surname>Cattin</surname><given-names>P</given-names></name></person-group>. (<year>2016</year>). &#x201C;Multi-dimensional gated recurrent units for the segmentation of biomedical 3D-data.&#x201D; in <italic>International Conference on Medical Image Computing and Computer Assisted Intervention</italic>.</citation></ref>
<ref id="ref33"><label>33.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Yip</surname><given-names>SSF</given-names></name> <name><surname>Aerts</surname><given-names>HJWL</given-names></name></person-group>. <article-title>Applications and limitations of radiomics</article-title>. <source>Phys Med Biol</source>. (<year>2016</year>) <volume>61</volume>:<fpage>R150</fpage>&#x2013;<lpage>66</lpage>. doi: <pub-id pub-id-type="doi">10.1088/0031-9155/61/13/R150</pub-id>, PMID: <pub-id pub-id-type="pmid">27269645</pub-id></citation></ref>
<ref id="ref34"><label>34.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Liu</surname><given-names>Z</given-names></name> <name><surname>Li</surname><given-names>Z</given-names></name> <name><surname>Qu</surname><given-names>J</given-names></name> <name><surname>Zhang</surname><given-names>R</given-names></name> <name><surname>Zhou</surname><given-names>X</given-names></name> <name><surname>Li</surname><given-names>L</given-names></name> <etal/></person-group>. <article-title>Radiomics of multiparametric MRI for pretreatment prediction of pathologic complete response to neoadjuvant chemotherapy in breast cancer: a multicenter study</article-title>. <source>Clin Cancer Res</source>. (<year>2019</year>) <volume>25</volume>:<fpage>3538</fpage>&#x2013;<lpage>47</lpage>. doi: <pub-id pub-id-type="doi">10.1158/1078-0432.CCR-18-3190</pub-id>, PMID: <pub-id pub-id-type="pmid">30842125</pub-id></citation></ref>
<ref id="ref35"><label>35.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Shen</surname><given-names>D</given-names></name> <name><surname>Wu</surname><given-names>G</given-names></name> <name><surname>Suk</surname><given-names>HI</given-names></name></person-group>. <article-title>Deep learning in medical image analysis</article-title>. <source>Annu Rev Biomed Eng</source>. (<year>2017</year>) <volume>19</volume>:<fpage>221</fpage>&#x2013;<lpage>48</lpage>. doi: <pub-id pub-id-type="doi">10.1146/annurev-bioeng-071516-044442</pub-id>, PMID: <pub-id pub-id-type="pmid">28301734</pub-id></citation></ref>
<ref id="ref36"><label>36.</label><citation citation-type="book"><person-group person-group-type="author"><name><surname>Sudre</surname><given-names>C H</given-names></name> <name><surname>Li</surname><given-names>W</given-names></name> <name><surname>Vercauteren</surname><given-names>T</given-names></name> <name><surname>Ourselin</surname><given-names>S.</given-names></name> <name><surname>Jorge Cardoso</surname><given-names>M.</given-names></name></person-group> (<year>2017</year>). <source>Generalised dice overlap as a deep learning loss function for highly unbalanced segmentations[M]//Deep learning in medical image analysis and multimodal learning for clinical decision support</source>. <publisher-name>Springer</publisher-name>, <publisher-loc>Cham</publisher-loc>, <fpage>240</fpage>&#x2013;<lpage>248</lpage>.</citation></ref>
<ref id="ref37"><label>37.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Liu</surname><given-names>Q</given-names></name> <name><surname>Dou</surname><given-names>Q</given-names></name> <name><surname>Yu</surname><given-names>L</given-names></name> <name><surname>Heng</surname><given-names>PA</given-names></name></person-group>. <article-title>MS-net: multi-site network for improving prostate segmentation with heterogeneous MRI data</article-title>. <source>IEEE Trans Med Imaging</source>. (<year>2020</year>) <volume>39</volume>:<fpage>2713</fpage>&#x2013;<lpage>24</lpage>. doi: <pub-id pub-id-type="doi">10.1109/TMI.2020.2974574</pub-id></citation></ref>
<ref id="ref38"><label>38.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Dou</surname><given-names>Q</given-names></name> <name><surname>Liu</surname><given-names>Q</given-names></name> <name><surname>Heng</surname><given-names>PA</given-names></name> <name><surname>Glocker</surname><given-names>B</given-names></name></person-group>. <article-title>Unpaired multi-modal segmentation via knowledge distillation</article-title>. <source>IEEE Trans Med Imaging</source>. (<year>2020</year>) <volume>39</volume>:<fpage>2415</fpage>&#x2013;<lpage>25</lpage>. doi: <pub-id pub-id-type="doi">10.1109/TMI.2019.2963882</pub-id>, PMID: <pub-id pub-id-type="pmid">32012001</pub-id></citation></ref>
<ref id="ref39"><label>39.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kingma</surname><given-names>DP</given-names></name> <name><surname>Ba</surname><given-names>J</given-names></name></person-group>. <article-title>Adam: a method for stochastic optimization</article-title>. <source>arXiv</source>. (<year>2014</year>) arXiv:1412.6980</citation></ref>
<ref id="ref40"><label>40.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Liu</surname><given-names>C</given-names></name> <name><surname>Zhao</surname><given-names>Z</given-names></name> <name><surname>Gu</surname><given-names>X</given-names></name> <name><surname>Sun</surname><given-names>L</given-names></name> <name><surname>Chen</surname><given-names>G</given-names></name> <name><surname>Zhang</surname><given-names>H</given-names></name> <etal/></person-group>. <article-title>Establishment and verification of a bagged-trees-based model for prediction of sentinel lymph node metastasis for early breast cancer patients</article-title>. <source>Front Oncol</source>. (<year>2019</year>) <volume>9</volume>:<fpage>282</fpage>. doi: <pub-id pub-id-type="doi">10.3389/fonc.2019.00282</pub-id></citation></ref>
</ref-list>
</back>
</article>