<?xml version="1.0" encoding="UTF-8" standalone="no"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Neurosci.</journal-id>
<journal-title>Frontiers in Neuroscience</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Neurosci.</abbrev-journal-title>
<issn pub-type="epub">1662-453X</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fnins.2023.1118340</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Neuroscience</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Unified ICH quantification and prognosis prediction in NCCT images using a multi-task interpretable network</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name><surname>Gong</surname> <given-names>Kai</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="author-notes" rid="fn002"><sup>&#x02020;</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/1460133/overview"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Dai</surname> <given-names>Qian</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<xref ref-type="author-notes" rid="fn002"><sup>&#x02020;</sup></xref>
</contrib>
<contrib contrib-type="author">
<name><surname>Wang</surname> <given-names>Jiacheng</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<xref ref-type="author-notes" rid="fn002"><sup>&#x02020;</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/2131015/overview"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Zheng</surname> <given-names>Yingbin</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
</contrib>
<contrib contrib-type="author">
<name><surname>Shi</surname> <given-names>Tao</given-names></name>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/1493432/overview"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Yu</surname> <given-names>Jiaxing</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
</contrib>
<contrib contrib-type="author">
<name><surname>Chen</surname> <given-names>Jiangwang</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name><surname>Huang</surname> <given-names>Shaohui</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>&#x0002A;</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/671452/overview"/>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name><surname>Wang</surname> <given-names>Zhanxiang</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="corresp" rid="c002"><sup>&#x0002A;</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/1337286/overview"/>
</contrib>
</contrib-group>
<aff id="aff1"><sup>1</sup><institution>The First Affiliated Hospital of Xiamen University, Xiamen University, Xiamen</institution>, <addr-line>Fujian</addr-line>, <country>China</country></aff>
<aff id="aff2"><sup>2</sup><institution>Department of Computer Science, School of Informatics, Xiamen University, Xiamen</institution>, <addr-line>Fujian</addr-line>, <country>China</country></aff>
<aff id="aff3"><sup>3</sup><institution>Ningbo Medical Center Lihuili Hospital, Ningbo</institution>, <addr-line>Zhejiang</addr-line>, <country>China</country></aff>
<author-notes>
<fn fn-type="edited-by"><p>Edited by: Baptiste Magnier, Mines-Telecom Institute Al&#x000E8;s, France</p></fn>
<fn fn-type="edited-by"><p>Reviewed by: Zhenghua Xu, Hebei University of Technology, China; Dan Wu, Shenzhen Institutes of Advanced Technology (CAS), China</p></fn>
<corresp id="c001">&#x0002A;Correspondence: Shaohui Huang <email>hsh&#x00040;xmu.edu.cn</email></corresp>
<corresp id="c002">Zhanxiang Wang <email>wangzx&#x00040;xmu.edu.cn</email></corresp>
<fn fn-type="other" id="fn001"><p>This article was submitted to Brain Imaging Methods, a section of the journal Frontiers in Neuroscience</p></fn>
<fn fn-type="equal" id="fn002"><p>&#x02020;These authors have contributed equally to this work</p></fn></author-notes>
<pub-date pub-type="epub">
<day>14</day>
<month>03</month>
<year>2023</year>
</pub-date>
<pub-date pub-type="collection">
<year>2023</year>
</pub-date>
<volume>17</volume>
<elocation-id>1118340</elocation-id>
<history>
<date date-type="received">
<day>07</day>
<month>12</month>
<year>2022</year>
</date>
<date date-type="accepted">
<day>23</day>
<month>02</month>
<year>2023</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x000A9; 2023 Gong, Dai, Wang, Zheng, Shi, Yu, Chen, Huang and Wang.</copyright-statement>
<copyright-year>2023</copyright-year>
<copyright-holder>Gong, Dai, Wang, Zheng, Shi, Yu, Chen, Huang and Wang</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p></license> </permissions>
<abstract>
<p>With the recent development of deep learning, the regression, classification, and segmentation tasks of Computer-Aided Diagnosis (CAD) using Non-Contrast head Computed Tomography (NCCT) for spontaneous IntraCerebral Hematoma (ICH) have become popular in the field of emergency medicine. However, a few challenges such as time-consuming of ICH volume manual evaluation, excessive cost demanding patient-level predictions, and the requirement for high performance in both accuracy and interpretability remain. This paper proposes a multi-task framework consisting of upstream and downstream components to overcome these challenges. In the upstream, a weight-shared module is trained as a robust feature extractor that captures global features by performing multi-tasks (regression and classification). In the downstream, two heads are used for two different tasks (regression and classification). The final experimental results show that the multi-task framework has better performance than single-task framework. And it also reflects its good interpretability in the heatmap generated by Gradient-weighted Class Activation Mapping (Grad-CAM), which is a widely used model interpretation method, and will be presented in subsequent sections.</p></abstract>
<kwd-group>
<kwd>intracerebral hematoma (ICH)</kwd>
<kwd>Non-Contrast head Computed Tomography (NCCT)</kwd>
<kwd>multi-task</kwd>
<kwd>ResNet</kwd>
<kwd>interpretability</kwd>
</kwd-group>
<counts>
<fig-count count="5"/>
<table-count count="2"/>
<equation-count count="4"/>
<ref-count count="69"/>
<page-count count="11"/>
<word-count count="8171"/>
</counts>
</article-meta>
</front>
<body>
<sec sec-type="intro" id="s1">
<title>1. Introduction</title>
<p>Spontaneous IntraCerebral Hematoma (ICH) is characterized by high incidence, high disability, and high mortality, accounting for approximately 10&#x02013;20% of all strokes (Raafat et al., <xref ref-type="bibr" rid="B40">2020</xref>). The initial hematoma volume is the strongest predictor of mortality and functional outcomes (Beslow et al., <xref ref-type="bibr" rid="B3">2014</xref>). Fast and precise evaluation of ICH volume through Non-Contrast Computed Tomography (NCCT) images is generally considered the first and the most critical step to making further clinical decisions including medical and surgical options. As the evaluation of ICH volume is relatively a time-consuming process, head NCCT reports usually provide only a general description of ICH location and shape, whereas the ICH volume is not calculated. In most cases, the ICH volume is assessed manually by clinicians such as neurosurgeons or neurology physicians, using various methods including ABC/2, ABC/2.4, ABC/3,1/2Sh, and 2/3Sh. Among them, ABC/2, also known as the Tada formula, is the most used one. For further explanation, A is the maximum axial hematoma diameter, B is the maximum axial diameter perpendicular to A on the same slice, and C is the vertical diameter of the hematoma (Beslow et al., <xref ref-type="bibr" rid="B4">2010</xref>). The reliability of the ABC/2 formula is shown good correlation with computerized ICH volume measurements for small and uniformly shaped ICHs, but the accuracy can be affected by observer variability and imprecision (Oge et al., <xref ref-type="bibr" rid="B33">2021</xref>). The method is prone to overestimate ICH volume by approximately 20% and even misestimate more for large and irregularly shaped hemorrhages (Patel et al., <xref ref-type="bibr" rid="B37">2019</xref>). The 3D slicer software (<ext-link ext-link-type="uri" xlink:href="http://www.slicer.org">http://www.slicer.org</ext-link>) provides a free open source software platform for biomedical research. It can identify hematoma pixels based on CT images and reconstruct blood clots in a three-dimensional manner, which is free from restriction by hematoma morphology and bleeding sites. Compared with manual methods, 3D slicer method is considered as a more stable and capable method of high precision for the volume calculation of most hematomas and is gradually accepted as an effective measurement method clinically (Xu et al., <xref ref-type="bibr" rid="B59">2014</xref>; Chen et al., <xref ref-type="bibr" rid="B9">2020</xref>; Gong et al., <xref ref-type="bibr" rid="B17">2021</xref>).</p>
<p>Newly automatic, versatile, easily deployable, and accurate hematoma volume assessment tools are highly needed giving their clinical significance. Numerous attempts have been conducted to develop 3D slicer as well as other computer-assisted automated tools for hematoma volume evaluation (Huttner et al., <xref ref-type="bibr" rid="B22">2006</xref>; Wang et al., <xref ref-type="bibr" rid="B56">2009</xref>; Zhao et al., <xref ref-type="bibr" rid="B65">2009</xref>; Yang et al., <xref ref-type="bibr" rid="B60">2013</xref>; Xu et al., <xref ref-type="bibr" rid="B59">2014</xref>), but the methods have not been fully automated yet, and still require a significant amount of decisions to be made by researchers, which remains a challenging problem.</p>
<p>With the recent growth of deep learning, the evaluation methods of hematomas with deep learning-based techniques have been advanced dramatically, including the ResNet model, the DenseNet model, and the H-DenseUNet model, see Zhao et al. (<xref ref-type="bibr" rid="B66">2021</xref>), Mantas (<xref ref-type="bibr" rid="B32">2020</xref>), Dawud et al. (<xref ref-type="bibr" rid="B10">2019</xref>), Zhou et al. (<xref ref-type="bibr" rid="B68">2022</xref>), and Gou and He (<xref ref-type="bibr" rid="B18">2021</xref>).</p>
<p>For the evaluation of intracranial hemorrhage volume, the deep learning method has been proved to be superior to the clinical method in stability and accuracy many times (Freeman et al., <xref ref-type="bibr" rid="B15">2008</xref>; Sharrock et al., <xref ref-type="bibr" rid="B49">2022</xref>). Most quantitative assessments of ICH volume are based on the segmentation or classification of ICH areas. Phaphuangwittayakul et al. (<xref ref-type="bibr" rid="B38">2022</xref>) propose a quantitative assessment algorithm to automatically measure both thickness and volume <italic>via</italic> the 3D shape mask combined with the output probabilities of the classification network. Xu et al. (<xref ref-type="bibr" rid="B58">2021</xref>) realize hematoma segmentation and volume evaluation by Dense-Unet (Cai et al., <xref ref-type="bibr" rid="B6">2020</xref>; Sharrock et al., <xref ref-type="bibr" rid="B48">2021</xref>), and is based on V-Net architecture. Much literature focuses on developing better neural network architectures and training strategies to optimize ICH segmentation, on which post-processing can be performed to automatically measure the volume of intracerebral hemorrhage. In this paper, we do not have the segmentation mask of the cerebral hemorrhage area. Instead, we have the measurement results of the volume of cerebral hemorrhage by multiple doctors. On the basis of training on this, we expect the model to be able to locate the bleeding location and even mark the bleeding area, which is what we intend to continue to study in the future.</p>
<p>This paper describes a lightweight multi-task learning framework firstly, which is specifically designed to identify and evaluate ICH hematoma volumes by training a large number of NCCT images collected from 258 patients with ICH. Then, based on the assessment of ICH hematoma volumes, further prognosis analysis of intracerebral patients through the multi-task framework is discussed. Finally, satisfactory results are obtained for both tasks. For the evaluation of ICH hematoma volumes, the effect of this model is even better than the assessments of some clinicians. To further improve the interpretability of the model, Gradient-weighted Class Activation Mapping (Grad-CAM) is utilized to visualize the prediction of the model, which focuses well on the bleeding area and provides a reliable basis for the predictive output of the model. This can make a lot of sense for clinical application.</p></sec>
<sec id="s2">
<title>2. Related work</title>
<sec>
<title>2.1. Deep learning in brain computerized tomography images</title>
<p>The application of neuroimaging technology plays an important role in the diagnosis and treatment of cerebrovascular diseases and is an indispensable auxiliary diagnostic tool. As the main imaging methods of the brain, computerized tomography (CT) and Magnetic Resonance Imaging (MRI) are widely used in clinical practice. Among them, CT examination is the first choice to find most brain diseases, including congenital brain development intracranial abnormalities, brain tumors, cerebral hemorrhage, and so on. The major advantages of CT images after reconstruction for medical image analysis are high density and clear image; it can assist clinicians to master the brain structure and abnormalities within the brain tissues easily (Padma Nanthagopal and Sukanesh Rajamony, <xref ref-type="bibr" rid="B34">2013</xref>; Sachdeva et al., <xref ref-type="bibr" rid="B44">2013</xref>; Vidyarthi and Mittal, <xref ref-type="bibr" rid="B55">2014</xref>). Researchers utilize different automated approaches for brain disease detection and type classification through brain radio images since medical images can be scanned and uploaded to computers with a fine-resolution. For a long time in the past, Support Vector Machine (SVM) and Neural Networks (NN) techniques have been widely used due to their stable and good performance (Pan et al., <xref ref-type="bibr" rid="B35">2015</xref>). But in recent years, due to the improvement of equipment computing power, Deep Learning (DL) models have created an exciting new trend in the field of machine learning, because deep architecture can efficiently represent more complex relationships without requiring a large number of nodes like traditional machine learning, such as SVM and K-Nearest Neighbor (K-NN). With the vigorous development of these technologies, they have become advanced technologies in different medical and health fields, such as bioinformatics, medical informatics, and medical image analysis. Among all kinds of deep architectures, the Convolutional Neural Network (CNN) is undoubtedly the most commonly used architecture now, which can use the convolution kernel to realize complex operations such as feature extraction (Pan et al., <xref ref-type="bibr" rid="B35">2015</xref>; Rav&#x00300;&#x00131; et al., <xref ref-type="bibr" rid="B41">2016</xref>; Litjens et al., <xref ref-type="bibr" rid="B30">2017</xref>). Usually, CNN is designed for image recognition tasks. The image is first processed by multiple convolutional layers (each convolutional layer is followed by a Rectified Linear Unit (ReLU) layer and a pooling layer), then input to the fully connected layer and the ReLU layer. Finally, the output layer produces the prediction of class probabilities. Although the CNN architecture does not require manual feature extraction compared with traditional machine learning methods, it is very difficult to train a CNN model from scratch, which requires a large number of labeled data sets for adequate learning, especially for tasks such as classification and regression: the data volume requirements are relatively large. Moreover, for processing a large number of filters, such as large-scale medical 3D images such as 256 &#x000D7; 256 &#x000D7; 32, the hardware requirements are very high (Ben Ahmed et al., <xref ref-type="bibr" rid="B1">2017</xref>).</p>
<p>The contribution of this paper is to apply deep learning to the prediction of intracranial hemorrhage. Due to the correlation between intracranial hemorrhage and prognosis survival, this paper further proposes a lightweight multi-task learning framework based on the prediction of ICH hematoma volume. The aim is to further effectively predict the prognosis of patients according to the relevant features of the blood loss learned by the model.</p></sec>
<sec>
<title>2.2. Interpretability</title>
<p>Deep learning is a multi-layer representation learning method built with simple nonlinear modules that transform the previous layer representation into a higher-level, more abstract representation, so that it is able to detect increasingly abstract features. Therefore, a large number of features are also generated in the process of generating the prediction results, which makes the information very compact, especially in the deep layers. It&#x00027;s hard to explain this process, so we don&#x00027;t know why the neural network makes this prediction. This is called the &#x0201C;black box&#x0201D; problem of deep neural networks (Castelvecchi, <xref ref-type="bibr" rid="B7">2016</xref>). Indeed, they are capable of producing extremely accurate predictions, but how can predictions based on features beyond comprehension be reliable? It is a challenging problem.</p>
<p>In medical images, the interpretability of the model is even more important: it determines whether the clinicians can trust and accept the model predictions which could be critical in clinical applications. Interpreting the predictions of deep models is hard without appropriate techniques, but a range of approaches has emerged over the past few years. One example is to show the texture of imaging features at a single layer to examine the hierarchical process of learning (Zeiler et al., <xref ref-type="bibr" rid="B63">2011</xref>). In addition, numerous researches and explorations have been carried out on generating meaningful heatmaps in order to highlight the importance of individual pixel regions in the input image for the final CNN prediction. Many techniques have demonstrated the feasibility of generating heatmaps, including strategies using deconvolution (Zeiler and Fergus, <xref ref-type="bibr" rid="B62">2014</xref>), layer-wise relevance propagation (Samek et al., <xref ref-type="bibr" rid="B45">2016</xref>), and saliency map construction (Simonyan et al., <xref ref-type="bibr" rid="B50">2013</xref>; Ghorbani et al., <xref ref-type="bibr" rid="B16">2020</xref>; Thomas et al., <xref ref-type="bibr" rid="B52">2020</xref>). However, these methods still have certain shortcomings, such as being susceptible to noise and artifacts, lacking sensitivity to input disturbances and qualitative criteria for evaluating the quality of back propagation (Samek et al., <xref ref-type="bibr" rid="B45">2016</xref>; Smilkov et al., <xref ref-type="bibr" rid="B51">2017</xref>; Ghorbani et al., <xref ref-type="bibr" rid="B16">2020</xref>). Class Activation Mapping (CAM) serves as a better alternative approach. But in its most basic implementation, CAM requires adding a pooling layer to the target model, limiting the interpretation to only one specific layer (Zhou et al., <xref ref-type="bibr" rid="B67">2016</xref>). Recently, two extended versions of CAM have emerged, namely, Gradient-weighted Class Activation Mapping (Grad-CAM) (Selvaraju et al., <xref ref-type="bibr" rid="B46">2017</xref>), and Grad-CAM&#x0002B;&#x0002B; (Chattopadhay et al., <xref ref-type="bibr" rid="B8">2018</xref>). Compared with CAM, both of them can interpret arbitrary layers of CNN without any architectural modification, thus increasing the flexibility of the use.</p>
<p>Grad-CAM is one of the most widely used techniques for prediction interpretation; it is chosen in this paper to interpret the prediction of the model and to increase the credibility of the model so that it has good interpretability in clinical application.</p></sec></sec>
<sec id="s3">
<title>3. Method</title>
<p>To obtain the IntraCerebral Hemorrhage (ICH) volume and provide interpretative decision details, we propose a novel multi-task learning framework to take the complementary advantages of ICH regression and classification. Besides improving the accuracy of both tasks, the fusion is also able to enhance the interpretability which is of great significance for the clinical application.</p>
<sec>
<title>3.1. Overall architecture</title>
<p>As described above, the proposed network adopts the multi-task learning manner allowing the model to learn from different annotations. It can not only boost knowledge extraction by sharing the convolutional parameters but also can accelerate the learning program by calculating gradients from multiple branches. The motivation is the correlation between ICH regression and classification tasks. The research in Brott et al. (<xref ref-type="bibr" rid="B5">1997</xref>) shows that the increase of bleeding volume in the early stage of ICH is an important factor determining the prognosis of ICH patients. Unfortunately, the mortality rate of ICH is about 40% per month, with 61&#x02013;88% of survivors having degrees of residual disability. In this context, the hematoma size is a key character for prognosis predictions. Therefore, it is more accurate and acceptable to predict the prognosis of patients based on a certain accurate assessment of the bleeding amount.</p>
<p>Our multi-task CNN architecture is proposed to jointly learn from both tasks, whose details are shown in <xref ref-type="fig" rid="F1">Figure 1</xref>. The major components of the proposed network include a shared module for brain image feature extraction and two heads corresponding to different tasks. The algorithms for regression and classification tasks are built on ResNet deep neural network architecture (He et al., <xref ref-type="bibr" rid="B19">2016</xref>), which includes the residual information extracted from the previous layer and mitigated the adverse performance by using a large number of layers.</p>
<fig id="F1" position="float">
<label>Figure 1</label>
<caption><p>The overview of our multi-task framework.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnins-17-1118340-g0001.tif"/>
</fig></sec>
<sec>
<title>3.2. Efficient feature extraction</title>
<p>Experience suggests that the multi-task model often achieves better results than the single-task model. Compared to single-task learning, multi-task learning has the following advantages: shared encoder can reduce parameter amounts and save computation, associated tasks can improve each other&#x00027;s performance by sharing information and complementing each other. Also, it is helpful for the bottom-sharing encoder to learn the common feature representation.</p>
<p>The essence of multi-task learning lies in sharing the presentation layer and making the tasks interact with each other when the correlation between the predicted goals is relatively high.<xref ref-type="fn" rid="fn0001"><sup>1</sup></xref> Consequently, the parameter sharing layer will not bring too much loss, the parameter sharing layer can strengthen the parameter sharing, multiple target models can be trained jointly, reduce the parameter scale of the model and prevent the model from overfitting. Our final experimental results also fully demonstrate the effectiveness of applying multi-task learning to our tasks.</p>
<p>For multi-task learning, learning a general feature representation is very important for the model. As shown in <xref ref-type="fig" rid="F1">Figure 1</xref>, the robust feature extraction module is designed to extract features of a 3D image and obtain a high-quality feature vector, which is used for downstream tasks. The module is based on 3D ResNet34 (He et al., <xref ref-type="bibr" rid="B19">2016</xref>) because the residual structure effectively solves the problem of model degradation due to its depth, skip connection and also enhances the information transfer between the upper and lower layers.</p>
<p>Given a 3D CT image of 224 &#x000D7; 224 &#x000D7; 32 pixel size, the features are extracted by one stem block and four stages of ResNet (see <xref ref-type="fig" rid="F1">Figure 1</xref>). Specifically, the first layer is a convolutional layer with a 7 &#x000D7; 7 &#x000D7; 7 kernel size and 64 filters. The pooling layer is set after that to decrease the image&#x00027;s size and save the computation cost. The max pooling layer with a kernel size of 3 &#x000D7; 3 &#x000D7; 3 and stride 2 is used here empirically. The second, third, fourth, and fifth phases are made up of multiple convolutional operations, corresponding to the four stages in <xref ref-type="fig" rid="F1">Figure 1</xref>. The first is a convolution layer with a filter size of 3 &#x000D7; 3 &#x000D7; 3 and the rest of the three modules are stacked on the top of each layer. The second is a convolution layer with the filter size of 3 &#x000D7; 3 &#x000D7; 3 and there are four residual blocks. Each convolutional layer in the network applies zero-padding, after each convolutional layer, a batch normalization layer (Ioffe and Szegedy, <xref ref-type="bibr" rid="B23">2015</xref>) is applied to speed up model training and convergence. To allow the application to images of arbitrary size, 3D feature maps are not flattened but dense layers are implemented as convolutions with the size of 1 &#x000D7; 1 &#x000D7; 1 (Long et al., <xref ref-type="bibr" rid="B31">2015</xref>).</p>
<p>After the entire feature extraction process, the final extracted feature map is obtained. The network has two output heads, one for ICH volume regression and another one concerns the classification of the prognosis. Finally, the features extracted by the last layer of the module are fed to each task-specific head.</p></sec>
<sec>
<title>3.3. Regression head</title>
<p>The first task-specific head for regression only has an average pooling layer and one fully connected layer. The final output is a scalar referring to the amount of ICH volume predicted by the model. This regression task is performed to predict the IntraCerebral Hemorrhage (ICH) volume in each 3D CT scan, task-specific head module to find the most intensive crux features of ICH. The regression loss is defined as the Mean Square Error Loss (MSE Loss) defined by <italic>L</italic><sub><italic>reg</italic></sub>:</p>
<disp-formula id="E1"><label>(1)</label><mml:math id="M1"><mml:mrow><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mi>r</mml:mi><mml:mi>e</mml:mi><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mfrac><mml:mn>1</mml:mn><mml:mi>N</mml:mi></mml:mfrac><mml:mstyle displaystyle='true'><mml:munderover><mml:mo>&#x02211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>N</mml:mi></mml:munderover><mml:mrow><mml:msup><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:msub><mml:mover accent='true'><mml:mi>y</mml:mi><mml:mo>&#x0005E;</mml:mo></mml:mover><mml:mrow><mml:mi>r</mml:mi><mml:mi>e</mml:mi><mml:msub><mml:mi>g</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow></mml:msub><mml:mo>&#x02212;</mml:mo><mml:msub><mml:mi>y</mml:mi><mml:mrow><mml:mi>r</mml:mi><mml:mi>e</mml:mi><mml:msub><mml:mi>g</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow></mml:msub><mml:mo stretchy='false'>)</mml:mo></mml:mrow><mml:mn>2</mml:mn></mml:msup></mml:mrow></mml:mstyle><mml:mo>,</mml:mo></mml:mrow></mml:math></disp-formula>
<p>Where <inline-formula><mml:math id="M6"><mml:mrow><mml:mover accent='true'><mml:mi>y</mml:mi><mml:mo>&#x0005E;</mml:mo></mml:mover></mml:mrow></mml:math></inline-formula><sub><italic>re</italic><sub><italic>g</italic></sub><sub><italic>i</italic></sub></sub> is the predicted output of the regression head, and <italic>N</italic> is the number of CT images.</p></sec>
<sec>
<title>3.4. Classification head</title>
<p>The second task-specific head for classification has a simple linear classifier which consists of a global average pooling, a linear layer, and a ReLU activation function. Finally, the softmax function is applied to the output to obtain the probability of the model prediction, which is between 0 and 1. Generally, we use 0.5 as the classification decision threshold: a probability greater than 0.5 is fixed to 1, which means that the class belongs to, whereas a probability less than 0.5 is fixed to 0, meaning that the class does not belong to. The upstream framework is the same and shares the weight of the feature extraction module of the regression task. The classification loss is defined as the Binary Cross-Entropy Loss (BCE Loss) called <italic>L</italic><sub><italic>cls</italic></sub>:</p>
<disp-formula id="E2"><label>(2)</label><mml:math id="M2"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>s</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:mfrac><mml:mo>&#x000B7;</mml:mo><mml:mstyle displaystyle="true"><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:munderover></mml:mstyle><mml:mrow><mml:mo stretchy="false">[</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>s</mml:mi></mml:mrow></mml:msub><mml:mo>&#x000B7;</mml:mo><mml:mo class="qopname">log</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mover accent='true'><mml:mi>y</mml:mi><mml:mo>&#x0005E;</mml:mo></mml:mover><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:msub><mml:mrow><mml:mi>s</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x0002B;</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mn>1</mml:mn><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:msub><mml:mrow><mml:mi>s</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x000B7;</mml:mo><mml:mo class="qopname">log</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mn>1</mml:mn><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mover accent='true'><mml:mi>y</mml:mi><mml:mo>&#x0005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:msub><mml:mrow><mml:mi>s</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">]</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>Where <inline-formula><mml:math id="M7"><mml:mrow><mml:mover accent='true'><mml:mi>y</mml:mi><mml:mo>&#x0005E;</mml:mo></mml:mover></mml:mrow></mml:math></inline-formula><sub><italic>cl</italic><sub><italic>s</italic></sub><sub><italic>i</italic></sub></sub> is the predicted output of the classification head, and <italic>N</italic> represents the number of CT images.</p>
<p>For our multi-task framework, the feature extractor is shared and the two heads fit the outputs of the two tasks separately. Therefore, the loss of multi-task framework is a combination of the losses of two tasks. Through the simple weighting of loss, the network weights are updated to optimize.</p></sec>
<sec>
<title>3.5. Interpretive design</title>
<p>Traditional deep learning applications lack interpretability and thus faces limitation in clinical practice. We propose to utilize the advanced Grad-CAM to explore how the results are obtained. Specifically, class weights derived in Grad-CAM used the equation:</p>
<disp-formula id="E3"><label>(3)</label><mml:math id="M3"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msubsup><mml:mrow><mml:mi>w</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow><mml:mrow><mml:mi>c</mml:mi></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>Z</mml:mi></mml:mrow></mml:mfrac><mml:mstyle displaystyle="true"><mml:munder class="msub"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:munder></mml:mstyle><mml:mstyle displaystyle="true"><mml:munder class="msub"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:munder></mml:mstyle><mml:mfrac><mml:mrow><mml:mi>&#x02202;</mml:mi><mml:msup><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mrow><mml:mi>c</mml:mi></mml:mrow></mml:msup></mml:mrow><mml:mrow><mml:mi>&#x02202;</mml:mi><mml:msubsup><mml:mrow><mml:mi>A</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>j</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:mfrac></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>Where <italic>Z</italic> denotes the total number of elements in a feature map <italic>A</italic>, <inline-formula><mml:math id="M4"><mml:msubsup><mml:mrow><mml:mi>A</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>j</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> represents the data of the feature map in channel <italic>k</italic>, and the coordinate is <italic>ij</italic>, <italic>y</italic><sup><italic>c</italic></sup> denotes the score predicted by the network for class <italic>c</italic>.</p>
<p>This equation is backpropagated through the prediction score of class <italic>c</italic> to obtain the gradient information that is backpropagated back to feature layer <italic>A</italic>, namely <inline-formula><mml:math id="M5"><mml:msubsup><mml:mrow><mml:mi>w</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow><mml:mrow><mml:mi>c</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula>, which represents the importance of each channel of feature map <italic>A</italic>, here for class <italic>c</italic>. The higher the value, the greater the contribution, and the more important it is considered by the model. Finally, weighted summation is performed and the final Grad-CAM heatmap is obtained by ReLU activation function. Consequently, the heatmap generation uses the formula:</p>
<disp-formula id="E4"><label>(4)</label><mml:math id="M9"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msubsup><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi><mml:mi>r</mml:mi><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mo>-</mml:mo><mml:mi>C</mml:mi><mml:mi>A</mml:mi><mml:mi>M</mml:mi></mml:mrow><mml:mrow><mml:mi>c</mml:mi></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:mi>R</mml:mi><mml:mi>e</mml:mi><mml:mi>L</mml:mi><mml:mi>U</mml:mi><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:mstyle displaystyle="true"><mml:munder class="msub"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:munder></mml:mstyle><mml:msubsup><mml:mrow><mml:mi>w</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow><mml:mrow><mml:mi>c</mml:mi></mml:mrow></mml:msubsup><mml:msup><mml:mrow><mml:mi>A</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msup></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>Where the ReLU (Rectified Linear Unit) function allows to evaluate features with positive impact only. In this paper, we use the last convolutional layer in computing the weights as suggested previously.</p></sec></sec>
<sec id="s4">
<title>4. Experiments</title>
<sec>
<title>4.1. Ethics approval and consent to participate</title>
<p>The study is undertaken in compliance with the principles of the Declaration of Helsinki and is approved by the ethics committees of the First Affiliated Hospital of Xiamen University (Approved ID: SL-2020KY034). All individual patient data used for the analysis are collected by providers after obtaining appropriate consent and agreements.</p></sec>
<sec>
<title>4.2. Data collection</title>
<p>This study involves 258 patients with spontaneous intracerebral hematoma (ICH) admitted to the First Affiliated Hospital of Xiamen University between April 2017 and February 2019 to develop the prognosis model. These axial brain Computed Tomography (CT) scans, taken on admission using a Philips Brilliance 64-row spiral CT scanner are exported from the Department of Neurosurgery through Picture Archiving and Communication Systems (PACS) and stored in Digital Imaging and Communications in Medicine (DICOM) format.</p></sec>
<sec>
<title>4.3. Data preprocessing</title>
<p>To evaluate the effectiveness of the developed network, we collect a new ICH dataset with 258 patients, of which 227 patients survived and 31 patients dead. All patients are given a table with the patient&#x00027;s name and measurements of the ICH volume (mililiter, ml) from eight professional neurosurgeons (with 2, 3, 3, 4, 4, 8, 15, and 20 years of experience) and the 3D slicer software, which is considered the gold standard.</p>
<p>For the regression labels, because of the uneven distribution of ICH volume (shown in <xref ref-type="fig" rid="F2">Figure 2</xref> left), we calculate the third root of the ICH volume and convert it to the ICH size (milimeter, mm), which is more balanced in distribution (shown in <xref ref-type="fig" rid="F2">Figure 2</xref> right) and more beneficial to model training. For the final prediction results of the regression model, we perform post-processing and then convert them into ICH volumes for further analysis and comparisons.</p>
<fig id="F2" position="float">
<label>Figure 2</label>
<caption><p>Histogram of the blood loss distribution.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnins-17-1118340-g0002.tif"/>
</fig>
<p>Considering the intensity range of brain characteristics (Fosbinder and Orth, <xref ref-type="bibr" rid="B14">2011</xref>), we apply a brain window that clips [40, 110] Hounsfield units (HU) and normalized the input to [0, 1]. After cutting out the blank part of the brain in the image, the image size is resized down to 32 &#x000D7; 224 &#x000D7; 224 by linear interpolation. Since ICH can occur anywhere in the brain with multiple subtypes simultaneously, we do not crop the images and don&#x00027;t use patch images for training, which may lead to unstable results and false positives (Hu et al., <xref ref-type="bibr" rid="B20">2011</xref>).</p></sec>
<sec>
<title>4.4. Evaluation metrics</title>
<p>For the regression task, we choose the following three indicators to evaluate, which include Mean Square Error (MSE), Mean Absolute Error (MAE), and Mean Absolute Percentage Error (MAPE). And MAPE is the relative error, which is the percentage of absolute error and truth. For classification task, three widely-used metrics are used for quantitative analysis, which are accuracy, ROC curve, and area under the curve(AUC).</p></sec>
<sec>
<title>4.5. Implementation details</title>
<p>Our network is implemented on Pytorch (Paszke et al., <xref ref-type="bibr" rid="B36">2019</xref>) and trains using SGD (Robbins and Monro, <xref ref-type="bibr" rid="B42">1951</xref>) with 300 epochs, an initial learning rate of 1 &#x000D7; 10<sup>&#x02212;4</sup>, a momentum of 0.9, and a weight decay of 5 &#x000D7; 10<sup>&#x02212;4</sup>. The whole architecture is trained on one GeForce RTX 3080 Ti GPU, and each GPU has a batch size of 8. For a fair comparison, our method and all models follow the same training settings.</p></sec>
<sec>
<title>4.6. Results</title>
<p>The performance of our proposed approach for regression and classification tasks is reported in <xref ref-type="table" rid="T1">Tables 1</xref>, <xref ref-type="table" rid="T2">2</xref>, respectively.</p>
<table-wrap position="float" id="T1">
<label>Table 1</label>
<caption><p>Quantitative comparisons for the effectiveness of regression model.</p></caption>
<table frame="box" rules="all">
<thead><tr style="background-color:#919497;color:#ffffff">
<th valign="top" align="left"><bold>Methods</bold></th>
<th valign="top" align="center"><bold>MAE</bold></th>
<th valign="top" align="center"><bold>MSE</bold></th>
<th valign="top" align="center"><bold>MAPE(%)</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Clinician 1</td>
<td valign="top" align="center">5.21</td>
<td valign="top" align="center">65.93</td>
<td valign="top" align="center">28.78</td>
</tr> <tr>
<td valign="top" align="left">Clinician 2</td>
<td valign="top" align="center">6.14</td>
<td valign="top" align="center">84.10</td>
<td valign="top" align="center">38.50</td>
</tr> <tr>
<td valign="top" align="left">clinician 3</td>
<td valign="top" align="center">4.46</td>
<td valign="top" align="center">39.15</td>
<td valign="top" align="center">38.68</td>
</tr> <tr>
<td valign="top" align="left">Clinician 4</td>
<td valign="top" align="center">4.77</td>
<td valign="top" align="center">50.05</td>
<td valign="top" align="center">34.13</td>
</tr> <tr>
<td valign="top" align="left">Clinician 5</td>
<td valign="top" align="center">4.88</td>
<td valign="top" align="center">72.04</td>
<td valign="top" align="center">29.60</td>
</tr> <tr>
<td valign="top" align="left">Clinician 6</td>
<td valign="top" align="center">5.30</td>
<td valign="top" align="center">56.08</td>
<td valign="top" align="center">44.65</td>
</tr> <tr>
<td valign="top" align="left">Clinician 7</td>
<td valign="top" align="center">3.86</td>
<td valign="top" align="center">38.84</td>
<td valign="top" align="center">20.14</td>
</tr> <tr>
<td valign="top" align="left">Clinician 8</td>
<td valign="top" align="center">4.02</td>
<td valign="top" align="center">40.98</td>
<td valign="top" align="center">24.42</td>
</tr> <tr>
<td valign="top" align="left">X3D (Feichtenhofer, <xref ref-type="bibr" rid="B12">2020</xref>)</td>
<td valign="top" align="center">18.28</td>
<td valign="top" align="center">490.19</td>
<td valign="top" align="center">99.98</td>
</tr> <tr>
<td valign="top" align="left">R(2&#x0002B;1)D (Tran et al., <xref ref-type="bibr" rid="B53">2018</xref>)</td>
<td valign="top" align="center">7.09</td>
<td valign="top" align="center">80.02</td>
<td valign="top" align="center">45.59</td>
</tr> <tr>
<td valign="top" align="left">P3D (Qiu et al., <xref ref-type="bibr" rid="B39">2017</xref>)</td>
<td valign="top" align="center">10.24</td>
<td valign="top" align="center">143.54</td>
<td valign="top" align="center">171.18</td>
</tr> <tr>
<td valign="top" align="left">TSM (Lin et al., <xref ref-type="bibr" rid="B29">2019</xref>)</td>
<td valign="top" align="center">9.51</td>
<td valign="top" align="center">130.69</td>
<td valign="top" align="center">160.93</td>
</tr> <tr>
<td valign="top" align="left">Times Former (Bertasius et al., <xref ref-type="bibr" rid="B2">2021</xref>)</td>
<td valign="top" align="center">10.49</td>
<td valign="top" align="center">158.13</td>
<td valign="top" align="center">171.09</td>
</tr> <tr>
<td valign="top" align="left">TIN-ResNet-18 (Shao et al., <xref ref-type="bibr" rid="B47">2020</xref>)</td>
<td valign="top" align="center">7.90</td>
<td valign="top" align="center">107.52</td>
<td valign="top" align="center">110.68</td>
</tr> <tr>
<td valign="top" align="left">TIN-ResNet-34 (Shao et al., <xref ref-type="bibr" rid="B47">2020</xref>)</td>
<td valign="top" align="center">9.71</td>
<td valign="top" align="center">134.73</td>
<td valign="top" align="center">159.54</td>
</tr> <tr>
<td valign="top" align="left">TIN-ResNet-50 (Shao et al., <xref ref-type="bibr" rid="B47">2020</xref>)</td>
<td valign="top" align="center">10.87</td>
<td valign="top" align="center">170.85</td>
<td valign="top" align="center">184.58</td>
</tr> <tr>
<td valign="top" align="left">Uniformizing-3D (Zunair et al., <xref ref-type="bibr" rid="B69">2020</xref>)</td>
<td valign="top" align="center">7.87</td>
<td valign="top" align="center">118.06</td>
<td valign="top" align="center">102.38</td>
</tr> <tr>
<td valign="top" align="left">Multi-task (ours)</td>
<td valign="top" align="center">4.91</td>
<td valign="top" align="center">52.65</td>
<td valign="top" align="center">43.00</td>
</tr></tbody>
</table>
<table-wrap-foot>
<p>The first eight lines denote the measurements of eight clinicians.</p>
</table-wrap-foot>
</table-wrap>
<table-wrap position="float" id="T2">
<label>Table 2</label>
<caption><p>Quantitative comparisions of the effectiveness of multi-task framework.</p></caption>
<table frame="box" rules="all">
<thead><tr style="background-color:#919497;color:#ffffff">
<th valign="top" align="left"><bold>Methods</bold></th>
<th valign="top" align="center"><bold>ACC (Accuracy)</bold></th>
<th valign="top" align="center"><bold>AUC</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">ResNet (single-task)</td>
<td valign="top" align="center">88.68</td>
<td valign="top" align="center">87.94</td>
</tr> <tr>
<td valign="top" align="left">ResNet (multi-task, ours)</td>
<td valign="top" align="center"><bold>94.34</bold></td>
<td valign="top" align="center"><bold>93.62</bold></td>
</tr></tbody>
</table>
<table-wrap-foot>
<p>&#x0201C;Single-task&#x0201D; denotes the classification model train from scratch.</p>
<p>Bold values indicate the optimal result.</p>
</table-wrap-foot>
</table-wrap>
<p>To verify the effectiveness and feasibility of our proposed framework, <xref ref-type="table" rid="T1">Table 1</xref> selects five existing methods widely used for 3D image classification tasks for comparison, including X3D (Feichtenhofer, <xref ref-type="bibr" rid="B12">2020</xref>), R(2&#x0002B;1)D (Tran et al., <xref ref-type="bibr" rid="B53">2018</xref>), P3D (Qiu et al., <xref ref-type="bibr" rid="B39">2017</xref>), TSM (Lin et al., <xref ref-type="bibr" rid="B29">2019</xref>), Times Former (Bertasius et al., <xref ref-type="bibr" rid="B2">2021</xref>), TIN (Shao et al., <xref ref-type="bibr" rid="B47">2020</xref>) with backbone of ResNet-18, 34, 50 and Uniformizing-3D (Zunair et al., <xref ref-type="bibr" rid="B69">2020</xref>), which is a 3D CNN for CT scans . For providing a fair comparison, we obtain the final classification results on the official implementations of these compared methods. We train these models on our dataset and set the same experimental parameters as ours. As shown in <xref ref-type="table" rid="T1">Table 1</xref>, the model trained by deep learning performs well. The method we choose based on ResNet architecture achieves the best results and is even more suitable for the tasks of regression and classification of 3D brain CT images, with MAE reaching 4.91, the MSE reaching 51.92, and MAPE reaching 43.00%. Compared to the evaluation results of eight clinicians (first eight rows in <xref ref-type="table" rid="T1">Table 1</xref>), our method also has competitive performance, outperforming the evaluation results of three clinicians (clinicians 1, 2, 6), while also having better interpretability.</p>
<p>To further show the stability and effectiveness of our model training, we plot the curve of the indicator change during the model training and validation process, including MAE and MSE, as shown in <xref ref-type="fig" rid="F3">Figure 3</xref>. From the figure, we can see that whether in the training or validation phase, the MAE and MSE loss functions of our model gradually decrease and converge to be stable in 300 epochs, which indicates that the model obtained by our training is stable and reliable. The training of other models also shows a similar trend.</p>
<fig id="F3" position="float">
<label>Figure 3</label>
<caption><p>The training curve of the regression loss in our multi-task framework.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnins-17-1118340-g0003.tif"/>
</fig>
<p>The results of the classification task shown in <xref ref-type="table" rid="T2">Table 2</xref> demonstrate the effectiveness of the multi-task network. Compared with the single-task model, that is, trained from scratch, the performance of the multi-task model is also greatly improved due to the shared encoder parameters, with accuracy improved by 5.66% and AUC improved by 5.68%, and both reached more than 90%.</p>
<p>More visually, <xref ref-type="fig" rid="F4">Figure 4</xref> shows the ROC curves of our multi-task model and single-task model. The performance of the multi-task model is better than others, and the area of AUC remains over 90%, which fully demonstrates its effectiveness.</p>
<fig id="F4" position="float">
<label>Figure 4</label>
<caption><p>ROC curves of our multi-task framework and compared method.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnins-17-1118340-g0004.tif"/>
</fig></sec>
<sec>
<title>4.7. Interpretability</title>
<p>To enhance the interpretability of the model for the prediction of ICH volume, we visualize the output of the last convolutional layer of the regression model. As mentioned earlier, we introduce Grad-CAM (Gradient-weighted Class Activation Mapping) (Selvaraju et al., <xref ref-type="bibr" rid="B46">2017</xref>) to generate heatmaps, which show the region of interest of the model.</p>
<p>In the heatmap, the brighter the color is, the more attention the model pays to. As shown in <xref ref-type="fig" rid="F5">Figure 5</xref>, the highlighted areas are concentrated in certain parts of the brain with some white spots. And according to our confirmation with a professional clinician, this is exactly the location of the cerebral hemorrhage in the image. Therefore, it is reasonable to say that in the prediction of ICH volume, our model accurately locates the location of bleeding, to make the prediction, and the result is acceptable to clinicians. At the same time, it also provides a basis for subsequent prognostic analysis tasks, as the ICH volume is a key factor affecting patient mortality.</p>
<fig id="F5" position="float">
<label>Figure 5</label>
<caption><p>Interpretable visualization of our multi-task framework. The images are five slices taken from a 3D CT scans at medium intervals (<italic>t</italic> = 5).</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnins-17-1118340-g0005.tif"/>
</fig></sec></sec>
<sec sec-type="discussion" id="s5">
<title>5. Discussion</title>
<sec>
<title>5.1. Clinical significance for proposed algorithm</title>
<p>Stroke is the second-leading cause of death, and the third-leading cause of death and disability combined in the world (Feigin et al., <xref ref-type="bibr" rid="B13">2021</xref>). Intracerebral hematoma (ICH) is a common and fatal subtype of stroke, which is characterized by rupture of arterial blood vessels and formation of hematoma within the brain parenchyma, such as the basal ganglia (Zhang et al., <xref ref-type="bibr" rid="B64">2020</xref>; Kinoshita et al., <xref ref-type="bibr" rid="B27">2021</xref>). The formation of a large hematoma compresses the surrounding brain tissue, resulting in neurological deterioration, intracranial hypertension, global cerebral ischemia, and brain herniation, which can affect vital autonomic structures (Yu et al., <xref ref-type="bibr" rid="B61">2020</xref>). The mortality rate of ICH is about 40% per month, with 61&#x02013;88% of survivors having high degrees of residual disability (Van Asch et al., <xref ref-type="bibr" rid="B54">2010</xref>). Management of increased intracranial pressure is the mainstay of ICH treatment in the acute phase (Kase and Hanley, <xref ref-type="bibr" rid="B24">2021</xref>). Patients with excessive bleeding and coma often require surgical treatment to relieve the hematoma, brain tissue compression, and secondary brain injury (Lin et al., <xref ref-type="bibr" rid="B28">2014</xref>). However, as an invasive process, iatrogenic injury is inevitable during hematoma evacuation, which means the accurate assessments of the patients&#x00027; condition and appropriate selections for surgical interventions are critically important. The hematoma size is a key character for prognosis predictions and clinical decision making. Supratentorial hematomas larger than 30 ml (Kerebel et al., <xref ref-type="bibr" rid="B25">2013</xref>; Huang et al., <xref ref-type="bibr" rid="B21">2021</xref>) and infratentorial hematomas larger than 15 ml (de Oliveira Manoel, <xref ref-type="bibr" rid="B11">2020</xref>) are often considered to result in poor prognosis and tend to require surgical evacuation. Thus, the accuracy of volume evaluation for ICH patients has significant clinical implications.</p>
<p>There are mainly four kinds of methods to evaluate ICH volume, including the mathematical formula method, tool measurement method, CT machine measurement method, and software method (Chen et al., <xref ref-type="bibr" rid="B9">2020</xref>). Among them, Tada formula, a kind of formula method, is the most commonly used one. However, the accuracy of Tada formula can be affected by both hematoma shape and volume, resulting in imprecise disease evaluation and potential improper treatments (Gong et al., <xref ref-type="bibr" rid="B17">2021</xref>). 3D Slicer is shown to be more precise than most methods (Chen et al., <xref ref-type="bibr" rid="B9">2020</xref>), but the modeling process is relatively time-consuming for clinicians. With the development of deep learning techniques, artificial intelligence has provided promising results in the medical field and may help with medical image interpretation and prognosis prediction, offering more reliable results than manual decisions by clinicians. Several previous studies have used deep learning-based techniques to evaluate ICH volumes (Roh et al., <xref ref-type="bibr" rid="B43">2019</xref>), but there are still some shortcomings. For instance, some studies use the manually estimated hematoma volume as the label to train the model, which is not accurate enough to evaluate the performance of the model because the relatively large error of manual estimation leads to the inaccuracy of the label. In addition, it does not take the problem of model interpretation into account.</p>
<p>For this reason, this study uses the results of 3D slicer as labels to improve the accuracy of model training and evaluation. At the same time, in addition to achieving competitive accuracy with clinicians, our model has good interpretability and is easier to be understood and accepted by clinicians. Intuitively, we attribute the accuracy of model prediction to the fact that the heatmap shows that it focuses on the information related to the bleeding area, which is in turn the target of the model being explained. This demonstrates the accuracy of the effect of our model, and its concerns are consistent with those of professional physicians. Furthermore, the evaluation of hematoma volume to predict patient prognosis is not accurate enough because the ICH density, shape, location, the middle line deviation, the effacement of Sylvian fissure and perimesencephalic cisterns (Kim, <xref ref-type="bibr" rid="B26">2012</xref>) are all related to the prognosis of ICH. Therefore, this study uses the complete CT images as the training objects and does not mark a specific region of interest, so that the results are more reasonable and robust.</p></sec>
<sec>
<title>5.2. Limitations and further directions</title>
<p>Our experimental results prove the effectiveness of our proposed framework well, but to better apply it clinically, we should test it on more data. This is also the current limitation of our method. First, our data are collected from a single center and the result should be further validated by data from multiple centers in terms of predicting prognosis due to differences in treatment levels. Second, given the variable performance of ICH, the sample size is still needed to be expanded, including hemotomas of different sizes, locations, shapes, and densities. Third, cerebral hemorrhage is a dynamic process, and a single NCCT can only be used to evaluate static intracranial conditions. Therefore, subsequent studies can use different concurrent images as time series for prognostic evaluation. Fourth, besides the radiological features, the clinical features including age, the severity of other underlying diseases, the GCS coma score, etc. are all associated with prognosis (Wang et al., <xref ref-type="bibr" rid="B57">2014</xref>). Using multimodal data may further advance the performance of such clinical prediction models.</p></sec></sec>
<sec sec-type="conclusions" id="s6">
<title>6. Conclusion</title>
<p>We introduce a multi-task framework for ICH volume (regression) and patient prognosis (classification) prediction in NCCT. First, we design a robust feature extractor through the ICH volume prediction task, which effectively locates and focuses on ICH regions. This is effectively demonstrated in the Grad-CAM. By exploring the relationship between ICH volume and patient prognosis, we find that the effectiveness of feature extraction on the former task affected the performance on the latter task. A feature extractor that can well locate the hemorrhage area and pay attention to the hematoma, which is very beneficial for the subsequent task of prognostic analysis. The experimental results also prove it well. Compared with existing methods, our framework achieves impressive results on both tasks, and its reliable interpretability also makes it possible for clinical application.</p></sec>
<sec sec-type="data-availability" id="s7">
<title>Data availability statement</title>
<p>The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.</p></sec>
<sec sec-type="author-contributions" id="s8">
<title>Author contributions</title>
<p>KG, ZW, and SH contributed to the study concept and design. KG and YZ collected the data. KG, QD, and JW performed the statistical analysis and wrote the manuscript. YZ, TS, JY, and JC revised the manuscript. All authors contributed to the article and agreed to the submitted version.</p></sec>
</body>
<back>
<sec sec-type="COI-statement" id="conf1">
<title>Conflict of interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="disclaimer" id="s9">
<title>Publisher&#x00027;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<fn-group>
<fn id="fn0001"><p><sup>1</sup>In our task, ICH volume regression and patient prognosis classification based on brain CT images. The underlying features of both should be similar, both at the bleeding site in the brain.</p></fn>
</fn-group>
<ref-list>
<title>References</title>
<ref id="B1">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ben Ahmed</surname> <given-names>K.</given-names></name> <name><surname>Hall</surname> <given-names>L.</given-names></name> <name><surname>Goldgof</surname> <given-names>D.</given-names></name> <name><surname>Liu</surname> <given-names>R.</given-names></name> <name><surname>Gatenby</surname> <given-names>R.</given-names></name></person-group> (<year>2017</year>). <article-title>Fine-tuning convolutional deep features for MRI based brain tumor classification</article-title>. <source>Med. Imaging Comput. Aided Diagn</source>. <volume>10134</volume>, <fpage>613</fpage>&#x02013;<lpage>619</lpage>. <pub-id pub-id-type="doi">10.1117/12.2253982</pub-id><pub-id pub-id-type="pmid">31150950</pub-id></citation></ref>
<ref id="B2">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Bertasius</surname> <given-names>G.</given-names></name> <name><surname>Wang</surname> <given-names>H.</given-names></name> <name><surname>Torresani</surname> <given-names>L.</given-names></name></person-group> (<year>2021</year>). <article-title>&#x0201C;Is space-time attention all you need for video understanding?&#x0201D;</article-title> in <source>International Conference on Machine Learning</source> (<publisher-loc>Honolulu</publisher-loc>), <volume>2</volume>, <fpage>4</fpage>.</citation>
</ref>
<ref id="B3">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Beslow</surname> <given-names>L. A.</given-names></name> <name><surname>Ichord</surname> <given-names>R. N.</given-names></name> <name><surname>Gindville</surname> <given-names>M. C.</given-names></name> <name><surname>Kleinman</surname> <given-names>J. T.</given-names></name> <name><surname>Bastian</surname> <given-names>R. A.</given-names></name> <name><surname>Smith</surname> <given-names>S. E.</given-names></name> <etal/></person-group>. (<year>2014</year>). <article-title>Frequency of hematoma expansion after spontaneous intracerebral hemorrhage in children</article-title>. <source>JAMA Neurol</source>. <volume>71</volume>, <fpage>165</fpage>&#x02013;<lpage>171</lpage>. <pub-id pub-id-type="doi">10.1001/jamaneurol.2013.4672</pub-id><pub-id pub-id-type="pmid">24296993</pub-id></citation></ref>
<ref id="B4">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Beslow</surname> <given-names>L. A.</given-names></name> <name><surname>Ichord</surname> <given-names>R. N.</given-names></name> <name><surname>Kasner</surname> <given-names>S. E.</given-names></name> <name><surname>Mullen</surname> <given-names>M. T.</given-names></name> <name><surname>Licht</surname> <given-names>D. J.</given-names></name> <name><surname>Smith</surname> <given-names>S. E.</given-names></name> <etal/></person-group>. (<year>2010</year>). <article-title>ABC/XYZ estimates intracerebral hemorrhage volume as a percent of total brain volume in children</article-title>. <source>Stroke</source> <volume>41</volume>, <fpage>691</fpage>&#x02013;<lpage>694</lpage>. <pub-id pub-id-type="doi">10.1161/STROKEAHA.109.566430</pub-id><pub-id pub-id-type="pmid">20181678</pub-id></citation></ref>
<ref id="B5">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Brott</surname> <given-names>T.</given-names></name> <name><surname>Broderick</surname> <given-names>J.</given-names></name> <name><surname>Kothari</surname> <given-names>R.</given-names></name> <name><surname>Barsan</surname> <given-names>W.</given-names></name> <name><surname>Tomsick</surname> <given-names>T.</given-names></name> <name><surname>Sauerbeck</surname> <given-names>L.</given-names></name> <etal/></person-group>. (<year>1997</year>). <article-title>Early hemorrhage growth in patients with intracerebral hemorrhage</article-title>. <source>Stroke</source> <volume>28</volume>, <fpage>1</fpage>&#x02013;<lpage>5</lpage>. <pub-id pub-id-type="doi">10.1161/01.STR.28.1.1</pub-id><pub-id pub-id-type="pmid">8996478</pub-id></citation></ref>
<ref id="B6">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Cai</surname> <given-names>S.</given-names></name> <name><surname>Tian</surname> <given-names>Y.</given-names></name> <name><surname>Lui</surname> <given-names>H.</given-names></name> <name><surname>Zeng</surname> <given-names>H.</given-names></name> <name><surname>Wu</surname> <given-names>Y.</given-names></name> <name><surname>Chen</surname> <given-names>G.</given-names></name></person-group> (<year>2020</year>). <article-title>Dense-UNet: a novel multiphoton in vivo cellular image segmentation model based on a convolutional neural network</article-title>. <source>Quant. Imaging Med. Surg</source>. <volume>10</volume>, <fpage>1275</fpage>. <pub-id pub-id-type="doi">10.21037/qims-19-1090</pub-id><pub-id pub-id-type="pmid">32550136</pub-id></citation></ref>
<ref id="B7">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Castelvecchi</surname> <given-names>D.</given-names></name></person-group> (<year>2016</year>). <article-title>Can we open the black box of AI?</article-title> <source>Nat. News</source> <volume>538</volume>, <fpage>20</fpage>. <pub-id pub-id-type="doi">10.1038/538020a</pub-id><pub-id pub-id-type="pmid">27708329</pub-id></citation></ref>
<ref id="B8">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Chattopadhay</surname> <given-names>A.</given-names></name> <name><surname>Sarkar</surname> <given-names>A.</given-names></name> <name><surname>Howlader</surname> <given-names>P.</given-names></name> <name><surname>Balasubramanian</surname> <given-names>V. N.</given-names></name></person-group> (<year>2018</year>). <article-title>&#x0201C;Grad-cam&#x0002B;&#x0002B;: Generalized gradient-based visual explanations for deep convolutional networks,&#x0201D;</article-title> in <source>IEEE Winter Conference on Applications of Computer Vision</source> (<publisher-loc>Lake Tahoe, NV</publisher-loc>: <publisher-name>IEEE</publisher-name>), <fpage>839</fpage>&#x02013;<lpage>847</lpage>.<pub-id pub-id-type="pmid">36050386</pub-id></citation></ref>
<ref id="B9">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chen</surname> <given-names>M.</given-names></name> <name><surname>Li</surname> <given-names>Z.</given-names></name> <name><surname>Ding</surname> <given-names>J.</given-names></name> <name><surname>Lu</surname> <given-names>X.</given-names></name> <name><surname>Cheng</surname> <given-names>Y.</given-names></name> <name><surname>Lin</surname> <given-names>J.</given-names></name></person-group> (<year>2020</year>). <article-title>Comparison of common methods for precision volume measurement of hematoma</article-title>. <source>Comput. Math. Methods Med</source>. <volume>2020</volume>, <fpage>1</fpage>&#x02013;<lpage>11</lpage>. <pub-id pub-id-type="doi">10.1155/2020/6930836</pub-id><pub-id pub-id-type="pmid">32724331</pub-id></citation></ref>
<ref id="B10">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Dawud</surname> <given-names>A. M.</given-names></name> <name><surname>Yurtkan</surname> <given-names>K.</given-names></name> <name><surname>Oztoprak</surname> <given-names>H.</given-names></name></person-group> (<year>2019</year>). <article-title>Application of deep learning in neuroradiology: brain haemorrhage classification using transfer learning</article-title>. <source>Comput. Intell. Neurosci</source>. <volume>2019</volume>, <fpage>1</fpage>&#x02013;<lpage>12</lpage>. <pub-id pub-id-type="doi">10.1155/2019/4629859</pub-id><pub-id pub-id-type="pmid">32908475</pub-id></citation></ref>
<ref id="B11">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>de Oliveira Manoel</surname> <given-names>A. L.</given-names></name></person-group> (<year>2020</year>). <article-title>Surgery for spontaneous intracerebral hemorrhage</article-title>. <source>Crit. Care</source> <volume>24</volume>, <fpage>1</fpage>&#x02013;<lpage>19</lpage>. <pub-id pub-id-type="doi">10.1186/s13054-020-2749-2</pub-id><pub-id pub-id-type="pmid">32033578</pub-id></citation></ref>
<ref id="B12">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Feichtenhofer</surname> <given-names>C.</given-names></name></person-group> (<year>2020</year>). <article-title>&#x0201C;X3D: expanding architectures for efficient video recognition,&#x0201D;</article-title> in <source>Proceedings of IEEE Conference on Computer Vision and Pattern Recognition</source> (<publisher-loc>Seattle, WA</publisher-loc>: <publisher-name>IEEE</publisher-name>), <fpage>203</fpage>&#x02013;<lpage>213</lpage>.</citation>
</ref>
<ref id="B13">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Feigin</surname> <given-names>V. L.</given-names></name> <name><surname>Stark</surname> <given-names>B. A.</given-names></name> <name><surname>Johnson</surname> <given-names>C. O.</given-names></name> <name><surname>Roth</surname> <given-names>G. A.</given-names></name> <name><surname>Bisignano</surname> <given-names>C.</given-names></name> <name><surname>Abady</surname> <given-names>G. G.</given-names></name> <etal/></person-group>. (<year>2021</year>). <article-title>Global, regional, and national burden of stroke and its risk factors, 1990-2019: a systematic analysis for the Global Burden of Disease Study 2019</article-title>. <source>Lancet Neurol</source>. <volume>20</volume>, <fpage>795</fpage>&#x02013;<lpage>820</lpage>. <pub-id pub-id-type="doi">10.1016/S1474-4422(21)00252-0</pub-id><pub-id pub-id-type="pmid">34487721</pub-id></citation></ref>
<ref id="B14">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Fosbinder</surname> <given-names>R.</given-names></name> <name><surname>Orth</surname> <given-names>D.</given-names></name></person-group> (<year>2011</year>). <source>Essentials of Radiologic Science</source>. <publisher-loc>Philadelphia, PA</publisher-loc>: <publisher-name>Lippincott Williams &#x00026; Wilkins</publisher-name>.</citation>
</ref>
<ref id="B15">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Freeman</surname> <given-names>W. D.</given-names></name> <name><surname>Barrett</surname> <given-names>K. M.</given-names></name> <name><surname>Bestic</surname> <given-names>J. M.</given-names></name> <name><surname>Meschia</surname> <given-names>J. F.</given-names></name> <name><surname>Broderick</surname> <given-names>D. F.</given-names></name> <name><surname>Brott</surname> <given-names>T. G.</given-names></name></person-group> (<year>2008</year>). <article-title>Computer-assisted volumetric analysis compared with ABC/2 method for assessing warfarin-related intracranial hemorrhage volumes</article-title>. <source>Neurocrit. Care</source> <volume>9</volume>, <fpage>307</fpage>&#x02013;<lpage>312</lpage>. <pub-id pub-id-type="doi">10.1007/s12028-008-9089-4</pub-id><pub-id pub-id-type="pmid">18392784</pub-id></citation></ref>
<ref id="B16">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ghorbani</surname> <given-names>A.</given-names></name> <name><surname>Ouyang</surname> <given-names>D.</given-names></name> <name><surname>Abid</surname> <given-names>A.</given-names></name> <name><surname>He</surname> <given-names>B.</given-names></name> <name><surname>Chen</surname> <given-names>J. H.</given-names></name> <name><surname>Harrington</surname> <given-names>R. A.</given-names></name> <etal/></person-group>. (<year>2020</year>). <article-title>Deep learning interpretation of echocardiograms</article-title>. <source>NPJ Digital Med</source>. <volume>3</volume>, <fpage>1</fpage>&#x02013;<lpage>10</lpage>. <pub-id pub-id-type="doi">10.1038/s41746-019-0216-8</pub-id><pub-id pub-id-type="pmid">31993508</pub-id></citation></ref>
<ref id="B17">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gong</surname> <given-names>K.</given-names></name> <name><surname>Shi</surname> <given-names>T.</given-names></name> <name><surname>Zhao</surname> <given-names>L.</given-names></name> <name><surname>Xu</surname> <given-names>Z.</given-names></name> <name><surname>Wang</surname> <given-names>Z.</given-names></name></person-group> (<year>2021</year>). <article-title>Comparing the inter-observer reliability of the Tada formula among neurosurgeons while estimating the intracerebral haematoma volume</article-title>. <source>Clin. Neurol Neurosurg</source>. <volume>205</volume>, <fpage>106668</fpage>. <pub-id pub-id-type="doi">10.1016/j.clineuro.2021.106668</pub-id><pub-id pub-id-type="pmid">33962148</pub-id></citation></ref>
<ref id="B18">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gou</surname> <given-names>X.</given-names></name> <name><surname>He</surname> <given-names>X.</given-names></name></person-group> (<year>2021</year>). <article-title>Deep learning-based detection and diagnosis of subarachnoid hemorrhage</article-title>. <source>J. Healthc Eng</source>. 2021. <pub-id pub-id-type="doi">10.1155/2021/9639419</pub-id><pub-id pub-id-type="pmid">34853673</pub-id></citation></ref>
<ref id="B19">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>He</surname> <given-names>K.</given-names></name> <name><surname>Zhang</surname> <given-names>X.</given-names></name> <name><surname>Ren</surname> <given-names>S.</given-names></name> <name><surname>Sun</surname> <given-names>J.</given-names></name></person-group> (<year>2016</year>). <article-title>&#x0201C;Deep residual learning for image recognition,&#x0201D;</article-title> in <source>Proceedings of IEEE Conference on Computer Vision and Pattern Recognition</source> (<publisher-loc>Las Vegas, NV</publisher-loc>: <publisher-name>IEEE</publisher-name>), <fpage>770</fpage>&#x02013;<lpage>778</lpage>.<pub-id pub-id-type="pmid">32166560</pub-id></citation></ref>
<ref id="B20">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Hu</surname> <given-names>S.</given-names></name> <name><surname>Coup&#x000E9;</surname> <given-names>P.</given-names></name> <name><surname>Pruessner</surname> <given-names>J. C.</given-names></name> <name><surname>Collins</surname> <given-names>L. D.</given-names></name></person-group> (<year>2011</year>). <article-title>&#x0201C;Validation of appearance-model based segmentation with patch-based refinement on medial temporal lobe structures,&#x0201D;</article-title> in <source>MICCAI Workshop on Multi-Atlas Labeling and Statistical Fusion</source> (<publisher-loc>Toronto, ON</publisher-loc>), <fpage>28</fpage>&#x02013;<lpage>37</lpage>.<pub-id pub-id-type="pmid">22987811</pub-id></citation></ref>
<ref id="B21">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Huang</surname> <given-names>X.</given-names></name> <name><surname>Jiang</surname> <given-names>L.</given-names></name> <name><surname>Chen</surname> <given-names>S.</given-names></name> <name><surname>Li</surname> <given-names>G.</given-names></name> <name><surname>Pan</surname> <given-names>W.</given-names></name> <name><surname>Peng</surname> <given-names>L.</given-names></name> <etal/></person-group>. (<year>2021</year>). <article-title>Comparison of the curative effect and prognosis of stereotactic drainage and conservative treatment for moderate and small basal ganglia haemorrhage</article-title>. <source>BMC Neurol</source>. <volume>21</volume>, <fpage>1</fpage>&#x02013;<lpage>8</lpage>. <pub-id pub-id-type="doi">10.1186/s12883-021-02293-7</pub-id><pub-id pub-id-type="pmid">34229606</pub-id></citation></ref>
<ref id="B22">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Huttner</surname> <given-names>H. B.</given-names></name> <name><surname>Steiner</surname> <given-names>T.</given-names></name> <name><surname>Hartmann</surname> <given-names>M.</given-names></name> <name><surname>K&#x000F6;hrmann</surname> <given-names>M.</given-names></name> <name><surname>Juettler</surname> <given-names>E.</given-names></name> <name><surname>Mueller</surname> <given-names>S.</given-names></name> <etal/></person-group>. (<year>2006</year>). <article-title>Comparison of ABC/2 estimation technique to computer-assisted planimetric analysis in warfarin-related intracerebral parenchymal hemorrhage</article-title>. <source>Stroke</source> <volume>37</volume>, <fpage>404</fpage>&#x02013;<lpage>408</lpage>. <pub-id pub-id-type="doi">10.1161/01.STR.0000198806.67472.5c</pub-id><pub-id pub-id-type="pmid">16373654</pub-id></citation></ref>
<ref id="B23">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Ioffe</surname> <given-names>S.</given-names></name> <name><surname>Szegedy</surname> <given-names>C.</given-names></name></person-group> (<year>2015</year>). <article-title>&#x0201C;Batch normalization: accelerating deep network training by reducing internal covariate shift,&#x0201D;</article-title> in <source>International Conference on Machine Learning</source> (<publisher-loc>Lille</publisher-loc>), <fpage>448</fpage>&#x02013;<lpage>456</lpage>.<pub-id pub-id-type="pmid">35496726</pub-id></citation></ref>
<ref id="B24">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kase</surname> <given-names>C. S.</given-names></name> <name><surname>Hanley</surname> <given-names>D. F.</given-names></name></person-group> (<year>2021</year>). <article-title>Intracerebral hemorrhage: advances in emergency care</article-title>. <source>Neurol Clin</source>. <volume>39</volume>, <fpage>405</fpage>&#x02013;<lpage>418</lpage>. <pub-id pub-id-type="doi">10.1016/j.ncl.2021.02.002</pub-id><pub-id pub-id-type="pmid">33896526</pub-id></citation></ref>
<ref id="B25">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kerebel</surname> <given-names>D.</given-names></name> <name><surname>Joly</surname> <given-names>L.-M.</given-names></name> <name><surname>Honnart</surname> <given-names>D.</given-names></name> <name><surname>Schmidt</surname> <given-names>J.</given-names></name> <name><surname>Galanaud</surname> <given-names>D.</given-names></name> <name><surname>Negrier</surname> <given-names>C.</given-names></name> <etal/></person-group>. (<year>2013</year>). <article-title>A french multicenter randomised trial comparing two dose-regimens of prothrombin complex concentrates in urgent anticoagulation reversal</article-title>. <source>Crit. Care</source> <volume>17</volume>, <fpage>1</fpage>&#x02013;<lpage>11</lpage>. <pub-id pub-id-type="doi">10.1186/cc11923</pub-id><pub-id pub-id-type="pmid">23305460</pub-id></citation></ref>
<ref id="B26">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kim</surname> <given-names>H. J.</given-names></name></person-group> (<year>2012</year>). <article-title>The prognostic factors related to traumatic brain stem injury</article-title>. <source>J. Korean Neurosurg. Soc</source>. <volume>51</volume>, <fpage>24</fpage>&#x02013;<lpage>30</lpage>. <pub-id pub-id-type="doi">10.3340/jkns.2012.51.1.24</pub-id><pub-id pub-id-type="pmid">22396839</pub-id></citation></ref>
<ref id="B27">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kinoshita</surname> <given-names>K.</given-names></name> <name><surname>Hamanaka</surname> <given-names>G.</given-names></name> <name><surname>Ohtomo</surname> <given-names>R.</given-names></name> <name><surname>Takase</surname> <given-names>H.</given-names></name> <name><surname>Chung</surname> <given-names>K. K.</given-names></name> <name><surname>Lok</surname> <given-names>J.</given-names></name> <etal/></person-group>. (<year>2021</year>). <article-title>Mature adult mice with exercise-preconditioning show better recovery after intracerebral hemorrhage</article-title>. <source>Stroke</source> <volume>52</volume>, <fpage>1861</fpage>&#x02013;<lpage>1865</lpage>. <pub-id pub-id-type="doi">10.1161/STROKEAHA.120.032201</pub-id><pub-id pub-id-type="pmid">33840224</pub-id></citation></ref>
<ref id="B28">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lin</surname> <given-names>B.</given-names></name> <name><surname>Yang</surname> <given-names>H.</given-names></name> <name><surname>Cui</surname> <given-names>M.</given-names></name> <name><surname>Li</surname> <given-names>Y.</given-names></name> <name><surname>Yu</surname> <given-names>J.</given-names></name></person-group> (<year>2014</year>). <article-title>Surgicel&#x02122; application in intracranial hemorrhage surgery contributed to giant-cell granuloma in a patient with hypertension: case report and review of the literature</article-title>. <source>World J. Surg. Oncol</source>. <volume>12</volume>, <fpage>1</fpage>&#x02013;<lpage>5</lpage>. <pub-id pub-id-type="doi">10.1186/1477-7819-12-101</pub-id><pub-id pub-id-type="pmid">24745805</pub-id></citation></ref>
<ref id="B29">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Lin</surname> <given-names>J.</given-names></name> <name><surname>Gan</surname> <given-names>C.</given-names></name> <name><surname>Han</surname> <given-names>S.</given-names></name></person-group> (<year>2019</year>). <article-title>&#x0201C;TSM: temporal shift module for efficient video understanding,&#x0201D;</article-title> in <source>Proceedings of the IEEE International Conference on Computer Vision</source> (<publisher-loc>Seoul</publisher-loc>: <publisher-name>IEEE</publisher-name>), <fpage>7083</fpage>&#x02013;<lpage>7093</lpage>.<pub-id pub-id-type="pmid">33035158</pub-id></citation></ref>
<ref id="B30">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Litjens</surname> <given-names>G.</given-names></name> <name><surname>Kooi</surname> <given-names>T.</given-names></name> <name><surname>Bejnordi</surname> <given-names>B. E.</given-names></name> <name><surname>Setio</surname> <given-names>A. A. A.</given-names></name> <name><surname>Ciompi</surname> <given-names>F.</given-names></name> <name><surname>Ghafoorian</surname> <given-names>M.</given-names></name> <etal/></person-group>. (<year>2017</year>). <article-title>A survey on deep learning in medical image analysis</article-title>. <source>Med. Image Anal</source>. <volume>42</volume>, <fpage>60</fpage>&#x02013;<lpage>88</lpage>. <pub-id pub-id-type="doi">10.1016/j.media.2017.07.005</pub-id><pub-id pub-id-type="pmid">28778026</pub-id></citation></ref>
<ref id="B31">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Long</surname> <given-names>J.</given-names></name> <name><surname>Shelhamer</surname> <given-names>E.</given-names></name> <name><surname>Darrell</surname> <given-names>T.</given-names></name></person-group> (<year>2015</year>). <article-title>&#x0201C;Fully convolutional networks for semantic segmentation,&#x0201D;</article-title> in <source>Proceedings of IEEE Conference on Computer Vision and Pattern Recognition</source> (<publisher-loc>Boston, MA</publisher-loc>: <publisher-name>IEEE</publisher-name>), <fpage>3431</fpage>&#x02013;<lpage>3440</lpage>.</citation>
</ref>
<ref id="B32">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Mantas</surname> <given-names>J.</given-names></name></person-group> (<year>2020</year>). <article-title>&#x0201C;Classification of intracranial hemorrhage subtypes using deep learning on CT scans,&#x0201D;</article-title> in <source>The Importance of Health Informatics in Public Health during a Pandemic, Vol. 272</source> (<publisher-loc>Clifton</publisher-loc>), <fpage>370</fpage>.<pub-id pub-id-type="pmid">32604679</pub-id></citation></ref>
<ref id="B33">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Oge</surname> <given-names>D. D.</given-names></name> <name><surname>Arsava</surname> <given-names>E. M.</given-names></name> <name><surname>Pektezel</surname> <given-names>M. Y.</given-names></name> <name><surname>Gocmen</surname> <given-names>R.</given-names></name> <name><surname>Topcuoglu</surname> <given-names>M. A.</given-names></name></person-group> (<year>2021</year>). <article-title>Intracerebral hemorrhage volume estimation: Is modification of the ABC/2 formula necessary according to the hematoma shape?</article-title> <source>Clin. Neurol Neurosurg</source>. <volume>207</volume>, <fpage>106779</fpage>. <pub-id pub-id-type="doi">10.1016/j.clineuro.2021.106779</pub-id><pub-id pub-id-type="pmid">34214866</pub-id></citation></ref>
<ref id="B34">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Padma Nanthagopal</surname> <given-names>A.</given-names></name> <name><surname>Sukanesh Rajamony</surname> <given-names>R.</given-names></name></person-group> (<year>2013</year>). <article-title>Classification of benign and malignant brain tumor CT images using wavelet texture parameters and neural network classifier</article-title>. <source>J. Visualizat</source>. <volume>16</volume>, <fpage>19</fpage>&#x02013;<lpage>28</lpage>. <pub-id pub-id-type="doi">10.1007/s12650-012-0153-y</pub-id></citation>
</ref>
<ref id="B35">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Pan</surname> <given-names>Y.</given-names></name> <name><surname>Huang</surname> <given-names>W.</given-names></name> <name><surname>Lin</surname> <given-names>Z.</given-names></name> <name><surname>Zhu</surname> <given-names>W.</given-names></name> <name><surname>Zhou</surname> <given-names>J.</given-names></name> <name><surname>Wong</surname> <given-names>J.</given-names></name> <etal/></person-group>. (<year>2015</year>). <article-title>&#x0201C;Brain tumor grading based on neural networks and convolutional neural networks,&#x0201D;</article-title> in <source>Annual International Conference of the IEEE Engineering in Medicine and Biology Society</source> (<publisher-loc>Milan</publisher-loc>: <publisher-name>IEEE</publisher-name>), <fpage>699</fpage>&#x02013;<lpage>702</lpage>.<pub-id pub-id-type="pmid">26736358</pub-id></citation></ref>
<ref id="B36">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Paszke</surname> <given-names>A.</given-names></name> <name><surname>Gross</surname> <given-names>S.</given-names></name> <name><surname>Massa</surname> <given-names>F.</given-names></name> <name><surname>Lerer</surname> <given-names>A.</given-names></name> <name><surname>Bradbury</surname> <given-names>J.</given-names></name> <name><surname>Chanan</surname> <given-names>G.</given-names></name> <etal/></person-group>. (<year>2019</year>). <article-title>Pytorch: an imperative style, high-performance deep learning library</article-title>. <source>Adv. Neural Inf. Process. Syst</source>. <volume>32</volume>, <fpage>8024</fpage>&#x02013;<lpage>8035</lpage>. <pub-id pub-id-type="doi">10.48550/arXiv.1912.01703</pub-id></citation>
</ref>
<ref id="B37">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Patel</surname> <given-names>A.</given-names></name> <name><surname>Schreuder</surname> <given-names>F. H.</given-names></name> <name><surname>Klijn</surname> <given-names>C. J.</given-names></name> <name><surname>Prokop</surname> <given-names>M.</given-names></name> <name><surname>Ginneken</surname> <given-names>B. v.</given-names></name><etal/></person-group>. (<year>2019</year>). <article-title>Intracerebral haemorrhage segmentation in non-contrast CT</article-title>. <source>Sci. Rep</source>. <volume>9</volume>, <fpage>1</fpage>&#x02013;<lpage>11</lpage>. <pub-id pub-id-type="doi">10.1038/s41598-019-54491-6</pub-id><pub-id pub-id-type="pmid">31780815</pub-id></citation></ref>
<ref id="B38">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Phaphuangwittayakul</surname> <given-names>A.</given-names></name> <name><surname>Guo</surname> <given-names>Y.</given-names></name> <name><surname>Ying</surname> <given-names>F.</given-names></name> <name><surname>Dawod</surname> <given-names>A. Y.</given-names></name> <name><surname>Angkurawaranon</surname> <given-names>S.</given-names></name> <name><surname>Angkurawaranon</surname> <given-names>C.</given-names></name></person-group> (<year>2022</year>). <article-title>An optimal deep learning framework for multi-type hemorrhagic lesions detection and quantification in head CT images for traumatic brain injury</article-title>. <source>Appl. Intell</source>. <volume>52</volume>, <fpage>7320</fpage>&#x02013;<lpage>7338</lpage>. <pub-id pub-id-type="doi">10.1007/s10489-021-02782-9</pub-id><pub-id pub-id-type="pmid">34764620</pub-id></citation></ref>
<ref id="B39">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Qiu</surname> <given-names>Z.</given-names></name> <name><surname>Yao</surname> <given-names>T.</given-names></name> <name><surname>Mei</surname> <given-names>T.</given-names></name></person-group> (<year>2017</year>). <article-title>&#x0201C;Learning spatio-temporal representation with pseudo-3d residual networks,&#x0201D;</article-title> in <source>Proceedings of the IEEE International Conference on Computer Vision</source> (<publisher-loc>Venice</publisher-loc>: <publisher-name>IEEE</publisher-name>), <fpage>5533</fpage>&#x02013;<lpage>5541</lpage>.</citation>
</ref>
<ref id="B40">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Raafat</surname> <given-names>M.</given-names></name> <name><surname>Ragab</surname> <given-names>O. A.</given-names></name> <name><surname>Abdelwahab</surname> <given-names>O. M.</given-names></name> <name><surname>Salama</surname> <given-names>M. M.</given-names></name> <name><surname>Hafez</surname> <given-names>M. A.</given-names></name></person-group> (<year>2020</year>). <article-title>Early versus delayed surgical evacuation of spontaneous supratentorial intracerebral hematoma: a prospective cohort study</article-title>. <source>Surg. Neurol Int</source>. <volume>11</volume>, <fpage>103</fpage>. <pub-id pub-id-type="doi">10.25259/SNI_103_2020</pub-id><pub-id pub-id-type="pmid">32547832</pub-id></citation></ref>
<ref id="B41">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Rav&#x000EC;</surname> <given-names>D.</given-names></name> <name><surname>Wong</surname> <given-names>C.</given-names></name> <name><surname>Deligianni</surname> <given-names>F.</given-names></name> <name><surname>Berthelot</surname> <given-names>M.</given-names></name> <name><surname>Andreu-Perez</surname> <given-names>J.</given-names></name> <name><surname>Lo</surname> <given-names>B.</given-names></name> <etal/></person-group>. (<year>2016</year>). <article-title>Deep learning for health informatics</article-title>. <source>IEEE J. Biomed. Health Inform</source>. <volume>21</volume>, <fpage>4</fpage>&#x02013;<lpage>21</lpage>. <pub-id pub-id-type="doi">10.1109/JBHI.2016.2636665</pub-id><pub-id pub-id-type="pmid">28055930</pub-id></citation></ref>
<ref id="B42">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Robbins</surname> <given-names>H.</given-names></name> <name><surname>Monro</surname> <given-names>S.</given-names></name></person-group> (<year>1951</year>). <article-title>&#x0201C;A stochastic approximation method</article-title>. <source>Ann. Math. Stat</source>. <volume>22</volume>, <fpage>400</fpage>&#x02013;<lpage>407</lpage>. <pub-id pub-id-type="doi">10.1214/aoms/1177729586</pub-id></citation>
</ref>
<ref id="B43">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Roh</surname> <given-names>D.</given-names></name> <name><surname>Sun</surname> <given-names>C.-H.</given-names></name> <name><surname>Murthy</surname> <given-names>S.</given-names></name> <name><surname>Elkind</surname> <given-names>M. S.</given-names></name> <name><surname>Bruce</surname> <given-names>S. S.</given-names></name> <name><surname>Melmed</surname> <given-names>K.</given-names></name> <etal/></person-group>. (<year>2019</year>). <article-title>Hematoma expansion differences in lobar and deep primary intracerebral hemorrhage</article-title>. <source>Neurocrit. Care</source> <volume>31</volume>, <fpage>40</fpage>&#x02013;<lpage>45</lpage>. <pub-id pub-id-type="doi">10.1007/s12028-018-00668-2</pub-id><pub-id pub-id-type="pmid">30756318</pub-id></citation></ref>
<ref id="B44">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Sachdeva</surname> <given-names>J.</given-names></name> <name><surname>Kumar</surname> <given-names>V.</given-names></name> <name><surname>Gupta</surname> <given-names>I.</given-names></name> <name><surname>Khandelwal</surname> <given-names>N.</given-names></name> <name><surname>Ahuja</surname> <given-names>C. K.</given-names></name></person-group> (<year>2013</year>). <article-title>Segmentation, feature extraction, and multiclass brain tumor classification</article-title>. <source>J. Digit. Imaging</source> <volume>26</volume>, <fpage>1141</fpage>&#x02013;<lpage>1150</lpage>. <pub-id pub-id-type="doi">10.1007/s10278-013-9600-0</pub-id><pub-id pub-id-type="pmid">23645344</pub-id></citation></ref>
<ref id="B45">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Samek</surname> <given-names>W.</given-names></name> <name><surname>Binder</surname> <given-names>A.</given-names></name> <name><surname>Montavon</surname> <given-names>G.</given-names></name> <name><surname>Lapuschkin</surname> <given-names>S.</given-names></name> <name><surname>M&#x000FC;ller</surname> <given-names>K.-R.</given-names></name></person-group> (<year>2016</year>). <article-title>Evaluating the visualization of what a deep neural network has learned</article-title>. <source>IEEE Trans. Neural Networks Learn. Syst</source>. <volume>28</volume>, <fpage>2660</fpage>&#x02013;<lpage>2673</lpage>. <pub-id pub-id-type="doi">10.1109/TNNLS.2016.2599820</pub-id><pub-id pub-id-type="pmid">27576267</pub-id></citation></ref>
<ref id="B46">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Selvaraju</surname> <given-names>R. R.</given-names></name> <name><surname>Cogswell</surname> <given-names>M.</given-names></name> <name><surname>Das</surname> <given-names>A.</given-names></name> <name><surname>Vedantam</surname> <given-names>R.</given-names></name> <name><surname>Parikh</surname> <given-names>D.</given-names></name> <name><surname>Batra</surname> <given-names>D.</given-names></name></person-group> (<year>2017</year>). <article-title>&#x0201C;Grad-cam: visual explanations from deep networks via gradient-based localization,&#x0201D;</article-title> in <source>Proceedings of the IEEE International Conference on Computer Vision</source> (<publisher-loc>Venice</publisher-loc>: <publisher-name>IEEE</publisher-name>), <fpage>618</fpage>&#x02013;<lpage>626</lpage>.</citation>
</ref>
<ref id="B47">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Shao</surname> <given-names>H.</given-names></name> <name><surname>Qian</surname> <given-names>S.</given-names></name> <name><surname>Liu</surname> <given-names>Y.</given-names></name></person-group> (<year>2020</year>). <article-title>Temporal interlacing network</article-title>. <source>Proc. AAAI Conf. Artif. Intell</source>. <volume>34</volume>, <fpage>11966</fpage>&#x02013;<lpage>11973</lpage>. <pub-id pub-id-type="doi">10.1609/aaai.v34i07.6872</pub-id></citation>
</ref>
<ref id="B48">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Sharrock</surname> <given-names>M. F.</given-names></name> <name><surname>Mould</surname> <given-names>W. A.</given-names></name> <name><surname>Ali</surname> <given-names>H.</given-names></name> <name><surname>Hildreth</surname> <given-names>M.</given-names></name> <name><surname>Awad</surname> <given-names>I. A.</given-names></name> <name><surname>Hanley</surname> <given-names>D. F.</given-names></name> <etal/></person-group>. (<year>2021</year>). <article-title>3D deep neural network segmentation of intracerebral hemorrhage: development and validation for clinical trials</article-title>. <source>Neuroinformatics</source> <volume>19</volume>, <fpage>403</fpage>&#x02013;<lpage>415</lpage>. <pub-id pub-id-type="doi">10.1007/s12021-020-09493-5</pub-id><pub-id pub-id-type="pmid">32980970</pub-id></citation></ref>
<ref id="B49">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Sharrock</surname> <given-names>M. F.</given-names></name> <name><surname>Mould</surname> <given-names>W. A.</given-names></name> <name><surname>Hildreth</surname> <given-names>M.</given-names></name> <name><surname>Ryu</surname> <given-names>E. P.</given-names></name> <name><surname>Walborn</surname> <given-names>N.</given-names></name> <name><surname>Awad</surname> <given-names>I. A.</given-names></name> <etal/></person-group>. (<year>2022</year>). <article-title>Bayesian deep learning outperforms clinical trial estimators of intracerebral and intraventricular hemorrhage volume</article-title>. <source>J. Neuroimaging</source> <volume>32</volume>, <fpage>968</fpage>&#x02013;<lpage>976</lpage>. <pub-id pub-id-type="doi">10.1111/jon.12997</pub-id><pub-id pub-id-type="pmid">35434846</pub-id></citation></ref>
<ref id="B50">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Simonyan</surname> <given-names>K.</given-names></name> <name><surname>Vedaldi</surname> <given-names>A.</given-names></name> <name><surname>Zisserman</surname> <given-names>A.</given-names></name></person-group> (<year>2013</year>). <article-title>Deep inside convolutional networks: visualising image classification models and saliency maps</article-title>. <source>arXiv preprint</source> arXiv:1312.6034. <pub-id pub-id-type="doi">10.48550/arXiv.1312.6034</pub-id></citation>
</ref>
<ref id="B51">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Smilkov</surname> <given-names>D.</given-names></name> <name><surname>Thorat</surname> <given-names>N.</given-names></name> <name><surname>Kim</surname> <given-names>B.</given-names></name> <name><surname>Vi&#x000E9;gas</surname> <given-names>F.</given-names></name> <name><surname>Wattenberg</surname> <given-names>M.</given-names></name></person-group> (<year>2017</year>). <article-title>Smoothgrad: removing noise by adding noise</article-title>. <source>arXiv preprint</source> arXiv:1706.03825. <pub-id pub-id-type="doi">10.48550/arXiv.1706.03825</pub-id></citation>
</ref>
<ref id="B52">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Thomas</surname> <given-names>K. A.</given-names></name> <name><surname>Kidzi&#x00144;ski</surname> <given-names>&#x00141;.</given-names></name> <name><surname>Halilaj</surname> <given-names>E.</given-names></name> <name><surname>Fleming</surname> <given-names>S. L.</given-names></name> <name><surname>Venkataraman</surname> <given-names>G. R.</given-names></name> <name><surname>Oei</surname> <given-names>E. H.</given-names></name> <etal/></person-group>. (<year>2020</year>). <article-title>Automated classification of radiographic knee osteoarthritis severity using deep neural networks</article-title>. <source>Radiol. Artif. Intell</source>. <volume>2</volume>, <fpage>e190065</fpage>. <pub-id pub-id-type="doi">10.1148/ryai.2020190065</pub-id><pub-id pub-id-type="pmid">32280948</pub-id></citation></ref>
<ref id="B53">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Tran</surname> <given-names>D.</given-names></name> <name><surname>Wang</surname> <given-names>H.</given-names></name> <name><surname>Torresani</surname> <given-names>L.</given-names></name> <name><surname>Ray</surname> <given-names>J.</given-names></name> <name><surname>LeCun</surname> <given-names>Y.</given-names></name> <name><surname>Paluri</surname> <given-names>M.</given-names></name></person-group> (<year>2018</year>). <article-title>&#x0201C;A closer look at spatiotemporal convolutions for action recognition,&#x0201D;</article-title> in <source>Proceedings of IEEE Conference on Computer Vision and Pattern Recognition</source> (<publisher-loc>Salt Lake City, UT</publisher-loc>: <publisher-name>IEEE</publisher-name>), <fpage>6450</fpage>&#x02013;<lpage>6459</lpage>.</citation>
</ref>
<ref id="B54">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Van Asch</surname> <given-names>C. J.</given-names></name> <name><surname>Luitse</surname> <given-names>M. J.</given-names></name> <name><surname>Rinkel</surname> <given-names>G. J.</given-names></name> <name><surname>van der Tweel</surname> <given-names>I.</given-names></name> <name><surname>Algra</surname> <given-names>A.</given-names></name> <name><surname>Klijn</surname> <given-names>C. J.</given-names></name></person-group> (<year>2010</year>). <article-title>Incidence, case fatality, and functional outcome of intracerebral haemorrhage over time, according to age, sex, and ethnic origin: a systematic review and meta-analysis</article-title>. <source>Lancet Neurol</source>. <volume>9</volume>, <fpage>167</fpage>&#x02013;<lpage>176</lpage>. <pub-id pub-id-type="doi">10.1016/S1474-4422(09)70340-0</pub-id><pub-id pub-id-type="pmid">20056489</pub-id></citation></ref>
<ref id="B55">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Vidyarthi</surname> <given-names>A.</given-names></name> <name><surname>Mittal</surname> <given-names>N.</given-names></name></person-group> (<year>2014</year>). <article-title>&#x0201C;Comparative study for brain tumor classification on MR/CT images,&#x0201D;</article-title> in <source>Proceedings of the International Conference on Soft Computing for Problem Solving</source> (<publisher-loc>New Delhi</publisher-loc>), <fpage>889</fpage>&#x02013;<lpage>897</lpage>.</citation>
</ref>
<ref id="B56">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>C.-W.</given-names></name> <name><surname>Juan</surname> <given-names>C.-J.</given-names></name> <name><surname>Liu</surname> <given-names>Y.-J.</given-names></name> <name><surname>Hsu</surname> <given-names>H.-H.</given-names></name> <name><surname>Liu</surname> <given-names>H.-S.</given-names></name> <name><surname>Chen</surname> <given-names>C.-Y.</given-names></name> <etal/></person-group>. (<year>2009</year>). <article-title>Volume-dependent overestimation of spontaneous intracerebral hematoma volume by the ABC/2 formula</article-title>. <source>Acta Radiol</source>. <volume>50</volume>, <fpage>306</fpage>&#x02013;<lpage>311</lpage>. <pub-id pub-id-type="doi">10.1080/02841850802647039</pub-id><pub-id pub-id-type="pmid">19173095</pub-id></citation></ref>
<ref id="B57">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>C.-W.</given-names></name> <name><surname>Liu</surname> <given-names>Y.-J.</given-names></name> <name><surname>Lee</surname> <given-names>Y.-H.</given-names></name> <name><surname>Hueng</surname> <given-names>D.-Y.</given-names></name> <name><surname>Fan</surname> <given-names>H.-C.</given-names></name> <name><surname>Yang</surname> <given-names>F.-C.</given-names></name> <etal/></person-group>. (<year>2014</year>). <article-title>Hematoma shape, hematoma size, Glasgow coma scale score and ICH score: which predicts the 30-day mortality better for intracerebral hematoma?</article-title> <source>PLoS ONE</source> <volume>9</volume>, <fpage>e102326</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pone.0102326</pub-id><pub-id pub-id-type="pmid">25029592</pub-id></citation></ref>
<ref id="B58">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Xu</surname> <given-names>J.</given-names></name> <name><surname>Zhang</surname> <given-names>R.</given-names></name> <name><surname>Zhou</surname> <given-names>Z.</given-names></name> <name><surname>Wu</surname> <given-names>C.</given-names></name> <name><surname>Gong</surname> <given-names>Q.</given-names></name> <name><surname>Zhang</surname> <given-names>H.</given-names></name> <etal/></person-group>. (<year>2021</year>). <article-title>Deep network for the automatic segmentation and quantification of intracranial hemorrhage on CT</article-title>. <source>Front. Neurosci</source>. 14, 541817. <pub-id pub-id-type="doi">10.3389/fnins.2020.541817</pub-id><pub-id pub-id-type="pmid">33505231</pub-id></citation></ref>
<ref id="B59">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Xu</surname> <given-names>X.</given-names></name> <name><surname>Chen</surname> <given-names>X.</given-names></name> <name><surname>Zhang</surname> <given-names>J.</given-names></name> <name><surname>Zheng</surname> <given-names>Y.</given-names></name> <name><surname>Sun</surname> <given-names>G.</given-names></name> <name><surname>Yu</surname> <given-names>X.</given-names></name> <etal/></person-group>. (<year>2014</year>). <article-title>Comparison of the Tada formula with software slicer: precise and low-cost method for volume assessment of intracerebral hematoma</article-title>. <source>Stroke</source> <volume>45</volume>, <fpage>3433</fpage>&#x02013;<lpage>3435</lpage>. <pub-id pub-id-type="doi">10.1161/STROKEAHA.114.007095</pub-id><pub-id pub-id-type="pmid">25316277</pub-id></citation></ref>
<ref id="B60">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Yang</surname> <given-names>W.</given-names></name> <name><surname>Feng</surname> <given-names>Y.</given-names></name> <name><surname>Zhang</surname> <given-names>Y.</given-names></name> <name><surname>Yan</surname> <given-names>J.</given-names></name> <name><surname>Fu</surname> <given-names>Y.</given-names></name> <name><surname>Chen</surname> <given-names>S.</given-names></name></person-group> (<year>2013</year>). <article-title>Volume quantification of acute infratentorial hemorrhage with computed tomography: validation of the formula 1/2ABC and 2/3Sh</article-title>. <source>PLoS ONE</source> <volume>8</volume>, <fpage>e62286</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pone.0062286</pub-id><pub-id pub-id-type="pmid">23638025</pub-id></citation></ref>
<ref id="B61">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Yu</surname> <given-names>H.</given-names></name> <name><surname>Cao</surname> <given-names>X.</given-names></name> <name><surname>Li</surname> <given-names>W.</given-names></name> <name><surname>Liu</surname> <given-names>P.</given-names></name> <name><surname>Zhao</surname> <given-names>Y.</given-names></name> <name><surname>Song</surname> <given-names>L.</given-names></name> <etal/></person-group>. (<year>2020</year>). <article-title>Targeting connexin 43 provides anti-inflammatory effects after intracerebral hemorrhage injury by regulating yap signaling</article-title>. <source>J. Neuroinflamm</source>. <volume>17</volume>, <fpage>1</fpage>&#x02013;<lpage>19</lpage>. <pub-id pub-id-type="doi">10.1186/s12974-020-01978-z</pub-id><pub-id pub-id-type="pmid">33115476</pub-id></citation></ref>
<ref id="B62">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Zeiler</surname> <given-names>M. D.</given-names></name> <name><surname>Fergus</surname> <given-names>R.</given-names></name></person-group> (<year>2014</year>). <article-title>&#x0201C;Visualizing and understanding convolutional networks,&#x0201D;</article-title> in <source>European Conference on Computer Vision</source> (<publisher-loc>Cham</publisher-loc>), <fpage>818</fpage>&#x02013;<lpage>833</lpage>.</citation>
</ref>
<ref id="B63">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Zeiler</surname> <given-names>M. D.</given-names></name> <name><surname>Taylor</surname> <given-names>G. W.</given-names></name> <name><surname>Fergus</surname> <given-names>R.</given-names></name></person-group> (<year>2011</year>). <article-title>&#x0201C;Adaptive deconvolutional networks for mid and high level feature learning,&#x0201D;</article-title> in <source>Proceedings of the IEEE International Conference on Computer Vision</source> (<publisher-loc>Barcelona</publisher-loc>: <publisher-name>IEEE</publisher-name>), <fpage>2018</fpage>&#x02013;<lpage>2025</lpage>.</citation>
</ref>
<ref id="B64">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>J.</given-names></name> <name><surname>Cheng</surname> <given-names>H.</given-names></name> <name><surname>Zhou</surname> <given-names>S.</given-names></name> <name><surname>Huang</surname> <given-names>L.</given-names></name> <name><surname>Lv</surname> <given-names>J.</given-names></name> <name><surname>Wang</surname> <given-names>P.</given-names></name> <etal/></person-group>. (<year>2020</year>). <article-title>3D-printed model-guided endoscopic evacuation for basal ganglia hemorrhage</article-title>. <source>Sci. Rep</source>. <volume>10</volume>, <fpage>1</fpage>&#x02013;<lpage>7</lpage>. <pub-id pub-id-type="doi">10.1038/s41598-020-62232-3</pub-id><pub-id pub-id-type="pmid">32251343</pub-id></citation></ref>
<ref id="B65">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhao</surname> <given-names>K.-J.</given-names></name> <name><surname>Liu</surname> <given-names>Y.</given-names></name> <name><surname>Zhang</surname> <given-names>R.-Y.</given-names></name> <name><surname>Wang</surname> <given-names>X.-Q.</given-names></name> <name><surname>Gao</surname> <given-names>C.</given-names></name> <name><surname>Shen</surname> <given-names>J.-K.</given-names></name></person-group> (<year>2009</year>). <article-title>A precise, simple, convenient and new method for estimation of intracranial hematoma volume-the formula 2/3 Sh</article-title>. <source>Neurol Res</source>. <volume>31</volume>, <fpage>1031</fpage>&#x02013;<lpage>1036</lpage>. <pub-id pub-id-type="doi">10.1179/174313209X385662</pub-id><pub-id pub-id-type="pmid">19426586</pub-id></citation></ref>
<ref id="B66">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhao</surname> <given-names>X.</given-names></name> <name><surname>Chen</surname> <given-names>K.</given-names></name> <name><surname>Wu</surname> <given-names>G.</given-names></name> <name><surname>Zhang</surname> <given-names>G.</given-names></name> <name><surname>Zhou</surname> <given-names>X.</given-names></name> <name><surname>Lv</surname> <given-names>C.</given-names></name> <etal/></person-group>. (<year>2021</year>). <article-title>Deep learning shows good reliability for automatic segmentation and volume measurement of brain hemorrhage, intraventricular extension, and peripheral edema</article-title>. <source>Eur. Radiol</source>. <volume>31</volume>, <fpage>5012</fpage>&#x02013;<lpage>5020</lpage>. <pub-id pub-id-type="doi">10.1007/s00330-020-07558-2</pub-id><pub-id pub-id-type="pmid">33409788</pub-id></citation></ref>
<ref id="B67">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Zhou</surname> <given-names>B.</given-names></name> <name><surname>Khosla</surname> <given-names>A.</given-names></name> <name><surname>Lapedriza</surname> <given-names>A.</given-names></name> <name><surname>Oliva</surname> <given-names>A.</given-names></name> <name><surname>Torralba</surname> <given-names>A.</given-names></name></person-group> (<year>2016</year>). <article-title>&#x0201C;Learning deep features for discriminative localization,&#x0201D;</article-title> in <source>Proceedings of IEEE Conference on Computer Vision and Pattern Recognition</source> (<publisher-loc>Las Vegas, NV</publisher-loc>: <publisher-name>IEEE</publisher-name>), <fpage>2921</fpage>&#x02013;<lpage>2929</lpage>.</citation>
</ref>
<ref id="B68">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhou</surname> <given-names>Q.</given-names></name> <name><surname>Zhu</surname> <given-names>W.</given-names></name> <name><surname>Li</surname> <given-names>F.</given-names></name> <name><surname>Yuan</surname> <given-names>M.</given-names></name> <name><surname>Zheng</surname> <given-names>L.</given-names></name> <name><surname>Liu</surname> <given-names>X.</given-names></name></person-group> (<year>2022</year>). <article-title>Transfer learning of the ResNet-18 and DenseNet-121 model used to diagnose intracranial hemorrhage in CT scanning</article-title>. <source>Curr. Pharm Des</source>. <volume>28</volume>, <fpage>287</fpage>&#x02013;<lpage>295</lpage>. <pub-id pub-id-type="doi">10.2174/1381612827666211213143357</pub-id><pub-id pub-id-type="pmid">34961458</pub-id></citation></ref>
<ref id="B69">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Zunair</surname> <given-names>H.</given-names></name> <name><surname>Rahman</surname> <given-names>A.</given-names></name> <name><surname>Mohammed</surname> <given-names>N.</given-names></name> <name><surname>Cohen</surname> <given-names>J. P.</given-names></name></person-group> (<year>2020</year>). <article-title>&#x0201C;Uniformizing techniques to process CT scans with 3D CNNs for tuberculosis prediction,&#x0201D;</article-title> in <source>International Workshop on PRedictive Intelligence in MEdicine</source> (<publisher-loc>Cham</publisher-loc>), <fpage>156</fpage>&#x02013;<lpage>168</lpage>.</citation>
</ref>
</ref-list>
</back>
</article> 