<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article article-type="research-article" dtd-version="2.3" xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Artif. Intell.</journal-id>
<journal-title>Frontiers in Artificial Intelligence</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Artif. Intell.</abbrev-journal-title>
<issn pub-type="epub">2624-8212</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">598932</article-id>
<article-id pub-id-type="doi">10.3389/frai.2021.598932</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Artificial Intelligence</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>COVID-FACT: A Fully-Automated Capsule Network-Based Framework for Identification of COVID-19 Cases from Chest CT Scans</article-title>
<alt-title alt-title-type="left-running-head">Heidarian et&#x20;al.</alt-title>
<alt-title alt-title-type="right-running-head">COVID-FACT</alt-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name>
<surname>Heidarian</surname>
<given-names>Shahin</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/1069922/overview"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Afshar</surname>
<given-names>Parnian</given-names>
</name>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Enshaei</surname>
<given-names>Nastaran</given-names>
</name>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Naderkhani</surname>
<given-names>Farnoosh</given-names>
</name>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Rafiee</surname>
<given-names>Moezedin Javad</given-names>
</name>
<xref ref-type="aff" rid="aff3">
<sup>3</sup>
</xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Babaki Fard</surname>
<given-names>Faranak</given-names>
</name>
<xref ref-type="aff" rid="aff4">
<sup>4</sup>
</xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Samimi</surname>
<given-names>Kaveh</given-names>
</name>
<xref ref-type="aff" rid="aff5">
<sup>5</sup>
</xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Atashzar</surname>
<given-names>S. Farokh</given-names>
</name>
<xref ref-type="aff" rid="aff6">
<sup>6</sup>
</xref>
<xref ref-type="aff" rid="aff7">
<sup>7</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/990532/overview"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Oikonomou</surname>
<given-names>Anastasia</given-names>
</name>
<xref ref-type="aff" rid="aff8">
<sup>8</sup>
</xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Plataniotis</surname>
<given-names>Konstantinos N.</given-names>
</name>
<xref ref-type="aff" rid="aff9">
<sup>9</sup>
</xref>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Mohammadi</surname>
<given-names>Arash</given-names>
</name>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<xref ref-type="corresp" rid="c001">&#x2a;</xref>
<uri xlink:href="https://loop.frontiersin.org/people/1069816/overview"/>
</contrib>
</contrib-group>
<aff id="aff1">
<label>
<sup>1</sup>
</label>Department of Electrical and Computer Engineering, Concordia University, <addr-line>Montreal</addr-line>, <addr-line>QC</addr-line>, <country>Canada</country>
</aff>
<aff id="aff2">
<label>
<sup>2</sup>
</label>Concordia Institute for Information Systems Engineering (CIISE), Concordia University, <addr-line>Montreal</addr-line>, <addr-line>QC</addr-line>, <country>Canada</country>
</aff>
<aff id="aff3">
<label>
<sup>3</sup>
</label>Department of Medicine and Diagnostic Radiology, McGill University Health Center-Research Institute, <addr-line>Montreal</addr-line>, <addr-line>QC</addr-line>, <country>Canada</country>
</aff>
<aff id="aff4">
<label>
<sup>4</sup>
</label>Biomedical Sciences Department, Faculty of Medicine, University of Montreal, <addr-line>Montreal</addr-line>, <addr-line>QC</addr-line>, <country>Canada</country>
</aff>
<aff id="aff5">
<label>
<sup>5</sup>
</label>Department of Radiology, Iran University of Medical Science, <addr-line>Tehran</addr-line>, <country>Iran</country>
</aff>
<aff id="aff6">
<label>
<sup>6</sup>
</label>Department of Electrical and Computer Engineering, New York University, New York, <addr-line>NY</addr-line>, <country>United&#x20;States</country>
</aff>
<aff id="aff7">
<label>
<sup>7</sup>
</label>Department of Mechanical and Aerospace Engineering, New York University, New York, <addr-line>NY</addr-line>, <country>United&#x20;States</country>
</aff>
<aff id="aff8">
<label>
<sup>8</sup>
</label>Department of Medical Imaging, Sunnybrook Health Sciences Centre, University of Toronto, <addr-line>Toronto</addr-line>, <addr-line>ON</addr-line>, <country>Canada</country>
</aff>
<aff id="aff9">
<label>
<sup>9</sup>
</label>Department of Electrical and Computer Engineering, University of Toronto, <addr-line>Toronto</addr-line>, <addr-line>ON</addr-line>, <country>Canada</country>
</aff>
<author-notes>
<fn fn-type="edited-by">
<p>
<bold>Edited by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/431624/overview">Jun Deng</ext-link>, Yale University, United&#x20;States</p>
</fn>
<fn fn-type="edited-by">
<p>
<bold>Reviewed by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/483672/overview">Shailesh Tripathi</ext-link>, Tampere University of Technology, Finland</p>
<p>
<ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1110848/overview">Yuliang Huang</ext-link>, Peking University Cancer Hospital, China</p>
</fn>
<corresp id="c001">&#x2a;Correspondence: Arash Mohammadi, <email>arash.mohammadi@concordia.ca</email>
</corresp>
<fn fn-type="other">
<p>This article was submitted to Medicine and Public Health, a section of the journal Frontiers in Artificial Intelligence</p>
</fn>
</author-notes>
<pub-date pub-type="epub">
<day>25</day>
<month>05</month>
<year>2021</year>
</pub-date>
<pub-date pub-type="collection">
<year>2021</year>
</pub-date>
<volume>4</volume>
<elocation-id>598932</elocation-id>
<history>
<date date-type="received">
<day>10</day>
<month>07</month>
<year>2020</year>
</date>
<date date-type="accepted">
<day>09</day>
<month>02</month>
<year>2021</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#xa9; 2021 Heidarian, Afshar, Enshaei, Naderkhani, Rafiee, Babaki Fard, Samimi, Atashzar, Oikonomou, Plataniotis and Mohammadi.</copyright-statement>
<copyright-year>2021</copyright-year>
<copyright-holder>Heidarian, Afshar, Enshaei, Naderkhani, Rafiee, Babaki Fard, Samimi, Atashzar, Oikonomou, Plataniotis and Mohammadi</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/">
<p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these&#x20;terms.</p>
</license>
</permissions>
<abstract>
<p>The newly discovered Coronavirus Disease 2019 (COVID-19) has been globally spreading and causing hundreds of thousands of deaths around the world as of its first emergence in late 2019. The rapid outbreak of this disease has overwhelmed health care infrastructures and arises the need to allocate medical equipment and resources more efficiently. The early diagnosis of this disease will lead to the rapid separation of COVID-19 and non-COVID cases, which will be helpful for health care authorities to optimize resource allocation plans and early prevention of the disease. In this regard, a growing number of studies are investigating the capability of deep learning for early diagnosis of COVID-19. Computed tomography (CT) scans have shown distinctive features and higher sensitivity compared to other diagnostic tests, in particular the current gold standard, i.e.,&#x20;the Reverse Transcription Polymerase Chain Reaction (RT-PCR) test. Current deep learning-based algorithms are mainly developed based on Convolutional Neural Networks (CNNs) to identify COVID-19 pneumonia cases. CNNs, however, require extensive data augmentation and large datasets to identify detailed spatial relations between image instances. Furthermore, existing algorithms utilizing CT scans, either extend slice-level predictions to patient-level ones using a simple thresholding mechanism or rely on a sophisticated infection segmentation to identify the disease. In this paper, we propose a two-stage fully automated CT-based framework for identification of COVID-19 positive cases referred to as the &#x201c;COVID-FACT&#x201d;. COVID-FACT utilizes Capsule Networks, as its main building blocks and is, therefore, capable of capturing spatial information. In particular, to make the proposed COVID-FACT independent from sophisticated segmentations of the area of infection, slices demonstrating infection are detected at the first stage and the second stage is responsible for classifying patients into COVID and non-COVID cases. COVID-FACT detects slices with infection, and identifies positive COVID-19 cases using an in-house CT scan dataset, containing COVID-19, community acquired pneumonia, and normal cases. Based on our experiments, COVID-FACT achieves an accuracy of <inline-formula id="inf1">
<mml:math id="m1">
<mml:mrow>
<mml:mn>90.82</mml:mn>
<mml:mtext>%</mml:mtext>
</mml:mrow>
</mml:math>
</inline-formula>, a sensitivity of <inline-formula id="inf2">
<mml:math id="m2">
<mml:mrow>
<mml:mn>94.55</mml:mn>
<mml:mtext>%</mml:mtext>
</mml:mrow>
</mml:math>
</inline-formula>, a specificity of <inline-formula id="inf3">
<mml:math id="m3">
<mml:mrow>
<mml:mn>86.04</mml:mn>
<mml:mtext>%</mml:mtext>
</mml:mrow>
</mml:math>
</inline-formula>, and an Area Under the Curve (AUC) of 0.98, while depending on far less supervision and annotation, in comparison to its counterparts.</p>
</abstract>
<kwd-group>
<kwd>capsule networks</kwd>
<kwd>COVID-19</kwd>
<kwd>computed tomography scans</kwd>
<kwd>fully automated classification</kwd>
<kwd>deep learning</kwd>
</kwd-group>
</article-meta>
</front>
<body>
<sec id="s1">
<title>1 Introduction</title>
<p>The recent outbreak of the novel coronavirus infection (COVID-19) has sparked an unforeseeable global crisis since its emergence in late 2019. Resulting COVID-19 pandemic is reshaping our societies and people&#x2019;s lives in many ways and caused more than half a million deaths so far. In spite of the global enterprise to prevent the rapid outbreak of the disease, there are still thousands of reported cases around the world on daily bases, which raised the concern of facing a major second wave of the pandemic. Early diagnosis of COVID-19, therefore, is of paramount importance, to assist health and government authorities with developing efficient resource allocations and breaking the transmission&#x20;chain.</p>
<p>Reverse Transcription Polymerase Chain Reaction (RT-PCR), which is currently the gold standard in diagnosing COVID-19, is time-consuming and prone to high false-negative rate (<xref ref-type="bibr" rid="B13">Fang et&#x20;al., 2020</xref>). Recently, chest Computed Tomography (CT) scans and Chest Radiographs (CR) of COVID-19 patients, have shown specific findings, such as bilateral and peripheral distribution of Ground Glass Opacities (GGO) mostly in the lung lower lobes, and patchy consolidations in some of the cases (<xref ref-type="bibr" rid="B18">Inui et&#x20;al., 2020</xref>). Diffuse distribution, vascular thickening, and fine reticular opacities are other commonly observed features of COVID-19 reported in (<xref ref-type="bibr" rid="B9">Bai et&#x20;al., 2020</xref>; <xref ref-type="bibr" rid="B10">Chung et&#x20;al., 2020</xref>; <xref ref-type="bibr" rid="B23">Ng et&#x20;al., 2020</xref>; <xref ref-type="bibr" rid="B29">Shi et&#x20;al., 2020</xref>). Although imaging studies and their results can be obtained in a timely fashion, such features can be seen in other viral or bacterial infections or other entities such as organizing pneumonia, leading to misclassification even by experienced radiologists.</p>
<p>With the increasing number of people in need of COVID-19 examination, health care professionals are experiencing a heavy workload reducing their concentration to properly diagnose COVID-19 cases and confirm the results. This arises the need to distinguish normal cases and non-COVID infections from COVID-19 cases in a timely fashion to put a higher focus on COVID-19 infected cases. Using deep learning-based algorithms to classify patients into COVID and non-COVID, health care professionals can exclude non-COVID cases quickly in the first step and allow for paying more attention and allocating more medical resources to COVID-19 identified cases. It is worth mentioning that although the RT-PCR, as a non-destructive diagnosis test, is commonly used for COVID-19 detection, in some countries with high number of COVID-19 cases, CT imaging is widely used as the primary detection technique. Therefore, there is an unmet need to develop advanced deep learning-based solutions based on CT images to speed up the diagnosis procedure.</p>
<sec id="s1-1">
<title>1.1 Literature Review</title>
<p>Convolutional Neural Networks (CNNs) have been widely used in several studies to account for the human-centered weaknesses in detecting COVID-19. CNNs are powerful models in related tasks and are capable of extracting distinguishing features from CT scans and chest radiographs (<xref ref-type="bibr" rid="B33">Yamashita et&#x20;al., 2018</xref>). In this regard, many studies have utilized CNNs to identify COVID-19 cases from medical images. The study by (<xref ref-type="bibr" rid="B31">Wang and Wong, 2020</xref>), is an example of the application of CNN in COVID-19 detection, where CNN is first pre-trained on the ImageNet dataset (<xref ref-type="bibr" rid="B19">Krizhevsky et&#x20;al., 2017</xref>). Fine-tuning is then performed using a CR dataset. Results show an accuracy of <inline-formula id="inf4">
<mml:math id="m4">
<mml:mrow>
<mml:mn>93.3</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> in distinguishing normal, non-COVID-19 pneumonia (viral and bacterial), and COVID-19 infection cases (<xref ref-type="bibr" rid="B28">Sethy et&#x20;al., 2020</xref>). have also explored the same problem with the difference that the CNN is followed by a Support Vector Machine (SVM), to identify positive COVID-19 cases. Their obtained results show an overall accuracy of <inline-formula id="inf5">
<mml:math id="m5">
<mml:mrow>
<mml:mn>95.38</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, sensitivity of <inline-formula id="inf6">
<mml:math id="m6">
<mml:mrow>
<mml:mn>97.29</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> and specificity of <inline-formula id="inf7">
<mml:math id="m7">
<mml:mrow>
<mml:mn>93.47</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>. Another study by (<xref ref-type="bibr" rid="B22">Mahmud et&#x20;al., 2020</xref>) proposed a CNN-based model utilizing depth-wise convolutions with varying dilation rates to extract more diversified features from chest radiographs. They used a pre-trained model on a dataset of normal, viral, and bacterial pneumonia patients followed by additional fine-tuned layers on a dataset of COVID-19 and other pneumonia patients, obtaining an overall accuracy of <inline-formula id="inf8">
<mml:math id="m8">
<mml:mrow>
<mml:mn>90.2</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, sensitivity of <inline-formula id="inf9">
<mml:math id="m9">
<mml:mrow>
<mml:mn>89.9</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, and specificity of&#x20;<inline-formula id="inf10">
<mml:math id="m10">
<mml:mrow>
<mml:mn>89.1</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>.</p>
<p>Chest radiograph acquisition is relatively simple with less radiation exposure than CT scans. However, a single CR image fails to incorporate details of infections in the lung and cannot provide a comprehensive view for the lung infection diagnosis. CT scan, on the other hand, is an alternative imaging modality that incorporates the detailed structure of the lung and infected areas. Unlike CR images, CT scans generate cross-sectional images (slices) to create a 3D representation of the body. Consequently, there has been a surge of interest on utilizing 2D and 3D CT images to identify COVID-19 infection. For instance (<xref ref-type="bibr" rid="B34">Yang et&#x20;al., 2020</xref>), proposed a DenseNet-based model to classify manually selected slices with COVID-19 manifestations and pulmonary parenchyma into COVID-19 and normal classes. The underlying study achieved an accuracy of <inline-formula id="inf11">
<mml:math id="m11">
<mml:mrow>
<mml:mn>92</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> on the patient-level classification by averaging slice-level probabilities followed by a threshold of 0.8 on the averaged values. Furthermore, the dataset used to train and test the model does not include other types of pneumonia. <italic>Identified Drawback 1</italic>: Such methods require manual selecting slices demonstrating infection to feed the model, which makes the overall process time-consuming and only partially automated. To extract features from all CT slices (<xref ref-type="bibr" rid="B20">Li et&#x20;al., 2020</xref>), first segmented the lung regions using a U-net based segmentation method (<xref ref-type="bibr" rid="B26">Ronneberger et&#x20;al., 2015</xref>), and then used them to fine-tune a ResNet50 model, which was pre-trained on natural images from the ImageNet dataset (<xref ref-type="bibr" rid="B12">Deng et&#x20;al., 2009</xref>). Extracted features are then combined using a max-pooling operation followed by a fully connected layer to generate probability scores for each disease type. Their proposed model achieved sensitivities of <inline-formula id="inf12">
<mml:math id="m12">
<mml:mrow>
<mml:mn>90</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf13">
<mml:math id="m13">
<mml:mrow>
<mml:mn>87</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, and <inline-formula id="inf14">
<mml:math id="m14">
<mml:mrow>
<mml:mn>94</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> for COVID-19, Community Acquired Pneumonia (CAP), and non-pneumonia cases respectively. <italic>Identified Drawback 2:</italic> Such methods combine extracted features from all slices of a patient, with or without infection, which potentially results in lower accuracy as there are numerous slices without evidence of infection in a volumetric CT scan of an infected patient.</p>
<p>In the study by (<xref ref-type="bibr" rid="B17">Hu et&#x20;al., 2020</xref>), segmented lungs are fed into a multi-scale CNN-based classification model, which utilizes intermediate CNN layers to obtain classification scores, and aggregates scores generated by intermediate layers to make the final prediction. Their proposed method achieves an overall accuracy of <inline-formula id="inf15">
<mml:math id="m15">
<mml:mrow>
<mml:mn>87.4</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> in the three-way classification (<xref ref-type="bibr" rid="B37">Zhang et&#x20;al., 2020</xref>). proposed a two-stage method consisting of a Deeplabv3-based lung-lesion segmentation model (<xref ref-type="bibr" rid="B11">Chen et al., 2017</xref>) followed by a 3D ResNet18 classification model (<xref ref-type="bibr" rid="B14">Hara et&#x20;al., 2017</xref>) to identify lung lesions and abnormalities and use them to classify patients into COVID-19, community acquired pneumonia, and normal findings. They manually annotated chest CT scans into seven regions to train their lung segmentation model, which is a time-consuming and sophisticated task requiring high level of thoracic radiology expertise to accomplish. Their proposed method achieves the overall accuracy of <inline-formula id="inf16">
<mml:math id="m16">
<mml:mrow>
<mml:mn>92.49</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> in both three-way and binary (COVID-19 vs. others) classifications.</p>
</sec>
<sec id="s1-2">
<title>1.2 Problem Statement</title>
<p>At one hand, we aim to address the two identified drawbacks of the aforementioned methods. More specifically, existing solutions either require a precise annotation/labeling of lung images, which is time-consuming and error-prone, especially when we are facing a new and unknown type of disease such as COVID-19, or assign the patient-level label to all the slices. On the other hand, CNN, which is widely adopted in COVID-19 studies, suffers from an important drawback that reduces its reliability in clinical practice. CNNs are required to be trained on different variations of the same object to fully capture the spatial relations and patterns. In other words, CNNs, commonly, fail to recognize an object when it is rotated or transformed. In practice, extensive data augmentation and/or adoption of huge data resources are needed to compensate for the lack of spatial interpretation. As COVID-19 is a relatively new phenomenon, large datasets are not easily accessible, especially due to strict privacy preserving constraints. Furthermore, most COVID-19 cases have been reported with a specific infection distribution in their image (<xref ref-type="bibr" rid="B9">Bai et&#x20;al., 2020</xref>; <xref ref-type="bibr" rid="B10">Chung et&#x20;al., 2020</xref>; <xref ref-type="bibr" rid="B23">Ng et&#x20;al., 2020</xref>; <xref ref-type="bibr" rid="B29">Shi et&#x20;al., 2020</xref>), which makes capturing spatial relations in the image highly important.</p>
</sec>
<sec id="s1-3">
<title>1.3 Contributions</title>
<p>As stated previously, structure of infection spread in the lung for COVID-19 is not yet fully understood given its recent and abrupt emergence. Furthermore, COVID-19 has a particular structure in affecting the lung, therefore, picking up those spatial structures are significantly important. Capsule Networks (CapsNets) (<xref ref-type="bibr" rid="B15">Hinton et&#x20;al., 2018</xref>), in contrast to CNNs, are equipped with routing by agreement process enabling them to capture such spatial patterns. Even without a large dataset, capsules interpret the object instantiation parameters, besides its existence, and by reaching a mutual agreement, higher-level objects are developed from lower-level ones. The superiority of Capsule Networks over their counterparts has been shown in different medial image processing problems (<xref ref-type="bibr" rid="B4">Afshar et&#x20;al., 2018</xref>; <xref ref-type="bibr" rid="B7">Afshar et&#x20;al., 2019a</xref>; <xref ref-type="bibr" rid="B8">Afshar et&#x20;al., 2019b</xref>; <xref ref-type="bibr" rid="B5">Afshar et&#x20;al., 2020b</xref>; <xref ref-type="bibr" rid="B6">Afshar et&#x20;al., 2020d</xref>; <xref ref-type="bibr" rid="B3">Afshar et&#x20;al., 2020c</xref>). Recently, we proposed a Capsule Network-based framework (<xref ref-type="bibr" rid="B2">Afshar et&#x20;al., 2020a</xref>), referred to as the COVID-CAPS, to identify COVID-19 cases from chest radiographs, which achieved an accuracy of <inline-formula id="inf17">
<mml:math id="m17">
<mml:mrow>
<mml:mn>98.3</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, a specificity of <inline-formula id="inf18">
<mml:math id="m18">
<mml:mrow>
<mml:mn>98.6</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, and a sensitivity of <inline-formula id="inf19">
<mml:math id="m19">
<mml:mrow>
<mml:mn>80</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>. As stated previously, CT imaging is superior for COVID-19 detection and diagnosis purposes when compared to chest radiographs. However, as in the case of CT imaging, we are dealing with 3D inputs and several slices per patient (compared to one chest radiograph per patient), the learning process is significantly more challenging. As such, accuracies of deep models trained over CT scans cannot be directly compared with those obtained based on chest radiographs.</p>
<p>Following our previous study on chest radiographs, in the present study, we take one step forward and propose a fully automated two-stage Capsule Network-based framework, referred to as the COVID-FACT, to identify COVID-19 patients using chest CT images. Based on our in-house dataset, COVID-FACT achieves an accuracy of <inline-formula id="inf20">
<mml:math id="m20">
<mml:mrow>
<mml:mn>90.82</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, sensitivity of <inline-formula id="inf21">
<mml:math id="m21">
<mml:mrow>
<mml:mn>94.55</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, specificity of <inline-formula id="inf22">
<mml:math id="m22">
<mml:mrow>
<mml:mn>86.04</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, and Area Under the Curve (AUC) of 0.98. We developed two variants of the COVID-FACT, one of which is fed with the whole chest CT image, while the other one utilizes the segmented lung area as the input. In the latter case, instead of using an original chest CT image, first a segmentation model (<xref ref-type="bibr" rid="B16">Hofmanninger et&#x20;al., 2020</xref>) is applied to extract the lung region, which is then provided as input to the COVID-FACT. This will be further clarified in <xref ref-type="sec" rid="s3">Section 3</xref>. Experimental results show that the model coupled with lung segmentation achieves the same overall accuracy compared to the other COVID-FACT variation working with original images. However, using the segmented lung regions increases the sensitivity and AUC from <inline-formula id="inf23">
<mml:math id="m23">
<mml:mrow>
<mml:mn>92.72</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> and 0.95 to <inline-formula id="inf24">
<mml:math id="m24">
<mml:mrow>
<mml:mn>94.55</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> and 0.98, respectively, while slightly decreasing the specificity from <inline-formula id="inf25">
<mml:math id="m25">
<mml:mrow>
<mml:mn>88.37</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> to <inline-formula id="inf26">
<mml:math id="m26">
<mml:mrow>
<mml:mn>86.04</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>.</p>
<p>COVID-FACT benefits from a two-stage design, which is of paramount importance in COVID-19 detection using CT scans, as a CT examination is typically associated with hundreds of slices that cannot be analyzed at once. At the first stage, the proposed COVID-FACT detects slices demonstrating infection in a 3D volumetric CT scan to be analyzed and classified at the next stage. At the second stage, candidate slices detected at the previous stage are classified into COVID and non-COVID (community acquired pneumonia and normal) cases and a voting mechanism is applied to generate the classification scores in the patient level. COVID-FACT&#x2019;s two-stage architecture has the advantage of being trained by even weakly labeled dataset, as errors at the first stage can be compensated at the second stage. As a result, COVID-FACT does not require any infection annotation or a very precise slice labeling, which is a valuable asset due to the limited knowledge and experience on the novel COVID-19 disease. In fact, manual annotation is completely removed from the COVID-FACT. The only information required from the radiologists to train the first stage is the slices containing evidence of infection. In other words, COVID-FACT is not dependent on the manual delineation of specific infected regions in the slices, which is a complicated and time-consuming task compared to only identifying slices with the evidence of infection. This issue is more critical in the case of a novel disease such as COVID-19, which requires comprehensive research to identify the disease manifestations. It is worth noting that the pre-trained lung segmentation model used as the pre-processing step in our study is related to the well-studied lung segmentation task, which is totally different from the infection segmentation. As a final note, we would like to mention that the radiologist&#x2019;s input is not required in the test phase of the COVID-FACT and the trained framework is fully automated.</p>
<p>The reminder of the paper is organized as follows: <xref ref-type="sec" rid="s2">Section 2</xref> describes the dataset and imaging protocol used in this study. <xref ref-type="sec" rid="s3">Section 3</xref> presents a brief description of Capsule Networks and explains the proposed COVID-FACT in details. Experimental results and model evaluation are presented in <xref ref-type="sec" rid="s4">Section 4</xref>. Finally, <xref ref-type="sec" rid="s5">Section 5</xref> concludes the&#x20;work.</p>
</sec>
</sec>
<sec id="s2">
<title>2 Materials and Equipment</title>
<p>In this section, we will explain the in-house dataset used in this study, along with the associated imaging protocol.</p>
<sec id="s2-1">
<title>2.1 Dataset</title>
<p>The dataset used in this study, referred to as the &#x201c;COVID-CT-MD&#x201d; <xref ref-type="bibr" rid="B1">Afshar et&#x20;al. (2021)</xref>, contains volumetric chest CT scans of 171 patients positive for COVID-19 infection, 60 patients with Community Acquired Pneumonia (CAP), and 76 normal patients acquired from April 2018 to May 2020. The average age of patients is <inline-formula id="inf27">
<mml:math id="m27">
<mml:mrow>
<mml:mn>50</mml:mn>
<mml:mo>&#xb1;</mml:mo>
<mml:mn>16</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> including 183 men and 124 women. This dataset and the related annotations are publicly available through Figshare at <ext-link ext-link-type="uri" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="https://figshare.com/s/c20215f3d42c98f09ad0">https://figshare.com/s/c20215f3d42c98f09ad0</ext-link>.</p>
<p>Diagnosis of COVID-19 infection is based on positive real-time reverse transcription polymerase chain reaction (rRT-PCR) test results, clinical parameters, and CT scan manifestations by a thoracic radiologist, with 20&#xa0;years of experience in thoracic imaging. CAP and normal cases were included from another study and the diagnosis was confirmed using clinical parameters, and CT scans. A subset of 55&#x20;COVID-19, and 25 community acquired pneumonia cases were analyzed by the radiologist to identify and label slices with evidence of infection as shown in <xref ref-type="fig" rid="F1">Figure&#x20;1</xref>. This labeling process focuses more on distinctive manifestations rather than slices with minimal findings. The labeled subset of the data contains <inline-formula id="inf28">
<mml:math id="m28">
<mml:mrow>
<mml:mn>4,962</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> number of slices demonstrating infection and <inline-formula id="inf29">
<mml:math id="m29">
<mml:mrow>
<mml:mn>18,447</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> number of slices without infection. The data is then used to train and validate the first stage of our proposed COVID-FACT model to extract slices demonstrating infection from volumetric CT scans to be used in the second classification stage. We have randomly divided this subset into three separate components for training, validation, and testing. <inline-formula id="inf30">
<mml:math id="m30">
<mml:mrow>
<mml:mn>60</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> of the labeled data is used for training, <inline-formula id="inf31">
<mml:math id="m31">
<mml:mrow>
<mml:mn>10</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> for validation, and <inline-formula id="inf32">
<mml:math id="m32">
<mml:mrow>
<mml:mn>30</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> for the test. The unlabeled subset is also randomly divided with the same proportion and used along with the labeled data to develop the second stage of the COVID-FACT model and evaluate the overall method. The data leakage between the train and test sets has been prevented. In other words, all slices related to a patient are included either in the train or the test dataset. This research work is performed based on the policy certification number 30013394 of Ethical acceptability for secondary use of medical data approved by Concordia University. Furthermore, informed consent is obtained from all the patients. Finally, the dataset is complied with the DICOM supplement 142 (Clinical Trial De-identification Profiles) <xref ref-type="bibr" rid="B35">DICOM Standards Committee, Working Group 18 Clinical Trials (2011)</xref>, indicating that all CT studies are de-identified by either removing or obfuscating the patient and center-related information such as names, UIDs, dates, times, and comments based on the directions specified in <xref ref-type="bibr" rid="B35">DICOM Standards Committee, Working Group 18 Clinical Trials (2011)</xref>.</p>
<fig id="F1" position="float">
<label>FIGURE 1</label>
<caption>
<p>(<bold>A</bold>,<bold>B)</bold>: Infected and non-infected sample slices in a COVID-19 case; <bold>(C</bold>,<bold>D)</bold>: Infected and non-infected sample slices in a non-COVID Pneumonia&#x20;case.</p>
</caption>
<graphic xlink:href="frai-04-598932-g001.tif"/>
</fig>
</sec>
<sec id="s2-2">
<title>2.2 Imaging Protocol</title>
<p>All CT examinations have been acquired using a single CT scanner with the same acquisition setting and technical parameters, which are presented in <xref ref-type="table" rid="T1">Table&#x20;1</xref>, where kVP (kiloVoltage Peak) and Exposure Time affect the radiation exposure dose, while Slice Thickness and Reconstruction Matrix represent the axial resolution and output size of the images, respectively <xref ref-type="bibr" rid="B25">Raman et&#x20;al. (2013)</xref>. Next, we describe the proposed COVID-FACT framework followed by the experimental results.</p>
<table-wrap id="T1" position="float">
<label>TABLE 1</label>
<caption>
<p>Imaging device and settings used to acquire the in-house dataset.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left">Scanner manufacturer and model</th>
<th align="center">Slice thickness (mm)</th>
<th align="center">Image type</th>
<th align="center">kVP (kV)</th>
<th align="center">Exposure time (ms)</th>
<th align="center">Reconstruction matrix</th>
<th align="center">Window center</th>
<th align="left">Window width</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">SIEMENS, SOMATOM scope</td>
<td align="char" char=".">2</td>
<td align="center">Axial</td>
<td align="char" char=".">110</td>
<td align="char" char=".">600</td>
<td align="center">
<inline-formula id="inf33">
<mml:math id="m33">
<mml:mrow>
<mml:mn>512</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>512</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf34">
<mml:math id="m34">
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mn>50</mml:mn>
<mml:mo>,</mml:mo>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>600</mml:mn>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="left">
<inline-formula id="inf35">
<mml:math id="m35">
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mn>350,1200</mml:mn>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
</sec>
<sec sec-type="methods" id="s3">
<title>3 Methods</title>
<p>The COVID-FACT framework is developed to automatically distinguish COVID-19 cases from other types of pneumonia and normal cases using volumetric chest CT scans. It utilizes a lung segmentation model at a pre-processing step to segment lung regions and pass them as the input to the two-stage Capsule Network-based classifier. The first stage of the COVID-FACT extracts slices demonstrating infection in a CT scan, while the second stage uses the detected slices in first stage to classify patients into COVID-19 and non-COVID cases. Finally, the Gradient-weighted Class Activation Mapping (Grad-CAM) localization approach (<xref ref-type="bibr" rid="B27">Selvaraju et&#x20;al., 2017</xref>) is incorporated into the model to highlight important components of a chest CT scan, that contribute the most to the final decision.</p>
<p>In this section, different components of the proposed COVID-FACT are explained. First, Capsule Network, which is the main building block of our proposed approach, is briefly introduced. Then the lung segmentation method is described, followed by the details related to the first and second stages of the COVID-FACT architecture. Finally, the Grad-CAM localization mapping approach is presented.</p>
<sec id="s3-1">
<title>3.1 Capsule Networks</title>
<p>A Capsule Network (CapsNet) is an alternative architecture for CNNs with the advantage of capturing hierarchical and spatial relations between image instances. Each Capsule layer utilizes several capsules to determine existence probability and pose of image instances using an instantiation vector. The length of the vector represents the existence probability and the orientation determines the pose. Each Capsule <italic>i</italic> consists of a set of neurons, which collectively create the instantiation vector <inline-formula id="inf36">
<mml:math id="m36">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">u</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> for the associated instance. Capsules in lower layers try to predict the output of Capsules in higher levels using a trainable weight matrix <inline-formula id="inf37">
<mml:math id="m37">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">W</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> as follows<disp-formula id="equ1">
<mml:math id="m38">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mi>u</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mo>&#x7c;</mml:mo>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold-italic">W</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mi mathvariant="bold-italic">u</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(1)</label>
</disp-formula>where <inline-formula id="inf38">
<mml:math id="m39">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mi mathvariant="bold-italic">u</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mo>&#x7c;</mml:mo>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> is the predicted output of Capsule <italic>j</italic> in the next layer by the Capsule <italic>i</italic> in the lower layer. The association between the prediction <inline-formula id="inf39">
<mml:math id="m40">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mi mathvariant="bold-italic">u</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mo>&#x7c;</mml:mo>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> and the actual output of Capsule <italic>j</italic>, denoted by <inline-formula id="inf40">
<mml:math id="m41">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">v</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, is determined by taking the inner product of <inline-formula id="inf41">
<mml:math id="m42">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mi mathvariant="bold-italic">u</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mo>&#x7c;</mml:mo>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf42">
<mml:math id="m43">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">v</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>. The higher the inner product, the more contribution of the lower level capsules to the higher level one. The contribution of Capsule <italic>i</italic> to the output of the Capsule <italic>j</italic> in the next layer is determined by a coupling coefficient <inline-formula id="inf43">
<mml:math id="m44">
<mml:mrow>
<mml:msub>
<mml:mi>c</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, trained over a course of few iterations known as the &#x201c;Routing by Agreement&#x201d; given by<disp-formula id="equ2">
<mml:math id="m45">
<mml:mrow>
<mml:msub>
<mml:mi>a</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold-italic">v</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
<mml:mo>.</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mi mathvariant="bold-italic">u</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mo>&#x7c;</mml:mo>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(2)</label>
</disp-formula>
<disp-formula id="equ3">
<mml:math id="m46">
<mml:mrow>
<mml:msub>
<mml:mi>b</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mi>b</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mi>a</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(3)</label>
</disp-formula>
<disp-formula id="equ4">
<mml:math id="m47">
<mml:mrow>
<mml:msub>
<mml:mi>c</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mi>exp</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>b</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:msub>
<mml:mo>&#x2211;</mml:mo>
<mml:mi>k</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mi>exp</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>b</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>k</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mstyle>
</mml:mrow>
</mml:mfrac>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(4)</label>
</disp-formula>
<disp-formula id="equ5">
<mml:math id="m48">
<mml:mrow>
<mml:msub>
<mml:mi>s</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:munder>
<mml:mstyle displaystyle="true">
<mml:mo>&#x2211;</mml:mo>
</mml:mstyle>
<mml:mi>i</mml:mi>
</mml:munder>
<mml:msub>
<mml:mi>c</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mi>u</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mo>&#x7c;</mml:mo>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(5)</label>
</disp-formula>and<disp-formula id="e6">
<mml:math id="m49">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">v</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mrow>
<mml:mo>&#x2016;</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">s</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>&#x2016;</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mn>2</mml:mn>
</mml:msup>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2b;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mrow>
<mml:mo>&#x2016;</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">s</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>&#x2016;</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mn>2</mml:mn>
</mml:msup>
</mml:mrow>
</mml:mfrac>
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">s</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:mrow>
<mml:mo>&#x2016;</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">s</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>&#x2016;</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mfrac>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(6)</label>
</disp-formula>where <inline-formula id="inf44">
<mml:math id="m50">
<mml:mrow>
<mml:msub>
<mml:mi>a</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> is referred to as the agreement coefficient between the prediction and actual output, and <inline-formula id="inf45">
<mml:math id="m51">
<mml:mrow>
<mml:msub>
<mml:mi>b</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> denotes the log prior of the coupling coefficient <inline-formula id="inf46">
<mml:math id="m52">
<mml:mrow>
<mml:msub>
<mml:mi>c</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>. Vector <inline-formula id="inf47">
<mml:math id="m53">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">s</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> denotes the Capsule output before applying the squashing function. As the length of output vectors represents probabilities, the ultimate output of Capsule <italic>j</italic> (<inline-formula id="inf48">
<mml:math id="m54">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">v</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>) is obtained by squashing <inline-formula id="inf49">
<mml:math id="m55">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">s</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> between 0 and 1 using the squashing function defined in <xref ref-type="disp-formula" rid="e6">Eq. 6</xref>. In order to update weight matrix <inline-formula id="inf50">
<mml:math id="m56">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">W</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> through a backward training process, the loss function is calculated for each Capsule <italic>k</italic> as follows<disp-formula id="equ6">
<mml:math id="m57">
<mml:mrow>
<mml:msub>
<mml:mi>l</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mi>T</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
<mml:mi>max</mml:mi>
<mml:msup>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mn>0</mml:mn>
<mml:mo>,</mml:mo>
<mml:msup>
<mml:mi>m</mml:mi>
<mml:mo>&#x2b;</mml:mo>
</mml:msup>
<mml:mo>&#x2212;</mml:mo>
<mml:mo>&#x7c;</mml:mo>
<mml:mo>&#x7c;</mml:mo>
<mml:msub>
<mml:mi>v</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
<mml:mo>&#x7c;</mml:mo>
<mml:mo>&#x7c;</mml:mo>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mn>2</mml:mn>
</mml:msup>
<mml:mo>&#x2b;</mml:mo>
<mml:mi>&#x3bb;</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mi>T</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mi>max</mml:mi>
<mml:msup>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mn>0</mml:mn>
<mml:mo>,</mml:mo>
<mml:mo>&#x7c;</mml:mo>
<mml:mo>&#x7c;</mml:mo>
<mml:msub>
<mml:mi>v</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
<mml:mo>&#x7c;</mml:mo>
<mml:mo>&#x7c;</mml:mo>
<mml:mo>&#x2212;</mml:mo>
<mml:msup>
<mml:mi>m</mml:mi>
<mml:mo>&#x2212;</mml:mo>
</mml:msup>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mn>2</mml:mn>
</mml:msup>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(7)</label>
</disp-formula>where <inline-formula id="inf51">
<mml:math id="m58">
<mml:mrow>
<mml:msub>
<mml:mi>T</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> is 1 when the class <italic>k</italic> is present and 0 otherwise. <inline-formula id="inf52">
<mml:math id="m59">
<mml:mrow>
<mml:msup>
<mml:mi>m</mml:mi>
<mml:mo>&#x2b;</mml:mo>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf53">
<mml:math id="m60">
<mml:mrow>
<mml:msup>
<mml:mi>m</mml:mi>
<mml:mo>&#x2212;</mml:mo>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula>, and &#x3bb; are hyper parameters of the model and are originally set to 0.9, 0.1, and 0.5, respectively. The overall loss is the summation of all losses calculated for all Capsules.</p>
</sec>
<sec id="s3-2">
<title>3.2 Proposed COVID-FACT</title>
<p>The overall architecture of the COVID-FACT is illustrated in <xref ref-type="fig" rid="F2">Figure&#x20;2</xref>, which consists of a lung segmentation model at the beginning followed by two Capsule Network-based models and an average voting mechanism coupled with a thresholding approach to generate patient-level classification results. The three components of the COVID-FACT are as follows:<list list-type="simple">
<list-item>
<p>&#x2022; Lung Segmentation: The input of the COVID-FACT is the segmented lung regions identified by a U-net based segmentation model (<xref ref-type="bibr" rid="B16">Hofmanninger et&#x20;al., 2020</xref>), referred to as the &#x201c;U-net (R231CovidWeb)&#x201d;, which has been initially trained on a large and diverse dataset including multiple pulmonary diseases, and fine-tuned on a small dataset of COVID-19 images. The Input of the U-net (R231CovidWeb) model is a single slice with the size of <inline-formula id="inf54">
<mml:math id="m61">
<mml:mrow>
<mml:mn>512</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>512</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>. The output is the lung tissues, which will be normalized between 0 and 1 to generalize the features and help the model to perform more effectively. Following the literature (<xref ref-type="bibr" rid="B17">Hu et&#x20;al., 2020</xref>; <xref ref-type="bibr" rid="B37">Zhang et&#x20;al., 2020</xref>), we down-sampled the output from <inline-formula id="inf55">
<mml:math id="m62">
<mml:mrow>
<mml:mn>512</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>512</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> to <inline-formula id="inf56">
<mml:math id="m63">
<mml:mrow>
<mml:mn>256</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>256</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> size to reduce the complexity and memory allocation without losing any significant information. Slices with no detected lung regions are removed and the remaining are fed into the first stage of the COVID-FACT&#x20;model.</p>
</list-item>
<list-item>
<p>&#x2022; COVID-FACT&#x2019;s Stage One: The first stage of the COVID-FACT, shown in <xref ref-type="fig" rid="F3">Figure&#x20;3</xref> is responsible to identify slices demonstrating infection (by COVID-19 or other types of pneumonia). Using this stage, we discard slices without infection and focus only on the ones with infection. Intuitively speaking, this process is similar in nature to the way that radiologists analyze a CT scan. When radiologists review a CT scan containing numerous consecutive cross-sectional slices of the body, they identify the slices with an abnormality in the first step, and analyze the abnormal ones to diagnose the disease in the next step. Existing CT-based deep learning processing methods either use all slices as a 3D input to a classifier, or classify individual slices and transform slice-level predictions to the patient-level ones using a threshold on the entire slices (<xref ref-type="bibr" rid="B24">Rahimzadeh et al., 2021</xref>). Determining a threshold on the number or percentage of slices demonstrating infection over the entire slices is not precise, as most pulmonary infections have different stages with involvement of different lung regions (<xref ref-type="bibr" rid="B36">Yu et&#x20;al., 2020</xref>). Furthermore, a CT scan may contain different number of slices depending on the acquisition setting, which makes it impossible to find such a threshold. In most methods passing all slices as a 3D input to the model, the input size is fixed and the model is trained to assign higher scores to slices demonstrating infection. However, the performance of such models will be reduced when testing on a dataset other than the dataset on which they are originally trained (<xref ref-type="bibr" rid="B37">Zhang et&#x20;al., 2020</xref>).</p>
</list-item>
</list>
</p>
<fig id="F2" position="float">
<label>FIGURE 2</label>
<caption>
<p>The two-stage architecture of the proposed COVID-FACT.</p>
</caption>
<graphic xlink:href="frai-04-598932-g002.tif"/>
</fig>
<fig id="F3" position="float">
<label>FIGURE 3</label>
<caption>
<p>Architecture of the COVID-FACT at stage&#x20;one.</p>
</caption>
<graphic xlink:href="frai-04-598932-g003.tif"/>
</fig>
<p>The model used in stage one of the proposed COVID-FACT is adapted from the COVID-CAPS model presented in our previous work (<xref ref-type="bibr" rid="B2">Afshar et&#x20;al., 2020a</xref>), which was developed to identify COVID cases from chest radiographs. The first stage consists of four convolutional layers and three capsule layers. The first and second layers are convolutional ones followed by a batch-normalization. Similarly, the third and fourth layers are convolutional ones followed by a max-pooling layer. The fourth layer, referred to as the primary Capsule layer, is reshaped to form the desired primary capsules. Afterwards, three capsule layers perform sequential routing processes. Finally, the last Capsule layer represents two classes of infected and non-infected slices. The input of stage one is set of CT slices corresponding to a patient, and the output is slices of the volumetric CT scan demonstrating infection. The output of stage one may vary in size for each patient due to different areas of lung involvement and phase of infection.</p>
<p>In order to cope with our imbalanced training dataset, we modified the loss function, so that a higher penalty rate is given to the false positive (infected slices) cases. The loss function is modified as follows<disp-formula id="e8">
<mml:math id="m64">
<mml:mrow>
<mml:mi>l</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>s</mml:mi>
<mml:mi>s</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:msup>
<mml:mi>N</mml:mi>
<mml:mo>&#x2b;</mml:mo>
</mml:msup>
</mml:mrow>
<mml:mrow>
<mml:msup>
<mml:mi>N</mml:mi>
<mml:mo>&#x2b;</mml:mo>
</mml:msup>
<mml:mo>&#x2b;</mml:mo>
<mml:msup>
<mml:mi>N</mml:mi>
<mml:mo>&#x2212;</mml:mo>
</mml:msup>
</mml:mrow>
</mml:mfrac>
<mml:mo>&#xd7;</mml:mo>
<mml:mi>l</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>s</mml:mi>
<mml:msup>
<mml:mi>s</mml:mi>
<mml:mo>&#x2212;</mml:mo>
</mml:msup>
<mml:mo>&#x2b;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:msup>
<mml:mi>N</mml:mi>
<mml:mo>&#x2212;</mml:mo>
</mml:msup>
</mml:mrow>
<mml:mrow>
<mml:msup>
<mml:mi>N</mml:mi>
<mml:mo>&#x2b;</mml:mo>
</mml:msup>
<mml:mo>&#x2b;</mml:mo>
<mml:msup>
<mml:mi>N</mml:mi>
<mml:mo>&#x2212;</mml:mo>
</mml:msup>
</mml:mrow>
</mml:mfrac>
<mml:mo>&#xd7;</mml:mo>
<mml:mi>l</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>s</mml:mi>
<mml:msup>
<mml:mi>s</mml:mi>
<mml:mo>&#x2b;</mml:mo>
</mml:msup>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(8)</label>
</disp-formula>where <inline-formula id="inf57">
<mml:math id="m65">
<mml:mrow>
<mml:msup>
<mml:mi>N</mml:mi>
<mml:mo>&#x2b;</mml:mo>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> is the number of positive samples, <inline-formula id="inf58">
<mml:math id="m66">
<mml:mrow>
<mml:msup>
<mml:mi>N</mml:mi>
<mml:mo>&#x2212;</mml:mo>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> is the number of negative samples, <inline-formula id="inf59">
<mml:math id="m67">
<mml:mrow>
<mml:mi>l</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>s</mml:mi>
<mml:msup>
<mml:mi>s</mml:mi>
<mml:mo>&#x2b;</mml:mo>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> denotes the loss associated with positive samples, and <inline-formula id="inf60">
<mml:math id="m68">
<mml:mrow>
<mml:mi>l</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>s</mml:mi>
<mml:msup>
<mml:mi>s</mml:mi>
<mml:mo>&#x2212;</mml:mo>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> denotes the loss associated with negative samples.<list list-type="simple">
<list-item>
<p>&#x2022; COVID-FACT&#x2019;s Stage Two: As mentioned earlier, we need to apply classification methods on a subset of slices demonstrating infection rather than on the entire slices in a CT scan. It is worth noting that, lung segmentation (i.e.,&#x20;extracting lung tissues) is performed in one of the variants of the COVID-FACT as a pre-processing step. The first stage of the COVID-FACT, on the other hand, is tasked with this specific issue of extracting slices demonstrating infections.</p>
</list-item>
</list>
</p>
<p>The second stage of the COVID-FACT takes candidate slices of a patient detected in stage one as the input, and classifies them into one of COVID-19 or non-COVID (including normal and pneumonia) classes, i.e.,&#x20;we consider a binary classification problem. Stage two is a stack of four convolutional and two capsule layers shown in <xref ref-type="fig" rid="F4">Figure&#x20;4</xref>. The output of the last capsule indicates classification probabilities in the slice-level. An average voting function is applied to the classification probabilities, in order to aggregate slice-level values and find the patient-level predictions as follows<disp-formula id="equ7">
<mml:math id="m69">
<mml:mrow>
<mml:mi>P</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
<mml:mo>&#x2208;</mml:mo>
<mml:mi>c</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mrow>
<mml:msub>
<mml:mi>L</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mfrac>
<mml:mstyle displaystyle="true">
<mml:munderover>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mi>L</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
</mml:mrow>
</mml:munderover>
<mml:mrow>
<mml:mi>P</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msubsup>
<mml:mi>s</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>k</mml:mi>
</mml:msubsup>
<mml:mo>&#x2208;</mml:mo>
<mml:mi>c</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mstyle>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(9)</label>
</disp-formula>where <inline-formula id="inf61">
<mml:math id="m70">
<mml:mrow>
<mml:mi>P</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
<mml:mo>&#x2208;</mml:mo>
<mml:mi>c</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> refers to the probability that patient <italic>k</italic> belongs to the target class <italic>c</italic> (e.g., COVID-19), <inline-formula id="inf62">
<mml:math id="m71">
<mml:mrow>
<mml:msub>
<mml:mi>L</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> is the total number of slices detected in stage one for patient <italic>k</italic>, and <inline-formula id="inf63">
<mml:math id="m72">
<mml:mrow>
<mml:mi>P</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msubsup>
<mml:mi>s</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>k</mml:mi>
</mml:msubsup>
<mml:mo>&#x2208;</mml:mo>
<mml:mi>c</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> refers to the probability that the <inline-formula id="inf64">
<mml:math id="m73">
<mml:mrow>
<mml:msup>
<mml:mi>i</mml:mi>
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mi>h</mml:mi>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> slice detected for patient <italic>k</italic> belongs to the target class <italic>c</italic>. It is worth noting that while, initially, the COVID-FACT performs slice-level classification in its second stage, the output is patient-level classification (through its voting mechanism), which is on par with other works that COVID-FACT is compared with. As a final note to our discussion, we would like to add that, corona virus infection is, typically, distributed across the lung volume as such manifests itself in several CT slices. Therefore, having a single slice identified as COVID-19 infection can not necessarily lead to a positive COVID-19 detection.</p>
<fig id="F4" position="float">
<label>FIGURE 4</label>
<caption>
<p>Architecture of the COVID-FACT at stage&#x20;two.</p>
</caption>
<graphic xlink:href="frai-04-598932-g004.tif"/>
</fig>
<p>Similar to stage one, the loss function modification in <xref ref-type="disp-formula" rid="e8">Eq. 8</xref> is used in the training phase of Stage two. The default cut-off probability of 0.5 is chosen in Stage two to distinguish COVID-19 and non-COVID cases. However, it is worth mentioning that the main concern in the clinical practice is to have a high sensitivity in identifying COVID-19 positive patients, even if the specificity is not very high. As such, the classification cut-off probability can be modified by physicians using the ROC curve shown in <xref ref-type="fig" rid="F5">Figure&#x20;5</xref> in order to provide a desired balance between the sensitivity and the specificity (e.g., having a high sensitivity while the specificity is also satisfying). In other words, physicians can decide how much certainty is required to consider a CT scan as a COVID-19 positive case. By choosing a cut-off value higher than 0.5, we can exclude those community acquired pneumonia cases that contain highly overlapped features with COVID-19 cases. On the other hand, by selecting a lower cut-off value, we will allow more cases to be identified as a COVID-19&#x20;case.</p>
<fig id="F5" position="float">
<label>FIGURE 5</label>
<caption>
<p>ROC curve of the proposed COVID-FACT.</p>
</caption>
<graphic xlink:href="frai-04-598932-g005.tif"/>
</fig>
<p>To further improve the ability of the proposed COVID-FACT model to distinguish COVID-19 and non-COVID cases and attenuate effects of errors in the first stage, we classify all patients with less than 3% of slices demonstrating infection in the entire volume as a non-COVID case. These cases are more likely normal cases without any slices with infection. The few slices with infection identified for these cases might be due to the model error in the first stage, non-infectious abnormalities such as pulmonary fibrosis, or motion artifacts in the original images, which will be covered by this threshold. Based on (<xref ref-type="bibr" rid="B36">Yu et&#x20;al., 2020</xref>), it can be interpreted that 4% lung involvement is the minimum percentage for COVID-19 positive cases. In addition, the minimum percentage of slices demonstrating infection detected by the radiologist in our dataset is 7%, and therefore 3% would be a safe threshold to prevent mis-classifying infected cases as normal.</p>
<p>As a final note, it is worth mentioning that the role of Stage 1 is critical to achieving a fully automated framework, which does not require any input from the radiologists, especially when an early and fast diagnosis is desired. However, the COVID-FACT framework is completely flexible and Stage 1 can be skipped if the slices demonstrating infections have already been identified by the radiologists, meaning that the normal cases are already identified in this case and Stage 2 merely separates COVID-19 and CAP cases.<list list-type="simple">
<list-item>
<p>&#x2022; Grad-CAM: Using the Grad-CAM approach, we can visually verify the relation between the model&#x2019;s prediction and the features extracted by the intermediate convolutional layers, which ultimately leads to a higher level of interpretability of the model. Grad-CAM&#x2019;s outcome is a weighted average of the feature maps of a convolutional layer, followed by a Rectified Linear Unit (ReLU) activation function,&#x20;i.e.,</p>
</list-item>
</list>
<disp-formula id="equ8">
<mml:math id="m74">
<mml:mrow>
<mml:msubsup>
<mml:mi>L</mml:mi>
<mml:mrow>
<mml:mi>G</mml:mi>
<mml:mi>r</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>d</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>C</mml:mi>
<mml:mi>A</mml:mi>
<mml:mi>M</mml:mi>
</mml:mrow>
<mml:mi>c</mml:mi>
</mml:msubsup>
<mml:mo>&#x3d;</mml:mo>
<mml:mi mathvariant="normal">Re</mml:mi>
<mml:mi>l</mml:mi>
<mml:mi>U</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:munder>
<mml:mstyle displaystyle="true">
<mml:mo>&#x2211;</mml:mo>
</mml:mstyle>
<mml:mi>k</mml:mi>
</mml:munder>
<mml:msubsup>
<mml:mi>&#x3b1;</mml:mi>
<mml:mi>k</mml:mi>
<mml:mi>c</mml:mi>
</mml:msubsup>
<mml:msup>
<mml:mi>A</mml:mi>
<mml:mi>k</mml:mi>
</mml:msup>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(10)</label>
</disp-formula>where <inline-formula id="inf65">
<mml:math id="m75">
<mml:mrow>
<mml:msubsup>
<mml:mi>L</mml:mi>
<mml:mrow>
<mml:mi>G</mml:mi>
<mml:mi>r</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>d</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>C</mml:mi>
<mml:mi>A</mml:mi>
<mml:mi>M</mml:mi>
</mml:mrow>
<mml:mi>c</mml:mi>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula> refers to the Grad-CAM&#x2019;s output for the target class <italic>c</italic>; <inline-formula id="inf66">
<mml:math id="m76">
<mml:mrow>
<mml:msubsup>
<mml:mi>&#x3b1;</mml:mi>
<mml:mi>k</mml:mi>
<mml:mi>c</mml:mi>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula> is the importance weight for the feature map <italic>k</italic> and the target class <italic>c</italic>, and; <inline-formula id="inf67">
<mml:math id="m77">
<mml:mrow>
<mml:msup>
<mml:mi>A</mml:mi>
<mml:mi>k</mml:mi>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> refers to the feature map <italic>k</italic> of a convolutional layer. The weights <inline-formula id="inf68">
<mml:math id="m78">
<mml:mrow>
<mml:msubsup>
<mml:mi>&#x3b1;</mml:mi>
<mml:mi>k</mml:mi>
<mml:mi>c</mml:mi>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula> are obtained based on the gradients of the probability score of the target class with respect to an intermediate convolutional layer followed by a global average pooling function as follows<disp-formula id="equ9">
<mml:math id="m79">
<mml:mrow>
<mml:msubsup>
<mml:mi>&#x3b1;</mml:mi>
<mml:mi>k</mml:mi>
<mml:mi>c</mml:mi>
</mml:msubsup>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mi>Z</mml:mi>
</mml:mfrac>
<mml:munder>
<mml:mstyle displaystyle="true">
<mml:mo>&#x2211;</mml:mo>
</mml:mstyle>
<mml:mi>i</mml:mi>
</mml:munder>
<mml:munder>
<mml:mstyle displaystyle="true">
<mml:mo>&#x2211;</mml:mo>
</mml:mstyle>
<mml:mi>j</mml:mi>
</mml:munder>
<mml:mfrac>
<mml:mrow>
<mml:mo>&#x2202;</mml:mo>
<mml:msup>
<mml:mi>y</mml:mi>
<mml:mi>c</mml:mi>
</mml:msup>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2202;</mml:mo>
<mml:msubsup>
<mml:mi>A</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
<mml:mi>k</mml:mi>
</mml:msubsup>
</mml:mrow>
</mml:mfrac>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(11)</label>
</disp-formula>where <inline-formula id="inf69">
<mml:math id="m80">
<mml:mrow>
<mml:msup>
<mml:mi>y</mml:mi>
<mml:mi>c</mml:mi>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> is the prediction value (probability) for target class <italic>c</italic>, and <italic>Z</italic> refers to the total number of feature maps in the convolutional&#x20;layer.</p>
</sec>
</sec>
<sec id="s4">
<title>4 Experimental Results</title>
<p>The proposed COVID-FACT is tested on the in-house dataset described earlier in <xref ref-type="sec" rid="s2">Section 2</xref>. The testing set contains 53&#x20;COVID-19 and 43&#x20;non-COVID cases (including 19 community acquired pneumonia and 24 normal cases). We used the Adam optimizer with the initial learning rate of <inline-formula id="inf70">
<mml:math id="m81">
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mi>e</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>4</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, batch size of 16, and 100 epochs. The model with the minimum loss value on the validation set was selected to evaluate the performance of the model on the test set. The proposed COVID-FACT method achieved an accuracy of <inline-formula id="inf71">
<mml:math id="m82">
<mml:mrow>
<mml:mn>90.82</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, sensitivity of <inline-formula id="inf72">
<mml:math id="m83">
<mml:mrow>
<mml:mn>94.55</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, specificity of <inline-formula id="inf73">
<mml:math id="m84">
<mml:mrow>
<mml:mn>86.04</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, and AUC of 0.97. The obtained ROC curve is shown in <xref ref-type="fig" rid="F5">Figure&#x20;5</xref>. The training and validation loss curves are also illustrated in <xref ref-type="fig" rid="F6">Figure&#x20;6</xref>.</p>
<fig id="F6" position="float">
<label>FIGURE 6</label>
<caption>
<p>Training and Validation loss curves obtained for the COVID-FACT stage one and stage&#x20;two.</p>
</caption>
<graphic xlink:href="frai-04-598932-g006.tif"/>
</fig>
<p>In a second experiment, we trained our model using the complete CT images without segmenting the lung regions. The obtained model reached an accuracy of <inline-formula id="inf74">
<mml:math id="m85">
<mml:mrow>
<mml:mn>90.82</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, sensitivity of <inline-formula id="inf75">
<mml:math id="m86">
<mml:mrow>
<mml:mn>92.72</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, specificity of <inline-formula id="inf76">
<mml:math id="m87">
<mml:mrow>
<mml:mn>88.37</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, and AUC of 0.95. The corresponding ROC curve is shown in <xref ref-type="fig" rid="F5">Figure&#x20;5</xref>. This experiment indicates that segmenting lung regions in the first step will increase the sensitivity and AUC from <inline-formula id="inf77">
<mml:math id="m88">
<mml:mrow>
<mml:mn>92.72</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> and 0.95 to <inline-formula id="inf78">
<mml:math id="m89">
<mml:mrow>
<mml:mn>90.82</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> and 0.98 respectively, while slightly decreases the specificity from <inline-formula id="inf79">
<mml:math id="m90">
<mml:mrow>
<mml:mn>88.37</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> to <inline-formula id="inf80">
<mml:math id="m91">
<mml:mrow>
<mml:mn>86.04</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>. Although the numerical results show a slight improvement achieved by segmenting the lung regions, further investigating the sources of errors demonstrates the superiority of using segmented lung regions over the original CT scans. In the COVID-FACT model using lung segmented regions, none of COVID-19 and community acquired pneumonia cases have been mis-classified as a normal case by the <inline-formula id="inf81">
<mml:math id="m92">
<mml:mrow>
<mml:mn>3</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> thresholding after the first stage, and <inline-formula id="inf82">
<mml:math id="m93">
<mml:mrow>
<mml:mn>95.84</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> (23/24) of normal cases have been identified correctly using this threshold, while for the model without the lung segmentation, there is one mis-classification of a COVID-19 case by the <inline-formula id="inf83">
<mml:math id="m94">
<mml:mrow>
<mml:mn>3</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> thresholding, and <inline-formula id="inf84">
<mml:math id="m95">
<mml:mrow>
<mml:mn>91.66</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> (22/24) of normal cases were identified correctly using this threshold.</p>
<p>Furthermore, we compared performance of the Capsule Network-based framework of COVID-FACT with a CNN-based alternative to demonstrate the effectiveness of Capsule Networks and their superiority over CNN in terms of number of trainable parameters and accuracy. In other words, the CNN-based alternative model has the same front-end (convolutional layers) as that of COVID-FACT in both stages. However, the Capsule layers are replaced by fully connected layers including 128 neurons for intermediate layers and two neurons for the last layer at each stage. The last fully connected layer in each stage is followed by a sigmoid activation function and the remaining modifications and hyper-parameters are kept the same as used in COVID-FACT. The CNN-based COVID-FACT achieved an accuracy of <inline-formula id="inf85">
<mml:math id="m96">
<mml:mrow>
<mml:mn>71.43</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, sensitivity of <inline-formula id="inf86">
<mml:math id="m97">
<mml:mrow>
<mml:mn>81.82</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, and specificity of <inline-formula id="inf87">
<mml:math id="m98">
<mml:mrow>
<mml:mn>58.14</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>. The COVID-FACT performance, and number of trainable parameters for examined models are presented in <xref ref-type="table" rid="T2">Table&#x20;2</xref>. It is worth noting that in designing the CNN-based COVID-FACT described above, the complexity and structure have been kept similar to its capsule-based version. The goal is to evaluate and illustrate potential advantages of capsule network design over its CNN-based counterpart. Alternative models using CNN architecture and fully connected layers such as the DenseNet model (<xref ref-type="bibr" rid="B34">Yang et&#x20;al., 2020</xref>), however, consist of several convolutional layers and a high degree of complexity, as such it is expected from such complex models to outperform the CNN-based COVID-FACT.</p>
<table-wrap id="T2" position="float">
<label>TABLE 2</label>
<caption>
<p>Results obtained from COVID-FACT and the alternative CNN-based&#x20;model.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left">Method</th>
<th align="center">Accuracy</th>
<th align="center">Sensitivity</th>
<th align="center">Specificity</th>
<th align="center">AUC</th>
<th align="center">Trainable parameters</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">COVID-FACT with lung segmentation</td>
<td align="center">90.82</td>
<td align="center">94.55</td>
<td align="center">
<inline-formula id="inf88">
<mml:math id="m99">
<mml:mrow>
<mml:mn>86.04</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">0.98</td>
<td align="center">406,880</td>
</tr>
<tr>
<td align="left">COVID-FACT without lung segmentation</td>
<td align="center">90.82</td>
<td align="center">
<inline-formula id="inf89">
<mml:math id="m100">
<mml:mrow>
<mml:mn>92.72</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">88.37</td>
<td align="center">0.95</td>
<td align="center">406,880</td>
</tr>
<tr>
<td align="left">CNN-based COVID-FACT</td>
<td align="center">
<inline-formula id="inf90">
<mml:math id="m101">
<mml:mrow>
<mml:mn>71.43</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf91">
<mml:math id="m102">
<mml:mrow>
<mml:mn>81.82</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf92">
<mml:math id="m103">
<mml:mrow>
<mml:mn>58.14</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">0.67</td>
<td align="center">
<inline-formula id="inf93">
<mml:math id="m104">
<mml:mrow>
<mml:mn>365,806,660</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>As mentioned earlier, the ROC curve provides physicians with a precious tool to modify the sensitivity/specificity balance based on their preference by changing the classification cut-off probability. To elaborate this point, we changed the default cut-off probability from 0.5 to 0.75 and reached an accuracy of <inline-formula id="inf94">
<mml:math id="m105">
<mml:mrow>
<mml:mn>91.83</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, a sensitivity of <inline-formula id="inf95">
<mml:math id="m106">
<mml:mrow>
<mml:mn>90.91</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, and a specificity of <inline-formula id="inf96">
<mml:math id="m107">
<mml:mrow>
<mml:mn>93.02</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>. Further increasing the cut-off probability to 0.8 results in the same accuracy of <inline-formula id="inf97">
<mml:math id="m108">
<mml:mrow>
<mml:mn>91.83</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, a lower sensitivity of <inline-formula id="inf98">
<mml:math id="m109">
<mml:mrow>
<mml:mn>89.01</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, and a higher specificity of <inline-formula id="inf99">
<mml:math id="m110">
<mml:mrow>
<mml:mn>95.34</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>. On the other hand, decreasing the cut-off probability from 0.5 to 0.35 will increase the accuracy and the sensitivity to <inline-formula id="inf100">
<mml:math id="m111">
<mml:mrow>
<mml:mn>91.83</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf101">
<mml:math id="m112">
<mml:mrow>
<mml:mn>98.18</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> respectively, while slightly decreases the specificity to <inline-formula id="inf102">
<mml:math id="m113">
<mml:mrow>
<mml:mn>83.72</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>. The performance of the COVID-FACT for different values of cut-off probability are presented in <xref ref-type="table" rid="T3">Table&#x20;3</xref>.</p>
<table-wrap id="T3" position="float">
<label>TABLE 3</label>
<caption>
<p>Performance of COVID-FACT for different values of cut-off probability.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left">Cut-off probability</th>
<th align="center">0.35</th>
<th align="center">0.5</th>
<th align="center">0.6</th>
<th align="center">0.7</th>
<th align="center">0.75</th>
<th align="center">0.8</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">Accuracy (%)</td>
<td align="center">91.83</td>
<td align="center">90.82</td>
<td align="center">91.83</td>
<td align="center">90.82</td>
<td align="center">91.83</td>
<td align="center">91.83</td>
</tr>
<tr>
<td align="left">Sensitivity (%)</td>
<td align="center">98.18</td>
<td align="center">94.55</td>
<td align="center">92.73</td>
<td align="center">90.91</td>
<td align="center">90.91</td>
<td align="center">89.01</td>
</tr>
<tr>
<td align="left">Specificity (%)</td>
<td align="center">83.72</td>
<td align="center">86.04</td>
<td align="center">90.70</td>
<td align="center">90.70</td>
<td align="center">93.02</td>
<td align="center">95.34</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>While performance of the COVID-FACT is evaluated by its final decision made in the second stage, the first stage plays a crucial role in the overall accuracy of the model. As such, performance of the COVID-FACT in the first stage is also reported in <xref ref-type="table" rid="T4">Table&#x20;4</xref>. As shown in <xref ref-type="table" rid="T4">Table&#x20;4</xref>, <inline-formula id="inf103">
<mml:math id="m114">
<mml:mrow>
<mml:mo>&#x223c;</mml:mo>
<mml:mn>91</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> of the slices demonstrating infection are identified correctly by the COVID-FACT at the first stage, while there are some mis-classified slices that will be passed to the next stage as the infectious slices. It is also evident that the CNN-based model cannot properly identify infectious slices, which in turn led to the low performance of the second stage. It is worth mentioning that stage one is only responsible to detect candidate slices, while stage two classifies the slices into COVID and non-COVID categories. The second stage is followed by an aggregation mechanism, which takes all the slices of a patient into account and consequently decreases the impact of mis-classified slices at the first stage. We have also investigated the performance of the model when the commonly used focal loss function (<xref ref-type="bibr" rid="B21">Lin et&#x20;al., 2017</xref>) is utilized to train the model. The COVID-FACT framework trained by the focal loss function (<inline-formula id="inf104">
<mml:math id="m115">
<mml:mrow>
<mml:mi>&#x3b3;</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf105">
<mml:math id="m116">
<mml:mrow>
<mml:mi>&#x3b1;</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.25</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>) achieved the same patient-level performance compared to our proposed model while the performance of the first stage was lower with the accuracy of <inline-formula id="inf106">
<mml:math id="m117">
<mml:mrow>
<mml:mn>92.79</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, sensitivity of <inline-formula id="inf107">
<mml:math id="m118">
<mml:mrow>
<mml:mn>87.69</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, and the specificity of <inline-formula id="inf108">
<mml:math id="m119">
<mml:mrow>
<mml:mn>97.03</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>. The lower sensitivity in the first stage shows benefits of using the modified loss function as the role of the first stage in the pipeline is to detect slices with the evidence of infection to be analyzed in the second stage. As such, the model, which is trained using our modified loss function has been selected as the final model due to its higher accuracy and sensitivity in detecting slices demonstrating infection.</p>
<table-wrap id="T4" position="float">
<label>TABLE 4</label>
<caption>
<p>The performance of stage one in diagnosis of slices demonstrating infection.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left">Method (stage one)</th>
<th align="center">Accuracy (%)</th>
<th align="center">Sensitivity (%)</th>
<th align="center">Specificity (%)</th>
<th align="center">AUC</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">COVID-FACT with lung segmentation</td>
<td align="center">93.14</td>
<td align="center">90.75</td>
<td align="center">
<inline-formula id="inf109">
<mml:math id="m120">
<mml:mrow>
<mml:mn>94.01</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">0.96</td>
</tr>
<tr>
<td align="left">COVID-FACT without lung segmentation</td>
<td align="center">
<inline-formula id="inf110">
<mml:math id="m121">
<mml:mrow>
<mml:mn>92.78</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf111">
<mml:math id="m122">
<mml:mrow>
<mml:mn>87.59</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">94.36</td>
<td align="center">0.96</td>
</tr>
<tr>
<td align="left">CNN-based COVID-FACT</td>
<td align="center">
<inline-formula id="inf112">
<mml:math id="m123">
<mml:mrow>
<mml:mn>79.74</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf113">
<mml:math id="m124">
<mml:mrow>
<mml:mn>33.00</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf114">
<mml:math id="m125">
<mml:mrow>
<mml:mn>91.28</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">0.64</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>As another experiment, performance of stage two is evaluated without applying the first stage to provide a better comparison of the models used in the second stage. More specifically, the stage two model is trained based on the infectious slices identified by the radiologist and evaluated on the labeled test set including 17&#x20;COVID-19 and 8 CAP cases. The numbers of correctly predicted cases in this experiment are presented in <xref ref-type="table" rid="T5">Table&#x20;5</xref>. The experimental results obtained by the COVID-FACT framework using the lung segmentation achieved quite a similar performance compared to the case in which the model was trained based on the outputs of stage one. This result further demonstrates that the Capsule Network and the aggregation mechanism used in stage two can cope with errors in the previous stage and achieve desirable performance. It is worth mentioning that this experiment was performed using only the labeled dataset, which consequently provided a smaller dataset to train the&#x20;model.</p>
<table-wrap id="T5" position="float">
<label>TABLE 5</label>
<caption>
<p>Correctly predicted cases using only stage two without applying the first&#x20;stage.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left">Model</th>
<th align="center">COVID-19</th>
<th align="center">CAP</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">Stage 2 with lung segmentation</td>
<td align="center">94.1% (16/17)</td>
<td align="center">87.5% (7/8)</td>
</tr>
<tr>
<td align="left">Stage 2 without lung segmentation</td>
<td align="center">
<inline-formula id="inf115">
<mml:math id="m126">
<mml:mrow>
<mml:mn>88.2</mml:mn>
<mml:mo>%</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mn>15</mml:mn>
<mml:mo>/</mml:mo>
<mml:mn>17</mml:mn>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf116">
<mml:math id="m127">
<mml:mrow>
<mml:mn>62.5</mml:mn>
<mml:mo>%</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mn>5</mml:mn>
<mml:mo>/</mml:mo>
<mml:mn>8</mml:mn>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left">Stage 2&#x20;CNN-based</td>
<td align="center">
<inline-formula id="inf117">
<mml:math id="m128">
<mml:mrow>
<mml:mn>82.4</mml:mn>
<mml:mo>%</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mn>14</mml:mn>
<mml:mo>/</mml:mo>
<mml:mn>17</mml:mn>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf118">
<mml:math id="m129">
<mml:mrow>
<mml:mn>25</mml:mn>
<mml:mo>%</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mn>2</mml:mn>
<mml:mo>/</mml:mo>
<mml:mn>8</mml:mn>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>The localization maps generated by the Grad-CAM method are illustrated in <xref ref-type="fig" rid="F7">Figure&#x20;7</xref> for the second and fourth convolutional layers in the first stage of the COVID-FACT. It is evident in <xref ref-type="fig" rid="F7">Figure&#x20;7</xref> that the COVID-FACT model is looking at the right infectious areas of the lung to make the final decision. Due to the inherent structure of the Capsule layers, which represent image instances separately, their outputs cannot be superimposed over the input image. Consequently, in this study, the Grad-CAM localization maps are presented only for convolutional layers.</p>
<fig id="F7" position="float">
<label>FIGURE 7</label>
<caption>
<p>Localization heatmaps for the second and forth convolutional layers of the first stage obtained by the Grad-CAM for two slices.</p>
</caption>
<graphic xlink:href="frai-04-598932-g007.tif"/>
</fig>
<sec id="s4-1">
<title>4.1&#x20;K-Fold Cross-Validation</title>
<p>We have evaluated the performance of the COVID-FACT and its variants based on the 5-fold cross-validation (<xref ref-type="bibr" rid="B30">Stone, 1974</xref>) to provide more objective assessments. In this experiment, the COVID-FACT achieves the accuracy of <inline-formula id="inf119">
<mml:math id="m130">
<mml:mrow>
<mml:mn>87.61</mml:mn>
<mml:mo>&#xb1;</mml:mo>
<mml:mn>2.00</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, the sensitivity of <inline-formula id="inf120">
<mml:math id="m131">
<mml:mrow>
<mml:mn>88.30</mml:mn>
<mml:mo>&#xb1;</mml:mo>
<mml:mn>3.22</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, and specificity of <inline-formula id="inf121">
<mml:math id="m132">
<mml:mrow>
<mml:mn>86.75</mml:mn>
<mml:mo>&#xb1;</mml:mo>
<mml:mn>1.91</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>. Using the same 5-fold cross-validation technique, the COVID-FACT without using the segmented lung areas achieves the accuracy of <inline-formula id="inf122">
<mml:math id="m133">
<mml:mrow>
<mml:mn>87.31</mml:mn>
<mml:mo>&#xb1;</mml:mo>
<mml:mn>3.37</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, sensitivity of <inline-formula id="inf123">
<mml:math id="m134">
<mml:mrow>
<mml:mn>88.32</mml:mn>
<mml:mo>&#xb1;</mml:mo>
<mml:mn>5.00</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, and specificity of <inline-formula id="inf124">
<mml:math id="m135">
<mml:mrow>
<mml:mn>86.03</mml:mn>
<mml:mo>&#xb1;</mml:mo>
<mml:mn>3.18</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>. Finally, the CNN-based COVID-FACT achieves the accuracy of <inline-formula id="inf125">
<mml:math id="m136">
<mml:mrow>
<mml:mn>64.49</mml:mn>
<mml:mo>&#xb1;</mml:mo>
<mml:mn>1.61</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, sensitivity of <inline-formula id="inf126">
<mml:math id="m137">
<mml:mrow>
<mml:mn>79.58</mml:mn>
<mml:mo>&#xb1;</mml:mo>
<mml:mn>6.61</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, and specificity of <inline-formula id="inf127">
<mml:math id="m138">
<mml:mrow>
<mml:mn>46.67</mml:mn>
<mml:mo>&#xb1;</mml:mo>
<mml:mn>8.48</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>. The results confirm the superiority of the COVID-FACT using the segmented lung areas over its variants as was demonstrated in the previous experiments based on randomly selected test dataset. Moreover, similar to the previous experiments, modifying the cut-off probability is beneficial in the cross-validation case to adjust the capability of the model to focus on COVID or non-COVID cases depending on radiologists&#x2019; priorities. More specifically, in the aforementioned 5-fold cross-validation, decreasing the cut-off probability to 0.35 increases the sensitivity to <inline-formula id="inf128">
<mml:math id="m139">
<mml:mrow>
<mml:mn>92.97</mml:mn>
<mml:mo>&#xb1;</mml:mo>
<mml:mn>2.96</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> while the overall accuracy remains the same. Increasing the cut-off probability to 0.6, on the other hand, increases the specificity to <inline-formula id="inf129">
<mml:math id="m140">
<mml:mrow>
<mml:mn>91.16</mml:mn>
<mml:mo>&#xb1;</mml:mo>
<mml:mn>3.73</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> and provides the same accuracy similar to the previous&#x20;case.</p>
</sec>
</sec>
<sec sec-type="discussion" id="s5">
<title>5 Discussion</title>
<p>In this study, we proposed a fully automated Capsule Network-based framework, referred to as the COVID-FACT, to diagnose COVID-19 disease based on chest CT scans. The proposed framework consists of two stages, each of which containing several layers of convolutional and Capsule layers. COVID-FACT is augmented with a thresholding method to classify CT scans with zero or very few slices demonstrating infection as non-COVID patients, and an average voting mechanism coupled with a thresholding approach is embedded to extend slice-level classification into patient-level ones. Experimental results indicate that the COVID-FACT achieves a satisfactory performance, in particular a high sensitivity with far less trainable parameters, supervision requirements, and annotations compared to its counterparts.</p>
<p>We further investigated mis-classified cases to determine the limitations and possible improvements. <xref ref-type="table" rid="T6">Table&#x20;6</xref> shows the number of the mis-classified cases for each type of the input disease (COVID-19, CAP, normal) obtained at stage two, as well as the number of normal cases that were not identified correctly by the <inline-formula id="inf130">
<mml:math id="m141">
<mml:mrow>
<mml:mn>3</mml:mn>
<mml:mtext>%</mml:mtext>
</mml:mrow>
</mml:math>
</inline-formula> threshold after the first stage. The low rate of errors obtained by the <inline-formula id="inf131">
<mml:math id="m142">
<mml:mrow>
<mml:mn>3</mml:mn>
<mml:mtext>%</mml:mtext>
</mml:mrow>
</mml:math>
</inline-formula> threshold in the first stage demonstrates the capability of COVID-FACT to identify normal cases in the first stage, which is very helpful for physicians and radiologists to exclude normal cases at the very beginning of their&#x20;study.</p>
<table-wrap id="T6" position="float">
<label>TABLE 6</label>
<caption>
<p>The number of the mis-classified cases for each type of the input disease and the number of cases that were not identified correctly by. the <inline-formula id="inf132">
<mml:math id="m143">
<mml:mrow>
<mml:mn>3</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> threshold.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left">Input</th>
<th align="center">Errors (thresholding)</th>
<th align="center">Errors (stage two)</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">COVID-19</td>
<td align="center">
<inline-formula id="inf133">
<mml:math id="m144">
<mml:mrow>
<mml:mn>0</mml:mn>
<mml:mo>/</mml:mo>
<mml:mn>55</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf134">
<mml:math id="m145">
<mml:mrow>
<mml:mn>3</mml:mn>
<mml:mo>/</mml:mo>
<mml:mn>55</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left">CAP</td>
<td align="center">
<inline-formula id="inf135">
<mml:math id="m146">
<mml:mrow>
<mml:mn>0</mml:mn>
<mml:mo>/</mml:mo>
<mml:mn>19</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf136">
<mml:math id="m147">
<mml:mrow>
<mml:mn>5</mml:mn>
<mml:mo>/</mml:mo>
<mml:mn>19</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left">Normal</td>
<td align="center">
<inline-formula id="inf137">
<mml:math id="m148">
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>/</mml:mo>
<mml:mn>24</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf138">
<mml:math id="m149">
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>/</mml:mo>
<mml:mn>24</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>As in the case of highly contagious diseases such as COVID-19, the False-Negative-Rate (FNR) is of utmost importance, we have further analyzed such errors to explore the possible sources of the mis-classification. As shown in <xref ref-type="table" rid="T6">Table&#x20;6</xref> there are <inline-formula id="inf139">
<mml:math id="m150">
<mml:mrow>
<mml:mn>3</mml:mn>
<mml:mo>/</mml:mo>
<mml:mn>55</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> COVID-19 cases that are mis-classified by the COVID-FACT framework. We found that one mis-classified COVID-19 case contains unifocal infection manifestation with consolidation predominance rather than GGO, which are more common in CAP cases rather than COVID-19 ones. One other case of error was identified as an incomplete CT scan with missing slices, which has consequently made the correct identification difficult for the framework. In addition, we have reviewed the aforementioned errors in the case of image quality and lung segmentation as other potential causes of the error. The assessment results showed that the image qualities are adequate and the segmentation model performed well without removing or cropping the infection manifestations. Therefore, some errors are likely to be caused by the similarities between the infection patterns in CAP and COVID-19 cases. It is worth noting that decreasing the cut-off probability from 0.5 to 0.35, as shown in <xref ref-type="table" rid="T3">Table&#x20;3</xref>, will result in the correct classification of the two false-negative cases, which contain similar characteristics to other infections. This can be considered as a remedy, when FNR is of the main concern.</p>
<p>We also identified that errors in stage one are mainly caused by non-infectious abnormalities such as pulmonary fibrosis and artifacts. In this regard, we have further explored slices with the evidence of artifact where no infection manifestation presents. In some cases, the motion artifact or the artifacts caused by the presence of metallic components inside the body have generated some components in the image that were mis-classified as infectious slices. <xref ref-type="fig" rid="F8">Figure&#x20;8</xref> illustrates 4 samples of such slices in which images A) and B) belong to a mis-classified normal case while images C) and D) are related to two CAP cases, where classified correctly in the second stage. It is worth mentioning that, the number of such slices is negligible especially when they appear in cases that have multiple infectious slices (caused by CAP or COVID-19). In those cases, the influence of such slices with the evidence of artifact will be diminished by the second stage and the following aggregation mechanism. Motion artifact reduction algorithms can be investigated as a future work to cope with undesired impacts of the artifacts on the final result. It is worth mentioning that during the labeling process accomplished by the radiologist to detect slices demonstrating infection, we noticed that in some cases the abnormalities are barely visible with the standard visualization setting (window center and window width). Those abnormalities have been detected by changing the image contrast (by adjusting the window center and width) manually by the radiologist. This limitation will arise the need to research on the optimal contrast and window level use in future studies. As another limitation, we can point to the retrospective study used in the data collection part of this research. Although the provided dataset is acquired with the utmost caution and inspection, a retrospective data collection might add inappropriate cases to the study at hand. The potential improvement to address this limitation could be the collaboration of more radiologists in analyzing and labeling the data to assess if the interobserver agreement is satisfying or&#x20;not.</p>
<fig id="F8" position="float">
<label>FIGURE 8</label>
<caption>
<p>Example of slices with the evidence of artifact where no infection manifestation presents.</p>
</caption>
<graphic xlink:href="frai-04-598932-g008.tif"/>
</fig>
<p>As a side note to our discussion, we would like to mention that while both CT and CR can decrease the false negative rate at the admission and discharge times, the CR is less sensitive, and less specific compared to CT. Some studies such as Reference (<xref ref-type="bibr" rid="B32">Wong et&#x20;al., 2020</xref>) report that CR often shows no lung infection in COVID-19 patients at early stages resulting in a low sensitivity of 69% for diagnosis of COVID-19. Therefore, chest CT has a key role for diagnosis of COVID-19 in the early stages of the infection and also to set up a prognosis. Furthermore, a single CR image fails to incorporate details of infections in the lung and cannot provide a comprehensive view for the lung infection diagnosis. Unlike CR images, CT scans generate cross-sectional images (slices) and create a 3D representation of the body (i.e.,&#x20;each patient is associated with several 2D slices). As a result, CT images can show detailed structure of the lung and infected areas. Consequently, CT is considered as the preferred modality for grading and evaluation of imaging manifestations for COVID-19 diagnosis. It is worth adding that as CT scans are 3D images, as opposed to 2D chest radiographs, they are more difficult to be processed using ML and DL techniques, as the currently available resources cannot efficiently process the whole volume at once. As such, slice-level and thresholding techniques are utilized to cope with such limitations, leading to a reduced performance compared to the models working with CR (e.g., the COVID-CAPS (<xref ref-type="bibr" rid="B6">Afshar et&#x20;al., 2020d</xref>), which deals with 2D chest radiographs). The focus of our ongoing research is to further enhance performance of CT-based COVID-19 diagnosis models to fill the gap between the radiologists&#x2019; performance and that of volumetric-based DL techniques.</p>
<p>As a final note, unlike our previous work on the chest radiographs (<xref ref-type="bibr" rid="B2">Afshar et&#x20;al., 2020a</xref>), where we used a more imbalanced public dataset, the dataset used in this study contains a substantial number of COVID-19 confirmed cases making our results more reliable. Upon receiving more data from medical centers and collaborators, we will continue to further modify and validate the COVID-FACT by incorporating new datasets.</p>
</sec>
</body>
<back>
<sec id="s6">
<title>Data Availability Statement</title>
<p>Publicly available datasets were analyzed in this study. This data&#x20;can be found here: Figshare, <ext-link ext-link-type="uri" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="https://figshare.com/s/c20215f3d42c98f09ad0">https://figshare.com/s/c20215f3d42c98f09ad0</ext-link>.</p>
</sec>
<sec id="s7">
<title>Ethics Statement</title>
<p>The studies involving human participants were reviewed and approved by Concordia University. The patients/participants provided their written informed consent to participate in this study.</p>
</sec>
<sec id="s8">
<title>Author Contributions</title>
<p>SH, PA, and NE implemented the deep learning models. SH and PA drafted the manuscript jointly with AM and FN. FB, KS, and MR supervised the clinical study and data collection. FB and MR annotated the CT images. AO contributed to interpretation analysis and edited the manuscript. SA and KP edited the manuscript, AM, FN, and MR directed and supervised the study. All authors reviewed the manuscript.</p>
</sec>
<sec id="s10">
<title>Funding</title>
<p>This work was supported by the Natural Sciences and Engineering Research Council (NSERC) of Canada through the NSERC Discovery Grant RGPIN-2016-04988. Atashzar&#x0027;s efforts were supported by US National Science Foundation, Award &#x0023; 2031594.</p>
</sec>
<sec sec-type="COI-statement" id="s9">
<title>Conflict of Interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<ref-list>
<title>References</title>
<ref id="B1">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Afshar</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Heidarian</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Enshaei</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Naderkhani</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Rafiee</surname>
<given-names>M. J.</given-names>
</name>
<name>
<surname>Oikonomou</surname>
<given-names>A.</given-names>
</name>
<etal/>
</person-group> (<year>2021</year>). <source>COVID-CT-MD: COVID-19 Computed Tomography (CT) Scan Dataset Applicable in Machine Learning and Deep Learning</source>. <publisher-name>Nature Scientific Data</publisher-name>
<comment>In Press</comment>.</citation>
</ref>
<ref id="B2">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Afshar</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Heidarian</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Naderkhani</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Oikonomou</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Plataniotis</surname>
<given-names>K. N.</given-names>
</name>
<name>
<surname>Mohammadi</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2020a</year>). <article-title>COVID-CAPS: A Capsule Network-Based Framework for Identification of COVID-19 Cases from X-Ray Images</article-title>. <source>Pattern Recognition Lett.</source> <volume>138</volume>, <fpage>638</fpage>&#x2013;<lpage>643</lpage>. <pub-id pub-id-type="doi">10.1016/j.patrec.2020.09.010</pub-id> </citation>
</ref>
<ref id="B3">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Afshar</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Mohammadi</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Plataniotis</surname>
<given-names>K. N.</given-names>
</name>
</person-group> (<year>2020c</year>). &#x201c;<article-title>BayesCap: A Bayesian Approach to Brain Tumor Classification Using Capsule Networks</article-title>,&#x201d; in <conf-name>Submitted to IEEE International Conference on Image&#x20;Processing (ICIP)</conf-name> (<publisher-name>IEEE</publisher-name>), <fpage>2024</fpage>&#x2013;<lpage>2028</lpage>. <pub-id pub-id-type="doi">10.1109/LSP.2020.3034858</pub-id> </citation>
</ref>
<ref id="B4">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Afshar</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Mohammadi</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Plataniotis</surname>
<given-names>K. N.</given-names>
</name>
</person-group> (<year>2018</year>). &#x201c;<article-title>Brain Tumor Type Classification via Capsule Networks</article-title>,&#x201d; in <conf-name>2018 25th IEEE International Conference on Image Processing (ICIP)</conf-name> (<publisher-name>IEEE</publisher-name>), <fpage>3129</fpage>&#x2013;<lpage>3133</lpage>. <pub-id pub-id-type="doi">10.1109/ICIP.2018.8451379</pub-id> </citation>
</ref>
<ref id="B5">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Afshar</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Oikonomou</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Naderkhani</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Tyrrell</surname>
<given-names>P. N.</given-names>
</name>
<name>
<surname>Plataniotis</surname>
<given-names>K. N.</given-names>
</name>
<name>
<surname>Farahani</surname>
<given-names>K.</given-names>
</name>
<etal/>
</person-group> (<year>2020b</year>). <article-title>3D-MCN: A 3D Multi-Scale Capsule Network for Lung Nodule Malignancy Prediction</article-title>. <source>Sci. Rep.</source> <volume>10</volume>, <fpage>7948</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-020-64824-5</pub-id> </citation>
</ref>
<ref id="B6">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Afshar</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Plataniotis</surname>
<given-names>K. N.</given-names>
</name>
<name>
<surname>Mohammadi</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2020d</year>). &#x201c;<article-title>BoostCaps: A Boosted Capsule Network for Brain Tumor Classification</article-title>,&#x201d; in <source>Accepted in IEEE Engineering in Medicine and Biology Society (EMBC)</source> (<publisher-name>IEEE</publisher-name>), <fpage>20</fpage>&#x2013;<lpage>24</lpage>. <pub-id pub-id-type="doi">10.1109/EMBC44109.2020.9175922</pub-id> </citation>
</ref>
<ref id="B7">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Afshar</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Plataniotis</surname>
<given-names>K. N.</given-names>
</name>
<name>
<surname>Mohammadi</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2019a</year>). &#x201c;<article-title>Capsule Networks for Brain Tumor Classification Based on MRI Images and Coarse Tumor Boundaries</article-title>,&#x201d; in <conf-name>ICASSP 2019 - 2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)</conf-name> (<publisher-name>IEEE</publisher-name>), <fpage>1368</fpage>&#x2013;<lpage>1372</lpage>. <pub-id pub-id-type="doi">10.1109/ICASSP.2019.8683759</pub-id> </citation>
</ref>
<ref id="B8">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Afshar</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Plataniotis</surname>
<given-names>K. N.</given-names>
</name>
<name>
<surname>Mohammadi</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2019b</year>). &#x201c;<article-title>Capsule Networks&#x27; Interpretability for Brain Tumor Classification via Radiomics Analyses</article-title>,&#x201d; in <conf-name>2019 IEEE International Conference on Image Processing (ICIP)</conf-name> (<publisher-name>IEEE</publisher-name>), <fpage>3816</fpage>&#x2013;<lpage>3820</lpage>. <pub-id pub-id-type="doi">10.1109/ICIP.2019.8803615</pub-id> </citation>
</ref>
<ref id="B9">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Bai</surname>
<given-names>H. X.</given-names>
</name>
<name>
<surname>Hsieh</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Xiong</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Halsey</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Choi</surname>
<given-names>J.&#x20;W.</given-names>
</name>
<name>
<surname>Tran</surname>
<given-names>T. M. L.</given-names>
</name>
<etal/>
</person-group> (<year>2020</year>). <article-title>Performance of Radiologists in Differentiating COVID-19 from Non-COVID-19 Viral Pneumonia at Chest CT</article-title>. <source>Radiol.</source> <volume>296</volume>, <fpage>E46</fpage>&#x2013;<lpage>E54</lpage>. <pub-id pub-id-type="doi">10.1148/radiol.2020200823</pub-id> </citation>
</ref>
<ref id="B11">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Chen</surname>
<given-names>L. C.</given-names>
</name>
<name>
<surname>Papandreou</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Schroff</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Adam</surname>
<given-names>H.</given-names>
</name>
</person-group> (<year>2017</year>). <article-title>Rethinking Atrous Convolution for Semantic Image Segmentation</article-title> <source>ArXiv: 1706.05587</source>.</citation>
</ref>
<ref id="B10">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Chung</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Bernheim</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Mei</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Huang</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Zeng</surname>
<given-names>X.</given-names>
</name>
<etal/>
</person-group> (<year>2020</year>). <article-title>CT Imaging Features of 2019 Novel Coronavirus (2019-nCoV)</article-title>. <source>Radiol.</source> <volume>295</volume>, <fpage>202</fpage>&#x2013;<lpage>207</lpage>. <pub-id pub-id-type="doi">10.1148/radiol.2020200230</pub-id> </citation>
</ref>
<ref id="B12">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Deng</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Dong</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Socher</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>L. J.</given-names>
</name>
<name>
<surname>Kai Li</surname>
<given-names>Kai.</given-names>
</name>
<name>
<surname>Li Fei-Fei</surname>
<given-names>Li.</given-names>
</name>
</person-group> (<year>2009</year>). &#x201c;<article-title>ImageNet: A Large-Scale Hierarchical Image Database</article-title>,&#x201d; in <conf-name>2009 IEEE Conference on Computer Vision and Pattern Recognition</conf-name> (<publisher-name>IEEE</publisher-name>), <fpage>248</fpage>&#x2013;<lpage>255</lpage>. <pub-id pub-id-type="doi">10.1109/CVPR.2009.5206848</pub-id> </citation>
</ref>
<ref id="B35">
<citation citation-type="journal">
<collab>DICOM Standards Committee, Working Group 18 Clinical Trials</collab> (<year>2011</year>). <source>Supplement 142: Clinical Trial De-identification Profiles</source>. <publisher-loc>Rosslyn, VI, United States</publisher-loc>:<publisher-name>DICOM Standard</publisher-name>, <fpage>1</fpage>&#x2013;<lpage>44</lpage>. </citation>
</ref>
<ref id="B13">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Fang</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Xie</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Lin</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Ying</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Pang</surname>
<given-names>P.</given-names>
</name>
<etal/>
</person-group> (<year>2020</year>). <article-title>Sensitivity of Chest CT for COVID-19: Comparison to RT-PCR</article-title>. <source>Radiol.</source> <volume>296</volume>, <fpage>E115</fpage>&#x2013;<lpage>E117</lpage>. <pub-id pub-id-type="doi">10.1148/radiol.2020200432</pub-id> </citation>
</ref>
<ref id="B14">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Hara</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Kataoka</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Satoh</surname>
<given-names>Y.</given-names>
</name>
</person-group> (<year>2017</year>). &#x201c;<article-title>Learning Spatio-Temporal Features with 3D Residual Networks for Action Recognition</article-title>,&#x201d; in <conf-name>Proceedings - 2017 IEEE International Conference on Computer Vision Workshops</conf-name>, <conf-loc>Venice, Italy</conf-loc> (<publisher-name>ICCVW</publisher-name>), <fpage>3154</fpage>&#x2013;<lpage>3160</lpage>. <pub-id pub-id-type="doi">10.1109/ICCVW.2017.373</pub-id> </citation>
</ref>
<ref id="B15">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Hinton</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Sabour</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Frosst</surname>
<given-names>N.</given-names>
</name>
</person-group> (<year>2018</year>). &#x201c;<article-title>Matrix Capsules with EM Routing</article-title>,&#x201d; in <conf-name>6th International Conference on Learning Representations, ICLR 2018 - Conference Track Proceedings</conf-name>, <fpage>1</fpage>&#x2013;<lpage>29</lpage>. </citation>
</ref>
<ref id="B16">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Hofmanninger</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Prayer</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Pan</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Rohrich</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Prosch</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Langs</surname>
<given-names>G.</given-names>
</name>
</person-group> (<year>2020</year>). <source>Automatic Lung Segmentation in Routine Imaging Is a Data Diversity Problem, Not a Methodology Problem</source>, <fpage>1</fpage>&#x2013;<lpage>10</lpage>.</citation>
</ref>
<ref id="B17">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Hu</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Gao</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Niu</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Jiang</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Xiao</surname>
<given-names>X.</given-names>
</name>
<etal/>
</person-group> (<year>2020</year>). <article-title>Weakly Supervised Deep Learning for COVID-19 Infection Detection and Classification from CT Images</article-title>. <source>IEEE Access</source> <volume>8</volume>, <fpage>118869</fpage>&#x2013;<lpage>118883</lpage>. <pub-id pub-id-type="doi">10.1109/ACCESS.2020.3005510</pub-id> </citation>
</ref>
<ref id="B18">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Inui</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Fujikawa</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Jitsu</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Kunishima</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Watanabe</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Suzuki</surname>
<given-names>Y.</given-names>
</name>
<etal/>
</person-group> (<year>2020</year>). <article-title>Chest CT Findings in Cases from the Cruise Ship Diamond Princess with Coronavirus Disease (COVID-19)</article-title>. <source>Radiol. Cardiothorac. Imaging</source> <volume>2</volume>, <fpage>e200110</fpage>. <pub-id pub-id-type="doi">10.1148/ryct.2020200110</pub-id> </citation>
</ref>
<ref id="B19">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Krizhevsky</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Sutskever</surname>
<given-names>I.</given-names>
</name>
<name>
<surname>Hinton</surname>
<given-names>G. E.</given-names>
</name>
</person-group> (<year>2017</year>). <article-title>ImageNet Classification with Deep Convolutional Neural Networks</article-title>. <source>Commun. ACM</source> <volume>60</volume>, <fpage>84</fpage>&#x2013;<lpage>90</lpage>. <pub-id pub-id-type="doi">10.1145/3065386</pub-id> </citation>
</ref>
<ref id="B20">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Li</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Qin</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Xu</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Yin</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Kong</surname>
<given-names>B.</given-names>
</name>
<etal/>
</person-group> (<year>2020</year>). <article-title>Using Artificial Intelligence to Detect COVID-19 and Community-Acquired Pneumonia Based on Pulmonary CT: Evaluation of the Diagnostic Accuracy</article-title>. <source>Radiol.</source> <volume>296</volume>, <fpage>E65</fpage>&#x2013;<lpage>E71</lpage>. <pub-id pub-id-type="doi">10.1148/radiol.2020200905</pub-id> </citation>
</ref>
<ref id="B21">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Lin</surname>
<given-names>T.-Y.</given-names>
</name>
<name>
<surname>Goyal</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Girshick</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>He</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Doll&#xe1;r</surname>
<given-names>P.</given-names>
</name>
</person-group> (<year>2017</year>). <source>Focal Loss for Dense Object Detection</source>. <pub-id pub-id-type="doi">10.1109/iccv.2017.324</pub-id> </citation>
</ref>
<ref id="B22">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Mahmud</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Rahman</surname>
<given-names>M. A.</given-names>
</name>
<name>
<surname>Fattah</surname>
<given-names>S. A.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>CovXNet: A Multi-Dilation Convolutional Neural Network for Automatic COVID-19 and Other Pneumonia Detection from Chest X-Ray Images with Transferable Multi-Receptive Feature Optimization</article-title>. <source>Comput. Biol. Med.</source> <volume>122</volume>, <fpage>103869</fpage>. <pub-id pub-id-type="doi">10.1016/j.compbiomed.2020.103869</pub-id> </citation>
</ref>
<ref id="B23">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ng</surname>
<given-names>M.-Y.</given-names>
</name>
<name>
<surname>Lee</surname>
<given-names>E. Y. P.</given-names>
</name>
<name>
<surname>Yang</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Yang</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>H.</given-names>
</name>
<etal/>
</person-group> (<year>2020</year>). <article-title>Imaging Profile of the COVID-19 Infection: Radiologic Findings and Literature Review</article-title>. <source>Radiol. Cardiothorac. Imaging</source> <volume>2</volume>, <fpage>e200034</fpage>. <pub-id pub-id-type="doi">10.1148/ryct.2020200034</pub-id> </citation>
</ref>
<ref id="B24">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Rahimzadeh</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Attar</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Sakhaei</surname>
<given-names>S. M.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>A fully automated deep learning-based network for detecting COVID-19 from a new and large lung CT scan dataset</article-title> <source>Biomed. Signal Process. Control</source>. <volume>68</volume>, <fpage>102588</fpage>. <pub-id pub-id-type="doi">10.1016/j.bspc.2021.102588</pub-id> </citation>
</ref>
<ref id="B25">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Raman</surname>
<given-names>S. P.</given-names>
</name>
<name>
<surname>Mahesh</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Blasko</surname>
<given-names>R. V.</given-names>
</name>
<name>
<surname>Fishman</surname>
<given-names>E. K.</given-names>
</name>
</person-group> (<year>2013</year>). <article-title>CT Scan Parameters and Radiation Dose: Practical Advice for Radiologists</article-title>. <source>J.&#x20;Am. Coll. Radiol.</source> <volume>10</volume>, <fpage>840</fpage>&#x2013;<lpage>846</lpage>. <pub-id pub-id-type="doi">10.1016/j.jacr.2013.05.032</pub-id> </citation>
</ref>
<ref id="B26">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Ronneberger</surname>
<given-names>O.</given-names>
</name>
<name>
<surname>Fischer</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Brox</surname>
<given-names>T.</given-names>
</name>
</person-group> (<year>2015</year>). <source>U-Net: Convolutional Networks for Biomedical Image Segmentation</source>, <fpage>234</fpage>&#x2013;<lpage>241</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-319-24574-4_28</pub-id> </citation>
</ref>
<ref id="B27">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Selvaraju</surname>
<given-names>R. R.</given-names>
</name>
<name>
<surname>Cogswell</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Das</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Vedantam</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Parikh</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Batra</surname>
<given-names>D.</given-names>
</name>
</person-group> (<year>2017</year>). &#x201c;<article-title>Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization</article-title>,&#x201d; in <conf-name>2017 IEEE International Conference on Computer Vision (ICCV)</conf-name> (<publisher-name>IEEE</publisher-name>), <fpage>618</fpage>&#x2013;<lpage>626</lpage>. <pub-id pub-id-type="doi">10.1109/ICCV.2017.74</pub-id> </citation>
</ref>
<ref id="B28">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Sethy</surname>
<given-names>P. K.</given-names>
</name>
<name>
<surname>Behera</surname>
<given-names>S. K.</given-names>
</name>
<name>
<surname>Ratha</surname>
<given-names>P. K.</given-names>
</name>
<name>
<surname>Biswas</surname>
<given-names>P.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Detection of Coronavirus Disease (COVID-19) Based on Deep Features</article-title>. <source>Int. J.&#x20;Math. Eng. Manag. Sci.</source> <volume>5</volume>, <fpage>643</fpage>&#x2013;<lpage>651</lpage>. <pub-id pub-id-type="doi">10.20944/preprints202003.0300.v1</pub-id> </citation>
</ref>
<ref id="B29">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Shi</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Han</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Jiang</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Cao</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Alwalid</surname>
<given-names>O.</given-names>
</name>
<name>
<surname>Gu</surname>
<given-names>J.</given-names>
</name>
<etal/>
</person-group> (<year>2020</year>). <article-title>Radiological Findings from 81 Patients with COVID-19 Pneumonia in Wuhan, China: a Descriptive Study</article-title>. <source>Lancet Infect. Dis.</source> <volume>20</volume>, <fpage>425</fpage>&#x2013;<lpage>434</lpage>. <pub-id pub-id-type="doi">10.1016/S1473-3099(20)30086-4</pub-id> </citation>
</ref>
<ref id="B30">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Stone</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>1974</year>). <article-title>Cross-Validatory Choice and Assessment of Statistical Predictions</article-title>. <source>J.&#x20;R. Stat. Soc. Ser. B (Methodological)</source> <volume>36</volume>, <fpage>111</fpage>&#x2013;<lpage>133</lpage>. <pub-id pub-id-type="doi">10.1111/j.2517-6161.1974.tb00994.x</pub-id> </citation>
</ref>
<ref id="B31">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Wong</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2020</year>). <source>COVID-Net: A Tailored Deep Convolutional Neural Network Design for Detection of COVID-19 Cases from Chest X-Ray Images</source>.</citation>
</ref>
<ref id="B32">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wong</surname>
<given-names>H. Y. F.</given-names>
</name>
<name>
<surname>Lam</surname>
<given-names>H. Y. S.</given-names>
</name>
<name>
<surname>Fong</surname>
<given-names>A. H.-T.</given-names>
</name>
<name>
<surname>Leung</surname>
<given-names>S. T.</given-names>
</name>
<name>
<surname>Chin</surname>
<given-names>T. W.-Y.</given-names>
</name>
<name>
<surname>Lo</surname>
<given-names>C. S. Y.</given-names>
</name>
<etal/>
</person-group> (<year>2020</year>). <article-title>Frequency and Distribution of Chest Radiographic Findings in Patients Positive for COVID-19</article-title>. <source>Radiol.</source> <volume>296</volume>, <fpage>E72</fpage>&#x2013;<lpage>E78</lpage>. <pub-id pub-id-type="doi">10.1148/radiol.2020201160</pub-id> </citation>
</ref>
<ref id="B33">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Yamashita</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Nishio</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Do</surname>
<given-names>R. K. G.</given-names>
</name>
<name>
<surname>Togashi</surname>
<given-names>K.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>Convolutional Neural Networks: an Overview and Application in Radiology</article-title>. <source>Insights Imaging</source> <volume>9</volume>, <fpage>611</fpage>&#x2013;<lpage>629</lpage>. <pub-id pub-id-type="doi">10.1007/s13244-018-0639-9</pub-id> </citation>
</ref>
<ref id="B34">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Yang</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Jiang</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Cao</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Cao</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Feng</surname>
<given-names>R.</given-names>
</name>
<etal/>
</person-group> (<year>2020</year>). <article-title>Deep Learning for Detecting Corona Virus Disease 2019 (COVID-19) on High-Resolution Computed Tomography: a Pilot Study</article-title>. <source>Ann. Transl Med.</source> <volume>8</volume>, <fpage>450</fpage>. <pub-id pub-id-type="doi">10.21037/atm.2020.03.132</pub-id> </citation>
</ref>
<ref id="B36">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Yu</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Shen</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Yu</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Dang</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Cai</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Guo</surname>
<given-names>Y.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Lung Involvement in Patients with Coronavirus Disease-19 (COVID-19): a Retrospective Study Based on Quantitative CT Findings</article-title>. <source>Chin. J.&#x20;Acad. Radiol.</source> <volume>3</volume>, <fpage>102</fpage>&#x2013;<lpage>107</lpage>. <pub-id pub-id-type="doi">10.1007/s42058-020-00034-2</pub-id> </citation>
</ref>
<ref id="B37">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhang</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Shen</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Sang</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Wu</surname>
<given-names>X.</given-names>
</name>
<etal/>
</person-group> (<year>2020</year>). <article-title>Clinically Applicable AI System for Accurate Diagnosis, Quantitative Measurements, and Prognosis of COVID-19 Pneumonia Using Computed Tomography</article-title>. <source>Cell</source> <volume>181</volume>, <fpage>1423</fpage>&#x2013;<lpage>1433</lpage>. <pub-id pub-id-type="doi">10.1016/j.cell.2020.04.045</pub-id> </citation>
</ref>
</ref-list>
</back>
</article>