ORIGINAL RESEARCH article

Front. Public Health, 25 May 2022

Sec. Infectious Diseases: Epidemiology and Prevention

Volume 10 - 2022 | https://doi.org/10.3389/fpubh.2022.886958

Automated Multi-View Multi-Modal Assessment of COVID-19 Patients Using Reciprocal Attention and Biomedical Transform


Yanhan Li&#x;Yanhan Li1Hongyun Zhao,&#x;Hongyun Zhao2,3Tian GanTian Gan4Yang LiuYang Liu5Lian ZouLian Zou1Ting XuTing Xu4Xuan ChenXuan Chen6Cien Fan
Cien Fan1*Meng Wu
Meng Wu4*
  • 1Electronic Information School, Wuhan University, Wuhan, China
  • 2Department of Gastroenterology, The Second Affiliated Hospital of Chongqing Medical University, Chongqing, China
  • 3Chongqing Key Laboratory of Ultrasound Molecular Imaging, The Second Affiliated Hospital of Chongqing Medical University, Chongqing, China
  • 4Department of Ultrasound, Zhongnan Hospital of Wuhan University, Wuhan, China
  • 5School of Economics and Management, Wuhan University, Wuhan, China
  • 6Beijing Genomics Institute (BGI) Research, Shenzhen, China

Automated severity assessment of coronavirus disease 2019 (COVID-19) patients can help rationally allocate medical resources and improve patients' survival rates. The existing methods conduct severity assessment tasks mainly on a unitary modal and single view, which is appropriate to exclude potential interactive information. To tackle the problem, in this paper, we propose a multi-view multi-modal model to automatically assess the severity of COVID-19 patients based on deep learning. The proposed model receives multi-view ultrasound images and biomedical indices of patients and generates comprehensive features for assessment tasks. Also, we propose a reciprocal attention module to acquire the underlying interactions between multi-view ultrasound data. Moreover, we propose biomedical transform module to integrate biomedical data with ultrasound data to produce multi-modal features. The proposed model is trained and tested on compound datasets, and it yields 92.75% for accuracy and 80.95% for recall, which is the best performance compared to other state-of-the-art methods. Further ablation experiments and discussions conformably indicate the feasibility and advancement of the proposed model.

1. Introduction

In December 2019, coronavirus disease 2019 (COVID-19) broke out and began spreading to many countries around the globe, causing the ongoing coronavirus pandemic (1, 2). During the COVID-19 pandemic, lung imaging has played a crucial role in clinical care and epidemiological studies. COVID-19 has a significant impact on patients' respiratory systems, causing changes in the parenchyma of the lungs. Therefore, lung imaging is not only an effective technique for detecting COVID-19 but can also provide important information for clinicians to judge the severity of COVID-19 in patients through imaging features (3, 4). Common medical imaging techniques include computed tomography (CT), X-ray, and ultrasound. Among those medical imaging modalities, ultrasound possesses many advantages over others, including a high equipment penetration depth, ease of operation, the absence of radiation exposure, portability, the ability to perform real-time diagnosis, affordability, etc. (58). It is widely used as an additional screening and practical imaging method.

However, utilizing lung ultrasound images to assess the severity of COVID-19 patients is a complex and time-consuming task. Besides, manual judgments of sonographers will be influenced by inevitable subjective factors, leading to the omittance of inapparent image features. Deep learning (9) has achieved great success in the medical imaging domain (1012). Deep neural network (DNN) models (9) even outperform human experts in the diagnosis of some diseases (1315). Recently, many researchers have applied deep learning to the diagnosis or evaluation of COVID-19 (1620). Wang et al. (21) proposed the COVID-Net to classify X-ray images into normal, pneumonia, and COVID-19. Chen et al. (22) established a deep-learning-based diagnostic system to identify COVID-19 pneumonia. Aboutalebi et al. (23) leveraged transfer learning to transfer representational knowledge for predicting the airspace severity of a SARS-CoV-2 positive patient based on CXR images. Amyar et al. (24) proposed a new multitask deep learning model to jointly identify COVID-19 patients and segment COVID-19 lesions from chest CT images. Park et al. (25) proposed a novel multi-task vision transformer that leverages low-level CXR feature corpus for COVID-19 diagnosis and severity quantification. Sharifrazi et al. (26) proposed a model fusing convolutional neural network (CNN), support vector machine (SVM), and Sobel filter to detect COVID-19 using X-ray images. Ayoobi et al. (27) proposed to predict new cases and death rates of COVID-19 patients in different time spans utilizing multiple deep learning methods. Asgharnezhad et al. (28) proposed to quantify the competency of DNNs for generating reliable uncertainty estimates for COVID-19 diagnosis by introducing novel performance metrics. Alizadehsani et al. (29) proposed to cope with insufficient labeled COVID-19 data by introducing a semi-supervised classification method relying on Sobel edge detection and generative adversarial networks (GANs). Similarly, to mitigate the shortage of medical resources, Joloudari et al. (30) proposed DNN-GFE which combined DNNs with a Global Feature Extractor (GFE) for COVID-19 diagnosis. Khozeimeh et al. (31) proposed to deal with unbalanced data by introducing a data augmentation procedure based on autoencoders (AEs) and constructing CNN-AE to automatically diagnose COVID-19 cases. However, there are still several flaws in existing methods. (1) Most studies (19, 21, 23, 24) focus on chest X-ray and CT imaging, and little work takes lung ultrasound images into consideration. (2) Existing methods (17, 19, 21, 23, 24, 32) mostly leverage single-view images as input while it is more rational to exploit multi-view ones. (3) Existing methods (17, 19, 21, 2325, 32) mostly utilize data of single modality (unitary CT, X-ray, ultrasound, or other modalities) while multi-modal data are conductive to offer more information.

To mitigate the aforementioned flaws, in this paper, we propose an automated multi-view multi-modal model to analyze the severity in COVID-19 patients. The proposed model receives dual-view ultrasound image pairs and biomedical indices of patients to automatically conduct the comprehensive severity assessment tasks. We also propose reciprocal attention module and biomedical transform module, respectively, to extract and integrate multi-view and multi-modal features.

Briefly, the main contributions of our model are as follows:

(1) A novel multi-view multi-modal DNN is proposed. To the best of our knowledge, we are one of the forerunners to use both multi-view and multi-modal model for severity assessment in COVID-19 patients. Our proposed model has been evaluated on this dataset and outperformed all other state-of-the-art methods.

(2) A novel reciprocal attention module is proposed. Reciprocal attention module embedded with attention mechanism (33) rationally explores the inherent connection between ultrasound images of multiple views, generating attention features.

(3) A novel biomedical transform module is proposed. The biomedical transform module incorporates information of biomedical indices into ultrasound features, producing comprehensive hybrid features for assessment.

2. Materials and Methods

2.1. Data Acquisition

We collected data from a total of 164 patients, ranging in age from 17 to 87 years old, with 48.17% males and 51.83% females from Zhongnan Hospital of Wuhan University and Leishenshan Hospital to form the two datasets of COVID-19 patients. We collected 1,712 ultrasound images from the patients and simultaneously, we collected the corresponding biochemical indices related to pneumonia, including lymphocytes, c-reactive protein, lactate dehydrogenase, procalcitonin, and interleukin 6. See Table 1 for detailed information.

TABLE 1
www.frontiersin.org

Table 1. The clinical information related to coronavirus disease 2019 (COVID-19) patients.

All lung ultrasound images were saved in .jpg format. The ultrasound equipment used was a Siemens ACUSON OXANA1 with 6C1HD and 9L4 probes. The probes were placed perpendicular to the chest wall and parallel to the frame on areas 1–6 on the left and right sides (for both the left and right sides: the upper and lower axillary areas, the front and side areas of the chest wall, and the area of breast attachment and of the shoulder blade angle on the posterior side). Notably, since patients in ICUs could not lie on their sides, only the anterior and lateral thorax (areas 1–4) were examined. According to a general ultrasound triage protocol (34), if an abnormal ultrasound sign was discovered (e.g., irregular pleural lines, B lines, consolidation, and pleural effusion), a static picture of the scan was saved. The examples of abnormal ultrasound cases are presented in Figure 1.

FIGURE 1
www.frontiersin.org

Figure 1. Typical examples of abnormal ultrasound cases related to coronavirus disease 2019 (COVID-19). (A) The pleural line is jagged or concave. (B) The pleural line is broken. (C) The scanning area shows a wide range of dense white areas, with or without large consolidation.

All biomedical indices were collected according to patients′ medical records, including Lymphocytes, C-reactive protein, Lactate dehydrogenase, Procalcitonin, and Interleukin 6. Table 1 shows that lymphocytes possess Q1 (1st Quartile) of 0.28 109/L, Q3 (3rd Quartile) of 3.68 109/L, and a median of 1.6 109/L. C-reactive protein possesses Q1 of 0.5 mg/L, Q3 of 278.11 mg/L, and a median of 2.06 mg/L. Lactate dehydrogenase possesses Q1 of 116 IU/L, Q3 of 847 IU/L, and a median of 193 IU/L. Procalcitonin possesses Q1 of 0.01 ng/mL, Q3 of 9.25 ng/mL, and a median of 0.04 ng/mL. Interleukin 6 possesses Q1 of 1.5 pg/mL, Q3 of 1716 pg/mL, and a median of 1.71 pg/mL. Given that one patient usually has one ultrasonic examination and multiple pathological examinations, to make our work more reliable, biochemical indices from the pathological examination closest to ultrasonic examination were collected. Utilizing the summarized information, four experienced doctors made clinical diagnoses and annotated the corresponding data.

2.2. Overall Architecture

The architecture of the proposed model is presented in Figure 2. Generally, the proposed model receives multi-modal information of COVID-19 patients, containing ultrasound image pairs of two views and biomedical indices. Afterward, two branches of features are extracted from ultrasound image pairs. The two sets of features then undergo the proposed reciprocal attention module to acquire their attention features. Subsequently, attention features are further processed to obtain high-dimensional features. After average pooling, the high-dimensional features undergo the proposed biomedical transform module, where biomedical indices of corresponding patients are integrated to generate hybrid features. Exploiting hybrid features, our model conducts the final decision. Detailed architectures of reciprocal attention module and biomedical transform module are discussed in the following sections.

FIGURE 2
www.frontiersin.org

Figure 2. Overall flowchart of the proposed model. The proposed model receives multi-view ultrasound image pairs and biomedical indices of COVID-19 to conduct severity assessment tasks. The proposed reciprocal attention module tackles the multi-view ultrasound data and the proposed biomedical transform module tackles the biomedical data.

2.3. Reciprocal Attention Module

Different views of lung ultrasound images may contain complementary information. Hence, it is feasible to explore their inherent connections. Based on this, we propose a reciprocal attention module to acquire the attention feature from source view to target view bidirectionally, i.e., from left view to right view and from right view to left view simultaneously.

The detailed architecture of the reciprocal attention module is presented in Figure 3. The module receives features of two views extracted from previous CNN networks and calculates attention features utilizing attention mechanism (33). Initially, features of two views {FT,FS}H×W×C are processed by an embedder to produce two independent embeddings {ET,ES}HW×D, where S and T stand for source view and target view; H, W and C are the channel, height, and width of features; D is the dimension of embeddings. Then, according to (33), a query matrix Q ∈ ℝHW×D, a key matrix K ∈ ℝHW×D, and a value matrix V ∈ ℝHW×D are obtained as following:

{Q=ETWQK=ESWKV=ESWV    (1)

where {WQ, WK, WV} are learnable transform weights of query, key, and value, respectively. Exploiting acquired Q, K, V, reciprocal attention RA(·) ∈ ℝHW×D is calculated as follows:

RA(Q,K,V)=softmax(QKTD)V    (2)

Similar to self-attention, reciprocal attention represents the weights from source embedding toward each element in target embedding. To conduct effective fusion with features of target view, reciprocal attention is firstly reshaped via convolution layers and then added to target features. Formally, attention feature is computed as follows:

FA=σ(FT+τ(ε(RA)))    (3)

where FA is the attention feature after fusion, ε stands for reshaping convolution, τ is the dropout operation, and σ denotes the layer normalization. During the inference procedure, each view serves both as target and source and the other view serves as the opposite, yielding two branches of attention features corresponding to multi-view networks.

FIGURE 3
www.frontiersin.org

Figure 3. Detailed architecture of reciprocal attention module. Reciprocal attention module receives ultrasound image pairs and generates bidirectional attention features utilizing attention mechanism (33).

2.4. Biomedical Transform Module

Ultrasound images can provide partial information about the pulmonary lesions caused by COVID-19. However, COVID-19 causes wider damages to various organs and tissues, resulting in abnormal biomedical indices. Note that image features after CNNs are usually high-dimensioned while biomedical indices are low-dimensioned. Hence, it's significant to deploy biomedical indices and incorporate them with ultrasonic features. On the basis of this, we propose a biomedical transform module to undergo legitimate fusions between extracted graphic features and biomedical indices.

The detailed architecture of biomedical transform module is illustrated in Figure 4. For any feature maps F ∈ ℝH×W×C in the network, X ∈ ℝB denotes the corresponding biomedical indices collected from the same patient and B is the dimension of biomedical indices. Let Fi, i ∈ {1, 2, ⋯ , C} be the channel-wise features of F. Namely, [F1, F2, ⋯ , FC] = F where [·] denotes the concatenation operator. We aim to combine X and F through affine transformation. Concretely, an auxiliary network is built to generate transform parameters Ψ ∈ ℝC and Φ ∈ ℝC, which can be indicated as follows:

{Φ,Ψ}=fan(X)    (4)

where fan denotes the auxiliary network. Specifically, the auxiliary network is composed of multiple linear layers with ReLU functions to project X to Φ and Ψ. The auxiliary is jointly learned during the training procedure. Subsequently, affine transforms with the two learned parameters are applied to feature maps from corresponding patients to acquire hybrid features FhH×W×C. Formally, Fh is obtained as follows:

Fh=[Fh,1Fh,1Fh,C]=[ψ1*F1+ϕ1ψ2*F2+ϕ2ψC*FC+ϕC]=Ψ*F+Φ    (5)

where Fh, i, i ∈ {1, 2, ⋯ , C} is the channel-wise features of Fh, * denotes scalar multiplication, and +denotes scalar addition.

FIGURE 4
www.frontiersin.org

Figure 4. Detailed architectures of biomedical transform module. The biomedical transform module receives biomedical indices and generates parameters of affine transformation for ultrasonic features to obtain hybrid features.

2.5. Evaluation

First, for severity assessment of COVID-19 patients, we define Positive for severe cases and Negative for mild cases. Then, TP, TN, FP, FN denote true positive, true negative, false positive, and false negative, respectively. To evaluate the performance of our proposed model, the following evaluating metrics are selected.

Accuracy: The most primitive evaluating metric in classification problems, defined as the percentage of correctly predicted results in the total sample. Formally, accuracy is defined as follows:

Accuracy=TP+TNTP+FP+FN+TN    (6)

Precision: It is defined as the proportion of correctly classified positive samples among all samples predicted to be positive and is a measure of how well the model can predict correct results among all predicted positive samples. Formally, precision is defined as follows:

Precision=TPTP+FP    (7)

Recall: The probability of predicting a correctly classified positive sample among all actually positive samples. Formally, recall is defined as follows:

Recall=TPTP+FN    (8)

F-score: The comprehensive measure of model precision and recall. We select two types of F-scores, F1-score and F1.5 score. F1-score treats precision and recall of equal significance and it is defined as follows:

F1-score=2×Recall×PrecisionRecall+Precision    (9)

F1.5-score emphasizes more on recall than precision, which is more rational in clinical practice (will be discussed in the following section). Formally, F1.5-score is defined as follows:

F1.5-score=13×Recall×Precision9×Recall+4×Precision    (10)

3. Experimental Results and Discussion

3.1. Implementation Details

To construct rational dataset for our multi-view multi-modal model, the aforementioned lung ultrasound images from 164 patients were separated into image pairs containing opposite views. Namely, ultrasound images of L1 were combined with those of R1 and the same for the remaining views. Under this strategy, one patient could produce multiple cases for training and test. Whereafter, biomedical indices were first normalized and then added into corresponding cases to form the complete dataset. In total, 834 cases were constructed, where 171 were severe and 663 were mild. Afterward, a set of 627 cases (498 mild and 129 severe) was selected as the training set, and the rest were selected as the test set (165 mild and 42 severe).

To reduce the influence of unusual data distributions on ultrasound images and improve the training efficiency of our model, original images were first normalized and then resized to 448 × 448 pixels. Common data augmentations like random flip and crop were applied to ultrasound images before entering the network.

The proposed network was implemented in Pytorch (35). The network was trained with a batch size of 32 and the total training epoch was set to 100. The initial learning rate was 1 × 10−4 and was adjusted according to ReduceLROnPlateau strategy (35). Specifically, the learning rate was reduced by a factor of 0.5 when the loss did not decline after 8 continuous epochs. Adam (36) with default parameters was adopted as the optimizer. Given that the number of mild cases and severe cases was disproportionate, Focal loss (37) was selected as the loss function to cope with imbalanced cases. Focal loss is defined as follows:

LFL={-α(1-ȳ)γlogȳ,y=1-(1-α)ȳγlog(1-ȳ),y=0    (11)

where ȳ is the output of networks, y is ground truth, α and γ are parameters constructed to alleviate the negative influence brought by imbalanced data. During the training procedure, we set α = 0.25 and γ = 3.

3.2. Results of Severity Assessment of COVID-19 Patients

In this section, we present the experimental results of severity assessment in COVID-19 patients. To verify the advancement of our model, the performance of our model is compared to several other mainstream methods, namely, VGG (38), ResNet (39), DenseNet (40), SENet (41), SEResNet (41), Xception (42), InceptionV4 (43), and Sharifrazi et al. (26). We trained and tested the comparative methods with the same strategy and hyperparameters as the proposed model.

Table 2 summarizes the comparison of assessment results of COVID-19 patients. Our proposed model was referred to as RAB in this section. The experimental results indicated that the performance of the proposed RAB exceeds other models in almost all measurements. As for accuracy, our model achieved 92.75%, outperforming all other models. This indicates that RAB performs efficiently in recognizing both mild and severe cases. In terms of precision, RAB achieved 82.93%, inferior to several models. However, in early screening of COVID-19 patients, it counts more to recognize severe cases as many as possible, and it is tolerable to diagnose a tiny proportion of mild cases as severe ones but disastrous in turn. Therefore, precision is of less significance and emphasis should be placed more on recall. For recall, RAB yielded 80.95%, outperforming all other models. High performance in recall indicates that RAB performs preeminently in distinguishing severe cases from general COVID-19 cases, which is crucial clinically. Besides, for both F1 and F1.5 scores, our model also yielded the best outcomes, demonstrating that the comprehensive performance of our model overmatches all state-of-the-art models.

TABLE 2
www.frontiersin.org

Table 2. Results of severity assessment in COVID-19 patients in terms of Accuracy (Acc), Precision (Pre), recall, and F-score.

To make the experimental results more intuitive, the confusion matrices of all models were calculated and are presented in Figure 5. It is apparent to find that comparative methods tended to leave out more severe cases. For example, VGG11BN (38) and DenseNet121 (40) failed to recognize 15 cases out of 42 severe cases (Figures 5A,B), ResNet18 (39), SEResNet (41), and InceptionV4 (43) failed to recognize 13 severe cases (Figures 5C,E,G), and SENet (41) failed to recognize 17 severe cases (Figure 5D). The poor ability to identify severe COVID-19 cases restrains their applications. While RAB yielded the first place in recognizing severe cases and merely misdeemed 8 cases. Besides, it is noteworthy that Sharifrazi et al. (26) achieved plain results in recognizing COVID-19 cases because their method is aiming at X-ray images. X-ray images contain distinct boundaries for tissues and organs while ultrasound images usually have blurry ones. Hence, edge detection tends to produce mediocre outcomes for ultrasound images. We owe the remarkable performance of RAB to the proposed reciprocal attention module and biomedical transform module. The advancement of the two modules is discussed in the following section.

FIGURE 5
www.frontiersin.org

Figure 5. Confuse matrices of severity assessment for COVID-19 patients. (A) VGG11BN (38), (B) DenseNet121 (40), (C) ResNet18 (39), (D) SENet (41), (E) SEResNet (41), (F) Xception (42), (G) InceptionV4 (43), (H) Sharifrazi et al. (26) and, (I) RAB (Ours).

3.3. Ablation Studies

In this section, to verify the advancement of the proposed reciprocal attention module and biomedical transform module, ablation experiments were conducted. The performance of our model is compared to several baseline methods. Except for our model, we also constructed 3 baseline models and 1 variant model. As shown in Figure 6, the 3 baseline models were as follows: Single View model (SV), Dual View model without reciprocal attention module and biomedical transform module (DV), and reciprocal attention model without biomedical transform module (RA). The 1 variant model is inserting the biomedical transform module before the reciprocal attention module (named RAB-early). Our proposed model is referred as RAB-late in this section. Similarly, we trained and tested the baseline methods with the same strategy and hyperparameters as RAB-late′s.

FIGURE 6
www.frontiersin.org

Figure 6. Structures of baseline models. (A) Single view (SV) Model. (B) Dual view model without reciprocal attention module and biomedical transform module (DV) Model. (C) Reciprocal attention model without biomedical transform module (RA) Model. (D) Biomedical transform module before reciprocal attention module (named RAB-early)-early Model.

Table 3 summarizes the comparison of ablation results of COVID-19 patients. The experimental results indicated that the performance of the RAB-late model exceeds other baseline models in most measurements. As for accuracy, the RAB-late model achieved 92.75%, outperforming SV, DV, and RAB-early, the same as RA. In terms of precision, the RAB-late model achieved 82.93%, outperforming SV and RAB-early while inferior to DV and RA. The same as mentioned before, we focus more on recall. For recall, the RAB-late model yielded 80.95%, overmatching most other models. Note that the RAB-early model achieved the best recall of 85.71% while it had only 59.01% of precision. Such a model is fallacious in clinical use. For both F1 and F1.5 scores, the RAB-late model yielded the best outcomes.

TABLE 3
www.frontiersin.org

Table 3. Results of ablation experiments in terms of Acc, Pre, recall, and F-score.

Additionally, the confusion matrices of all 5 models are shown in Figure 7. As shown in Figure 7A, among 42 severe cases, the SV model failed to recognize 13 cases, leading to a disappointing recall. DV model surpassed the SV model in both recall and precision, demonstrating the validity of a dual-view strategy (Figure 7B). Moreover, the RA model strengthened the model's ability in identifying mild cases and improving model's precision, confirming the advancement of our proposed reciprocal attention module (Figure 7C). Notably, as presented in Figure 7D, the RAB-early model identified 36 severe cases out of the total 42 testing cases, yielding the best performance in recall, whereas it failed to classify 25 mild cases, far inferior to other models. In contrast, the RAB-late model maintained the most remarkable in both precision and recall and subsequent F-scores. The reason account for such gaps lies in the intrinsic structures of RAB-early and RAB-late, namely, the sequence of reciprocal attention and biomedical transform. According to aforementioned methodology, the reciprocal attention module seeks for connections of multiple views on the hypothesis that different views may possess complementary information. Nevertheless, such mechanism is futile when confronting identical features. In the RAB-early model, before the reciprocal attention module, biomedical transform is conducted to dual-view features with parameters generated from the same biomarker data, bringing identical factors to subsequent blocks and those identical factors restrain reciprocal attention mechanism. Whereas, the DAVB-late model encounters no similar restraints.

FIGURE 7
www.frontiersin.org

Figure 7. Confuse matrices of ablation experiments. (A) SV Model, (B) DV Model, (C) RA Model, (D) RAB-early Model, and (E) RAB-late Model.

Apart from the two proposed modules, the adopted focal loss function is also discussed. We replaced the focal loss with a more general loss function, BCE Loss, to train and test our model. Table 4 summarizes the comparison of different loss functions. The result indicates that the focal loss apparently elevated our model's overall performance. Note that the BCE Loss model generated extremely uneven precision and recall. The confusion matrices are shown in Figure 8. BCE Loss model failed to recognize 18 severe cases out of 42, distinctly inferior to the focal loss model. Technically, focal loss reduces the weights of categories with vast majority, thus applicable for the imbalanced COVID-19 data.

TABLE 4
www.frontiersin.org

Table 4. Results of different loss functions in terms of Acc, Pre, recall, and F-score.

FIGURE 8
www.frontiersin.org

Figure 8. Confuse Matrices of different loss functions. (A) BCE Loss model. (B) Focal Loss model.

3.4. Visualization

To interpret that our model has indeed learned certain abnormal signs to undergo an assessment of COVID-19 patients, grad class activation map (grad-CAM) (44) was used to visualize the most disease-indicative image areas learned by our model. Figure 9 presents the heat maps relevant to COVID-19 generated by the grad-CAM of our model. Distinctly, it demonstrates in Figure 9 that our model has detected several abnormal signs like irregular pleural lines, B lines, consolidation, and pleural effusion as standards to conduct downstream tasks.

FIGURE 9
www.frontiersin.org

Figure 9. Nidus-related visualization of ultrasound images. Using grad-CAM (44), the proposed model could highlight the image areas that are most relevant to COVID-19.

Additionally, to further interpret the advancement of biomedical indices in an assessment task, the statistical characteristics of biomedical indices are also considered. As shown in Figure 10, the mean value and SD of those data are demonstrated with × and bars with different colors indicate mild and severe cases. In Figure 10, the red and blue bars represent the SD for all biomedical indices, and the length of each bar represents the data range from meanstd to mean + std, standing for the approximate distribution of biomedical indices for mild and severe cases. For C-reactive protein, lactate dehydrogenase, procalcitonin, and interleukin 6, the data distributions of mild and severe patients overlap completely. For absolute lymphocyte value, there also exists overlaps. Data distributions of biomedical indices indicate that mild and severe cases possess certain but no sheer disparity, which brings challenges for manual judgments. Whereas, our model successfully exploits the meritorious information in biomedical indices and incorporates it into multi-modal data.

FIGURE 10
www.frontiersin.org

Figure 10. Statistical characteristics of biomedical indices. × stands for the mean value. The red and blue bar stands for the standard deviation for severe and mild cases, respectively. (A) Lymphocyte (absolute value). (B) C-reactive protein. (C) Lactate dehydrogenase. (D) Procalcitonin. (E) Interleukin-6.

Still, our model owns several flaws. Subject to insufficient laboratory data, the scale of our dataset is not as abundant as million-level ones, imposing restrictions on our model. In addition, for the same reason, not all patients have complete biochemical indices available and we simply replaced those void data with average ones. Those missing data may lead to restraints to on our model as well. Experiments with data of larger-scale and better integrity will be carried out in future work.

4. Conclusion

In this paper, we present a novel automated multi-view multi-modal model to assess the severity of COVID-19 patients exploiting ultrasound images and biomedical indices. The model has a dual-view structure and receives ultrasound images along with biomedical indices to generate comprehensive features. Specifically, the proposed reciprocal attention module acquires inherent connections between multiple views. and the proposed biomedical transform module integrates biomedical indices with extracted ultrasonic features to form hybrid features.

We have evaluated our model on compound datasets composed of ultrasound images and biomedical indices of COVID-19 patients. Experimental results demonstrate that our method outperforms all other state-of-the-art methods with better comprehensive performance. Further ablation studies and discussions consistently substantiate the rationality and advancement of our model. In the future, the model will be extended to wider ranges of modalities and larger scales of data.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Ethics Statement

The studies involving human participants were reviewed and approved by the Institutional Ethics Board of Zhongnan Hospital of Wuhan University (No. 2020042). The patients/participants provided their written informed consent to participate in this study.

Author Contributions

YLi, HZ, LZ, CF, and MW: conceptualization. YLi, HZ, and TG: methodology and software. YLi, TG, YLiu, LZ, TX, and CF: validation. YLi, HZ, TG, and TX: formal analysis. YLi, HZ, TG, and YLiu: investigation and visualization. YLi, HZ, TG, YLiu, TX, XC, and MW: resources and data curation. YLi, TG, YLiu, LZ, TX, XC, CF, and MW: writing—original draft preparation. YLi, HZ, TG, YLiu, LZ, TX, XC, CF, and MW: writing—review and editing. HZ, LZ, CF, and MW: supervision and project administration. HZ and MW: funding acquisition. All authors have read and agreed to the published version of the manuscript.

Funding

This study was funded by Hubei Provincial Natural Science Foundation (Grant No. 2020CFB729), Chinese Ultrasound Doctors Association, Science and Technology New Star Project (Grant No. KJXX2020002), Health Commission of Hubei Province Youth Talent Project (Grant No. WJ2021Q044), Zhongnan Hospital of Wuhan University Science, Technology and Innovation Seed Fund (Grant No. 2019090), research fund from the Medical Sci-Tech Innovation Platform of Zhongnan Hospital, Wuhan University (Grant No. PTXM2020027), Kuanren Talents Program of The Second Affiliated Hospital of Chongqing Medical University (Grant No. 2021240308), and National Natural Science Foundation Youth Project (Grant No. 81801714).

Conflict of Interest

XC was employed by BGI Research.

The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

1. Zhu N, Zhang D, Wang W, Li X, Yang B, Song J, et al. A novel coronavirus from patients with pneumonia in China 2019. N Engl J Med. (2020) 382:727–33. doi: 10.1056/NEJMoa2001017

PubMed Abstract | CrossRef Full Text | Google Scholar

2. Holshue ML, DeBolt C, Lindquist S, Lofy KH, Wiesman J, Bruce H, et al. First case of 2019 novel coronavirus in the United States. N Engl J Med. (2020) 382:929–36. doi: 10.1056/NEJMoa2001191

PubMed Abstract | CrossRef Full Text | Google Scholar

3. Ai T, Yang Z, Hou H, Zhan C, Chen C, Lv W, et al. Correlation of chest CT and RT-PCR testing for coronavirus disease 2019 (COVID-19) in China: a report of 1014 cases. Radiology. (2020) 296:E32–40. doi: 10.1148/radiol.2020200642

PubMed Abstract | CrossRef Full Text | Google Scholar

4. Ng MY, Lee EY, Yang J, Yang F, Li X, Wang H, et al. Imaging profile of the COVID-19 infection: radiologic findings and literature review. Radiol Cardiothor Imaging. (2020) 2:e200034. doi: 10.1148/ryct.2020200034

PubMed Abstract | CrossRef Full Text | Google Scholar

5. Jain PK, Sharma N, Giannopoulos AA, Saba L, Nicolaides A, Suri JS. Hybrid deep learning segmentation models for atherosclerotic plaque in internal carotid artery B-mode ultrasound. Comput Biol Med. (2021) 136:104721. doi: 10.1016/j.compbiomed.2021.104721

PubMed Abstract | CrossRef Full Text | Google Scholar

6. Buda N, Segura-Grau E, Cylwik J, Wełnicki M. Lung ultrasound in the diagnosis of COVID-19 infection-A case series and review of the literature. Adv Med Sci. (2020) 65:378–85. doi: 10.1016/j.advms.2020.06.005

PubMed Abstract | CrossRef Full Text | Google Scholar

7. Kim K, Wagner WR. Non-invasive and non-destructive characterization of tissue engineered constructs using ultrasound imaging technologies: a review. Ann Biomed Eng. (2016) 44:621–35. doi: 10.1007/s10439-015-1495-0

PubMed Abstract | CrossRef Full Text | Google Scholar

8. Buonsenso D, Pata D, Chiaretti A. COVID-19 outbreak: less stethoscope, more ultrasound. Lancet Respirat Med. (2020) 8:e27. doi: 10.1016/S2213-2600(20)30120-X

PubMed Abstract | CrossRef Full Text | Google Scholar

9. Nielsen MA. Neural Networks and Deep Learning. Vol. 25. San Francisco, CA: Determination Press (2015).

Google Scholar

10. Xia C, Li X, Wang X, Kong B, Chen Y, Yin Y, et al. A multi-modality network for cardiomyopathy death risk prediction with CMR images and clinical information. In: International Conference on Medical Image Computing and Computer-Assisted Intervention. Shenzhen: Springer (2019). p. 577–85. doi: 10.1007/978-3-030-32245-8_64

CrossRef Full Text | Google Scholar

11. Kong B, Wang X, Bai J, Lu Y, Gao F, Cao K, et al. Learning tree-structured representation for 3D coronary artery segmentation. Comput Med Imaging Graph. (2020) 80:101688. doi: 10.1016/j.compmedimag.2019.101688

PubMed Abstract | CrossRef Full Text | Google Scholar

12. Ye H, Gao F, Yin Y, Guo D, Zhao P, Lu Y, et al. Precise diagnosis of intracranial hemorrhage and subtypes using a three-dimensional joint convolutional and recurrent neural network. Eur Radiol. (2019) 29:6191–201. doi: 10.1007/s00330-019-06163-2

PubMed Abstract | CrossRef Full Text | Google Scholar

13. Gulshan V, Peng L, Coram M, Stumpe MC, Wu D, Narayanaswamy A, et al. Development and validation of a deep learning algorithm for detection of diabetic retinopathy in retinal fundus photographs. JAMA. (2016) 316:2402–10. doi: 10.1001/jama.2016.17216

PubMed Abstract | CrossRef Full Text | Google Scholar

14. Ting DSW, Cheung CYL, Lim G, Tan GSW, Quang ND, Gan A, et al. Development and validation of a deep learning system for diabetic retinopathy and related eye diseases using retinal images from multiethnic populations with diabetes. JAMA. (2017) 318:2211–23. doi: 10.1001/jama.2017.18152

PubMed Abstract | CrossRef Full Text | Google Scholar

15. Kermany DS, Goldbaum M, Cai W, Valentim CC, Liang H, Baxter SL, et al. Identifying medical diagnoses and treatable diseases by image-based deep learning. Cell. (2018) 172:1122–31. doi: 10.1016/j.cell.2018.02.010

PubMed Abstract | CrossRef Full Text | Google Scholar

16. Fang Y, Zhang H, Xie J, Lin M, Ying L, Pang P, et al. Sensitivity of chest CT for COVID-19: comparison to RT-PCR. Radiology. (2020) 296:E115–7. doi: 10.1148/radiol.2020200432

PubMed Abstract | CrossRef Full Text | Google Scholar

17. Yan L, Zhang HT, Goncalves J, Xiao Y, Wang M, Guo Y, et al. An interpretable mortality prediction model for COVID-19 patients. Nat Mach Intell. (2020) 2:283–8. doi: 10.1038/s42256-020-0180-7

CrossRef Full Text | Google Scholar

18. Shen B, Yi X, Sun Y, Bi X, Du J, Zhang C, et al. Proteomic and metabolomic characterization of COVID-19 patient sera. Cell. (2020) 182:59–72. doi: 10.1016/j.cell.2020.05.032

PubMed Abstract | CrossRef Full Text | Google Scholar

19. Harmon SA, Sanford TH, Xu S, Turkbey EB, Roth H, Xu Z, et al. Artificial intelligence for the detection of COVID-19 pneumonia on chest CT using multinational datasets. Nat Commun. (2020) 11:1–7. doi: 10.1038/s41467-020-17971-2

PubMed Abstract | CrossRef Full Text | Google Scholar

20. Liang W, Yao J, Chen A, Lv Q, Zanin M, Liu J, et al. Early triage of critically ill COVID-19 patients using deep learning. Nat Commun. (2020) 11:1–7. doi: 10.1038/s41467-020-17280-8

PubMed Abstract | CrossRef Full Text | Google Scholar

21. Wang L, Lin ZQ, Wong A. Covid-Net: a tailored deep convolutional neural network design for detection of covid-19 cases from chest x-ray images. Sci Rep. (2020) 10:1–12. doi: 10.1038/s41598-020-76550-z

PubMed Abstract | CrossRef Full Text | Google Scholar

22. Chen J, Wu L, Zhang J, Zhang L, Gong D, Zhao Y, et al. Deep learning-based model for detecting 2019 novel coronavirus pneumonia on high-resolution computed tomography. Sci Rep. (2020) 10:1–11. doi: 10.1038/s41598-020-76282-0

PubMed Abstract | CrossRef Full Text | Google Scholar

23. Aboutalebi H, Pavlova M, Shafiee MJ, Sabri A, Alaref A, Wong A. Covid-net cxr-s: Deep convolutional neural network for severity assessment of covid-19 cases from chest x-ray images. Diagnostics. (2022) 12:25. doi: 10.3390/diagnostics12010025

PubMed Abstract | CrossRef Full Text | Google Scholar

24. Amyar A, Modzelewski R, Li H, Ruan S. Multi-task deep learning based CT imaging analysis for COVID-19 pneumonia: classification and segmentation. Comput Biol Med. (2020) 126:104037. doi: 10.1016/j.compbiomed.2020.104037

PubMed Abstract | CrossRef Full Text | Google Scholar

25. Park S, Kim G, Oh Y, Seo JB, Lee SM, Kim JH, et al. Multi-task vision transformer using low-level chest X-ray feature corpus for COVID-19 diagnosis and severity quantification. Med Image Anal. (2022) 75:102299. doi: 10.1016/j.media.2021.102299

PubMed Abstract | CrossRef Full Text | Google Scholar

26. Sharifrazi D, Alizadehsani R, Roshanzamir M, Joloudari JH, Shoeibi A, Jafari M, et al. Fusion of convolution neural network, support vector machine and Sobel filter for accurate detection of COVID-19 patients using X-ray images. Biomed Signal Process Control. (2021) 68:102622. doi: 10.1016/j.bspc.2021.102622

PubMed Abstract | CrossRef Full Text | Google Scholar

27. Ayoobi N, Sharifrazi D, Alizadehsani R, Shoeibi A, Gorriz JM, Moosaei H, et al. Time series forecasting of new cases and new deaths rate for COVID-19 using deep learning methods. Results Phys. (2021) 27:104495. doi: 10.1016/j.rinp.2021.104495

PubMed Abstract | CrossRef Full Text | Google Scholar

28. Asgharnezhad H, Shamsi A, Alizadehsani R, Khosravi A, Nahavandi S, Sani ZA, et al. Objective evaluation of deep uncertainty predictions for covid-19 detection. Sci Rep. (2022) 12:1–11. doi: 10.1038/s41598-022-05052-x

PubMed Abstract | CrossRef Full Text | Google Scholar

29. Alizadehsani R, Sharifrazi D, Izadi NH, Joloudari JH, Shoeibi A, Gorriz JM, et al. Uncertainty-aware semi-supervised method using large unlabeled and limited labeled COVID-19 data. ACM Trans Multimedia Comput Commun Appl. (2021) 17:1–24. doi: 10.1145/3462635

CrossRef Full Text | Google Scholar

30. Joloudari JH, Azizi F, Nodehi I, Nematollahi MA, Kamrannejhad F, Mosavi A, et al. DNN-GFE: a deep neural network model combined with global feature extractor for COVID-19 diagnosis based on ct scan images. In: Easychair. Manchester (2021).

Google Scholar

31. Khozeimeh F, Sharifrazi D, Izadi NH, Joloudari JH, Shoeibi A, Alizadehsani R, et al. Combining a convolutional neural network with autoencoders to predict the survival chance of COVID-19 patients. Sci Rep. (2021) 11:1–18. doi: 10.1038/s41598-021-93543-8

PubMed Abstract | CrossRef Full Text | Google Scholar

32. Bhattacharyya A, Bhaik D, Kumar S, Thakur P, Sharma R, Pachori RB. A deep learning based approach for automatic detection of COVID-19 cases using chest X-ray images. Biomed Signal Process Control. (2022) 71:103182. doi: 10.1016/j.bspc.2021.103182

PubMed Abstract | CrossRef Full Text | Google Scholar

33. Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, et al. Attention is all you need. Advances in Neural Information Processing Systems. Vancouver, BC: Neural Information Processing Systems (NIPS) (2017). p. 30.

Google Scholar

34. Soldati G, Smargiassi A, Inchingolo R, Buonsenso D, Perrone T, Briganti DF, et al. Proposal for international standardization of the use of lung ultrasound for patients with COVID-19: a simple, quantitative, reproducible method. J Ultrasound Med. (2020) 39:1413–9. doi: 10.1002/jum.15285

PubMed Abstract | CrossRef Full Text | Google Scholar

35. Paszke A, Gross S, Massa F, Lerer A, Bradbury J, Chanan G, et al. Pytorch: an imperative style, high-performance deep learning library. Advances in Neural Information Processing Systems. Long Beach, CA: Neural Information Processing Systems (NIPS) (2019). p. 32.

Google Scholar

36. Kingma DP, Ba J. Adam: a method for stochastic optimization. arXiv preprint arXiv:14126980. (2014). doi: 10.48550/arXiv.1412.6980

CrossRef Full Text | Google Scholar

37. Lin TY, Goyal P, Girshick R, He K, Dollár P. Focal loss for dense object detection. In: Proceedings of the IEEE International Conference on Computer Vision. Los Alamitos, CA (2017) p. 2980–8. doi: 10.1109/ICCV.2017.324

PubMed Abstract | CrossRef Full Text | Google Scholar

38. Simonyan K, Zisserman A. Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:14091556. (2014). doi: 10.48550/arXiv.1409.1556

CrossRef Full Text | Google Scholar

39. He K, Zhang X, Ren S, Sun J. Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Seattle, WA (2016). p. 770–8. doi: 10.1109/CVPR.2016.90

PubMed Abstract | CrossRef Full Text | Google Scholar

40. Huang G, Liu Z, Van Der Maaten L, Weinberger KQ. Densely connected convolutional networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Honolulu, HI (2017). p. 4700–08. doi: 10.1109/CVPR.2017.243

CrossRef Full Text | Google Scholar

41. Hu J, Shen L, Sun G. Squeeze-and-excitation networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Los Alamitos, CA (2018). p. 7132–41. doi: 10.1109/CVPR.2018.00745

CrossRef Full Text | Google Scholar

42. Chollet F. Xception: Deep learning with depthwise separable convolutions. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Honolulu, HI (2017). p. 1251–8. doi: 10.1109/CVPR.2017.195

CrossRef Full Text | Google Scholar

43. Szegedy C, Ioffe S, Vanhoucke V, Alemi AA. Inception-v4, inception-resnet and the impact of residual connections on learning. In: Thirty-First AAAI Conference on Artificial Intelligence. San Francisco, CA (2017).

Google Scholar

44. Selvaraju RR, Cogswell M, Das A, Vedantam R, Parikh D, Batra D. Grad-cam: Visual explanations from deep networks via gradient-based localization. In: Proceedings of the IEEE International Conference on Computer Vision. Venice (2017). p. 618–26. doi: 10.1109/ICCV.2017.74

CrossRef Full Text | Google Scholar

Keywords: COVID-19, deep learning, multi-view, multi-modal, computer aided diagnosis

Citation: Li Y, Zhao H, Gan T, Liu Y, Zou L, Xu T, Chen X, Fan C and Wu M (2022) Automated Multi-View Multi-Modal Assessment of COVID-19 Patients Using Reciprocal Attention and Biomedical Transform. Front. Public Health 10:886958. doi: 10.3389/fpubh.2022.886958

Received: 01 March 2022; Accepted: 20 April 2022;
Published: 25 May 2022.

Edited by:

Sadiq Hussain, Dibrugarh University, India

Reviewed by:

Roohallah Alizadehsani, Deakin University, Australia
Javad Hassannataj Joloudari, University of Birjand, Iran

Copyright © 2022 Li, Zhao, Gan, Liu, Zou, Xu, Chen, Fan and Wu. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Cien Fan, ZmNlQHdodS5lZHUuY24=; Meng Wu, d2IwMDA3MTNAd2h1LmVkdS5jbg==

These authors have contributed equally to this work

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.