Skip to main content

ORIGINAL RESEARCH article

Front. Pediatr., 18 January 2022
Sec. Pediatric Cardiology
Volume 9 - 2021 | https://doi.org/10.3389/fped.2021.770182

Standard Echocardiographic View Recognition in Diagnosis of Congenital Heart Defects in Children Using Deep Learning Based on Knowledge Distillation

Lanping Wu1 Bin Dong2 Xiaoqing Liu3 Wenjing Hong1 Lijun Chen1 Kunlun Gao3 Qiuyang Sheng3 Yizhou Yu3 Liebin Zhao2* Yuqi Zhang1*
  • 1Department of Pediatric Cardiology, Shanghai Children's Medical Center, School of Medicine, Shanghai Jiao Tong University, Shanghai, China
  • 2Shanghai Engineering Research Center of Intelligence Pediatrics, Shanghai, China
  • 3Deepwise Artificial Intelligence Laboratory, Beijing, China

Standard echocardiographic view recognition is a prerequisite for automatic diagnosis of congenital heart defects (CHDs). This study aims to evaluate the feasibility and accuracy of standard echocardiographic view recognition in the diagnosis of CHDs in children using convolutional neural networks (CNNs). A new deep learning-based neural network method was proposed to automatically and efficiently identify commonly used standard echocardiographic views. A total of 367,571 echocardiographic image slices from 3,772 subjects were used to train and validate the proposed echocardiographic view recognition model where 23 standard echocardiographic views commonly used to diagnose CHDs in children were identified. The F1 scores of a majority of views were all ≥0.90, including subcostal sagittal/coronal view of the atrium septum, apical four-chamber view, apical five-chamber view, low parasternal four-chamber view, sax-mid, sax-basal, parasternal long-axis view of the left ventricle (PSLV), suprasternal long-axis view of the entire aortic arch, M-mode echocardiographic recording of the aortic (M-AO) and the left ventricle at the level of the papillary muscle (M-LV), Doppler recording from the mitral valve (DP-MV), the tricuspid valve (DP-TV), the ascending aorta (DP-AAO), the pulmonary valve (DP-PV), and the descending aorta (DP-DAO). This study provides a solid foundation for the subsequent use of artificial intelligence (AI) to identify CHDs in children.

Introduction

Congenital heart defects (CHDs) are the most common birth defects in China, with an incidence rate of approximately 0.9% in live births, and are the main causes of death in children aged 0–5 years (1). Precise preoperative diagnosis helps to develop the most reasonable surgery and interventional treatment plan for CHDs. Transthoracic echocardiography (TTE) can provide sufficient information about the structure of the heart to evaluate hemodynamics and cardiac function. It is currently the most commonly used non-invasive examination method for CHDs. Many factors contribute to the heterogeneity of echocardiographic diagnosis, such as inherent pulse variability in the heart, speckled noise and artifacts in echocardiogram, and inter- and intra-class differences in standard echocardiographic views (2). The anatomical structure and spatial configuration of CHDs are complex and changeable, and accurate diagnosis through TTE is complicated and time-consuming, and it is heavily dependent on the accurate judgment of each echocardiographic view by experienced cardiologists. In China, there is a lack of experienced cardiologists at the grassroots level, especially in rural areas. It takes a great effort to train an experienced doctor. In addition, many cardiologists lack experience in diagnosing CHDs. Experienced cardiac surgery centers usually have long waiting lists for appointments, and cardiologists are overworked, leading to an imbalance between supply and demand. Therefore, it is very necessary to establish an automatic diagnosis system of CHDs to alleviate the difficulty of CHD diagnosis at the grassroots level.

In recent years, artificial intelligence (AI) based image recognition technology, especially deep learning via convolutional neural networks (CNNs), have been greatly improved and increasingly applied to diagnostic imaging in medical fields (311). Currently, AI based analytic software has been applied to the research of valvular diseases. Jin et al. (12) has compared the accuracy of anatomical intelligence ultrasound (AIUS) and manual depiction of the mitral valve prolapse degree and found that AIUS can significantly improve accuracy and reduce analysis time. According to the research of Choi et al. (13), the aortic valve (AV) regurgitation volume measured by three-dimensional full volume color Doppler echocardiography has a high correlation and consistency with those measured by magnetic resonance spectrum, which can be used to quantitatively evaluate the AV disease.

Standard echocardiographic view recognition provides the foundation for the clinical diagnosis of heart diseases. Recent CNNs based on deep learning methods have been applied in standard echocardiographic view recognition and achieved superior performance than traditional methods. Extensive research has been focused on fetal ultrasound (14, 15). Baumgartner et al. (14) proposed a two-dimensional full CNN containing six convolutional modules, which replaced the commonly used full connection layer with the convolutional and global average pooling layers, and recognized 12 standard views of fetal ultrasound with an average accuracy of 69% and average recall rate of 80%. Sridar et al. (15) also started with fetal ultrasound data and used the AlexNet neural network pre-trained with a natural image dataset to identify 14 views. They designed a strategy for simultaneous training using two parallel networks. The input of one network was the entire ultrasonic image used to learn the global semantic information of the image. The input of the other network was random local segmentation of ultrasonic images used to learn the local features of the image. The average accuracy rate and average recall rate were 76.47 and 75.41%, respectively. Zhang et al. (16) attempted to develop a fully automated echocardiographic clinical diagnosis system consisting of standard view recognition, cardiac structure segmentation, and heart disease diagnosis. They applied a 13-layer CNN network to standard view recognition, and obtained a global accuracy of 84% on standard view recognition using 277 cases through a five-fold cross validation. Their model can be used to detect hypertrophic cardiomyopathy, pulmonary arterial hypertension, and cardiac amyloidosis. However, their work was far from the diagnosis of common CHDs since the standard views they used were mainly derived from five views including parasternal long axis (PSLA), parasternal short axis, apical 2-chamber (A2C), apical 3-chamber (A3C), apical 4-chamber (A4C).

In this study, we used 367,571 echocardiographic images from 3,772 subjects to build a 24-category classification model that can identify 23 standard echocardiographic views and one “other” view. To our knowledge, this is the first work that covers the most commonly used CHD diagnostic views. As we all know, there is a trade-off between the complexity and accuracy of the model. Achieving high accuracy and timeliness simultaneously is challenging, especially when a large number of views need to be identified. Our experimental results demonstrate that the proposed knowledge distillation method maximizes the accuracy of the network while keeping the complexity of the network unchanged and has important clinical significance. Ultrasound examination of CHDs is a dynamic scanning process and images collected are mostly non-standard views and unevenly distributed. Compared with existing work, the data set used in our study is the largest taken from a real historical examination database, closer to the routine CHD diagnosis, which is another clinical contribution of this work.

Methods

Subjects

The experimental data consists of a training data set (3,409 cases and 247,750 TTE images) and a test data set (363 cases and 119,821 TTE images), both from the Shanghai Children's Medical Center. The TTE images are retrieved from a historical real examination database in PACS system (Table 1). The 363 cases of the test data set include 193 male cases and 170 female cases, with an average age of 5.6 years. There were 170 healthy children, 102 children with CHDs, and 91 children with CHDs after surgery. Table 2 lists the characteristics of the test dataset. The machines used were mainly Philips iE33 and EPIC 7C. Six professional ultrasound doctors participated in the data annotation, of which four ultrasound doctors participated in the annotation of the training data set, and two ultrasound doctors participated in the annotation of the test data set. To avoid human bias, all images were annotated and confirmed by two doctors. More specifically, each image was annotated by a senior resident or attending physician with more than 5 years of experience, and then reviewed and confirmed by an associate chief physician or chief physician with at least 10 years of experience. All data was tailored to anonymize patient information.

TABLE 1
www.frontiersin.org

Table 1. View distribution of our data.

TABLE 2
www.frontiersin.org

Table 2. Characteristics comparison between the CHD group and the normal groups of the test data set.

Data Preprocessing

Usually, the boundary area of each frame of echocardiogram contains a large amount of background information that is not useful for diagnosis. In order to extract the discriminative features embedded in the image, the central fan-shaped area was cropped, and each cropped image frame was resized to a fixed size of 192 × 256 pixels.

Standard Echocardiographic View Recognition

In clinical practice, the diagnosis of various CHDs involves a series of standard scans. Standard echocardiographic view recognition is a prerequisite for automatic CHD diagnosis. In this study, we established a 24-category classification model that can recognize 23 standard echocardiographic views commonly used to diagnose CHDs in children and one “others.” As shown in Table 1, the 23 views include subcostal four-chamber view (sub4C), subcostal short-axis view of the left ventricle (subSALV), subcostal sagittal view of the atrium septum (subSAS), subcostal coronal view of the atrium septum (subCAS), subcostal short-axis view through the right ventricular outflow tract (subRVOT), apical four-chamber view (A4C), apical five-chamber view (A5C), low parasternal four-chamber view (LPS4C), low parasternal five-chamber view (LPS5C), parasternal short-axis view at the base of the heart (basal short axis, sax-basal), parasternal short-axis view at the level of the mitral valve (short axis at mid, sax-mid), parasternal long-axis view of the left ventricle (PSLV), parasternal view of the pulmonary artery (PSPA), suprasternal long-axis view of the entire aortic arch (supAO), Doppler recording from the mitral valve (DP-MV), the tricuspid valve (DP-TV), the ascending aorta (DP-AAO), the pulmonary valve (DP-PV), the descending aorta (DP-DAO), other Doppler recordings (DP-OTHER), and M-mode echocardiographic recording of the aortic (M-AO), the left ventricle at the level of the papillary muscle (M-LV), other M-mode echocardiographic recordings (M-OTHER). Except for these 23 views, all other views are classified as “others.”

Figure 1 demonstrated the example images of the 24 standard echocardiographic views. Among these views, subcostal views are of great significance for the diagnosis of congenital heart disease in infants and young children, especially for the diagnosis of atrial septal defect (ASD). Apical and low parasternal views are mainly used for the diagnosis of ASD, ventricular septal defect (VSD), tetralogy of Fallot (TOF), and atrioventricular valve diseases. The parasternal views are mainly used for the diagnosis of AV, pulmonary valve (PV) diseases, ASD, VSD, TOF, and patent ductus arteriosus (PDA). M-mode ultrasound is mainly measured next to the parasternal to assess the size of the heart chamber and the systolic function of the left ventricle. Pulse Doppler is mainly used to measure the blood flow velocity of various valves and blood vessels to determine whether there is stenosis.

FIGURE 1
www.frontiersin.org

Figure 1. Example images of the 24 standard echocardiographic views: (a) LPS4C; (b) LPS5C; (c) subCAS; (d) subSAS; (e) sub4C; (f) subRVOT; (g) subSALV; (h) DP-MV; (i) DP-PV; (j) DP-DAO; (k) DP-OTHER; (l) DP-TV; (m) DP-AAO; (n) OTHER; (o) A4C; (p) A5C; (q) sax-basal; (r) sax-mid; (s) PSLA; (t) PSPA; (u) supAO; (v) M-OTHER; (w) M-AO; (x) M-LV.

As we all know, there is a trade-off between the complexity and accuracy of the model. Achieving high accuracy and timeliness simultaneously is challenging, especially when a large number of views need to be identified. Inspired by the idea of probabilistic knowledge transfer method (17), which considered knowledge distillation as a metric learning problem, we proposed a CNN method that performed standard echocardiographic view recognition through knowledge distillation. Knowledge distillation enables the student network learn the generalization ability of the teacher network, through replacing the hard original one-hot label with the soft label, and learn the ability to distinguish similar features (1720). Therefore, on the one hand, the knowledge distillation method compresses the model, on the other hand, it enhances the generalization ability of the model. As shown in Figure 2, ResNet-34 (21) was used as the student model and ResNeSt-200 (22) was applied as the teacher model. Our distillation method works by matching the probability distribution of logits of the student and the teacher. In each iteration of the training phase, the student learn from data by minimizing the cross entropy loss. And meanwhile with Kullback-Leibler divergence as the loss function, the output logits of the student will be constantly close to that of the teacher. As a result, we balanced the performance of the model and the complexity of running time through realizing the knowledge transfer from the teacher model to the student model.

FIGURE 2
www.frontiersin.org

Figure 2. The proposed network architecture for standard echocardiographic view recognition.

Model Training

We trained the teacher model and the student model simultaneously and also applied transfer learning to the pre-trained backbone network using the large-scale natural image dataset ImageNet (23). To enhance the diversity and robustness of data, a series of online data augmentations including horizontal and/or vertical flipping with a random probability of p = 0.5 and random rotation with a random probability of p = 0.5 were also implemented. Stochastic gradient descent (SGD) was used as the optimizer with an initial learning rate of 0.001, a decay factor of 0.1 for every 10 epochs, and a mini-batch size of 128.

All codes were implemented using Python 3.7 and Pytorch 1.4.0. The experiments were carried out on a workstation platform with 8 NVIDIA TITAN RTX GPUs, 24GB GPU memory, 256G RAM, and 80 Intel(R) Xeon(R) Gold 6248 CPU @ 2.50GHz, using Ubuntu 16.04.

Results

In this study, four performance evaluation metrics including Precision, Sensitivity, Specificity, and F1 Score were applied to validate the performance of the proposed view recognition model and they are defined as follows:

Precision = TPTP + FP    (1)
Sensitivity = Recall = TPTP + FN    (2)
Specificity = TNTN + FP    (3)
F1 = 2xPrecision * RecallPrecision + Recall    (4)

where TP, FP, TN, and FN are the true positive, false positive, true negative, and false negative rates, respectively. TP and TN represent the positives and negatives of correct predictions with respect to the ground truth. FP and FN represent the positives and negatives of incorrect predictions with respect to the ground truth. F1 Score is the harmonic average of precision and recall with a value ranged in [0–1]. The higher the value means that the model has better performance.

Common CHDs in children include ASDs, VSD, PDA, pulmonary stenosis (PS), aortic stenosis (AS), TOF, and so on. Table 3 lists the diagnosable CHDs corresponding to each standard echocardiographic view. More specifically, the most commonly used ASD diagnostic views include sub4C, subSAS, subCAS, LPS4C, and sax-basal. The most commonly used views for the diagnosis of VSD are sub4C, subRVOT, subSALV, A4C, A5C, LPS4C, LPS5C, and PSLA. The most commonly used views for PVS diagnosis include the subRVOT, sax-basal, and PSPA. The most commonly used views for diagnosing PDA include PSPA and supAO. The most commonly used views to diagnose TOF include sub4C, subRVOT, A4C, A5C, LPS4C, LPS5C, sax-basal, PSPA, and PSLA. M-AO and M-LV are used for heart size measurement and heart function calculation. The blood flow spectrum of the four valves and descending aorta in these 23 views provide the basis for the subsequent diagnosis of valvular and aortic arch diseases. As shown in Table 3, the F1 scores of the subCAS+SAS, A4C, A5C, PSLV, LPS4C, sax-mid, sax-basal, supAO, M-AO, M-LV, M-other, DP-MV, DP-TV, DP-AO, and DP-DAO are all higher or close to 0.90.

TABLE 3
www.frontiersin.org

Table 3. Performance evaluation for different views.

To help clinicians better understand the decision-making mechanism behind the model, the activation map can be visualized. As shown in Figure 3, the activation map reveals the attention distribution of the network, where different colors represent the weights of different pixels in the model's prediction process. The weight of the red part is larger, and the weight of the blue part is smaller. The t-distributed stochastic neighbor embedding (t-SNE) scatter plot in Figure 4 shows how our model clusters different views, where each point represents an image of a view in the test data, and its color represents its true view category. Each point in the scatter plot is a two-dimensional projection of the 512-dimensional feature vector generated by the last hidden layer of the model. We observe that most views can be separated well, and only a few clusters are mixed, which means our model can perform generally well on most views. The confusion matrix shown in Figure 5 also demonstrates that the proposed model performs well on most echocardiographic views. There are some confusion between subRVOT and subSALV, sub4C and subCAS+subSAS, LPS4C and LPS5C, and sax-basal and PSPA.

FIGURE 3
www.frontiersin.org

Figure 3. The activation maps of the apical four-chamber view and the subcostal sagittal view of the atrium septum. Different colors in the activation map represent different weights in model prediction. The red part has a higher weight and the blue part has a lower weight.

FIGURE 4
www.frontiersin.org

Figure 4. t-SNE visualization of CNN feature clusters for 24 echocardiographic views. Different views are represented with colored clusters and labels. The images are sampled from the test set data and 256 samples were randomly sampled for each view. For views whose total number are <256, all samples are applied.

FIGURE 5
www.frontiersin.org

Figure 5. The confusion matrix between different echocardiographic views.

To further evaluate our model, we compared our model with ResNet-34 (the baseline student model), ResNeSt-200 (the teacher model), and two other widely used deep learning models, ResNeXt-50 (24) and DenseNet-161 (25). As shown in Table 4, our model and the baseline ResNet-34 model have the smallest parameters of 21.3 M. The parameters of the ResNeXt-50 model, Densenet-161 model, and ResNeSt-200 model are 23.0 M, 26.5 M, and 68.2 M, respectively. Our model achieved the highest sensitivity of 0.865 and specificity of 0.994, and the second best F1 score of 0.853, slightly lower than the highest F1 score of 0.854 of the teacher model. The precision rates of our model and the teacher model are 0.848 and 0.856, respectively, and the highest precision is 0.865 of DenseNet-161. With a comparable F1 score, the teacher network ResNeSt-200 is almost three times that of our model. Therefore, we can conclude that we achieved the best overall performance through knowledge distillation using the smallest model. In addition, an ablation study on pre-training using image-net was also carried out. Experimental results show that there are significant differences in precision (0.848 vs. 0.833), sensitivity (0.865 vs. 0.856), specificity (0.994 vs. 0.993), and F1 score (0.853 vs. 0.841) with and without pre-training, as shown in Table 4.

TABLE 4
www.frontiersin.org

Table 4. Performance comparison.

Discussion

In our study, we established a 24-category classification model to accurately and efficiently identify 23 commonly used standard echocardiographic views in the diagnosis of CHDs in children by using an effective deep learning-based neural network method through knowledge distillation. The most common CHDs in clinics include ASD, VSD, PS, PDA, TOF, and so on. One of the future work that we will conduct in the near future is to investigate how well the proposed model performs for specific congenital heart diseases (such as ASD).

Most of the ultrasound examinations for CHDs are the process of dynamic scanning where images acquired are mostly non-standard views and the image views are unevenly distributed. The images used in our study are all retrieved from the historical real examination database which is closer to the routine diagnosis of CHDs. This requires a trade-off between the complexity and accuracy of the model. The higher the model complexity, the higher the accuracy, but the operating efficiency will decrease. Therefore, we proposed a knowledge distillation method to maximize the accuracy of the network while keeping the complexity of the network unchanged. After training, the student network can be directly used to recognize the input ultrasonic image and achieve performance comparable to that of the teacher network without adding additional model complexity and time consumption.

Table 3 shows that F1 scores of the subCAS+SAS, A4C, A5C, LPS4C, sax-mid, sax-basal, PSLA, supAO, M-AO, M-LV, DP-MV, DP-TV, DP-AAO, and DP-DAO are all above or close to 0.90. Based on these views, common CHDs can be diagnosed, including ASD, VSD, PS, and TOF. Initially, subCAS and subSAS were treated as two separate views. However, due to the similarity of their image appearance, it is difficult to distinguish them. In addition, these two views are mainly used for the diagnosis of ASD and the scanning process is often between the two views, which makes the distinction more difficult. On the other hand, the confusion between the two views has no significant difference in the diagnosis of the disease. Therefore, we merged subCAS and subSAS into a new subCAS+SAS view.

As shown in Figure 5, the sub4C view was easily classified as subCAS, the subSALV view was easy to get misclassified as subRVOT, LPS5C view was easy to be misclassified as LPS4C and sax-basal. The main reason for these misclassifications was that these views were relatively close to each other. Moreover, in the scanning process, image slices were collected between these views, therefore it was more likely to cause misclassification. In addition, the training samples of the above views were also relatively small. Subcostal short-axis view of the left ventricle is primarily used to diagnose muscular VSDs, the information provided in this view can be supplemented by sax-mid. Parasternal view of the pulmonary artery was easily misclassified as sax-basal and “other” views. Through further analysis of misclassified images, we found that images that clearly showed the aortic root and pulmonary artery branches were easily misclassified. The low F1 score of the PV flow spectrum was caused by confusion with the blood flow spectrum of the pulmonary artery branches. In addition, the training data for the blood flow spectrum and the M-mode ultrasound was also small. In the future, we will add more training data to improve the performance of the proposed model.

Conclusions

This study proposes an effective deep learning based method to identify echocardiographic views commonly used in the diagnosis of CHDs in children. Experimental results show that the proposed method can maximize the accuracy of the network while maintaining the complexity of the network. Compared with existing standard echocardiographic view recognition research, the images used in this study are all from the historical real examination database which are more realistic and authentic, providing foundation for subsequent AI identification of various types of CHDs with greater clinical significance.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Ethics Statement

The studies involving human participants were reviewed and approved by the Medical Ethics Committee of Shanghai Children Medical Center. Written informed consent from the participants' legal guardian/next of kin was not required to participate in this study in accordance with the national legislation and the institutional requirements.

Author Contributions

LW and XL analyzed the data and draft the manuscript draft. WH and LC reanalyzed the data. KG, QS, and YY did the statistical analysis. YZ, BD, and LZ conceived and designed the study, edited and reviewed the manuscript draft. All authors revised, read, and approval the final version of the manuscript.

Funding

This work was supported by the ZheJiang Provincial Key Research Development Program (No. 2020C03073), the National Natural Science Foundation of China (No. 82001825), the Health and Family Planning Commission of China (No. 1132 GDEK201706), the Science and Technology Innovation-Biomedical Supporting Program of Shanghai Science and Technology Committee (No. 19441904400), and Program for artificial intelligence innovation and development of Shanghai Municipal Commission of Economy and Information, and the Biomedical and Engineering (Science) Interdisciplinary Study Fund of Shanghai Jiaotong University (No. YG2019QNB03).

Conflict of Interest

XL, KG, QS, and YY were employed by Deepwise Healthcare.

The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

1. Zhao QM, Liu F, Wu L, Ma XJ, Niu C, Huang GY. Prevalence of congenital heart disease at live birth in China. J Pediatr. (2019) 204:53–8. doi: 10.1016/j.jpeds.2018.08.040

PubMed Abstract | CrossRef Full Text | Google Scholar

2. Braunwald E. The rise of cardiovascular medicine. Eur Heart J. (2012) 33:838–45, 845a. doi: 10.1093/eurheartj/ehr452

PubMed Abstract | CrossRef Full Text | Google Scholar

3. Zhang X, Lin X, Zhang Z, Dong L, Sun X, Sun D, et al. Artificial intelligence medical ultrasound equipment: application of breast lesions detection. Ultrason Imaging. (2020) 42:191–202. doi: 10.1177/0161734620928453

PubMed Abstract | CrossRef Full Text | Google Scholar

4. Lv J, Dong B, Lei H, Shi G, Chen H. Artificial intelligence-assisted auscultation in detecting congenital heart disease. Eur Heart J Digit Health. (2021) 2:119–24. doi: 10.1093/ehjdh/ztaa017

CrossRef Full Text | Google Scholar

5. Cao W, An X, Cong L, Lyu C, Zhou Q, Guo R. Application of deep learning in quantitative analysis of 2-dimensional ultrasound imaging of nonalcoholic fatty liver disease. J Ultrasound Med. (2020) 39:51–9. doi: 10.1002/jum.15070

PubMed Abstract | CrossRef Full Text | Google Scholar

6. Hirasawa T, Aoyama K, Tanimoto T, Ishihara S, Shichijo S, Ozawa T, et al. Application of artificial intelligence using a convolutional neural network for detecting gastric cancer in endoscopic images. Gastric Cancer. (2018) 21:653–60. doi: 10.1007/s10120-018-0793-2

PubMed Abstract | CrossRef Full Text | Google Scholar

7. Zhou LQ, Wang JY, Yu SY, Wu GG, Wei Q, Deng YB, et al. Artificial intelligence in medical imaging of the liver. World J Gastroenterol. (2019) 25:672–82. doi: 10.3748/wjg.v25.i6.672

PubMed Abstract | CrossRef Full Text | Google Scholar

8. Hosny A, Parmar C, Quackenbush J, Schwartz LH, Aerts HJWL. Artificial intelligence in radiology. Nat Rev Cancer. (2018) 18:500–10. doi: 10.1038/s41568-018-0016-5

PubMed Abstract | CrossRef Full Text | Google Scholar

9. Wang L, Yang S, Yang S, Zhao C, Tian G, Gao Y, et al. Automatic thyroid nodule recognition and diagnosis in ultrasound imaging with the YOLOv2 neural network. World J Surg Oncol. (2019) 17:12. doi: 10.1186/s12957-019-1558-z

PubMed Abstract | CrossRef Full Text | Google Scholar

10. Han S, Kang HK, Jeong JY, Park MH, Kim W, Bang WC, et al. A deep learning framework for supporting the classification of breast lesions in ultrasound images. Phys Med Biol. (2017) 62:7714–28. doi: 10.1088/1361-6560/aa82ec

PubMed Abstract | CrossRef Full Text | Google Scholar

11. Song J, Chai YJ, Masuoka H, Park SW, Kim SJ, Choi JY, et al. Ultrasound image analysis using deep learning algorithm for the diagnosis of thyroid nodules. Medicine (Baltimore). (2019) 98:e15133. doi: 10.1097/MD.0000000000015133

PubMed Abstract | CrossRef Full Text | Google Scholar

12. Jin CN, Salgo IS, Schneider RJ, Kam KK, Chi WK, So CY, et al. Using anatomic intelligence to localize mitral valve prolapse on three-dimensional echocardiography. J Am Soc Echocardiogr. (2016) 29:938–45. doi: 10.1016/j.echo.2016.07.002

PubMed Abstract | CrossRef Full Text | Google Scholar

13. Choi J, Hong GR, Kim M, Cho IJ, Shim CY, Chang HJ, et al. Automatic quantification of aortic regurgitation using 3D full volume color doppler echocardiography: a validation study with cardiac magnetic resonance imaging. Int J Cardiovasc Imaging. (2015) 31:1379–89. doi: 10.1007/s10554-015-0707-x

PubMed Abstract | CrossRef Full Text | Google Scholar

14. Baumgartner CF, Kamnitsas K, Matthew J, Fletcher TP, Smith S, Koch LM, et al. SonoNet: real-time detection and localisation of fetal standard scan planes in freehand ultrasound. IEEE Trans Med Imaging. (2017) 36:2204–15. doi: 10.1109/TMI.2017.2712367

PubMed Abstract | CrossRef Full Text | Google Scholar

15. Sridar P, Kumar A, Quinton A, Nanan R, Kim J, Krishnakumar R. Decision fusion-based fetal ultrasound image plane classification using convolutional neural networks. Ultrasound Med Biol. (2019) 45:1259–73. doi: 10.1016/j.ultrasmedbio.2018.11.016

PubMed Abstract | CrossRef Full Text | Google Scholar

16. Zhang J, Gajjala S, Agrawal P, Tison GH, Hallock LA, Beussink-Nelson L, et al. Fully automated echocardiogram interpretation in clinical practice. Circulation. (2018) 138:1623–35. doi: 10.1161/CIRCULATIONAHA.118.034338

PubMed Abstract | CrossRef Full Text | Google Scholar

17. Passalis N, Tzelepi M, Tefas A. Probabilistic knowledge transfer for lightweight deep representation learning. IEEE Trans Neural Netw Learn Syst. (2021) 32:2030–9. doi: 10.1109/TNNLS.2020.2995884

PubMed Abstract | CrossRef Full Text | Google Scholar

18. Cho J, Lee M. Building a compact convolutional neural network for embedded intelligent sensor systems using group sparsity and knowledge distillation. Sensors (Basel). (2019) 19:4307. doi: 10.3390/s19194307

PubMed Abstract | CrossRef Full Text | Google Scholar

19. Chen H, Wang Y, Xu C, Xu C, Tao D. Learning student networks via feature embedding. IEEE Trans Neural Netw Learn Syst. (2021) 32:25–35. doi: 10.1109/TNNLS.2020.2970494

PubMed Abstract | CrossRef Full Text | Google Scholar

20. Zhao H, Sun X, Dong J, Chen C, Dong Z. Highlight every step: knowledge distillation via collaborative teaching. IEEE Trans Cybern. (2020). doi: 10.1109/TCYB.2020.3007506. [Epub ahead of print].

PubMed Abstract | CrossRef Full Text | Google Scholar

21. He K, Zhang X, Ren S, Sun J. Deep residual learning for image recognition. In: Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). Las Vegas, NV (2016). p. 770–8. doi: 10.1109/CVPR.2016.90

PubMed Abstract | CrossRef Full Text | Google Scholar

22. Zhang H, Wu C, Zhang Z, Zhu Y, Zhang Z, Lin H, et al. Resnest: split-attention networks. arXiv preprint. (2020) arXiv:2004.08955.

Google Scholar

23. Deng J, Dong W, Socher R, Li LJ, Li K, Li FF. Imagenet: a large-scale hierarchical image database. In: Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). Miami, FL (2009). p. 248–55. doi: 10.1109/CVPR.2009.5206848

PubMed Abstract | CrossRef Full Text | Google Scholar

24. Xie S, Girshick R, Dollár P, Tu Z, He K. Aggregated residual transformations for deep neural networks. In: Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). Honolulu, HI (2017). p. 5987–95. doi: 10.1109/CVPR.2017.634

PubMed Abstract | CrossRef Full Text | Google Scholar

25. Huang G, Liu Z, Van Der Maaten L, Weinberger K. Densely connected convolutional networks. In: Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). Honolulu, HI (2017) p. 4700–8. doi: 10.1109/CVPR.2017.243

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: standard echocardiographic view, congenital heart defect, deep learning, convolutional neural network, knowledge distillation

Citation: Wu L, Dong B, Liu X, Hong W, Chen L, Gao K, Sheng Q, Yu Y, Zhao L and Zhang Y (2022) Standard Echocardiographic View Recognition in Diagnosis of Congenital Heart Defects in Children Using Deep Learning Based on Knowledge Distillation. Front. Pediatr. 9:770182. doi: 10.3389/fped.2021.770182

Received: 03 September 2021; Accepted: 20 December 2021;
Published: 18 January 2022.

Edited by:

Diego Gallo, Politecnico di Torino, Italy

Reviewed by:

Xin Zhang, Capital Medical University, China
Michele Lioncino, University of Campania Luigi Vanvitelli, Italy

Copyright © 2022 Wu, Dong, Liu, Hong, Chen, Gao, Sheng, Yu, Zhao and Zhang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Liebin Zhao, zhaoleibin@126.com; Yuqi Zhang, zhangyuqi@hotmail.com

These authors have contributed equally to this work and share first authorship

Download